With regard to 'Direct Path read' expected

As we know, the direct path read waits is a new feature of Oracle 11 g. Oracle Documents or other objects, when it's the full table scan, direct path reading took place. But why it happened?  I don't know clearly.

Here is the description of the Oracle Document:

http://docs.Oracle.com/CD/E18283_01/server.112/e17110/waitevents003.htm#sthref3849

During operations of Direct Reading of data asynchronously access path in the database files. At some point, the session must ensure that all the asynchronous i/o in circulation have been realized on the drive. This can also happen if during a direct reading without more slots is available for storing exceptional load requests (a charge application might consist of several e/s).


Question:

1 during operations of Direct data asynchronously reading path in database files.  > > this statement means what? What of "asynchronous reading '?

2 describe above description for me in details.

3. who can clearly explain "why read direct path wait to happen ' for me?


Thanks in advance.

Lonion


Lonion wrote:

Question:

1. to path operations Direct playback of data asynchronously in the database files.  > This statement means what? What of "asynchronous reading '?

2 describe above description for me in details.

3. who can clearly explain "why read direct path wait to happen ' for me?

If you want to get very technical, Frits Hoogland wrote a lot about the implementation:

http://fritshoogland.WordPress.com/2013/05/09/direct-path-read-and-fast-full-index-scans/

http://fritshoogland.WordPress.com/2013/01/04/Oracle-11-2-and-the-direct-path-read-event/

http://fritshoogland.WordPress.com/2012/12/27/Oracle-11-2-0-1-and-the-KFK-async-disk-IO-wait-event/

http://www.UKOUG.org/what-we-offer/library/about-Multiblock-reads/about-Multiblock-reads.PDF

Concerning

Jonathan Lewis

Tags: Database

Similar Questions

  • Insert adds parallel sessions expect "direct path read temp".

    Hi all.

    The database is 11.2.0.3 on a linux machine.

    I published the following query, and I found that parallel query sessions were waiting for "direct path read temp.

    And SDPCSM. Table SDP_CHILD_SVC_PROFATTR_HIS does NOT index and is in nologging mode.

    Why the pq sessions expect "direct path read/write temp?
    insert /*+ append parallel (8) */ into SDPCSM.SDP_CHILD_SVC_PROFATTR_HIS
    select /*+ parallel (SDP_CHILD_SVC_PROFATTR_HIS_E, 8) */ * from SDPCSM.SDP_CHILD_SVC_PROFATTR_HIS_E@tb_link;
    Thanks in advance.
    Best regards.

    Please check this blog:
    http://www.Confio.com/blog/Oracle-wait-event-explained-direct-path-read-temp

  • Is it reasonable to require a Full table Scan (Direct path read) on a large Table

    Hello

    I have an Oracle 11.2.0.3 on a Sun Solaris 10 sparc with SGA 25 GB machine

    One of my made SQL an analysis of indexes on the table to 45 GB where the size of the index is 14FR and the thourghput for the sequential read is 2MBPS.

    so now (Index) 14 000 Mo, 14 GB / 2 MB = dry 7000 = 2 hours to scan the index.

    Flow of the direct path read is 500 Mbps for an another SQL and reads assimilate them all live.

    and because of this flow, a read (FTS) of direct path on the table of 7 GB out in 12-13 seconds. So in my case, it probably takes 100 seconds to scan the entire table. It is a monthly report and does not often.

    Should I try to use a trick to force sql to do a full scan and exploit the possibilities of direct reading and is done in a few minutes? Assuming it to do a FTS (direct play)

    Of course, it will be tested and verified, but the question is, is this a sensible apprioach? is - anyone has experience by forcing it?

    Any suggestions would be helpful... I'll test this on a different machine tomorrow.

    Best regards

    Maran

    82million lines and a grouping of 18million factor?  And really 17million from lines to recover?  However, your result set after the hash join is 3500 lines, although the optimizer expects 16 lines.

    I would say that the statistics are not representative or not in use for this SQL.  Certainly, the index does not match query predicates.

    The fact that the index is also using the virtual columns only adds to confusion.

    Hemant K Collette

  • Direct path read event

    Hi all

    We use 10.2.0.3 on sparc solaris 10 64 bit OS.

    I wondered about two events in my environment that will take a long time. Yes, we do a lot of grouping, sorting and hash joins that temp on spill data.

    For the most part, all time will "lead the read temp path. I took route and seen it's just 1 block reading all time (p3) vs other reads like scattered or direct path write 32 blocks read/write.
      direct path read temp                      241389        4.19       2511.08
      direct path write                               2        0.00          0.00
    So, is there a reason why the direct path read temp to read than a single block at the time? Can we define something that can read multiple blocks in an extraction

    Nico wrote:
    Thank you Greg for more information.

    Have you found any workaround to 10.2.0.3

    It is not a work around. The changes in the fixed a bug how temp blocks are read for positional kinds. It should all but remove the one-piece temp readings. Much more effective with this fix. I saw about 2 x increase in performance in many cases actual test using this hotfix.

    You can try and see if Oracle Support will backport of 10.2.0.3, but I recommend you do 10.2.0.4 anyway due to the considerable number of bug fixes. It won't take long and 10.2.0.5 comes out as well.

    --
    Kind regards
    Greg Rahn
    http://structureddata.org

  • Wait events "direct path write" and "direct path read".

    Hello

    We have a query that takes more than 2 minutes. It's a 9.2.0.7 database. We took the request trace/tkprof and identified there so manay 'direct entry path' and 'direct path read' wait for events in the trace file.

    WAITING #3: nam = "Write" direct path ela = 5 201 p1 = p2 = p3 70710 = 15
    WAITING #3: nam = "direct path read" ela = 170 201 p1 = p2 = 71719 p3 = 15

    In the light of the foregoing, "p1 = 201" is a the file_id, but we could not find any data file, the temporary file, the control file with this id # 201.
    Can you please let us know what "p1 = 201" here, how to identify the file that is causing the problem.

    Thank you
    Sravan

    Whatever it is:

    show parameter db_files
    

    back? I think, is that it returns 200.

    Read the file live and direct file writing events are reads and writes of tablespace TEMP. Wait events, folder # is reported as db_files + id of a temporary file. So, 201 means temp #1 file.

    Now, as to your real performance issue.

    Without seeing the SQL and the corresponding implementation plan, it is impossible to be sure. However, the most frequent causes of temporary entries are the operations of sorting and group by operations.

    If you decide to display your plan and SQL execution, please be sure to make it readable by formatting it. Information on how to do this can be found here.

    Hope that helps,

    -Mark

    Published by: mbobak on May 1st, 2011 01:50

  • Why a parallel query use direct path read?

    I think because the cache must Access pads, lock and lock on buffer block if parallel query do not reading of the direct path, parallel queries will be affected by the serialization as the latch and oracle .so lock mechanism choose direct path read to avoid what he.

    that someone has a good idea?

    Published by: Jeremiah on December 8, 2008 07:52

    Jinyu wrote:
    I think because the cache must Access pads, lock and lock on buffer block if parallel query do not reading of the direct path, parallel queries will be affected by the serialization as the latch and oracle .so lock mechanism choose direct path read to avoid what he.

    Jinyu,

    actually, Yes, I think that's it. The parallel query is designed to scan very much, because the load of communication between processes and maintenance/commissioning the parallel slave makes inefficient operation for small segments.

    So I guess that the assumption is that the segment to analyze is probably very large, the fraction of the blocks in the buffer cache is very low compared to the blocks to scan and so the fresh reduction General read directly blocks without moving through all the questions of serialization of the buffer cache should prevail on the issue of blocks "unbuffered" and save the buffer for objects cache more benefit from development caching.

    Kind regards
    Randolf

    Oracle related blog stuff:
    http://Oracle-Randolf.blogspot.com/

    SQLTools ++ for Oracle (Open source Oracle GUI for Windows):
    http://www.sqltools-plusplus.org:7676 /.
    http://sourceforge.NET/projects/SQLT-pp/

  • message to the subject of the defective hosts SMTP and HTTP, with regard to the two reading: 'No. Socket Error 11001 error number Ox800CCCOD. »

    To set up my Outlook Express, questioned, by sending an e-mail with the download, to provide the above hosts "addresses", I have inserted the wrong information ignorance.  Now cannot send and open downloads, I try to access the page in question again and RE-ENTER the RIGHT "addresses", but all in vain.  Can someone help me to access this critical page, I could revisit the functional plan online?  PLEASE?

    Open Outlook Express, click the Tools menu, click accounts, click Mail, click Properties, verify each tab and enter the correct information. If you don't know, your ISP will give you the correct information.

  • DB file sequential read and read of the direct path

    Hello

    Could someone please clear my doubts about 'db file sequential read' and ' path direct reading. And help me understand the tkprof report correctly.
    Please suggest if my understanding for scenario below is correct.

    We have a 11.2.0.1 ' cluster 2 node rac version + asm' production and environment of his test environment that is a stand-alone database.
    The query performs well in production compared to the test database.
    The table is to have 254 + columns (264) with many lobs coulumns however LOB is not currently selected in the query.
    I read in metalink this 254 table + column a intra-line-chaining, causing "db file sequential read" full table Scan.

    Here are some details on the table which is similar to the prod and test, block size is 8 k:
    TABLE                             UNUSED BLOCKS     TOTAL BLOCKS  HIGH WATER MARK
    ------------------------------  ---------------  ---------------  ---------------
    PROBSUMMARYM1                               0          17408          17407
    What I understand less tkprof in production environment for a given session is:
    1 - the request resulted in disk 19378 readings and 145164 consistent readings.
    2 19378 disc bed, 2425 reads disc has given rise to the wait event 'db file sequential read'.
    This statement is correct this disc remaining readings were "db file sequential reads" but real quick so didn't wait event tied to it?
    3 - 183 'direct path read' there also. Is it because of the order by clause of the query?

    SQL ID: 72tvt5h4402c9
    Plan Hash: 1127048874
    select "NUMBER" num 
    from
     smprd.probsummarym1 where flag ='f' and affected_item = 'PAUSRWVP39486' 
      order by num asc
    
    
    call     count       cpu    elapsed       disk      query    current        rows
    ------- ------  -------- ---------- ---------- ---------- ----------  ----------
    Parse        1      0.00       0.00          0          0          0           0
    Execute      1      0.00       0.00          0          0          0           0
    Fetch        1      0.53       4.88      19378     145164          0           0
    ------- ------  -------- ---------- ---------- ---------- ----------  ----------
    total        3      0.53       4.88      19378     145164          0           0
    
    Misses in library cache during parse: 0
    Optimizer mode: ALL_ROWS
    Parsing user id: SYS
    
    Rows     Row Source Operation
    -------  ---------------------------------------------------
          0  SORT ORDER BY (cr=145164 pr=19378 pw=0 time=0 us cost=4411 size=24 card=2)
          0   TABLE ACCESS FULL PROBSUMMARYM1 (cr=145164 pr=19378 pw=0 time=0 us cost=4410 size=24 card=2)
    
    
    Elapsed times include waiting on following events:
      Event waited on                             Times   Max. Wait  Total Waited
      ----------------------------------------   Waited  ----------  ------------
      SQL*Net message to client                       1        0.00          0.00
      ges message buffer allocation                   3        0.00          0.00
      enq: KO - fast object checkpoint                2        0.00          0.00
      reliable message                                1        0.00          0.00
      KJC: Wait for msg sends to complete             1        0.00          0.00
      Disk file operations I/O                        1        0.00          0.00
      kfk: async disk IO                            274        0.00          0.00
      direct path read                              183        0.01          0.72
      db file sequential read                      2425        0.05          3.71
      SQL*Net message from client                     1        2.45          2.45
    The same query when ran in no no rac - asm test stand alone database has given below tkprof.
    Does this mean that:
    1-here too, reads happen through "db file sequential read", but they were so fast that has failed to the wait event?
    2. "direct path read," it's because of the order clause in the query. "
    SQL ID: 72tvt5h4402c9
    Plan Hash: 1127048874
    select "NUMBER" num 
    from
     smprd.probsummarym1 where flag ='f' and affected_item = 'PAUSRWVP39486' 
      order by num asc
    
    
    call     count       cpu    elapsed       disk      query    current        rows
    ------- ------  -------- ---------- ---------- ---------- ----------  ----------
    Parse        1      0.00       0.00          0          0          0           0
    Execute      1      0.00       0.06          0          0          0           0
    Fetch        1      0.10       0.11      17154      17298          0           0
    ------- ------  -------- ---------- ---------- ---------- ----------  ----------
    total        3      0.10       0.18      17154      17298          0           0
    
    Misses in library cache during parse: 0
    Optimizer mode: ALL_ROWS
    Parsing user id: SYS
    
    Rows     Row Source Operation
    -------  ---------------------------------------------------
          0  SORT ORDER BY (cr=17298 pr=17154 pw=0 time=0 us cost=4694 size=12 card=1)
          0   TABLE ACCESS FULL PROBSUMMARYM1 (cr=17298 pr=17154 pw=0 time=0 us cost=4693 size=12 card=1)
    
    
    Elapsed times include waiting on following events:
      Event waited on                             Times   Max. Wait  Total Waited
      ----------------------------------------   Waited  ----------  ------------
      SQL*Net message to client                       1        0.00          0.00
      Disk file operations I/O                        1        0.00          0.00
      db file sequential read                         3        0.00          0.00
      direct path read                              149        0.00          0.03
      SQL*Net message from client                     1        2.29          2.29
    For trace files in the database Production and Test, I see that 'direct path read' is against the same data file that's stored table.
    Then how come 'this direct path read' because of the order by clause of the query and would have been in sort field size or pga?
    Or this direct path read extracts real PGA disk data, and "db file sequential read" do not extract data?
    I know, it's 'direct path read' is wait event when data are asked in PGA drive or what kind segment or temp tablespace is used.

    Here is the example of trace file in the Production database:
    *** 2013-01-04 13:49:15.109
    WAIT #1: nam='SQL*Net message from client' ela= 11258483 driver id=1650815232 #bytes=1 p3=0 obj#=-1 tim=1357278555109496
    CLOSE #1:c=0,e=9,dep=0,type=1,tim=1357278555109622
    =====================
    PARSING IN CURSOR #1 len=113 dep=0 uid=0 oct=3 lid=0 tim=1357278555109766 hv=138414473 ad='cfc02ab8' sqlid='72tvt5h4402c9'
    select "NUMBER" num from smprd.probsummarym1 where flag ='f' and affected_item = 'PAUSRWVP39486' order by num asc
    END OF STMT
    PARSE #1:c=0,e=98,p=0,cr=0,cu=0,mis=0,r=0,dep=0,og=1,plh=1127048874,tim=1357278555109765
    EXEC #1:c=0,e=135,p=0,cr=0,cu=0,mis=0,r=0,dep=0,og=1,plh=1127048874,tim=1357278555109994
    WAIT #1: nam='SQL*Net message to client' ela= 2 driver id=1650815232 #bytes=1 p3=0 obj#=-1 tim=1357278555110053
    WAIT #1: nam='ges message buffer allocation' ela= 3 pool=0 request=1 allocated=0 obj#=-1 tim=1357278555111630
    WAIT #1: nam='enq: KO - fast object checkpoint' ela= 370 name|mode=1263468550 2=65610 0=1 obj#=-1 tim=1357278555112098
    WAIT #1: nam='reliable message' ela= 1509 channel context=3691837552 channel handle=3724365720 broadcast message=3692890960 obj#=-1 tim=1357278555113975
    WAIT #1: nam='ges message buffer allocation' ela= 2 pool=0 request=1 allocated=0 obj#=-1 tim=1357278555114051
    WAIT #1: nam='enq: KO - fast object checkpoint' ela= 364 name|mode=1263468550 2=65610 0=1 obj#=-1 tim=1357278555114464
    WAIT #1: nam='KJC: Wait for msg sends to complete' ela= 9 msg=3686348728 dest|rcvr=65536 mtype=8 obj#=-1 tim=1357278555114516
    WAIT #1: nam='ges message buffer allocation' ela= 2 pool=0 request=1 allocated=0 obj#=-1 tim=1357278555114680
    WAIT #1: nam='Disk file operations I/O' ela= 562 FileOperation=2 fileno=6 filetype=2 obj#=85520 tim=1357278555115710
    WAIT #1: nam='kfk: async disk IO' ela= 5 count=1 intr=0 timeout=4294967295 obj#=85520 tim=1357278555117332
    
    *** 2013-01-04 13:49:15.123
    WAIT #1: nam='direct path read' ela= 6243 file number=6 first dba=11051 block cnt=5 obj#=85520 tim=1357278555123628
    WAIT #1: nam='db file sequential read' ela= 195 file#=6 block#=156863 blocks=1 obj#=85520 tim=1357278555123968
    WAIT #1: nam='db file sequential read' ela= 149 file#=6 block#=156804 blocks=1 obj#=85520 tim=1357278555124216
    WAIT #1: nam='db file sequential read' ela= 155 file#=6 block#=156816 blocks=1 obj#=85520 tim=1357278555124430
    WAIT #1: nam='db file sequential read' ela= 4826 file#=6 block#=156816 blocks=1 obj#=85520 tim=1357278555129317
    WAIT #1: nam='db file sequential read' ela= 987 file#=6 block#=156888 blocks=1 obj#=85520 tim=1357278555130427
    WAIT #1: nam='db file sequential read' ela= 3891 file#=6 block#=156888 blocks=1 obj#=85520 tim=1357278555134394
    WAIT #1: nam='db file sequential read' ela= 155 file#=6 block#=156912 blocks=1 obj#=85520 tim=1357278555134645
    WAIT #1: nam='db file sequential read' ela= 145 file#=6 block#=156920 blocks=1 obj#=85520 tim=1357278555134866
    WAIT #1: nam='db file sequential read' ela= 234 file#=6 block#=156898 blocks=1 obj#=85520 tim=1357278555135332
    WAIT #1: nam='db file sequential read' ela= 204 file#=6 block#=156907 blocks=1 obj#=85520 tim=1357278555135666
    WAIT #1: nam='kfk: async disk IO' ela= 4 count=1 intr=0 timeout=4294967295 obj#=85520 tim=1357278555135850
    WAIT #1: nam='direct path read' ela= 6894 file number=6 first dba=72073 block cnt=15 obj#=85520 tim=1357278555142774
    WAIT #1: nam='db file sequential read' ela= 4642 file#=6 block#=156840 blocks=1 obj#=85520 tim=1357278555147574
    WAIT #1: nam='db file sequential read' ela= 162 file#=6 block#=156853 blocks=1 obj#=85520 tim=1357278555147859
    WAIT #1: nam='db file sequential read' ela= 6469 file#=6 block#=156806 blocks=1 obj#=85520 tim=1357278555154407
    WAIT #1: nam='db file sequential read' ela= 182 file#=6 block#=156826 blocks=1 obj#=85520 tim=1357278555154660
    WAIT #1: nam='db file sequential read' ela= 147 file#=6 block#=156830 blocks=1 obj#=85520 tim=1357278555154873
    WAIT #1: nam='db file sequential read' ela= 145 file#=6 block#=156878 blocks=1 obj#=85520 tim=135727855515
    Here are the trace file for the test database:
    *** 2013-01-04 13:46:11.354
    WAIT #1: nam='SQL*Net message from client' ela= 10384792 driver id=1650815232 #bytes=1 p3=0 obj#=-1 tim=1357278371354075
    CLOSE #1:c=0,e=3,dep=0,type=3,tim=1357278371354152
    =====================
    PARSING IN CURSOR #1 len=113 dep=0 uid=0 oct=3 lid=0 tim=1357278371363427 hv=138414473 ad='c7bd8d00' sqlid='72tvt5h4402c9'
    select "NUMBER" num from smprd.probsummarym1 where flag ='f' and affected_item = 'PAUSRWVP39486' order by num asc
    END OF STMT
    PARSE #1:c=0,e=9251,p=0,cr=0,cu=0,mis=0,r=0,dep=0,og=1,plh=1127048874,tim=1357278371363426
    EXEC #1:c=0,e=63178,p=0,cr=0,cu=0,mis=0,r=0,dep=0,og=1,plh=1127048874,tim=1357278371426691
    WAIT #1: nam='SQL*Net message to client' ela= 2 driver id=1650815232 #bytes=1 p3=0 obj#=-1 tim=1357278371426766
    WAIT #1: nam='Disk file operations I/O' ela= 1133 FileOperation=2 fileno=55 filetype=2 obj#=93574 tim=1357278371428069
    WAIT #1: nam='db file sequential read' ela= 51 file#=55 block#=460234 blocks=1 obj#=93574 tim=1357278371428158
    WAIT #1: nam='direct path read' ela= 31 file number=55 first dba=460235 block cnt=5 obj#=93574 tim=1357278371428956
    WAIT #1: nam='direct path read' ela= 47 file number=55 first dba=136288 block cnt=8 obj#=93574 tim=1357278371429099
    WAIT #1: nam='direct path read' ela= 80 file number=55 first dba=136297 block cnt=15 obj#=93574 tim=1357278371438529
    WAIT #1: nam='direct path read' ela= 62 file number=55 first dba=136849 block cnt=15 obj#=93574 tim=1357278371438653
    WAIT #1: nam='direct path read' ela= 17 file number=55 first dba=136881 block cnt=7 obj#=93574 tim=1357278371438750
    WAIT #1: nam='direct path read' ela= 35 file number=55 first dba=136896 block cnt=8 obj#=93574 tim=1357278371438855
    WAIT #1: nam='direct path read' ela= 22 file number=55 first dba=136913 block cnt=7 obj#=93574 tim=1357278371438936
    WAIT #1: nam='direct path read' ela= 19 file number=55 first dba=137120 block cnt=8 obj#=93574 tim=1357278371439029
    WAIT #1: nam='direct path read' ela= 36 file number=55 first dba=137145 block cnt=7 obj#=93574 tim=1357278371439114
    WAIT #1: nam='direct path read' ela= 18 file number=55 first dba=137192 block cnt=8 obj#=93574 tim=1357278371439193
    WAIT #1: nam='direct path read' ela= 16 file number=55 first dba=137201 block cnt=7 obj#=93574 tim=1357278371439252
    WAIT #1: nam='direct path read' ela= 17 file number=55 first dba=137600 block cnt=8 obj#=93574 tim=1357278371439313
    WAIT #1: nam='direct path read' ela= 15 file number=55 first dba=137625 block cnt=7 obj#=93574 tim=1357278371439369
    WAIT #1: nam='direct path read' ela= 22 file number=55 first dba=137640 block cnt=8 obj#=93574 tim=1357278371439435
    WAIT #1: nam='direct path read' ela= 702 file number=55 first dba=801026 block cnt=126 obj#=93574 tim=1357278371440188
    WAIT #1: nam='direct path read' ela= 1511 file number=55 first dba=801154 block cnt=126 obj#=93574 tim=1357278371441763
    WAIT #1: nam='direct path read' ela= 263 file number=55 first dba=801282 block cnt=126 obj#=93574 tim=1357278371442547
    WAIT #1: nam='direct path read' ela= 259 file number=55 first dba=801410 block cnt=126 obj#=93574 tim=1357278371443315
    WAIT #1: nam='direct path read' ela= 294 file number=55 first dba=801538 block cnt=126 obj#=93574 tim=1357278371444099
    WAIT #1: nam='direct path read' ela= 247 file number=55 first dba=801666 block cnt=126 obj#=93574 tim=1357278371444843
    WAIT #1: nam='direct path read' ela= 266 file number=55 first dba=801794 block cnt=126 obj#=93574 tim=1357278371445619
    Thanks & Rgds,
    Vijay

    911786 wrote:

    Direct path readings can be done on the series tablescans in your version of Oracle, but if you have chained rows in the table and then Oracle can read read read the beginning of the line in the path directly, but must make a single block in the cache (the db file sequential read) to get the next part of the line.

    It is possible that your production system has a lot of chained rows, while your test system is not. As corroboration (though not convincing) indicator of what you might notice that if you take (reads disc - db file sequential reads) - who might get you close to all the blocks direct read - the numbers are very similar.

    I'm not 100% convinced that it's the answer for the difference in behavior, but worth a visit. If you can force and indexed access path in the table, do something like "select / * + index (use {pk}) * / max table (last_column_in_table)" and check the number of "table fetch continued lines" could be close to the number of db file sequential reads you. (There are other options for the counting of the chained rows that could be faster).

    Concerning
    Jonathan Lewis

  • Multithreading works only with the direct path load

    Oracle DB version: Oracle Database 11 g Enterprise Edition Release 11.2.0.4.0 - 64 bit Production

    I'm on my way live load to load data from a flat file into a table using SQL * Loader. I also kept as parallel. However, I can't in multithreading is used at all, based on the report of the log file.

    I use the settings according to the true value in the sqlldr: -.

    parallel=true , multithreading=true , skip_index_maintenance=true

    Output in the journal sqlldr:-

    Path used: Direct Insert option in effect for this table: APPEND

    Trigger DEV."R_TM_BK_BORROWER" was disabled before the load.

    DEV."R_TM_BK_BORROWER" was re-enabled.

    The following index(es) on table "YO"."TM_BK_BORROWER" were processed:

    index DEV.I_NK_TM_BK_BORR_1 loaded successfully with 1554238 keys

    index DEV.I_NK_TM_BK_BORR_2 loaded successfully with 1554238 keys

    index DEV.I_NK_TM_BK_BORR_3 loaded successfully with 1554238 keys

    index DEV.I_NK_TM_BK_BORR_31 loaded successfully with 1554238 keys

     

    Bind array size not used in direct path.

    Column array rows : 5000

    Stream buffer bytes: 256000

    Read buffer bytes: 1048576


    Total logical records skipped: 1

    Total logical records read: 1554241

    Total logical records rejected: 48

    Total logical records discarded: 2

    Total stream buffers loaded by SQL*Loader main thread: 7695

    Total stream buffers loaded by SQL*Loader load thread: 0

    So, I can still see the newspaper sqlldr that all data flow buffers loaded by the main thread and load wire is not always used.

    SQL * Loader load wire do not unload the SQL * Loader main thread. If the load wire supports the current stream buffers, then it allows the primary thread to build the buffer to the next stream while the thread of load load the current stream to the server. We have a server CPU 24.

    I'm not able to find a clue on Google too. Any help is appreciated.

    People, Tom Kyte has finally responded to my message. Here's the thread on asktom-

    http://asktom.Oracle.com/pls/Apex/f?p=100:11:0:P11_QUESTION_ID:1612304461350 #7035459900346550399

  • Redo generation for Direct path Inserts

    Hello, I'm trying to understand some test results confused I see this morning on the generation of redo direct path for the pads. Based on my understanding of Tom to ask several discussions directly inserts path on a set of data to force the record should generate much less remake a traditional insert because the insertion of the direct route does not generate as much cancel which in turn should generate less do on cancellation.

    https://asktom.Oracle.com/pls/asktom/f?p=100:11:0:P11_QUESTION_ID:5280714813869

    Of course, always connect the actual inserted rows but I expected to the remake was less because less cancel has been generated. Instead the roll forward is actually bigger and I don't know why.

    Here's a test case to prove my example.

    set autotrace traceonly;
    create table test_redo as select * from all_tables where 1=0;
    insert into test_redo select * from all_tables;
    rollback;
    insert /*+ append */ into test_redo select * from all_tables;
    rollback;
    

    Stats without Append Hint
    
    Statistics
    ----------------------------------------------------------
            387  recursive calls
           1275  db block gets
          19604  consistent gets
              9  physical reads
        2409204  redo size
            501  bytes sent via SQL*Net to client
            897  bytes received via SQL*Net from client
              4  SQL*Net roundtrips to/from client
              4  sorts (memory)
              0  sorts (disk)
           9031  rows processed
    
    Stats with Append Hint
    
    Statistics
    ----------------------------------------------------------
             59  recursive calls
            162  db block gets
          18675  consistent gets
              0  physical reads
        2596904  redo size
            490  bytes sent via SQL*Net to client
            911  bytes received via SQL*Net from client
              4  SQL*Net roundtrips to/from client
              2  sorts (memory)
              0  sorts (disk)
           9031  rows processed
    

    Any ideas on what I'm missing?

    Thank you

    The / * + append * / copies all blocks Oracle in roll forward, with a little extra for recording etc. headers.

    The standard insert creates change descriptions which save odd odd little bits of space and adds little bits of information links (as well as some redo to describe a very small amount to cancel)...

    The difference between the two has always been pretty low (assuming that you are running in archivelog mode, or force logging) on a right append.  It is possible that curious little details of efficiencies in future versions of the average of the Oracle that the standard Insert wins a place a bit more efficient - it used to be the other way around in earlier versions.

    Concerning

    Jonathan Lewis

  • Archiving log / nologging / direct path insert

    Could you please confirm if the following are true or correct me if my interpretation is wrong:

    Log mode archive 1) and logging is necessary to deal with the media recovery; It was not necessary for the recovery of the instance.

    (2) IF the insert is no. APPEND mode, redo is generated, even if the table is in nologging mode and database is in log noachive mode. This recovery is necessary for recovery of the instance.

    (3) direct path insert ignore cancel the generation and can jump redo generation if the object is in nologging mode.

    Thank you.

    In case if it is relevant, I'm using Oracle 11.2.0.3.

    (1) Yes, archive logs is required for media recovery.
    2 and 3) even if the table is in nologging mode, it generates little recovery data dictionary and index maintenance. After a reboot of a failure - Oracle reads the online redo logs and replay any what transaction it finds in there. It's the little "roll forward." The binary redo information are used to re-read everything that did not get written to the data files. This review included refreshing the information for CANCELLATION (UNDO is protected by roll forward).

    After the restoration has been applied, the database is generally available for use now - and restore begins. For any transaction that was being processed when the instance failed - we have to revert his changes, roll back. We do this by treating the cancellation of all uncommitted transactions.

    The database is now fully recovered.

    Also read he following link
    http://docs.Oracle.com/CD/B19306_01/server.102/b14220/startup.htm
    http://asktom.Oracle.com/pls/asktom/f?p=100:11:0:P11_QUESTION_ID:5280714813869

  • Voodoo Omen with regard, water pump is longer supported by HP - help! : p

    Hi all

    I have an old omen Voodoo (water cooling system) - here is a link to a review of this version: http://reviews.cnet.com/desktops/voodoo-omen/4505-3118_7-31292135.html which was purchased from voodoo before its acquisition by HP.  HP supports more Voodoo telephone products.   My former client ID attached to the omen is 37047 (HP Voodoo team recognized that a few years - if you were on this team, I really appreciate your help!).

    Problem: when I start, there is a horrible noise that I've isolated as coming from the pump to water.  After a bit, he calms down, but I think that it is a 'sign' of bad things to come.    Looking closely, I suspect that the cooling liquid does not flow even in the tubes.  Thus, the pc stopped several times during the boot sequence.

    Question: I think it is a good idea to replace the pump sick before he died.  I'm having a lot of trouble finding any computer repair shop in the san francisco Bay area (I'm in the east Bay) who knows even that there is a such thing as the water cooled pc.

    1. How can I find the mark of the pump that was installed so that I can order a replacement?
    2. I'm looking for a local preference pc technician who is at ease of repair / replacement of water cooling systems.
    3. Is anyone out there?

    It is possible that I just need to replace the pump - hose seems fine.

    Thank you!

    Robert

    The solution involves a complete rebuild of the system with all new parts, except the excellent Voodoo chassis.  I would have added a cooling system, the water but there was no need - and here is a picture, finally.   There is some dust, but you get the idea - it's an approach minimalist and clean wiring.  You can see the fan of collmax takes a lot of space, sitting on top of the cpu - but the entire Setup is silent and ultra reliable.

    It was perfect for my needs, and very quickly - I push it more difficult with several processes going on in lightroom, photoshop and the plugins.  I tried a few games as well and was very impressed - there was no need to overclock.

    I always have this problem with the locations of card reader is not recognized - don't know how to solve this one - and as I said, I can live with that by importing directly from the camera.

    System information:

    Case of Voodoo curse

    ASUS rampage motherboard

    Intel cpu i7 - 4930K

    32 GB of fast ram

    Graphics GeForce GTX 780 Ti

    Power 1200 Watt

    in the solid state boot drive.

    For cooling, I went later with a generic double fan coolmax system instead of water.

    Verdict: it is silent, reliable and fast without overclocking; the system is basically brand new.   The only thing that remained was the case of Voodoo Omen very excellent and very heavy.

  • Common problem with Windows 7, what causes errors with itunes, the software adobe reader, TurboTax 12

    Is there a common problem with Windows 7, what causes errors with itunes, the software adobe reader, TurboTax 12?  I get error messages when I try to use one of these programs

    Hello

    Thanks for the reply with the details.
     
    Problem with opening PDF documents, please see the link and follow the steps in the article:
     
    Error "Windows cannot access the specified device, path or file" when you try to install, update or launch a program or file
    Important: Disable your antivirus or changed settings can make your PC vulnerable to viral, fraudulent, or malicious attacks. Microsoft does not recommend that you disable your antivirus software or change settings. If you need to temporarily disable your anti-virus software, you should reactivate as soon as you are finished.
     
    See the link for the question with idevices:
     
    iTunes for Windows: "Could not load the data class" or "Cannot load provider data" alert synchronization services
     
    Problem with TurboTax 12, check out the links and follow the steps:
     
    5639 error: Cannot configure the product update system
     
     
    If the troubleshooting steps above do not help, check in a new user account:
    Reference:
     
    If it works in the new user account, then the current may be corrupt user account.
    To fix the damaged user account, follow the steps in the following article:
    http://Windows.Microsoft.com/en-us/Windows7/fix-a-corrupted-user-profile
     
    Hope this solves the problem. If the problem persists, you can write to us and we will be happy to help you further.
  • I tried Windows 10 and after 10 days of hell, I asked them to uninstall.  But went with is my old Acrobat Reader.  I tried to use it, but it says that my browser plug is missing and that it cannot start because MSVCP120 is missing on the computer too.

    I tried Windows 10 and after 10 days of hell, I asked them to uninstall.  But went with is my old Acrobat Reader.  I tried to use it, but it says that my browser plug is missing and that it cannot start because MSVCP120 is missing on the computer too.  A friend told me that I have to uninstall and then re-install again.  I am a 71 years who does not know what to do.  Help

    Hi jimm13886845,

    Of course I'll help you get this problem is resolved.

    Run this cleanup tool to uninstall any old version of Adobe Reader Download Adobe Reader and Acrobat tool - Adobe Labs.

    Restart your computer and then installer you old version of Adobe Reader using this link Adobe - Adobe Acrobat Reader DC Distribution.

    Let me know if you encounter a problem, we will be happy to help you.

    Kind regards

    Nicos

  • We can convert a VM with 2 snapshots directly

    Hi team

    I have a vm with 2 snapshot server and there are only 20 left GB in storage. My snapshot file size is 200 GB and we do not have enough space to remove the snapshot. I intend to buy 1 TB HARD drive and attach to my laptopp vm converter and install in my laptop, then convert the vsphere VM VM workstation file on my external HARD drive. My question is can you convert a virtual machine with 2 snapshots directly VM workstation, it will create another cliché of ESX server to do the conversion? If a snapshot is created more conversion will fail because we have only 20 GB free space. Please provide a good solution for this

    Concerning

    Cherif

    I hope that everything will work as expected. Good luck.

    BTW there is a snapshot on the virtual machine. Where did you see the other? If there is a difference between graphical snapshot Manager and user interface that you see in the browser to store data (files such as - 00000x.vmdk are instant), always approve the data browser store!

    André

Maybe you are looking for