Context_file and CIM log file

Hi all

EBS R12.1.3

I was wondering, is it a specific context_file variable for the log ICM file?

Thank you!

I don't - see MOS Doc 458282.1

HTH
Srini

Tags: Oracle Applications

Similar Questions

  • difference between online redolog files and archived log files

    What is the fundamental difference between archived log files and online redo log files? .. .plz keep things simple as m a newbie

    Online redo log:

    Redo online Oracle log files are filled with records of restoration by progression. A record of remake, also called entrance to roll forward, is composed of a group of agents of change, each of which is a description of a change to a single block in the database. Each instance of an Oracle database has an associated online redo log to protect data in case of failure of the instance.

    See here,
    http://download.Oracle.com/docs/CD/B10500_01/server.920/a96521/onlineredo.htm

    Archived redo log:

    Oracle Database allows you to save files to one or more destinations offline newspapers filled with repeat groups, known collectively as the archived redo log. An archived redo log file is a copy of one of the members of a group of redo log filled.

    See here,
    http://download.Oracle.com/docs/CD/B19306_01/server.102/b14231/archredo.htm

    Thank you

  • Recovery of database of old backup and archive log file complete

    Hi expert Oracle!
    I encountered the problem where my oracle database restore.

    In my case:
    -My version of the database: 10.2.0.2
    -J' have a full database backup (01 - Nov)
    -J' have all the archived since (01 - Nov-> 05 - Nov) log file
    -My database down to 05 - Nov with disk error (no data file, not again...).
    -I have no backup RMAN (01 - Nov-> 05 - Nov)

    How to restore my database to 05 - Nov?

    Thank you

    After you RESTORE the DATABASE,

    Use the VALUE UNTIL to set the Point of recovery to 05 - Nov, and then RECOVER the DATABASE
    OR
    Use the SQL command "RECOVER database with the HELP of BACKUP CONTROLFILE until CANCEL" that will keep asking 'future' ArchiveLogs backup.

    Hemant K Collette

  • LOG_ARCHIVE_FORMAT and old log files

    Hello

    in 10g documentation, it is said:
    LOG_ARCHIVE_FORMAT initialization parameter, using %t 
    to include the thread number as part of
     the file name, %s to include the log sequence number,
     and %r to include the resetlogs ID 
    What is the point of having resetlogs ID in the name of archived newspaper? How he would leave to avoid replacing old log files?

    Thank you.

    Very easy. From 10 gR 1, including the resetlogs Id made recoveries by resetlogs much easier than before.

    Werner

  • imageFORMULA CR-180 check the log file to scan

    Hello

    I work with an imageFORMULA CR-180 check Scanner and when executing a good amount (longer than 100) of the controls, I noticed that the scan control stops and a pop-up window appears on the Analysis ToolPak asked if I would like to replace my log file and start over.

    The software that I use is called "Utility scan for CR-180II" and the log file contains the information of MICR line controls scanned, as well as the date of the analysis, and the location of the file to the images captured from the parser.

    Log file options under "Options-> MICR data parameters.

    What I do, it is by clicking on ' no, I want to overwirte the file ", stop the scan, moving the log of the 'old' file, restart the scanning and backup to a new log file.

    Is there anyway to work around this problem and record continuously in the same log file?

    Is there a memory limit on the scanner?

    This is the Web page of the scanner that I use:

    http://www.USA.Canon.com/Cusa/support/Office/imageformula_scanners/imageformula_cr_180_cr_180ii/imag...

    Hi lendjones!

    Thanks for posting in the Forum! Canon does not provide direct support for imageFORMULA product series, but your dealer will be able to help you! If you do not have a reseller, please call us at 1-800-OKCANON (652-2666) and we will be happy to provide you with dealers who are in your area.

  • Try to uninstall Earthlink Accelerator but I get an error message that INSTALL. LOG file cannot be found

    I got earthlink.net as a dial-up internet service and changed to netzero.net, so I decided to uninstall all software from Earthlink.  Everything uninstalled great, except for the Earthlink accelerator.  When I go into control panel > programs > earthlink Accelerator > uninstall I get an error message: cannot find INSTALL. LOG file.  I did a search as an administrator to INSTALL fond. Not and the LOG file to get anything.  How can I find this file so that I can uninstall Earthlink accelerator.  I'm very technically inclined but still can't find the problem.

    We will try to remove the registry key associated with Earthlink accelerator.

    Before that, please make a return register by following the steps in http://windows.microsoft.com/en-in/windows-vista/back-up-the-registry

    Now, perform the following steps:

    Type regedit.exe in the Start search and press on enter.

    Go to the following location and delete the keys. (Also, note the location of the Earthlink accelerator installation file and delete them)

    HKEY_LOCAL_MACHINE\SOFTWARE\Microsoft\Windows\CurrentVersion\Run

    HKEY_LOCAL_MACHINE\Software\Microsoft\Windows\CurrentVersion\RunOnce

    Hkey_current_user\\software\\microsoft\\windows\\currentversion\\run

    HKEY_CURRENT_USER\Software\Microsoft\Windows\CurrentVersion\Runonce

    Hope this information helps. Let us know the results.

  • View the log files in the configuration of the cluster (HTTPS UI)

    Hi all

    In my stand-alone configuration, I can view and download log files from HTTPS-Webgui under "Journal subscriptions"-> "log files".

    But in a cluster configuration, I can't see the column "Log Files" under "journal subscriptions. How can I access the files?

    Thank you!

    Christoph

    Hi Christoph,

    try to change the mode of the machine to access the files.

    Best regards

    Enrico

  • Expectations extended on synchronization of log file while the parallel writing journal is fine

    We have 9.2.0.8 as experiences of long waits on the database log file sync (average waiting time = 46 ms) while waiting for the log file write Parallels is very good (average waiting time is less than 1 millisecond).

    The application is of type middleware, it connects to several other applications. A single user in a single application action train several requests to send back through this middleware, so he needs response time of db in milliseconds.

    The database is quite simple:

    -It has a few config tables that the application reads but rarely updated

    -She has table TRANSACTION_HISTORY: the application inserts records into this table using Insert einreihig (about 100 lines per second); each insert is followed by a validation.

    Records are kept for several months and then purged. The table has only column VARCHAR2/NUMBER/DATE, no LOBS, LONG, etc. The table has 4 non-unique single-column index.

    The average line length is 100 bytes.

    The load profile does not appear something unusual, the main figures: 110 transactions per second average transaction = 1.5 KB size.

    The data below are to 1 hour interval (purge wasn't running during this interval), physical reads or writes physical rate is low:

    Load profile

    ~ ~ ~ Per second per Transaction

    ---------------       ---------------

    Size: 160,164.75 1,448.42

    Logical reads: 521,58 57 675,25

    Block changes: 934,90 8.45

    Physical reads: 76,27 0.69

    Physical writings: 86,10 0.78

    Calls of the user: 491,69 4.45

    Analysis: 321,24 2.91

    Hard analysis: 0.09 0.00

    Kinds: 126.96 1.15

    Logons: 0.06 0.00

    Runs: 17.70 1 956,91

    Operations: 110,58


    Top 5 events are dominated by the synchronization of log file:

    Top 5 timed events

    ~~~~~~~~~~~~~~~~~~                                                     % Total

    Event expects Ela time (s)

    -------------------------------------------- ------------ ----------- --------

    401 608 18 448 59.94 file synchronization log

    db file parallel write 124 044 3 404 11.06

    CPU time                                                        3,097    10.06

    Enqueue 10 476 2 916 9.48

    DB file sequential read 261 947 2 435 7.91


    Section events:

    AVG

    Total wait wait wait

    Hour of wait time wait for the event (s) (ms) /txn

    ---------------------------- ------------ ---------- ---------- ------ --------

    Synchronize 0 401 608 46 18 448 1.0 file log

    db file parallel write 124 044 0 3 404 27 0.3

    Enqueue 10 476 277 2 916 278 0.0

    DB file sequential read 261 947 0 2 435 9 0.7

    buffer busy waits 11 467 67 173 15 0.0

    SQL * Net more data to the client 1 565 619 0 79 0 3.9

    lock row cache 2 800 0 52 18 0.0

    control file parallel write 1 294 0 45 35 0,0

    Log end of switch file 261 0 36 138 0.0

    latch free 2 087 1 446 24 12 0.0

    PL/SQL 1 1 20 19531 0,0 lock timer

    log file parallel write 0 143 739 17 0.4 0

    db file scattered read 1 644 0 17 10 0.0

    sequential log file read 636 0 8 13 0.0


    Log buffer is about 1.3 MB. We could increase the log buffer, but there is no log buffer space waits, so I doubt this will help.


    Newspapers in recovery have their own file systems, not shared with the data files. This explains the difference between waiting avg on parallel writing of log (less than 1 ms) file and db file parallel write (27 ms).

    Restoring logs is 100 MB, there are about 120 journal switches per day.


    What has changed: the pads/validations rate grew. Several months ago there were 25 inserts/validations per second in the TRANSACTION_HISTORY table, now get us 110 inserts/validation per second.


    What problem it causes application: due to slow down the reaction of the basis of the application (Java-based) requires discussions more and more.


    MOS documents on synchronization of log file (for example, 1376916,1 waits for troubleshooting "log file sync") recommend to compare the average waiting time on synchronization of log file and the log file parallel write.

    If the values are close (for example log file sync = 20 ms and log file parallel write = 10 ms) so expectations are caused by nits IO. However, it is not the case here.


    There was a bug (2669566) in 9.2 which resulted in underreporting lgwr parallel time of writing to the log file. I was talking about September 2005, during which the bug was present in 9.2.0.6, reported 10.1 fixed in: file parallel journal written (JL Comp) it is possible that your problem IS written to the log file.

    Concerning

    Jonathan Lewis

  • That redo log files waiting?

    Hello Experts,

    I read articles on the log redo and undo segment files. I was wondering something very simple. That redo log files waiting in there? It stores the sql statements?

    Lets say that my update statement to modify 800 blocks of data. A unique single update statement can modify different data 800 right blocks? Yes, it may be true. I think that these data blocks can not hold buffers to the log to roll forward, right? I mean I know exactly what to do redo log buffer and redo log file. And I know that the task of backgrounding LGWR. But, I wonder if she she holds the data blocks? It is not supposed to hold data like cache buffer blocks, right?

    My second question is, rollback isn't effect to restore the newspaper to the right buffer? Because it does not need log buffer for effect do it again. Conversely, the restoration; statement is included in the restore log buffer by progression when someone isse, am I right?

    As far as I know, rollback interact directly with UNDO TABLESPACE?

    I hope that I have to express myself clearly.

    Thanks in advance.

    Here's my question:

    My second question is, rollback isn't effect to restore the newspaper to the right buffer? Because it does not need log buffer for effect do it again. Conversely, the restoration; statement is included in the restore log buffer by progression when someone isse, am I right?

    As far as I know, rollback interact directly with UNDO TABLESPACE?

    Yes, where else would the undo data come from? Undo tablespace contains the Undo segments that contain the Undo data required for the restoration of your transaction.

    I can say that rollback does not alter the data of the log buffer rede to the past. In other words, change vectors will be remain the same before restoration. Conversely, rollback command is also recorded in the log file of restoration by progression. As the name, all orders are saved in the REDO LOGS.

    I hope that I am wrong so far?

    Not sure why you even the buffer log roll forward for Rollback? This is the reason why I asked you it was for, where occurs the dose the cancellation? And the answer for this is that it happens in the buffer cache. Before you worry about the drivers of change, you must understand that it is not serious what contains where as long as there is no transaction recorded in the operating of the Undo segment table. If the operating table indicates that the transaction is longer there, there must be a cancellation of the transaction. Vectors of change are saved in the file log roll forward, while the restore happens on blocks of data stored in the file "data" undo blocks stored in the undo file "data".

    At the same time I read an article about redo and undo. In this article process transaction is explained. Here is the link http://pavandba.files.wordpress.com/2009/11/undo_redo1.pdf

    I found some interesting information in this article as follows.

    It is worth noting that during the restore process, recovery logs never participate. The only time where redo logs are read is retrieving and archiving. This is the concept of tuning key: redo logs are written on. Oracle does not read during normal processing. As long as you have sufficient devices so that when the ARC is reading a file, LGWR's writing to a different device, then there no contention for redo logs.

    If redo logs are never involved in the restoration process, how is it Oracle will then know the order of the transaction? As far as I know it is only written in redo logs.

    I have thoughts very amazed to Aman.

    Why you ask?

    Now, before giving a response, I say two things. One, I know Pavan and he is a regular contributor to this forum and on several other forums Facebook and two, with all due respect to him, a little advice for you, when you try to understand a concept, to stick to the Oracle documentation and do not read and merge articles/blog-posts from the web. Everone, which publishes on the web, has their own way to express things and many times, the context of the writing makes it more confusing things. Maybe we can erase the doubts that you can get after reading the various search results on the web.

    Redo logs used for the restoration, not to restore. The reason is the redo log files are applied in sequential order, and this is not the case when we look for the restoration. A restore is required to do for a few blocks away. Basically, what happens in a restoration, is that the records of cancellation required for a block of data are sought in the reverse order of their creation. The entry of the transaction is in the slot ITL of the block of data that point to the necessary undo bytes Address (UBA) using which oracle also knows what that undo the blocks would be necessary for the restoration of your transaction. As soon as the blocks of data will be cancelled, the ITL slots would be cleared as well.

    In addition, you must remember, until the transaction is not qualified as finished, using either a commit or a rollback, the cancellation of this data would remain intact. The reason for this is that oracle would ensure that undo data would be available to make the cancellation of the transaction. The reason why Undo data are also recorded in the journals of recovery is to ensure that in the event of the loss of the cancellation of the data file, retrieving them would be possible. Because it would also require changes that's happened on the blocks cancel, restore the vectors change associated with blocks of cancellation are also saved in the buffer log roll forward and, in the redo log files.

    HTH

    Aman...

  • Analysis of the log files from vCenter

    Hello, a virtual machine has been removed in our area, and I need to analyse what/how it happened. Unfortunately, the vSphere and vCenter log files have since turned. However, in regards to the vcenter, I have backups of the log files, and these have been restored. When you look at the log files I don't let you content or find the name of the virtual machine when you perform a search.

    Can anyone tell how to examine the log for a deleted VM files?

    Thank you

    The entry in the journals of vpxd looks like this:

    (Not very helpful, I know - but at least there machine virtual-did you ID the identifier for the missing VM of old scripts of the inventory by chance?))

    [2011-03-01 15:49:44.093 04732 info "App"] [VpxLRO] - task-354389 - 142778-vm BEGIN - vim. ManagedEntity.destroy-

    [2011-03-01 15:49:46.937 04732 info "App"] [VpxLRO] - FINISHING task-354389 - 142778-vm - vim. ManagedEntity.destroy-

    useful to have a look at this thread:

    http://communities.VMware.com/message/720890

  • To access the log file entries in Oracle 10 g with respect

    Hi all
    How can I access log files entries in oracle 10 g (linux) by questioning... and where it is stored to the origin
    Thank you...

    user8843114 wrote:
    In fact, I do not know the difference between the log files and Audit information... are they equal?
    I know how to obtain audit information... that is to say using a table sys.aud$... is it possible to access the log files if the two above are not equal

    It really depends.

    "Log file" is a generic term. Most of the time it refers to record information in a flat file outside the database. And there are many types of log files. An Oracle database's redo log files and archived log files. An application will have its own log usually on a middle-tier server. An Oracle database can be configured to generate an audit trail in an external file or in a database table (SYS. AUD$). And there are dozens of other ways that people could use the term "log file."

    Depending on how you use the term, it may or may not be possible to access the information in the log file of the database. You will need to know what log file you were interested and where this log file. If the log file is accessible to the database, it would probably be possible to use external tables or same UTL_FILE to access information, depending on the format of the log file.

    Justin

  • synchronization of log file event

    Hi all


    We use Oracle 9.2.0.4 on SUSE Linux 10. In the statspack report, one of the best timed event is
    log file sysnc
    We are in the process. We do not use any storage.IS this a bug of 9.2.0.4 or what is the solution of it
    STATSPACK report for
    
    DB Name         DB Id    Instance     Inst Num Release     Cluster Host
    ------------ ----------- ------------ -------- ----------- ------- ------------
    ai          1495142514 ai                1 9.2.0.4.0   NO      ai-oracle
    
                Snap Id     Snap Time      Sessions Curs/Sess Comment
                ------- ------------------ -------- --------- -------------------
    Begin Snap:     241 03-Sep-09 12:17:17      255      63.2
      End Snap:     242 03-Sep-09 12:48:50      257      63.4
       Elapsed:               31.55 (mins)
    
    Cache Sizes (end)
    ~~~~~~~~~~~~~~~~~
                   Buffer Cache:     1,280M      Std Block Size:         8K
               Shared Pool Size:       160M          Log Buffer:     1,024K
    
    Load Profile
    ~~~~~~~~~~~~                            Per Second       Per Transaction
                                       ---------------       ---------------
                      Redo size:              7,881.17              8,673.87
                  Logical reads:             14,016.10             15,425.86
                  Block changes:                 44.55                 49.04
                 Physical reads:              3,421.71              3,765.87
                Physical writes:                  8.97                  9.88
                     User calls:                254.50                280.10
                         Parses:                 27.08                 29.81
                    Hard parses:                  0.46                  0.50
                          Sorts:                  8.54                  9.40
                         Logons:                  0.12                  0.13
                       Executes:                139.47                153.50
                   Transactions:                  0.91
    
      % Blocks changed per Read:    0.32    Recursive Call %:    42.75
     Rollback per transaction %:   13.66       Rows per Sort:   120.84
    
    Instance Efficiency Percentages (Target 100%)
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
                Buffer Nowait %:  100.00       Redo NoWait %:  100.00
                Buffer  Hit   %:   75.59    In-memory Sort %:   99.99
                Library Hit   %:   99.55        Soft Parse %:   98.31
             Execute to Parse %:   80.58         Latch Hit %:  100.00
    Parse CPU to Parse Elapsd %:   67.17     % Non-Parse CPU:   99.10
    
     Shared Pool Statistics        Begin   End
                                   ------  ------
                 Memory Usage %:   95.32   96.78    
        % SQL with executions>1:   74.91   74.37
      % Memory for SQL w/exec>1:   68.59   69.14
    
    Top 5 Timed Events
    ~~~~~~~~~~~~~~~~~~                                                     % Total
    Event                                               Waits    Time (s) Ela Time
    -------------------------------------------- ------------ ----------- --------
    log file sync                                      11,558      10,488    67.52
    db file sequential read                           611,828       3,214    20.69
    control file parallel write                           436         541     3.48
    buffer busy waits                                     626         522     3.36
    CPU time                                                          395     2.54
              -------------------------------------------------------------
    ^LWait Events for DB: ai  Instance: ai  Snaps: 241 -242
    -> s  - second
    -> cs - centisecond -     100th of a second
    -> ms - millisecond -    1000th of a second
    -> us - microsecond - 1000000th of a second
    -> ordered by wait time desc, waits desc (idle events last)
    
                                                                       Avg
                                                         Total Wait   wait    Waits
    Event                               Waits   Timeouts   Time (s)   (ms)     /txn
    ---------------------------- ------------ ---------- ---------- ------ --------                
    log file sync                      11,558      9,981     10,488    907      6.7
    db file sequential read           611,828          0      3,214      5    355.7
    control file parallel write           436          0        541   1241      0.3
    buffer busy waits                     626        518        522    834      0.4
    control file sequential read          661          0        159    241      0.4
    BFILE read                            734          0        110    151      0.4
    db file scattered read            595,462          0         81      0    346.2
    enqueue                                15          5         19   1266      0.0
    latch free                            109         22          1      8      0.1
    db file parallel read                 102          0          1      6      0.1
    log file parallel write             1,498      1,497          1      0      0.9
    BFILE get length                      166          0          0      3      0.1
    SQL*Net break/reset to clien          199          0          0      1      0.1
    SQL*Net more data to client         5,139          0          0      0      3.0
    BFILE open                             76          0          0      0      0.0
    row cache lock                          5          0          0      0      0.0
    BFILE internal seek                   734          0          0      0      0.4
    BFILE closure                          76          0          0      0      0.0
    db file parallel write                173          0          0      0      0.1
    direct path read                       18          0          0      0      0.0
    direct path write                       4          0          0      0      0.0
    SQL*Net message from client       480,888          0    284,247    591    279.6
    virtual circuit status                 64         64      1,861  29072      0.0
    wakeup time manager                    59         59      1,757  29781      0.0

    Your elapsed time is about 2000 seconds (31: 55 rounded up) - and your log file sync time is 10,000 - which is 5 seconds per second for the duration. Otherwise your session count is about 250 at the beginning and end of snapshot - so if we assume that the number of sessions is stable for the duration, each session has undergone 40 seconds synchronization log file in the meantime. You have saved roughly 1 500 operations in the meantime (0.91 per second, about 13 per cent of restorations) - so synchronize your time log file was on average more than 6.5 seconds by validation.

    Regardless of how you look at it, this suggests that numbers of synchronization of the log file are false, or you had a temporary outage. Given that you had some expectations occupied buffer and control file write expects about 900 m/s each, the hardware failure seems likely.

    Check log file parallel write times to see if this helps to confirm the hypothesis. (Unfortunately some platforms report liog, parallel wriite time of the files properly for earlier versions of 9.2 - so this may not help.)

    You also 15 enqueue waits with an average of 1.2 seconds - check the enqueue statistics in the report section to see what enqueue it was: if it was for example (CF - control file), then it also helps confirm the hypothesis of material.

    It is possible that you had a couple of resets of material or something like this in the meantime that stopped your system quite dramatically for a minute or two.

    Concerning
    Jonathan Lewis
    http://jonathanlewis.WordPress.com
    http://www.jlcomp.demon.co.UK

    "Science is more than a body of knowledge; It's a way of thinking. "
    Carl Sagan

  • Mail creation of massive log files and stop the synchronization with the server

    Since the upgrade to El Capitan (10.11.1), Mail presents two problems persist:

    (1) it generates massive log files, e.g. 2015-12 - 03_IMAPSyncActivity.log these files regularly exceed 10 GB and block Mail and freeze the Mac. Mail can be over 8 GB of RAM. Once removed, Mail recreates the file and fills again, ends up doing more massive files that must also delete.

    (2) mail stop sync with my different IMAP e-mail servers. The only solution I've found that it completely delete the or the e-mail accounts and reinstalls the account. It works for a day (or even a few hours) before it just stops the synchronization, even if a connection test shows he connects with the server. This problem seems to occur on two e-mail accounts separately on two different servers / hosting companies, but not on my Gmail accounts.

    These two problems occur on both my iMac (21.5 ", mid-2011, 2.5 GHz Intel Core i5, 8 GB RAM) and MacBookPro (13', mid-2011, 2.5 GHz Intel core i5, 4 GB RAM).

    Full of ideas gratefully received!

    Thank you!

    Please take these steps to remove the mail folders 'sandbox '.

    Step 1

    Back up all data.

    Triple-click anywhere in the line below on this page to select this option:

    ~/Library/Containers/com.apple.mail

    Right-click or Ctrl-click on the highlighted line and select

    Services ▹ Reveal

    the contextual menu.* A Finder window should open up with a folder named "com.apple.mail" selected. If this is the case, move the selected folder - not only its content - on the desktop. Open the window Finder for now.

    Restart the computer. Launch Mail and test. If the problem is resolved, you may have to re-create some of your e-mail settings. Any writing paper custom that you created may be lost. Ask for instructions if you want to keep these data. You can then delete the folder that you moved and close Finder.

    CAUTION: If you change the content of the sandbox, but leave the folder itself in place, Mail may hang or starts any. Remove the tray to sand everything will be rebuilt automatically.

    Step 2

    If step 1 does not resolve the problem, repeat with this line:

    ~/Library/Containers/com.apple.MailServiceAgent

    * If you do not see the item context menu copy the selected text in the Clipboard by pressing Control-C key combination. In the Finder, select

    Go ▹ go to the folder...

    from the menu bar and paste it into the box that opens by pressing command + V. You won't see what you pasted a newline being included. Press return.

  • AoE3 cannot create log files and will not run.

    AOE3 I installed on my machine. The game is stored on an external hard drive that is almost new and has yet to give me any problems. Whenever I try to launch the game, I get the following error "could not create the log file. Ensure that you have all rights in the directory where the Age of Empires III, and that you have available disk space. \n Erroe: access denied. »

    I am the only user of this computer, my account has full administrator access and there is more than a terabyte of disk space. I don't knopw not why there is no problem running this game - especially as all the other files on the road very well works.

    Any help would be greatly appreciated.

    Thank you.

    Delete the log file, "Age3Log.txt", which could give this error message from My Documents\My Games\Age of Empires 3 'folder.

    Try to delete this file as an administrator, and then click Connect.

    Girls. You never know what they think.

  • Files dump debugging and log files to install in the disk cleanup window

    The two points above appear in my disk cleanup window, and I wonder if it is o: k to remove. The first occupies a space of 283 KB and the other 2537 KB. The description of these two files in the disk cleanup window is "files created by Windows.

    I have a second question: I have about 5 GB left as free space on my hard drive of 'C '. If this is considered sufficient, or should I get more? I am a senior retired and only use the computer for basic items.

    I have a computer Dell Dimension 4100 with Service Pack 3 for Windows XP - Home Edition Operating System - desktop, 111 Pentium with 930 MHZ, 20 GB hard drive and 512 MB RAM. The file system is FAT 32.

    Thank you very much for your cooperation and your response.

    "Debug dump files" are copies of the contents of your computer's memory.

    When a program crashes, it will sometimes "dump" all or part of the
    contents of the RAM to a file on your hard drive. The file is useful to
    a technical support person who tries to understand why the program
    crashed. (Most of the tech support people have no idea what to do with a dump
    file.)

    Unless you have been asked to submit a dump for inspection file, feel free
    to get rid of them.

    Setup logs files are "log" installation of a program. If you
    Watch a (they are really just text files - open with Notepad) you will have
    see that they give a blow by blow description of how the program has been
    installed, glorious detail.

    If a program does not properly settled, the Setup log will tell you what that
    was wrong. If you have not had a facility wrong, do not hesitate to ditch
    These files as well.

    On the question of how much free space on the disk is enough... There is no
    standards. 15% of your drive must be empty in order for Windows
    The Defragmenter built into the work. In addition, it is to you.
     
    Your processor is another story. It's not easy to find effective security
    software that will work on a P3. Even if you only use the computer
    for the basic elements, an internet connection is pretty basic for a online
    criminal to find you.

    Another response of the community of Windows XP newsgroups

Maybe you are looking for