Moving oracle data file while the database is broken down

I have a database in montage mode while I restore the last of my data files through rman, which won't be finished for a day and a half, and then there's a mountain of archivelogs to apply. I need to pass the data to another disk files so that I have room to restore this file. I looked around a few instructions, but they are all different. Many of them tell me to stop first of all, the database is not a good option with the restoration goes. More say to use the operating system to copy the file, but a site ignored that. It is even said that I can't log out of my session between saying oracle sql to move the file and open the database. All I get is a list of steps that all take my database work.

Currently, I copy the file to the new destination of Linux, and I'm confused about what my options are and how to get the data file of movement before Oracle is running out of space for my restoration, without doing anything to make things worse. I guess I can just wait for the copy at the end and then run something like "alter rename file ' / old/my_file.dbf ' to ' / new/my_file.dbf '; If I have to cancel the copy for a reason, it is not a big deal, as long as I don't have the space in time.

A quick look, it seems that you should be able to move the data file and to a change of name as you have suggested, as long as you are in editing mode. I hope you have good backup (or an easier way to recover) in case it does not work.

MOS Doc 115424.1 - how to rename or move data files and log files

HTH
Srini

Tags: Database

Similar Questions

  • put the offline data file when the database is in NOARCHIVELOG mode

    My question is when the database is in log mode Archive No. I am not able to take the database offline.

    When I tried in my computer I noticed set-aside.

    CASES1:

    SYS > alter database datafile 5 offline;

    ERROR at line 1;

    ORA-01145: offline immediately rejected at least media recovery enabled.

    case 2:

    SYS > alter database datafile 5 immediate offline;

    ERROR at line 1;

    ORA - 00933:SQL not correctly completed order

    CASE3:

    I tried the command drop alter database datafile offline 6; (in NOARCHIVELOG mode) and it is to show the same effect that alter database datafile offline 6; (in ARCHIVELOG mode).

    * In NOARCHIVELOG mode are we really drop the data file to use the offline data file? You please tell me the effect of the dropkeyword.

    Oracle protects you. Please review ARCHIVELOG and NOARCHIVELOG mode.

    When you take the database offline in NOARCHIVELOG mode:

    (1) the file is offline

    (2) the RCS in which you took the database offline will be is more to be in the online redo logs

    (3) at this point, you need to perform a media recovery to enable again the file online (ie. restore the database)

    So you can see why you shouldn't do that.

    The DROP Offline command does not actually remove the data file. It removes the control file file details and updates the offline data file. It can be used for purposes of recovery or when you actually plan on deleting data files.

  • In my production, size of data file for the database system is 98.2%

    Hi all

    Can u please help me what I do, it's my production database and system tablespace is show 98.2% full... Details

    TablespaceSYSTEM
    StatusOnline
    Size of the file (MB)3,380.00
    Auto expandYes
    Increment (MB)10 h 00
    Maximum file size (MB)32,767.00

    Hello

    you don't need to do anything about it either because the auto extend turned on, it will go to maxsize 32 GB

  • When OMF add the data file in the tablespace

    Hi friends,

    We use oracle 11.2 with CMS (oracle managed file) for the database on the linux platform. the parameter - DB_CREATE_FILE_DEST had been set. at the present time, we received a message from global warming this tablespace criterion is 85% full.

    According to the document of the oracle, OMF auto will add a data file in the tablespace tast. more than a week, I do not see any data file has been added by OMF.

    I want to know when OMF adds the data file in the tablespace? 85% or 95% or some (parameter) setting this action?

    Thank you

    newdba

    OMF does not automatically add a new data file.  You must explicitly add a new data file with the ALTER TABLESPACE ADD DATAFILE command tbsname.

    What is OMF is to provide a Unique name and a default size (100 MB) for the data file.  That is why the ALTER TABLESPACE... Command ADD DATAFILE that you run didn't need to specify the file size or file name.

    Hemant K Collette

  • ORA-01157: cannot identify/lock data file error in database pending.

    Hello

    I have a back-end database and the base ensures (11.2.0.1.0) that runs in ASM with the names of different diskgroup. I applied an incremental backup on a standby database to solve the gap newspaper archive and generated a controlfile to standby in the primary database and restore the standby database controlfile. But when I started the MRP process his starts not and lift error in the alerts log ORA-01157: cannot identify/lock file. When I questioned the standby database file it shows the location on primary data filenames not the database pending.

    ******************************

    PRIMARY DATABASE

    *****************************

    SQL > select name from v$ datafile;


    NAME

    --------------------------------------------------------------------------------

    +Data/OraDB/datafile/system.256.788911005

    +Data/OraDB/datafile/SYSAUX.257.788911005

    +Data/OraDB/datafile/undotbs1.258.788911005

    +Data/OraDB/datafile/users.259.788911005

    ****************************************

    BACKUP DATABASE

    ****************************************

    SQL > select name from v$ datafile;


    NAME

    --------------------------------------------------------------------------------

    +STDBY/OraDB/datafile/system.256.788911005

    +STDBY/OraDB/datafile/SYSAUX.257.788911005

    +STDBY/OraDB/datafile/undotbs1.258.788911005

    +STDBY/OraDB/datafile/users.259.788911005

    The actual physical location of files of database Eve in ASM in the standby server is shown below

    ASMCMD > pwd

    + STDBY/11gdb/DATAFILE

    ASMCMD >

    ASMCMD > ls

    SYSAUX.259.805921967

    SYSTEM.258.805921881

    UNDOTBS1.260.805922023

    USERS.261.805922029

    ASMCMD >

    ASMCMD > pwd

    + STDBY/11gdb/DATAFILE

    I even tried to rename the data files in the database backup, but it throws error

    ERROR on line 1:

    ORA-01511: Error renaming data/log files

    ORA-01275: RENAME Operation is not allowed if management undo file is

    Automatic.

    Kind regards

    007

    You must specify the complete location

    *.db_file_name_convert='+data/OraDB/datafile/,'+STDBY/11gdb/DATAFILE /'

    and to rename the data file, your standby_file_management parameter must be set to MANUAL.

  • After you rename the data file, cannot start database

    Oracle 11 g 2
    OEL 5 (ASM)
    Network infrastructure (cluster install) - no CARS yet

    Something interesting happened. Perhaps this question might be more suited to the "ASM" section, but that's.

    I gave a data in ASM file an alias with the following command (the ASM instance):
    SQL> alter diskgroup DATA add alias '+DATA/oradb/datafile/users01.dbf' for '+DATA/oradb/datafile/users.253.795630487';
    
    Diskgroup altered.
    Then, as mentioned in Note: 564993.1, we need to update the database as well with the new alias. However, when I went to the stopping and starting of the database, I received the following:
    SQL> shutdown immediate;
    Database closed.
    Database dismounted.
    ORACLE instance shut down.
    SQL> startup mount;
    ORA-12514: TNS:listener does not currently know of service requested in connect descriptor
    It's strange. Nothing has changed except the alias being added to the data file. This is all before operation. That's happened?

    I think that it may have to do with the different ASMOPER, ASMADMIN, ASMDBA groups that have been installed as part of the installation of the grid Infrastructure. In addition, the listener is running out of the GI home.

    All my variables environment (e.g. ORACLE_SID, ORACLE_HOME, etc.) are defined.

    Any ideas?

    Thank you.

    Hello
    When you connect with tnsnames alias while the database is down the listner did not know the service, so you must start the database on the server database with sqlplus / as sysdba. After starting the registry of data base with the listener and your connection works again.
    Alternativ you can add a service to your listener.ora to connect with him while he is arrested, as long as you do not have it you must use local connect which means sqlplus / as sysdba.

    concerning
    Peter

  • How do I reduce the data file after the moving images to another data file

    All,

    I did a test of simepl to see how should I reduce my data file after the transfer of the objects thereof.

    I had an image in my data file USER_image, which is about 9 GB in size. I created the new data under the new tablpespace file and moved all the images in this new space of tables. So my old tablespace of image with its data file is now empty. At least there is not all user objects.

    Now I tried to resize the old datafile (USER_image) to a smaller value, but I keep this error
    "ORA-03297: file contains data beyond the value required to RESIZE.

    The file size is 9 GB and I tried to resize to 8 GB but it still does not allow me.

    Is it possible to resize a data file after moving all its purpose?
    I'm something wrong or missing something?

    Thanks in advance.

    If you old image tablespace is indeed empty (you need to run a statement against DBA_SEGMENTS to check) then you can bring it, data files included.
    After this, you can create a 'new' tablespace with the old name and desired data file sizes.

  • Why not place the Oracle data files on local disks

    Hi, I want to ask a fundamental question.

    Almost any Oracle installation I've seen, data files have been placed on Mount points, disk etc..

    Is there a reason for this? Why no one does place for local disks oracle data files? I couldn't find the answer on the installation guide

    I ask this because I will install oracle (10g) on vmware and trying to decide whether to put data on the vmdk (virtual computer local disk) files...
    Is - this recommended? If not, why it is recommended to put data files to a disc of fixation for point/external etc..


    Thanks in advance

    Hello

    People store data files on separate disks
    (1) bettter performance
    (2) the data can be in a centralized storage
    (3) if the OS crashes always files are safe
    (4) configurable with raid options
    (5) hot Swapping
    (6) easily can add additional storage

    Kind regards
    Delphine K

  • Download csv file into the database using apex

    Hi all

    I use apex 4 and express oracle 10g, I need to upload the .csv file in the database for one of my queries, I just discussion forum for solutions, I found too, but some how its does not work for me.

    below mentioned is error and the code


    ERROR:

    ORA-06550: line 38, column 8: PLS-00221: 'V_DATA_ARRAY' is not a procedure or is undefined ORA-06550: line 38, column 8: PL/SQL: statement ignored ORA-06550: line 39, column 8: PLS-00221: 'V_DATA_ARRAY' is not a procedure or is undefined ORA-06550: line 39, column 8: PL/SQL: statement ignored ORA-06550: line 40, column 8: PLS-00221: 'V_DATA_ARRAY' is not a procedure or is undefined ORA-06550 : line 40, column 8: PL/SQL: statement ignored ORA-06550: line 41, column 8: PLS-00221: 'V_DATA_ARRAY' is not a proc
    Error
    Ok


    CODE:
    DECLARE
    v_blob_data BLOB;
    v_blob_len NUMBER;
    V_POSITION NUMBER;
    v_raw_chunk RAW (10000);
    v_char char (1);
    number of c_chunk_len: = 1;
    v_line VARCHAR2 (32767): = NULL;
    v_data_array wwv_flow_global.vc_arr2;
    BEGIN
    -Read data from wwv_flow_files
    Select blob_content from v_blob_data
    of wwv_flow_files where filename = "DDNEW.csv";

    v_blob_len: = dbms_lob.getlength (v_blob_data);
    V_POSITION: = 1;

    -Read and convert binary to a char
    WHILE (v_position < = v_blob_len) LOOP
    v_raw_chunk: = dbms_lob.substr(v_blob_data,c_chunk_len,v_position);
    v_char: = chr (hex_to_decimal (rawtohex (v_raw_chunk)));
    v_line: = v_line | v_char;
    V_POSITION: = v_position + c_chunk_len;
    -When a whole line is retrieved
    IF v_char = Chr (10) THEN
    -Convert comma to: use of wwv_flow_utilities
    v_line: = REPLACE (v_line, ', ' :'); ")
    -Converting each column separated by: in the data table
    v_data_array: = wwv_flow_utilities.string_to_table (v_line);
    -Insert data into the target table
    EXECUTE IMMEDIATE ' insert into TABLE_X (v1, v2, v3, v4, v5, v6, v7, v8, v9, v10, v11)
    "values (: 1,: 2,: 3: 4:5:6,: 7,: 8,: 9,: 10:11).
    USING
    v_data_array (1),
    v_data_array (2),
    v_data_array (3),
    v_data_array (4);
    v_data_array (5);
    v_data_array (6).
    v_data_array (7);
    v_data_array (8);
    v_data_array (9);
    v_data_array (10);
    v_data_array (11);

    -Remove
    v_line: = NULL;
    END IF;
    END LOOP;
    END;



    what I understand of it's system does not identify v_data_array as table for some reasons, please help me.



    initially the system was in error for hex_to_decimal, but I managed to get this feature on the discussion forum, and now it seems to be ok. but v_data_array problem is still there.



    Thanks in advance

    concerning

    Uday

    Hello

    Correct errors in your sample, I have

    Problem 1

    select blob_content into v_blob_data
    from wwv_flow_files where filename = 'DDNEW.csv'; 
    

    TO

    select blob_content into v_blob_data
    from wwv_flow_files where name = :P1_FILE;
    

    Problem 2

    EXECUTE IMMEDIATE 'insert into TABLE_X (v1, v2, v3, v4 ,v5, v6, v7,v8 ,v9, v10, v11)
    values (:1,:2,:3,:4,:5,:6,:7,:8,:9,:10,:11)'
    USING
    v_data_array(1),
    v_data_array(2),
    v_data_array(3),
    v_data_array(4);
    v_data_array(5);
    v_data_array(6);
    v_data_array(7);
    v_data_array(8);
    v_data_array(9);
    v_data_array(10);
    v_data_array(11);  
    

    TO

    EXECUTE IMMEDIATE 'insert into TABLE_X (v1, v2, v3, v4 ,v5, v6, v7,v8 ,v9, v10, v11)
    values (:1,:2,:3,:4,:5,:6,:7,:8,:9,:10,:11)'
    USING
    v_data_array(1),
    v_data_array(2),
    v_data_array(3),
    v_data_array(4),
    v_data_array(5),
    v_data_array(6),
    v_data_array(7),
    v_data_array(8),
    v_data_array(9),
    v_data_array(10),
    v_data_array(11);  
    

    And I have created table missing

    CREATE TABLE TABLE_X
      (
        v1  VARCHAR2(255),
        v2  VARCHAR2(255),
        v3  VARCHAR2(255),
        v4  VARCHAR2(255),
        v5  VARCHAR2(255),
        v6  VARCHAR2(255),
        v7  VARCHAR2(255),
        v8  VARCHAR2(255),
        v9  VARCHAR2(255),
        v10 VARCHAR2(255),
        v11 VARCHAR2(255)
      );
    

    Kind regards
    Jari

    Published by: jarola November 19, 2010 15:03

  • Distribute the data file with the cod file?

    Hi all

    My application runs from a file of database on camera flash memory.  While the application can transfer data from my server to complete the database, the first download is quite large and takes a long time to run (I'm working on it).

    It occurred to me that an alternative might be to provide a database of pre-populated with the app file, while the app will run straight out of the box without needing to first synchronization.

    Does anyone know if the App World store supports this - distribution of data, and the app?

    Thank you very much...

    You're right about the wasted space.  How about you always include a resource file, but make sure that if your program needs to read data from "base", read from the resource file, and if it's new data, read from the DB.  I don't know what your app does so I don't know if it's feasible or not, but if this is the case, that would solve your problem.  What do you think?

  • Download any file in the database column

    Hi all

    Is it possible that I can download any file in a column of data?

    I use oracle forms 6i and oracle database 10g.

    Kind regards

    Atif Zafar

    Dear Sir

    You can select the built-in file using GET_FILE_NAME. And then copy the file into the database using HOST Directory integrated (using the CMD). After you copy the file in the database directory, you can call the procedure of database to update the database file.

    Manu.

  • ORA-19846: cannot read the header of the data file of the remote site 21

    Hello

    I have a situation or I can say a scenario. It is purely for testing base. Database is on 12.1.0.1 on a Linux box using ASM (OMF).

    Standby is created on another machine with the same platform and who also uses ASM (OMF) and is in phase with the primary. Now, suppose I have create a PDB file on the primary of the SEED and it is created successfully.

    After that is a couple of log, do it again passes to the waiting, but MRP fails because of naming conventions. Agree with that! Now, on the primary, I remove the newly created PDB (coward the PDB newly created). Once again a couple of switches of newspapers which is passed on to the wait. Of course, the wait is always out of sync.

    Now, how to get back my watch in sync with the primary? I can't roll method until the required data (new PDB) file does not exist on the main site as well. I get the following error:

    RMAN > recover database service prim noredo using backupset compressed;

    To go back to November 8, 15

    using the control file of the target instead of recovery catalog database

    allocated channel: ORA_DISK_1

    channel ORA_DISK_1: SID = 70 = device = DISK stby type instance

    RMAN-00571: ===========================================================

    RMAN-00569: = ERROR MESSAGE STACK FOLLOWS =.

    RMAN-00571: ===========================================================

    RMAN-03002: failure of the command recover at the 18:55:32 08/11/2015

    ORA-19846: cannot read the header of the data file of the remote site 21

    The clues on how to I go ahead? Of course, recreating the eve is an option as its only based on test, but I don't want recreation.

    Thank you.

    I tried like below:

    1 a incremental backup of the primary of the CNS where off the eve also taken primary backup controlfile as Eve format.

    2 copy the backup of the watch parts, catalogged them on the day before.

    3 recovered Eve with noredo option - it fails here with the same error pointing to the 21 data file.

    OK, understood. Try not to get back the day before first, rather than restore the controlfile and then perform the restoration.

    Make it like:

    1. take incremental backup of primary SNA, also ensures the backup controlfile format.

    2. copy pending, get the location of the data file (names) by querying v$ datafile on the eve. Restore the controlfile ensures from the backup controlfile you took on primary and mount.

    3. Since you are using OMF, the path of primary and standby data file will be different. (/). If you require catalog data from the database files pending.

    (Reason: you restore controlfile from elementary to step 2, which takes place from the main access road). Use the details that you obtained in step 2 and catalog them.

    4. turn the database copy by RMAN. (RMAN > switch database to copy ;))

    5 Catalog backup items that you copied in step 2.

    6. recover the standby database using 'noredo' option.

    7. finally start the MRP. This should solve your problem.

    The reason I say this works is because here, you restore the controlfile to primary first, which will not have details 21, datafile, and then you are recovering. So it must succeed.

    In the previous method, you tried to first collect all the day before, and then restore the controlfile. While remedial classes, always watch seeks datafile 21 as he controlfile is not yet updated.

    HTH

    -Jonathan Rolland

  • Automatically deleted from the storage of the DSO after data files using the command cp ASMCMD... why?

    I installed Oracle database on Linux 6.6, Virtualbox 11.2.0.1.0.

    I used ASM for database storage.

    When I want to perform a backup in offline mode. I stop the database data files and copy to the filesystem as follows (for example).

    ASMCMD [+] > cp +DATA/ora11g/datafile/system.270.883592533/u01/app/oracle

    The files copied with success and also any other data files, logs and controlfiles files. Then, I also rename all the files in editing mode.

    After that, I used the backup files to start the oracle of backup (storage non - ASM with new pfile edited)

    Open successfully from a backup database.

    But when I want to use old spfile to open the database for the storage of the DSO again, a few errors have occurred as no data files in ASM storage.

    I check the contents of the ASM storage with ASMCMD commands. And realize that there only spfile and controlfile located ASM storage and other files: data files and online redo logfiles automatically deleted.

    Why the data files and log files deleted ASM storage? Is this normal? I have no delete all files of the DSO.

    It is actually deleted from the use of the cp command?

    Exactly what we say.  Data in ASM files are OMF (Oracle managed files).  The RENAME translates the DSO by deleting the original file.

    Hemant K Collette

  • Data files to the tablespace sysaux listed several times v $ datafile

    When I run the query - select name, number of status from v$ datafile;

    the data files for system tablespaces are included several times.

    For example, for sysaux, output a below entries. Why is this?

    It is the same with the space of system tables and users.

    The configuration is Oracle 12 c

    I've used this in earlier versions of oracle, but this is something new that I am observing with 12 c.

    + ASM_DG/gdb/DataFile/SYSAUX.269.838407323, 3

    + ASM_DG/gdb/DD7C48AA5A4404A2E04325AAE80A403C/DataFile/SYSAUX.273.838407639, 7

    + ASM_DG/gdb/F159D742BCE602C0E0438B59D10A581B/DataFile/SYSAUX.258.838409993, 10

    When I use the query-

    SELECT FILE_NAME, BLOCKS, NOM_TABLESPACE

    FROM DBA_DATA_FILES;

    Then only the single entry appears

    + ASM_DG/gdb/DataFile/SYSAUX.269.838407323, 108800, SYSAUX

    Any help will be great.

    Hello

    It's the new architecture shared in action.

    You get a sysaux/system for each plug-in database data file, but the views dba only show data for the current container.

    This is probably the biggest architectural change I've seen in oracle (and I worked with her since 7.2...)

    I recommend you to read the oracle documentation (and then read them again when it totally confuses you until you get it).

    See you soon,.

    Rich

  • When loading, error: field in the data file exceeds the maximum length

    Oracle Database 11 g Enterprise Edition Release 11.2.0.3.0 - 64 bit Production

    PL/SQL Release 11.2.0.3.0 - Production

    CORE Production 11.2.0.3.0

    AMT for Solaris: 11.2.0.3.0 - Production Version

    NLSRTL Version 11.2.0.3.0 - Production

    I am trying to load a table, small size (110 lines, 6 columns).  One of the columns, called NOTES is less error when I run the load.  That is to say that the size of the column exceeds the limit max.  As you can see here, the column of the table is equal to 4000 bytes)

    CREATE TABLE NRIS. NRN_REPORT_NOTES

    (

    Sys_guid() NOTES_CN VARCHAR2 (40 BYTE) DEFAULT is NOT NULL.

    REPORT_GROUP VARCHAR2 (100 BYTE) NOT NULL,

    POSTCODE VARCHAR2 (50 BYTE) NOT NULL,

    ROUND NUMBER (3) NOT NULL,

    VARCHAR2 (4000 BYTE) NOTES,

    LAST_UPDATE TIMESTAMP (6) WITH ZONE SCHEDULE systimestamp NOT NULL default

    )

    TABLESPACE USERS

    RESULT_CACHE (DEFAULT MODE)

    PCTUSED 0

    PCTFREE 10

    INITRANS 1

    MAXTRANS 255

    STORAGE)

    80K INITIAL

    ACCORDING TO 1 M

    MINEXTENTS 1

    MAXEXTENTS UNLIMITED

    PCTINCREASE 0

    DEFAULT USER_TABLES

    DEFAULT FLASH_CACHE

    DEFAULT CELL_FLASH_CACHE

    )

    LOGGING

    NOCOMPRESS

    NOCACHE

    NOPARALLEL

    MONITORING;

    I did a little investigating, and it does not match.

    When I run

    Select max (lengthb (notes)) in NRIS. NRN_REPORT_NOTES

    I got a return of

    643

    .

    Which tells me that the larger size of this column is only 643 bytes.  But EACH insert is a failure.

    Here is the header of the file loader and first couple of inserts:

    DOWNLOAD THE DATA

    INFILE *.

    BADFILE '. / NRIS. NRN_REPORT_NOTES. BAD'

    DISCARDFILE '. / NRIS. NRN_REPORT_NOTES. DSC"

    ADD IN THE NRIS TABLE. NRN_REPORT_NOTES

    Fields ended by '; '. Eventually framed by ' |'

    (

    NOTES_CN,

    REPORT_GROUP,

    Zip code

    ALL ABOUT NULLIF (R = 'NULL'),

    NOTES,

    LAST_UPDATE TIMESTAMP WITH TIME ZONE ' MM/DD/YYYY HH24:MI:SS. FF9 TZR' NULLIF (LAST_UPDATE = 'NULL')

    )

    BEGINDATA

    | E2ACF256F01F46A7E0440003BA0F14C2; | | DEMOGRAPHIC DATA |; A01003; | 3 ; | demographic results show that 46% of visits are made by women.  Among racial and ethnic minorities, the most often encountered are native American (4%) and Hispanic / Latino (2%).  The breakdown by age shows that the Bitterroot has a relatively low of children under 16 (14%) proportion in the population of visit.  People over 60 represent about 22% of visits.   Most of the visitation comes from the region.  More than 85% of the visits come from people who live within 50 miles. | ; 29/07/2013 0, 16:09:27.000000000 - 06:00

    | E2ACF256F02046A7E0440003BA0F14C2; | | DESCRIPTION OF THE VISIT; | | A01003; | 3 ; | most visits to the Bitterroot are relatively short.  More than half of the visits last less than 3 hours.  The median duration of visiting sites for the night is about 43 hours, or about 2 days.  The average Wilderness visit lasts only about 6 hours, although more than half of these visits are shorter than the duration of 3 hours.   Most of the visits come from people who are frequent visitors.  Over thirty percent are made by people who visit between 40 and 100 times a year.  Another 8% of visits from people who say they visit more than 100 times a year. | ; 29/07/2013 0, 16:09:27.000000000 - 06:00

    | E2ACF256F02146A7E0440003BA0F14C2; | | ACTIVITIES |. A01003; | 3 ; | most often reported the main activity is hiking (42%), followed by alpine skiing (12%) and hunting (8%).  More than half of the report visits participating in the relaxation and the display landscape. | ; 29/07/2013 0, 16:09:27.000000000 - 06:00

    Here's the full start of log loader, ending after the return of the first row.  (They ALL say the same error)

    SQL * Loader: Release 10.2.0.4.0 - Production Thu Aug 22 12:09:07 2013

    Copyright (c) 1982, 2007, Oracle.  All rights reserved.

    Control file: NRIS. NRN_REPORT_NOTES. CTL

    Data file: NRIS. NRN_REPORT_NOTES. CTL

    Bad File:. / NRIS. NRN_REPORT_NOTES. BAD

    Discard File:. / NRIS. NRN_REPORT_NOTES. DSC

    (Allow all releases)

    Number of loading: ALL

    Number of jump: 0

    Authorized errors: 50

    Link table: 64 lines, maximum of 256000 bytes

    Continuation of the debate: none is specified

    Path used: classics

    NRIS table. NRN_REPORT_NOTES, loaded from every logical record.

    Insert the option in effect for this table: APPEND

    Column Position Len term Encl. Datatype name

    ------------------------------ ---------- ----- ---- ---- ---------------------

    FIRST NOTES_CN *;  O (|) CHARACTER

    REPORT_GROUP NEXT *;  O (|) CHARACTER

    AREA CODE FOLLOWING *;  O (|) CHARACTER

    ROUND                                NEXT     *   ;  O (|) CHARACTER

    NULL if r = 0X4e554c4c ('NULL' character)

    NOTES                                NEXT     *   ;  O (|) CHARACTER

    LAST_UPDATE NEXT *;  O (|) DATETIME MM/DD/YYYY HH24:MI:SS. FF9 TZR

    NULL if LAST_UPDATE = 0X4e554c4c ('NULL' character)

    Sheet 1: Rejected - error in NRIS table. NRN_REPORT_NOTES, information ABOUT the column.

    Field in the data file exceeds the maximum length.

    I don't see why this should be failed.

    Hello

    the problem is bounded by default, char (255) data... Very useful, I know...

    you need two, IE sqlldr Hat data is longer than this.

    so change notes to notes char (4000) you control file and it should work.

    see you soon,

    Harry

Maybe you are looking for