ORA-39139: Data Pump

As he tried to export certain patterns of mine, he ended up showing...
'complete with 1 error (s)'

Looking at the export log file that I found:
ORA-39139: data pump does not support the XMLSchema objects. TABLE_DATA: "BETUSR." "' NSC_SERVICES ' will be ignored.

What it means?

expdp does not support the columns based on the XMLSchema before 11.x or XMLSchemas.
To fix this, you can use the "old" exp utility

Take a look at this note:
How to move the XML schema based on data from xmltype 10 g to another schema / 11 g [1318012.1 ID]

Tags: Database

Similar Questions

  • ORA-39097: Data Pump job encountered the error unexpected-39076

    Hello world

    Today, I tried to take a pump dump to export my test database (specific table), the version is 10.2.0.4 on Solaris10(64-bit) and I got the following error message

    Work 'SYSTEM '. "" SYS_EXPORT_TABLE_23 "carried out at 09:51:36
    ORA-39097: Data Pump job encountered the error unexpected-39076
    ORA-39065: exception of unexpected master process in KUPV$ FT_INT. DELETE_JOB
    ORA-39076: cannot remove SYS_EXPORT_TABLE_23 work for the SYSTEM user
    ORA-06512: at "SYS." DBMS_SYS_ERROR', line 95
    ORA-06512: at "SYS." "KUPV$ FT_INT", line 934
    ORA-31632: main table "of the SYSTEM. SYS_EXPORT_TABLE_23"not found, invalid or unreachable
    ORA-06512: at "SYS." DBMS_SYS_ERROR', line 95
    ORA-06512: at "SYS." "KUPV$ FT_INT", line 1079
    ORA-20000: failed to send the e-mail message from pl/sql because of:
    ORA-29260: network error: Connect failed because target host or object does not exist
    ORA-39097: Data Pump job encountered the error unexpected-39076
    ORA-39065: exception unexpected master process in HAND
    ORA-39076: cannot remove SYS_EXPORT_TABLE_23 work for the SYSTEM user
    ORA-06512: at "SYS." DBMS_SYS_ERROR', line 95
    ORA-06512: at "SYS." "KUPV$ FT_INT", line 934
    ORA-31632: main table "of the SYSTEM. SYS_EXPORT_TABLE_23"not found, invalid or unreachable
    ORA-06512: at "SYS." DBMS_SYS_ERROR', line 95
    ORA-06512: at "SYS." "KUPV$ FT_INT", line 1079
    ORA-20000: failed to send the e-mail message from pl/sql because of:
    ORA-29260: network error: Connect failed because target host or object does not exist

    I hope that the export dumpfile is valid, but I don't know why I get this error message. One faced this kind of problem. please me tips

    Thank you

    Shan

    Once you see this:

    Work 'SYSTEM '. "" SYS_EXPORT_TABLE_23 "carried out at 09:51:36

    The Data Pump task is done with the dumpfile. It is some clean that is needed and it looks like something in the cleaning failed. Don't know what it was, but you dumpfile should be good. An easy way to test is to run impdp with sqlfile. This will make all import will do, but instead to create objects, he writes the ddl in sql file.

    Impdp directory of the user/password sqlfile = my_test.sql = your_dir dupmfile = your_dump.dmp...

    If it works, then your dumpfile should be fine. The last action of export, it is write the main table Data Pump in the dumpfile. The first thing that import is read this table. So, if you can read it in (which sqlfile impdp) your dump is good.

    Dean

  • Error data pump using the concept of the tablespace of transport.

    Hello

    I am facing the following error when I try to use the concept of transport tablespace.


    ORA-39123: Data Pump tablespace portable job abandoned
    ORA-01141: Error renaming file /LP2SANdataR53/oraData/Data/TTSIDX5.dbf - new file data "not found
    ORA-27046: file size is not a multiple of the logical block size
    Additional information: 1
    Additional information: 1073724207
    Additional information: 8192


    For less data (for 50-100 MB) size, this concept works very well. But when I increase the size of the tablespace to 1 GB and when I try to use the concept of transport tablespace I get this error.

    More than three times, I tried. Small, it works fine.

    Can someone give the solution for this.

    Mark the correct answer that helped you for future reference.

    concerning

  • migration from 10g to 12 c using the data pump in

    Hi, while I used the data pump at the level of the schema before, I'm relatively new to the full database import.

    We are trying a full database migration to 10.2.0.4 to 12 c using the complete method of database data pump over db link.

    the DBA has indicated to avoid move SYSAUX and SYSTEM objects. but initially during the documentation review, it appeared that these objects are not exported since the TRANSPORTABLE given target = system NEVER. If anyone can confirm this? done import and export log refers to the objects I thought would not:

    ...

    19:41:11.684 23 FEBRUARY 15:Estimated TABLE_DATA 3718 objects in 77 seconds

    19:41:12.450 23 February 15: total estimation using BLOCKS method: 52,93 GB

    19:41:14.058 23 February 15: object DATABASE_EXPORT/TABLESPACE of treatment type

    20:10:33.185 23 February 15: ORA-31684: TABLESPACE object type: 'UNDOTBS1' already exists

    20:10:33.185 23 February 15: ORA-31684: TABLESPACE object type: 'SYSAUX' already exists

    20:10:33.185 23 February 15: ORA-31684: TABLESPACE object type: 'TEMP' already exists

    20:10:33.185 23 February 15: ORA-31684: TABLESPACE object type: 'USERS' existing

    20:10:33.200 23 FEBRUARY 15:96 objects TABLESPACE finished in 1759 seconds

    20:10:33.208 23 February 15: treatment of type of object DATABASE_EXPORT/PROFILE

    20:10:33.445 23 FEBRUARY 15:7 PROFILE items finished in 1 seconds

    20:10:33.453 23 February 15: treatment of DATABASE_EXPORT/SYS_USER/USER object type

    20:10:33.842 23 FEBRUARY 15:1 USER objects ended in 0 seconds

    20:10:33.852 23 February 15: treatment of DATABASE_EXPORT/SCHEMA/USER object type

    20:10:52.368 23 February 15: ORA-31684: USER object type: 'OUTLN' already exists

    20:10:52.368 23 February 15: ORA-31684: USER object type: 'ANONYMOUS' already exists

    20:10:52.368 23 February 15: ORA-31684: USER object type: 'OLAPSYS' already exists

    20:10:52.368 23 February 15: ORA-31684: USER object type: 'MDDATA' already exists

    20:10:52.368 23 February 15: ORA-31684: USER object type: 'SCOTT' already exists

    20:10:52.368 23 February 15: ORA-31684: USER object type: 'LLTEST' already exists

    20:10:52.372 23 FEBRUARY 15:Finished objects USER 1140 in 19 seconds

    20:10:52.375 23 February 15: object DATABASE_EXPORT/ROLE of treatment type

    20:10:55.255 23 February 15: ORA-31684: object ROLE type: 'SELECT_CATALOG_ROLE' already exists

    20:10:55.255 23 February 15: ORA-31684: object ROLE type: 'EXECUTE_CATALOG_ROLE' already exists

    20:10:55.255 23 February 15: ORA-31684: object ROLE type: 'DELETE_CATALOG_ROLE' already exists

    20:10:55.256 23 February 15: ORA-31684: object ROLE type: 'RECOVERY_CATALOG_OWNER' already exists

    ...

    the most insight.

    The schema SYS, CTXSYS and MDSYS ORDSYS are not exported using exp/expdp

    DOC - ID: Note: 228482.1

    I guess that he has already installed a 12 c software and created an itseems database - so when you have imported you have this "already exists."

    Every time the database is created and the software installed by default system, sys, sysaux will be created.

  • Data Pump: export/import tables in different schemas

    Hi all

    I use Oracle 11.2 and I would like to use the data pump to export / import data into the tables between the different schemas. The table already exists in the source and target schemas. Here are the commands that I use:


    Script working export:

    expdp scott/tiger@db12 schema = source include = TABLE:------"IN (\'TB_TEST1 ', \'TB_ABC')\ 'directory = datapump_dir dumpfile = test.dump logfile = test_exp.log

    Script to import all the desks:

    Impdp scott/tiger@db12 remap_schemas = rΘpertoire source: target = datapump_dir dumpfile = test.dump logfile = test_imp.log content = data_only table_exists_action = truncate

    Script error for some tables to import:

    Impdp scott/tiger@db12 remap_schemas = source: target include = TABLE:------"IN (\'TB_TEST1')\' directory = datapump_dir dumpfile = test.dump logfile = test_imp.log content = data_only

    Export is good, I got the folling error when I try to import the table TB_TEST1 only: "ORA-31655: no metadata data_or objectsz selected for employment. The user scott is a DBA role, and it works fine when I try to import all tables being export without the include clause.

    It is possble to import some tables but not all tables in the export file?

    Thanks for the help!

    942572 wrote:

    It's good to import all tables exported through Scott, who do NOT have the clause 'include '.

    I have the error only when I try to import some tables with clause "include".

    Can I import only some table since the export dump file? Thank you!

    you use INCLUDE incorrectly!

    do below yourself

    Impdp help = yes

    INCLUDE

    Include the specific object types.

    For example, INCLUDE TABLE_DATA =.

  • ODI LKM Oracle for Oracle Data Pump question

    Hi all

    I have a weird problem, ODI.

    I associate myself with per_all_people_f, fnd_user to load the w_user_ds using Oracle Data Integrator. The used LKM is LKM Oracle for Oracle Data Pump.

    Fine when I run the interface. I am getting below error

    ODI-1227: task failed USER_DATA_SET (load) on the source of connection ORACLE EBS.

    Caused by: java.sql.SQLSyntaxErrorException: ORA-00923: KEYWORD not found where expected

    at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:462)

    at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:405)

    at oracle.jdbc.driver.T4C8Oall.processError(T4C8Oall.java:931)

    at oracle.jdbc.driver.T4CTTIfun.receive(T4CTTIfun.java:481)

    at oracle.jdbc.driver.T4CTTIfun.doRPC(T4CTTIfun.java:205)

    at oracle.jdbc.driver.T4C8Oall.doOALL(T4C8Oall.java:548)

    at oracle.jdbc.driver.T4CPreparedStatement.doOall8(T4CPreparedStatement.java:217)

    at oracle.jdbc.driver.T4CPreparedStatement.executeForRows(T4CPreparedStatement.java:1115)

    at oracle.jdbc.driver.OracleStatement.doExecuteWithTimeout(OracleStatement.java:1488)

    at oracle.jdbc.driver.OraclePreparedStatement.executeInternal(OraclePreparedStatement.java:3769)

    at oracle.jdbc.driver.OraclePreparedStatement.execute(OraclePreparedStatement.java:3954)

    at oracle.jdbc.driver.OraclePreparedStatementWrapper.execute(OraclePreparedStatementWrapper.java:1539)

    at oracle.odi.runtime.agent.execution.sql.SQLCommand.execute(SQLCommand.java:163)

    at oracle.odi.runtime.agent.execution.sql.SQLExecutor.execute(SQLExecutor.java:102)

    at oracle.odi.runtime.agent.execution.sql.SQLExecutor.execute(SQLExecutor.java:1)

    at oracle.odi.runtime.agent.execution.TaskExecutionHandler.handleTask(TaskExecutionHandler.java:50)

    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.processTask(SnpSessTaskSql.java:2913)

    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java:2625)

    at com.sunopsis.dwg.dbobj.SnpSessStep.treatAttachedTasks(SnpSessStep.java:577)

    at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java:468)

    at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java:2128)

    to oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$ 2.doAction(StartSessRequestProcessor.java:366)

    at oracle.odi.core.persistence.dwgobject.DwgObjectTemplate.execute(DwgObjectTemplate.java:216)

    at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.doProcessStartSessTask(StartSessRequestProcessor.java:300)

    to oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.access$ 0 (StartSessRequestProcessor.java:292)

    to oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$ StartSessTask.doExecute (StartSessRequestProcessor.java:855)

    at oracle.odi.runtime.agent.processor.task.AgentTask.execute(AgentTask.java:126)

    to oracle.odi.runtime.agent.support.DefaultAgentTaskExecutor$ 2.run(DefaultAgentTaskExecutor.java:82)

    at java.lang.Thread.run(Thread.java:662)

    The generated code is

    create the table 780021 X

    (

    C1_FIRST_NAME,

    C2_MID_NAME,

    C3_LAST_NAME,

    C4_FULL_NAME,

    C5_NAME_SUFFIX,

    C6_SEX_MF_CODE,

    C7_SEX_MF_NAME,

    C8_COUNTRY_NAME,

    C9_LOGIN,

    C10_CREATED_BY_ID,

    C11_CHANGED_BY_ID,

    C12_CREATED_ON_DT,

    C13_CHANGED_ON_DT,

    C14_AUX1_CHANGED_ON_DT,

    C15_SRC_EFF_TO_DT,

    C16_INTEGRATION_ID,

    C17_EFFECTIVE_START_DATE

    )

    (EXTERNAL) ORGANIZATION

    TYPE oracle_datapump

    Dat_dir default DIRECTORY

    LOCATION ("X780021.exp")

    )

    PARALLEL

    in SELECT

    ALL_PEOPLE_F.FIRST_NAME,

    ALL_PEOPLE_F.MIDDLE_NAMES,

    ALL_PEOPLE_F.LAST_NAME,

    ALL_PEOPLE_F.FULL_NAME,

    ALL_PEOPLE_F.SUFFIX,

    ALL_PEOPLE_F.SEX,

    ALL_PEOPLE_F.SEX,

    ALL_PEOPLE_F.NATIONALITY,

    USER. USER_NAME,

    ALL_PEOPLE_F.CREATED_BY,

    ALL_PEOPLE_F.LAST_UPDATED_BY,

    ALL_PEOPLE_F.CREATION_DATE,

    ALL_PEOPLE_F.LAST_UPDATE_DATE,

    ALL_PEOPLE_F.CREATION_DATE,

    ALL_PEOPLE_F.EFFECTIVE_END_DATE,

    USER. USER_ID,

    ALL_PEOPLE_F.EFFECTIVE_START_DATE

    from APPS. FND_USER USER, APPS. PER_ALL_PEOPLE_F ALL_PEOPLE_F

    where (1 = 1)

    And (ALL_PEOPLE_F.PERSON_ID = USER. EMPLOYEE_ID)

    I don't see what is the problem here.

    Someone can help me.

    Thank you and best regards,

    Krishna Prasad

    I found the problem, its with the way ODI generated alias for the FND_USER table, by default it produces USER as an alias, which is a keyword from oracle. We just need to rename it to something else, and it worked.

  • Error data pump import

    Hi all

    I get errors below when you try to use the data pump import a table into the dump of the file (of a pre-production database) in a different database (prod_db) even version. I used the expdp to generate the dump file.

    Gettings errors-
    ORA-39083
    ORA-00959
    ORA-39112

    Any suggestions or advice would be appreciated.

    Thank you


    Import: Release 10.2.0.1.0 - 64 bit Production

    Copyright (c) 2003, 2005, Oracle. All rights reserved.
    ;;;
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - 64 bit Production
    With partitioning, OLAP and Data Mining options
    Main table 'SYSTEM '. "' SYS_IMPORT_TABLE_01 ' properly load/unloaded
    Start "SYSTEM". "" SYS_IMPORT_TABLE_01 ": System / * DIRECTORY = DATA_PUMP_DIR DUMPFILE = EXP_Mxxxx_1203.DMP LOGFILE = Mxxxx_CLAIMS. LOG TABLES = CLAIMS
    Object type SCHEMA_EXPORT/TABLE/TABLE processing
    ORA-39083: object of type THAT TABLE could not create with the error:
    ORA-00959: tablespace "OXFORD_DATA_01" does not exist
    Because sql is:
    CREATE TABLE 'xxx '. "" CLAIMS "("CLAIM_NUM"VARCHAR2 (25) NOT NULL ACTIVATE, ACTIVATE THE"LINE_NUM"NUMBER (3.0) NOT NULL, 'PAID_FLAG' CHAR (1), CHAR (1) 'ADJ_FLAG', 'CLAIM_TYPE' VARCHAR2 (20), 'MEM_ID' VARCHAR2 (20), VARCHAR2 (20) 'MEM_SUF', DATE 'MEM_BEG_DATE', 'MEM_REL' CHAR (2), 'MEM_NAME' VARCHAR2 (40), DATE 'MEM_DOB', 'MEM_SEX' CHAR (1),"REC_DATE"DATE, DATE OF THE"PAID_DATE") 'FROM_DATE' DATE "
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/INDEX
    ORA-39112: type of dependent object INDEX: 'xxx '. «CLAIMS_IDX1 ' ignored, base object type TABLE: 'Xxx'.» "" Creation of CLAIMS ' failed
    ORA-39112: type of dependent object INDEX: 'xxx '. «CLAIMS_PK ' ignored, base object type TABLE: 'Xxx'.» "" Creation of CLAIMS ' failed
    Object type SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS of treatment
    ORA-39112: dependent object type skipped INDEX_STATISTICS, object-based INDEX type: 'xxx '. "' CLAIMS_IDX1 ' failed to create
    ORA-39112: dependent object type skipped INDEX_STATISTICS, object-based INDEX type: 'xxx '. "" Creation of CLAIMS_PK ' failed
    Object type SCHEMA_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS treatment
    ORA-39112: dependent object type skipped TABLE_STATISTICS, object-based ARRAY type: "xxx". "" Creation of CLAIMS ' failed
    Work 'SYSTEM '. "" SYS_IMPORT_TABLE_01 "completed with error (s) from 6 to 13:49:33


    Impdp name username/password DIRECTORY = DATA_PUMP_DIR DUMPFILE = EXP_Mxxxx_1203.DMP LOGFILE is Mxxxx_CLAIMS. LOG TABLES = CLAIMS

    Before you create the tablespace or use clause REMAP_TABLESPACE for the import.

  • Export data pump job scheduler does not load

    I have a pump a 10.2.0.2.0 data dump file database, and then import on 11.2.0.3.0 fails with this error:
    ORA-39083: Object type PROCOBJ failed to create with error:
    ORA-06502: PL/SQL: numeric or value error: character to number conversion error
    Failing sql is:
    BEGIN 
    dbms_scheduler.create_job('"MY_JOB_NAME"',
    job_type=>'STORED_PROCEDURE', job_action=>
    'MY_SCHEMA.MY_PROCEDURE'
    , number_of_arguments=>0,
    start_date=>'31-JUL-12 02.05.13.782000 PM EUROPE/BERLIN', repeat_interval=> 
    'FREQ=WEEKLY;BYDAY=FRI;BYHOUR=7;BYMINUTE=0;BYSECOND=0'
    , end_date=>NULL,
    job_class=>'"DEFAULT_JOB_CL
    {code}
    
    I extracted the SQL Code from the dump file and it looks like this:
    {code}
    BEGIN 
    dbms_scheduler.create_job('"MY_JOB"',
    job_type=>'STORED_PROCEDURE', job_action=>
    'MY_SCHEMA.MY_PROCEDURE'
    , number_of_arguments=>0,
    start_date=>'31-JUL-12 02.05.13.782000 PM EUROPE/BERLIN', repeat_interval=> 
    'FREQ=WEEKLY;BYDAY=FRI;BYHOUR=7;BYMINUTE=0;BYSECOND=0'
    , end_date=>NULL,
    job_class=>'"DEFAULT_JOB_CLASS"', enabled=>FALSE, auto_drop=>FALSE,comments=>
    'bla bla comment'
    );
    dbms_scheduler.set_attribute('"MY_JOB"','logging_level','HIGH');
    dbms_scheduler.enable('"MY_JOB"');
    COMMIT; 
    END; 
    / 
    {code}
    
    
    
    After the job is defined the second statement fails:
    
    {code}
    SQL> exec dbms_scheduler.set_attribute('"MY_JOB"','logging_level','HIGH');
    BEGIN dbms_scheduler.set_attribute('"MY_JOB"','logging_level','HIGH'); END;
    
    *
    ERROR at line 1:
    ORA-06502: PL/SQL: numeric or value error: character to number conversion error
    ORA-06512: at "SYS.DBMS_SCHEDULER", line 2851
    ORA-06512: at line 1
    {code}
    
    
    From the source I see:
    
    {code}
    SQL> select logging_level from dba_scheduler_jobs where job_name = 'MY_JOB';
    
    LOGG
    ----
    FULL
    {code}
    
    In the docs I see these valid LOGGING_LEVELs:
    http://docs.oracle.com/cd/E14072_01/server.112/e10595/scheduse008.htm#CHDFDFAB
    
    DBMS_SCHEDULER.LOGGING_OFF
    DBMS_SCHEDULER.LOGGING_FAILED_RUNS
    DBMS_SCHEDULER.LOGGING_RUNS
    DBMS_SCHEDULER.LOGGING_FULL
    
    So please help me, I can't find something useful on MOS, what is Data Pump exporting there, and can not import it again
    Maybe I have overseen a known bug?                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            

    Finally I found the bug myself:

    MOS: Impdp of objects of treatment fails with ORA-39083 and ORA-06502 [ID 780174.1]

    https://support.Oracle.com/epmos/faces/DocContentDisplay?ID=780174.1

  • Export schema through Oracle data pump with question database Vault enabled

    Hello

    I installed and configured Vault database on an Oracle 11 g - r2 - 11.2.0.3 to protect a specific schema (SCHEMA_NAME) via a Kingdom. I followed the following doc:
    http://www.Oracle.com/technetwork/database/security/TWP-databasevault-DBA-BestPractices-199882.PDF
    to ensure that the system and the network user has sufficient rights to complete a schedule pump data Oracle export operation.

    I.e. I gave sys and system the following:
    execute dvsys.dbms_macadm.authorize_scheduler_user ('sys', 'SCHEMA_NAME');
    execute dvsys.dbms_macadm.authorize_scheduler_user ('system', 'SCHEMA_NAME');

    execute dvsys.dbms_macadm.authorize_datapump_user ('sys', 'SCHEMA_NAME');
    execute dvsys.dbms_macadm.authorize_datapump_user ('system', 'SCHEMA_NAME');

    I also create a second domain on the same schema (SCHEMA_NAME) to allow sys and system to manage indexes for the actual tables protected, in order to allow a system to manage indexes for the tables of protected area and sys. This separate realm was created for all types of indexes: Index, Partition of Index and Indextype, sys and system were allowed as the OWNER of this Kingdom.

    However, when I try and complete an operation of Oracle Data Pump export on the schematic, I get two errors directly after the following line appears in the export log:

    Processing object type SCHEMA_EXPORT/TABLE/INDEX/DOMAIN_INDEX/INDEX:
    ORA-39127: unexpected error of the call to export_string: = SYS. DBMS_TRANSFORM_EXIMP. INSTANCE_INFO_EXP('AQ$_MGMT_NOTIFY_QTABLE_S','SYSMAN',1,1,'11.02.00.00.00',newBlock)
    ORA-01031: insufficient privileges
    ORA-06512: at "SYS." DBMS_TRANSFORM_EXIMP', line 197
    ORA-06512: at line 1
    ORA-06512: at "SYS." Dbms_metadata", line 9081
    ORA-39127: unexpected error of the call to export_string: = SYS. DBMS_TRANSFORM_EXIMP. INSTANCE_INFO_EXP('AQ$_MGMT_LOADER_QTABLE_S','SYSMAN',1,1,'11.02.00.00.00',newBlock)
    ORA-01031: insufficient privileges
    ORA-06512: at "SYS." DBMS_TRANSFORM_EXIMP', line 197
    ORA-06512: at line 1
    ORA-06512: at "SYS." Dbms_metadata", line 9081

    The export is completed, but this error.

    Any help, pointers, suggestions, etc. in fact no matter what will be very welcome at this stage.

    Thank you

    I moved the forum thread on the "database - security. If the document does not help, pl open an SR with Support

    HTH
    Srini

  • Question about data pump and directory/sub-directories

    I have a process that makes these calls to Data Pump:
    dbms_datapump.add_file (dph, file_name |) '.log', ' DATA_PUMP_DIR/test', filetype = > DBMS_DATAPUMP. KU$ _FILE_TYPE_LOG_FILE);
    dbms_datapump.add_file (dph, file_name |) ".dmp", ' DATA_PUMP_DIR/test', filetype = > DBMS_DATAPUMP. KU$ _FILE_TYPE_DUMP_FILE);

    He failed with the subdirectory "/ test". If I delete it, as in these calls it works.
    dbms_datapump.add_file (dph, file_name |) '.log', "DATA_PUMP_DIR" filetype = > DBMS_DATAPUMP. KU$ _FILE_TYPE_LOG_FILE);
    dbms_datapump.add_file (dph, file_name |) ".dmp", "DATA_PUMP_DIR" filetype = > DBMS_DATAPUMP. KU$ _FILE_TYPE_DUMP_FILE);

    I even did a test 'chmod 777 '.

    No idea what I am doing wrong?

    I'm on 11.0.1.6

    It's the pile of error I get:
    ORA-39002: invalid operation
    ORA-06512: at "SYS." DBMS_SYS_ERROR', line 79
    ORA-06512: at "SYS." DBMS_DATAPUMP', line 3043
    ORA-06512: at "SYS." DBMS_DATAPUMP', line 3292
    ORA-06512: at "ORA_ADMIN. DATAPUMP_UTIL", line 46
    ORA-06512: at line 1

    Chris wrote:
    I have a process that makes these calls to Data Pump:
    dbms_datapump.add_file (dph, file_name |) '.log', ' DATA_PUMP_DIR/test', filetype => DBMS_DATAPUMP. KU$ _FILE_TYPE_LOG_FILE);
    dbms_datapump.add_file (dph, file_name |) ".dmp", ' DATA_PUMP_DIR/test', filetype => DBMS_DATAPUMP. KU$ _FILE_TYPE_DUMP_FILE);

    He failed with the subdirectory "/ test". If I delete it, as in these calls it works.
    dbms_datapump.add_file (dph, file_name |) '.log', "DATA_PUMP_DIR", filetype => DBMS_DATAPUMP. KU$ _FILE_TYPE_LOG_FILE);
    dbms_datapump.add_file (dph, file_name |) ".dmp", "DATA_PUMP_DIR", filetype => DBMS_DATAPUMP. KU$ _FILE_TYPE_DUMP_FILE);

    I even did a test 'chmod 777 '.

    No idea what I am doing wrong?

    invalid syntax

    DATA_PUMP_DIR

    above is DIRECTORY Oracle
    If you want a different location, then CREATE DIRECTORY DATA_PUMP_DIR_TEST

  • Data pump import

    I cannot using the data pump import tool. I work with Oracle 10.2.4.0 on Windows Server 2008. I am trying to import a large amount of data (several patterns, thousands of paintings, hundreds of GB of data) of the same version of Oracle on Linux. Since there is so much data and will be if long to import, I try to make sure that I can make it work for a table or a diagram before I try. So I will try to import the TEST_USER. Table 1 table using the following command:

    Dumpfile = logfile = big_file_import.log big_file.dmp Impdp.exe tables = test_user.table1 remap_datafile=\'/oradata1/data1.dbf\':\'D:\oracle\product\10.2.0\oradata\orcl\data1.dbf\'

    I provide sys as sysdba to connect when he asks (not sure how you get this info in the original order). However, I get the following error:

    the "TEST_USER" user does not exist

    My understanding is that the data pump utility would create all the necessary patterns for me. Is it not the case? The target database is a new facility for the source schemas do not exist here.

    Even if I create the schema of test_user by hand, then I get an error message indicating that the tablespace does not exist:

    ORA-00959: tablespace "TS_1" does not exist

    Then even to do it, I don't want to have to do first, does not work. Then he gives me something about how the user has no privileges on tablespace.

    The data pump utility is not supposed to do that sort of thing automatic? What I really need to create all schemas and storage by hand? That will take a long time. I'm doing something wrong here?

    Thank you
    Dave

    tables = test_user.table1

    The "TABLES" mode does NOT create database accounts.

    The FULL mode creates space storage and database accounts before importing the data.

    PATTERNS mode creates the database accounts before importing data - but expect Tablespaces to exist before so that tables and indexes can be created in the correct storage spaces.

    Hemant K Collette
    http://hemantoracledba.blogspot.com

  • "Resume" a Data Pump import Execution failure

    First of all, here is a "failure" Data Pump Import I want to "resume":
    ] $ impdp directory dumpfile=mySchema%U.dmp "" / as sysdba "" = DATA_PUMP_DIR logfile = Import.log patterns parallel = MYSCHEMA = 2

    Import: Release 10.2.0.3.0 - 64 bit Production Tuesday, February 16, 2010 14:35:15

    Copyright (c) 2003, 2005, Oracle. All rights reserved.

    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.3.0 - 64 bit Production
    With the options of Real Application Clusters, partitioning, OLAP and Data Mining
    Table main 'SYS '. "' SYS_IMPORT_SCHEMA_01 ' properly load/unloaded
    Departure 'SYS '. "' SYS_IMPORT_SCHEMA_01 ': ' / * AS SYSDBA"dumpfile=mySchema%U.dmp directory = DATA_PUMP_DIR logfile = Import.log patterns parallel = MYSCHEMA = 2
    Processing object type SCHEMA_EXPORT/PRE_SCHEMA/PROCACT_SCHEMA
    Object type SCHEMA_EXPORT/SYNONYM/SYNONYM of treatment
    Object type SCHEMA_EXPORT/TYPE/TYPE_SPEC of treatment
    Object type SCHEMA_EXPORT/SEQUENCE/SEQUENCE of treatment
    Processing object type SCHEMA_EXPORT/SEQUENCE/EXCHANGE/OWNER_GRANT/OBJECT_GRANT
    Object type SCHEMA_EXPORT/TABLE/TABLE processing
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    . . imported "MYSCHE"...
    ...
    ... 0 KB 0 rows
    Processing object type SCHEMA_EXPORT/TABLE/SCHOLARSHIP/OWNER_GRANT/OBJECT_GRANT
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/INDEX
    ORA-39126: worker unexpected fatal worker error of $ MAIN. LOAD_METADATA [INDEX: "MONSCHEMA".] ["" OBJECT_RELATION_I2 "]
    SELECT process_order, flags, xml_clob, NVL (dump_fileid,: 1), NVL (dump_position,: 2), dump_length, dump_allocation, constituent, object_row, object_schema, object_long_name, processing_status, processing_state, base_object_type, base_object_schema, base_object_name, goods, size_estimate, in_progress 'SYS '. "" SYS_IMPORT_SCHEMA_01 "process_order where: 3: 4 AND processing_state <>: 5 AND double = 0 ORDER BY process_order
    ORA-06502: PL/SQL: digital error or value
    ORA-06512: at "SYS." "MAIN$ WORKER", line 12280
    ORA-12801: error reported in the parallel query P001, instance pace2.capitolindemnity.com:bondacc2 Server (2)
    ORA-30032: the suspended order (trade-in) has expired
    ORA-01652: unable to extend segment temp of 128 in tablespace OBJECT_INDX
    ORA-06512: at "SYS." DBMS_SYS_ERROR', line 105
    ORA-06512: at "SYS." "MAIN$ WORKER", line 6272
    -PL/SQL call stack-
    the line object
    serial number of handle
    package body 14916 0x1f9ac8d50 SYS. MAIN$ WORKER
    body 0x1f9ac8d50 6293 package SYS. MAIN$ WORKER
    body 0x1f9ac8d50 3511 package SYS. MAIN$ WORKER
    body 0x1f9ac8d50 6882 package SYS. MAIN$ WORKER
    body 0x1f9ac8d50 1259 package SYS. MAIN$ WORKER
    0x1f8431598 anonymous block 2
    Job 'SYS '. "' SYS_IMPORT_SCHEMA_01 ' stopped because of the fatal 23:05:45
    As you can see this work is not without treatment statistics, constraints, etc. from PL/SQL. I want to do is make another impdp command but ignore objects that were imported with success as shown above. Is it possible to do (using the EXCLUDE parameter maybe?) with impdp? If so, what would it be?

    Thank you

    John

    Why you just restart this work. It ignores everything that has already been imported.

    Impdp "" / as sysdba "" attach = "SYS." "" SYS_IMPORT_SCHEMA_01 ".

    Import > keep

    Dean

  • default data pump

    Hi all

    I use Oracle 10 G. Every day my data pump fails with the following error code.
    Please guide me.

    ORA-31693: Data Table object 'username '. "' Test_ERROR_LOG ' failed to load/unload and is skip
    Developing Countries due to the error:
    ORA-19502: write error on file ' / export/test_dataPump.dmp ', blockno 736810
    (blocksize = 4096)
    ORA-27063: number of bytes read or written is incorrect.

    Thanks and greetings
    Muthu

    Hello

    There can be two reasons.

    1. number of open files: check ulimit for your oracle username. (Operating system are you on?)

    2 possible issue of I/O: look for the file system, get help for UNIX admin and check the possible problem of I/O on the disk/machine.

    Piesset.

  • ORA-00904: "DATA". "" AM ": invalid identifier

    Hello
    I am facing upwards a situation I can't find why is the error ora-00904 appears.
    I wrote the request as follows:

    (A)

    with data as
      (
      select uc.* 
        from user_computer uc
         where computer_name like '%'||:perioxi||'%'
      )
      select uc.* 
        from user_computer uc
        where exists (select 1
                       from data
                       where data.am = uc.am)
         and exists (select 1
                          from v_misthotoi_orga
                          where uc.am = v_misthotoi_orga.matr) 
    

    Where the utilisateur_ordinateur is a table, and the v_misthotoi_orga is a db view (a view that is based on another view that is based on a table that is accessed through a dblink).
    A query runs without error.

    (B)

    with data as
      (
      select uc.* 
        from user_computer uc
         where computer_name like '%'||:perioxi||'%'
      )
      select uc.* 
        from user_computer uc
        where exists (select 1
                       from data
                       where data.am = uc.am)                       <.... THE DATACOLUMN DATA.AM IS USED, WITHOUT ERROR                
         and exists (select 1
                          from v_misthotoi_orga
                          where data.am = v_misthotoi_orga.matr)    <.....THIS IS WHERE THE ERROR IS LOCATED (DATA.AM)
    

    The above query (application B) stops with error: ORA-00904: "DATA". "" AM ": invalid identifier.

    You can identify the cause of the error, since the datacolumn 'am' in 'data' datasource is used twice in query B.
    In the first instance without any error in the second with the error.

    Note: Oracle db v.2
    Thank you
    SIM

    Hey Sim,

    Change

    select 1
    from v_misthotoi_orga
    where data.am = v_misthotoi_orga.matr
    

    TO

    select 1
    from v_misthotoi_orga,
           data  
    
    where data.am = v_misthotoi_orga.matr
    

    In this query, Oracle knows no data unless you specify in the From clause that he should use.

    HTH,

    Stefan

  • ORA-31693: Data Table object 'AWSTEMPUSER '. "' TEMPMANUALMAPRPT_273 ' failed to load/unload and being ignored because of the error:

    Dear all,

    OS - Windows server 2012 R2

    version - 11.2.0.1.0

    Server: production server

    ORA-31693: Data Table object 'AWSTEMPUSER '. "' TEMPMANUALMAPRPT_273 ' failed to load/unload and being ignored because of the error:

    ORA-02354: Error exporting/importing data

    ORA-00942: table or view does not exist

    When taken expdp and faced error mentioned above. but expdp completed successfully with waring as below.

    Work "AWSCOMMONMASTER". "" FULLEXPJOB26SEP15_053001 "finished with 6 errors at 09:30:54

    (1) what is the error

    (2) is there any problem in the dump because file as above of the error. If Yes, then I'll resume expdp.

    Please suggest me. Thanks in advance

    Hello

    I suspect that what has happened, is that demand has dropped a temporary table to during the time that you run the export - consider this series of events

    (1) temp table created by application

    (2) start expdp work - including this table

    (3) the extracted table metadata

    (4) the application deletes the table

    (5) expdp is trying to retrieve data from the table - and gets the above error.

    Just to confirm with the enforcement team that the table is just a temporary thing - it certainly seems it name.

    See you soon,.

    Rich

Maybe you are looking for