Oracle Data Pump

oracle11g oracle introduced the new feature called oracle data dump. What is the difference in comparing utility imp/exp? can someone explain how to use this new feature

for example, by using the user scott to export some tables in my diagram.

Data pump is very fast, parallel importation is made in 11g. exp/imp sometimes fails due to problems of space on the server. but by using method of parallel import is performed. parallel importation is nothing but imported when 1 row have exported the corresponding row in the table, thus saving disk space. With Data Pump Import, a single stream of data load is about 15 to 45 times faster than the initial import.

Tags: Database

Similar Questions

  • ODI LKM Oracle for Oracle Data Pump question

    Hi all

    I have a weird problem, ODI.

    I associate myself with per_all_people_f, fnd_user to load the w_user_ds using Oracle Data Integrator. The used LKM is LKM Oracle for Oracle Data Pump.

    Fine when I run the interface. I am getting below error

    ODI-1227: task failed USER_DATA_SET (load) on the source of connection ORACLE EBS.

    Caused by: java.sql.SQLSyntaxErrorException: ORA-00923: KEYWORD not found where expected

    at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:462)

    at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:405)

    at oracle.jdbc.driver.T4C8Oall.processError(T4C8Oall.java:931)

    at oracle.jdbc.driver.T4CTTIfun.receive(T4CTTIfun.java:481)

    at oracle.jdbc.driver.T4CTTIfun.doRPC(T4CTTIfun.java:205)

    at oracle.jdbc.driver.T4C8Oall.doOALL(T4C8Oall.java:548)

    at oracle.jdbc.driver.T4CPreparedStatement.doOall8(T4CPreparedStatement.java:217)

    at oracle.jdbc.driver.T4CPreparedStatement.executeForRows(T4CPreparedStatement.java:1115)

    at oracle.jdbc.driver.OracleStatement.doExecuteWithTimeout(OracleStatement.java:1488)

    at oracle.jdbc.driver.OraclePreparedStatement.executeInternal(OraclePreparedStatement.java:3769)

    at oracle.jdbc.driver.OraclePreparedStatement.execute(OraclePreparedStatement.java:3954)

    at oracle.jdbc.driver.OraclePreparedStatementWrapper.execute(OraclePreparedStatementWrapper.java:1539)

    at oracle.odi.runtime.agent.execution.sql.SQLCommand.execute(SQLCommand.java:163)

    at oracle.odi.runtime.agent.execution.sql.SQLExecutor.execute(SQLExecutor.java:102)

    at oracle.odi.runtime.agent.execution.sql.SQLExecutor.execute(SQLExecutor.java:1)

    at oracle.odi.runtime.agent.execution.TaskExecutionHandler.handleTask(TaskExecutionHandler.java:50)

    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.processTask(SnpSessTaskSql.java:2913)

    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java:2625)

    at com.sunopsis.dwg.dbobj.SnpSessStep.treatAttachedTasks(SnpSessStep.java:577)

    at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java:468)

    at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java:2128)

    to oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$ 2.doAction(StartSessRequestProcessor.java:366)

    at oracle.odi.core.persistence.dwgobject.DwgObjectTemplate.execute(DwgObjectTemplate.java:216)

    at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.doProcessStartSessTask(StartSessRequestProcessor.java:300)

    to oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.access$ 0 (StartSessRequestProcessor.java:292)

    to oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$ StartSessTask.doExecute (StartSessRequestProcessor.java:855)

    at oracle.odi.runtime.agent.processor.task.AgentTask.execute(AgentTask.java:126)

    to oracle.odi.runtime.agent.support.DefaultAgentTaskExecutor$ 2.run(DefaultAgentTaskExecutor.java:82)

    at java.lang.Thread.run(Thread.java:662)

    The generated code is

    create the table 780021 X

    (

    C1_FIRST_NAME,

    C2_MID_NAME,

    C3_LAST_NAME,

    C4_FULL_NAME,

    C5_NAME_SUFFIX,

    C6_SEX_MF_CODE,

    C7_SEX_MF_NAME,

    C8_COUNTRY_NAME,

    C9_LOGIN,

    C10_CREATED_BY_ID,

    C11_CHANGED_BY_ID,

    C12_CREATED_ON_DT,

    C13_CHANGED_ON_DT,

    C14_AUX1_CHANGED_ON_DT,

    C15_SRC_EFF_TO_DT,

    C16_INTEGRATION_ID,

    C17_EFFECTIVE_START_DATE

    )

    (EXTERNAL) ORGANIZATION

    TYPE oracle_datapump

    Dat_dir default DIRECTORY

    LOCATION ("X780021.exp")

    )

    PARALLEL

    in SELECT

    ALL_PEOPLE_F.FIRST_NAME,

    ALL_PEOPLE_F.MIDDLE_NAMES,

    ALL_PEOPLE_F.LAST_NAME,

    ALL_PEOPLE_F.FULL_NAME,

    ALL_PEOPLE_F.SUFFIX,

    ALL_PEOPLE_F.SEX,

    ALL_PEOPLE_F.SEX,

    ALL_PEOPLE_F.NATIONALITY,

    USER. USER_NAME,

    ALL_PEOPLE_F.CREATED_BY,

    ALL_PEOPLE_F.LAST_UPDATED_BY,

    ALL_PEOPLE_F.CREATION_DATE,

    ALL_PEOPLE_F.LAST_UPDATE_DATE,

    ALL_PEOPLE_F.CREATION_DATE,

    ALL_PEOPLE_F.EFFECTIVE_END_DATE,

    USER. USER_ID,

    ALL_PEOPLE_F.EFFECTIVE_START_DATE

    from APPS. FND_USER USER, APPS. PER_ALL_PEOPLE_F ALL_PEOPLE_F

    where (1 = 1)

    And (ALL_PEOPLE_F.PERSON_ID = USER. EMPLOYEE_ID)

    I don't see what is the problem here.

    Someone can help me.

    Thank you and best regards,

    Krishna Prasad

    I found the problem, its with the way ODI generated alias for the FND_USER table, by default it produces USER as an alias, which is a keyword from oracle. We just need to rename it to something else, and it worked.

  • Export schema through Oracle data pump with question database Vault enabled

    Hello

    I installed and configured Vault database on an Oracle 11 g - r2 - 11.2.0.3 to protect a specific schema (SCHEMA_NAME) via a Kingdom. I followed the following doc:
    http://www.Oracle.com/technetwork/database/security/TWP-databasevault-DBA-BestPractices-199882.PDF
    to ensure that the system and the network user has sufficient rights to complete a schedule pump data Oracle export operation.

    I.e. I gave sys and system the following:
    execute dvsys.dbms_macadm.authorize_scheduler_user ('sys', 'SCHEMA_NAME');
    execute dvsys.dbms_macadm.authorize_scheduler_user ('system', 'SCHEMA_NAME');

    execute dvsys.dbms_macadm.authorize_datapump_user ('sys', 'SCHEMA_NAME');
    execute dvsys.dbms_macadm.authorize_datapump_user ('system', 'SCHEMA_NAME');

    I also create a second domain on the same schema (SCHEMA_NAME) to allow sys and system to manage indexes for the actual tables protected, in order to allow a system to manage indexes for the tables of protected area and sys. This separate realm was created for all types of indexes: Index, Partition of Index and Indextype, sys and system were allowed as the OWNER of this Kingdom.

    However, when I try and complete an operation of Oracle Data Pump export on the schematic, I get two errors directly after the following line appears in the export log:

    Processing object type SCHEMA_EXPORT/TABLE/INDEX/DOMAIN_INDEX/INDEX:
    ORA-39127: unexpected error of the call to export_string: = SYS. DBMS_TRANSFORM_EXIMP. INSTANCE_INFO_EXP('AQ$_MGMT_NOTIFY_QTABLE_S','SYSMAN',1,1,'11.02.00.00.00',newBlock)
    ORA-01031: insufficient privileges
    ORA-06512: at "SYS." DBMS_TRANSFORM_EXIMP', line 197
    ORA-06512: at line 1
    ORA-06512: at "SYS." Dbms_metadata", line 9081
    ORA-39127: unexpected error of the call to export_string: = SYS. DBMS_TRANSFORM_EXIMP. INSTANCE_INFO_EXP('AQ$_MGMT_LOADER_QTABLE_S','SYSMAN',1,1,'11.02.00.00.00',newBlock)
    ORA-01031: insufficient privileges
    ORA-06512: at "SYS." DBMS_TRANSFORM_EXIMP', line 197
    ORA-06512: at line 1
    ORA-06512: at "SYS." Dbms_metadata", line 9081

    The export is completed, but this error.

    Any help, pointers, suggestions, etc. in fact no matter what will be very welcome at this stage.

    Thank you

    I moved the forum thread on the "database - security. If the document does not help, pl open an SR with Support

    HTH
    Srini

  • import schema data using oracle data pump

    Hi all

    I want to import a schema on a single server using IMPDP. But on this server, the user does not exists.is it is possible to import all the objects that break including the creation of the schema. In exp/imp, it may be possible. But in data pump I can't do the same. Its giving error that the user does not exist. For your reference, I'll give you the code below the import.

    Impdp system directory = DIR_DUMP logfile = expdp_tsadmin_tcgadmin_11182013_2_2.log

    dumpfile = expdp_tsadmin_tcgadmin_11182013_2.dmp PATTERNS = TS_ADMIN version = 10.1.0.2.0

    Pls help and thanks in advance.

    Thank you

    Piku

    Hello

    All errors are due to the fact that the tablespace that ts_data does not exist, everything else is hit because of this.

    create tablespace ts_data datafile 'xxxxxx' size xM;

    then try again

    See you soon,.

    Harry

  • Differences between Data Pump and always by inheritance, import and export

    Hi all

    I work as a junior in my organization dba and I saw that my organization still uses legacy import and export utility instead of Oracle Data Pump.

    I want to convence my manager to change the existing deal with Oracle Data Pump, I have a meeting with them to keep my points and convence them for Oracle utility pump data.

    I have a week very convencing power but I won't miss to discover myself, I can't work myself and really a need with differences of strength against import and export, it would be really appreciated if someone can put strong points against Oracle Data pump on legacy import and export.

    Thank you

    Cabbage

    Hello

    a other people have already said the main advantage of datapump is performance - it is not just a little more fast exp/imp it is massively faster (especially when combined with parallel).

    It is also more flexible (once much more) - it will even create users with exports level schema which imp can't do for you (and it is very annoying that he could never).

    It is reusable

    It has an api plsql

    It supports all types of objects and new features (exp is not - and that alone is probably reason to spend)

    There even a 'legacy' at 11.2 mode where most of your old exp parameter file will still work with him - just change exp expdp and impdp print.

    The main obstacle to the transition to datapump seems to be all "what do you mean I have to create a directory so that it works", well who and where is my dumpfile why can't it be on my local machine. These are minor things to go well beyond.

    I suggest do you some sort of demo with real data of one of your large databases - do a full exp and a full expdp with parallel and show them the runtimes for them to compare...

    See you soon,.

    Rich

  • is it possible to exclude some tables of data pump?

    Hello

    I want to exclude some tables while import/export to a single system using data pump. Is this all possible for this?

    If this isn't the case, please suggest me for all other ideas too.

    Iqbal

    Use the exclude parameter and check the link below.

    http://www.Oracle-base.com/articles/10G/Oracle-data-pump-10G.php

    Concerning
    Asif Kabir

  • Kill an Oracle 10 g data pump work

    I have a snap task data pump import that loop forever. Probably because by mistake I started it as SYS instead of a particular schema owner. She correctly creates objects in the schema of right, because of the setting of PATTERNS, but he won't finish.
    How can I kill it? I can see it in select * from dba_datapump_jobs; but not in dba_jobs.
    I can stop it without having to restart the database?

    Impdp comes with aid = clause of y.
    When you issue, you will see the pricing clause and clause kill_job.
    The documentation online at http://tahiti.oracle.com for your unknown version provides more information.

    You try to read the documentation before asking a volunteer to make abstraction of the documentation on your behalf is strongly recommended.

    --------
    Sybrand Bakker
    Senior Oracle DBA

  • Oracle 11 g 2 Standard Edition Data Pump

    Hello

    I understand that Data Pump is fully supported in Oracle 11 g 2, Standard Edition, and the one feature that is not included is "parallel." According to the http://download.oracle.com/docs/cd/B28359_01/license.111/b28287/editions.htm, I noticed that for itself, Flashback Database is "N".

    That being the case, anyone know if I can use expdp with the parameter 'FLASHBACK_TIME' or 'FLASHACBK_SCN' with Oracle 11 g 2 SE? We are a boutique EE here so I will dl/install SE and test it in-house, but if someone has tried already and you know the answer, it will save me a lot of time.

    Thanks in advance.

    Yes, you can use flashback_time and flashback_scn in itself. It is similar to using a query flashback that is supported and self-esteem (SE1).

    Here is an example of flashback_time, but 10 gr 2 (don't have 11 GR 2 run immediately)

    Connected to: Oracle Database 10g Release 10.2.0.4.0 - Production
    Starting "SYSTEM"."SYS_EXPORT_SCHEMA_01":  system/******** dumpfile=dptest_fb.dmp schemas=sample include=table:" = 'CLASS'" flashback_time='2010-03-02:11:30:00'
    Estimate in progress using BLOCKS method...
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 128 KB
    Processing object type SCHEMA_EXPORT/TABLE/TABLE
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/INDEX
    Processing object type SCHEMA_EXPORT/TABLE/CONSTRAINT/CONSTRAINT
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    Processing object type SCHEMA_EXPORT/TABLE/TRIGGER
    Processing object type SCHEMA_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS
    . . exported "SAMPLE"."CLASS"                          87.64 KB     696 rows
    Master table "SYSTEM"."SYS_EXPORT_SCHEMA_01" successfully loaded/unloaded
    
  • Data Pump Export/Import

    Hello Forum,

    I have a question regarding imports and exports of data pump, perhaps what I should already know.

    I need to empty a table that has about 200 million lines, that I need to get rid of about three quarters of data.

    My intention is to use the data pump to export the table and and indexes and constraints etc..

    The table has no relationship to any other table, it is composed of approximately 8 columns with constraints not null.

    My plan is

    1 truncate table

    2. disable or remove index

    3 leave the constraints in place?

    4. use data pump to import a lines to keep.

    My question

    will be my clues and imported too much constraints I want to import only a subset of my exported table?

    or

    If I dropped the table after truncation, I'll be able to import my table and indexes, even if I use the sub as part of my statement to import query functionality?

    My table using the query sub in data pump functionality must exist in the database before doing the import

    or handful of data pump import as usual IE will create table indices grants and statistics etc.?

    Thank you for your comments.

    Concerning

    Your approach is ineffective.

    What you need to do is to

    create the table in select foo * bar where the

    bar of truncate table;

    Insert / * + APPEND * / into select bar * foo.

    Rebuild the indexes on the table.

    Fact.

    This whole thing with expdp and impdp is only a waste of resources. My approach generates redo a minimum.

    ----------

    Sybrand Bakker

    Senior Oracle DBA

  • Data Pump: export/import tables in different schemas

    Hi all

    I use Oracle 11.2 and I would like to use the data pump to export / import data into the tables between the different schemas. The table already exists in the source and target schemas. Here are the commands that I use:


    Script working export:

    expdp scott/tiger@db12 schema = source include = TABLE:------"IN (\'TB_TEST1 ', \'TB_ABC')\ 'directory = datapump_dir dumpfile = test.dump logfile = test_exp.log

    Script to import all the desks:

    Impdp scott/tiger@db12 remap_schemas = rΘpertoire source: target = datapump_dir dumpfile = test.dump logfile = test_imp.log content = data_only table_exists_action = truncate

    Script error for some tables to import:

    Impdp scott/tiger@db12 remap_schemas = source: target include = TABLE:------"IN (\'TB_TEST1')\' directory = datapump_dir dumpfile = test.dump logfile = test_imp.log content = data_only

    Export is good, I got the folling error when I try to import the table TB_TEST1 only: "ORA-31655: no metadata data_or objectsz selected for employment. The user scott is a DBA role, and it works fine when I try to import all tables being export without the include clause.

    It is possble to import some tables but not all tables in the export file?

    Thanks for the help!

    942572 wrote:

    It's good to import all tables exported through Scott, who do NOT have the clause 'include '.

    I have the error only when I try to import some tables with clause "include".

    Can I import only some table since the export dump file? Thank you!

    you use INCLUDE incorrectly!

    do below yourself

    Impdp help = yes

    INCLUDE

    Include the specific object types.

    For example, INCLUDE TABLE_DATA =.

  • Error data pump import

    Hi all

    I get errors below when you try to use the data pump import a table into the dump of the file (of a pre-production database) in a different database (prod_db) even version. I used the expdp to generate the dump file.

    Gettings errors-
    ORA-39083
    ORA-00959
    ORA-39112

    Any suggestions or advice would be appreciated.

    Thank you


    Import: Release 10.2.0.1.0 - 64 bit Production

    Copyright (c) 2003, 2005, Oracle. All rights reserved.
    ;;;
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - 64 bit Production
    With partitioning, OLAP and Data Mining options
    Main table 'SYSTEM '. "' SYS_IMPORT_TABLE_01 ' properly load/unloaded
    Start "SYSTEM". "" SYS_IMPORT_TABLE_01 ": System / * DIRECTORY = DATA_PUMP_DIR DUMPFILE = EXP_Mxxxx_1203.DMP LOGFILE = Mxxxx_CLAIMS. LOG TABLES = CLAIMS
    Object type SCHEMA_EXPORT/TABLE/TABLE processing
    ORA-39083: object of type THAT TABLE could not create with the error:
    ORA-00959: tablespace "OXFORD_DATA_01" does not exist
    Because sql is:
    CREATE TABLE 'xxx '. "" CLAIMS "("CLAIM_NUM"VARCHAR2 (25) NOT NULL ACTIVATE, ACTIVATE THE"LINE_NUM"NUMBER (3.0) NOT NULL, 'PAID_FLAG' CHAR (1), CHAR (1) 'ADJ_FLAG', 'CLAIM_TYPE' VARCHAR2 (20), 'MEM_ID' VARCHAR2 (20), VARCHAR2 (20) 'MEM_SUF', DATE 'MEM_BEG_DATE', 'MEM_REL' CHAR (2), 'MEM_NAME' VARCHAR2 (40), DATE 'MEM_DOB', 'MEM_SEX' CHAR (1),"REC_DATE"DATE, DATE OF THE"PAID_DATE") 'FROM_DATE' DATE "
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/INDEX
    ORA-39112: type of dependent object INDEX: 'xxx '. «CLAIMS_IDX1 ' ignored, base object type TABLE: 'Xxx'.» "" Creation of CLAIMS ' failed
    ORA-39112: type of dependent object INDEX: 'xxx '. «CLAIMS_PK ' ignored, base object type TABLE: 'Xxx'.» "" Creation of CLAIMS ' failed
    Object type SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS of treatment
    ORA-39112: dependent object type skipped INDEX_STATISTICS, object-based INDEX type: 'xxx '. "' CLAIMS_IDX1 ' failed to create
    ORA-39112: dependent object type skipped INDEX_STATISTICS, object-based INDEX type: 'xxx '. "" Creation of CLAIMS_PK ' failed
    Object type SCHEMA_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS treatment
    ORA-39112: dependent object type skipped TABLE_STATISTICS, object-based ARRAY type: "xxx". "" Creation of CLAIMS ' failed
    Work 'SYSTEM '. "" SYS_IMPORT_TABLE_01 "completed with error (s) from 6 to 13:49:33


    Impdp name username/password DIRECTORY = DATA_PUMP_DIR DUMPFILE = EXP_Mxxxx_1203.DMP LOGFILE is Mxxxx_CLAIMS. LOG TABLES = CLAIMS

    Before you create the tablespace or use clause REMAP_TABLESPACE for the import.

  • Data Pump 11g Network Import

    I have a need to perform a network import a 10.2.0.4 DataPump mode database on my old server (HP - UX 11.11) to my new 11.2.0.3 database on a new server (HP - UX 11.31). What I would REALLY like to do is to import directly from my database physical standby (running in mode READ_ONLY while I do importation) rather than having calm my database of production for a couple of hours while I do the import from there.

    What I want to know is if the importation of network mode Data Pump running on 11.2.0.3 the new server creates a task Data Pump to extract in the old database as part of this importation of direct network link. If so, I won't be able to use the physical Standby as the source of my import because Data Pump will not be able to create the main table in the old database. I can't find anything in any Oracle documentation on the use of a physical Standby as a source. I know that I can't do a Data Pump export regularly this database, but I would really like to know if anyone has experience doing this.

    Any comments would be greatly appreciated.

    Bad news, Harry - it worked for me on a standard database open in READ ONLY mode. Not sure what is different between your environment and mine, but there must be something... Read only database is a 10.2.0.4 database running on a box of HP PA-RISC under HP - UX 11.11. The target database runs under 11.2.0.3, on a box of HP Itanium under HP - UX 11.31. The user, I am connected to the database target IMP_FULL_DATABASE privs, and this is the user id used for the DB_LINK and also the same user id on the source database (which, of course, follows!). This user also has the privs permission. My file by looks like this:

    TABLES = AC_LIAB_ %
    NETWORK_LINK = ARCH_LINK
    DIRECTORY = DATA_PUMP_DIR
    JOB_NAME = AMIBASE_IMPDP_ARCHDB
    LOGFILE = DATA_PUMP_DIR:base_impdp_archdb.log
    REMAP_TABLESPACE = ARCHIVE_BEFORE2003:H_DATA
    REMAP_TABLESPACE = ARCHIVE_2003:H_DATA
    REMAP_TABLESPACE = ARCHIVE_2004:H_DATA
    REMAP_TABLESPACE = ARCHIVE_2005:H_DATA
    REMAP_TABLESPACE = ARCHIVE_2006:H_DATA
    REMAP_TABLESPACE = ARCHIVE_2007:H_DATA
    REMAP_TABLESPACE = ARCHIVE_2008:H_DATA
    REMAP_TABLESPACE = ARCHIVE_INDEXES:H_INDEXES
    REUSE_DATAFILES = NO
    SKIP_UNUSABLE_INDEXES = Y
    TABLE_EXISTS_ACTION = REPLACE

  • Export data pump job scheduler does not load

    I have a pump a 10.2.0.2.0 data dump file database, and then import on 11.2.0.3.0 fails with this error:
    ORA-39083: Object type PROCOBJ failed to create with error:
    ORA-06502: PL/SQL: numeric or value error: character to number conversion error
    Failing sql is:
    BEGIN 
    dbms_scheduler.create_job('"MY_JOB_NAME"',
    job_type=>'STORED_PROCEDURE', job_action=>
    'MY_SCHEMA.MY_PROCEDURE'
    , number_of_arguments=>0,
    start_date=>'31-JUL-12 02.05.13.782000 PM EUROPE/BERLIN', repeat_interval=> 
    'FREQ=WEEKLY;BYDAY=FRI;BYHOUR=7;BYMINUTE=0;BYSECOND=0'
    , end_date=>NULL,
    job_class=>'"DEFAULT_JOB_CL
    {code}
    
    I extracted the SQL Code from the dump file and it looks like this:
    {code}
    BEGIN 
    dbms_scheduler.create_job('"MY_JOB"',
    job_type=>'STORED_PROCEDURE', job_action=>
    'MY_SCHEMA.MY_PROCEDURE'
    , number_of_arguments=>0,
    start_date=>'31-JUL-12 02.05.13.782000 PM EUROPE/BERLIN', repeat_interval=> 
    'FREQ=WEEKLY;BYDAY=FRI;BYHOUR=7;BYMINUTE=0;BYSECOND=0'
    , end_date=>NULL,
    job_class=>'"DEFAULT_JOB_CLASS"', enabled=>FALSE, auto_drop=>FALSE,comments=>
    'bla bla comment'
    );
    dbms_scheduler.set_attribute('"MY_JOB"','logging_level','HIGH');
    dbms_scheduler.enable('"MY_JOB"');
    COMMIT; 
    END; 
    / 
    {code}
    
    
    
    After the job is defined the second statement fails:
    
    {code}
    SQL> exec dbms_scheduler.set_attribute('"MY_JOB"','logging_level','HIGH');
    BEGIN dbms_scheduler.set_attribute('"MY_JOB"','logging_level','HIGH'); END;
    
    *
    ERROR at line 1:
    ORA-06502: PL/SQL: numeric or value error: character to number conversion error
    ORA-06512: at "SYS.DBMS_SCHEDULER", line 2851
    ORA-06512: at line 1
    {code}
    
    
    From the source I see:
    
    {code}
    SQL> select logging_level from dba_scheduler_jobs where job_name = 'MY_JOB';
    
    LOGG
    ----
    FULL
    {code}
    
    In the docs I see these valid LOGGING_LEVELs:
    http://docs.oracle.com/cd/E14072_01/server.112/e10595/scheduse008.htm#CHDFDFAB
    
    DBMS_SCHEDULER.LOGGING_OFF
    DBMS_SCHEDULER.LOGGING_FAILED_RUNS
    DBMS_SCHEDULER.LOGGING_RUNS
    DBMS_SCHEDULER.LOGGING_FULL
    
    So please help me, I can't find something useful on MOS, what is Data Pump exporting there, and can not import it again
    Maybe I have overseen a known bug?                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            

    Finally I found the bug myself:

    MOS: Impdp of objects of treatment fails with ORA-39083 and ORA-06502 [ID 780174.1]

    https://support.Oracle.com/epmos/faces/DocContentDisplay?ID=780174.1

  • Degree parallel data pump

    Hello

    I have an impdp job running a long time and I think it's maybe due to too high degree of parallelism. Currently it takes about a minute for a connection session to create.

    Is there a way to reduce this number in the middle of an import? I see the degree DBA_DATAPUMP_JOBS and the corresponding sessions. Can I update this table Data Pump to reduce the number of sessions created by Data Pump? What comes to mind is by updating the column of DEGREE, but it seems risky.

    On a related topic, Hat are general rules to decide on a number of parallel processes?

    It's about Oracle 11.2.0.3.

    Thank you and best regards.

    I see the degree DBA_DATAPUMP_JOBS and the corresponding sessions. Can I update this table Data Pump to reduce the number of sessions created by Data Pump? What comes to mind is by updating the column of DEGREE, but it seems risky.

    You shouldn't be updated any table data pump.
    You can use the command line interface and connect to the session of import do the work.

    Impdp-help (see the bottom section)
    The following commands are valid in interactive mode. Note: the abbreviations are allowed.
    PARALLEL: Change the number of active workers for work in progress.

    I have not done, but it seems that the PARALLEL parameter can help you.

  • How to exclude statistics using Data Pump API?

    How to exclude all statistics when exporting data by using the data pump API Oracle (package DBMS_DATAPUMP)?

    You would call the api metadata filter as follows:

    dbms_datapump. () METADATA_FILTER
    manage = your_handle_here,
    name = "EXCLUDE_PATH_LIST"
    value = 'STATISTICS');

    I hope this helps.

    Dean

Maybe you are looking for