10g to 11 GR 2 upgrade using Data Pump Import

Hello

I intend to move a database 10g a windows server to another. However, there is also a requirement to upgrade this database to GR 11, 2. That's why I was going to combine the 2 in one movement.

1. take an export of complete data to source 10 g data pump
2. create a new database empty 11 g on the target environment
3 import the dump file into the target database

However, I have a couple of queries running in my mind about this approach-

Q1. What happens with the SYSTEM and SYS SYSTEM and SYSAUX objects when importing? Given that I have in fact already created a new dictionary on the empty target database - import SYS or SYSTEM will simply produce errror messages that should be ignored?

Q2. should I use EXCLUDE on SYS and SYSTEM (is it EXCLUDE better export or import side)?

Q3. What happens if there are things like scheduled tasks, etc. on the source system - since these would be stored in the property SYSTEM tables, how I would bring them across to the target 11 g database?

Thank you
Jim

This approach is included in the Upgrade Guide 11 GR 2 - http://docs.oracle.com/cd/E11882_01/server.112/e23633/expimp.htm

PL, ensure that you use SYSDBA privileges to run commands expdp and impdp. See here the first sections of "Note".

http://docs.Oracle.com/CD/B19306_01/server.102/b14215/dp_export.htm#sthref57
http://docs.Oracle.com/CD/E11882_01/server.112/e22490/dp_import.htm#i1012504

As mentioned, sown schemas (for example SYS etc.) are not exported. See http://docs.oracle.com/cd/B19306_01/server.102/b14215/dp_export.htm#i1006790

HTH
Srini

Tags: Database

Similar Questions

  • Data pump import only indexes

    Hello I dumpfile having 100 million documents. I was loading to another server using data pump import utility. It loads the table of data in half an hour or less but it take 7hr when importing indexes. Finally I have to kill jobs and to reveal the index Btree alone was created successfully on all the table any failure of plan B to create now I have the dump file which have all data and metadata. is it possible in utility Datapump to import only the missing indexes or import all indexes only no table and other things? If I use the Include command simply index? Please suggest me a solution

    Oracle Studnet wrote:
    Right, but there is only one way to solve the problem I want to do? can I extract only the index of the dump file and import them rather table and other objects

    Impdp HR/hr dumpfile = directory = data_pump_dir hrdp.dmp include = index content = metadata_only

  • Upgrade of database using Data Pump

    Hello

    I'm moving my database to a Windows 2003 server to Windows 2007 server. At the same time I get to level this database from 10g to 11 g 2 (11.2.0.3).
    so, I use the export / import method of upgrade (via Data Pump not the old exp/imp).

    I have exported successfully by the source database and created the empty shell ready to take import database. However, I have a couple of queries

    Q1. with regard to all objects in the source database SYSTEM. How they will import as target the new database already has a SYSTEM tablespace
    I guess I need to use the TABLE_EXISTS_ACTION option for import. However should I set this Append, SKIP, REPLACE or TRUNCATE - which is best?

    Q2. I intend to slightly modify the directory structure on the new database server - it would be preferable to pre-create the tablespaces or leave that to import but use the REMAPPER DATAFILE option - what is all the world experience to know what is the best way to go? Once again if I pre-create the tablespaces, how should I inform the import to ignore the creation of the tablespaces

    Q3. These 2 databases are on the same network, so in theorey instead of a manual export, copy of the dump for the new server file, then import, I could use a link to network for import. I was wondering where it all stupid this method using the explicit export dump file?

    Thank you
    Jim

    Jim,

    Q1. with regard to all objects in the source database SYSTEM. How they will import as target the new database already has a SYSTEM tablespace
    I guess I need to use the TABLE_EXISTS_ACTION option for import. However should I set this Append, SKIP, REPLACE or TRUNCATE - which is best?

    What if all you have is the basic database and nothing created, then you can do the full = y. In fact, it is probably what you want. The system tablespace will be there then when Data Pump attempts to create, it will fail just as the create statement. Anything else will fail. In most cases, your system tables will already be there, and that's ok too. If you import schema view, you will miss out on some of the other things.

    Q2. I intend to slightly modify the directory structure on the new database server - it would be better to pre-create the tablespaces or leave cela import but use the REMAPPING > DATAFILE option - what is everyones experience to know what is the best way to go? Once again if I pre-create the tablespaces, how should I inform the import to ignore the creation of the tablespaces

    If the directory structure is different (which they usually are), then there is no easier way. You can run impdp but with sqlfile and you can tell - include = tablespace. This will give you all orders of tablespace to create in a txt file and you can edit the text file to change what you want to change. You can tell datapump to ignore the creation of tablespace using--exclude = tablespace

    Q3. These 2 databases are on the same network, so in theorey instead of a manual export, copy of the dump for the new server file, then import, I could use a link to network for import. I > I was just wondering where it all with this method using the explicit export dump file?

    The only con might be if you have a slow network. This will make it slower, but if you need to copy the dumpfile on the same network, then you will always see the same basic traffic. The benefits are that we should not have additional disk space. Here's how I look at it.

    1. you have XX GB for the source database
    2. you are looking for the source dumpfile YY GB
    3. you must YY GB for the dumpfile target that you copy
    4. you have XX GB for the target database.

    By network get rid if YY * 2 GB for the dumpfiles.

    Dean

  • Export the whole (10 GB) using Data Pump utility export base

    Hello

    I have a requirement that we have to export all of the base (10 GB) using Data Pump export utility because it is not possible to send the discharge of 10 GB in a CD/DVD for the seller of our application system (to analyze a few problems that we have).

    Now when I checked online full export is available but not able to understand how it works, as we never used this data pump utility, we use the method of normal export. In addition, the data pump will reduce the size of the dump file so it can fit on a DVD or can we use utility export parallel DB full to split the files and include them in a DVD, it is possible.

    Please correct me if I am wrong and kindly help.

    Thank you for your help in advance.

    Pravin,

    The server saves files in the directory object that you specify on the command line. So what you want to do is:

    1. from your operating system, find an existing directory or create a new directory. In case you, C:/Dump is as good a place as any.

    2. connect to sqlplus and create the directory object. Just use the path. I use linux, so my directory looks like/scratch/xxx/yyy
    If you use Windows, the path to your directory would look like C:/Dump you should not attack.

    3. don't forget to grant access to this directory. You can grant access to a single user, group of users, or the public. Just like
    any other object.

    If it helps, or if she has answered your question, please mark messages with the appropriate tag.

    Thank you

    Dean

  • Select table when import using Data Pump API

    Hello
    Sorry for the trivial question, I export the data using Data Pump API, with the mode "TABLE".
    If all the tables will be exported in a .dmp file.

    So, my question is how to import a few tables using Data Pump API?, how to set the "TABLES" property as a command line interface?
    can I use procedures DATA_FILTER?, if so how?

    Really thanks in advance

    Kind regards

    Kahlil

    Hello

    You should use the procedure of metadata_filter for the same thing.
    for example:

    dbms_datapump.metadata_filter
                (handle1
                 ,'NAME_EXPR'
                 ,'IN (''TABLE1'', '"TABLE2'')'
                );
    {code}
    
    Regards
    Anurag                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        
    
  • Error data pump import

    Hi all

    I get errors below when you try to use the data pump import a table into the dump of the file (of a pre-production database) in a different database (prod_db) even version. I used the expdp to generate the dump file.

    Gettings errors-
    ORA-39083
    ORA-00959
    ORA-39112

    Any suggestions or advice would be appreciated.

    Thank you


    Import: Release 10.2.0.1.0 - 64 bit Production

    Copyright (c) 2003, 2005, Oracle. All rights reserved.
    ;;;
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - 64 bit Production
    With partitioning, OLAP and Data Mining options
    Main table 'SYSTEM '. "' SYS_IMPORT_TABLE_01 ' properly load/unloaded
    Start "SYSTEM". "" SYS_IMPORT_TABLE_01 ": System / * DIRECTORY = DATA_PUMP_DIR DUMPFILE = EXP_Mxxxx_1203.DMP LOGFILE = Mxxxx_CLAIMS. LOG TABLES = CLAIMS
    Object type SCHEMA_EXPORT/TABLE/TABLE processing
    ORA-39083: object of type THAT TABLE could not create with the error:
    ORA-00959: tablespace "OXFORD_DATA_01" does not exist
    Because sql is:
    CREATE TABLE 'xxx '. "" CLAIMS "("CLAIM_NUM"VARCHAR2 (25) NOT NULL ACTIVATE, ACTIVATE THE"LINE_NUM"NUMBER (3.0) NOT NULL, 'PAID_FLAG' CHAR (1), CHAR (1) 'ADJ_FLAG', 'CLAIM_TYPE' VARCHAR2 (20), 'MEM_ID' VARCHAR2 (20), VARCHAR2 (20) 'MEM_SUF', DATE 'MEM_BEG_DATE', 'MEM_REL' CHAR (2), 'MEM_NAME' VARCHAR2 (40), DATE 'MEM_DOB', 'MEM_SEX' CHAR (1),"REC_DATE"DATE, DATE OF THE"PAID_DATE") 'FROM_DATE' DATE "
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/INDEX
    ORA-39112: type of dependent object INDEX: 'xxx '. «CLAIMS_IDX1 ' ignored, base object type TABLE: 'Xxx'.» "" Creation of CLAIMS ' failed
    ORA-39112: type of dependent object INDEX: 'xxx '. «CLAIMS_PK ' ignored, base object type TABLE: 'Xxx'.» "" Creation of CLAIMS ' failed
    Object type SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS of treatment
    ORA-39112: dependent object type skipped INDEX_STATISTICS, object-based INDEX type: 'xxx '. "' CLAIMS_IDX1 ' failed to create
    ORA-39112: dependent object type skipped INDEX_STATISTICS, object-based INDEX type: 'xxx '. "" Creation of CLAIMS_PK ' failed
    Object type SCHEMA_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS treatment
    ORA-39112: dependent object type skipped TABLE_STATISTICS, object-based ARRAY type: "xxx". "" Creation of CLAIMS ' failed
    Work 'SYSTEM '. "" SYS_IMPORT_TABLE_01 "completed with error (s) from 6 to 13:49:33


    Impdp name username/password DIRECTORY = DATA_PUMP_DIR DUMPFILE = EXP_Mxxxx_1203.DMP LOGFILE is Mxxxx_CLAIMS. LOG TABLES = CLAIMS

    Before you create the tablespace or use clause REMAP_TABLESPACE for the import.

  • "Resume" a Data Pump import Execution failure

    First of all, here is a "failure" Data Pump Import I want to "resume":
    ] $ impdp directory dumpfile=mySchema%U.dmp "" / as sysdba "" = DATA_PUMP_DIR logfile = Import.log patterns parallel = MYSCHEMA = 2

    Import: Release 10.2.0.3.0 - 64 bit Production Tuesday, February 16, 2010 14:35:15

    Copyright (c) 2003, 2005, Oracle. All rights reserved.

    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.3.0 - 64 bit Production
    With the options of Real Application Clusters, partitioning, OLAP and Data Mining
    Table main 'SYS '. "' SYS_IMPORT_SCHEMA_01 ' properly load/unloaded
    Departure 'SYS '. "' SYS_IMPORT_SCHEMA_01 ': ' / * AS SYSDBA"dumpfile=mySchema%U.dmp directory = DATA_PUMP_DIR logfile = Import.log patterns parallel = MYSCHEMA = 2
    Processing object type SCHEMA_EXPORT/PRE_SCHEMA/PROCACT_SCHEMA
    Object type SCHEMA_EXPORT/SYNONYM/SYNONYM of treatment
    Object type SCHEMA_EXPORT/TYPE/TYPE_SPEC of treatment
    Object type SCHEMA_EXPORT/SEQUENCE/SEQUENCE of treatment
    Processing object type SCHEMA_EXPORT/SEQUENCE/EXCHANGE/OWNER_GRANT/OBJECT_GRANT
    Object type SCHEMA_EXPORT/TABLE/TABLE processing
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    . . imported "MYSCHE"...
    ...
    ... 0 KB 0 rows
    Processing object type SCHEMA_EXPORT/TABLE/SCHOLARSHIP/OWNER_GRANT/OBJECT_GRANT
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/INDEX
    ORA-39126: worker unexpected fatal worker error of $ MAIN. LOAD_METADATA [INDEX: "MONSCHEMA".] ["" OBJECT_RELATION_I2 "]
    SELECT process_order, flags, xml_clob, NVL (dump_fileid,: 1), NVL (dump_position,: 2), dump_length, dump_allocation, constituent, object_row, object_schema, object_long_name, processing_status, processing_state, base_object_type, base_object_schema, base_object_name, goods, size_estimate, in_progress 'SYS '. "" SYS_IMPORT_SCHEMA_01 "process_order where: 3: 4 AND processing_state <>: 5 AND double = 0 ORDER BY process_order
    ORA-06502: PL/SQL: digital error or value
    ORA-06512: at "SYS." "MAIN$ WORKER", line 12280
    ORA-12801: error reported in the parallel query P001, instance pace2.capitolindemnity.com:bondacc2 Server (2)
    ORA-30032: the suspended order (trade-in) has expired
    ORA-01652: unable to extend segment temp of 128 in tablespace OBJECT_INDX
    ORA-06512: at "SYS." DBMS_SYS_ERROR', line 105
    ORA-06512: at "SYS." "MAIN$ WORKER", line 6272
    -PL/SQL call stack-
    the line object
    serial number of handle
    package body 14916 0x1f9ac8d50 SYS. MAIN$ WORKER
    body 0x1f9ac8d50 6293 package SYS. MAIN$ WORKER
    body 0x1f9ac8d50 3511 package SYS. MAIN$ WORKER
    body 0x1f9ac8d50 6882 package SYS. MAIN$ WORKER
    body 0x1f9ac8d50 1259 package SYS. MAIN$ WORKER
    0x1f8431598 anonymous block 2
    Job 'SYS '. "' SYS_IMPORT_SCHEMA_01 ' stopped because of the fatal 23:05:45
    As you can see this work is not without treatment statistics, constraints, etc. from PL/SQL. I want to do is make another impdp command but ignore objects that were imported with success as shown above. Is it possible to do (using the EXCLUDE parameter maybe?) with impdp? If so, what would it be?

    Thank you

    John

    Why you just restart this work. It ignores everything that has already been imported.

    Impdp "" / as sysdba "" attach = "SYS." "" SYS_IMPORT_SCHEMA_01 ".

    Import > keep

    Dean

  • Data pump import

    I cannot using the data pump import tool. I work with Oracle 10.2.4.0 on Windows Server 2008. I am trying to import a large amount of data (several patterns, thousands of paintings, hundreds of GB of data) of the same version of Oracle on Linux. Since there is so much data and will be if long to import, I try to make sure that I can make it work for a table or a diagram before I try. So I will try to import the TEST_USER. Table 1 table using the following command:

    Dumpfile = logfile = big_file_import.log big_file.dmp Impdp.exe tables = test_user.table1 remap_datafile=\'/oradata1/data1.dbf\':\'D:\oracle\product\10.2.0\oradata\orcl\data1.dbf\'

    I provide sys as sysdba to connect when he asks (not sure how you get this info in the original order). However, I get the following error:

    the "TEST_USER" user does not exist

    My understanding is that the data pump utility would create all the necessary patterns for me. Is it not the case? The target database is a new facility for the source schemas do not exist here.

    Even if I create the schema of test_user by hand, then I get an error message indicating that the tablespace does not exist:

    ORA-00959: tablespace "TS_1" does not exist

    Then even to do it, I don't want to have to do first, does not work. Then he gives me something about how the user has no privileges on tablespace.

    The data pump utility is not supposed to do that sort of thing automatic? What I really need to create all schemas and storage by hand? That will take a long time. I'm doing something wrong here?

    Thank you
    Dave

    tables = test_user.table1

    The "TABLES" mode does NOT create database accounts.

    The FULL mode creates space storage and database accounts before importing the data.

    PATTERNS mode creates the database accounts before importing data - but expect Tablespaces to exist before so that tables and indexes can be created in the correct storage spaces.

    Hemant K Collette
    http://hemantoracledba.blogspot.com

  • Data pump import - can I can the name of the table when importing only?

    It is possible to use data pump export on a table name, and then import (add lines) at a table of different table with the same structure of exact column? I don't see a way to change only the name of the table when importing.

    Hello

    From 11.1 you can remap a table name. Use:

    remap_table = old_name:new_name

    Thank you

    Dean

  • Migration using data pump for Oracle 10 g-&gt; Oracle 11 g

    Hi all

    1)
    At the moment I am using Oracle 11g. I have a plan to import data from Oracle 10 g. I would like to know if its possible to import data that has been exported by datapump on Oracle 10 g?

    Can I convert somehow expdp out of Oracle 10 g Oracle 11 g format?





    2)
    The next question is. If I use expdp to create the dump of the database complete. Can I use *.dmp for import selected users? Or only the complete database can be restored?

    Yes, you can import dump 10g in an 11g database.

    Maybe you should take the time and read the section on datapump in the Oracle thin® [Database Utilities | http://download.oracle.com/docs/cd/B28359_01/server.111/b28319/dp_import.htm#i1007324] Manual.
    : p

  • How to exclude statistics using Data Pump API?

    How to exclude all statistics when exporting data by using the data pump API Oracle (package DBMS_DATAPUMP)?

    You would call the api metadata filter as follows:

    dbms_datapump. () METADATA_FILTER
    manage = your_handle_here,
    name = "EXCLUDE_PATH_LIST"
    value = 'STATISTICS');

    I hope this helps.

    Dean

  • Update of data by Data Pump Import

    Hi all!

    I want to update the data in my database using Full Import Data Pump to a different database. But I do not know which option I should use when executing import for the second time? Or I can run full import without encompassing any option again?

    Thank you

    Tien Lai

    What if all you want to do is update of data and if the tables already exist, you can use this:

    content of user/password = data_only =(truncate or append) table_exists_action Impdp...

    If you use Add, then add new data to the existing data. If you use truncate then existing data will be deleted and the data in the dumpfile will be imported.

    There will be problems if you have a referential constraint. Say that table 1 references given in table 2, if datapump loads data into Database2 first, and then it truncates the data again. This truncation may fail because there is a referential constraint on it. If you have a ref, then you must disable them before you run impdp and then allow them once impdp is completed.

    If all you want after the first import only the data, you can add the

    content = data_only

    your order expdp. He would complete much faster.

    Do not forget that your statistics on the tables and the indexes will not be reloaded if you use table_exists_action = truncate or replace existing statistics so would probably obsolete.

    If you want to replace the table and the data, then the command would be:

    Impdp username/password table_exists_action = replace...

    This will remove the table and re-create it, and then load the data, then create all dependent objects for tables.

    I hope this helps.

    Dean

  • Data pump import a table in to another schema in 11g

    Hi all

    I have an Oracle 11.2 database and I have a requirement to import a few tables in a new scheme using my export from the previous month. I can't import any scheme as it is very large. I check REMAP_TABLE option, but it is enough to create the table in the same pattern and only rename the table.


    For example, I TABLE GAS. EMPLOYE_DATA I want to import the to GASTEST. EMPLOYEE_DATA

    Is there a way to do it using datapump?



    Appriciate any advice.

    Hello

    You can use the parameter INCLUDE in order to only select one Table:

    REMAP_SCHEMA=SCOTT:SCOTT_TEST
    INCLUDE=TABLE:"IN ('EMP')"
    

    Hope this helps.
    Best regards
    Jean Valentine

  • I want to learn more about the data pump and table space transferable

    Please notify easy tutorials as I want to know how to import and export between oracle 10 and 11.
    Thank you

    Hello
    Please check this oracle tutorial:
    http://www.exforsys.com/tutorials/Oracle-10G/Oracle-10G-using-data-pump-import.html
    about transportable table spaces, you may consult:
    http://www.rampant-books.com/art_otn_transportable_tablespace_tricks.htm
    Kind regards
    Mohamed
    Oracle DBA

  • is it possible to exclude some tables of data pump?

    Hello

    I want to exclude some tables while import/export to a single system using data pump. Is this all possible for this?

    If this isn't the case, please suggest me for all other ideas too.

    Iqbal

    Use the exclude parameter and check the link below.

    http://www.Oracle-base.com/articles/10G/Oracle-data-pump-10G.php

    Concerning
    Asif Kabir

Maybe you are looking for

  • Unable to find Mail 10.0.2

    Since the upgrade to 10.0.2, I was able to pick up my mail. The normal field of search box at the top of the list of messages does not appear. Very frustrating. Any suggestions on how to do this come back, so I can get my emails?

  • I'll try to find a list of bookmarks in SQ.lite profiles, but none appears

    I deleted bookmarks today by mistake and was hoping to get back them. To this end, I looked at the instructions on how to find my profile. When I chose bookmarkbackups it shows files that are not able to be opened with any application I have, and whe

  • Satellite U200-112: switching between monitors

    Satellite U200-112. To the passage of the mobile display on the external monitor, the resolution is exposed to the display, instead of the external display. Who faced such a problem?

  • Dual boot Linux - Windows on the Satellite U940

    Hello to everyone. I'm pretty new to this forum and I have a Toshiba Satellite nine U940-10N? Using Linux for the past 2 years, I would install a debian wheezy on this distribution. Only Debian has been installed on my old computer, so I'm net really

  • I can't play my Apple music albums downloaded?

    A few days ago, I decided to sign up for the Apple Music 3 month free trial.  Then, I went and downloaded about 4 albums, so that I wouldn't have to have the Internet connection to be able to listen to them.  However, the next night, when I tried to