Export schema through Oracle data pump with question database Vault enabled

Hello

I installed and configured Vault database on an Oracle 11 g - r2 - 11.2.0.3 to protect a specific schema (SCHEMA_NAME) via a Kingdom. I followed the following doc:
http://www.Oracle.com/technetwork/database/security/TWP-databasevault-DBA-BestPractices-199882.PDF
to ensure that the system and the network user has sufficient rights to complete a schedule pump data Oracle export operation.

I.e. I gave sys and system the following:
execute dvsys.dbms_macadm.authorize_scheduler_user ('sys', 'SCHEMA_NAME');
execute dvsys.dbms_macadm.authorize_scheduler_user ('system', 'SCHEMA_NAME');

execute dvsys.dbms_macadm.authorize_datapump_user ('sys', 'SCHEMA_NAME');
execute dvsys.dbms_macadm.authorize_datapump_user ('system', 'SCHEMA_NAME');

I also create a second domain on the same schema (SCHEMA_NAME) to allow sys and system to manage indexes for the actual tables protected, in order to allow a system to manage indexes for the tables of protected area and sys. This separate realm was created for all types of indexes: Index, Partition of Index and Indextype, sys and system were allowed as the OWNER of this Kingdom.

However, when I try and complete an operation of Oracle Data Pump export on the schematic, I get two errors directly after the following line appears in the export log:

Processing object type SCHEMA_EXPORT/TABLE/INDEX/DOMAIN_INDEX/INDEX:
ORA-39127: unexpected error of the call to export_string: = SYS. DBMS_TRANSFORM_EXIMP. INSTANCE_INFO_EXP('AQ$_MGMT_NOTIFY_QTABLE_S','SYSMAN',1,1,'11.02.00.00.00',newBlock)
ORA-01031: insufficient privileges
ORA-06512: at "SYS." DBMS_TRANSFORM_EXIMP', line 197
ORA-06512: at line 1
ORA-06512: at "SYS." Dbms_metadata", line 9081
ORA-39127: unexpected error of the call to export_string: = SYS. DBMS_TRANSFORM_EXIMP. INSTANCE_INFO_EXP('AQ$_MGMT_LOADER_QTABLE_S','SYSMAN',1,1,'11.02.00.00.00',newBlock)
ORA-01031: insufficient privileges
ORA-06512: at "SYS." DBMS_TRANSFORM_EXIMP', line 197
ORA-06512: at line 1
ORA-06512: at "SYS." Dbms_metadata", line 9081

The export is completed, but this error.

Any help, pointers, suggestions, etc. in fact no matter what will be very welcome at this stage.

Thank you

I moved the forum thread on the "database - security. If the document does not help, pl open an SR with Support

HTH
Srini

Tags: Database

Similar Questions

  • ODI LKM Oracle for Oracle Data Pump question

    Hi all

    I have a weird problem, ODI.

    I associate myself with per_all_people_f, fnd_user to load the w_user_ds using Oracle Data Integrator. The used LKM is LKM Oracle for Oracle Data Pump.

    Fine when I run the interface. I am getting below error

    ODI-1227: task failed USER_DATA_SET (load) on the source of connection ORACLE EBS.

    Caused by: java.sql.SQLSyntaxErrorException: ORA-00923: KEYWORD not found where expected

    at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:462)

    at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:405)

    at oracle.jdbc.driver.T4C8Oall.processError(T4C8Oall.java:931)

    at oracle.jdbc.driver.T4CTTIfun.receive(T4CTTIfun.java:481)

    at oracle.jdbc.driver.T4CTTIfun.doRPC(T4CTTIfun.java:205)

    at oracle.jdbc.driver.T4C8Oall.doOALL(T4C8Oall.java:548)

    at oracle.jdbc.driver.T4CPreparedStatement.doOall8(T4CPreparedStatement.java:217)

    at oracle.jdbc.driver.T4CPreparedStatement.executeForRows(T4CPreparedStatement.java:1115)

    at oracle.jdbc.driver.OracleStatement.doExecuteWithTimeout(OracleStatement.java:1488)

    at oracle.jdbc.driver.OraclePreparedStatement.executeInternal(OraclePreparedStatement.java:3769)

    at oracle.jdbc.driver.OraclePreparedStatement.execute(OraclePreparedStatement.java:3954)

    at oracle.jdbc.driver.OraclePreparedStatementWrapper.execute(OraclePreparedStatementWrapper.java:1539)

    at oracle.odi.runtime.agent.execution.sql.SQLCommand.execute(SQLCommand.java:163)

    at oracle.odi.runtime.agent.execution.sql.SQLExecutor.execute(SQLExecutor.java:102)

    at oracle.odi.runtime.agent.execution.sql.SQLExecutor.execute(SQLExecutor.java:1)

    at oracle.odi.runtime.agent.execution.TaskExecutionHandler.handleTask(TaskExecutionHandler.java:50)

    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.processTask(SnpSessTaskSql.java:2913)

    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java:2625)

    at com.sunopsis.dwg.dbobj.SnpSessStep.treatAttachedTasks(SnpSessStep.java:577)

    at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java:468)

    at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java:2128)

    to oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$ 2.doAction(StartSessRequestProcessor.java:366)

    at oracle.odi.core.persistence.dwgobject.DwgObjectTemplate.execute(DwgObjectTemplate.java:216)

    at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.doProcessStartSessTask(StartSessRequestProcessor.java:300)

    to oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.access$ 0 (StartSessRequestProcessor.java:292)

    to oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$ StartSessTask.doExecute (StartSessRequestProcessor.java:855)

    at oracle.odi.runtime.agent.processor.task.AgentTask.execute(AgentTask.java:126)

    to oracle.odi.runtime.agent.support.DefaultAgentTaskExecutor$ 2.run(DefaultAgentTaskExecutor.java:82)

    at java.lang.Thread.run(Thread.java:662)

    The generated code is

    create the table 780021 X

    (

    C1_FIRST_NAME,

    C2_MID_NAME,

    C3_LAST_NAME,

    C4_FULL_NAME,

    C5_NAME_SUFFIX,

    C6_SEX_MF_CODE,

    C7_SEX_MF_NAME,

    C8_COUNTRY_NAME,

    C9_LOGIN,

    C10_CREATED_BY_ID,

    C11_CHANGED_BY_ID,

    C12_CREATED_ON_DT,

    C13_CHANGED_ON_DT,

    C14_AUX1_CHANGED_ON_DT,

    C15_SRC_EFF_TO_DT,

    C16_INTEGRATION_ID,

    C17_EFFECTIVE_START_DATE

    )

    (EXTERNAL) ORGANIZATION

    TYPE oracle_datapump

    Dat_dir default DIRECTORY

    LOCATION ("X780021.exp")

    )

    PARALLEL

    in SELECT

    ALL_PEOPLE_F.FIRST_NAME,

    ALL_PEOPLE_F.MIDDLE_NAMES,

    ALL_PEOPLE_F.LAST_NAME,

    ALL_PEOPLE_F.FULL_NAME,

    ALL_PEOPLE_F.SUFFIX,

    ALL_PEOPLE_F.SEX,

    ALL_PEOPLE_F.SEX,

    ALL_PEOPLE_F.NATIONALITY,

    USER. USER_NAME,

    ALL_PEOPLE_F.CREATED_BY,

    ALL_PEOPLE_F.LAST_UPDATED_BY,

    ALL_PEOPLE_F.CREATION_DATE,

    ALL_PEOPLE_F.LAST_UPDATE_DATE,

    ALL_PEOPLE_F.CREATION_DATE,

    ALL_PEOPLE_F.EFFECTIVE_END_DATE,

    USER. USER_ID,

    ALL_PEOPLE_F.EFFECTIVE_START_DATE

    from APPS. FND_USER USER, APPS. PER_ALL_PEOPLE_F ALL_PEOPLE_F

    where (1 = 1)

    And (ALL_PEOPLE_F.PERSON_ID = USER. EMPLOYEE_ID)

    I don't see what is the problem here.

    Someone can help me.

    Thank you and best regards,

    Krishna Prasad

    I found the problem, its with the way ODI generated alias for the FND_USER table, by default it produces USER as an alias, which is a keyword from oracle. We just need to rename it to something else, and it worked.

  • Oracle Data Guard with EBS R12

    Dear gurus,

    I want to know about Oracle Data Guard with Oracle EBS Release 12, claims we separate license with Oracle E-Business Suite R12 of application Data Guard? (i.e. separate premises, same place).


    Please me advice in this regard.


    -Mickaël

    AFAIK, you don't need to have a separate license for Data Guard. However, I would say that you call your Oracle representative.

  • import schema data using oracle data pump

    Hi all

    I want to import a schema on a single server using IMPDP. But on this server, the user does not exists.is it is possible to import all the objects that break including the creation of the schema. In exp/imp, it may be possible. But in data pump I can't do the same. Its giving error that the user does not exist. For your reference, I'll give you the code below the import.

    Impdp system directory = DIR_DUMP logfile = expdp_tsadmin_tcgadmin_11182013_2_2.log

    dumpfile = expdp_tsadmin_tcgadmin_11182013_2.dmp PATTERNS = TS_ADMIN version = 10.1.0.2.0

    Pls help and thanks in advance.

    Thank you

    Piku

    Hello

    All errors are due to the fact that the tablespace that ts_data does not exist, everything else is hit because of this.

    create tablespace ts_data datafile 'xxxxxx' size xM;

    then try again

    See you soon,.

    Harry

  • Oracle Data Pump

    oracle11g oracle introduced the new feature called oracle data dump. What is the difference in comparing utility imp/exp? can someone explain how to use this new feature

    for example, by using the user scott to export some tables in my diagram.

    Data pump is very fast, parallel importation is made in 11g. exp/imp sometimes fails due to problems of space on the server. but by using method of parallel import is performed. parallel importation is nothing but imported when 1 row have exported the corresponding row in the table, thus saving disk space. With Data Pump Import, a single stream of data load is about 15 to 45 times faster than the initial import.

  • Data Pump - expdp / question utility impdp

    HiAll,

    As a basis for learning exercise Data pump, I am trying to export the schema scott and then you want to load the dump file into a test_t diagram in the same database.

    -1. from dir object sysdba
    CREATE a DIRECTORY dpump_dir1 AS 'C:\output_dir ';

    -2. created dir on the operating system as c:/output_dit

    -3. run expdp
    schemas system/***@orcl expdp = scott DIRECTORY = JOB_NAME = hr DUMPFILE = scott_orcl_nov5.dmp PARALLEL dpump_dir1 = 4

    -4. create a schema test_t and allows the dba to it.

    -5. impdp execution
    Impdp system/***@orcl patterns = DIRECTORY = JOB_NAME = hr DUMPFILE = scott_orcl_nov5.dmp PARALLEL dpump_dir1 test_t = 8

    He does not here as ORA39165: test_t schema not found. However, the schema test_t exist.

    So, I don't know why she should give to this error. It seems that the Scott schema in the expdp dump file may not be charged to any other schema but only to a schema named scott... Is this good? If Yes, how can I load all objects in the schema scott say to another schema say test_t? It would be useful that you can show the respective expdp and impdp command please.

    Thank you very much
    KS

    Hello

    You must specify remap_schema when you import

    do impdp system/***@orcl = dpump_dir1 = scott_orcl_nov5.dmp remap_schema = scott DUMPFILE directory: test_t

    See you soon

  • register the database with the database Vault

    Version: 11.2.0.2.0

    Operating system: Solaris

    Can someone tell me please how to enter Oracle Database Vault with a database from command line without using DBCA.

    The doc oracle E23090-10 provides only through the DBCA Wizard. I'm having some trouble with my software x - manager, so I cannot use the DBCA Wizard and seeks to enter the vault of the database with the database by command line.


    Oracle database Vault is already enabled through 'chopt enable lbac' and 'chopt enable dv. Now all I do is register with the databases on the server.

    Thank you

    Hi DK2010, actually I found my answer in the metalink Doc ID 803948.1

    I have to run the dbca in silent mode and at the end of this doc, it's exactly what I need.

    "

    DBCA-silent - configureDatabase

    sourceDB-

    [- sysDBAUserName

    -sysDBAPassword ]

    -addDBOption DV, WHO

    -dvUserName

    -dvUserPassword

    -dvAccountManagerName

    -dvAccountManagerPassword

    "

    Now, the keystone of the database is registered with the database.

  • To update the schema in the data pump

    Hello

    Database version: 11.2.0.1.0

    Is that we can update the schema in the database without creating the similar to the network mode dump file

    for example

    I have a DB 1 database to say, it contains two schema Test and production, we can update the test with production data without creating the dump file in the database server.

    the whole idea is to perform the export and import data in a single step here to reduce the time and human intervention.

    Currently, I've followed the steps below:

    (1) export of production data

    (2) file the test schema

    (3) create a diagram of the test

    (4) import production data in the test schema

    Thank you

    Sery

    Hello

    It is possible.,.

    SQL > create public database link impdpt connect to the SYSTEM identified by abc123 using 'DG1 ';

    Database link created.

    SQL >

    -----

    Directory of system/abc123 impdp [oracle@prima admin] $ = network_link = impdpt schemas = hr remap_schema hr:hrtn logfile = test_same.log = DATA_PUMP_DIR

    Import: Release 11.2.0.4.0 - Production on Tue Dec 29 02:53:24 2015

    Copyright (c) 1982, 2011, Oracle and/or its affiliates.  All rights reserved.

    Connected to: Oracle Database 11 g Enterprise Edition Release 11.2.0.4.0 - 64 bit Production

    With partitioning, OLAP, Data Mining and Real Application Testing options

    Start "SYSTEM". "" SYS_IMPORT_SCHEMA_01 ": System / * Directory = network_link = impdpt schemas = hr remap_schema hr:hrtn logfile = test_same.log = DATA_PUMP_DIR

    Current estimation using BLOCKS method...

    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA

    Total estimation using BLOCKS method: 448 KB

    Processing object type SCHEMA_EXPORT/USER

    Processing object type SCHEMA_EXPORT/SYSTEM_GRANT

    Processing object type SCHEMA_EXPORT/ROLE_GRANT

    Processing object type SCHEMA_EXPORT/DEFAULT_ROLE

    Processing object type SCHEMA_EXPORT/PRE_SCHEMA/PROCACT_SCHEMA

    Object type SCHEMA_EXPORT/SEQUENCE/SEQUENCE of treatment

    Object type SCHEMA_EXPORT/TABLE/TABLE processing

    . . imported "HRTN. "" COUNTRY "25 lines

    . . imported "HRTN. "' DEPARTMENTS ' 27 lines

    . . imported "HRTN. "' EMPLOYEES ' 107 lines

    . . imported "HRTN. "JOBS"19 ranks. "

    . . imported "HRTN. "" JOB_HISTORY "10 lines

    . . imported "HRTN. "" LOCATIONS "23 lines

    . . imported "HRTN. "The REGIONS"4 lines.

    Processing object type SCHEMA_EXPORT/TABLE/SCHOLARSHIP/OWNER_GRANT/OBJECT_GRANT

    Object type SCHEMA_EXPORT/TABLE/COMMENT of treatment

    Object type SCHEMA_EXPORT/PROCEDURE/treatment PROCEDURE

    Object type SCHEMA_EXPORT/PROCEDURE/ALTER_PROCEDURE processing

    Processing object type SCHEMA_EXPORT/TABLE/INDEX/INDEX

    Object type SCHEMA_EXPORT/TABLE/CONSTRAINT/treatment

    Object type SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS of treatment

    Object type SCHEMA_EXPORT/VIEW/VIEW processing

    Object type SCHEMA_EXPORT/TABLE/CONSTRAINT/REF_CONSTRAINT of treatment

    Object type SCHEMA_EXPORT/TABLE/TRIGGER processing

    Object type SCHEMA_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS treatment

    Work 'SYSTEM '. "' SYS_IMPORT_SCHEMA_01 ' completed Fri Dec 29 02:54:09 2015 elapsed 0 00:00:44

    [oracle@prima admin] $

  • Failure of the load of the inventory... OPatch failed to load the inventory for the Home.OPatch Oracle data failed with the error code = 73

    Hello

    I will apply patch bundle and my oracle database 11.2.0.2 and Microsoft Windows x 86 platform. As read the file Readme Oracle recommends to use OPatch utility version 11.2.0.1.3 or later. To check the inventory availability I use the command lsinventory % ORACLE_HOME%/OPatch/opatch.

    It ended to give details below.

    Citing O 11.2 patch .0.1.1

    OPatch could not create/open history for writing file: * my home oracle path***\cfgtoollogs\opatch\opatch_history.txt

    **ORACLE_HOME***\cfgtoollogs\opatch\opatch_history.txt (access is denied)

    Setup Oracle interim Patch version 11.2.0.1.1

    Copyright (c) 2009, Oracle Corporation.  All rights reserved.

    OPatch could open the log file, the recording is not possible

    Home Oracle: *.

    Inventory Center: C:\Program Files\Oracle\Inventory

    from: n/a

    OPatch version: 11.2.0.1.1

    YES version: 11.2.0.2.0

    YES location: * ORACLE_HOME * \oui

    Location of the log file: **ORACLE_HOME**\cfgtoollogs\opatch\opatch2015-04-09_16-27-49PM.log

    History of patch file: **ORACLE_HOME**\cfgtoollogs\opatch\opatch_history.txt

    Failure of the load of the inventory... OPatch failed to load the inventory for the given Oracle Home.

    LsInventorySession failed: LsInventory failed to create the log directory * ORACLE_HOME * \cfgtoollogs\opatch\lsinv\lsinventory2015-04 -.

    09_16-27 - 49 PM .txt

    OPatch failed with error code = 73

    Can someone please help me solve this problem?

    Do I need to install the last opatch utility?

    Hello

    This error message " LsInventorySession failed: LsInventory failed to create the log directory * ORACLE_HOME * \cfgtoollogs\opatch\lsinv\lsinventory2015-04 - .

    09_16-27 - 49 PM .txt "

    indicates that you have a permission problem with your file systems: cfgtoollogs/ORACLE_HOME/opatch

    Check the owner of this file system.

    Kind regards

  • Line of Oracle data object with a statement select column entry.

    create table mainTable)

    Offer number,

    emp_name VARCHAR2 (15)

    );

    Insert in mainTable values(1,'ifti');

    CREATE or REPLACE TYPE table1_objtyp () AS OBJECT

    Offer number,

    emp_name VARCHAR2 (15)

    );

    create table historical table

    {

    number of TID,

    empObj table1_objtyp

    };

    I want to insert mainTable recode in field empObj historical table with a select as below.

    Insert Into table values history (1, (Select * from mainTable where bid = 1));

    How can I achieve this?

    insert into the historical table SELECT 1, table1_objtyp (IDB, emp_name) OF MAINTABLE.

  • Oracle 12 c with RMAN database backup

    Hi All-

    I started back to the top of one of my PDB using utility RMAN following the instructions on the link below.

    http://www.Oracle.com/WebFolder/technetwork/tutorials/OBE/DB/12C/R1/PDB/pdb_pitr/pdb_pitr.html

    --Connect to Oracle RMAN.
    rman> Connect target /
    --Set the Oracle RMAN backup device type and backup file location.
    rman> configure channel device type disk format '/stage/db_backup_files/cdb1/%U';
    --Turn on automatic backup of control files.
    rman> configure controlfile autobackup on;
    --Back up the database and archive logs.
    rman>backup database plus archivelog;
    rman> exit
    

    Back to the top has been done successfully, but when I started with the recovery process he gave me error reporting collection.

    Run {set until what SNA = 2263440; restore database connectable to pdb2; retrieve auxiliary destination of the pluggable database pdb2 = "/ stage/db_backup_files/cdb1 '; alter connectable database to pdb2 ;} open resetlogs}

    This is the track to RMAN

    Recovery Manager: release 12.1.0.2.0 - Production on sat 22 Aug 12:22:44 2015

    Copyright (c) 1982, 2014, Oracle and/or its affiliates.  All rights reserved.

    RMAN > Connect target /.

    connected to target database: RELIABLE (DBID = 1028507357)

    RMAN > run {}

    2 > ALTER DATABASE connectable to MultiPlay NARROW;

    3 > set up what SNA = 14899894;

    4 > restore database pluggable MultiPlay;

    5 > restore database pluggable MultiPlay;

    6 > alter database connectable to MultiPlay open resetlogs;

    {7 >}

    using the control file of the target instead of recovery catalog database

    Statement processed

    executing command: SET until clause

    From restoration to 22 August 15

    allocated channel: ORA_DISK_1

    channel ORA_DISK_1: SID = 22 type of device = DISK

    channel ORA_DISK_1: from the restore backup set data file

    channel ORA_DISK_1: specifying datafile (s) to restore from backup set

    channel ORA_DISK_1: restore datafile 00035 to E:\APP\ORADATA\RELIABLE\MULTIPLAY\SYSTEM01. DBF

    channel ORA_DISK_1: restore datafile 00036 to E:\APP\ORADATA\RELIABLE\MULTIPLAY\SYSAUX01. DBF

    channel ORA_DISK_1: restore datafile 00037 to E:\APP\ORADATA\RELIABLE\MULTIPLAY\MULTIPLAY_USERS01. DBF

    channel ORA_DISK_1: backup E:\APP\RECOVERY_AREA\RELIABLE\40A433F530464F3BBF2E7C0B1553ECE8\BACKUPSET\2015_08_22\O1_MF_NNNDF_TAG20150 piece reading

    822T115328_BXJ5C182_. BKP

    channel ORA_DISK_1: piece handle = E:\APP\RECOVERY_AREA\RELIABLE\40A433F530464F3BBF2E7C0B1553ECE8\BACKUPSET\2015_08_22\O1_MF_NNNDF_TAG20150822T115328_BX

    J5C182_. Tag BKP = TAG20150822T115328

    channel ORA_DISK_1: restored the backup part 1

    channel ORA_DISK_1: restore complete, duration: 00:01:05

    Restoration finished in 22 August 15

    From pick up to August 22, 15

    using channel ORA_DISK_1

    RMAN-05026: WARNING: following tablespaces set applies to the Point in time specified assuming

    List of tablespaces should have UNDO segments

    SYSTEM tablespace

    Tablespace UNDOTBS1

    Creation of automatic instance with SID = "zbnt."

    RMAN-00571: ===========================================================

    RMAN-00569: = ERROR MESSAGE STACK FOLLOWS =.

    RMAN-00571: ===========================================================

    RMAN-03002: failure of the command recover at 22/08/2015 12:24:56

    ORA-19852: error creating services for the instance auxiliary zbnt (error 0)

    ORA-27302: failure occurred:

    ORA-27303: additional information: could not start the instance

    Thank you

    Siradji

    Just a wild guess

    You do not use this script in a command as Administrator window

    Therefore, you won't be able to create any service.

    I also have a question: your only troubleshooting activity consisted of hitting CTRL + C and CTRL-V, is not it?

    It's so boring to most of the people here do not even TRY to solve their own problems.

    -------------

    Sybrand Bakker

    Senior Oracle DBA

  • Moving oracle data file while the database is broken down

    I have a database in montage mode while I restore the last of my data files through rman, which won't be finished for a day and a half, and then there's a mountain of archivelogs to apply. I need to pass the data to another disk files so that I have room to restore this file. I looked around a few instructions, but they are all different. Many of them tell me to stop first of all, the database is not a good option with the restoration goes. More say to use the operating system to copy the file, but a site ignored that. It is even said that I can't log out of my session between saying oracle sql to move the file and open the database. All I get is a list of steps that all take my database work.

    Currently, I copy the file to the new destination of Linux, and I'm confused about what my options are and how to get the data file of movement before Oracle is running out of space for my restoration, without doing anything to make things worse. I guess I can just wait for the copy at the end and then run something like "alter rename file ' / old/my_file.dbf ' to ' / new/my_file.dbf '; If I have to cancel the copy for a reason, it is not a big deal, as long as I don't have the space in time.

    A quick look, it seems that you should be able to move the data file and to a change of name as you have suggested, as long as you are in editing mode. I hope you have good backup (or an easier way to recover) in case it does not work.

    MOS Doc 115424.1 - how to rename or move data files and log files

    HTH
    Srini

  • How to use oracle APEX remote with a database located on a remote server

    Hello

    I am a new user with the Oracle Technologies. I installed ORACLE 10 g XE and APEX upgrade version 4.0 on my laptop.
    My Manager has setup a remote server with a Linux Red Hat. He has installed on this server EBS R12 with a 11i database.

    As a training, I would like to play with the tables on the server with APEX 4.0 installed on my laptop.
    The server is on the same local network and as a result of the description above, it is possible to connect my local APEX in the database?

    If this is not the case, this means that I have to install APEX on the 11i Server database, and then connect to the server to play with these tables?

    Thank you very much for your time and support the members of the CCSE.

    Kind regards

    Ulrich

    Hi, Ulrich,.

    Yes, you can access another database from your database to the APEX. The mechanism used for this is called a "database connection". You create such a link of database in your local database, then you can access remote data like this:

    SELECT *
      FROM databaselink@EMP
    ;
    

    If you want to use remote objects in APEX you create local views to encapsulate this (search the forum for apex and database links, you will find some interesting discussions).

    brgds,
    Peter

    -----
    Blog: http://www.oracle-and-apex.com
    ApexLib: http://apexlib.oracleapex.info
    BuilderPlugin: http://builderplugin.oracleapex.info
    Work: http://www.click-click.at

  • Differences between Data Pump and always by inheritance, import and export

    Hi all

    I work as a junior in my organization dba and I saw that my organization still uses legacy import and export utility instead of Oracle Data Pump.

    I want to convence my manager to change the existing deal with Oracle Data Pump, I have a meeting with them to keep my points and convence them for Oracle utility pump data.

    I have a week very convencing power but I won't miss to discover myself, I can't work myself and really a need with differences of strength against import and export, it would be really appreciated if someone can put strong points against Oracle Data pump on legacy import and export.

    Thank you

    Cabbage

    Hello

    a other people have already said the main advantage of datapump is performance - it is not just a little more fast exp/imp it is massively faster (especially when combined with parallel).

    It is also more flexible (once much more) - it will even create users with exports level schema which imp can't do for you (and it is very annoying that he could never).

    It is reusable

    It has an api plsql

    It supports all types of objects and new features (exp is not - and that alone is probably reason to spend)

    There even a 'legacy' at 11.2 mode where most of your old exp parameter file will still work with him - just change exp expdp and impdp print.

    The main obstacle to the transition to datapump seems to be all "what do you mean I have to create a directory so that it works", well who and where is my dumpfile why can't it be on my local machine. These are minor things to go well beyond.

    I suggest do you some sort of demo with real data of one of your large databases - do a full exp and a full expdp with parallel and show them the runtimes for them to compare...

    See you soon,.

    Rich

  • Moving database from 1 server to another via Data Pump - some queries

    Hello

    I am planing to move my database from one windows server to another. Part of the obligation is also to update this database 10g to 11.2.0.3. So I'm combining 2 tasks using the export / import method (via Data Pump) upgrade.

    Regarding export / import (which will be a pump full data of the database export) my plan is.

    create an empty 11.2.0.3 target database on the server (same number of control files, and redo logs etc.) ready for an import of the source database
    Q1. This will create tablespaces SYSTEM and UNDO - I presume the datapump export doesn't include these spaces of storage anyway?

    For export, I intend to simulate CONSISTENT = Y using FLASHBACK_TIME = SYSTIMESTAMP
    Q2. What is the return of flame characteristics must be active on the source database to use this knowledge should I Flashback Database enabled on the source (as opposed to other flashback features) database?

    My target is a virtual server with a single virtual processor
    Q3. Is there any point PARALLEL usinng in the import settings file (normally I would fix this number of processors - however in the case of this virtual server, it is actually onely a virtual processor)?

    For the import, I intend to use the REMAP_DATAFILE option to change the location of the data files on the target server

    Q4. If the import fails before the end, what is the best course of action? for example I just let go of data storage spaces and remake a complete import?

    Thank you
    Jim

    Jim,

    I'll take a pass on your questions:

    create an empty 11.2.0.3 target database on the server (same number of control files, and redo logs etc.) ready for an import > source database
    Q1. This will create tablespaces SYSTEM and UNDO - I presume the datapump export does not include these storage spaces > anyway?

    The system tablespace is created when you create a database, but the Data Pump will export and try to import. It will fail with tablespace exists. I am sure that the undo tablespace will be also exported and imported. If they are there, then just import will report that they already exist.

    For export, I intend to simulate CONSISTENT = Y using FLASHBACK_TIME = SYSTIMESTAMP
    Q2. What is the return of flame characteristics must be active on the source database to use this knowledge should I Flashback Database active > on the source (as opposed to other flashback features) database?

    I know not true about it. I thought that you need just enough cancel, but I hope that others will run in.

    My target is a virtual server with a single virtual processor
    Q3. Y at - it no PARALLEL point usinng in the import settings file (normally I put this on the number of processors - > however in the case of this virtual server, it is actually onely a virtual processor)?

    We recommend usually 2 times the number of processes, so 2 parallel should be ok.

    For the import, I intend to use the REMAP_DATAFILE option to change the location of the data files on the target server

    Q4. If the import fails before the end, what is the best course of action? that is, do I just give up storage of data and redo a > full import?

    It depends what is failure. Most of the failures will not stop work, but if this is the case, then most of these jobs may simply be restarted. To restart a job, you just need to know the name of the task, which is printed as you start to export/import, or you name the task in your order Data Pump. To restart, follow these steps:

    password/user Impdp attach = job_name

    If you do not name the work, the name of the job will be something like

    User.sys_import_full_01

    Hope that helps - and good luck with your migration.

    Dean

Maybe you are looking for

  • Satellite L40 - nothing happens if I want to start it

    I have a Toshiba Satelite L40 - 12z. If I want to start the laptop, nothing happens just the hd activity led flashes very quickly. What can be the problem?

  • E reader vs. software Adobe digital editions...

    Sinc does not work. Both are approved for me. How can I down 11 books of my software Adobe digital Editions to my reader (PRS-350)?

  • How to implement point filters fixed for FPGA

    Hello! I am applying the bandpass filter Butterworth for FPGA. I managed to develop filters of fixed point that suits my needs, but I don't know how to implement on a table in a project after the FPGA code generation. I tried to understand it in the

  • 2003 Server freezes / crashes

    I have a HP Proliant ML310 running Microsoft Server 2003 which will be freeze/hang up randomly, usually after several hours running (3 to 5 hours). When this happens, it brings my entire network with a brake.  Initially, the server seems to be upward

  • problem with the window update

    After putting to date of windows 8 to 8.1, my pc window has been activated to turn on. I own a laptop of pavilion g6 C5T69EA #BH5, when this was the case a message appeared on my screen saying that my pc has a problem. Then my previous window has bee