Table view is different in expdp/impdp?

11.2 / Linux

With exp/imp original: OWNER, TABLES and FULL modes are mutually exclusive. IE in original exp, if you want to make a backup of the EMP table in the scott schema, the following will be error
exp / tables= EMP owner=SCOTT file=selected_tables.dmp log=selected_tables.log

But this will end up in
EXP-00026: conflicting modes specified
EXP-00000: Export terminated unsuccessfully
You must prefix the name of the owner (tables = < owner >. < table_name >) as below for the prescription above
exp / tables= SCOTT.EMP file=selected_tables.dmp log=selected_tables.log
But in impdp things work differently

I provided a dump expdp of REG_TRACK_EVT of a CLS_USR_PROD scheme. I wanted to import that table in the schema of development CLS_USR_DEV. So, I did. But launch error ORA - 39166 TI kept as below
$ impdp userid=\'/ as sysdba\' DIRECTORY=DPUMP_DIR DUMPFILE= reg_track_evt.dmp LOGFILE=reg_track_evt-imp.log tables=CLS_USR_DEV.REG_TRACK_EVT

Import: Release 11.2.0.1.0 - Production on Mon Jul 30 18:24:32 2012

Copyright (c) 1982, 2009, Oracle and/or its affiliates.  All rights reserved.

Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
With the Partitioning, OLAP, Data Mining and Real Application Testing options
ORA-39002: invalid operation
ORA-39166: Object CLS_USR_DEV.REG_TRACK_EVT was not found.
To solve this problem, I had to remove the parameter TABLES (!) and use the REMAP_SCHEMA parameter
impdp userid=\'/ as sysdba\' DIRECTORY=DPUMP_DIR DUMPFILE= reg_track_evt.dmp LOGFILE=reg_track_evt-imp.log REMAP_SCHEMA=CLS_USR_PROD:CLS_USR_DEV
If this is the case, then what's the point of having the TABLES parameter in impdp?

It's really the same thing regarding conflicting modes. Of what you have used in your import command:

Impdp userid =------"/ ACE sysdba\ ' DIRECTORY = DUMPFILE DPUMP_DIR = LOGFILE reg_track_evt.dmp = reg_track_evt - imp.log REMAP_SCHEMA = CLS_USR_PROD:CLS_USR_DEV

If you are remapping CLS_USR_PROD:CLS_USR_DEV then the table in the dumpfile is called CLS_USR_PROD. REG_TRACK_EVT and you tried to import CLS_USR_DEV. REG_TRACK_EVT and this table did not exist.

What you specify on the tables = command is the table in the desired dumpfile. What you specify in the remapping of the color settings is where you want the table created, therefore, your first order was close, you need this:

Impdp userid =------"/ ACE sysdba\ ' DIRECTORY = DUMPFILE DPUMP_DIR = LOGFILE reg_track_evt.dmp = reg_track_evt - imp.log tables = CLS_USR_PROD. REG_TRACK_EVT REMAP_SCHEMA = CLS_USR_PROD:CLS_USR_DEV

I hope this helps.

Dean

Tags: Database

Similar Questions

  • expdp/impdp method in the database upgrade from 11.2.0.4 to 12.1.0.2 an error in logic of "command line"?

    Hi DBA:

    We are evaluating different methods of upgrading 11 GR 2 DB to 12 c (12.1.0.2) this year, it is very important for us to use the right method.  It was reported that when DB version was 9i and has been updated to 11g, there were some order of tables 'line' was wrong when execute SELECT statements in application (Java application) - at the moment I didn't if it is ' select * from... "or any specific part with a"where"clause, and what type of data was wrong (maybe sequence CLOB data, etc.).

    I think I know, ' exp/imp', "expdp/impdp" logical backup of DB methods.  Therefore, Oracle automatically import data and fix it in each table.  If no error in impdp connects after it's over, should not be any mistake, right?

    If the use of the method of 'creates a new data + expdp/impdp basis for schemas user' 11g 12 c port data, questions:

    1. this method will lead to erroneous sequence data when ordering?  If so, how to re - generate the sequence number?

    2. this method can bring CLOB, BLOB data and without any errors?  If error, how to detect and repair?

    I use the same method in the 10.2.0.5 to 11g port data, no problem.

    I'll keep this thread posted after getting more information from application.

    Thank you very much!

    Dania

    INSERT... SELECT, CREATE TABLE in the SELECT and Export-Import will all "re-organize" the table by assigning new blocks.

    In place--update (using DBUA or command line) don't re-create the user tables (although it DOES not rebuild the data dictionary tables).  Therefore, does not change the physical location of lines for user tables.  However, there is no guarantee that a further reason for INSERT, DELETE, UPDATE instructions does not change the physical location of ranks.  Nor is it a guarantee that a SELECT statement will return always stored in the same order unless an ORDER BY is specified explicitly.

    If you export the data while it is still used, you don't avoid update sequences in the source database.  Usually results in a mismatching number sequence between the source database and the new database (imported). In fact, even the numbers may be different, unless you use CONSISTENT or FLASHBACK_SCN.  (These two options in exp / expdp do NOT prevent update sequences).   The right way to manage sequences is to rebuild or increase them in data imported with the values in the data source, once the import is complete.

    Hemant K Collette

  • problem in expdp/impdp

    Hi, I'm currently expdp/impdp but every time the same mistakes. OS win Server 03 on ORACLE 10 g.

    errors:

    C:\Documents and Settings\Administrateur > impdp rΘpertoire "/ as sysdba" = "data_pump_dir" dumpfile = logfile = scott.log'scott.dmp ' '

    Import: Release 10.2.0.3.0 - Production on Tuesday, March 25, 2014 10:43:18

    Copyright (c) 2003, 2005, Oracle.  All rights reserved.

    Connected to: Oracle Database 10 g Enterprise Edition Release 10.2.0.3.0 - Production

    With partitioning, OLAP and Data Mining options

    ORA-31626: there is no job

    ORA-04063: package body 'SYS. DBMS_INTERNAL_LOGSTDBY"contains errors

    ORA-06508: PL/SQL: called program unit is not found: 'SYS. "DBMS_INTERNAL_LOGSTDBY '

    ORA-06512: at "SYS." "KUPV$ FT", line 834

    ORA-04063: package body 'SYS. DBMS_INTERNAL_LOGSTDBY"contains errors

    ORA-06508: PL/SQL: called program unit is not found: 'SYS. DBMS_INTERNAL_LOGSTDBY ".

    SQL > alter the body of package DBMS_INTERNAL_LOGSTDBY compilation;

    WARNING: The bodies of Package modified with compilation errors.

    SQL > show error

    PACKAGE BODY DBMS_INTERNAL_LOGSTDBY errors:

    LINE/COL ERROR

    -------- -----------------------------------------------------------------

    2563/4 PL/SQL: statement ignored

    2563/20 PL/SQL: ORA-00942: table or view does not exist

    2625/3 PL/SQL: statement ignored

    2625/19 PL/SQL: ORA-00942: table or view does not exist

    2809/2 PL/SQL: statement ignored

    2809/13 PL/SQL: ORA-00942: table or view does not exist

    HELP ME PLEASE

    Since you're using Oracle 10 g version...

    Check if you have columns that contain "TEMESTAMP (6) ' data by using this query type:"

    Select dba_tab_columns owner, table_name, column_name data_type where column_name = 'TIMESTAMP' and data_type like '% TIMESTAMP ';

    If the previous query lists for you a list of tables, you must do the following for each table:

    ALTER table system.table_name change (date timestamp);

    Then recompile the package "DBMS_INTERNAL_LOGSTDBY" again.

    I hope this helps.

    Kind regards

  • With the help of expdp/impdp in pump data

    Hi all

    I am newbie in oracle database through the use of data import and export oracle data pump. I use Oracle 10 g 2 on Windows 7

    After you create the directory object 'test_dir' and the granting of read, write privilege to user scott, I connect like scott to the database and create the table 'test' with two rows.

    Then I run expdp command prompt as follows:

    C:\users\administrator > expdp scott/tiger@orcl tables = happy test = all = TEST_DIR dumpfile = test.dmp expdp_test.log = logfile directory

    Export: Release 10.2.0.3.0 - Production on Monday, June 13, 2011 20:20:54

    Copyright (c) 2003, 2005, Oracle. All rights reserved.

    Connected to: Oracle Database 10 g Enterprise Edition release 10.2.0.3.0 - production
    tion
    With partitioning, OLAP and Data Mining options
    Departure 'SCOTT '. "SYS_EXPORT_TABLE_01": scott/***@orcl tables = test content.
    = all = TEST_DIR dumpfile = test.dmp expdp_test.log = logfile directory
    Current estimation using BLOCKS method...
    Processing object type TABLE_EXPORT/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 0 KB
    Object type TABLE_EXPORT/TABLE/TABLE processing
    . . "exported"SCOTT" TEST' 0Ko 0 rows
    Table main 'SCOTT '. "" SYS_EXPORT_TABLE_01 "properly load/unloaded
    ******************************************************************************
    Empty the files together for SCOTT. SYS_EXPORT_TABLE_01 is:
    D:\DAMP\TEST. DMP
    Job 'SCOTT '. "" SYS_EXPORT_TABLE_01 "conducted at 20:21:02

    My question is why data pump seem to export the table 'test' without lines (that is to say the line: exported 'SCOTT'.) ("' TEST ' 0Ko 0 lines)? How can I do it with the associated export lines?


    I dropped the table test, then I ran the command impdp as follows:

    C:\users\administrator > impdp scott/tiger tables = content test = all = TEST_DIR dumpfile = Test.dmp impdp_test.log = logfile directory

    Import: Release 10.2.0.3.0 - Production on Monday, June 13, 2011 20:23:18

    Copyright (c) 2003, 2005, Oracle. All rights reserved.

    Connected to: Oracle Database 10 g Enterprise Edition release 10.2.0.3.0 - production
    tion
    With partitioning, OLAP and Data Mining options
    Table main 'SCOTT '. "' SYS_IMPORT_TABLE_01 ' properly load/unloaded
    Departure 'SCOTT '. "" SYS_IMPORT_TABLE_01 ": scott / * tables = happy test = all
    Directory = TEST_DIR dumpfile = Test.dmp logfile = impdp_test.log
    Object type TABLE_EXPORT/TABLE/TABLE processing
    Processing object type TABLE_EXPORT/TABLE/TABLE_DATA
    . . imported 'SCOTT '. "' TEST ' 0Ko 0 rows
    Job 'SCOTT '. "" SYS_IMPORT_TABLE_01 "carried out at 20:23:21


    Then, after selection * test. No rows returned

    Please someone with an idea to this topic... What I expected after operation data in table view pump it could export and import table with the data if I'm not mistaken.

    Concerning
    Sadik

    Sadik says:
    He had two rows

    Export is disagreeing with you.
    have you COMMITTED after INSERT & before export?

  • fromuser/touser equivanet in expdp/impdp?

    Hello.
    I got a dump of the following cmmand file.
    expdp dumpfile = logfile = studentinfo.log studentinfo.dmp system/123456@SCHDB tables = school.studentmaster, school.studentmarks, school.studentleave directory is mydir


    now I want to import these 2 tables in a different schema (old_school) using impdp

    What is the fromuser / equivalent touser in impdp/expdp?

    REMAP_DATA
    Specify a data conversion function.
    For example, REMAP_DATA = EMP. EMPNO:REMAPPKG. EMPNO.

    REMAP_DATAFILE
    Redefine the references of data file in all DDL statements.

    REMAP_SCHEMA
    A schema objects are loaded in another schema.

    REMAP_TABLE
    Table names are remapped to another table.
    For example, REMAP_TABLE = h. EMPLOYEES: EMPS.

    REMAP_TABLESPACE
    Tablespace objects are remapped to a different tablespace.

    REUSE_DATAFILES
    Tablespace will be initialized if it exists already [N].

  • How can I import tables to a different schema in the relational model existing... to add these tables in the existing model? PLSS help

    How can I import tables from a different schema in the relational model... to add these tables in the existing relational/logic model? PLSS help

    Notes; I already have a schema ready relational/logic model... and I need to add more tables to this relational/logic model

    can I import the same way like I did before?

    But even if I do the same how can I add it in the template? as the logic model has been designed...

    Help, please...

    Thank you

    To view the diagram of logic model in Bachman notation, you right-click on a blank area of the diagram and select the rating > Bachman Notation.

    David

  • expdp, impdp - redo generation?

    Friends...

    Oracle 11 g 2 SE

    OS: Linux

    I had tried to search in the documentation, but could not find an answer.

    I'm dropping some tables but prior to export for backup, db is in archivelog mode.

    Before you delete tables, trying to find if expdp job generates a lot of repeat or not given that the file system have 20 GB of free space.

    Issue.

    1. If I'm the scene expdp some tables with size 80 GB (size.. indexes on table not included), does expdp will generate a similar amount of redo or it will generate a lot of repeat?

    2 impdp job also generate a lot of repeat or only during the creation of index in impdp?

    Thank you

    Mike

    Export do NOT generate all over again. It's just dumping data in binary files. Very little again just to keep the export data table, rocking but very very minimal. IF I had to guess not more than 1 MB.

    Import generates recovery logs. Imports perform "Insert" statement behind the scenes and it does not generate a lot of newspapers of recovery. You can use sqlfile to create the definition of the index and to run manually with NOLOGGING which will reduce some redo. something like

    exclude the user/password Impdp = index - this will import everything except the index.

    Impdp username/password include sqlfile = index.sql index = - this will create a file named index.sql that will contain all the instructions for the index create

  • expdp/impdp: constraints in the child-Parent relationship

    Hello

    I have a table parent1 and child1, child2 and chld3 tables have foreign key created on this parent1.

    Now, I want to do a delete on parent1. But as the number of records is very high on parent1, we go with expdp / impdp with option querry.

    I took the expdp parent1 level query. Now I dropped parent1 with option of cascade constraints and all foreign keys created by child1, 2 and 3 that parent1 references are automatically deleted.

    Now, if I have the impdp to the query of fire level dump file, are these foreign key constraints will be created automatically on child1, 2 and 3 or I need to manually recreate it?

    Kind regards

    ANU

    Hello
    The CF will not be in the dumpfile - see the code example below where I generate a sqlfile following pretty much the process that you would have done. This is because FK belongs to the DDL for the child table not the parent.

    Connected to:
    Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    With the Partitioning option
    
    OPS$ORACLE@EMZA3>create table a (col1 number);
    
    Table created.
    
    OPS$ORACLE@EMZA3>alter table a add primary key (col1);
    
    Table altered.
    
    OPS$ORACLE@EMZA3>create table b (col1 number);
    
    Table created.
    
    OPS$ORACLE@EMZA3>alter table b add constraint x foreign key (col1) references a(col1);
    
    Table altered.
    
    OPS$ORACLE@EMZA3>
    
    EMZA3:[/oracle/11.2.0.1.2.DB/bin]# expdp / include=TABLE:\"=\'A\'\"
    
    Export: Release 11.2.0.3.0 - Production on Fri May 17 15:45:50 2013
    
    Copyright (c) 1982, 2011, Oracle and/or its affiliates.  All rights reserved.
    
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    With the Partitioning option
    Starting "OPS$ORACLE"."SYS_EXPORT_SCHEMA_04":  /******** include=TABLE:"='A'"
    Estimate in progress using BLOCKS method...
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 0 KB
    Processing object type SCHEMA_EXPORT/TABLE/TABLE
    Processing object type SCHEMA_EXPORT/TABLE/CONSTRAINT/CONSTRAINT
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    . . exported "OPS$ORACLE"."A"                                0 KB       0 rows
    Master table "OPS$ORACLE"."SYS_EXPORT_SCHEMA_04" successfully loaded/unloaded
    ******************************************************************************
    Dump file set for OPS$ORACLE.SYS_EXPORT_SCHEMA_04 is:
      /oracle/11.2.0.3.0.DB/rdbms/log/expdat.dmp
    Job "OPS$ORACLE"."SYS_EXPORT_SCHEMA_04" successfully completed at 15:45:58
    
    Import: Release 11.2.0.3.0 - Production on Fri May 17 15:46:16 2013
    
    Copyright (c) 1982, 2011, Oracle and/or its affiliates.  All rights reserved.
    
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    With the Partitioning option
    Master table "OPS$ORACLE"."SYS_SQL_FILE_FULL_01" successfully loaded/unloaded
    Starting "OPS$ORACLE"."SYS_SQL_FILE_FULL_01":  /******** sqlfile=a.sql
    Processing object type SCHEMA_EXPORT/TABLE/TABLE
    Processing object type SCHEMA_EXPORT/TABLE/CONSTRAINT/CONSTRAINT
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    Job "OPS$ORACLE"."SYS_SQL_FILE_FULL_01" successfully completed at 15:46:17
    
    -- CONNECT OPS$ORACLE
    ALTER SESSION SET EVENTS '10150 TRACE NAME CONTEXT FOREVER, LEVEL 1';
    ALTER SESSION SET EVENTS '10904 TRACE NAME CONTEXT FOREVER, LEVEL 1';
    ALTER SESSION SET EVENTS '25475 TRACE NAME CONTEXT FOREVER, LEVEL 1';
    ALTER SESSION SET EVENTS '10407 TRACE NAME CONTEXT FOREVER, LEVEL 1';
    ALTER SESSION SET EVENTS '10851 TRACE NAME CONTEXT FOREVER, LEVEL 1';
    ALTER SESSION SET EVENTS '22830 TRACE NAME CONTEXT FOREVER, LEVEL 192 ';
    -- new object type path: SCHEMA_EXPORT/TABLE/TABLE
    CREATE TABLE "OPS$ORACLE"."A"
       (    "COL1" NUMBER
       ) SEGMENT CREATION IMMEDIATE
      PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255
     NOCOMPRESS LOGGING
      STORAGE(INITIAL 16384 NEXT 16384 MINEXTENTS 1 MAXEXTENTS 505
      PCTINCREASE 50 FREELISTS 1 FREELIST GROUPS 1
      BUFFER_POOL DEFAULT FLASH_CACHE DEFAULT CELL_FLASH_CACHE DEFAULT)
      TABLESPACE "SYSTEM" ;
    -- new object type path: SCHEMA_EXPORT/TABLE/CONSTRAINT/CONSTRAINT
    ALTER TABLE "OPS$ORACLE"."A" ADD PRIMARY KEY ("COL1")
      USING INDEX PCTFREE 10 INITRANS 2 MAXTRANS 255
      STORAGE(INITIAL 16384 NEXT 16384 MINEXTENTS 1 MAXEXTENTS 505
      PCTINCREASE 50 FREELISTS 1 FREELIST GROUPS 1
      BUFFER_POOL DEFAULT FLASH_CACHE DEFAULT CELL_FLASH_CACHE DEFAULT)
      TABLESPACE "SYSTEM"  ENABLE;
    -- new object type path: SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    DECLARE I_N VARCHAR2(60);
      I_O VARCHAR2(60);
      NV VARCHAR2(1);
      c DBMS_METADATA.T_VAR_COLL;
      df varchar2(21) := 'YYYY-MM-DD:HH24:MI:SS';
     stmt varchar2(300) := ' INSERT INTO "SYS"."IMPDP_STATS" (type,version,flags,c1,c2,c3,c5,n1,n2,n3,n4,n5,n6,n7,n8,n9,n10,n11,n12,d1,cl1) VALUES (''I'',6,:1,:2,:3,:4,:5,:6,:7,:8,:9,:10,:11,:12,:13,NULL,:14,:15,NULL,:16,:17)';
    BEGIN
      DELETE FROM "SYS"."IMPDP_STATS";
      c(1) := 'COL1';
    
      DBMS_METADATA.GET_STAT_INDNAME('OPS$ORACLE','A',c,1,i_o,i_n);
      EXECUTE IMMEDIATE stmt USING 0,I_N,NV,NV,I_O,0,0,0,0,0,0,0,0,NV,NV,TO_DATE('2013-05-17 15:43:24',df),NV;
    
      DBMS_STATS.IMPORT_INDEX_STATS('"' || i_o || '"','"' || i_n || '"',NULL,'"IMPDP_STATS"',NULL,'"SYS"');
      DELETE FROM "SYS"."IMPDP_STATS";
    END;
    /
    

    Kind regards
    Harry

    http://dbaharrison.blogspot.com/

  • Join of two tables in two different schemas

    Hi all

    I have an obligation to join two tables on two different schemas. How to join these two tables in the object view.

    Thanks in advance.

    Concerning
    Kaushik Guillaumin

    You can do just that using schema name in the +' view object.table name ' + according to the query object and also grant select the another schema to this schema user

    ex
    you need schem is test and another Act you need to a view object based on the test to join a table on shcema Act

    you write semply object sql view code:

    Act.table name

    and you can also give him select statement on Bill schem table to test

    concerning

  • IMP/exp expdp/impdp vs

    Hi all;
    which is better in performance the traditional exp/imp or expdp/impdp?

    Hello

    Oracle Data Pump is a more recent, rapid and flexible alternative to utilities 'exp' and 'imp' used in previous versions of Oracle. In addition to export and import of basic feature data pump provides a PL/SQL API and support for external tables

    If Oracle is recommended for all 10g databases and more use of Data Pump and pre 10g use old exp/imp :)

  • [OBIEE 11 g] Create customized to group columns in a table view column header

    Hello world

    I got a requirement to group certain columns in the table view with a custom column header name (for example to create a two-level header).

    The layout I have now:

    [Jan]     [February]     [Mar]     [Apr]     [May]     [June]     [Jul]     [August]     [Total sales]

    -----------------------------------------------------------------------------------------------------------------------

    1          2          3            4          5           6           7          8           36

    The layout I want to achieve:

    [                  Q1                    ] [                   Q2                   ]    

    [Jan]     [February]     [Mar]     [Apr]     [May]     [June]     [Jul]     [August]     [Total sales]

    -----------------------------------------------------------------------------------------------------------------------

    1          2          3            4          5           6           7          8           36

    I use Table View, is it possible to get the above provision? Any comments appreciated. Thank you.

    See you soon,.

    Kayip

    Hello

    It is a simple parameter, it is unpacking already...

    The property "of the topic folder" of your columns allows you to enter the required text.

    PS: Do you really have columns in your field with the name of the month? Is it not something better managed with a hierarchy of real time and using a pivot to display the month columns? In this case the grouping of quarters is already there...

  • Complete export dashboard in excel gives error generate document in the table view of subtotals

    Hi all

    We have dashboard with several tabs and you try to export the whole dashboard in excel.

    Excel file is generated with all tabs on the dashboard as the leaves, but few of these leaves are empty with error message of document generation in the first cell.

    After further analysis, we concluded that the cause is seen with subtotals on table. PivotTables with subtotals are generated correctly.

    Workaround obvious to re-create the table views in the articulation points is not possible due to a large number of reports.

    We run Oracle Business Intelligence Product Version 11.1.1.7.150120 (Build 150113.1200 64-bit)

    Any ideas or solutions to this problem would be appreciated.

    Thank you;

    Klaudio

    We just found guilty

    Lines per page to display in the table must be 999999999 max which is more than enough for our needs.

  • Line display only 1 in a table view

    Pls... is there a way to display only 1 row in a table view (OBIEE 11 g)?

    (I know it's possible in the narrative view...)

    TXS

    A.L.

    Try with pagination option in a table view. You can display a single default line.

    Thank you

    Prasanna

  • Please can someone tell if google chrome development tool that emulates different devices provide a view Specifies how the site will be viewed on different devices

    Please can someone tell if google chrome development tool that emulates different devices provide a view Specifies how the site will be viewed on different devices

    All the emulators fail at something, so it is always best to test the real devices when you can.

    That said, Chrome and Firefox are quite close. One of the things bigger than don't miss the emulators in the browser is the function of auto zoom of smaller devices. Be sure to include the viewport with an initial scale of 1.0 and you should be fine...

    Generally emulators you what a phone with auto zoom see when this tag is missing. Some phone browsers also automatically link and add styles to things like phone numbers or dates, if there is no style associated with them are coming from css to your site. Dev tools will miss that also and sometimes it's a little surprising to see a bunch of links on your page when you never coded in a first time.

    Again, it is always better to use actual devices, you intend to support, to see what will actually make your site.

  • Accelerate exp/imp or expdp/impdp

    Hello

    Is it possible to speed up and exp/imp or expdp/impdp that already works? Is is possible to speed up a run RMAN backup or restore RMAN process?

    Kind regards

    007

    To accelerate a datapump export market and import you can attach to employment and increase the level of parallelism... impdp attach = job

    I don't know any way to speed up a running RMAN backup.

    To expedite an RMAN restore, you can kill the restoration and re - run using several channels.  The restoration should take up where he left off and can run faster with many channels.  It is relevant only if you have several items from backup.

Maybe you are looking for