data file header = tablespace, it belongs to

Hi guys,.

>
If storage space is read-only headers in the data file are not changed to reflect the name change and a message is written to the log of alerts to warn you of this fact. The impact on the recovery, it's that the tablespace will be recovered at its old name if the controlfile is re-created and the data files containing the old headers are used.
>

is there a way to check if the data header here mention the new name of the tablespace of score / existing name?

Published by: flaskvacuum on 5 March 2012 21:39

So that means you want to know the current name of tablespace_name file... so it is in dba_data_files view an example: 11.2.0.1

SQL> create tablespace test datafile 'c:\test.dbf' size 1m;

Tablespace created.

SQL> select file#,status,tablespace_name from v$datafile_header;

     FILE# STATUS  TABLESPACE_NAME
---------- ------- ------------------------------
         1 ONLINE  SYSTEM
         2 ONLINE  SYSAUX
         3 ONLINE  UNDOTBS1
         4 ONLINE  USERS
         5 ONLINE  EXAMPLE
         6 ONLINE  TEST

6 rows selected.

SQL> alter tablespace test read only;

Tablespace altered.

SQL> alter tablespace test rename to newtest;

Tablespace altered.

SQL> select file#,status,tablespace_name from v$datafile_header;

     FILE# STATUS  TABLESPACE_NAME
---------- ------- ------------------------------
         1 ONLINE  SYSTEM
         2 ONLINE  SYSAUX
         3 ONLINE  UNDOTBS1
         4 ONLINE  USERS
         5 ONLINE  EXAMPLE
         6 ONLINE  TEST

6 rows selected.

SQL> desc dba_data_files;
 Name                                      Null?    Type
 ----------------------------------------- -------- ----------------------------
 FILE_NAME                                          VARCHAR2(513)
 FILE_ID                                            NUMBER
 TABLESPACE_NAME                                    VARCHAR2(30)
 BYTES                                              NUMBER
 BLOCKS                                             NUMBER
 STATUS                                             VARCHAR2(9)
 RELATIVE_FNO                                       NUMBER
 AUTOEXTENSIBLE                                     VARCHAR2(3)
 MAXBYTES                                           NUMBER
 MAXBLOCKS                                          NUMBER
 INCREMENT_BY                                       NUMBER
 USER_BYTES                                         NUMBER
 USER_BLOCKS                                        NUMBER
 ONLINE_STATUS                                      VARCHAR2(7)

SQL> select file_id,tablespace_name from dba_data_files;

   FILE_ID TABLESPACE_NAME
---------- ------------------------------
         4 USERS
         3 UNDOTBS1
         2 SYSAUX
         1 SYSTEM
         5 EXAMPLE
         6 NEWTEST

6 rows selected.

SQL> alter tablespace test read write;
alter tablespace test read write
*
ERROR at line 1:
ORA-00959: tablespace 'TEST' does not exist

SQL> alter tablespace newtest read write;

Tablespace altered.

SQL> select file#,status,tablespace_name from v$datafile_header;

     FILE# STATUS  TABLESPACE_NAME
---------- ------- ------------------------------
         1 ONLINE  SYSTEM
         2 ONLINE  SYSAUX
         3 ONLINE  UNDOTBS1
         4 ONLINE  USERS
         5 ONLINE  EXAMPLE
         6 ONLINE  NEWTEST

6 rows selected.

SQL>

Concerning
Girish Sharma

Tags: Database

Similar Questions

  • Data in the data file header

    RDBMS version: 10.2.0.4.0, 11.2.0.2

    During the backup hot (RMAN and user managed backup), data can be written and read from data files. But its headers are frozen. Then, where will be data for the headers of data stored up to that hot backup is made?

    Why must 'means for the headers of data data' be 'stored' somewhere?

    The header is not updated. Header updates are Checkpoint SCN, LogSequence (which is not visible to us) etc. These updates are applied to the data file when the database tablespace backup is complete. Intermediate of the SCN values need not be "stored". When the ends of backup and a control point is completed, Oracle updates the header with the current SNA Checkpoint.

    Hemant K Collette

  • data file update the header by various processes

    Hi master,

    It is certainly not a question any. I just wanted to clear some doubts in my mind. We all know SNA are generated and updated in the controlfiles and datafile headers by various oracle process.

    System change number are updated by DBWr in the header of the data files and of ckpt in controlfiles... Please correct me if I'm wrong.


    I am little confuse on when the database writer writes dirty blocks in the data files update the header of this particular data file or updated all the header of data file in the tablespace or all the headers of data file?

    and it is, but it is obvious that if the instance crash, then database will be inconsistent state and application of recovery of incompatible data files as YVERT in the control file does not match the data file header...

    treat means only when good checkpoint with CKPT then only the database will be a consistent state after reboot... it means making ckpt YVERT in all the datafile header match with in controlfiles?

    dealing with writing or update YVERT in other than CKPT controlfiles? at the time, only respective datafile is updated with the new change number or all in this tablespace data files are updated?

    Although many change of system is not timed... It may often happen and it does... but how to find change number perfect if we want to recover a file of data or database with incomplete recovery until the sequence or YVERT? I tried to find the list for the smon_scn_time , but don't know how to use...

    I'm looking also for document on this point of view I have Google, but have not found anything successful.


    any suggestions would be much appreciated...

    Thanks and greetings
    VD

    Published by: vikrant dixit on April 2, 2009 12:05 AM

    gurantee Oracle with CKPT SNA who wrote to datafile is completed and will be contradicted with system files, tablespace and other
    tabelspace load datafiles, each tablespace datafiles CKPT SNA is circumscribed in controlfile is why when you take line/hot
    Save then your backup is inconsistent on recovery makes inconsistent backups consistent oracle by applying all archived and online
    redo logs, oracle makes the recovery by reading more early/more former SNA to one of the headers of data file (media recovery) and apply the
    changes of newspapers in the data file.

    After the completion of checkpoint process CKPT process increment header controlfile CKPT SNA and broad cast than SNA CKPT to others
    header data file. This SNA CKPT is not individual for alls datafile within the database or tablespace datafiles, each CKPT SNA data file is the same.
    You can verify this behavior.

    SQL> select substr(name,1,50) fname,checkpoint_change#,last_change#,status
      2    from v$datafile
      3  /
    
    FNAME                                              CHECKPOINT_CHANGE# LAST_CHANGE# STATUS
    -------------------------------------------------- ------------------ ------------ -------
    F:\ORACLE\PRODUCT\10.1.0\ORADATA\PROD\SYSTEM01.DBF             327957              SYSTEM
    F:\ORACLE\PRODUCT\10.1.0\ORADATA\PROD\UNDOTBS01.DB             327957              ONLINE
    F:\ORACLE\PRODUCT\10.1.0\ORADATA\PROD\SYSAUX01.DBF             327957              ONLINE
    F:\ORACLE\PRODUCT\10.1.0\ORADATA\PROD\USERS01.DBF              327957              ONLINE<---
    F:\ORACLE\PRODUCT\10.1.0\ORADATA\PROD\USERS02.DBF              327957              ONLINE<---
    
    SQL> alter system checkpoint
      2  /
    
    System altered.
    
    SQL> select substr(name,1,50) fname,checkpoint_change#,last_change#,status
      2    from v$datafile
      3  /
    
    FNAME                                              CHECKPOINT_CHANGE# LAST_CHANGE# STATUS
    -------------------------------------------------- ------------------ ------------ -------
    F:\ORACLE\PRODUCT\10.1.0\ORADATA\PROD\SYSTEM01.DBF             327960              SYSTEM
    F:\ORACLE\PRODUCT\10.1.0\ORADATA\PROD\UNDOTBS01.DB             327960              ONLINE
    F:\ORACLE\PRODUCT\10.1.0\ORADATA\PROD\SYSAUX01.DBF             327960              ONLINE
    F:\ORACLE\PRODUCT\10.1.0\ORADATA\PROD\USERS01.DBF              327960              ONLINE
    F:\ORACLE\PRODUCT\10.1.0\ORADATA\PROD\USERS02.DBF              327960              ONLINE
    
    SQL> alter system checkpoint
      2  /
    
    System altered.
    
    SQL> select substr(name,1,50) fname,checkpoint_change#,last_change#,status
      2    from v$datafile
      3  /
    
    FNAME                                              CHECKPOINT_CHANGE# LAST_CHANGE# STATUS
    -------------------------------------------------- ------------------ ------------ -------
    F:\ORACLE\PRODUCT\10.1.0\ORADATA\PROD\SYSTEM01.DBF             327965              SYSTEM
    F:\ORACLE\PRODUCT\10.1.0\ORADATA\PROD\UNDOTBS01.DB             327965              ONLINE
    F:\ORACLE\PRODUCT\10.1.0\ORADATA\PROD\SYSAUX01.DBF             327965              ONLINE
    F:\ORACLE\PRODUCT\10.1.0\ORADATA\PROD\USERS01.DBF              327965              ONLINE
    F:\ORACLE\PRODUCT\10.1.0\ORADATA\PROD\USERS02.DBF              327965              ONLINE
    

    Khurram

  • Find the time to restore a data file

    RDBMS version: 11.2.0.4

    Platform: Oracle Linux 6.4

    To test our backup RMAN Tape (Netbackup), we have created a table called BACKUPTEST with a single space data file.

    Then, we have removed the data file. Then was restored and recovered with RMAN's data file and tablespace was again online.

    I wanted to show the evidence to my manager that the data file has been restored. But v$ datafile. CREATION_TIME will show the time when the tablespace is imagined. Is there another way, I could find the time when the data file has been restored other than log RMAN.

    Below is an excerpt after the data file has been restored.

    $ sqlplus / as sysdba

    SQL * more: Production of the version 11.2.0.4.0 Fri dec 19 15:08:03 2013

    Copyright (c) 1982, 2013, Oracle.  All rights reserved.

    Connected to:

    Oracle Database 11 g Enterprise Edition Release 11.2.0.4.0 - 64 bit Production

    With partitioning, Real Application Clusters, Automatic Storage Management, OLAP,.

    Options of Data Mining and Real Application Testing

    SQL > select TABLESPACE_NAME, status of dba_tablespaces where nom_tablespace = 'BACKUPTEST ';

    STATUS TABLESPACE_NAME

    ------------------------------ ---------

    BACKUPTEST ONLINE

    SQL > alter session set nls_date_format = "DD-MM-YYYY HH24."

    Modified session.

    SQL > select CREATION_TIME of v$ datafile where FILE # = 78;

    CREATION_TIME

    ----------------

    19/12/2013-12:16

    SQL > select CREATION_TIME, last_time v$ datafile where FILE # = 78;

    CREATION_TIME LAST_TIME

    ---------------- ----------------

    19/12/2013-12:16

    SQL >

    19/12/2013 12:16 is the hour when the tablespace has been created with this data file no time, it has been restored from a backup RMAN.

    Check the alert.log.  There will be messages about the RESTORATION and RECOVERY actions.

    Hemant K Collette

  • data file and size

    I have potentially 400 GB of data to be stored. There are two options that I can think of:

    (1) use several small data files;

    create tablespace TEST
    DataFile ' / oradata/test.dbf ' size 2048M
    Local extended management

    (2) use bigfile tablespace unique.


    PRO for 1) is easy to backup and restore; CON is too many files. For the size of 400 GB, I need about 200 files of data with 2 GB each (?)

    PRO for 2) is the only data file to use. CON, is that the size will be very large. It will make it difficult to backup and restore.

    You please provide a few suggestions or advice?

    Thank you


    Scott

    I understand that with the 8K block by default, can I have a data file as large as 32 G. For the small tablespace with multiple data option files, should I keep the size< 2g="" or=""><>

    Small corrections necessary.

    The maximum size of the single data file or tempfile is 128 tera-bytes (to) for a storage space with blocks of 32 KB and 32 TB for a tablespace with 8 K blocks.

    http://docs.Oracle.com/CD/B19306_01/server.102/b14237/limits002.htm#sthref2833

    As Tom (in the same link) and Mark above say "but over time, things change. Today, in 2006, I wouldn't have a problem with a file more 2 Giga - tools have all taken up to the fact that the files can and receive more. »

    So, just start by petit_fichier tablespace and add data files if necessary. And yet one thing, there is no relationship of big/small file performance given tablespace, because storage space is a logical term of database Oracle. E/s are related with datafile and OS not tablespace.

    If someone is looking for listening to I/O then I think it is doing a scientific approach to the problem rather than seeking performance vs tablespace problems. I'm not finding a link where Sir Thomas kyte said that the same, something like that, same docs are wrong on this issue.

    Concerning
    Girish Sharma

  • ORA-19846: cannot read the header of the data file of the remote site 21

    Hello

    I have a situation or I can say a scenario. It is purely for testing base. Database is on 12.1.0.1 on a Linux box using ASM (OMF).

    Standby is created on another machine with the same platform and who also uses ASM (OMF) and is in phase with the primary. Now, suppose I have create a PDB file on the primary of the SEED and it is created successfully.

    After that is a couple of log, do it again passes to the waiting, but MRP fails because of naming conventions. Agree with that! Now, on the primary, I remove the newly created PDB (coward the PDB newly created). Once again a couple of switches of newspapers which is passed on to the wait. Of course, the wait is always out of sync.

    Now, how to get back my watch in sync with the primary? I can't roll method until the required data (new PDB) file does not exist on the main site as well. I get the following error:

    RMAN > recover database service prim noredo using backupset compressed;

    To go back to November 8, 15

    using the control file of the target instead of recovery catalog database

    allocated channel: ORA_DISK_1

    channel ORA_DISK_1: SID = 70 = device = DISK stby type instance

    RMAN-00571: ===========================================================

    RMAN-00569: = ERROR MESSAGE STACK FOLLOWS =.

    RMAN-00571: ===========================================================

    RMAN-03002: failure of the command recover at the 18:55:32 08/11/2015

    ORA-19846: cannot read the header of the data file of the remote site 21

    The clues on how to I go ahead? Of course, recreating the eve is an option as its only based on test, but I don't want recreation.

    Thank you.

    I tried like below:

    1 a incremental backup of the primary of the CNS where off the eve also taken primary backup controlfile as Eve format.

    2 copy the backup of the watch parts, catalogged them on the day before.

    3 recovered Eve with noredo option - it fails here with the same error pointing to the 21 data file.

    OK, understood. Try not to get back the day before first, rather than restore the controlfile and then perform the restoration.

    Make it like:

    1. take incremental backup of primary SNA, also ensures the backup controlfile format.

    2. copy pending, get the location of the data file (names) by querying v$ datafile on the eve. Restore the controlfile ensures from the backup controlfile you took on primary and mount.

    3. Since you are using OMF, the path of primary and standby data file will be different. (/). If you require catalog data from the database files pending.

    (Reason: you restore controlfile from elementary to step 2, which takes place from the main access road). Use the details that you obtained in step 2 and catalog them.

    4. turn the database copy by RMAN. (RMAN > switch database to copy ;))

    5 Catalog backup items that you copied in step 2.

    6. recover the standby database using 'noredo' option.

    7. finally start the MRP. This should solve your problem.

    The reason I say this works is because here, you restore the controlfile to primary first, which will not have details 21, datafile, and then you are recovering. So it must succeed.

    In the previous method, you tried to first collect all the day before, and then restore the controlfile. While remedial classes, always watch seeks datafile 21 as he controlfile is not yet updated.

    HTH

    -Jonathan Rolland

  • Tablespace does not use a certain data file

    Hi all, I am new to Oracle DB. Reported by developer that storage space is full, and then I found that there are two data files associated with the storage space by checking dba_data_files. The first data file size is already full of 32GB, and the second is 8GB. However, it seems that the tablespace does not use the data file second when I check the free space of the tablespace because I cannot find its use. That's why Oracle DB shows ORA-01654 and tell me to add another data file. I fix this by adding another data file. But I'm just wondering why it seems that the tablespace does not use the second data file, I checked the status of second in the data file is online of v$ datafile. Could someone give me an idea about this problem? Thank you very much.

    c115bca7-E432-460f-a0c5-273b2157dcbe wrote:

    Hi all, I am new to Oracle DB. Reported by developer that storage space is full, and then I found that there are two data files associated with the storage space by checking dba_data_files. The first data file size is already full of 32GB, and the second is 8GB. However, it seems that the tablespace does not use the data file second when I check the free space of the tablespace because I cannot find its use. That's why Oracle DB shows ORA-01654 and tell me to add another data file. I fix this by adding another data file. But I'm just wondering why it seems that the tablespace does not use the second data file, I checked the status of second in the data file is online of v$ datafile. Could someone give me an idea about this problem? Thank you very much.

    1 cause the second data file - is not activate with auto-extend and its fixed size, 8 GB, which could deliver the extensions are not free - filled and that you may not able to find any free extends from dba_free_space
    2. Please activate auto extend for this data file and add the new data file based on space allocation.

    -Pavan Kumar N

  • Increase the size of the tablespace: which of the data file for the tablespaces ine that I have to increase?

    Hello

    I need to increase the size of the tablespaces in EBS R12 Database 11.2.0.3.0:

    APPS_TS_TX_IDX

    APPS_TS_SEED

    APPS_TS_TX_DATA

    But they all have multiple datafiles:

    Example for APPS_TS_SEED:

    [data oracle@ebs122dev] $ of the hs - a_ref0 *.

    1, 5G a_ref01.dbf

    1, 2G a_ref02.dbf

    508M a_ref03.dbf

    2, 0G a_ref04.dbf

    2, 0G a_ref05.dbf

    2, 0G a_ref06.dbf

    Many of the datafiles should I resize to increase the size?

    Thank you

    Luis

    Hello

    It does not matter what file you resize.

    Personally, I prefer to start with the first file and reach the largest file size and resize then the following.

    In addition, having the files with a size that is a multiple of 1 GB is a good practice for me.

    Another option is to activate the autoextend for data files option.

    You can set the maximum size of 4 GB for example all files in storage areas that are growing rapidly.

    Kind regards

    Bashar

  • Data files to the tablespace sysaux listed several times v $ datafile

    When I run the query - select name, number of status from v$ datafile;

    the data files for system tablespaces are included several times.

    For example, for sysaux, output a below entries. Why is this?

    It is the same with the space of system tables and users.

    The configuration is Oracle 12 c

    I've used this in earlier versions of oracle, but this is something new that I am observing with 12 c.

    + ASM_DG/gdb/DataFile/SYSAUX.269.838407323, 3

    + ASM_DG/gdb/DD7C48AA5A4404A2E04325AAE80A403C/DataFile/SYSAUX.273.838407639, 7

    + ASM_DG/gdb/F159D742BCE602C0E0438B59D10A581B/DataFile/SYSAUX.258.838409993, 10

    When I use the query-

    SELECT FILE_NAME, BLOCKS, NOM_TABLESPACE

    FROM DBA_DATA_FILES;

    Then only the single entry appears

    + ASM_DG/gdb/DataFile/SYSAUX.269.838407323, 108800, SYSAUX

    Any help will be great.

    Hello

    It's the new architecture shared in action.

    You get a sysaux/system for each plug-in database data file, but the views dba only show data for the current container.

    This is probably the biggest architectural change I've seen in oracle (and I worked with her since 7.2...)

    I recommend you to read the oracle documentation (and then read them again when it totally confuses you until you get it).

    See you soon,.

    Rich

  • When OMF add the data file in the tablespace

    Hi friends,

    We use oracle 11.2 with CMS (oracle managed file) for the database on the linux platform. the parameter - DB_CREATE_FILE_DEST had been set. at the present time, we received a message from global warming this tablespace criterion is 85% full.

    According to the document of the oracle, OMF auto will add a data file in the tablespace tast. more than a week, I do not see any data file has been added by OMF.

    I want to know when OMF adds the data file in the tablespace? 85% or 95% or some (parameter) setting this action?

    Thank you

    newdba

    OMF does not automatically add a new data file.  You must explicitly add a new data file with the ALTER TABLESPACE ADD DATAFILE command tbsname.

    What is OMF is to provide a Unique name and a default size (100 MB) for the data file.  That is why the ALTER TABLESPACE... Command ADD DATAFILE that you run didn't need to specify the file size or file name.

    Hemant K Collette

  • Transportable tablespaces and data files

    Hello

    I wonder if the following is possible.

    With the help of 11.2.0.3

    Export datapump job to export an entire table every day to dumpfile daily functioning

    Then, slide old partitions for example 1 day

    About destination databased days later (imgaeine 4) if need import current dumpfile and datafile for the affected tablespace that will import the deleted partition (that is, the 1st day

    Thank you

    I'm a little confused as to what you're asking, but since you're on 11.2.0.3 you have called transportable table mode.  You can export a single partition of a table using transportable tablespace data movement.  It helps a lot when you create a new storage space for each partition.  It also works if you do not have, but there are a lot of additional data you need to store.  Let me give you an example:

    table foo with:

    partition 1 - tablespace 1 - day 1

    partition 2 - tablespace 2 - day 2

    partition 3 - tablespace 3 - day 3

    partition 4 - tablespace 4 - day 4

    If you were to run this export:

    tables of user/password = foo:partition1 transportable expdp = always dumpfile = foo_part1.dmp...

    you would get a dumpfile with just partition1.  You must save the dumpfile and the data file if you ever need to restore.  To restore it, run a command like:

    Impdp username/password DB11G = dumpfile = foo_part1.dmp...

    It would be important only the partition, and it would create a table called foo_partition1.  You can then use the swap partition to put back in the table where you need restored.

    If all of these scores have been in the same table, you could do the same thing for export and import, but you would need to keep a copy of the dumpfile with a copy of the storage space for every day.  Then

    day 1 - would thus dumpfile datafile with just data from day 1

    day 2 - would thus dumpfile data file with data from day 1 and the day2

    day 3 - would thus dumpfile datafile on day 1, day 2 + data of the day 3

    etc.

    The reason for this is because the import checks that the tablespace that has the same features when it was exported.  It uses a checksum or something to verify this.  So, if you've added data to day 2, the data file would not be the same and the import fails.

    I hope this helps.

    Dean

  • doubt in the creation of tables in different data in a tablespace files

    Hi good Eve everyone,

    in my scenario, I must create nearly 50 paintings with 3 data files in a table space.

    I have to save the top 15 in datafile 1 and 15 next in 2nd datafile and the rest in 3rd datafile.

    How do I create and tables in a data file of the group.

    looking for a helping hand in need...

    thanking you in advance

    Prakash

    You can do this, but it's awkward. Here is an example:

    ALTER session set deferred_segment_creation = true;

    create tablespace twofiles datafile 'c:\tmp\tf1.dbf' size 1m,'c:\tmp\tf2.dbf' size 1 m;

    create table infile1 (c1 date) tablespace twofiles;

    create table infile2 (c1 date) tablespace twofiles;

    ALTER table infile1 allocate extended (datafile 'c:\tmp\tf1.dbf');

    ALTER table infile2 allocate extended (datafile 'c:\tmp\tf2.dbf');

    Select nom_segment, file_id from dba_extents where nom_tablespace = 'TWOFILES ';

    However, it is simpler to let the Oracle uncle to decide how to distribute the scopes.

  • Users tablespace - data file is 24.5 GB, only 1.2 GB used - what should I do?

    Hi all

    In my database I have users tablespace and its data file is users01.dbf. File size is 24.5 GB, but the data used are of 1.2 GB. I got this information from OEM, see image below:
    https://picasaweb.Google.com/100555146235257804098/Oracle_a1#5759303022195932370

    How I can free unused space of 23GB?

    Just to let you know to my knowledge, I am a developer of report OBI.

    Thank you!

    >
    I executed the script sql on your link http://www.oracle-ckpt.com/how-to-shrink-datafiles-adjusting-hwm/

    It gives only 7 GB. If I can save only 7 GB so it will not help me. This is the maximum possible savings, you can do.

    Do you mean if I load 20 GB of data now in the post above, the database uses the current data file to increase the size of the file? Or it adds to the size of the file, that is to say 25 + 20 = 45 GB?

    It seems CanGrow is enabled for the USERS tablespace: CANGROW - YES

    If it uses the size of the file, then I can deal with the current structure. If we add to the file so I can't move forward.

    Your data file already occupied 23 GB, if you load another 20 GB of data. Of course, it can take another 20 GB, if all 23 GB already existing aid. Have you talked about the link that is mentioned by Sir Jonathan Lewis?

  • How to move a segment in the SYSAUX tablespace data file

    Hello

    I had a problem with my tablespace SYSAUX growing due to the history of Stats being not properly purged (lots of hits and articles against this known issue). I have since corrected this problem and I did my SYSAUX tablespace so that it is now less than 3 GB. However, the data file unique original behind this tablespace is still 10 GB and I want to reduce again to below 6 GB.

    However as a suspect when I try to reduce the data return file I get ORA-03297 file contains data beyond the requested value of RESIZING.

    I checked the table dba_extents for this tablespace, which reveals there are approximately 765 segments over my desired fix of 6 GB.

    Q1. If there is only 1 or 2 segments above this mark of 6 GB, how would manually move their bottom?

    Q2. I gave have 765 segments above this 6 GB brand, what are the recommendations to achieve the mass movement required of these segments?

    Thank you
    Jim

    The Alter Table or Alter Index statement will pass all these types of segment?

    Lol some are special items.

    Here is what I make sure you test something you care to.

    And the double test. Frankly, I would like to change the least possible and if I didn't do it, I would avoid.
    It might be easily really much, much more trouble than its worth.

    LOBINDEX

    ALTER table move lob () store (tablespace );

    NESTED TABLE

    A nested table can be moved to a different tablespace only by moving his table containing storage target.
    (Move the parent table, which is only necessary.)

    PARTITON TABLE

    ALTER partitioned table move partition part_2 tablespace ;

    LOB SEGMENT

    ALTER TABLE owner.table_name MOVE LOB (column_name) STORE AS (nom_tablespace);

    CLUSTER

    I'm not sure.

    INDEX PARTITION

    example of

    ALTER INDEX SYS. I_STREAMS_APPLY_SPILL_MSGS_PT1 CHANGE DEFAULT USERS TABLESPACE ATTRIBUTES;

    PARTITION OF LOB

    ALTER table

    partition of moving lob () store as (tablespace );

    See also - online redefinition of tables

    http://Oracle.Su/docs/11g/server.112/e10595/tables007.htm

    Oracle Database 10 g Release 2 online data reorganization & redefinition

    http://www.Oracle.com/technetwork/database/features/availability/HA-10gR2-online-reorg-TWP-131644.PDF

    http://jhdba.WordPress.com/2009/05/19/purging-statistics-from-the-SYSAUX-tablespace/

    Best regards

    mseberg

  • Reduction of size of the Tablespace SYSTEM or adding data file?

    Hello

    I read the MOS notes on aud$ table, etc and the purge it (truncation). But what happened in this situation. Please take a look. # Oracle Database 11 g Release 11.2.0.2.0 - 64 bit Production
     ## size in MB
    TS Name                                 Total_size   Free space     %age_free    %age_free
    
    SYSTEM                               2000       409.875              20         80
    SYS@AP AS SYSDBA> select owner, segment_name, segment_type, bytes/1024/1024 "MB" from dba_segments
      2  where tablespace_name = 'SYSTEM' AND rownum <=20 AND bytes/1024/1024 > 1 order by bytes desc;
    
    OWNER                      SEGMENT_NAME                                              SEGMENT_TYPE              MB
    ------------------------------ --------------------------------------------------------------------------------- ------------------ ----------
    SYS                      OBJ$                                                   TABLE                   25
    SYS                      I_OBJ2                                                   INDEX                   22
    SYS                      I_OBJ5                                                   INDEX                   22
    SYS                      DEPENDENCY$                                              TABLE                   15
    SYS                      C_OBJ#                                                   CLUSTER              14
    SYS                      I_OBJ1                                                   INDEX                   11
    SYS                      I_OBJ4                                                   INDEX                   11
    SYS                      I_SYN2                                                   INDEX                   10
    SYS                      SYN$                                                   TABLE                    8
    # System Tablespace has a data file:
    /Oracle/oradata/AP/datafile/o1_mf_system_6l9549kc_.dbf

    What are the possibilities for me here, please? Can I purge/truncate above segments, or add a data file?

    Thank you

    Concerning

    ALTER database... datafile autoextend on next maxsize 64M

    So yes, you can change this. Easily!

    --------
    Sybrand Bakker
    Senior Oracle DBA

Maybe you are looking for