collect statistics for the tablespace

Friends...

OS: Linux

DB: 11 GR 2

Data size: 1 TB

I spend monthly multiple partitioned table spaces and bring together in a single annual partition. (for example tbs_2014_01, tbs_2014_02 - tbs_2014_12... all combine them into tbs_2014 as a tablespace)

Over the weekend, work of database gets executed that collects statistics that are obsolete, it collects all the segments that have been moved from the storage.

Given that the collection of statistics at the end of the week takes too long, I tried to find a smart way to collect statistics after each tablespace move rather than waiting for job to weekend which will take two or three days to complete.

1. is there a way to gather statistics at the tablespace level and collect statistics for all objects in this table space?

2. how to determine the overall stats of collection of statistics part?

That is, suppose I have move the tbs_2014_01 tablespace and collect statistics with global stats that could take 2 hours but it will be difficult to spend 2 hours for each stats global tablespace which in my opinion is not good and we should be collecting global stats only once.

3. any other advice?

977272 wrote:

@sol.beach... Thanks for your comments...

I've not been asked to collect statistics to the tablespace but level to collect statistics after that finish objects move in storage.

Given the size of the data, it is difficult to gather all the statistics at the weekend so trying to understand another method to collect the statistics the weekend load will be less.

You can collect statistics object on an object by object basis level after that each object has been moved.

Tags: Database

Similar Questions

  • What is advised to collect statistics for the huge tables?

    We have a staging database, some tables are huge, hundreds GB in size.  Auto stats tasks are performed, but sometimes it will miss deadlines.

    We would like to know the best practices or tips.

    Thank you.

    Improvement of the efficiency of the collection of statistics can be achieved with:

    1. Parallelism using
    2. Additional statistics

    Parallelism using

    Parallelism can be used in many ways for the collection of statistics

    1. Parallelism object intra
    2. Internal parallelism of the object
    3. Inner and Intra object jointly parallelism

    Parallelism object intra

    The DBMS_STATS package contains the DEGREE parameter. This setting controls the intra parallelism, it controls the number of parallel processes to gather statistics. By default, this parameter has the value is equal to 1. You can increase it by using the DBMS_STATS.SET_PARAM procedure. If you do not set this number, you can allow oracle to determine the optimal number of parallel processes that will be used to collect the statistics. It can be achieved if you set the DEGREE with the DBMS_STATS. Value AUTO_DEGREE.

    Internal parallelism of the object

    If you have the 11.2.0.2 version of Oracle database you can set SIMULTANEOUS preferences that are responsible for the collection of statistics, preferably. When there is TRUE value at the same TIME, Oracle uses the Scheduler and Advanced Queuing to simultaneously manage several jobs statistics. The number of parallel jobs is controlled by the JOB_QUEUE_PROCESSES parameter. This parameter must be equal to two times a number of your processor cores (if you have two CPU with 8 cores of each, then the JOB_QUEUE_PROCESSES parameter must be equal to 2 (CPU) x 8 (cores) x 2 = 32). You must set this parameter at the level of the system (ALTER SYSTEM SET...).

    Additional statistics

    This best option corresponds to a partitioned table. If the INCREMENTAL for a partitioned table parameter is set to TRUE and the DBMS_STATS. GATHER_TABLE_STATS GRANULARITY setting is set to GLOBAL and the parameter of DBMS_STATS ESTIMATE_PERCENT. GATHER_TABLE_STATS is set to AUTO_SAMPLE_SIZE, Oracle will scan only the partitions that have changes.

    For more information, read this document and DBMS_STATS

  • Job GATHER_STATS collect statistics for the tables 'static '.

    Oracle version: 10 gr 2

    If a corporate table has not changed (No. DML) in the last 10 days, will be the collection of default oracle job stats
    DBMS_STATS.GATHER_DATABASE_STATS_JOB_PROC
    yet collect statistics in this table?

    The answer is no, unless you have changed the default optimizer stats collection of statistics because approximately 10% of the data must have undergone change before that table is elgible for new statistics.

    See the next topic in the Performance and Tuning section 14.2.1 GATHER_STATS_JOB Manual:

    http://download.Oracle.com/docs/CD/B19306_01/server.102/b14211/stats.htm#sthref1068

    HTH - Mark D Powell.

  • How to collect statistics on the table for tables in a different schema

    Hi all

    I have a table in a schema, and I want to collect statistics for the table in a different schema.

    I gave GRANT ALL ON SCHEMA1. T1 TO SCHEMA2;

    And when I tried to run the command to collect statistics to help

    DBMS_STATS. GATHER_TABLE_STATS (OWNNAME = > 'SCHMEA1', TABNAME = > 'T1');

    The function will fail.

    Is there a way we can collect statistics of the table for tables in a schema into another schema.

    Thank you
    MK.

    You must grant analyze to schema2.

    SY.

  • Collect statistics of the table in windows

    Hello

    Could you please help me to collect statistics for the tables below in Windows. I tried collection using dbms_scheduler job for a table, but it is not the collection of statistics for tables.

    TABLE_NAME OWNERNUM_ROWS BLOCKSSize

    MQRDW DWREFERENCETRACE GO 2891985937 26372904, 17.86

    TRAY PON_SERIAL 563722072 5135734 GB 8,12

    TRAY PON_PSN2358851732009064 2.67 GB

    TRAY PON_BOM_LOG 37199475 212936 601 MB

    TRAY UNIQUE_ITEM_LOTS 6633907 79710 160 MB

    TRAY PON_BOM5921377 41717328 MB

    SQL > start

    (2 dbms_scheduler.create_job)

    job_name 3 = > "SCOTT_JOB_SCHEDULE"

    job_type 4 = > "EXECUTABLE."

    5 job_action = > ' begin dbms_stats.gather_table_stats (ownname = > "TRACE3")

    6 tabname = > 'PON_PSN ',.

    estimate_percent 7 = > 30,

    Cascade 8 = > true,

    method_opt 9 = > 'for all THE COLUMNS of SIZE 1.

    10 degrees = > 5,

    11 options = > "BRINGING OUTDATED"); end;',

    12 repeat_interval = > ' freq = daily; byhour = 04; byminute = 0; bysecond = 0; »,

    13 active = > TRUE,

    14 comments = > ' custom stats collection for engine risk ");

    15 end;

    17 N

    PL/SQL procedure successfully completed .


    Total number of CPU available on this server is 8

    Oracle - 10.2.0.4.0

    OS - windows

    Kind regards

    Bala

    52 million lines?

    and how to change every day or be inserted?

    It is possible that you can never bring his stats again

    and everything will probably continue to work as expected

    read the guide to setting performance to understand why you need to collect statistics and when

    https://docs.Oracle.com/CD/E11882_01/server.112/e41573/stats.htm#PFGRF94714

    After reading this guide, you'll be ready for any question, you get all that against this team

    Sometimes you need to say 'no' and why the answer is 'no '.

  • How to collect statistics for a partition?

    Hi all

    I create a table partition. I need to collect statistics for this partition only. Before I used to analyze it, but now I need to analyze using DBMS_STATS.

    What is the best way to analyze the partition using DBMS_STATS?

    How long will it take to complete?

    How can I estimate the time of accomplishment for DBMS_STATS before starting?

    Thank you

    I create a table partition. I need to collect statistics for this partition only. Before I used to analyze it, but now I need to analyze using DBMS_STATS.

    What is the best way to analyze the partition using DBMS_STATS?

    Follow the documented instructions: INCREMENTIELLE TRUE and GRANULARITY on AUTO.

    See the section "Partitioned objects statistics" the doc of performance tuning

    http://docs.Oracle.com/CD/B28359_01/server.111/b28274/stats.htm#i42218

    With partitioned tables, the new data is usually loaded into a new partition. As new partitions are added and loaded, statistical data must be collected on the new partition and statistics need to be updated. If the INCREMENTAL for a partition table is set to the value TRUE , and collect you statistics on the table with the GRANULARITY parameter defined on AUTO , Oracle will collect statistics on the new partition and update statistics on the overall table by scanning only those partitions which have been modified and not the entire table. If the INCREMENTAL for the partitioned table is set to the value FALSE (the default), then a full table scan is used to maintain the global statistics. It is a highly resource intensive and time-consuming for large party.

    How long will it take to complete?

    No way to know - using an estimate of 10% takes less time than with an estimated of 40%, which takes less time than using 100%.

    How can I estimate the time of accomplishment for DBMS_STATS before starting?

    By comparing the amount of data and the percentage of estimate for the data that you have in the other partitions and the time required to collect statistics on other partitions.,.

  • Collects statistics for tables using jobs in Oracle 11 G 11.2.0.1.0?

    Hello

    My query is regarding the collection of statistics for tables.

    My Version of Oracle DB is:
    BANNER
    --------------------------------------------------------------------------------
    Oracle Database 11 g Enterprise Edition Release 11.2.0.1.0 - 64 bit Production
    PL/SQL Release 11.2.0.1.0 - Production
    CORE 11.2.0.1.0 Production
    AMT for Linux: Version 11.2.0.1.0 - Production
    NLSRTL Version 11.2.0.1.0 - Production

    In versions of prior oracle db, we used to schedule tasks to run on a daily basis for collecting statistics. Especially for tables that are frequent and huge inserts.

    I read that in 11g stats for all of the schema on a database are automatically make every night jobs. I checked these jobs and I see that they are running on a monthly basis [joined query]. This job is enabled and is scheduled to run monthly.
    This means that my diagram will be analyzed on a monthly basis. My understanding is correct?

    Can I still plan jobs to collect statistics for specific tables on every week? This will diminish the performance?
    We expect 100000 documents to insert on a daily basis.
    SELECT  JOB_NAME,  
       START_DATE, REPEAT_INTERVAL, 
         LAST_START_DATE, 
        NEXT_RUN_DATE,ENABLED
    FROM dba_scheduler_jobs
    WHERE job_name LIKE '%STAT%'
    ORDER BY 1;
    JOB_NAME                       START_DATE                             REPEAT_INTERVAL                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  LAST_START_DATE                        NEXT_RUN_DATE                          ENABLED
    ------------------------------ -------------------------------------- ---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- -------------------------------------- -------------------------------------- -------
    BSLN_MAINTAIN_STATS_JOB        16-AUG-09 12.00.00.000000000 AM -07:00                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  14-APR-13 12.00.00.427370000 AM -07:00 21-APR-13 12.00.00.400000000 AM -07:00 TRUE    
    MGMT_STATS_CONFIG_JOB          15-AUG-09 12.24.04.694342000 AM -07:00 freq=monthly;interval=1;bymonthday=1;byhour=01;byminute=01;bysecond=01                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           01-APR-13 01.01.01.710280000 AM -07:00 01-MAY-13 01.01.01.700000000 AM -07:00 TRUE    
    Thank you
    Somiya

    Your understanding is not correct. These jobs are in dba_autotask_task.
    And they will be run every day.

    HTH

    -------------
    Sybrand Bakker
    Senior Oracle DBA

  • I want to collect rain for the day, month and year; Can what formula I use?

    I want to collect rain for the day, month and year; What formula or expression that I can use.

    I use a Rain Wise product that converts impulses to an analog value.  The rain wise device can be

    "the value measure up to 1", 5", or 10".  I will be setting the unit at 10 inches in increments of 0.01 inch.

    What I would do is at each time increments of signal I would consider it as 0.01 then after reaching

    a period of time a day or a month, back to zero.

    Need some advice on this problem.

    How about you, using a counter to count the number of times where the signal is less than 0.01?

    For example,.

    Counter.Count = signal<=>

    The total number of rain is

    Counter * 10 + signal

    Use TimeOfDay or TimeOfWeek to reset the counter.

  • missing metric collections officer for the duration of downtime

    Hello

    We had to stop the FMS, and after we grew up, we discovered that metric collections officer for the duration of the downtime was missing and not populated dashboard FMS,

    We have a hypothesis, that when FMS is down or not accessible by FGLAM on the monitored hosts, that FGLAM still collects its measures and it will grow to FMS when it is returned to the online state.

    The questions are, is it true the assumption? And if it is true, how restore us these metrics data for the mentioned period?

    Thank you

    Hi Waseem-

    fglam there a store and the capacity before. The amount of agent Manager data stores is controlled

    with some configuration settings. You can read about some details here:

    http://eDOCS.quest.com/Foglight/555/doc/wwhelp/wwhimpl/common/HTML/frameset.htm?context=field&file=StoreAndForward-field/StoreAndForwardProcessing.17.4.php&single=true

    HTH

    Robert Statsinger

  • What is the reason for the tablespaces exist?

    Hello

    If I can store in a tablespace of data of different users / schemas / apps and also a user data / shema / application in several storage spaces. What is the meaning of a tablespace?

    What is the reason for the tablespaces exist? Only to assign the storage space and know where are the objects of the user? Is it not possible to work with data directly files instead of doing with the tablespaces?

    I'm afraid I'm too philosophical today but I do not understand why the structure of the tablespace.

    Hello

    Apart from the above, I'll add the storage capacity of a Tablespace.

    A Tablespace usually (default is a petit_fichier Tablespace) can have up to * 1022 * data files. And (in the case
    petit_fichier Tablespace) each data files can be big 4 MBlocks.

    So if your db_block_size is 8 KB your Tablespace can store up to 1022 x 4 M x 8 KB ~ 32 to data.

    More, these 1022 data files can be located on separate disks. So it can give you plenty of opportunity.

    However, from 10g you have a new type of Tablespace: bigfile Tablespace.

    Bigfile Tablespace has only 1 data file, but it can reach 4 GBlocks. So finally this Tablespace can store
    also 32 to (for 8 KB db_block_size) but, ensure you that your OS platform and system disk can support
    a big Datafile.

    The syntax for the statements that manage the data of a bigfile Tablespace file are too different. In fact, you can
    manage this data at the level of the Tablespace file.

    Please find attached a link to bigfile Tablespace:

    [http://download.oracle.com/docs/cd/E11882_01/server.112/e10595/tspaces002.htm#ADMIN01102]

    Best regards
    Jean Valentine

  • collect statistics of the table

    Hi all

    DB version: 10.2.0.4
    OS: AIX 6.1

    I want to collect the stats table for a table as the query that uses that table is slow. Also, I noticed that this table is using full table scan, and it has been analyzed in last 2 months back.

    I intend to run the query to collect the statistics below. The table has 50 million records.
    COUNT (*)
    ----------
    51364617

    I expect that this will take time if I run the query as below.


    DBMS_STATS.gather_table_stats EXEC ('schema_name', 'table_name');

    My doubts specified below.

    1. can I use the also estimate_percent parameter to collect statistics?

    2 should. how much we indicate for the estimate_percent parameter?

    3. what difference it will make if I use the estimate_percent parameter?

    Thanks in advance

    Published by: user13364377 on March 27, 2012 13:28

    If you are worried about the stats of collection process for a long time running, consider collecting statistics in parallel.

    1. can you use estimate_percent? Sure! Go ahead.
    2. what percentage of use? Why not let the data decide with auto_sample_size? Various "rules of thumb" were thrown around over the years, usually about 10 to 20%.
    3. what difference it will make? Very little, no doubt. Occasionally, you can see where a small sample makes a difference, but in general that it's perfectly ok to estimate its stats.

    Maybe something like this:

    BEGIN
      dbms_stats.gather_table_stats(ownname => user, tabname => 'MY_TABLE',
       estimate_percent => dbms_stats.auto_sample_size, method_opt=>'for all columns size auto',
       cascade=>true,degree=>8);
    END;
    
  • What happends with the statistics after the tablespace change?

    Hi all.

    I'm on Oracle 10 g R2. I want to spend a few tables and its indices of a tablespace to another. I would like to know what
    happends with statistics and I have to recalculate the.

    Thank you.

    http://download.Oracle.com/docs/CD/B19306_01/server.102/b14231/tables.htm#sthref2297


    Move a table changes the ROWID of the rows in the table. This causes indexes on the table must be worn UNUSABLE and DML accessing the table using these indices will receive an ORA-01502 error. The indexes on the table must be deleted or rebuilt. Similarly, all the statistics in the table become invalid and new statistics should be collected after the removal of the table.

  • New on Foglight - better statistics for the dashboard alerts

    I would like to delete alerts of history that really make sense to keep such as a user having connection problems.  However, I would like to personalize history such as keep statistics for CPU, memory, disk usage.  I called quest and opened a ticket that they responded that they do not support this type of customization.  We bought Foglight to make the area more transparent DBA which means other groups will be watching the dashboards, for once that we recognize and disable alerts/alarms, I want these numbers to go.  Is there a way to do this?

    Hi Carolyn-

    First of all, welcome to our Community Site and Foglight.

    There are a couple things to point out on top:

    1. the alarm has a dashboard filter that you can assign only to view the current alarms or show historical alarms (e.g. alarms that have been deleted or that have taken place in the past). So, you can use this to set how you want old alarms to be displayed, and remove those that you don't care.

    2. If you really want to * remove * old alarms of the Foglight repository, you can do it too, but note if you do that they will not

    be available for display, and they will not be able to be used when we want to establish a correlation between a current alarm (for example when this type of alarm occurred before). Is actually deleting old alarms of the repository of something you want to do?

    3. Note that the * data * keeps Foglight can also be customized. If you look at the Administration-> data-> manage retention policies you

    can set policies on how long Foglight retains certain types of data and how Foglight roll (e.g. aggregates) data points while they age over time.

    If you look at Management Server-> servers-> data management, you will see the current * global * applied to all data entered data retention policy. You can customize this type of data in the view to manage retention policies.

    Does that help?

    Robert Statsinger

  • Increase the size of the tablespace: which of the data file for the tablespaces ine that I have to increase?

    Hello

    I need to increase the size of the tablespaces in EBS R12 Database 11.2.0.3.0:

    APPS_TS_TX_IDX

    APPS_TS_SEED

    APPS_TS_TX_DATA

    But they all have multiple datafiles:

    Example for APPS_TS_SEED:

    [data oracle@ebs122dev] $ of the hs - a_ref0 *.

    1, 5G a_ref01.dbf

    1, 2G a_ref02.dbf

    508M a_ref03.dbf

    2, 0G a_ref04.dbf

    2, 0G a_ref05.dbf

    2, 0G a_ref06.dbf

    Many of the datafiles should I resize to increase the size?

    Thank you

    Luis

    Hello

    It does not matter what file you resize.

    Personally, I prefer to start with the first file and reach the largest file size and resize then the following.

    In addition, having the files with a size that is a multiple of 1 GB is a good practice for me.

    Another option is to activate the autoextend for data files option.

    You can set the maximum size of 4 GB for example all files in storage areas that are growing rapidly.

    Kind regards

    Bashar

  • Collect statistics on the table with indexes of text only?

    I gathered statistics for a table that contains a text index
    EXEC DBMS_STATS. GATHER_TABLE_STATS (USER, 'CONADDR', estimate_percent = > 10, block_sample = > TRUE, cascade = > TRUE);

    There are a lot of tables/indexes not monitored (e.g. DR$ TI_CONADDR$ I). Do I have to analyse the tables there, too? The Guide Tuning Oracle text mentions just to analyze the table of "base".

    Oracle DB version is 10.2.0.4.
    select table_name, last_analyzed, num_rows from dba_tables where table_name like '%CONADDR%';
    CONADDR     11.08.2010 10:29:37     17944660
    DR$TI_CONADDR$I          
    DR$TI_CONADDR$R          
    DR$TI_CONADDR$K          
    DR$TI_CONADDR$N          
    
    select index_name, table_name, last_analyzed, num_rows from dba_indexes where table_name like '%CONADDR%';
    SYS_IL0003730268C00004$$     CONADDR          
    IDX_CONADDR                     CONADDR     11.08.2010 10:29:46     17106050
    SYS_IL0003731165C00006$$     DR$TI_CONADDR$I          
    SYS_IOT_TOP_3731168             DR$TI_CONADDR$K          
    SYS_IL0003731170C00002$$     DR$TI_CONADDR$R          
    SYS_IOT_TOP_3731173             DR$TI_CONADDR$N          
    DR$TI_CONADDR$X                     DR$TI_CONADDR$I     11.08.2010 10:05:05     67585029
    TI_CONADDR                     CONADDR     11.08.2010 10:29:46     

    DR$ table do NOT need to be analysed - and should not be.

    As "secondary objects", they will not be analyzed by orders based on patterns, and it is strongly recommended to not analyze manually. All commands that access these tables are set correctly without the input of the optimizer.

Maybe you are looking for