分区表的统计信息收集策略

#####1 如果每天产生一个分区,

1.10g库如果是一个日分区表,每天产生20到30万 笔数据,可以考虑采用分区复制的方式来缓解10g 晚上22点的统计信息造成的I/O 高峰期的

2.11g库可以采用“Incremental Statistic 的方式只收集增量数据。

https://blogs.oracle.com/optimizer/maintaining-statistics-on-large-partitioned-tables

Maintaining statistics on large partitioned tables

By: Sunil Chakkappen

We have gotten a lot of questions recently regarding how to gather and maintain optimizer statistics on large partitioned tables. The majority of these questions can be summarized into two topics:

  1. When queries access a single partition with stale or non-existent partition level statistics I get a sub optimal plan due to "Out of Range" values
  2. Global statistics collection is extremely expensive in terms of time and system resources

This article will describe both of these issues and explain how you can address them both.

This is big topic so I recommend that you also check out the three-part series of posts on maintaining incremental statistics in partitioned tables.

Out of Range
Large tables are often decomposed into smaller pieces called partitions in order to improve query performance and ease of data management. The Oracle query optimizer relies on both the statistics of the entire table (global statistics) and the statistics of the individual partitions (partition statistics) to select a good execution plan for a SQL statement. If the query needs to access only a single partition, the optimizer uses only the statistics of the accessed partition. If the query access more than one partition, it uses a combination of global and partition statistics.

"Out of Range" means that the value supplied in a where clause predicate is outside the domain of values represented by the [minimum, maximum] column statistics. The optimizer prorates the selectivity based on the distance between the predicate value and the maximum value (assuming the value is higher than the max), that is, the farther the value is from the maximum value, the lower the selectivity will be. This situation occurs most frequently in tables that are range partitioned by a date column, a new partition is added, and then queried while rows are still being loaded in the new partition. The partition statistics will be stale very quickly due to the continuous trickle feed load even if the statistics get refreshed periodically. The maximum value known to the optimizer is not correct leading to the "Out of Range" condition. The under-estimation of selectivity often leads the query optimizer to pick a sub optimal plan. For example, the query optimizer would pick an index access path while a full scan is a better choice.

The "Out of Range" condition can be prevented by using the new copy table statistics procedure available in Oracle Database10.2.0.4 and 11g. This procedure copies the statistics of the source [sub] partition to the destination [sub] partition. It also copies the statistics of the dependent objects: columns, local (partitioned) indexes etc. It adjusts the minimum and maximum values of the partitioning column as follows; it uses the high bound partitioning value as the maximum value of the first partitioning column (it is possible to have concatenated partition columns) and high bound partitioning value of the previous partition as the minimum value of the first partitioning column for range partitioned table. It can optionally scale some of the other statistics like the number of blocks, number of rows etc. of the destination partition.

Assume we have a table called SALES that is ranged partitioned by quarter on the SALES_DATE column. At the end of every day data is loaded into latest partition. However, statistics are only gathered at the end of every quarter when the partition is fully loaded. Assuming global and partition level statistics (for all fully loaded partitions) are up to date, use the following steps in order to prevent getting a sub-optimal plan due to "out of range".

1. Lock the table statistics using LOCK_TABLE_STATS procedure in DBMS_STATS. This is to avoid interference from auto statistics job.

EXEC DBMS_STATS.LOCK_TABLE_STATS(‘SH‘,‘SALES‘);

2. Before beginning the initial load into each new partition (say SALES_Q4_2000) copy the statistics from the previous partition (say SALES_Q3_2000) using COPY_TABLE_STATS. You need to specify FORCE=>TRUE to override the statistics lock.

EXEC DBMS_STATS.COPY_TABLE_STATS (‘SH‘, ‘SALES‘, ‘SALES_Q3_2000‘, ‘SALES_Q4_2000‘, FORCE=>TRUE);

Expensive global statistics collection

In data warehouse environment it is very common to do a bulk load directly into one or more empty partitions. This will make the partition statistics stale and may also make the global statistics stale. Re-gathering statistics for the effected partitions and for the entire table can be very time consuming. Traditionally, statistics collection is done in a two-pass approach:

  • In the first pass we will scan the table to gather the global statistics
  • In the second pass we will scan the partitions that have been changed to gather their partition level statistics.

The full scan of the table for global statistics collection can be very expensive depending on the size of the table. Note that the scan of the entire table is done even if we change a small subset of partitions.

We avoid scanning the whole table when computing global statistics by deriving the global statistics from the partition statistics. Some of the statistics can be derived easily and accurately from partition statistics. For example, number of rows at global level is the sum of number of rows of partitions. Even global histogram can be derived from partition histograms. But the number of distinct values (NDV) of a column cannot be derived from partition level NDVs. So, Oracle maintains another structure called a synopsis for each column at the partition level. A synopsis can be considered as sample of distinct values. The NDV can be accurately derived from synopses. We can also merge multiple synopses into one. The global NDV is derived from the synopsis generated by merging all of the partition level synopses. To summarize:

  1. Gather statistics and create synopses for the changed partitions only
  2. Oracle automatically merges partition level synopses into a global synopsis
  3. The global statistics are automatically derived from the partition level statistics and global synopses

Incremental maintenance feature is disabled by default. It can be enabled by changing the INCREMENTAL table preference to true. It can also be enabled for a particular schema or at the database level.

Assume we have table called SALES that is range partitioned by day on the SALES_DATE column. At the end of every day data is loaded into latest partition and partition statistics are gathered. Global statistics are only gathered at the end of every month because gathering them is very time and resource intensive. Use the following steps in order to maintain global statistics after every load.

Turn on incremental feature for the table.

EXEC DBMS_STATS.SET_TABLE_PREFS(‘SH‘,‘SALES‘,‘INCREMENTAL‘,‘TRUE‘);

At the end of every load gather table statistics using GATHER_TABLE_STATS command. You don‘t need to specify the partition name. Also, do not specify the granularity parameter. The command will collect statistics for partitions where data has change or statistics are missing and update the global statistics based on the partition level statistics and synopsis.

EXEC DBMS_STATS.GATHER_TABLE_STATS(‘SH‘,‘SALES‘);  (11g)

Note: that the incremental maintenance feature was introduced in Oracle Database 11g Release 1. However, we also provide a solution in Oracle Database10g Release 2 (10.2.0.4) that simulates the same behavior. The 10g solution is a new value, ‘APPROX_GLOBAL AND PARTITION‘ for the GRANULARITY parameter of the GATHER_TABLE_STATS procedures. It behaves the same as the incremental maintenance feature except that we don‘t update the NDV for non-partitioning columns and number of distinct keys of the index at the global level. For partitioned column we update the NDV as the sum of NDV at the partition levels. Also we set the NDV of columns of unique indexes as the number of rows of the table. In general, non-partitioning column NDV at the global level becomes stale less often. It may be possible to collect global statistics less frequently then the default (when table changes 10%) since approx_global option maintains most of the global statistics accurately.

#########2 11G 新特性:

http://blog.itpub.net/90618/viewspace-1296970/

Applying “Incremental Statistic” for Oracle Big Partition Table

分类: Linux操作系统

2014-10-13 11:17:09

原文地址:Applying “Incremental Statistic” for Oracle Big Partition Table 作者:realkid4

In CBO, Statistic is critical important for Optimizer. Precise and timely statistics will truly reflect the data distribution and volume, and generate wiser SQL Execution Plan (SEP). After years of improvement, CBO is already widely accepted as the default and desirable solution for Oracle Optimizer.

As I mention in many blogs before, statistic is like the raw material for CBO optimizer. Without accurate statistic, it’s impossible to generate optimal SEP for CBO. So in CBO era, DBAs’ concern is about how to collect data statistic and finding out the proper frequency to collect them.

Since 10g, Oracle introduces a new feature of data statistic job to collect statistic automatically. The job will run at daily interval to collect statistic, and provide most database object statistic to CBO. The new feature solves many problems in most cases. The statistic collection job and mature dynamic sampling sweep the barrier of CBO’s usage.

  1. 1.      Big Partition Table Statistic

Frankly speaking, Oracle internal functions are already enough for ordinary system requirement. But for some special cases, life would be tougher.

For many OLTP/OLAP systems, they usually load bulk of data into some extremely huge table (Mostly Partition Tables) in the night and some other business-free time, and then doing the processing work. The problem is that Oracle would have few chances to collect statistic after the data loading, which would influence the data volume and distribution information.

The problem would finally lead to the bad performance in the next processing work. In production environments, we often receive complains about some job processing is extremely slow. But after clearly examination, the old statistic is main reason. After collection, the job would be better. And then in the next loading, errors would happen again.

If we insert the statistic collection statement between data loading and processing, there would be some time consuming drawbacks for some big partition tables. For most partitions in these tables, the data is stable, and loading work only affects one or two partitions. But for normal partition statistic works, Oracle will collect all partition including inactive partitions, which consume a lot of resource.

In 11g, Oracle introduces a new feature named “Incremental Statistic”, which will only collect the partitions which are new or undergoing huge data changes. So it will save the time of collecting inactive partitions and make collection work shorter.

  1. 2.      Environment Introduction

The new feature of “Incremental Statistic” is first introduced in the update edition of 10.2.0.4, and become formal in Oracle 11gR1. So we chose the Oracle 11gR2 as the test environment.

SQL> select * from v$version;

BANNER

--------------------------------------------------------------------------------

Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - Production

PL/SQL Release 11.2.0.1.0 - Production

CORE        11.2.0.1.0         Production

I prepare a normal list partition table named T_PART.

SQL> create table t_part

2  ( owner varchar2(100),

3    object_id number,

4    object_name varchar2(100),

5    create_date date

6  )

7  partition by list(owner)

8  (partition t_part_pub values (‘PUBLIC‘),

9   partition t_part_sys values (‘SYS‘),

10   partition t_part_big1 values (‘APEX_030200‘,‘SYSMAN‘),

11   partition t_part_big2 values (‘ORDSYS‘,‘MDSYS‘),

12   partition t_part_other values(default));

Table created

  1. 3.      No-Incremental Statistic Behavior

Firstly, let’s see the default Oracle behavior. for partition statistic.

SQL> insert into t_part select owner, object_id, object_name, sysdate from dba_objects where owner in (‘PUBLIC‘,‘ORDSYS‘);

30228 rows inserted

SQL> commit;

Commit complete

SQL> exec dbms_stats.gather_table_stats(user,‘T_PART‘,cascade => true);

PL/SQL procedure successfully completed

SQL> select partition_name, NUM_ROWS, BLOCKS, LAST_ANALYZED, global_stats from dba_tab_partitions where table_owner=‘SYS‘ and table_name=‘T_PART‘;

PARTITION_NAME    NUM_ROWS     BLOCKS LAST_ANALYZED        GLOBAL_STATS

--------------- ---------- ---------- -------------------- ------------

T_PART_PUB           27696        197 2012/10/9 22:08:11   YES

T_PART_SYS               0          0 2012/10/9 22:08:11   YES

T_PART_BIG1              0          0 2012/10/9 22:08:10   YES

T_PART_BIG2           2532         20 2012/10/9 22:08:10   YES

T_PART_OTHER             0          0 2012/10/9 22:08:10   YES

The default behavior. is that Oracle will collect all partitions according to the last_analyzed column in view dba_tab_partition.

Doing some modification focus on some partitions, let’s see the collection results.

SQL> insert into t_part select owner, object_id, object_name, sysdate from dba_objects where owner in (‘SYS‘,‘APEX_030200‘);

33241 rows inserted

SQL> commit;

Commit complete

SQL> exec dbms_stats.gather_table_stats(user,‘T_PART‘,cascade => true);

PL/SQL procedure successfully completed

SQL> select partition_name, NUM_ROWS, BLOCKS, LAST_ANALYZED, global_stats from dba_tab_partitions where table_owner=‘SYS‘ and table_name=‘T_PART‘;

PARTITION_NAME    NUM_ROWS     BLOCKS LAST_ANALYZED        GLOBAL_STATS

--------------- ---------- ---------- -------------------- ------------

T_PART_PUB           27696        197 2012/10/9 22:13:31   YES

T_PART_SYS           30835        197 2012/10/9 22:13:31   YES

T_PART_BIG1           2406         20 2012/10/9 22:13:30   YES

T_PART_BIG2           2532         20 2012/10/9 22:13:31   YES

T_PART_OTHER             0          0 2012/10/9 22:13:31   YES

The insert statement only affects partition T_PART_SYS and T_PART_BIG1. But when we run the collection statement, all partitions will getting involved in the collection work.

  1. 4.      Incremental Statistic Setting

It’s easy to change the statistic strategy for big partition table in Oracle 11g. Using the package dbms_stats will help setting the parameters.

The default strategy for Oracle partition table is not incremental setting. Only three parameters will affect the behavior.

SQL> select dbms_stats.get_prefs(‘PUBLISH‘,‘SYS‘,‘T_PART‘) from dual;

DBMS_STATS.GET_PREFS(‘PUBLISH‘

--------------------------------------------------------------------------------

TRUE

SQL> select dbms_stats.get_prefs(‘INCREMENTAL‘,‘SYS‘,‘T_PART‘) from dual;

DBMS_STATS.GET_PREFS(‘INCREMEN

--------------------------------------------------------------------------------

FALSE

SQL> select dbms_stats.get_prefs(‘GRANULARITY‘,‘SYS‘,‘T_PART‘) from dual;

DBMS_STATS.GET_PREFS(‘GRANULAR

--------------------------------------------------------------------------------

AUTO

Parameter “PUBLISH” is used to determinate whether CBO use the newest statistic when new statistic coming. It relates to pending statistic feature in most cases. And we need to ensure its value is true (Default Value).

Parameter “Incremental” default value is false, which means it would not adopt incremental statistic strategy by default.

Parameter “GRANULARITY” is not related to incremental statistic strategy in Oracle 11g, but related in Oracle 10gR2. So we need to keep it “AUTO” in Oracle 11g environment.

The estimated_precent should be keep to “AUTO”, in order to maintain the incremental statistic.

In the next section, we will see the effect of the new feature.

  1. 5.      Incremental Statistic Behavior

First, we need to change the setting for Big partition table.

SQL> exec dbms_stats.set_table_prefs(user,‘T_PART‘,pname => ‘INCREMENTAL‘,pvalue => ‘TRUE‘);

PL/SQL procedure successfully completed

SQL> select dbms_stats.get_prefs(‘INCREMENTAL‘,‘SYS‘,‘T_PART‘) from dual;

DBMS_STATS.GET_PREFS(‘INCREMEN

--------------------------------------------------------------------------------

TRUE

Do some modification work to T_PART, and see the effect of collection.

SQL> select partition_name, NUM_ROWS, BLOCKS, LAST_ANALYZED, global_stats from dba_tab_partitions where table_owner=‘SYS‘ and table_name=‘T_PART‘;

PARTITION_NAME    NUM_ROWS     BLOCKS LAST_ANALYZED        GLOBAL_STATS

--------------- ---------- ---------- -------------------- ------------

T_PART_PUB           27696        197 2012/10/9 22:13:31   YES

T_PART_SYS           30835        197 2012/10/9 22:13:31   YES

T_PART_BIG1           2406         20 2012/10/9 22:13:30   YES

T_PART_BIG2           2532         20 2012/10/9 22:13:31   YES

T_PART_OTHER             0          0 2012/10/9 22:13:31   YES

SQL> insert into t_part select owner, object_id, object_name, sysdate from dba_objects where owner not in (‘SYS‘,‘APEX_030200‘,‘PUBLIC‘,‘SYSMAN‘,‘ORDSYS‘,‘MDSYS‘);

3465 rows inserted

SQL> insert into t_part select owner, object_id, object_name, sysdate from dba_objects where owner in (‘SYS‘);

30835 rows inserted

SQL> commit;

Commit complete

SQL> exec dbms_stats.gather_table_stats(user,‘T_PART‘,cascade => true);

PL/SQL procedure successfully completed

SQL> select partition_name, NUM_ROWS, BLOCKS, LAST_ANALYZED, global_stats from dba_tab_partitions where table_owner=‘SYS‘ and table_name=‘T_PART‘;

PARTITION_NAME    NUM_ROWS     BLOCKS LAST_ANALYZED        GLOBAL_STATS

--------------- ---------- ---------- -------------------- ------------

T_PART_PUB           27696        197 2012/10/9 22:29:26   YES

T_PART_SYS           61670        398 2012/10/9 22:29:21   YES

T_PART_BIG1           2406         20 2012/10/9 22:29:29   YES

T_PART_BIG2           2532         20 2012/10/9 22:29:22   YES

T_PART_OTHER          3465         23 2012/10/9 22:29:28   YES

The result is weird. Only two partitions are affected in the insert serial, but after collection, we find all partitions are collected. Not the things should be~

Let’s do more continue.

--Delete only One Partition Data

SQL> delete t_part where wner=‘SYS‘;

61670 rows deleted

SQL> commit;

Commit complete

SQL> exec dbms_stats.gather_table_stats(user,‘T_PART‘);

PL/SQL procedure successfully completed

SQL> select partition_name, NUM_ROWS, BLOCKS, LAST_ANALYZED, global_stats from dba_tab_partitions where table_owner=‘SYS‘ and table_name=‘T_PART‘;

PARTITION_NAME    NUM_ROWS     BLOCKS LAST_ANALYZED        GLOBAL_STATS

--------------- ---------- ---------- -------------------- ------------

T_PART_PUB           27696        197 2012/10/9 22:29:26   YES

T_PART_SYS               0        398 2012/10/9 22:42:47   YES

T_PART_BIG1           2406         20 2012/10/9 22:29:29   YES

T_PART_BIG2           2532         20 2012/10/9 22:29:22   YES

T_PART_OTHER          3465         23 2012/10/9 22:29:28   YES

The desired result comes out. Only the affected partitions is collected according to the last_analyzed column, and the table analyzed time is also changed.

SQL> select table_name, last_analyzed from dba_tables where wner=‘SYS‘ and table_name=‘T_PART‘;

TABLE_NAME                     LAST_ANALYZED

------------------------------ --------------------

T_PART                         2012/10/9 22:42:49

If we affect more than one partition at once, how things would be.

SQL> insert into t_part select owner, object_id, object_name, sysdate from dba_objects where owner in (‘PUBLIC‘);

27696 rows inserted

SQL> insert into t_part select owner, object_id, object_name, sysdate from dba_objects where owner in (‘APEX_030200‘);

2406 rows inserted

SQL> commit;

Commit complete

SQL> exec dbms_stats.gather_table_stats(user,‘T_PART‘,cascade => true);

PL/SQL procedure successfully completed

SQL> select partition_name, NUM_ROWS, BLOCKS, LAST_ANALYZED, global_stats from dba_tab_partitions where table_owner=‘SYS‘ and table_name=‘T_PART‘;

PARTITION_NAME    NUM_ROWS     BLOCKS LAST_ANALYZED        GLOBAL_STATS

--------------- ---------- ---------- -------------------- ------------

T_PART_PUB           55392        388 2012/10/9 22:47:07   YES

T_PART_SYS           30874        398 2012/10/9 22:45:21   YES

T_PART_BIG1           4812         36 2012/10/9 22:47:08   YES

T_PART_BIG2           2532         20 2012/10/9 22:29:22   YES

T_PART_OTHER          3465         23 2012/10/9 22:29:28   YES

The truncate statement will affect the partitions and let’s see the results.

SQL> truncate table t_part;

Table truncated

SQL> exec dbms_stats.gather_table_stats(user,‘T_PART‘,cascade => true);

PL/SQL procedure successfully completed

SQL> select partition_name, NUM_ROWS, BLOCKS, LAST_ANALYZED, global_stats from dba_tab_partitions where table_owner=‘SYS‘ and table_name=‘T_PART‘;

PARTITION_NAME    NUM_ROWS     BLOCKS LAST_ANALYZED        GLOBAL_STATS

--------------- ---------- ---------- -------------------- ------------

T_PART_PUB               0          0 2012/10/9 22:55:31   YES

T_PART_SYS               0          0 2012/10/9 22:55:31   YES

T_PART_BIG1              0          0 2012/10/9 22:55:31   YES

T_PART_BIG2              0          0 2012/10/9 22:55:31   YES

T_PART_OTHER             0          0 2012/10/9 22:55:31   YES

SQL> insert into t_part select owner, object_id, object_name, sysdate from dba_objects where owner in (‘SYS‘,‘APEX_030200‘);

33280 rows inserted

SQL> commit;

Commit complete

SQL> exec dbms_stats.gather_table_stats(user,‘T_PART‘,cascade => true);

PL/SQL procedure successfully completed

SQL> select partition_name, NUM_ROWS, BLOCKS, LAST_ANALYZED, global_stats from dba_tab_partitions where table_owner=‘SYS‘ and table_name=‘T_PART‘;

PARTITION_NAME    NUM_ROWS     BLOCKS LAST_ANALYZED        GLOBAL_STATS

--------------- ---------- ---------- -------------------- ------------

T_PART_PUB               0          0 2012/10/9 22:55:31   YES

T_PART_SYS           30874        202 2012/10/9 22:58:05   YES

T_PART_BIG1           2406         20 2012/10/9 22:58:07   YES

T_PART_BIG2              0          0 2012/10/9 22:55:31   YES

T_PART_OTHER             0          0 2012/10/9 22:55:31   YES

Ok, the result is correct.

Only one thing we still wandering: Why we first switch the table to incremental statistic mode, and then we do some modification on some partitions. After collection, we found Oracle will collect the statistic for all partition, and then incremental will happen in the next collection.

The reasonable explanation is: Like Oracle incremental backup strategy, although you have a full mode backup, Oracle would do the incremental full backup work at first time. As for incremental statistic, things are similar. Oracle first needs a full incremental statistic, and then collects the increments.

  1. 6.      Conclusion

The advantage of incremental statistic is that they can collect only active partitions’ data in the shorter time. The inactive partitions which do not have significant data modification operation will not getting involved in the collection work.

For the load data and then process jobs, it is the desirable collection strategy.

##############3  gather_stats_job - GRANULARITY AUTO for Partitions and Subpartitions

Hi,

we have a few tables with partitions and subpartitions and use the "auto optimizer stats collection" to collect the stats.

My question is about the "GRANULARITY" parameter:

SELECT   DBMS_STATS.get_prefs (‘GRANULARITY‘) GRANULARITY FROM DUAL;

GRANULARITY

------------------------------

AUTO

The documentations says: "determines the granularity based on the partitioning type. This is the default value." What does "AUTO" mean, does oracle the stats collection for global, partition and subpartition if they exist?

See my other parameter:

SELECT   DBMS_STATS.get_prefs (‘cascade‘) CASCADE FROM DUAL;

CASCADE

------------------------------

DBMS_STATS.AUTO_CASCADE

SELECT   DBMS_STATS.get_prefs (‘estimate_percent‘) ESTIMATE_PESQL> RCENT FROM DUAL;

ESTIMATE_PERCENT

------------------------------

DBMS_STATS.AUTO_SAMPLE_SIZE

SELECT DBMS_STATS.get_prefs (‘stale_percent‘) STALE_PERCENT FROM DUAL;

STALE_PERCENT

------------------------------

10

Regards,

Sebastian S.

我有同样的问题 显示 0 喜欢(0)

  • 204 查看
  • 标签:
  • 回复

平均用户评级: 无评分 (0 评级)

平均用户评级

无评分

(0 评级)

您的评级:

评级 差(1,最高值为 5)评级 中下(2,最高值为 5)评级 中等(3,最高值为 5)评级 中上(4,最高值为 5)评级 优(5,最高值为 5)

###########3

深入分区表的增量统计信息收集技术(incremetal statistics collection)

http://blog.itpub.net/53956/viewspace-1372944

############5

Oracle Database - Enterprise Edition - Version 11.2.0.2 and later
Information in this document applies to any platform.

SYMPTOMS

  • A large database was upgraded (RAC or non-RAC)  with large AWR repository
  • DBMS_STATS.GATHER_DICTIONARY_STATS() command was ran and took many hours without generation of data dictionary statistics
  • Stats gathering seems to be stuck in following command on "SYS"."WRH$_ACTIVE_SESSION_HISTORY" table as per 10046 event tracing at level 12 on the session:

    select substrb(dump(val,16,0,32),1,120) ep, cnt
    from
    (select /*+ no_expand_table(t) index_rs(t) no_parallel(t) no_parallel_index(t) dbms_stats cursor_sharing_exact use_weak_name_resl dynamic_sampling(0) no_monitoring no_substrb_pad */max(substrb("CLIENT_ID",1,32)) val,count(*) cnt
    from "SYS"."WRH$_ACTIVE_SESSION_HISTORY" sample ( .3361644492) t
    where substrb("CLIENT_ID",1,32) is not null ...

CAUSE

Depending on how much AWR data has been kept in the database, "SYS"."WRH$_ACTIVE_SESSION_HISTORY"  table can be one of the largest SYS objects in some databases.  As a result, it can take long time to gather statistics for this table.

SOLUTION

1.Exclude any AWR repository table like  WRH$_ACTIVE_SESSION_HISTORY which is known to take a long time to gather statistics:

$ sqlplus / as sysdba

EXEC DBMS_STATS.LOCK_TABLE_STATS(‘SYS‘, ‘WRH$_ACTIVE_SESSION_HISTORY‘);

2. Gather stats only for the objects with 10% or more data changes using options => ‘GATHER STALE option:

$ sqlplus / as sysdba

EXEC DBMS_STATS.GATHER_DICTIONARY_STATS ( options => ‘GATHER STALE‘);

原文地址:https://www.cnblogs.com/feiyun8616/p/9193034.html

时间: 2024-08-29 16:30:12

分区表的统计信息收集策略的相关文章

Oracle执行计划突变诊断之统计信息收集问题

Oracle执行计划突变诊断之统计信息收集问题 1.  情形描述 DB version:11.2.0.4 WITH SQL1 AS  (SELECT LAC,          CI,          TO_NUMBER(C.LONGITUDE) LONGITUDE,          TO_NUMBER(C.LATITUDE) LATITUDE     FROM MB_SYS_CELL_INFO C    WHERE C.CONTY_NAME = '道孚县'), SQL2 AS  (SELE

SQL Server 统计信息维护策略的选择

SQL Server 统计信息维护策略的选择 问题描述: 在对OLTP系统的一个上千万的表做归档后,循环分批删除源表数据时,业务应用收到超时告警,如下: V1.1.1.1: ****Process - QueryTransactionFor****: 23075129 Timeout expired.   The timeout period elapsed prior to completion of the operation or the server is not responding.

Oracle 统计信息收集

统计信息收集 1:创建分析表 $ cd $ORACLE_HOME/rdbms/admin $ sqlplus / as sysdba > @utlxplan.sql 2:为了方便,可以创建一个同义表(默认情况下只有sys用户可以使用) SQL> create public synonym plan_tables for plan_table; 3:把这个表的权限给所有人,也可以给指定的人 SQL> grant all on plan_tables to public; 4:创建plust

10G之后统计信息收集后为什么执行计划不会被立马淘汰

在10G之前,使用DBMS_STATS收集统计信息将会导致与此对象相关的游标失效,下次执行此 的时候将会进行HARD PARSE,除非收集的时候NO_INVALIDATE设置为TRUE. 由于硬解析会消耗大量的CPU,还会导致大量的library cache 和 shared pool 的LATCH竞争,因此 如果由于统计信息收集导致大量的的游标失效,可能会带来HARD PARSE风暴,造成系统的负担. 但是如果采用NO_INVALIDATE=TRUE的方法,由于游标不失效,游标无法利用到新的统

MySQL 5.6为什么关闭元数据统计信息自动更新&统计信息收集源代码探索

问题描述: MySQL 5.5.15 原sql如下: select constraint_schema,table_name,constraint_name,constraint_type from information_schema.table_constraints where table_schema not in ('information_schema', 'mysql', 'test','performance_schema'); 不只是上面提到的table_constraints

oracle 12c 关闭统计信息收集和启用统计信息收集

oracle 12c 关闭统计信息收集和启用统计信息收集 --关闭统计信息 col client_name for a60 select client_name,status from DBA_AUTOTASK_CLIENT; CLIENT_NAME STATUS ------------------------------------------------------------ ---------------- auto optimizer stats collection ENABLED

Oracle 分区表的统计信息实例

ORACLE的统计信息在执行SQL的过程中扮演着非常重要的作用,而且ORACLE在表的各个层次都会有不同的统计信息,通过这些统计信息来描述表的,列的各种各样的统计信息.下面通过一个复合分区表来说明一些常见的和不常见的统计信息. 搭建测试用例 --创建表. SQL> create table test partition by range(object_id) subpartition by hash(object_type) subpartitions 4 (partition p1 value

ORACLE 收集统计信息

原文链接:http://blog.itpub.net/2317695/viewspace-1226138/ 今天网上看到一篇关于收集统计信息的文章,还不错,特转载下来. 1.     理解什么是统计信息 优化器统计信息就是一个更加详细描述数据库和数据库对象的集合,这些统计信息被用于查询优化器,让其为每条SQL语句选择最佳的执行计划.优化器统计信息包括: ·         表的统计信息 o   行数 o   Block数 o   行平均长度 ·         列的统计信息 o   列中不同值的

Oracle里收集与查看统计信息的方法

Oracle数据库里的统计信息是这样的一组数据:它存储在数据字典里,且从多个维度描述了Oracle数据库里对象的详细信息.CBO会利用这些统计信息来计算目标SQL各种可能的.不同的执行路径的成本,并从中选择一条成本值最小的执行路径来作为目标SQL的执行计划. Oracle数据库里的统计信息可以分为如下6种类型: 表的统计信息 索引的统计信息 列的统计信息 系统统计信息 数据字典统计信息 内部对象统计信息 表的统计信息用于描述Oracle数据库里表的详细信息,它包含了一些典型的维度,如记录数.表块