8 Monitoring Performance and Troubleshooting

The Global Cache Block Access Latency chart shows data for two different types of data block requests: current and consistent-read (CR) blocks. When you update data in the database, Oracle Database must locate the most recent version of the data block that
contains the data, which is called the current block. If you perform a query, then only data committed before the query began is visible to the query. Data blocks that were changed after the start of the query are reconstructed from data in the undo segments.
The reconstructed data is made available to the query in the form of a consistent-read block.

-----这里官档说了什么叫当前块,什么叫CR块

当前块:发生在更新数据库的时候,如果更新数据到数据库,oracle需要知道包含这个数据的数据库的最近的版本,这就叫做current block.

CR块:如果你执行了查询,只有在查询之前提交的数据是可见的.在查询之后数据块开始改变的,就需要从undo中构造一致块.

Concurrent read and write activity on shared data in a cluster is a frequently occurring activity. Depending on the service requirements, this activity does not usually cause performance problems. However, when global cache requests cause a performance problem,
optimizing SQL plans and the schema to improve the rate at which data blocks are located in the local buffer cache, and minimizing I/O is a successful strategy for performance tuning. If the latency for consistent-read and current block requests reaches 10
milliseconds, then your first step in resolving the problem should be to go to the Cluster Cache Coherency page for more detailed information.

-----如果一致地读取和当前块请求的延迟到达10毫秒,就需要进一步查看原因了。

Using CRSCTL to Diagnose Cluster Issues

About the Oracle Clusterware Alert Log

The Oracle Clusterware should be the first place to look for serious errors. It often contains references to other diagnostic logs that provide detailed information on a specific component. The location of the Oracle Clusterware log file is CRS_home/log/hostname/alerthostname.log,
where CRS_home is the directory in which Oracle Clusterware was installed and hostname is the host name of the local node.

----CRS的alert日志默认放在CRS_home/log/hostname/alerthostname.log,hostname为本地节点的名字

Running the Oracle Clusterware Diagnostics Collection Script

To run the Oracle Clusterware Diagnostics Collection script:

In a command window, log in to the operating system as the root user.

Run the diagcollection.pl script from the operating system prompt as follows, where Grid_home is the home directory of your Oracle Grid Infrastructure for a cluster installation:

# Grid_home/bin/diagcollection.pl --collect

执行例子如下:

[[email protected] bin]# ./diagcollection.pl --collect

Production Copyright 2004, 2010, Oracle.  All rights reserved

Cluster Ready Services (CRS) diagnostic collection tool

The following CRS diagnostic archives will be created in the local directory.

crsData_elvis111_20150504_0728.tar.gz -> logs,traces and cores from CRS home. Note: core files will be packaged only with the --core option.

ocrData_elvis111_20150504_0728.tar.gz -> ocrdump, ocrcheck etc

coreData_elvis111_20150504_0728.tar.gz -> contents of CRS core files in text format

osData_elvis111_20150504_0728.tar.gz -> logs from Operating System

lsInventory_elvis111_20150504_0728 ->Opatch lsinventory details

Collecting crs data

/bin/tar: log/elvis111/gipcd/gipcd.log: file changed as we read it

/bin/tar: log/elvis111/agent/crsd/orarootagent_root/orarootagent_root.log: file changed as we read it

Collecting OCR data

Collecting information from core files

warning: exec file is newer than core file.

Cannot access memory at address 0x642520726f727265

warning: exec file is newer than core file.

Cannot access memory at address 0x642520726f727265

Collecting lsinventory details

The following diagnostic archives will be created in the local directory.

acfsData_elvis111_20150504_0728.tar.gz -> logs from acfs log.

Collecting acfs data

Collecting OS logs

Collecting sysconfig data

-rw-r--r-- 1 root root      11933011 May  4 07:29 crsData_elvis111_20150504_0728.tar.gz

-rw-r--r-- 1 root root         25331 May  4 07:30 ocrData_elvis111_20150504_0728.tar.gz

-rw-r--r-- 1 root root         58268 May  4 07:30 coreData_elvis111_20150504_0728.tar.gz

-rw-r--r-- 1 root root          7634 May  4 07:31 lsInventory_elvis111_20150504_0728

-rw-r--r-- 1 root root           847 May  4 07:31 acfsData_elvis111_20150504_0728.tar.gz

-rw-r--r-- 1 root root        133061 May  4 07:31 osData_elvis111_20150504_0728.tar.gz

-rw-r--r-- 1 root root         29113 May  4 07:31 sysconfig_elvis111_20150504_0728.txt

------------可以看到这里系统自动打包了很多CRS日志,系统版本等信息,在故障处理或者提交SR应该非常有用.

-------下面是debug的方法,在对CRS的进行问题定位的时候会非常有用,建议好好仔细看下

Enabling Debugging of Oracle Clusterware Components

To enable debugging of Oracle Clusterware components:

1.In a command window, log in to the operating system as the root user.

2.Use the following command to obtain the module names for a component, where component_name is crs, evm, css or the name of the component for which you want to enable debugging:

# crsctl lsmodules component_name

For example, viewing the modules of the css component might return the following results:

# crsctl lsmodules css

The following are the CSS modules ::

CSSD

COMMCRS

COMMNS

例子输出如下:

[[email protected] grid]# crsctl lsmodules css

List CSSD Debug Module: CLSF

List CSSD Debug Module: CSSD

List CSSD Debug Module: GIPCCM

List CSSD Debug Module: GIPCGM

List CSSD Debug Module: GIPCNM

List CSSD Debug Module: GPNP

List CSSD Debug Module: OLR

List CSSD Debug Module: SKGFD

[[email protected] grid]# crsctl lsmodules crs

List CRSD Debug Module: AGENT

List CRSD Debug Module: AGFW

List CRSD Debug Module: CLSFRAME

List CRSD Debug Module: CLSVER

List CRSD Debug Module: CLUCLS

List CRSD Debug Module: COMMCRS

List CRSD Debug Module: COMMNS

List CRSD Debug Module: CRSAPP

List CRSD Debug Module: CRSCCL

List CRSD Debug Module: CRSCEVT

List CRSD Debug Module: CRSCOMM

List CRSD Debug Module: CRSD

List CRSD Debug Module: CRSEVT

List CRSD Debug Module: CRSMAIN

List CRSD Debug Module: CRSOCR

List CRSD Debug Module: CRSPE

List CRSD Debug Module: CRSPLACE

List CRSD Debug Module: CRSRES

List CRSD Debug Module: CRSRPT

List CRSD Debug Module: CRSRTI

List CRSD Debug Module: CRSSE

List CRSD Debug Module: CRSSEC

List CRSD Debug Module: CRSTIMER

List CRSD Debug Module: CRSUI

List CRSD Debug Module: CSSCLNT

List CRSD Debug Module: OCRAPI

List CRSD Debug Module: OCRASM

List CRSD Debug Module: OCRCAC

List CRSD Debug Module: OCRCLI

List CRSD Debug Module: OCRMAS

List CRSD Debug Module: OCRMSG

List CRSD Debug Module: OCROSD

List CRSD Debug Module: OCRRAW

List CRSD Debug Module: OCRSRV

List CRSD Debug Module: OCRUTL

List CRSD Debug Module: SuiteTes

List CRSD Debug Module: UiServer

[[email protected] grid]#

3.Use CRSCTL as follows, where component_name is the name of the Oracle Clusterware component for which you want to enable debugging, module is the name of module, and debugging_level is a number from 1 to 5:

# crsctl debug log component module:debugging_level

For example, to enable the lowest level of tracing for the CSSD module of the css component, you would use the following command:

# crsctl debug log css CSSD:1

--------------进行这个debug之后,会在相应的log里面能看到debug的日志

Enabling and Disabling Oracle Clusterware Daemons

When the Oracle Clusterware daemons are enabled, they start automatically when the node is started. To prevent the daemons from starting automatically, you can disable them using crsctl commands.

To enable automatic startup for all Oracle Clusterware daemons:

1.In a command window, log in to the operating system as the root user.

2.Run the following CRSCTL command:

# crsctl enable crs

To disable automatic startup for all Oracle Clusterware daemons:

1.In a command window, log in to the operating system as the root user.

2.Run the following CRSCTL command:

# crsctl disable crs

----通过上述命令可以开启或者关闭CRS自动启动的,默认是可以自动启动的.

-------利用cluvy在诊断RAC数据库配置方面非常有用,建议仔细研究下

Using the Cluster Verification Utility to Diagnose Problems

使用CVU可以诊断各种各样的配置问题

This section contains the following topics:

Verifying the Existence of Node Applications

Verifying the Integrity of Oracle Clusterware Components

Verifying the Integrity of the Oracle Cluster Registry

Verifying the Integrity of Your Entire Cluster

Checking the Settings for the Interconnect

Enabling Tracing

可以诊断如下问题:

1.验证节点应用程序的存在性

2.验证Oracle集群组件的完整性

3.验证OCR的完整性

4.验证整个集群的完整性

5.检查内网的设置

6.可以开启trace功能

To verify the existence of node applications:

1.In a command window, log in to the operating system as the user who owns the Oracle Clusterware software installation.

2.Use the comp nodeapp command of CVU, using the following syntax:

cluvfy comp nodeapp [ -n node_list] [-verbose]

where node_list represents the nodes to check.

例子输出如下:

如果不知道node_list是什么,可以使用olsnodes命令看下

[[email protected] ~]$ cluvfy comp nodeapp -n elvis111,elvis112 -verbose

Verifying node application existence

Checking node application existence...

Checking existence of VIP node application (required)

Node Name     Required                  Running?                  Comment

------------  ------------------------  ------------------------  ----------

elvis112      yes                       yes                       passed

elvis111      yes                       yes                       passed

VIP node application check passed

Checking existence of NETWORK node application (required)

Node Name     Required                  Running?                  Comment

------------  ------------------------  ------------------------  ----------

elvis112      yes                       yes                       passed

elvis111      yes                       yes                       passed

NETWORK node application check passed

Checking existence of GSD node application (optional)

Node Name     Required                  Running?                  Comment

------------  ------------------------  ------------------------  ----------

elvis112      no                        no                        exists

elvis111      no                        no                        exists

GSD node application is offline on nodes "elvis112,elvis111"

Checking existence of ONS node application (optional)

Node Name     Required                  Running?                  Comment

------------  ------------------------  ------------------------  ----------

elvis112      no                        yes                       passed

elvis111      no                        yes                       passed

ONS node application check passed

Verification of node application existence was successful.

[[email protected] ~]$

3.If the cluvfy command returns the value of UNKNOWN for a particular node, then CVU cannot determine whether a check passed or failed. Determine if the failure was caused by one of the following reasons:

The node is down.

Executable files that CVU requires are missing in the CRS_home/bin directory or the Oracle_home/bin directory.

The user account that ran CVU does not have permissions to run common operating system executable files on the node.

The node is missing an operating system patch or required package.

The kernel parameters on that node were not configured correctly and CVU cannot obtain the operating system resources required to perform its checks.

Verifying the Integrity of Oracle Clusterware Components

You use the CVU comp crs command to verify the existence of all the Oracle Clusterware components.

To verify the integrity of Oracle Clusterware components:

1.In a command window, log in to the operating system as the user who owns the Oracle Clusterware software installation.

2.Use the comp crs command of CVU, using the following syntax:

cluvfy comp crs [ -n node_list] [-verbose]

where node_list represents the nodes to check.

Verifying the Integrity of the Oracle Cluster Registry

You use the CVU comp ocr command to verify the integrity of the Oracle Clusterware registry.

To verify the integrity of the Oracle Clusterware registry:

1.In a command window, log in to the operating system as the user who owns the Oracle Clusterware software installation.

2.Use the comp ocr command of CVU, using the following syntax:

cluvfy comp ocr [ -n node_list] [-verbose]

where node_list represents the nodes to check.

例子输入如下:

[[email protected] ~]$ cluvfy comp crs -n elvis111,elvis112 -verbose

Verifying CRS integrity

Checking CRS integrity...

Clusterware version consistency passed

The Oracle Clusterware is healthy on node "elvis112"

The Oracle Clusterware is healthy on node "elvis111"

CRS integrity check passed

Verification of CRS integrity was successful.

Verifying the Integrity of Your Entire Cluster

You use the CVU comp clu command to check that all nodes in the cluster have the same view of the cluster configuration.

To verify the integrity of your cluster:

1.In a command window, log in to the operating system as the user who owns the Oracle Clusterware software installation.

2.Use the comp clu command of CVU, using the following syntax:

cluvfy comp clu [-verbose]

例子输入如下:

[[email protected] ~]$ cluvfy comp clu -verbose

Verifying cluster integrity

Checking cluster integrity...

Node Name

------------------------------------

elvis112

elvis111

Cluster integrity check passed

Verification of cluster integrity was successful.

[[email protected] ~]$

Checking the Settings for the Interconnect

To check the settings for the interconnect:

1.In a command window, log in to the operating system as the user who owns the Oracle Clusterware software installation.

2.To verify the accessibility of the cluster nodes, specified by node_list, from the local node or from any other cluster node, specified by srcnode, use the component verification command nodereach as follows:

cluvfy comp nodereach -n node_list [ -srcnode node ] [-verbose]

例子输入如下:

[[email protected] ~]$ cluvfy comp nodereach -n elvis111,elvis112 -srcnode elvis112 -verbose

Verifying node reachability

Checking node reachability...

Check: Node reachability from node "elvis112"

Destination Node                      Reachable?

------------------------------------  ------------------------

elvis111                              yes

elvis112                              yes

Result: Node reachability check passed from node "elvis112"

Verification of node reachability was successful.

3.To verify the connectivity among the nodes, specified by node_list, through the available network interfaces from the local node or from any other cluster node, use the comp nodecon command as shown in the following example:

cluvfy comp nodecon -n node_list -verbose

例子输出如下:

[[email protected] ~]$ cluvfy comp nodecon -n elvis111,elvis112 -verbose

Verifying node connectivity

Checking node connectivity...

Checking hosts config file...

Node Name                             Status

------------------------------------  ------------------------

elvis112                              passed

elvis111                              passed

Verification of the hosts config file successful

Interface information for node "elvis112"

Name   IP Address      Subnet          Gateway         Def. Gateway    HW Address        MTU

------ --------------- --------------- --------------- --------------- ----------------- ------

eth0   192.168.56.112  192.168.56.0    0.0.0.0         192.168.56.1    08:00:27:72:C8:E4 1500

eth0   192.168.56.115  192.168.56.0    0.0.0.0         192.168.56.1    08:00:27:72:C8:E4 1500

eth0   192.168.56.114  192.168.56.0    0.0.0.0         192.168.56.1    08:00:27:72:C8:E4 1500

eth1   172.168.1.112   172.168.1.0     0.0.0.0         192.168.56.1    08:00:27:ED:96:2A 1500

eth1   169.254.215.143 169.254.0.0     0.0.0.0         192.168.56.1    08:00:27:ED:96:2A 1500

Interface information for node "elvis111"

Name   IP Address      Subnet          Gateway         Def. Gateway    HW Address        MTU

------ --------------- --------------- --------------- --------------- ----------------- ------

eth0   192.168.56.111  192.168.56.0    0.0.0.0         192.168.56.1    08:00:27:FA:68:C3 1500

eth0   192.168.56.113  192.168.56.0    0.0.0.0         192.168.56.1    08:00:27:FA:68:C3 1500

eth1   172.168.1.111   172.168.1.0     0.0.0.0         192.168.56.1    08:00:27:91:EA:DE 1500

eth1   169.254.250.101 169.254.0.0     0.0.0.0         192.168.56.1    08:00:27:91:EA:DE 1500

Check: Node connectivity of subnet "192.168.56.0"

Source                          Destination                     Connected?

------------------------------  ------------------------------  ----------------

elvis112[192.168.56.112]        elvis112[192.168.56.115]        yes

elvis112[192.168.56.112]        elvis112[192.168.56.114]        yes

elvis112[192.168.56.112]        elvis111[192.168.56.111]        yes

elvis112[192.168.56.112]        elvis111[192.168.56.113]        yes

elvis112[192.168.56.115]        elvis112[192.168.56.114]        yes

elvis112[192.168.56.115]        elvis111[192.168.56.111]        yes

elvis112[192.168.56.115]        elvis111[192.168.56.113]        yes

elvis112[192.168.56.114]        elvis111[192.168.56.111]        yes

elvis112[192.168.56.114]        elvis111[192.168.56.113]        yes

elvis111[192.168.56.111]        elvis111[192.168.56.113]        yes

Result: Node connectivity passed for subnet "192.168.56.0" with node(s) elvis112,elvis111

Check: TCP connectivity of subnet "192.168.56.0"

Source                          Destination                     Connected?

------------------------------  ------------------------------  ----------------

elvis111:192.168.56.111         elvis112:192.168.56.112         passed

elvis111:192.168.56.111         elvis112:192.168.56.115         passed

elvis111:192.168.56.111         elvis112:192.168.56.114         passed

elvis111:192.168.56.111         elvis111:192.168.56.113         passed

Result: TCP connectivity check passed for subnet "192.168.56.0"

Check: Node connectivity of subnet "172.168.1.0"

Source                          Destination                     Connected?

------------------------------  ------------------------------  ----------------

elvis112[172.168.1.112]         elvis111[172.168.1.111]         yes

Result: Node connectivity passed for subnet "172.168.1.0" with node(s) elvis112,elvis111

Check: TCP connectivity of subnet "172.168.1.0"

Source                          Destination                     Connected?

------------------------------  ------------------------------  ----------------

elvis111:172.168.1.111          elvis112:172.168.1.112          passed

Result: TCP connectivity check passed for subnet "172.168.1.0"

Check: Node connectivity of subnet "169.254.0.0"

Source                          Destination                     Connected?

------------------------------  ------------------------------  ----------------

elvis112[169.254.215.143]       elvis111[169.254.250.101]       yes

Result: Node connectivity passed for subnet "169.254.0.0" with node(s) elvis112,elvis111

Check: TCP connectivity of subnet "169.254.0.0"

Source                          Destination                     Connected?

------------------------------  ------------------------------  ----------------

elvis111:169.254.250.101        elvis112:169.254.215.143        passed

Result: TCP connectivity check passed for subnet "169.254.0.0"

Interfaces found on subnet "192.168.56.0" that are likely candidates for VIP are:

elvis112 eth0:192.168.56.112 eth0:192.168.56.115 eth0:192.168.56.114

elvis111 eth0:192.168.56.111 eth0:192.168.56.113

Interfaces found on subnet "172.168.1.0" that are likely candidates for VIP are:

elvis112 eth1:172.168.1.112

elvis111 eth1:172.168.1.111

Interfaces found on subnet "169.254.0.0" that are likely candidates for VIP are:

elvis112 eth1:169.254.215.143

elvis111 eth1:169.254.250.101

WARNING:

Could not find a suitable set of interfaces for the private interconnect

Checking subnet mask consistency...

Subnet mask consistency check passed for subnet "192.168.56.0".

Subnet mask consistency check passed for subnet "172.168.1.0".

Subnet mask consistency check passed for subnet "169.254.0.0".

Subnet mask consistency check passed.

Result: Node connectivity check passed

Verification of node connectivity was successful.

When you issue the nodecon command as shown in the previous example, it instructs CVU to perform the following tasks:

1.Discover all the network interfaces that are available on the specified cluster nodes.

2.Review the corresponding IP addresses and subnets for the interfaces.

3.Obtain the list of interfaces that are suitable for use as VIPs and the list of interfaces to private interconnects.

4.Verify the connectivity among all the nodes through those interfaces.

When you run the nodecon command in verbose mode, it identifies the mappings between the interfaces, IP addresses, and subnets.

4.To verify the connectivity among the nodes through specific network interfaces, use the comp nodecon command with the -i option and specify the interfaces to be checked with the interface_list argument:

cluvfy comp nodecon -n node_list -i interface_list [-verbose]

例子如下:

[[email protected] ~]$ cluvfy comp nodecon -n elvis111,elvis112 -i interface_list -verbose

Verifying node connectivity

Checking node connectivity...

Checking hosts config file...

Node Name                             Status

------------------------------------  ------------------------

elvis112                              passed

elvis111                              passed

Verification of the hosts config file successful

Interface information for node "elvis112"

Name   IP Address      Subnet          Gateway         Def. Gateway    HW Address        MTU

------ --------------- --------------- --------------- --------------- ----------------- ------

eth0   192.168.56.112  192.168.56.0    0.0.0.0         192.168.56.1    08:00:27:72:C8:E4 1500

eth0   192.168.56.115  192.168.56.0    0.0.0.0         192.168.56.1    08:00:27:72:C8:E4 1500

eth0   192.168.56.114  192.168.56.0    0.0.0.0         192.168.56.1    08:00:27:72:C8:E4 1500

eth1   172.168.1.112   172.168.1.0     0.0.0.0         192.168.56.1    08:00:27:ED:96:2A 1500

eth1   169.254.215.143 169.254.0.0     0.0.0.0         192.168.56.1    08:00:27:ED:96:2A 1500

Interface information for node "elvis111"

Name   IP Address      Subnet          Gateway         Def. Gateway    HW Address        MTU

------ --------------- --------------- --------------- --------------- ----------------- ------

eth0   192.168.56.111  192.168.56.0    0.0.0.0         192.168.56.1    08:00:27:FA:68:C3 1500

eth0   192.168.56.113  192.168.56.0    0.0.0.0         192.168.56.1    08:00:27:FA:68:C3 1500

eth1   172.168.1.111   172.168.1.0     0.0.0.0         192.168.56.1    08:00:27:91:EA:DE 1500

eth1   169.254.250.101 169.254.0.0     0.0.0.0         192.168.56.1    08:00:27:91:EA:DE 1500

ERROR:

PRVG-11049 : Interface "interface_list" does not exist on nodes "elvis112,elvis111"

Check: Node connectivity for interface "interface_list"

Result: Node connectivity failed for interface "interface_list"

Checking subnet mask consistency...

Subnet mask consistency check passed for subnet "192.168.56.0".

Subnet mask consistency check passed for subnet "172.168.1.0".

Subnet mask consistency check passed for subnet "169.254.0.0".

Subnet mask consistency check passed.

Result: Node connectivity check failed

Verification of node connectivity was unsuccessful on all the specified nodes.

[[email protected] ~]$

For example, you can verify the connectivity among the nodes racnode1, racnode2, and racnode3, through the specific network interface eth0 by running the following command:

cluvfy comp nodecon -n racnode1,racnode2,racnode3 -i eth0 -verbose

例子如下:

[[email protected] ~]$ cluvfy comp nodecon -n elvis111,elvis112 -i eth0 -verbose

Verifying node connectivity

Checking node connectivity...

Checking hosts config file...

Node Name                             Status

------------------------------------  ------------------------

elvis112                              passed

elvis111                              passed

Verification of the hosts config file successful

Interface information for node "elvis112"

Name   IP Address      Subnet          Gateway         Def. Gateway    HW Address        MTU

------ --------------- --------------- --------------- --------------- ----------------- ------

eth0   192.168.56.112  192.168.56.0    0.0.0.0         192.168.56.1    08:00:27:72:C8:E4 1500

eth0   192.168.56.115  192.168.56.0    0.0.0.0         192.168.56.1    08:00:27:72:C8:E4 1500

eth0   192.168.56.114  192.168.56.0    0.0.0.0         192.168.56.1    08:00:27:72:C8:E4 1500

eth1   172.168.1.112   172.168.1.0     0.0.0.0         192.168.56.1    08:00:27:ED:96:2A 1500

eth1   169.254.215.143 169.254.0.0     0.0.0.0         192.168.56.1    08:00:27:ED:96:2A 1500

Interface information for node "elvis111"

Name   IP Address      Subnet          Gateway         Def. Gateway    HW Address        MTU

------ --------------- --------------- --------------- --------------- ----------------- ------

eth0   192.168.56.111  192.168.56.0    0.0.0.0         192.168.56.1    08:00:27:FA:68:C3 1500

eth0   192.168.56.113  192.168.56.0    0.0.0.0         192.168.56.1    08:00:27:FA:68:C3 1500

eth1   172.168.1.111   172.168.1.0     0.0.0.0         192.168.56.1    08:00:27:91:EA:DE 1500

eth1   169.254.250.101 169.254.0.0     0.0.0.0         192.168.56.1    08:00:27:91:EA:DE 1500

Check: Node connectivity for interface "eth0"

Source                          Destination                     Connected?

------------------------------  ------------------------------  ----------------

elvis112[192.168.56.112]        elvis112[192.168.56.115]        yes

elvis112[192.168.56.112]        elvis112[192.168.56.114]        yes

elvis112[192.168.56.112]        elvis111[192.168.56.111]        yes

elvis112[192.168.56.112]        elvis111[192.168.56.113]        yes

elvis112[192.168.56.115]        elvis112[192.168.56.114]        yes

elvis112[192.168.56.115]        elvis111[192.168.56.111]        yes

elvis112[192.168.56.115]        elvis111[192.168.56.113]        yes

elvis112[192.168.56.114]        elvis111[192.168.56.111]        yes

elvis112[192.168.56.114]        elvis111[192.168.56.113]        yes

elvis111[192.168.56.111]        elvis111[192.168.56.113]        yes

Result: Node connectivity passed for interface "eth0"

Check: TCP connectivity of subnet "192.168.56.0"

Source                          Destination                     Connected?

------------------------------  ------------------------------  ----------------

elvis111:192.168.56.111         elvis112:192.168.56.112         passed

elvis111:192.168.56.111         elvis112:192.168.56.115         passed

elvis111:192.168.56.111         elvis112:192.168.56.114         passed

elvis111:192.168.56.111         elvis111:192.168.56.113         passed

Result: TCP connectivity check passed for subnet "192.168.56.0"

Checking subnet mask consistency...

Subnet mask consistency check passed for subnet "192.168.56.0".

Subnet mask consistency check passed for subnet "172.168.1.0".

Subnet mask consistency check passed for subnet "169.254.0.0".

Subnet mask consistency check passed.

Result: Node connectivity check passed

Verification of node connectivity was successful.

Enabling Tracing

CVU does not generate trace files unless you enable tracing. The CVU trace files are created in the CRS_home/cv/log directory. Oracle RAC automatically rotates the log files, and the most recently created log file has the name cvutrace.log.0. You should remove
unwanted log files or archive them to reclaim disk space, if needed.

To enable tracing using CVU:

1.In a command window, log in to the operating system as the root user.

2.Set the environment variable SRVM_TRACE to true.

# set SRVM_TRACE=true; export SRVM_TRACE

Run the command to trace.

时间: 2024-11-13 06:57:00

8 Monitoring Performance and Troubleshooting的相关文章

hbase官方文档(转)

Apache HBase™ 参考指南  HBase 官方文档中文版 Copyright © 2012 Apache Software Foundation.保留所有权利. Apache Hadoop, Hadoop, MapReduce, HDFS, Zookeeper, HBase 及 HBase项目 logo 是Apache Software Foundation的商标. Revision History Revision 0.95-SNAPSHOT 2012-12-03T13:38 中文版

Migrating Oracle on UNIX to SQL Server on Windows

Appendices Published: April 27, 2005 On This Page Appendix A: SQL Server for Oracle Professionals Appendix B: Getting the Best Out of SQL Server 2000 and Windows Appendix C: Baselining Appendix D: Installing Common Drivers and Applications Installing

MYSQL术语表

MYSQL术语表 http://dev.mysql.com/doc/refman/5.6/en/glossary.html MySQL Glossary These terms are commonly used in information about the MySQL database server. This glossary originated as a reference for terminology about the InnoDB storage engine, and th

[转载]系统运维秘诀大分享专题

系统运维秘诀大分享专题 本专题整合收录了有关系统运维/系统管理员工作和个人成长方面的各种心得分享.经验总结.以及必须牢记的一些准则,适合所有在运维领域有追求的技术人阅读.有些分享的层次比较深,有些则是运维的基础课,但通过翻看他人的心得,相信你总能有所收获. 1 Dormando的系统运维秘诀三部曲... 4 1.1 技术篇... 4 1.1.1 为变化而设计.... 4 1.1.2 使用自动的,可重复的构建过程.... 4 1.1.3 使用冗余.... 4 1.1.4 使用备份.... 5 1.

.NetChajian

Code generation(代码自动生成) NVelocity CodeSmith X-Code .NET XGoF - NMatrix / DEVerest Compilation(编译工具) eXtensible C# - ResolveCorp Mono DotGNU - GNU Obfuscation(混淆加密) LSW-IL-Obfuscator - Lesser Software Demeanor for .NET - Wise Owl Salamander .NET Obfus

net辅助工具列表

最近对.net的辅助工具比较感兴趣,网上也已经有.net的十个必备工具的帖子了,这里提供了一份全面的工具列表: Code generation NVelocity CodeSmith X-Code .NET XGoF - NMatrix / DEVerest Compilation eXtensible C# - ResolveCorp Mono DotGNU - GNU Obfuscation LSW-IL-Obfuscator - Lesser Software Demeanor for .

ThreadPool has stuck threads

weblogic 10后台出现警告,原因:ThreadPool has stuck threads 在WEBLOGIC中如果一个线程执行时间超过了Stuck Thread Max Time规定的时间, WEBLOGIC会把它认为是STUCK线程,并记录在日志中. 我们也可以通过'Dump Thread Stacks' 来捕获STUCK状态的进程. Home > Summary of Servers > AdminServer>Monitoring>Performance>'D

Do waiting or suspended tasks tie up a worker thread?

https://blogs.msdn.microsoft.com/askjay/2012/07/29/do-waiting-or-suspended-tasks-tie-up-a-worker-thread/ jamesaskJuly 29, 20120 I had a discussion the other day with someone about worker threads and their relation to tasks.  I thought a quick demo mi

libvirtsAPI

mongodb远程服务器连接 mongo -uroot -p321 master.puppet.org:27017/admin 基于libvcirt的虚拟化guest domain overview: A guest domain may be transient, or persistent. A transient guest domain can only be managed whileit is running on the host and, when powered off, al