Skip Navigation Links | |
Exit Print View | |
Oracle Solaris Cluster System Administration Guide Oracle Solaris Cluster 4.1 |
1. Introduction to Administering Oracle Solaris Cluster
2. Oracle Solaris Cluster and RBAC
3. Shutting Down and Booting a Cluster
4. Data Replication Approaches
5. Administering Global Devices, Disk-Path Monitoring, and Cluster File Systems
7. Administering Cluster Interconnects and Public Networks
Overview of Administering the Cluster
How to Change the Cluster Name
How to Map Node ID to Node Name
How to Work With New Cluster Node Authentication
How to Reset the Time of Day in a Cluster
SPARC: How to Display the OpenBoot PROM (OBP) on a Node
How to Change the Node Private Hostname
How to Put a Node Into Maintenance State
How to Bring a Node Out of Maintenance State
How to Uninstall Oracle Solaris Cluster Software From a Cluster Node
Troubleshooting a Node Uninstallation
Unremoved Cluster File System Entries
Unremoved Listing in Device Groups
Creating, Setting Up, and Managing the Oracle Solaris Cluster SNMP Event MIB
How to Enable an SNMP Event MIB
How to Disable an SNMP Event MIB
How to Change an SNMP Event MIB
How to Enable an SNMP Host to Receive SNMP Traps on a Node
How to Disable an SNMP Host From Receiving SNMP Traps on a Node
How to Add an SNMP User on a Node
How to Remove an SNMP User From a Node
How to Configure Load Limits on a Node
Changing Port Numbers for Services or Management Agents
How to Use the Common Agent Container to Change the Port Numbers for Services or Management Agents
Performing Zone Cluster Administrative Tasks
How to Add a Network Address to a Zone Cluster
How to Remove a File System From a Zone Cluster
How to Remove a Storage Device From a Zone Cluster
Running an Application Outside the Global Cluster
How to Take a Solaris Volume Manager Metaset From Nodes Booted in Noncluster Mode
How to Save the Solaris Volume Manager Software Configuration
How to Purge the Corrupted Diskset
How to Recreate the Solaris Volume Manager Software Configuration
10. Configuring Control of CPU Usage
This section contains troubleshooting procedures that you can use for testing purposes.
Use this procedure to run an application outside the global cluster for testing purposes.
phys-schost# clquorum show
phys-schost# clquorum add did
phys-schost# clqorum remove did
The following command finds the Persistent Group Reservation Emulation (PGRE) keys. If there are no keys on the disk, an errno=22 message is displayed.
# /usr/cluster/lib/sc/pgre -c pgre_inkeys -d /dev/did/rdsk/dids2
After you locate the keys, scrub the PGRE keys.
# /usr/cluster/lib/sc/pgre -c pgre_scrub -d /dev/did/rdsk/dids2
Caution - If you scrub the active quorum device keys from the disk, the cluster will panic on the next reconfiguration with a Lost operational quorum message. |
phys-schost# clresourcegroup evacuate -n targetnode
phys-schost# clresourcegroup offline resourcegroupname
phys-schost# clresource disable resourcename
phys-schost# clresourcegroup unmanage resourcegroupname
phys-schost# cldevicegroup offline devicegroupname
phys-schost# cldevicegroup disable devicegroupname
phys-schost# reboot -x
phys-schost# svcs -x
Run the following command on all disks in the metasets.
phys-schost# /usr/cluster/lib/sc/scsi -c inkeys -d /dev/did/rdsk/dids2
phys-schost# /usr/cluster/lib/sc/scsi -c scrub -d /dev/did/rdsk/dids2
phys-schost# metaset -s name -C take -f
phys-schost# mount device mountpoint
phys-schost# reboot
phys-schost# cldevicegroup online -e devicegroupname
phys-schost# clresourcegroup online -eM resourcegroupname
Use this procedure if a diskset is corrupted or in a state that the nodes in the cluster are unable to take ownership of the diskset. If your attempts to clear the state have failed, use this procedure as a last attempt to fix the diskset.
These procedures work for Solaris Volume Manager metasets and multi-owner Solaris Volume Manager metasets.
Restoring a disk set from scratch can be time-consuming and error prone. A better alternative is to use the metastat command to regularly back up replicas or use Oracle Explorer (SUNWexplo) to create a backup. You can then use the saved configuration to recreate the diskset. You should save the current configuration into files (using the prtvtoc and the metastat commands), and then recreate the disk set and its components. See How to Recreate the Solaris Volume Manager Software Configuration.
# /usr/sbin/prtvtoc /dev/global/rdsk/diskname > /etc/lvm/diskname.vtoc
# /bin/cp /etc/lvm/md.tab /etc/lvm/md.tab_ORIGINAL
# /usr/sbin/metastat -p -s setname >> /etc/lvm/md.tab
Note - Other configuration files, such as the /etc/vfstab file, might reference the Solaris Volume Manager software. This procedure assumes that an identical Solaris Volume Manager software configuration is rebuilt and therefore, the mount information is the same. If Oracle Explorer (SUNWexplo) is run on a node that owns the set, it retrieves the prtvtoc and metaset —p information.
Purging a set from a node or all nodes removes the configuration. To purge a diskset from a node, the node must not have ownership of the diskset.
# /usr/sbin/metaset -s setname -P
Running this command removes the diskset information from the database replicas, as well as the Oracle Solaris Cluster repository. The -P and -C options allow a diskset to be purged without the need to completely rebuild the Solaris Volume Manager environment.
Note - If a multi-owner diskset is purged while the nodes were booted out of cluster mode, you might need to remove the information from the dcs configuration files.
# /usr/cluster/lib/sc/dcs_config -c remove -s setname
For more information, see the dcs_config(1M) man page.
# /usr/sbin/metaset -s setname -C purge
You should generally use the -P option, rather than the -C option. Using the -C option can cause a problem recreating the diskset because the Oracle Solaris Cluster software still recognizes the diskset.
# /usr/cluster/lib/sc/dcs_config -c remove -s setname
If the purge options fail, verify that you installed the latest kernel and metadevice updates and contact My Oracle Support.
Use this procedure only if you experience a complete loss of your Solaris Volume Manager software configuration. The steps assume that you have saved your current Solaris Volume Manager configuration and its components and purged the corrupted diskset.
Note - Mediators should be used only on two-node clusters.
# /usr/sbin/metaset -s setname -a -h nodename1 nodename2
If this is a multi-owner diskset, use the following command to create a new diskset.
/usr/sbin/metaset -s setname -aM -h nodename1 nodename2
/usr/sbin/metaset -s setname -a -m nodename1 nodename2
/usr/sbin/metaset -s setname -a /dev/did/rdsk/diskname /dev/did/rdsk/diskname
However, if you are recreating a set to recover, you should format the disks according to a saved configuration in the /etc/lvm/diskname.vtoc file. For example:
# /usr/sbin/fmthard -s /etc/lvm/d4.vtoc /dev/global/rdsk/d4s2
# /usr/sbin/fmthard -s /etc/lvm/d8.vtoc /dev/global/rdsk/d8s2
You can run this command on any node.
# /usr/sbin/metainit -s setname -n -a metadevice
# /usr/sbin/metainit -s setname -a metadevice
# /usr/sbin/fsck -n /dev/md/setname/rdsk/metadevice
If the fsck command displays only a few errors, such as superblock count, then the device was probably reconstructed correctly. You can then run the fsck command without the -n option. If multiple errors appear, verify that you reconstructed the metadevice correctly. If you have, review the fsck errors to determine if the filesystem can be recovered. If it cannot, you should restore the data from a backup.
# /usr/sbin/metastat -p >> /etc/lvm/md.tab