Perform this procedure to prepare a cluster for live upgrade.
Perform the following tasks:
Ensure that the configuration meets the requirements for upgrade. See Upgrade Requirements and Software Support Guidelines.
Have available the installation media, documentation, and patches for all software products that you are upgrading, including the following software:
Sun Cluster 3.2 framework
Sun Cluster 3.2 data services (agents)
Applications that are managed by Sun Cluster 3.2 data-services
VERITAS Volume Manager, if applicable
See Patches and Required Firmware Levels in Sun Cluster 3.2 Release Notes for Solaris OS for the location of patches and installation instructions.
If you use role-based access control (RBAC) instead of superuser to access the cluster nodes, ensure that you can assume an RBAC role that provides authorization for all Sun Cluster commands. This series of upgrade procedures requires the following Sun Cluster RBAC authorizations if the user is not superuser:
See Role-Based Access Control (Overview) in System Administration Guide: Security Services for more information about using RBAC roles. See the Sun Cluster man pages for the RBAC authorization that each Sun Cluster subcommand requires.
Ensure that the cluster is functioning normally.
View the current status of the cluster by running the following command from any node.
See the scstat(1M) man page for more information.
Search the /var/adm/messages log on the same node for unresolved error messages or warning messages.
Check the volume-manager status.
If necessary, notify users that cluster services will be temporarily interrupted during the upgrade.
Service interruption will be approximately the amount of time that your cluster normally takes to switch services to another node.
Become superuser on a node of the cluster.
For uninstallation procedures, see the documentation for your version of Sun Cluster Geographic Edition software.
For a two-node cluster that uses Sun StorEdge Availability Suite software or Sun StorageTek Availability Suite software, ensure that the configuration data for availability services resides on the quorum disk.
The configuration data must reside on a quorum disk to ensure the proper functioning of Availability Suite after you upgrade the cluster software.
Become superuser on a node of the cluster that runs Availability Suite software.
Identify the device ID and the slice that is used by the Availability Suite configuration file.
phys-schost# /usr/opt/SUNWscm/sbin/dscfg /dev/did/rdsk/dNsS
In this example output, N is the device ID and S the slice of device N.
Identify the existing quorum device.
phys-schost# scstat -q -- Quorum Votes by Device -- Device Name Present Possible Status ----------- ------- -------- ------ Device votes: /dev/did/rdsk/dQsS 1 1 Online
In this example output, dQsS is the existing quorum device.
If the quorum device is not the same as the Availability Suite configuration-data device, move the configuration data to an available slice on the quorum device.
phys-schost# dd if=`/usr/opt/SUNWesm/sbin/dscfg` of=/dev/did/rdsk/dQsS
You must use the name of the raw DID device, /dev/did/rdsk/, not the block DID device, /dev/did/dsk/.
If you moved the configuration data, configure Availability Suite software to use the new location.
As superuser, issue the following command on each node that runs Availability Suite software.
phys-schost# /usr/opt/SUNWesm/sbin/dscfg -s /dev/did/rdsk/dQsS
Ensure that all shared data is backed up.
Ensure that each system disk is backed up.
Perform a live upgrade of the Solaris OS, Sun Cluster 3.2 software, and other software. Go to How to Upgrade the Solaris OS and Sun Cluster 3.2 Software (Live Upgrade).