This chapter provides information about and procedures for shutting down and booting a global cluster, a zone cluster, and individual nodes. For information about booting a non-global zone, see Chapter 18, Planning and Configuring Non-Global Zones (Tasks), in System Administration Guide: Solaris Containers-Resource Management and Solaris Zones.
For a high-level description of the related procedures in this chapter, see How to Boot a Node in Noncluster Mode and Table 3–2.
The Sun Cluster cluster(1CL) shutdown command stops global cluster services in an orderly fashion and cleanly shuts down an entire global cluster. You can use the cluster shutdown command when moving the location of a global cluster, or to shut down the global cluster if an application error causes data corruption. The clzonecluster halt command stops a zone cluster that is running on a specific node or an entire zone cluster on all configured nodes. (You can also use the cluster shutdown command within a zone cluster.)
In the procedures in this chapter, phys-schost# reflects a global-cluster prompt. The clzonecluster interactive shell prompt is clzc:schost>.
Use the cluster shutdown command to ensure proper shutdown of the entire global cluster. The Solaris shutdown command is used with the clnode(1CL) evacuate command to shut down individual nodes. See How to Shut Down a Cluster or Shutting Down and Booting a Single Node in a Cluster for more information.
The cluster shutdown and the clzonecluster halt commands stop all nodes in a global cluster or zone cluster, respectively, by performing the following actions:
Takes all running resource groups offline.
Unmounts all cluster file systems for a global cluster or a zone cluster.
The cluster shutdown command shuts down active device services on a global cluster or a zone cluster.
The cluster shutdown command runs init 0 and brings all nodes on the cluster to the OpenBootTM PROM ok prompt on a SPARC based system or the Press any key to continue message on the GRUB menu of an x86 based system. The GRUB menus are described in more detail in Booting an x86 Based System by Using GRUB (Task Map) in System Administration Guide: Basic Administration. The clzonecluster halt command performs the zoneadm -z zoneclustername halt command to stop (but not shut down) the zones of the zone cluster.
If necessary, you can boot a node in noncluster mode so that the node does not participate in cluster membership. Noncluster mode is useful when installing cluster software or for performing certain administrative procedures. See How to Boot a Node in Noncluster Mode for more information.
Task |
Instructions |
---|---|
Stop the cluster. | |
Start the cluster by booting all nodes. The nodes must have a working connection to the cluster interconnect to attain cluster membership. | |
Reboot the cluster. |
You can shut down a global cluster, a zone cluster, or all zone clusters.
Do not use send brk on a cluster console to shut down a global-cluster node or a zone-cluster node. The command is not supported within a cluster.
The phys-schost# prompt reflects a global-cluster prompt. Perform this procedure on a global cluster.
This procedure provides the long forms of the Sun Cluster commands. Most commands also have short forms. Except for the long and short forms of the command names, the commands are identical. For a list of the commands and their short forms, see Appendix B, Sun Cluster Object-Oriented Commands.
If your global cluster or zone cluster is running Oracle Real Application Clusters (RAC), shut down all instances of the database on the cluster you are shutting down.
Refer to the Oracle RAC product documentation for shutdown procedures.
Become superuser or assume a role that provides solaris.cluster.admin RBAC authorization on any node in the cluster. Perform all steps in this procedure from a node of the global cluster.
Shut down the global cluster, the zone cluster, or all zone clusters.
Shut down the global cluster. This action also shuts down all zone clusters.
phys-schost# cluster shutdown -g0 -y |
Shut down a specific zone cluster.
phys-schost# clzonecluster halt zoneclustername |
Shut down all zone clusters.
phys-schost# clzonecluster halt + |
You can also use the cluster shutdown command within a zone cluster to shut down all zone clusters.
Verify that all nodes on the global cluster or zone cluster are showing the ok prompt on a SPARC based system or a GRUB menu on an x86 based system.
Do not power off any nodes until all nodes are at the ok prompt on a SPARC based system or in a boot subsystem on an x86 based system.
If necessary, power off the nodes of the global cluster.
The following example shuts down a zone cluster called sparse-sczone.
phys-schost# clzonecluster halt sparse-sczone Waiting for zone halt commands to complete on all the nodes of the zone cluster "sparse-sczone"... Sep 5 19:06:01 schost-4 cl_runtime: NOTICE: Membership : Node 2 of cluster 'sparse-sczone' died. Sep 5 19:06:01 schost-4 cl_runtime: NOTICE: Membership : Node 4 of cluster 'sparse-sczone' died. Sep 5 19:06:01 schost-4 cl_runtime: NOTICE: Membership : Node 3 of cluster 'sparse-sczone' died. Sep 5 19:06:01 schost-4 cl_runtime: NOTICE: Membership : Node 1 of cluster 'sparse-sczone' died. phys-schost# |
The following example shows the console output when normal global-cluster operation is stopped and all nodes are shut down, enabling the ok prompt to be shown. The -g 0 option sets the shutdown grace period to zero, and the -y option provides an automatic yes response to the confirmation question. Shutdown messages also appear on the consoles of the other nodes in the global cluster.
phys-schost# cluster shutdown -g0 -y Wed Mar 10 13:47:32 phys-schost-1 cl_runtime: WARNING: CMM monitoring disabled. phys-schost-1# INIT: New run level: 0 The system is coming down. Please wait. System services are now being stopped. /etc/rc0.d/K05initrgm: Calling scswitch -S (evacuate) The system is down. syncing file systems... done Program terminated ok |
The following example shows the console output when normal global-cluster operation is stopped and all nodes are shut down. In this example, the ok prompt is not displayed on all of the nodes. The -g 0 option sets the shutdown grace period to zero, and the -y option provides an automatic yes response to the confirmation question. Shutdown messages also appear on the consoles of the other nodes in the global cluster.
phys-schost# cluster shutdown -g0 -y May 2 10:32:57 phys-schost-1 cl_runtime: WARNING: CMM: Monitoring disabled. root@phys-schost-1# INIT: New run level: 0 The system is coming down. Please wait. System services are now being stopped. /etc/rc0.d/K05initrgm: Calling scswitch -S (evacuate) failfasts already disabled on node 1 Print services already stopped. May 2 10:33:13 phys-schost-1 syslogd: going down on signal 15 The system is down. syncing file systems... done Type any key to continue |
See How to Boot a Cluster to restart a global cluster or a zone cluster that was shut down.
This procedure explains how to start a global cluster or zone cluster whose nodes have been shut down. For global-cluster nodes, the system displays the ok prompt on SPARC systems or the Press any key to continue message on the GRUB based x86 systems.
The phys-schost# prompt reflects a global-cluster prompt. Perform this procedure on a global cluster.
This procedure provides the long forms of the Sun Cluster commands. Most commands also have short forms. Except for the long and short forms of the command names, the commands are identical. For a list of the commands and their short forms, see Appendix B, Sun Cluster Object-Oriented Commands.
To create a zone cluster, follow the instructions in Configuring a Zone Cluster in Sun Cluster Software Installation Guide for Solaris OS.
Boot each node into cluster mode. Perform all steps in this procedure from a node of the global cluster.
On SPARC based systems, run the following command.
ok boot |
On x86 based systems, run the following commands.
When the GRUB menu is displayed, select the appropriate Solaris entry and press Enter. The GRUB menu appears similar to the following:
GNU GRUB version 0.95 (631K lower / 2095488K upper memory) +-------------------------------------------------------------------------+ | Solaris 10 /sol_10_x86 | | Solaris failsafe | | | +-------------------------------------------------------------------------+ Use the ^ and v keys to select which entry is highlighted. Press enter to boot the selected OS, 'e' to edit the commands before booting, or 'c' for a command-line. |
For more information about GRUB based booting, see Booting an x86 Based System by Using GRUB (Task Map) in System Administration Guide: Basic Administration.
Nodes must have a working connection to the cluster interconnect to attain cluster membership.
If you have a zone cluster, you can boot the entire zone cluster.
phys-schost# clzonecluster boot zoneclustername |
If you have more than one zone cluster, you can boot all zone clusters. Use + instead of the zoneclustername.
Verify that the nodes booted without error and are online.
The cluster(1CL) status command reports the global-cluster nodes' status.
phys-schost# cluster status -t node |
When you run the clzonecluster(1CL) status command from a global-cluster node, the command reports the state of the zone-cluster node.
phys-schost# clzonecluster status |
If a node's /var file system fills up, Sun Cluster might not be able to restart on that node. If this problem arises, see How to Repair a Full /var File System.
The following example shows the console output when node phys-schost-1 is booted into the global cluster. Similar messages appear on the consoles of the other nodes in the global cluster. When the autoboot property of a zone cluster is set to true, the system automatically boots the zone-cluster node after booting the global-cluster node on that machine.
When a global-cluster node reboots, all zone cluster nodes on that machine halt. Any zone-cluster node on that same machine with the autoboot property set to true boots after the global-cluster node restarts.
ok boot Rebooting with command: boot ... Hostname: phys-schost-1 Booting as part of a cluster NOTICE: Node phys-schost-1 with votecount = 1 added. NOTICE: Node phys-schost-2 with votecount = 1 added. NOTICE: Node phys-schost-3 with votecount = 1 added. ... NOTICE: Node phys-schost-1: attempting to join cluster ... NOTICE: Node phys-schost-2 (incarnation # 937690106) has become reachable. NOTICE: Node phys-schost-3 (incarnation # 937690290) has become reachable. NOTICE: cluster has reached quorum. NOTICE: node phys-schost-1 is up; new incarnation number = 937846227. NOTICE: node phys-schost-2 is up; new incarnation number = 937690106. NOTICE: node phys-schost-3 is up; new incarnation number = 937690290. NOTICE: Cluster members: phys-schost-1 phys-schost-2 phys-schost-3. ... |
To shut down a global cluster, run the cluster shutdown command and then boot the global cluster with the boot command on each node. To shut down a zone cluster, use the clzonecluster halt command and then use the clzonecluster boot command to boot the zone cluster. You can also use the clzonecluster reboot command. For more information, see the cluster(1CL), boot(1M), and clzonecluster(1CL) man pages.
The phys-schost# prompt reflects a global-cluster prompt. Perform this procedure on a global cluster.
This procedure provides the long forms of the Sun Cluster commands. Most commands also have short forms. Except for the long and short forms of the command names, the commands are identical. For a list of the commands and their short forms, see Appendix B, Sun Cluster Object-Oriented Commands.
If your cluster is running Oracle RAC, shut down all instances of the database on the cluster you are shutting down.
Refer to the Oracle RAC product documentation for shutdown procedures.
Become superuser or assume a role that provides solaris.cluster.admin RBAC authorization on any node in the cluster. Perform all steps in this procedure from a node of the global cluster.
Shut down the cluster.
Shut down the global cluster.
phys-schost# cluster shutdown -g0 -y |
If you have a zone cluster, shut down the zone cluster from a global-cluster node.
phys-schost# clzonecluster halt zoneclustername |
Each node is shut down. You can also use the cluster shutdown command within a zone cluster to shut down the zone cluster.
Nodes must have a working connection to the cluster interconnect to attain cluster membership.
Boot each node.
The order in which the nodes are booted is irrelevant unless you make configuration changes between shutdowns. If you make configuration changes between shutdowns, start the node with the most current configuration first.
For a global-cluster node on a SPARC based system, run the following command.
ok boot |
For a global-cluster node on an x86 based system, run the following commands.
When the GRUB menu is displayed, select the appropriate Solaris entry and press Enter. The GRUB menu appears similar to the following:
GNU GRUB version 0.95 (631K lower / 2095488K upper memory) +-------------------------------------------------------------------------+ | Solaris 10 /sol_10_x86 | | Solaris failsafe | | | +-------------------------------------------------------------------------+ Use the ^ and v keys to select which entry is highlighted. Press enter to boot the selected OS, 'e' to edit the commands before booting, or 'c' for a command-line. |
Nodes must have a working connection to the cluster interconnect to attain cluster membership.
For more information about GRUB-based booting, see Booting an x86 Based System by Using GRUB (Task Map) in System Administration Guide: Basic Administration.
For a zone cluster, type the following command on a single node of the global cluster to boot the zone cluster.
phys-schost# clzonecluster boot zoneclustername |
Messages appear on the booted nodes' consoles as cluster components are activated.
Verify that the nodes booted without error and are online.
The clnode status command reports the status of the nodes on the global cluster.
phys-schost# clnode status |
Running the clzonecluster status command on a global-cluster node reports the status of the zone-cluster nodes.
phys-schost# clzonecluster status |
You can also run the cluster status command within a zone cluster to see the status of the nodes.
If a node's /var file system fills up, Sun Cluster might not be able to restart on that node. If this problem arises, see How to Repair a Full /var File System.
The following example shows how to halt and boot a zone cluster called sparse-sczone. You can also use the clzonecluster reboot command.
phys-schost# clzonecluster halt sparse-sczone Waiting for zone halt commands to complete on all the nodes of the zone cluster "sparse-sczone"... Sep 5 19:17:46 schost-4 cl_runtime: NOTICE: Membership : Node 4 of cluster 'sparse-sczone' died. Sep 5 19:17:46 schost-4 cl_runtime: NOTICE: Membership : Node 2 of cluster 'sparse-sczone' died. Sep 5 19:17:46 schost-4 cl_runtime: NOTICE: Membership : Node 1 of cluster 'sparse-sczone' died. Sep 5 19:17:46 schost-4 cl_runtime: NOTICE: Membership : Node 3 of cluster 'sparse-sczone' died. phys-schost# phys-schost# clzonecluster boot sparse-sczone Waiting for zone boot commands to complete on all the nodes of the zone cluster "sparse-sczone"... phys-schost# Sep 5 19:18:23 schost-4 cl_runtime: NOTICE: Membership : Node 1 of cluster 'sparse-sczone' joined. Sep 5 19:18:23 schost-4 cl_runtime: NOTICE: Membership : Node 2 of cluster 'sparse-sczone' joined. Sep 5 19:18:23 schost-4 cl_runtime: NOTICE: Membership : Node 3 of cluster 'sparse-sczone' joined. Sep 5 19:18:23 schost-4 cl_runtime: NOTICE: Membership : Node 4 of cluster 'sparse-sczone' joined. phys-schost# phys-schost# clzonecluster status === Zone Clusters === --- Zone Cluster Status --- Name Node Name Zone HostName Status Zone Status ---- --------- ------------- ------ ----------- sparse-sczone schost-1 sczone-1 Online Running schost-2 sczone-2 Online Running schost-3 sczone-3 Online Running schost-4 sczone-4 Online Running phys-schost# |
The following example shows the console output when normal global-cluster operation is stopped, all nodes are shut down to the ok prompt, and the global cluster is restarted. The -g 0 option sets the grace period to zero, and the -y option provides an automatic yes response to the confirmation question. Shutdown messages also appear on the consoles of other nodes in the global cluster.
phys-schost# cluster shutdown -g0 -y Wed Mar 10 13:47:32 phys-schost-1 cl_runtime: WARNING: CMM monitoring disabled. phys-schost-1# INIT: New run level: 0 The system is coming down. Please wait. ... The system is down. syncing file systems... done Program terminated ok boot Rebooting with command: boot ... Hostname: phys-schost-1 Booting as part of a cluster ... NOTICE: Node phys-schost-1: attempting to join cluster ... NOTICE: Node phys-schost-2 (incarnation # 937690106) has become reachable. NOTICE: Node phys-schost-3 (incarnation # 937690290) has become reachable. NOTICE: cluster has reached quorum. ... NOTICE: Cluster members: phys-schost-1 phys-schost-2 phys-schost-3. ... NOTICE: Node phys-schost-1: joined cluster ... The system is coming up. Please wait. checking ufs filesystems ... reservation program successfully exiting Print services started. volume management starting. The system is ready. phys-schost-1 console login: NOTICE: Node phys-schost-1: joined cluster ... The system is coming up. Please wait. checking ufs filesystems ... reservation program successfully exiting Print services started. volume management starting. The system is ready. phys-schost-1 console login: |
You can shut down a global-cluster node, a zone-cluster node, or a non-global zone. This section provides instructions for shutting down a global-cluster node and a zone-cluster node.
To shut down a global-cluster node, use the clnode evacuate command with the Solaris shutdown command. Use the cluster shutdown command only when shutting down an entire global cluster.
On a zone-cluster node, use the clzonecluster halt command on a global cluster to shut down a single zone-cluster node or an entire zone cluster. You can also use the clnode evacuate and shutdown commands to shut down a zone-cluster node.
For information about shutting down and booting a non-global zone, see Chapter 20, Installing, Booting, Halting, Uninstalling, and Cloning Non-Global Zones (Tasks), in System Administration Guide: Solaris Containers-Resource Management and Solaris Zones. See also clnode(1CL), shutdown(1M), and clzonecluster(1CL).
In the procedures in this chapter, phys-schost# reflects a global-cluster prompt. The clzonecluster interactive shell prompt is clzc:schost>.
Table 3–2 Task Map: Shutting Down and Booting a Node
Task |
Tool |
Instructions |
---|---|---|
Stop a node. |
For a global-cluster node, use clnode(1CL) evacuate and shutdown. For a zone-cluster node, use clzonecluster(1CL) halt. | |
Start a node. The node must have a working connection to the cluster interconnect to attain cluster membership. |
For a global-cluster node, use boot or b. For a zone-cluster node, use clzonecluster(1CL) boot. | |
Stop and restart (reboot) a node on a cluster. The node must have a working connection to the cluster interconnect to attain cluster membership. |
For a global-cluster node, use clnode evacuate and shutdown, followed by boot or b. For a zone-cluster node, use clzonecluster(1CL) reboot. | |
Boot a node so that the node does not participate in cluster membership. |
For a global-cluster node, use clnode evacuate and shutdown commands, followed by boot -x on SPARC or GRUB menu entry editing on x86. If the underlying global cluster is booted in noncluster mode, the zone cluster node is automatically in noncluster mode. |
The phys-schost# prompt reflects a global-cluster prompt. Perform this procedure on a global cluster.
This procedure provides the long forms of the Sun Cluster commands. Most commands also have short forms. Except for the long and short forms of the command names, the commands are identical. For a list of the commands and their short forms, see Appendix B, Sun Cluster Object-Oriented Commands.
Do not use send brk on a cluster console to shut down a node on a global cluster or a zone cluster. The command is not supported within a cluster.
If your cluster is running Oracle RAC, shut down all instances of the database on the cluster you are shutting down.
Refer to the Oracle RAC product documentation for shutdown procedures.
Become superuser or assume a role that provides solaris.cluster.admin RBAC authorization on the cluster node to be shut down. Perform all steps in this procedure from a node of the global cluster.
If you want to halt a specific zone cluster member, skip Steps 4 - 6 and execute the following command from a global-cluster node:
phys-schost# clzonecluster halt -n physical-name zoneclustername |
When you specify a particular zone-cluster node, you stop only that node. By default, the halt command stops the zone clusters on all nodes.
Switch all resource groups, resources, and device groups from the node being shut down to other global cluster members.
On the global-cluster node to shut down, type the following command. The clnode evacuate command switches over all resource groups and device groups, including all non-global zones, from the specified node to the next-preferred node. (You can also run clnode evacuate within a zone-cluster node.)
phys-schost# clnode evacuate node |
Specifies the node from which you are switching resource groups and device groups.
Shut down the node.
Specify the global-cluster node you want to shut down.
phys-schost# shutdown -g0 -y -i0 |
Verify that the global-cluster node is showing the ok prompt on a SPARC based system or the Press any key to continue message on the GRUB menu on an x86 based system.
If necessary, power off the node.
The following example shows the console output when node phys-schost-1 is shut down. The -g0 option sets the grace period to zero, and the -y option provides an automatic yes response to the confirmation question. Shutdown messages for this node appear on the consoles of other nodes in the global cluster.
phys-schost# clnode evacuate nodename phys-schost# shutdown -g0 -y Wed Mar 10 13:47:32 phys-schost-1 cl_runtime: WARNING: CMM monitoring disabled. phys-schost-1# INIT: New run level: 0 The system is coming down. Please wait. Notice: rgmd is being stopped. Notice: rpc.pmfd is being stopped. Notice: rpc.fed is being stopped. umount: /global/.devices/node@1 busy umount: /global/phys-schost-1 busy The system is down. syncing file systems... done Program terminated ok |
The following example shows the console output when node phys-schost-1 is shut down. The -g0 option sets the grace period to zero, and the -y option provides an automatic yes response to the confirmation question. Shutdown messages for this node appear on the consoles of other nodes in the global cluster.
phys-schost# clnode evacuate phys-schost-1 phys-schost# shutdown -g0 -y Shutdown started. Wed Mar 10 13:47:32 PST 2004 Changing to init state 0 - please wait Broadcast Message from root (console) on phys-schost-1 Wed Mar 10 13:47:32... THE SYSTEM phys-schost-1 IS BEING SHUT DOWN NOW ! ! ! Log off now or risk your files being damaged phys-schost-1# INIT: New run level: 0 The system is coming down. Please wait. System services are now being stopped. /etc/rc0.d/K05initrgm: Calling scswitch -S (evacuate) failfasts disabled on node 1 Print services already stopped. Mar 10 13:47:44 phys-schost-1 syslogd: going down on signal 15 umount: /global/.devices/node@2 busy umount: /global/.devices/node@1 busy The system is down. syncing file systems... done WARNING: CMM: Node being shut down. Type any key to continue |
The following example shows how use the clzonecluster halt to shut down a node on a zone cluster called sparse-sczone. (You can also run the clnode evacuate and shutdown commands in a zone-cluster node.)
phys-schost# clzonecluster status === Zone Clusters === --- Zone Cluster Status --- Name Node Name Zone HostName Status Zone Status ---- --------- ------------- ------ ----------- sparse-sczone schost-1 sczone-1 Online Running schost-2 sczone-2 Online Running schost-3 sczone-3 Online Running schost-4 sczone-4 Online Running phys-schost# phys-schost# clzonecluster halt -n schost-4 sparse-sczone Waiting for zone halt commands to complete on all the nodes of the zone cluster "sparse-sczone"... Sep 5 19:24:00 schost-4 cl_runtime: NOTICE: Membership : Node 3 of cluster 'sparse-sczone' died. phys-host# phys-host# clzonecluster status === Zone Clusters === --- Zone Cluster Status --- Name Node Name Zone HostName Status Zone Status ---- --------- ------------- ------ ----------- sparse-sczone schost-1 sczone-1 Online Running schost-2 sczone-2 Online Running schost-3 sczone-3 Offline Installed schost-4 sczone-4 Online Running phys-schost# |
See How to Boot a Node to restart a global-cluster node that was shut down.
If you intend to shut down or reboot other active nodes in the global cluster or zone cluster, wait until the node you are booting has reached at least the following status:
SPARC: If you are running Solaris 9 OS, wait for the login prompt.
If you are running Solaris 10 OS, wait for the multiuser-server milestone to come online.
Otherwise, the node will not be available to take over services from other nodes in the cluster that you shut down or reboot. For information about booting a non-global zone, see Chapter 20, Installing, Booting, Halting, Uninstalling, and Cloning Non-Global Zones (Tasks), in System Administration Guide: Solaris Containers-Resource Management and Solaris Zones.
Starting a node can be affected by the quorum configuration. In a two-node cluster, you must have a quorum device configured so that the total quorum count for the cluster is three. You should have one quorum count for each node and one quorum count for the quorum device. In this situation, if the first node is shut down, the second node continues to have quorum and runs as the sole cluster member. For the first node to come back in the cluster as a cluster node, the second node must be up and running. The required cluster quorum count (two) must be present.
The phys-schost# prompt reflects a global-cluster prompt. Perform this procedure on a global cluster.
This procedure provides the long forms of the Sun Cluster commands. Most commands also have short forms. Except for the long and short forms of the command names, the commands are identical. For a list of the commands and their short forms, see Appendix B, Sun Cluster Object-Oriented Commands.
Nodes must have a working connection to the cluster interconnect to attain cluster membership.
To start a global-cluster node or zone-cluster node that has been shut down, boot the node. Perform all steps in this procedure from a node of the global cluster.
On SPARC based systems, run the following command.
ok boot |
On x86 based systems, run the following commands.
When the GRUB menu is displayed, select the appropriate Solaris entry and press Enter. The GRUB menu appears similar to the following:
GNU GRUB version 0.95 (631K lower / 2095488K upper memory) +-------------------------------------------------------------------------+ | Solaris 10 /sol_10_x86 | | Solaris failsafe | | | +-------------------------------------------------------------------------+ Use the ^ and v keys to select which entry is highlighted. Press enter to boot the selected OS, 'e' to edit the commands before booting, or 'c' for a command-line. |
Messages appear on the booted nodes' consoles as cluster components are activated.
If you have a zone cluster, you can specify a node to boot.
phys-schost# clzonecluster boot -n node zoneclustername |
Verify that the node booted without error, and is online.
Running the cluster status command reports the status of a global-cluster node.
phys-schost# cluster status -t node |
Running the clzonecluster status command from a node on the global cluster reports the status of all zone-cluster nodes.
phys-schost# clzonecluster status |
A zone-cluster node can only be booted in cluster mode when the node hosting the node is booted in cluster mode.
If a node's /var file system fills up, Sun Cluster might not be able to restart on that node. If this problem arises, see How to Repair a Full /var File System.
The following example shows the console output when node phys-schost-1 is booted into the global cluster.
ok boot Rebooting with command: boot ... Hostname: phys-schost-1 Booting as part of a cluster ... NOTICE: Node phys-schost-1: attempting to join cluster ... NOTICE: Node phys-schost-1: joined cluster ... The system is coming up. Please wait. checking ufs filesystems ... reservation program successfully exiting Print services started. volume management starting. The system is ready. phys-schost-1 console login: |
To shut down or reboot other active nodes in the global cluster or zone cluster, wait until the node that you are rebooting has reached at least the following status:
SPARC: If you are running Solaris 9 OS, wait for the login prompt.
If you are running Solaris 10 OS, wait for the multiuser-server milestone to come online.
Otherwise, the node will not be available to take over services from other nodes in the cluster that you shut down or reboot. For information about rebooting a non-global zone, see Chapter 20, Installing, Booting, Halting, Uninstalling, and Cloning Non-Global Zones (Tasks), in System Administration Guide: Solaris Containers-Resource Management and Solaris Zones.
The phys-schost# prompt reflects a global-cluster prompt. Perform this procedure on a global cluster.
This procedure provides the long forms of the Sun Cluster commands. Most commands also have short forms. Except for the long and short forms of the command names, the commands are identical. For a list of the commands and their short forms, see Appendix B, Sun Cluster Object-Oriented Commands.
If the global-cluster or zone-cluster node is running Oracle RAC, shut down all instances of the database on the node that you are shutting down.
Refer to the Oracle RAC product documentation for shutdown procedures.
Become superuser or assume a role that provides solaris.cluster.admin RBAC authorization on the node to shut down. Perform all steps in this procedure from a node of the global cluster.
Shut down the global-cluster node by using the clnode evacuate and shutdown commands. Shut down the zone cluster with the clzonecluster halt command executed on a node of the global cluster. (The clnode evacuate and shutdown commands also work in a zone cluster.)
For a global cluster, type the following commands on the node to shut down. The clnode evacuate command switches over all device groups from the specified node to the next-preferred node. The command also switches all resource groups from global or non-global zones on the specified node to the next-preferred global or non-global zones on other nodes.
On a SPARC based system, run the following commands.
phys-schost# clnode evacuate node |
phys-schost# shutdown -g0 -y -i6 |
On an x86 based system, run the following commands.
phys-schost# clnode evacuate node |
phys-schost# shutdown -g0 -y -i6 |
When the GRUB menu is displayed, select the appropriate Solaris entry and press Enter. The GRUB menu appears similar to the following:
GNU GRUB version 0.95 (631K lower / 2095488K upper memory) +-------------------------------------------------------------------------+ | Solaris 10 /sol_10_x86 | | Solaris failsafe | | | +-------------------------------------------------------------------------+ Use the ^ and v keys to select which entry is highlighted. Press enter to boot the selected OS, 'e' to edit the commands before booting, or 'c' for a command-line. |
Specify the zone-cluster node to shut down and reboot.
phys-schost# clzonecluster reboot - node zoneclustername |
Nodes must have a working connection to the cluster interconnect to attain cluster membership.
Verify that the node booted without error and is online.
The following example shows the console output when node phys-schost-1 is rebooted. Messages for this node, such as shutdown and startup notification, appear on the consoles of other nodes in the global cluster.
phys-schost# clnode evacuate phys-schost-1 phys-schost# shutdown -g0 -y -i6 Shutdown started. Wed Mar 10 13:47:32 phys-schost-1 cl_runtime: WARNING: CMM monitoring disabled. phys-schost-1# INIT: New run level: 6 The system is coming down. Please wait. System services are now being stopped. Notice: rgmd is being stopped. Notice: rpc.pmfd is being stopped. Notice: rpc.fed is being stopped. umount: /global/.devices/node@1 busy umount: /global/phys-schost-1 busy The system is down. syncing file systems... done rebooting... Resetting ... ,,, Sun Ultra 1 SBus (UltraSPARC 143MHz), No Keyboard OpenBoot 3.11, 128 MB memory installed, Serial #5932401. Ethernet address 8:8:20:99:ab:77, Host ID: 8899ab77. ... Rebooting with command: boot ... Hostname: phys-schost-1 Booting as part of a cluster ... NOTICE: Node phys-schost-1: attempting to join cluster ... NOTICE: Node phys-schost-1: joined cluster ... The system is coming up. Please wait. The system is ready. phys-schost-1 console login: |
The following example shows the console output when rebooting node phys-schost-1. Messages for this node, such as shutdown and startup notification, appear on the consoles of other nodes in the global cluster.
phys-schost# clnode evacuate phys-schost-1 phys-schost # shutdown -y -g0 -i6 GNU GRUB version 0.95 (631K lower / 2095488K upper memory) +-------------------------------------------------------------------------+ | Solaris 10 /sol_10_x86 | | Solaris failsafe | | | +-------------------------------------------------------------------------+ Use the ^ and v keys to select which entry is highlighted. Press enter to boot the selected OS, 'e' to edit the commands before booting, or 'c' for a command-line. Hostname: phys-schost-1 Booting as part of a cluster ... NOTICE: Node phys-schost-1: attempting to join cluster ... NOTICE: Node phys-schost-1: joined cluster ... The system is coming up. Please wait. checking ufs filesystems ... reservation program successfully exiting Print services started. volume management starting. The system is ready. phys-schost-1 console login: |
The following example shows how to reboot a node on a zone cluster.
phys-schost# clzonecluster reboot -n schost-4 sparse-sczone Waiting for zone reboot commands to complete on all the nodes of the zone cluster "sparse-sczone"... Sep 5 19:40:59 schost-4 cl_runtime: NOTICE: Membership : Node 3 of cluster 'sparse-sczone' died. phys-schost# Sep 5 19:41:27 schost-4 cl_runtime: NOTICE: Membership : Node 3 of cluster 'sparse-sczone' joined. phys-schost# phys-schost# clzonecluster status === Zone Clusters === --- Zone Cluster Status --- Name Node Name Zone HostName Status Zone Status ---- --------- ------------- ------ ----------- sparse-sczone schost-1 sczone-1 Online Running schost-2 sczone-2 Online Running schost-3 sczone-3 Online Running schost-4 sczone-4 Online Running phys-schost# |
You can boot a global-cluster node in noncluster mode, where the node does not participate in the cluster membership. Noncluster mode is useful when installing the cluster software or performing certain administrative procedures, such as patching a node. A zone-cluster node cannot be in a boot state that is different from the state of the underlying global-cluster node. If the global-cluster node is booted in noncluster mode, the zone-cluster node is automatically in noncluster mode.
The phys-schost# prompt reflects a global-cluster prompt. Perform this procedure on a global cluster.
This procedure provides the long forms of the Sun Cluster commands. Most commands also have short forms. Except for the long and short forms of the command names, the commands are identical. For a list of the commands and their short forms, see Appendix B, Sun Cluster Object-Oriented Commands.
Become superuser or assume a role that provides solaris.cluster.admin RBAC authorization on the cluster to be started in noncluster mode. Perform all steps in this procedure from a node of the global cluster.
Shut down the zone-cluster node by running the clzonecluster halt command on a node of the global cluster. Shut down the global-cluster node by using the clnode evacuate and shutdown commands.
The clnode evacuate command switches over all device groups from the specified node to the next-preferred node. The command also switches all resource groups from global or non-global zones on the specified node to the next-preferred global or non-global zones on other nodes.
Shut down a specific global cluster.
phys-schost# clnode evacuate node |
phys-schost# shutdown -g0 -y |
Shut down a specific zone-cluster node from a global-cluster node.
phys-schost# clzonecluster halt -n node zoneclustername |
You can also use the clnode evacuate and shutdown commands within a zone cluster.
Verify that the global-cluster node is showing the ok prompt on a Solaris based system or the Press any key to continue message on a GRUB menu on an x86 based system.
Boot the global-cluster node in noncluster mode.
On SPARC based systems, run the following command.
ok boot -xs |
On x86 based systems, run the following commands.
In the GRUB menu, use the arrow keys to select the appropriate Solaris entry and type e to edit its commands.
The GRUB menu appears similar to the following:
GNU GRUB version 0.95 (631K lower / 2095488K upper memory) +-------------------------------------------------------------------------+ | Solaris 10 /sol_10_x86 | | Solaris failsafe | | | +-------------------------------------------------------------------------+ Use the ^ and v keys to select which entry is highlighted. Press enter to boot the selected OS, 'e' to edit the commands before booting, or 'c' for a command-line. |
For more information about GRUB based booting, see Booting an x86 Based System by Using GRUB (Task Map) in System Administration Guide: Basic Administration.
In the boot parameters screen, use the arrow keys to select the kernel entry and type e to edit the entry.
The GRUB boot parameters screen appears similar to the following:
GNU GRUB version 0.95 (615K lower / 2095552K upper memory) +----------------------------------------------------------------------+ | root (hd0,0,a) | | kernel /platform/i86pc/multiboot | | module /platform/i86pc/boot_archive | +----------------------------------------------------------------------+ Use the ^ and v keys to select which entry is highlighted. Press 'b' to boot, 'e' to edit the selected command in the boot sequence, 'c' for a command-line, 'o' to open a new line after ('O' for before) the selected line, 'd' to remove the selected line, or escape to go back to the main menu. |
Add -x to the command to specify system boot in noncluster mode.
[ Minimal BASH-like line editing is supported. For the first word, TAB lists possible command completions. Anywhere else TAB lists the possible completions of a device/filename. ESC at any time exits. ] grub edit> kernel /platform/i86pc/multiboot -x |
Press the Enter key to accept the change and return to the boot parameters screen.
The screen displays the edited command.
GNU GRUB version 0.95 (615K lower / 2095552K upper memory) +----------------------------------------------------------------------+ | root (hd0,0,a) | | kernel /platform/i86pc/multiboot -x | | module /platform/i86pc/boot_archive | +----------------------------------------------------------------------+ Use the ^ and v keys to select which entry is highlighted. Press 'b' to boot, 'e' to edit the selected command in the boot sequence, 'c' for a command-line, 'o' to open a new line after ('O' for before) the selected line, 'd' to remove the selected line, or escape to go back to the main menu.- |
Type b to boot the node into noncluster mode.
This change to the kernel boot parameter command does not persist over the system boot. The next time you reboot the node, it will boot into cluster mode. To boot into noncluster mode instead, perform these steps again to add the -x option to the kernel boot parameter command.
The following example shows the console output when node phys-schost-1 is shut down and restarted in noncluster mode. The -g0 option sets the grace period to zero, the -y option provides an automatic yes response to the confirmation question, and the -i0 option invokes run level 0 (zero). Shutdown messages for this node appear on the consoles of other nodes in the global cluster.
phys-schost# clnode evacuate phys-schost-1 phys-schost# cluster shutdown -g0 -y Shutdown started. Wed Mar 10 13:47:32 phys-schost-1 cl_runtime: WARNING: CMM monitoring disabled. phys-schost-1# ... rg_name = schost-sa-1 ... offline node = phys-schost-2 ... num of node = 0 ... phys-schost-1# INIT: New run level: 0 The system is coming down. Please wait. System services are now being stopped. Print services stopped. syslogd: going down on signal 15 ... The system is down. syncing file systems... done WARNING: node phys-schost-1 is being shut down. Program terminated ok boot -x ... Not booting as part of cluster ... The system is ready. phys-schost-1 console login: |
Both Solaris software and Sun Cluster software write error messages to the /var/adm/messages file, which over time can fill the /var file system. If a cluster node's /var file system fills up, Sun Cluster might not be able to restart on that node. Additionally, you might not be able to log in to the node.
If a node reports a full /var file system and continues to run Sun Cluster services, use this procedure to clear the full file system. Refer to Viewing System Messages in System Administration Guide: Advanced Administration in System Administration Guide: Advanced Administrationfor more information.