Go to main content

Administering an Oracle® Solaris Cluster 4.4 Configuration

Exit Print View

Updated: March 2019
 
 

How to Shut Down a Node

The phys-schost# prompt reflects a global-cluster prompt. Perform this procedure on a global cluster.

This procedure provides the long forms of the Oracle Solaris Cluster commands. Most commands also have short forms. Except for the long and short forms of the command names, the commands are identical.


Caution

Caution  -  Do not use send brk on a cluster console to shut down a node on a global cluster or a zone cluster. The command is not supported within a cluster.



Note -  You can also use the Oracle Solaris Cluster Manager browser interface to evacuate a global-cluster node and switch all resource groups and device groups to the next-preferred node. You can also shut down a zone-cluster node. For Oracle Solaris Cluster Manager log-in instructions, see How to Access Oracle Solaris Cluster Manager.
  1. If your cluster is running Oracle RAC, shut down all instances of the database on the cluster you are shutting down.

    Refer to the Oracle RAC product documentation for shutdown procedures.

  2. Assume the root role or a role that provides solaris.cluster.admin authorization on the cluster node to be shut down.

    Perform all steps in this procedure from a node of the global cluster.

  3. If you want to halt a specific zone cluster member, skip Steps 4 - 6 and execute the following command from a global-cluster node:
    phys-schost# clzonecluster halt -n physical-name zone-cluster-name

    When you specify a particular zone-cluster node, you stop only that node. By default, the halt command stops the zone clusters on all nodes.

  4. Switch all resource groups, resources, and device groups from the node being shut down to other global cluster members.

    On the global-cluster node to shut down, type the following command. The clnode evacuate command switches over all resource groups and device groups from the specified node to the next-preferred node. (You can also run clnode evacuate within a zone-cluster node.)

    phys-schost# clnode evacuate node
    node

    Specifies the node from which you are switching resource groups and device groups.

  5. Shut down the node.

    Execute the shutdown command on the global-cluster node you want to shut down.

    phys-schost# shutdown -g0 -y -i0

    Verify that the global-cluster node is showing the ok prompt on a SPARC based system or the Press any key to continue message on the GRUB menu on an x86 based system.

  6. If necessary, power off the node.
Example 22  SPARC: Shutting Down a Global-Cluster Node

The following example shows the console output when node phys-schost-1 is shut down. The –g0 option sets the grace period to zero, and the –y option provides an automatic yes response to the confirmation question. Shutdown messages for this node appear on the consoles of other nodes in the global cluster.

phys-schost# clnode evacuate phys-schost-1
phys-schost# shutdown -g0 -y -i0
Wed Mar 10 13:47:32 phys-schost-1 cl_runtime:
WARNING: CMM monitoring disabled.
phys-schost-1#
INIT: New run level: 0
The system is coming down.  Please wait.
Notice: rgmd is being stopped.
Notice: rpc.pmfd is being stopped.
Notice: rpc.fed is being stopped.
umount: /global/.devices/node@1 busy
umount: /global/phys-schost-1 busy
The system is down.
syncing file systems... done
Program terminated
ok 
Example 23  x86: Shutting Down a Global-Cluster Node

The following example shows the console output when node phys-schost-1 is shut down. The –g0 option sets the grace period to zero, and the –y option provides an automatic yes response to the confirmation question. Shutdown messages for this node appear on the consoles of other nodes in the global cluster.

phys-schost# clnode evacuate phys-schost-1
phys-schost# shutdown -g0 -y -i0
Shutdown started.    Wed Mar 10 13:47:32 PST 2004

Changing to init state 0 - please wait
Broadcast Message from root (console) on phys-schost-1 Wed Mar 10 13:47:32...
THE SYSTEM phys-schost-1 IS BEING SHUT DOWN NOW ! ! !
Log off now or risk your files being damaged

phys-schost-1#
INIT: New run level: 0
The system is coming down.  Please wait.
System services are now being stopped.
/etc/rc0.d/K05initrgm: Calling clnode evacuate
failfasts disabled on node 1
Print services already stopped.
Mar 10 13:47:44 phys-schost-1 syslogd: going down on signal 15
umount: /global/.devices/node@2 busy
umount: /global/.devices/node@1 busy
The system is down.
syncing file systems... done
WARNING: CMM: Node being shut down.
Type any key to continue 
Example 24  Shutting Down a Zone-Cluster Node

The following example shows how use the clzonecluster halt to shut down a node on a zone cluster called sparse-sczone. (You can also run the clnode evacuate and shutdown commands in a zone-cluster node.)

phys-schost# clzonecluster status

=== Zone Clusters ===

--- Zone Cluster Status ---

Name            Node Name   Zone HostName   Status   Zone Status
----            ---------   -------------   ------   -----------
sparse-sczone   schost-1    sczone-1        Online   Running
                schost-2    sczone-2        Online   Running
                schost-3    sczone-3        Online   Running
                schost-4    sczone-4        Online   Running

phys-schost#
phys-schost# clzonecluster halt –n schost-4 sparse-sczone
Waiting for zone halt commands to complete on all the nodes of the zone cluster "sparse-sczone"...
Sep  5 19:24:00 schost-4 cl_runtime: NOTICE: Membership : Node 3 of cluster 'sparse-sczone' died.
phys-host#
phys-host# clzonecluster status

=== Zone Clusters ===

--- Zone Cluster Status ---

Name            Node Name   Zone HostName   Status    Zone Status
----            ---------   -------------   ------    -----------
sparse-sczone   schost-1    sczone-1        Online    Running
                schost-2    sczone-2        Online    Running
                schost-3    sczone-3        Offline   Installed
                schost-4    sczone-4        Online    Running

phys-schost# 

See Also

See How to Boot a Node to restart a global-cluster node that was shut down.