Sun Cluster 3.1 10/03 System Administration Guide

Shutting Down and Booting a Single Cluster Node


Note –

Use the scswitch(1M) command in conjunction with the Solaris shutdown(1M) command to shut down an individual node. Use the scshutdown command only when shutting down an entire cluster.


Table 3–2 Task Map: Shutting Down and Booting a Cluster Node

Task 

For Instructions 

Stop a cluster node  

    - Use scswitch(1M) and shutdown(1M)

How to Shut Down a Cluster Node

Start a node 

The node must have a working connection to the cluster interconnect to attain cluster membership. 

How to Boot a Cluster Node

Stop and restart (reboot) a cluster node 

    - Use scswitch and shutdown

The node must have a working connection to the cluster interconnect to attain cluster membership. 

How to Reboot a Cluster Node

Boot a node so that the node does not participate in cluster membership 

    - Use scswitch and shutdown, then boot -x

How to Boot a Cluster Node in Non-Cluster Mode

How to Shut Down a Cluster Node


Caution – Caution –

Do not use send brk on a cluster console to shut down a cluster node. Using send brk and entering go at the ok prompt to reboot causes a node to panic. This functionality is not supported within a cluster.


  1. If you are running Oracle Parallel Server/Real Application Clusters, shut down all instances of the database.

    Refer to the Oracle Parallel Server/Real Application Clusters product documentation for shutdown procedures.

  2. Become superuser on the cluster node to be shut down.

  3. Switch all resource groups, resources, and device groups from the node being shut down to other cluster members.

    On the node to be shut down, type the following command.


    # scswitch -S -h node
    

    -S

    Evacuates all device services and resource groups from the specified node.

    -h node

    Specifies the node from which you are switching resource groups and device groups.

  4. Shut down the cluster node to OBP.

    On the node to be shut down, type the following command.


    # shutdown -g0 -y -i0
    

  5. Verify that the cluster node has reached the ok prompt.

  6. If necessary, power off the node.

Example—Shutting Down a Cluster Node

The following example shows the console output when shutting down node phys-schost-1. The -g0 option sets the grace period to zero, -y provides an automatic yes response to the confirmation question, and -i0 invokes run level 0 (zero). Shutdown messages for this node appear on the consoles of other nodes in the cluster.


# scswitch -S -h phys-schost-1
# shutdown -g0 -y -i0
May 2 10:08:46 phys-schost-1 cl_runtime:
 WARNING: CMM monitoring disabled.phys-schost-1# 
INIT: New run level: 0
The system is coming down.  Please wait.
Notice: rgmd is being stopped.
Notice: rpc.pmfd is being stopped.
Notice: rpc.fed is being stopped.
umount: /global/.devices/node@1 busy
umount: /global/phys-schost-1 busy
The system is down.
syncing file systems... done
Program terminated
ok 

Where to Go From Here

See How to Boot a Cluster Node to restart a cluster node that has been shut down.

How to Boot a Cluster Node


Note –

Starting a cluster node can be affected by the quorum configuration. In a two-node cluster, you must have a quorum device configured so that the total quorum count for the cluster is three. You should have one quorum count for each node and one quorum count for the quorum device. In this situation, if the first node is shut down, the second node continues to have quorum and runs as the sole cluster member. For the first node to come back in the cluster as a cluster node, the second node must be up and running. The required cluster quorum count (two) must be present.


  1. To start a cluster node that has been shut down, boot the node.


    ok boot
    

    Messages are displayed on all node consoles as cluster components are activated.


    Note –

    A cluster node must have a working connection to the cluster interconnect to attain cluster membership.


  2. Verify that the node has booted without error, and is online.

    The scstat(1M) command reports the status of a node.


    # scstat -n
    


    Note –

    If a cluster node's /var file system fills up, Sun Cluster might not be able to restart on that node. If this problem arises, see How to Repair a Full /var File System.


Example—Booting a Cluster Node

The following example shows the console output when booting node phys-schost-1 into the cluster.


ok boot
Rebooting with command: boot 
...
Hostname: phys-schost-1
Booting as part of a cluster
...
NOTICE: Node 1: attempting to join cluster
...
NOTICE: Node 1: joined cluster
...
The system is coming up.  Please wait.
checking ufs filesystems
...
reservation program successfully exiting
Print services started.
volume management starting.
The system is ready.
phys-schost-1 console login:

How to Reboot a Cluster Node

  1. If the cluster node is running Oracle Parallel Server/Real Application Clusters, shut down all instances of the database.

    Refer to the Oracle Parallel Server/Real Application Clusters product documentation for shutdown procedures.

  2. Become superuser on the cluster node to be shut down.

  3. Shut down the cluster node by using the scswitch and shutdown commands.

    Enter these commands on the node to be shut down. The -i 6 option with the shutdown command causes the node to reboot after the node shuts down to the ok prompt.


    # scswitch -S -h node
    # shutdown -g0 -y -i6
    

    Note –

    Cluster nodes must have a working connection to the cluster interconnect to attain cluster membership.


  4. Verify that the node has booted without error, and is online.


    # scstat -n
    

Example—Rebooting a Cluster Node

The following example shows the console output when rebooting node phys-schost-1. Messages for this node, such as shutdown and startup notification, appear on the consoles of other nodes in the cluster.


# scswitch -S -h phys-schost-1
# shutdown -g0 -y -i6
May 2 10:08:46 phys-schost-1 cl_runtime: 
WARNING: CMM monitoring disabled.
phys-schost-1# 
INIT: New run level: 6
The system is coming down.  Please wait.
System services are now being stopped.
Notice: rgmd is being stopped.
Notice: rpc.pmfd is being stopped.
Notice: rpc.fed is being stopped.
umount: /global/.devices/node@1 busy
umount: /global/phys-schost-1 busy
The system is down.
syncing file systems... done
rebooting...
Resetting ... 
,,,
Sun Ultra 1 SBus (UltraSPARC 143MHz), No Keyboard
OpenBoot 3.11, 128 MB memory installed, Serial #5932401.
Ethernet address 8:8:20:99:ab:77, Host ID: 8899ab77.
...
Rebooting with command: boot
...
Hostname: phys-schost-1
Booting as part of a cluster
...
NOTICE: Node 1: attempting to join cluster
...
NOTICE: Node 1: joined cluster
...
The system is coming up.  Please wait.
The system is ready.
phys-schost-1 console login: 

How to Boot a Cluster Node in Non-Cluster Mode

You can boot a node so that the node does not participate in the cluster membership, that is, in non-cluster mode. Non-cluster mode is useful when installing the cluster software or performing certain administrative procedures, such as patching a node.

  1. Become superuser on the cluster node to be started in non-cluster mode.

  2. Shut down the node by using the scswitch and shutdown commands.


    # scswitch -S -h node
    # shutdown -g0 -y -i0
    

  3. Verify that the node is at the ok prompt.

  4. Boot the node in non-cluster mode by using the boot(1M) command with the -x option.


    ok boot -x
    

    Messages appear on the node's console stating that the node is not part of the cluster.

Example—Booting a Cluster Node in Non-Cluster Mode

The following example shows the console output when shutting down node phys-schost-1 then restarting the node in non-cluster mode. The -g0 option sets the grace period to zero, -y provides an automatic yes response to the confirmation question, and -i0 invokes run level 0 (zero). Shutdown messages for this node appear on the consoles of other nodes in the cluster.


# scswitch -S -h phys-schost-1
# shutdown -g0 -y -i0
May 2 10:08:46 phys-schost-1 cl_runtime: 
WARNING: CMM monitoring disabled.
phys-schost-1# 
...
rg_name = schost-sa-1 ...
offline node = phys-schost-2 ...
num of node = 0 ...
phys-schost-1# 
INIT: New run level: 0
The system is coming down.  Please wait.
System services are now being stopped.
Print services stopped.
syslogd: going down on signal 15
...
The system is down.
syncing file systems... done
WARNING: node 1 is being shut down.
Program terminated

ok boot -x
...
Not booting as part of cluster
...
The system is ready.
phys-schost-1 console login: