Sun Cluster System Administration Guide for Solaris OS

Overview of Shutting Down and Booting a Cluster

The Sun Cluster cluster(1CL) shutdown command stops cluster services in an orderly fashion and cleanly shuts down the entire cluster. You can use the cluster shutdown command when moving the location of a cluster. You can also use the command to shut down the cluster if an application error causes data corruption.


Note –

Use the cluster shutdown command instead of the shutdown or halt commands to ensure proper shutdown of the entire cluster. The Solaris shutdown command is used with the clnode(1CL) evacuate command to shut down individual nodes. See How to Shut Down a Cluster or Shutting Down and Booting a Single Cluster Node for more information.


The cluster shutdown command stops all nodes in a cluster by performing the following actions:

  1. Takes offline all running resource groups.

  2. Unmounts all cluster file systems.

  3. Shuts down active device services.

  4. Runs init 0 and brings all nodes to the OpenBootTM PROM ok prompt on a SPARC based system or to the GRUB menu on an x86 based system. The GRUB menus are described in more detail in Chapter 11, GRUB Based Booting (Tasks), in System Administration Guide: Basic Administration.


Note –

If necessary, you can boot a node in noncluster mode so that the node does not participate in cluster membership. Noncluster mode is useful when installing cluster software or for performing certain administrative procedures. See How to Boot a Cluster Node in Noncluster Mode for more information.


Table 3–1 Task List: Shutting Down and Booting a Cluster

Task 

For Instructions 

Stop the cluster. 

    -Use cluster(1CL) shutdown

See How to Shut Down a Cluster

Start the cluster by booting all nodes. 

The nodes must have a working connection to the cluster interconnect to attain cluster membership. 

See How to Boot a Cluster

Reboot the cluster. 

    - Use cluster shutdown.

At the Press any key to continue message, boot each node individually by pressing a key.

The nodes must have a working connection to the cluster interconnect to attain cluster membership. 

See How to Reboot a Cluster

ProcedureHow to Shut Down a Cluster


Caution – Caution –

Do not use send brk on a cluster console to shut down a cluster node. The command is not supported within a cluster.


This procedure provides the long forms of the Sun Cluster commands. Most commands also have short forms. Except for the long and short forms of the command names, the commands are identical. For a list of the commands and their short forms, see Appendix A, Sun Cluster Object-Oriented Commands.

  1. SPARC: If your cluster is running Oracle Parallel Server or Oracle Real Application Clusters (RAC), shut down all instances of the database.

    Refer to the Oracle Parallel Server or Oracle RAC product documentation for shutdown procedures.

  2. Become superuser or assume a role that provides solaris.cluster.admin RBAC authorization on any node in the cluster.

  3. Shut down the cluster immediately.

    From a single node in the cluster, type the following command.


    # cluster shutdown -g0 -y
    
  4. Verify that all nodes are showing the ok prompt on a SPARC-based system or a GRUB menu on an x86 based system.

    Do not power off any nodes until all cluster nodes are at the ok prompt on a SPARC-based system or in a Boot Subsystem on an x86 based system.


    # cluster status -t node
    
  5. If necessary, power off the nodes.


Example 3–1 SPARC: Shutting Down a Cluster

The following example shows the console output when normal cluster operation is stopped and all nodes are shut down so that the ok prompt is shown. The -g 0 option sets the shutdown grace period to zero, and the -y option provides an automatic yes response to the confirmation question. Shutdown messages also appear on the consoles of the other nodes in the cluster.


# cluster shutdown -g0 -y
Wed Mar 10 13:47:32 phys-schost-1 cl_runtime: 
WARNING: CMM monitoring disabled.
phys-schost-1# 
INIT: New run level: 0
The system is coming down.  Please wait.
System services are now being stopped.
/etc/rc0.d/K05initrgm: Calling scswitch -S (evacuate)
The system is down.
syncing file systems... done
Program terminated
ok 


Example 3–2 x86: Shutting Down a Cluster

The following example shows the console output when normal cluster operation is stopped all nodes are shut down. In this example, the ok prompt is not displayed on all of the nodes. The -g 0 option sets the shutdown grace period to zero, and the -y option provides an automatic yes response to the confirmation question. Shutdown messages also appear on the consoles of the other nodes in the cluster.


# cluster shutdown -g0 -y
May  2 10:32:57 phys-schost-1 cl_runtime: 
WARNING: CMM: Monitoring disabled.  
root@phys-schost-1#
INIT: New run level: 0
The system is coming down.  Please wait.
System services are now being stopped.
/etc/rc0.d/K05initrgm: Calling scswitch -S (evacuate)
failfasts already disabled on node 1
Print services already stopped.
May  2 10:33:13 phys-schost-1 syslogd: going down on signal 15
The system is down.
syncing file systems... done
Type any key to continue 

See Also

See How to Boot a Cluster to restart a cluster that has been shut down.

ProcedureHow to Boot a Cluster

This procedure explains how to start a cluster whose nodes have been shut down and are at the ok prompt on SPARC systems or at the Press any key to continue message on the GRUB-based x86 systems.

This procedure provides the long forms of the Sun Cluster commands. Most commands also have short forms. Except for the long and short forms of the command names, the commands are identical. For a list of the commands and their short forms, see Appendix A, Sun Cluster Object-Oriented Commands.

  1. Boot each node into cluster mode.

    • On SPARC based systems, do the following:


      ok boot
      
    • On x86 based systems, do the following:

      When the GRUB menu is displayed, select the appropriate Solaris entry and press Enter. The GRUB menu appears similar to the following:


      GNU GRUB version 0.95 (631K lower / 2095488K upper memory)
      +-------------------------------------------------------------------------+
      | Solaris 10 /sol_10_x86                                                  |
      | Solaris failsafe                                                        |
      |                                                                         |
      +-------------------------------------------------------------------------+
      Use the ^ and v keys to select which entry is highlighted.
      Press enter to boot the selected OS, 'e' to edit the
      commands before booting, or 'c' for a command-line.

    Note –

    Cluster nodes must have a working connection to the cluster interconnect to attain cluster membership.


    For more information about GRUB based booting, see Chapter 11, GRUB Based Booting (Tasks), in System Administration Guide: Basic Administration.

  2. Verify that the nodes booted without error and are online.

    The cluster(1CL) status command reports the nodes' status.


    # cluster status -t node
    

    Note –

    If a cluster node's /var file system fills up, Sun Cluster might not be able to restart on that node. If this problem arises, see How to Repair a Full /var File System.



Example 3–3 SPARC: Booting a Cluster

The following example shows the console output when node phys-schost-1 is booted into the cluster. Similar messages appear on the consoles of the other nodes in the cluster.


ok boot
Rebooting with command: boot 
...
Hostname: phys-schost-1
Booting as part of a cluster
NOTICE: Node phys-schost-1 with votecount = 1 added.
NOTICE: Node phys-schost-2 with votecount = 1 added.
NOTICE: Node phys-schost-3 with votecount = 1 added.
...
NOTICE: Node phys-schost-1: attempting to join cluster
...
NOTICE: Node phys-schost-2 (incarnation # 937690106) has become reachable.
NOTICE: Node phys-schost-3 (incarnation # 937690290) has become reachable.
NOTICE: cluster has reached quorum.
NOTICE: node phys-schost-1 is up; new incarnation number = 937846227.
NOTICE: node phys-schost-2 is up; new incarnation number = 937690106.
NOTICE: node phys-schost-3 is up; new incarnation number = 937690290.
NOTICE: Cluster members: phys-schost-1 phys-schost-2 phys-schost-3.
...

ProcedureHow to Reboot a Cluster

Run the cluster(1CL) shutdown command to shut down the cluster, then boot the cluster with the boot(1M) command on each node.

This procedure provides the long forms of the Sun Cluster commands. Most commands also have short forms. Except for the long and short forms of the command names, the commands are identical. For a list of the commands and their short forms, see Appendix A, Sun Cluster Object-Oriented Commands.

  1. SPARC: If your cluster is running Oracle Parallel Server or Oracle RAC, shut down all instances of the database.

    Refer to the Oracle Parallel Server or Oracle RAC product documentation for shutdown procedures.

  2. Become superuser or assume a role that provides solaris.cluster.admin RBAC authorization on any node in the cluster.

  3. Shut down the cluster.

    From a single node in the cluster, type the following command.


    # cluster shutdown -g0 -y 
    

    Each node is shut down.


    Note –

    Cluster nodes must have a working connection to the cluster interconnect to attain cluster membership.


  4. Boot each node.

    The order in which the nodes are booted is irrelevant unless you make configuration changes between shutdowns. If you make configuration changes between shutdowns, start the node with the most current configuration first.

    • On SPARC based systems, do the following:


      ok boot
      
    • On x86 based systems, do the following:

      When the GRUB menu is displayed, select the appropriate Solaris entry and press Enter. The GRUB menu appears similar to the following:


      GNU GRUB version 0.95 (631K lower / 2095488K upper memory)
      +-------------------------------------------------------------------------+
      | Solaris 10 /sol_10_x86                                                  |
      | Solaris failsafe                                                        |
      |                                                                         |
      +-------------------------------------------------------------------------+
      Use the ^ and v keys to select which entry is highlighted.
      Press enter to boot the selected OS, 'e' to edit the
      commands before booting, or 'c' for a command-line.

    Note –

    Cluster nodes must have a working connection to the cluster interconnect to attain cluster membership.


    For more information about GRUB based booting, see Chapter 11, GRUB Based Booting (Tasks), in System Administration Guide: Basic Administration.

    Messages appear on the booted nodes' consoles as cluster components are activated.

  5. Verify that the nodes booted without error and are online.

    The scstat command reports the nodes' status.


    # cluster status -t node
    

    Note –

    If a cluster node's /var file system fills up, Sun Cluster might not be able to restart on that node. If this problem arises, see How to Repair a Full /var File System.



Example 3–4 SPARC: Rebooting a Cluster

The following example shows the console output when normal cluster operation is stopped, all nodes are shut down to the ok prompt, and the cluster is restarted. The -g 0 option sets the grace period to zero, and -y provides an automatic yes response to the confirmation question. Shutdown messages also appear on the consoles of other nodes in the cluster.


# cluster shutdown -g0 -y
Wed Mar 10 13:47:32 phys-schost-1 cl_runtime: 
WARNING: CMM monitoring disabled.
phys-schost-1# 
INIT: New run level: 0
The system is coming down.  Please wait.
...
The system is down.
syncing file systems... done
Program terminated
ok boot
Rebooting with command: boot 
...
Hostname: phys-schost-1
Booting as part of a cluster
...
NOTICE: Node phys-schost-1: attempting to join cluster
...
NOTICE: Node phys-schost-2 (incarnation # 937690106) has become reachable.
NOTICE: Node phys-schost-3 (incarnation # 937690290) has become reachable.
NOTICE: cluster has reached quorum.
...
NOTICE: Cluster members: phys-schost-1 phys-schost-2 phys-schost-3.
...
NOTICE: Node phys-schost-1: joined cluster
...
The system is coming up.  Please wait.
checking ufs filesystems
...
reservation program successfully exiting
Print services started.
volume management starting.
The system is ready.
phys-schost-1 console login:
NOTICE: Node phys-schost-1: joined cluster
...
The system is coming up.  Please wait.
checking ufs filesystems
...
reservation program successfully exiting
Print services started.
volume management starting.
The system is ready.
phys-schost-1 console login: