Sun Cluster System Administration Guide for Solaris OS

Chapter 7 Administering the Cluster

This chapter provides the procedures for administering items that affect the entire cluster.

This is a list of the procedures in this chapter.

Administering the Cluster Overview

Table 7–1 Task List: Administering the Cluster

Task 

For Instructions, Go To 

Change the name of the cluster 

How to Change the Cluster Name

List node IDs and their corresponding node names 

How to Map Node ID to Node Name

Permit or deny new nodes to add themselves to the cluster 

How to Work With New Cluster Node Authentication

Change the time for a cluster using the Network Time Protocol (NTP) 

How to Reset the Time of Day in a Cluster

Bring down a node to the OpenBoot PROM ok prompt on a SPARC based system or to a Boot Subsystem on an x86 based system

SPARC: How to Enter the OpenBoot PROM (OBP) on a Node

Change the private hostname 

How to Change the Private Hostname

Put a cluster node in maintenance state 

How to Put a Node Into Maintenance State

Bring a cluster node out of maintenance state 

How to Bring a Node Out of Maintenance State

Add a node to a cluster 

How to Add a Node to the Authorized Node List

Remove a node from a cluster 

How to Remove a Node From the Cluster Software Configuration

How to Change the Cluster Name

If necessary, you can change the cluster name after initial installation.

  1. Become superuser on any node in the cluster.

  2. Type the scsetup command.


    # scsetup
    

    The Main Menu is displayed.

  3. To change the cluster name, type 7 (Other cluster properties).

    The Other Cluster Properties menu is displayed.

  4. Make your selection from the menu and follow the onscreen instructions.

Example—Changing the Cluster Name

The following example shows the scconf(1M) command generated from the scsetup(1M) utility to change to the new cluster name, dromedary.


# scconf -c -C cluster=dromedary

How to Map Node ID to Node Name

During Sun Cluster installation, each node is automatically assigned a unique node ID number. The node ID number is assigned to a node in the order in which it joins the cluster for the first time; once assigned, the number cannot be changed. The node ID number is often used in error messages to identify which cluster node the message concerns. Use this procedure to determine the mapping between node IDs and node names.

You do not need to be superuser to list configuration information.

  1. Use the scconf(1M) command to list the cluster configuration information.


    % scconf -pv | grep "Node ID"
    

Example—Mapping the Node ID to the Node Name

The following example shows the node ID assignments


% scconf -pv | grep “Node ID”
(phys-schost-1) Node ID:                    1
(phys-schost-2) Node ID:                    2
(phys-schost-3) Node ID:                    3

How to Work With New Cluster Node Authentication

Sun Cluster enables you to determine if new nodes can add themselves to the cluster and with what type of authentication. You can permit any new node to join the cluster over the public network, deny new nodes from joining the cluster, or indicate a specific node that can join the cluster. New nodes can be authenticated by using either standard UNIX or Diffie-Hellman (DES) authentication. If you select DES authentication, you must also configure all necessary encryption keys before a node can join. See the keyserv(1M) and publickey(4) man pages for more information.

  1. Become superuser on any node in the cluster.

  2. Enter the scsetup(1M) utility.


    # scsetup
    

    The Main Menu is displayed.

  3. To work with cluster authentication, type 6 (New nodes).

    The New Nodes menu is displayed.

  4. Make your selection from the menu and follow the onscreen instructions.

Examples—Preventing New Machines From Being Added to the Cluster

The following example shows the scconf command generated from the scsetup utility that would prevent new machines from being added to the cluster.


# scconf -a -T node=.

Examples—Permitting All New Machines to Be Added to the Cluster

The following example shows the scconf command generated from the scsetup utility that would enable all new machines to be added to the cluster.


# scconf -r -T all

Examples—Specifying a New Machine to Be Added to the Cluster

The following example shows the scconf command generated from the scsetup utility to enable a single new machine to be added to the cluster.


# scconf -a -T node=phys-schost-4

Examples—Setting the Authentication to Standard UNIX

The following example shows the scconf command generated from the scsetup utility to reset to standard UNIX authentication for new nodes joining the cluster.


# scconf -c -T authtype=unix

Examples—Setting the Authentication to DES

The following example shows the scconf command generated from the scsetup utility to use DES authentication for new nodes joining the cluster.


# scconf -c -T authtype=des

Note –

When using DES authentication, you need to also configure all necessary encryption keys before a node can join the cluster. See the keyserv(1M) and publickey(4) man pages for more information.


How to Reset the Time of Day in a Cluster

Sun Cluster uses the Network Time Protocol (NTP) to maintain time synchronization between cluster nodes. Adjustments in the cluster occur automatically as needed when nodes synchronize their time. See the Sun Cluster Concepts Guide for Solaris OS and the Network Time Protocol User's Guide for more information.


Caution – Caution –

When using NTP, do not attempt to adjust the cluster time while the cluster is up and running. This includes using the date(1), rdate(1M), or xntpd(1M) commands interactively or within cron(1M) scripts.


  1. Become superuser on any node in the cluster.

  2. Shut down the cluster.


    # scshutdown -g0 -y
    

  3. Verify that the node is showing the ok prompt or the Select (b)oot or (i)nterpreter prompt on the Current Boot Parameters screen.

  4. Boot the node in non-cluster mode by using the boot(1M) or the b command with the -x option.

    • SPARC:


      ok boot -x
      

    • x86:


                            <<< Current Boot Parameters >>>
      Boot path: /pci@0,0/pci8086,2545@3/pci8086,1460@1d/pci8086,341a@7,1/
      sd@0,0:a
      Boot args:
      
      Type    b [file-name] [boot-flags] <ENTER>  to boot with options
      or      i <ENTER>                           to enter boot interpreter
      or      <ENTER>                             to boot with defaults
      
                        <<< timeout in 5 seconds >>>
      Select (b)oot or (i)nterpreter: b -x
      

  5. On a single node, set the time of day by running the date command.


    # date HHMM.SS
    

  6. On the other machines, synchronize the time to that node by running the rdate(1M) command.


    # rdate hostname
    

  7. Boot each node to restart the cluster.


    # reboot
    

  8. Verify that the change took place on all cluster nodes.

    On each node, run the date command.


    # date
    

SPARC: How to Enter the OpenBoot PROM (OBP) on a Node

Use this procedure if you need to configure or change OpenBoot PROM settings.

  1. Connect to the terminal concentrator port.


    # telnet tc_name tc_port_number
    

    tc_name

    Specifies the name of the terminal concentrator.

    tc_port_number

    Specifies the port number on the terminal concentrator. Port numbers are configuration dependent. Typically, ports 2 and 3 (5002 and 5003) are used for the first cluster installed at a site.

  2. Shut down the cluster node gracefully by using the scswitch(1M) command to evacuate any resource or disk device groups and then shutdown.


    # scswitch -S -h node[,...]
    # shutdown -g0 -y -i0
    


    Caution – Caution –

    Do not use send brk on a cluster console to shut down a cluster node.


  3. Execute the OBP commands.

How to Change the Private Hostname

Use this procedure to change the private hostname of a cluster node after installation has been completed.

Default private hostnames are assigned during initial cluster installation. The default private hostname takes the form clusternode<nodeid>-priv, for example: clusternode3-priv. You should only change a private hostname if the name is already in use in the domain.


Caution – Caution –

Do not attempt to assign IP addresses to new private hostnames. The clustering software assigns them.


  1. Disable, on all nodes in the cluster, any Data Service resources or other applications that might cache private hostnames.


    # scswitch -n -j resource1, resource2
    

    Include the following in the applications you disable.

    • HA-DNS and HA-NFS services, if configured.

    • Any application which has been custom configured to use the private hostname.

    • Any application which is being used by clients over the private interconnect.

    See the scswitch(1M) man page and the Sun Cluster Data Services Planning and Administration Guide for Solaris OS for information about using the scswitch command.

  2. Bring down the Network Time Protocol (NTP) daemon on each node of the cluster.

    See the xntpd(1M) man page for more information about the NTP daemon.


    # /etc/initd./xntpd.cluster stop
    

  3. Run the scsetup(1M) utility to change the private hostname of the appropriate node.

    It is only necessary to do this from one of the nodes in the cluster.


    Note –

    When selecting a new private hostname, be sure the name is unique to the cluster node.


  4. Select 5, Private Hostnames, from the Main Menu.

  5. Select 1, Change a Private Hostname, from the Private Hostnames Menu.

    Answer the questions when prompted. You will be asked the name of the node whose private hostname is being changed (clusternode<nodeid>-priv), and the new private hostname.

  6. Flush the name service cache.

    Do this on each node in the cluster. This prevents the cluster applications and data services from trying to access the old private hostname.


    # nscd -i hosts
    

  7. Edit the ntp.conf.cluster file on each node to change the private hostname to the new one.

    Use whatever editing tool you prefer.

    If this is done at install time, also remember to remove names for nodes which are configured; the default template comes pre-configured with sixteen nodes. Typically, the ntp.conf.cluster file will be identical on each cluster node.

  8. Verify that you can successfully ping the new private hostname from all cluster nodes.

  9. Restart the NTP daemon.

    Do this on each node of the cluster.


    # /etc/initd./xntpd.cluster start
    

  10. Enable all Data Service resources and other applications that were disabled in Step 1.


    # scswitch -e -j resource1, resource2
    

    See the scswitch man page and the Sun Cluster Data Services Planning and Administration Guide for Solaris OS for information about using the scswitch command.

Example—Changing the Private Hostname

The following example changes the private hostname from clusternode2-priv to clusternode4-priv, on node phys-schost-2.


[Disable all applications and data services as necessary.]
phys-schost-1# /etc/init.d/xntpd stop
phys-schost-1# scconf -p | grep node
 ...
 Cluster nodes:                  phys-schost-1 phys-schost-2 phys-
 schost-3
 Cluster node name:                                 phys-schost-1
  Node private hostname:                           clusternode1-priv
 Cluster node name:                                 phys-schost-2
  Node private hostname:                           clusternode2-priv
 Cluster node name:                                 phys-schost-3
  Node private hostname:                           clusternode3-priv
 ...
phys-schost-1# scsetup
phys-schost-1# nscd -i hosts
phys-schost-1# vi /etc/inet/ntp.conf
 ...
 peer clusternode1-priv
 peer clusternode4-priv
 peer clusternode3-priv
phys-schost-1# ping clusternode4-priv
phys-schost-1# /etc/init.d/xntpd start
[Enable all applications and data services disabled at the beginning of the procedure.]

How to Put a Node Into Maintenance State

Put a cluster node into maintenance state when taking the node out of service for an extended period of time. This way, the node does not contribute to the quorum count while it is being serviced. To put a cluster node into maintenance state, the node must be brought down using scswitch(1M) and shutdown(1M).


Note –

Use the Solaris shutdown command to shut down a single node. The scshutdown command should be used only when shutting down an entire cluster.


When a cluster node is brought down and put into maintenance state, all quorum devices that are configured with ports to the node have their quorum vote counts decremented by one. The node and quorum device vote counts are incremented by one when the node is taken out of maintenance mode and brought back online.

You need to use the scconf(1M) command to put a cluster node into maintenance state. The scsetup(1M) utility does not include the functionality for putting a quorum device into maintenance state.

  1. Become superuser on the node to be put into maintenance state.

  2. Evacuate any resource groups and disk device groups from the node.


    # scswitch -S -h node[,...]
    

    -S

    Evacuates all device services and resource groups from the specified node.

    -h node[,...]

    Specifies the node from which you are switching resource groups and devices groups.

  3. Shut down the node that you evacuated.


    # shutdown -g0 -y -i0
    

  4. Become superuser on another node in the cluster and put the node that you shut down in Step 3 into maintenance state.


    # scconf -c -q node=node,maintstate
    

    -c

    Specifies the change form of the scconf command.

    -q

    Manages the quorum options.

    node=node

    Specifies the node name or node ID of the node to change.

    maintstate

    Puts the node into maintenance state.

  5. Verify that the cluster node is now in maintenance state.


    # scstat -q
    

    The node you put into maintenance state should have a Status of offline and 0 (zero) for Present and Possible quorum votes.

Example—Putting a Cluster Node Into Maintenance State

The following example moves a cluster node into maintenance state and verifies the results. The scstat -q output shows the Node votes for phys-schost-1 to be 0 (zero) and the status to be Offline. The Quorum Summary should also show reduced vote counts. Depending on your configuration, the Quorum Votes by Device output might indicate that some quorum disk devices are offline as well.


[On the node to be put into maintenance state:]
phys-schost-1# scswitch -S -h phys-schost-1
phys-schost-1# shutdown -g0 -y -i0

[On another node in the cluster:]
phys-schost-2# scconf -c -q node=phys-schost-1,maintstate
phys-schost-2# scstat -q

-- Quorum Summary --
  Quorum votes possible:      3
  Quorum votes needed:        2
  Quorum votes present:       3

-- Quorum Votes by Node --
                    Node Name           Present Possible Status
                    ---------           ------- -------- ------
  Node votes:       phys-schost-1       0        0       Offline
  Node votes:       phys-schost-2       1        1       Online
  Node votes:       phys-schost-3       1        1       Online

-- Quorum Votes by Device --
                    Device Name         Present Possible Status
                    -----------         ------- -------- ------
  Device votes:     /dev/did/rdsk/d3s2  0        0       Offline
  Device votes:     /dev/did/rdsk/d17s2 0        0       Offline
  Device votes:     /dev/did/rdsk/d31s2 1        1       Online

Where to Go From Here

To bring a node back online, see How to Bring a Node Out of Maintenance State.

How to Bring a Node Out of Maintenance State

Use the following procedure to bring a node back online and reset the quorum vote count to the default. For cluster nodes, the default quorum count is one. For quorum devices, the default quorum count is N-1, where N is the number of nodes with non-zero vote counts that have ports to the quorum device.

When a node has been put into maintenance state, the node's quorum vote count is decremented by one. All quorum devices that are configured with ports to the node will also have their quorum vote counts decremented. When the quorum vote count is reset and a node is brought back out of maintenance state, both the node's quorum vote count and the quorum device vote count are incremented by one.

Run this procedure any time a node has been put into maintenance state and you are bringing it out of maintenance state.


Caution – Caution –

If you do not specify either the globaldev or node options, the quorum count is reset for the entire cluster.


  1. Become superuser on any node of the cluster, other than the one in maintenance state.

  2. Are you bringing a node out of maintenance state in a two-node cluster?

  3. If using quorum, reset the cluster quorum count from a node other than the one in maintenance state.

    You must reset the quorum count from a node other than the node in maintenance state before rebooting the node, or it might hang waiting for quorum.


    # scconf -c -q node=node,reset
    

    -c

    Specifies the change form of the scconf command.

    -q

    Manages the quorum options.

    node=node

    Specifies the name of the node to be reset, for example, phys-schost-1.

    reset

    The change flag that resets quorum.

  4. Boot the node that you want to bring out of maintenance state.

  5. Verify the quorum vote count.


    # scstat -q
    

    The node you brought out of maintenance state should have a status of online and show the appropriate vote count for Present and Possible quorum votes.

Example—Bringing a Cluster Node Out of Maintenance State and Resetting the Quorum Vote Count

The following example resets the quorum count for a cluster node and its quorum devices to their defaults and verifies the result. The scstat -q output shows the Node votes for phys-schost-1 to be 1 and the status to be online. The Quorum Summary should also show an increase in vote counts.


phys-schost-2# scconf -c -q node=phys-schost-1,reset

On phys-schost-1:


phys-schost-1# scstat -q

-- Quorum Summary --

  Quorum votes possible:      6
  Quorum votes needed:        4
  Quorum votes present:       6

-- Quorum Votes by Node --

                    Node Name           Present Possible Status
                    ---------           ------- -------- ------
  Node votes:       phys-schost-1       1        1       Online
  Node votes:       phys-schost-2       1        1       Online
  Node votes:       phys-schost-3       1        1       Online

-- Quorum Votes by Device --

                    Device Name         Present Possible Status
                    -----------         ------- -------- ------
  Device votes:     /dev/did/rdsk/d3s2  1        1       Online
  Device votes:     /dev/did/rdsk/d17s2 1        1       Online
  Device votes:     /dev/did/rdsk/d31s2 1        1       Online

Adding and Removing a Cluster Node

The following table lists the tasks to perform when adding a node to an existing cluster. To complete the procedure correctly, these tasks must be performed in the order shown.

Table 7–2 Task Map: Adding a Cluster Node to an Existing Cluster

Task 

For Instructions, Go To 

Install the host adapter on the node and verify that the existing cluster interconnects can support the new node 

Sun Cluster Hardware Administration Manual for Solaris OS

Add shared storage 

Sun Cluster Hardware Administration Manual for Solaris OS

Add the node to the authorized node list 

   - Use scsetup.

How to Add a Node to the Authorized Node List

Install and configure the software on the new cluster node 

   - Install the Solaris Operating Environment and Sun Cluster software 

   - Configure the node as part of the cluster 

“Installing and Configuring Sun Cluster Software” in Sun Cluster Software Installation Guide for Solaris OS

The following table lists the tasks to perform when removing a node from an existing cluster. To complete the procedure correctly, the tasks must be performed in the order shown.


Caution – Caution –

Do not use this procedure if your cluster is running an OPS configuration. At this time, removing a node in an OPS configuration might cause nodes to panic at reboot.


Table 7–3 Task Map: Removing a Cluster Node (5/02)

Task 

For Instructions, Go To 

Move all resource groups and disk device groups off of the node to be removed. 

   - Use scswitch(1M)

# scswitch -S -h from-node

Remove the node from all resource groups. 

   - Use scrgadm(1M)

Sun Cluster Data Services Planning and Administration Guide for Solaris OS

Remove node from all disk device groups 

   - Use scconf(1M), metaset(1M), and scsetup(1M)

How to Remove a Node From a Disk Device Group (Solstice DiskSuite/Solaris Volume Manager)

SPARC: How to Remove a Node From a Disk Device Group (VERITAS Volume Manager)

SPARC: How to Remove a Node From a Raw Disk Device Group

Caution: If the number of desired secondaries is configured as 2 or more, it must be decreased to 1.

Remove all fully connected quorum devices. 

   - Use scsetup.

Caution: Do not remove the quorum device if you are removing a node from a two-node cluster.

 

How to Remove a Quorum Device

 

Note that although you must remove the quorum device before you remove the storage device in the next step, you can add the quorum device back immediately afterward. 

Remove all fully connected storage devices from the node.  

  - Use devfsadm(1M), scdidadm(1M).

Caution: Do not remove the quorum device if you are removing a node from a two-node cluster.How to Remove Connectivity Between an Array and a Single Node, in a Cluster With Greater Than Two-Node Connectivity

Add back the quorum devices (to only the nodes that are intended to remain in the cluster). 

  - Use scconf -a -q globaldev=d[n],node=node1,node=node2

scconf(1M)

Place the node being removed into maintenance state. 

   - Use scswitch(1M), shutdown(1M), and scconf(1M).

How to Put a Node Into Maintenance State

Remove all logical transport connections (transport cables and adapters) to the node being removed. 

   - Use scsetup.

How to Remove Cluster Transport Cables, Transport Adapters, and Transport Junctions

Remove all quorum devices connected to the node being removed. 

   - Use scsetup, scconf(1M).

How to Remove the Last Quorum Device From a Cluster

Remove node from the cluster software configuration. 

   - Use scconf(1M).

How to Remove a Node From the Cluster Software Configuration

How to Add a Node to the Authorized Node List

Before adding a machine to an existing cluster, be sure the node has all of the necessary hardware correctly installed and configured, including a good physical connection to the private cluster interconnect.

For hardware installation information, refer to the Sun Cluster Hardware Administration Manual for Solaris OS or the hardware documentation that shipped with your server.

This procedure permits a machine to install itself into a cluster by adding its node name to the list of authorized nodes for that cluster.

You must be superuser on a current cluster member to complete this procedure.

  1. Be sure you have correctly completed all prerequisite hardware installation and configuration tasks listed in the task map for Adding and Removing a Cluster Node.

  2. Type the scsetup command.


    # scsetup
    

    The Main Menu is displayed.

  3. To access the New Nodes Menu, type 7 at the Main Menu.

  4. To modify the authorized list, type 3 at the New Nodes Menu, Specify the name of a machine which may add itself.

    Follow the prompts to add the node's name to the cluster. You will be asked for the name of the node to be added.

  5. Verify that the task has been performed successfully.

    The scsetup utility prints a “Command completed successfully” message if it completes the task without error.

  6. Quit the scsetup utility.

  7. Install and configure the software on the new cluster node.

    Use either scinstall or JumpStartTM to complete the installation and configuration of the new node, as described in the Sun Cluster Software Installation Guide for Solaris OS.

  8. To prevent any new machines from being added to the cluster, type 1 at the New Nodes Menu.

    Follow the scsetup prompts. This option tells the cluster to ignore all requests coming in over the public network from any new machine trying to add itself to the cluster.

Example—Adding a Cluster Node to the Authorized Node List

The following example shows how to add a node named phys-schost-3 to the authorized node list in an existing cluster.


[Become superuser and execute the scsetup utility.]
# scsetup
Select New nodes>Specify the name of a machine which may add itself.
Answer the questions when prompted.
Verify that the scconf command completed successfully.
 
scconf -a -T node=phys-schost-3
 
    Command completed successfully.
Select Prevent any new machines from being added to the cluster.
Quit the scsetup New Nodes Menu and Main Menu.
[Install the cluster software.]

Where to Go From Here

For an overall list of tasks for adding a cluster node, see Table 7–2, “Task Map: Adding a Cluster Node.”

To add a node to an existing resource group, see the Sun Cluster Data Services Planning and Administration Guide for Solaris OS.

How to Remove a Node From the Cluster Software Configuration

Perform this procedure to remove a node from the cluster.

  1. Be sure you have correctly completed all prerequisite tasks listed in the “Removing a Cluster Node” task map in Adding and Removing a Cluster Node.


    Note –

    Be sure you have removed the node from all resource groups, disk device groups, and quorum device configurations and placed it in maintenance state before you continue with this procedure.


  2. Become superuser on a node in the cluster other than the node to remove.

  3. Remove the node from the cluster.


    # scconf -r -h node=node-name
    

  4. Verify the node removal by using scstat(1M).


    # scstat -n
    

  5. Do you intend to uninstall Sun Cluster software from the removed node?

Example—Removing a Node From the Cluster Software Configuration

This example shows how to remove a node (phys-schost-2) from a cluster. All commands are run from another node of the cluster (phys-schost-1).


[Remove the node from the cluster:]
phys-schost-1# scconf -r -h node=phys-schost-2
[Verify node removal:]
phys-schost-1# scstat -n
-- Cluster Nodes --
                    Node name           Status
                    ---------           ------
  Cluster node:     phys-schost-1       Online

Where to Go From Here

To uninstall Sun Cluster software from the removed node, see How to Uninstall Sun Cluster Software From a Cluster Node

For hardware procedures, see theSun Cluster Hardware Administration Manual for Solaris OS .

For an overall list of tasks for removing a cluster node, see Table 7–3.

To add a node to an existing cluster, see How to Add a Node to the Authorized Node List.

How to Remove Connectivity Between an Array and a Single Node, in a Cluster With Greater Than Two-Node Connectivity

Use this procedure to detach a storage array from a single cluster node, in a cluster that has three- or four-node connectivity.

  1. Back up all database tables, data services, and volumes that are associated with the storage array that you are removing.

  2. Determine the resource groups and device groups that are running on the node to be disconnected.


    # scstat
    
  3. If necessary, move all resource groups and device groups off the node to be disconnected.


    Caution (SPARC only) – Caution (SPARC only) –

    If your cluster is running Oracle Parallel Server/Real Application Clusters software, shut down the Oracle Parallel Server/Real Application Clusters database instance that is running on the node before you move the groups off the node. For instructions see the Oracle Database Administration Guide.



    # scswitch -S -h from-node
    
  4. Put the device groups into maintenance state.

    For the procedure on acquiescing I/O activity to Veritas shared disk groups, see your VERITAS Volume Manager documentation.

    For the procedure on putting a device group in maintenance state, see the “Administering the Cluster”.

  5. Remove the node from the device groups.

    • If you use VERITAS Volume Manager or raw disk, use the scconf(1M) command to remove the device groups.

    • If you use Solstice DiskSuite, use the metaset command to remove the device groups.

  6. If the cluster is running HAStorage or HAStoragePlus, remove the node from the resource group's nodelist.


    # scrgadm -a -g resource-group -h nodelist 
    

    See the Sun Cluster Data Services Planning and Administration Guide for Solaris OS for more information on changing a resource group's nodelist.


    Note –

    Resource type, resource group, and resource property names are case insensitive when executing scrgadm.


  7. If the storage array you are removing is the last storage array that is connected to the node, disconnect the fiber-optic cable between the node and the hub or switch that is connected to this storage array (otherwise, skip this step).

  8. Do you want to remove the host adapter from the node you are disconnecting?

    • If yes, shut down and power off the node.

    • If no, skip to Step 11.

  9. Remove the host adapter from the node.

    For the procedure on removing host adapters, see the documentation that shipped with your node.

  10. Without allowing the node to boot, power on the node.

  11. Boot the node into non-cluster mode.

    • SPARC:


      ok boot -x
      

    • x86:


                            <<< Current Boot Parameters >>>
      Boot path: /pci@0,0/pci8086,2545@3/pci8086,1460@1d/pci8086,341a@7,1/
      sd@0,0:a
      Boot args:
      
      Type    b [file-name] [boot-flags] <ENTER>  to boot with options
      or      i <ENTER>                           to enter boot interpreter
      or      <ENTER>                             to boot with defaults
      
                        <<< timeout in 5 seconds >>>
      Select (b)oot or (i)nterpreter: b -x
      


    Caution (SPARC only) – Caution (SPARC only) –

    The node must be in non-cluster mode before you remove Oracle Parallel Server/Real Application Clusters software in the next step or the node panics and potentially causes a loss of data availability.


  12. SPARC: If Oracle Parallel Server/Real Application Clusters software has been installed, remove the Oracle Parallel Server/Real Application Clusters software package from the node that you are disconnecting.


    # pkgrm SUNWscucm 
    

    Caution (SPARC only) – Caution (SPARC only) –

    If you do not remove the Oracle Parallel Server/Real Application Clusters software from the node you disconnected, the node will panic when the node is reintroduced to the cluster and potentially cause a loss of data availability.


  13. Boot the node into cluster mode.

    • SPARC:


      ok boot
      

    • x86:


                            <<< Current Boot Parameters >>>
      Boot path: /pci@0,0/pci8086,2545@3/pci8086,1460@1d/pci8086,341a@7,1/
      sd@0,0:a
      Boot args:
      
      Type    b [file-name] [boot-flags] <ENTER>  to boot with options
      or      i <ENTER>                           to enter boot interpreter
      or      <ENTER>                             to boot with defaults
      
                        <<< timeout in 5 seconds >>>
      Select (b)oot or (i)nterpreter: b
      

  14. On the node, update the device namespace by updating the /devices and /dev entries.


    # devfsadm -C 
    # scdidadm -C
    
  15. Bring the device groups back online.

    For procedures about bringing a VERITAS shared disk group online, see your VERITAS Volume Manager documentation.

    For the procedure on bringing a device group online, see the procedure on putting a device group into maintenance state.

How to Uninstall Sun Cluster Software From a Cluster Node

Perform this procedure to uninstall Sun Cluster software from a cluster node before you disconnect it from a fully established cluster configuration. You can use this procedure to uninstall software from the last remaining node of a cluster.


Note –

To uninstall Sun Cluster software from a node that has not yet joined the cluster or is still in install mode, do not perform this procedure. Instead, go to “How to Uninstall Sun Cluster Software to Correct Installation Problems” in the Sun Cluster Software Installation Guide for Solaris OS.


  1. Be sure you have correctly completed all prerequisite tasks listed in the task map for removing a cluster node.

    See Adding and Removing a Cluster Node.


    Note –

    Be sure you have removed the node from all resource groups, device groups, and quorum device configurations, placed it in maintenance state, and removed it from the cluster before you continue with this procedure.


  2. Become superuser on an active cluster member other than the node you will uninstall.

  3. From the active cluster member, add the node you intend to uninstall to the cluster's node authentication list.


    # scconf -a -T node=nodename
    
    -a

    Add

    -T

    Specifies authentication options

    node=nodename

    Specifies the name of the node to add to the authentication list

    Alternately, you can use the scsetup(1M) utility. See How to Add a Node to the Authorized Node List for procedures.

  4. Become superuser on the node to uninstall.

  5. Reboot the node into non-cluster mode.

    • SPARC:


      # shutdown -g0 -y -i0
      ok boot -x
      

    • x86:


      # shutdown -g0 -y -i0
      ...
                            <<< Current Boot Parameters >>>
      Boot path: /pci@0,0/pci8086,2545@3/pci8086,1460@1d/pci8086,341a@7,1/
      sd@0,0:a
      Boot args:
      
      Type    b [file-name] [boot-flags] <ENTER>  to boot with options
      or      i <ENTER>                           to enter boot interpreter
      or      <ENTER>                             to boot with defaults
      
                        <<< timeout in 5 seconds >>>
      Select (b)oot or (i)nterpreter: b -x
      

  6. In the /etc/vfstab file, remove all globally mounted file system entries except the /global/.devices global mounts.

  7. Uninstall Sun Cluster software from the node.

    Run the command from a directory that is not associated with any Sun Cluster packages.


    # cd /
    # scinstall -r
    

    See the scinstall(1M) man page for more information. If scinstall returns error messages, see Unremoved Cluster File System Entries.

  8. Disconnect the transport cables and the transport junction, if any, from the other cluster devices.

    1. If the uninstalled node is connected to a storage device that uses a parallel SCSI interface, install a SCSI terminator to the open SCSI connector of the storage device after you disconnect the transport cables.

      If the uninstalled node is connected to a storage device that uses Fibre Channel interfaces, no termination is necessary.

    2. Follow the documentation that shipped with your host adapter and server for disconnection procedures.

Correcting Error Messages

To correct the error messages in the previous sections, perform this procedure.

  1. Attempt to rejoin the node to the cluster.


    # boot
    

  2. Did the node successfully rejoin the cluster?

    • If no, proceed to Step 3.

    • If yes, perform the following steps to remove the node from disk device groups.

    1. If the node successfully rejoins the cluster, remove the node from the remaining disk device group(s).

      Follow procedures in How to Remove a Node From All Disk Device Groups.

    2. After you remove the node from all disk device groups, return to How to Uninstall Sun Cluster Software From a Cluster Node and repeat the procedure.

  3. If the node could not rejoin the cluster, rename the node's /etc/cluster/ccr file to any other name you choose, for example, ccr.old.


    # mv /etc/cluster/ccr /etc/cluster/ccr.old
    

  4. Return to How to Uninstall Sun Cluster Software From a Cluster Node and repeat the procedure.

Troubleshooting a Node Uninstallation

This section describes error messages you might receive when you run the scinstall -r command and the corrective actions to take.

Unremoved Cluster File System Entries

The following error messages indicate that the node you removed still has cluster file systems referenced in its vfstab file.


Verifying that no unexpected global mounts remain in /etc/vfstab ... failed
scinstall:  global-mount1 is still configured as a global mount.
scinstall:  global-mount1 is still configured as a global mount.
scinstall:  /global/dg1 is still configured as a global mount.
 
scinstall:  It is not safe to uninstall with these outstanding errors.
scinstall:  Refer to the documentation for complete uninstall instructions.
scinstall:  Uninstall failed.

To correct this error, return to How to Uninstall Sun Cluster Software From a Cluster Node and repeat the procedure. Ensure that you successfully complete Step 6 in the procedure before you rerun the scinstall -r command.

Unremoved Listing in Disk Device Groups

The following error messages indicate that the node you removed is still listed with a disk device group.


Verifying that no device services still reference this node ... failed
scinstall:  This node is still configured to host device service "service".
scinstall:  This node is still configured to host device service "service2".
scinstall:  This node is still configured to host device service "service3".
scinstall:  This node is still configured to host device service "dg1".
 
scinstall:  It is not safe to uninstall with these outstanding errors.          
scinstall:  Refer to the documentation for complete uninstall instructions.
scinstall:  Uninstall failed.