Perform this procedure to add a new node to an existing global cluster. To use JumpStart to add a new node, instead follow procedures in How to Install Solaris and Sun Cluster Software (JumpStart).
This procedure uses the interactive form of the scinstall command. To use the noninteractive forms of the scinstall command, such as when developing installation scripts, see the scinstall(1M) man page.
Ensure that Sun Cluster software packages are installed on the node, either manually or by using the silent-mode form of the Java ES installer program, before you run the scinstall command. For information about running the Java ES installer program from an installation script, see Chapter 5, Installing in Silent Mode, in Sun Java Enterprise System 5 Installation Guide for UNIX.
Perform the following tasks:
Ensure that the Solaris OS is installed to support Sun Cluster software.
If Solaris software is already installed on the node, you must ensure that the Solaris installation meets the requirements for Sun Cluster software and any other software that you intend to install on the cluster. See How to Install Solaris Software for more information about installing Solaris software to meet Sun Cluster software requirements.
SPARC: If you are configuring Sun Logical Domains (LDoms) I/O domains or guest domains as cluster nodes, ensure that Sun LDoms software is installed on each physical machine and that the domains meet Sun Cluster requirements. See SPARC: How to Install Sun Logical Domains Software and Create Domains.
Ensure that Sun Cluster software packages and patches are installed on the node. See How to Install Sun Cluster Framework and Data-Service Software Packages.
Ensure that the cluster is prepared for the addition of the new node. See How to Prepare the Cluster for Additional Global-Cluster Nodes.
Determine which mode of the scinstall utility you will use, Typical or Custom. For the Typical installation of Sun Cluster software, scinstall automatically specifies the following configuration defaults.
Component |
Default Value |
---|---|
Cluster-transport switches |
switch1 and switch2 |
Global-devices file-system name |
/globaldevices |
Complete one of the following configuration planning worksheets. See Planning the Solaris OS and Planning the Sun Cluster Environment for planning guidelines.
Typical Mode Worksheet – If you will use Typical mode and accept all defaults, complete the following worksheet.
Custom Mode Worksheet – If you will use Custom mode and customize the configuration data, complete the following worksheet.
Follow these guidelines to use the interactive scinstall utility in this procedure:
Interactive scinstall enables you to type ahead. Therefore, do not press the Return key more than once if the next menu screen does not appear immediately.
Unless otherwise noted, you can press Control-D to return to either the start of a series of related questions or to the Main Menu.
Default answers or answers to previous sessions are displayed in brackets ([ ]) at the end of a question. Press Return to enter the response that is in brackets without typing it.
On the cluster node to configure, become superuser.
Start the scinstall utility.
phys-schost-new# /usr/cluster/bin/scinstall |
The scinstall Main Menu is displayed.
Type the option number for Create a New Cluster or Add a Cluster Node and press the Return key.
*** Main Menu *** Please select from one of the following (*) options: * 1) Create a new cluster or add a cluster node 2) Configure a cluster to be JumpStarted from this install server 3) Manage a dual-partition upgrade 4) Upgrade this cluster node * 5) Print release information for this cluster node * ?) Help with menu options * q) Quit Option: 1 |
The New Cluster and Cluster Node Menu is displayed.
Type the option number for Add This Machine as a Node in an Existing Cluster and press the Return key.
Follow the menu prompts to supply your answers from the configuration planning worksheet.
The scinstall utility configures the node and boots the node into the cluster.
Unload the Sun Java Availability Suite DVD-ROM from the DVD-ROM drive.
Repeat this procedure on any other node to add to the cluster until all additional nodes are fully configured.
For the Solaris 10 OS, verify on each node that multiuser services for the Service Management Facility (SMF) are online.
If services are not yet online for a node, wait until the state becomes online before you proceed to the next step.
phys-schost# svcs multi-user-server node STATE STIME FMRI online 17:52:55 svc:/milestone/multi-user-server:default |
From an active cluster member, prevent any other nodes from joining the cluster.
phys-schost# claccess deny-all |
Alternately, you can use the clsetup utility. See How to Add a Node to the Authorized Node List in Sun Cluster System Administration Guide for Solaris OS for procedures.
On one node, become superuser.
Verify that all nodes have joined the cluster.
phys-schost# clnode status |
Output resembles the following.
=== Cluster Nodes === --- Node Status --- Node Name Status --------- ------ phys-schost-1 Online phys-schost-2 Online phys-schost-3 Online |
For more information, see the clnode(1CL) man page.
Verify that all necessary patches are installed.
phys-schost# showrev -p |
(Optional) Enable automatic node reboot if all monitored disk paths fail.
Enable the automatic reboot feature.
phys-schost# clnode set -p reboot_on_path_failure=enabled |
Specifies the property to set
Specifies that the node will reboot if all monitored disk paths fail, provided that at least one of the disks is accessible from a different node in the cluster.
Verify that automatic reboot on disk-path failure is enabled.
phys-schost# clnode show === Cluster Nodes === Node Name: node … reboot_on_path_failure: enabled … |
If you intend to use Sun Cluster HA for NFS on a highly available local file system, ensure that the loopback file system (LOFS) is disabled.
To disable LOFS, add the following entry to the /etc/system file on each node of the cluster.
exclude:lofs |
The change to the /etc/system file becomes effective after the next system reboot.
You cannot have LOFS enabled if you use Sun Cluster HA for NFS on a highly available local file system and have automountd running. LOFS can cause switchover problems for Sun Cluster HA for NFS. If you choose to add Sun Cluster HA for NFS on a highly available local file system, you must make one of the following configuration changes.
However, if you configure non-global zones in your cluster, you must enable LOFS on all cluster nodes. If Sun Cluster HA for NFS on a highly available local file system must coexist with LOFS, use one of the other solutions instead of disabling LOFS.
Disable LOFS.
Disable the automountd daemon.
Exclude from the automounter map all files that are part of the highly available local file system that is exported by Sun Cluster HA for NFS. This choice enables you to keep both LOFS and the automountd daemon enabled.
See The Loopback File System in System Administration Guide: Devices and File Systems (Solaris 9 or Solaris 10) for more information about loopback file systems.
The following example shows the node phys-schost-3 added to the cluster schost. The sponsoring node is phys-schost-1.
*** Adding a Node to an Existing Cluster *** Fri Feb 4 10:17:53 PST 2005 scinstall -ik -C schost -N phys-schost-1 -A trtype=dlpi,name=qfe2 -A trtype=dlpi,name=qfe3 -m endpoint=:qfe2,endpoint=switch1 -m endpoint=:qfe3,endpoint=switch2 Checking device to use for global devices file system ... done Adding node "phys-schost-3" to the cluster configuration ... done Adding adapter "qfe2" to the cluster configuration ... done Adding adapter "qfe3" to the cluster configuration ... done Adding cable to the cluster configuration ... done Adding cable to the cluster configuration ... done Copying the config from "phys-schost-1" ... done Copying the postconfig file from "phys-schost-1" if it exists ... done Copying the Common Agent Container keys from "phys-schost-1" ... done Setting the node ID for "phys-schost-3" ... done (id=1) Setting the major number for the "did" driver ... Obtaining the major number for the "did" driver from "phys-schost-1" ... done "did" driver major number set to 300 Checking for global devices global file system ... done Updating vfstab ... done Verifying that NTP is configured ... done Initializing NTP configuration ... done Updating nsswitch.conf ... done Adding clusternode entries to /etc/inet/hosts ... done Configuring IP Multipathing groups in "/etc/hostname.<adapter>" files Updating "/etc/hostname.hme0". Verifying that power management is NOT configured ... done Ensure that the EEPROM parameter "local-mac-address?" is set to "true" ... done The "local-mac-address?" parameter setting has been changed to "true". Ensure network routing is disabled ... done Updating file ("ntp.conf.cluster") on node phys-schost-1 ... done Updating file ("hosts") on node phys-schost-1 ... done Rebooting ... |
Unsuccessful configuration – If one or more nodes cannot join the cluster, or if the wrong configuration information was specified, first attempt to rerun this procedure. If that does not correct the problem, perform the procedure How to Unconfigure Sun Cluster Software to Correct Installation Problems on each misconfigured node to remove it from the cluster configuration. You do not need to uninstall the Sun Cluster software packages. Then rerun this procedure.
If you added a node to an existing cluster that uses a quorum device, go to How to Update Quorum Devices After Adding a Node to a Global Cluster.
Otherwise, go to How to Verify the Quorum Configuration and Installation Mode.