Perform this procedure from one node of the global cluster to configure Sun Cluster software on all nodes of the cluster.
This procedure uses the interactive form of the scinstall command. To use the noninteractive forms of the scinstall command, such as when developing installation scripts, see the scinstall(1M) man page.
Ensure that Sun Cluster software packages are installed on the node, either manually or by using the silent-mode form of the Java ES installer program, before you run the scinstall command. For information about running the Java ES installer program from an installation script, see Chapter 5, Installing in Silent Mode, in Sun Java Enterprise System 5 Update 1 Installation Guide for UNIX.
Perform the following tasks:
Ensure that the Solaris OS is installed to support Sun Cluster software.
If Solaris software is already installed on the node, you must ensure that the Solaris installation meets the requirements for Sun Cluster software and any other software that you intend to install on the cluster. See How to Install Solaris Software for more information about installing Solaris software to meet Sun Cluster software requirements.
SPARC: If you are configuring Sun Logical Domains (LDoms) I/O domains or guest domains as cluster nodes, ensure that Sun LDoms software is installed on each physical machine and that the domains meet Sun Cluster requirements. See SPARC: How to Install Sun Logical Domains Software and Create Domains.
Ensure that Sun Cluster software packages and patches are installed on each node. See How to Install Sun Cluster Framework and Data-Service Software Packages.
Determine which mode of the scinstall utility you will use, Typical or Custom.
For the Typical installation of Sun Cluster software, scinstall automatically specifies the following configuration defaults.
Component |
Default Value |
---|---|
Private-network address |
172.16.0.0 |
Private-network netmask |
Solaris 9: 255.255.248.0Solaris 10: 255.255.240.0 |
Cluster-transport adapters |
Exactly two adapters |
Cluster-transport switches |
switch1 and switch2 |
Global fencing |
Enabled |
Global-devices file-system name |
/globaldevicesIf scinstall does not find a mounted /globaldevices file system in /etc/vfstab on a node, it prompts you to instead configure a lofi device or specify another file-system name |
Installation security (DES) |
Limited |
Complete one of the following cluster configuration worksheets, depending on whether you run the scinstall utility in Typical mode or Custom mode.
Typical Mode Worksheet – If you will use Typical mode and accept all defaults, complete the following worksheet.
Custom Mode Worksheet – If you will use Custom mode and customize the configuration data, complete the following worksheet.
If you are installing a single-node cluster, the scinstall utility automatically assigns the default private network address and netmask, even though the cluster does not use a private network.
Follow these guidelines to use the interactive scinstall utility in this procedure:
Interactive scinstall enables you to type ahead. Therefore, do not press the Return key more than once if the next menu screen does not appear immediately.
Unless otherwise noted, you can press Control-D to return to either the start of a series of related questions or to the Main Menu.
Default answers or answers to previous sessions are displayed in brackets ([ ]) at the end of a question. Press Return to enter the response that is in brackets without typing it.
If you disabled remote configuration during Sun Cluster software installation, re-enable remote configuration.
Enable remote shell (rsh(1M)) or secure shell (ssh(1)) access for superuser to all cluster nodes.
If you are using switches in the private interconnect of your new cluster, ensure that Neighbor Discovery Protocol (NDP) is disabled.
Follow the procedures in the documentation for your switches to determine whether NDP is enabled and to disable NDP.
During cluster configuration, the software checks that there is no traffic on the private interconnect. If NDP sends any packages to a private adapter when the private interconnect is being checked for traffic, the software will assume that the interconnect is not private and cluster configuration will be interrupted. NDP must therefore be disabled during cluster creation.
After the cluster is established, you can re-enable NDP on the private-interconnect switches if you want to use that feature.
From one cluster node, start the scinstall utility.
phys-schost# /usr/cluster/bin/scinstall |
Type the option number for Create a New Cluster or Add a Cluster Node and press the Return key.
*** Main Menu *** Please select from one of the following (*) options: * 1) Create a new cluster or add a cluster node * 2) Configure a cluster to be JumpStarted from this install server 3) Manage a dual-partition upgrade 4) Upgrade this cluster node * 5) Print release information for this cluster node * ?) Help with menu options * q) Quit Option: 1 |
The New Cluster and Cluster Node Menu is displayed.
Type the option number for Create a New Cluster and press the Return key.
The Typical or Custom Mode menu is displayed.
Type the option number for either Typical or Custom and press the Return key.
The Create a New Cluster screen is displayed. Read the requirements, then press Control-D to continue.
Follow the menu prompts to supply your answers from the configuration planning worksheet.
The scinstall utility installs and configures all cluster nodes and reboots the cluster. The cluster is established when all nodes have successfully booted into the cluster. Sun Cluster installation output is logged in a /var/cluster/logs/install/scinstall.log.N file.
For the Solaris 10 OS, verify on each node that multiuser services for the Service Management Facility (SMF) are online.
If services are not yet online for a node, wait until the state changes to online before you proceed to the next step.
phys-schost# svcs multi-user-server node STATE STIME FMRI online 17:52:55 svc:/milestone/multi-user-server:default |
From one node, verify that all nodes have joined the cluster.
phys-schost# clnode status |
Output resembles the following.
=== Cluster Nodes === --- Node Status --- Node Name Status --------- ------ phys-schost-1 Online phys-schost-2 Online phys-schost-3 Online |
For more information, see the clnode(1CL) man page.
(Optional) Enable the automatic node reboot feature.
This feature automatically reboots a node if all monitored shared-disk paths fail, provided that at least one of the disks is accessible from a different node in the cluster.
Enable automatic reboot.
phys-schost# clnode set -p reboot_on_path_failure=enabled |
Specifies the property to set
Enables automatic node reboot if failure of all monitored shared-disk paths occurs.
Verify that automatic reboot on disk-path failure is enabled.
phys-schost# clnode show === Cluster Nodes === Node Name: node … reboot_on_path_failure: enabled … |
If you intend to use Sun Cluster HA for NFS on a highly available local file system, ensure that the loopback file system (LOFS) is disabled.
To disable LOFS, add the following entry to the /etc/system file on each node of the cluster.
exclude:lofs |
The change to the /etc/system file becomes effective after the next system reboot.
You cannot have LOFS enabled if you use Sun Cluster HA for NFS on a highly available local file system and have automountd running. LOFS can cause switchover problems for Sun Cluster HA for NFS. If you choose to add Sun Cluster HA for NFS on a highly available local file system, you must make one of the following configuration changes.
However, if you configure non-global zones in your cluster, you must enable LOFS on all cluster nodes. If Sun Cluster HA for NFS on a highly available local file system must coexist with LOFS, use one of the other solutions instead of disabling LOFS.
Disable LOFS.
Disable the automountd daemon.
Exclude from the automounter map all files that are part of the highly available local file system that is exported by Sun Cluster HA for NFS. This choice enables you to keep both LOFS and the automountd daemon enabled.
See The Loopback File System in System Administration Guide: Devices and File Systems (Solaris 9 or Solaris 10) for more information about loopback file systems.
The following example shows the scinstall progress messages that are logged as scinstall completes configuration tasks on the two-node cluster, schost. The cluster is installed from phys-schost-1 by using the scinstall utility in Typical Mode. The other cluster node is phys-schost-2. The adapter names are qfe2 and qfe3. The automatic selection of a quorum device is enabled. Both nodes use the partition /globaldevices for the global-devices namespace.
Installation and Configuration Log file - /var/cluster/logs/install/scinstall.log.24747 Testing for "/globaldevices" on "phys-schost-1" … done Testing for "/globaldevices" on "phys-schost-2" … done Checking installation status … done The Sun Cluster software is already installed on "phys-schost-1". The Sun Cluster software is already installed on "phys-schost-2". Starting discovery of the cluster transport configuration. The following connections were discovered: phys-schost-1:qfe2 switch1 phys-schost-2:qfe2 phys-schost-1:qfe3 switch2 phys-schost-2:qfe3 Completed discovery of the cluster transport configuration. Started cluster check on "phys-schost-1". Started cluster check on "phys-schost-2". cluster check completed with no errors or warnings for "phys-schost-1". cluster check completed with no errors or warnings for "phys-schost-2". Removing the downloaded files … done Configuring "phys-schost-2" … done Rebooting "phys-schost-2" … done Configuring "phys-schost-1" … done Rebooting "phys-schost-1" … Log file - /var/cluster/logs/install/scinstall.log.24747 Rebooting … |
Unsuccessful configuration – If one or more nodes cannot join the cluster, or if the wrong configuration information was specified, first attempt to rerun this procedure. If that does not correct the problem, perform the procedure How to Unconfigure Sun Cluster Software to Correct Installation Problems on each misconfigured node to remove it from the cluster configuration. You do not need to uninstall the Sun Cluster software packages. Then rerun this procedure.
If you installed a single-node cluster, cluster establishment is complete. Go to Creating Cluster File Systems to install volume management software and configure the cluster.
If you installed a multiple-node cluster and chose automatic quorum configuration, postinstallation setup is complete. Go to How to Verify the Quorum Configuration and Installation Mode.
If you installed a multiple-node cluster and declined automatic quorum configuration, perform postinstallation setup. Go to How to Configure Quorum Devices.
If you intend to configure any quorum devices in your cluster, go to How to Configure Quorum Devices.
Otherwise, go to How to Verify the Quorum Configuration and Installation Mode.