How to Install and Configure Oracle Solaris and Oracle Solaris Cluster Software (Unified Archives)

You will use the AI server to install a cluster from the Unified Archives and configure its nodes. This procedure retains all the software packages that are contained in the Unified Archives, but you must provide the new cluster configuration that you designed in the worksheet. Before you perform this procedure, you must first create the archive. See Step 1 below for instructions on creating the recovery archive.

The AI server sets up installation of the nodes from the Unified Archives and creates the cluster with the new configuration. Only a Unified Archive created in the global zone is accepted. You can use either a clone archive or a recovery archive. The following list describes the differences between the two archives:

  • When you install from a clone archive, only the global zone is installed. Any zones in the archive are not installed. When you install from a recovery archive, both the global zone and the zones contained in the archive are installed.

  • A clone archive does not contain system configuration, including IPMPs, VLANs, and VNICs.

  • A clone archive only contains the BE that is active when the archive is created, therefore only that BE is installed. A recovery archive can contain multiple BEs, but only the active BE is updated with the new cluster configuration.

This procedure prompts you for the cluster name, node names and their MAC addresses, the path to the Unified Archives, and the cluster configuration you designed in the worksheet.

  1. Assume the root role on a node of the global cluster and create an archive.
    phys-schost# archiveadm create -r archive-location

    Use the create command to create a clone archive or the create -r option to create a recovery archive. For more information on using the archiveadm command, see the archiveadm(8) man page.

  2. Set up your Automated Installer (AI) install server and DHCP server.

    Ensure that the AI install server meets the following requirements.

    • The install server is on the same subnet as the cluster nodes.

    • The install server is not itself a cluster node.

    • The install server runs a release of the Oracle Solaris OS that is supported by the Oracle Solaris Cluster software.

    • Each new cluster node is configured as a custom AI installation client that uses the custom AI directory that you set up for Oracle Solaris Cluster installation.

    Follow the appropriate instructions for your software platform and OS version to set up the AI install server and DHCP server. See Chapter 4, Setting Up the AI Server in Automatically Installing Oracle Solaris 11.4 Systems and Working With DHCP in Oracle Solaris 11.4.

  3. Log into the Automated Installer server and assume the root role.
  4. On the AI install server, install the Oracle Solaris Cluster AI support package.
    1. Ensure that the solaris and ha-cluster publishers are valid.
      installserver# pkg publisher
      PUBLISHER        TYPE     STATUS   URI
      solaris          origin   online   solaris-repository
      ha-cluster       origin   online   ha-cluster-repository
    2. Install the cluster AI support package.
      installserver# pkg install ha-cluster/system/install
  5. On the AI install server, start the scinstall utility.
    installserver# /usr/cluster/bin/scinstall

    The scinstall Main Menu is displayed.

  6. Type the option number and press Return.
    *** Main Menu ***
    
    Please select from one of the following (*) options:
    
    * 1) Install, restore, or replicate a cluster from this Automated Installer server
    * 2) Securely install, restore, or replicate a cluster from this Automated Installer server
    * 3) Print release information for this Automated Installer install server
    
    * ?) Help with menu options
    * q) Quit
    
    Option:  2

    Choose Option 1 if you want to install a cluster using a non-secure AI server installation. Choose Option 2 for a secure AI installation.

    The Custom Automated Installer Menu or Custom Secure Automated Installer Menu is displayed.

  7. Type the option number to Install and Configure a New Cluster from Unified Archives and press Return.

    The Custom Automated Installer User screen is displayed.

  8. Type the password and press Return.

    Type the password a second time to confirm it. The Typical or Customer Mode screen is displayed.

  9. Type the option number for the install mode you will use.

    The Cluster Name screen is displayed.

  10. Type the name of the cluster you want to install and press Return.

    The Cluster Nodes screen is displayed.

  11. Types the names of the cluster nodes that you plan to install from the Unified Archives and press Return.

    If the scinstall utility is unable to find the MAC address of the nodes, type in each address when prompted and press Return. You can then choose to install all the nodes from the same archive, or use a different archive for each node.

  12. Type the full path to the archive and press Return.

    The archive can either be a recovery archive or a clone archive.

    The Cluster Transport Adapters and Cables screen is displayed.

  13. Type the names of the cluster transport adapters and press Return.

    Select the type of each transport adapter. The Resource Security Configuration screen is displayed.

  14. Choose to enable to disable automatic quorum device selection and press Return.

    The Confirmation screen is displayed.

  15. For each node, confirm the options you chose so that the scinstall utility performs the necessary configuration to install the cluster nodes from this AI server.

    The utility also prints instructions to add the DHCP macros on the DHCP server, and adds (if you chose secure installation) or clears (if you chose non-secure installation) the security keys for SPARC nodes. Follow those instructions.

  16. (Optional) To customize the target device, update the AI manifest for each node.

    The AI manifest is located in the following directory:

    /var/cluster/logs/install/autoscinstall.d/ \
    cluster-name/node-name/node-name_aimanifest.xml
    1. To customize the target device, update the target element in the manifest file.

      scinstall assumes the existing boot disk in the manifest file to be the target device. To customize the target device, update the target element in the manifest file based on how you want to use the supported criteria to locate the target device for the installation. For example, you can specify the disk_name sub-element.

      For more information, see Configuring an AI Server in Automatically Installing Oracle Solaris 11.4 Systems and the ai_manifest(5) man page.

    2. Run the installadm command for each node.
      # installadm update-manifest -n cluster-name-{sparc|i386} \ 
      -f /var/cluster/logs/install/autoscinstall.d/cluster-name/node-name/node-name_aimanifest.xml \
      -m node-name_manifest

    Note that SPARC and i386 is the architecture of the cluster node.

  17. If you are using a cluster administrative console, display a console screen for each node in the cluster.
    • If pconsole software is installed and configured on your administrative console, use the pconsole utility to display the individual console screens.

      As the root role, use the following command to start the pconsole utility:

      adminconsole# pconsole host[:port] [...] &

      The pconsole utility also opens a master window from which you can send your input to all individual console windows at the same time.

    • If you do not use the pconsole utility, connect to the consoles of each node individually.

  18. Shut down and boot each node to start the AI installation.

    The Oracle Solaris software is installed with the default configuration.

    Note:

    You cannot use this method if you want to customize the Oracle Solaris installation. If you choose the Oracle Solaris interactive installation, the Automated Installer is bypassed and Oracle Solaris Cluster software is not installed and configured. To customize Oracle Solaris during installation, instead follow instructions in How to Install Oracle Solaris Software, then install and configure the cluster by following instructions in How to Install Oracle Solaris Cluster Software (pkg).
    • SPARC:

      1. Shut down each node.

        phys-schost# shutdown -g0 -y -i0
      2. Boot the node with the following command

        ok boot net:dhcp - install

        Note:

        Surround the dash (-) in the command with a space on each side.
    • x86:

      1. Reboot the node.

        # reboot -p
      2. During PXE boot, press Control-N.

        The GRUB menu is displayed.

      3. Immediately select the Automated Install entry and press Return.

        Note:

        If you do not select the Automated Install entry within 20 seconds, installation proceeds using the default interactive text installer method, which will not install and configure the Oracle Solaris Cluster software.

        Each node will be automatically rebooted a few times before the node completely joins the cluster. Ignore any error messages from SMF services on the console. On each node, the Automated Installer installs the software that is contained in the Unified Archives. When the installation is successfully completed, each node is fully installed as a new cluster node. Oracle Solaris Cluster installation output is logged in the /var/cluster/logs/install/scinstall.log.N file and the /var/cluster/logs/install/sc_ai_config.log file on each node.

  19. From one node, verify that all nodes have joined the cluster.
    phys-schost# clnode status

    Output resembles the following.

    === Cluster Nodes ===
    
    --- Node Status ---
    
    Node Name                                       Status
    ---------                                       ------
    phys-schost-1                                   Online
    phys-schost-2                                   Online
    phys-schost-3                                   Online

    For more information, see the clnode(8CL) man page.

  20. If you plan to enable RPC use of TCP wrappers, add all clprivnet0 IP addresses to the /etc/hosts.allow file on each cluster node.

    Without this addition to the /etc/hosts.allow file, TCP wrappers prevent internode communication over RPC for cluster administration utilities.

    1. On each node, display the IP addresses for all clprivnet0 devices on the node.

      # /usr/sbin/ipadm show-addr
      ADDROBJ           TYPE     STATE        ADDR
      clprivnet0/N      static   ok           ip-address/netmask-length
    2. On each cluster node, add to the /etc/hosts.allow file the IP addresses of all clprivnet0 devices in the cluster.
  21. If you use the LDAP naming service, you must manually configure it on the cluster nodes after they boot.