C H A P T E R  3

Installing the Software on Solaris

This chapter explains how to install Sun HPC ClusterTools software on Solaris using the installation utilities. For information about how to install Sun HPC ClusterTools software on a Linux-based system, see Chapter 4.

The Sun HPC ClusterTools software command-line interface (CLI) utilities are:

This chapter contains the following topics:


Download and Extract the Software



Note - You must log in as root to install and configure the Sun HPC ClusterTools 8.2 software. This is because ctinstall and the other commands use the Solaris OS pkg* commands; the pkg* commands must be executed as root.



Before you can install and configure the software, you must make sure you have downloaded the correct software archive for your hardware platform. Once you have downloaded the software, you need to extract it to the correct directory. If you have installed a previous version of the software, there are additional steps you need to do to prepare for installation. The following procedure explains these steps.

procedure icon  To Download and Extract the Software

  1. Boot the cluster nodes.

  2. Download and extract the archive file containing the Sun HPC ClusterTools software to a location (such as a shared file system) that is visible to all the nodes in the cluster.

    1. Make sure that all the compute and administrative nodes have access to the shared file system.

    2. Make sure that the file system is readable by superuser and accessible through a common path from all nodes.

    For centralized installations, do this on the central host as well.

    You can obtain the correct HPC ClusterTools archive file for your platform from the following location:

    http://www.sun.com/clustertools/

  3. Log in as superuser on the system from which you will be initiating the command: the central host or, if initiating the command locally, directly on a cluster node.

  4. If the Sun HPC ClusterTools software has not been installed yet, change directory to one of the following:

    1. If you are installing the software on a SPARC-based system, change directory to /sun-hpc-ct8.2-sparc/Product/Install_Utilities/bin

    2. If you are installing on an x64-based system, change directory to /sun-hpc-ct8.2-i386/Product/Install_Utilities/bin

    If the software has already been installed, go to Step 5.

  5. If the software was previously installed, remove the previous version of the software by changing the directory to $INSTALL_LOC/SUNWhpc/HPC8.2/bin/Install_Utilities/bin.

    $INSTALL_LOC is the location where the software was installed.

    To remove a previous version of the software, see Remove the Previous Sun HPC ClusterTools Software Installation.

    You can now start using the CLI commands. They are described separately below, with examples of common applications given for each.

    For usage information on any command, either enter the command without options or with the –h option.


    ./command
    or ./command –h
    


Remove the Previous Sun HPC ClusterTools Software Installation

If you have previously installed Sun HPC ClusterTools software on your cluster nodes, you must remove it before you can reinstall the software.

Use the ctremove command to remove Sun HPC ClusterTools software from cluster nodes.

ctremove Options

See Table 1 for a summary of the ctremove options.



Note - If the nodes are active at the time ctremove is initiated, they will be deactivated automatically before the removal process begins.




TABLE 3-1   ctremove Options  
Options Description
General  

–h

Command help.

–l

Execute the command on the local node only.

–R

Specify the full path to be used as the root path.

–x

Turn on command debug at the specified nodes.
Command Specific

–p

List of packages to be selectively removed. Separate names by comma.
Centralized Operations Only

–g

Generate node lists of successful and unsuccessful removals.

–k

Specify a central location for storing copies of local log files.

–n

List of nodes targeted for removal. Separate names by comma.

–N

File containing list of nodes targeted for removal. One node per line.

–r

Remote connection method: rsh, ssh, or telnet.

–S

Specify full path to an alternate ssh executable.

Removing Nodes From a Central Host

This section shows examples of software removal in which the ctremove command is initiated from a central host.

To Remove Software From Specified Cluster Nodes


./ctremove –N /tmp/nodelist –r rsh

This command removes the software from the nodes listed in /tmp/nodelist. The remote connection method is rsh.

To Remove the Software and Generate Log Files


./ctremove –N /tmp/nodelist –r rsh –k /tmp/cluster-logs –g

This command is the same as the command in the previous section, except that it specifies the options –k and –g.

To Remove Specified Software Packages


./ctremove –N /tmp/nodelist –r rsh –p SUNWompi,SUNWompimn

This command removes the packages SUNWompi and SUNWompimn from the nodes listed in /tmp/nodelist. The remote connection method is rsh.

Removing Software From the Local Node

This section shows software removal from the local node.

To Remove Software Locally


./ctremove –l

This command removes the software on the local node.

To Remove Specified Software Packages


./ctremove –l –p SUNWompi,SUNWompimn

This command removes the packages SUNWompi and SUNWompimn from the local node.


Install the Software

ctinstall Options

Use the ctinstall command to install Sun HPC ClusterTools software on cluster nodes. See TABLE 3-2 for a summary of the ctinstall options.


TABLE 3-2   ctinstall Options  
Options Description
General  

–h

Command help.

–l

Execute the command on the local node only.

–R

Specify the full path to be used as the root path.

–x

Turn on command debug at the specified nodes.
Command Specific

-c

Specify the server and mount path for the software.

–d

Specify a non-default install from location. The default is distribution/Product, relative to the directory where ctinstall is invoked.

–p

List of packages to be installed. Separate names by comma.

–t

Specify a nondefault install to location. The default is /opt.
Centralized Operations Only

–g

Generate node lists of successful and unsuccessful installations.

–k

Specify a central location for storing log files of all specified nodes.

–n

List of nodes targeted for installation. Separate names by comma.

–N

File containing list of nodes targeted for installation. One node per line.

–r

Remote connection method: rsh, ssh, or telnet.

–S

Specify full path to an alternate ssh executable.



Note - The options –g, –k, –n, –N, –r, and –S are incompatible with local (non-centralized) installations. If the –l option is used with any of these options, an error message is displayed. 



Install the Software from a Central Host

This section shows examples of software installations in which the ctinstall command is initiated from a central host. 

To Install From a Central Host Using rsh


./ctinstall –n node1,node2 –r rsh

This command installs the full Sun HPC ClusterTools software suite on node1 and node2 from a central host. The node list is specified on the command line. The remote connection method is ssh. This requires a trusted hosts setup.

The software will be ready for use when the installation process completes.

To Install From a Central Host Using ssh


./ctinstall –n node1,node2 –r ssh

This example is the same as that in the previous section, except that the remote connection method is ssh. This method requires that the initiating node be able to log in as superuser to the target nodes without being prompted for any interaction, such as a password.

To Install From a Central Host Using telnet


./ctinstall –N /tmp/nodelist –r telnet

This command installs the full Sun HPC ClusterTools software suite on the set of nodes listed in the file /tmp/nodelist from a central host. A node list file is particularly useful when you have a large set of nodes or you want to run operations on the same set of nodes repeatedly.

The node list file has the following contents:


# Node list for the above example
 
node1
node2

The remote connection method is telnet. All cluster nodes must share the same password. If some nodes do not use the same password as others, install the software in groups, each group consisting of nodes that use a common password.

The software will be ready for use when the installation process completes.

To Install the Software and Save the Log Files


./ctinstall –N /tmp/nodelist –r telnet –k /tmp/cluster-logs –g

The command in this section is the same as that shown in the previous section, except that it includes the –k and –g options.

In this example, the –k option causes the local log files of all specified nodes to be saved in /tmp/cluster-logs on the central host.

The –g option causes a pair of node list files to be created on the central host in /var/sadm/system/logs/hpc/nodelists. One file, ctinstall.pass$$, contains a list of the nodes on which the installation was successful. The other file, ctinstall.fail$$, lists the nodes on which the installation was unsuccessful. The $$ symbol is replaced by the process number associated with the installation.

These generated node list files can then be used for command retries or in subsequent operations using the –N switch.



Note - Specify a directory that is local to the central host (for example, /tmp). This will avoid unnecessary network traffic in the transfer of log files and will result in faster execution of the operation.



To Install Specified Software Packages

The following command installs only the specified software packages.


./ctinstall –N /tmp/nodelist –r telnet –p SUNWompi

This command installs the packages SUNWompi and SUNWompimn on the set of nodes listed in the file /tmp/nodelist. No other packages are installed. The remote connection method is telnet.

Solaris OS Packages lists the packages in the Sun HPC ClusterTools 8.2 installation.

The –p option can be useful if individual packages were not installed on the nodes by ctinstall.

 


./ctinstall –N /tmp/nodelist –r rsh

This command installs and activates the full Sun HPC ClusterTools software suite on the nodes listed in the file /tmp/nodelist. The remote connection method is rsh.

Solaris OS Packages

The following is the Solaris OS package breakdown for the Sun HPC ClusterTools 8.2 (Open MPI) release.


TABLE 3-3   Solaris OS Packages in the Sun HPC ClusterTools 8.2 Installation
Package Name Contents
SUNWompi Open MPI Message Passing Interface files
SUNWompiat Open MPI installer utilities
SUNWompimn Open MPI Message Passing Interface man pages
SUNWomsc Extra package to include miscellaneous files
SUNWompir Open MPI Root Filesystems files

Install the Software Locally

This section shows examples of software installations in which the ctinstall command is initiated on the local node. The installation options shown in this example include:



Note - The options –g, –k, –n, –N, –r, and –S are incompatible with local (non-centralized) installations. If the –l option is used with any of these options, an error message is displayed. 



To Install the Complete Software Suite Locally


./ctinstall –l

This command installs the full Sun HPC ClusterTools software suite on the local node only.

To Install Specified Software Packages Locally


./ctinstall –l –p SUNWompi,SUNWompimn

The command in this section installs the packages SUNWompi and SUNWompimn on the local node.

Solaris OS Packages lists the packages in the Sun HPC ClusterTools 8.2 installation.


Additional Steps

This chapter describes the final steps needed to get your Sun HPC system ready for use after installing the software, and information about viewing installation log files.

Enabling Close Integration With Batch Processing Systems

Open Run-Time Environment (ORTE) provides close integration with several distributed resource management (DRM) systems, such as Sun Grid Engine. For information on how close integration works and how to set up the integration for each of the supported resource managers, refer to the Sun HPC ClusterTools Software Administrator’s Guide.

The Sun HPC ClusterTools 8.2 Software User’s Guide contains information about how to set up close integration with Sun Grid Engine.

Sun HPC ClusterTools 8.2 Installation Log Files

The Sun HPC ClusterTools 8.2 installation tools log information about installation-related tasks locally on the nodes where installation tasks are performed. The default location for the log files is /var/sadm/system/logs/hpc. If installation tasks are initiated from a central host, a summary log file is also created on the central host.

Local, Node-Specific Log Files

Two types of log files are created locally on each cluster node where installation operations take place.

  • Task-specific logs – Separate log files are created for each installation-related task. They are:

    • ct_install.log

    • ct_remove.log

    • ct_act.log

    • ct_deact.log

    • ct_startd.log

    • ct_stopd.log

    These log files contain detailed logging information for the most recent associated task. Each time a task is repeated, its log file is overwritten.

  • History log – A ct_history.log file is created to store all installation-related tasks performed on the local node. This provides a convenient record of the Sun HPC ClusterTools 8.2 software installation history on the local node. Each time a new installation task is performed on the node, a new log entry is appended to the history log.

These node specific installation log files are created regardless of the installation method used, local or centralized.

Central Node Summary Log

When installation tasks are initiated from a central host, a summary log file named ct_summary.log is created on the central host. This log file records the final summary report that is generated by the CLI. The ct_summary.log is not overwritten when a new task is performed. As with the ct_history.log file, new entries are appended to the summary log file.