C H A P T E R  3

Installing the Software in a Non-NFS Configuration

This chapter explains how to install and configure Sun HPC ClusterTools software on a Solaris OS-based non-NFS configuration using the installation utilities. For information about how to install the software in a Solaris OS-based NFS configuration, see Chapter 4. For information about how to install Sun HPC ClusterTools software on a Linux-based system, see Chapter 6.

The Sun HPC ClusterTools software CLI utilities are:



Note - If you use rsh connections for centralized operations on hundreds of nodes at a time, the operations may encounter system resource limitations that prevent the connections from being established to all the nodes. For clusters with hundreds of nodes, it is best to perform these operations on subsets of nodes, one subset at a time, with no more than 200 nodes in a subset.


This chapter contains the following topics:


Installing Sun HPC ClusterTools Software



Note - You must log in as root to install and configure the Sun HPC ClusterTools 8.1 software. This is because ctinstall and the other commands use the Solaris OS pkg* commands; the pkg* commands must be executed as root.


Before you can install and configure the software, you must make sure you have downloaded the correct software archive for your hardware platform. Once you have downloaded the software, you need to extract it to the correct directory. If you have installed a previous version of the software, there are additional steps you need to do to prepare for installation. The following procedure explains these steps.


procedure icon  To Download and Extract the Software



Note - If you have previously installed Sun HPC ClusterTools 8.1 software on your cluster nodes, you must remove it before you can reinstall the software.


1. Boot the cluster nodes.

2. Download and extract the archive file containing the Sun HPC ClusterTools software to a location (such as a shared file system) that is visible to all the nodes in the cluster.

a. Make sure that all the compute and administrative nodes have access to the shared file system.

b. Make sure that the file system is readable by superuser and accessible through a common path from all nodes.

For centralized installations, do this on the central host as well.

You can obtain the correct HPC ClusterTools archive file for your platform from the following location:

http://www.sun.com/clustertools/

3. Log in as superuser on the system from which you will be initiating the command: the central host or, if initiating the command locally, directly on a cluster node.

4. If the Sun HPC ClusterTools software has not been installed yet, change directory to one of the following:

a. If you are installing the software on a SPARC-based system, change directory to /sun-hpc-ct8.1-sparc/Product/Install_Utilities/bin

b. If you are installing on an x64-based system, change directory to /sun-hpc-ct-i386/Product/Install_Utilities/bin

If the software has already been installed, go to Step 5.

5. If the software was previously installed and you intend to perform such tasks as activation, deactivation, or removal of the software, change directory to $INSTALL_LOC/SUNWhpc/HPC8.1/bin/Install_Utilities/bin.

$INSTALL_LOC is the location where the software was installed.

You can now start using the CLI commands. They are described separately below, with examples of common applications given for each.

For usage information on any command, either enter the command without options or with the -h option.


# ./command

or

# ./command -h

ctinstall Options

Use the ctinstall command to install Sun HPC ClusterTools software on cluster nodes. See TABLE 3-1 for a summary of the ctinstall options.


TABLE 3-1 ctinstall Options

Options

Description

General

 

-h

Command help.

-l

Execute the command on the local node only.

-R

Specify the full path to be used as the root path.

-x

Turn on command debug at the specified nodes.

Command Specific

 

-a

Activate automatically after installation completes.

-c

Specify the server and mount path for the software if you are performing an NFS installation.

-d

Specify a non-default install from location. The default is distribution/Product, relative to the directory where ctinstall is invoked.

-p

List of packages to be installed. Separate names by comma.

-t

Specify a nondefault install to location. The default is /opt.

Centralized Operations Only

 

-g

Generate node lists of successful and unsuccessful installations.

-k

Specify a central location for storing log files of all specified nodes.

-n

List of nodes targeted for installation. Separate names by comma.

-N

File containing list of nodes targeted for installation. One node per line.

-r

Remote connection method: rsh, ssh, or telnet.

-S

Specify full path to an alternate ssh executable.




Note - The options -g, -k, -n, -N, -r, and -S are incompatible with local (non-centralized) installations. If the -l option is used with any of these options, an error message is displayed.


Installing the Software From a Central Host in a Non-NFS Configuration

This section discusses the following options for installing the software from a central host:


procedure icon  To Install From a Central Host Using rsh

This section shows examples of software installations in which the ctinstall command is initiated from a central host.


# ./ctinstall -n node1,node2 -r rsh


procedure icon  To Install From a Central Host Using ssh

The following command installs the full Sun HPC ClusterTools software suite on node1 and node2 from a central host. The node list is specified on the command line. The remote connection method is rsh. This requires a trusted hosts setup.

The software will be ready for use when the installation process completes.


# ./ctinstall -n node1,node2 -r ssh

This example is the same as that in the previous section, except that the remote connection method is ssh. This method requires that the initiating node be able to log in as superuser to the target nodes without being prompted for any interaction, such as a password.


procedure icon  To Install From a Central Host Using telnet


# ./ctinstall -N /tmp/nodelist -r telnet

This command installs the full Sun HPC ClusterTools software suite on the set of nodes listed in the file /tmp/nodelist from a central host. A node list file is particularly useful when you have a large set of nodes or you want to run operations on the same set of nodes repeatedly.

The node list file has the following contents:


# Node list for the above example
 
node1
node2

The remote connection method is telnet. All cluster nodes must share the same password. If some nodes do not use the same password as others, install the software in groups, each group consisting of nodes that use a common password.

The software will be ready for use when the installation process completes.


procedure icon  To Install the Software and Save the Log Files


# ./ctinstall -N /tmp/nodelist -r telnet -k /tmp/cluster-logs -g

The command in this section is the same as that shown in the previous section, except that it includes the -k and -g options.

In this example, the -k option causes the local log files of all specified nodes to be saved in /tmp/cluster-logs on the central host.

The -g option causes a pair of node list files to be created on the central host in /var/sadm/system/logs/hpc/nodelists. One file, ctinstall.pass$$, contains a list of the nodes on which the installation was successful. The other file, ctinstall.fail$$, lists the nodes on which the installation was unsuccessful. The $$ symbol is replaced by the process number associated with the installation.

These generated node list files can then be used for command retries or in subsequent operations using the -N switch.



Note - Specify a directory that is local to the central host rather than an NFS-mounted directory. This will avoid unnecessary network traffic in the transfer of log files and will result in faster execution of the operation.



procedure icon  To Install Specified Software Packages

The following command installs only the specified software packages.


# ./ctinstall -N /tmp/nodelist -r telnet -p SUNWompi,SUNWompimn

This command installs the packages SUNWompi and SUNWompimn on the set of nodes listed in the file /tmp/nodelist. No other packages are installed. The remote connection method is telnet.

Appendix A lists the packages in the Sun HPC ClusterTools 8.1 installation.

The -p option can be useful if individual packages were not installed on the nodes by ctinstall.


# ./ctinstall -N /tmp/nodelist -r rsh

This command installs and activates the full Sun HPC ClusterTools software suite on the nodes listed in the file /tmp/nodelist. The remote connection method is rsh.

Installing Software Locally in a Non-NFS Configuration

This section shows examples of software installations in which the ctinstall command is initiated on the local node. The installation options shown in this example include:



Note - The options -g, -k, -n, -N, -r, and -S are incompatible with local (non-centralized) installations. If the -l option is used with any of these options, an error message is displayed.



procedure icon  To Install the Complete Software Suite


# ./ctinstall -l

This command installs the full Sun HPC ClusterTools software suite on the local node only.


procedure icon  To Install Specified Software Packages


# ./ctinstall -l -p SUNWompi,SUNWompimn

The command in this section installs the packages SUNWompi and SUNWompimn on the local node.

Appendix A lists the packages in the Sun HPC ClusterTools 8.1 installation.


Activating Sun HPC ClusterTools Software

In Sun HPC ClusterTools 8.1 software, the activation step sets up symbolic links to the program binaries. If you are running Sun HPC ClusterTools 8.1 only, you may skip the activation step and run Sun HPC ClusterTools 8.1 software from the directory in which it is installed (by default /opt/SUNWhpc/HPC8.1/bin).

ctact Options

Use the ctact command to activate Sun HPC ClusterTools software on cluster nodes. See TABLE 3-2 for a summary of the ctact options.



Note - The general options and options specific to centralized operations serve essentially the same role for ctact as for ctinstall. Consequently, fewer examples are used to illustrate ctact than were used for ctinstall.



TABLE 3-2 ctact Options

Options

Description

General

 

-h

Command help.

-l

Execute the command on the local node only.

-R

Specify the full path to be used as the root path.

-x

Turn on command debug at the specified nodes.

Command Specific

 

-c

Specify that you are activating on an NFS client node.

Centralized Operations Only

 

-g

Generate node lists of successful and unsuccessful activation.

-k

Specify a central location for storing copies of local log files.

-n

List of nodes targeted for activation. Separate names by comma.

-N

File containing list of nodes targeted for activation. One node per line.

-r

Remote connection method: rsh, ssh, or telnet.

-S

Specify full path to an alternate ssh executable.


Activating Nodes From a Central Host

This section shows examples of software activation in which the ctact command is initiated from a central host.


procedure icon  To Activate Specified Cluster Nodes From a Central Host in a Non-NFS Configuration


# ./ctact -n node1,node2 -r rsh

This command activates the software on node1 and node2. The remote connection method is rsh.


procedure icon  To Activate Cluster Nodes and Create Log Files


# ./ctact -n node1,node2 -r rsh -k /tmp/cluster-logs -g

This command is the same as the one in the previous section, except that it specifies the options -k and -g.

In this example, the -k option causes the local log files of all specified nodes to be saved in /tmp/cluster-logs on the central host.



Note - Specify a directory that is local to the central host rather than an NFS-mounted directory. This avoids unnecessary network traffic and results in faster execution of the operation.


The -g option causes files ctact.pass$$ and ctact.fail$$ to be created on the central host in /var/sadm/system/logs/hpc/nodelists. ctact.pass$$ lists the cluster nodes on which software activation was successful and ctact.fail$$ lists the nodes on which activation was unsuccessful. The $$ symbol is replaced by the process number associated with the activation.

These generated node list files can then be used for command retries or in subsequent operations using the -N switch.

Activating the Local Node

This section shows an example of software activation on the local node.


procedure icon  To Activate Locally


# ./ctact -l 

This command activates the software on the local node.


Deactivating Sun HPC ClusterTools Software

Use the ctdeact command to deactivate Sun HPC ClusterTools software on cluster nodes.

ctdeact Options

See TABLE 3-3 for a summary of the ctdeact options.


TABLE 3-3 ctdeact Options

Options

Description

General

 

-h

Command help.

-l

Execute the command on the local node only.

-R

Specify the full path to be used as the root path.

-x

Turn on command debug at the specified nodes.

Centralized Operations Only

 

-g

Generate node lists of successful and unsuccessful deactivation.

-k

Specify a central location for storing copies of local log files.

-n

List of nodes targeted for deactivation. Separate names by comma.

-N

File containing list of nodes to be deactivated. One node per line.

-r

Remote connection method: rsh, ssh, or telnet.

-S

Specify full path to an alternate ssh executable.


Deactivating Software From a Central Host

This section shows examples of software deactivation in which the ctdeact command is initiated from a central host.


procedure icon  To Deactivate Specified Cluster Nodes in a Non-NFS Configuration


# ./ctdeact -N /tmp/nodelist -r rsh

This command deactivates the software on the nodes listed in /tmp/nodelist. The remote connection method is rsh.


procedure icon  To Deactivate Cluster Nodes and Create Log Files


# ./ctdeact -N /tmp/nodelist -r rsh -k /tmp/cluster-logs -g

This command is the same as the command in the previous section, except that it specifies the options -k and -g.

In this example, the -k option causes the local log files of all specified nodes to be saved in /tmp/cluster-logs on the central host.



Note - Specify a directory that is local to the central host rather than an NFS-mounted directory. This will avoid unnecessary network traffic in the transfer of log files and will result in faster execution of the operation.


The -g option causes files ctdeact.pass$$ and ctdeact.fail$$ to be created on the central host. ctdeact.pass$$ lists the cluster nodes where software deactivation was successful. ctdeact.fail$$ lists the nodes where deactivation was unsuccessful. The $$ symbol is replaced by the process number associated with the software deactivation.

These generated node list files can then be used for command retries or in subsequent operations using the -N switch.

Deactivating the Local Node

This section shows software deactivation on the local node.


procedure icon  To Deactivate Locally


# ./ctdeact -l

This command deactivates the software on the local node.


Removing Sun HPC ClusterTools Software

Use the ctremove command to remove Sun HPC ClusterTools software from cluster nodes.

ctremove Options

See TABLE 3-4 for a summary of the ctremove options.



Note - If the nodes are active at the time ctremove is initiated, they will be deactivated automatically before the removal process begins.



TABLE 3-4 ctremove Options

Options

Description

General

 

-h

Command help.

-l

Execute the command on the local node only.

-R

Specify the full path to be used as the root path.

-x

Turn on command debug at the specified nodes.

Command Specific

 

-p

List of packages to be selectively removed. Separate names by comma.

-c 

Specify that you are removing Sun HPC ClusterTools 8.1 from the NFS client node.

Centralized Operations Only

 

-g

Generate node lists of successful and unsuccessful removals.

-k

Specify a central location for storing copies of local log files.

-n

List of nodes targeted for removal. Separate names by comma.

-N

File containing list of nodes targeted for removal. One node per line.

-r

Remote connection method: rsh, ssh, or telnet.

-S

Specify full path to an alternate ssh executable.


Removing Nodes From a Central Host

This section shows examples of software removal in which the ctremove command is initiated from a central host.


procedure icon  To Remove Software From Specified Cluster Nodes in a Non-NFS Configuration


# ./ctremove -N /tmp/nodelist -r rsh

This command removes the software from the nodes listed in /tmp/nodelist. The remote connection method is rsh.


procedure icon  To Remove the Software and Generate Log Files


# ./ctremove -N /tmp/nodelist -r rsh -k /tmp/cluster-logs -g

This command is the same as the command in the previous section, except that it specifies the options -k and -g.


procedure icon  To Remove Specified Software Packages


# ./ctremove -N /tmp/nodelist -r rsh -p SUNWompi,SUNWompimn

This command removes the packages SUNWompi and SUNWompimn from the nodes listed in /tmp/nodelist. The remote connection method is rsh.

Removing Software From the Local Node

This section shows software removal from the local node.


procedure icon  To Remove Software Locally


# ./ctremove -l

This command removes the software on the local node.


procedure icon  To Remove Specified Software Packages


# ./ctremove -l -p SUNWompi,SUNWompimn

This command removes the packages SUNWompi and SUNWompimn from the local node.