C H A P T E R 3 |
Installing the Software in a Non-NFS Configuration |
This chapter explains how to install and configure Sun HPC ClusterTools software on a Solaris OS-based non-NFS configuration using the installation utilities. For information about how to install the software in a Solaris OS-based NFS configuration, see Chapter 4. For information about how to install Sun HPC ClusterTools software on a Linux-based system, see Chapter 6.
The Sun HPC ClusterTools software CLI utilities are:
This chapter contains the following topics:
Before you can install and configure the software, you must make sure you have downloaded the correct software archive for your hardware platform. Once you have downloaded the software, you need to extract it to the correct directory. If you have installed a previous version of the software, there are additional steps you need to do to prepare for installation. The following procedure explains these steps.
To Download and Extract the Software |
Note - If you have previously installed Sun HPC ClusterTools 8 software on your cluster nodes, you must remove it before you can reinstall the software. |
2. Download and extract the archive file containing the Sun HPC ClusterTools software to a location (such as a shared file system) that is visible to all the nodes in the cluster.
a. Make sure that all the compute and administrative nodes have access to the shared file system.
b. Make sure that the file system is readable by superuser and accessible through a common path from all nodes.
For centralized installations, do this on the central host as well.
You can obtain the correct HPC ClusterTools archive file for your platform from the following location:
http://www.sun.com/clustertools/
3. Log in as superuser on the system from which you will be initiating the command: the central host or, if initiating the command locally, directly on a cluster node.
4. If the Sun HPC ClusterTools software has not been installed yet, change directory to one of the following:
a. If you are installing the software on a SPARC-based system, change directory to /sun-hpc-ct8.0-sparc/Product/Install_Utilities/bin
b. If you are installing on an x64-based system, change directory to /sun-hpc-ct-i386/Product/Install_Utilities/bin
If the software has already been installed, go to Step 5.
5. If the software was previously installed and you intend to perform such tasks as activation, deactivation, or removal of the software, change directory to $INSTALL_LOC/SUNWhpc/HPC8.0/bin/Install_Utilities/bin.
$INSTALL_LOC is the location where the software was installed.
You can now start using the CLI commands. They are described separately below, with examples of common applications given for each.
For usage information on any command, either enter the command without options or with the -h option.
Use the ctinstall command to install Sun HPC ClusterTools software on cluster nodes. See TABLE 3-1 for a summary of the ctinstall options.
Note - The options -g, -k, -n, -N, -r, and -S are incompatible with local (non-centralized) installations. If the -l option is used with any of these options, an error message is displayed. |
This section discusses the following options for installing the software from a central host:
This section shows examples of software installations in which the ctinstall command is initiated from a central host.
CODE EXAMPLE 3-1 installs the full Sun HPC ClusterTools software suite on node1 and node2 from a central host. The node list is specified on the command line. The remote connection method is rsh. This requires a trusted hosts setup.
The software will be ready for use when the installation process completes.
CODE EXAMPLE 3-2 is the same as CODE EXAMPLE 3-1, except that the remote connection method is ssh. This method requires that the initiating node be able to log in as superuser to the target nodes without being prompted for any interaction, such as a password.
CODE EXAMPLE 3-3 installs the full Sun HPC ClusterTools software suite on the set of nodes listed in the file /tmp/nodelist from a central host. A node list file is particularly useful when you have a large set of nodes or you want to run operations on the same set of nodes repeatedly.
The node list file has the following contents:
The remote connection method is telnet. All cluster nodes must share the same password. If some nodes do not use the same password as others, install the software in groups, each group consisting of nodes that use a common password.
The software will be ready for use when the installation process completes.
CODE EXAMPLE 3-4 is the same as CODE EXAMPLE 3-3, except it includes the -k and -g options.
In this example, the -k option causes the local log files of all specified nodes to be saved in /tmp/cluster-logs on the central host.
The -g option causes a pair of node list files to be created on the central host in /var/sadm/system/logs/hpc/nodelists. One file, ctinstall.pass$$, contains a list of the nodes on which the installation was successful. The other file, ctinstall.fail$$, lists the nodes on which the installation was unsuccessful. The $$ symbol is replaced by the process number associated with the installation.
These generated node list files can then be used for command retries or in subsequent operations using the -N switch.
The following command installs only the specified software packages.
CODE EXAMPLE 3-5 installs the packages SUNWompi and SUNWompimn on the set of nodes listed in the file /tmp/nodelist. No other packages are installed. The remote connection method is telnet.
Appendix A lists the packages in the Sun HPC ClusterTools 8 installation.
The -p option can be useful if individual packages were not installed on the nodes by ctinstall.
CODE EXAMPLE 3-6 installs and activates the full Sun HPC ClusterTools software suite on the nodes listed in the file /tmp/nodelist. The remote connection method is rsh.
This section shows examples of software installations in which the ctinstall command is initiated on the local node. The installation options shown in this example include:
Note - The options -g, -k, -n, -N, -r, and -S are incompatible with local (non-centralized) installations. If the -l option is used with any of these options, an error message is displayed. |
CODE EXAMPLE 3-7 installs the full Sun HPC ClusterTools software suite on the local node only.
CODE EXAMPLE 3-8 installs the packages SUNWompi and SUNWompimn on the local node.
Appendix A lists the packages in the Sun HPC ClusterTools 8 installation.
In Sun HPC ClusterTools 8 software, the activation step sets up symbolic links to the program binaries. If you are running Sun HPC ClusterTools 8 only, you may skip the activation step and run Sun HPC ClusterTools 8 software from the directory in which it is installed (by default /opt/SUNWhpc/HPC8.0/bin).
Use the ctact command to activate Sun HPC ClusterTools software on cluster nodes. See TABLE 3-2 for a summary of the ctact options.
This section shows examples of software activation in which the ctact command is initiated from a central host.
To Activate Specified Cluster Nodes From a Central Host in a Non-NFS Configuration |
CODE EXAMPLE 3-9 activates the software on node1 and node2 . The remote connection method is rsh.
CODE EXAMPLE 3-10 is the same as CODE EXAMPLE 3-9, except it specifies the options -k and -g.
In this example, the -k option causes the local log files of all specified nodes to be saved in /tmp/cluster-logs on the central host.
Note - Specify a directory that is local to the central host rather than an NFS-mounted directory. This avoids unnecessary network traffic and results in faster execution of the operation. |
The -g option causes files ctact.pass$$ and ctact.fail$$ to be created on the central host in /var/sadm/system/logs/hpc/nodelists. ctact.pass$$ lists the cluster nodes on which software activation was successful and ctact.fail$$ lists the nodes on which activation was unsuccessful. The $$ symbol is replaced by the process number associated with the activation.
These generated node list files can then be used for command retries or in subsequent operations using the -N switch.
This section shows an example of software activation on the local node.
CODE EXAMPLE 3-11 activates the software on the local node.
Use the ctdeact command to deactivate Sun HPC ClusterTools software on cluster nodes.
See TABLE 3-3 for a summary of the ctdeact options.
This section shows examples of software deactivation in which the ctdeact command is initiated from a central host.
To Deactivate Specified Cluster Nodes in a Non-NFS Configuration |
CODE EXAMPLE 3-12 deactivates the software on the nodes listed in /tmp/nodelist. The remote connection method is rsh.
CODE EXAMPLE 3-13 is the same as CODE EXAMPLE 3-12, except it specifies the options -k and -g.
In this example, the -k option causes the local log files of all specified nodes to be saved in /tmp/cluster-logs on the central host.
The -g option causes files ctdeact.pass$$ and ctdeact.fail$$ to be created on the central host. ctdeact.pass$$ lists the cluster nodes where software deactivation was successful. ctdeact.fail$$ lists the nodes where deactivation was unsuccessful. The $$ symbol is replaced by the process number associated with the software deactivation.
These generated node list files can then be used for command retries or in subsequent operations using the -N switch.
This section shows software deactivation on the local node.
CODE EXAMPLE 3-14 deactivates the software on the local node.
Use the ctremove command to remove Sun HPC ClusterTools software from cluster nodes.
See TABLE 3-4 for a summary of the ctremove options.
Note - If the nodes are active at the time ctremove is initiated, they will be deactivated automatically before the removal process begins. |
This section shows examples of software removal in which the ctremove command is initiated from a central host.
To Remove Software From Specified Cluster Nodes in a Non-NFS Configuration |
CODE EXAMPLE 3-15 removes the software from the nodes listed in /tmp/nodelist. The remote connection method is rsh.
CODE EXAMPLE 3-16 is the same as CODE EXAMPLE 3-15, except it specifies the options -k and -g.
CODE EXAMPLE 3-17 removes the packages SUNWompi and SUNWompimn from the nodes listed in /tmp/nodelist. The remote connection method is rsh.
This section shows software removal from the local node.
CODE EXAMPLE 3-18 deactivates the software on the local node.
CODE EXAMPLE 3-19 removes the packages SUNWompi and SUNWompimn from the local node.
Copyright © 2008 Sun Microsystems, Inc. All Rights Reserved.