Skip Headers
Oracle® Clusterware Installation Guide
11g Release 1 (11.1) for Linux

B28263-13
Go to Documentation Home
Home
Go to Book List
Book List
Go to Table of Contents
Contents
Go to Index
Index
Go to Master Index
Master Index
Go to Feedback page
Contact Us

Go to previous page
Previous
Go to next page
Next
PDF · Mobi · ePub

6 Installing Oracle Clusterware

This chapter describes the procedures for installing Oracle Clusterware for Linux. If you are installing Oracle Database with Oracle Real Application Clusters (Oracle RAC), then this is phase one of a two-phase installation.

This chapter contains the following topics:

6.1 Verifying Oracle Clusterware Requirements with CVU

Using the following command syntax, log in as the installation owner user (oracle or crs), and start Cluster Verification Utility (CVU) to check system requirements for installing Oracle Clusterware:

/mountpoint/runcluvfy.sh stage -pre crsinst -n node_list 

In the preceding syntax example, replace the variable mountpoint with the installation media mountpoint, and replace the variable node_list with the names of the nodes in your cluster, separated by commas.

For example, for a cluster with mountpoint /mnt/dvdrom/, and with nodes node1, node2, and node3, enter the following command:

$ /mnt/dvdrom/runcluvfy.sh stage -pre crsinst -n node1,node2,node3

The Cluster Verification Utility Oracle Clusterware preinstallation stage check verifies the following:

6.1.1 Interpreting CVU Messages About Oracle Clusterware Setup

If the Cluster Verification Utility report indicates that your system fails to meet the requirements for Oracle Clusterware installation, then use the topics in this section to correct the problem or problems indicated in the report, and run Cluster Verification Utility again.

User Equivalence Check Failed
Cause: Failure to establish user equivalency across all nodes. This can be due to not creating the required users, or failing to complete secure shell (SSH) configuration properly.
Action: Cluster Verification Utility provides a list of nodes on which user equivalence failed. For each node listed as a failure node, review the oracle user configuration to ensure that the user configuration is properly completed, and that SSH configuration is properly completed.

Use the command su - oracle and check user equivalence manually by running the ssh command on the local node with the date command argument using the following syntax:

$ ssh node_name date

The output from this command should be the timestamp of the remote node identified by the value that you use for node_name. If ssh is in the default location, the /usr/bin directory, then use ssh to configure user equivalence. You can also use rsh to confirm user equivalence.

If you have not attempted to use SSH to connect to the host node before running, then Cluster Verification Utility indicates a user equivalence error. If you see a message similar to the following when entering the date command with SSH, then this is the probable cause of the user equivalence error:

The authenticity of host 'node1 (140.87.152.153)' can't be established.
RSA key fingerprint is 7z:ez:e7:f6:f4:f2:4f:8f:9z:79:85:62:20:90:92:z9.
Are you sure you want to continue connecting (yes/no)?

Enter yes, and then run Cluster Verification Utility again to determine if the user equivalency error is resolved.

If ssh is in a location other than the default, /usr/bin, then Cluster Verification Utility reports a user equivalence check failure. To avoid this error, navigate to the directory $CV_HOME/cv/admin, open the file cvu_config with a text editor, and add or update the key ORACLE_SRVM_REMOTESHELL to indicate the ssh path location on your system. For example:

# Locations for ssh and scp commands
ORACLE_SRVM_REMOTESHELL=/usr/local/bin/ssh
ORACLE_SRVM_REMOTECOPY=/usr/local/bin/scp

Note the following rules for modifying the cvu_config file:

  • Key entries have the syntax name=value

  • Each key entry and the value assigned to the key defines one property only

  • Lines beginning with the number sign (#) are comment lines, and are ignored

  • Lines that do not follow the syntax name=value are ignored

When you have changed the path configuration, run Cluster Verification Utility again. If ssh is in another location than the default, you also need to start OUI with additional arguments to specify a different location for the remote shell and remote copy commands. Enter runInstaller -help to obtain information about how to use these arguments.

Note:

When you or OUI run ssh or rsh commands, including any login or other shell scripts they start, you may see errors about invalid arguments or standard input if the scripts generate any output. You should correct the cause of these errors.

To stop the errors, remove all commands from the oracle user's login scripts that generate output when you run ssh or rsh commands.

If you see messages about X11 forwarding, then complete the task "Setting Display and X11 Forwarding Configuration" to resolve this issue.

If you see errors similar to the following:

stty: standard input: Invalid argument
stty: standard input: Invalid argument

These errors are produced if hidden files on the system (for example, .bashrc or .cshrc) contain stty commands. If you see these errors, then refer to Chapter 2, "Preventing Oracle Clusterware Installation Errors Caused by stty Commands" to correct the cause of these errors.

Node Reachability Check or Node Connectivity Check Failed
Cause: One or more nodes in the cluster cannot be reached using TCP/IP protocol, through either the public or private interconnects.
Action: Use the command /bin/ping address to check each node address. When you find an address that cannot be reached, check your list of public and private addresses to make sure that you have them correctly configured. If you use third-party vendor clusterware, then refer to the vendor documentation for assistance. Ensure that the public and private network interfaces have the same interface names on each node of your cluster.
User Existence Check or User-Group Relationship Check Failed
Cause: The administrative privileges for users and groups required for installation are missing or incorrect.
Action: Use the id command on each node to confirm that the oracle user is created with the correct group membership. Ensure that you have created the required groups, and create or modify the user account on affected nodes to establish required group membership.

See Also:

"Creating Standard Configuration Operating System Groups and Users" in Chapter 3 for instructions about how to create required groups, and how to configure the oracle user

6.2 Preparing to Install Oracle Clusterware with OUI

Before you install Oracle Clusterware with Oracle Universal Installer (OUI), use the following checklist to ensure that you have all the information you will need during installation, and to ensure that you have completed all tasks that must be done before starting to install Oracle Clusterware. Mark the box for each task as you complete it, and write down the information needed, so that you can provide it during installation.

6.3 Installing Oracle Clusterware with OUI

This section provides you with information about how to use Oracle Universal Installer (OUI) to install Oracle Clusterware. It contains the following sections:

6.3.1 Running OUI to Install Oracle Clusterware

Complete the following steps to install Oracle Clusterware on your cluster. At any time during installation, if you have a question about what you are being asked to do, click the Help button on the OUI page.

  1. Unless you have the same terminal window open that you used to set up SSH, enter the following commands:

    $ exec /usr/bin/ssh-agent $SHELL
    $ /usr/bin/ssh-add
    
  2. Start the runInstaller command from the /Disk1 directory on the Oracle Database 11g release 1 (11.1) installation media. For example:

    $ cd /Disk1
    ./runInstaller
    
  3. Provide information or run scripts as root when prompted by OUI. If you need assistance during installation, click Help.

    Note:

    You must run root.sh scripts one at a time. Do not run root.sh scripts simultaneously.
  4. After you run root.sh on all the nodes, OUI runs the Oracle Notification Server Configuration Assistant, Oracle Private Interconnect Configuration Assistant, and Cluster Verification Utility. These programs run without user intervention.

When you have verified that your Oracle Clusterware installation is completed successfully, you can either use it to maintain high availability for other applications, or you can install an Oracle database.

If you intend to install Oracle Database 11g release 1 (11.1) with Oracle RAC, then refer to Oracle Real Application Clusters Installation Guide for Linux. If you intend to use Oracle Clusterware by itself, then refer to the single-instance Oracle Database installation guide.

See Also:

Oracle Real Application Clusters Administration and Deployment Guide for information about using cloning and node addition procedures, and Oracle Clusterware Administration and Deployment Guide for cloning Oracle Clusterware

6.3.2 Installing Oracle Clusterware Using a Cluster Configuration File

During installation of Oracle Clusterware, on the Specify Cluster Configuration page, you are given the option either of providing cluster configuration information manually, or of using a cluster configuration file. A cluster configuration file is a text file that you can create before starting OUI, which provides OUI with information about the cluster name and node names that it requires to configure the cluster.

Oracle suggests that you consider using a cluster configuration file if you intend to perform repeated installations on a test cluster, or if you intend to perform an installation on many nodes.

To create a cluster configuration file:

  1. On the installation media, navigate to the directory Disk1/response.

  2. Using a text editor, open the response file crs.rsp, and find the section CLUSTER_CONFIGURATION_FILE.

  3. Follow the directions in that section for creating a cluster configuration file.

6.3.3 Troubleshooting OUI Error Messages for Oracle Clusterware

The following is a list of some common Oracle Clusterware installation issues, and how to resolve them.

PRKC-1044 Failed to check remote command execution
Cause: SSH keys need to be loaded into memory, or there is a user equivalence error.
Action: Run the following commands to load SSH keys into memory:
$ exec /usr/bin/ssh-agent $SHELL
$ /usr/bin/ssh-add

Note that you must have the passphrase used to set up SSH. If you are not the person who set up SSH, then obtain the passphrase. Note also that the .ssh folder in the user home that is performing the installation must be set with 600 permissions.

In addition, confirm group membership by entering the id command, and entering ID username. For example:

$ id
$ id oracle
Incorrect permissions on partitions used for OCR or Voting Disks
Cause: The user account performing the installation (oracle or crs) does not have permission to write to these partitions
Action: Make the partitions writable by the user performing installation. For example, use the command chown user to make the selected partitions writable by the user (oracle or crs) performing the installation. During installation, these permissions are changed to root ownership.

6.4 Confirming Oracle Clusterware Function

After installation, log in as root, and use the following command syntax to confirm that your Oracle Clusterware installation is installed and running correctly:

CRS_home/bin/crs_stat -t -v

For example:

[root@node1 /]:/u01/app/crs/bin/crs_stat -t -v
Name       a Type           R/RA   F/FT    Target    State       Host
crs....ac3.gsd application  0/5    0/0     Online    Online      node1
crs....ac3.ons application  0/5    0/0     Online    Online      node1
crs....ac3.vip application  0/5    0/0     Online    Online      node1
crs....ac3.gsd application  0/5    0/0     Online    Online      node2
crs....ac3.ons application  0/5    0/0     Online    Online      node2
crs....ac3.vip application  0/5    0/0     Online    Online      node2

You can also use the command crsctl check crs for a less detailed system check. for example:

[root@node1 bin] $ ./crsctl check crs
Cluster Synchronization Services appears healthy
Cluster Ready Services appears healthy
Event Manager appears healthy

Caution:

After installation is complete, do not remove manually or run cron jobs that remove /tmp/.oracle or /var/tmp/.oracle or its files while Oracle Clusterware is up. If you remove these files, then Oracle Clusterware could encounter intermittent hangs, and you will encounter error CRS-0184: Cannot communicate with the CRS daemon.