Perform the following procedures to install the packages and patches for all software products and set up the user environment.
You install the Oracle software later in this manual.
Have the following available:
The Java Availability Suite DVD-ROM that contains Sun Cluster 3.2 software.
Access to the SunSolveSM web site to download patches.
Become superuser on the administrative console.
Configure the preinstalled Solaris 10 11/06 software, if you have not already done so.
For more information, see the Sun Ultra 20 Workstation Getting Started Guide (819–2148).
Download, install, and configure Sun Update Connection.
See http://www.sun.com/service/sunupdate/gettingstarted.html for details. Documentation for Sun Update Connection is available at http://docs.sun.com/app/docs/coll/1320.2.
Download and apply any Solaris 10 patches by using Sun Update Connection.
Load the Java Availability Suite DVD-ROM into the DVD-ROM drive.
Change to the Solaris_sparc/Product/sun_cluster/Solaris_10/Packages/ directory.
Install software packages for the Cluster Control Panel and man pages.
admincon# pkgadd -d . SUNWccon SUNWscman |
Change to a directory that does not reside on the DVD-ROM and eject the DVD-ROM.
host# cd / host# eject cdrom |
Create an /etc/cluster file that contains the cluster name and the two node names.
admincon# vi /etc/clusters sccluster phys-sun phys-moon |
Create an /etc/serialports file that contains both node names and the hostname and port number that each node uses to connect to the management network.
admincon# vi /etc/serialports phys-sun phys-sun 46 phys-moon phys-moon 47 |
Add the Sun Cluster PATH and MANPATH to the .cshrc user initialization file.
To the PATH entry, add /opt/SUNWcluster/bin.
To the MANPATH entry, add /opt/SUNWcluster/man and /usr/cluster/man.
Initialize your modifications.
admincon# cd admincon# source .cshrc |
This procedure describes how to install the Solaris 10 OS to meet Sun Cluster software installation requirements.
If your system comes with the Solaris OS preinstalled but does not meet Sun Cluster software installation requirements, perform this procedure to reinstall Solaris software to meet installation requirements.
Have the following available:
The Solaris 10 11/06 Software for SPARC platforms DVD-ROM.
Access to the SunSolve web site to download patches.
Add all public hostnames and logical addresses for the cluster to the naming service.
The IP addresses in this step are for example only and are not valid for use on the public network. Substitute your own IP addresses when you perform this step.
192.168.10.1 phys-sun 192.168.10.2 phys-moon 192.168.10.3 apache-lh 192.168.10.4 nfs-lh 192.168.10.5 oracle-lh 192.168.10.6 admincon 192.168.11.1 phys-sun-11 192.168.11.2 phys-moon-11 192.168.11.3 se3510fc 192.168.11.4 admincon-11 |
For more information about naming services, see System Administration Guide: Naming and Directory Services (DNS, NIS, and LDAP).
From the administrative console, start the cconsole(1M) utility.
admincon# cconsole & |
Use the cconsole utility to communicate with each individual cluster node or use the master window to send commands to both nodes simultaneously.
Insert the Solaris 10 11/06 DVD-ROM in the DVD-ROM drive of phys-sun.
Access the console window for phys-sun.
Boot phys-sun.
If the system is new, turn on the system.
If the system is currently running, shut down the system.
phys-sun# init 0 |
The ok prompt is displayed.
ok setenv auto-boot? false |
Disabling automatic reboot prevents continuous boot cycling.
Create an alias for each disk.
The assignment of aliases to the disks enables you to access and boot from the second disk if you cannot boot from the default disk.
Display the disks and choose the boot disk.
ok show-disks … Enter selection, q to quit: X |
Assign the alias name rootdisk to the disk that you chose.
ok nvalias rootdisk Control-Y |
The Control-Y keystroke combination enters the disk name that you chose from the show-disks menu.
Save the disk alias.
ok nvstore |
Repeat the preceding steps to identify and assign the alias name backup_root to the alternate boot disk.
Set the boot-device environment variable to the aliases for the default boot disk and backup boot disk.
ok setenv boot-device rootdisk backup_root |
For more information, see OpenBoot 4.x Command Reference Manual.
Start the Solaris installation program.
ok boot cdrom |
Follow the prompts.
Make the following installation choices:
Prompt |
Value |
---|---|
Solaris Software Group |
Entire Plus OEM Support |
Partitions |
Manual formatting |
Root password |
Same password on both nodes |
Automatic reboot |
No |
Enable network services for remote clients |
Yes |
Set the following partition sizes and file-system names, if not already set:
Slice |
Size |
File System Name |
---|---|---|
0 |
remaining free space |
/ |
1 |
2 Gbyte |
swap |
4 |
512 Mbyte |
/globaldevices |
5 |
2 Gbyte |
/var |
7 |
32 Mbyte |
for Solaris Volume Manager use |
Return to Step 3 and repeat these steps on phys-moon.
On both nodes, download, install, and configure Sun Update Connection.
See http://www.sun.com/service/sunupdate/gettingstarted.html for details. Documentation for Sun Update Connection is available at http://docs.sun.com/app/docs/coll/1320.2.
On both nodes, download and apply any Solaris 10 patches by using Sun Update Connection.
Perform this procedure on both nodes. The steps in this procedure use the C shell environment. If you are using a different shell, perform the equivalent tasks for your preferred shell environment.
For more information, see Customizing a User’s Work Environment in System Administration Guide: Basic Administration.
Open the cconsole master console window, if it is not already open.
Use the master console window to perform the steps in this procedure on both nodes at the same time.
Display the settings for the umask and the environment variables.
phys-X# umask phys-X# env | more |
If not already set, set the umask to 22.
This entry sets the default permissions for newly created files.
umask 022 |
Ensure that the PATH includes the following paths.
/usr/bin
/usr/cluster/bin
/usr/sbin
/oracle/oracle/product/10.2.0/bin
(Optional) Add the following paths to the MANPATH.
/usr/cluster/man
/usr/apache/man
Set the ORACLE_BASE and ORACLE_SID environment variables.
ORACLE_BASE=/oracle ORACLE_SID=orasrvr |
Verify the setting changes that you made.
phys-X# umask phys-X# env | more |
This procedure describes how to modify certain system settings to support the Quick Start configuration.
On both nodes, enable Solaris multipathing functionality.
phys-X# /usr/sbin/stmsboot -e |
Enables Solaris I/O multipathing
For more information, see the stmsboot(1M) man page.
On both nodes, update the /etc/inet/ipnodes file with all public hostnames and logical addresses for the cluster.
Except for the loghost entries, these entries are the same on both nodes.
The IP addresses in this step are for example only and are not valid for use on the public network. Substitute your own IP addresses when you perform this step.
phys-X# vi /etc/inet/ipnodes |
On phys-sun, add the following entries:
127.0.0.1 localhost 192.168.10.1 phys-sun loghost 192.168.10.2 phys-moon 192.168.10.3 apache-lh 192.168.10.4 nfs-lh 192.168.10.5 oracle-lh 192.168.10.6 admincon 192.168.11.1 phys-sun-11 192.168.11.2 phys-moon-11 192.168.11.3 se3510fc-11 192.168.11.4 admincon-11 |
On phys-moon, add the following entries:
127.0.0.1 localhost 192.168.10.1 phys-sun 192.168.10.2 phys-moon loghost 192.168.10.3 apache-lh 192.168.10.4 nfs-lh 192.168.10.5 oracle-lh 192.168.10.6 admincon 192.168.11.1 phys-sun-11 192.168.11.2 phys-moon-11 192.168.11.3 se3510fc-11 192.168.11.4 admincon-11 |
On both nodes, ensure that the following kernel parameters are set to at least the minimum values that Oracle requires.
Display the settings for the default project.
phys-X# prctl -i project default |
If no kernel parameters are set, or if any kernel parameters are not set to the minimum required value for Oracle as shown in the following table, set the parameter.
phys-X# projmod -s -K "parameter=(priv,value,deny)" default |
Oracle Kernel Parameter |
Minimum Required Value |
---|---|
process.max-sem-nsems |
256 |
project.max-sem-ids |
100 |
project.max-shm-ids |
100 |
project.max-shm-memory |
4294967295 |
Verify the new settings.
phys-X# prctl -i project default |
These settings are the minimum required values to support the Oracle software in a Sun Cluster Quick Start configuration. For more information about these parameters, see the Oracle10g Installation Guide.
On both nodes, add the following entries to the /etc/system file.
phys-X# vi /etc/system set ce:ce_taskq_disable=1 exclude:lofs |
The first entry supports ce adapters for the private interconnect.
The second entry disables the loopback file system (LOFS), which must be disabled when Sun Cluster HA for NFS is configured on a highly available local file system. For more information and alternatives to disabling LOFS when Sun Cluster HA for NFS is configured, see the information about loopback file systems in Solaris OS Feature Restrictions in Sun Cluster Software Installation Guide for Solaris OS.
These changes take effect at the next system reboot.
On both nodes, update the /devices and /dev entries.
phys-X# devfsadm -C |
On both nodes, confirm that the storage array is visible.
phys-X# luxadm probe |
This procedure assumes that the specified disks are available for creation of database replicas. Substitute your own disk names in this procedure.
On both nodes, create state database replicas.
Create three replicas on each of the two internal disks.
phys-X# metadb -af -c 3 c0t0d0s7 phys-X# metadb -a -c 3 c0t1d0s7 |
On both nodes, verify the replicas.
phys-X# metadb flags first blk block count a u 16 8192 /dev/dsk/c0t0d0s7 a u 8208 8192 /dev/dsk/c0t0d0s7 a u 16400 8192 /dev/dsk/c0t0d0s7 a u 16 8192 /dev/dsk/c0t1d0s7 a u 8208 8192 /dev/dsk/c0t1d0s7 a u 16400 8192 /dev/dsk/c0t1d0s7 |
Perform this procedure on one node at a time.
This procedure assumes that the cluster node contains the internal nonshared disks c0t0d0 and c0t1d0. Substitute your own internal disk names if necessary in the steps of this procedure.
On phys-sun, place the root slice c0t0d0s0 in a single-slice (one-way) concatenation.
phys-sun# metainit -f d10 1 1 c0t0d0s0 |
Create a second concatenation with the other internal disk, c0t1d0s0.
phys-sun# metainit d20 1 1 c0t1d0s0 |
Create a one-way mirror with one submirror.
phys-sun# metainit d0 -m d10 |
Set up the system files for the root directory.
phys-sun# metaroot d0 |
The metaroot command edits the /etc/vfstab and /etc/system files so that the system can be booted with the root (/) file system on a metadevice or volume. For more information, see the metaroot(1M) man page.
Flush all file systems.
phys-sun# lockfs -fa |
The lockfs command flushes all transactions from the log and writes the transactions to the master file system on all mounted UFS file systems. For more information, see the lockfs(1M) man page.
Reboot the node to remount the newly mirrored root (/) file system.
phys-sun# init 6 |
Attach the second submirror to the mirror.
phys-sun# metattach d0 d20 |
For more information, see the metattach(1M) man page.
Record the alternate boot path for possible future use.
If the primary boot device fails, you can then boot from this alternate boot device. For more information about alternate boot devices, see Creating a RAID-1 Volume in Solaris Volume Manager Administration Guide.
phys-sun# ls -l /dev/rdsk/c0t1d0s0 |
This procedure installs software packages for the Sun Cluster framework and for the Sun Cluster HA for Apache, Sun Cluster HA for NFS, and Sun Cluster HA for Oracle data services.
Have available the following:
The Java Availability Suite DVD-ROM that contains Sun Cluster 3.2 software.
Access to the SunSolve web site to download patches.
On phys-sun, load the Java Availability Suite DVD-ROM in the DVD-ROM drive.
Start the Java Enterprise System (ES) installer program.
phys-sun# ./installer |
For more information about using the Java ES installer program, see the Sun Java Enterprise System 5 Installation Guide for UNIX.
Follow the onscreen instructions to install the Sun Cluster framework packages.
After the installation is finished, the installer program provides an installation summary. This summary enables you to view logs that the program created during the installation. These logs are located in the /var/sadm/install/logs/ directory.
Change to a directory that does not reside on the DVD-ROM and eject the DVD-ROM.
host# cd / host# eject cdrom |
Return to Step 1 and repeat all steps on phys-moon.
On both nodes, use Sun Update Connection to download and apply any needed patches.
Perform the steps in this procedure on both nodes.
Open the cconsole master console window, if it is not already open.
Use the master console window to perform the steps in this procedure on both nodes at the same time.
Create the Oracle Inventory group, oinstall, and the database administrator group, dba.
phys-X# groupadd oinstall phys-X# groupadd dba |
Create the Oracle user account, oracle.
Specify the Oracle home directory, /oracle/oracle/product/10.2.0. Set dba as the primary group and set oinstall as the secondary group.
phys-X# useradd -g dba -G oinstall -d /oracle/oracle/product/10.2.0 oracle |
phys-X# passwd -r files oracle |