Sun Cluster 2.2 Software Installation Guide

Chapter 2 Planning the Configuration

This chapter provides information and procedures for planning your Sun Cluster configuration, and includes the following sections.

Configuration Planning Overview

Configuration planning includes making decisions about:

Before you develop your configuration plan, consider the reliability issues described in "Configuration Rules for Improved Reliability". Also, the Sun Cluster environment imposes some configuration restrictions that you should consider before completing your configuration plan. These are described in "Configuration Restrictions".

"Configuration Worksheets", provides worksheets to help you plan your configuration.

Configuration Planning Tasks

The following sections describe the tasks and issues associated with planning your configuration. You are not required to perform the tasks in the order shown here, but you should address each task as part of your configuration plan.

Planning the Administrative Workstation

You must decide whether to use a dedicated SPARCTM workstation, known as the administrative workstation, for administering the active cluster. The administrative workstation is not a cluster node. The administrative workstation can be any SPARC machine capable of running a telnet session to the Terminal Concentrator to facilitate console logins. Alternatively, on E10000 platforms, you must have the ability from the administrative workstation to log into the System Service Processor (SSP) and connect using the netcon command.

Sun Cluster does not require a dedicated administrative workstation, but using one provides you these advantages:

Establishing Names and Naming Conventions

Before configuring the cluster, you must decide on names for the following:

The network interface names (and associated IP addresses) are necessary for each logical host on each public network. Although you are not required to use a particular naming convention, the following naming conventions are used throughout the documentation and are recommended. Use the configuration worksheets included in "Configuration Worksheets".

Cluster - As part of the configuration process, you will be prompted for the name of the cluster. You can choose any name; there are no restrictions imposed by Sun Cluster.

Physical Hosts - Physical host names are created by adding the prefix phys- to the logical host names (for physical hosts that master only one logical host each). For example, the physical host that masters a logical host named hahost1 would be named phys-hahost1 by default. There is no Sun Cluster naming convention or default for physical hosts that master more than one logical host.


Caution - Caution -

If you are using DNS as your name service, do not use an underscore in your physical or logical host names. DNS will not recognize a host name containing an underscore.


Logical Hosts and Disk Groups - Logical host names can be different from disk group names in Sun Cluster. However, using the same names is the Sun Cluster convention and eases administration. Refer to "Planning Your Logical Host Configuration", for more information.

Public Network - The names by which physical hosts are known on the public network are their primary physical host names. The names by which physical hosts are known on a secondary public network are their secondary physical host names. Assign these names using the following conventions, as illustrated in Figure 2-1:

Private Interconnect - There is no default naming convention for the private interconnect.

Naming convention examples are illustrated in Figure 2-1.

Figure 2-1 Public and Private Network Naming Conventions

Graphic

Planning Network Connections

You must have at least one public network connection to a local area network and exactly two private interconnects between the cluster nodes. Refer to Chapter 1, Understanding the Sun Cluster Environment, for overviews of Sun Cluster network configurations, and to "Configuration Worksheets", for network planning worksheets.

Public Network Connections

Consider these points when planning your public network configuration:

Private Network Connections

Sun Cluster requires two private networks for normal operation. You must decide whether to use 100 Mbit/sec Ethernet or 1 Gbit/sec Scalable Coherent Interface (SCI) connections for the private networks.

In two-node configurations, these networks may be implemented with point-to-point cables between the cluster nodes. In three- or four-node configurations, they are implemented using hubs or switches. Only private traffic between Sun Cluster nodes is transported on these networks.

If you connect nodes by using SCI switches, each node must be connected to the same port number on both switches. During the installation, note that the port numbers on the switches correspond to node numbers. For example, node 0 is the host physically connect to port 0 on the switch, and so on.

A class C network number (204.152.64) is reserved for private network use by the Sun Cluster nodes. The same network number is used by all Sun Cluster systems.

Terminal Concentrator and Administrative Workstation Network Connections

The Terminal Concentrator and administrative workstation connect to a public network with access to the Sun Cluster nodes. You must assign IP addresses and host names for them to enable access to the cluster nodes over the public network.


Note -

E10000 systems use a System Service Processor (SSP) instead of a Terminal Concentrator. You will need to assign the SSP a host name, IP address, and root password. You will also need to create a user named "ssp" on the SSP and provide a password for user "ssp" during Sun Cluster installation.


Planning Your Solaris Operating Environment Installation

All nodes in a cluster must be installed with the same version of the Solaris operating environment (Solaris 2.6, Solaris 7, or Solaris 8) before you can install the Sun Cluster software. When you install Solaris on cluster nodes, follow the general rules in this section.

Disabling Solaris Interface Groups

A new feature called interface groups was added to the Solaris 2.6 operating environment. This feature is implemented as default behavior in Solaris 2.6 software, but as optional behavior in subsequent releases.

As described in the ifconfig(1M) man page, logical or physical interfaces that share an IP prefix are collected into an interface group. IP uses the interface group to rotate source address selections when the source address is unspecified. An interface group made up of multiple physical interfaces is used to distribute traffic across different IP addresses on a per-IP-destination basis (see netstat(1M) for information about per-IP-destination).

When enabled, the interface groups feature causes a problem with switchover of logical hosts. The system will experience RPC timeouts and the switchover will fail, causing the logical host to remain mastered on its current host. Therefore, interface groups must be disabled on all cluster nodes. The status of interface groups is determined by the value of ip_enable_group_ifs in /etc/system.

The value for this parameter can be checked with the following ndd command:


# ndd /dev/ip ip_enable_group_ifs

If the value returned is 1 (enabled), disable interface groups by editing the /etc/system file to include the following line


set ip:ip_enable_group_ifs=0
:


Caution - Caution -

Whenever you modify the /etc/system file, you must reboot the system for the changes to take effect.


Partitioning System Disks

When Solaris is installed, the system disk is partitioned into slices for root (/), /usr, and other standard file systems. You must change the partition configuration to meet the requirements your volume manager. Use the guidelines in the following sections to allocate disk space accordingly.

File System Slices

See your Solaris documentation for file system sizing guidelines. Sun Cluster imposes no additional requirements for file system slices.

Volume Manager Slices

If you will be using Solstice DiskSuite, set aside a 10 Mbyte slice on the system disk for metadevice state database replicas. See your Solstice DiskSuite documentation for more information about replicas.

If you will be using VxVM, designate a disk for the root disk group (rootdg). See your VERITAS documentation for guidelines and details about creating the rootdg. Refer also to "VERITAS Volume Manager Considerations", for more information.

The Root (/) Slice

The root (/) slice on your local disk must have enough space for the various files and directories as well as space for the device inodes in /devices and symbolic links in /dev.

The root slice also must be large enough to hold the following:

Your cluster might require a larger root file system if it contains large numbers of disk drives.


Note -

If you run out of free space, you must reinstall the operating environment on all cluster nodes to obtain additional free space in the root slice. Make sure at least 20 percent of the total space on the root slice is left free.


The /usr, /var, and /opt Slices

The /usr slice holds the user file system. The /var slice holds the system log files. The /opt slice holds the Sun Cluster and data service software packages. See your Solaris advanced system administration documentation for details about changing the allocation values when installing Solaris software.

Volume Management

Sun Cluster uses volume management software to group disks into disk groups that can then be administered as one unit. Sun Cluster supports Solstice DiskSuite and VERITAS Volume Manager (VxVM). Sun Cluster supports the cluster feature of VERITAS Volume Manager for use with the Oracle Parallel Server data service.

You can use Solstice DiskSuite and VERITAS Volume Manager together only if you use Solstice DiskSuite to manage the local disks and VERITAS Volume Manager to control the multihost disks. In such a configuration, plan your physical disk needs accordingly. You might need additional disks to use for the VERITAS Volume Manager root disk group, for example. See your volume manager documentation for more information.

You must install the volume management software after you install the Solaris operating environment. You can install the volume management software either before or after you install Sun Cluster software. Refer to your volume manager software documentation and to Chapter 3, Installing and Configuring Sun Cluster Software, for instructions on installing the volume management software.

Use these guidelines when configuring your disks:

See "Volume Manager Slices" for disk layout recommendations related to volume management, and consult your volume manager documentation for any additional restrictions.

Solstice DiskSuite Considerations

Consider these points when planning Solstice DiskSuite configurations:

VERITAS Volume Manager Considerations

Consider these points when planning VxVM configurations:

File System Logging

One important aspect of high availability is the ability to bring file systems back online quickly in the event of a node failure. This aspect is best served by using a logging file system. Sun Cluster supports three logging file systems: VxFS logging from VERITAS, DiskSuite UFS logging, and Solaris UFS logging. VERITAS Volume Manager cluster feature, when used with Oracle Parallel Server, uses raw partitions so does not use a logging file system. However, you can also run VxVM with cluster feature in a cluster with both OPS and HA data services. In this configuration, the OPS shared disk groups would use raw partitions, but the HA disk groups could use either VxFS or Solaris UFS logging file systems. Excluding the co-existent VxVM with cluster feature configuration described above, Sun Cluster supports the following combinations of volume managers and logging file systems.

Table 2-1 Supported Logging File Systems

Solaris Operating Environment 

Volume Manager 

Supported File Systems 

Solaris 2.6 

VERITAS Volume Manager 

VxFS, UFS (no logging) 

Solstice DiskSuite 

DiskSuite UFS logging 

Solaris 7 

Solstice DiskSuite 

DiskSuite UFS logging, Solaris UFS logging 

Solaris 8 

Solstice DiskSuite 

DiskSuite UFS logging, Solaris UFS logging 

VxVM with cluster feature uses Dirty Region Logging to aid in fast recovery after a reboot, similar to what the logging file systems provide. For information about the VxVM cluster feature, refer to your VERITAS Volume Manager documentation. For information on DiskSuite UFS logging, refer to the Solstice DiskSuite documentation. For information on VxFS logging, see the VERITAS documentation. Solaris UFS logging is described briefly below. See the mount_ufs(1M) for more details.

Solaris UFS logging is a new feature in the Solaris 7 operating environment.

Solaris UFS logging uses a circular log to journal the changes made to a UFS file system. As the log fills up, changes are "rolled" into the actual file system. The advantage of logging is that the UFS file system is never left in an inconsistent state, that is, with a half-completed operation. After a system crash, fsck has nothing to fix, so you boot up much faster.

Solaris UFS logging is enabled using the "logging" mount option. To enable logging on a UFS file system, you either add -o logging to the mount command or add the word "logging" to the /etc/opt/SUNWcluster/conf/hanfs/vfstab.logicalhost entry (the rightmost column).

Solaris UFS logging always allocates the log using free space on the UFS file system. The log takes up 1 MByte on file systems less than 1 GByte in size, and 1 MByte per GByte on larger file systems, up to a maximum of 64 MBytes.

Solaris UFS logging always puts the log files on the same disk as the file system. If you use this logging option, you are limited to the size of the disk. DiskSuite UFS logging allows the log to be separated on a different disk. This has the effect of reducing a bit of the I/O that is associated with the log.

With DiskSuite UFS logging, the trans device used for logging creates a metadevice. The log is yet another metadevice which can be mirrored and striped. Furthermore, you can create up to a 1TByte logging file system with Solstice DiskSuite.

The "logging" mount option will not work if you already have logging provided by Solstice DiskSuite--you will receive a warning message explaining you already have logging on that file system. If you require more control over the size or location of the log, you should use DiskSuite UFS logging.

Depending on the file system usage, Solaris UFS logging gives you performance that is the same or better than running without logging.

There is currently no support for converting from DiskSuite UFS logging to Solaris UFS logging.

Determining Your Multihost Disk Requirements

Unless you are using a RAID5 configuration, all multihost disks must be mirrored in Sun Cluster configurations. This enables the configuration to tolerate single-disk failures. Refer to "Mirroring Guidelines", and to your volume management documentation, for more information.

Determine the amount of data that you want to move to the Sun Cluster configuration. If you are not using RAID5, double that amount to allow disk space for mirroring. With RAID5, you need extra space equal to 1/(# of devices -1). Use the worksheets in "Configuration Worksheets", to help plan your disk requirements.

Consider these points when planning your disk requirements:

Disk Space Growth

Consider these points when planning for disk space growth:

Size and Number of Disk Drives

Several sizes of disks are supported in multihost disk expansion units. Consider these points when deciding which size drives to use:

Planning Your File System Layout on the Multihost Disks

Sun Cluster does not require any specific disk layout or file system size. The requirements for the file system hierarchy are dependent on the volume management software you are using.

Regardless of your volume management software, Sun Cluster requires at least one file system per disk group to serve as the HA administrative file system. This administrative file system should always be mounted on /logicalhost, and must be a minimum of 10 Mbytes. It is used to store private directories containing data service configuration information.

For clusters using Solstice DiskSuite, you need to create a metadevice to contain the HA administrative file system. The HA administrative file system should be configured the same as your other multihost file systems, that is, it should be mirrored and set up as a trans device.

For clusters using VxVM, Sun Cluster creates the HA administrative file system on a volume named dg-stat where dg is the name of the disk group in which the volume is created. dg is usually the first disk group in the list of disk groups specified when defining a logical host.

Consider these points when planning file system size and disk layout:

File Systems With Solstice DiskSuite

Solstice DiskSuite software requires some additional space on the multihost disks and imposes some restrictions on its use. For example, if you are using UNIX file system (UFS) logging under Solstice DiskSuite, one to two percent of each multihost disk must be reserved for metadevice state database replicas and UFS logging. Refer to Appendix B, Configuring Solstice DiskSuite, and to the Solstice DiskSuite documentation for specific guidelines and restrictions.

All metadevices used by each shared diskset are created in advance, at reconfiguration boot time, based on settings found in the md.conf file. The fields in md.conf file are described in the Solstice DiskSuite documentation. The two fields that are used in the Sun Cluster configuration are md_nsets and nmd. The md_nsets field defines the number of disksets and the mnd field defines the number of metadevices to create for each diskset. You should set these fields at install time to allow for all predicted future expansion of the cluster.

Extending the Solstice DiskSuite configuration after the cluster is in production is time consuming because it requires a reconfiguration reboot for each node and always carries the risk that there will not be enough space allocated in the root (/) file system to create all of the requested devices.

The value of md_nsets must be set to the expected number of logical hosts in the cluster, plus one to allow Solstice DiskSuite to manage the private disks on the local host (that is, those metadevices that are not in the local diskset).

The value of nmd must be set to the predicted largest number of metadevices used by any one of the disksets in the cluster. For example, if a cluster uses 10 metadevices in its first 15 disksets, but 1000 metadevices in the 16th diskset, nmd must be set to at least 1000.


Caution - Caution -

All cluster nodes (or cluster pairs in the cluster pair topology) must have identical md.conf files, regardless of the number of logical hosts served by each node. Failure to follow this guideline can result in serious Solstice DiskSuite errors and possible loss of data.


Consider these points when planning your Solstice DiskSuite file system layout:

Table 2-2 Multihost Disk Partitioning for Most Drives

Slice 

Description 

2 Mbytes, reserved for Solstice DiskSuite 

UFS logs 

Remainder of the disk 

Overlaps Slice 6 and 0 

Table 2-3 Multihost Disk Partitioning for the First Drive on the First Two Controllers

Slice 

Description 

2 Mbytes, reserved for Solstice DiskSuite 

2 Mbytes, UFS log for HA administrative file systems 

9 Mbytes, UFS master for HA administrative file systems 

UFS logs 

Remainder of the disk 

Overlaps Slice 6 and 0 

File Systems With VERITAS Volume Manager

You can create UNIX File System (UFS) or VERITAS File System (VxFS) file systems in the disk groups of logical hosts. When a logical host is mastered on a cluster node, the file systems associated with the disk groups of the logical host are mounted on the specified mount points of the mastering node.

When you reconfigure logical hosts, Sun Cluster must check the file systems before mounting them, by running the fsck command. Even though the fsck command checks the UFS file systems in non-interactive parallel mode on UFS file systems, this still consumes some time, and this affects the reconfiguration process. VxFS drastically cuts down on the file system check time, especially if the configuration contains large file systems (greater than 500 Mbytes) used for data services.

When setting up mirrored volumes, always add a Dirty Region Log (DRL) to decrease volume recovery time in the event of a system crash. When mirrored volumes are used in clusters, DRL must be assigned for volumes greater than 500 Mbytes.

With VxVM, it is important to estimate the maximum number of volumes that will be used by any given disk group at the time the disk group is created. If the number is less than 1000, default minor numbering can be used. Otherwise, you must carefully plan the way in which minor numbers are assigned to disk group volumes. It is important that no two disk groups shared by the same nodes have overlapping minor number assignments.

As long as default numbering can be used and all disk groups are currently imported, it is not necessary to use the minor option to the vxdg init command at disk group creation time. Otherwise, the minor option must be used to prevent overlapping the volume minor number assignments. It is possible to modify the minor numbering later, but doing so might require you to reboot and import the disk group again. Refer to the vxdg(1M) man page for details.

Mount Information

The /etc/vfstab file contains the mount points of file systems residing on local devices. For a multihost file system used for a logical host, all the nodes that can potentially master the logical host should possess the mount information.

The mount information for a logical host's file system is kept in a separate file on each node, named /etc/opt/SUNWcluster/conf/hanfs/vfstab.logicalhost. The format of this file is identical to the /etc/vfstab file for ease of maintenance, though not all fields are used.


Note -

You must keep the vfstab.logicalhost file consistent among all nodes of the cluster. Use the rcp command or file transfer protocol (FTP) to copy the file to the other nodes of the cluster. Alternatively, edit all of the files simultaneously by using crlogin or ctelnet.


The same file system cannot be mounted by more than one node at the same time, because a file system can be mounted only if the corresponding disk group has been imported by the node. The consistency and uniqueness of the disk group imports and logical host mastery is enforced by the cluster framework logical host reconfiguration sequence.

Booting From a SPARCstorage Array

Sun Cluster supports booting from a private or shared disk inside a SPARCstorage Array.

Consider these points when using a boot disk in an SSA:

Planning Your Logical Host Configuration

A disk group stores the data for one or more data services. Generally, several data services share a logical host, and therefore fail over together. If you want to enable a particular data service to fail over independently of all other data services, then assign a logical host to that data service alone, and do not allow any other data services to share it.

As part of the installation and configuration, you need to establish the following for each logical host:

Use the logical host worksheet in "Configuration Worksheets", to record this information.

Consider these points when planning your logical host configuration:

Planning the Cluster Configuration Database Volume

As part of Sun Cluster installation and configuration, you configure a Cluster Configuration Database (CCD) volume to store internal configuration data. In a two-node cluster using VxVM, this volume can be shared between the nodes thereby increasing the availability of the CCD. In clusters with more than two nodes, a copy of the CCD is local to each node. See "Configuring the Shared CCD Volume" for details on configuring a shared CCD.


Note -

You cannot used a shared CCD in a two-node cluster using Solstice DiskSuite.


If each node keeps its own copy of the CCD, then updates to the CCD are disabled by default when one node is not part of the cluster. This prevents the database from getting out of synchronization when only a single node is up.

The CCD requires two disks as part of a disk group for a shared volume. These disks are dedicated for CCD use and cannot be used by any other application, file system, or database.

The CCD should be mirrored for maximum availability. The two disks comprising the CCD should be on separate controllers.

In clusters using VxVM, the scinstall(1M) script will ask you how you want to set up the CCD on a shared volume in your configuration.

Refer to Chapter 1, Understanding the Sun Cluster Environment, for a general overview of the CCD. Refer to the chapter on general cluster administration in the Sun Cluster 2.2 System Administration Guide for procedures used to administer the CCD.


Note -

Although the installation procedure does not prevent you from choosing disks on the same controller, this would introduce a possible single point of failure and is not recommended.


Planning the Quorum Device (VxVM)

If you are using VERITAS Volume Manager (with or without the cluster feature), you must configure a quorum device regardless of the number of cluster nodes. During the Sun Cluster installation process, scinstall(1M) will prompt you to configure a quorum device.

The quorum device is either an array controller or a disk.

During the cluster software installation, you will need to make decisions concerning:

Cluster Topology Considerations

Before you select the quorum device for your cluster, be aware of the implications of your selection. Any node pair of the cluster must have a quorum device. That is, one quorum device must be specified for every node set that share multihost disks. Each node in the cluster must be informed of all quorum devices in the cluster, not just the quorum device connected to it. During cluster installation, the scinstall(1M) command displays all possible node pairs in sequence and displays any common devices that are quorum device candidates.

In two-node clusters with dual-ported disks, a single quorum device needs to specified.

In greater than two-node clusters with dual-ported disks, not all of the cluster nodes have access to the entire disk subsystem. In such configurations, you must specify one quorum device for each set of nodes that shares disks.

Sun Cluster configurations can consist of disk storage units (such as the Sun StorEdge A5000) that can be connected to all nodes in the cluster. This allows for applications such as OPS to run on clusters of greater than two nodes. A disk storage unit that is physically connected to all nodes in the cluster is referred to as a direct attached device. In this type of cluster a single quorum device needs to be selected from a direct attached device.

In clusters with direct attached devices, if the cluster interconnect fails, one of the following will happen:

In clusters without direct attached devices to all nodes of the cluster, you will, by definition, have multiple quorum devices (one for each node pair that share disks). In this configuration, the quorum device only comes into play where only two nodes are remaining and they share a common quorum device.

In the event of a node failure, the node that is able to reserve the quorum device remains as the sole survivor of the cluster. This is necessary to ensure the integrity of data on the shared disks.

Planning a Data Migration Strategy

Consider these points when deciding how to migrate existing data to the Sun Cluster environment.

Selecting a Multihost Backup Strategy

Before you load data onto the multihost disks in a Sun Cluster configuration, you should have a plan for backing up the data. Sun recommends using Solstice Backup(TM), ufsdump, or VERITAS NetBackup to back up your Sun Cluster configuration.

If you are converting your backup method from Online: Backup(TM) to Solstice Backup, special considerations exist because the two products are not compatible. The primary decision for the system administrator is whether or not the files backed up with Online:Backup will be available online after Solstice Backup is in use. Refer to the Solstice Backup documentation for details on conversion.

VERITAS NetBackup provides a powerful, scalable backup solution. The NetBackup master server can be made highly available by placing it under the control of Sun Cluster HA for NetBackup. Refer to Chapter 14, Installing and Configuring Sun Cluster HA for NetBackup for more information about VERITAS NetBackup.

Planning for Problem Resolution

The following files should be saved after the system is configured and running. In the unlikely event that the cluster should experience problems, these files can help service providers debug and solve cluster problems.

Selecting a Solaris Install Method

You can install Solaris software from a local CD-ROM or from a network install server using JumpStart. If you are installing several Solaris machines, consider a network install. Otherwise, use the local CD-ROM.


Note -

Configurations using FDDI as the primary public network cannot be network-installed directly using JumpStart because the FDDI drivers are unbundled and are not available in "mini-unix." If you use FDDI as the primary public network, you must install Solaris software from CD-ROM.


Licensing

You will receive a paper license for the Sun Cluster 2.2 framework, one for each hardware platform on which Sun Cluster 2.2 will run. You will also receive a paper license for each Sun Cluster data service, one per node. The Sun Cluster 2.2 framework does not enforce these licenses, but you should retain the paper licenses as proof of ownership when you need technical support or other support services.

You do not need licenses to run Solstice DiskSuite with a licensed Sun Cluster 2.2 configuration. However, you need a license for VERITAS Volume Manager (VxVM) and optionally for VERITAS Volume Manager cluster functionality. The base VxVM license certificates are included with Sun Cluster Server license kits, and VxVM cluster functionality license certificates are bundled with Oracle Parallel Server RTU license kits. The Sun Cluster and Oracle Parallel Server license kits are available from Sun. Follow the instructions printed on the license certificates to obtain active license keys.

You may need to obtain licenses for DBMS products and other third-party products. Contact your third-party service provider for third-party product licenses. See http://www.sun.com/licensing/ for more information.

Configuration Rules for Improved Reliability

The rules discussed in this section help ensure that your Sun Cluster configuration is highly available. These rules also help determine the appropriate hardware for your configuration.

Configuring redundant hardware is not always possible--some configurations might contain only one system board--but some of the concerns can still be addressed easily with hardware options. For example, in an Ultra Enterprise(TM) 2 Cluster with two SPARCstorage Arrays, one private network can be connected to a Sun Quad FastEthernet(TM) Controller card (SQEC), while the other private network can be connected to the on-board interface.

Mirroring Guidelines

Unless you are using a RAID5 configuration, all multihost disks must be mirrored in Sun Cluster configurations. This enables the configuration to tolerate single-disk failures.

Consider these points when mirroring multihost disks:

Mirroring Root (/)

For maximum availability, you should mirror root (/), /usr, /var, /opt, and swap on the local disks. Under VERITAS Volume Manager, this means encapsulating the root disk and mirroring the generated subdisks. However, mirroring the root disk is not a requirement of Sun Cluster.

You should consider the risks, complexity, cost, and service time for the various alternatives concerning the root disk. There is not one answer for all configurations. You might want to consider your local Enterprise Services representative's preferred solution when deciding whether to mirror root.

Refer to your volume manager documentation for instructions on mirroring root.

Consider the following issues when deciding whether to mirror the root file system.

Note, however, that there is nothing in the Sun Cluster software that guarantees an immediate takeover. In fact, the takeover might not occur at all. For example, presume some sectors of a disk are bad. Presume they are all in the user data portions of a file that is crucial to some data service. The data service will start getting I/O errors, but the Sun Cluster node will stay up.

At a later point the primary root disk might return to service (perhaps after a power cycle or transient I/O errors) and subsequent boots are performed using the primary root disk specified in the OpenBoot(TM) PROM boot-device field. Note that a Solstice DiskSuite resync has not occurred--that requires a manual step when the drive is returned to service.

In this situation there was no manual repair task--the drive simply started working "well enough" to boot.

If there were changes to any files on the secondary (mirror) root device, they would not be reflected on the primary root device during boot time (causing a stale submirror). For example, changes to /etc/system would be lost. It is possible that some Solstice DiskSuite administrative commands changed /etc/system while the primary root device was out of service.

The boot program does not know whether it is booting from a mirror or an underlying physical device, and the mirroring becomes active part way through the boot process (after the metadevices are loaded). Before this point the system is vulnerable to stale submirror problems.

Solstice DiskSuite Mirroring Alternatives

Consider the following alternatives when deciding whether to mirror root (/) file systems under Solstice DiskSuite. The issues mentioned in this section are not applicable to VERITAS Volume Manager configurations.

Configuration Restrictions

This section describes Sun Cluster configuration restrictions.

Service and Application Restrictions

Note the following restrictions related to services and applications.

Default Port Numbers Reserved by Sun Cluster

Sun Cluster reserves certain port numbers for internal use. These port numbers are stored in the clustername.cdb file. Note the following reserved port numbers when planning your configuration, data services, and applications.

In addition, note that Solaris reserves ports 6000 to 6031 for UNIX Distributed Lock Manager (UDLM). UDLM is used with Oracle Parallel Server configurations.

Table 2-4 Default Ports Reserved by Sun Cluster

Port Number 

Reserved For... 

5556 

Cluster Membership Monitor 

5568-5599 

VERITAS Volume Manager cluster feature, vxclust

5559 

VERITAS Volume Manager cluster feature, vxkmsgd

5560 

VERITAS Volume Manager cluster feature, vxconfigd

603 

sm_configd and smad (for TCP and UDP)

Changing these port numbers is not allowed, except in the case of sm_configd and smad. To change the port numbers used by sm_configd or smad, edit the /etc/services files on all nodes. The files must be identical on all nodes. For all cases other than sm_configd and smad, you must change the application's port use rather than the cluster's.

Sun Cluster HA for NFS Restrictions

Note the following restrictions related to Sun Cluster HA for NFS.

Hardware Restrictions

Note the following hardware-related restrictions.

For more information about Sun Cluster 2.2 hardware considerations and procedures, see the Sun Cluster 2.2 Hardware Site Preparation, Planning, and Installation Guide and Sun Cluster 2.2 Hardware Service Manual.

Solstice DiskSuite Restrictions

Note the following restrictions related to Solstice DiskSuite.

Other Restrictions