Skip Navigation Links | |
Exit Print View | |
Oracle Solaris Cluster Software Installation Guide Oracle Solaris Cluster 4.1 |
1. Planning the Oracle Solaris Cluster Configuration
2. Installing Software on Global-Cluster Nodes
Overview of Installing the Software
How to Prepare for Cluster Software Installation
How to Install Oracle Solaris Software
How to Install pconsole Software on an Administrative Console
How to Install and Configure Oracle Solaris Cluster Quorum Server Software
How to Configure Internal Disk Mirroring
SPARC: How to Install Oracle VM Server for SPARC Software and Create Domains
How to Install Oracle Solaris Cluster Framework and Data Service Software Packages
How to Install the Availability Suite Feature of Oracle Solaris 11
3. Establishing the Global Cluster
4. Configuring Solaris Volume Manager Software
5. Creating a Cluster File System
This section provides information and procedures to install software on the cluster nodes.
How to Install pconsole Software on an Administrative Console
How to Install and Configure Oracle Solaris Cluster Quorum Server Software
SPARC: How to Install Oracle VM Server for SPARC Software and Create Domains
How to Install Oracle Solaris Cluster Framework and Data Service Software Packages
How to Install the Availability Suite Feature of Oracle Solaris 11
See Cluster Nodes in Oracle Solaris Cluster Concepts Guide for information about physical and virtual machines that are supported as cluster nodes.
See Oracle Solaris Cluster 4 Compatibility Guide for information about supported software and hardware for this release.
Contact your Oracle sales representative for the most current information about supported cluster configurations.
Oracle Solaris Cluster 4.1 Release Notes – Restrictions, bug workarounds, and other late-breaking information.
Oracle Solaris Cluster Concepts Guide - Overview of the Oracle Solaris Cluster product.
Oracle Solaris Cluster Software Installation Guide (this manual) – Planning guidelines and procedures for installing and configuring Oracle Solaris, Oracle Solaris Cluster, and volume manager software.
Oracle Solaris Cluster Data Services Planning and Administration Guide – Planning guidelines and procedures to install and configure data services.
The following is a partial list of products whose documentation you might need to reference during cluster installation:
Oracle Solaris OS
Solaris Volume Manager software
Third-party applications
Use the planning guidelines in Chapter 1, Planning the Oracle Solaris Cluster Configuration and in the Oracle Solaris Cluster Data Services Planning and Administration Guide to determine how to install and configure your cluster.
Caution - Plan your cluster installation completely. Identify requirements for all data services and third-party products before you begin Oracle Solaris and Oracle Solaris Cluster software installation. Failure to do so might result in installation errors that require you to completely reinstall the Oracle Solaris and Oracle Solaris Cluster software. |
See Chapter 11, Updating Your Software, in Oracle Solaris Cluster System Administration Guide for installation instructions.
Next Steps
If you want to install a machine as a quorum server to use as the quorum device in your cluster, go next to How to Install and Configure Oracle Solaris Cluster Quorum Server Software.
Otherwise, if you want to use an administrative console to communicate with the cluster nodes, go to How to Install pconsole Software on an Administrative Console.
Otherwise, choose the Oracle Solaris installation procedure to use.
To configure Oracle Solaris Cluster software by using the scinstall(1M) utility, go to How to Install Oracle Solaris Software to first install Oracle Solaris software.
To install and configure both Oracle Solaris and Oracle Solaris Cluster software in the same operation (Automated Installer method), go to How to Install and Configure Oracle Solaris and Oracle Solaris Cluster Software (Automated Installer).
Use this procedure to install the Oracle Solaris OS on the following systems, as applicable to your cluster configuration:
1. (Optional) An administrative console that you will install with pconsole software. For more information, see How to Install pconsole Software on an Administrative Console.
2. (Optional) A quorum server. For more information, see How to Install and Configure Oracle Solaris Cluster Quorum Server Software.
3. Each node in the global cluster, if you will not use the scinstall custom Automated Installer method to install software. For more information about Automated Installer installation of a cluster, see How to Install and Configure Oracle Solaris and Oracle Solaris Cluster Software (Automated Installer).
If your nodes are already installed with the Oracle Solaris OS but do not meet Oracle Solaris Cluster installation requirements, you might need to reinstall the Oracle Solaris software. Follow the steps in this procedure to ensure subsequent successful installation of the Oracle Solaris Cluster software. See Planning the Oracle Solaris OS for information about required root-disk partitioning and other Oracle Solaris Cluster installation requirements.
Before You Begin
Perform the following tasks:
Ensure that the hardware setup is complete and that connections are verified before you install Oracle Solaris software. See the Oracle Solaris Cluster 4.1 Hardware Administration Manual and your server and storage device documentation for details.
Ensure that your cluster configuration planning is complete. See How to Prepare for Cluster Software Installation for requirements and guidelines.
If you use a naming service, add address-to-name mappings for all public hostnames and logical addresses to any naming services that clients use for access to cluster services. See Public-Network IP Addresses for planning guidelines. See your Oracle Solaris system administrator documentation for information about using Oracle Solaris naming services.
Follow installation instructions in Installing Oracle Solaris 11.1 Systems.
Note - You must install all nodes in a cluster with the same version of the Oracle Solaris OS.
You can use any method that is normally used to install the Oracle Solaris software. During Oracle Solaris software installation, perform the following steps:
# pkg publisher PUBLISHER TYPE STATUS URI solaris origin online solaris-repository
For information about setting the solaris publisher, see Adding and Updating Oracle Solaris 11.1 Software Packages.
This series of installation procedures requires the following Oracle Solaris Cluster RBAC authorizations if the user is not the root role:
solaris.cluster.modify
solaris.cluster.admin
solaris.cluster.read
See Role-Based Access Control (Overview) in Oracle Solaris 11.1 Administration: Security Services for more information about using RBAC roles. See the Oracle Solaris Cluster man pages for the RBAC authorization that each Oracle Solaris Cluster subcommand requires.
phys-schost-1# mount | grep global | egrep -v node@ | awk '{print $1}'
phys-schost-new# mkdir -p mountpoint
For example, if the mount command returned the file system name /global/dg-schost-1, run mkdir -p /global/dg-schost-1 on the new node you are adding to the cluster.
Include those updates for storage array support. Also download any needed firmware that is contained in the hardware updates.
See Chapter 11, Updating Your Software, in Oracle Solaris Cluster System Administration Guide for installation instructions.
The setting of this value enables you to reboot the node if you are unable to access a login prompt.
grub edit> kernel /platform/i86pc/kernel/amd64/unix -B $ZFS-BOOTFS -k
For more information, see How to Boot a System With the Kernel Debugger Enabled (kmdb) in Booting and Shutting Down Oracle Solaris on x86 Platforms.
Perform this step regardless of whether you are using a naming service.
Note - During establishment of a new cluster or new cluster node, the scinstall utility automatically adds the public IP address of each node that is being configured to the /etc/inet/hosts file.
If you do not want to use the multiple-adapter IPMP groups that the scinstall utility configures during cluster creation, configure custom IPMP groups as you would in a stand-alone system. See Chapter 6, Administering IPMP (Tasks), in Managing Oracle Solaris 11.1 Network Performance for details.
During cluster creation, the scinstall utility configures each set of public-network adapters that use the same subnet and are not already configured in an IPMP group into a single multiple-adapter IPMP group. The scinstall utility ignores any existing IPMP groups.
Caution - If the Oracle Solaris Cluster software is already installed, do not issue this command. Running the stmsboot command on an active cluster node might cause Oracle Solaris services to go into the maintenance state. Instead, follow instructions in the stmsboot(1M) man page for using the stmsboot command in an Oracle Solaris Cluster environment. |
phys-schost# /usr/sbin/stmsboot -e
Enables Oracle Solaris I/O multipathing.
See How to Enable Multipathing in Oracle Solaris 11.1 Administration: SAN Configuration and Multipathing and the stmsboot(1M) man page for more information.
Next Steps
If you want to use the pconsole utility, go to How to Install pconsole Software on an Administrative Console.
If you want to use a quorum server, go to How to Install and Configure Oracle Solaris Cluster Quorum Server Software.
If your cluster nodes support the mirroring of internal hard drives and you want to configure internal disk mirroring, go to How to Configure Internal Disk Mirroring.
SPARC: If you want to install Oracle VM Server for SPARC, go to SPARC: How to Install Oracle VM Server for SPARC Software and Create Domains.
Otherwise, install the Oracle Solaris Cluster software on the cluster nodes.
If you already installed the Oracle Solaris OS on the cluster nodes, go to How to Install Oracle Solaris Cluster Framework and Data Service Software Packages.
If you want to use the scinstall custom Automated Installer (AI) method to install both Oracle Solaris OS and Oracle Solaris Cluster software on the cluster nodes, go to How to Install and Configure Oracle Solaris and Oracle Solaris Cluster Software (Automated Installer).
See Also
See the Oracle Solaris Cluster System Administration Guide for procedures to perform dynamic reconfiguration tasks in an Oracle Solaris Cluster configuration.
Note - You are not required to use an administrative console. If you do not use an administrative console, perform administrative tasks from one designated node in the cluster.
You cannot use this software to connect to Oracle VM Server for SPARC guest domains.
This procedure describes how to install the Parallel Console Access (pconsole) software on an administrative console. The pconsole utility is part of the Oracle Solaris 11 terminal/pconsole package.
The pconsole utility creates a host terminal window for each remote host that you specify on the command line. The utility also opens a central, or master, console window that you can use to send input to all nodes at one time. For additional information, see the pconsole(1) man page that is installed with the terminal/pconsole package.
You can use any desktop machine that runs a version of the Oracle Solaris OS that is supported by Oracle Solaris Cluster 4.1 software as an administrative console.
Before You Begin
Ensure that a supported version of the Oracle Solaris OS and any Oracle Solaris software updates are installed on the administrative console.
# pkg publisher PUBLISHER TYPE STATUS URI solaris origin online solaris-repository ha-cluster origin online ha-cluster-repository
For information about setting the solaris publisher, see Set the Publisher Origin to the File Repository URI in Copying and Creating Oracle Solaris 11.1 Package Repositories.
adminconsole# pkg install terminal/pconsole
adminconsole# pkg install pkgname …
|
When you install the Oracle Solaris Cluster man page packages on the administrative console, you can view them from the administrative console before you install Oracle Solaris Cluster software on the cluster nodes or on a quorum server.
Specify in the command each node that you want to connect to.
adminconsole# pconsole host[:port] […] &
See the procedures Logging Into the Cluster Remotely in Oracle Solaris Cluster System Administration Guide and How to Connect Securely to Cluster Consoles in Oracle Solaris Cluster System Administration Guide for additional information about how to use the pconsole utility. Also see the pconsole(1) man page that is installed as part of the Oracle Solaris 11 terminal/pconsole package.
Next Steps
If you want to use a quorum server, go to How to Install and Configure Oracle Solaris Cluster Quorum Server Software.
If your cluster nodes support the mirroring of internal hard drives and you want to configure internal disk mirroring, go to How to Configure Internal Disk Mirroring.
SPARC: If you want to install Oracle VM Server for SPARC, go to SPARC: How to Install Oracle VM Server for SPARC Software and Create Domains.
Otherwise, install the Oracle Solaris Cluster software on the cluster nodes.
If you already installed the Oracle Solaris OS on the cluster nodes, go to How to Install Oracle Solaris Cluster Framework and Data Service Software Packages.
If you want to use the scinstall custom Automated Installer (AI) method to install both Oracle Solaris OS and Oracle Solaris Cluster software on the cluster nodes, go to How to Install and Configure Oracle Solaris and Oracle Solaris Cluster Software (Automated Installer)
Perform this procedure to configure a host server as a quorum server.
Before You Begin
Perform the following tasks:
Ensure that the quorum server machine is connected to a public network that is accessible to the cluster nodes.
Disable the spanning tree algorithm on the Ethernet switches for the ports that are connected to the cluster public network where the quorum server will run.
# pkg publisher PUBLISHER TYPE STATUS URI solaris origin online solaris-repository ha-cluster origin online ha-cluster-repository
For information about setting the solaris publisher, see Set the Publisher Origin to the File Repository URI in Copying and Creating Oracle Solaris 11.1 Package Repositories.
quorumserver# pkg install ha-cluster-quorum-server-full
quorumserver# PATH=$PATH:/usr/cluster/bin
Identify the quorum server by specifying the port number and optionally the instance name.
If you provide an instance name, that name must be unique among your quorum servers.
If you do not provide an instance name, always refer to this quorum server by the port on which it listens.
The format for the entry is as follows:
/usr/cluster/lib/sc/scqsd [-d quorum-directory] [-i instance-name] -p port
The path to the directory where the quorum server can store quorum data.
The quorum server process creates one file per cluster in this directory to store cluster-specific quorum information.
By default, the value of this option is /var/scqsd. This directory must be unique for each quorum server that you configure.
A unique name that you choose for the quorum-server instance.
The port number on which the quorum server listens for requests from the cluster.
quorumserver# /usr/cluster/bin/clquorumserver start quorum-server
Identifies the quorum server. You can use the port number on which the quorum server listens. If you provided an instance name in the configuration file, you can use that name instead.
To start a single quorum server, provide either the instance name or the port number.
To start all quorum servers when you have multiple quorum servers configured, use the + operand.
Troubleshooting
Oracle Solaris Cluster Quorum Server software consists of the following packages:
ha-cluster/service/quorum-server
ha-cluster/service/quorum-server/locale
ha-cluster/service/quorum-server/manual
ha-cluster/service/quorum-server/manual/locale
These packages are contained in the ha-cluster/group-package/ha-cluster-quorum-server-full and ha-cluster/group-package/ha-cluster-quorum-server-l10n group packages.
The installation of these packages adds software to the /usr/cluster and /etc/scqsd directories. You cannot modify the location of the Oracle Solaris Cluster Quorum Server software.
If you receive an installation error message regarding the Oracle Solaris Cluster Quorum Server software, verify that the packages were properly installed.
Next Steps
If your cluster nodes support the mirroring of internal hard drives and you want to configure internal disk mirroring, go to How to Configure Internal Disk Mirroring.
SPARC: If you want to install Oracle VM Server for SPARC, go to SPARC: How to Install Oracle VM Server for SPARC Software and Create Domains.
Otherwise, install the Oracle Solaris Cluster software on the cluster nodes.
If you already installed the Oracle Solaris OS on the cluster nodes, go to How to Install Oracle Solaris Cluster Framework and Data Service Software Packages.
If you want to use the scinstall custom Automated Installer (AI) method to install both Oracle Solaris OS and Oracle Solaris Cluster software on the cluster nodes, go to How to Install and Configure Oracle Solaris and Oracle Solaris Cluster Software (Automated Installer).
Perform this procedure on each node of the global cluster to configure internal hardware RAID disk mirroring to mirror the system disk. This procedure is optional.
Note - Do not perform this procedure under either of the following circumstances:
Your servers do not support the mirroring of internal hard drives.
You have already established the cluster.
Instead, perform Mirroring Internal Disks on Servers that Use Internal Hardware Disk Mirroring or Integrated Mirroring in Oracle Solaris Cluster 4.1 Hardware Administration Manual.
Before You Begin
Ensure that the Oracle Solaris operating system and any necessary updates are installed.
phys-schost# raidctl -c clt0d0 clt1d0
Creates the mirror of primary disk to the mirror disk. Provide the name of your primary disk as the first argument and the name of the mirror disk as the second argument.
For specifics about how to configure your server's internal disk mirroring, refer to the documents that shipped with your server and the raidctl(1M) man page.
Next Steps
SPARC: If you want to install Oracle VM Server for SPARC, go to SPARC: How to Install Oracle VM Server for SPARC Software and Create Domains.
Otherwise, install the Oracle Solaris Cluster software on the cluster nodes.
If you already installed the Oracle Solaris OS on the cluster nodes, go to How to Install Oracle Solaris Cluster Framework and Data Service Software Packages.
If you want to use the scinstall custom Automated Installer (AI) method to install both Oracle Solaris OS and Oracle Solaris Cluster software on the cluster nodes, go to How to Install and Configure Oracle Solaris and Oracle Solaris Cluster Software (Automated Installer)
Perform this procedure to install Oracle VM Server for SPARC software on a physically clustered machine and to create I/O and guest domains.
Before You Begin
Perform the following tasks:
Ensure that the machine is SPARC hypervisor capable.
Have available Oracle VM Server for SPARC 2.1 Administration Guide and Oracle VM Server for SPARC 2.1 Release Notes.
Read the requirements and guidelines in SPARC: Guidelines for Oracle VM Server for SPARC in a Cluster.
Observe the following special instructions:
If you create guest domains, adhere to the Oracle Solaris Cluster guidelines for creating guest domains in a cluster.
Use the mode=sc option for all virtual switch devices that connect the virtual network devices that are used as the cluster interconnect.
For shared storage, map only the full SCSI disks into the guest domains.
Next Steps
If your server supports the mirroring of internal hard drives and you want to configure internal disk mirroring, go to How to Configure Internal Disk Mirroring.
Otherwise, install the Oracle Solaris Cluster software packages. Go to How to Install Oracle Solaris Cluster Framework and Data Service Software Packages.
Follow this procedure to perform one or more of the following installation tasks:
To install the Oracle Solaris Cluster framework and data service software packages on each node in the global cluster. These nodes can be physical machines or (SPARC only) Oracle VM Server for SPARC I/O domains or guest domains, or a combination of any of these types of nodes.
Note - If your physically clustered machines are configured with Oracle VM Server for SPARC, install Oracle Solaris Cluster software only in I/O domains or guest domains.
To install data services.
Note - You cannot add or remove individual packages that are part of the ha-cluster-minimal group package except by complete reinstallation or uninstallation. See How to Unconfigure Oracle Solaris Cluster Software to Correct Installation Problems and How to Uninstall Oracle Solaris Cluster Software From a Cluster Node in Oracle Solaris Cluster System Administration Guide for procedures to remove the cluster framework packages.
However, you can add or remove other, optional packages without removing the ha-cluster-minimal group package.
Before You Begin
Perform the following tasks:
Ensure that the Oracle Solaris OS is installed to support Oracle Solaris Cluster software.
If the Oracle Solaris software is already installed on the node, you must ensure that the Oracle Solaris installation meets the requirements for the Oracle Solaris Cluster software and any other software that you intend to install on the cluster. See How to Install Oracle Solaris Software for more information about installing the Oracle Solaris software to meet Oracle Solaris Cluster software requirements.
Choose which Oracle Solaris Cluster software packages to install.
The following table lists the primary group packages for the Oracle Solaris Cluster 4.1 software and the principal features that each group package contains. You must install at least the ha-cluster-framework-minimal group package.
|
As the root role, use the following command to start the pconsole utility:
adminconsole# pconsole host[:port] […] &
The pconsole utility also opens a master window from which you can send your input to all individual console windows at the same time.
During the installation of the Oracle Solaris OS, a restricted network profile is used that disables external access for certain network services. The restricted services include the RPC communication service, which is required for cluster communication.
Perform the following commands to restore external access to RPC communication.
# svccfg svc:> select network/rpc/bind svc:/network/rpc/bind> setprop config/local_only=false svc:/network/rpc/bind> quit # svcadm refresh network/rpc/bind:default # svcprop network/rpc/bind:default | grep local_only
The output of the last command should show that the local_only property is now set to false.
Alternatively, if your user account is assigned the System Administrator profile, issue commands as nonroot through a profile shell, or prefix the command with the pfexec command.
NWAM activates a single network interface and disables all others. For this reason, NWAM cannot coexist with the Oracle Solaris Cluster software and you must disable it before you configure or run your cluster. To disable NWAM, you enable the defaultfixed profile.
# netadm enable -p ncp defaultfixed # netadm list -p ncp defaultfixed
The certification page is displayed with download buttons for the key and the certificate.
In the following example the repository name is https://pkg.oracle.com/repository-location/.
# pkg set-publisher \ -k /var/pkg/ssl/Oracle_Solaris_Cluster_4.0.key.pem \ -c /var/pkg/ssl/Oracle_Solaris_Cluster_4.0.certificate.pem \ -O https://pkg.oracle.com/repository-location/ ha-cluster
Specifies the full path to the downloaded certificate file.
Specifies the URL to the Oracle Solaris Cluster 4.1 package repository.
For more information, see the pkg(1) man page.
Note - A valid Oracle license is required to access Oracle Software Delivery Cloud.
Oracle Solaris Cluster software is part of the Oracle Solaris Product Pack. Follow online instructions to complete selection of the media pack and download the software.
# lofiadm -a path-to-iso-image /dev/lofi/N # mount -F hsfs /dev/lofi/N /mnt
Specifies the full path and file name of the ISO image.
# pkg set-publisher -g file:///mnt/repo ha-cluster
# pkg publisher PUBLISHER TYPE STATUS URI solaris origin online solaris-repository ha-cluster origin online ha-cluster-repository
For information about setting the solaris publisher, see Set the Publisher Origin to the File Repository URI in Copying and Creating Oracle Solaris 11.1 Package Repositories.
# /usr/bin/pkg install package
$ pkg info -r package
Package installation succeeded if the state is Installed.
See Chapter 11, Updating Your Software, in Oracle Solaris Cluster System Administration Guide for installation instructions.
Next Steps
If you want to use the Availability Suite feature of Oracle Solaris software, install the Availability Suite software. Go to How to Install the Availability Suite Feature of Oracle Solaris 11.
Otherwise, to set up the root user environment, go to How to Set Up the Root Environment.
Before You Begin
Ensure that a minimum of Oracle Solaris 11 SRU 1 is installed.
# pkg publisher PUBLISHER TYPE STATUS URI solaris origin online solaris-repository
For information about setting the solaris publisher, see Set the Publisher Origin to the File Repository URI in Copying and Creating Oracle Solaris 11.1 Package Repositories.
# /usr/bin/pkg install storage/avs
For details, see Initial Configuration Settings in Sun StorageTek Availability Suite 4.0 Software Installation and Configuration Guide.
Next Steps
To set up the root user environment, go to How to Set Up the Root Environment.
Note - In an Oracle Solaris Cluster configuration, user initialization files for the various shells must verify that they are run from an interactive shell. The files must verify this before they attempt to output to the terminal. Otherwise, unexpected behavior or interference with data services might occur. See Customizing a User’s Work Environment in Managing User Accounts and User Environments in Oracle Solaris 11.1 for more information.
Perform this procedure on each node in the global cluster.
Note - Always make /usr/cluster/bin the first entry in the PATH. This placement ensures that Oracle Solaris Cluster commands take precedence over any other binaries that have the same name, thus avoiding unexpected behavior.
See your Oracle Solaris OS documentation, volume manager documentation, and other application documentation for additional file paths to set.
Next Steps
If you want to use the IP Filter feature of Oracle Solaris, go to How to Configure IP Filter.
Otherwise, configure Oracle Solaris Cluster software on the cluster nodes. Go to Establishing a New Global Cluster or New Global-Cluster Node.
Perform this procedure to configure the IP Filter feature of Oracle Solaris software on the global cluster.
Note - Only use IP Filter with failover data services. The use of IP Filter with scalable data services is not supported.
For more information about the IP Filter feature, see Chapter 4, IP Filter in Oracle Solaris (Overview), in Securing the Network in Oracle Solaris 11.1.
Before You Begin
Read the guidelines and restrictions to follow when you configure IP Filter in a cluster. See the “IP Filter” bullet item in Oracle Solaris OS Feature Restrictions.
Observe the following guidelines and requirements when you add filter rules to Oracle Solaris Cluster nodes.
In the ipf.conf file on each node, add rules to explicitly allow cluster interconnect traffic to pass unfiltered. Rules that are not interface specific are applied to all interfaces, including cluster interconnects. Ensure that traffic on these interfaces is not blocked mistakenly. If interconnect traffic is blocked, the IP Filter configuration interferes with cluster handshakes and infrastructure operations.
For example, suppose the following rules are currently used:
# Default block TCP/UDP unless some later rule overrides block return-rst in proto tcp/udp from any to any # Default block ping unless some later rule overrides block return-rst in proto icmp all
To unblock cluster interconnect traffic, add the following rules. The subnets used are for example only. Derive the subnets to use by using the ifconfig show-addr | grep interface command.
# Unblock cluster traffic on 172.16.0.128/25 subnet (physical interconnect) pass in quick proto tcp/udp from 172.16.0.128/25 to any pass out quick proto tcp/udp from 172.16.0.128/25 to any # Unblock cluster traffic on 172.16.1.0/25 subnet (physical interconnect) pass in quick proto tcp/udp from 172.16.1.0/25 to any pass out quick proto tcp/udp from 172.16.1.0/25 to any # Unblock cluster traffic on 172.16.4.0/23 (clprivnet0 subnet) pass in quick proto tcp/udp from 172.16.4.0/23 to any pass out quick proto tcp/udp from 172.16.4.0/23 to any
You can specify either the adapter name or the IP address for a cluster private network. For example, the following rule specifies a cluster private network by its adapter's name:
# Allow all traffic on cluster private networks. pass in quick on net1 all …
Oracle Solaris Cluster software fails over network addresses from node to node. No special procedure or code is needed at the time of failover.
All filtering rules that reference IP addresses of logical hostname and shared address resources must be identical on all cluster nodes.
Rules on a standby node will reference a nonexistent IP address. This rule is still part of the IP filter's active rule set and will become effective when the node receives the address after a failover.
All filtering rules must be the same for all NICs in the same IPMP group. In other words, if a rule is interface-specific, the same rule must also exist for all other interfaces in the same IPMP group.
For more information about IP Filter rules, see the ipf(4) man page.
phys-schost# svcadm enable /network/ipfilter:default
Next Steps
Configure Oracle Solaris Cluster software on the cluster nodes. Go to Establishing a New Global Cluster or New Global-Cluster Node.