Sun Cluster System Administration Guide for Solaris OS

Chapter 1 Introduction to Administering Sun Cluster

This chapter provides the following information about administering the cluster and the procedures for using Sun Cluster administration tools:

Overview of Administering Sun Cluster

Sun Cluster's highly available environment ensures that critical applications are available to end users. The system administrator's job is to make sure that the Sun Cluster configuration is stable and operational.

Familiarize yourself with the planning information in the Sun Cluster Software Installation Guide for Solaris OS and the Sun Cluster Concepts Guide for Solaris OS before beginning administration tasks. Sun Cluster administration is organized into tasks among the following manuals.

For the most part, you can perform Sun Cluster administration tasks while the cluster is operational. If you need take a node out of the cluster or even shut down the node, you can so while the rest of the nodes continue cluster operations. For Solaris 10, unless otherwise indicated, Sun Cluster administrative tasks should be performed in the global zone. For those procedures that require the entire cluster to be shut down, schedule downtime outside normal working hours, to impose minimal impact on the system. If you plan to shut down the cluster or a cluster node, notify users in advance.

Solaris Operating System (Solaris OS) Feature Restrictions

Do not enable or disable the following Sun Cluster services by using the Solaris 10 Service Management Facility (SMF) management interface.

Sun Cluster Service 

FMRI 

pnm

svc:/system/cluster/pnm:default

cl_event

svc:/system/cluster/cl_event:default

cl_eventlog

svc:/system/cluster/cl_eventlog:default

rpc_pmf

svc:/system/cluster/rpc_pmf:default

rpc_fed

svc:/system/cluster/rpc_fed:default

rgm

svc:/system/cluster/rgm:default

scdpm

svc:/system/cluster/scdpm:default

cl_ccra

svc:/system/cluster/cl_ccra:default

scsymon_srv

svc:/system/cluster/scsymon_srv:default

spm

svc:/system/cluster/spm:default

cl_svc_cluster_milestone

svc:/system/cluster/cl_svc_cluster_milestone:default

cl_svc_enable

svc:/system/cluster/cl_svc_enable:default

network-multipathing

svc:/system/cluster/network-multipathing

Administration Tools

You can perform administrative tasks on a Sun Cluster configuration by using a graphical user interface (GUI) or by using the command-line. The following sections provides an overview of the GUI and command line tools.

Graphical User Interface

Sun Cluster software supports GUI tools that you can use to perform various administrative tasks on your cluster. These GUI tools are Sun Cluster Manager and, if you are using Sun Cluster software on a SPARC based system, Sun Management Center. See Chapter 12, Administering Sun Cluster With the Graphical User Interfaces for more information and for procedures about configuring Sun Cluster Manager and Sun Management Center. For specific information about how to use Sun Cluster Manager, see the online help for the GUI.

Command-Line Interface

You can perform most Sun Cluster administration tasks interactively through the clsetup(1CL) utility. Whenever possible, administration procedures in this guide use the clsetup utility.

You can administer the following Main Menu items through the clsetup utility.

Other commands that you use to administer a Sun Cluster configuration are provided in the following list. See the man pages for more detailed information.

ccp(1M)

Starts remote console access to the cluster.

if_mpadm(1M)

Switches IP addresses from one adapter to another in an IP Network Multipathing group.

claccess(1CL)

Manages Sun Cluster access policies for adding nodes.

cldevice(1CL)

Manages Sun Cluster devices.

cldevicegroup(1CL)

Manages Sun Cluster device groups.

clinterconnect(1CL)

Manages the Sun Cluster interconnect.

clnasdevice(1CL)

Manages access to NAS devices for a Sun Cluster configuration.

clnode(1CL)

Manages Sun Cluster nodes.

clquorum(1CL)

Manages Sun Cluster quorum.

clreslogicalhostname(1CL)

Manages Sun Cluster resources for logical host names.

clresource(1CL)

Manages resources for Sun Cluster data services.

clresourcegroup(1CL)

Manages resources for Sun Cluster data services.

clresourcetype(1CL)

Manages resources for Sun Cluster data services.

clressharedaddress(1CL)

Manages Sun Cluster resources for shared addresses.

clsetup(1CL)

Configures a Sun Cluster configuration interactively.

clsnmphost(1CL)

Administers Sun Cluster SNMP hosts.

clsnmpmib(1CL)

Administers Sun Cluster SNMP MIB.

clsnmpuser(1CL)

Administers Sun Cluster SNMP users.

cltelemetryattribute(1CL)

Configures system resource monitoring.

cluster(1CL)

Manages the global configuration and the global status of the Sun Cluster configuration.

clvxvm

Initializes VERITAS Volume Manager (VxVM) on a Sun Cluster node and optionally performs root-disk encapsulation.

In addition, you can use commands to administer the volume manager portion of a Sun Cluster configuration. These commands depend on the specific volume manager that your cluster uses, either Solstice DiskSuiteTM, VERITAS Volume Manager, or Solaris Volume ManagerTM.

Preparing to Administer the Cluster

This section describes what to do to prepare for administering your cluster.

Documenting a Sun Cluster Hardware Configuration

Document the hardware aspects that are unique to your site as your Sun Cluster configuration is scaled. To save administration labor, refer to your hardware documentation when you change or upgrade the cluster. Labeling cables and connections between the various cluster components can also make administration easier.

Reduce the time required by a third-party service provider when servicing your cluster by keeping records of your original cluster configuration, and subsequent changes.

Using an Administrative Console

You can use either a dedicated workstation or a workstation connected through a management network as the administrative console, to administer the active cluster. Typically, you install and run the Cluster Control Panel (CCP) and graphical user interface (GUI) tools on the administrative console. For more information about the CCP, see How to Log In to the Cluster Remotely. For instructions on installing the Cluster Control Panel module for Sun Management Center and Sun Cluster Manager GUI tools, see the Sun Cluster Software Installation Guide for Solaris OS.

The administrative console is not a cluster node. The administrative console is used for remote access to the cluster nodes, either over the public network or through a network-based terminal concentrator.

If your SPARC cluster consists of a Sun EnterpriseTM 10000 server, you must log in from the administrative console to the System Service Processor (SSP). Connect by using the netcon command. The default method for netcon to connect with a Sun Enterprise 10000 domain is through the network interface. If the network is inaccessible, you can use netcon in “exclusive” mode by setting the -f option. You can also send ~* during a normal netcon session. Either of the previous solutions gives you the option of toggling to the serial interface if the network becomes unreachable.

Sun Cluster does not require a dedicated administrative console, but using a console provides the following benefits:

Backing Up the Cluster

Back up your cluster on a regular basis. Even though Sun Cluster software provides a highly available environment, with mirrored copies of data on the storage devices, Sun Cluster software is not a replacement for regular backups. A Sun Cluster configuration can survive multiple failures, but does not protect against user or program error, or catastrophic failure. Therefore, you must have a backup procedure in place to protect against data loss.

The following information should be included as part of your backup.

Beginning to Administer the Cluster

Table 1–1 provides a starting point for administering your cluster.

Table 1–1 Sun Cluster 3.2 Administration Tools

Task 

Tool 

Instructions 

Log in to the cluster remotely 

Use the ccp command to launch the Cluster Control Panel (CCP). Then select one of the following icons: cconsolecrlogin, cssh, or ctelnet.

How to Log In to the Cluster Remotely

How to Connect Securely to Cluster Consoles

Configure the cluster interactively 

Start the clsetup(1CL) utility.

How to Access the clsetup Utility

Display Sun Cluster release number and version information 

Use the clnode(1CL) command with the show-rev --v -node subcommand and option.

How to Display Sun Cluster Release and Version Information

Display installed resources, resource groups, and resource types 

Use the following commands to display the resource information: 

How to Display Configured Resource Types, Resource Groups, and Resources

Monitor cluster components graphically 

Use Sun Cluster Manager. 

See online help 

Administer some cluster components graphically 

Use Sun Cluster Manager or the Sun Cluster module for Sun Management Center (which is available with Sun Cluster on SPARC based systems only). 

For Sun Cluster Manager, see online help 

For Sun Management Center, see the Sun Management Center documentation 

Check the status of cluster components 

Use the cluster(1CL) command with the status subcommand.

How to Check the Status of Cluster Components

Check the status of IP network multipathing groups on the public network 

Use the clnode(1CL) status command with the -m option.

How to Check the Status of the Public Network

View the cluster configuration 

Use the cluster(1CL) command with the show subcommand.

How to View the Cluster Configuration

Check global mount points 

Use the sccheck(1M) command.

How to Validate a Basic Cluster Configuration

Look at the contents of Sun Cluster command logs 

Examine the /var/cluster/logs/commandlog file.

How to View the Contents of Sun Cluster Command Logs

Look at Sun Cluster system messages 

Examine the /var/adm/messages file.

Viewing System Messages in System Administration Guide: Advanced Administration

Monitor the status of Solstice DiskSuite 

Use the metastat commands.

Solaris Volume Manager documentation 

Monitor the status of Solaris Volume Manager if running Solaris 9 or Solaris 10 

Use the metastat command.

Solaris Volume Manager Administration Guide

ProcedureHow to Log In to the Cluster Remotely

The Cluster Control Panel (CCP) provides a launch pad for the cconsole, crlogin, cssh, and ctelnet tools. All tools start a multiple-window connection to a set of specified nodes. The multiple-window connection consists of a host window for each of the specified nodes and a common window. Input to the common window is sent to each of the host windows, enabling you to run commands simultaneously on all nodes of the cluster.

You can also start cconsole, crlogin, cssh, or ctelnet sessions from the command line.

By default, the cconsole utility uses a telnet connection to the node consoles. To establish Secure Shell connections to the consoles instead, enable the Use SSH checkbox in the Options menu of the cconsole window. Or, specify the -s option when you issue the ccp or cconsole command.

See the ccp(1M) and cconsole(1M) man pages for more information.

This procedure provides the long forms of the Sun Cluster commands. Most commands also have short forms. Except for the long and short forms of the command names, the commands are identical. For a list of the commands and their short forms, see Appendix A, Sun Cluster Object-Oriented Commands.

Before You Begin

Verify that the following prerequisites are met before starting the CCP:

  1. If you have a Sun Enterprise 10000 server platform, log in to the System Service Processor (SSP).

    1. Connect by using the netcon command.

    2. After the connection is made, type Shift~@ to unlock the console and gain write access.

  2. From the administrative console, start the CCP launch pad.


    # ccp clustername
    

    The CCP launch pad is displayed.

  3. To start a remote session with the cluster, click either the cconsole, crlogin, cssh, or ctelnet icon in the CCP launch pad.

ProcedureHow to Connect Securely to Cluster Consoles

Perform this procedure to establish Secure Shell connections to the consoles of the cluster nodes.

Before You Begin

Configure the clusters file, the serialports file, and the nsswitch.conf file if you are using a terminal concentrator. The files can be either /etc files or NIS or NIS+ databases.


Note –

In the serialports file, use the port number to use for secure connection to each console-access device. The default port number for Secure Shell connection is 22.


See the clusters(4) and serialports(4) man pages for more information.

  1. Become superuser on the administrative console.

  2. Start the cconsole utility in secure mode.


    # cconsole -s [-l username] [-p ssh-port]
    
    -s

    Enables secure shell connection.

    -l username

    Specifies the user name for the remote connections. If the -l option is not specified, the user name that launched the cconsole utility is used.

    -p ssh-port

    Specifies the Secure Shell port number to use. If the -p option is not specified, the default port number 22 is used for the secure connections.

ProcedureHow to Access the clsetup Utility

The clsetup(1CL) utility enables you to interactively configure quorum, resource group, cluster transport, private host name, device group, and new node options for the cluster.

This procedure provides the long forms of the Sun Cluster commands. Most commands also have short forms. Except for the long and short forms of the command names, the commands are identical. For a list of the commands and their short forms, see Appendix A, Sun Cluster Object-Oriented Commands.

  1. Become superuser on any node in the cluster.

  2. Start the clsetup utility.


    # clsetup
    

    The Main Menu is displayed.

  3. Choose your configuration from the menu. Follow the onscreen instructions to complete a task.

See Also

See the clsetup online help for more information.

ProcedureHow to Display Sun Cluster Patch Information

You do not need to be logged in as superuser to perform this procedure.

This procedure provides the long forms of the Sun Cluster commands. Most commands also have short forms. Except for the long and short forms of the command names, the commands are identical. For a list of the commands and their short forms, see Appendix A, Sun Cluster Object-Oriented Commands.

  1. Display the Sun Cluster patch information:


    # showrev -p
    

    Sun Cluster update releases are identified by the main product patch number plus the update version.


Example 1–1 Displaying Sun Cluster Patch Information

The following example displays information about patch 110648-05.


# showrev -p | grep 110648
Patch: 110648-05 Obsoletes:  Requires:  Incompatibles:  Packages: 

ProcedureHow to Display Sun Cluster Release and Version Information

You do not need to be logged in as superuser to perform this procedure.

This procedure provides the long forms of the Sun Cluster commands. Most commands also have short forms. Except for the long and short forms of the command names, the commands are identical. For a list of the commands and their short forms, see Appendix A, Sun Cluster Object-Oriented Commands.

  1. Display Sun Cluster release and version information:


    # clnode show-rev -v -node
    

    This command displays Sun Cluster release number and version strings for all Sun Cluster packages.


Example 1–2 Displaying Sun Cluster Release and Version Information

The following example displays the cluster's release information and version information for all packages.


# clnode show-rev
3.2

#% clnode show-rev -v
Sun Cluster 3.2 for Solaris 9 sparc

SUNWscr:       3.2.0,REV=2006.02.17.18.11
SUNWscu:       3.2.0,REV=2006.02.17.18.11
SUNWsczu:      3.2.0,REV=2006.02.17.18.11
SUNWscsck:     3.2.0,REV=2006.02.17.18.11
SUNWscnm:      3.2.0,REV=2006.02.17.18.11
SUNWscdev:     3.2.0,REV=2006.02.17.18.11
SUNWscgds:     3.2.0,REV=2006.02.17.18.11
SUNWscman:     3.2.0,REV=2005.10.18.08.42
SUNWscsal:     3.2.0,REV=2006.02.17.18.11
SUNWscsam:     3.2.0,REV=2006.02.17.18.11
SUNWscvm:      3.2.0,REV=2006.02.17.18.11
SUNWmdm:       3.2.0,REV=2006.02.17.18.11
SUNWscmasa:    3.2.0,REV=2006.02.17.18.11
SUNWscmautil:  3.2.0,REV=2006.02.17.18.11
SUNWscmautilr: 3.2.0,REV=2006.02.17.18.11
SUNWjfreechart: 3.2.0,REV=2006.02.17.18.11
SUNWscva:      3.2.0,REV=2006.02.17.18.11
SUNWscspm:     3.2.0,REV=2006.02.17.18.11
SUNWscspmu:    3.2.0,REV=2006.02.17.18.11
SUNWscspmr:    3.2.0,REV=2006.02.17.18.11
SUNWscderby:   3.2.0,REV=2006.02.17.18.11
SUNWsctelemetry: 3.2.0,REV=2006.02.17.18.11
SUNWscrsm:     3.2.0,REV=2006.02.17.18.11
SUNWcsc:       3.2.0,REV=2006.02.21.10.16
SUNWcscspm:    3.2.0,REV=2006.02.21.10.16
SUNWcscspmu:   3.2.0,REV=2006.02.21.10.16
SUNWdsc:       3.2.0,REV=2006.02.21.10.09
SUNWdscspm:    3.2.0,REV=2006.02.21.10.09
SUNWdscspmu:   3.2.0,REV=2006.02.21.10.09
SUNWesc:       3.2.0,REV=2006.02.21.10.11
SUNWescspm:    3.2.0,REV=2006.02.21.10.11
SUNWescspmu:   3.2.0,REV=2006.02.21.10.11
SUNWfsc:       3.2.0,REV=2006.02.21.10.06
SUNWfscspm:    3.2.0,REV=2006.02.21.10.06
SUNWfscspmu:   3.2.0,REV=2006.02.21.10.06
SUNWhsc:       3.2.0,REV=2006.02.21.10.20
SUNWhscspm:    3.2.0,REV=2006.02.21.10.20
SUNWhscspmu:   3.2.0,REV=2006.02.21.10.20
SUNWjsc:       3.2.0,REV=2006.02.21.10.22
SUNWjscman:    3.2.0,REV=2006.02.21.10.22
SUNWjscspm:    3.2.0,REV=2006.02.21.10.22
SUNWjscspmu:   3.2.0,REV=2006.02.21.10.22
SUNWksc:       3.2.0,REV=2006.02.21.10.14
SUNWkscspm:    3.2.0,REV=2006.02.21.10.14
SUNWkscspmu:   3.2.0,REV=2006.02.21.10.14

ProcedureHow to Display Configured Resource Types, Resource Groups, and Resources

You can also accomplish this procedure by using the Sun Cluster Manager GUI. Refer to Chapter 12, Administering Sun Cluster With the Graphical User Interfaces or see the Sun Cluster Manager online help for more information.

This procedure provides the long forms of the Sun Cluster commands. Most commands also have short forms. Except for the long and short forms of the command names, the commands are identical. For a list of the commands and their short forms, see Appendix A, Sun Cluster Object-Oriented Commands.

Before You Begin

Users other than superuser require solaris.cluster.read RBAC authorization to use this subcommand.

  1. Display the cluster's configured resource types, resource groups, and resources.


    % cluster show -t resource,resourcetype,resourcegroup
    

    For information about individual resources, resource groups, and resource types, use the show subcommand with one of the following commands:

    • resource

    • resource group

    • resourcetype


Example 1–3 Displaying Configured Resource Types, Resource Groups, and Resources

The following example shows the resource types (RT Name), resource groups (RG Name), and resources (RS Name ) configured for the cluster schost.


% cluster show -t resource,resourcetype,resourcegroup


=== Registered Resource Types ===

Resource Type:                                  SUNW.qfs
  RT_description:                                  SAM-QFS Agent on SunCluster
  RT_version:                                      3.1
  API_version:                                     3
  RT_basedir:                                      /opt/SUNWsamfs/sc/bin
  Single_instance:                                 False
  Proxy:                                           False
  Init_nodes:                                      All potential masters
  Installed_nodes:                                 <All>
  Failover:                                        True
  Pkglist:                                         <NULL>
  RT_system:                                       False

=== Resource Groups and Resources ===

Resource Group:                                 qfs-rg
  RG_description:                                  <NULL>
  RG_mode:                                         Failover
  RG_state:                                        Managed
  Failback:                                        False
  Nodelist:                                        phys-schost-2 phys-schost-1

  --- Resources for Group qfs-rg ---

  Resource:                                     qfs-res
    Type:                                          SUNW.qfs
    Type_version:                                  3.1
    Group:                                         qfs-rg
    R_description:                                 
    Resource_project_name:                         default
    Enabled{phys-schost-2}:                        True
    Enabled{phys-schost-1}:                        True
    Monitored{phys-schost-2}:                      True
    Monitored{phys-schost-1}:                      True

ProcedureHow to Check the Status of Cluster Components

You can also accomplish this procedure by using the Sun Cluster Manager GUI. See the Sun Cluster Manager online help for more information.

This procedure provides the long forms of the Sun Cluster commands. Most commands also have short forms. Except for the long and short forms of the command names, the commands are identical. For a list of the commands and their short forms, see Appendix A, Sun Cluster Object-Oriented Commands.

Before You Begin

Users other than superuser require solaris.cluster.read RBAC authorization to use the status subcommand.

  1. Check the status of cluster components.


    % cluster status
    

Example 1–4 Checking the Status of Cluster Components

The following example provides a sample of status information for cluster components returned by cluster(1CL) status.


% cluster status
=== Cluster Nodes ===

--- Node Status ---

Node Name                                       Status
---------                                       ------
phys-schost-1                                   Online
phys-schost-2                                   Online


=== Cluster Transport Paths ===

Endpoint1               Endpoint2               Status
---------               ---------               ------
phys-schost-1:qfe1      phys-schost-4:qfe1      Path online
phys-schost-1:hme1      phys-schost-4:hme1      Path online


=== Cluster Quorum ===

--- Quorum Votes Summary ---

            Needed   Present   Possible
            ------   -------   --------
            3        3         4


--- Quorum Votes by Node ---

Node Name       Present       Possible       Status
---------       -------       --------       ------
phys-schost-1   1             1              Online
phys-schost-2   1             1              Online


--- Quorum Votes by Device ---

Device Name           Present      Possible          Status
-----------               -------      --------      ------
/dev/did/rdsk/d2s2      1            1                Online
/dev/did/rdsk/d8s2      0            1                Offline


=== Cluster Device Groups ===

--- Device Group Status ---

Device Group Name     Primary          Secondary    Status
-----------------     -------          ---------    ------
schost-2              phys-schost-2     -           Degraded


--- Spare, Inactive, and In Transition Nodes ---

Device Group Name   Spare Nodes   Inactive Nodes   In Transistion Nodes
-----------------   -----------   --------------   --------------------
schost-2            -             -                -


=== Cluster Resource Groups ===

Group Name        Node Name      Suspended      Status
----------        ---------      ---------      ------
test-rg           phys-schost-1       No             Offline
                  phys-schost-2       No             Online

test-rg           phys-schost-1       No             Offline
                  phys-schost-2       No             Error--stop failed

test-rg           phys-schost-1       No             Online
                  phys-schost-2       No             Online


=== Cluster Resources ===

Resource Name     Node Name     Status               Message
-------------     ---------     ------               -------
test_1            phys-schost-1      Offline         Offline
                  phys-schost-2      Online          Online

test_1            phys-schost-1      Offline         Offline
                  phys-schost-2      Stop failed     Faulted

test_1            phys-schost-1      Online          Online
                  phys-schost-2      Online          Online


Device Instance             Node                     Status
---------------             ----                     ------
/dev/did/rdsk/d2            phys-schost-1            Ok

/dev/did/rdsk/d3            phys-schost-1            Ok
                            phys-schost-2            Ok

/dev/did/rdsk/d4            phys-schost-1            Ok
                            phys-schost-2            Ok

/dev/did/rdsk/d6            phys-schost-2            Ok 

ProcedureHow to Check the Status of the Public Network

You can also accomplish this procedure by using the Sun Cluster Manager GUI. See the Sun Cluster Manager online help for more information.

This procedure provides the long forms of the Sun Cluster commands. Most commands also have short forms. Except for the long and short forms of the command names, the commands are identical. For a list of the commands and their short forms, see Appendix A, Sun Cluster Object-Oriented Commands.

To check the status of the IP Network Multipathing groups, use the clnode(1CL) command with the status subcommand.

Before You Begin

Users other than superuser require solaris.cluster.read RBAC authorization to use this subcommand.

  1. Check the status of cluster components.


    % clnode status -m
    

Example 1–5 Checking the Public Network Status

The following example provides a sample of status information for cluster components returned by the clnode status command.


% clnode status -m
--- Node IPMP Group Status ---

Node Name         Group Name    Status    Adapter    Status
---------         ----------    ------    -------    ------
phys-schost-1     test-rg       Online    qfe1       Online
phys-schost-2     test-rg       Online    qfe1       Online 

ProcedureHow to View the Cluster Configuration

You can also accomplish this procedure by using the Sun Cluster Manager GUI. See the Sun Cluster Manager online help for more information.

This procedure provides the long forms of the Sun Cluster commands. Most commands also have short forms. Except for the long and short forms of the command names, the commands are identical. For a list of the commands and their short forms, see Appendix A, Sun Cluster Object-Oriented Commands.

Before You Begin

Users other than superuser require solaris.cluster.read RBAC authorization to use the status subcommand.

  1. View the cluster configuration.


    % cluster show

    To display more information by using the cluster command, use the verbose options. See the cluster(1CL) man page for details.


Example 1–6 Viewing the Cluster Configuration

The following example lists the cluster configuration.


% cluster show

=== Cluster ===                                

Cluster Name:                                   cluster-1
  installmode:                                     disabled
  heartbeat_timeout:                               10000
  heartbeat_quantum:                               1000
  private_netaddr:                                 172.16.0.0
  private_netmask:                                 255.255.248.0
  max_nodes:                                       64
  max_privatenets:                                 10
  global_fencing:                                  Unknown
  Node List:                                       phys-schost-1
  Node Zones:                                      phys_schost-2:za

  === Host Access Control ===                  

  Cluster name:                                 clustser-1
    Allowed hosts:                                 phys-schost-1, phys-schost-2:za
    Authentication Protocol:                       sys

  === Cluster Nodes ===                        

  Node Name:                                    phys-schost-1
    Node ID:                                       1
    Type:                                          cluster
    Enabled:                                       yes
    privatehostname:                               clusternode1-priv
    reboot_on_path_failure:                        disabled
    globalzoneshares:                              3
    defaultpsetmin:                                1
    quorum_vote:                                   1
    quorum_defaultvote:                            1
    quorum_resv_key:                               0x43CB1E1800000001
    Transport Adapter List:                        qfe3, hme0

    --- Transport Adapters for phys-schost-1 ---    

    Transport Adapter:                          qfe3
      Adapter State:                               Enabled
      Adapter Transport Type:                      dlpi
      Adapter Property(device_name):               qfe
      Adapter Property(device_instance):           3
      Adapter Property(lazy_free):                 1
      Adapter Property(dlpi_heartbeat_timeout):    10000
      Adapter Property(dlpi_heartbeat_quantum):    1000
      Adapter Property(nw_bandwidth):              80
      Adapter Property(bandwidth):                 10
      Adapter Property(ip_address):                172.16.1.1
      Adapter Property(netmask):                   255.255.255.128
      Adapter Port Names:                          0
      Adapter Port State(0):                       Enabled

    Transport Adapter:                          hme0
      Adapter State:                               Enabled
      Adapter Transport Type:                      dlpi
      Adapter Property(device_name):               hme
      Adapter Property(device_instance):           0
      Adapter Property(lazy_free):                 0
      Adapter Property(dlpi_heartbeat_timeout):    10000
      Adapter Property(dlpi_heartbeat_quantum):    1000
      Adapter Property(nw_bandwidth):              80
      Adapter Property(bandwidth):                 10
      Adapter Property(ip_address):                172.16.0.129
      Adapter Property(netmask):                   255.255.255.128
      Adapter Port Names:                          0
      Adapter Port State(0):                       Enabled

    --- SNMP MIB Configuration on phys-schost-1 --- 

    SNMP MIB Name:                              Event
      State:                                       Disabled
      Protocol:                                    SNMPv2

    --- SNMP Host Configuration on phys-schost-1 ---

    --- SNMP User Configuration on phys-schost-1 ---

    SNMP User Name:                             foo
      Authentication Protocol:                     MD5
      Default User:                                No

  Node Name:                                    phys-schost-2:za
    Node ID:                                       2
    Type:                                          cluster
    Enabled:                                       yes
    privatehostname:                               clusternode2-priv
    reboot_on_path_failure:                        disabled
    globalzoneshares:                              1
    defaultpsetmin:                                2
    quorum_vote:                                   1
    quorum_defaultvote:                            1
    quorum_resv_key:                               0x43CB1E1800000002
    Transport Adapter List:                        hme0, qfe3

    --- Transport Adapters for phys-schost-2 ---    

    Transport Adapter:                          hme0
      Adapter State:                               Enabled
      Adapter Transport Type:                      dlpi
      Adapter Property(device_name):               hme
      Adapter Property(device_instance):           0
      Adapter Property(lazy_free):                 0
      Adapter Property(dlpi_heartbeat_timeout):    10000
      Adapter Property(dlpi_heartbeat_quantum):    1000
      Adapter Property(nw_bandwidth):              80
      Adapter Property(bandwidth):                 10
      Adapter Property(ip_address):                172.16.0.130
      Adapter Property(netmask):                   255.255.255.128
      Adapter Port Names:                          0
      Adapter Port State(0):                       Enabled

    Transport Adapter:                          qfe3
      Adapter State:                               Enabled
      Adapter Transport Type:                      dlpi
      Adapter Property(device_name):               qfe
      Adapter Property(device_instance):           3
      Adapter Property(lazy_free):                 1
      Adapter Property(dlpi_heartbeat_timeout):    10000
      Adapter Property(dlpi_heartbeat_quantum):    1000
      Adapter Property(nw_bandwidth):              80
      Adapter Property(bandwidth):                 10
      Adapter Property(ip_address):                172.16.1.2
      Adapter Property(netmask):                   255.255.255.128
      Adapter Port Names:                          0
      Adapter Port State(0):                       Enabled

    --- SNMP MIB Configuration on phys-schost-2 --- 

    SNMP MIB Name:                              Event
      State:                                       Disabled
      Protocol:                                    SNMPv2

    --- SNMP Host Configuration on phys-schost-2 ---

    --- SNMP User Configuration on phys-schost-2 ---

  === Transport Cables ===                     

  Transport Cable:                              phys-schost-1:qfe3,switch2@1
    Cable Endpoint1:                               phys-schost-1:qfe3
    Cable Endpoint2:                               switch2@1
    Cable State:                                   Enabled

  Transport Cable:                              phys-schost-1:hme0,switch1@1
    Cable Endpoint1:                               phys-schost-1:hme0
    Cable Endpoint2:                               switch1@1
    Cable State:                                   Enabled

  Transport Cable:                              phys-schost-2:hme0,switch1@2
    Cable Endpoint1:                               phys-schost-2:hme0
    Cable Endpoint2:                               switch1@2
    Cable State:                                   Enabled

  Transport Cable:                              phys-schost-2:qfe3,switch2@2
    Cable Endpoint1:                               phys-schost-2:qfe3
    Cable Endpoint2:                               switch2@2
    Cable State:                                   Enabled

  === Transport Switches ===                   

  Transport Switch:                             switch2
    Switch State:                                  Enabled
    Switch Type:                                   switch
    Switch Port Names:                             1 2
    Switch Port State(1):                          Enabled
    Switch Port State(2):                          Enabled

  Transport Switch:                             switch1
    Switch State:                                  Enabled
    Switch Type:                                   switch
    Switch Port Names:                             1 2
    Switch Port State(1):                          Enabled
    Switch Port State(2):                          Enabled


  === Quorum Devices ===                       

  Quorum Device Name:                           d3
    Enabled:                                       yes
    Votes:                                         1
    Global Name:                                   /dev/did/rdsk/d3s2
    Type:                                          scsi
    Access Mode:                                   scsi2
    Hosts (enabled):                               phys-schost-1, phys-schost-2

  Quorum Device Name:                           qs1
    Enabled:                                       yes
    Votes:                                         1
    Global Name:                                   qs1
    Type:                                          quorum_server
    Hosts (enabled):                               phys-schost-1, phys-schost-2
    Quorum Server Host:                            10.11.114.83
    Port:                                          9000


  === Device Groups ===                        

  Device Group Name:                            testdg3
    Type:                                          SVM
    failback:                                      no
    Node List:                                     phys-schost-1, phys-schost-2
    preferenced:                                   yes
    numsecondaries:                                1
    diskset name:                                  testdg3

  === Registered Resource Types ===            

  Resource Type:                                SUNW.LogicalHostname:2
    RT_description:                                Logical Hostname Resource Type
    RT_version:                                    2
    API_version:                                   2
    RT_basedir:                                    /usr/cluster/lib/rgm/rt/hafoip
    Single_instance:                               False
    Proxy:                                         False
    Init_nodes:                                    All potential masters
    Installed_nodes:                               <All>
    Failover:                                      True
    Pkglist:                                       SUNWscu
    RT_system:                                     True

  Resource Type:                                SUNW.SharedAddress:2
    RT_description:                                HA Shared Address Resource Type
    RT_version:                                    2
    API_version:                                   2
    RT_basedir:                                    /usr/cluster/lib/rgm/rt/hascip
    Single_instance:                               False
    Proxy:                                         False
    Init_nodes:                                    <Unknown>
    Installed_nodes:                              <All>
    Failover:                                      True
    Pkglist:                                       SUNWscu
    RT_system:                                     True

  Resource Type:                                SUNW.HAStoragePlus:4
    RT_description:                                HA Storage Plus
    RT_version:                                    4
    API_version:                                   2
    RT_basedir:                                    /usr/cluster/lib/rgm/rt/hastorageplus
    Single_instance:                               False
    Proxy:                                         False
    Init_nodes:                                    All potential masters
    Installed_nodes:                               <All>
    Failover:                                      False
    Pkglist:                                       SUNWscu
    RT_system:                                     False

  Resource Type:                                SUNW.haderby
    RT_description:                                haderby server for Sun Cluster
    RT_version:                                    1
    API_version:                                   7
    RT_basedir:                                    /usr/cluster/lib/rgm/rt/haderby
    Single_instance:                               False
    Proxy:                                         False
    Init_nodes:                                    All potential masters
    Installed_nodes:                               <All>
    Failover:                                      False
    Pkglist:                                       SUNWscderby
    RT_system:                                     False

  Resource Type:                                SUNW.sctelemetry
    RT_description:                                sctelemetry service for Sun Cluster
    RT_version:                                    1
    API_version:                                   7
    RT_basedir:                                    /usr/cluster/lib/rgm/rt/sctelemetry
    Single_instance:                               True
    Proxy:                                         False
    Init_nodes:                                    All potential masters
    Installed_nodes:                               <All>
    Failover:                                      False
    Pkglist:                                       SUNWsctelemetry
    RT_system:                                     False

  === Resource Groups and Resources ===        

  Resource Group:                               HA_RG
    RG_description:                                <Null>
    RG_mode:                                       Failover
    RG_state:                                      Managed
    Failback:                                      False
    Nodelist:                                      phys-schost-1 phys-schost-2

    --- Resources for Group HA_RG ---          

    Resource:                                   HA_R
      Type:                                        SUNW.HAStoragePlus:4
      Type_version:                                4
      Group:                                       HA_RG
      R_description:                               
      Resource_project_name:                       SCSLM_HA_RG
      Enabled{phys-schost-1}:                      True
      Enabled{phys-schost-2}:                      True
      Monitored{phys-schost-1}:                    True
      Monitored{phys-schost-2}:                    True

  Resource Group:                               cl-db-rg
    RG_description:                                <Null>
    RG_mode:                                       Failover
    RG_state:                                      Managed
    Failback:                                      False
    Nodelist:                                      phys-schost-1 phys-schost-2

    --- Resources for Group cl-db-rg ---       

    Resource:                                   cl-db-rs
      Type:                                        SUNW.haderby
      Type_version:                                1
      Group:                                       cl-db-rg
      R_description:                               
      Resource_project_name:                       default
      Enabled{phys-schost-1}:                      True
      Enabled{phys-schost-2}:                      True
      Monitored{phys-schost-1}:                    True
      Monitored{phys-schost-2}:                    True

  Resource Group:                               cl-tlmtry-rg
    RG_description:                                <Null>
    RG_mode:                                       Scalable
    RG_state:                                      Managed
    Failback:                                      False
    Nodelist:                                      phys-schost-1 phys-schost-2

    --- Resources for Group cl-tlmtry-rg ---   

    Resource:                                   cl-tlmtry-rs
      Type:                                        SUNW.sctelemetry
      Type_version:                                1
      Group:                                       cl-tlmtry-rg
      R_description:                               
      Resource_project_name:                       default
      Enabled{phys-schost-1}:                      True
      Enabled{phys-schost-2}:                      True
      Monitored{phys-schost-1}:                    True
      Monitored{phys-schost-2}:                    True

  === DID Device Instances ===                 

  DID Device Name:                              /dev/did/rdsk/d1
    Full Device Path:                              phys-schost-1:/dev/rdsk/c0t2d0
    Replication:                                   none
    default_fencing:                               global

  DID Device Name:                              /dev/did/rdsk/d2
    Full Device Path:                              phys-schost-1:/dev/rdsk/c1t0d0
    Replication:                                   none
    default_fencing:                               global

  DID Device Name:                              /dev/did/rdsk/d3
    Full Device Path:                              phys-schost-2:/dev/rdsk/c2t1d0
    Full Device Path:                              phys-schost-1:/dev/rdsk/c2t1d0
    Replication:                                   none
    default_fencing:                               global

  DID Device Name:                              /dev/did/rdsk/d4
    Full Device Path:                              phys-schost-2:/dev/rdsk/c2t2d0
    Full Device Path:                              phys-schost-1:/dev/rdsk/c2t2d0
    Replication:                                   none
    default_fencing:                               global

  DID Device Name:                              /dev/did/rdsk/d5
    Full Device Path:                              phys-schost-2:/dev/rdsk/c0t2d0
    Replication:                                   none
    default_fencing:                               global

  DID Device Name:                              /dev/did/rdsk/d6
    Full Device Path:                              phys-schost-2:/dev/rdsk/c1t0d0
    Replication:                                   none
    default_fencing:                               global

  === NAS Devices ===                          

  Nas Device:                                   nas_filer1
    Type:                                          netapp
    User ID:                                       root

  Nas Device:                                   nas2
    Type:                                          netapp
    User ID:                                       llai

ProcedureHow to Validate a Basic Cluster Configuration

The sccheck(1M) command runs a set of checks to validate the basic configuration that is required for a cluster to function properly. If no checks fail, sccheck returns to the shell prompt. If a check fails, sccheck produces reports in either the specified or the default output directory. If you run sccheck against more than one node, sccheck will produce a report for each node and a report for multi-node checks.

The sccheck command runs in two steps: data collection and analysis. Data collection can be time consuming, depending on the system configuration. You can run sccheck in verbose mode with the -v1 flag to print progress messages. Alternately, you can use the -v2 flag to run sccheck in highly verbose mode, which prints more detailed progress messages, especially during data collection.


Note –

Run sccheck after performing an administration procedure that might result in changes to devices, volume management components, or the Sun Cluster configuration.


  1. Become superuser on any node in the cluster.


    % su
    
  2. Verify the cluster configuration.


    # sccheck
    

Example 1–7 Checking the Cluster Configuration With All Checks Passing

The following example shows sccheck being run in verbose mode against nodes phys-schost-1 and phys-schost-2 with all checks passing.


# sccheck -v1 -h phys-schost-1,phys-schost-2

sccheck: Requesting explorer data and node report from phys-schost-1.
sccheck: Requesting explorer data and node report from phys-schost-2.
sccheck: phys-schost-1: Explorer finished.
sccheck: phys-schost-1: Starting single-node checks.
sccheck: phys-schost-1: Single-node checks finished.
sccheck: phys-schost-2: Explorer finished.
sccheck: phys-schost-2: Starting single-node checks.
sccheck: phys-schost-2: Single-node checks finished.
sccheck: Starting multi-node checks.
sccheck: Multi-node checks finished
# 


Example 1–8 Checking the Cluster Configuration With a Failed Check

The following example shows the node phys-schost-2 in the cluster suncluster missing the mount point /global/phys-schost-1. Reports are created in the output directory /var/cluster/sccheck/myReports/.


# sccheck -v1 -h phys-schost-1,phys-schost-2 -o /var/cluster/sccheck/myReports

sccheck: Requesting explorer data and node report from phys-schost-1.
sccheck: Requesting explorer data and node report from phys-schost-2.
sccheck: phys-schost-1: Explorer finished.
sccheck: phys-schost-1: Starting single-node checks.
sccheck: phys-schost-1: Single-node checks finished.
sccheck: phys-schost-2: Explorer finished.
sccheck: phys-schost-2: Starting single-node checks.
sccheck: phys-schost-2: Single-node checks finished.
sccheck: Starting multi-node checks.
sccheck: Multi-node checks finished.
sccheck: One or more checks failed.
sccheck: The greatest severity of all check failures was 3 (HIGH).
sccheck: Reports are in /var/cluster/sccheck/myReports.
# 
# cat /var/cluster/sccheck/myReports/sccheck-results.suncluster.txt
...
===================================================
= ANALYSIS DETAILS =
===================================================
------------------------------------
CHECK ID : 3065
SEVERITY : HIGH
FAILURE  : Global filesystem /etc/vfstab entries are not consistent across 
all Sun Cluster 3.x nodes.
ANALYSIS : The global filesystem /etc/vfstab entries are not consistent across 
all nodes in this cluster.
Analysis indicates:
FileSystem '/global/phys-schost-1' is on 'phys-schost-1' but missing from 'phys-schost-2'.
RECOMMEND: Ensure each node has the correct /etc/vfstab entry for the 
filesystem(s) in question.
...
 #

ProcedureHow to Check the Global Mount Points

The sccheck(1M) command includes checks that examine the /etc/vfstab file for configuration errors with the cluster file system and its global mount points.


Note –

Run sccheck after making cluster configuration changes that have affected devices or volume management components.


  1. Become superuser on any node in the cluster.


    % su
    
  2. Verify the cluster configuration.


    # sccheck
    

Example 1–9 Checking the Global Mount Points

The following example shows the node phys-schost-2 of the cluster suncluster missing the mount point /global/schost-1. Reports are being sent to the output directory, /var/cluster/sccheck/myReports/.


# sccheck -v1 -h phys-schost-1,phys-schost-2 -o /var/cluster/sccheck/myReports

sccheck: Requesting explorer data and node report from phys-schost-1.
sccheck: Requesting explorer data and node report from phys-schost-2.
sccheck: phys-schost-1: Explorer finished.
sccheck: phys-schost-1: Starting single-node checks.
sccheck: phys-schost-1: Single-node checks finished.
sccheck: phys-schost-2: Explorer finished.
sccheck: phys-schost-2: Starting single-node checks.
sccheck: phys-schost-2: Single-node checks finished.
sccheck: Starting multi-node checks.
sccheck: Multi-node checks finished.
sccheck: One or more checks failed.
sccheck: The greatest severity of all check failures was 3 (HIGH).
sccheck: Reports are in /var/cluster/sccheck/myReports.
# 
# cat /var/cluster/sccheck/myReports/sccheck-results.suncluster.txt

...
===================================================
= ANALYSIS DETAILS =
===================================================
------------------------------------
CHECK ID : 3065
SEVERITY : HIGH
FAILURE  : Global filesystem /etc/vfstab entries are not consistent across 
all Sun Cluster 3.x nodes.
ANALYSIS : The global filesystem /etc/vfstab entries are not consistent across 
all nodes in this cluster.
Analysis indicates:
FileSystem '/global/phys-schost-1' is on 'phys-schost-1' but missing from 'phys-schost-2'.
RECOMMEND: Ensure each node has the correct /etc/vfstab entry for the 
filesystem(s) in question.
...
#
# cat /var/cluster/sccheck/myReports/sccheck-results.phys-schost-1.txt

...
===================================================
= ANALYSIS DETAILS =
===================================================
------------------------------------
CHECK ID : 1398
SEVERITY : HIGH
FAILURE  : An unsupported server is being used as a Sun Cluster 3.x node.
ANALYSIS : This server may not been qualified to be used as a Sun Cluster 3.x node.  
Only servers that have been qualified with Sun Cluster 3.x are supported as 
Sun Cluster 3.x nodes.
RECOMMEND: Because the list of supported servers is always being updated, check with 
your Sun Microsystems representative to get the latest information on what servers 
are currently supported and only use a server that is supported with Sun Cluster 3.x.
...
#

ProcedureHow to View the Contents of Sun Cluster Command Logs

The /var/cluster/logs/commandlog ASCII text file contains records of selected Sun Cluster commands that are executed in a cluster. The logging of commands starts automatically when you set up the cluster and ends when you shut down the cluster. Commands are logged on all nodes that are up and booted in cluster mode.

Commands that are not logged in this file include those that display the configuration and current state of the cluster.

Commands that are logged in this file include those that configure and change the current state of the cluster:

Records in the commandlog file can contain the following elements:

By default, the commandlog file is regularly archived once a week. To change the archiving policies for the commandlog file, on each node in the cluster, use the crontab command. See the crontab(1) man page for more information.

Sun Cluster software maintains up to eight previously archived commandlog files on each cluster node at any given time. The commandlog file for the current week is named commandlog. The most recent complete week's file is named commandlog.0. The oldest complete week's file is named commandlog.7.

    View the contents of the current week's commandlog file, one screen at a time.


    # more /var/cluster/logs/commandlog
    

Example 1–10 Viewing the Contents of Sun Cluster Command Logs

The following example shows the contents of the commandlog file that are displayed by the more command.


more -lines10 /var/cluster/logs/commandlog
11/11/2006 09:42:51 phys-schost-1 5222 root START - clsetup
11/11/2006 09:43:36 phys-schost-1 5758 root START - clrg add "app-sa-1"
11/11/2006 09:43:36 phys-schost-1 5758 root END 0
11/11/2006 09:43:36 phys-schost-1 5760 root START - clrg set -y
"RG_description=Department Shared Address RG" "app-sa-1"
11/11/2006 09:43:37 phys-schost-1 5760 root END 0
11/11/2006 09:44:15 phys-schost-1 5810 root START - clrg online "app-sa-1"
11/11/2006 09:44:15 phys-schost-1 5810 root END 0
11/11/2006 09:44:19 phys-schost-1 5222 root END -20988320
12/02/2006 14:37:21 phys-schost-1 5542 jbloggs START - clrg -c -g "app-sa-1"
-y "RG_description=Joe Bloggs Shared Address RG"
12/02/2006 14:37:22 phys-schost-1 5542 jbloggs END 0