JavaScript is required to for searching.
Skip Navigation Links
Exit Print View
Oracle Solaris Cluster Software Installation Guide     Oracle Solaris Cluster 4.1
search filter icon
search icon

Document Information

Preface

1.  Planning the Oracle Solaris Cluster Configuration

2.  Installing Software on Global-Cluster Nodes

3.  Establishing the Global Cluster

4.  Configuring Solaris Volume Manager Software

5.  Creating a Cluster File System

Creating Cluster File Systems

How to Create Cluster File Systems

6.  Creating Zone Clusters

7.  Uninstalling Software From the Cluster

Index

Creating Cluster File Systems

This section provides procedures to create cluster file systems to support data services.

How to Create Cluster File Systems

Perform this procedure for each cluster file system that you want to create. Unlike a local file system, a cluster file system is accessible from any node in the global cluster.

Before You Begin

Perform the following tasks:

  1. Assume the root role on any node in the cluster.

    Tip - For faster file system creation, become the root role on the current primary of the global device for which you create a file system.


  2. Create a UFS file system by using the newfs command.

    Caution

    Caution - Any data on the disks is destroyed when you create a file system. Be sure that you specify the correct disk device name. If you specify the wrong device name, you might erase data that you did not intend to delete.


    phys-schost# newfs raw-disk-device

    The following table shows examples of names for the raw-disk-device argument. Note that naming conventions differ for each volume manager.


    Volume Manager
    Sample Disk Device Name
    Description
    Solaris Volume Manager
    /dev/md/nfs/rdsk/d1
    Raw disk device d1 within the nfs disk set
    None
    /dev/global/rdsk/d1s3
    Raw disk device d1s3
  3. On each node in the cluster, create a mount-point directory for the cluster file system.

    A mount point is required on each node, even if the cluster file system is not accessed on that node.


    Tip - For ease of administration, create the mount point in the /global/device-group/ directory. This location enables you to easily distinguish cluster file systems, which are globally available, from local file systems.


    phys-schost# mkdir -p /global/device-group/mount-point/
    device-group

    Name of the directory that corresponds to the name of the device group that contains the device.

    mount-point

    Name of the directory on which to mount the cluster file system.

  4. On each node in the cluster, add an entry to the /etc/vfstab file for the mount point.

    See the vfstab(4) man page for details.

    1. In each entry, specify the required mount options for the type of file system that you use.
    2. To automatically mount the cluster file system, set the mount at boot field to yes.
    3. For each cluster file system, ensure that the information in its /etc/vfstab entry is identical on each node.
    4. Ensure that the entries in each node's /etc/vfstab file list devices in the same order.
    5. Check the boot order dependencies of the file systems.

      For example, consider the scenario where phys-schost-1 mounts disk device d0 on /global/oracle/ and phys-schost-2 mounts disk device d1 on /global/oracle/logs/. With this configuration, phys-schost-2 can boot and mount /global/oracle/logs/ only after phys-schost-1 boots and mounts /global/oracle/.

  5. On any node in the cluster, run the configuration check utility.
    phys-schost# cluster check -k vfstab

    The configuration check utility verifies that the mount points exist. The utility also verifies that /etc/vfstab file entries are correct on all nodes of the cluster. If no errors occur, no output is returned.

    For more information, see the cluster(1CL) man page.

  6. Mount the cluster file system from any node in the cluster.
    phys-schost# mount /global/device-group/mountpoint/
  7. On each node of the cluster, verify that the cluster file system is mounted.

    You can use either the df command or mount command to list mounted file systems. For more information, see the df(1M) man page or mount(1M) man page.

Example 5-1 Creating a UFS Cluster File System

The following example creates a UFS cluster file system on the Solaris Volume Manager volume /dev/md/oracle/rdsk/d1. An entry for the cluster file system is added to the vfstab file on each node. Then from one node the cluster check command is run. After configuration check processing is completed successfully, the cluster file system is mounted from one node and verified on all nodes.

phys-schost# newfs /dev/md/oracle/rdsk/d1
…
phys-schost# mkdir -p /global/oracle/d1
phys-schost# vi /etc/vfstab
#device           device        mount   FS      fsck    mount   mount
#to mount         to fsck       point   type    pass    at boot options
#                     
/dev/md/oracle/dsk/d1 /dev/md/oracle/rdsk/d1 /global/oracle/d1 ufs 2 yes global,logging
…
phys-schost# cluster check -k vfstab
phys-schost# mount /global/oracle/d1
phys-schost# mount
…
/global/oracle/d1 on /dev/md/oracle/dsk/d1 read/write/setuid/global/logging/largefiles
on Sun Oct 3 08:56:16 2005

Next Steps

To find out how to install third-party applications, register resource types, set up resource groups, and configure data services, see the documentation that is supplied with the application software and the Oracle Solaris Cluster Data Services Planning and Administration Guide.