Skip Navigation Links | |
Exit Print View | |
Oracle Solaris Cluster Software Installation Guide Oracle Solaris Cluster 3.3 3/13 |
1. Planning the Oracle Solaris Cluster Configuration
2. Installing Software on Global-Cluster Nodes
3. Establishing the Global Cluster
4. Configuring Solaris Volume Manager Software
5. Creating a Cluster File System
Creating Oracle ACFS File Systems
How to Register and Configure the Framework Resource Groups
How to Create an Oracle ACFS File System
How to Register and Configure the Scalable Device-Group Resource Group
How to Register and Configure the Oracle ASM Resource Group
How to Register and Configure the Oracle ACFS Proxy Resource Group
How to Create an Oracle Grid Infrastructure Resource for Interoperation With Oracle Solaris Cluster
6. Creating Non-Global Zones and Zone Clusters
This section provides procedures to create cluster file systems to support data services.
Perform this procedure for each cluster file system that you want to create. Unlike a local file system, a cluster file system is accessible from any node in the global cluster.
Before You Begin
Perform the following tasks:
Ensure that you installed software packages for the Oracle Solaris OS, Oracle Solaris Cluster framework, and other products as described in Installing the Software.
Ensure that you established the new cluster or cluster node as described in Establishing a New Global Cluster or New Global-Cluster Node.
If you are using Solaris Volume Manager software, ensure that volume-management software is configured. For procedures, see Configuring Solaris Volume Manager Software.
Determine the mount options to use for each cluster file system that you want to create. See Choosing Mount Options for UFS Cluster File Systems.
Perform this procedure from the global zone if non-global zones are configured in the cluster.
Tip - For faster file-system creation, become superuser on the current primary of the global device for which you create a file system.
Caution - Any data on the disks is destroyed when you create a file system. Be sure that you specify the correct disk device name. If you specify the wrong device name, you might erase data that you did not intend to delete. |
phys-schost# newfs raw-disk-device
The following table shows examples of names for the raw-disk-device argument. Note that naming conventions differ for each volume manager.
|
A mount point is required on each node, even if the cluster file system is not accessed on that node.
Tip - For ease of administration, create the mount point in the /global/device-group/ directory. This location enables you to easily distinguish cluster file systems, which are globally available, from local file systems.
phys-schost# mkdir -p /global/device-group/mountpoint/
Name of the directory that corresponds to the name of the device group that contains the device.
Name of the directory on which to mount the cluster file system.
See the vfstab(4) man page for details.
Note - If non-global zones are configured in the cluster, ensure that you mount cluster file systems in the global zone on a path in the global zone's root directory.
For example, consider the scenario where phys-schost-1 mounts disk device d0 on /global/oracle/, and phys-schost-2 mounts disk device d1 on /global/oracle/logs/. With this configuration, phys-schost-2 can boot and mount /global/oracle/logs/ only after phys-schost-1 boots and mounts /global/oracle/.
phys-schost# cluster check -k vfstab
The configuration check utility verifies that the mount points exist. The utility also verifies that /etc/vfstab file entries are correct on all nodes of the cluster. If no errors occur, nothing is returned.
For more information, see the cluster(1CL) man page.
For UFS and QFS, mount the cluster file system from any node in the cluster.
phys-schost# mount /global/device-group/mountpoint/
You can use either the df command or mount command to list mounted file systems. For more information, see the df(1M) man page or mount(1M) man page.
Cluster file systems are accessible from both the global zone and the non-global zone.
Example 5-1 Creating a UFS Cluster File System
The following example creates a UFS cluster file system on the Solaris Volume Manager volume /dev/md/oracle/rdsk/d1. An entry for the cluster file system is added to the vfstab file on each node. Then from one node the cluster check command is run. After configuration check processing is completes successfully, the cluster file system is mounted from one node and verified on all nodes.
phys-schost# newfs /dev/md/oracle/rdsk/d1 … phys-schost# mkdir -p /global/oracle/d1 phys-schost# vi /etc/vfstab #device device mount FS fsck mount mount #to mount to fsck point type pass at boot options # /dev/md/oracle/dsk/d1 /dev/md/oracle/rdsk/d1 /global/oracle/d1 ufs 2 yes global,logging … phys-schost# cluster check -k vfstab phys-schost# mount /global/oracle/d1 phys-schost# mount … /global/oracle/d1 on /dev/md/oracle/dsk/d1 read/write/setuid/global/logging/largefiles on Sun Oct 3 08:56:16 2005
Next Steps
Determine from the following list the next task to perform that applies to your cluster configuration. If you need to perform more than one task from this list, go to the first of those tasks in this list.
To create non-global zones on a node, go to How to Create a Non-Global Zone on a Global-Cluster Node.
Install third-party applications, register resource types, set up resource groups, and configure data services. See the documentation that is supplied with the application software and the Oracle Solaris Cluster Data Services Planning and Administration Guide.