You remove a cluster file system by merely unmounting it. If you want to also remove or delete the data, remove the underlying disk device (or metadevice or volume) from the system.
Cluster file systems are automatically unmounted as part of the system shutdown that occurs when you run scshutdown(1M) to stop the entire cluster. A cluster file system is not unmounted when you run shutdown to stop a single node. However, if the node being shut down is the only node with a connection to the disk, any attempt to access the cluster file system on that disk results in an error.
The prerequisites to unmount cluster file systems are:
Superuser privilege on a node in the cluster.
The file system cannot be busy. A file system is considered busy if a user is in a directory in the file system, or if a program has a file open in that file system. The user or program could be running on any node in the cluster.
Become superuser on any node in the cluster.
Determine which cluster file systems are mounted.
# mount -v
On each node, list all processes that are using the cluster file system, so you know which processes you are going to stop.
# fuser -c [ -u ] mountpoint
Reports on files that are mount points for file systems and any files within those mounted file systems.
(Optional) Displays the user login name for each process ID.
Specifies the name of the cluster file system for which you want to stop processes.
On each node, stop all processes for the cluster file system.
Use your preferred method for stopping processes. If necessary, use the following command to force termination of processes associated with the cluster file system.
# fuser -c -k mountpoint
A SIGKILL is sent to each process using the cluster file system.
On each node, verify that no processes are using the file system.
# fuser -c mountpoint
From just one node, umount the file system.
# umount mountpoint
Specifies the name of the cluster file system you want to unmount. This can be either the directory name where the cluster file system is mounted, or the device name path of the file system.
(Optional) Edit the /etc/vfstab file to delete the entry for the cluster file system being removed.
Perform this step on each cluster node that has an entry for this cluster file system in its /etc/vfstab file.
(Optional) Remove the disk device group/metadevice/volume/plex .
See your volume manager documentation for more information.
The following example removes a UFS cluster file system mounted on the Solstice DiskSuite/Solaris Volume Manager metadevice or volume/dev/md/oracle/rdsk/d1.
# mount -v ... /global/oracle/d1 on /dev/md/oracle/dsk/d1 read/write/setuid/global/logging/largefiles # fuser -c /global/oracle/d1 /global/oracle/d1: 4006c # fuser -c -k /global/oracle/d1 /global/oracle/d1: 4006c # fuser -c /global/oracle/d1 /global/oracle/d1: # umount /global/oracle/d1 (on each node, remove the highlighted entry:) # vi /etc/vfstab #device device mount FS fsck mount mount #to mount to fsck point type pass at boot options # /dev/md/oracle/dsk/d1 /dev/md/oracle/rdsk/d1 /global/oracle/d1 ufs 2 yes global,logging [Save and exit.]
To remove the data on the cluster file system, remove the underlying device. See your volume manager documentation for more information.