Go to main content

Administering an Oracle® Solaris Cluster 4.4 Configuration

Exit Print View

Updated: November 2019

How to Remove a UFS Cluster File System

You remove a UFS cluster file system by merely unmounting it. To also remove or delete the data, remove the underlying disk device (or metadevice or volume) from the system.

Note -  UFS Cluster file systems are automatically unmounted as part of the system shutdown that occurs when you run cluster shutdown to stop the entire cluster. A UFS cluster file system is not unmounted when you run shutdown to stop a single node. However, if the node being shut down is the only node with a connection to the disk, any attempt to access the UFS cluster file system on that disk results in an error.

Ensure that the following prerequisites have been completed prior to unmounting UFS cluster file systems:

  • The root role privilege is established on a node in the cluster.

  • The file system is not busy. A file system is considered busy if a user is working in a directory in the file system, or if a program has a file open in that file system. The user or program could be running on any node in the cluster.

Note -  You can also remove a zone-cluster file system by using the Oracle Solaris Cluster Manager browser interface. For Oracle Solaris Cluster Manager log-in instructions, see How to Access Oracle Solaris Cluster Manager.
  1. Assume the root role on any node in the cluster.
  2. Determine which cluster file systems are mounted.
    # mount -v
  3. On each node, list all processes that are using the cluster file system, so that you know which processes you are going to stop.
    # fuser -c [ -u ] mountpoint

    Reports on files that are mount points for file systems and any files within those mounted file systems.


    (Optional) Displays the user login name for each process ID.


    Specifies the name of the cluster file system for which you want to stop processes.

  4. On each node, stop all processes for the cluster file system.

    Use your preferred method for stopping processes. If necessary, use the following command to force termination of processes associated with the cluster file system.

    # fuser -c -k mountpoint

    A SIGKILL is sent to each process that uses the cluster file system.

  5. On each node, verify that no processes are using the file system.
    # fuser -c mountpoint
  6. From just one node, unmount the file system.
    # umount mountpoint

    Specifies the name of the cluster file system you want to unmount. This can be either the directory name where the cluster file system is mounted, or the device name path of the file system.

  7. (Optional) Edit the /etc/vfstab file to delete the entry for the cluster file system being removed.

    Perform this step on each cluster node that has an entry for this cluster file system in its /etc/vfstab file.

  8. (Optional) Remove the disk device group/metadevice/volume/plex.

    See your volume manager documentation for more information.

Example 45  Removing a Cluster File System

The following example removes a UFS cluster file system that is mounted on the Solaris Volume Manager metadevice or volume/dev/md/oracle/rdsk/d1.

# mount -v
/global/oracle/d1 on /dev/md/oracle/dsk/d1 read/write/setuid/global/logging/largefiles 
# fuser -c /global/oracle/d1
/global/oracle/d1: 4006c
# fuser -c -k /global/oracle/d1
/global/oracle/d1: 4006c
# fuser -c /global/oracle/d1
# umount /global/oracle/d1

On each node, remove the highlighted entry
# pfedit /etc/vfstab
#device           device        mount   FS      fsck    mount   mount
#to mount         to fsck       point   type    pass    at boot options
/dev/md/oracle/dsk/d1 /dev/md/oracle/rdsk/d1 /global/oracle/d1 ufs 2 yes global,logging

Save and exit

To remove the data on the cluster file system, remove the underlying device. See your volume manager documentation for more information.