Sun Cluster 3.1 - 3.2 With Sun StorEdge T3 or T3+ Array Manual for Solaris OS

ProcedureHow to Remove a Storage Array in a Single-Controller Configuration

Use this procedure to remove a storage array and its submirrors from a running cluster. This procedure provides the flexibility to remove the host adapters from the nodes for the storage array that you are removing. To remove a partner group from the cluster, see How to Remove a Partner Group.


Note –

When you upgrade firmware on a storage device or on an enclosure, redefine the stripe size of a LUN, or perform other LUN operations, a device ID might change unexpectedly. When you perform a check of the device ID configuration by running the cldevice check or scdidadm -c command, the following error message appears on your console if the device ID changed unexpectedly.


device id for nodename:/dev/rdsk/cXtYdZsN does not match physical 
device's id for ddecimalnumber, device may have been replaced.

To fix device IDs that report this error, run the cldevice repair or scdidadm -R command for each affected device.


This procedure defines Node A as the node with which you begin working. Node B is another node in the cluster.


Caution – Caution –

During this procedure, you lose access to the data that resides on the storage array that you are removing.


This procedure provides the long forms of the Sun Cluster commands. Most commands also have short forms. Except for the forms of the command names, the commands are identical. For a list of the commands and their short forms, see Appendix A, Sun Cluster Object-Oriented Commands, in Sun Cluster 3.1 - 3.2 Hardware Administration Manual for Solaris OS.

Before You Begin

To perform this procedure, become superuser or assume a role that provides solaris.cluster.read and solaris.cluster.modify role-based access control (RBAC) authorization.

  1. Back up all database tables, data services, and volumes that are associated with the storage array. This storage array is the storage array you are removing.

  2. Detach the submirrors from the storage array that you are removing. Detach the submirrors to stop all I/O activity to the storage array.

    For more information, see your Solaris Volume Manager or Veritas Volume Manager documentation.

  3. Remove the references to the LUN(s) from any diskset or disk group.

    For more information, see your Solaris Volume Manager or Veritas Volume Manager documentation.

  4. Determine the resource groups and device groups that are running on Node B.

    Record this information because you use this information in Step 18 and Step 19 of this procedure to return resource groups and device groups to these nodes.

    • If you are using Sun Cluster 3.2, use the following commands:


      # clresourcegroup status + 
      # cldevicegroup status +
      
    • If you are using Sun Cluster 3.1, use the following command:


      # scstat
      
  5. Shut down Node A.

    For the procedure about how to shut down and power off a node, see your Sun Cluster system administration documentation.

  6. If the storage array that you are removing is the last storage array that is connected to Node A, disconnect the fiber-optic cable between Node A and the FC hub/switch that is connected to this storage array. Afterward, disconnect the fiber-optic cable between the FC hub/switch and this storage array.

    If this is not the last storage array. skip this step.

    For the procedure about how to remove a fiber-optic cable, see the Sun StorEdge T3 and T3+ Array Configuration Guide.


    Note –

    If you are use your storage array in a SAN-configured cluster, you must keep two FC switches configured in parallel. This configuration maintains cluster availability. See SAN Solutions in a Sun Cluster Environment in Sun Cluster 3.1 - 3.2 Hardware Administration Manual for Solaris OS for more information.


  7. If you want to remove the host adapter from Node A, power off the node.

    If you do not want to remove host adapters, skip to Step 10.

  8. Remove the host adapter from Node A.

    For the procedure about how to remove host adapters, see the documentation that shipped with your nodes.

  9. Without enabling the node to boot, power on Node A.

    For more information, see your Sun Cluster system administration documentation.

  10. Boot Node A into cluster mode.

    For more information about how to boot nodes, see your Sun Cluster system administration documentation.

  11. Shut down Node B.

    For the procedure about how to shut down and power off a node, see your Sun Cluster system administration documentation.

  12. If the storage array that you are removing is the last storage array that is connected to the FC hub/switch, disconnect the fiber-optic cable that connects this FC hub/switch and Node B.

    For the procedure about how to remove a fiber-optic cable, see the Sun StorEdge T3 and T3+ Array Configuration Guide.


    Note –

    If you are use your storage array in a SAN-configured cluster, you must keep two FC switches configured in parallel. This configuration maintains cluster availability. See SAN Solutions in a Sun Cluster Environment in Sun Cluster 3.1 - 3.2 Hardware Administration Manual for Solaris OS for more information.


  13. If you want to remove the host adapter from Node B, power off the node.

    If you do not want to remove host adapters, skip to Step 16.

  14. Remove the host adapter fromNode B.

    For the procedure about how to remove host adapters, see the documentation that shipped with your nodes.

  15. Without enabling the node to boot, power on Node B

    For more information, see your Sun Cluster system administration documentation.

  16. Boot Node B into cluster mode.

    For more information about how to boot nodes, see your Sun Cluster system administration documentation.

  17. On all nodes, update the /devices and /dev entries.

    • If you are using Sun Cluster 3.2, use the following command:


      # devfsadm -C
      # cldevice clear
      
    • If you are using Sun Cluster 3.1, use the following command:


      # devfsadm -C
      # scdidadm -C
      
  18. (Optional) Restore the device groups to the original node.

    Perform the following step for each device group you want to return to the original node.

    • If you are using Sun Cluster 3.2, use the following command:


      # cldevicegroup switch -n nodenamedevicegroup1[ devicegroup2 …]
      
      -n nodename

      The node to which you are restoring device groups.

      devicegroup1[ devicegroup2 …]

      The device group or groups that you are restoring to the node.

    • If you are using Sun Cluster 3.1, use the following command:


      # scswitch -z -D devicegroup -h nodename
      
  19. (Optional) Restore the resource groups to the original node.

    Perform the following step for each resource group you want to return to the original node.

    • If you are using Sun Cluster 3.2, use the following command:


      # clresourcegroup switch -n nodename resourcegroup1[ resourcegroup2 …]
      
      nodename

      For failover resource groups, the node to which the groups are returned. For scalable resource groups, the node list to which the groups are returned.

      resourcegroup1[ resourcegroup2 …]

      The resource group or groups that you are returning to the node or nodes.

    • If you are using Sun Cluster 3.1, use the following command:


      # scswitch -z -g resourcegroup -h nodename