Sun Cluster 3.1 - 3.2 With Fibre Channel JBOD Storage Device Manual

ProcedureHow to Replace a Disk Drive

For conceptual information about quorum, quorum devices, global devices, and device IDs, see your Sun Cluster concepts documentation.

Before You Begin

This procedure assumes that your cluster is operational.

This procedure provides the long forms of the Sun Cluster commands. Most commands also have short forms. Except for the forms of the command names, the commands are identical. For a list of the commands and their short forms, see Appendix A, Sun Cluster Object-Oriented Commands, in Sun Cluster 3.1 - 3.2 Hardware Administration Manual for Solaris OS.

To perform this procedure, become superuser or assume a role that provides solaris.cluster.read and solaris.cluster.modify RBAC authorization.

  1. Identify the disk drive that needs replacement.

    If the disk error message reports the drive problem by device ID (DID), determine the Solaris logical device name. If the disk error message reports the drive problem by the Solaris physical device name, use your Solaris documentation to map the Solaris physical device name to the Solaris logical device name. Use this Solaris logical device name and DID throughout this procedure.

    • If you are using Sun Cluster 3.2, use the following command:


      # cldevice list -v 
      
    • If you are using Sun Cluster 3.1, use the following command:


      # scdidadm -l deviceID
      
  2. Determine whether the disk drive that you want to replace is configured as a quorum device.

    • If you are using Sun Cluster 3.2, use the following command:


      # clquorum show
      
    • If you are using Sun Cluster 3.1, use the following command:


      # scstat -q
      
  3. If the disk drive that you want to replace is configured as a quorum device, add a new quorum device on a different storage array. Remove the old quorum device.

    For procedures about how to add and remove quorum devices, see Sun Cluster system administration documentation.

  4. If possible, back up the metadevice or volume.

    For more information, see your Solstice DiskSuite/Solaris Volume Manager or VERITAS Volume Manager documentation.

  5. Identify the failed disk drive's physical DID.

    Use this physical DID in Step 14 to verify that the failed disk drive has been replaced with a new disk drive. The DID and the world wide name (WWN) for the disk drive are the same.

    • If you are using Sun Cluster 3.2, use the following command:


      # cldevice list -v
      
    • If you are using Sun Cluster 3.1, use the following command:


      # scdidadm -o diskid -l cNtXdY
      
  6. If you are using VERITAS Volume Manager, proceed to Step 8.

  7. If you are using Solstice DiskSuite/Solaris Volume Manager, save the disk partitioning information to partition the new disk drive.


    # prtvtoc /dev/rdsk/cNtXdYs2 > filename
    

    Note –

    You can also use the format utility to save the disk's partition information.


  8. On any node that is connected to the storage array, remove the disk drive when prompted.


    # luxadm remove_device -F /dev/rdsk/cNtXdYs2
    

    After running the command, warning messages might display. These messages can be ignored.

  9. On any node that is connected to the storage array, run the luxadm insert_device command. Add the new disk drive when prompted.


    # luxadm insert_device boxname,fslotnumber
    

    or


    # luxadm insert_device boxname,fslotnumber
    

    If you are inserting a front disk drive, use the fslotnumber parameter. If you are inserting a rear disk drive, use the rslotnumber parameter.

  10. On all other nodes that are attached to the storage array, probe all devices. Write the new disk drive to the /dev/rdsk directory.

    The amount of time that the devfsadm command requires to complete depends on the number of devices that are connected to the node. Expect at least five minutes.


    # devfsadm -C
    
  11. If you are using If VERITAS Volume Manager, proceed to Step 13.

  12. If you are using Solstice DiskSuite/Solaris Volume Manager, on one node that is connected to the storage array, partition the new disk drive. Use the partitioning information you saved in Step 6.


    # fmthard -s filename /dev/rdsk/cNtXdYs2
    

    Note –

    You can also use the format utility to partition the new disk drive.


  13. From all nodes that are connected to the storage array, update the DID database and driver.

    • If you are using Sun Cluster 3.2, use the following command:


      # cldevice repair
      
    • If you are using Sun Cluster 3.1, use the following command:


      # scdidadm -R deviceID
      

      Note –

      After running scdidadm —R on the first node, each subsequent node that you run the command on might display the warning, device id for the device matches the database. Ignore this warning.


  14. On any node, confirm that the failed disk drive has been replaced. Compare the following physical DID to the physical DID in Step 5.

    If the following physical DID is different from the physical DID in Step 5, you successfully replaced the failed disk drive with a new disk drive.

    • If you are using Sun Cluster 3.2, use the following command:


      # cldevice list -v
      
    • If you are using Sun Cluster 3.1, use the following command:


      # scdidadm -o diskid -l cNtXdY
      
  15. Perform volume management administration to add the disk drive back to its diskset or disk group.

    For more information, see your Solstice DiskSuite/Solaris Volume Manager or VERITAS Volume Manager documentation.

  16. If you want this new disk drive to be a quorum device, add the quorum device.

    For the procedure about how to add a quorum device, see Sun Cluster system administration documentation.