Sun Cluster 3.0 Hardware Guide

How to Install a StorEdge MultiPack

Use this procedure for an initial installation of a StorEdge MultiPack enclosure, prior to installing the Solaris operating environment and Sun Cluster software. Perform this procedure in conjunction with the procedures in Sun Cluster 3.0 Installation Guide and your server hardware manual.

Multihost storage in clusters uses the multi-initiator capability of the SCSI (Small Computer System Interface) specification. For conceptual information on multi-initiator capability, see Sun Cluster 3.0 Concepts.

  1. Ensure that each device in the SCSI chain has a unique SCSI address.

    The default SCSI address for host adapters is 7. Reserve SCSI address 7 for one host adapter in the SCSI chain. This procedure refers to the host adapter you choose for SCSI address 7 as the host adapter on the second node. To avoid conflicts, in Step 7 you will change the scsi-initiator-id of the remaining host adapter in the SCSI chain to an available SCSI address. This procedure refers to the host adapter with an available SCSI address as the host adapter on the first node. Depending on the device and configuration settings of the device, either SCSI address 6 or 8 is usually available.


    Caution - Caution -

    Even though a slot in the enclosure might not be in use, you should avoid setting the scsi-initiator-id for the first node to the SCSI address for that disk slot. This precaution minimizes future complications if you install additional disk drives.


    For more information, see the OpenBoot 3.x Command Reference Manual and the labels inside the storage device.

  2. Install the host adapters in the nodes that will be connected to the enclosure.

    For the procedure on installing host adapters, see the documentation that shipped with your host adapters and nodes.

  3. Connect the cables to the enclosure, as shown in Figure 4-1.

    Make sure that the entire SCSI bus length to each enclosure is less than 6 m. This measurement includes the cables to both nodes, as well as the bus length internal to each enclosure, node, and host adapter. Refer to the documentation that shipped with the enclosure for other restrictions regarding SCSI operation.

    Figure 4-1 Example of a StorEdge MultiPack enclosure mirrored pair

    Graphic

  4. Connect the AC power cord for each enclosure of the pair to a different power source.

  5. Without allowing the node to boot, power on the first node. If necessary, abort the system to continue with OpenBoot PROM Monitor tasks.

  6. Find the paths to the host adapters.


    {0} ok show-disks
    a) /pci@1f,4000/pci@4/SUNW,isptwo@4/sd
    b) /pci@1f,4000/pci@2/SUNW,isptwo@4/sd

    Identify and record the two controllers that will be connected to the storage devices, and record these paths. Use this information to change the SCSI addresses of these controllers in the nvramrc script. Do not include the /sd directories in the device paths.

  7. Edit the nvramrc script to set the scsi-initiator-id for the host adapters on the first node.

    For a list of nvramrc editor and nvedit keystroke commands, see Appendix B, NVRAMRC Editor and NVEDIT Keystroke Commands.

    The following example sets the scsi-initiator-id to 6. The OpenBoot PROM Monitor prints the line numbers (0:, 1:, and so on).


    Caution - Caution -

    Insert exactly one space after the first double quote and before scsi-initiator-id.



    {0} ok nvedit 
    0: probe-all
    1: cd /pci@1f,4000/pci@4/SUNW,isptwo@4
    2: 6 " scsi-initiator-id" integer-property 
    3: device-end 
    4: cd /pci@1f,4000/pci@2/SUNW,isptwo@4 
    5: 6 " scsi-initiator-id" integer-property 
    6: device-end 
    7: install-console 
    8: banner <Control C> 
    {0} ok
  8. Store the changes.

    The changes you make through the nvedit command are done on a temporary copy of the nvramrc script. You can continue to edit this copy without risk. After you complete your edits, save the changes. If you are not sure about the changes, discard them.

    • To store the changes, type:


      {0} ok nvstore
      {0} ok 

    • To discard the changes, type:


      {0} ok nvquit
      {0} ok 
  9. Verify the contents of the nvramrc script you created in Step 7.

    If the contents of the nvramrc script are incorrect, use the nvedit command to make corrections.


    {0} ok printenv nvramrc 
    nvramrc =             probe-all
                          cd /pci@1f,4000/pci@4/SUNW,isptwo@4
                          6 " scsi-initiator-id" integer-property 
                          device-end 
                          cd /pci@1f,4000/pci@2/SUNW,isptwo@4
                          6 " scsi-initiator-id" integer-property  
                          device-end  
                          install-console
                          banner
    {0} ok
  10. Instruct the OpenBoot PROM Monitor to use the nvramrc script.


    {0} ok setenv use-nvramrc? true
    use-nvramrc? = true
    {0} ok 
  11. Without allowing the node to boot, power on the second node. If necessary, abort the system to continue with OpenBoot PROM Monitor tasks.

  12. Verify that the scsi-initiator-id for the host adapter on the second node is set to 7.

    Use the show-disks command to find the paths to the host adapters connected to these enclosures. Select each host adapter's device tree node, and display the node's properties to confirm that the scsi-initiator-id for each host adapter is set to 7.


    {0} ok cd /pci@1f,4000/pci@4/SUNW,isptwo@4
    {0} ok .properties
    ...
    scsi-initiator-id        00000007
    ...
    {0} ok cd /pci@1f,4000/pci@2/SUNW,isptwo@4
    {0} ok .properties
    ...
    scsi-initiator-id        00000007
  13. Continue with the Solaris operating environment, Sun Cluster software, and volume management software installation tasks.

    For software installation procedures, see Sun Cluster 3.0 Installation Guide.