Sun Cluster System Administration Guide for Solaris OS

ProcedureHow to Verify That Replication Is Configured Correctly

Before You Begin

Complete the procedure How to Perform a Point-in-Time Snapshot.

  1. Access nodeA and nodeC as superuser or assume a role that provides solaris.cluster.admin RBAC authorization.

  2. Verify that the primary cluster is in replicating mode, with autosynchronization on.

    • For Sun StorEdge Availability Suite software:


      nodeA# /usr/opt/SUNWesm/sbin/sndradm -P
      
    • For Sun StorageTek Availability Suite software:


      nodeA# /usr/sbin/sndradm -P
      

    The output should resemble the following:


    /dev/vx/rdsk/devgrp/vol01 ->
    lhost-reprg-sec:/dev/vx/rdsk/devgrp/vol01
    autosync: on, max q writes:4194304, max q fbas:16384, mode:sync,ctag:
    devgrp, state: replicating

    In replicating mode, the state is replicating, and the active state of autosynchronization is on. When the primary volume is written to, the secondary volume is updated by Sun StorageTek Availability Suite software.

  3. If the primary cluster is not in replicating mode, put it into replicating mode.

    • For Sun StorEdge Availability Suite software:


      nodeA# /usr/opt/SUNWesm/sbin/sndradm -n -u lhost-reprg-prim \
      /dev/vx/rdsk/devgrp/vol01 \
      /dev/vx/rdsk/devgrp/vol04 lhost-reprg-sec \
      /dev/vx/rdsk/devgrp/vol01 \
      /dev/vx/rdsk/devgrp/vol04 ip sync
      
    • For Sun StorageTek Availability Suite software:


      nodeA# /usr/sbin/sndradm -n -u lhost-reprg-prim \
      /dev/vx/rdsk/devgrp/vol01 \
      /dev/vx/rdsk/devgrp/vol04 lhost-reprg-sec \
      /dev/vx/rdsk/devgrp/vol01 \
      /dev/vx/rdsk/devgrp/vol04 ip sync
      
  4. Create a directory on a client machine.

    1. Log in to a client machine as superuser.

      You see a prompt that resembles the following:


      client-machine#
    2. Create a directory on the client machine.


      client-machine# mkdir /dir
      
  5. Mount the directory to the application on the primary cluster, and display the mounted directory.

    1. Mount the directory to the application on the primary cluster.


      client-machine# mount -o rw lhost-nfsrg-prim:/global/mountpoint /dir
      
    2. Display the mounted directory.


      client-machine# ls /dir
      
  6. Mount the directory to the application on the secondary cluster, and display the mounted directory.

    1. Unmount the directory from the application on the primary cluster.


      client-machine# umount /dir
      
    2. Take the application resource group offline on the primary cluster.


      nodeA# clresource disable -n nodeA nfs-rs
      nodeA# clresource disable -n nodeA nfs-dg-rs
      nodeA# clresource disable -n nodeA lhost-nfsrg-prim
      nodeA# clresourcegroup online -n "" nfs-rg
      
    3. Change the primary cluster to logging mode.

      • For Sun StorEdge Availability Suite software:


        nodeA# /usr/sbin/sndradm -n -l lhost-reprg-prim \
        /dev/vx/rdsk/devgrp/vol01 \
        /dev/vx/rdsk/devgrp/vol04 lhost-reprg-sec \
        /dev/vx/rdsk/devgrp/vol01 \
        /dev/vx/rdsk/devgrp/vol04 ip sync
        
      • For Sun StorageTek Availability Suite software:


        nodeA# /usr/sbin/sndradm -n -l lhost-reprg-prim \
        /dev/vx/rdsk/devgrp/vol01 \
        /dev/vx/rdsk/devgrp/vol04 lhost-reprg-sec \
        /dev/vx/rdsk/devgrp/vol01 \
        /dev/vx/rdsk/devgrp/vol04 ip sync
        

      When the data volume on the disk is written to, the bitmap file on the same disk is updated. No replication occurs.

    4. Ensure that the PathPrefix directory is available.


      nodeC# mount | grep /global/etc
      
    5. Bring the application resource group online on the secondary cluster.


      nodeC# clresourcegroup online -n nodeC nfs-rg
      
    6. Access the client machine as superuser.

      You see a prompt that resembles the following:


      client-machine#
    7. Mount the directory that was created in Step 4 to the application on the secondary cluster.


      client-machine# mount -o rw lhost-nfsrg-sec:/global/mountpoint /dir
      
    8. Display the mounted directory.


      client-machine# ls /dir
      
  7. Ensure that the directory displayed in Step 5 is the same as the directory displayed in Step 6.

  8. Return the application on the primary cluster to the mounted directory.

    1. Take the application resource group offline on the secondary cluster.


      nodeC# clresource disable -n nodeC nfs-rs
      nodeC# clresource disable -n nodeC nfs-dg-rs
      nodeC# clresource disable -n nodeC lhost-nfsrg-sec
      nodeC# clresourcegroup online -n "" nfs-rg
      
    2. Ensure that the global volume is unmounted from the secondary cluster.


      nodeC# umount /global/mountpoint
      
    3. Bring the application resource group online on the primary cluster.


      nodeA# clresourcegroup online -n nodeA nfs-rg
      
    4. Change the primary cluster to replicating mode.

      • For Sun StorEdge Availability Suite software:


        nodeA# /usr/opt/SUNWesm/sbin/sndradm -n -u lhost-reprg-prim \
        /dev/vx/rdsk/devgrp/vol01 \
        /dev/vx/rdsk/devgrp/vol04 lhost-reprg-sec \
        /dev/vx/rdsk/devgrp/vol01 \
        /dev/vx/rdsk/devgrp/vol04 ip sync
        
      • For Sun StorageTek Availability Suite software:


        nodeA# /usr/sbin/sndradm -n -u lhost-reprg-prim \
        /dev/vx/rdsk/devgrp/vol01 \
        /dev/vx/rdsk/devgrp/vol04 lhost-reprg-sec \
        /dev/vx/rdsk/devgrp/vol01 \
        /dev/vx/rdsk/devgrp/vol04 ip sync
        

      When the primary volume is written to, the secondary volume is updated by Sun StorageTek Availability Suite software.

See Also

Example of How to Manage a Failover