Cluster Server 7.4 Agent for EMC SRDF Configuration Guide - Windows

Last Published:
Product(s): InfoScale & Storage Foundation (7.4.1)
Platform: Windows
  1. Introducing the agent for EMC SRDF
    1.  
      About the agent for EMC SRDF
    2.  
      Supported hardware for EMC SRDF
    3.  
      Supported software
    4.  
      Typical EMC SRDF setup in a VCS cluster
    5. EMC SRDF agent functions
      1.  
        About the EMC SRDF agent's online function
      2.  
        About dynamic swap support for the EMC SRDF agent
    6.  
      Installing the agent for EMC SRDF
  2. Configuring the agent for EMC SRDF
    1. Configuration concepts for the EMC SRDF agent
      1.  
        Resource type definition for the EMC SRDF agent
      2. Attribute definitions for the SRDF agent
        1.  
          Required attributes
        2.  
          Optional attributes
        3.  
          Internal attributes
      3.  
        Sample configuration for the EMC SRDF agent
    2. Before you configure the agent for EMC SRDF
      1.  
        About cluster heartbeats
      2.  
        About configuring system zones in replicated data clusters
      3.  
        About preventing split-brain
    3. Configuring the agent for EMC SRDF
      1. Configuring the agent manually in a global cluster
        1.  
          Configuring the Symm heartbeat on each cluster
      2.  
        Configuring the agent manually in a replicated data cluster
      3.  
        Setting the OnlineTimeout attribute for the SRDF resource
      4.  
        Additional configuration considerations for the SRDF agent
  3. Testing VCS disaster recovery support with EMC SRDF
    1. How VCS recovers from various disasters in an HA/DR setup with EMC SRDF
      1.  
        Failure scenarios in global clusters
      2.  
        Failure scenarios in replicated data clusters
    2.  
      Testing the global service group migration
    3.  
      Testing disaster recovery after host failure
    4.  
      Testing disaster recovery after site failure
    5.  
      Performing failback after a node failure or an application failure
    6.  
      Performing failback after a site failure
  4. Setting up fire drill
    1.  
      About fire drills
    2.  
      Fire drill configurations
    3. About the SRDFSnap agent
      1.  
        SRDFSnap agent functions
      2.  
        Resource type definition for the SRDFSnap agent
      3.  
        Attribute definitions for the SRDFSnap agent
      4.  
        About the Snapshot attributes
      5.  
        Sample configuration for a fire drill service group
    4.  
      Additional considerations for running a fire drill
    5.  
      Before you configure the fire drill service group
    6. Configuring the fire drill service group
      1.  
        About the Fire Drill wizard
    7.  
      Verifying a successful fire drill

Performing failback after a site failure

After a site failure at the primary site, the hosts and the storage at the primary site are down. VCS brings the global service group online at the secondary site and the EMC SRDF agent write enables the R2 devices.

The device state is PARTITIONED.

Review the details on site failure and how VCS and the agent for EMC SRDF behave in response to the failure.

See Failure scenarios in global clusters.

See Failure scenarios in replicated data clusters.

When the hosts and the storage at the primary site are restarted and the replication link is restored, the SRDF devices attain SPLIT state at both the sites. The devices are write-enabled at both sites. You can now perform a failback of the global service group to the primary site.

To perform failback after a site failure in global cluster

  1. Take the global service group offline at the secondary site. On a node at the secondary site, run the following command:
    hagrp -offline global_group -any
  2. Resync the devices using the symrdf restore command.

    The symrdf restore command write disables the devices at both the R1 and R2 sites.

    After the resync is complete, the device state is CONSISTENT or SYNCHRONIZED at both the sites.

    The devices are write-enabled at the primary site and write-disabled at the secondary site.

  3. Bring the global service group online at the primary site. On a node in the primary site, run the following command:
    hagrp -online global_group -any

    This again swaps the role of R1 and R2.

To perform failback after a site failure in replicated data cluster

  1. Take the global service group offline at the secondary site. On a node in the secondary site, run the following command:
    hagrp -offline service_group -sys sys_name
  2. Resync the devices using the symrdf restore command.

    The symrdf restore command write disables the devices at both the R1 and R2 sites.

    After the resync is complete, the device state is CONSISTENT or SYNCHRONIZED at both the sites. The devices are write-enabled at the primary site and write-disabled at the secondary site.

  3. Bring the global service group online at the primary site. On a node in the primary site, run the following command:
    hagrp -online service_group -sys sys_name

    This again swaps the role of R1 and R2.