InfoScale™ 9.0 Storage Foundation and High Availability Solutions Microsoft Clustering Solutions Guide for Microsoft SQL Server - Windows

Last Published:
Product(s): InfoScale & Storage Foundation (9.0)
Platform: Windows
  1. Introducing SFW solutions for a Microsoft cluster
    1.  
      About Microsoft clustering solutions with SFW
    2.  
      Advantages of using SFW in a Microsoft cluster
    3.  
      About high availability clusters
    4.  
      About campus clusters
    5.  
      About disaster recovery clusters
  2. Planning for deploying SQL Server with SFW in a Microsoft cluster
    1.  
      InfoScale requirements for Microsoft clustering solutions
    2. Planning your SQL Server high availability configuration
      1.  
        Sample high availability configuration for SQL Server with SFW
      2.  
        Configuring the quorum device for high availability
    3. Planning your campus cluster configuration
      1.  
        Microsoft campus cluster failure scenarios
      2. Microsoft cluster quorum and quorum arbitration
        1.  
          Quorum
        2.  
          Cluster ownership of the quorum resource
        3.  
          The vxclus utility
    4. Planning your disaster recovery configuration
      1.  
        Sample disaster recovery configuration for SQL Server with SFW and Volume Replicator
  3. Workflows for deploying SQL Server with SFW in a Microsoft cluster
    1.  
      Workflow for a high availability (HA) configuration
    2. Workflow for a campus cluster configuration
      1.  
        Campus cluster: Connecting the two nodes
    3.  
      Workflow for a disaster recovery configuration
    4.  
      Using the Solutions Configuration Center workflow
    5.  
      Configuring the storage hardware and network
  4. Configuring SFW storage
    1.  
      Tasks for configuring InfoScale Storage
    2. Planning for SFW cluster disk groups and volumes
      1.  
        Sample SQL Server high-availability cluster storage configuration
      2.  
        Sample campus cluster storage configuration
      3.  
        Sample SQL Server disaster recovery storage configuration
    3.  
      Considerations when creating disk groups and volumes for a campus cluster
    4.  
      Considerations when creating volumes for a DR configuration using Volume Replicator replication
    5.  
      Viewing the available disk storage
    6.  
      Creating dynamic cluster disk groups
    7.  
      Adding disks to campus cluster sites
    8.  
      Creating dynamic volumes for high availability clusters
    9.  
      Creating dynamic volumes for campus clusters
  5. Implementing a dynamic mirrored quorum resource
    1.  
      Tasks for implementing a dynamic mirrored quorum resource
    2.  
      Creating a dynamic cluster disk group and a mirrored volume for the quorum resource
    3.  
      Adding a Volume Manager Disk Group resource for the quorum
    4.  
      Changing the quorum resource to a dynamic mirrored quorum resource
  6. Installing SQL Server and configuring resources
    1.  
      Tasks for installing and configuring SQL Server
    2.  
      Creating the resource group for the SQL Server instance
    3.  
      Prerequisites for installing SQL Server
    4.  
      Installing SQL Server in an InfoScale Storage environment
    5.  
      Dependency graph for SQL Server
    6.  
      Verifying the SQL Server group in the Microsoft cluster
  7. Configuring disaster recovery
    1.  
      Tasks for configuring the secondary site for disaster recovery for SQL Server
    2.  
      Verifying the primary site configuration
    3.  
      Creating a parallel environment for SQL Server on the secondary site
    4.  
      Volume Replicator components overview
    5.  
      Setting up security for Volume Replicator
    6.  
      Creating resources for Volume Replicator
    7. Configuring Volume Replicator: Setting up an RDS
      1.  
        Prerequisites for setting up the RDS
      2.  
        Creating a Replicated Data Set (RDS)
    8.  
      Creating the RVG resource
    9.  
      Setting the SQL server resource dependency on the RVG resource
    10. Normal Volume Replicator operations and recovery procedures
      1.  
        Monitoring the status of the replication
      2.  
        Performing planned migration
      3. Replication recovery procedures
        1.  
          Bringing up the application on the secondary host
        2.  
          Restoring the primary host
  8. Appendix A. Configure InfoScale Storage in an existing Microsoft Failover Cluster
    1.  
      Configuring InfoScale Storage in an existing Microsoft Failover Cluster

Cluster ownership of the quorum resource

The Microsoft clustering challenge/defense protocol uses a low-level bus reset of the SCSI buses between the computers to attempt to gain control of the quorum resource.

After a SCSI bus reset, the reservation that each server had been holding on the quorum disk is lost. Each server has about 10 seconds to re-establish that reservation, which would in turn let the other servers know that it is still functioning, even though the other servers would not necessarily be able to communicate with it.

If the active cluster server does not re-establish the SCSI reservation on the quorum resource within the time limit, the applications that were on the server transfer to the server that establishes the SCSI reservation first. The new server servicing the application may now be a bit slower, but clients still get their applications serviced. The Internet Protocol (IP) address and network names move, applications are reconstituted according to the defined dependencies, and clients are still serviced, without any question as to the state of the cluster.

The challenge/defense protocol is more complex when the quorum device is a volume in a Storage Foundation disk group. For a server to take ownership of the disk group containing the cluster quorum device, SFW on that server must successfully import the disk group, obtaining SCSI reservations on more than half of its disks.

Because a campus cluster configuration has an even number of disks on each site, failover cannot occur automatically. After a site failure, you must use the manual CLI command vxclus enable to bring the cluster disk groups online on the secondary node.