Veritas InfoScale™ 7.3.1 Virtualization Guide - Solaris

Last Published:
Product(s): InfoScale & Storage Foundation (7.3.1)
Platform: Solaris
  1. Section I. Overview of Veritas InfoScale Solutions used in Solaris virtualization
    1. Overview of Veritas InfoScale Virtualization Solutions
      1.  
        Overview of the Veritas InfoScale Products Virtualization Guide
      2. About Veritas InfoScale support for Solaris virtualization technology
        1. About SmartIO in the Solaris virtualized environments
          1.  
            Performing live migration between LDOMs in the SmartIO environment
      3.  
        Virtualization use cases addressed by Veritas InfoScale products
  2. Section II. Zones and Projects
    1. Storage Foundation and High Availability Solutions support for Solaris Zones
      1.  
        About Solaris Zones
      2. About VCS support for zones
        1.  
          Overview of how VCS works with zones
        2.  
          About the ContainerInfo service group attribute
        3.  
          About the ContainerOpts resource type attribute
        4.  
          About the ResContainerInfo resource type attribute
        5.  
          Zone-aware resources
        6. About the Mount agent
          1.  
            Sample configurations for Physical-to-Virtual (P2V) scenarios
          2.  
            Bringing a Mount resource online in the zone
          3.  
            Setting the attribute values for a Mount resource for NFS mounts
          4.  
            Configuring a direct mount of VxFS file system in a non-global zone with VCS
        7.  
          About networking agents
        8.  
          About the Zone agent
        9.  
          About configuring failovers among physical and virtual servers
      3. Configuring VCS in zones
        1. Prerequisites for configuring VCS in zones
          1.  
            Method for file system access inside non-global zone
          2.  
            Using custom agents in zones
        2. Deciding on the zone root location
          1.  
            Creating a zone with root on local disk
          2.  
            Creating a zone with root on shared storage
        3.  
          Performing the initial internal zone configuration
        4.  
          About installing applications in a zone
        5. Configuring the service group for the application
          1.  
            Resource dependency diagrams: zone root on local disks
          2.  
            Resource dependency diagrams: zone root on shared disks
        6.  
          Configuring a zone resource in a failover service group with the hazonesetup utility
        7.  
          Configuring zone resource in a parallel service group with the hazonesetup utility
        8.  
          Configuring multiple zone resources using same VCS user for password less communication
        9.  
          Modifying the service group configuration
        10.  
          Verifying the zone configuration
        11.  
          Synchronizing the zone configuration across cluster nodes
        12.  
          Performing maintenance tasks
        13.  
          Troubleshooting zones
        14.  
          Configuring for physical to virtual and virtual to physical failovers - a typical setup
      4.  
        Adding VxFS file systems to a non-global zone
      5.  
        Mounting VxFS as lofs into a non-global zone
      6.  
        Mounting VxFS directly into a non-global zone from global zone
      7.  
        Mounting VxFS as VxFS inside a non-global zone
      8.  
        Adding a direct mount to a zone's configuration
      9.  
        Benefits of a VxFS mount in a non-global zone over VxFS mount from global zone into the non-global zone
      10.  
        SFCFSHA mounts
      11.  
        Concurrent I/O access in non-global zones
      12.  
        Veritas Extension for Oracle Disk Manager
      13. Exporting VxVM volumes to a non-global zone
        1.  
          VxVM devices in Oracle Solaris global zones
        2.  
          Removing a VxVM volume from a non-global zone
      14. About SF Oracle RAC support for Oracle RAC in a zone environment
        1.  
          Supported configuration
        2. Known issues with supporting SF Oracle RAC in a zone environment
          1.  
            CFS mount agent does not support mounting VxVM devices inside non-global zones
          2.  
            Issue with VCS agents
          3.  
            Stopping non-global zones configured with direct-mount file systems from outside VCS causes the corresponding zone resource to fault or go offline
          4.  
            Error message displayed for PrivNIC resource if zone is not running
          5.  
            Warning messages displayed when VCS restarts
          6.  
            The installer log of non-global zone contains warning messages
          7.  
            Issue with CFS mounts
      15.  
        Configuring Solaris non-global zones for disaster recovery
      16. Software limitations of Storage Foundation support of non-global zones
        1.  
          Administration commands are not supported in non-global zone
        2.  
          VxFS file system is not supported as the root of a non-global zone
        3.  
          QIO and CQIO are not supported
        4.  
          Package installation in non-global zones
        5.  
          Package removal with non-global zone configurations
        6.  
          Root volume cannot be added to non-global zones
        7.  
          Some Veritas Volume Manager operations can cause volume device names to go out of sync
    2. Storage Foundation and High Availability Solutions support for Solaris Projects
      1.  
        About Solaris Projects
      2. About VCS support for Solaris projects
        1. Overview of how VCS works with Solaris projects
          1.  
            How VCS models containers
          2.  
            Installing and configuring projects in a VCS environment
          3.  
            Configuring the ContainerInfo attribute
        2.  
          About the ContainerInfo service group attribute
        3.  
          About the ContainerOpts resource type attribute
        4.  
          Project-aware resources
        5.  
          About the Project agent
      3. Configuring VCS in Solaris projects
        1. Prerequisites for configuring VCS in projects
          1.  
            Using custom agents in projects
  3. Section III. Oracle VM Server for SPARC
    1. Storage Foundation and High Availability Solutions support for Oracle VM Server for SPARC
      1.  
        About Oracle VM Server for SPARC
      2.  
        Terminology for Oracle VM Server for SPARC
      3. Oracle VM Server for SPARC deployment models
        1.  
          Split Storage Foundation stack
        2.  
          Guest-based Storage Foundation stack
        3.  
          Layered Storage Foundation stack
      4. Benefits of deploying Storage Foundation High Availability solutions in Oracle VM server for SPARC
        1.  
          Standardization of tools
        2.  
          Array migration
        3.  
          Moving storage between physical and virtual environments
        4.  
          Boot Image Management
      5. Features
        1. Storage Foundation features
          1. The vxloadm utility enables access to a file system contained in a VxVM volume from the Control Domain
            1.  
              Examples of using the vxloadm utility
          2. The vxformat utility automatically relabels the virtual disk backed by a VxVM volume in the guest domain
            1.  
              Example of using the vxformat utility
        2. Oracle VM Server for SPARC features
          1.  
            Guest domain migration
          2.  
            Virtual I/O dynamic reconfiguration
      6. Split Storage Foundation stack model
        1.  
          How Storage Foundation and High Availability Solutions works in the Oracle VM Server for SPARC
        2.  
          Storage Foundation features restrictions
      7. Guest-based Storage Foundation stack model
        1.  
          How Storage Foundation and High Availability Solutions works in the guest domains
        2. About SFCFSHA in an Oracle VM Server for SPARC environment
          1.  
            Supported configurations with SFCFSHA and multiple I/O Domains
          2.  
            SFCFSHA limitations
        3. Storage Foundation features restrictions
          1.  
            Veritas Volume Replicator bunker replication
          2.  
            Mirroring across controllers using vxassist the mirror=ctlr option
          3.  
            DMP SCSI bypass
          4.  
            Event Source Daemon (vxesd) fabric monitoring capabilities
          5.  
            Physical WWN for a path
      8. Layered Storage Foundation stack model
        1. DMP as a multi-pathing solution in an Oracle VM Server for SPARC environment
          1.  
            Enabling DMP in the control and alternate I/O domains
          2.  
            Enabling DMP path failover in the guest domain
          3.  
            Enabling I/O fencing in the guest domain
        2.  
          How Storage Foundation and High Availability Solutions works in the Oracle VM Server for SPARC
        3.  
          Storage Foundation features restrictions
      9. System requirements
        1.  
          Hardware requirements
      10.  
        Product release notes
      11.  
        Product licensing
      12. Installing Storage Foundation in a Oracle VM Server for SPARC environment
        1.  
          Installing and configuring Oracle VM Server for SPARC and domains
        2. Installing Storage Foundation in the control domain or guest
          1.  
            Installing the split Storage Foundation stack model
          2.  
            Installing the guest-based Storage Foundation stack model
        3.  
          Installing Veritas File System in the guest domain
        4.  
          Verifying the configuration
      13.  
        Exporting a Veritas volume to a guest domain from the control domain
      14. Provisioning storage for a guest domain
        1.  
          Provisioning Veritas Volume Manager volumes as data disks for guest domains
        2. Provisioning Veritas Volume Manager volumes as boot disks for guest domains
          1. Using VxVM mirrored volumes as boot devices for Ldoms
            1.  
              Provisioning boot disks for a guest domain
            2.  
              Using VxVM snapshot as a backup copy of the boot image during the upgrade
            3.  
              Restore the original environment
      15.  
        Using Veritas Volume Manager snapshots for cloning logical domain boot disks
      16.  
        Support of live migration for Solaris LDOMs with fencing configured in DMP mode
      17.  
        Configuring Oracle VM Server for SPARC guest domains for disaster recovery
      18. Software limitations
        1.  
          When an I/O domain fails, the vxdisk scandisks or vxdctl enable command take a long time to complete (2791127)
        2.  
          Resizing a Veritas Volume Manager volume (exported as a slice or full disk) does not dynamically reflect the new size of the volume in the guest
      19. Known issues
        1. Guest-based known issues
          1.  
            Encapsulating a non-scsi disk may fail (1745409)
          2.  
            Guest domain node shows only one PGR key instead of two after rejecting the other node in the cluster
          3.  
            Disk paths intermittently go offline while performing I/O on a mirrored volume
        2. Split Storage Foundation stack known issues
          1.  
            Caching of data writes on the backend volume in the service domain
          2.  
            A volume can become inaccessible from the guest in the event of control domain reboot
    2. Cluster Server support for using CVM with multiple nodes in a Oracle VM Server for SPARC environment
      1.  
        Clustering using Cluster Volume Manager
      2. Installing Storage Foundation on multiple nodes in a Logical Domain
        1. Reconfiguring the clustering agents for Cluster Volume Manager
          1.  
            Removing the vxfsckd resource
          2.  
            Creating CVMVolDg in a group
      3. Cluster Volume Manager in the control domain for providing high availability
        1.  
          Provisioning storage to guests with Flexible Storage Sharing volumes of control domain
    3. VCS: Configuring Oracle VM Server for SPARC for high availability
      1. About VCS in a Oracle VM Server for SPARC environment
        1.  
          Benefits of using VCS in Oracle VM server for SPARC environments to manage logical domains
        2.  
          Dynamic reconfiguration of memory and CPU of a guest domain
        3.  
          AdaptiveHA for logical domains
        4.  
          Cluster Server requirements
        5.  
          Cluster Server limitations
        6. Cluster Server known issues
          1.  
            Shutting down the control domain may cause the guest domain to crash (1631762)
      2. About Cluster Server configuration models in an Oracle VM Server for SPARC environment
        1. Cluster Server setup to fail over a logical domain on a failure of logical domain
          1. Configuration scenarios
            1.  
              Network configuration
            2. Storage configurations
              1.  
                Veritas Volume Manager exposed volumes
              2.  
                Image files
            3.  
              Configuring logical domain
            4.  
              Installing Cluster Server inside the control domain
          2.  
            Creating the VCS service groups for logical domains
          3. Verifying a logical domain service group failover
            1.  
              Using hagrp -switch command
            2.  
              Other verification scenarios
        2. Cluster Server setup to fail over an Application running inside logical domain on a failure of Application
          1.  
            Configuring VCS to fail over an application on a failure
      3. Oracle VM Server for SPARC guest domain migration in VCS environment
        1.  
          Overview of a warm migration
        2. Overview of a live migration
          1. VCS initiated migration of service groups that have LDom resource configured to monitor Oracle VM for SPARC guest domains
            1.  
              Troubleshooting migration-related issues
          2.  
            User initiated migration of Oracle VM guest domains managed by VCS
        3.  
          Prerequisites before you perform domain migration
        4.  
          Supported deployment models for Oracle VM Server for SPARC domain migration with VCS
        5.  
          Migrating Oracle VM guest when VCS is installed in the control domain that manages the guest domain
        6.  
          Migrating Oracle VM guest when VCS is installed in the control domain and single-node VCS is installed inside the guest domain to monitor applications inside the guest domain
        7.  
          Migrating Oracle VM guest when VCS cluster is installed in the guest domains to manage applications for Oracle VM Server for SPARC version 2.1 and above
        8.  
          Migrating Oracle VM guest when VCS cluster is installed in the guest domains to manage applications for Oracle VM Server for SPARC version 2.0
      4. About configuring VCS for Oracle VM Server for SPARC with multiple I/O domains
        1.  
          About Alternate I/O domain
        2.  
          Setting up the Alternate I/O domain
        3.  
          Configuring VCS to manage a Logical Domain with multiple I/O domains
      5. Configuring VCS to manage a Logical Domain using services from multiple I/O domains
        1.  
          A typical setup for a Logical Domain with multiple I/O services
        2.  
          Identify supported storage and network services
        3.  
          Determine the number of nodes to form VCS cluster
        4.  
          Install and configure VCS inside the control domain and alternate I/O domain
        5. Configuring storage services
          1.  
            Configuring virtual disk with LUN, Shared Disk, or CVM Volume
          2.  
            Virtual disk configurations with DMP
          3.  
            Configuring storage services when back-end device is a ZFS volume
        6.  
          Configure storage service groups
        7.  
          Configure network service groups
        8. Configure a service group to monitor services from multiple I/O domains
          1.  
            Type of service group configuration for the AlternateIO resource
        9. Configure the AlternateIO resource
          1.  
            Sample service group configuration for the AlternateIO resource
        10. Configure the service group for a Logical Domain
          1.  
            Sample configuration for LDom service group
        11.  
          Failover scenarios
        12.  
          Recommendations while configuring VCS and Oracle VM Server for SPARC with multiple I/O domains
        13.  
          Sample VCS configuration for AlternateIO resource configured as a fail over type
      6.  
        Configuring VCS on logical domains to manage applications using services from multiple I/O domains
    4. SF Oracle RAC support for Oracle VM Server for SPARC environments
      1.  
        About deploying SF Oracle RAC in Oracle VM Server for SPARC environments
      2.  
        Sample configuration scenarios
      3.  
        Preparing to deploy SF Oracle RAC in logical domain environments
      4.  
        SF Oracle RAC with Oracle RAC database on I/O domains of two hosts
      5.  
        SF Oracle RAC with Oracle RAC database on guest domains of two hosts
      6.  
        SF Oracle RAC with Oracle RAC database on guest domains of single host
      7.  
        SF Oracle RAC with Oracle RAC database on I/O domain and guest domain of single host
    5. Support for live migration in FSS environments
      1.  
        About live migration in Flexible Storage Sharing (FSS) environments
      2.  
        Performing live migration of Oracle VM Server for SPARC systems in Flexible Storage Sharing (FSS) environments
  4. Section IV. Reference
    1. Appendix A. Where to find more information
      1.  
        Veritas InfoScale documentation
      2.  
        Solaris virtualization documentation
      3.  
        Service and support
      4.  
        About Veritas Services and Operations Readiness Tools (SORT)

Configuring a direct mount of VxFS file system in a non-global zone with VCS

Typical steps to configure a direct mount inside a non-global zone.

To configure a direct mount inside a non-global zone

  1. Create a VxVM disk group and volume:

    • Create a VxVM disk group from a device:

      global# vxdg init data_dg c0t0d1
    • Create a volume from a disk group:

      global# vxassist -g data_dg make data_vol 5G

      For more information, see the Storage Foundation Administrator's Guide.

  2. Create a zone:

    • Create a root directory for a zone local-zone and change its permission to 700:

      global# mkdir -p /zones/local-zone
      global# chmod 700 /zones/local-zone
    • On Solaris 11, configure a zone local-zone:

      global# zonecfg -z local-zone
      
      local-zone: No such zone configured
      Use 'create' to begin configuring a new zone.
      
      zonecfg:local-zone> create         
      zonecfg:local-zone> set zonepath=/zones/local-zone
      zonecfg:local-zone> set ip-type=shared
      zonecfg:local-zone> add net
      zonecfg:local-zone:net> set physical=eri0
      zonecfg:local-zone:net> set address=192.168.5.59
      zonecfg:local-zone:net> end
      zonecfg:local-zone > verify
      zonecfg:local-zone > commit
      zonecfg:local-zone > exit

      The zone is in configured state.

    • Install the zone:

      global# zoneadm -z local-zone install
    • Login to the zone console to setup the zone from terminal 1:

      global# zlogin -C local-zone
    • Boot the zone from another terminal:

      global# zoneadm -z local-zone boot
    • Follow the steps on terminal 1 on the zone console to setup the zone.

      See the Oracle documentation for more information about creating a zone.

  3. Add VxVM volumes to the zone configuration:

    • Check the zone status and halt the zone, if it is running:

      global# zoneadm list -cv
      ID NAME       STATUS   PATH           BRAND    IP
      0 global      running  /              native    shared
      2 local-zone  running  /zones/myzone  native    shared
      
      global# zoneadm -z myzone halt
    • Add the VxVM devices to the zone's configuration:

      global# zonecfg -z local-zone
      zonecfg:local-zone:fs> add device
      zonecfg:local-zone:fs> set match=/dev/vxportal
      zonecfg:local-zone:fs> end
      zonecfg:local-zone:fs> add device
      zonecfg:local-zone:fs> set match=/dev/fdd
      zonecfg:local-zone:fs> end
      zonecfg:local-zone:fs> add device
      zonecfg:local-zone:fs> set match=/dev/vx/rdsk/data_dg/data_vol
      zonecfg:local-zone:fs> end
      zonecfg:local-zone:fs> add device
      zonecfg:local-zone:fs> set match=/dev/vx/dsk/data_dg/data_vol
      zonecfg:local-zone:fs> end
      zonecfg:local-zone:fs> add fs
      zonecfg:local-zone:fs> set dir=/etc/vx/licenses/lic
      zonecfg:local-zone:fs> set special=/etc/vx/licenses/lic
      zonecfg:local-zone:fs> set type=lofs
      zonecfg:local-zone:fs> end
      zonecfg:local-zone> verify
      zonecfg:local-zone> commit
      zonecfg:local-zone> exit
    • On Solaris 11, you must set fs-allowed to vxfs and odm in the zone's configuration:

      global# zonecfg -z myzone
      zonecfg:myzone> set fs-allowed=vxfs,odm
      zonecfg:myzone> commit
      zonecfg:myzone> exit

      See Adding a direct mount to a zone's configuration.

      See Mounting VxFS as VxFS inside a non-global zone.

    • Boot the zone:

      global# zoneadm -z myzone boot
  4. Create a VxFS file system on the volume inside a non-global zone:

    • Login to the local-zone:

      global# zlogin myzone
    • Create a VxFS file system on the block device:

      bash-3.00# mkfs -F vxfs /dev/vx/dsk/data_dg/data_vol
  5. Create a mount point inside the zone:

    • Login to the local-zone:

      global# zlogin myzone
    • Create a mount point inside the non-global zone:

      bash-3.00# mkdir -p /mydata
    • Mount the VxFS file system on the mount point:

      bash-3.00# mount -F vxfs /dev/vx/dsk/data_dg/data_vol /mydata
  6. Configure the zone service group:

    • On the first node, create the service group with password-less communication with global zone:

      global# hazonesetup -g zone_grp -r zone_res -z myzone \
      -p password -s sysA,sysB
    • Switch the service group from the first node to the second node and run the hazonesetup command to setup password-less communication from the next node.

    • Repeat step 6 for all the nodes in the cluster where the zone can go online.

    See Configuring a zone resource in a failover service group with the hazonesetup utility.
  7. Create a mount, disk group, and volume resources into the service group:

    • Add a disk group resource to the service group:

      global# hares -add dg_res DiskGroup zone_grp
      global# hares -modify dg_res DiskGroup data_dg
      global# hares -modify dg_res Enabled 1
    • Add a volume resource to the service group:

      global# hares -add vol_res Volume zone_grp
      global# hares -modify vol_res Volume data_vol
      global# hares -modify vol_res DiskGroup data_dg
      global# hares -modify vol_res Enabled 1
    • Add a Mount resource to the service group:

      global# hares -add mnt_res Mount zone_grp
      global# hares -modify mnt_res BlockDevice \
      /dev/vx/dsk/data_dg/data_vol
      global# hares -modify mnt_res MountPoint /mydata
      global# hares -modify mnt_res FSType vxfs
      global# hares -modify mnt_res FsckOpt %-y
      global# hares -modify mnt_res Enabled 1
    • Create a resource dependency between the resources in the service group:

      global# hares -link zone_res vol_res
      global# hares -link vol_res dg_res
      global# hares -link mnt_res zone_res
  8. Set the ContainerOpts attribute for the Mount resource for VxFS direct mount:

    • Override the ContainerOpts attribute at the resource level for mnt_res:

      global# hares -override mnt_res ContainerOpts
    • Set the value of the RunInContainer key to 1:

      global# hares -modify mnt_res ContainerOpts RunInContainer \
      1 PassCInfo 0
    For information on overriding resource type static attributes: see the Cluster Server Administrator's Guide.
  9. Here is a sample configuration for the VxFS direct mount service groups in the main.cf file:
    group zone_grp (
           SystemList = {sysA = 0, sysB = 1 }
           ContainerInfo = { Name = local-zone, Type = Zone, Enabled = 1 }
           Administrators = { z_zoneres_sysA, z_zoneres_sysB }
            )
            
           Mount mnt_res (
                    BlockDevice = "/dev/vx/dsk/data_dg/data_vol"
                    MountPoint = "/mydata"
                    FSType = vxfs
                    FsckOpt = "-y"
                    ContainerOpts = { RunInContainer = 1, PassCInfo = 0 }
                     )
                                   
           DiskGroup dg_res (
                    DiskGroup = data_dg
            )
    
           Volume vol_res (
                    Volume = data_vol
                    DiskGroup = data_dg
            )
                                  
           Zone zone_res (
            )
    
           zone_res requires vol_res
           vol_res requires dg_res 
           mnt_res requires zone_res