Please enter search query.
Search <book_title>...
Veritas InfoScale™ 7.3.1 Virtualization Guide - Solaris
Last Published:
2018-08-22
Product(s):
InfoScale & Storage Foundation (7.3.1)
Platform: Solaris
- Section I. Overview of Veritas InfoScale Solutions used in Solaris virtualization
- Section II. Zones and Projects
- Storage Foundation and High Availability Solutions support for Solaris Zones
- About VCS support for zones
- About the Mount agent
- Configuring VCS in zones
- Prerequisites for configuring VCS in zones
- Deciding on the zone root location
- Configuring the service group for the application
- Exporting VxVM volumes to a non-global zone
- About SF Oracle RAC support for Oracle RAC in a zone environment
- Known issues with supporting SF Oracle RAC in a zone environment
- Software limitations of Storage Foundation support of non-global zones
- Storage Foundation and High Availability Solutions support for Solaris Projects
- Storage Foundation and High Availability Solutions support for Solaris Zones
- Section III. Oracle VM Server for SPARC
- Storage Foundation and High Availability Solutions support for Oracle VM Server for SPARC
- Oracle VM Server for SPARC deployment models
- Benefits of deploying Storage Foundation High Availability solutions in Oracle VM server for SPARC
- Features
- Split Storage Foundation stack model
- Guest-based Storage Foundation stack model
- Layered Storage Foundation stack model
- System requirements
- Installing Storage Foundation in a Oracle VM Server for SPARC environment
- Provisioning storage for a guest domain
- Software limitations
- Known issues
- Cluster Server support for using CVM with multiple nodes in a Oracle VM Server for SPARC environment
- VCS: Configuring Oracle VM Server for SPARC for high availability
- About VCS in a Oracle VM Server for SPARC environment
- About Cluster Server configuration models in an Oracle VM Server for SPARC environment
- Cluster Server setup to fail over a logical domain on a failure of logical domain
- Cluster Server setup to fail over an Application running inside logical domain on a failure of Application
- Oracle VM Server for SPARC guest domain migration in VCS environment
- Overview of a live migration
- About configuring VCS for Oracle VM Server for SPARC with multiple I/O domains
- Configuring VCS to manage a Logical Domain using services from multiple I/O domains
- Configuring storage services
- Configure a service group to monitor services from multiple I/O domains
- Configure the AlternateIO resource
- Configure the service group for a Logical Domain
- SF Oracle RAC support for Oracle VM Server for SPARC environments
- Support for live migration in FSS environments
- Storage Foundation and High Availability Solutions support for Oracle VM Server for SPARC
- Section IV. Reference
About the Mount agent
You may need to modify the ContainerOpts values for the Mount resource in certain situations.
In certain situations where the block device is not exported to zone, you can make the file system available inside local zone. Mount the block device on the directory that has a path that includes the zone root from global zone, for example:
BlockDevice = /dev/vx/dsk/dg/vol1 MountPoint = /zones/zone-test/root/mntpt
Where /zones/zone-test
is the zone root of the local zone.
Mount agent supports the following configuration for mount points
- Direct mount of file system with mount point as full path seen from global zone. Typical mount resource configuration for this type of mount is shown below:
group mntgrp ( SystemList = { Sys1 = 0, Sys1 = 1 } ) Mount mnt-direct ( MountPoint = "/zones/zone-test/root/mnt" BlockDevice = "/dev/vx/dsk/dg/vol" FSType = vxfs FsckOpt = "-y" )
- Loop-back file system mount inside non-global zone for file system mounted in global zone. Typical mount resource configuration for this type of mount is shown below:
group loopbacksg ( SystemList = { sysA = 0, sysB = 1 } ContainerInfo@sysA = { Name = zone1, Type = Zone, Enabled = 1 } ContainerInfo@sysB = { Name = zone1, Type = Zone, Enabled = 1 } ) Mount zone_mnt ( MountPoint = "/export/home/zone1/root/lofs_mnt" BlockDevice = "/mnt1/m1" FSType = lofs ) Zone z1 ( ) Mount global_mnt ( MountPoint = "/mnt1" BlockDevice = "/dev/vx/dsk/tdg/tvol1" FSType = vxfs FsckOpt = "-y" ) zone_mnt requires z1 zone_mnt requires global_mnt
- Direct mount of NFS based file system inside non-global zone. Typical mount resource configuration for this type of mount is shown below:
group mntgrp ( SystemList = { Sys1 = 0, Sys1 = 1 } ContainerInfo = { Name = zone-test, Type = Zone, Enabled = 1 } ) Mount mntnfs ( MountPoint = "/mnt" BlockDevice = "system:/shared-dir" FSType = nfs FsckOpt = "-n" ContainerOpts = { RunInContainer = 1, PassCInfo = 0 } )
- Support for direct mount of VxFS file system inside non-global zone. VCS Mount agent supports the direct mount of VxFS file system inside non-global zone. Typical mount resource configuration for this type of mount is shown below:
group mntgrp ( SystemList = { sys1 = 0, sys2 = 1 } ContainerInfo = { Name = zone-test, Type = Zone, Enabled = 1 } Administrators = { z_zoneres_sys1, z_zoneres_sys2 } ) Mount mnt-zone-direct ( BlockDevice = "/dev/vx/dsk/data_dg/data_vol" MountPoint = "/mnt1" FSType = vxfs FsckOpt = "-y" ContainerOpts = { RunInContainer = 1, PassCInfo = 0 } )