Veritas InfoScale™ 7.3.1 Virtualization Guide - Solaris
- Section I. Overview of Veritas InfoScale Solutions used in Solaris virtualization
- Section II. Zones and Projects
- Storage Foundation and High Availability Solutions support for Solaris Zones
- About VCS support for zones
- About the Mount agent
- Configuring VCS in zones
- Prerequisites for configuring VCS in zones
- Deciding on the zone root location
- Configuring the service group for the application
- Exporting VxVM volumes to a non-global zone
- About SF Oracle RAC support for Oracle RAC in a zone environment
- Known issues with supporting SF Oracle RAC in a zone environment
- Software limitations of Storage Foundation support of non-global zones
- Storage Foundation and High Availability Solutions support for Solaris Projects
- Storage Foundation and High Availability Solutions support for Solaris Zones
- Section III. Oracle VM Server for SPARC
- Storage Foundation and High Availability Solutions support for Oracle VM Server for SPARC
- Oracle VM Server for SPARC deployment models
- Benefits of deploying Storage Foundation High Availability solutions in Oracle VM server for SPARC
- Features
- Split Storage Foundation stack model
- Guest-based Storage Foundation stack model
- Layered Storage Foundation stack model
- System requirements
- Installing Storage Foundation in a Oracle VM Server for SPARC environment
- Provisioning storage for a guest domain
- Software limitations
- Known issues
- Cluster Server support for using CVM with multiple nodes in a Oracle VM Server for SPARC environment
- VCS: Configuring Oracle VM Server for SPARC for high availability
- About VCS in a Oracle VM Server for SPARC environment
- About Cluster Server configuration models in an Oracle VM Server for SPARC environment
- Cluster Server setup to fail over a logical domain on a failure of logical domain
- Cluster Server setup to fail over an Application running inside logical domain on a failure of Application
- Oracle VM Server for SPARC guest domain migration in VCS environment
- Overview of a live migration
- About configuring VCS for Oracle VM Server for SPARC with multiple I/O domains
- Configuring VCS to manage a Logical Domain using services from multiple I/O domains
- Configuring storage services
- Configure a service group to monitor services from multiple I/O domains
- Configure the AlternateIO resource
- Configure the service group for a Logical Domain
- SF Oracle RAC support for Oracle VM Server for SPARC environments
- Support for live migration in FSS environments
- Storage Foundation and High Availability Solutions support for Oracle VM Server for SPARC
- Section IV. Reference
Exporting a Veritas volume to a guest domain from the control domain
Use the following procedure to migrate a VxVM disk group from a non-logical domain environment to a Oracle VM Server for SPARC environment.
Note:
This section applies to only the Split Storage Foundation model.
In the following example control domain is named primary
and the guest domain is named ldom1
The prompts in each step show in which domain to run the command.
To create virtual disks on top of the Veritas Volume Manager data volumes using the ldm command
- The VxVM diskgroup on the target host is imported in the control domain, after which volumes are visible from inside the control domain.
See the Storage Foundation Administrator's Guide to move disk groups between systems.
- In the control domain (primary), configure a service exporting the VxVM volume containing a VxFS or UFS filesystem as a slice using the options=slice option:
primary# ldm add-vdiskserverdevice options=slice \ /dev/vx/dsk/dg-name/volume_name \ volume_name volume_name@primary-vds0
Caution:
A volume by default shows up as a full disk in the guest. The Virtual Disk Client driver writes a VTOC on block 0 of the virtual disk, which will end up as a WRITE on block 0 of the VxVM volume. This can potentially cause data corruption, because block 0 of the VxVM volume contains user data. Using options=slice exports a volume as a slice to the guest and does not cause any writes to block 0, therefore preserving user data.
- Add the exported disk to a guest domain:
primary# ldm add-vdisk vdisk1 volume_name volume_name@primary-vds0 ldom1
- Start the guest domain, and ensure that the new virtual disk is visible.
primary# ldm bind ldom1
primary# ldm start ldom1
- If the new virtual disk device node entires do not show up in the
/dev/[r]dsk
directories, then run the devfsadm command in the guest domain:ldom1# devfsadm -C
In this example, the new disk appears as /dev/[r]dsk/c0d1s0.
ldom1# ls -l /dev/dsk/c0d1s0
lrwxrwxrwx 1 root root 62 Sep 11 13:30 /dev/dsk/c0d1s0 -> ../../devices/virtual-devices@100/channel-devices@200/disk@1:a
- Mount the file system on the disk to access the application data:
ldom1# mount -F vxfs /dev/dsk/c0d1s0 /mnt
ldom1# mount -F ufs /dev/dsk/c0d1s0 /mnt