Veritas InfoScale™ 8.0 Virtualization Guide - Solaris
- Section I. Overview of Veritas InfoScale Solutions used in Solaris virtualization
- Section II. Zones
- InfoScale Enterprise Solutions support for Solaris Native Zones
- About VCS support for zones
- Configuring VCS in zones
- Prerequisites for configuring VCS in zones
- Deciding on the zone root location
- Configuring the service group for the application
- Exporting VxVM volumes to a non-global zone
- About InfoScale SFRAC component support for Oracle RAC in a zone environment
- Known issues with supporting in a InfoScale SFRAC component zone environment
- Software limitations of InfoScale support of non-global zones
- InfoScale Enterprise Solutions support for Solaris Native Zones
- Section III. Oracle VM Server for SPARC
- InfoScale Enterprise Solutions support for Oracle VM Server for SPARC
- Oracle VM Server for SPARC deployment models
- Benefits of deploying solutions in Oracle VM server for SPARC
- Features
- Split InfoScale stack model
- Guest-based InfoScale stack model
- Layered InfoScale stack model
- System requirements
- Installing InfoScale in a Oracle VM Server for SPARC environment
- Provisioning storage for a guest domain
- Software limitations
- Known issues
- Cluster Server support for using CVM with multiple nodes in a Oracle VM Server for SPARC environment
- VCS: Configuring Oracle VM Server for SPARC for high availability
- About VCS in a Oracle VM Server for SPARC environment
- About Cluster Server configuration models in an Oracle VM Server for SPARC environment
- Cluster Server setup to fail over a logical domain on a failure of logical domain
- Cluster Server setup to fail over an Application running inside logical domain on a failure of Application
- Oracle VM Server for SPARC guest domain migration in VCS environment
- Overview of a live migration
- About configuring VCS for Oracle VM Server for SPARC with multiple I/O domains
- Configuring VCS to manage a Logical Domain using services from multiple I/O domains
- Configuring storage services
- Configure a service group to monitor services from multiple I/O domains
- Configure the AlternateIO resource
- Configure the service group for a Logical Domain
- SFRAC support for Oracle VM Server for SPARC environments
- Support for live migration in FSS environments
- Using SmartIO in the virtualized environment
- InfoScale Enterprise Solutions support for Oracle VM Server for SPARC
- Section IV. Reference
Exporting a VeritasVeritas volume to a guest domain from the control domain
Use the following procedure to migrate a VxVM disk group from a non-logical domain environment to a Oracle VM Server for SPARC environment.
Note:
This section applies to only the Split InfoScale model.
In the following example control domain is named primary
and the guest domain is named LDOM1
The prompts in each step show in which domain to run the command.
To create virtual disks on top of the Veritas Volume Manager data volumes using the ldm command
- The VxVM diskgroup on the target host is imported in the control domain, after which volumes are visible from inside the control domain.
See the Storage Foundation Administrator's Guide to move disk groups between systems.
- In the control domain (primary), configure a service exporting the VxVM volume containing a VxFS or UFS filesystem as a slice using the options=slice option:
primary# ldm add-vdiskserverdevice options=slice \ /dev/vx/dsk/dg-name/volume_name \ volume_name volume_name@primary-vds0
Caution:
A volume by default shows up as a full disk in the guest. The Virtual Disk Client driver writes a VTOC on block 0 of the virtual disk, which will end up as a WRITE on block 0 of the VxVM volume. This can potentially cause data corruption, because block 0 of the VxVM volume contains user data. Using options=slice exports a volume as a slice to the guest and does not cause any writes to block 0, therefore preserving user data.
- Add the exported disk to a guest domain:
primary# ldm add-vdisk vdisk1 volume_name volume_name@primary-vds0 ldom1
- Start the guest domain, and ensure that the new virtual disk is visible.
primary# ldm bind ldom1
primary# ldm start ldom1
- If the new virtual disk device node entries do not show up in the
/dev/[r]dsk
directories, then run the devfsadm command in the guest domain:ldom1# devfsadm -C
In this example, the new disk appears as /dev/[r]dsk/c0d1s0.
ldom1# ls -l /dev/dsk/c0d1s0
lrwxrwxrwx 1 root root 62 Sep 11 13:30 /dev/dsk/c0d1s0 -> ../../devices/virtual-devices@100/channel-devices@200/disk@1:a
- Mount the file system on the disk to access the application data:
ldom1# mount -F vxfs /dev/dsk/c0d1s0 /mnt
ldom1# mount -F ufs /dev/dsk/c0d1s0 /mnt