InfoScale™ 9.0 Solutions Guide - Linux

Last Published:
Product(s): InfoScale & Storage Foundation (9.0)
Platform: Linux
  1. Section I. Introducing Veritas InfoScale
    1. Introducing Veritas InfoScale
      1.  
        About the Arctera InfoScale product suite
      2.  
        Components of the Arctera InfoScale product suite
  2. Section II. Solutions for Veritas InfoScale products
    1. Solutions for Veritas InfoScale products
      1.  
        Use cases for Veritas InfoScale products
      2.  
        Feature support across Veritas InfoScale 9.0 products
      3.  
        Using SmartMove and Thin Provisioning with Sybase databases
      4.  
        Running multiple parallel applications within a single cluster using the application isolation feature
      5.  
        Scaling FSS storage capacity with dedicated storage nodes using application isolation feature
      6.  
        Finding Veritas InfoScale product use cases information
  3. Section III. Stack-level migration to IPv6 or dual stack
    1. Stack-level migration to IPv6 or dual stack
      1.  
        Migrating Veritas InfoScale products to support IPv6/dual-stack
  4. Section IV. Improving database performance
    1. Overview of database accelerators
      1.  
        About Arctera InfoScale™ product components database accelerators
    2. Improving database performance with Veritas Concurrent I/O
      1. About Concurrent I/O
        1.  
          How Concurrent I/O works
      2. Tasks for enabling and disabling Concurrent I/O
        1.  
          Enabling Concurrent I/O for Sybase
        2.  
          Disabling Concurrent I/O for Sybase
    3. Improving database performance with atomic write I/O
      1.  
        About the atomic write I/O
      2.  
        Requirements for atomic write I/O
      3.  
        Restrictions on atomic write I/O functionality
      4.  
        How the atomic write I/O feature of Storage Foundation helps MySQL databases
      5.  
        VxVM and VxFS exported IOCTLs
      6.  
        Configuring atomic write I/O support for MySQL on VxVM raw volumes
      7.  
        Configuring atomic write I/O support for MySQL on VxFS file systems
      8.  
        Dynamically growing the atomic write capable file system
      9.  
        Disabling atomic write I/O support
  5. Section V. Using point-in-time copies
    1. Understanding point-in-time copy methods
      1. About point-in-time copies
        1.  
          Implementing point-in time copy solutions on a primary host
        2.  
          Implementing off-host point-in-time copy solutions
      2.  
        When to use point-in-time copies
      3. About Storage Foundation point-in-time copy technologies
        1. Volume-level snapshots
          1.  
            Persistent FastResync of volume snapshots
          2.  
            Data integrity in volume snapshots
        2.  
          Storage Checkpoints
    2. Backing up and recovering
      1.  
        Storage Foundation and High Availability Solutions backup and recovery methods
      2. Preserving multiple point-in-time copies
        1.  
          Setting up multiple point-in-time copies
        2.  
          Refreshing point-in-time copies
        3.  
          Recovering from logical corruption
        4.  
          Off-host processing using refreshed snapshot images
      3. Online database backups
        1. Making a backup of an online database on the same host
          1.  
            Preparing a full-sized instant snapshot for a backup
          2.  
            Preparing a space-optimized snapshot for a database backup
          3.  
            Backing up a Sybase database on the same host
          4.  
            Resynchronizing a volume
        2. Making an off-host backup of an online database
          1.  
            Making an off-host backup of an online Sybase database
          2.  
            Resynchronizing a volume
      4. Backing up on an off-host cluster file system
        1.  
          Mounting a file system for shared access
        2.  
          Preparing a snapshot of a mounted file system with shared access
        3.  
          Backing up a snapshot of a mounted file system with shared access
        4.  
          Resynchronizing a volume from its snapshot volume
        5.  
          Reattaching snapshot plexes
      5. Database recovery using Storage Checkpoints
        1.  
          Creating Storage Checkpoints
        2.  
          Rolling back a database
    3. Backing up and recovering in a NetBackup environment
      1.  
        About Veritas NetBackup
      2.  
        About using NetBackup for backup and restore for Sybase
      3. Using NetBackup in an SFHA Solutions product environment
        1.  
          Clustering a NetBackup Master Server
        2.  
          Backing up and recovering a VxVM volume using NetBackup
        3.  
          Recovering a VxVM volume using NetBackup
    4. Off-host processing
      1.  
        Veritas InfoScale Storage Foundation off-host processing methods
      2. Using a replica database for decision support
        1. Creating a replica database on the same host
          1.  
            Preparing for the replica database
          2.  
            Creating a replica database
        2. Creating an off-host replica database
          1.  
            Setting up a replica database for off-host decision support
          2.  
            Resynchronizing the data with the primary host
          3.  
            Updating a warm standby Sybase ASE 12.5 database
          4.  
            Reattaching snapshot plexes
      3.  
        What is off-host processing?
      4.  
        About using VVR for off-host processing
    5. Creating and refreshing test environments
      1.  
        About test environments
      2.  
        Creating a test environment
      3.  
        Refreshing a test environment
    6. Creating point-in-time copies of files
      1. Using FileSnaps to create point-in-time copies of files
        1.  
          Using FileSnaps to provision virtual desktops
        2.  
          Using FileSnaps to optimize write intensive applications for virtual machines
        3.  
          Using FileSnaps to create multiple copies of data instantly
  6. Section VI. Maximizing storage utilization
    1. Optimizing storage tiering with SmartTier
      1.  
        About SmartTier
      2.  
        About VxFS multi-volume file systems
      3.  
        About VxVM volume sets
      4.  
        About volume tags
      5.  
        SmartTier use cases for Sybase
      6.  
        Setting up a filesystem for storage tiering with SmartTier
      7.  
        Relocating old archive logs to tier two storage using SmartTier
      8.  
        Relocating inactive tablespaces or segments to tier two storage
      9.  
        Relocating active indexes to premium storage
      10.  
        Relocating all indexes to premium storage
    2. Optimizing storage with Flexible Storage Sharing
      1. About Flexible Storage Sharing
        1.  
          Limitations of Flexible Storage Sharing
      2.  
        About use cases for optimizing storage with Flexible Storage Sharing
      3.  
        Setting up an SFRAC clustered environment with shared nothing storage
      4.  
        Implementing the SmartTier feature with hybrid storage
      5.  
        Configuring a campus cluster without shared storage
  7. Section VII. Migrating data
    1. Understanding data migration
      1.  
        Types of data migration
    2. Offline migration from LVM to VxVM
      1.  
        About migration from LVM
      2.  
        Converting unused LVM physical volumes to VxVM disks
      3. LVM volume group to VxVM disk group conversion
        1.  
          Volume group conversion limitations
        2.  
          Converting LVM volume groups to VxVM disk groups
        3. Examples of second stage failure analysis
          1.  
            Snapshot in the volume group
          2.  
            dm_mirror module not loaded in the kernel
          3.  
            Conversion requires extent movement on an LVM1 volume group
          4.  
            Unrecognized partition in volume group
      4. LVM volume group restoration
        1.  
          Restoring an LVM volume group
    3. Offline conversion of native file system to VxFS
      1.  
        About the offline conversion of native file system to VxFS
      2.  
        Requirements for offline conversion of a native file system to VxFS
      3.  
        Converting the native file system to VxFS
    4. Online migration of a native file system to the VxFS file system
      1.  
        About online migration of a native file system to the VxFS file system
      2.  
        Administrative interface for online migration of a native file system to the VxFS file system
      3.  
        Migrating a native file system to the VxFS file system
      4. Migrating a source file system to the VxFS file system over NFS v4
        1.  
          Restrictions of NFS v4 migration
      5.  
        Backing out an online migration of a native file system to the VxFS file system
      6. VxFS features not available during online migration
        1.  
          Limitations of online migration
    5. Migrating storage arrays
      1.  
        Array migration for storage using Linux
      2.  
        Overview of storage mirroring for migration
      3.  
        Allocating new storage
      4.  
        Initializing the new disk
      5.  
        Checking the current VxVM information
      6.  
        Adding a new disk to the disk group
      7.  
        Mirroring
      8.  
        Monitoring
      9.  
        Mirror completion
      10.  
        Removing old storage
      11.  
        Post-mirroring steps
    6. Migrating data between platforms
      1. Overview of the Cross-Platform Data Sharing (CDS) feature
        1.  
          Shared data across platforms
        2.  
          Disk drive sector size
        3.  
          Block size issues
        4.  
          Operating system data
      2. CDS disk format and disk groups
        1. CDS disk access and format
          1. CDS disk types
            1.  
              Private and public regions
            2.  
              Disk access type auto
            3.  
              Platform block
            4.  
              AIX coexistence label
            5.  
              HP-UX coexistence label
            6.  
              VxVM ID block
          2. About Cross-platform Data Sharing (CDS) disk groups
            1.  
              Device quotas
            2.  
              Minor device numbers
        2.  
          Non-CDS disk groups
        3. Disk group alignment
          1. Alignment values
            1.  
              Dirty region log alignment
          2.  
            Object alignment during volume creation
      3. Setting up your system to use Cross-platform Data Sharing (CDS)
        1. Creating CDS disks from uninitialized disks
          1.  
            Creating CDS disks by using vxdisksetup
          2.  
            Creating CDS disks by using vxdiskadm
        2. Creating CDS disks from initialized VxVM disks
          1.  
            Creating a CDS disk from a disk that is not in a disk group
          2.  
            Creating a CDS disk from a disk that is already in a disk group
        3. Creating CDS disk groups
          1.  
            Creating a CDS disk group by using vxdg init
          2.  
            Creating a CDS disk group by using vxdiskadm
        4.  
          Converting non-CDS disks to CDS disks
        5.  
          Converting a non-CDS disk group to a CDS disk group
        6.  
          Verifying licensing
        7.  
          Defaults files
      4. Maintaining your system
        1. Disk tasks
          1.  
            Changing the default disk format
          2.  
            Restoring CDS disk labels
        2. Disk group tasks
          1.  
            Changing the alignment of a disk group during disk encapsulation
          2.  
            Changing the alignment of a non-CDS disk group
          3.  
            Splitting a CDS disk group
          4.  
            Moving objects between CDS disk groups and non-CDS disk groups
          5.  
            Moving objects between CDS disk groups
          6.  
            Joining disk groups
          7.  
            Changing the default CDS setting for disk group creation
          8.  
            Creating non-CDS disk groups
          9.  
            Upgrading an older version non-CDS disk group
          10.  
            Replacing a disk in a CDS disk group
          11.  
            Setting the maximum number of devices for CDS disk groups
          12.  
            Changing the DRL map and log size
          13.  
            Creating a volume with a DRL log
          14.  
            Setting the DRL map length
        3. Displaying information
          1.  
            Determining the setting of the CDS attribute on a disk group
          2.  
            Displaying the maximum number of devices in a CDS disk group
          3.  
            Displaying map length and map alignment of traditional DRL logs
          4.  
            Displaying the disk group alignment
          5.  
            Displaying the log map length and alignment
          6.  
            Displaying offset and length information in units of 512 bytes
        4.  
          Default activation mode of shared disk groups
        5.  
          Additional considerations when importing CDS disk groups
      5. File system considerations
        1.  
          Considerations about data in the file system
        2.  
          File system migration
        3. Specifying the migration target
          1.  
            Examples of target specifications
        4. Using the fscdsadm command
          1.  
            Checking that the metadata limits are not exceeded
          2. Maintaining the list of target operating systems
            1.  
              Adding an entry to the list of target operating systems
            2.  
              Removing an entry from the list of target operating systems
            3.  
              Removing all entries from the list of target operating systems
            4.  
              Displaying the list of target operating systems
          3.  
            Enforcing the established CDS limits on a file system
          4.  
            Ignoring the established CDS limits on a file system
          5.  
            Validating the operating system targets for a file system
          6.  
            Displaying the CDS status of a file system
        5.  
          Migrating a file system one time
        6. Migrating a file system on an ongoing basis
          1.  
            Stopping ongoing migration
        7.  
          When to convert a file system
        8. Converting the byte order of a file system
          1.  
            Importing and mounting a file system from another system
      6.  
        Alignment value and block size
      7.  
        Migrating a snapshot volume
    7. Migrating from Oracle ASM to Veritas File System
      1.  
        About the migration
      2.  
        Pre-requisites for migration
      3.  
        Preparing to migrate
      4.  
        Migrating Oracle databases from Oracle ASM to VxFS
  8. Section VIII. Veritas InfoScale 4K sector device support solution
    1. Veritas InfoScale 4k sector device support solution
      1.  
        About 4K sector size technology
      2.  
        InfoScale unsupported configurations
      3.  
        Migrating VxFS file system from 512-bytes sector size devices to 4K sector size devices
  9. Section IX. REST API support
    1. Support for configurations and operations using REST APIs
      1.  
        Support for InfoScale operations using REST APIs
      2.  
        Supported operations
      3.  
        Configuring the REST server
      4.  
        Security considerations for REST API management
      5.  
        Authorization of users for performing operations using REST APIs
      6.  
        Reconfiguring the REST server
      7.  
        Configuring HA for the REST server
      8.  
        Accessing the InfoScale REST API documentation
      9.  
        Unconfiguring the REST server
      10.  
        Troubleshooting information
      11.  
        Limitations
  10. Section X. Reference
    1. Appendix A. Veritas AppProtect logs and operation states
      1.  
        Log files
      2.  
        Plan states
    2. Appendix B. Troubleshooting Veritas AppProtect
      1.  
        Troubleshooting Just In Time Availability

Preparing for the replica database

To prepare a snapshot for a replica database on the primary host

  1. If you have not already done so, prepare the host to use the snapshot volume that contains the copy of the database tables. Set up any new database logs and configuration files that are required to initialize the database. On the master node, verify that the volume has an instant snap data change object (DCO) and DCO volume, and FastResync is enabled on the volume:
    # vxprint -g database_dg -F%instant database_vol
    # vxprint -g database_dg -F%fastresync database_vol

    If both commands return the value as ON, proceed to step 3. Otherwise, continue with step 2.

  2. Use the following command to prepare a volume for instant snapshots:
    # vxsnap -g database_dg prepare database_vol [regionsize=size] \ 
    [ndcomirs=number] [alloc=storage_attributes]
  3. Use the following command to make a full-sized snapshot, snapvol, of the tablespace volume by breaking off plexes from the original volume:
    # vxsnap -g database_dg make \
      source=volume/newvol=snapvol/nmirror=N

    The nmirror attribute specifies the number of mirrors, N, in the snapshot volume.

    If the volume does not have any available plexes, or its layout does not support plex break-off, prepare an empty volume for the snapshot.

  4. Use the vxprint command on the original volume to find the required size for the snapshot volume.
    # LEN=`vxprint [-g diskgroup] -F%len volume`

    Note:

    The command shown in this and subsequent steps assumes that you are using a Bourne-type shell such as sh, ksh or bash. You may need to modify the command for other shells such as csh or tcsh. These steps are valid only for an instant snap DCO.

  5. Use the vxprint command on the original volume to discover the name of its DCO:
    #  DCONAME=`vxprint [-g diskgroup] -F%dco_name volume`
  6. Use the vxprint command on the DCO to discover its region size (in blocks):
    # RSZ=`vxprint [-g diskgroup] -F%regionsz $DCONAME`
  7. Use the vxassist command to create a volume, snapvol, of the required size and redundancy. You can use storage attributes to specify which disks should be used for the volume. The init=active attribute makes the volume available immediately.
    # vxassist [-g diskgroup] make snapvol $LEN \
      [layout=mirror nmirror=number] init=active \
      [storage_attributes]
  8. Prepare the snapshot volume for instant snapshot operations as shown here:
    # vxsnap [-g diskgroup] prepare snapvol [ndcomirs=number] \
      regionsz=$RSZ [storage_attributes]

    It is recommended that you specify the same number of DCO mirrors (ndcomirror) as the number of mirrors in the volume (nmirror).

  9. To create the snapshot, use the following command:
    # vxsnap -g database_dg make source=volume/snapvol=snapvol

    If a database spans more than one volume, specify all the volumes and their snapshot volumes as separate tuples on the same line, for example:

    # vxsnap -g database_dg make \
    source=vol1/snapvol=svol1/nmirror=2 \
    source=vol2/snapvol=svol2/nmirror=2 \
    source=vol3/snapvol=svol3/nmirror=2

    If you want to save disk space, you can use the following command to create a space-optimized snapshot instead:

    # vxsnap -g database_dg make \
      source=volume/newvol=snapvol/cache=cacheobject

    The argument cacheobject is the name of a pre-existing cache that you have created in the disk group for use with space-optimized snapshots. To create the cache object, follow step 10 through step 13.

    If several space-optimized snapshots are to be created at the same time, these can all specify the same cache object as shown in this example:

    # vxsnap -g database_dg make \
      source=vol1/newvol=svol1/cache=dbaseco \
      source=vol2/newvol=svol2/cache=dbaseco \
      source=vol3/newvol=svol3/cache=dbaseco

  10. Decide on the following characteristics that you want to allocate to the cache volume that underlies the cache object:

    • The size of the cache volume should be sufficient to record changes to the parent volumes during the interval between snapshot refreshes. A suggested value is 10% of the total size of the parent volumes for a refresh interval of 24 hours.

    • If redundancy is a desired characteristic of the cache volume, it should be mirrored. This increases the space that is required for the cache volume in proportion to the number of mirrors that it has.

    • If the cache volume is mirrored, space is required on at least as many disks as it has mirrors. These disks should not be shared with the disks used for the parent volumes. The disks should also be chosen to avoid impacting I/O performance for critical volumes, or hindering disk group split and join operations.

  11. Having decided on its characteristics, use the vxassist command to create the volume that is to be used for the cache volume. The following example creates a mirrored cache volume, cachevol, with size 1GB in the disk group, mydg, on the disks disk16 and disk17:
    # vxassist -g mydg make cachevol 1g layout=mirror \
      init=active disk16 disk17

    The attribute init=active is specified to make the cache volume immediately available for use.

  12. Use the vxmake cache command to create a cache object on top of the cache volume that you created in the previous step:
    #  vxmake [-g diskgroup] cache cache_object \
      cachevolname=volume [regionsize=size] [autogrow=on] \
      [highwatermark=hwmk] [autogrowby=agbvalue] \
      [maxautogrow=maxagbvalue]]

    If you specify the region size, it must be a power of 2, and be greater than or equal to 16KB (16k). If not specified, the region size of the cache is set to 64KB.

    Note:

    All space-optimized snapshots that share the cache must have a region size that is equal to or an integer multiple of the region size set on the cache. Snapshot creation also fails if the original volume's region size is smaller than the cache's region size.

    If the cache is not allowed to grow in size as required, specify autogrow=off. By default, the ability to automatically grow the cache is turned on.

    In the following example, the cache object, cobjmydg, is created over the cache volume, cachevol, the region size of the cache is set to 32KB, and the autogrow feature is enabled:

    # vxmake -g mydg cache cobjmydg cachevolname=cachevol \
      regionsize=32k autogrow=on
  13. Having created the cache object, use the following command to enable it:
    # vxcache [-g diskgroup] start cache_object

    For example to start the cache object, cobjmydg:

    # vxcache -g mydg start cobjmydg

    Note:

    This step sets up the snapshot volumes, and starts tracking changes to the original volumes.