Veritas InfoScale™ 7.3.1 Troubleshooting Guide - Solaris
- Introduction
- Section I. Troubleshooting Veritas File System
- Section II. Troubleshooting Veritas Volume Manager
- Recovering from hardware failure
- Failures on RAID-5 volumes
- Recovery from failure of a DCO volume
- Recovering from instant snapshot failure
- Recovering from failed vxresize operation
- Recovering from boot disk failure
- Hot-relocation and boot disk failure
- Recovery from boot failure
- Repair of root or /usr file systems on mirrored volumes
- Replacement of boot disks
- Recovery by reinstallation
- Managing commands, tasks, and transactions
- Backing up and restoring disk group configurations
- Troubleshooting issues with importing disk groups
- Recovering from CDS errors
- Logging and error messages
- Troubleshooting Veritas Volume Replicator
- Recovery from configuration errors
- Errors during an RLINK attach
- Errors during modification of an RVG
- Recovery on the Primary or Secondary
- Recovering from Primary data volume error
- Primary SRL volume error cleanup and restart
- Primary SRL header error cleanup and recovery
- Secondary data volume error cleanup and recovery
- Troubleshooting issues in cloud deployments
- Recovering from hardware failure
- Section III. Troubleshooting Dynamic Multi-Pathing
- Section IV. Troubleshooting Storage Foundation Cluster File System High Availability
- Troubleshooting Storage Foundation Cluster File System High Availability
- Troubleshooting CFS
- Troubleshooting fenced configurations
- Troubleshooting Cluster Volume Manager in Veritas InfoScale products clusters
- Troubleshooting Storage Foundation Cluster File System High Availability
- Section V. Troubleshooting Cluster Server
- Troubleshooting and recovery for VCS
- VCS message logging
- Gathering VCS information for support analysis
- Troubleshooting the VCS engine
- Troubleshooting Low Latency Transport (LLT)
- Troubleshooting Group Membership Services/Atomic Broadcast (GAB)
- Troubleshooting VCS startup
- Troubleshooting service groups
- Troubleshooting resources
- Troubleshooting I/O fencing
- System panics to prevent potential data corruption
- Fencing startup reports preexisting split-brain
- Troubleshooting CP server
- Troubleshooting server-based fencing on the Veritas InfoScale products cluster nodes
- Issues during online migration of coordination points
- Troubleshooting notification
- Troubleshooting and recovery for global clusters
- Troubleshooting licensing
- Licensing error messages
- VCS message logging
- Troubleshooting and recovery for VCS
- Section VI. Troubleshooting SFDB
Validating license keys
The installvcs script handles most license key validations. However, if you install a VCS key outside of installvcs (using vxlicinst, for example), you can validate the key using the procedure described below.
- The vxlicinst command handles some of the basic validations:
node lock: Ensures that you are installing a node-locked key on the correct system
demo hard end date: Ensures that you are not installing an expired demo key
- Run the vxlicrep command to make sure a VCS key is installed on the system. The output of the command resembles:
License Key = XXXX-XXXX-XXXX-XXXX-XXXX-XXXX Product Name = VERITAS Cluster Server Serial number = XXXX License Type = PERMANENT OEM ID = XXXX Site License = YES Editions Product = YES Features : Platform: = Unused Version = 7.3.1 Tier = Tiern3 Reserved = 0 Mode = VCS Global Cluster Option = Enabled CPU_TIER = 0
- Look for the following in the command output:
Make sure the Product Name lists the name of your purchased component, for example, Veritas Infoscale products. If the command output does not return the product name, you do not have a VCS key installed.
If the output shows the License Type for a VCS key as DEMO, ensure that the Demo End Date does not display a past date.
Make sure the Mode attribute displays the correct value.
If you have purchased a license key for the Global Cluster Option, make sure its status is Enabled.
- Start VCS. If HAD rejects a license key, see the licensing error message at the end of the engine_A log file.