Please enter search query.
Search <book_title>...
Veritas InfoScale™ 7.3.1 Troubleshooting Guide - Solaris
Last Published:
2018-08-22
Product(s):
InfoScale & Storage Foundation (7.3.1)
Platform: Solaris
- Introduction
- Section I. Troubleshooting Veritas File System
- Section II. Troubleshooting Veritas Volume Manager
- Recovering from hardware failure
- Failures on RAID-5 volumes
- Recovery from failure of a DCO volume
- Recovering from instant snapshot failure
- Recovering from failed vxresize operation
- Recovering from boot disk failure
- Hot-relocation and boot disk failure
- Recovery from boot failure
- Repair of root or /usr file systems on mirrored volumes
- Replacement of boot disks
- Recovery by reinstallation
- Managing commands, tasks, and transactions
- Backing up and restoring disk group configurations
- Troubleshooting issues with importing disk groups
- Recovering from CDS errors
- Logging and error messages
- Troubleshooting Veritas Volume Replicator
- Recovery from configuration errors
- Errors during an RLINK attach
- Errors during modification of an RVG
- Recovery on the Primary or Secondary
- Recovering from Primary data volume error
- Primary SRL volume error cleanup and restart
- Primary SRL header error cleanup and recovery
- Secondary data volume error cleanup and recovery
- Troubleshooting issues in cloud deployments
- Recovering from hardware failure
- Section III. Troubleshooting Dynamic Multi-Pathing
- Section IV. Troubleshooting Storage Foundation Cluster File System High Availability
- Troubleshooting Storage Foundation Cluster File System High Availability
- Troubleshooting CFS
- Troubleshooting fenced configurations
- Troubleshooting Cluster Volume Manager in Veritas InfoScale products clusters
- Troubleshooting Storage Foundation Cluster File System High Availability
- Section V. Troubleshooting Cluster Server
- Troubleshooting and recovery for VCS
- VCS message logging
- Gathering VCS information for support analysis
- Troubleshooting the VCS engine
- Troubleshooting Low Latency Transport (LLT)
- Troubleshooting Group Membership Services/Atomic Broadcast (GAB)
- Troubleshooting VCS startup
- Troubleshooting service groups
- Troubleshooting resources
- Troubleshooting I/O fencing
- System panics to prevent potential data corruption
- Fencing startup reports preexisting split-brain
- Troubleshooting CP server
- Troubleshooting server-based fencing on the Veritas InfoScale products cluster nodes
- Issues during online migration of coordination points
- Troubleshooting notification
- Troubleshooting and recovery for global clusters
- Troubleshooting licensing
- Licensing error messages
- VCS message logging
- Troubleshooting and recovery for VCS
- Section VI. Troubleshooting SFDB
Removing preexisting keys
If you encountered a split-brain condition, use the vxfenclearpre utility to remove CP Servers, SCSI-3 registrations, and reservations on the coordinator disks, Coordination Point servers, as well as on the data disks in all shared disk groups.
You can also use this procedure to remove the registration and reservation keys of another node or other nodes on shared disks or CP server.
To clear keys after split-brain
- Stop VCS on all nodes.
# hastop -all
- Make sure that the port h is closed on all the nodes. Run the following command on each node to verify that the port h is closed:
# gabconfig -a
Port h must not appear in the output.
- Stop I/O fencing on all nodes. Enter the following command on each node:
# svcadm disable -t vxfen
- If you have any applications that run outside of VCS control that have access to the shared storage, then shut down all other nodes in the cluster that have access to the shared storage. This prevents data corruption.
- Start the vxfenclearpre script:
# /opt/VRTSvcs/vxfen/bin/vxfenclearpre
- Read the script's introduction and warning. Then, you can choose to let the script run.
Do you still want to continue: [y/n] (default : n) y
In some cases, informational messages resembling the following may appear on the console of one of the nodes in the cluster when a node is ejected from a disk/LUN. You can ignore these informational messages.
<date> <system name> scsi: WARNING: /sbus@3,0/lpfs@0,0/ sd@0,1(sd91): <date> <system name> Error for Command: <undecoded cmd 0x5f> Error Level: Informational <date> <system name> scsi: Requested Block: 0 Error Block 0 <date> <system name> scsi: Vendor: <vendor> Serial Number: 0400759B006E <date> <system name> scsi: Sense Key: Unit Attention <date> <system name> scsi: ASC: 0x2a (<vendor unique code 0x2a>), ASCQ: 0x4, FRU: 0x0
The script cleans up the disks and displays the following status messages.
Cleaning up the coordinator disks... Cleared keys from n out of n disks, where n is the total number of disks. Successfully removed SCSI-3 persistent registrations from the coordinator disks. Cleaning up the Coordination Point Servers... ................... [10.209.80.194]:50001: Cleared all registrations [10.209.75.118]:443: Cleared all registrations Successfully removed registrations from the Coordination Point Servers. Cleaning up the data disks for all shared disk groups ... Successfully removed SCSI-3 persistent registration and reservations from the shared data disks. See the log file /var/VRTSvcs/log/vxfen/vxfen.log You can retry starting fencing module. In order to restart the whole product, you might want to reboot the system.
- Start the fencing module on all the nodes.
# svcadm enable vxfen
- Start VCS on all nodes.
# hastart