Please enter search query.
Search <book_title>...
InfoScale™ 9.0 Storage Foundation Cluster File System High Availability Configuration and Upgrade Guide - Solaris
Last Published:
2025-04-18
Product(s):
InfoScale & Storage Foundation (9.0)
Platform: Solaris
- Section I. Introduction to SFCFSHA
- Introducing Storage Foundation Cluster File System High Availability
- Section II. Configuration of SFCFSHA
- Preparing to configure
- Preparing to configure SFCFSHA clusters for data integrity
- About planning to configure I/O fencing
- Setting up the CP server
- Configuring the CP server manually
- Configuring SFCFSHA
- Configuring a secure cluster node by node
- Verifying and updating licenses on the system
- Configuring SFCFSHA clusters for data integrity
- Setting up disk-based I/O fencing using installer
- Setting up server-based I/O fencing using installer
- Performing an automated SFCFSHA configuration using response files
- Performing an automated I/O fencing configuration using response files
- Configuring CP server using response files
- Manually configuring SFCFSHA clusters for data integrity
- Setting up disk-based I/O fencing manually
- Setting up server-based I/O fencing manually
- Configuring server-based fencing on the SFCFSHA cluster manually
- Setting up non-SCSI-3 fencing in virtual environments manually
- Setting up majority-based I/O fencing manually
- Section III. Upgrade of SFCFSHA
- Planning to upgrade SFCFSHA
- Preparing to upgrade SFCFSHA
- Performing a full upgrade of SFCFSHA using the installer
- Performing a rolling upgrade of SFCFSHA
- Performing a phased upgrade of SFCFSHA
- About phased upgrade
- Performing a phased upgrade using the product installer
- Performing an automated SFCFSHA upgrade using response files
- Upgrading Volume Replicator
- Upgrading VirtualStore
- Upgrading SFCFSHA using Boot Environment upgrade
- Performing post-upgrade tasks
- Planning to upgrade SFCFSHA
- Section IV. Post-configuration tasks
- Section V. Configuration of disaster recovery environments
- Section VI. Adding and removing nodes
- Adding a node to SFCFSHA clusters
- Adding the node to a cluster manually
- Setting up the node to run in secure mode
- Adding a node using response files
- Configuring server-based fencing on the new node
- Removing a node from SFCFSHA clusters
- Adding a node to SFCFSHA clusters
- Section VII. Configuration and Upgrade reference
- Appendix A. Installation scripts
- Appendix B. Configuration files
- Appendix C. Configuring the secure shell or the remote shell for communications
- Appendix D. High availability agent information
- Appendix E. Sample SFCFSHA cluster setup diagrams for CP server-based I/O fencing
- Appendix F. Reconciling major/minor numbers for NFS shared disks
- Appendix G. Configuring LLT over UDP
- Using the UDP layer for LLT
- Manually configuring LLT over UDP using IPv4
- Using the UDP layer of IPv6 for LLT
- Manually configuring LLT over UDP using IPv6
Checking the major and minor number for VxVM volumes
The following sections describe checking and changing, if necessary, the major and minor numbers for the VxVM volumes that cluster systems use.
To check major and minor numbers on VxVM volumes
- Place the VCS command directory in your path. For example:
# export PATH=$PATH:/usr/sbin:/sbin:/opt/VRTS/bin
- To list the devices, use the ls -lL block_device command on each node:
# ls -lL /dev/vx/dsk/shareddg/vol3
On Node A, the output may resemble:
brw------- 1 root root 32,43000 Mar 22 16:4 1 /dev/vx/dsk/shareddg/vol3
On Node B, the output may resemble:
brw------- 1 root root 36,43000 Mar 22 16:4 1 /dev/vx/dsk/shareddg/vol3
- Import the associated shared disk group on each node.
- Use the following command on each node exporting an NFS file system. The command displays the major numbers for vxio and vxspec that Volume Manager (VxVM) uses . Note that other major numbers are also displayed, but only vxio and vxspec are of concern for reconciliation:
# grep vx /etc/name_to_major
Output on Node A:
vxdmp 30 vxio 32 vxspec 33 vxfen 87 vxglm 91
Output on Node B:
vxdmp 30 vxio 36 vxspec 37 vxfen 87 vxglm 91
- To change Node B's major numbers for vxio and vxspec to match those of Node A, use the command:
haremajor -vx major_number_vxio major_number_vxspec
For example, enter:
# haremajor -vx 32 33
If the command succeeds, proceed to step 8. If this command fails, you receive a report similar to the following:
Error: Preexisting major number 32 These are available numbers on this system: 128... Check /etc/name_to_major on all systems for available numbers.
- If you receive this report, use the haremajor command on Node A to change the major number (32/33) to match that of Node B (36/37). For example, enter:
# haremajor -vx 36 37
If the command fails again, you receive a report similar to the following:
Error: Preexisting major number 36 These are available numbers on this node: 126... Check /etc/name_to_major on all systems for available numbers.
- If you receive the second report, choose the larger of the two available numbers (in this example, 128). Use this number in the haremajor command to reconcile the major numbers. Type the following command on both nodes:
# haremajor -vx 128 129
- Reboot each node on which haremajor was successful.
- If the minor numbers match, proceed to reconcile the major and minor numbers of your next NFS block device.
- If the block device on which the minor number does not match is a volume, consult the vxdg(1M) manual page. The manual page provides instructions on reconciling the Volume Manager (VxVM) minor numbers, and gives specific reference to the reminor option.
Node where the vxio driver number have been changed require rebooting.