Please enter search query.
Search <book_title>...
InfoScale™ 9.0 Storage Foundation and High Availability Configuration and Upgrade Guide - Linux
Last Published:
2025-04-21
Product(s):
InfoScale & Storage Foundation (9.0)
Platform: Linux
- Section I. Introduction to SFHA
- Section II. Configuration of SFHA
- Preparing to configure
- Preparing to configure SFHA clusters for data integrity
- About planning to configure I/O fencing
- Setting up the CP server
- Configuring the CP server manually
- Configuring CP server using response files
- Configuring SFHA
- Configuring Storage Foundation High Availability using the installer
- Configuring a secure cluster node by node
- Completing the SFHA configuration
- Verifying and updating licenses on the system
- Configuring Storage Foundation High Availability using the installer
- Configuring SFHA clusters for data integrity
- Setting up disk-based I/O fencing using installer
- Setting up server-based I/O fencing using installer
- Manually configuring SFHA clusters for data integrity
- Setting up disk-based I/O fencing manually
- Setting up server-based I/O fencing manually
- Configuring server-based fencing on the SFHA cluster manually
- Setting up non-SCSI-3 fencing in virtual environments manually
- Setting up majority-based I/O fencing manually
- Performing an automated SFHA configuration using response files
- Performing an automated I/O fencing configuration using response files
- Section III. Upgrade of SFHA
- Planning to upgrade SFHA
- Preparing to upgrade SFHA
- Upgrading Storage Foundation and High Availability
- Performing a rolling upgrade of SFHA
- Performing a phased upgrade of SFHA
- About phased upgrade
- Performing a phased upgrade using the product installer
- Performing an automated SFHA upgrade using response files
- Upgrading SFHA using YUM
- Performing post-upgrade tasks
- Post-upgrade tasks when VCS agents for VVR are configured
- About enabling LDAP authentication for clusters that run in secure mode
- Planning to upgrade SFHA
- Section IV. Post-installation tasks
- Section V. Adding and removing nodes
- Adding a node to SFHA clusters
- Adding the node to a cluster manually
- Adding a node using response files
- Configuring server-based fencing on the new node
- Removing a node from SFHA clusters
- Removing a node from a SFHA cluster
- Removing a node from a SFHA cluster
- Adding a node to SFHA clusters
- Section VI. Configuration and upgrade reference
- Appendix A. Installation scripts
- Appendix B. SFHA services and ports
- Appendix C. Configuration files
- Appendix D. Configuring the secure shell or the remote shell for communications
- Appendix E. Sample SFHA cluster setup diagrams for CP server-based I/O fencing
- Appendix F. Configuring LLT over UDP
- Using the UDP layer for LLT
- Manually configuring LLT over UDP using IPv4
- Using the UDP layer of IPv6 for LLT
- Manually configuring LLT over UDP using IPv6
- About configuring LLT over UDP multiport
- Appendix G. Using LLT over RDMA
- Configuring LLT over RDMA
- Configuring RDMA over an Ethernet network
- Configuring RDMA over an InfiniBand network
- Tuning system performance
- Manually configuring LLT over RDMA
- Troubleshooting LLT over RDMA
Moving the service groups to the second subcluster
Perform the following steps to establish the service group's status and to switch the service groups.
To move service groups to the second subcluster
- On the first subcluster, determine where the service groups are online.
# hagrp -state
The output resembles:
#Group Attribute System Value sg1 State node01 |ONLINE| sg1 State node02 |ONLINE| sg1 State node03 |ONLINE| sg1 State node04 |ONLINE| sg2 State node01 |ONLINE| sg2 State node02 |ONLINE| sg2 State node03 |ONLINE| sg2 State node04 |ONLINE| sg3 State node01 |ONLINE| sg3 State node02 |OFFLINE| sg3 State node03 |OFFLINE| sg3 State node04 |OFFLINE| sg4 State node01 |OFFLINE| sg4 State node02 |ONLINE| sg4 State node03 |OFFLINE| sg4 State node04 |OFFLINE|
- Offline the parallel service groups (sg1 and sg2) from the first subcluster. Switch the failover service groups (sg3 and sg4) from the first subcluster (node01 and node02) to the nodes on the second subcluster (node03 and node04). For SFHA, vxfen sg is the parallel service group.
# hagrp -offline sg1 -sys node01 # hagrp -offline sg2 -sys node01 # hagrp -offline sg1 -sys node02 # hagrp -offline sg2 -sys node02 # hagrp -switch sg3 -to node03 # hagrp -switch sg4 -to node04
- On the nodes in the first subcluster, unmount all the VxFS file systems that VCS does not manage, for example:
# df -h Filesystem Size Used Avail Use% Mounted on /dev/sda1 26G 3.3G 22G 14% / udev 1007M 352K 1006M 1% /dev tmpfs 4.0K 0 4.0K 0% /dev/vx /dev/vx/dsk/dg2/dg2vol1 3.0G 18M 2.8G 1% /mnt/dg2/dg2vol1 /dev/vx/dsk/dg2/dg2vol2 1.0G 18M 944M 2% /mnt/dg2/dg2vol2 /dev/vx/dsk/dg2/dg2vol3 10G 20M 9.4G 1% /mnt/dg2/dg2vol3 # umount /mnt/dg2/dg2vol1 # umount /mnt/dg2/dg2vol2 # umount /mnt/dg2/dg2vol3
- On the nodes in the first subcluster, use the following command to take all the cache area offline:
# sfcache offline cachename
- On the nodes in the first subcluster, stop all VxVM volumes (for each disk group) that VCS does not manage.
- Make the configuration writable on the first subcluster.
# haconf -makerw
- Freeze the nodes in the first subcluster.
# hasys -freeze -persistent node01 # hasys -freeze -persistent node02
- Dump the configuration and make it read-only.
# haconf -dump -makero
- Verify that the service groups are offline on the first subcluster that you want to upgrade.
# hagrp -state
Output resembles:
#Group Attribute System Value sg1 State node01 |OFFLINE| sg1 State node02 |OFFLINE| sg1 State node03 |ONLINE| sg1 State node04 |ONLINE| sg2 State node01 |OFFLINE| sg2 State node02 |OFFLINE| sg2 State node03 |ONLINE| sg2 State node04 |ONLINE| sg3 State node01 |OFFLINE| sg3 State node02 |OFFLINE| sg3 State node03 |ONLINE| sg3 State node04 |OFFLINE| sg4 State node01 |OFFLINE| sg4 State node02 |OFFLINE| sg4 State node03 |OFFLINE| sg4 State node04 |ONLINE|