Please enter search query.
Search <book_title>...
Storage Foundation for Oracle® RAC 7.4.1 Administrator's Guide - Linux
Last Published:
2019-10-17
Product(s):
InfoScale & Storage Foundation (7.4.1)
Platform: Linux
- Section I. SF Oracle RAC concepts and administration
- Overview of Storage Foundation for Oracle RAC
- About Storage Foundation for Oracle RAC
- Component products and processes of SF Oracle RAC
- About Virtual Business Services
- Administering SF Oracle RAC and its components
- Administering SF Oracle RAC
- Starting or stopping SF Oracle RAC on each node
- Administering VCS
- Administering I/O fencing
- About the vxfentsthdw utility
- Testing the coordinator disk group using the -c option of vxfentsthdw
- About the vxfenadm utility
- About the vxfenclearpre utility
- About the vxfenswap utility
- Administering the CP server
- Administering CFS
- Administering CVM
- Changing the CVM master manually
- Administering Flexible Storage Sharing
- Backing up and restoring disk group configuration data
- Administering SF Oracle RAC global clusters
- Administering SF Oracle RAC
- Overview of Storage Foundation for Oracle RAC
- Section II. Performance and troubleshooting
- Troubleshooting SF Oracle RAC
- About troubleshooting SF Oracle RAC
- Troubleshooting I/O fencing
- Fencing startup reports preexisting split-brain
- Troubleshooting CP server
- Troubleshooting server-based fencing on the SF Oracle RAC cluster nodes
- Issues during online migration of coordination points
- Troubleshooting Cluster Volume Manager in SF Oracle RAC clusters
- Troubleshooting CFS
- Troubleshooting interconnects
- Troubleshooting Oracle
- Troubleshooting ODM in SF Oracle RAC clusters
- Prevention and recovery strategies
- Tunable parameters
- Troubleshooting SF Oracle RAC
- Section III. Reference
Shared disk group cannot be imported in SF Oracle RAC cluster
If you see a message resembling:
vxvm:vxconfigd:ERROR:vold_vcs_getnodeid(/dev/vx/rdmp/disk_name): local_node_id < 0 Please make sure that CVM and vxfen are configured and operating correctly
First, make sure that CVM is running. You can see the CVM nodes in the cluster by running the vxclustadm nidmap command.
# vxclustadm nidmap Name CVM Nid CM Nid State sys1 1 0 Joined: Master sys2 0 1 Joined: Slave
This above ouput shows that CVM is healthy, with system sys1 as the CVM master. If CVM is functioning correctly, then the output above is displayed when CVM cannot retrieve the node ID of the local system from the vxfen driver. This usually happens when port b is not configured.
To verify vxfen driver is configured
- Check the GAB ports with the command:
# gabconfig -a
Port b must exist on the local system.