Storage Foundation for Sybase ASE CE 7.4.1 Configuration and Upgrade Guide - Linux
- Section I. Configuring SF Sybase ASE CE
- Preparing to configure SF Sybase CE
- Configuring SF Sybase CE
- Configuring the SF Sybase CE components using the script-based installer
- Configuring the SF Sybase CE cluster
- Configuring SF Sybase CE in secure mode
- Configuring a secure cluster node by node
- Configuring the SF Sybase CE cluster
- Configuring SF Sybase CE clusters for data integrity
- Setting up disk-based I/O fencing using installer
- Performing an automated SF Sybase CE configuration
- Performing an automated I/O fencing configuration using response files
- Configuring a cluster under VCS control using a response file
- Section II. Post-installation and configuration tasks
- Section III. Upgrade of SF Sybase CE
- Planning to upgrade SF Sybase CE
- Performing a full upgrade of SF Sybase CE using the product installer
- Performing an automated full upgrade of SF Sybase CE using response files
- Performing a phased upgrade of SF Sybase CE
- Performing a phased upgrade of SF Sybase CE from version 6.2.1 and later release
- Performing a rolling upgrade of SF Sybase CE
- Performing post-upgrade tasks
- Section IV. Installation and upgrade of Sybase ASE CE
- Installing, configuring, and upgrading Sybase ASE CE
- Preparing to configure the Sybase instances under VCS control
- Installing, configuring, and upgrading Sybase ASE CE
- Section V. Adding and removing nodes
- Adding a node to SF Sybase CE clusters
- Adding the node to a cluster manually
- Setting up the node to run in secure mode
- Adding the new instance to the Sybase ASE CE cluster
- Removing a node from SF Sybase CE clusters
- Adding a node to SF Sybase CE clusters
- Section VI. Configuration of disaster recovery environments
- Section VII. Installation reference
- Appendix A. Installation scripts
- Appendix B. Sample installation and configuration values
- Appendix C. Tunable files for installation
- Appendix D. Configuration files
- Sample main.cf files for Sybase ASE CE configurations
- Appendix E. Configuring the secure shell or the remote shell for communications
- Appendix F. High availability agent information
Coordinator disk requirements for I/O fencing
Make sure that the I/O fencing coordinator disks meet the following requirements:
For disk-based I/O fencing, you must have at least three coordinator disks or there must be odd number of coordinator disks.
The coordinator disks must be DMP devices.
Each of the coordinator disks must use a physically separate disk or LUN.
Veritas recommends using the smallest possible LUNs for coordinator disks.
Each of the coordinator disks should exist on a different disk array, if possible.
The coordinator disks must support SCSI-3 persistent reservations.
Coordinator devices can be attached over iSCSI protocol but they must be DMP devices and must support SCSI-3 persistent reservations.
Veritas recommends using hardware-based mirroring for coordinator disks.
Coordinator disks must not be used to store data or must not be included in disk groups that store user data.
Coordinator disks cannot be the special devices that array vendors use. For example, you cannot use EMC gatekeeper devices as coordinator disks.
The coordinator disk size must be at least 128 MB.