Veritas™ Volume Manager Administrator's Guide
- Understanding Veritas Volume Manager
- VxVM and the operating system
- How VxVM handles storage management
- Volume layouts in VxVM
- Online relayout
- Volume resynchronization
- Dirty region logging
- Volume snapshots
- FastResync
- Provisioning new usable storage
- Administering disks
- Disk devices
- Discovering and configuring newly added disk devices
- Discovering disks and dynamically adding disk arrays
- How to administer the Device Discovery Layer
- Changing the disk-naming scheme
- Adding a disk to VxVM
- Rootability
- Displaying disk information
- Removing disks
- Removing and replacing disks
- Administering Dynamic Multi-Pathing
- How DMP works
- Administering DMP using vxdmpadm
- Gathering and displaying I/O statistics
- Specifying the I/O policy
- Online dynamic reconfiguration
- Reconfiguring a LUN online that is under DMP control
- Creating and administering disk groups
- About disk groups
- Displaying disk group information
- Creating a disk group
- Importing a disk group
- Moving disk groups between systems
- Handling cloned disks with duplicated identifiers
- Handling conflicting configuration copies
- Reorganizing the contents of disk groups
- Destroying a disk group
- Creating and administering subdisks and plexes
- Displaying plex information
- Reattaching plexes
- Creating volumes
- Types of volume layouts
- Creating a volume
- Using vxassist
- Creating a volume on specific disks
- Creating a mirrored volume
- Creating a striped volume
- Creating a volume using vxmake
- Initializing and starting a volume
- Using rules and persistent attributes to make volume allocation more efficient
- Administering volumes
- Displaying volume information
- Monitoring and controlling tasks
- Reclamation of storage on thin reclamation arrays
- Stopping a volume
- Resizing a volume
- Adding a mirror to a volume
- Preparing a volume for DRL and instant snapshots
- Adding traditional DRL logging to a mirrored volume
- Enabling FastResync on a volume
- Performing online relayout
- Adding a RAID-5 log
- Creating and administering volume sets
- Configuring off-host processing
- Administering hot-relocation
- How hot-relocation works
- Moving relocated subdisks
- Administering cluster functionality (CVM)
- Overview of clustering
- Multiple host failover configurations
- CVM initialization and configuration
- Dirty region logging in cluster environments
- Administering VxVM in cluster environments
- Changing the CVM master manually
- Importing disk groups as shared
- Administering sites and remote mirrors
- About sites and remote mirrors
- Fire drill - testing the configuration
- Changing the site name
- Administering the Remote Mirror configuration
- Failure and recovery scenarios
- Performance monitoring and tuning
- Appendix A. Using Veritas Volume Manager commands
- Appendix B. Configuring Veritas Volume Manager
Displaying cumulative I/O statistics
Use the groupby
clause of the vxdmpadm iostat command to display cumulative I/O statistics listings per DMP node, controller, array port id, or host-array controller pair and enclosure. If the groupby
clause is not specified, then the statistics are displayed per path.
To group by DMP node:
# vxdmpadm iostat show groupby=dmpnode [all | dmpnodename=dmpnodename | enclosure=enclr-name]
To group by controller:
# vxdmpadm iostat show groupby=ctlr [ all | ctlr=ctlr ]
For example:
# vxdmpadm iostat show groupby=ctlr ctlr=c5
OPERATIONS BLOCKS AVG TIME(ms) CTLRNAME READS WRITES READS WRITES READS WRITES c5 224 14 54 7 4.20 11.10
To group by arrayport:
# vxdmpadm iostat show groupby=arrayport [ all | pwwn=array_pwwn | enclosure=enclr portid=array-port-id ]
For example:
# vxdmpadm iostat show groupby=arrayport enclosure=HDS9500-ALUA0 \ portid=1A
OPERATIONS BLOCKS AVG TIME(ms) PORTNAME READS WRITES READS WRITES READS WRITES 1A 224 14 54 7 4.20 11.10
To group by enclosure:
# vxdmpadm iostat show groupby=enclosure [ all | enclosure=enclr ]
For example:
# vxdmpadm iostat show groupby=enclosure enclosure=EMC_CLARiiON0
OPERATIONS BLOCKS AVG TIME(ms) ENCLRNAME READS WRITES READS WRITES READS WRITES EMC_CLARiiON 0 0 0 0 0.00 0.00
You can also filter out entities for which all data entries are zero. This option is especially useful in a cluster environment which contains many failover devices. You can display only the statistics for the active paths.
To filter all zero entries from the output of the iostat show command:
# vxdmpadm -z iostat show [all|ctlr=ctlr_name | dmpnodename=dmp_device_name | enclosure=enclr_name [portid=portid] | pathname=path_name|pwwn=port_WWN][interval=seconds [count=N]]
For example:
# vxdmpadm -z iostat show dmpnodename=c2t16d4
OPERATIONS BLOCKS AVG TIME(ms) PATHNAME READS WRITES READS WRITES READS WRITES c3t16d4 10 110 2 25 12.00 27.96 c2t17d4 20 126 4 29 9.50 19.41
You can now specify the units in which the statistics data is displayed. By default, the read/write times are displayed in milliseconds up to 2 decimal places. The throughput data is displayed in terms of 'BLOCKS' and the output is scaled, meaning that the small values are displayed in small units and the larger values are displayed in bigger units, keeping significant digits constant.The -u option accepts the following options:
| Displays throughput in kiloblocks. |
| Displays throughput in megablocks. |
| Displays throughput in gigablocks. |
| Displays throughput in exact number of bytes. |
| Displays average read/write time in microseconds. |
For example: To display average read/write times in microseconds.
# vxdmpadm -u us iostat show pathname=c2t17d4
OPERATIONS BLOCKS AVG TIME(microsec) PATHNAME READS WRITES READS WRITES READS WRITES c2t17d4 20 126 4 29 9500.00 19413.79