NetBackup IT Analytics Data Collector Installation and Configuration Guide
- Section I. Introduction
- Introduction
- Install and configure a Data Collector
- Step-1: Choose operating system and complete prerequisites
- Installing the Data Collector software
- Configure SSL
- Section II. Data Protection
- Configuration for Veritas Backup Exec
- Configuration for Cohesity DataProtect
- Configuration for Commvault Simpana
- Open TCP/IP access to the Commvault database
- Set up a read-only user in the CommServe server
- Configuration for EMC Avamar
- Import EMC Avamar server information
- Configuration for EMC Data Domain backup
- Configuration for Dell EMC NetWorker backup & recovery
- Importing generic backup data
- Configuration for generic backup
- CSV format specification
- Configuration for HP Data Protector
- Architecture overview (HP Data Protector)
- Configure the Data Collector server in Cell Manager (HP Data Protector)
- Configuration for IBM Spectrum Protect (TSM)
- Architecture overview (IBM Spectrum Protect -TSM)
- Import IBM Spectrum Protect (TSM) information
- Configuration for NAKIVO Backup & Replication
- Configuration for Oracle Recovery Manager (RMAN)
- Configuration for Rubrik Cloud Data Management
- Configuration for Veeam Backup & Replication
- Configuration for Veritas Backup Exec
- Section III. Storage (Capacity)
- Configuration for Compute Resources
- Configuration for DELL EMC Elastic Cloud Storage (ECS)
- Configuration for Dell EMC Unity
- Configuration for EMC data domain storage
- Configuration for EMC Isilon
- Configuration for EMC Symmetrix
- Configuration for EMC VNX Celerra
- Configuration for EMC VNX CLARiiON
- Configuration for EMC VPLEX
- Configuration for EMC XtremIO
- Configuration for FUJITSU ETERNUS Data Collector
- Configuration for Hitachi Block
- Configuring a Hitachi Device manager user
- Configuration for Hitachi Content Platform (HCP)
- Hitachi content platform system management console
- Hitachi content platform tenant management console
- Configuration for Hitachi NAS
- Configuration for Hitachi Vantara All-Flash and Hybrid Flash Storage
- Configuration of Host inventory
- Host Access Privileges, Sudo Commands, Ports, and WMI Proxy Requirements
- Configure host Discovery policies to populate the host Inventory
- Validate host connectivity
- Host Inventory search and host Inventory export
- Configure and edit host probes
- Propagate Probe Settings: Copy Probes, Paste Probes
- Configuration for HP 3PAR
- Configuration for HP EVA
- Configuration for HPE Nimble Storage
- Configuration for HPE StoreOnce
- Configuration for IBM Enterprise
- Configuration for IBM COS
- Configuration for IBM SVC
- Configuration for IBM XIV
- Configuration for Microsoft Windows server
- Configuration for NetApp-7
- Configuration for NetApp StorageGRID
- Configuration for NetApp Cluster
- Configuration for NetApp E-Series
- Configuration for NEC HYDRAstor
- Configuration for Pure Storage FlashArray
- Section IV. Compute (Virtualization and Host Collection)
- Configuration for VMware
- Configuration for IBM VIO
- Configuration for Microsoft Hyper-V
- Section V. Cloud
- Configuration for Amazon Web Services (AWS)
- Mandatory probe user privileges
- Link AWS accounts for Collection of consolidated billing data
- Configuration for Google Cloud Platform
- Configuration for OpenStack Ceilometer
- Configuration for OpenStack Swift
- Configuration for Microsoft Azure
- Configuration for Amazon Web Services (AWS)
- Section VI. Fabric
- Configuration for Brocade switch
- Configuration for Cisco switch
- Configuration for Brocade Zone alias
- Configuration for Cisco Zone alias
- Configuration for Brocade switch
- Section VII. File Analytics
- Configuration for File Analytics
- Host Discovery and Collection File Analytics probe
- Adding a File Analytics Data Collector policy
- File Analytics Export Folder Size and Folder Depth
- Configuration for File Analytics
- Section VIII. Data Collection Validation and Troubleshooting
- Validate data collection
- Data Collector Troubleshooting
- Host resources: Check host connectivity using standard SSH
- Host resources: Generating host resource configuration files
- Configuring parameters for SSH
- Uninstalling the Data Collector
- Appendix A. Firewall Configuration: Default Ports
- Appendix B. Load historic events
- Load Veritas NetBackup events
- Appendix C. CRON Expressions for Policy and Report Schedules
- Appendix D. Maintenance Scenarios for Message Relay Server Certificate Generation
Manually loading the CSV file
Use the following procedure to manually load the Generic Backup CSV file into the Portal database.
Prerequisites:
These scripts must be run on the
.The
script must be run first to register the event collector ID.
List the Data Collectors to get the
and the , which will be used in step 2.Windows:
C:\opt\APTARE\mbs\bin\listcollectors.bat
Linux:
/opt/aptare/mbs/bin/listcollectors.sh
In the output, look for the Event Collectors section associated with the Software Home--the location of the CSV file (the path that was specified when the Data Collector Policy was created). Find the
and .==== Event Collectors ===
Event Collector Id: EVENT_1029161_9
Active: true Active: true Software Home: C:\gkgenericBackup.csv Server Address: 102961 Domain: gkdomain Group Id: 102961 Sub-system/Server Instance/Device Manager Id: 102961 Schedule: */10 * * * *
Use the following commands to load the data from the CSV file into the Portal database.
Windows:
C:\opt\APTARE\mbs\bin\loadGenericBackupData.bat <EventCollectorID> <ServerID> [verbose]
Linux:
/opt/aptare/mbs/bin/loadGenericBackupData.sh <EventCollectorID> <ServerID> [verbose]
Note:
If you run the command with no parameters, it will display the syntax.
The load script will check to see if the backup server and client already exist; if not, they will be added to the database. The script then checks for a backup job with the exact same backup server, client, start date and finish date. If no matches are found, the job will be added; otherwise, it will be ignored. This prevents duplicate entries and allows the import of the script to be repeated, if it has not been updated. Once the load is complete, these clients and jobs will be visible via the NetBackup IT Analytics Portal and the data will be available for reporting.