NetBackup™ Snapshot Manager Install and Upgrade Guide
- Introduction
- Section I. NetBackup Snapshot Manager installation and configuration
- Preparing for NetBackup Snapshot Manager installation
- Deploying NetBackup Snapshot Manager using container images
- Deploying NetBackup Snapshot Manager extensions
- Installing the NetBackup Snapshot Manager extension on a VM
- Installing the NetBackup Snapshot Manager extension on a managed Kubernetes cluster (AKS) in Azure
- Installing the NetBackup Snapshot Manager extension on a managed Kubernetes cluster (EKS) in AWS
- Installing the NetBackup Snapshot Manager extension on a managed Kubernetes cluster (GKE) in GCP
- NetBackup Snapshot Manager cloud providers
- Configuration for protecting assets on cloud hosts/VM
- Protecting assets with NetBackup Snapshot Manager's on-host agent feature
- Installing and configuring NetBackup Snapshot Manager agent
- Configuring the NetBackup Snapshot Manager application plug-in
- Microsoft SQL plug-in
- Oracle plug-in
- Protecting assets with NetBackup Snapshot Manager's agentless feature
- Snapshot Manager catalog backup and recovery
- NetBackup Snapshot Manager assets protection
- Volume Encryption in NetBackup Snapshot Manager
- NetBackup Snapshot Manager security
- Preparing for NetBackup Snapshot Manager installation
- Section II. NetBackup Snapshot Manager maintenance
- NetBackup Snapshot Manager logging
- Upgrading NetBackup Snapshot Manager
- Migrating and upgrading NetBackup Snapshot Manager
- Post-upgrade tasks
- Uninstalling NetBackup Snapshot Manager
- Troubleshooting NetBackup Snapshot Manager
Backup and restore jobs fail with timeout error
Due to reduced availability of resources on NetBackup Snapshot Manager server, backup and restore jobs fail as the jobs are in continuos search of memory due to which other services may also fail with the timeout error. This issue may be due to multiple jobs running together beyond the capacity of the host. On a cluster setup, the jobs may fail to schedule on nodes because of the maximum pods per node setting. The backup or restore jobs may fail, if the maximum pods per nodes are set to a lower number than the recommended value according to the node capability.
Workaround:
To resolve this issue, manually configure the following to set the maximum jobs that can run on a single node at a time:
host using the
/cloudpoint/flexsnap.conf
fileOr
cluster using the
flexsnap-conf
config map
[capability_limit] max_jobs = <num>
where, <num> is the maximum number of jobs that can run at a time on a node.
In case of multiple jobs running in parallel, if any service fails due to non availability of resources then reduce the number of parallel jobs that can be performed on the provided node type.