NetBackup™ Deployment Guide for Kubernetes Clusters
- Introduction
- Section I. Deployment
- Prerequisites for Kubernetes cluster configuration
- Deployment with environment operators
- Deploying NetBackup
- Primary and media server CR
- Deploying NetBackup using Helm charts
- Deploying MSDP Scaleout
- Deploying Snapshot Manager
- Section II. Monitoring and Management
- Monitoring NetBackup
- Monitoring MSDP Scaleout
- Monitoring Snapshot Manager
- Managing the Load Balancer service
- Managing MSDP Scaleout
- Performing catalog backup and recovery
- Section III. Maintenance
- MSDP Scaleout Maintenance
- Upgrading
- Uninstalling
- Troubleshooting
- Troubleshooting AKS and EKS issues
- Troubleshooting AKS-specific issues
- Troubleshooting EKS-specific issues
- Troubleshooting AKS and EKS issues
- Appendix A. CR template
Taint, Toleration, and Node affinity related issues in cpServer
If one of the following cpServer control pool pod is in pending state, then perform the steps that follow:
flexsnap-agent, flexsnap-api-gateway, flexsnap-certauth, flexsnap-coordinator, flexsnap-idm, flexsnap-nginx, flexsnap-notification, flexsnap-scheduler, flexsnap-, flexsnap-, flexsnap-fluentd-, flexsnap-fluentd
Obtain the pending pod's toleration and affinity status using the following command:
kubectl get pods <pod name>
Check if the node-affinity and tolerations of pod are matching with:
fields listed in
or in theenvironment.yaml
file.taint and label of node pool, mentioned in
or in theenvironment.yaml
file.
If all the above fields are correct and matching and still the control pool pod is in pending state, then the issue may be due to all the nodes in nodepool running at maximum capacity and cannot accommodate new pods. In such case the noodpool must be scaled properly.
If one of the following cpServer data pool pod is in pending state, then perform the steps that follow:
flexsnap-listener,flexsnap-workflow,flexsnap-datamover
Obtain the pending pod's toleration and affinity status using the following command:
kubectl get pods <pod name>
Check if the node-affinity and tolerations of pod are matching with:
fields listed in
in theenvironment.yaml
file.taint and label of node pool, mentioned in
in theenvironment.yaml
file.
If all the above fields are correct and matching and still the control pool pod is in pending state, then the issue may be due to all the nodes in nodepool running at maximum capacity and cannot accommodate new pods. In such case the noodpool must be scaled properly.
Obtain the pending pod's toleration and affinity status using the following command:
kubectl get pods <pod name>
Check if the node-affinity and tolerations of pod are matching with:
fields listed in
file.taint and label of node pool, mentioned in above values.
If all the above fields are correct and matching and still the control pool pod is in pending state, then the issue may be due to all the nodes in nodepool running at maximum capacity and cannot accommodate new pods. In such case the noodpool must be scaled properly.
If the nodes are configured with incorrect taint and label values, the user can edit them using the following command provided for AKS as an example:
az aks nodepool update \ --resource-group <resource_group> \ --cluster-name <cluster_name> \ --name <nodepool_name> \ --node-taints <key>=<value>:<effect> \ --no-wait
az aks nodepool update \ --resource-group <resource_group> \ --cluster-name <cluster_name> \ --name <cluster_name> \ --labels <key>=<value>