NetBackup™ Snapshot Manager Install and Upgrade Guide

Last Published:
Product(s): NetBackup (10.2)
  1. Introduction
    1.  
      About the deployment approach
    2.  
      Deciding where to run NetBackup Snapshot Manager
    3.  
      About deploying NetBackup Snapshot Manager in the cloud
  2. Section I. NetBackup Snapshot Manager installation and configuration
    1. Preparing for NetBackup Snapshot Manager installation
      1.  
        Meeting system requirements
      2.  
        NetBackup Snapshot Manager host sizing recommendations
      3.  
        NetBackup Snapshot Manager extension sizing recommendations
      4.  
        Creating an instance or preparing the host to install NetBackup Snapshot Manager
      5.  
        Installing container platform (Docker, Podman)
      6.  
        Creating and mounting a volume to store NetBackup Snapshot Manager data
      7.  
        Verifying that specific ports are open on the instance or physical host
      8.  
        Preparing NetBackup Snapshot Manager for backup from snapshot jobs
    2. Deploying NetBackup Snapshot Manager using container images
      1.  
        Before you begin installing NetBackup Snapshot Manager
      2.  
        Installing NetBackup Snapshot Manager in the Docker/Podman environment
      3.  
        Verifying that NetBackup Snapshot Manager is installed successfully
      4.  
        Restarting NetBackup Snapshot Manager
    3. Deploying NetBackup Snapshot Manager extensions
      1.  
        Before you begin installing NetBackup Snapshot Manager extensions
      2.  
        Downloading the NetBackup Snapshot Manager extension
      3. Installing the NetBackup Snapshot Manager extension on a VM
        1.  
          Prerequisites to install the extension on VM
        2.  
          Installing the extension on a VM
      4. Installing the NetBackup Snapshot Manager extension on a managed Kubernetes cluster (AKS) in Azure
        1.  
          Prerequisites to install the extension on a managed Kubernetes cluster in Azure
        2.  
          Installing the extension on Azure (AKS)
      5. Installing the NetBackup Snapshot Manager extension on a managed Kubernetes cluster (EKS) in AWS
        1.  
          Prerequisites to install the extension on a managed Kubernetes cluster in AWS
        2. Installing the extension on AWS (EKS)
          1.  
            Install extension using the extension script
      6. Installing the NetBackup Snapshot Manager extension on a managed Kubernetes cluster (GKE) in GCP
        1.  
          Prerequisites to install the extension on a managed Kubernetes cluster in GCP
        2.  
          Installing the extension on GCP (GKE)
      7.  
        Install extension using the Kustomize and CR YAMLs
      8.  
        Managing the extensions
    4. NetBackup Snapshot Manager cloud providers
      1.  
        Why to configure the NetBackup Snapshot Manager cloud providers?
      2. AWS plug-in configuration notes
        1.  
          Prerequisites for configuring the AWS plug-in
        2.  
          Configuring AWS permissions for NetBackup Snapshot Manager
        3.  
          AWS permissions required by NetBackup Snapshot Manager
        4.  
          Before you create a cross account configuration
      3. Google Cloud Platform plug-in configuration notes
        1.  
          Google Cloud Platform permissions required by NetBackup Snapshot Manager
        2.  
          Configuring a GCP service account for NetBackup Snapshot Manager
        3.  
          Preparing the GCP service account for plug-in configuration
        4.  
          GCP cross-project restore configuration
      4. Microsoft Azure plug-in configuration notes
        1.  
          Configuring permissions on Microsoft Azure
        2.  
          About Azure snapshots
      5. Microsoft Azure Stack Hub plug-in configuration notes
        1.  
          Configuring permissions on Microsoft Azure Stack Hub
        2.  
          Configuring staging location for Azure Stack Hub VMs to restore from backup
    5. Configuration for protecting assets on cloud hosts/VM
      1.  
        Deciding which feature (on-host agent or agentless) of NetBackup Snapshot Manager is to be used for protecting the assets
      2. Protecting assets with NetBackup Snapshot Manager's on-host agent feature
        1. Installing and configuring NetBackup Snapshot Manager agent
          1.  
            Downloading and installing the NetBackup Snapshot Manager agent
          2. Linux-based agent
            1.  
              Preparing to install the Linux-based agent
            2.  
              Registering the Linux-based agent
          3. Windows-based agent
            1.  
              Preparing to install the Windows-based agent
            2.  
              Registering the Windows-based agent
        2. Configuring the NetBackup Snapshot Manager application plug-in
          1.  
            Configuring an application plug-in
          2. Microsoft SQL plug-in
            1.  
              Microsoft SQL plug-in configuration requirements
            2.  
              Restore requirements and limitations for Microsoft SQL Server
            3.  
              Steps required before restoring SQL AG databases
            4.  
              Additional steps required after restoring SQL AG databases
            5. Additional steps required after a SQL Server instance snapshot restore
              1.  
                Steps required after a SQL Server host-level restore
              2.  
                Steps required after a SQL Server instance disk-level snapshot restore to new location
          3. Oracle plug-in
            1. Oracle plug-in configuration requirements
              1.  
                Optimizing your Oracle database data and metadata files
            2.  
              Restore requirements and limitations for Oracle
            3.  
              Additional steps required after an Oracle snapshot restore
      3. Protecting assets with NetBackup Snapshot Manager's agentless feature
        1. Prerequisites for the agentless configuration
          1.  
            Configuring SMB for Windows (Optional)
          2.  
            Configuring WMI security for Windows (optional)
        2.  
          Configuring the agentless feature
        3.  
          Configuring the agentless feature after upgrading NetBackup Snapshot Manager
    6. NetBackup Snapshot Manager assets protection
      1. NetBackup protection plan
        1.  
          Creating a NetBackup protection plan for cloud assets
        2.  
          Subscribing cloud assets to a NetBackup protection plan
      2.  
        Configuring VSS to store shadow copies on the originating drive
      3.  
        Additional steps required after restoring an AWS RDS database instance
    7. Volume Encryption in NetBackup Snapshot Manager
      1.  
        About volume encryption support in NetBackup Snapshot Manager
      2.  
        Volume encryption for Azure
      3.  
        Volume encryption for GCP
      4.  
        Volume encryption for AWS
    8. NetBackup Snapshot Manager security
      1.  
        Configuring security for Azure Stack
      2.  
        Configuring the cloud connector for Azure Stack
      3.  
        CA configuration for Azure Stack
      4.  
        Securing the connection to NetBackup Snapshot Manager
  3. Section II. NetBackup Snapshot Manager maintenance
    1. NetBackup Snapshot Manager logging
      1.  
        About NetBackup Snapshot Manager logging mechanism
      2. How Fluentd-based NetBackup Snapshot Manager logging works
        1.  
          About the NetBackup Snapshot Manager fluentd configuration file
        2.  
          Modifying the fluentd configuration file
      3.  
        NetBackup Snapshot Manager logs
      4.  
        Agentless logs
      5.  
        Troubleshooting NetBackup Snapshot Manager logging
    2. Upgrading NetBackup Snapshot Manager
      1.  
        About NetBackup Snapshot Manager upgrades
      2.  
        Supported upgrade path
      3.  
        Upgrade scenarios
      4.  
        Preparing to upgrade NetBackup Snapshot Manager
      5.  
        Upgrading NetBackup Snapshot Manager
      6.  
        Upgrading NetBackup Snapshot Manager using patch or hotfix
      7. Migrating and upgrading NetBackup Snapshot Manager
        1.  
          Before you begin migrating NetBackup Snapshot Manager
        2.  
          Migrate and upgrade NetBackup Snapshot Manager on RHEL 8.6 or 8.4
      8.  
        GCP configuration for migration from zone to region
      9. Post-upgrade tasks
        1.  
          Upgrading NetBackup Snapshot Manager extensions
      10.  
        Post-migration tasks
    3. Uninstalling NetBackup Snapshot Manager
      1.  
        Preparing to uninstall NetBackup Snapshot Manager
      2.  
        Backing up NetBackup Snapshot Manager
      3.  
        Unconfiguring NetBackup Snapshot Manager plug-ins
      4.  
        Unconfiguring NetBackup Snapshot Manager agents
      5.  
        Removing the NetBackup Snapshot Manager agents
      6.  
        Removing NetBackup Snapshot Manager from a standalone Docker host environment
      7.  
        Removing NetBackup Snapshot Manager extensions - VM-based or managed Kubernetes cluster-based
      8.  
        Restoring NetBackup Snapshot Manager
    4. Troubleshooting NetBackup Snapshot Manager
      1.  
        Troubleshooting NetBackup Snapshot Manager
      2.  
        SQL snapshot or restore and granular restore operations fail if the Windows instance loses connectivity with the NetBackup Snapshot Manager host
      3.  
        Disk-level snapshot restore fails if the original disk is detached from the instance
      4.  
        Discovery is not working even after assigning system managed identity to the control node pool
      5.  
        Performance issue with GCP backup from snapshot
      6.  
        Post migration on host agents fail with an error message
      7.  
        File restore job fails with an error message
      8.  
        Acknowledgment not received for datamover
      9.  
        Upgrade of extension on AWS (EKS) fails when upgrading through script
      10.  
        Backup from snapshot job fails with timeout error

Troubleshooting NetBackup Snapshot Manager

Refer to the following troubleshooting scenarios:

  • NetBackup Snapshot Manager agent fails to connect to the NetBackup Snapshot Manager server if the agent host is restarted abruptly.

    This issue may occur if the host where the NetBackup Snapshot Manager agent is installed is shut down abruptly. Even after the host restarts successfully, the agent fails to establish a connection with the NetBackup Snapshot Manager server and goes into an offline state.

    The agent log file contains the following error:

    Flexsnap-agent-onhost[4972] mainthread flexsnap.connectors.rabbitmq: error - channel 1 closed unexpectedly: (405) resource_locked - cannot obtain exclusive access to locked queue ' flexsnap-agent.a1f2ac945cd844e393c9876f347bd817' in vhost '/'

    This issue occurs because the RabbitMQ connection between the agent and the NetBackup Snapshot Manager server does not close even in case of an abrupt shutdown of the agent host. The NetBackup Snapshot Manager server cannot detect the unavailability of the agent until the agent host misses the heartbeat poll. The RabbitMQ connection remains open until the next heartbeat cycle. If the agent host reboots before the next heartbeat poll is triggered, the agent tries to establish a new connection with the NetBackup Snapshot Manager server. However, as the earlier RabbitMQ connection already exists, the new connection attempt fails with a resource locked error.

    As a result of this connection failure, the agent goes offline and leads to a failure of all snapshot and restore operations performed on the host.

    Workaround:

    Restart the Veritas NetBackup Snapshot Manager Agent service on the agent host.

    • On a Linux hosts, run the following command:

      # sudo systemctl restart flexsnap-agent.service

    • On Windows hosts:

      Restart the Veritas NetBackup Snapshot Manager™ Agent service from the Windows Services console.

  • NetBackup Snapshot Manager agent registration on Windows hosts may time out or fail.

    For protecting applications on Windows, you need to install and then register the NetBackup Snapshot Manager agent on the Windows host. The agent registration may sometimes take longer than usual and may either time out or fail.

    Workaround:

    To resolve this issue, try the following steps:

    • Re-register the agent on the Windows host using a fresh token.

    • If the registration process fails again, restart the NetBackup Snapshot Manager services on the NetBackup Snapshot Manager server and then try registering the agent again.

    Refer to the following for more information:

    See Registering the Windows-based agent.

    See Restarting NetBackup Snapshot Manager.

  • Disaster recovery when DR package is lost or passphrase is lost.

    This issue may occur if the DR package is lost or the passphrase is lost.

    In case of Catalog backup, 2 backup packages are created:

    • DR package which contains all the certs

    • Catalog package which contains the data base

    The DR package contains the NetBackup UUID certs and Catalog DB also has the UUID. When you perform disaster recovery using the DR package followed by catalog recovery, both the UUID cert and the UUID are restored. This allows NetBackup to communicate with NetBackup Snapshot Manager since the UUID is not changed.

    However if the DR package is lost or the Passphrase is lost the DR operation cannot be completed. You can only recover the catalog without DR package after you reinstall NetBackup. In this case, a new UUID is created for NetBackup which is not recognised by NetBackup Snapshot Manager. The one-to-one mapping of NetBackup and NetBackup Snapshot Manager is lost.

    Workaround:

    To resolve this issue, you must update the new NBU UUID and Version Number after NetBackup primary is created.

    • The NetBackup administrator must be logged on to the NetBackup Web Management Service to perform this task. Use the following command to log on:

      /usr/openv/netbackup/bin/bpnbat -login -loginType WEB

    • Execute the following command on the primary server to get the NBU UUID:

      /usr/openv/netbackup/bin/admincmd/nbhostmgmt -list -host <primary server host name> | grep "Host ID"

    • Execute the following command to get the Version Number:

      /usr/openv/netbackup/bin/admincmd/bpgetconfig -g <primary Ssrver host name> -L

    After you get the NBU UUID and Version number, execute the following command on the NetBackup Snapshot Manager host to update the mapping:

    /cloudpoint/scripts/cp_update_nbuuid.sh -i <NBU UUID> -v <Version Number>

  • The snapshot job is successful but backup job fails with error "The NetBackup Snapshot Managers certificate is not valid or doesn't exist.(9866)" when ECA_CRL_CHECK disabled on master server.

    If ECA_CRL_CHECK is configured on master server and is disabled then it must be configured in bp.conf on NetBackup Snapshot Manager setup with same value.

    For example, considering a scenario of backup from snapshot where NetBackup is configured with external certificate and certificate is revoked. In this case, if ECA_CRL_CHECK is set as DISABLE on master then set the same value in bp.conf of NetBackup Snapshot Manager setup, otherwise snapshot operation will be successful and backup operation will fail with the certificate error.

    See Configuring security for Azure Stack .

  • NetBackup Snapshot Manager fails to establish connection using agentless to the Windows cloud instance

    Error 1: <Instance_name>: network connection timed out.

    Case 1: NetBackup Snapshot Manager server log message:

    WARNING - Cannot connect to the remote host. SMB Connection timeout
     <IP address> <user>
    
    …
    
    flexsnap.OperationFailed: Could not connect to the remote server 
    <IP address>

    Workaround:

    To resolve this issue, try the following steps:

    • Verify if the SMB port 445 is added in the Network security group and is accessible from the NetBackup Snapshot Manager.

    • Verify if the SMB port 445 is allowed through cloud instance firewall.

    Case 2: NetBackup Snapshot Manager log message:

    WARNING - Cannot connect to the remote host. WMI Connection 
    timeout <IP address> <user>
    
    …
    
    flexsnap.OperationFailed: Could not connect to the remote 
    server <IP address>

    Workaround:

    To resolve this issue, try the following steps:

    • Verify if the DCOM port (135) is added in the Network security group and is accessible from NetBackup Snapshot Manager.

    • Verify if the port 135 is allowed through cloud instance firewall.

    Case 3: NetBackup Snapshot Manager log message:

    Exception while opening SMB connection, [Errno Connection error 
    (<IP address>:445)] [Errno 113] No route to host.

    Workaround:: Verify if the cloud instance is up and running or not in inconsistent state.

    Case 4: NetBackup Snapshot Manager log message:

    Error when closing dcom connection: 'Thread-xxxx'"

    Where, xxxx is the thread number.

    Workaround::

    To resolve this issue, try the following steps:

    • Verify if the WMI-IN dynamic port range or the fixed port as configured is added in the Network security group.

    • Verify and enable WMI-IN port from the cloud instance firewall.

    Error 2: <Instance_name>: Could not connect to the virtual machine.

    NetBackup Snapshot Manager log message:

    Error: Cannot connect to the remote host. <IP address> Access denied. 

    Workaround::

    To resolve this issue, try the following steps:

    • Verify if the user is having administrative rights.

    • Verify if the UAC is disabled for the user.

  • NetBackup Snapshot Manager cloud operations fail on a RHEL system if a firewall is disabled

    The NetBackup Snapshot Manager operations fail for all the supported cloud plugins on a RHEL system, if a firewall is disabled on that system when the NetBackup Snapshot Manager services are running. This is a network configuration issue that prevents the NetBackup Snapshot Manager from accessing the cloud provider REST API endpoints.

    Workaround:

    • Stop NetBackup Snapshot Manager

      # docker run --rm -it -u 0 -v /var/run/docker.sock:/var/run/docker.sock -v /cloudpoint:/cloudpoint veritas/flexsnap-deploy:<version> stop

      For RHEL 8.x, stop NetBackup Snapshot Manager by using the following podman command:

      podman run -it --rm -u 0 -v /cloudpoint:/cloudpoint -v /run/podman/podman.sock:/run/podman/podman.sock veritas/flexsnap-deploy: <version> stop

    • Restart Docker

      # systemctl restart docker

    • Restart NetBackup Snapshot Manager

      # docker run --rm -it -u 0 -v /var/run/docker.sock:/var/run/docker.sock -v /cloudpoint:/cloudpoint veritas/flexsnap-deploy:<version> start

      For RHEL 8.x, restart NetBackup Snapshot Manager by using the following podman command:

      podman run -it --rm -u 0 -v /cloudpoint:/cloudpoint -v /run/podman/podman.sock:/run/podman/podman.sock veritas/flexsnap-deploy: <version> start

  • Backup from Snapshot job and Indexing job fails with the errors

    Jun 10, 2021 2:17:48 PM - Error mqclient (pid=1054) SSL
    Connection failed with string, broker:<hostname>
    Jun 10, 2021 2:17:48 PM - Error mqclient (pid=1054) Failed SSL
    handshake, broker:<hostname>
    Jun 10, 2021 2:19:16 PM - Error nbcs (pid=29079) Invalid
    operation for asset: <asset_id>
    Jun 10, 2021 2:19:16 PM - Error nbcs (pid=29079) Acknowledgement
    not received for datamover <datamover_id>

    and/or

    Jun 10, 2021 3:06:13 PM - Critical bpbrm (pid=32373) from client
    <asset_id>: FTL - Cannot retrieve the exported snapshot details
    for the disk with UUID:<disk_asset_id>
    Jun 10, 2021 3:06:13 PM - Info bptm (pid=32582) waited for full
    buffer 1 times, delayed 220 times
    Jun 10, 2021 3:06:13 PM - Critical bpbrm (pid=32373) from client
    <asset_id>: FTL - cleanup() failed, status 6

    This can happen when the inbound access to NetBackup Snapshot Manager on port 5671 and 443 port gets blocked at the OS firewall level (firewalld). Hence, from the datamover container (used for the Backup from Snapshot and Indexing jobs), communication to NetBackup Snapshot Manager gets blocked. This results in the datamover container not being able to start the backup or indexing.

    Workaround:

    Modify the rules in OS firewall to allow the inbound connection from 5671 and 443 port.

  • Agentless connection fails for a VM with an error message.

    Agentless connection fails for a VM with the following error message when user changes the authentication type from SSH Key based to password based for a VM through the portal:

    User does not have the required privileges to establish an 
    agentless connection

    This issue occurs when the permissions are not defined correctly for the user in the sudoers file as mentioned in the above error message.

    Workaround:

    Resolve the sudoers file issue for the user by providing the required permissions to perform the passwordless sudo operations.

  • When NetBackup Snapshot Manager is deployed in private subnet (without internet) NetBackup Snapshot Manager function fails

    This issue occurs when NetBackup Snapshot Manager is deployed in private network where firewall is enabled or public IP which is disabled. The customer's information security team would not allow full internet access to the virtual machine's.

    Workaround:

    Enable the ports from the firewall command line using the following commands:

    firewall-cmd --add-port=22/tcp

    firewall-cmd --add-port=5671/tcp

    firewall-cmd --add-port=443/tcp

  • Restoring asset from backup copy fails

    In some of the scenarios it is observed that the connection resets intermittently in Docker container. Due to this the server sends more tcp payload than the advertised client window. Sometimes Docker container drops SYN+ACK packet from new TCP connection handshake. To allow these packets, use the nf_conntrack_tcp_be_liberal option.

    If nf_conntrack_tcp_be_liberal = 1 then the following packets are allowed:

    • ACK is under the lower bound (possible overly delayed ACK)

    • ACK is over the upper bound (ACKed data not seen yet)

    • SEQ is under the lower bound (already ACKed data retransmitted)

    • SEQ is over the upper bound (over the window of the receiver)

    If nf_conntrack_tcp_be_liberal = 0 then those are also rejected as invalid.

    Workaround:

    To resolve the issue of restore from backup copy, use the nf_conntrack_tcp_be_liberal = 1 option and set this value on node where datamover container is running.

    Use the following command for setting the value of nf_conntrack_tcp_be_liberal:

    sysctl -w net.netfilter.nf_conntrack_tcp_be_liberal=1

  • Some pods on Kubernetes extension progressed to completed state

    Workaround:

    Disable Kubernetes extension.

    Delete listener pod using the following command:

    #kubectl delete pod flexnsap-listener-xxxxx -n <namespace>

    Enable Kubernetes extension.

  • User is not able to customize a cloud protection plan

    Workaround:

    Create a new protection plan with the desired configuration and assign it to the asset.

  • Podman container not starting or containers are not up after reboot

    On RHEL 8.x platform, restarting container or machine reboot, the container displays the following error message:

    # podman restart flexsnap-coordinator 47ca97002e53de808cb8d0526ae033d4b317d5386ce085a8bce4cd434264afdf:
     "2022-02-05T04:53:42.265084989+00:00 Feb 05 04:53:42 flexsnap-coordinator flexsnap-coordinator[7] 
    agent_container_health_check flexsnap.container_manager: INFO - Response: b'{""cause"":""that name is already in use"",""message"":
    ""error creating container storage: the container name \\""flexsnap-agent.15bd0aea11164f7ba29e944115001d69\\"" is already in use by 
    \\""30f031d586b1ab524511601aad521014380752fb127a9440de86a81b327b6777\\"". You have to remove that container to be able to reuse that
     name.: that name is already in use"",""response"":500}\n'"

    Workaround:

    Check if there is a file with IP address entry mapping to the container that could not be started at/var/lib/cni/networks/flexsnap-network/ file system location.

    [ec2-user@ip-172-31-44-163 ~]$ ls -latr /var/lib/cni/networks/flexsnap-network/ total 16 -rwxr-x---. 1 root root 0 Jan 22 12:30 lock drwxr-xr-x. 4 root root 44 Jan 22 12:30 .. -rw-r--r--. 1 root root 70 Feb 4 14:47 10.89.0.150 -rw-r--r--. 1 root root 70 Feb 4 14:47 10.89.0.151 -rw-r--r--. 1 root root 70 Feb 4 14:47 10.89.0.152 -rw-r--r--. 1 root root 11 Feb 7 11:09 last_reserved_ip.0 drwxr-xr-x. 2 root root 101 Feb 7 11:13 . [ec2-user@ip-172-31-44-163 ~]$

    From the above directory , delete the duplicate IP address file and perform the stop and start operation as follows:

    Stop the container: #podman stop <container_name>

    Start the container:#podman start <container_name>

  • After starting the start/stop services, NetBackup Snapshot Manager, RabbitMQ and MongoDB containers are still in the starting state

    It was observed that flexsnap-mongodb and flexsnap-rabbitmq containers did not go into healthy state. Following is the state of flexsnap-mongodb container:

    [ec2-user@ip-172-31-23-60 log]$ sudo podman container inspect --format='{{json .Config.Healthcheck}}'
    flexsnap-mongodb {"Test":["CMD-SHELL","echo 'db.runCommand({ping: 1}).ok' 
    | mongo --ssl --sslCAFile /cloudpoint/keys/cacert.pem 
    --sslPEMKeyFile /cloudpoint/keys/mongodb.pem flexsnap-mongodb:27017/zenbrain --quiet"],
    "Interval":60,"Timeout":30000000000,"Retries":3} [ec2-user@ip-172-31-23-60 log]$ sudo podman container inspect --format='
    {{json .State.Healthcheck}}' flexsnap-mongodb {"Status":"starting","FailingStreak":0,"Log":null} [ec2-user@ip-172-31-23-60 log]$

    Workaround:

    Run the following #podman CLI(s) command:

    [ec2-user@ip-172-31-23-60 log]$ sudo podman healthcheck run flexsnap-mongodb
    
    [ec2-user@ip-172-31-23-60 log]$ sudo podman ps -a
    
    CONTAINER ID   IMAGE                        COMMAND                         CREATED      STATUS                  PORTS                           NAMES
    
    fe8cf001032b  localhost/veritas/   flexsnap-fluentd:10.0.0.0.9817         2 days ago    Up 45 hours ago      0.0.0.0:24224->24224/tcp        flexsnap-fluentd
    
    2c00500c1ac6  localhost/veritas/   flexsnap-mongodb:10.0.0.0.9817         2 days ago    Up 45 hours ago (healthy)                            flexsnap-mongodb
    
    7ab3e248024a  localhost/veritas/   flexsnap-rabbitmq:10.0.0.0.9817        2 days ago    Up 45 hours ago (starting)                           flexsnap-rabbitmq
    
    [ec2-user@ip-172-31-23-60 log]$ sudo podman healthcheck run flexsnap-rabbitmq
    
    [ec2-user@ip-172-31-23-60 log]$ sudo podman ps -a
    
    CONTAINER ID   IMAGE                        COMMAND                        CREATED       STATUS                      PORTS                         NAMES
    
    fe8cf001032b  localhost/veritas/    flexsnap-fluentd:10.0.0.0.9817        2 days ago    Up 45 hours ago          0.0.0.0:24224->24224/tcp     flexsnap-fluentd
    
    2c00500c1ac6  localhost/veritas/    flexsnap-mongodb:10.0.0.0.9817        2 days ago    Up 45 hours ago (healthy)                             flexsnap-mongodb
    
    7ab3e248024a  localhost/veritas/    flexsnap-rabbitmq:10.0.0.0.9817       2 days ago    Up 45 hours ago (healthy)                             flexsnap-rabbitmq
    
     [ec2-user@ip-172-31-23-60 log]$ sudo podman container inspect --format='{{json .State.Healthcheck}}' flexsnap-mongodb
    
    {"Status":"healthy","FailingStreak":0,"Log":[{"Start":"2022-02-14T07:32:13.051150432Z","End":"2022-02-14T07:32:13.444636429Z","ExitCode":0,"Output":""}]}
    
    [ec2-user@ip-172-31-23-60 log]$ sudo podman container inspect --format='{{json .State.Healthcheck}}' flexsnap-rabbitmq
    
    {"Status":"healthy","FailingStreak":0,"Log":[{"Start":"2022-02-14T07:32:46.537804403Z","End":"2022-02-14T07:32:47.293695744Z","ExitCode":0,"Output":""}]}
    
    [ec2-user@ip-172-31-23-60 log]$
  • Certificate generation would fail while registering NetBackup Snapshot Manager with NetBackup

    Starting NetBackup Snapshot Manager release 9.1.2, NetBackup certificate generation will happen synchronously with registration in register API of NetBackup Snapshot Manager. Hence, any failure in certificate generation will cause failure while registering NetBackup Snapshot Manager with NetBackup, that is ading or editing the NetBackup Snapshot Manager entry from Web UI. These certificates are used for datamover which is launched for operations like backup from snapshot, restore from backup, indexing (VxMS based), and so on. Hence, if certificate generation fails, these jobs cannot be performed. Hence NetBackup Snapshot Manager on cloud VMs cannot connect to NetBackup on lab VMs, hence the registration will fail, and hence NetBackup Snapshot Manager cannot be added to NetBackup.

    Workaround:

    To add NetBackup Snapshot Manager in such scenario requires to skip certificate generation on NetBackup Snapshot Manager by adding the following entry in /cloudpoint/flexsnap.conf file:

    [client_registration] skip_certificate_generation = yes

  • Default timeout of 6 hours is not allowing restore of larger database (size more than 300 GB)

    Workaround:

    Configurable timeout parameter value can be set to restore larger database. The timeout value can be specified in /etc/flexsnap.conf file of flexsnap-coordinator container. It does not require restart of the coordinator container. Timeout value would be picked up in next database restore job.

    User must specify the timeout value in seconds as follows:

    docker exec -it flexsnap-coordinator bash root@flexsnap-coordinator:/# cat /etc/flexsnap.conf [global] target = flexsnap-rabbitmq grt_timeout = 39600

  • Agentless connection and granular restore to restored host fails when the VM restored from backup has 50 tags attached to it

    Workaround:

    (For AWS) If a Windows VM restored from backup has 50 tags and platform tag does not exists, user can remove any tag that is not required and add the Platform: windows tag.

  • For few GKE versions, failed pod issues are observed in namespace

    Following few failed pods in namespace is observed with failure status as NodeAffinity:

    $ kubectl get pods -n <cp_extension_namespace>
    
    NAME                                        READY     STATUS       RESTARTS     AGE
    flexsnap-datamover-
    2fc2967943ba4ded8ef653318107f49c-664tm        0/1     Terminating    0          4d14h
    flexsnap-fluentd-collector-c88f8449c-5jkqh    0/1     NodeAffinity   0          3d15h
    flexsnap-fluentd-collector-c88f8449c-ph8mx    0/1     NodeAffinity   0          39h
    flexsnap-fluentd-collector-c88f8449c-rqw7w    1/1     Running        0          10h
    flexsnap-fluentd-collector-c88f8449c-sswzr    0/1     NodeAffinity   0          5d18h
    flexsnap-fluentd-ftlnv                        1/1     Running        3 (10h ago)10h
    flexsnap-listener-84c66dd4b8-6l4zj            1/1     Running        0          10h
    flexsnap-listener-84c66dd4b8-ls4nb            0/1     NodeAffinity   0          17h
    flexsnap-listener-84c66dd4b8-x84q8            0/1     NodeAffinity   0          3d15h
    flexsnap-listener-84c66dd4b8-z7d5m            0/1     NodeAffinity   0          5d18h
    flexsnap-operator-6b7dd6c56c-cf4pc            1/1     Running        0          10h
    flexsnap-operator-6b7dd6c56c-qjsbs            0/1     NodeAffinity   0          5d18h
    flexsnap-operator-6b7dd6c56c-xcsgj            0/1     NodeAffinity   0          3d15h
    flexsnap-operator-6b7dd6c56c-z86tc            0/1     NodeAffinity   0          39h

    However, these failures do not affect the functionality of NetBackup Snapshot Manager Kubernetes extension.

    Workaround:

    Manually clean-up the failed pods using the following command:

    kubectl get pods -n <cp_extension_namespace> | grep NodeAffinity | awk '{print $1}' | xargs kubectl delete pod -n <cp_extension_namespace>

  • Plugin information is duplicated, if NetBackup Snapshot Manager registration has failed in previous attempts

    This occurs only when NetBackup Snapshot Manager has been deployed using the MarketPlace Deployment Mechanism. This issue is observed when the plugin information is added before the registration. This issue creates duplicate plugin information in the CloudPoint_plugin.conf file.

    Workaround:

    Manually delete the duplicated plugin information from the CloudPoint_plugin.conf file.

    For example, consider the following example where the duplicate entry for GCP plugin config is visible (in bold) in CloudPoint_plugin.conf file:

    {
        "CPServer1": [
          {
            "Plugin_ID": "test",
            "Plugin_Type": "aws",
            "Config_ID": "aws.8dda1bf5-5ead-4d05-912a-71bdc13f55c4",
            "Plugin_Category": "Cloud",
            "Disabled": false
          }
        ]
      },
      {
        "CPServer2": [
          {
            "Plugin_ID": "gcp.2080179d-c149-498a-bf1f-4c9d9a76d4dd",
            "Plugin_Type": "gcp",
            "Config_ID": "gcp.2080179d-c149-498a-bf1f-4c9d9a76d4dd",
            "Plugin_Category": "Cloud",
            "Disabled": false
          },
          {
            "Plugin_ID": "gcp.2080179d-c149-498a-bf1f-4c9d9a76d4dd",
            "Plugin_Type": "gcp",
            "Config_ID": "gcp.2080179d-c149-498a-bf1f-4c9d9a76d4dd",
            "Plugin_Category": "Cloud",
            "Disabled": false
          }
        ]
      }
  • Plugin information is duplicated, if cloned NetBackup Snapshot Manager is added into NetBackup

    This occurs only when cloned NetBackup Snapshot Manager is added into NetBackup during migration of NetBackup Snapshot Manager to RHEL 8.6 VM. Cloning of NetBackup Snapshot Manager uses existing NetBackup Snapshot Manager volume to create new NetBackup Snapshot Manager. This creates duplicate entry into CloudPoint_plugin.conf file.

    Workaround:

    Manually edit and delete the duplicated plugin information from the CloudPoint_plugin.conf file.

    For example, consider the following example where the duplicate entry for Azure plugin config is visible (in bold) in CloudPoint_plugin.conf file:

    {
        "CPServer1": [
          {
            "Plugin_ID": "config10",
            "Plugin_Type": "azure",
            "Config_ID": "azure.327ec7fc-7a2d-4e94-90a4-02769a2ba521",
            "Plugin_Category": "Cloud",
            "Disabled": false
          }
        ]
      },
      {
        "CPServer2": [
          {
            "Plugin_ID": "azure.327ec7fc-7a2d-4e94-90a4-02769a2ba521",
            "Plugin_Type": "azure",
            "Config_ID": "azure.327ec7fc-7a2d-4e94-90a4-02769a2ba521",
            "Plugin_Category": "Cloud",
            "Disabled": false
          },
           {
        "cpserver101.yogesh.joshi2-dns-zone": [
          {
            "Plugin_ID": "azure.327ec7fc-7a2d-4e94-90a4-02769a2ba521",
            "Plugin_Type": "azure",
            "Config_ID": "azure.327ec7fc-7a2d-4e94-90a4-02769a2ba521",
            "Plugin_Category": "Cloud",
            "Disabled": false
          },
          {
            "Plugin_ID": "AZURE_PLUGIN",
            "Plugin_Type": "azure",
            "Config_ID": "azure.4400a00a-8d2b-4985-854a-74f48cd4567e",
            "Plugin_Category": "Cloud",
            "Disabled": false
          }
        ]
      }
     ]
    }
  • Backup from Snapshot operation using Snapshot Manager version 10.0 deployed in Azure fails due to SSL cert error

    Backup from Snapshot operation using Snapshot Manager version 10.2 deployed in Azure fails due to SSL cert error related to CRL (curl).

    Workaround:

    Add ECA_CRL_CHECK = 0 in Snapshot Manager bp.conf file and ensure that Azure endpoints are accessible from media server.