Restoring a backup results in restoring the state of the network at the time of the backup. In addition, the configurations maintained by NSX Manager or Global Manager appliances are also restored. For NSX Manager, any changes, such as adding or deleting nodes, that were made to the fabric since the backup was taken, are reconciled.

You must restore the backup to a new NSX Manager or Global Manager appliance.

If you had a cluster of the NSX Manager appliance when the backup was taken, the restore process restores one node first and then prompts you to add the other nodes. You can add the other nodes during the restore process or after the first node is restored.

If you had a cluster of Global Manager appliances, you can only restore one node using the restore process. You must create the cluster after the restore of the first node completes.

Important:

If any nodes in the appliance cluster are still available, you must power them off before you start the restore.

Prerequisites

  • Verify that you have the login credentials for the backup file server.
  • Verify that you have the SSH fingerprint of the backup file server. Only an SHA256 hashed ECDSA key is accepted as a fingerprint.
  • Verify that you have the passphrase of the backup file.
  • Identify which backup you want to restore by following the procedure in Listing Available Backups. Take note of the IP or FQDN of the NSX-T Data Center appliance that took the backup.
  • If you had an NSX Intelligence appliance installed on the NSX Manager when the backup was taken, you must upload the OVA file of that same version of the NSX Intelligence appliance on the new NSX Manager before you restore the backup. See Installing and Upgrading VMware NSX Intelligence for the installation information.

Procedure

  1. If any nodes in the appliance cluster that you are restoring are online, power them off.
  2. Install one new appliance node on which to restore the backup.
    • If the backup listing for the backup you are restoring contains an IP address, you must deploy the new NSX Manager or Global Manager node with the same IP address. Do not configure the node to publish its FQDN.

    • If the backup listing for the backup you are restoring contains an FQDN, you must configure the new appliance node with this FQDN and publish the FQDN. Note that only lowercase FQDN is supported for backup and restore.

      Use this API to publish the NSX Manager or Global Manager FQDN.
      Note: Until the FQDN is configured and published, the Restore button for the backup is disabled in the newly deployed NSX Manager or Global Manager UI.

      Example request:

      PUT https://<nsx-mgr OR global-mgr>/api/v1/configs/management
      
      {
        "publish_fqdns": true,
        "_revision": 0
      }

      See the NSX-T Data Center API Guide for API details.

      In addition, if the new manager node has a different IP address than the original one, you must update the DNS server's forward and reverse lookup entries for the manager node with the new IP address.

    After the new manager node is running and online, you can proceed with the restore.

  3. From a browser, log in with admin privileges to the NSX Manager or Global Manager at https://<manager-ip-address>.
  4. Select System > Backup & Restore.
  5. To configure the backup file server, click Edit.
  6. Enter the IP address or FQDN.
  7. Change the port number, if necessary.
    The default is 22.
  8. To log in to the server, enter the user name and password.
  9. In the Destination Directory text box, enter the absolute directory path where the backups are stored.
  10. Enter the passphrase that was used to encrypt the backup data.
  11. You can leave the SSH Fingerprint blank and accept or reject the fingerprint provided by the server after you click Save in a later step. If necessary, you can retrieve the SSH fingerprint by using this API: POST /api/v1/cluster/backups?action=retrieve_ssh_fingerprint. Note that only SHA256 hashed ECDSA (256 bit, 384 bit, 521 bit) host key is accepted as a fingerprint.
  12. Click Save.
  13. Select a backup.
  14. Click Restore.
  15. The restore process prompts you to take action, if necessary, as it progresses.
    Note: If you are restoring a Global Manager appliance, these steps do not appear. After restoring the first Global Manager node, you must manually join the other nodes to form the cluster.
    1. Confirm CM/VC Connectivity: If you want to restore existing compute managers, ensure that they are registered with the new NSX Manager node and available during the restore process.
    2. If you have deleted or added fabric nodes or transport nodes, you are prompted to take certain actions, for example, log in to a node and run a script. If you have created a logical switch or segment since the backup, the logical switch or segment will not appear after the restore.
    3. If the backup has information about a manager cluster, you are prompted to add other nodes. If you decide not to add nodes, you can still proceed with the restore and manually add other nodes to form the cluster after the restore of this node completes.
    4. If there are fabric nodes that did not discover the new manager node, you are provided a list of them.
    A progress bar displays the status of the restore operation noting the step the restore process is on.
    Note: During the restore process, services on the manager appliance get restarted and the control plane becomes unavailable until restore completes.
    After the restore operation is finished, the Restore Complete screen shows the result of the restore, the timestamp of the backup file, and the start and end time of the restore operation.
    Note: Any segments created after the backup was taken are not restored.

    If the restore fails, the screen displays the step where the failure occurred, for example, Current Step: Restoring Cluster (DB) or Current Step: Restoring Node. If either cluster restore or node restore fails, the error might be transient. In that case, there is no need to click Retry. You can restart or reboot the manager and the restore continues.

    You can also determine if there was a cluster restore or node restore failure by selecting the log files. Run get log-file syslog to view the system log file and search for the strings Cluster restore failed and Node restore failed.

    To restart the manager, run the restart service manager command.

    To reboot the manager, run the reboot command.

    Note: If you added a compute manager after the backup, and you try to add the compute manager again after the restore, you will get an error message indicating that registration failed. You can click the Resolve button to resolve the error and successfully add the compute manager. For more information, see Add a Compute Manager, step 4. If you want to remove information about NSX-T Data Center that is stored in a vCenter Server, follow the steps in Remove NSX-T Data Center Extension from vCenter Server

    If the vCenter Server was registered with custom ports in the backup, you must manually open all the custom ports on the restored manager appliances.

  16. If you have only one node deployed, after the restored manager node is up and functional, you can deploy additional nodes to form a cluster.
    See the NSX-T Data Center Installation Guide for instructions.
  17. If you had other manager cluster VMs that you powered down in Step 1, delete them after the new manager cluster is deployed.