VMware Cloud Foundation 3.10.2 | 14 APR 2021 | Build 17854560

VMware Cloud Foundation | 25 MAY 2021 | Build 18015401

Read about what’s new, learn about what was fixed, and find workarounds for known issues in VMware Cloud Foundation 3.10.2

VMware Cloud Foundation 3.10.2 can either be upgraded from VMware Cloud Foundation (sequential upgrade) or from VMware Cloud Foundation 3.5 (skip-level upgrade). It cannot be deployed as a new release. For more information, see Upgrade Information below.


What's in the Release Notes

The release notes cover the following topics:

What's New

The VMware Cloud Foundation 3.10.2 release includes the following:

  • Skip-Level Upgrade Tool Enhancements: The skip-level upgrade tool enables you to upgrade to VMware Cloud Foundation 3.10.2 from VMware Cloud Foundation 3.5 and later releases. Enhancements to the skip-level upgrade tool include:
    • Improvements in prechecks, help and error messages, and guardrails.
    • Improvement to the offline upgrade bundle download and upload method.
    • Advanced options to download specific upgrade bundles and verify bundle availability.
    • Details of available bundles for an upgrade and list of missing bundles before starting an upgrade.
    • Integrity checks on upgrade bundle file size after downloading from the depot and before uploading to the SDDC Manager.
    • Options to mitigate file permission issues.
    • Single tool to upgrade to VMware Cloud Foundation 3.10.2 and VMware Cloud Foundation
  • Enterprise PKS Deployment: Starting from VMware Cloud Foundation 3.10.2, Enterprise PKS cannot be deployed from SDDC Manager.
  • BOM Updates for the 3.10.2 ReleaseUpdated Bill of Materials with new product versions.

Cloud Foundation Bill of Materials (BOM)

The Cloud Foundation software product is comprised of the following software Bill-of-Materials (BOM). The components in the BOM are interoperable and compatible.

Software Component Version Date Build Number
SDDC Manager 3.10.2 14 APR 2021


VMware vCenter Server Appliance 6.7 Update 3m 18 MAR 2021


VMware ESXi ESXi670-202103001 18 MAR 2021


VMware NSX Data Center for vSphere 6.4.10 18 FEB 2021


VMware NSX-T Data Center 2.5.3 11 FEB 2021


VMware vRealize Suite Lifecycle Manager 2.1 Patch 2 04 MAY 2020


VMware vRealize Log Insight 4.8 11 APR 2019 13036238
vRealize Log Insight Content Pack for NSX for vSphere 3.9 n/a n/a
vRealize Log Insight Content Pack for Linux 2.0.1 n/a n/a
vRealize Log Insight Content Pack for vRealize Automation 7.5+ 1.0 n/a n/a
vRealize Log Insight Content Pack for vRealize Orchestrator 7.0.1+ 2.1 n/a n/a
vRealize Log insight Content Pack for NSX-T 3.8.2 n/a n/a
vSAN Content Pack for Log Insight 2.2 n/a n/a
vRealize Operations Manager 7.5 11 APR 2019 13165949
vRealize Automation 7.6 11 APR 2019 13027280
VMware Horizon 7 7.10.0 17 SEP 2019



  • vRealize Log Insight Content Packs are deployed during the workload domain creation.
  • VMware Solution Exchange and the vRealize Log Insight in-product marketplace store only the latest versions of the content packs for vRealize Log Insight. The Bill of Materials table contains the latest versions of the packs that were available at the time VMware Cloud Foundation is released. When you deploy the Cloud Foundation components, it is possible that the version of a content pack within the in-product marketplace for vRealize Log Insight is newer than the one used for this release.
  • To remediate VMSA-2020-0007 (CVE-2020-3953 and CVE-2020-3954) for vRealize Log Insight 4.8, you must apply the vRealize Log Insight 4.8 security patch. For information on the security patch, see KB article 79168. ​
  • For this release, you can install the vRealize Suite 2019 products instead of those listed in the BOM. These include vRealize Suite Lifecycle Manager 8.1, vRealize Log Insight 8.1.1, vRealize Operations 8.1, and vRealize Automation 8.1 along with Workspace ONE Access.. For prescriptive guidance on deploying and configuring these products with VMware Cloud Foundation 3.10, see Deployment of VMware vRealize Suite 2019 on VMware Cloud Foundation 3.10.
  • SDDC Manager 3.10.2 contains security fixes for Photon OS packages for PHSA-2020-3.0-0163 to PHSA-2021-3.0-0197 published here: https://github.com/vmware/photon/wiki/Security-Advisories-3.

VMware Software Edition License Information

The SDDC Manager software is licensed under the Cloud Foundation license. As part of this product, the SDDC Manager software deploys specific VMware software products.

The following VMware software components deployed by SDDC Manager are licensed under the Cloud Foundation license:

  • VMware ESXi
  • VMware vSAN
  • VMware NSX Data Center for vSphere

The following VMware software components deployed by SDDC Manager are licensed separately:

  • VMware vCenter Server
    NOTE Only one vCenter Server license is required for all vCenter Servers deployed in a Cloud Foundation system.
  • VMware NSX-T
  • VMware Enterprise PKS
  • VMware Horizon 7
  • VMware vRealize Automation
  • VMware vRealize Operations
  • VMware vRealize Log Insight and content packs
    NOTE Cloud Foundation permits limited use of vRealize Log Insight for the management domain without the purchase of a vRealize Log Insight license.

For details about the specific VMware software editions that are licensed under the licenses you have purchased, see the Cloud Foundation Bill of Materials (BOM) section above.

For general information about the product, see VMware Cloud Foundation.

Supported Hardware

For details on vSAN Ready Nodes in Cloud Foundation, see VMware Compatibility Guide (VCG) for vSAN and the Hardware Requirements section in the VMware Cloud Foundation Planning and Preparation Guide.


To access the Cloud Foundation 3.10.2 documentation, go to the VMware Cloud Foundation product documentation

To access the documentation for VMware software products that SDDC Manager can deploy, see the product documentation and use the drop-down menu on the page to choose the appropriate version:

Browser Compatibility and Screen Resolutions

The Cloud Foundation web-based interface supports the latest two versions of the following web browsers except the Internet Explorer:

  • Google Chrome
  • Mozilla Firefox
  • Microsoft Edge
  • Internet Explorer: Version 11

For the Web-based user interfaces, the supported standard resolution is 1024 by 768 pixels. For best results, use a screen resolution within these tested resolutions:

  • 1024 by 768 pixels (standard)
  • 1366 by 768 pixels
  • 1280 by 1024 pixels
  • 1680 by 1050 pixels

Resolutions below 1024 by 768, such as 640 by 960 or 480 by 800, are not supported.

Upgrade Information

VMware Cloud Foundation 3.10.2 can either be upgraded from VMware Cloud Foundation (sequential upgrade) or from VMware Cloud Foundation 3.5 (skip-level upgrade). It cannot be deployed as a new release. For upgrade information, refer to the VMware Cloud Foundation Upgrade Guide.

VMware Cloud Foundation Release Information

VMware Cloud Foundation includes bug and security fixes. You can upgrade to VMware Cloud Foundation from VMware Cloud Foundation 3.10.2.

To upgrade the management domain, apply the following bundles, in order:

  1. VMware Cloud Foundation update bundle.
  2. vCenter Server bundle.

To upgrade VI workload domains, apply the following bundle:

  • vCenter Server bundle.

VMware Cloud Foundation contains the following BOM updates:

Software Component Version Date Build Number
SDDC Manager 25 MAY 2021 18015401
VMware vCenter Server Appliance 6.7 Update 3n 25 MAY 2021 18010531

For more information about VMware vCenter Server Appliance 6.7 Update 3n, see the release notes.

Resolved Issues

  • SDDC Manager cannot manage the passwords for the NSX Edges and UDLR/DLR deployed to support application virtual networking

    These passwords are not managed through the SDDC Manager Dashboard.

Known Issues

The known issues are grouped as follows.

Upgrade Known Issues
  • The vRealize Automation upgrade reports the "Precheck Execution Failure : Make sure the latest version of VMware Tools is installed" message

    The vRealize Automation IaaS VMs must have the same version of VMware Tools as the ESXi hosts on which the VMs reside.

    Workaround: Upgrade VMware Tools on the vRealize Automation IaaS VMs.

  • Error upgrading vRealize Automation

    Under certain circumstances, upgrading vRealize Automation may fail with a message similar to:

    An automated upgrade has failed. Manual intervention is required.
    vRealize Suite Lifecycle Manager Pre-upgrade checks for vRealize Automation have failed:
    vRealize Automation Validations : iaasms1.rainpole.local : RebootPending : Check if reboot is pending : Reboot the machine.
    vRealize Automation Validations : iaasms2.rainpole.local : RebootPending : Check if reboot is pending : Reboot the machine.
    Please retry the upgrade once the upgrade is available again. 


    1. Log-in into the first VM listed in the error message using RDP or the VMware Remote Console.
    2. Reboot the VM.
    3. Wait 5 minutes after the login screen of the VM appears.
    4. Repeat steps 1-3 for the next VM listed in the error message.
    5. Once you have restarted all the VMs listed in the error message, retry the vRealize Automation upgrade.

  • When there is no associated workload domain to vRealize Automation, the VRA VM NODES CONSISTENCY CHECK upgrade precheck fails

    This upgrade precheck compares the content in the logical inventory on the SDDC Manager and the content in the vRealize Lifecycle Manager environment. When there is no associated workload domain, the vRealize Lifecycle Manager environment does not contain information about the iaasagent1.rainpole.local and iaasagent2.rainpole.local nodes. Therefore the check fails.

    Workaround: None. You can safely ignore a failed VRA VM NODES CONSISTENCY CHECK during the upgrade precheck. The upgrade will succeed even with this error.

  • NSX Data Center for vSphere upgrade fails with the message "Host Prep remediation failed"

    After addressing the issue, the NSX Data Center for vSphere bundle no longer appears as an available update.

    Workaround: To complete the upgrade, manually enable the anti-affinity rules.

    1. Log in to the management vCenter Server using the vSphere Client.
    2. Click Menu > Hosts and Clusters and select the cluster on which host prep remediation failed (for example SDDC-Cluster1).
    3. Click Configure > Configuration > VM/Host Rules.
    4. Select NSX Controller Anti-Affinity Rule and click Edit.
    5. Select Enable rule and click OK.

    This completes the NSX Data Center for vSphere upgrade.

  • Lifecycle Management displays fatal error

    When the user password in the /opt/vmware/vcf/lcm/lcm-app/conf/application.properties file contains a backslash (\), Lifecycle Manager does not start and displays the fatal error Password authentication failed for user lcm.

    Workaround: Follow the steps below to resolve the error:

    1. SSH to the SDDC Manager VM.
    2. Type su to switch to root user.
    3. Open the /opt/vmware/vcf/lcm/lcm-app/conf/application.properties file, remove all backslashes (\) from the lcm.datasource.password field, and save the file.
    4. Run the command systemctl restart lcm-db.
  • Task panel does not show correct upgrade tasks for NSX-T workload domain upgrades

    When you upgrade NSX-T workload domains. the task panel does not show upgrade status correctly. This is a UI issue only and there is no impact on the upgrade workflow.

    Workaround: Monitor upgrade status by navigating to the Update/Patches tab of the relevant workload domain:

    1. On the SDDC Manager Dashboard, click Inventory -> Workload Domains.
    2. In the Domain column, click the appropriate workload domain name.
    3. Click the Update/Patches tab.
    4. Monitor upgrade status.
  • Exception displayed when a scheduled NSX-T upgrade begins during an idle SDDC Manager session

    When a scheduled NSX-T upgrade begins during an idle SDDC Manager session, the following UI exception is displayed:
    Retrieving NSXT upgrade failed with unknown exception

    This is a UI issue only. There is no impact on the upgrade workflow.

    Workaround: Refresh the web browser.

  • Inapplicable ESXi upgrade bundles are displayed after upgrade has been scheduled

    After you schedule an ESXi upgrade on a workload domain, upgrade bundles are displayed until ESXi has been upgraded on all clusters. You can ignore these bundles.

    Workaround: None.

vRealize Integration Known Issues
  • The password update for vRealize Automation and vRealize Operations Manager may run infinitely or may fail when the password contains special character "%"

    Password management uses the vRealize Lifecycle Manager API to update the password of vRealize Automation and vRealize Operations Manager. When there is special character "%" in either of SSH or API or Administrator credential types of the vRealize Automation and vRealize Operations Manager users, then the vRealize Lifecycle Manager API hangs and doesn't respond to password management. There is a timeout of 5 mins and password management marks the operation as failed.

    Workaround:Retry the password update operation without the special character "%". Ensure that the passwords for all other vRealize Automation and vRealize Operations Manager accounts don't contain the "%" special character.

  • vRealize Log Insight installation gets stuck due to incorrect MTU configuration

    During deployment, Edge Service Gateways send frames with the MTU specified in the Universal Distributed Logical Router - MTU Size field in the deployment parameters file to the Top of Rack switches. If this MTU size is not configured correctly in your infrastructure, the vRealize Log Insight deployment may hang on an installation task after the Apply vRealize Log Insight License task. If any of these tasks remain incomplete for more than 30 minutes. follow the workaround below.

    1. Fix the routing between the Edge-vTEP network and ESXi-vTEP network
    2. SSH to  the Cloud Builder VM.
    3. Switch to root:
      sudo -i
    4. Restart the bring-up service:
      systemctl restart vcf-bringup
    5. Wait five minutes to get all services online.
    6. Retry the bring-up process.
Networking Known Issues
  • NSX Manager is not visible in the vSphere Web Client.

    In addition to NSX Manager not being visible in the vSphere Web Client, the following error message displays in the NSX Home screen: "No NSX Managers available. Verify current user has role assigned on NSX Manager." This issue occurs when vCenter Server is not correctly configured for the account that is logged in.

    Workaround: To resolve this issue, follow the procedure detailed in Knowledge Base article 2080740 "No NSX Managers available" error in the vSphere Web Client.

SDDC Manager Known Issues
  • Unable to delete VI workload domain enabled for vRealize Operations Manager from SDDC Manager.

    Attempts to delete the vCenter adapter also fail, and return an SSL error.

    Workaround: Use the following procedure to resolve this issue.

    1. Create a vCenter adapter instance in vRealize Operations Manager, as described in Configure a vCenter Adapter Instance in vRealize Operations Manager.
      This step is required because the existing adapter was deleted by the failed workload domain deletion.
    2. Follow the procedure described in Knowledge Base article 56946.
    3. Restart the failed VI workload domain deletion workflow from the SDDC Manager interface.
  • APIs for managing SDDC cannot be executed from the SDDC Manager Dashboard

    You cannot use the API Explorer in the SDDC Manager Dashboard to execute the APIs for managing SDDC (/v1/sddc). 

    Workaround: None. These APIs can only be executed using the Cloud Builder as the host.

  • Add host workflow fails

    Add host workflow fails with the following error message:

    Unable to update transport node: {\n    messages = [],\n    data = struct => {error_message=General error has occurred., details=Discovered node with id:3e9c025a-d5ae-4e7e-ab5d-a30b5269689d:host-51 is already prepared having fabric node id:Node/dd6562e9-34ce-45ae-aff7-7624c01df788., error_code=100, module_name=common-services}\n}"

    Workaround: On the NSX-T UI, uninstall NSX for the failed transport node.

Workload Domain Known Issues
  • Adding host fails when host is on a different VLAN

    A host add operation can sometimes fail if the host is on a different VLAN.


    1. Before adding the host, add a new portgroup to the VDS for that cluster.
    2. Tag the new portgroup with the VLAN ID of the host to be added.
    3. Add the Host. This workflow fails at the "Migrate host vmknics to dvs" operation.
    4. Locate the failed host in vCenter, and migrate the vmk0 of the host to the new portgroup you created in step 1.
      For more information, see Migrate VMkernel Adapters to a vSphere Distributed Switch in the vSphere product documentation.
    5. Retry the Add Host operation.

    NOTE: If you later remove this host in the future, you must manually remove the portgroup as well if it is not being used by any other host.

  • NSX Manager for VI workload domain is not displayed in vCenter

    Although NFS-based VI workload domains are created successfully, the NSX Manager VM is not registered in vCenter Server and is not displayed in vCenter.

    Workaround: To resolve this issue, use the following procedure:

    1. Log in to NSX Manager (http://<nsxmanager IP>).
    2. Navigate to Manage > NSX Management Service.
    3. Un-register the lookup service and vCenter, then re-register.
    4. Close the browser and log in to vCenter.
  • A vCenter Server on which certificates have been rotated is not accessible from a Horizon workload domain

    Cloud Foundation does not support the certificate rotation on the Horizon workload domains.

    Workaround: Refer to https://kb.vmware.com/s/article/70956.

  • Deploying partner services on an NSX-T workload domain displays an error

    Deploying partner services on an NSX-T workload domain such as McAfee or Trend displays the “Configure NSX at cluster level to deploy Service VM” error.

    Workaround: Attach the Transport node profile to the cluster and try deploying the partner service. After the service is deployed, detach the transport node profile from the cluster.

  • If the witness ESXi version does not match with the host ESXi version in the cluster, vSAN cluster partition may occur

    vSAN stretch cluster workflow does not check the ESXi version of the witness host. If the witness ESXi version does not match the host version in the cluster, then vSAN cluster partition may happen.


    1. Upgrade the witness host manually with the matching ESXi version using the vCenter VUM functionality.
    2. Replace or deploy the witness appliance matching with the ESXi version.

  • vSAN partition and critical alerts are generated when the witness MTU is not set to 9000

    If the MTU of the witness switch in the witness appliance is not set to 9000, the vSAN stretch cluster partition may occur.

    Workaround: Set the MTU of the witness switch in the witness appliance to 9000 MTU.

  • The certificate rotate operation on the second NSX-T domain fails

    Certificate rotation works on the first NSX-T workload domain in your environment, but fails on all subsequent NSX-T workload domains.

    Workaround: None

  • Operations on NSX-T workload domains fails if their host FQDNs include uppercase letters

    If the FQDNs of ESXi hosts in an NSX-T workload domain include uppercase letters, then the following operations may fail for the workload domain:

    • Add a host
    • Remove a host
    • Add a cluster
    • Remove a cluster
    • Delete the workload domain

    Workaround: See KB 76553.

  • Creating an NSX-T workload domain fails on the task "Add management domain vCenter as compute manager"

    This can happen if a previous attempt to create an NSX-T workload domain failed and Cloud Foundation was unable to clean up after the failed task.

    Workaround: Manually remove the NSX-T Data Center extension from the management vCenter Server and try to create the NSX-T workload domain again. See https://docs.vmware.com/en/VMware-NSX-T-Data-Center/2.5/administration/GUID-E6E2F017-1106-48C5-ABCA-3D3E9130A863.html.

  • VI workload domain creation or expansion operations fail

    If there is a mismatch between the letter case (upper or lower) of an ESXi host's FQDN and the FQDN used when the host was commissioned, then workload domain creation and expansion may fail.

    Workaround: ESXi hosts should have lower case FQDNs and should be commissioned using lower case FQDNs.

  • Cluster is deleted even if VMs are up and running on the cluster

    When you delete a cluster, it gets deleted even if there are VMs running on the cluster. This includes critical VMs such as Edge VMs, which may prevent you from accessing your environment after the cluster gets deleted.

    Migrate the VMs to a different cluster before deleting the cluster.

  • Workload domain operations fail if cluster upgrade is in progress

    Workload domain operations cannot be performed if one or more clusters are being upgraded. The UI does not block such oeprations during an upgrade.

    Workaround: Do not perform any operations on the workload domain when a cluster upgrade is in progress.

Security Operations Known Issues
  • Addition of members from PKS UAA to Harbor library fails when the certificate verification is enabled

     This issue occurs when Harbor does not honor the certificate chain under System Settings > Registry Root Certificate.


    1. SSH into the SDDC Manager VM as the vcf user.

    2. Run the following command. Make sure to update the password of the admin user and the Harbor URL:
    curl -k -H'Content-type: application/json' -u admin:"< >" -XPUT https://harbor.vrack.vsphere.local/api/configurations -d '{"uaa_verify_cert":"false"}'

    Harbor is in the UAA authentication mode and it uses members from PKS UAA.

    To create a user in UAA:
    1. Connect through SSH to Ops Manager appliance
    2. Run the following:

    uaac target https://pks.vrack.vsphere.local:8443 --skip-ssl-validation

    uaac token client get admin

    uaac user add <<user-name> > --emails <<email> >

Multi-Instance Management Known Issues
  • Federation creation information not displayed if you leave the Multi-Instance Management Dashboard

    Federation creation progress is displayed on the Multi-Instance Management Dashboard. If you navigate to another screen and then return to the Multi-Instance Management Dashboard, progress messages are not displayed. Instead, an empty map with no Cloud Foundation instances are displayed until the federation is created.

    Workaround: Stay on the Multi-Instance Dashboard till the task is complete. If you have navigated away, wait for around 20 minutes and then return to the dashboard by which time the operation should have completed.

  • The federation creation progress is not displayed

    While federation creation is in progress, the SDDC manager UI displays the progress on the multi-site page. If you navigate into any other screen and come back to the multi-site screen, the progress messages are not displayed. An empty map with no VMware Cloud Foundation instances is displayed until the federation creation process completes.

    Workaround: None

  • Multi-Instance Management Dashboard operation fails

    After a controller joins or leaves a federation, Kafka is restarted on all controllers in the federation. It can take up to 15 minutes for the federation to stabilize. Any operations performed on the dashboard during this time may fail.

    Workaround: Re-try the operation.

API Known Issues
  • Unversionsed APIs are not Supported

    Unversioned APIs in Cloud Foundation have been deprecated.

    Use Cloud Foundation public APIs.

check-circle-line exclamation-circle-line close-line
Scroll to top icon