VMware Cloud Director Container Service Extension 4.0.1 | 19 JAN 2023 | Build: 21109756 Check for additions and updates to these release notes. |
VMware Cloud Director Container Service Extension 4.0.1 | 19 JAN 2023 | Build: 21109756 Check for additions and updates to these release notes. |
VMware Cloud Director Container Service Extension version 4.0.1 release provides bug fixes, updates to the VMware Cloud Director Container Service Extension server and Kubernetes Container Clusters plug-in, and updates to product interoperability.
To access the full set of product documentation, go to VMware Cloud Director Container Service Extension.
New - VMware Cloud Director Container Service Extension Server 4.0.1
Service providers can now upgrade the VMware Cloud Director Container Service Extension Server from 4.0.0 to 4.0.1 through the CSE Management tab in Kubernetes Container Clusters UI plug-in of VMware Cloud Director.
For instructions on how to upgrade the VMware Cloud Director Container Service Extension Server from 4.0.0 to 4.0.1, see Update the VMware Cloud Director Container Service Extension Server.
You can download VMware Cloud Director Container Service Extension Server 4.0.1 from the VMware Cloud Director Container Service Extension Downloads page.
New - Kubernetes Container Clusters UI Plug-in 4.0.1 for VMware Cloud Director
A new version of Kubernetes Container Clusters UI plug-in is now available to use with VMware Cloud Director.
You can upgrade the Kubernetes Container Clusters UI plug-in before or after you upgrade the VMware Cloud Director Container Service Extension server.
The following steps outline how to upgrade the Kubernetes Container Clusters UI plug-in from 4.0.0 to 4.0.1:
Download the Kubernetes Container Clusters UI plug-in 4.0.1 from the VMware Cloud Director Container Service Extension Downloads page.
In the VMware Cloud Director Portal, from the top navigation bar, select More > Customize Portal.
Select the check box next to Kubernetes Container Clusters UI plug-in 4.0, and click Disable.
Click Upload > Select plugin file, and upload the Kubernetes Container Clusters UI plug-in 4.0.1 file.
Refresh the browser to start using the new plug-in.
For more information, refer to Managing Plug-Ins.
Important - In the Kubernetes Container Clusters 4.0.1 UI plug-in, it is necessary to use Cloud Storage Interface version 1.3.2.
From April 3, 2023, the k8s.gcr.io
registry is frozen and results in Cloud Storage Interface failures in the Kubernetes clusters. Such failures can occur when you attempt to scale a cluster or create a new cluster. Therefore, a new Cloud Storage Interface patch is released. Service providers must update the Cloud Storage Interface image version to avoid such failures when creating new clusters. Service providers can perform this task in the Update Configuration workflow in Server Details tab of Kubernetes Container Clusters 4.0.1 UI plug-in. For more information, see Update the VMware Cloud Director Container Service Extension Server. For instructions to upgrade Cloud Storage Interface to 1.3.2 on pre-existing clusters, see Cloud Storage Interface. For more information on k8s.gcr.io
registry freezing, see https://kubernetes.io/blog/2023/02/06/k8s-gcr-io-freeze-announcement/.
New - VMware Cloud Director Container Service Extension 4.0.1 interoperability updates
To view the interoperability of VMware Cloud Director Container Service Extension 4.0.1 and previous versions with VMware Cloud Director, and additional product interoperability, refer to the Product Interoperability Matrix.
The following table displays the interoperability between VMware Cloud Director Container Service Extension 4.0.1 and Kubernetes resources.
Kubernetes Resources |
Supported Versions |
Documentation |
---|---|---|
Kubernetes External Cloud Provider for VMware Cloud Director |
1.3.0, 1.2.0 |
|
Container Storage Interface driver for VMware Cloud Director Named Independent Disks |
1.3.2. From April 3, 2023, this is the only supported version of Cloud Storage Interface due to the freezing of the |
|
Kubernetes Cluster API Provider Cloud Director |
1.0.0 |
https://github.com/vmware/cluster-api-provider-cloud-director |
Service providers can update Kubernetes resources through the following workflow:
In VMware Cloud Director UI, from the top navigation bar, select More > Kubernetes Container Clusters.
In Kubernetes Container Clusters UI plug-in 4.0.0/4.0.1, select CSE Management > Server Details > Update Server.
In the Update CSE Server window, in the Current CSE Server Components section, update the Kubernetes resources configuration.
Click Submit Changes.
For more information, see Update the VMware Cloud Director Container Service Extension Server.
New - When there are two clusters of the same name in the same organization, and you attempt to delete one of the clusters, it results in the vApp of the other cluster with the same name being deleted.
Resolution: This bug is fixed in VMware Cloud Director Container Service Extension 4.0.1.
New - VMware Cloud Director Container Service Extension 4.0 cluster deployment fails with manual IP for Control Plane IP.
If a user inputs a value for Control Plane IP or Virtual IP Subnet, and then deletes the value from the input field, the cluster creation fails due to the UI sending an empty string instead of a null value.
Resolution: This bug is fixed in VMware Cloud Director Container Service Extension 4.0.1.
New - Update default NO_PROXY to use cluster.local
Current default settings for NO_PROXY contain k8s.test. The default should be cluster.local instead to match what Tanzu Kubernetes Grid clusters use.
Resolution: In Kubernetes Container Clusters 4.0.1 UI plug-in, the default value list now includes cluster.local, instead of k8s.test.
New - In Kubernetes Container Clusters UI, the cluster information page shows Fetching Upgrades while cluster is in pending state.
If you submit a Tanzu Kubernetes Grid cluster creation using Kubernetes Container Clusters UI plug-in, and the VMware Cloud Director Container Service Extension server has not yet started creating the cluster, the cluster's status is Pending. However, when you visit the cluster information page in the Kubernetes Container Clusters UI plug-in, you can see a spinner, and Fetching Upgrades for the upgrade availability value. This is incorrect as the upgrade availability value should be a hyphen.
Resolution: In the Kubernetes Container Clusters UI plug-in 4.0.1, this is fixed so that when the cluster's status is pending, the upgrade availability value is a hyphen.
New - In the Tanzu Kubernetes Grid cluster creation window of the Kubernetes Container Cluster UI plug-in, the Input validation help message does not appear when the control plane number of nodes is 0.
In the Tanzu Kubernetes Grid cluster creation wizard, if a user inputs 0 for control plane or worker number of nodes, the input validation help message does not appear to indicate that 0 is an invalid value.
Resolution: The validation help message now appears in Kubernetes Container Cluster UI plug-in 4.0.1.
New - In Kubernetes Container Clusters UI plug-in, the CSE Management workflow in a multi-site VMware Cloud Director setup only allows for a single server config entity. This results in CSE Management workflows failing in multi-site environments.
Resolution: The CSE Management workflow in Kubernetes Container Clusters UI plug-in now only fetches the server config entity that belongs to the site of the user that is currently logged in. Each site in a multi-site environment can now create, and maintain its own server config entity.
New - VMware Cloud Director Container Service Extension 4.0 fails to create a Tanzu Kubernetes Grid cluster when a proxy server is configured for downloading the binaries and images from external repositories.
Error message logged: Could not reach archive.ubuntu.com
.
Resolution: Proxy activation for Ubuntu package updates are present when a proxy is configured.
New - Misleading and false error log statements in VMware Cloud Director Container Service Extension 4.0.
Resolution: These redundant log statements are now rectified to present more relevant log messages.
In Kubernetes Container Clusters UI plug-in, the cluster delete operation can fail when the cluster status is Error.
To delete a cluster that is in Error status, it is necessary to force delete the cluster.
Log in to VMware Cloud Director, and from the top navigation bar, select More > Kubernetes Container Clusters.
Select a cluster, and in the cluster information page, click Delete.
In the Delete Cluster page, select the Force Delete checkbox, and click Delete.
In VMware Cloud Director Container Service Extension, the creation of Tanzu Kubernetes Grid clusters can fail due to a script execution error.
The following error appears in the Events tab of the cluster info page in Kubernetes Container Clusters UI:
ScriptExecutionTimeout
with the following details:
error while bootstrapping the machine [cluster-name/EPHEMERAL_TEMP_VM]; timeout for post customization phase [phase name of script execution]
Workaround:
When this error occurs, it is recommended to activate Auto Repair on Errors from cluster settings. This instructs VMware Cloud Director Container Service Extension to reattempt cluster creation.
Log in to VMware Cloud Director, and from the top navigation bar, select More > Kubernetes Container Clusters.
Select a cluster, and in the cluster information page, click Settings, and activate the Auto Repair on Errors toggle.
Click Save.
It is recommended to deactivate the Auto Repair on Errors toggle when troubleshooting cluster creation issues.
The cluster creation for multi-control plane or multi-worker node goes into an error state. The Events tab in the cluster details page shows an EphemeralVMError
event due to the failure to delete ephemeralVM in VMware Cloud Director.
The same error events can appear repeatedly if the Auto Repair on Errors setting is activated on the cluster. If the Auto Repair on Errors setting is off, sometimes the cluster can show an error state due to the failure to delete the ephemeralVM in VMware Cloud Director even though the control plane and worker nodes are created successfully.
This issue is visible in any release and patch release after but not including VMware Cloud Director 10.3.3.3, and any release and patch release starting with VMware Cloud Director 10.4.1.
Workaround:
Create the cluster with one control plane and one worker node, and then resize the cluster to the desired node count.
This issue is fixed for VMware Cloud Director Container Service Extension 4.0.3 release.
In some instances, nodes cannot join clusters. This occurs randomly due to intermittent issues, even when the cluster is in an available state.
The following error appears in the Events tab of the cluster info page in Kubernetes Container Clusters UI:
VcdMachineScriptExecutionError
with the following details:
script failed with status [x] and reason [Date Time 1 /root/node.sh: exit
[x]]
Workaround:
Perform this workaround for clusters that have nodes that fail to join. This workaround does not resolve the problem for future cluster creations.
It is necessary to perform this manual workaround when the issue occurs. There may be one or more nodes that fail to join, and it is necessary to perform the below steps starting from step 2 for every node that has not joined.
Download the kube config of the cluster that does not have all the nodes.
Log in to VMware Cloud Director, and from the top navigation bar, select More > Kubernetes Container Clusters.
Select a cluster, and in the cluster information page, click Download Kube Config.
For more information on Kube Config file, refer to the Kubernetes website.
In the Kubernetes Container Clusters UI plug-in, in the Events tab of the cluster information page, click on VcdMachineScriptExecutionError
to view the error details, and note the Resource Name.
In kubectl, enter the following command to fetch all the machines on the cluster:
kubectl get machines -A --kubeconfig=<path of downloaded kubeconfig>
The node that could not join the cluster should be stuck in a Provisioning state. To identify this node, look for the machine name with the resource name that was present in the VcdMachineScriptExecutionError
.
In kubectl, enter the following command:
Run `kubectl --kubeconfig=<path of downloaded kubeconfig> delete machine -n clusterNamespace machineName
Ensure the machine name that is being deleted matches the resource name that was present in the VcdMachineScriptExecutionError
.
Once the VM is deleted, it is recreated and the node reattempts to join the cluster.
ERROR: failed to create cluster: failed to pull image
failure
This error occurs in the following circumstances:
When a user attempts to create a Tanzu Kubernetes Grid Cluster using VMware Cloud Director Container Service Extension 4.0, and it fails intermittently.
An image pull error due to a HTTP 408 response is reported.
This issue can occur if there is difficulty reaching the Internet from the EPHEMERAL_TEMP_VM to pull the required images.
Potential causes:
Slow or intermittent Internet connectivity.
The network IP Pool cannot resolve DNS (docker pull error).
The network MTU behind a firewall must set lower.
To resolve the issue, ensure that there are no networking connectivity issues stopping the EPHEMERAL_TEMP_VM from reaching the Internet.
For more information, refer to https://kb.vmware.com/s/article/90326.
Users may encounter authorization errors when executing cluster operations in Kubernetes Container Clusters UI plug-in if a Legacy Rights Bundle exists for their organization.
After you upgrade VMware Cloud Director from version 9.1 or earlier, the system may create a Legacy Rights Bundle for each organization. This Legacy Rights Bundle includes the rights that are available in the associated organization at the time of the upgrade and is published only to this organization. To begin using the rights bundles model for an existing organization, you must delete the corresponding Legacy Rights Bundle. For more information, see Managing Rights and Roles.
In the Administration tab in the service provider portal, you can delete Legacy Rights Bundles. For more information, see Delete a Rights Bundle. Kubernetes Container Clusters UI plug-in CSE Management has a server setup process that automatically creates and publishes Kubernetes Clusters Rights Bundle to all tenants. The rights bundle contains all rights that are involved in Kubernetes cluster management in VMware Cloud Director Container Service Extension 4.0.
Updated - Resizing or upgrading Tanzu Kubernetes Grid cluster using kubectl.
After a cluster has been created in the Kubernetes Container Clusters UI plug-in, you can use kubectl to manage workloads on Tanzu Kubernetes Grid clusters.
If you also want to lifecycle manage, resize and upgrade the cluster through kubectl instead of the Kubernetes Container Clusters UI plug-in, complete the following steps:
Delete the RDE-Projector operator from the cluster kubectl delete deployment -n rdeprojector-system rdeprojector-controller-manager
Detach the Tanzu Kubernetes Grid cluster from Kubernetes Container Clusters UI plug-in.
In the VMware Cloud Director UI, in the Cluster Overview page, retrieve the cluster ID of the cluster.
Update the RDE with entity.spec.vcdKe.isVCDKECluster
to false.
Get the payload of the cluster - GET https://<vcd>/cloudapi/1.0.0/entities/<Cluster ID>
Copy and update the json path in the payload. - entity.spec.vcdKe.isVCDKECluster
to false.
PUT https://<vcd>/cloudapi/1.0.0/entities/<Cluster ID>
with the modified payload. It is necessary to include the entire payload as the body of PUT operation.
At this point the cluster is detached from VMware Cloud Director Container Service Extension 4.0.0 and 4.0.1, and it is not possible to manage the cluster through VMware Cloud Director Container Service Extension 4.0.0 and 4.0.1. It is now possible to use kubectl to manage, resize or upgrade the cluster by applying CAPI yaml, the cluster API specification, directly.
Cluster creation fails in VMware Cloud Director Container Service Extension due to invalid GitHub Token with Error: 401 Bad Credentials
This is the expected error during cluster creation. If customers set invalid Github access token, the cluster creation fails and the following error appears:
error creating the GitHub repository client: failed to get GitHub latest version: failed to get repository
versions: failed to get repository versions: failed to get the list of releases: GET
https://api.github.com/repos/kubernetes-sigs/cluster-api/releases: 401 Bad credentials
When you configure the VMware Cloud Director Container Service Extension server, enter an accurate Github access token.
Policies selection in VMware Cloud Director Container Service Extension 4 plug-in does not populate the full list after selection for the purpose of policy modification.
When a user selects a sizing policy in the Kubernetes Container Clusters plug-in and they want to change it, the dropdown menu only displays the selected sizing policy, and does not automatically load alternative sizing policies.
The user has to delete the text manually to allow the alternative sizing policies to appear. This also occurs in the dropdown menu when the user selects of placement policies and storage policies.
This is intentional. This is how the combobox html, Clarity, web component works.
Note:Clarity is the web framework that VMware Cloud Director UI is built on.
The dropdown box uses the input text as a filter. When nothing is in the input field, you can see all selections, and the selections filter as you type.
When you create a VMware Cloud Director Container Service Extension cluster, a character capitalization error appears.
In the Kubernetes Container Clusters UI, if you use capital letters, the following error appears:
Name must start with a letter, end with an alphanumeric, and only contain alphanumeric or hyphen (-) characters. (Max 63 characters)
This is a restriction set by Kubernetes. Object names are validated under RFC 1035 labels. For more information, refer to Kubernetes website.
Kubernetes Container Clusters UI-Plugin 4.0 does not interoperate with other Kubernetes Container Clusters UI plug-ins, such as 3.5.0.
The ability to operate these two plug-ins simultaneously without conflict is a known VMware Cloud Director UI limitation. You can only have one plug-in activated at any given time.
When a node of the cluster is deleted due to failure in vSphere or other underlying infrastructure, VMware Cloud Director Container Service Extension does not inform the user, and it does not auto-heal the cluster.
When the node of a cluster is deleted, basic cluster operations, such as cluster resize and cluster upgrade, continue to work. The deleted node remains in deleted state, and is included in computations regarding size of the cluster.
Download the Kubeconfig of the cluster.
Use the following command to delete the machine that continues to use the deleted node configuration:
kubectl --kubeconfig=<path to downloaded kubeconfig> get machines -A # try to match the machine name
here; also get namespace
kubectl -n <namespace name from above> --kubeconfig=<path to downloaded kubeconfig> delete machine
<machine name>
# wait for machine to get deleted
The above command deletes the machine, and CAPVCD automatically creates a new machine.
VMware Cloud Director Container Service Extension fails to deploy clusters with TKG templates that have an unmodifiable placement policy set on them.
Log in to the VMware Cloud Director Tenant Portal as an administrator.
Click Libraries > vApp Templates.
In the vApp Templates window, select the radio button to the left of the template.
In the top ribbon, click Tag with Compute Policies.
Select the Modifiable checkboxes, and click Tag.
In VMware Cloud Director 10.4, service providers are unable to log-in to the VMware Cloud Director Container Service Extension virtual machine by default.
In VMware Cloud Director 10.4, after deploying the VMware Cloud Director Container Service Extension virtual machine from OVA file, the following two checkboxes in the VM settings page are not selected by default:
Allow local administrator password
Auto-generate password
It is necessary to select these checkboxes to allow providers to log-in to the VMware Cloud Director Container Service Extension virtual machine in future to perform troubleshooting tasks.
Log in to VMware Cloud Director UI as a service provider, and create a vApp from the VMware Cloud Director Container Service Extension OVA file. For more information, see Create a vApp from VMware Cloud Director Container Service Extension server OVA file.
Once you deploy the vApp, and before you power it on, go to VM details > Guest OS Customization > Select Allow local administrator password and Auto-generate password.
After the vApp update task finishes, power on the vApp.
Fast provisioning must be deactivated in Organization VDC in order to resize disks.
Log in to VMware Cloud Director UI as a provider, and select Resources.
In the Cloud Resources tab, select Organization VDCs, and select an organization VDC.
In the organization VDC window, under Policies, select Storage.
Click Edit, and deactivate the Fast provisioning toggle.
Click Save.
When you log in as a service provider, after you upload the latest UI plug-in, the CSE Management tab does not display.
Workaround:
Deactivate the previous UI plug-in that is built into VMware Cloud Director.
Log in to VMware Cloud Director UI as a provider, and select More > Customize Portal.
Select the check box next to the names of the target plug-ins, and click Enable or Disable.
To start using the newly activated plug-in, refresh the Internet browser page.
If there are multiple activated plugins with the same name or id but different version, the lowest version plug-in is used. Therefore, only activate the highest version plug-in. Deactivate all other version plug-ins.
For more information on managing plug-ins, see Managing Plug-Ins.