Duplicate AI for workload clusters can be created by TKGm/TMC auto discover

Workaround 1: To auto manage and avoid any duplication of the newly created workload cluster by Supervisor cluster, you need to delete the manually created workload adapter instance. This will also clear the metric history on pods/container objects for the manually created adapter instance.

Workaround 2: Stop the collection of auto created K8s adapter instances. Stopping the adapter instance will not cause any memory or CPU usage in the collector but prevents duplication and monitoring (which happens after deleting the auto-created workload AI).

Some Prometheus exporter metrics are propagated to parent objects as the PromQL matches the same metric

Some Prometheus exporter metrics are propagated to parent objects as the PromQL matches the same metric. So, the the same set of metrics are grouped under parent Kubernetes objects. For example: The Container metrics are grouped under Pod/Node/Namespace.

Workaround: None

If you upgrade from VMware vRealize Operations Management Pack for Kubernetes 1.0 to 1.2.1, the collection state displays a status called Not Collecting for all the adapter instances

This occurs because of the addition of new settings and credential types in the 1.1 version of VMware vRealize Operations Management Pack for Kubernetes.

Note: If you upgrade from VMware vRealize Operations Management Pack for Kubernetes 1.1 to 1.2.1, you do not have to complete the steps listed below.

Workaround: All the adapter instances must be deleted and recreated. This will lead to creation of new objects. However, you can retain the old objects to keep historical data intact.

  1. From the main menu of vRealize Operations Manager, click Administration, and then in the left pane click Solutions.

  2. From the Solutions page, select VMware vRealize Operations Management Pack for Kubernetes.

  3. Click the Configure icon. The Manage Solution dialog box appears.

    • Select an adapter instance.

    • Click the Delete icon.

    • When the Confirmation dialog box appears and if you want to retain historical data, deselect the option Remove related objects.

  4. Recreate the adapter instance by following steps provided in User Guide.

  5. Repeat the above steps for all adapter instances.

During configuration, VMware vRealize Operations Management Pack for Kubernetes verifies if the cAdvisor service is accessible on every node. An error message similar to the following may appear: Unable to establish a valid connection to the target system. cAdvisor service on following nodes is either not reachable OR of a lower version than v2.1

The error occurs if the cAdvisor service is inaccessible or if the API version is lesser than 2.1. You may sometimes receive this error if the cAdvisor service temporarily throws a gateway error at the time of verification.

Workaround:

  1. Verify if the cAdvisor service is up and running on the affected nodes and responds to API calls.

  2. Verify if the API version of the cAdvisor service is later than 2.1. If not, deploy the latest version of the cAdvisor service.

If you have completed the above two steps, you can ignore the error message and continue to save the adapter instance.

Under recommendations, the Defined by column is displayed as KubernetesAdapter3

Under recommendations, the Defined by column is displayed as KubernetesAdapter3.

Deleting the VMware PKS adapter instance does not remove the Kubernetes adapter instances created by the VMware PKS adapter instance

When you delete the VMware PKS adapter instance, the Kubernetes adapter instances created by the VMware PKS adapter instance will not be removed.

Workaround : Manually delete the adapter instances related to the VMware PKS adapter instance

The Environment Overview dashboard does not display the relationship between the vCenter Hosts/Virtual Machines and the Kubernetes nodes

If the vRealize Operations Manager accesses the K8s through proxy, the vCenter adapter instance does not provide a provision to specify proxy. So, the Environment Overview dashboard may not display the relationship between the vCenter Hosts/Virtual Machines and the Kubernetes nodes.

Workaround: None

Data collection fails for the K8s adapters that are auto-configured by the VMware PKS adapter

The auto-configuredK8s adapter instances that presents the untrusted SSL certificates will have the collection status as 'Failed'.

Workaround: Manually accept the untrusted certificate for the auto-configured K8s adapter instances for which data collection has failed.

Adding the VMware PKS adapter will configure the K8s instances but does not create the vCenter adapter instances

Adding the VMware PKS adapter will configure the K8s instances but does not create the vCenter adapter instances or associate it with the vCenters that the Kubernetes cluster nodes are deployed in.

Workaround: Manually configure the vCenter adapter instances and then add the details to the K8s adapters that are auto-configured by VMware PKS.

Expired certificates are also auto accepted when the 'Auto-accept Kubernetes Cluster SSL Certificate' option is enabled

When you enable the 'Auto-accept Kubernetes Cluster SSL Certificate' under Advanced Settings, the expired certificates are also auto accepted and there is no prompt.

Workaround: None

Container File System base usage (MB) should be referred as file system base limit

The Container File System base usage (MB) should be referred as file system base limit. This is applicable only for Prometheus monitored containers.

Workaround: None

DiskIO|Sync and DiskIO|Async Container metrics will not be shown in vRealize Operations Manager 

The DiskIO|Sync and DiskIO|Async Container metrics will not be shown in vRealize Operations Manager for Prometheus monitored containers. 

Workaround: None

The Events metric under the namespace object is not shown in vRealize Operations Manager

This issue occurs due to the huge response of 100 MB that is received for events api call against the namespaces.

Workaround: None

Adapter instance is in the warning state if customer upgrades Kubernetes MP from 1.5.1/1.5.2 to 1.6

Due to refresh token expiry in TKGI setup, a popup with error message, "Error in accessing the base url of Kubernetes master - /api/v1", is displayed.

  1. Edit the adapter instance which is in the warning state.

  2. Perform Validate Connection.

  3. Accept the certificates:

    • TKGI (for Kubernetes clusters deployed on TKGI)

    • Accept the Kubernetes Cluster

  4. Click Ok when the popup error message is displayed. Click Save.

  5. Click Yes on the confirmation popup message window.

  6. Wait for 2 - 3 collection cycles until the adapter instance is in data receiving state.

Basic Auth does not work on Kubernetes MP for Kubernetes 1.19 and above version

Kubernetes deprecates the support of Basic authentication model from Kubernetes 1.19 onwards.

Use Certificate /Token auth to configure adapter instance for Kubernetes 1.19 and above versions.

Deprecation of cAdvisor

cAdvisor integration is deprecated from the latest version of TKG.

Deploy cAdvisor as Daemon or use Prometheus as a collector service with supported exporters.

Time-out error while performing test connection on selecting cAdvisor collector service

Kubernetes Cluster nodes are not routable for performance metrics to be collected from vROps using cAdvisor.

Use Prometheus as a collector service with supported exporters.

Performance metrics may not be collected for container runtimes except for docker

Few performance metrics may not be collected using cAdvisor collector service for container runtimes, except for docker.

Use Prometheus as a collector service with supported exporters.

Default alerts from Kubernetes Management Pack may not work

Default alerts from Kubernetes Management Pack may not work when Prometheus is used as a collector service.

Workaround: None

Localization cannot be found for Prometheus metrics

Due to the dynamic nature of Prometheus, localization cannot be performed for the Prometheus metrics.

Workaround: None