This section describes how to configure NSX-T Data Center networking for Kubernetes master and worker nodes.
Each node must have at least two network interfaces. The first is a management interface which might or might not be on the NSX-T Data Center fabric. The other interfaces provide networking for the pods, are on the NSX-T Data Center fabric, and connected to a logical switch which is referred to as the node logical switch. The management and pod IP addresses must be routable for Kubernetes health check to work. For communication between the management interface and the pods, NCP automatically creates a DFW rule to allow health check and other management traffic. You can see details of this rule in the NSX Manager GUI. This rule should not be changed or deleted.
For each node VM, ensure that the vNIC that is designated for container networking is attached to the node logical switch.
Tag | Scope |
---|---|
Node name | ncp/node_name |
Cluster name | ncp/cluster |
You can identify the logical switch port for a node VM by navigating to NSX Manager GUI.
from theIf the Kubernetes node name changes, you must update the tag ncp/node_name and restart NCP. You can use the following command to get the node names:
kubectl get nodes
If you add a node to a cluster while NCP is running, you must add the tags to the logical switch port before you run the kubeadm join command. Otherwise, the new node will not have network connectivity. If the tags are incorrect or missing, you can take the following steps to resolve the issue:
- Apply the correct tags to the logical switch port.
- Restart NCP.