One-Arm and multi-arm deployments route load balancer traffic differently.
In one-arm deployment, the load balancer is not physically in line of the traffic, which means that the load balancer’s ingress and egress traffic goes through the same network interface. Traffic from the client through the load balancer is network address translated (NAT) with the load balancer as its source address. The nodes send their return traffic to the load balancer before being passed back to the client. Without this reverse packet flow, return traffic would try to reach the client directly, causing connections to fail.
In a multi-arm configuration, the traffic is routed through the load balancer. The end devices typically have the load balancer as their default gateway.
The most common deployment is a one-arm configuration. The same principles apply to multi-arm deployments, and they both work with F5 and NetScaler. For this document, the vRealize Automation, vRealize Orchestrator or Workspace ONE Access components are deployed as a one-arm configuration. Multi-arm deployments are also supported and their configuration should be similar to the one-arm configuration.