To avoid interruption of vRealize Hyperic server operation in the case of failure, you can configure a cluster of vRealize Hyperic servers.

The failover configuration uses:

  • EHCache's distributed caching for replicating changes throughout the cluster.

  • The nodeStatus.hqu plug-in for monitoring the availability of nodes.

  • A hardware load balancer for managing failover when an node becomes unavailable. The load balancer checks the status of each node every 10 seconds, by issuing an HTTP request to the node's nodeStatus.hqu plug-in. The check returns a response of master=true with a return code of 200 for the primary node. The check returns master=false with a return code of 404 inside the body of the response for other nodes in the cluster.

A vRealize Hyperic server cluster contains multiple nodes. Two nodes are generally sufficient. One vRealize Hyperic server, automatically selected by vRealize Hyperic, serves as the primary node. The other node or nodes serve as hot backups. The hot backups do not share the workload with the primary node.

A failover configuration exists that is transparent to users and vRealize Hyperic administrators. This configuration starts a new session that requires you to log in again. It is not apparent that the active vRealize Hyperic server instance is clustered, or which node is currently active.