This topic describes how to enable High Availability (HA) for Tanzu Postgres. HA offers automatic failover ensuring that any application requests operate continuously and without downtime.
Tanzu Postgres uses the pg_auto_failover extension to provide a highly available Tanzu Postgres cluster on Kubernetes. For detailed information about pg_auto_failover features, see the pg_auto_failover documentation.
In the Tanzu Postgres HA cluster configuration, the topology consists of three pods: one monitor, one primary and one hot standby mirror. pg_auto_failover ensures that the data is synchronously replicated from the primary to the mirror node. If the primary node is unresponsive, the application requests are re-directed to the mirror node, which gets promoted to the primary. All application requests continue to the promoted primary, while a new postgres instance is started which becomes the new mirror. If the monitor pod fails, operations continue as normal. The Postgres operator redeploys a monitor pod, and when ready it resumes monitoring of the primary and secondary.
Ensure that you have completed the Installing a Postgres Operator procedures before proceeding. Also review Deploying a New Postgres Instance.
To enable Tanzu Postgres high availability (cluster mode), edit your copy of the instance yaml
file you created during Deploying a New Postgres Instance. In the yaml
, alter the highAvailability
field:
apiVersion: sql.tanzu.vmware.com/v1
kind: Postgres
metadata:
name: postgres-sample
spec:
memory: 800Mi
cpu: "0.8"
storageClassName: standard
storageSize: 10G
serviceType: CluterIP
pgConfig:
dbname:
username:
highAvailability:
enabled: true
highAvailability
values can be enabled: <true|false>
. If this field is left empty, the postgres instance is by default a single node configuration.
Execute this command to deploy or redeploy the cluster with the new highAvailability
setting:
kubectl apply -f postgres.yaml
where postgres.yaml
is the Kubernetes manifest created for this instance.
The command output is similar to:
postgres.sql.tanzu.vmware.com/postgres-sample created
where postgres-sample is the Postgres instance name
defined in the yaml
file.
At this point, the Postgres operator deploys the three Postgres instance pods: the monitor, the primary, and the mirror.
To confirm your HA configuration is ready for access, use kubectl get
to review the STATUS field and confirm that it shows "Running". Initially STATUS will show Created, until all artifacts are deployed. Use Ctr-C to escape the watch command.
watch kubectl get postgres/postgres-sample
NAME STATUS BACKUP LOCATION AGE
postgres-sample Pending backuplocation-sample 17m
To view the created pods, use:
kubectl get pods
NAME READY STATUS RESTARTS AGE
pod/postgres-sample-0 1/1 Running 0 11m
pod/postgres-sample-monitor 1/1 Running 0 12m
pod/postgres-sample-1 1/1 Running 0 4m28s
You can now log into the primary pod using kubectl exec -it <pod-name> -- bash
:
kubectl exec -it postgres-sample-0 -- bash
You can log into any pod with kubectl exec
and use the pg_autoctl
tool to inspect the state of the cluster. Run pg_autoctl show state
to see which pod is currently the primary:
kubectl exec -ti pod/postgres-sample-1 -- pg_autoctl show state
Name | Node | Host:Port | TLI: LSN | Connection | Current State | Assigned State
-------+-------+------------------------------------------------------------------------+----------------+--------------+---------------------+--------------------
node_1 | 1 | postgres-sample-0.postgres-sample-agent.default.svc.cluster.local:5432 | 2: 0/3002690 | read-only | secondary | secondary
node_2 | 2 | postgres-sample-1.postgres-sample-agent.default.svc.cluster.local:5432 | 2: 0/3002690 | read-write | primary | primary
The pg_autoctl
set of commands manage the pg_autofailover services. For further information, refer to the pg_autoctl reference documentation.
Note: VMware supports a limited range of pg_autoctl
commands, involving inspecting the nodes and performing a manual failover.
If the primary is unreachable, during the primary to mirror failover, the Current State
and Assigned State
status columns toggle between demoted, catching_up, wait_primary, secondary, and primary. You can monitor the states using pg_autoctl
:
watch pg_autoctl show state
Name | Node | Host:Port | TLI: LSN | Connection | Current State | Assigned State
-------+-------+------------------------------------------------------------------------+----------------+--------------+---------------------+--------------------
node_1 | 1 | postgres-sample-0.postgres-sample-agent.default.svc.cluster.local:5432 | 2: 0/3002690 | read-only | demoted | catching_up
node_2 | 2 | postgres-sample-1.postgres-sample-agent.default.svc.cluster.local:5432 | 2: 0/3002690 | read-write | primary | primary
Name | Node | Host:Port | TLI: LSN | Connection | Current State | Assigned State
-------+-------+------------------------------------------------------------------------+----------------+--------------+---------------------+--------------------
node_1 | 1 | postgres-sample-0.postgres-sample-agent.default.svc.cluster.local:5432 | 2: 0/3002690 | read-only | secondary | secondary
node_2 | 2 | postgres-sample-1.postgres-sample-agent.default.svc.cluster.local:5432 | 2: 0/3002690 | read-write | primary | primary