Monitoring RabbitMQ in Kubernetes

This guide describes how to monitor RabbitMQ instances deployed by the Kubernetes Cluster Operator.


Cluster Operator deploys RabbitMQ clusters with the rabbitmq_prometheus plugin enabled. The plugin exposes a Prometheus-compatible metrics endpoint.

For a detailed guide on RabbitMQ Prometheus configuration, check the Prometheus guide.

The following sections assume Prometheus is deployed and functional. How to configure Prometheus to monitor RabbitMQ depends on whether Prometheus is installed by Prometheus Operator or by other means.

Monitor RabbitMQ with Prometheus Operator

The Prometheus Operator defines the custom resource definitions (CRDs) ServiceMonitor, PodMonitor, and PrometheusRule. ServiceMonitor and PodMonitor CRDs allow to declaratively define how a dynamic set of services and pods should be monitored.

Check if the Kubernetes cluster has Prometheus Operator deployed:

kubectl get

If this command returns an error, Prometheus Operator is not deployed.

To monitor all RabbitMQ clusters, run:

kubectl apply --filename

To monitor RabbitMQ Cluster Operator, run:

kubectl apply --filename

ServiceMonitor and PodMonitor can be created in any namespace, as long as the Prometheus Operator has permissions to find it. For more information about these permissions, see Configure Permissions for the Prometheus Operator below.

Prometheus Operator will detect ServiceMonitor and PodMonitor objects and automatically configure and reload Prometheus' scrape config.

To validate whether Prometheus successfully scrapes metrics, open the Prometheus web UI in your browser and navigate to the Status -> Targets page where you should see an entry for the Cluster Operator (e.g. podMonitor/<podMonitorNamespace>/rabbitmq-cluster-operator/0 (1/1 up)) and one entry for each deployed RabbitMQ cluster (e.g. serviceMonitor/<serviceMonitorNamespace>/rabbitmq/0 (1/1 up)).

Prometheus Alerts

The custom resource PrometheusRule allows to declaratively define alerting rules. To install RabbitMQ alerting rules, first ensure that kube-state-metrics is installed. To deploy Prometheus rules for RabbitMQ, kubectl apply all YAML files in directory rules/rabbitmq. To deploy Prometheus rules for Cluster Operator, kubectl apply the YAML files in directory rules/rabbitmq-cluster-operator.

The ruleSelector from the Prometheus custom resource must match the labels of the deployed PrometheusRules. For example, if the Prometheus custom resource contains below ruleSelector, a label release: my-prometheus needs to be added to the PrometheusRules.

kind: Prometheus
  ruleNamespaceSelector: {}
      release: my-prometheus
  version: v2.26.0

To get notified on firing alerts (e.g. via Email or PagerDuty) configure a notification receiver in Alertmanager. To receive Slack notifications, deploy the Kubernetes Secret in directory alertmanager.

(Optional) Configure Permissions for the Prometheus Operator

If no RabbitMQ clusters appear in Prometheus, it might be necessary to adjust permissions for the Prometheus Operator.

The following steps have been tested with a kube-prometheus deployment.

To configure permissions for the Prometheus Operator, first create a file named prometheus-roles.yaml with the following contents:

kind: ClusterRole
  name: prometheus
- apiGroups: [""]
  - nodes
  - services
  - endpoints
  - pods
  verbs: ["get", "list", "watch"]
- apiGroups: [""]
  - configmaps
  verbs: ["get"]
- nonResourceURLs: ["/metrics"]
  verbs: ["get"]
kind: ClusterRoleBinding
  name: prometheus
  kind: ClusterRole
  name: prometheus
- kind: ServiceAccount
  name: prometheus-k8s
  namespace: monitoring

Then apply the permissions listed in prometheus-roles.yaml by running

kubectl apply -f prometheus-roles.yaml

Monitor RabbitMQ Without Prometheus Operator

If Prometheus is not installed by Prometheus Operator, but by other means, the CRDs ServiceMonitor, PodMonitor, and PrometheusRule are not available. Therefore, Prometheus needs to be configured via a config file. To monitor all RabbitMQ clusters and RabbitMQ Cluster Operator, use the scrape targets defined in Prometheus config file for RabbitMQ.

To set up RabbitMQ alerting rules, first configure Prometheus to receive metrics from the kube-state-metrics agent. Thereafter, configure Prometheus to use the Prometheus rule file. To receive Slack notifications, use the same alertmanager.yaml as provided in alertmanager/slack.yml for the Alertmanager configuration file.

Import Dashboards to Grafana

RabbitMQ provides Grafana dashboards to visualize the metrics scraped by Prometheus.

Follow the instructions in the Prometheus guide to import dashboards to Grafana.

Alternatively, if Grafana is deployed by the Grafana Helm chart, kubectl apply the ConfigMaps in directory grafana/dashboards to import RabbitMQ Grafana dashboards using a sidecar container.

The RabbitMQ-Alerts dashboard provides a history of all past RabbitMQ alerts across all RabbitMQ clusters in Kubernetes.

check-circle-line exclamation-circle-line close-line
Scroll to top icon