The RabbitMQ management plugin provides an HTTP-based API for management and monitoring of RabbitMQ nodes and clusters, along with a browser-based UI and a command line tool, rabbitmqadmin.
It periodically collects and aggregates data about many aspects of the system. Those metrics are exposed to human operators in the UI. The API it provides can be used by monitoring systems, however, Prometheus is the recommended option for long term storage, alerting, visualisation, chart analysis and so on.
The plugin also provides tools for analysing memory usage of the node, and other features related to monitoring, metrics, user, permission, and topology management. Previously it also provided definition export and import functionality. Those are now core RabbitMQ features and do not require or rely on this plugin.
This guide covers:
as well as other related topics.
The plugin also provides extension points that other plugins, such as rabbitmq-top or rabbitmq-shovel-management, use to extend the UI.
The management UI and its HTTP API is a built-in monitoring option for RabbitMQ. This is a convenient option for development and in environments where external monitoring is difficult or impossible to introduce.
However, the management UI has a number of limitations:
Long term metric storage and visualisation services such as Prometheus and Grafana or the ELK stack are more suitable options for production systems. They offer:
RabbitMQ provides first class support for Prometheus and Grafana as of 3.8. It is recommended for production environments.
The management plugin is included in the RabbitMQ distribution. Like any other plugin, it must be enabled before it can be used. That's done using rabbitmq-plugins:
rabbitmq-plugins enable rabbitmq_management
Node restart is not required after plugin activation.
During automated deployments, the plugin can be enabled via enabled plugin file.
The management UI can be accessed using a Web browser at http://:15672/
.
For example, for a node running on a machine with the hostname of warp10.local
, it can be accessed by users with sufficient privileges at either http://warp10.local:15672/
or http://localhost:15672/
(provided that localhost
resolves correctly).
Note that the UI and HTTP API port — typically 15672 — does not support AMQP 0-9-1, AMQP 1.0, STOMP or MQTT connections. Separate ports should be used by those clients.
Users must be granted permissions for management UI access.
The management UI is implemented as a single page application which relies on the HTTP API. Some of the features include:
The UI application supports recent versions of Google Chrome, Safari, Firefox, and Microsoft Edge browsers.
Any cluster node with rabbitmq_management
plugin enabled can be used for management UI access or data collection by monitoring tools. It will reach out to other nodes and collect their stats, then aggregate and return a response to the client.
To access management UI the user has to authenticate and have certain permissions (be authorised). This is covered in the following section.
The management UI requires authentication and authorisation, much like RabbitMQ requires it from connecting clients. In addition to successful authentication, management UI access is controlled by user tags. The tags are managed using rabbitmqctl. Newly created users do not have any tags set on them by default.
See Production Checklist for general recommendations on user and credential management.
Tag | Capabilities |
---|---|
(None) | No access to the management plugin |
management | Anything the user could do via messaging protocols plus:
|
policymaker | Everything "management" can plus:
|
monitoring | Everything "management" can plus:
|
administrator | Everything "policymaker" and "monitoring" can plus:
|
Note that since "administrator" does everything "monitoring" does, and "monitoring" does everything "management" does, each user often needs a maximum of one tag.
Normal RabbitMQ permissions to resources still apply to monitors and administrators; just because a user is a monitor or administrator does not grant them full access to exchanges, queues and bindings through the management plugin or other means.
All users can only list objects within the virtual hosts they have any permissions for.
If access to management UI is impossible to due the lack of users with sufficient permissions or forgotten/incorrect permissions, CLI tools must be used to manage the users and their credentials. rabbitmqctl add_user should be used to create a user, rabbitmqctl set_permissions to grant the user the desired permissions and finally, rabbitmqctl set_user_tags should be used to give the user management UI access permissions.
The following example creates a user with complete access to the management UI/HTTP API (as in, all virtual hosts and management features):
# create a user rabbitmqctl add_user full_access s3crEt # tag the user with "administrator" for full management UI and HTTP API access rabbitmqctl set_user_tags full_access administrator
RabbitMQ can be configured to use JWT-encoded OAuth 2.0 access tokens to authenticate client applications and management UI users. When doing so, the management UI does not automatically redirect users to authenticate against the OAuth 2 server, this must be configured separately. Currently, Authorization code flow with PKCE is tested with the following Authorization servers:
Important: from the OAuth 2.0 point of view, the RabbitMQ Management UI is a public app which means it cannot securely store credentials such as the client_secret. This means that RabbitMQ does not need to present a client_secret when authenticating users.
It is usually possible to configure the OAuth client as a public app with the authorization server that you are using. If target authorization server only supports a confidential app or it requires a client_secret, then a client_secret must be configured using the oauth_client_secret
setting.
To redirect users to the UAA server to authenticate, use the following configuration:
management.oauth_enabled = true management.oauth_client_id = rabbit_user_client management.oauth_client_secret = rabbit_user_client management.oauth_provider_url = https://my-uaa-server-host:8443/uaa management.oauth_scopes = openid profile rabbitmq.*
Important: UAA supports regular expression in scopes, e.g.
rabbitmq.*
. The above configuration assumes that theresource_server_id
configured in the oauth2 backend matches the valuerabbitmq
.
Important: Since RabbitMQ 3.10, RabbitMQ uses
authorization_code
grant type.implicit
flow is deprecated.
Important:
management.oauth_client_secret
is an optional setting. UAA 75.21.0 and earlier versions requireoauth_client_secret
regardless if the oauth client is configured as confidental.
When using management.oauth_enabled = true
, it is still possible to authenticate with HTTP basic authentication against the HTTP API. This means both of the following examples will work:
# swap <token> for an actual token curl -i -u ignored:<token> http://localhost:15672/api/vhosts
as well as
curl -i --header "authorization: Basic <encoded credentials>" http://localhost:15672/api/vhosts
To switch to authenticate using OAuth 2 exclusively for management UI access, set the management.disable_basic_auth
configuration key to true
:
management.disable_basic_auth = true management.oauth_client_id = rabbit_user_client management.oauth_client_secret = rabbit_user_client management.oauth_provider_url = https://my-uaa-server-host:8443/uaa management.oauth_scopes = openid profile rabbitmq.*
When setting management.disable_basic_auth
to true
, only the Bearer
(token-based) authorization method will work, for example:
# swap <token> for an actual token curl -i --header "authorization: Bearer <token>" http://localhost:15672/api/vhosts
This is true for all endpoints except GET /definitions
and POST /definitions
. Those endpoints require the token to be passed in the token
query string parameter.
It is possible to configure which OAuth 2.0 scopes RabbitMQ should claim when redirecting the user to the authorization server.
If management.oauth_enabled = true
and management.oauth_scopes
is not set, RabbitMQ default to openid profile
.
Depending on the Authorization server, we may use regular expression in scopes, e.g. <resource_server_id>.*
, or instead we have to explicitly ask for them, e.g.:
.tag:administrator
.read:*/*/*
By default, RabbitMQ assumes the OpenID Connect Discovery endpoint is at <management.oauth_provider_url>/.well-known/openid-configuration
. If your endpoint differs, you can set yours via the management.oauth_metadata_url
setting.
RabbitMQ uses this endpoint to discover other endpoints such as token endpoint, logout endpoint, and others.
RabbitMQ follows the OpenID Connect RP-Initiated Logout 1.0 specification to implement the logout workflow. This means that the logout workflow is triggered from the Management UI when the user clicks on the Logout button. Logging out from RabbitMQ management UI not only logs the user out from the management UI itself but also from the Identity Provider.
There are other two additional scenarios which can trigger a logout. One scenario occurs when the OAuth Token expires. Although RabbitMQ renews the token in the background before it expires, if the token expires, the user is logged out. The second scenario is when the management UI session exceeds the maximum allowed time configured on the Login Session Timeout.
connect-src
To support the OAuth 2.0 protocol, RabbitMQ makes asynchronous REST calls to the OpenId Connect Discovery endpoint. If you override the default CSP headers, you have to make sure that the connect-src
CSP directive whitelists the OpenId Connect Discovery endpoint.
For instance, if you configured the CSP header with the value default-src 'self'
you are, by default, setting connect-src 'self'
which means you are denying RabbitMQ access to any external endpoint; hence disabling OAuth 2.0.
In addition to the connect-src
CSP header, RabbitMQ also needs the CSP directives unsafe-eval
unsafe-inline
, otherwise the OAuth 2.0 functionality may not work.
By default, the RabbitMQ Management UI uses the OAuth 2.0 authorization code flow to authenticate and authorize users. However, there are scenarios where users prefer to be automatically redirected to RabbitMQ without getting involved in additional logon flows. By using OAuth2 proxies and web portals, these additional logon flows can be avoided. With a single click, users navigate straight to a RabbitMQ Management UI with a token obtained under the covers. This is known as Identity-Provider initiated logon.
RabbitMQ exposes a new setting called management.oauth_initiated_logon_type
whose default value sp_initiated
. To enable an Identity-Provider initiated logon you set it to idp_initiated
.
management.oauth_enabled = true management.oauth_initiated_logon_type = idp_initiated management.oauth_provider_url = https://my-web-portal
With the previous settings, the management UI exposes the HTTP endpoint /login
which accepts content-type: application/x-www-form-urlencoded
and it expects the JWT token in the access_token
form field. This is the endpoint where the Web portal will redirect users to the management UI. Additionally, RabbitMQ also accepts a JWT token in the HTTP Authorization
header when the user lands on the management UI.
When activated, the management plugin provides an HTTP API at http://server-name:15672/api/
by default. Browse to that location for more information on the API. For convenience the same API reference is available on GitHub.
The API is intended to be used for basic observability tasks. Prometheus and Grafana are recommended for long term metric storage, alerting, anomaly detection, and so on.
Any cluster node with rabbitmq-management
plugin enabled can be used for management UI access or HTTP API access. It will reach out to other nodes and collect their stats, then aggregate and return a response to the client.
When using the API in a cluster of nodes, there is no need to contact each node via HTTP API individually. Instead, contact a random node or a load balancer that sits in front of the cluster.
rabbitmqadmin is a Python command line tool that interacts with the HTTP API. It can be downloaded from any RabbitMQ node that has the management plugin enabled at http://:15672/cli/
.
For HTTP API clients in several languages, see Developer Tools.
Some API endpoints return a lot of information. The volume can be reduced by filtering what columns are returned by HTTP GET
requests. See latest HTTP API documentation for details.
It may be necessary to put a reverse proxy in front of a RabbitMQ cluster. Reverse proxy setup for RabbitMQ may require careful handling of encoded slashes in paths if default virtual host (/
) is used.
If default virtual host is not used, the additional settings to support encoded URIs will not be necessary. In other words, both Nginx and Apache configuration will require the standard minimum for any HTTP-based service.
If RabbitMQ HTTP API access is configured for the root location (/
), the location must not have a slash at the end:
# trailing slash in the location must be omitted only if default RabbitMQ virtual host is used location / { proxy_pass http://rabbitmq-host:15672; }
If a different location will be used to proxy requests to the HTTP API, a URI rewrite rule must be used:
# these rewrites are only if default RabbitMQ virtual host is used location ~* /rabbitmq/api/(.*?)/(.*) { proxy_pass http://rabbitmq-host:15672/api/$1/%2F/$2?$query_string; } location ~* /rabbitmq/(.*) { rewrite ^/rabbitmq/(.*)$ /$1 break; proxy_pass http://rabbitmq-host:15672; }
To support encoded slashes in URIs, Apache requires users to explicitly enable AllowEncodedSlashes
.
# required only if default RabbitMQ virtual host is used AllowEncodedSlashes On
for the Apache virtual host. Apache needs both mod_proxy and mod_proxy_http enabled. The location also needs a nocanon
setting:
ProxyPassReverse / http://rabbitmq-host:15672/ # "nocanon" is required only if default RabbitMQ virtual host is used ProxyPass / http://rabbitmq-host:15672/ nocanon
There are several configuration options which affect the management plugin. These are managed through the main RabbitMQ configuration file.
It is possible to configure HTTP API and management UI to use a different port or network interface, enable HTTPS and so on.
While rarely needed, it is possible to configure multiple listeners (ports), e.g. to both enable HTTPS and retain support for clients that can only use HTTP (without TLS).
The port is configured using the management.tcp.port
key:
management.tcp.port = 15672
It is possible to configure what interface the API endpoint will use, similarly to messaging protocol listeners, using the management.tcp.ip
key:
management.tcp.ip = 0.0.0.0
To check what interface and port is used by a running node, use rabbitmq-diagnostics
:
rabbitmq-diagnostics -s listeners # => Interface: [::], port: 15672, protocol: http, purpose: HTTP API # => Interface: [::], port: 15671, protocol: https, purpose: HTTP API over TLS (HTTPS)
or tools such as lsof
, ss
or netstat
.
The management plugin can be configured to use HTTPS. See the guide on TLS to learn more about certificate authorities, certificates and private key files.
management.ssl.port = 15671 management.ssl.cacertfile = /path/to/ca_certificate.pem management.ssl.certfile = /path/to/server_certificate.pem management.ssl.keyfile = /path/to/server_key.pem ## This key must only be used if private key is password protected # management.ssl.password = bunnies
More TLS options can be configured for the HTTPS listener.
management.ssl.port = 15671 management.ssl.cacertfile = /path/to/ca_certificate.pem management.ssl.certfile = /path/to/server_certificate.pem management.ssl.keyfile = /path/to/server_key.pem ## This key must only be used if private key is password protected # management.ssl.password = bunnies # For RabbitMQ 3.7.10 and later versions management.ssl.honor_cipher_order = true management.ssl.honor_ecc_order = true management.ssl.client_renegotiation = false management.ssl.secure_renegotiate = true management.ssl.versions.1 = tlsv1.2 management.ssl.versions.2 = tlsv1.1 management.ssl.ciphers.1 = ECDHE-ECDSA-AES256-GCM-SHA384 management.ssl.ciphers.2 = ECDHE-RSA-AES256-GCM-SHA384 management.ssl.ciphers.3 = ECDHE-ECDSA-AES256-SHA384 management.ssl.ciphers.4 = ECDHE-RSA-AES256-SHA384 management.ssl.ciphers.5 = ECDH-ECDSA-AES256-GCM-SHA384 management.ssl.ciphers.6 = ECDH-RSA-AES256-GCM-SHA384 management.ssl.ciphers.7 = ECDH-ECDSA-AES256-SHA384 management.ssl.ciphers.8 = ECDH-RSA-AES256-SHA384 management.ssl.ciphers.9 = DHE-RSA-AES256-GCM-SHA384 ## Usually RabbitMQ nodes do not perform peer verification of HTTP API clients ## but it can be enabled if needed. Clients then will have to be configured with ## a certificate and private key pair. ## ## See https://www.rabbitmq.com/ssl.html#peer-verification for details. # management.ssl.verify = verify_peer # management.ssl.fail_if_no_peer_cert = true
The above example in the classic config format:
[ {rabbitmq_management, [ {listener, [{port, 15671}, {ssl, true}, {ssl_opts, [{cacertfile, "/path/to/ca_certificate.pem"}, {certfile, "/path/to/server_certificate.pem"}, {keyfile, "/path/to/server_key.pem"}, %% don't do peer verification to HTTPS clients {verify, verify_none}, {fail_if_no_peer_cert, false}, {client_renegotiation, false}, {secure_renegotiate, true}, {honor_ecc_order, true}, {honor_cipher_order, true}, {versions,['tlsv1.1', 'tlsv1.2']}, {ciphers, ["ECDHE-ECDSA-AES256-GCM-SHA384", "ECDHE-RSA-AES256-GCM-SHA384", "ECDHE-ECDSA-AES256-SHA384", "ECDHE-RSA-AES256-SHA384", "ECDH-ECDSA-AES256-GCM-SHA384", "ECDH-RSA-AES256-GCM-SHA384", "ECDH-ECDSA-AES256-SHA384", "ECDH-RSA-AES256-SHA384", "DHE-RSA-AES256-GCM-SHA384" ]} ]} ]} ]} ].
It is possible to use both HTTP and HTTPS on different ports:
management.tcp.port = 15672 management.ssl.port = 15671 management.ssl.cacertfile = /path/to/ca_certificate.pem management.ssl.certfile = /path/to/server_certificate.pem management.ssl.keyfile = /path/to/server_key.pem
The same configuration keys can be used to configure a single listener (just HTTP or HTTPS) and match those used by the Web STOMP and Web MQTT.
Cowboy, the embedded Web server used by the management plugin, provides a number of options that can be used to customize the behavior of the server. Most of the options were introduced in RabbitMQ 3.7.9.
Response compression is enabled by default. To enable it explicitly, use management.tcp.compress
:
# For RabbitMQ 3.7.9 and later versions management.tcp.compress = true
Some HTTP API endpoints respond quickly, others may need to return or stream a sizeable data set to the client (e.g. many thousands of connections) or perform an operation that takes time proportionally to the input (e.g. import a large definitions file). In those cases the amount of time it takes to process the request can exceed certain timeouts in the Web server as well as HTTP client.
It is possible to bump Cowboy timeouts using the management.tcp.idle_timeout
, management.tcp.inactivity_timeout
, management.tcp.request_timeout
options.
management.tcp.inactivity_timeout
controls HTTP(S) client's TCP connection inactivity timeout. When it is reached, the connection will be closed by the HTTP server.management.tcp.request_timeout
controls the window of time in which the client has to send an HTTP request.management.tcp.idle_timeout
controls the window of time in which the client has to send more data (if any) within the context of an HTTP request.If a load balancer or proxy is used between HTTP clients and the management HTTP server, the inactivity_timeout
and idle_timeout
values should be at least as large, and often greater than, the timeout and inactivity values used by the load balancer.
Here are some example configuration snippets that modify the timeouts:
# For RabbitMQ 3.7.9 and later versions. # # Configures HTTP (non-encrypted) listener timeouts management.tcp.idle_timeout = 120000 management.tcp.inactivity_timeout = 120000 management.tcp.request_timeout = 10000
# For RabbitMQ 3.7.9 and later versions. # # Configures HTTPS (TLS-enabled) listener timeouts management.ssl.idle_timeout = 120000 management.ssl.inactivity_timeout = 120000 management.ssl.request_timeout = 10000
All values are in milliseconds. Their defaults vary:
management.tcp.inactivity_timeout
has the default of 300 secondsmanagement.tcp.request_timeout
has the default of 60 secondsmanagement.tcp.idle_timeout
has the default of 5 secondsIt is recommended that if the inactivity or idle timeout need changing, management.tcp.inactivity_timeout
value should match or be greater than that of management.tcp.idle_timeout
.
management.tcp.request_timeout
typically does not need increasing as clients send a request shortly after establishing a TCP connection.
To create simple access logs of requests to the HTTP API, set the value of the management.http_log_dir
key to the path of a directory in which logs can be created:
management.http_log_dir = /path/to/folder
For the change to have an effect, restart the plugin or the node.
By default the server will emit statistics events every 5 seconds (5000
ms). The message rate values shown in the management plugin are calculated over this period.
Increasing this value will reduce CPU resource consumption of stats collection in environments with a large number of stats emitting entities such as connections, channels, queues.
In order to do so, set the value of the collect_statistics_interval
configuration key to the desired interval in milliseconds and restart the node:
# 15s collect_statistics_interval = 15000
The management plugin by default shows message rates globally, and for each queue, channel, exchange, and vhost. These are known as the basic message rates.
It can also show message rates for all the combinations of channel to exchange, exchange to queue, and queue to channel. These are known as detailed message rates. Detailed message rates are disabled by default as they can have a large memory footprint when there are a large number of combinations of channels, queues and exchanges.
Alternatively, the message rates can be disabled altogether. This can help get reduce CPU resource consumption of the plugin.
The message rate mode is controlled by the management.rates_mode
configuration key:
# supported values: basic, detailed, none management.rates_mode = basic
Supported values are basic
(the default), detailed
, and none
.
The management plugin will retain samples of some data such as message rates and queue lengths. Depending on how long the data is retained, some time range options on UI charts may be incomplete or unavailable.
There are three policies:
global
: how long to retain data for the overview and virtual host pagesbasic
: how long to retain data for individual connections, channels, exchanges and queuesdetailed
: how long to retain data for message rates between pairs of connections, channels, exchanges and queues (as shown under "Message rates breakdown")Below is a configuration example:
management.sample_retention_policies.global.minute = 5 management.sample_retention_policies.global.hour = 60 management.sample_retention_policies.global.day = 1200 management.sample_retention_policies.basic.minute = 5 management.sample_retention_policies.basic.hour = 60 management.sample_retention_policies.detailed.10 = 5
The configuration in the example above retains global data at a 5 second resolution (sampling happens every 5 seconds) for a minute, then at a 1 minute (60 second) resolution for 1 hour, then at a 20 minute resolution for one day. It retains basic data at a 5 second resolution for 1 minute, at a 1 minute (60 second) resolution for 1 hour, and detailed data only for 10 seconds.
All three policies are mandatory, and must contain at least one retention setting (period).
It is possible to disable the statistics in the UI and HTTP API in order for these to be used only for operations. This can be a useful feature if external monitoring solutions such as Prometheus and Grafana are being used. If statistics are disabled in any of the following ways, all charts and detailed statistics will be hidden in the UI.
In order to completely disable the internal metrics collection, the disable_metrics_collector
flag must be set in the rabbitmq_management_agent
plugin. The Prometheus plugin will still work even if collection is disabled.
management_agent.disable_metrics_collector = true
Disabling the metrics collection is the preferred option if it is being used with an external monitoring system, as this reduced the overhead that statistics collection and aggregation causes in the broker. If the statistics are only temporary disabled, or are not required in some HTTP API queries, the aggregation of the stats can be disabled in the rabbitmq_management
plugin. The disable flag can be also passed as part of the query string in the URI.
As at the moment the Prometheus plugin cannot report individual queue totals, there is a configuration option that allows to list messages
, messages_ready
and messages_unacknowledged
in the queues
endpoint.
Below is a configuration example that disables the statistics but returns individual queue totals in the queues
page:
management.disable_stats = true management.enable_queue_totals = true
It is possible to configure what CSP header value is used by HTTP API responses. The default value is script-src 'self' 'unsafe-eval' 'unsafe-inline'; object-src 'self'
:
management.csp.policy = script-src 'self' 'unsafe-eval' 'unsafe-inline'; object-src 'self'
The value can be any valid CSP header string:
management.csp.policy = default-src https://rabbitmq.eng.example.local
Wildcards are also allowed:
management.csp.policy = default-src 'self' *.eng.example.local
A CSP policy frame-ancestors
directive can be used to prevent frame embedding of the management UI, mitigating certain types of cross-frame scripting attacks:
# prohibits iframe embedding of the UI management.csp.policy = frame-ancestors 'none'
It is possible to configure what Strict Transport Security header value is used by HTTP API responses:
management.hsts.policy = max-age=31536000; includeSubDomains
The management UI application will by default refuse access to websites hosted on origins different from its own using the Cross-Origin Resource Sharing mechanism, also known as CORS. It is possible to white list origins:
management.cors.allow_origins.1 = https://origin1.org management.cors.allow_origins.2 = https://origin2.org
It is possible to allow any origin to use the API using a wildcard. This is highly discouraged for deployments where the UI application may be exposed to the public.
management.cors.allow_origins.1 = *
The CORS pre-flight requests are cached by the browser. The management plugin defines a timeout of 30 minutes by default. The value can be changed. It is configured in seconds:
management.cors.allow_origins.1 = https://origin1.org management.cors.allow_origins.2 = https://origin2.org management.cors.max_age = 3600
It is possible to set a few more security-related headers for management UI and HTTP API responses. Note that some of them have been superseded by CORS and other newer development in the browser security space.
The supported headers are:
management.headers.content_type_options = nosniff management.headers.xss_protection = 1; mode=block management.headers.frame_options = DENY
They can be combined with the aforementioned CORS, HSTS, CSP headers:
management.hsts.policy = max-age=31536000; includeSubDomains management.csp.policy = default-src 'self'; script-src 'self' 'unsafe-eval' management.headers.content_type_options = nosniff management.headers.xss_protection = 1; mode=block management.headers.frame_options = DENY
After the user logs in, the user's web UI login session will expire after 8 hours by default. It is possible to configure a different timeout using the login_session_timeout
setting.
The value should be an integer: it controls the length of login session in minutes. When the time is up, the user will be signed out.
The following example sets the session timeout to 1 hour:
management.login_session_timeout = 60
Some environments require the use of a custom prefix for all HTTP requests to the management plugin. The management.path_prefix
setting allows an arbitrary prefix to be set for all HTTP request handlers in the management plugin.
Setting management.path_prefix
to /my-prefix
specifies all API requests to use the URI host:port/my-prefix/api/[...]
The management UI login page will have the URI host:port/my-prefix/
- note that the trailing slash is required in this case.
management.path_prefix = /my-prefix
An example configuration file for RabbitMQ that switches on request logging, increases the statistics interval to 10 seconds and explicitly sets some other relevant parameters to their default values, would look like this:
listeners.tcp.default = 5672 collect_statistics_interval = 10000 ## Note: this uses the core `load_definitions` key over ## now deprecated `management.load_definitions` # load_definitions = /path/to/exported/definitions.json management.tcp.port = 15672 management.tcp.ip = 0.0.0.0 management.ssl.port = 15671 management.ssl.ip = 0.0.0.0 management.ssl.cacertfile = /path/to/ca_certificate.pem management.ssl.certfile = /path/to/server_certificate.pem management.ssl.keyfile = /path/to/server_key.pem management.http_log_dir = /path/to/rabbit/logs/http management.rates_mode = basic # Configure how long aggregated data (such as message rates and queue # lengths) is retained. # Your can use 'minute', 'hour' and 'day' keys or integer key (in seconds) management.sample_retention_policies.global.minute = 5 management.sample_retention_policies.global.hour = 60 management.sample_retention_policies.global.day = 1200 management.sample_retention_policies.basic.minute = 5 management.sample_retention_policies.basic.hour = 60 management.sample_retention_policies.detailed.10 = 5
Nodes and clusters store information that can be thought of schema, metadata or topology. Users, vhosts, queues, exchanges, bindings, runtime parameters all fall into this category.
Definitions can be exported and imported via the rabbitmqctl
or the HTTP API provided by this plugin, including rabbitmqadmin
.
Please refer to the Definitions guide.
The management plugin is aware of clusters. It can be enabled on one or more nodes in a cluster, and see information pertaining to the entire cluster no matter which node you connect to.
It is possible to deploy management plugin only on a subset of cluster nodes. In that case only the nodes running the plugin would be able to serve client HTTP API requests. For every cluster node to have its metrics collected, it is still required that the rabbitmq-management-agent
plugin is enabled on each node, otherwise the metrics from the node won't be available.
In cluster, HTTP API performs cluster-wide queries when handling client requests, which means it can be affected by network partitions, network slowdowns, and unresponsive nodes.
Timeouts for inter-node aggregation queries are controlled via the net tick mechanism. Lowering the value may help reduce the delay introduced by nodes that have recently become unresponsive. Values lower than 10s can produce false positives and must be avoided.
In contrast to the HTTP API, the Prometheus monitoring endpoint only serves node-local data and is generally not affected by failures or unavailability of other nodes in the cluster.
It is possible to make the web UI available via any proxy that conforms with RFC 1738. The following sample Apache configuration illustrates the minimum necessary directives to coax Apache into conformance. It assumes a management web UI on the default port of 15672:
AllowEncodedSlashes NoDecode ProxyPass "/api" http://localhost:15672/api nocanon ProxyPass "/" http://localhost:15672/ ProxyPassReverse "/" http://localhost:15672/
Statistics database is stored entirely in memory. All of its contents is transient and should be treated as such.
In modern versions, ach node has its own statistics database containing a fraction of stats recorded on this node.
It is possible to restart the stats database on a given node using rabbitmqctl
or an HTTP API endpoint:
DELETE /api/reset/:node
rabbitmqctl eval 'rabbit_mgmt_storage:reset().'
To reset the entire statistics database on all nodes, use
DELETE /api/reset
rabbitmqctl eval 'rabbit_mgmt_storage:reset_all().'
Management UI can be used to inspect node's memory use, including displaying a per-category breakdown. See the Memory Use Analysis guide for details.
Management database builds around periodically emitted stats, regulated by the statistics interval described above, or when certain components are created/declared (e.g. a new connection or channel is opened, or a queue declared) or closed/deleted. Message rates do not directly affect management database memory usage.
Total amount of memory consumed by the stats database depends on the topology size (e.g. the number of queues), number of concurrent connections and channels, event emission interval, effective rates mode and retention policies.
Entities that emit stats (connections, channels, queues, nodes) do so periodically. The interval can be configured using the collect_statistics_interval
key:
# sets the interval to 30 seconds collect_statistics_interval = 30000
Increasing the interval value to 30-60s will reduce CPU footprint and peak memory consumption for systems with large amounts of connections, channels and queues. This comes with a downside: metrics of said entities will refresh every 30-60 seconds. This can be perfectly reasonable in an externally monitored production system but will make management UI less convenient to use for operators.
The memory usage of the channel and stats collector processes can be limited by setting the maximum backlog queue size using the parameter stats_event_max_backlog
. If the backlog queue is full, new channel and queue stats will be dropped until the previous ones have been processed.
The statistics interval can also be changed at runtime. Doing so will have no effect on existing connections, channels or queues. Only new stats emitting entities are affected.
rabbitmqctl eval 'application:set_env(rabbit, collect_statistics_interval, 60000).'
The statistics database can be restarted (see above) and thus forced to release all memory. Management UI's Overview page contains buttons that reset stats database for individual nodes as well as all nodes in the cluster.
It is possible to publish and consume messages using the HTTP API. This way of messaging is discouraged: prefer one of the binary messaging protocols supported by RabbitMQ. Publishing and consuming that way will be significantly more efficient and will provide access to various messaging protocol features such as confirmations.
Publishing over HTTP API can be useful in environments where long lived messaging protocol connections is not an option.