Bitnami package for Valkey

Valkey is an open source (BSD) high-performance key/value datastore that supports a variety workloads such as caching, message queues, and can act as a primary database.

Overview of Valkey

Trademarks: This software listing is packaged by Bitnami. The respective trademarks mentioned in the offering are owned by the respective companies, and use of them does not imply any affiliation or endorsement.

TL;DR

helm install my-release oci://REGISTRY_NAME/REPOSITORY_NAME/valkey

Note: You need to substitute the placeholders REGISTRY_NAME and REPOSITORY_NAME with a reference to your Helm chart registry and repository.

Introduction

This chart bootstraps a Valkey deployment on a Kubernetes cluster using the Helm package manager.

Bitnami charts can be used with Kubeapps for deployment and management of Helm Charts in clusters.

Prerequisites

  • Kubernetes 1.23+
  • Helm 3.8.0+
  • PV provisioner support in the underlying infrastructure

Installing the Chart

To install the chart with the release name my-release:

helm install my-release oci://REGISTRY_NAME/REPOSITORY_NAME/valkey

Note: You need to substitute the placeholders REGISTRY_NAME and REPOSITORY_NAME with a reference to your Helm chart registry and repository. For example, in the case of Bitnami, you need to use REGISTRY_NAME=registry-1.docker.io and REPOSITORY_NAME=bitnamicharts.

The command deploys Valkey on the Kubernetes cluster in the default configuration. The Parameters section lists the parameters that can be configured during installation.

Tip: List all releases using helm list

Configuration and installation details

Resource requests and limits

Bitnami charts allow setting resource requests and limits for all containers inside the chart deployment. These are inside the resources value (check parameter table). Setting requests is essential for production workloads and these should be adapted to your specific use case.

To make this process easier, the chart contains the resourcesPreset values, which automatically sets the resources section according to different presets. Check these presets in the bitnami/common chart. However, in production workloads using resourcePreset is discouraged as it may not fully adapt to your specific needs. Find more information on container resource management in the official Kubernetes documentation.

Rolling VS Immutable tags

It is strongly recommended to use immutable tags in a production environment. This ensures your deployment does not change automatically if the same tag is updated with a different image.

Bitnami will release a new chart updating its containers if a new version of the main container, significant changes, or critical vulnerabilities exist.

Use a different Valkey version

To modify the application version used in this chart, specify a different version of the image using the image.tag parameter and/or a different repository using the image.repository parameter.

Bootstrapping with an External Cluster

This chart is equipped with the ability to bring online a set of Pods that connect to an existing Valkey deployment that lies outside of Kubernetes. This effectively creates a hybrid Valkey Deployment where both Pods in Kubernetes and Instances such as Virtual Machines can partake in a single Valkey Deployment. This is helpful in situations where one may be migrating Valkey from Virtual Machines into Kubernetes, for example. To take advantage of this, use the following as an example configuration:

replica:
  externalMaster:
    enabled: true
    host: external-valkey-0.internal
sentinel:
  externalMaster:
    enabled: true
    host: external-valkey-0.internal

:warning: This is currently limited to clusters in which Sentinel and Valkey run on the same node! :warning:

Please also note that the external sentinel must be listening on port 26379, and this is currently not configurable.

Once the Kubernetes Valkey Deployment is online and confirmed to be working with the existing cluster, the configuration can then be removed and the cluster will remain connected.

External DNS

This chart is equipped to allow leveraging the ExternalDNS project. Doing so will enable ExternalDNS to publish the FQDN for each instance, in the format of <pod-name>.<release-name>.<dns-suffix>. Example, when using the following configuration:

useExternalDNS:
  enabled: true
  suffix: prod.example.org
  additionalAnnotations:
    ttl: 10

On a cluster where the name of the Helm release is a, the hostname of a Pod is generated as: a-valkey-node-0.a-valkey.prod.example.org. The IP of that FQDN will match that of the associated Pod. This modifies the following parameters of the Valkey/Sentinel configuration using this new FQDN:

  • replica-announce-ip
  • known-sentinel
  • known-replica
  • announce-ip

:warning: This requires a working installation of external-dns to be fully functional. :warning:

See the official ExternalDNS documentation for additional configuration options.

Cluster topologies

Default: Master-Replicas

When installing the chart with architecture=replication, it will deploy a Valkey master StatefulSet and a Valkey replicas StatefulSet. The replicas will be read-replicas of the master. Two services will be exposed:

  • Valkey Master service: Points to the master, where read-write operations can be performed
  • Valkey Replicas service: Points to the replicas, where only read operations are allowed by default.

In case the master crashes, the replicas will wait until the master node is respawned again by the Kubernetes Controller Manager.

Standalone

When installing the chart with architecture=standalone, it will deploy a standalone Valkey StatefulSet. A single service will be exposed:

  • Valkey Master service: Points to the master, where read-write operations can be performed

Master-Replicas with Sentinel

When installing the chart with architecture=replication and sentinel.enabled=true, it will deploy a Valkey master StatefulSet (only one master allowed) and a Valkey replicas StatefulSet. In this case, the pods will contain an extra container with Valkey Sentinel. This container will form a cluster of Valkey Sentinel nodes, which will promote a new master in case the actual one fails.

On graceful termination of the Valkey master pod, a failover of the master is initiated to promote a new master. The Valkey Sentinel container in this pod will wait for the failover to occur before terminating. If sentinel.valkeyShutdownWaitFailover=true is set (the default), the Valkey container will wait for the failover as well before terminating. This increases availability for reads during failover, but may cause stale reads until all clients have switched to the new master.

In addition to this, only one service is exposed:

  • Valkey service: Exposes port 6379 for Valkey read-only operations and port 26379 for accessing Valkey Sentinel.

For read-only operations, access the service using port 6379. For write operations, it’s necessary to access the Valkey Sentinel cluster and query the current master using the command below (using valkey-cli or similar):

SENTINEL get-master-addr-by-name <name of your MasterSet. e.g: mymaster>

This command will return the address of the current master, which can be accessed from inside the cluster.

In case the current master crashes, the Sentinel containers will elect a new master node.

master.replicaCount greater than 1 is not designed for use when sentinel.enabled=true.

Multiple masters (experimental)

When master.replicaCount is greater than 1, special care must be taken to create a consistent setup.

An example of use case is the creation of a redundant set of standalone masters or master-replicas per Kubernetes node where you must ensure:

  • No more than 1 master can be deployed per Kubernetes node
  • Replicas and writers can only see the single master of their own Kubernetes node

One way of achieving this is by setting master.service.internalTrafficPolicy=Local in combination with a master.affinity.podAntiAffinity spec to never schedule more than one master per Kubernetes node.

It’s recommended to only change master.replicaCount if you know what you are doing. master.replicaCount greater than 1 is not designed for use when sentinel.enabled=true.

Using a password file

To use a password file for Valkey you need to create a secret containing the password and then deploy the chart using that secret. Follow these instructions:

  • Create the secret with the password. It is important that the file with the password must be called valkey-password.
kubectl create secret generic valkey-password-secret --from-file=valkey-password.yaml
  • Deploy the Helm Chart using the secret name as parameter:
usePassword=true
usePasswordFile=true
existingSecret=valkey-password-secret
sentinels.enabled=true
metrics.enabled=true

Securing traffic using TLS

TLS support can be enabled in the chart by specifying the tls. parameters while creating a release. The following parameters should be configured to properly enable the TLS support in the cluster:

  • tls.enabled: Enable TLS support. Defaults to false
  • tls.existingSecret: Name of the secret that contains the certificates. No defaults.
  • tls.certFilename: Certificate filename. No defaults.
  • tls.certKeyFilename: Certificate key filename. No defaults.
  • tls.certCAFilename: CA Certificate filename. No defaults.

For example:

First, create the secret with the certificates files:

kubectl create secret generic certificates-tls-secret --from-file=./cert.pem --from-file=./cert.key --from-file=./ca.pem

Then, use the following parameters:

tls.enabled="true"
tls.existingSecret="certificates-tls-secret"
tls.certFilename="cert.pem"
tls.certKeyFilename="cert.key"
tls.certCAFilename="ca.pem"

Metrics

The chart optionally can start a metrics exporter for prometheus. The metrics endpoint (port 9121) is exposed in the service. Metrics can be scraped from within the cluster using something similar as the described in the example Prometheus scrape configuration. If metrics are to be scraped from outside the cluster, the Kubernetes API proxy can be utilized to access the endpoint.

If you have enabled TLS by specifying tls.enabled=true you also need to specify TLS option to the metrics exporter. You can do that via metrics.extraArgs. You can find the metrics exporter CLI flags for TLS here. For example:

You can either specify metrics.extraArgs.skip-tls-verification=true to skip TLS verification or providing the following values under metrics.extraArgs for TLS client authentication:

tls-client-key-file
tls-client-cert-file
tls-ca-cert-file

Deploy a custom metrics script in the sidecar

A custom Lua script can be added to the redis-exporter sidecar by way of the metrics.extraArgs.script parameter. The pathname of the script must exist on the container, or the redis_exporter process (and therefore the whole pod) will refuse to start. The script can be provided to the sidecar containers via the metrics.extraVolumes and metrics.extraVolumeMounts parameters:

metrics:
  extraVolumeMounts:
    - name: '{{ printf "%s-metrics-script-file" (include "common.names.fullname" .) }}'
      mountPath: '{{ printf "/mnt/%s/" (include "common.names.name" .) }}'
      readOnly: true
  extraVolumes:
    - name: '{{ printf "%s-metrics-script-file" (include "common.names.fullname" .) }}'
      configMap:
        name: '{{ printf "%s-metrics-script" (include "common.names.fullname" .) }}'
  extraArgs:
    script: '{{ printf "/mnt/%s/my_custom_metrics.lua" (include "common.names.name" .) }}'

Then deploy the script into the correct location via extraDeploy:

extraDeploy:
  - apiVersion: v1
    kind: ConfigMap
    metadata:
      name: '{{ printf "%s-metrics-script" (include "common.names.fullname" .) }}'
    data:
      my_custom_metrics.lua: |
        -- LUA SCRIPT CODE HERE, e.g.,
        return {'bitnami_makes_the_best_charts', '1'}

Host Kernel Settings

Valkey may require some changes in the kernel of the host machine to work as expected, in particular increasing the somaxconn value and disabling transparent huge pages. To do so, you can set securityContext.sysctls which will configure sysctls for master and replica pods. Example:

securityContext:
  sysctls:
  - name: net.core.somaxconn
    value: "10000"

Note that this will not disable transparent huge tables.

Backup and restore

To backup and restore Valkey deployments on Kubernetes, you will need to create a snapshot of the data in the source cluster, and later restore it in a new cluster with the new parameters. Follow the instructions below:

Step 1: Backup the deployment

  • Connect to one of the nodes and start the Valkey CLI tool. Then, run the commands below:

    $ kubectl exec -it my-release-master-0 bash
    $ valkey-cli
    127.0.0.1:6379> auth your_current_valkey_password
    OK
    127.0.0.1:6379> save
    OK
    
  • Copy the dump file from the Valkey node:

    kubectl cp my-release-master-0:/data/dump.rdb dump.rdb -c valkey
    

Step 2: Restore the data on the destination cluster

To restore the data in a new cluster, you will need to create a PVC and then upload the dump.rdb file to the new volume.

Follow the following steps:

  • In the values.yaml file set the appendonly parameter to no. You can skip this step if it is already configured as no

    commonConfiguration: |-
       # Enable AOF https://valkey.io/topics/persistence#append-only-file
       appendonly no
       # Disable RDB persistence, AOF persistence already enabled.
       save ""
    

    Note that the Enable AOF comment belongs to the original config file and what you’re actually doing is disabling it. This change will only be neccessary for the temporal cluster you’re creating to upload the dump.

  • Start the new cluster to create the PVCs. Use the command below as an example:

    helm install new-valkey  -f values.yaml .  --set cluster.enabled=true  --set cluster.slaveCount=3
    
  • Now that the PVC were created, stop it and copy the dump.rdp file on the persisted data by using a helping pod.

    $ helm delete new-valkey
    
    $ kubectl run --generator=run-pod/v1 -i --rm --tty volpod --overrides='
    {
        "apiVersion": "v1",
        "kind": "Pod",
        "metadata": {
            "name": "valkeyvolpod"
        },
        "spec": {
            "containers": [{
               "command": [
                    "tail",
                    "-f",
                    "/dev/null"
               ],
               "image": "bitnami/os-shell",
               "name": "mycontainer",
               "volumeMounts": [{
                   "mountPath": "/mnt",
                   "name": "valkeydata"
                }]
            }],
            "restartPolicy": "Never",
            "volumes": [{
                "name": "valkeydata",
                "persistentVolumeClaim": {
                    "claimName": "valkey-data-new-valkey-master-0"
                }
            }]
        }
    }' --image="bitnami/os-shell"
    
    $ kubectl cp dump.rdb valkeyvolpod:/mnt/dump.rdb
    $ kubectl delete pod volpod
    
  • Restart the cluster:

    INFO: The appendonly parameter can be safely restored to your desired value.

    helm install new-valkey  -f values.yaml .  --set cluster.enabled=true  --set cluster.slaveCount=3
    

NetworkPolicy

To enable network policy for Valkey, install a networking plugin that implements the Kubernetes NetworkPolicy spec, and set networkPolicy.enabled to true.

With NetworkPolicy enabled, only pods with the generated client label will be able to connect to Valkey. This label will be displayed in the output after a successful install.

With networkPolicy.ingressNSMatchLabels pods from other namespaces can connect to Valkey. Set networkPolicy.ingressNSPodMatchLabels to match pod labels in matched namespace. For example, for a namespace labeled valkey=external and pods in that namespace labeled valkey-client=true the fields should be set:

networkPolicy:
  enabled: true
  ingressNSMatchLabels:
    valkey: external
  ingressNSPodMatchLabels:
    valkey-client: true

Setting Pod’s affinity

This chart allows you to set your custom affinity using the XXX.affinity parameter(s). Find more information about Pod’s affinity in the Kubernetes documentation.

As an alternative, you can use of the preset configurations for pod affinity, pod anti-affinity, and node affinity available at the bitnami/common chart. To do so, set the XXX.podAffinityPreset, XXX.podAntiAffinityPreset, or XXX.nodeAffinityPreset parameters.

Persistence

By default, the chart mounts a Persistent Volume at the /data path. The volume is created using dynamic volume provisioning. If a Persistent Volume Claim already exists, specify it during installation.

Existing PersistentVolumeClaim

  1. Create the PersistentVolume
  2. Create the PersistentVolumeClaim
  3. Install the chart
helm install my-release --set master.persistence.existingClaim=PVC_NAME oci://REGISTRY_NAME/REPOSITORY_NAME/valkey

Note: You need to substitute the placeholders REGISTRY_NAME and REPOSITORY_NAME with a reference to your Helm chart registry and repository. For example, in the case of Bitnami, you need to use REGISTRY_NAME=registry-1.docker.io and REPOSITORY_NAME=bitnamicharts.

Parameters

Global parameters

Name Description Value
global.imageRegistry Global Docker image registry ""
global.imagePullSecrets Global Docker registry secret names as an array []
global.storageClass Global StorageClass for Persistent Volume(s) ""
global.valkey.password Global Valkey password (overrides auth.password) ""
global.compatibility.openshift.adaptSecurityContext Adapt the securityContext sections of the deployment to make them compatible with Openshift restricted-v2 SCC: remove runAsUser, runAsGroup and fsGroup and let the platform use their allowed default IDs. Possible values: auto (apply if the detected running cluster is Openshift), force (perform the adaptation always), disabled (do not perform adaptation) auto

Common parameters

Name Description Value
kubeVersion Override Kubernetes version ""
nameOverride String to partially override common.names.fullname ""
fullnameOverride String to fully override common.names.fullname ""
namespaceOverride String to fully override common.names.namespace ""
commonLabels Labels to add to all deployed objects {}
commonAnnotations Annotations to add to all deployed objects {}
secretAnnotations Annotations to add to secret {}
clusterDomain Kubernetes cluster domain name cluster.local
extraDeploy Array of extra objects to deploy with the release []
useHostnames Use hostnames internally when announcing replication. If false, the hostname will be resolved to an IP address true
nameResolutionThreshold Failure threshold for internal hostnames resolution 5
nameResolutionTimeout Timeout seconds between probes for internal hostnames resolution 5
diagnosticMode.enabled Enable diagnostic mode (all probes will be disabled and the command will be overridden) false
diagnosticMode.command Command to override all containers in the deployment ["sleep"]
diagnosticMode.args Args to override all containers in the deployment ["infinity"]

Valkey Image parameters

Name Description Value
image.registry Valkey image registry REGISTRY_NAME
image.repository Valkey image repository REPOSITORY_NAME/valkey
image.digest Valkey image digest in the way sha256:aa…. Please note this parameter, if set, will override the tag ""
image.pullPolicy Valkey image pull policy IfNotPresent
image.pullSecrets Valkey image pull secrets []
image.debug Enable image debug mode false

Valkey common configuration parameters

Name Description Value
architecture Valkey architecture. Allowed values: standalone or replication replication
auth.enabled Enable password authentication true
auth.sentinel Enable password authentication on sentinels too true
auth.password Valkey password ""
auth.existingSecret The name of an existing secret with Valkey credentials ""
auth.existingSecretPasswordKey Password key to be retrieved from existing secret ""
auth.usePasswordFiles Mount credentials as files instead of using an environment variable false
auth.usePasswordFileFromSecret Mount password file from secret true
commonConfiguration Common configuration to be added into the ConfigMap ""
existingConfigmap The name of an existing ConfigMap with your custom configuration for Valkey nodes ""

Valkey master configuration parameters

Name Description Value
master.replicaCount Number of Valkey master instances to deploy (experimental, requires additional configuration) 1
master.configuration Configuration for Valkey master nodes ""
master.disableCommands Array with Valkey commands to disable on master nodes ["FLUSHDB","FLUSHALL"]
master.command Override default container command (useful when using custom images) []
master.args Override default container args (useful when using custom images) []
master.enableServiceLinks Whether information about services should be injected into pod’s environment variable true
master.preExecCmds Additional commands to run prior to starting Valkey master []
master.extraFlags Array with additional command line flags for Valkey master []
master.extraEnvVars Array with extra environment variables to add to Valkey master nodes []
master.extraEnvVarsCM Name of existing ConfigMap containing extra env vars for Valkey master nodes ""
master.extraEnvVarsSecret Name of existing Secret containing extra env vars for Valkey master nodes ""
master.containerPorts.valkey Container port to open on Valkey master nodes 6379
master.startupProbe.enabled Enable startupProbe on Valkey master nodes false
master.startupProbe.initialDelaySeconds Initial delay seconds for startupProbe 20
master.startupProbe.periodSeconds Period seconds for startupProbe 5
master.startupProbe.timeoutSeconds Timeout seconds for startupProbe 5
master.startupProbe.failureThreshold Failure threshold for startupProbe 5
master.startupProbe.successThreshold Success threshold for startupProbe 1
master.livenessProbe.enabled Enable livenessProbe on Valkey master nodes true
master.livenessProbe.initialDelaySeconds Initial delay seconds for livenessProbe 20
master.livenessProbe.periodSeconds Period seconds for livenessProbe 5
master.livenessProbe.timeoutSeconds Timeout seconds for livenessProbe 5
master.livenessProbe.failureThreshold Failure threshold for livenessProbe 5
master.livenessProbe.successThreshold Success threshold for livenessProbe 1
master.readinessProbe.enabled Enable readinessProbe on Valkey master nodes true
master.readinessProbe.initialDelaySeconds Initial delay seconds for readinessProbe 20
master.readinessProbe.periodSeconds Period seconds for readinessProbe 5
master.readinessProbe.timeoutSeconds Timeout seconds for readinessProbe 1
master.readinessProbe.failureThreshold Failure threshold for readinessProbe 5
master.readinessProbe.successThreshold Success threshold for readinessProbe 1
master.customStartupProbe Custom startupProbe that overrides the default one {}
master.customLivenessProbe Custom livenessProbe that overrides the default one {}
master.customReadinessProbe Custom readinessProbe that overrides the default one {}
master.resourcesPreset Set container resources according to one common preset (allowed values: none, nano, micro, small, medium, large, xlarge, 2xlarge). This is ignored if master.resources is set (master.resources is recommended for production). nano
master.resources Set container requests and limits for different resources like CPU or memory (essential for production workloads) {}
master.podSecurityContext.enabled Enabled Valkey master pods’ Security Context true
master.podSecurityContext.fsGroupChangePolicy Set filesystem group change policy Always
master.podSecurityContext.sysctls Set kernel settings using the sysctl interface []
master.podSecurityContext.supplementalGroups Set filesystem extra groups []
master.podSecurityContext.fsGroup Set Valkey master pod’s Security Context fsGroup 1001
master.containerSecurityContext.enabled Enabled Valkey master containers’ Security Context true
master.containerSecurityContext.seLinuxOptions Set SELinux options in container {}
master.containerSecurityContext.runAsUser Set Valkey master containers’ Security Context runAsUser 1001
master.containerSecurityContext.runAsGroup Set Valkey master containers’ Security Context runAsGroup 1001
master.containerSecurityContext.runAsNonRoot Set Valkey master containers’ Security Context runAsNonRoot true
master.containerSecurityContext.allowPrivilegeEscalation Is it possible to escalate Valkey pod(s) privileges false
master.containerSecurityContext.readOnlyRootFilesystem Set container’s Security Context read-only root filesystem true
master.containerSecurityContext.seccompProfile.type Set Valkey master containers’ Security Context seccompProfile RuntimeDefault
master.containerSecurityContext.capabilities.drop Set Valkey master containers’ Security Context capabilities to drop ["ALL"]
master.kind Use either Deployment, StatefulSet (default) or DaemonSet StatefulSet
master.schedulerName Alternate scheduler for Valkey master pods ""
master.updateStrategy.type Valkey master statefulset strategy type RollingUpdate
master.minReadySeconds How many seconds a pod needs to be ready before killing the next, during update 0
master.priorityClassName Valkey master pods’ priorityClassName ""
master.automountServiceAccountToken Mount Service Account token in pod false
master.hostAliases Valkey master pods host aliases []
master.podLabels Extra labels for Valkey master pods {}
master.podAnnotations Annotations for Valkey master pods {}
master.shareProcessNamespace Share a single process namespace between all of the containers in Valkey master pods false
master.podAffinityPreset Pod affinity preset. Ignored if master.affinity is set. Allowed values: soft or hard ""
master.podAntiAffinityPreset Pod anti-affinity preset. Ignored if master.affinity is set. Allowed values: soft or hard soft
master.nodeAffinityPreset.type Node affinity preset type. Ignored if master.affinity is set. Allowed values: soft or hard ""
master.nodeAffinityPreset.key Node label key to match. Ignored if master.affinity is set ""
master.nodeAffinityPreset.values Node label values to match. Ignored if master.affinity is set []
master.affinity Affinity for Valkey master pods assignment {}
master.nodeSelector Node labels for Valkey master pods assignment {}
master.tolerations Tolerations for Valkey master pods assignment []
master.topologySpreadConstraints Spread Constraints for Valkey master pod assignment []
master.dnsPolicy DNS Policy for Valkey master pod ""
master.dnsConfig DNS Configuration for Valkey master pod {}
master.lifecycleHooks for the Valkey master container(s) to automate configuration before or after startup {}
master.extraVolumes Optionally specify extra list of additional volumes for the Valkey master pod(s) []
master.extraVolumeMounts Optionally specify extra list of additional volumeMounts for the Valkey master container(s) []
master.sidecars Add additional sidecar containers to the Valkey master pod(s) []
master.initContainers Add additional init containers to the Valkey master pod(s) []
master.persistence.enabled Enable persistence on Valkey master nodes using Persistent Volume Claims true
master.persistence.medium Provide a medium for emptyDir volumes. ""
master.persistence.sizeLimit Set this to enable a size limit for emptyDir volumes. ""
master.persistence.path The path the volume will be mounted at on Valkey master containers /data
master.persistence.subPath The subdirectory of the volume to mount on Valkey master containers ""
master.persistence.subPathExpr Used to construct the subPath subdirectory of the volume to mount on Valkey master containers ""
master.persistence.storageClass Persistent Volume storage class ""
master.persistence.accessModes Persistent Volume access modes ["ReadWriteOnce"]
master.persistence.size Persistent Volume size 8Gi
master.persistence.annotations Additional custom annotations for the PVC {}
master.persistence.labels Additional custom labels for the PVC {}
master.persistence.selector Additional labels to match for the PVC {}
master.persistence.dataSource Custom PVC data source {}
master.persistence.existingClaim Use a existing PVC which must be created manually before bound ""
master.persistentVolumeClaimRetentionPolicy.enabled Controls if and how PVCs are deleted during the lifecycle of a StatefulSet false
master.persistentVolumeClaimRetentionPolicy.whenScaled Volume retention behavior when the replica count of the StatefulSet is reduced Retain
master.persistentVolumeClaimRetentionPolicy.whenDeleted Volume retention behavior that applies when the StatefulSet is deleted Retain
master.service.type Valkey master service type ClusterIP
master.service.ports.valkey Valkey master service port 6379
master.service.nodePorts.valkey Node port for Valkey master ""
master.service.externalTrafficPolicy Valkey master service external traffic policy Cluster
master.service.extraPorts Extra ports to expose (normally used with the sidecar value) []
master.service.internalTrafficPolicy Valkey master service internal traffic policy (requires Kubernetes v1.22 or greater to be usable) Cluster
master.service.clusterIP Valkey master service Cluster IP ""
master.service.loadBalancerIP Valkey master service Load Balancer IP ""
master.service.loadBalancerClass master service Load Balancer class if service type is LoadBalancer (optional, cloud specific) ""
master.service.loadBalancerSourceRanges Valkey master service Load Balancer sources []
master.service.externalIPs Valkey master service External IPs []
master.service.annotations Additional custom annotations for Valkey master service {}
master.service.sessionAffinity Session Affinity for Kubernetes service, can be “None” or “ClientIP” None
master.service.sessionAffinityConfig Additional settings for the sessionAffinity {}
master.terminationGracePeriodSeconds Integer setting the termination grace period for the valkey-master pods 30
master.serviceAccount.create Specifies whether a ServiceAccount should be created true
master.serviceAccount.name The name of the ServiceAccount to use. ""
master.serviceAccount.automountServiceAccountToken Whether to auto mount the service account token false
master.serviceAccount.annotations Additional custom annotations for the ServiceAccount {}
master.pdb.create Enable/disable a Pod Disruption Budget creation true
master.pdb.minAvailable Minimum number/percentage of pods that should remain scheduled {}
master.pdb.maxUnavailable Maximum number/percentage of pods that may be made unavailable. Defaults to 1 if both master.pdb.minAvailable and master.pdb.maxUnavailable are empty. {}

Valkey replicas configuration parameters

Name Description Value
replica.kind Use either DaemonSet or StatefulSet (default) StatefulSet
replica.replicaCount Number of Valkey replicas to deploy 3
replica.configuration Configuration for Valkey replicas nodes ""
replica.disableCommands Array with Valkey commands to disable on replicas nodes ["FLUSHDB","FLUSHALL"]
replica.command Override default container command (useful when using custom images) []
replica.args Override default container args (useful when using custom images) []
replica.enableServiceLinks Whether information about services should be injected into pod’s environment variable true
replica.preExecCmds Additional commands to run prior to starting Valkey replicas []
replica.extraFlags Array with additional command line flags for Valkey replicas []
replica.extraEnvVars Array with extra environment variables to add to Valkey replicas nodes []
replica.extraEnvVarsCM Name of existing ConfigMap containing extra env vars for Valkey replicas nodes ""
replica.extraEnvVarsSecret Name of existing Secret containing extra env vars for Valkey replicas nodes ""
replica.externalMaster.enabled Use external master for bootstrapping false
replica.externalMaster.host External master host to bootstrap from ""
replica.externalMaster.port Port for Valkey service external master host 6379
replica.containerPorts.valkey Container port to open on Valkey replicas nodes 6379
replica.startupProbe.enabled Enable startupProbe on Valkey replicas nodes true
replica.startupProbe.initialDelaySeconds Initial delay seconds for startupProbe 10
replica.startupProbe.periodSeconds Period seconds for startupProbe 10
replica.startupProbe.timeoutSeconds Timeout seconds for startupProbe 5
replica.startupProbe.failureThreshold Failure threshold for startupProbe 22
replica.startupProbe.successThreshold Success threshold for startupProbe 1
replica.livenessProbe.enabled Enable livenessProbe on Valkey replicas nodes true
replica.livenessProbe.initialDelaySeconds Initial delay seconds for livenessProbe 20
replica.livenessProbe.periodSeconds Period seconds for livenessProbe 5
replica.livenessProbe.timeoutSeconds Timeout seconds for livenessProbe 5
replica.livenessProbe.failureThreshold Failure threshold for livenessProbe 5
replica.livenessProbe.successThreshold Success threshold for livenessProbe 1
replica.readinessProbe.enabled Enable readinessProbe on Valkey replicas nodes true
replica.readinessProbe.initialDelaySeconds Initial delay seconds for readinessProbe 20
replica.readinessProbe.periodSeconds Period seconds for readinessProbe 5
replica.readinessProbe.timeoutSeconds Timeout seconds for readinessProbe 1
replica.readinessProbe.failureThreshold Failure threshold for readinessProbe 5
replica.readinessProbe.successThreshold Success threshold for readinessProbe 1
replica.customStartupProbe Custom startupProbe that overrides the default one {}
replica.customLivenessProbe Custom livenessProbe that overrides the default one {}
replica.customReadinessProbe Custom readinessProbe that overrides the default one {}
replica.resourcesPreset Set container resources according to one common preset (allowed values: none, nano, micro, small, medium, large, xlarge, 2xlarge). This is ignored if replica.resources is set (replica.resources is recommended for production). nano
replica.resources Set container requests and limits for different resources like CPU or memory (essential for production workloads) {}
replica.podSecurityContext.enabled Enabled Valkey replicas pods’ Security Context true
replica.podSecurityContext.fsGroupChangePolicy Set filesystem group change policy Always
replica.podSecurityContext.sysctls Set kernel settings using the sysctl interface []
replica.podSecurityContext.supplementalGroups Set filesystem extra groups []
replica.podSecurityContext.fsGroup Set Valkey replicas pod’s Security Context fsGroup 1001
replica.containerSecurityContext.enabled Enabled Valkey replicas containers’ Security Context true
replica.containerSecurityContext.seLinuxOptions Set SELinux options in container {}
replica.containerSecurityContext.runAsUser Set Valkey replicas containers’ Security Context runAsUser 1001
replica.containerSecurityContext.runAsGroup Set Valkey replicas containers’ Security Context runAsGroup 1001
replica.containerSecurityContext.runAsNonRoot Set Valkey replicas containers’ Security Context runAsNonRoot true
replica.containerSecurityContext.allowPrivilegeEscalation Set Valkey replicas pod’s Security Context allowPrivilegeEscalation false
replica.containerSecurityContext.readOnlyRootFilesystem Set container’s Security Context read-only root filesystem true
replica.containerSecurityContext.seccompProfile.type Set Valkey replicas containers’ Security Context seccompProfile RuntimeDefault
replica.containerSecurityContext.capabilities.drop Set Valkey replicas containers’ Security Context capabilities to drop ["ALL"]
replica.schedulerName Alternate scheduler for Valkey replicas pods ""
replica.updateStrategy.type Valkey replicas statefulset strategy type RollingUpdate
replica.minReadySeconds How many seconds a pod needs to be ready before killing the next, during update 0
replica.priorityClassName Valkey replicas pods’ priorityClassName ""
replica.podManagementPolicy podManagementPolicy to manage scaling operation of %%MAIN_CONTAINER_NAME%% pods ""
replica.automountServiceAccountToken Mount Service Account token in pod false
replica.hostAliases Valkey replicas pods host aliases []
replica.podLabels Extra labels for Valkey replicas pods {}
replica.podAnnotations Annotations for Valkey replicas pods {}
replica.shareProcessNamespace Share a single process namespace between all of the containers in Valkey replicas pods false
replica.podAffinityPreset Pod affinity preset. Ignored if replica.affinity is set. Allowed values: soft or hard ""
replica.podAntiAffinityPreset Pod anti-affinity preset. Ignored if replica.affinity is set. Allowed values: soft or hard soft
replica.nodeAffinityPreset.type Node affinity preset type. Ignored if replica.affinity is set. Allowed values: soft or hard ""
replica.nodeAffinityPreset.key Node label key to match. Ignored if replica.affinity is set ""
replica.nodeAffinityPreset.values Node label values to match. Ignored if replica.affinity is set []
replica.affinity Affinity for Valkey replicas pods assignment {}
replica.nodeSelector Node labels for Valkey replicas pods assignment {}
replica.tolerations Tolerations for Valkey replicas pods assignment []
replica.topologySpreadConstraints Spread Constraints for Valkey replicas pod assignment []
replica.dnsPolicy DNS Policy for Valkey replica pods ""
replica.dnsConfig DNS Configuration for Valkey replica pods {}
replica.lifecycleHooks for the Valkey replica container(s) to automate configuration before or after startup {}
replica.extraVolumes Optionally specify extra list of additional volumes for the Valkey replicas pod(s) []
replica.extraVolumeMounts Optionally specify extra list of additional volumeMounts for the Valkey replicas container(s) []
replica.sidecars Add additional sidecar containers to the Valkey replicas pod(s) []
replica.initContainers Add additional init containers to the Valkey replicas pod(s) []
replica.persistence.enabled Enable persistence on Valkey replicas nodes using Persistent Volume Claims true
replica.persistence.medium Provide a medium for emptyDir volumes. ""
replica.persistence.sizeLimit Set this to enable a size limit for emptyDir volumes. ""
replica.persistence.path The path the volume will be mounted at on Valkey replicas containers /data
replica.persistence.subPath The subdirectory of the volume to mount on Valkey replicas containers ""
replica.persistence.subPathExpr Used to construct the subPath subdirectory of the volume to mount on Valkey replicas containers ""
replica.persistence.storageClass Persistent Volume storage class ""
replica.persistence.accessModes Persistent Volume access modes ["ReadWriteOnce"]
replica.persistence.size Persistent Volume size 8Gi
replica.persistence.annotations Additional custom annotations for the PVC {}
replica.persistence.labels Additional custom labels for the PVC {}
replica.persistence.selector Additional labels to match for the PVC {}
replica.persistence.dataSource Custom PVC data source {}
replica.persistence.existingClaim Use a existing PVC which must be created manually before bound ""
replica.persistentVolumeClaimRetentionPolicy.enabled Controls if and how PVCs are deleted during the lifecycle of a StatefulSet false
replica.persistentVolumeClaimRetentionPolicy.whenScaled Volume retention behavior when the replica count of the StatefulSet is reduced Retain
replica.persistentVolumeClaimRetentionPolicy.whenDeleted Volume retention behavior that applies when the StatefulSet is deleted Retain
replica.service.type Valkey replicas service type ClusterIP
replica.service.ports.valkey Valkey replicas service port 6379
replica.service.nodePorts.valkey Node port for Valkey replicas ""
replica.service.externalTrafficPolicy Valkey replicas service external traffic policy Cluster
replica.service.internalTrafficPolicy Valkey replicas service internal traffic policy (requires Kubernetes v1.22 or greater to be usable) Cluster
replica.service.extraPorts Extra ports to expose (normally used with the sidecar value) []
replica.service.clusterIP Valkey replicas service Cluster IP ""
replica.service.loadBalancerIP Valkey replicas service Load Balancer IP ""
replica.service.loadBalancerClass replicas service Load Balancer class if service type is LoadBalancer (optional, cloud specific) ""
replica.service.loadBalancerSourceRanges Valkey replicas service Load Balancer sources []
replica.service.annotations Additional custom annotations for Valkey replicas service {}
replica.service.sessionAffinity Session Affinity for Kubernetes service, can be “None” or “ClientIP” None
replica.service.sessionAffinityConfig Additional settings for the sessionAffinity {}
replica.terminationGracePeriodSeconds Integer setting the termination grace period for the valkey-replicas pods 30

Autoscaling

Name Description Value
replica.autoscaling.vpa.enabled Enable VPA false
replica.autoscaling.vpa.annotations Annotations for VPA resource {}
replica.autoscaling.vpa.controlledResources VPA List of resources that the vertical pod autoscaler can control. Defaults to cpu and memory []
replica.autoscaling.vpa.maxAllowed VPA Max allowed resources for the pod {}
replica.autoscaling.vpa.minAllowed VPA Min allowed resources for the pod {}

VPA update policy

Name Description Value
replica.autoscaling.vpa.updatePolicy.updateMode Autoscaling update policy Specifies whether recommended updates are applied when a Pod is started and whether recommended updates are applied during the life of a Pod Auto
replica.autoscaling.hpa.enabled Enable HPA false
replica.autoscaling.hpa.minReplicas Minimum number of replicas ""
replica.autoscaling.hpa.maxReplicas Maximum number of replicas ""
replica.autoscaling.hpa.targetCPU Target CPU utilization percentage ""
replica.autoscaling.hpa.targetMemory Target Memory utilization percentage ""
replica.serviceAccount.create Specifies whether a ServiceAccount should be created true
replica.serviceAccount.name The name of the ServiceAccount to use. ""
replica.serviceAccount.automountServiceAccountToken Whether to auto mount the service account token false
replica.serviceAccount.annotations Additional custom annotations for the ServiceAccount {}
replica.pdb.create Enable/disable a Pod Disruption Budget creation true
replica.pdb.minAvailable Minimum number/percentage of pods that should remain scheduled {}
replica.pdb.maxUnavailable Maximum number/percentage of pods that may be made unavailable. Defaults to 1 if both replica.pdb.minAvailable and replica.pdb.maxUnavailable are empty. {}

Valkey Sentinel configuration parameters

Name Description Value
sentinel.enabled Use Valkey Sentinel on Valkey pods. false
sentinel.image.registry Valkey Sentinel image registry REGISTRY_NAME
sentinel.image.repository Valkey Sentinel image repository REPOSITORY_NAME/valkey-sentinel
sentinel.image.digest Valkey Sentinel image digest in the way sha256:aa…. Please note this parameter, if set, will override the tag ""
sentinel.image.pullPolicy Valkey Sentinel image pull policy IfNotPresent
sentinel.image.pullSecrets Valkey Sentinel image pull secrets []
sentinel.image.debug Enable image debug mode false
sentinel.annotations Additional custom annotations for Valkey Sentinel resource {}
sentinel.masterSet Master set name mymaster
sentinel.quorum Sentinel Quorum 2
sentinel.getMasterTimeout Amount of time to allow before get_sentinel_master_info() times out. 90
sentinel.automateClusterRecovery Automate cluster recovery in cases where the last replica is not considered a good replica and Sentinel won’t automatically failover to it. false
sentinel.valkeyShutdownWaitFailover Whether the Valkey master container waits for the failover at shutdown (in addition to the Valkey Sentinel container). true
sentinel.downAfterMilliseconds Timeout for detecting a Valkey node is down 60000
sentinel.failoverTimeout Timeout for performing a election failover 180000
sentinel.parallelSyncs Number of replicas that can be reconfigured in parallel to use the new master after a failover 1
sentinel.configuration Configuration for Valkey Sentinel nodes ""
sentinel.command Override default container command (useful when using custom images) []
sentinel.args Override default container args (useful when using custom images) []
sentinel.enableServiceLinks Whether information about services should be injected into pod’s environment variable true
sentinel.preExecCmds Additional commands to run prior to starting Valkey Sentinel []
sentinel.extraEnvVars Array with extra environment variables to add to Valkey Sentinel nodes []
sentinel.extraEnvVarsCM Name of existing ConfigMap containing extra env vars for Valkey Sentinel nodes ""
sentinel.extraEnvVarsSecret Name of existing Secret containing extra env vars for Valkey Sentinel nodes ""
sentinel.externalMaster.enabled Use external master for bootstrapping false
sentinel.externalMaster.host External master host to bootstrap from ""
sentinel.externalMaster.port Port for Valkey service external master host 6379
sentinel.containerPorts.sentinel Container port to open on Valkey Sentinel nodes 26379
sentinel.startupProbe.enabled Enable startupProbe on Valkey Sentinel nodes true
sentinel.startupProbe.initialDelaySeconds Initial delay seconds for startupProbe 10
sentinel.startupProbe.periodSeconds Period seconds for startupProbe 10
sentinel.startupProbe.timeoutSeconds Timeout seconds for startupProbe 5
sentinel.startupProbe.failureThreshold Failure threshold for startupProbe 22
sentinel.startupProbe.successThreshold Success threshold for startupProbe 1
sentinel.livenessProbe.enabled Enable livenessProbe on Valkey Sentinel nodes true
sentinel.livenessProbe.initialDelaySeconds Initial delay seconds for livenessProbe 20
sentinel.livenessProbe.periodSeconds Period seconds for livenessProbe 10
sentinel.livenessProbe.timeoutSeconds Timeout seconds for livenessProbe 5
sentinel.livenessProbe.failureThreshold Failure threshold for livenessProbe 6
sentinel.livenessProbe.successThreshold Success threshold for livenessProbe 1
sentinel.readinessProbe.enabled Enable readinessProbe on Valkey Sentinel nodes true
sentinel.readinessProbe.initialDelaySeconds Initial delay seconds for readinessProbe 20
sentinel.readinessProbe.periodSeconds Period seconds for readinessProbe 5
sentinel.readinessProbe.timeoutSeconds Timeout seconds for readinessProbe 1
sentinel.readinessProbe.failureThreshold Failure threshold for readinessProbe 6
sentinel.readinessProbe.successThreshold Success threshold for readinessProbe 1
sentinel.customStartupProbe Custom startupProbe that overrides the default one {}
sentinel.customLivenessProbe Custom livenessProbe that overrides the default one {}
sentinel.customReadinessProbe Custom readinessProbe that overrides the default one {}
sentinel.persistence.enabled Enable persistence on Valkey sentinel nodes using Persistent Volume Claims (Experimental) false
sentinel.persistence.storageClass Persistent Volume storage class ""
sentinel.persistence.accessModes Persistent Volume access modes ["ReadWriteOnce"]
sentinel.persistence.size Persistent Volume size 100Mi
sentinel.persistence.annotations Additional custom annotations for the PVC {}
sentinel.persistence.labels Additional custom labels for the PVC {}
sentinel.persistence.selector Additional labels to match for the PVC {}
sentinel.persistence.dataSource Custom PVC data source {}
sentinel.persistence.medium Provide a medium for emptyDir volumes. ""
sentinel.persistence.sizeLimit Set this to enable a size limit for emptyDir volumes. ""
sentinel.persistentVolumeClaimRetentionPolicy.enabled Controls if and how PVCs are deleted during the lifecycle of a StatefulSet false
sentinel.persistentVolumeClaimRetentionPolicy.whenScaled Volume retention behavior when the replica count of the StatefulSet is reduced Retain
sentinel.persistentVolumeClaimRetentionPolicy.whenDeleted Volume retention behavior that applies when the StatefulSet is deleted Retain
sentinel.resourcesPreset Set container resources according to one common preset (allowed values: none, nano, micro, small, medium, large, xlarge, 2xlarge). This is ignored if sentinel.resources is set (sentinel.resources is recommended for production). nano
sentinel.resources Set container requests and limits for different resources like CPU or memory (essential for production workloads) {}
sentinel.containerSecurityContext.enabled Enabled Valkey Sentinel containers’ Security Context true
sentinel.containerSecurityContext.seLinuxOptions Set SELinux options in container {}
sentinel.containerSecurityContext.runAsUser Set Valkey Sentinel containers’ Security Context runAsUser 1001
sentinel.containerSecurityContext.runAsGroup Set Valkey Sentinel containers’ Security Context runAsGroup 1001
sentinel.containerSecurityContext.runAsNonRoot Set Valkey Sentinel containers’ Security Context runAsNonRoot true
sentinel.containerSecurityContext.readOnlyRootFilesystem Set container’s Security Context read-only root filesystem true
sentinel.containerSecurityContext.allowPrivilegeEscalation Set Valkey Sentinel containers’ Security Context allowPrivilegeEscalation false
sentinel.containerSecurityContext.seccompProfile.type Set Valkey Sentinel containers’ Security Context seccompProfile RuntimeDefault
sentinel.containerSecurityContext.capabilities.drop Set Valkey Sentinel containers’ Security Context capabilities to drop ["ALL"]
sentinel.lifecycleHooks for the Valkey sentinel container(s) to automate configuration before or after startup {}
sentinel.extraVolumes Optionally specify extra list of additional volumes for the Valkey Sentinel []
sentinel.extraVolumeMounts Optionally specify extra list of additional volumeMounts for the Valkey Sentinel container(s) []
sentinel.service.type Valkey Sentinel service type ClusterIP
sentinel.service.ports.valkey Valkey service port for Valkey 6379
sentinel.service.ports.sentinel Valkey service port for Valkey Sentinel 26379
sentinel.service.nodePorts.valkey Node port for Valkey ""
sentinel.service.nodePorts.sentinel Node port for Sentinel ""
sentinel.service.externalTrafficPolicy Valkey Sentinel service external traffic policy Cluster
sentinel.service.extraPorts Extra ports to expose (normally used with the sidecar value) []
sentinel.service.clusterIP Valkey Sentinel service Cluster IP ""
sentinel.service.createMaster Enable master service pointing to the current master (experimental) false
sentinel.service.loadBalancerIP Valkey Sentinel service Load Balancer IP ""
sentinel.service.loadBalancerClass sentinel service Load Balancer class if service type is LoadBalancer (optional, cloud specific) ""
sentinel.service.loadBalancerSourceRanges Valkey Sentinel service Load Balancer sources []
sentinel.service.annotations Additional custom annotations for Valkey Sentinel service {}
sentinel.service.sessionAffinity Session Affinity for Kubernetes service, can be “None” or “ClientIP” None
sentinel.service.sessionAffinityConfig Additional settings for the sessionAffinity {}
sentinel.service.headless.annotations Annotations for the headless service. {}
sentinel.terminationGracePeriodSeconds Integer setting the termination grace period for the valkey-node pods 30

Other Parameters

Name Description Value
serviceBindings.enabled Create secret for service binding (Experimental) false
networkPolicy.enabled Enable creation of NetworkPolicy resources true
networkPolicy.allowExternal Don’t require client label for connections true
networkPolicy.allowExternalEgress Allow the pod to access any range of port and all destinations. true
networkPolicy.extraIngress Add extra ingress rules to the NetworkPolicy []
networkPolicy.extraEgress Add extra egress rules to the NetworkPolicy []
networkPolicy.ingressNSMatchLabels Labels to match to allow traffic from other namespaces {}
networkPolicy.ingressNSPodMatchLabels Pod labels to match to allow traffic from other namespaces {}
networkPolicy.metrics.allowExternal Don’t require client label for connections for metrics endpoint true
networkPolicy.metrics.ingressNSMatchLabels Labels to match to allow traffic from other namespaces to metrics endpoint {}
networkPolicy.metrics.ingressNSPodMatchLabels Pod labels to match to allow traffic from other namespaces to metrics endpoint {}
podSecurityPolicy.create Whether to create a PodSecurityPolicy. WARNING: PodSecurityPolicy is deprecated in Kubernetes v1.21 or later, unavailable in v1.25 or later false
podSecurityPolicy.enabled Enable PodSecurityPolicy’s RBAC rules false
rbac.create Specifies whether RBAC resources should be created false
rbac.rules Custom RBAC rules to set []
serviceAccount.create Specifies whether a ServiceAccount should be created true
serviceAccount.name The name of the ServiceAccount to use. ""
serviceAccount.automountServiceAccountToken Whether to auto mount the service account token false
serviceAccount.annotations Additional custom annotations for the ServiceAccount {}
pdb DEPRECATED Please use master.pdb and replica.pdb values instead {}
tls.enabled Enable TLS traffic false
tls.authClients Require clients to authenticate true
tls.autoGenerated Enable autogenerated certificates false
tls.existingSecret The name of the existing secret that contains the TLS certificates ""
tls.certFilename Certificate filename ""
tls.certKeyFilename Certificate Key filename ""
tls.certCAFilename CA Certificate filename ""
tls.dhParamsFilename File containing DH params (in order to support DH based ciphers) ""

Metrics Parameters

Name Description Value
metrics.enabled Start a sidecar prometheus exporter to expose Valkey metrics false
metrics.image.registry Valkey Exporter image registry REGISTRY_NAME
metrics.image.repository Valkey Exporter image repository REPOSITORY_NAME/valkey-exporter
metrics.image.digest Valkey Exporter image digest in the way sha256:aa…. Please note this parameter, if set, will override the tag ""
metrics.image.pullPolicy Valkey Exporter image pull policy IfNotPresent
metrics.image.pullSecrets Valkey Exporter image pull secrets []
metrics.containerPorts.http Metrics HTTP container port 9121
metrics.startupProbe.enabled Enable startupProbe on Valkey replicas nodes false
metrics.startupProbe.initialDelaySeconds Initial delay seconds for startupProbe 10
metrics.startupProbe.periodSeconds Period seconds for startupProbe 10
metrics.startupProbe.timeoutSeconds Timeout seconds for startupProbe 5
metrics.startupProbe.failureThreshold Failure threshold for startupProbe 5
metrics.startupProbe.successThreshold Success threshold for startupProbe 1
metrics.livenessProbe.enabled Enable livenessProbe on Valkey replicas nodes true
metrics.livenessProbe.initialDelaySeconds Initial delay seconds for livenessProbe 10
metrics.livenessProbe.periodSeconds Period seconds for livenessProbe 10
metrics.livenessProbe.timeoutSeconds Timeout seconds for livenessProbe 5
metrics.livenessProbe.failureThreshold Failure threshold for livenessProbe 5
metrics.livenessProbe.successThreshold Success threshold for livenessProbe 1
metrics.readinessProbe.enabled Enable readinessProbe on Valkey replicas nodes true
metrics.readinessProbe.initialDelaySeconds Initial delay seconds for readinessProbe 5
metrics.readinessProbe.periodSeconds Period seconds for readinessProbe 10
metrics.readinessProbe.timeoutSeconds Timeout seconds for readinessProbe 1
metrics.readinessProbe.failureThreshold Failure threshold for readinessProbe 3
metrics.readinessProbe.successThreshold Success threshold for readinessProbe 1
metrics.customStartupProbe Custom startupProbe that overrides the default one {}
metrics.customLivenessProbe Custom livenessProbe that overrides the default one {}
metrics.customReadinessProbe Custom readinessProbe that overrides the default one {}
metrics.command Override default metrics container init command (useful when using custom images) []
metrics.valkeyTargetHost A way to specify an alternative Valkey hostname localhost
metrics.extraArgs Extra arguments for Valkey exporter, for example: {}
metrics.extraEnvVars Array with extra environment variables to add to Valkey exporter []
metrics.containerSecurityContext.enabled Enabled Valkey exporter containers’ Security Context true
metrics.containerSecurityContext.seLinuxOptions Set SELinux options in container {}
metrics.containerSecurityContext.runAsUser Set Valkey exporter containers’ Security Context runAsUser 1001
metrics.containerSecurityContext.runAsGroup Set Valkey exporter containers’ Security Context runAsGroup 1001
metrics.containerSecurityContext.runAsNonRoot Set Valkey exporter containers’ Security Context runAsNonRoot true
metrics.containerSecurityContext.allowPrivilegeEscalation Set Valkey exporter containers’ Security Context allowPrivilegeEscalation false
metrics.containerSecurityContext.readOnlyRootFilesystem Set container’s Security Context read-only root filesystem true
metrics.containerSecurityContext.seccompProfile.type Set Valkey exporter containers’ Security Context seccompProfile RuntimeDefault
metrics.containerSecurityContext.capabilities.drop Set Valkey exporter containers’ Security Context capabilities to drop ["ALL"]
metrics.extraVolumes Optionally specify extra list of additional volumes for the Valkey metrics sidecar []
metrics.extraVolumeMounts Optionally specify extra list of additional volumeMounts for the Valkey metrics sidecar []
metrics.resourcesPreset Set container resources according to one common preset (allowed values: none, nano, micro, small, medium, large, xlarge, 2xlarge). This is ignored if metrics.resources is set (metrics.resources is recommended for production). nano
metrics.resources Set container requests and limits for different resources like CPU or memory (essential for production workloads) {}
metrics.podLabels Extra labels for Valkey exporter pods {}
metrics.podAnnotations Annotations for Valkey exporter pods {}
metrics.service.enabled Create Service resource(s) for scraping metrics using PrometheusOperator ServiceMonitor, can be disabled when using a PodMonitor true
metrics.service.type Valkey exporter service type ClusterIP
metrics.service.ports.http Valkey exporter service port 9121
metrics.service.externalTrafficPolicy Valkey exporter service external traffic policy Cluster
metrics.service.extraPorts Extra ports to expose (normally used with the sidecar value) []
metrics.service.loadBalancerIP Valkey exporter service Load Balancer IP ""
metrics.service.loadBalancerClass exporter service Load Balancer class if service type is LoadBalancer (optional, cloud specific) ""
metrics.service.loadBalancerSourceRanges Valkey exporter service Load Balancer sources []
metrics.service.annotations Additional custom annotations for Valkey exporter service {}
metrics.service.clusterIP Valkey exporter service Cluster IP ""
metrics.serviceMonitor.port the service port to scrape metrics from http-metrics
metrics.serviceMonitor.enabled Create ServiceMonitor resource(s) for scraping metrics using PrometheusOperator false
metrics.serviceMonitor.namespace The namespace in which the ServiceMonitor will be created ""
metrics.serviceMonitor.interval The interval at which metrics should be scraped 30s
metrics.serviceMonitor.scrapeTimeout The timeout after which the scrape is ended ""
metrics.serviceMonitor.relabelings Metrics RelabelConfigs to apply to samples before scraping. []
metrics.serviceMonitor.metricRelabelings Metrics RelabelConfigs to apply to samples before ingestion. []
metrics.serviceMonitor.honorLabels Specify honorLabels parameter to add the scrape endpoint false
metrics.serviceMonitor.additionalLabels Additional labels that can be used so ServiceMonitor resource(s) can be discovered by Prometheus {}
metrics.serviceMonitor.podTargetLabels Labels from the Kubernetes pod to be transferred to the created metrics []
metrics.serviceMonitor.sampleLimit Limit of how many samples should be scraped from every Pod false
metrics.serviceMonitor.targetLimit Limit of how many targets should be scraped false
metrics.serviceMonitor.additionalEndpoints Additional endpoints to scrape (e.g sentinel) []
metrics.podMonitor.port the pod port to scrape metrics from metrics
metrics.podMonitor.enabled Create PodMonitor resource(s) for scraping metrics using PrometheusOperator false
metrics.podMonitor.namespace The namespace in which the PodMonitor will be created ""
metrics.podMonitor.interval The interval at which metrics should be scraped 30s
metrics.podMonitor.scrapeTimeout The timeout after which the scrape is ended ""
metrics.podMonitor.relabelings Metrics RelabelConfigs to apply to samples before scraping. []
metrics.podMonitor.metricRelabelings Metrics RelabelConfigs to apply to samples before ingestion. []
metrics.podMonitor.honorLabels Specify honorLabels parameter to add the scrape endpoint false
metrics.podMonitor.additionalLabels Additional labels that can be used so PodMonitor resource(s) can be discovered by Prometheus {}
metrics.podMonitor.podTargetLabels Labels from the Kubernetes pod to be transferred to the created metrics []
metrics.podMonitor.sampleLimit Limit of how many samples should be scraped from every Pod false
metrics.podMonitor.targetLimit Limit of how many targets should be scraped false
metrics.podMonitor.additionalEndpoints Additional endpoints to scrape (e.g sentinel) []
metrics.prometheusRule.enabled Create a custom prometheusRule Resource for scraping metrics using PrometheusOperator false
metrics.prometheusRule.namespace The namespace in which the prometheusRule will be created ""
metrics.prometheusRule.additionalLabels Additional labels for the prometheusRule {}
metrics.prometheusRule.rules Custom Prometheus rules []

Init Container Parameters

Name Description Value
volumePermissions.enabled Enable init container that changes the owner/group of the PV mount point to runAsUser:fsGroup false
volumePermissions.image.registry OS Shell + Utility image registry REGISTRY_NAME
volumePermissions.image.repository OS Shell + Utility image repository REPOSITORY_NAME/os-shell
volumePermissions.image.digest OS Shell + Utility image digest in the way sha256:aa…. Please note this parameter, if set, will override the tag ""
volumePermissions.image.pullPolicy OS Shell + Utility image pull policy IfNotPresent
volumePermissions.image.pullSecrets OS Shell + Utility image pull secrets []
volumePermissions.resourcesPreset Set container resources according to one common preset (allowed values: none, nano, micro, small, medium, large, xlarge, 2xlarge). This is ignored if volumePermissions.resources is set (volumePermissions.resources is recommended for production). nano
volumePermissions.resources Set container requests and limits for different resources like CPU or memory (essential for production workloads) {}
volumePermissions.containerSecurityContext.seLinuxOptions Set SELinux options in container {}
volumePermissions.containerSecurityContext.runAsUser Set init container’s Security Context runAsUser 0
kubectl.image.registry Kubectl image registry REGISTRY_NAME
kubectl.image.repository Kubectl image repository REPOSITORY_NAME/kubectl
kubectl.image.digest Kubectl image digest in the way sha256:aa…. Please note this parameter, if set, will override the tag ""
kubectl.image.pullPolicy Kubectl image pull policy IfNotPresent
kubectl.image.pullSecrets Kubectl pull secrets []
kubectl.command kubectl command to execute ["/opt/bitnami/scripts/kubectl-scripts/update-master-label.sh"]
kubectl.containerSecurityContext.enabled Enabled kubectl containers’ Security Context true
kubectl.containerSecurityContext.seLinuxOptions Set SELinux options in container {}
kubectl.containerSecurityContext.runAsUser Set kubectl containers’ Security Context runAsUser 1001
kubectl.containerSecurityContext.runAsGroup Set kubectl containers’ Security Context runAsGroup 1001
kubectl.containerSecurityContext.runAsNonRoot Set kubectl containers’ Security Context runAsNonRoot true
kubectl.containerSecurityContext.allowPrivilegeEscalation Set kubectl containers’ Security Context allowPrivilegeEscalation false
kubectl.containerSecurityContext.readOnlyRootFilesystem Set container’s Security Context read-only root filesystem true
kubectl.containerSecurityContext.seccompProfile.type Set kubectl containers’ Security Context seccompProfile RuntimeDefault
kubectl.containerSecurityContext.capabilities.drop Set kubectl containers’ Security Context capabilities to drop ["ALL"]
kubectl.resources.limits The resources limits for the kubectl containers {}
kubectl.resources.requests The requested resources for the kubectl containers {}

useExternalDNS Parameters

Name Description Value
useExternalDNS.enabled Enable various syntax that would enable external-dns to work. Note this requires a working installation of external-dns to be usable. false
useExternalDNS.additionalAnnotations Extra annotations to be utilized when external-dns is enabled. {}
useExternalDNS.annotationKey The annotation key utilized when external-dns is enabled. Setting this to false will disable annotations. external-dns.alpha.kubernetes.io/
useExternalDNS.suffix The DNS suffix utilized when external-dns is enabled. Note that we prepend the suffix with the full name of the release. ""

Specify each parameter using the --set key=value[,key=value] argument to helm install. For example,

helm install my-release \
  --set auth.password=secretpassword \
    oci://REGISTRY_NAME/REPOSITORY_NAME/valkey

Note: You need to substitute the placeholders REGISTRY_NAME and REPOSITORY_NAME with a reference to your Helm chart registry and repository. For example, in the case of Bitnami, you need to use REGISTRY_NAME=registry-1.docker.io and REPOSITORY_NAME=bitnamicharts.

The above command sets the Valkey server password to secretpassword.

NOTE: Once this chart is deployed, it is not possible to change the application’s access credentials, such as usernames or passwords, using Helm. To change these application credentials after deployment, delete any persistent volumes (PVs) used by the chart and re-deploy it, or use the application’s built-in administrative tools if available.

Alternatively, a YAML file that specifies the values for the parameters can be provided while installing the chart. For example,

helm install my-release -f values.yaml oci://REGISTRY_NAME/REPOSITORY_NAME/valkey

Note: You need to substitute the placeholders REGISTRY_NAME and REPOSITORY_NAME with a reference to your Helm chart registry and repository. For example, in the case of Bitnami, you need to use REGISTRY_NAME=registry-1.docker.io and REPOSITORY_NAME=bitnamicharts. Tip: You can use the default values.yaml

Troubleshooting

Find more information about how to deal with common errors related to Bitnami’s Helm charts in this troubleshooting guide.

License

Copyright © 2024 Broadcom. The term “Broadcom” refers to Broadcom Inc. and/or its subsidiaries.

Licensed under the Apache License, Version 2.0 (the “License”); you may not use this file except in compliance with the License. You may obtain a copy of the License at

http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an “AS IS” BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.

check-circle-line exclamation-circle-line close-line
Scroll to top icon