Redis Enterprise for Kubernetes release notes 6.2.18-41 (Dec 2022)

This is a maintenance release for 6.2.18 and includes bug fixes.


The Redis Enterprise K8s 6.2.18-41 is a maintenance release which supports the Redis Enterprise Software release 6.2.18 and contains bug fixes.

The key bug fixes and known limitations are described below.


Due to a change in the SCC, on OpenShift clusters running version 6.2.12 or earlier upgrading to version 6.2.18 or later, where node:1 is not the master node, the upgrade might get stuck.

For more info and steps to prevent this issue, see upgrade a Redis Enterprise cluster (REC).


  • Redis Enterprise: redislabs/redis:6.2.18-65
  • Operator: redislabs/operator:6.2.18-41
  • Services Rigger: redislabs/k8s-controller:6.2.18-41

OpenShift images

OLM operator version: v6.2.18-41a

  • Redis Enterprise: (or redislabs/redis-enterprise:6.2.18-65.rhel7-openshift if upgrading from RHEL 7)
  • Operator:
  • Services Rigger:

Bug fixes

  • Fixed issues with incorrect image digest (OLM/OpenShift) (RED-88863)
  • Fixed upgrade issue for using NGINX-based ingress with Active-Active databases. (RED-88882)
  • Upgraded to Redis Software 6.2.18-65 (RED-88985)
  • Fixed upgrade issue with OLM (RED-89170)
  • Removed non-core Python dependencies from the log collector (RED-90108)
  • Fixed issues with running log collector on OLM (RED-90129

Compatibility notes

Below is a table showing supported distributions at the time of this release. See Supported Kubernetes distributions for the current list of supported distributions.

Kubernetes version 1.21 1.22 1.23 1.24 1.25
Community Kubernetes supported supported supported supported
Amazon EKS deprecated supported supported
Azure AKS supported supported supported
Google GKE deprecated supported supported supported supported
Rancher 2.6 supported supported supported supported
VMware TKG 1.6 supported supported
OpenShift version 4.8 4.9 4.10 4.11
deprecated supported supported
VMware TKGI version 1.12 1.13 1.14 1.15
supported supported supported

* Support added in this release

Before upgrading

Be aware the following changes included in this release affect the upgrade process. Please read carefully before upgrading.

Supported upgrade paths

If you are using a version earlier than 6.2.10-45, you cannot upgrade directly to this release. You must upgrade to 6.2.10-45 before you can upgrade to versions 6.2.18 or later.

Known limitations

Large clusters

On clusters with more than 9 REC nodes, a Kubernetes upgrade can render the Redis cluster unresponsive in some cases. A fix is available in the 6.4.2-5 release. Upgrade your operator version to 6.4.2-5 or later before upgrading your Kubernetes cluster. (RED-93025)

  • Long cluster names cause routes to be rejected (RED-25871)

    A cluster name longer than 20 characters will result in a rejected route configuration because the host part of the domain name will exceed 63 characters. The workaround is to limit cluster name to 20 characters or less.

  • Cluster CR (REC) errors are not reported after invalid updates (RED-25542)

    A cluster CR specification error is not reported if two or more invalid CR resources are updated in sequence.

  • An unreachable cluster has status running (RED-32805)

    When a cluster is in an unreachable state, the state is still running instead of being reported as an error.

  • Readiness probe incorrect on failures (RED-39300)

    STS Readiness probe does not mark a node as "not ready" when running rladmin status on node failure.

  • Internal DNS and Kubernetes DNS may have conflicts (RED-37462)

    DNS conflicts are possible between the cluster mdns_server and the K8s DNS. This only impacts DNS resolution from within cluster nodes for Kubernetes DNS names.

  • 5.4.10 negatively impacts 5.4.6 (RED-37233)

    Kubernetes-based 5.4.10 deployments seem to negatively impact existing 5.4.6 deployments that share a Kubernetes cluster.

  • Node CPU usage is reported instead of pod CPU usage (RED-36884)

    In Kubernetes, the node CPU usage we report on is the usage of the Kubernetes worker node hosting the REC pod.

  • Clusters must be named "rec" in OLM-based deployments (RED-39825)

    In OLM-deployed operators, the deployment of the cluster will fail if the name is not "rec". When the operator is deployed via the OLM, the security context constraints (scc) are bound to a specific service account name (i.e., "rec"). The workaround is to name the cluster "rec".

  • REC clusters fail to start on Kubernetes clusters with unsynchronized clocks (RED-47254)

    When REC clusters are deployed on Kubernetes clusters with unsynchronized clocks, the REC cluster does not start correctly. The fix is to use NTP to synchronize the underlying K8s nodes.

  • Deleting an OpenShift project with an REC deployed may hang (RED-47192)

    When a REC cluster is deployed in a project (namespace) and has REDB resources, the REDB resources must be deleted first before the REC can be deleted. As such, until the REDB resources are deleted, the project deletion will hang. The fix is to delete the REDB resources first and the REC second. Afterwards, you may delete the project.

  • Hashicorp Vault integration - no support for Gesher (RED-55080)

    There is no workaround at this time.

  • REC might report error states on initial startup (RED-61707)

    There is no workaround at this time except to ignore the errors.

  • PVC size issues when using decimal value in spec (RED-62132)

    The workaround for this issue is to make sure you use integer values for the PVC size.

  • Following old revision of quick start guide causes issues creating an REDB due to unrecognized memory field name (RED-69515)

    The workaround is to use the newer (current) revision of the quick start document available online.

  • autoUpgrade set to true by operator might cause unexpected bdb upgrades when redisUpgradePolicy is set to true (RED-72351)

    Contact support if your deployment is impacted.

Back to top ↑