Rancher Release - v2.3.9


  • Please review the v2.3.0 release notes for important updates/ breaking changes if you are upgrading from a v2.2 release.

Please review our version documentation for more details on versioning and tagging conventions.


Updated Kubernetes Versions

Note: These were made available earlier, but are officially packaged into this release.

  • Updated to use v1.16.13-rancher1-1 and v1.17.9-rancher1-1
  • Kubernetes related CVEs
  1. CVE-2020-8557 - Node disk DOS by writing to container /etc/hosts
  2. CVE-2020-8558 - Node setting allows for neighboring hosts to bypass localhost boundary
  3. CVE-2020-8559 - Privilege escalation from compromised node to cluster
  • Updated Canal to use Flannel v0.12.0 [#27577]
  • Increase memory limit for Minio [#28025]

EKS Version Updates #28385 - Ability to provision k8s 1.15, 1.16, or 1.17 EKS clusters

Updated Tools

  • Istio v1.4.9 [#27222]
  • Monitoring v0.1.1 [#27465]
    • Updated Grafana to v6.7.4
    • Fixed an issue to address broken persistent volume

Experimental Features

We have the ability to turn on and off experimental components inside Rancher. You can manage feature flags through our UI. Alternatively, you can refer to our docs on how to turn on the features when starting Rancher.

Major Bugs Fixed Since v2.3.8

  • Fixed an issue with the k8s auth proxy where traffic was not being forwarded properly [#25488]
  • Fixed an issue where cluster provisioning failed when using a private registry with authentication [#28004]
  • Fixed a UI issue where users were unable to add a registry for all namespaces in a project [#27182]
  • Fixed a UI issue where cloned node templates were unusable [#27424]

Other notes

Air Gap Installations and Upgrades

In v2.3.0, an air gap install no longer requires mirroring the systems chart git repo. Please follow the directions on how to install Rancher to use the packaged systems chart.

Known Major Issues

  • NGINX ingress controller 0.25.0 doesn’t work on CPUs without SSE4.2 instruction set support [#23307]
  • Windows Limitations - There are a couple of known limitations with Windows due to upstream issues:
    • Windows pods cannot access the Kubernetes API when using VXLAN (Overlay) backend for the flannel network provider. The workaround is to use the Host Gateway (L2bridge) backend for the flannel network provider. [#20968]
    • Logging only works on Host Gateway (L2bridge) backend for the flannel network provider [#20510]
  • HPA Limitation - HPA UI doesn’t work on GKE clusters as GKE doesn’t support the v2beta2.autoscaling API [#22292]
  • Hardening Guide Limitations - If you have used Rancher’s hardening guide, there are some known issues
    • kubectl in UI doesn’t work [#19439]
    • Pipelines don’t work [#22844]
  • Adding taints to existing node templates from an upgraded setup will not be applied unless a reconcile is triggered on the cluster. When scaling up/down worker nodes, no reconcile is triggered, but scaling up/down either control plane/etcd nodes or editing a cluster (like upgrading to the latest Kubernetes version) would update to support taints on the nodes. [#22672]
  • Cluster alerting and logging can get stuck in Updating state after upgrading Rancher. Workaround steps are provided in the issue [21480]
  • If you have Rancher cluster with OpenStack cloud provider having LoadBalancer set, and the cluster was provisioned on version 2.2.3 or less, the upgrade to the Rancher version v2.2.4 and up will fail. Steps to mitigate can be found in the comment to [20699]
  • In clusters that have a Kubernetes cloud provider configured and have agents registered with hostname or FQDN (so not valid IP addresses), kube-proxy will fail to start. This can be checked in the API output for the node (customConfig -> address or internalAddress) [RKE#1725]
  • Rancher log collection format changed when upgrading the Fluentd Kubernetes metadata plugin. A json log is no longer parsed and put into the log as top level keys. Issue to optionally bring back this behavior[23646]



  • rancher/rancher:v2.3.9
  • rancher/rancher-agent:v2.3.9



Upgrades and Rollbacks

Rancher supports both upgrade and rollback. Please note the version you would like to upgrade or rollback to change the Rancher version.

Please be aware that upon an upgrade to v2.3.0+, any edits to a Rancher launched Kubernetes cluster will cause all system components to restart due to added tolerations to Kubernetes system components. Plan accordingly.

Recent changes to cert-manager require an upgrade if you have an HA install of Rancher using self-signed certificates. If you are using cert-manager older than v0.9.1, please see the documentation on how to upgrade cert-manager.

Important: When rolling back, we are expecting you to rollback to the state at the time of your upgrade. Any changes post upgrade would not be reflected.