Skip to content

Latest commit

 

History

History
71 lines (54 loc) · 5.33 KB

File metadata and controls

71 lines (54 loc) · 5.33 KB

Major Themes

Action Required

Notable Features

  • Added K8s 1.16 to the test matrix and removed K8s 1.11 from the test matrix.
  • Discover daemon:
    • When the Storage Operator is deleted, the Discover Daemon will also be deleted, as well as its Config Map
    • Device filtering is now configurable for the user by adding an environment variable
      • A new environment variable DISCOVER_DAEMON_UDEV_BLACKLIST is added through which the user can blacklist the devices
      • If no device is specified, the default values will be used to blacklist the devices

Ceph

  • The job for detecting the Ceph version can be started with node affinity or tolerations according to the same settings in the Cluster CR as the mons.
  • A new CR property skipUpgradeChecks has been added, which allows you force an upgrade by skipping daemon checks. Use this at YOUR OWN RISK, only if you know what you're doing. To understand Rook's upgrade process of Ceph, read the upgrade doc.
  • Mon Quorum Disaster Recovery guide has been updated to work with the latest Rook and Ceph release.
  • A new CRD property PreservePoolsOnDelete has been added to Filesystem(fs) and Object Store(os) resources in order to increase protection against data loss. if it is set to true, associated pools won't be deleted when the main resource(fs/os) is deleted. Creating again the deleted fs/os with the same name will reuse the preserved pools.
  • A new ceph-crashcollector controller has been added. These new deployments will run on any node where a Ceph pod is running. Read more about this in the doc
  • PriorityClassNames can now be added to the Rook/Ceph components to influence the scheduler's pod preemption.
  • When the operator is upgraded, the mgr and OSDs (not running on PVC) won't be restarted if the Rook binary version changes
  • Rook is now able to create and manage Ceph clients client crd.
  • OSDs:
    • Rook will no longer automatically remove OSDs if nodes are removed from the cluster CR to avoid the risk of destroying OSDs unintentionally. To remove OSDs manually, see the new doc on OSD Management
    • Ceph OSD's admin socket is now placed under Ceph's default system location /run/ceph.
    • The on-host log directory for OSDs was set incorrectly to <dataDirHostPath>/<namespace>/log; fix this to be <dataDirHostPath>/log/<namespace>, the same as other daemons.
    • Do not generate a config (during pod init) for directory-based or legacy filestore OSDs
    • Add a new CRD property devicePathFilter to support device filtering with path names, e.g. /dev/disk/by-path/pci-.*-sas-.*.
    • Support PersistentVolume backed by LVM Logical Volume for "OSD on PVC".
    • Creation of new Filestore OSDs on disks is now deprecated. Filestore is in sustaining mode in Ceph.
      • The storeType storage config setting is now ignored
        • New OSDs created in directories are always Filestore type
        • New OSDs created on disks are always Bluestore type
      • Preexisting disks provisioned as Filestore OSDs will remain as Filestore OSDs
  • RGWs:
    • Ceph Object Gateway are automatically configured to not run on the same host if hostNetwork is activated

EdgeFS

  • Rook EdgeFS operator adds support for single node, single device deployments. This is to enable embedded and remote developer use cases.
  • Support for new EdgeFS backend, rtkvs, enables ability to operate on top of any key-value capable interface. Initial integration adds support for Samsung KV-SSD devices.
  • Enhanced support for running EdgeFS in the AWS cloud. It is now possible to store data payload chunks directly in AWS S3 buckets, thus greatly reducing storage billing cost. Metadata chunks still will be in AWS EBS, thus provide low-latency and high-performance.
  • It is now possible to configure ISGW Full-Mesh functionality without the need to create multiple ISGW services. Please read more about ISGW Full-Mesh functionality here.
  • EdgeFS now capable of creating instant snapshots of S3 buckets. It supports billion of objects per-bucket use cases. A snapshot's metadata gets distributed among all the connected EdgeFS segments, such that cloning or accessing of snapshotted objects can be done without the need of full-delta transferring, i.e. on-demand.

YugabyteDB

Breaking Changes

Ceph

  • The topology setting has been removed from the CephCluster CR. To configure the OSD topology, node labels must be applied. See the OSD topology topic. This setting only affects OSDs when they are first created, thus OSDs will not be impacted during upgrade. The topology settings only apply to bluestore OSDs on raw devices. The topology labels are not applied to directory-based OSDs.

Known Issues

Deprecations