Re: v15.2.4 Octopus released

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



True. That said, the blog post points to http://download.ceph.com/tarballs/ where all the tarballs, including 15.2.4, live.

 On 6/30/2020 5:57 PM, Sasha Litvak wrote:
David,

Download link points to 14.2.10 tarball.

On Tue, Jun 30, 2020, 3:38 PM David Galloway <dgallowa@xxxxxxxxxx> wrote:

We're happy to announce the fourth bugfix release in the Octopus series.
In addition to a security fix in RGW, this release brings a range of fixes
across all components. We recommend that all Octopus users upgrade to this
release. For a detailed release notes with links & changelog please
refer to the official blog entry at
https://ceph.io/releases/v15-2-4-octopus-released

Notable Changes
---------------
* CVE-2020-10753: rgw: sanitize newlines in s3 CORSConfiguration's
ExposeHeader
   (William Bowling, Adam Mohammed, Casey Bodley)

* Cephadm: There were a lot of small usability improvements and bug fixes:
   * Grafana when deployed by Cephadm now binds to all network interfaces.
   * `cephadm check-host` now prints all detected problems at once.
   * Cephadm now calls `ceph dashboard set-grafana-api-ssl-verify false`
     when generating an SSL certificate for Grafana.
   * The Alertmanager is now correctly pointed to the Ceph Dashboard
   * `cephadm adopt` now supports adopting an Alertmanager
   * `ceph orch ps` now supports filtering by service name
   * `ceph orch host ls` now marks hosts as offline, if they are not
     accessible.

* Cephadm can now deploy NFS Ganesha services. For example, to deploy NFS
with
   a service id of mynfs, that will use the RADOS pool nfs-ganesha and
namespace
   nfs-ns::

     ceph orch apply nfs mynfs nfs-ganesha nfs-ns

* Cephadm: `ceph orch ls --export` now returns all service specifications
in
   yaml representation that is consumable by `ceph orch apply`. In addition,
   the commands `orch ps` and `orch ls` now support `--format yaml` and
   `--format json-pretty`.

* Cephadm: `ceph orch apply osd` supports a `--preview` flag that prints a
preview of
   the OSD specification before deploying OSDs. This makes it possible to
   verify that the specification is correct, before applying it.

* RGW: The `radosgw-admin` sub-commands dealing with orphans --
   `radosgw-admin orphans find`, `radosgw-admin orphans finish`, and
   `radosgw-admin orphans list-jobs` -- have been deprecated. They have
   not been actively maintained and they store intermediate results on
   the cluster, which could fill a nearly-full cluster.  They have been
   replaced by a tool, currently considered experimental,
   `rgw-orphan-list`.

* RBD: The name of the rbd pool object that is used to store
   rbd trash purge schedule is changed from "rbd_trash_trash_purge_schedule"
   to "rbd_trash_purge_schedule". Users that have already started using
   `rbd trash purge schedule` functionality and have per pool or namespace
   schedules configured should copy "rbd_trash_trash_purge_schedule"
   object to "rbd_trash_purge_schedule" before the upgrade and remove
   "rbd_trash_purge_schedule" using the following commands in every RBD
   pool and namespace where a trash purge schedule was previously
   configured::

     rados -p <pool-name> [-N namespace] cp rbd_trash_trash_purge_schedule
rbd_trash_purge_schedule
     rados -p <pool-name> [-N namespace] rm rbd_trash_trash_purge_schedule

   or use any other convenient way to restore the schedule after the
   upgrade.

Getting Ceph
------------
* Git at git://github.com/ceph/ceph.git
* Tarball at http://download.ceph.com/tarballs/ceph-14.2.10.tar.gz
* For packages, see http://docs.ceph.com/docs/master/install/get-packages/
* Release git sha1: 7447c15c6ff58d7fce91843b705a268a1917325c

--
David Galloway
Systems Administrator, RDU
Ceph Engineering
IRC: dgalloway
_______________________________________________
Dev mailing list -- dev@xxxxxxx
To unsubscribe send an email to dev-leave@xxxxxxx

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux