Re: v14.2.8 Nautilus released

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



No didn't get answer to this. 

Yes I thought also, but recently there has been an issue here with an 
upgrade to Octopus, where osd's are being changed automatically and 
consume huge amounts of memory during this. Furthermore if you have a 
cluster with hundreds of osds, it is not really acceptable to have to 
recreate them.


 

-----Original Message-----
From: Brent Kennedy [mailto:bkennedy@xxxxxxxxxx] 
Sent: 05 April 2020 04:26
To: Marc Roos; 'abhishek'; 'ceph-users'
Subject: RE:  Re: v14.2.8 Nautilus released

Did you get an answer for this?  My original thought when I read it was 
that the osd would need to be recreated(as you noted).

-Brent

-----Original Message-----
From: Marc Roos <M.Roos@xxxxxxxxxxxxxxxxx>
Sent: Tuesday, March 3, 2020 10:58 AM
To: abhishek <abhishek@xxxxxxxx>; ceph-users <ceph-users@xxxxxxx>
Subject:  Re: v14.2.8 Nautilus released

 
This bluestore_min_alloc_size_ssd=4K, do I need to recreate these osd's? 

Or does this magically change? What % performance increase can be 
expected?


-----Original Message-----
To: ceph-announce@xxxxxxx; ceph-users@xxxxxxx; dev@xxxxxxx; 
ceph-devel@xxxxxxxxxxxxxxx
Subject:  v14.2.8 Nautilus released


This is the eighth update to the Ceph Nautilus release series. This 
release fixes issues across a range of subsystems. We recommend that all 

users upgrade to this release. Please note the following important 
changes in this release; as always the full changelog is posted at:
https://ceph.io/releases/v14-2-8-nautilus-released

Notable Changes
---------------

* The default value of `bluestore_min_alloc_size_ssd` has been changed
  to 4K to improve performance across all workloads.

* The following OSD memory config options related to bluestore cache 
autotuning can now
  be configured during runtime:

    - osd_memory_base (default: 768 MB)
    - osd_memory_cache_min (default: 128 MB)
    - osd_memory_expected_fragmentation (default: 0.15)
    - osd_memory_target (default: 4 GB)

  The above options can be set with::

    ceph config set osd <option> <value>

* The MGR now accepts `profile rbd` and `profile rbd-read-only` user 
caps.
  These caps can be used to provide users access to MGR-based RBD 
functionality
  such as `rbd perf image iostat` an `rbd perf image iotop`.

* The configuration value `osd_calc_pg_upmaps_max_stddev` used for upmap
  balancing has been removed. Instead use the mgr balancer config
  `upmap_max_deviation` which now is an integer number of PGs of 
deviation
  from the target PGs per OSD.  This can be set with a command like
  `ceph config set mgr mgr/balancer/upmap_max_deviation 2`.  The default
  `upmap_max_deviation` is 1.  There are situations where crush rules
  would not allow a pool to ever have completely balanced PGs.  For 
example, if
  crush requires 1 replica on each of 3 racks, but there are fewer OSDs 
in 1 of
  the racks.  In those cases, the configuration value can be increased.

* RGW: a mismatch between the bucket notification documentation and the 
actual
  message format was fixed. This means that any endpoints receiving 
bucket
  notification, will now receive the same notifications inside a JSON 
array
  named 'Records'. Note that this does not affect pulling bucket 
notification
  from a subscription in a 'pubsub' zone, as these are already wrapped 
inside
  that array.

* CephFS: multiple active MDS forward scrub is now rejected. Scrub 
currently
  only is permitted on a file system with a single rank. Reduce the 
ranks to one
  via `ceph fs set <fs_name> max_mds 1`.

* Ceph now refuses to create a file system with a default EC data pool. 
For
  further explanation, see:
  https://docs.ceph.com/docs/nautilus/cephfs/createfs/#creating-pools

* Ceph will now issue a health warning if a RADOS pool has a `pg_num`
  value that is not a power of two. This can be fixed by adjusting
  the pool to a nearby power of two::

    ceph osd pool set <pool-name> pg_num <new-pg-num>

  Alternatively, the warning can be silenced with::

    ceph config set global mon_warn_on_pool_pg_num_not_power_of_two 
false

Getting Ceph
------------

* Git at git://github.com/ceph/ceph.git
* Tarball at http://download.ceph.com/tarballs/ceph-14.2.8.tar.gz
* For packages, see 
http://docs.ceph.com/docs/master/install/get-packages/
* Release git sha1: 2d095e947a02261ce61424021bb43bd3022d35cb

--
Abhishek Lekshmanan
SUSE Software Solutions Germany GmbH
GF: Felix Imendörffer HRB 21284 (AG Nürnberg) 
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an 
email to ceph-users-leave@xxxxxxx

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx


_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux