We're happy to announce the 1st backport release in the Squid series. https://ceph.io/en/news/blog/2025/v19-2-1-squid-released/ Notable Changes --------------- * CephFS: The command `fs subvolume create` now allows tagging subvolumes by supplying the option `--earmark` with a unique identifier needed for NFS or SMB services. The earmark string for a subvolume is empty by default. To remove an already present earmark, an empty string can be assigned to it. Additionally, the commands `ceph fs subvolume earmark set`, `ceph fs subvolume earmark get`, and `ceph fs subvolume earmark rm` have been added to set, get and remove earmark from a given subvolume. * CephFS: Expanded removexattr support for CephFS virtual extended attributes. Previously one had to use setxattr to restore the default in order to "remove". You may now properly use removexattr to remove. You can also now remove layout on the root inode, which then will restore the layout to the default. * RADOS: A performance bottleneck in the balancer mgr module has been fixed. Related Tracker: https://tracker.ceph.com/issues/68657 * RADOS: Based on tests performed at scale on an HDD-based Ceph cluster, it was found that scheduling with mClock was not optimal with multiple OSD shards. For example, in the test cluster with multiple OSD node failures, the client throughput was found to be inconsistent across test runs coupled with multiple reported slow requests. However, the same test with a single OSD shard and with multiple worker threads yielded significantly better results in terms of consistency of client and recovery throughput across multiple test runs. Therefore, as an interim measure until the issue with multiple OSD shards (or multiple mClock queues per OSD) is investigated and fixed, the following change to the default HDD OSD shard configuration is made: - `osd_op_num_shards_hdd = 1` (was 5) - `osd_op_num_threads_per_shard_hdd = 5` (was 1) For more details, see https://tracker.ceph.com/issues/66289. * mgr/REST: The REST manager module will trim requests based on the 'max_requests' option. Without this feature, and in the absence of manual deletion of old requests, the accumulation of requests in the array can lead to Out Of Memory (OOM) issues, resulting in the Manager crashing. Getting Ceph ------------ * Git at git://github.com/ceph/ceph.git * Tarball at https://download.ceph.com/tarballs/ceph-19.2.1.tar.gz * Containers at https://quay.io/repository/ceph/ceph * For packages, see https://docs.ceph.com/en/latest/install/get-packages/ * Release git sha1: 58a7fab8be0a062d730ad7da874972fd3fba59fb _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx