On Tue, Jan 11, 2022 at 5:29 AM Dan van der Ster <dan@xxxxxxxxxxxxxx> wrote: > > Hi, > > Yes it's confusing -- the release notes are normally only published in > master, which is shown as "latest", and are rarely backported to a > release branch. > The notes you're looking for are here: > https://docs.ceph.com/en/latest/releases/pacific/#v16-2-7-pacific > > Zac is in cc -- maybe we can make this more friendly? Yeah, we need to push minor-version release notes to their relevant branch so users see it in the docs that are relevant to their version. -Greg > > -- dan > > On Tue, Jan 11, 2022 at 2:11 PM Frank Schilder <frans@xxxxxx> wrote: > > > > Hi Dan, > > > > I seem to have a problem finding correct release notes and upgrade instructions on the ceph docs. If I open https://docs.ceph.com/en/pacific/ and go to "Ceph Releases", pacific is not even listed even though these doc pages are supposedly about pacific. I also can't find the "upgrading from nautilus or octopus" pages anywhere. If I go to v:latest, I can see such information. Is there a link error that, when selecting version pacific shows the pages for octopus? This would be very confusing, because the link "https://docs.ceph.com" forwards to "https://docs.ceph.com/en/pacific/" which would show information on octopus. > > > > Best regards, > > ================= > > Frank Schilder > > AIT Risø Campus > > Bygning 109, rum S14 > > > > ________________________________________ > > From: Dan van der Ster <dan@xxxxxxxxxxxxxx> > > Sent: 09 December 2021 17:58:22 > > To: David Galloway > > Cc: ceph-announce@xxxxxxx; Ceph Users; Ceph Developers; ceph-maintainers@xxxxxxx; Patrick Donnelly > > Subject: Re: v16.2.7 Pacific released > > > > Hi all, > > > > The release notes are missing an upgrade step that is needed only for > > clusters *not* managed by cephadm. > > This was noticed in > > https://lists.ceph.io/hyperkitty/list/ceph-users@xxxxxxx/thread/7KSPSUE4VO274H5XQYNFCT7HKWT75BCY/ > > > > If you are not using cephadm, you must disable FSMap sanity checks > > *before starting the upgrade*: > > > > ceph config set mon mon_mds_skip_sanity 1 > > > > After the upgrade is finished and the cluster is stable, please remove > > that setting: > > > > ceph config rm mon mon_mds_skip_sanity > > > > Clusters upgraded by cephadm take care of this step automatically. > > > > Best Regards, > > > > Dan > > > > > > > > On Wed, Dec 8, 2021 at 1:12 AM David Galloway <dgallowa@xxxxxxxxxx> wrote: > > > > > > We're happy to announce the 7th backport release in the Pacific series. > > > We recommend all users upgrade to this release. > > > > > > Notable Changes > > > --------------- > > > > > > * Critical bug in OMAP format upgrade is fixed. This could cause data > > > corruption (improperly formatted OMAP keys) after pre-Pacific cluster > > > upgrade if bluestore-quick-fix-on-mount parameter is set to true or > > > ceph-bluestore-tool's quick-fix/repair commands are invoked. Relevant > > > tracker: https://tracker.ceph.com/issues/53062. > > > bluestore-quick-fix-on-mount continues to be set to false, by default. > > > > > > * MGR: The pg_autoscaler will use the 'scale-up' profile as the default > > > profile. 16.2.6 changed the default profile to 'scale-down' but we ran > > > into issues with the device_health_metrics pool consuming too many PGs, > > > which is not ideal for performance. So we will continue to use the > > > 'scale-up' profile by default, until we implement a limit on the number > > > of PGs default pools should consume, in combination with the > > > 'scale-down' profile. > > > > > > * Cephadm & Ceph Dashboard: NFS management has been completely reworked > > > to ensure that NFS exports are managed consistently across the different > > > Ceph components. Prior to this, there were 3 incompatible > > > implementations for configuring the NFS exports: Ceph-Ansible/OpenStack > > > Manila, Ceph Dashboard and 'mgr/nfs' module. With this release the > > > 'mgr/nfs' way becomes the official interface, and the remaining > > > components (Cephadm and Ceph Dashboard) adhere to it. While this might > > > require manually migrating from the deprecated implementations, it will > > > simplify the user experience for those heavily relying on NFS exports. > > > > > > * Dashboard: "Cluster Expansion Wizard". After the 'cephadm bootstrap' > > > step, users that log into the Ceph Dashboard will be presented with a > > > welcome screen. If they choose to follow the installation wizard, they > > > will be guided through a set of steps to help them configure their Ceph > > > cluster: expanding the cluster by adding more hosts, detecting and > > > defining their storage devices, and finally deploying and configuring > > > the different Ceph services. > > > > > > * OSD: When using mclock_scheduler for QoS, there is no longer a need to > > > run any manual benchmark. The OSD now automatically sets an appropriate > > > value for osd_mclock_max_capacity_iops by running a simple benchmark > > > during initialization. > > > > > > * MGR: The global recovery event in the progress module has been > > > optimized and a sleep_interval of 5 seconds has been added between stats > > > collection, to reduce the impact of the progress module on the MGR, > > > especially in large clusters. > > > > > > > > > Getting Ceph > > > ------------ > > > * Git at git://github.com/ceph/ceph.git > > > * Tarball at https://download.ceph.com/tarballs/ceph-16.2.7.tar.gz > > > * Containers at https://quay.io/repository/ceph/ceph > > > * For packages, see https://docs.ceph.com/docs/master/install/get-packages/ > > > * Release git sha1: dd0603118f56ab514f133c8d2e3adfc983942503 > > > > > > _______________________________________________ > > > ceph-users mailing list -- ceph-users@xxxxxxx > > > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > _______________________________________________ > > ceph-users mailing list -- ceph-users@xxxxxxx > > To unsubscribe send an email to ceph-users-leave@xxxxxxx > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx