Re: [Ceph-maintainers] v16.2.0 Pacific released

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

This is awesome news! =).

I did hear mention before about Crimson and Pacific - does anybody know
what the current state of things is?

I see there's a doc page for it here -
https://docs.ceph.com/en/latest/dev/crimson/crimson/

Are we able to use Crimson yet in Pacific? (As in, do we need to rebuild
from source, or is it available in the packages just announced?) Are we
likely to see any improvements with U.2/NVMe drives yet?

Regards,
Victor

On Fri, Apr 2, 2021 at 7:07 AM Martin Verges <martin.verges@xxxxxxxx> wrote:

> Hello,
>
> thanks for a very interesting new Ceph Release.
>
> Are there any plans to build for Debian bullseye as well? It's in
> "Hard Freeze" since 2021-03-12 and at the moment it comes with a
> Nautilus release that will be EOL when Debian bullseye will be
> official stable. That will be a pain for Debian users and if it's
> still possible we should try to avoid that. Is there something we
> could help to make it happen?
>
> --
> Martin Verges
> Managing director
>
> Mobile: +49 174 9335695 <+49%20174%209335695>
> E-Mail: martin.verges@xxxxxxxx
> Chat: https://t.me/MartinVerges
>
> croit GmbH, Freseniusstr. 31h, 81247 Munich
> CEO: Martin Verges - VAT-ID: DE310638492
> Com. register: Amtsgericht Munich HRB 231263
>
> Web: https://croit.io
> YouTube: https://goo.gl/PGE1Bx
>
> On Thu, 1 Apr 2021 at 16:27, David Galloway <dgallowa@xxxxxxxxxx> wrote:
> >
> > We're glad to announce the first release of the Pacific v16.2.0 stable
> > series. There have been a lot of changes across components from the
> > previous Ceph releases, and we advise everyone to go through the release
> > and upgrade notes carefully.
> >
> > Major Changes from Octopus
> > --------------------------
> >
> > General
> > ~~~~~~~
> >
> > * Cephadm can automatically upgrade an Octopus cluster to Pacific with a
> single
> >   command to start the process.
> >
> > * Cephadm has improved significantly over the past year, with improved
> >   support for RGW (standalone and multisite), and new support for NFS
> >   and iSCSI.  Most of these changes have already been backported to
> >   recent Octopus point releases, but with the Pacific release we will
> >   switch to backporting bug fixes only.
> >
> > * Packages are built for the following distributions:
> >
> >   - CentOS 8
> >   - Ubuntu 20.04 (Focal)
> >   - Ubuntu 18.04 (Bionic)
> >   - Debian Buster
> >   - Container image (based on CentOS 8)
> >
> >   With the exception of Debian Buster, packages and containers are
> >   built for both x86_64 and aarch64 (arm64) architectures.
> >
> >   Note that cephadm clusters may work on many other distributions,
> >   provided Python 3 and a recent version of Docker or Podman is
> >   available to manage containers.  For more information, see
> >   `cephadm-host-requirements`.
> >
> >
> > Dashboard
> > ~~~~~~~~~
> >
> > The `mgr-dashboard` brings improvements in the following management
> areas:
> >
> > * Orchestrator/Cephadm:
> >
> >   - Host management: maintenance mode, labels.
> >   - Services: display placement specification.
> >   - OSD: disk replacement, display status of ongoing deletion, and
> improved
> >     health/SMART diagnostics reporting.
> >
> > * Official `mgr ceph api`:
> >
> >   - OpenAPI v3 compliant.
> >   - Stability commitment starting from Pacific release.
> >   - Versioned via HTTP `Accept` header (starting with v1.0).
> >   - Thoroughly tested (>90% coverage and per Pull Request validation).
> >   - Fully documented.
> >
> > * RGW:
> >
> >   - Multi-site synchronization monitoring.
> >   - Management of multiple RGW daemons and their resources (buckets and
> users).
> >   - Bucket and user quota usage visualization.
> >   - Improved configuration of S3 tenanted users.
> >
> > * Security (multiple enhancements and fixes resulting from a pen testing
> conducted by IBM):
> >
> >   - Account lock-out after a configurable number of failed log-in
> attempts.
> >   - Improved cookie policies to mitigate XSS/CSRF attacks.
> >   - Reviewed and improved security in HTTP headers.
> >   - Sensitive information reviewed and removed from logs and error
> messages.
> >   - TLS 1.0 and 1.1 support disabled.
> >   - Debug mode when enabled triggers HEALTH_WARN.
> >
> > * Pools:
> >
> >   - Improved visualization of replication and erasure coding modes.
> >   - CLAY erasure code plugin supported.
> >
> > * Alerts and notifications:
> >
> >   - Alert triggered on MTU mismatches in the cluster network.
> >   - Favicon changes according cluster status.
> >
> > * Other:
> >
> >   - Landing page: improved charts and visualization.
> >   - Telemetry configuration wizard.
> >   - OSDs: management of individual OSD flags.
> >   - RBD: per-RBD image Grafana dashboards.
> >   - CephFS: Dirs and Caps displayed.
> >   - NFS: v4 support only (v3 backward compatibility planned).
> >   - Front-end: Angular 10 update.
> >
> >
> > RADOS
> > ~~~~~
> >
> > * Pacific introduces RocksDB sharding, which reduces disk space
> requirements.
> >
> > * Ceph now provides QoS between client I/O and background operations via
> the
> >   mclock scheduler.
> >
> > * The balancer is now on by default in upmap mode to improve
> distribution of
> >   PGs across OSDs.
> >
> > * The output of `ceph -s` has been improved to show recovery progress in
> >   one progress bar. More detailed progress bars are visible via the
> >   `ceph progress` command.
> >
> >
> > RBD block storage
> > ~~~~~~~~~~~~~~~~~
> >
> > * Image live-migration feature has been extended to support external data
> >   sources.  Images can now be instantly imported from local files, remote
> >   files served over HTTP(S) or remote S3 buckets in `raw` (`rbd export
> v1`)
> >   or basic `qcow` and `qcow2` formats.  Support for `rbd export v2`
> >   format, advanced QCOW features and `rbd export-diff` snapshot
> differentials
> >   is expected in future releases.
> >
> > * Initial support for client-side encryption has been added.  This is
> based
> >   on LUKS and in future releases will allow using per-image encryption
> keys
> >   while maintaining snapshot and clone functionality -- so that parent
> image
> >   and potentially multiple clone images can be encrypted with different
> keys.
> >
> > * A new persistent write-back cache is available.  The cache operates in
> >   a log-structured manner, providing full point-in-time consistency for
> the
> >   backing image.  It should be particularly suitable for PMEM devices.
> >
> > * A Windows client is now available in the form of `librbd.dll` and
> >   `rbd-wnbd` (Windows Network Block Device) daemon.  It allows mapping,
> >   unmapping and manipulating images similar to `rbd-nbd`.
> >
> > * librbd API now offers quiesce/unquiesce hooks, allowing for coordinated
> >   snapshot creation.
> >
> >
> > RGW object storage
> > ~~~~~~~~~~~~~~~~~~
> >
> > * Initial support for S3 Select. See `s3-select-feature-table` for
> supported queries.
> >
> > * Bucket notification topics can be configured as `persistent`, where
> events
> >   are recorded in rados for reliable delivery.
> >
> > * Bucket notifications can be delivered to SSL-enabled AMQP endpoints.
> >
> > * Lua scripts can be run during requests and access their metadata.
> >
> > * SSE-KMS now supports KMIP as a key management service.
> >
> > * Multisite data logs can now be deployed on `cls_fifo` to avoid large
> omap
> >   cluster warnings and make their trimming cheaper. See
> `rgw_data_log_backing`.
> >
> >
> > CephFS distributed file system
> > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> >
> > * The CephFS MDS modifies on-RADOS metadata such that the new format is
> no
> >   longer backwards compatible. It is not possible to downgrade a file
> system from
> >   Pacific (or later) to an older release.
> >
> > * Multiple file systems in a single Ceph cluster is now stable. New Ceph
> >   clusters enable support for multiple file systems by default. Existing
> clusters
> >   must still set the "enable_multiple" flag on the FS. See also
> >   `cephfs-multifs`.
> >
> > * A new `mds_autoscaler` `ceph-mgr` plugin is available for automatically
> >   deploying MDS daemons in response to changes to the `max_mds`
> configuration.
> >   Expect further enhancements in the future to simplify and automate MDS
> scaling.
> >
> > * `cephfs-top` is a new utility for looking at performance metrics from
> CephFS
> >   clients. It is development preview quality and will have bugs. For more
> >   information, see `cephfs-top`.
> >
> > * A new `snap_schedule` `ceph-mgr` plugin provides a command toolset for
> >   scheduling snapshots on a CephFS file system. For more information, see
> >   `snap-schedule`.
> >
> > * First class NFS gateway support in Ceph is here! It's now possible to
> create
> >   scale-out ("active-active") NFS gateway clusters that export CephFS
> using
> >   a few commands. The gateways are deployed via cephadm (or Rook, in the
> future).
> >   For more information, see `cephfs-nfs`.
> >
> > * Multiple active MDS file system scrub is now stable. It is no longer
> necessary
> >   to set `max_mds` to 1 and wait for non-zero ranks to stop. Scrub
> commands
> >   can only be sent to rank 0: `ceph tell mds.<fs_name>:0 scrub start
> /path ...`.
> >   For more information, see `mds-scrub`.
> >
> > * Ephemeral pinning -- policy based subtree pinning -- is considered
> stable.
> >   `mds_export_ephemeral_random` and `mds_export_ephemeral_distributed`
> now
> >   default to true. For more information, see `cephfs-ephemeral-pinning`.
> >
> > * A new `cephfs-mirror` daemon is available to mirror CephFS file
> systems to
> >   a remote Ceph cluster.
> >
> > * A Windows client is now available for connecting to CephFS. This is
> offered
> >   through a new `ceph-dokan` utility which operates via the Dokan
> userspace
> >   API, similar to FUSE. For more information, see `ceph-dokan`.
> >
> >
> > Upgrading from Octopus or Nautilus
> > ----------------------------------
> >
> > Before starting, make sure your cluster is stable and healthy (no down or
> > recovering OSDs).  (This is optional, but recommended.)
> >
> > Upgrading cephadm clusters
> > ~~~~~~~~~~~~~~~~~~~~~~~~~~
> >
> > If your cluster is deployed with cephadm (first introduced in Octopus),
> then
> > the upgrade process is entirely automated.  To initiate the upgrade,
> >
> >     ceph orch upgrade start --ceph-version 16.2.0
> >
> > The same process is used to upgrade to future minor releases.
> >
> > Upgrade progress can be monitored with `ceph -s` (which provides a simple
> > progress bar) or more verbosely with
> >
> >     ceph -W cephadm
> >
> > The upgrade can be paused or resumed with
> >
> >     ceph orch upgrade pause   # to pause
> >     ceph orch upgrade resume  # to resume
> >
> > or canceled with
> >
> >     ceph orch upgrade stop
> >
> > Note that canceling the upgrade simply stops the process; there is no
> ability to
> > downgrade back to Octopus.
> >
> >
> > Upgrading non-cephadm clusters
> > ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> >
> >    If you cluster is running Octopus (15.2.x), you might choose
> >    to first convert it to use cephadm so that the upgrade to Pacific
> >    is automated (see above).  For more information, see
> >    `cephadm-adoption`.
> >
> > #. Set the `noout` flag for the duration of the upgrade. (Optional,
> >    but recommended.):
> >
> >      # ceph osd set noout
> >
> > #. Upgrade monitors by installing the new packages and restarting the
> >    monitor daemons.  For example, on each monitor host,:
> >
> >      # systemctl restart ceph-mon.target
> >
> >    Once all monitors are up, verify that the monitor upgrade is
> >    complete by looking for the `octopus` string in the mon
> >    map.  The command:
> >
> >      # ceph mon dump | grep min_mon_release
> >
> >    should report:
> >
> >      min_mon_release 16 (pacific)
> >
> >    If it doesn't, that implies that one or more monitors hasn't been
> >    upgraded and restarted and/or the quorum does not include all
> monitors.
> >
> > #. Upgrade `ceph-mgr` daemons by installing the new packages and
> >    restarting all manager daemons.  For example, on each manager host,:
> >
> >      # systemctl restart ceph-mgr.target
> >
> >    Verify the `ceph-mgr` daemons are running by checking `ceph
> >    -s`
> >
> >      # ceph -s
> >
> >      ...
> >        services:
> >         mon: 3 daemons, quorum foo,bar,baz
> >         mgr: foo(active), standbys: bar, baz
> >      ...
> >
> > #. Upgrade all OSDs by installing the new packages and restarting the
> >    ceph-osd daemons on all OSD hosts:
> >
> >      # systemctl restart ceph-osd.target
> >
> >    Note that if you are upgrading from Nautilus, the first time each
> >    OSD starts, it will do a format conversion to improve the
> >    accounting for "omap" data.  This may take a few minutes to as much
> >    as a few hours (for an HDD with lots of omap data).  You can
> >    disable this automatic conversion with:
> >
> >      # ceph config set osd bluestore_fsck_quick_fix_on_mount false
> >
> >    You can monitor the progress of the OSD upgrades with the
> >    `ceph versions` or `ceph osd versions` commands:
> >
> >      # ceph osd versions
> >      {
> >         "ceph version 14.2.5 (...) nautilus (stable)": 12,
> >         "ceph version 16.2.0 (...) pacific (stable)": 22,
> >      }
> >
> > #. Upgrade all CephFS MDS daemons. For each CephFS file system,
> >
> >    #. Disable standby_replay:
> >
> >    # ceph fs set <fs_name> allow_standby_replay false
> >
> >    #. Reduce the number of ranks to 1.  (Make note of the original
> >       number of MDS daemons first if you plan to restore it later.):
> >
> >         # ceph status
> >         # ceph fs set <fs_name> max_mds 1
> >
> >    #. Wait for the cluster to deactivate any non-zero ranks by
> >       periodically checking the status:
> >
> >         # ceph status
> >
> >    #. Take all standby MDS daemons offline on the appropriate hosts with:
> >
> >         # systemctl stop ceph-mds@<daemon_name>
> >
> >    #. Confirm that only one MDS is online and is rank 0 for your FS:
> >
> >         # ceph status
> >
> >    #. Upgrade the last remaining MDS daemon by installing the new
> >       packages and restarting the daemon:
> >
> >         # systemctl restart ceph-mds.target
> >
> >    #. Restart all standby MDS daemons that were taken offline:
> >
> >         # systemctl start ceph-mds.target
> >
> >    #. Restore the original value of `max_mds` for the volume:
> >
> >         # ceph fs set <fs_name> max_mds <original_max_mds>
> >
> > #. Upgrade all radosgw daemons by upgrading packages and restarting
> >    daemons on all hosts:
> >
> >      # systemctl restart ceph-radosgw.target
> >
> > #. Complete the upgrade by disallowing pre-Pacific OSDs and enabling
> >    all new Pacific-only functionality:
> >
> >      # ceph osd require-osd-release pacific
> >
> > #. If you set `noout` at the beginning, be sure to clear it with:
> >
> >      # ceph osd unset noout
> >
> > #. Consider transitioning your cluster to use the cephadm deployment
> >    and orchestration framework to simplify cluster management and
> >    future upgrades.  For more information on converting an existing
> >    cluster to cephadm, see `cephadm-adoption`.
> >
> >
> > Post-upgrade
> > ~~~~~~~~~~~~
> >
> > #. Verify the cluster is healthy with `ceph health`.
> >
> >    If your CRUSH tunables are older than Hammer, Ceph will now issue a
> >    health warning.  If you see a health alert to that effect, you can
> >    revert this change with:
> >
> >      ceph config set mon mon_crush_min_required_version firefly
> >
> >    If Ceph does not complain, however, then we recommend you also
> >    switch any existing CRUSH buckets to straw2, which was added back
> >    in the Hammer release.  If you have any 'straw' buckets, this will
> >    result in a modest amount of data movement, but generally nothing
> >    too severe.:
> >
> >      ceph osd getcrushmap -o backup-crushmap
> >      ceph osd crush set-all-straw-buckets-to-straw2
> >
> >    If there are problems, you can easily revert with:
> >
> >      ceph osd setcrushmap -i backup-crushmap
> >
> >    Moving to 'straw2' buckets will unlock a few recent features, like
> >    the `crush-compat` `balancer <balancer>` mode added back in Luminous.
> >
> > #. If you did not already do so when upgrading from Mimic, we
> >    recommened you enable the new `v2 network protocol <msgr2>`,
> >    issue the following command:
> >
> >      ceph mon enable-msgr2
> >
> >    This will instruct all monitors that bind to the old default port
> >    6789 for the legacy v1 protocol to also bind to the new 3300 v2
> >    protocol port.  To see if all monitors have been updated,:
> >
> >      ceph mon dump
> >
> >    and verify that each monitor has both a `v2:` and `v1:` address
> >    listed.
> >
> > #. Consider enabling the `telemetry module <telemetry>` to send
> >    anonymized usage statistics and crash information to the Ceph
> >    upstream developers.  To see what would be reported (without actually
> >    sending any information to anyone),:
> >
> >      ceph mgr module enable telemetry
> >      ceph telemetry show
> >
> >    If you are comfortable with the data that is reported, you can opt-in
> to
> >    automatically report the high-level cluster metadata with:
> >
> >      ceph telemetry on
> >
> >    The public dashboard that aggregates Ceph telemetry can be found at
> >    https://telemetry-public.ceph.com.
> >
> >    For more information about the telemetry module, see `the
> >    telemetry documentation.
> >
> >
> > Upgrade from pre-Nautilus releases (like Mimic or Luminous)
> > -----------------------------------------------------------
> >
> > You must first upgrade to Nautilus (14.2.z) or Octopus (15.2.z) before
> > upgrading to Pacific.
> >
> >
> > Notable Changes
> > ---------------
> >
> > * A new library is available, libcephsqlite. It provides a SQLite
> Virtual File
> >   System (VFS) on top of RADOS. The database and journals are striped
> over
> >   RADOS across multiple objects for virtually unlimited scaling and
> throughput
> >   only limited by the SQLite client. Applications using SQLite may
> change to
> >   the Ceph VFS with minimal changes, usually just by specifying the
> alternate
> >   VFS. We expect the library to be most impactful and useful for
> applications
> >   that were storing state in RADOS omap, especially without striping
> which
> >   limits scalability.
> >
> > * New `bluestore_rocksdb_options_annex` config parameter. Complements
> >   `bluestore_rocksdb_options` and allows setting rocksdb options without
> >   repeating the existing defaults.
> >
> > * $pid expansion in config paths like `admin_socket` will now properly
> expand
> >   to the daemon pid for commands like `ceph-mds` or `ceph-osd`.
> Previously
> >   only `ceph-fuse`/`rbd-nbd` expanded `$pid` with the actual daemon pid.
> >
> > * The allowable options for some `radosgw-admin` commands have been
> changed.
> >
> >   * `mdlog-list`, `datalog-list`, `sync-error-list` no longer accepts
> >     start and end dates, but does accept a single optional start marker.
> >   * `mdlog-trim`, `datalog-trim`, `sync-error-trim` only accept a
> >     single marker giving the end of the trimmed range.
> >   * Similarly the date ranges and marker ranges have been removed on
> >     the RESTful DATALog and MDLog list and trim operations.
> >
> > * ceph-volume: The `lvm batch` subcommand received a major rewrite. This
> >   closed a number of bugs and improves usability in terms of size
> specification
> >   and calculation, as well as idempotency behaviour and disk replacement
> >   process.
> >   Please refer to https://docs.ceph.com/en/latest/ceph-volume/lvm/batch/
> for
> >   more detailed information.
> >
> > * Configuration variables for permitted scrub times have changed.  The
> legal
> >   values for `osd_scrub_begin_hour` and `osd_scrub_end_hour` are 0 - 23.
> >   The use of 24 is now illegal.  Specifying `0` for both values causes
> every
> >   hour to be allowed.  The legal values for `osd_scrub_begin_week_day`
> and
> >   `osd_scrub_end_week_day` are 0 - 6.  The use of 7 is now illegal.
> >   Specifying `0` for both values causes every day of the week to be
> allowed.
> >
> > * volume/nfs: Recently "ganesha-" prefix from cluster id and nfs-ganesha
> common
> >   config object was removed, to ensure consistent namespace across
> different
> >   orchestrator backends. Please delete any existing nfs-ganesha clusters
> prior
> >   to upgrading and redeploy new clusters after upgrading to Pacific.
> >
> > * A new health check, DAEMON_OLD_VERSION, will warn if different
> versions of Ceph are running
> >   on daemons. It will generate a health error if multiple versions are
> detected.
> >   This condition must exist for over mon_warn_older_version_delay (set
> to 1 week by default) in order for the
> >   health condition to be triggered.  This allows most upgrades to proceed
> >   without falsely seeing the warning.  If upgrade is paused for an
> extended
> >   time period, health mute can be used like this
> >   "ceph health mute DAEMON_OLD_VERSION --sticky".  In this case after
> >   upgrade has finished use "ceph health unmute DAEMON_OLD_VERSION".
> >
> > * MGR: progress module can now be turned on/off, using the commands:
> >   `ceph progress on` and `ceph progress off`.
> >
> > * An AWS-compliant API: "GetTopicAttributes" was added to replace the
> existing "GetTopic" API. The new API
> >   should be used to fetch information about topics used for bucket
> notifications.
> >
> > * librbd: The shared, read-only parent cache's config option
> `immutable_object_cache_watermark` now has been updated
> >   to property reflect the upper cache utilization before space is
> reclaimed. The default `immutable_object_cache_watermark`
> >   now is `0.9`. If the capacity reaches 90% the daemon will delete cold
> cache.
> >
> > * OSD: the option `osd_fast_shutdown_notify_mon` has been introduced to
> allow
> >   the OSD to notify the monitor it is shutting down even if
> `osd_fast_shutdown`
> >   is enabled. This helps with the monitor logs on larger clusters, that
> may get
> >   many 'osd.X reported immediately failed by osd.Y' messages, and
> confuse tools.
> >
> > * The mclock scheduler has been refined. A set of built-in profiles are
> now available that
> >   provide QoS between the internal and external clients of Ceph. To
> enable the mclock
> >   scheduler, set the config option "osd_op_queue" to "mclock_scheduler".
> The
> >   "high_client_ops" profile is enabled by default, and allocates more
> OSD bandwidth to
> >   external client operations than to internal client operations (such as
> background recovery
> >   and scrubs). Other built-in profiles include "high_recovery_ops" and
> "balanced". These
> >   built-in profiles optimize the QoS provided to clients of mclock
> scheduler.
> >
> > * The balancer is now on by default in upmap mode. Since upmap mode
> requires
> >   `require_min_compat_client` luminous, new clusters will only support
> luminous
> >   and newer clients by default. Existing clusters can enable upmap
> support by running
> >   `ceph osd set-require-min-compat-client luminous`. It is still
> possible to turn
> >   the balancer off using the `ceph balancer off` command. In earlier
> versions,
> >   the balancer was included in the `always_on_modules` list, but needed
> to be
> >   turned on explicitly using the `ceph balancer on` command.
> >
> > * Version 2 of the cephx authentication protocol (`CEPHX_V2` feature
> bit) is
> >   now required by default.  It was introduced in 2018, adding replay
> attack
> >   protection for authorizers and making msgr v1 message signatures
> stronger
> >   (CVE-2018-1128 and CVE-2018-1129).  Support is present in Jewel
> 10.2.11,
> >   Luminous 12.2.6, Mimic 13.2.1, Nautilus 14.2.0 and later; upstream
> kernels
> >   4.9.150, 4.14.86, 4.19 and later; various distribution kernels, in
> particular
> >   CentOS 7.6 and later.  To enable older clients, set
> `cephx_require_version`
> >   and `cephx_service_require_version` config options to 1.
> >
> > * `blacklist` has been replaced with `blocklist` throughout.  The
> following commands have changed:
> >
> >   - `ceph osd blacklist ...` are now `ceph osd blocklist ...`
> >   - `ceph <tell|daemon> osd.<NNN> dump_blacklist` is now `ceph
> <tell|daemon> osd.<NNN> dump_blocklist`
> >
> > * The following config options have changed:
> >
> >   - `mon osd blacklist default expire` is now `mon osd blocklist default
> expire`
> >   - `mon mds blacklist interval` is now `mon mds blocklist interval`
> >   - `mon mgr blacklist interval` is now ''mon mgr blocklist interval`
> >   - `rbd blacklist on break lock` is now `rbd blocklist on break lock`
> >   - `rbd blacklist expire seconds` is now `rbd blocklist expire seconds`
> >   - `mds session blacklist on timeout` is now `mds session blocklist on
> timeout`
> >   - `mds session blacklist on evict` is now `mds session blocklist on
> evict`
> >
> > * The following librados API calls have changed:
> >
> >   - `rados_blacklist_add` is now `rados_blocklist_add`; the former will
> issue a deprecation warning and be removed in a future release.
> >   - `rados.blacklist_add` is now `rados.blocklist_add` in the C++ API.
> >
> > * The JSON output for the following commands now shows `blocklist`
> instead of `blacklist`:
> >
> >   - `ceph osd dump`
> >   - `ceph <tell|daemon> osd.<N> dump_blocklist`
> >
> > * Monitors now have config option `mon_allow_pool_size_one`, which is
> disabled
> >   by default. However, if enabled, user now have to pass the
> >   `--yes-i-really-mean-it` flag to `osd pool set size 1`, if they are
> really
> >   sure of configuring pool size 1.
> >
> > * `ceph pg #.# list_unfound` output has been enhanced to provide
> >   might_have_unfound information which indicates which OSDs may
> >   contain the unfound objects.
> >
> > * OSD: A new configuration option `osd_compact_on_start` has been added
> which triggers
> >   an OSD compaction on start. Setting this option to `true` and
> restarting an OSD
> >   will result in an offline compaction of the OSD prior to booting.
> >
> > * OSD: the option named `bdev_nvme_retry_count` has been removed. Because
> >   in SPDK v20.07, there is no easy access to bdev_nvme options, and this
> >   option is hardly used, so it was removed.
> >
> > * Alpine build related script, documentation and test have been removed
> since
> >   the most updated APKBUILD script of Ceph is already included by Alpine
> Linux's
> >   aports repository.
> >
> >
> >
> > Getting Ceph
> > ------------
> > * Git at git://github.com/ceph/ceph.git
> > * Tarball at http://download.ceph.com/tarballs/ceph-16.2.0.tar.gz
> > * For packages, see
> http://docs.ceph.com/docs/master/install/get-packages/
> > * Release git sha1: 0c2054e95bcd9b30fdd908a79ac1d8bbc3394442
> > _______________________________________________
> > Ceph-maintainers mailing list -- ceph-maintainers@xxxxxxx
> > To unsubscribe send an email to ceph-maintainers-leave@xxxxxxx
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux