Re: [ceph-users] v15.2.0 Octopus released

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I modified our sync-push script a little so the symlinks are relative
instead of absolute.  (https://github.com/ceph/ceph-build/pull/1534)

In addition to that change, if you add `--copy-links` to your rsync
command, it syncs as expected now.

On 3/24/20 1:48 PM, David Galloway wrote:
> I have tried on-and-off all morning to get the magical combination of
> rsync flags to get this working and I don't think it's possible.
>
> The plan is, for each version released, have that version go in its
> respective /rpm-X.X.X directory.  Once all the packages are pushed, the
> /rpm-octopus symlink is then pointed to the new version rpm-X.X.X dir.
>
> @Sage what are your thoughts on rsyncd currently being broken due to
> this new directory structure?  It has also broken our mirror syncing
> procedure.  e.g., http://au.ceph.com/ doesn't have {deb,rpm}-octopus dirs.
>
> On 3/24/20 1:32 PM, Eric Goirand wrote:
>> Thanks Abhishek for the rpm-15.2.0 link.
>> @David are there any plans to have rsync still work with rpm-octopus as
>> it is the case for http access ?
>> If you plan to move to rpm-15.2.0, would it still be the same directory
>> when you go to 15.2.1 or 15.2.x as it will be difficult to change the
>> DNF or YUM URLs to cope with these changes.
>> Thanks and Regards,
>> Eric.
>>
>> On Tue, Mar 24, 2020 at 3:10 PM Abhishek Lekshmanan <abhishek@xxxxxxxx
>> <mailto:abhishek@xxxxxxxx>> wrote:
>>
>>     Eric Goirand <egoirand@xxxxxxxxxx <mailto:egoirand@xxxxxxxxxx>> writes:
>>
>>     > Hello Abhishek,
>>     Hi Eric
>>     >
>>     > Many Many Thanks to all the contributors for this awesome milestone !
>>     > Two thumbs up to you all !
>>
>>     The directory structure changed (buried somewhere in the release
>>     announcment) mainly so that we can make it easy to retire a bad release
>>     etc. so rpm-octopus should be a symlink to
>>     https://download.ceph.com/rpm-15.2.0/
>>
>>     I've CC'ed David who's in charge of the infra.
>>
>>
>>     > Note however that the Octopus repository via rsync is no longer
>>     accessible :
>>     > rsync -aiv --dry-run rsync://download.ceph.com/ceph/rpm-octopus/
>>     <http://download.ceph.com/ceph/rpm-octopus/> .
>>     > receiving incremental file list
>>     > rsync: change_dir "/rpm-octopus" (in ceph) failed: No such file or
>>     > directory (2)
>>     >
>>     > But, is still accessible using http :
>>     http://download.ceph.com/rpm-octopus/
>>     >
>>     > Would you know who to contact to have the rsync access work again ?
>>     >
>>     > Many Thanks,
>>     >
>>     > Eric.
>>     >
>>     >
>>     > On Tue, Mar 24, 2020 at 12:41 PM Abhishek Lekshmanan
>>     <abhishek@xxxxxxxx <mailto:abhishek@xxxxxxxx>>
>>     > wrote:
>>     >
>>     >>
>>     >> We're happy to announce the first stable release of Octopus v15.2.0.
>>     >> There are a lot of changes and new features added, we advise
>>     everyone to
>>     >> read the release notes carefully, and in particular the upgrade
>>     notes,
>>     >> before upgrading. Please refer to the official blog entry
>>     >> https://ceph.io/releases/v15-2-0-octopus-released/ for a detailed
>>     >> version with links & changelog.
>>     >>
>>     >> This release wouldn't have been possible without the support of the
>>     >> community, this release saw contributions from over 330
>>     developers & 80
>>     >> organizations, and we thank everyone for making this release happen.
>>     >>
>>     >> Major Changes from Nautilus
>>     >> ---------------------------
>>     >> General
>>     >> ~~~~~~~
>>     >> * A new deployment tool called **cephadm** has been introduced that
>>     >>   integrates Ceph daemon deployment and management via containers
>>     >>   into the orchestration layer.
>>     >> * Health alerts can now be muted, either temporarily or permanently.
>>     >> * Health alerts are now raised for recent Ceph daemons crashes.
>>     >> * A simple 'alerts' module has been introduced to send email
>>     >>   health alerts for clusters deployed without the benefit of an
>>     >>   existing external monitoring infrastructure.
>>     >> * Packages are built for the following distributions:
>>     >>   - CentOS 8
>>     >>   - CentOS 7 (partial--see below)
>>     >>   - Ubuntu 18.04 (Bionic)
>>     >>   - Debian Buster
>>     >>   - Container images (based on CentOS 8)
>>     >>
>>     >>   Note that the dashboard, prometheus, and restful manager modules
>>     >>   will not work on the CentOS 7 build due to Python 3 module
>>     >>   dependencies that are missing in CentOS 7.
>>     >>
>>     >>   Besides this packages built by the community will also
>>     available for the
>>     >>   following distros:
>>     >>   - Fedora (33/rawhide)
>>     >>   - openSUSE (15.2, Tumbleweed)
>>     >>
>>     >> Dashboard
>>     >> ~~~~~~~~~
>>     >> The mgr-dashboard has gained a lot of new features and functionality:
>>     >>
>>     >> * UI Enhancements
>>     >>   - New vertical navigation bar
>>     >>   - New unified sidebar: better background task and events
>>     notification
>>     >>   - Shows all progress mgr module notifications
>>     >>   - Multi-select on tables to perform bulk operations
>>     >>
>>     >> * Dashboard user account security enhancements
>>     >>   - Disabling/enabling existing user accounts
>>     >>   - Clone an existing user role
>>     >>   - Users can change their own password
>>     >>   - Configurable password policies: Minimum password
>>     complexity/length
>>     >>     requirements
>>     >>   - Configurable password expiration
>>     >>   - Change password after first login
>>     >>
>>     >> New and enhanced management of Ceph features/services:
>>     >>
>>     >> * OSD/device management
>>     >>   - List all disks associated with an OSD
>>     >>   - Add support for blinking enclosure LEDs via the orchestrator
>>     >>   - List all hosts known by the orchestrator
>>     >>   - List all disks and their properties attached to a node
>>     >>   - Display disk health information (health prediction and SMART
>>     data)
>>     >>   - Deploy new OSDs on new disks/hosts
>>     >>   - Display and allow sorting by an OSD's default device class in
>>     the OSD
>>     >>     table
>>     >>   - Explicitly set/change the device class of an OSD, display and
>>     sort
>>     >> OSDs by
>>     >>     device class
>>     >>
>>     >> * Pool management
>>     >>   - Viewing and setting pool quotas
>>     >>   - Define and change per-pool PG autoscaling mode
>>     >>
>>     >> * RGW management enhancements
>>     >>   - Enable bucket versioning
>>     >>   - Enable MFA support
>>     >>   - Select placement target on bucket creation
>>     >>
>>     >> * CephFS management enhancements
>>     >>   - CephFS client eviction
>>     >>   - CephFS snapshot management
>>     >>   - CephFS quota management
>>     >>   - Browse CephFS directory
>>     >>
>>     >> * iSCSI management enhancements
>>     >>   - Show iSCSI GW status on landing page
>>     >>   - Prevent deletion of IQNs with open sessions
>>     >>   - Display iSCSI "logged in" info
>>     >>
>>     >> * Prometheus alert management
>>     >>   - List configured Prometheus alerts
>>     >>
>>     >> RADOS
>>     >> ~~~~~
>>     >> * Objects can now be brought in sync during recovery by copying only
>>     >>   the modified portion of the object, reducing tail latencies during
>>     >>   recovery.
>>     >> * Ceph will allow recovery below *min_size* for Erasure coded pools,
>>     >>   wherever possible.
>>     >> * The PG autoscaler feature introduced in Nautilus is enabled for
>>     >>   new pools by default, allowing new clusters to autotune *pg num*
>>     >>   without any user intervention.  The default values for new pools
>>     >>   and RGW/CephFS metadata pools have also been adjusted to perform
>>     >>   well for most users.
>>     >> * BlueStore has received several improvements and performance
>>     >>   updates, including improved accounting for "omap" (key/value)
>>     >>   object data by pool, improved cache memory management, and a
>>     >>   reduced allocation unit size for SSD devices.  (Note that by
>>     >>   default, the first time each OSD starts after upgrading to octopus
>>     >>   it will trigger a conversion that may take from a few minutes to a
>>     >>   few hours, depending on the amount of stored "omap" data.)
>>     >> * Snapshot trimming metadata is now managed in a more efficient and
>>     >>   scalable fashion.
>>     >>
>>     >> RBD block storage
>>     >> ~~~~~~~~~~~~~~~~~
>>     >> * Mirroring now supports a new snapshot-based mode that no longer
>>     requires
>>     >>   the journaling feature and its related impacts in exchange for
>>     the loss
>>     >>   of point-in-time consistency (it remains crash consistent).
>>     >> * Clone operations now preserve the sparseness of the underlying
>>     RBD image.
>>     >> * The trash feature has been improved to (optionally) automatically
>>     >>   move old parent images to the trash when their children are all
>>     >>   deleted or flattened.
>>     >> * The trash can be configured to automatically purge on a defined
>>     schedule.
>>     >> * Images can be online re-sparsified to reduce the usage of
>>     zeroed extents.
>>     >> * The `rbd-nbd` tool has been improved to use more modern kernel
>>     >> interfaces.
>>     >> * Caching has been improved to be more efficient and performant.
>>     >> * `rbd-mirror` automatically adjusts its per-image memory usage based
>>     >>   upon its memory target.
>>     >> * A new persistent read-only caching daemon is available to
>>     offload reads
>>     >> from
>>     >>   shared parent images.
>>     >>
>>     >> RGW object storage
>>     >> ~~~~~~~~~~~~~~~~~~
>>     >> * New `Multisite Sync Policy` primitives for per-bucket replication.
>>     >> (EXPERIMENTAL)
>>     >> * S3 feature support:
>>     >>     - Bucket Replication (EXPERIMENTAL)
>>     >>     - `Bucket Notifications`_ via HTTP/S, AMQP and Kafka
>>     >>     - Bucket Tagging
>>     >>     - Object Lock
>>     >>     - Public Access Block for buckets
>>     >> * Bucket sharding:
>>     >>     - Significantly improved listing performance on buckets with many
>>     >> shards.
>>     >>     - Dynamic resharding prefers prime shard counts for improved
>>     >> distribution.
>>     >>     - Raised the default number of bucket shards to 11.
>>     >> * Added `HashiCorp Vault Integration`_ for SSE-KMS.
>>     >> * Added Keystone token cache for S3 requests.
>>     >>
>>     >> CephFS distributed file system
>>     >> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>>     >> * Inline data support in CephFS has been deprecated and will
>>     likely be
>>     >>   removed in a future release.
>>     >> * MDS daemons can now be assigned to manage a particular file
>>     system via
>>     >> the
>>     >>   new `mds_join_fs` option.
>>     >> * MDS now aggressively asks idle clients to trim caps which improves
>>     >> stability
>>     >>   when file system load changes.
>>     >> * The mgr volumes plugin has received numerous improvements to
>>     support
>>     >> CephFS
>>     >>   via CSI, including snapshots and cloning.
>>     >> * cephfs-shell has had numerous incremental improvements and bug
>>     fixes.
>>     >>
>>     >>
>>     >> Upgrading from Mimic or Nautilus
>>     >> --------------------------------
>>     >>   You can monitor the progress of your upgrade at each stage with the
>>     >>   `ceph versions` command, which will tell you what ceph
>>     version(s) are
>>     >>   running for each type of daemon.
>>     >>
>>     >> Instructions
>>     >> ~~~~~~~~~~~~
>>     >> #. Make sure your cluster is stable and healthy (no down or
>>     >>    recovering OSDs).  (Optional, but recommended.)
>>     >>
>>     >> #. Set the `noout` flag for the duration of the upgrade. (Optional,
>>     >>    but recommended.)::
>>     >>
>>     >>      # ceph osd set noout
>>     >>
>>     >> #. Upgrade monitors by installing the new packages and restarting the
>>     >>    monitor daemons.  For example, on each monitor host,::
>>     >>
>>     >>      # systemctl restart ceph-mon.target
>>     >>
>>     >>    Once all monitors are up, verify that the monitor upgrade is
>>     >>    complete by looking for the `octopus` string in the mon
>>     >>    map.  The command::
>>     >>
>>     >>      # ceph mon dump | grep min_mon_release
>>     >>
>>     >>    should report::
>>     >>
>>     >>      min_mon_release 15 (nautilus)
>>     >>
>>     >>    If it doesn't, that implies that one or more monitors hasn't been
>>     >>    upgraded and restarted and/or the quorum does not include all
>>     monitors.
>>     >>
>>     >> #. Upgrade `ceph-mgr` daemons by installing the new packages and
>>     >>    restarting all manager daemons.  For example, on each manager
>>     host,::
>>     >>
>>     >>      # systemctl restart ceph-mgr.target
>>     >>
>>     >>    Verify the `ceph-mgr` daemons are running by checking `ceph
>>     >>    -s`::
>>     >>
>>     >>      # ceph -s
>>     >>
>>     >>      ...
>>     >>        services:
>>     >>         mon: 3 daemons, quorum foo,bar,baz
>>     >>         mgr: foo(active), standbys: bar, baz
>>     >>      ...
>>     >>
>>     >> #. Upgrade all OSDs by installing the new packages and restarting the
>>     >>    ceph-osd daemons on all OSD hosts::
>>     >>
>>     >>      # systemctl restart ceph-osd.target
>>     >>
>>     >>    Note that the first time each OSD starts, it will do a format
>>     >>    conversion to improve the accounting for "omap" data.  This may
>>     >>    take a few minutes to as much as a few hours (for an HDD with lots
>>     >>    of omap data).  You can disable this automatic conversion with::
>>     >>
>>     >>      # ceph config set osd bluestore_fsck_quick_fix_on_mount false
>>     >>
>>     >>    You can monitor the progress of the OSD upgrades with the
>>     >>    `ceph versions` or `ceph osd versions` commands::
>>     >>
>>     >>      # ceph osd versions
>>     >>      {
>>     >>         "ceph version 13.2.5 (...) mimic (stable)": 12,
>>     >>         "ceph version 15.2.0 (...) octopus (stable)": 22,
>>     >>      }
>>     >>
>>     >> #. Upgrade all CephFS MDS daemons.  For each CephFS file system,
>>     >>
>>     >>    #. Reduce the number of ranks to 1.  (Make note of the original
>>     >>       number of MDS daemons first if you plan to restore it later.)::
>>     >>
>>     >>         # ceph status
>>     >>         # ceph fs set <fs_name> max_mds 1
>>     >>
>>     >>    #. Wait for the cluster to deactivate any non-zero ranks by
>>     >>       periodically checking the status::
>>     >>
>>     >>         # ceph status
>>     >>
>>     >>    #. Take all standby MDS daemons offline on the appropriate
>>     hosts with::
>>     >>
>>     >>         # systemctl stop ceph-mds@<daemon_name>
>>     >>
>>     >>    #. Confirm that only one MDS is online and is rank 0 for your FS::
>>     >>
>>     >>         # ceph status
>>     >>
>>     >>    #. Upgrade the last remaining MDS daemon by installing the new
>>     >>       packages and restarting the daemon::
>>     >>
>>     >>         # systemctl restart ceph-mds.target
>>     >>
>>     >>    #. Restart all standby MDS daemons that were taken offline::
>>     >>
>>     >>         # systemctl start ceph-mds.target
>>     >>
>>     >>    #. Restore the original value of `max_mds` for the volume::
>>     >>
>>     >>         # ceph fs set <fs_name> max_mds <original_max_mds>
>>     >>
>>     >> #. Upgrade all radosgw daemons by upgrading packages and restarting
>>     >>    daemons on all hosts::
>>     >>
>>     >>      # systemctl restart ceph-radosgw.target
>>     >>
>>     >> #. Complete the upgrade by disallowing pre-Octopus OSDs and enabling
>>     >>    all new Octopus-only functionality::
>>     >>
>>     >>      # ceph osd require-osd-release octopus
>>     >>
>>     >> #. If you set `noout` at the beginning, be sure to clear it with::
>>     >>
>>     >>      # ceph osd unset noout
>>     >>
>>     >> #. Verify the cluster is healthy with `ceph health`.
>>     >>
>>     >>    If your CRUSH tunables are older than Hammer, Ceph will now
>>     issue a
>>     >>    health warning.  If you see a health alert to that effect, you can
>>     >>    revert this change with::
>>     >>
>>     >>      ceph config set mon mon_crush_min_required_version firefly
>>     >>
>>     >>    If Ceph does not complain, however, then we recommend you also
>>     >>    switch any existing CRUSH buckets to straw2, which was added back
>>     >>    in the Hammer release.  If you have any 'straw' buckets, this will
>>     >>    result in a modest amount of data movement, but generally nothing
>>     >>    too severe.::
>>     >>
>>     >>      ceph osd getcrushmap -o backup-crushmap
>>     >>      ceph osd crush set-all-straw-buckets-to-straw2
>>     >>
>>     >>    If there are problems, you can easily revert with::
>>     >>
>>     >>      ceph osd setcrushmap -i backup-crushmap
>>     >>
>>     >>    Moving to 'straw2' buckets will unlock a few recent features, like
>>     >>    the `crush-compat` :ref:`balancer <balancer>` mode added back in
>>     >> Luminous.
>>     >>
>>     >> #. If you are upgrading from Mimic, or did not already do so when you
>>     >>    upgraded to Nautlius, we recommened you enable the new :ref:`v2
>>     >>    network protocol <msgr2>`, issue the following command::
>>     >>
>>     >>      ceph mon enable-msgr2
>>     >>
>>     >>    This will instruct all monitors that bind to the old default port
>>     >>    6789 for the legacy v1 protocol to also bind to the new 3300 v2
>>     >>    protocol port.  To see if all monitors have been updated,::
>>     >>
>>     >>      ceph mon dump
>>     >>
>>     >>    and verify that each monitor has both a `v2:` and `v1:` address
>>     >>    listed.
>>     >>
>>     >> #. Consider enabling the :ref:`telemetry module <telemetry>` to send
>>     >>    anonymized usage statistics and crash information to the Ceph
>>     >>    upstream developers.  To see what would be reported (without
>>     actually
>>     >>    sending any information to anyone),::
>>     >>
>>     >>      ceph mgr module enable telemetry
>>     >>      ceph telemetry show
>>     >>
>>     >>    If you are comfortable with the data that is reported, you can
>>     opt-in to
>>     >>    automatically report the high-level cluster metadata with::
>>     >>
>>     >>      ceph telemetry on
>>     >>
>>     >>    For more information about the telemetry module, see :ref:`the
>>     >>    documentation <telemetry>`.
>>     >>
>>     >>
>>     >> Upgrading from pre-Mimic releases (like Luminous)
>>     >> -------------------------------------------------
>>     >> You *must* first upgrade to Mimic (13.2.z) or Nautilus (14.2.z)
>>     before
>>     >> upgrading to Octopus.
>>     >>
>>     >>
>>     >> Upgrade compatibility notes
>>     >> ---------------------------
>>     >> * Starting with Octopus, there is now a separate repository directory
>>     >>   for each version on `download.ceph.com
>>     <http://download.ceph.com>` (e.g., `rpm-15.2.0` and
>>     >>   `debian-15.2.0`).  The traditional package directory that is named
>>     >>   after the release (e.g., `rpm-octopus` and `debian-octopus`) is
>>     >>   now a symlink to the most recently bug fix version for that
>>     release.
>>     >>   We no longer generate a single repository that combines all bug fix
>>     >>   versions for a single named release.
>>     >>
>>     >> * The RGW "num_rados_handles" has been removed.
>>     >>   If you were using a value of "num_rados_handles" greater than 1
>>     >>   multiply your current "objecter_inflight_ops" and
>>     >>   "objecter_inflight_op_bytes" paramaeters by the old
>>     >>   "num_rados_handles" to get the same throttle behavior.
>>     >>
>>     >> * Ceph now packages python bindings for python3.6 instead of
>>     >>   python3.4, because python3 in EL7/EL8 is now using python3.6
>>     >>   as the native python3. see the `announcement`_
>>     >>   for more details on the background of this change.
>>     >>
>>     >> * librbd now uses a write-around cache policy be default,
>>     >>   replacing the previous write-back cache policy default.
>>     >>   This cache policy allows librbd to immediately complete
>>     >>   write IOs while they are still in-flight to the OSDs.
>>     >>   Subsequent flush requests will ensure all in-flight
>>     >>   write IOs are completed prior to completing. The
>>     >>   librbd cache policy can be controlled via a new
>>     >>   "rbd_cache_policy" configuration option.
>>     >>
>>     >> * librbd now includes a simple IO scheduler which attempts to
>>     >>   batch together multiple IOs against the same backing RBD
>>     >>   data block object. The librbd IO scheduler policy can be
>>     >>   controlled via a new "rbd_io_scheduler" configuration
>>     >>   option.
>>     >>
>>     >> * RGW: radosgw-admin introduces two subcommands that allow the
>>     >>   managing of expire-stale objects that might be left behind after a
>>     >>   bucket reshard in earlier versions of RGW. One subcommand lists
>>     such
>>     >>   objects and the other deletes them. Read the troubleshooting
>>     section
>>     >>   of the dynamic resharding docs for details.
>>     >>
>>     >> * RGW: Bucket naming restrictions have changed and likely to cause
>>     >>   InvalidBucketName errors. We recommend to set
>>     >> `rgw_relaxed_s3_bucket_names`
>>     >>   option to true as a workaround.
>>     >>
>>     >> * In the Zabbix Mgr Module there was a typo in the key being send
>>     >>   to Zabbix for PGs in backfill_wait state. The key that was sent
>>     >>   was 'wait_backfill' and the correct name is 'backfill_wait'.
>>     >>   Update your Zabbix template accordingly so that it accepts the
>>     >>   new key being send to Zabbix.
>>     >>
>>     >> * zabbix plugin for ceph manager now includes osd and pool
>>     >>   discovery. Update of zabbix_template.xml is needed
>>     >>   to receive per-pool (read/write throughput, diskspace usage)
>>     >>   and per-osd (latency, status, pgs) statistics
>>     >>
>>     >> * The format of all date + time stamps has been modified to fully
>>     >>   conform to ISO 8601.  The old format (`YYYY-MM-DD
>>     >>   HH:MM:SS.ssssss`) excluded the `T` separator between the date and
>>     >>   time and was rendered using the local time zone without any
>>     explicit
>>     >>   indication.  The new format includes the separator as well as a
>>     >>   `+nnnn` or `-nnnn` suffix to indicate the time zone, or a `Z`
>>     >>   suffix if the time is UTC.  For example,
>>     >>   `2019-04-26T18:40:06.225953+0100`.
>>     >>
>>     >>   Any code or scripts that was previously parsing date and/or time
>>     >>   values from the JSON or XML structure CLI output should be checked
>>     >>   to ensure it can handle ISO 8601 conformant values.  Any code
>>     >>   parsing date or time values from the unstructured human-readable
>>     >>   output should be modified to parse the structured output
>>     instead, as
>>     >>   the human-readable output may change without notice.
>>     >>
>>     >> * The `bluestore_no_per_pool_stats_tolerance` config option has been
>>     >>   replaced with `bluestore_fsck_error_on_no_per_pool_stats`
>>     >>   (default: false).  The overall default behavior has not changed:
>>     >>   fsck will warn but not fail on legacy stores, and repair will
>>     >>   convert to per-pool stats.
>>     >>
>>     >> * The disaster-recovery related 'ceph mon sync force' command has
>>     been
>>     >>   replaced with 'ceph daemon <...> sync_force'.
>>     >>
>>     >> * The `osd_recovery_max_active` option now has
>>     >>   `osd_recovery_max_active_hdd` and `osd_recovery_max_active_ssd`
>>     >>   variants, each with different default values for HDD and SSD-backed
>>     >>   OSDs, respectively.  By default `osd_recovery_max_active` now
>>     >>   defaults to zero, which means that the OSD will conditionally use
>>     >>   the HDD or SSD option values.  Administrators who have customized
>>     >>   this value may want to consider whether they have set this to a
>>     >>   value similar to the new defaults (3 for HDDs and 10 for SSDs) and,
>>     >>   if so, remove the option from their configuration entirely.
>>     >>
>>     >> * monitors now have a `ceph osd info` command that will provide
>>     information
>>     >>   on all osds, or provided osds, thus simplifying the process of
>>     having to
>>     >>   parse `osd dump` for the same information.
>>     >>
>>     >> * The structured output of `ceph status` or `ceph -s` is now more
>>     >>   concise, particularly the `mgrmap` and `monmap` sections, and the
>>     >>   structure of the `osdmap` section has been cleaned up.
>>     >>
>>     >> * A health warning is now generated if the average osd heartbeat ping
>>     >>   time exceeds a configurable threshold for any of the intervals
>>     >>   computed.  The OSD computes 1 minute, 5 minute and 15 minute
>>     >>   intervals with average, minimum and maximum values.  New
>>     >>   configuration option `mon_warn_on_slow_ping_ratio` specifies a
>>     >>   percentage of `osd_heartbeat_grace` to determine the threshold.  A
>>     >>   value of zero disables the warning.  New configuration option
>>     >>   `mon_warn_on_slow_ping_time` specified in milliseconds over-rides
>>     >>   the computed value, causes a warning when OSD heartbeat pings take
>>     >>   longer than the specified amount.  New admin command `ceph daemon
>>     >>   mgr.# dump_osd_network [threshold]` command will list all
>>     >>   connections with a ping time longer than the specified threshold or
>>     >>   value determined by the config options, for the average for any of
>>     >>   the 3 intervals.  New admin command `ceph daemon osd.#
>>     >>   dump_osd_network [threshold]` will do the same but only including
>>     >>   heartbeats initiated by the specified OSD.
>>     >>
>>     >> * Inline data support for CephFS has been deprecated. When
>>     setting the
>>     >> flag,
>>     >>   users will see a warning to that effect, and enabling it now
>>     requires the
>>     >>   `--yes-i-really-really-mean-it` flag. If the MDS is started on a
>>     >>   filesystem that has it enabled, a health warning is generated.
>>     Support
>>     >> for
>>     >>   this feature will be removed in a future release.
>>     >>
>>     >> * `ceph {set,unset} full` is not supported anymore. We have been
>>     using
>>     >>   `full` and `nearfull` flags in OSD map for tracking the
>>     fullness status
>>     >>   of a cluster back since the Hammer release, if the OSD map is
>>     marked
>>     >> `full`
>>     >>   all write operations will be blocked until this flag is
>>     removed. In the
>>     >>   Infernalis release and Linux kernel 4.7 client, we introduced the
>>     >> per-pool
>>     >>   full/nearfull flags to track the status for a finer-grained
>>     control, so
>>     >> the
>>     >>   clients will hold the write operations if either the
>>     cluster-wide `full`
>>     >>   flag or the per-pool `full` flag is set. This was a compromise,
>>     as we
>>     >>   needed to support the cluster with and without per-pool `full`
>>     flags
>>     >>   support. But this practically defeated the purpose of
>>     introducing the
>>     >>   per-pool flags. So, in the Mimic release, the new flags finally
>>     took the
>>     >>   place of their cluster-wide counterparts, as the monitor
>>     started removing
>>     >>   these two flags from OSD map. So the clients of Infernalis and
>>     up can
>>     >> benefit
>>     >>   from this change, as they won't be blocked by the full pools
>>     which they
>>     >> are
>>     >>   not writing to. In this release, `ceph {set,unset} full` is now
>>     >> considered
>>     >>   as an invalid command. And the clients will continue honoring
>>     both the
>>     >>   cluster-wide and per-pool flags to be backward comaptible with
>>     >> pre-infernalis
>>     >>   clusters.
>>     >>
>>     >> * The telemetry module now reports more information.
>>     >>
>>     >>   First, there is a new 'device' channel, enabled by default, that
>>     >>   will report anonymized hard disk and SSD health metrics to
>>     >>   telemetry.ceph.com <http://telemetry.ceph.com> in order to
>>     build and improve device failure
>>     >>   prediction algorithms.  If you are not comfortable sharing device
>>     >>   metrics, you can disable that channel first before re-opting-in::
>>     >>
>>     >>     ceph config set mgr mgr/telemetry/channel_device false
>>     >>
>>     >>   Second, we now report more information about CephFS file systems,
>>     >>   including:
>>     >>
>>     >>   - how many MDS daemons (in total and per file system)
>>     >>   - which features are (or have been) enabled
>>     >>   - how many data pools
>>     >>   - approximate file system age (year + month of creation)
>>     >>   - how many files, bytes, and snapshots
>>     >>   - how much metadata is being cached
>>     >>
>>     >>   We have also added:
>>     >>
>>     >>   - which Ceph release the monitors are running
>>     >>   - whether msgr v1 or v2 addresses are used for the monitors
>>     >>   - whether IPv4 or IPv6 addresses are used for the monitors
>>     >>   - whether RADOS cache tiering is enabled (and which mode)
>>     >>   - whether pools are replicated or erasure coded, and
>>     >>     which erasure code profile plugin and parameters are in use
>>     >>   - how many hosts are in the cluster, and how many hosts have
>>     each type
>>     >> of daemon
>>     >>   - whether a separate OSD cluster network is being used
>>     >>   - how many RBD pools and images are in the cluster, and how
>>     many pools
>>     >> have RBD mirroring enabled
>>     >>   - how many RGW daemons, zones, and zonegroups are present;
>>     which RGW
>>     >> frontends are in use
>>     >>   - aggregate stats about the CRUSH map, like which algorithms
>>     are used,
>>     >> how
>>     >>     big buckets are, how many rules are defined, and what
>>     tunables are in
>>     >>     use
>>     >>
>>     >>   If you had telemetry enabled, you will need to re-opt-in with::
>>     >>
>>     >>     ceph telemetry on
>>     >>
>>     >>   You can view exactly what information will be reported first with::
>>     >>
>>     >>     $ ceph telemetry show        # see everything
>>     >>     $ ceph telemetry show basic  # basic cluster info (including
>>     all of
>>     >> the new info)
>>     >>
>>     >> * Following invalid settings now are not tolerated anymore
>>     >>   for the command `ceph osd erasure-code-profile set xxx`.
>>     >>   * invalid `m` for "reed_sol_r6_op" erasure technique
>>     >>   * invalid `m` and invalid `w` for "liber8tion" erasure technique
>>     >>
>>     >> * New OSD daemon command dump_recovery_reservations which reveals the
>>     >>   recovery locks held (in_progress) and waiting in priority queues.
>>     >>
>>     >> * New OSD daemon command dump_scrub_reservations which reveals the
>>     >>   scrub reservations that are held for local (primary) and remote
>>     >> (replica) PGs.
>>     >>
>>     >> * Previously, `ceph tell mgr ...` could be used to call commands
>>     >>   implemented by mgr modules.  This is no longer supported.  Since
>>     >>   luminous, using `tell` has not been necessary: those same commands
>>     >>   are also accessible without the `tell mgr` portion (e.g., `ceph
>>     >>   tell mgr influx foo` is the same as `ceph influx foo`.  `ceph
>>     >>   tell mgr ...` will now call admin commands--the same set of
>>     >>   commands accessible via `ceph daemon ...` when you are logged into
>>     >>   the appropriate host.
>>     >>
>>     >> * The `ceph tell` and `ceph daemon` commands have been unified,
>>     >>   such that all such commands are accessible via either interface.
>>     >>   Note that ceph-mgr tell commands are accessible via either `ceph
>>     >>   tell mgr ...` or `ceph tell mgr.<id> ...`, and it is only
>>     >>   possible to send tell commands to the active daemon (the
>>     standbys do
>>     >>   not accept incoming connections over the network).
>>     >>
>>     >> * Ceph will now issue a health warning if a RADOS pool as a `pg_num`
>>     >>   value that is not a power of two.  This can be fixed by adjusting
>>     >>   the pool to a nearby power of two::
>>     >>
>>     >>     ceph osd pool set <pool-name> pg_num <new-pg-num>
>>     >>
>>     >>   Alternatively, the warning can be silenced with::
>>     >>
>>     >>     ceph config set global
>>     mon_warn_on_pool_pg_num_not_power_of_two false
>>     >>
>>     >> * The format of MDSs in `ceph fs dump` has changed.
>>     >>
>>     >> * The `mds_cache_size` config option is completely removed. Since
>>     luminous,
>>     >>   the `mds_cache_memory_limit` config option has been preferred to
>>     >> configure
>>     >>   the MDS's cache limits.
>>     >>
>>     >> * The `pg_autoscale_mode` is now set to `on` by default for newly
>>     >>   created pools, which means that Ceph will automatically manage the
>>     >>   number of PGs.  To change this behavior, or to learn more about PG
>>     >>   autoscaling, see :ref:`pg-autoscaler`.  Note that existing pools in
>>     >>   upgraded clusters will still be set to `warn` by default.
>>     >>
>>     >> * The `upmap_max_iterations` config option of mgr/balancer has been
>>     >>   renamed to `upmap_max_optimizations` to better match its behaviour.
>>     >>
>>     >> * `mClockClientQueue` and `mClockClassQueue` OpQueue
>>     >>   implementations have been removed in favor of of a single
>>     >>   `mClockScheduler` implementation of a simpler OSD interface.
>>     >>   Accordingly, the `osd_op_queue_mclock*` family of config options
>>     >>   has been removed in favor of the `osd_mclock_scheduler*` family
>>     >>   of options.
>>     >>
>>     >> * The config subsystem now searches dot ('.') delimited prefixes for
>>     >>   options.  That means for an entity like `client.foo.bar`, its
>>     >>   overall configuration will be a combination of the global options,
>>     >>   `client`, `client.foo`, and `client.foo.bar`.  Previously,
>>     >>   only global, `client`, and `client.foo.bar` options would apply.
>>     >>   This change may affect the configuration for clients that include a
>>     >>   `.` in their name.
>>     >>
>>     >> Getting Ceph
>>     >> ------------
>>     >> * Git at git://github.com/ceph/ceph.git
>>     <http://github.com/ceph/ceph.git>
>>     >> * Tarball at http://download.ceph.com/tarballs/ceph-15.2.0.tar.gz
>>     >> * For packages, see
>>     http://docs.ceph.com/docs/master/install/get-packages/
>>     >> * Release git sha1: dc6a0b5c3cbf6a5e1d6d4f20b5ad466d76b96247
>>     >>
>>     >> --
>>     >> Abhishek Lekshmanan
>>     >> SUSE Software Solutions Germany GmbH
>>     >> GF: Felix Imendörffer
>>     >> _______________________________________________
>>     >> ceph-users mailing list -- ceph-users@xxxxxxx
>>     <mailto:ceph-users@xxxxxxx>
>>     >> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>     <mailto:ceph-users-leave@xxxxxxx>
>>     >>
>>
>>     -- 
>>     Abhishek Lekshmanan
>>     SUSE Software Solutions Germany GmbH
>>     GF: Felix Imendörffer, Mary Higgins, Sri Rasiah, HRB 21284 (AG Nürnberg)
>>
_______________________________________________
Dev mailing list -- dev@xxxxxxx
To unsubscribe send an email to dev-leave@xxxxxxx




[Index of Archives]     [CEPH Users]     [Ceph Devel]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux