On Tue, Dec 13, 2016 at 12:18 PM, Dietmar Rieder <dietmar.rieder@xxxxxxxxxxx> wrote: > Hi John, > > Thanks for your answer. > The mentioned modification of the pool validation would than allow for > CephFS having the data pools on EC while keeping the metadata on a > replicated pool, right? I would expect so. John > > Dietmar > > On 12/13/2016 12:35 PM, John Spray wrote: >> On Tue, Dec 13, 2016 at 7:35 AM, Dietmar Rieder >> <dietmar.rieder@xxxxxxxxxxx> wrote: >>> Hi, >>> >>> this is good news! Thanks. >>> >>> As far as I see the RBD supports (experimentally) now EC data pools. Is >>> this true also for CephFS? It is not stated in the announce, so I wonder >>> if and when EC pools are planned to be supported by CephFS. >> >> Nobody has worked on this so far. For EC data pools, it should mainly >> be a case of modifying the pool validation in MDSMonitor that >> currently prevents assigning an EC pool. I strongly suspect we'll get >> around to this before Luminous. >> >> John >> >>> ~regards >>> Dietmar >>> >>> On 12/13/2016 03:28 AM, Abhishek L wrote: >>>> Hi everyone, >>>> >>>> This is the first release candidate for Kraken, the next stable >>>> release series. There have been major changes from jewel with many >>>> features being added. Please note the upgrade process from jewel, >>>> before upgrading. >>>> >>>> Major Changes from Jewel >>>> ------------------------ >>>> >>>> - *RADOS*: >>>> >>>> * The new *BlueStore* backend now has a stable disk format and is >>>> passing our failure and stress testing. Although the backend is >>>> still flagged as experimental, we encourage users to try it out >>>> for non-production clusters and non-critical data sets. >>>> * RADOS now has experimental support for *overwrites on >>>> erasure-coded* pools. Because the disk format and implementation >>>> are not yet finalized, there is a special pool option that must be >>>> enabled to test the new feature. Enabling this option on a cluster >>>> will permanently bar that cluster from being upgraded to future >>>> versions. >>>> * We now default to the AsyncMessenger (``ms type = async``) instead >>>> of the legacy SimpleMessenger. The most noticeable difference is >>>> that we now use a fixed sized thread pool for network connections >>>> (instead of two threads per socket with SimpleMessenger). >>>> * Some OSD failures are now detected almost immediately, whereas >>>> previously the heartbeat timeout (which defaults to 20 seconds) >>>> had to expire. This prevents IO from blocking for an extended >>>> period for failures where the host remains up but the ceph-osd >>>> process is no longer running. >>>> * There is a new ``ceph-mgr`` daemon. It is currently collocated with >>>> the monitors by default, and is not yet used for much, but the basic >>>> infrastructure is now in place. >>>> * The size of encoded OSDMaps has been reduced. >>>> * The OSDs now quiesce scrubbing when recovery or rebalancing is in progress. >>>> >>>> - *RGW*: >>>> >>>> * RGW now supports a new zone type that can be used for metadata indexing >>>> via Elasticseasrch. >>>> * RGW now supports the S3 multipart object copy-part API. >>>> * It is possible now to reshard an existing bucket. Note that bucket >>>> resharding currently requires that all IO (especially writes) to >>>> the specific bucket is quiesced. >>>> * RGW now supports data compression for objects. >>>> * Civetweb version has been upgraded to 1.8 >>>> * The Swift static website API is now supported (S3 support has been added >>>> previously). >>>> * S3 bucket lifecycle API has been added. Note that currently it only supports >>>> object expiration. >>>> * Support for custom search filters has been added to the LDAP auth >>>> implementation. >>>> * Support for NFS version 3 has been added to the RGW NFS gateway. >>>> * A Python binding has been created for librgw. >>>> >>>> - *RBD*: >>>> >>>> * RBD now supports images stored in an *erasure-coded* RADOS pool >>>> using the new (experimental) overwrite support. Images must be >>>> created using the new rbd CLI "--data-pool <ec pool>" option to >>>> specify the EC pool where the backing data objects are >>>> stored. Attempting to create an image directly on an EC pool will >>>> not be successful since the image's backing metadata is only >>>> supported on a replicated pool. >>>> * The rbd-mirror daemon now supports replicating dynamic image >>>> feature updates and image metadata key/value pairs from the >>>> primary image to the non-primary image. >>>> * The number of image snapshots can be optionally restricted to a >>>> configurable maximum. >>>> * The rbd Python API now supports asynchronous IO operations. >>>> >>>> - *CephFS*: >>>> >>>> * libcephfs function definitions have been changed to enable proper >>>> uid/gid control. The library version has been increased to reflect the >>>> interface change. >>>> * Standby replay MDS daemons now consume less memory on workloads >>>> doing deletions. >>>> * Scrub now repairs backtrace, and populates `damage ls` with >>>> discovered errors. >>>> * A new `pg_files` subcommand to `cephfs-data-scan` can identify >>>> files affected by a damaged or lost RADOS PG. >>>> * The false-positive "failing to respond to cache pressure" warnings have >>>> been fixed. >>>> >>>> >>>> Upgrading from Jewel >>>> -------------------- >>>> >>>> * All clusters must first be upgraded to Jewel 10.2.z before upgrading >>>> to Kraken 11.2.z (or, eventually, Luminous 12.2.z). >>>> >>>> * The ``sortbitwise`` flag must be set on the Jewel cluster before upgrading >>>> to Kraken. The latest Jewel (10.2.4+) releases issue a health warning if >>>> the flag is not set, so this is probably already set. If it is not, Kraken >>>> OSDs will refuse to start and will print and error message in their log. >>>> >>>> >>>> Upgrading >>>> --------- >>>> >>>> * The list of monitor hosts/addresses for building the monmap can now be >>>> obtained from DNS SRV records. The service name used in when querying the DNS >>>> is defined in the "mon_dns_srv_name" config option, which defaults to >>>> "ceph-mon". >>>> >>>> * The 'osd class load list' config option is a list of object class names that >>>> the OSD is permitted to load (or '*' for all classes). By default it >>>> contains all existing in-tree classes for backwards compatibility. >>>> >>>> * The 'osd class default list' config option is a list of object class >>>> names (or '*' for all classes) that clients may invoke having only >>>> the '*', 'x', 'class-read', or 'class-write' capabilities. By >>>> default it contains all existing in-tree classes for backwards >>>> compatibility. Invoking classes not listed in 'osd class default >>>> list' requires a capability naming the class (e.g. 'allow class >>>> foo'). >>>> >>>> * The 'rgw rest getusage op compat' config option allows you to dump >>>> (or not dump) the description of user stats in the S3 GetUsage >>>> API. This option defaults to false. If the value is true, the >>>> reponse data for GetUsage looks like:: >>>> >>>> "stats": { >>>> "TotalBytes": 516, >>>> "TotalBytesRounded": 1024, >>>> "TotalEntries": 1 >>>> } >>>> >>>> If the value is false, the reponse for GetUsage looks as it did before:: >>>> >>>> { >>>> 516, >>>> 1024, >>>> 1 >>>> } >>>> >>>> * The 'osd out ...' and 'osd in ...' commands now preserve the OSD >>>> weight. That is, after marking an OSD out and then in, the weight >>>> will be the same as before (instead of being reset to 1.0). >>>> Previously the mons would only preserve the weight if the mon >>>> automatically marked and OSD out and then in, but not when an admin >>>> did so explicitly. >>>> >>>> * The 'ceph osd perf' command will display 'commit_latency(ms)' and >>>> 'apply_latency(ms)'. Previously, the names of these two columns are >>>> 'fs_commit_latency(ms)' and 'fs_apply_latency(ms)'. We remove the >>>> prefix 'fs_', because they are not filestore specific. >>>> >>>> * Monitors will no longer allow pools to be removed by default. The >>>> setting mon_allow_pool_delete has to be set to true (defaults to >>>> false) before they allow pools to be removed. This is a additional >>>> safeguard against pools being removed by accident. >>>> >>>> * If you have manually specified the monitor user rocksdb via the >>>> ``mon keyvaluedb = rocksdb`` option, you will need to manually add a >>>> file to the mon data directory to preserve this option:: >>>> >>>> echo rocksdb > /var/lib/ceph/mon/ceph-`hostname`/kv_backend >>>> >>>> New monitors will now use rocksdb by default, but if that file is >>>> not present, existing monitors will use leveldb. The ``mon >>>> keyvaluedb`` option now only affects the backend chosen when a >>>> monitor is created. >>>> >>>> * The 'osd crush initial weight' option allows you to specify a CRUSH >>>> weight for a newly added OSD. Previously a value of 0 (the default) >>>> meant that we should use the size of the OSD's store to weight the >>>> new OSD. Now, a value of 0 means it should have a weight of 0, and >>>> a negative value (the new default) means we should automatically >>>> weight the OSD based on its size. If your configuration file >>>> explicitly specifies a value of 0 for this option you will need to >>>> change it to a negative value (e.g., -1) to preserve the current >>>> behavior. >>>> >>>> * The `osd crush location` config option is no longer supported. Please >>>> update your ceph.conf to use the `crush location` option instead. >>>> >>>> * The static libraries are no longer included by the debian >>>> development packages (lib*-dev) as it is not required per debian >>>> packaging policy. The shared (.so) versions are packaged as before. >>>> >>>> * The libtool pseudo-libraries (.la files) are no longer included by >>>> the debian development packages (lib*-dev) as they are not required >>>> per https://wiki.debian.org/ReleaseGoals/LAFileRemoval and >>>> https://www.debian.org/doc/manuals/maint-guide/advanced.en.html. >>>> >>>> * The jerasure and shec plugins can now detect SIMD instruction at >>>> runtime and no longer need to be explicitly configured for different >>>> processors. The following plugins are now deprecated: >>>> jerasure_generic, jerasure_sse3, jerasure_sse4, jerasure_neon, >>>> shec_generic, shec_sse3, shec_sse4, and shec_neon. If you use any of >>>> these plugins directly you will see a warning in the mon log file. >>>> Please switch to using just 'jerasure' or 'shec'. >>>> >>>> * The librados omap get_keys and get_vals operations include a start key and a >>>> limit on the number of keys to return. The OSD now imposes a configurable >>>> limit on the number of keys and number of total bytes it will respond with, >>>> which means that a librados user might get fewer keys than they asked for. >>>> This is necessary to prevent careless users from requesting an unreasonable >>>> amount of data from the cluster in a single operation. The new limits are >>>> configured with `osd_max_omap_entries_per_request`, defaulting to 131,072, and >>>> 'osd_max_omap_bytes_per_request', defaulting to 4MB. >>>> >>>> >>>> >>>> Due to the really long changelog in this release, please read the >>>> detailed feature list here: >>>> http://ceph.com/releases/v11-1-0-kraken-released/ >>>> >>>> The debian and rpm packages are available at the usual locations at >>>> http://download.ceph.com/debian-kraken/ and >>>> http://download.ceph.com/rpm-kraken respectively. For more details refer >>>> below. >>>> >>>> >>>> Getting Ceph >>>> ------------ >>>> >>>> * Git at git://github.com/ceph/ceph.git >>>> * Tarball at http://download.ceph.com/tarballs/ceph-11.1.0.tar.gz >>>> * For packages, see http://ceph.com/docs/master/install/get-packages >>>> * For ceph-deploy, see http://ceph.com/docs/master/install/install-ceph-deploy >>>> >>>> Best, >>>> Abhishek >>>> _______________________________________________ >>>> ceph-users mailing list >>>> ceph-users@xxxxxxxxxxxxxx >>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>> >>> >>> >>> -- >>> _________________________________________ >>> D i e t m a r R i e d e r, Mag.Dr. >>> Innsbruck Medical University >>> Biocenter - Division for Bioinformatics >>> Innrain 80, 6020 Innsbruck >>> Phone: +43 512 9003 71402 >>> Fax: +43 512 9003 73100 >>> Email: dietmar.rieder@xxxxxxxxxxx >>> Web: http://www.icbi.at >>> >>> >>> >>> _______________________________________________ >>> ceph-users mailing list >>> ceph-users@xxxxxxxxxxxxxx >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>> > > > -- > _________________________________________ > D i e t m a r R i e d e r, Mag.Dr. > Innsbruck Medical University > Biocenter - Division for Bioinformatics > Innrain 80, 6020 Innsbruck > Phone: +43 512 9003 71402 > Fax: +43 512 9003 73100 > Email: dietmar.rieder@xxxxxxxxxxx > Web: http://www.icbi.at > > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com