RE: v9.1.0 Infernalis release candidate released

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, 14 Oct 2015, Deneau, Tom wrote:
> I tried an rpmbuild on Fedora21 from the tarball which seemed to work ok.
> But having trouble doing "ceph-deploy --overwrite-conf mon create-initial" with 9.1.0".
> This is using ceph-deploy version 1.5.24.
> Is this part of the "needs Fedora 22 or later" story?

Yeah I think so, but it's probably mostly a "tested fc22 and it worked" 
situation.  THis is probably what is failing:

https://github.com/ceph/ceph-deploy/blob/master/ceph_deploy/hosts/fedora/__init__.py#L21

So maybe the specfile isn't using systemd for fc21?

sage


> 
> -- Tom
> 
> [myhost][DEBUG ] create a done file to avoid re-doing the mon deployment
> [myhost][DEBUG ] create the init path if it does not exist
> [myhost][DEBUG ] locating the `service` executable...
> [myhost][INFO  ] Running command: /usr/sbin/service ceph -c /etc/ceph/ceph.conf start mon.myhost
> [myhost][WARNIN] The service command supports only basic LSB actions (start, stop, restart, try-restart, reload, force-reload, sta\
> tus). For other actions, please try to use systemctl.
> [myhost][ERROR ] RuntimeError: command returned non-zero exit status: 2
> [ceph_deploy.mon][ERROR ] Failed to execute command: /usr/sbin/service ceph -c /etc/ceph/ceph.conf start mon.myhost
> [ceph_deploy][ERROR ] GenericError: Failed to create 1 monitors
> 
> 
> > -----Original Message-----
> > From: ceph-devel-owner@xxxxxxxxxxxxxxx [mailto:ceph-devel-
> > owner@xxxxxxxxxxxxxxx] On Behalf Of Sage Weil
> > Sent: Tuesday, October 13, 2015 4:02 PM
> > To: ceph-announce@xxxxxxxx; ceph-devel@xxxxxxxxxxxxxxx; ceph-
> > users@xxxxxxxx; ceph-maintainers@xxxxxxxx
> > Subject: v9.1.0 Infernalis release candidate released
> > 
> > This is the first Infernalis release candidate.  There have been some
> > major changes since hammer, and the upgrade process is non-trivial.
> > Please read carefully.
> > 
> > Getting the release candidate
> > -----------------------------
> > 
> > The v9.1.0 packages are pushed to the development release repositories::
> > 
> >   http://download.ceph.com/rpm-testing
> >   http://download.ceph.com/debian-testing
> > 
> > For for info, see::
> > 
> >   http://docs.ceph.com/docs/master/install/get-packages/
> > 
> > Or install with ceph-deploy via::
> > 
> >   ceph-deploy install --testing HOST
> > 
> > Known issues
> > ------------
> > 
> > * librbd and librados ABI compatibility is broken.  Be careful
> >   installing this RC on client machines (e.g., those running qemu).
> >   It will be fixed in the final v9.2.0 release.
> > 
> > Major Changes from Hammer
> > -------------------------
> > 
> > * *General*:
> >   * Ceph daemons are now managed via systemd (with the exception of
> >     Ubuntu Trusty, which still uses upstart).
> >   * Ceph daemons run as 'ceph' user instead root.
> >   * On Red Hat distros, there is also an SELinux policy.
> > * *RADOS*:
> >   * The RADOS cache tier can now proxy write operations to the base
> >     tier, allowing writes to be handled without forcing migration of
> >     an object into the cache.
> >   * The SHEC erasure coding support is no longer flagged as
> >     experimental. SHEC trades some additional storage space for faster
> >     repair.
> >   * There is now a unified queue (and thus prioritization) of client
> >     IO, recovery, scrubbing, and snapshot trimming.
> >   * There have been many improvements to low-level repair tooling
> >     (ceph-objectstore-tool).
> >   * The internal ObjectStore API has been significantly cleaned up in
> > order
> >     to faciliate new storage backends like NewStore.
> > * *RGW*:
> >   * The Swift API now supports object expiration.
> >   * There are many Swift API compatibility improvements.
> > * *RBD*:
> >   * The ``rbd du`` command shows actual usage (quickly, when
> >     object-map is enabled).
> >   * The object-map feature has seen many stability improvements.
> >   * Object-map and exclusive-lock features can be enabled or disabled
> >     dynamically.
> >   * You can now store user metadata and set persistent librbd options
> >     associated with individual images.
> >   * The new deep-flatten features allows flattening of a clone and all
> >     of its snapshots.  (Previously snapshots could not be flattened.)
> >   * The export-diff command command is now faster (it uses aio).  There is
> > also
> >     a new fast-diff feature.
> >   * The --size argument can be specified with a suffix for units
> >     (e.g., ``--size 64G``).
> >   * There is a new ``rbd status`` command that, for now, shows who has
> >     the image open/mapped.
> > * *CephFS*:
> >   * You can now rename snapshots.
> >   * There have been ongoing improvements around administration,
> > diagnostics,
> >     and the check and repair tools.
> >   * The caching and revocation of client cache state due to unused
> >     inodes has been dramatically improved.
> >   * The ceph-fuse client behaves better on 32-bit hosts.
> > 
> > Distro compatibility
> > --------------------
> > 
> > We have decided to drop support for many older distributions so that we
> > can move to a newer compiler toolchain (e.g., C++11).  Although it is
> > still possible to build Ceph on older distributions by installing
> > backported development tools, we are not building and publishing release
> > packages for ceph.com.
> > 
> > In particular,
> > 
> > * CentOS 7 or later; we have dropped support for CentOS 6 (and other
> >   RHEL 6 derivatives, like Scientific Linux 6).
> > * Debian Jessie 8.x or later; Debian Wheezy 7.x's g++ has incomplete
> >   support for C++11 (and no systemd).
> > * Ubuntu Trusty 14.04 or later; Ubuntu Precise 12.04 is no longer
> >   supported.
> > * Fedora 22 or later.
> > 
> > Upgrading from Firefly
> > ----------------------
> > 
> > Upgrading directly from Firefly v0.80.z is not possible.  All clusters
> > must first upgrade to Hammer v0.94.4 or a later v0.94.z release; only then
> > is it possible to upgrade to Infernalis 9.2.z.
> > 
> > Note that v0.94.4 isn't released yet, but you can upgrade to a test build
> > from gitbuilder with::
> > 
> >   ceph-deploy install --dev hammer HOST
> > 
> > The v0.94.4 Hammer point release will be out before v9.2.0 Infernalis is.
> > 
> > Upgrading from Hammer
> > ---------------------
> > 
> > * For all distributions that support systemd (CentOS 7, Fedora, Debian
> >   Jessie 8.x, OpenSUSE), ceph daemons are now managed using native systemd
> >   files instead of the legacy sysvinit scripts.  For example,::
> > 
> >     systemctl start ceph.target       # start all daemons
> >     systemctl status ceph-osd@12      # check status of osd.12
> > 
> >   The main notable distro that is *not* yet using systemd is Ubuntu trusty
> >   14.04.  (The next Ubuntu LTS, 16.04, will use systemd instead of
> > upstart.)
> > 
> > * Ceph daemons now run as user and group ``ceph`` by default.  The
> >   ceph user has a static UID assigned by Fedora and Debian (also used
> >   by derivative distributions like RHEL/CentOS and Ubuntu).  On SUSE
> >   the ceph user will currently get a dynamically assigned UID when the
> >   user is created.
> > 
> >   If your systems already have a ceph user, upgrading the package will
> > cause
> >   problems.  We suggest you first remove or rename the existing 'ceph'
> > user
> >   before upgrading.
> > 
> >   When upgrading, administrators have two options:
> > 
> >    #. Add the following line to ``ceph.conf`` on all hosts::
> > 
> >         setuser match path = /var/lib/ceph/$type/$cluster-$id
> > 
> >       This will make the Ceph daemons run as root (i.e., not drop
> >       privileges and switch to user ceph) if the daemon's data
> >       directory is still owned by root.  Newly deployed daemons will
> >       be created with data owned by user ceph and will run with
> >       reduced privileges, but upgraded daemons will continue to run as
> >       root.
> > 
> >    #. Fix the data ownership during the upgrade.  This is the preferred
> > option,
> >       but is more work.  The process for each host would be to:
> > 
> >       #. Upgrade the ceph package.  This creates the ceph user and group.
> > For
> > 	 example::
> > 
> > 	   ceph-deploy install --stable infernalis HOST
> > 
> >       #. Stop the daemon(s).::
> > 
> > 	   service ceph stop           # fedora, centos, rhel, debian
> > 	   stop ceph-all               # ubuntu
> > 
> >       #. Fix the ownership::
> > 
> > 	   chown -R ceph:ceph /var/lib/ceph
> > 
> >       #. Restart the daemon(s).::
> > 
> > 	   start ceph-all                # ubuntu
> > 	   systemctl start ceph.target   # debian, centos, fedora, rhel
> > 
> > * The on-disk format for the experimental KeyValueStore OSD backend has
> >   changed.  You will need to remove any OSDs using that backend before you
> >   upgrade any test clusters that use it.
> > 
> > Upgrade notes
> > -------------
> > 
> > * When a pool quota is reached, librados operations now block
> > indefinitely,
> >   the same way they do when the cluster fills up.  (Previously they would
> > return
> >   -ENOSPC).  By default, a full cluster or pool will now block.  If your
> >   librados application can handle ENOSPC or EDQUOT errors gracefully, you
> > can
> >   get error returns instead by using the new librados OPERATION_FULL_TRY
> > flag.
> > 
> > Notable changes
> > ---------------
> > 
> > NOTE: These notes are somewhat abbreviated while we find a less time-
> > consuming process for generating them.
> > 
> > * build: C++11 now supported
> > * build: many cmake improvements
> > * build: OSX build fixes (Yan, Zheng)
> > * build: remove rest-bench
> > * ceph-disk: many fixes (Loic Dachary)
> > * ceph-disk: support for multipath devices (Loic Dachary)
> > * ceph-fuse: mostly behave on 32-bit hosts (Yan, Zheng)
> > * ceph-objectstore-tool: many improvements (David Zafman)
> > * common: bufferlist performance tuning (Piotr Dalek, Sage Weil)
> > * common: make mutex more efficient
> > * common: some async compression infrastructure (Haomai Wang)
> > * librados: add FULL_TRY and FULL_FORCE flags for dealing with full
> > clusters or pools (Sage Weil)
> > * librados: fix notify completion race (#13114 Sage Weil)
> > * librados, libcephfs: randomize client nonces (Josh Durgin)
> > * librados: pybind: fix binary omap values (Robin H. Johnson)
> > * librbd: fix reads larger than the cache size (Lu Shi)
> > * librbd: metadata filter fixes (Haomai Wang)
> > * librbd: use write_full when possible (Zhiqiang Wang)
> > * mds: avoid emitting cap warnigns before evicting session (John Spray)
> > * mds: fix expected holes in journal objects (#13167 Yan, Zheng)
> > * mds: fix SnapServer crash on deleted pool (John Spray)
> > * mds: many fixes (Yan, Zheng, John Spray, Greg Farnum)
> > * mon: add cache over MonitorDBStore (Kefu Chai)
> > * mon: 'ceph osd metadata' can dump all osds (Haomai Wang)
> > * mon: detect kv backend failures (Sage Weil)
> > * mon: fix CRUSH map test for new pools (Sage Weil)
> > * mon: fix min_last_epoch_clean tracking (Kefu Chai)
> > * mon: misc scaling fixes (Sage Weil)
> > * mon: streamline session handling, fix memory leaks (Sage Weil)
> > * mon: upgrades must pass through hammer (Sage Weil)
> > * msg/async: many fixes (Haomai Wang)
> > * osd: cache proxy-write support (Zhiqiang Wang, Samuel Just)
> > * osd: configure promotion based on write recency (Zhiqiang Wang)
> > * osd: don't send dup MMonGetOSDMap requests (Sage Weil, Kefu Chai)
> > * osd: erasure-code: fix SHEC floating point bug (#12936 Loic Dachary)
> > * osd: erasure-code: update to ISA-L 2.14 (Yuan Zhou)
> > * osd: fix hitset object naming to use GMT (Kefu Chai)
> > * osd: fix misc memory leaks (Sage Weil)
> > * osd: fix peek_queue locking in FileStore (Xinze Chi)
> > * osd: fix promotion vs full cache tier (Samuel Just)
> > * osd: fix replay requeue when pg is still activating (#13116 Samuel Just)
> > * osd: fix scrub stat bugs (Sage Weil, Samuel Just)
> > * osd: force promotion for ops EC can't handle (Zhiqiang Wang)
> > * osd: improve behavior on machines with large memory pages (Steve Capper)
> > * osd: merge multiple setattr calls into a setattrs call (Xinxin Shu)
> > * osd: newstore prototype (Sage Weil)
> > * osd: ObjectStore internal API refactor (Sage Weil)
> > * osd: SHEC no longer experimental
> > * osd: throttle evict ops (Yunchuan Wen)
> > * osd: upgrades must pass through hammer (Sage Weil)
> > * osd: use SEEK_HOLE / SEEK_DATA for sparse copy (Xinxin Shu)
> > * rbd: rbd-replay-prep and rbd-replay improvements (Jason Dillaman)
> > * rgw: expose the number of unhealthy workers through admin socket (Guang
> > Yang)
> > * rgw: fix casing of Content-Type header (Robin H. Johnson)
> > * rgw: fix decoding of X-Object-Manifest from GET on Swift DLO (Radslow
> > Rzarzynski)
> > * rgw: fix sysvinit script
> > * rgw: fix sysvinit script w/ multiple instances (Sage Weil, Pavan
> > Rallabhandi)
> > * rgw: improve handling of already removed buckets in expirer (Radoslaw
> > Rzarzynski)
> > * rgw: log to /var/log/ceph instead of /var/log/radosgw
> > * rgw: rework X-Trans-Id header to be conform with Swift API (Radoslaw
> > Rzarzynski)
> > * rgw: s3 encoding-type for get bucket (Jeff Weber)
> > * rgw: set max buckets per user in ceph.conf (Vikhyat Umrao)
> > * rgw: support for Swift expiration API (Radoslaw Rzarzynski, Yehuda
> > Sadeh)
> > * rgw: user rm is idempotent (Orit Wasserman)
> > * selinux policy (Boris Ranto, Milan Broz)
> > * systemd: many fixes (Sage Weil, Owen Synge, Boris Ranto, Dan van der
> > Ster)
> > * systemd: run daemons as user ceph
> > 
> > Getting Ceph
> > ------------
> > 
> > * Git at git://github.com/ceph/ceph.git
> > * Tarball at http://download.ceph.com/tarballs/ceph-9.1.0.tar.gz
> > * For packages, see http://ceph.com/docs/master/install/get-packages
> > * For ceph-deploy, see http://ceph.com/docs/master/install/install-ceph-
> > deploy
> > --
> > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> > the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at
> > http://vger.kernel.org/majordomo-info.html
> 
> 
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux