Re: v0.53 released

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Sage, *,

after having some trouble with the journals - had to erase the partition and redo a ceph... --mkjournal - I started my testing... Everything fine.

--- 8-< ---
2012-10-17 12:54:11.167782 7febab24a780 0 filestore(/data/osd0) mount: enabling PARALLEL journal mode: btrfs, SNAP_CREATE_V2 detected and 'filestore btrfs snap' mode is enabled
2012-10-17 12:54:11.191723 7febab24a780  0 journal  kernel version is 3.5.0
2012-10-17 12:54:11.191907 7febab24a780 1 journal _open /dev/sdb1 fd 27: 1073741824 bytes, block size 4096 bytes, directio = 1, aio = 1
2012-10-17 12:54:11.201764 7febab24a780  0 journal  kernel version is 3.5.0
2012-10-17 12:54:11.201924 7febab24a780 1 journal _open /dev/sdb1 fd 27: 1073741824 bytes, block size 4096 bytes, directio = 1, aio = 1
--- 8-< ---

And the other minute I started my fairly destructive testing, 0.52 never ever failed on that. And then a loop started with
--- 8-< ---

2012-10-17 12:59:15.403247 7feba5fed700 0 -- 10.0.0.11:6801/29042 >> 10.0.0.12:6801/17706 pipe(0x55a2240 sd=34 :57922 pgs=3 cs=1 l=0).fault, initiating reconnect 2012-10-17 12:59:17.280143 7feb950cc700 0 -- 10.0.0.11:6801/29042 >> 10.0.0.12:6804/17972 pipe(0x17f2240 sd=29 :49431 pgs=3 cs=1 l=0).fault with nothing to send, going to standby 2012-10-17 12:59:18.288902 7feb951cd700 0 -- 10.0.0.11:6801/29042 >> 10.0.0.12:6801/17706 pipe(0x55a2240 sd=34 :37519 pgs=3 cs=2 l=0).connect claims to be 0.0.0.0:6801/5738 not 10.0.0.12:6801/17706 - wrong node! 2012-10-17 12:59:18.297663 7feb951cd700 0 -- 10.0.0.11:6801/29042 >> 10.0.0.12:6801/17706 pipe(0x55a2240 sd=34 :34833 pgs=3 cs=2 l=0).connect claims to be 0.0.0.0:6801/5738 not 10.0.0.12:6801/17706 - wrong node! 2012-10-17 12:59:18.303215 7feb951cd700 0 -- 10.0.0.11:6801/29042 >> 10.0.0.12:6801/17706 pipe(0x55a2240 sd=34 :35169 pgs=3 cs=2 l=0).connect claims to be 0.0.0.0:6801/5738 not 10.0.0.12:6801/17706 - wrong node!
--- 8-< ---

leading to high CPU-load on node2 ( IP 10.0.0.11). The destructive part happens on node3 ( IP 10.0.0.12).

Procedure is as always just kill some OSDs and start over again... Happened now twice, so I would call it reproducable ;)

Kind regards,

Oliver.


On 10/17/2012 01:48 AM, Sage Weil wrote:
Another development release of Ceph is ready, v0.53. We are getting pretty
close to what will be frozen for the next stable release (bobtail), so if
you would like a preview, give this one a go. Notable changes include:

  * librbd: image locking
  * rbd: fix list command when more than 1024 (format 2) images
  * osd: backfill reservation framework (to avoid flooding new osds with
    backfill data)
  * osd, mon: honor new 'nobackfill' and 'norecover' osdmap flags
  * osd: new 'deep scrub' will compare object content across replicas (once
    per week by default)
  * osd: crush performance improvements
  * osd: some performance improvements related to request queuing
  * osd: capability syntax improvements, bug fixes
  * osd: misc recovery fixes
  * osd: fix memory leak on certain error paths
  * osd: default journal size to 1 GB
  * crush: default root of tree type is now 'root' instead of 'pool' (to
    avoid confusiong wrt rados pools)
  * ceph-fuse: fix handling for .. in root directory
  * librados: some locking fixes
  * mon: some election bug fixes
  * mon: some additional on-disk metadata to facilitate future mon changes
    (post-bobtail)
  * mon: throttle osd flapping based on osd history (limits osdmap
    "thrashing" on overloaded or unhappy clusters)
  * mon: new 'osd crush create-or-move ...' command
  * radosgw: fix copy-object vs attributes
  * radosgw: fix bug in bucket stat updates
  * mds: fix ino release on abort session close, relative getattr path, mds
    shutdown, other misc items
  * upstart: stop jobs on shutdown
  * common: thread pool sizes can now be adjusted at runtime
  * build fixes for Fedora 18, CentOS/RHEL 6

The big items are locking support in RBD, and OSD improvements like deep
scrub (which verify object data across replicas) and backfill reservations
(which limit load on expanding clusters). And a huge swath of bugfixes and
cleanups, many due to feeding the code through scan.coverity.com (they
offer free static code analysis for open source projects).

v0.54 is now frozen, and will include many deployment-related fixes
(including a new ceph-deploy tool to replace mkcephfs), more bugfixes for
libcephfs, ceph-fuse, and the MDS, and the fruits of some performance work
on the OSD.

You can get v0.53 from the usual locations:

  * Git at git://github.com/ceph/ceph.git
  * Tarball at http://ceph.com/download/ceph-0.53.tar.gz
  * For Debian/Ubuntu packages, see http://ceph.com/docs/master/install/debian
  * For RPMs, see http://ceph.com/docs/master/install/rpm
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


--

Oliver Francke

filoo GmbH
Moltkestraße 25a
33330 Gütersloh
HRB4355 AG Gütersloh

Geschäftsführer: S.Grewing | J.Rehpöhler | C.Kunz

Folgen Sie uns auf Twitter: http://twitter.com/filoogmbh

--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux