Re: v0.53 released

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Josh,

On 10/19/2012 07:42 AM, Josh Durgin wrote:
On 10/17/2012 04:26 AM, Oliver Francke wrote:
Hi Sage, *,

after having some trouble with the journals - had to erase the partition
and redo a ceph... --mkjournal - I started my testing... Everything fine.

This would be due to the change in default osd journal size. In 0.53
it's 1024MB, even for block devices. Previously it defaulted to
the entire block device.

I already fixed this to use the entire block device in 0.54, and
didn't realize the fix wasn't included in 0.53.

You can restore the correct behaviour for block devices by setting
this in the [osd] section of your ceph.conf:

osd journal size = 0

thnx for the explanation, gives me a better feeling for the next stable to come to the stores ;) Uhm, may it be impertinant to bring http://tracker.newdream.net/issues/2573 to your attention, as it's still ongoing at least in 0.48.2argonaut?

Thnx in advance,

Oliver.


Josh


--- 8-< ---
2012-10-17 12:54:11.167782 7febab24a780  0 filestore(/data/osd0) mount:
enabling PARALLEL journal mode: btrfs, SNAP_CREATE_V2 detected and
'filestore btrfs snap' mode is enabled
2012-10-17 12:54:11.191723 7febab24a780 0 journal kernel version is 3.5.0
2012-10-17 12:54:11.191907 7febab24a780  1 journal _open /dev/sdb1 fd
27: 1073741824 bytes, block size 4096 bytes, directio = 1, aio = 1
2012-10-17 12:54:11.201764 7febab24a780 0 journal kernel version is 3.5.0
2012-10-17 12:54:11.201924 7febab24a780  1 journal _open /dev/sdb1 fd
27: 1073741824 bytes, block size 4096 bytes, directio = 1, aio = 1
--- 8-< ---

And the other minute I started my fairly destructive testing, 0.52 never
ever failed on that. And then a loop started with
--- 8-< ---

2012-10-17 12:59:15.403247 7feba5fed700  0 -- 10.0.0.11:6801/29042 >>
10.0.0.12:6801/17706 pipe(0x55a2240 sd=34 :57922 pgs=3 cs=1 l=0).fault,
initiating reconnect
2012-10-17 12:59:17.280143 7feb950cc700  0 -- 10.0.0.11:6801/29042 >>
10.0.0.12:6804/17972 pipe(0x17f2240 sd=29 :49431 pgs=3 cs=1 l=0).fault
with nothing to send, going to standby
2012-10-17 12:59:18.288902 7feb951cd700  0 -- 10.0.0.11:6801/29042 >>
10.0.0.12:6801/17706 pipe(0x55a2240 sd=34 :37519 pgs=3 cs=2 l=0).connect
claims to be 0.0.0.0:6801/5738 not 10.0.0.12:6801/17706 - wrong node!
2012-10-17 12:59:18.297663 7feb951cd700  0 -- 10.0.0.11:6801/29042 >>
10.0.0.12:6801/17706 pipe(0x55a2240 sd=34 :34833 pgs=3 cs=2 l=0).connect
claims to be 0.0.0.0:6801/5738 not 10.0.0.12:6801/17706 - wrong node!
2012-10-17 12:59:18.303215 7feb951cd700  0 -- 10.0.0.11:6801/29042 >>
10.0.0.12:6801/17706 pipe(0x55a2240 sd=34 :35169 pgs=3 cs=2 l=0).connect
claims to be 0.0.0.0:6801/5738 not 10.0.0.12:6801/17706 - wrong node!
--- 8-< ---

leading to high CPU-load on node2 ( IP 10.0.0.11). The destructive part
happens on node3 ( IP 10.0.0.12).

Procedure is as always just kill some OSDs and start over again...
Happened now twice, so I would call it reproducable ;)

Kind regards,

Oliver.


On 10/17/2012 01:48 AM, Sage Weil wrote:
Another development release of Ceph is ready, v0.53. We are getting
pretty
close to what will be frozen for the next stable release (bobtail), so if
you would like a preview, give this one a go. Notable changes include:

  * librbd: image locking
  * rbd: fix list command when more than 1024 (format 2) images
* osd: backfill reservation framework (to avoid flooding new osds with
    backfill data)
  * osd, mon: honor new 'nobackfill' and 'norecover' osdmap flags
  * osd: new 'deep scrub' will compare object content across replicas
(once
    per week by default)
  * osd: crush performance improvements
  * osd: some performance improvements related to request queuing
  * osd: capability syntax improvements, bug fixes
  * osd: misc recovery fixes
  * osd: fix memory leak on certain error paths
  * osd: default journal size to 1 GB
* crush: default root of tree type is now 'root' instead of 'pool' (to
    avoid confusiong wrt rados pools)
  * ceph-fuse: fix handling for .. in root directory
  * librados: some locking fixes
  * mon: some election bug fixes
  * mon: some additional on-disk metadata to facilitate future mon
changes
    (post-bobtail)
  * mon: throttle osd flapping based on osd history (limits osdmap
    "thrashing" on overloaded or unhappy clusters)
  * mon: new 'osd crush create-or-move ...' command
  * radosgw: fix copy-object vs attributes
  * radosgw: fix bug in bucket stat updates
  * mds: fix ino release on abort session close, relative getattr
path, mds
    shutdown, other misc items
  * upstart: stop jobs on shutdown
  * common: thread pool sizes can now be adjusted at runtime
  * build fixes for Fedora 18, CentOS/RHEL 6

The big items are locking support in RBD, and OSD improvements like deep
scrub (which verify object data across replicas) and backfill
reservations
(which limit load on expanding clusters). And a huge swath of bugfixes
and
cleanups, many due to feeding the code through scan.coverity.com (they
offer free static code analysis for open source projects).

v0.54 is now frozen, and will include many deployment-related fixes
(including a new ceph-deploy tool to replace mkcephfs), more bugfixes for
libcephfs, ceph-fuse, and the MDS, and the fruits of some performance
work
on the OSD.

You can get v0.53 from the usual locations:

  * Git at git://github.com/ceph/ceph.git
  * Tarball at http://ceph.com/download/ceph-0.53.tar.gz
  * For Debian/Ubuntu packages, see
http://ceph.com/docs/master/install/debian
  * For RPMs, see http://ceph.com/docs/master/install/rpm
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html





--

Oliver Francke

filoo GmbH
Moltkestraße 25a
33330 Gütersloh
HRB4355 AG Gütersloh

Geschäftsführer: S.Grewing | J.Rehpöhler | C.Kunz

Folgen Sie uns auf Twitter: http://twitter.com/filoogmbh

--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux