Re: v0.53 released

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Sage,

Am 19.10.2012 um 17:48 schrieb Sage Weil <sage@xxxxxxxxxxx>:

> On Fri, 19 Oct 2012, Oliver Francke wrote:
>> Hi Josh,
>> 
>> On 10/19/2012 07:42 AM, Josh Durgin wrote:
>>> On 10/17/2012 04:26 AM, Oliver Francke wrote:
>>>> Hi Sage, *,
>>>> 
>>>> after having some trouble with the journals - had to erase the partition
>>>> and redo a ceph... --mkjournal - I started my testing... Everything fine.
>>> 
>>> This would be due to the change in default osd journal size. In 0.53
>>> it's 1024MB, even for block devices. Previously it defaulted to
>>> the entire block device.
>>> 
>>> I already fixed this to use the entire block device in 0.54, and
>>> didn't realize the fix wasn't included in 0.53.
>>> 
>>> You can restore the correct behaviour for block devices by setting
>>> this in the [osd] section of your ceph.conf:
>>> 
>>> osd journal size = 0
>> 
>> thnx for the explanation, gives me a better feeling for the next stable to
>> come to the stores ;)
>> Uhm, may it be impertinant to bring http://tracker.newdream.net/issues/2573 to
>> your attention, as it's still ongoing at least in 0.48.2argonaut?
> 
> Do you mean these messages?
> 
> 2012-10-11 10:51:25.879084 7f25d08dc700 0 osd.13 1353 pg[6.5( v 
> 1353'2567562 (1353'2566561,1353'2567562] n=1857 ec=390 les/c 1347/1349 
> 1340/1347/1333) [13,33] r=0 lpr=1347 mlcod 1353'2567561 active+clean] 
> watch: ctx->obc=0x6381000 cookie=1 oi.version=2301953 
> ctx->at_version=1353'2567563
> 2012-10-11 10:51:25.879133 7f25d08dc700 0 osd.13 1353 pg[6.5( v 
> 1353'2567562 (1353'2566561,1353'2567562] n=1857 ec=390 les/c 1347/1349 
> 1340/1347/1333) [13,33] r=0 lpr=1347 mlcod 1353'2567561 active+clean] 
> watch: oi.user_version=2301951
> 
> They're fixed in master; I'll backport the cleanup to stable.  It's 
> useless noise.
> 

uhm, more into the following:

Oct 19 15:28:13 fcmsnode1 kernel: [1483536.141269] libceph: osd13 10.10.10.22:6812 socket closed
Oct 19 15:43:13 fcmsnode1 kernel: [1484435.176280] libceph: osd13 10.10.10.22:6812 socket closed
Oct 19 15:58:13 fcmsnode1 kernel: [1485334.382798] libceph: osd13 10.10.10.22:6812 socket closed

It's kind of "new", cause I would have got aware before. And we have 4 ODS's on every node, so why only from one OSD?
Same picture on two other nodes, If I read the ticket, no data is lost, just closing a socket? But then a kern.log entry is far too much? ;)

Oliver.


> sage
> 
> 
> 
>> 
>> Thnx in advance,
>> 
>> Oliver.
>> 
>>> 
>>> Josh
>>> 
>>>> 
>>>> --- 8-< ---
>>>> 2012-10-17 12:54:11.167782 7febab24a780  0 filestore(/data/osd0) mount:
>>>> enabling PARALLEL journal mode: btrfs, SNAP_CREATE_V2 detected and
>>>> 'filestore btrfs snap' mode is enabled
>>>> 2012-10-17 12:54:11.191723 7febab24a780  0 journal  kernel version is
>>>> 3.5.0
>>>> 2012-10-17 12:54:11.191907 7febab24a780  1 journal _open /dev/sdb1 fd
>>>> 27: 1073741824 bytes, block size 4096 bytes, directio = 1, aio = 1
>>>> 2012-10-17 12:54:11.201764 7febab24a780  0 journal  kernel version is
>>>> 3.5.0
>>>> 2012-10-17 12:54:11.201924 7febab24a780  1 journal _open /dev/sdb1 fd
>>>> 27: 1073741824 bytes, block size 4096 bytes, directio = 1, aio = 1
>>>> --- 8-< ---
>>>> 
>>>> And the other minute I started my fairly destructive testing, 0.52 never
>>>> ever failed on that. And then a loop started with
>>>> --- 8-< ---
>>>> 
>>>> 2012-10-17 12:59:15.403247 7feba5fed700  0 -- 10.0.0.11:6801/29042 >>
>>>> 10.0.0.12:6801/17706 pipe(0x55a2240 sd=34 :57922 pgs=3 cs=1 l=0).fault,
>>>> initiating reconnect
>>>> 2012-10-17 12:59:17.280143 7feb950cc700  0 -- 10.0.0.11:6801/29042 >>
>>>> 10.0.0.12:6804/17972 pipe(0x17f2240 sd=29 :49431 pgs=3 cs=1 l=0).fault
>>>> with nothing to send, going to standby
>>>> 2012-10-17 12:59:18.288902 7feb951cd700  0 -- 10.0.0.11:6801/29042 >>
>>>> 10.0.0.12:6801/17706 pipe(0x55a2240 sd=34 :37519 pgs=3 cs=2 l=0).connect
>>>> claims to be 0.0.0.0:6801/5738 not 10.0.0.12:6801/17706 - wrong node!
>>>> 2012-10-17 12:59:18.297663 7feb951cd700  0 -- 10.0.0.11:6801/29042 >>
>>>> 10.0.0.12:6801/17706 pipe(0x55a2240 sd=34 :34833 pgs=3 cs=2 l=0).connect
>>>> claims to be 0.0.0.0:6801/5738 not 10.0.0.12:6801/17706 - wrong node!
>>>> 2012-10-17 12:59:18.303215 7feb951cd700  0 -- 10.0.0.11:6801/29042 >>
>>>> 10.0.0.12:6801/17706 pipe(0x55a2240 sd=34 :35169 pgs=3 cs=2 l=0).connect
>>>> claims to be 0.0.0.0:6801/5738 not 10.0.0.12:6801/17706 - wrong node!
>>>> --- 8-< ---
>>>> 
>>>> leading to high CPU-load on node2 ( IP 10.0.0.11). The destructive part
>>>> happens on node3 ( IP 10.0.0.12).
>>>> 
>>>> Procedure is as always just kill some OSDs and start over again...
>>>> Happened now twice, so I would call it reproducable ;)
>>>> 
>>>> Kind regards,
>>>> 
>>>> Oliver.
>>>> 
>>>> 
>>>> On 10/17/2012 01:48 AM, Sage Weil wrote:
>>>>> Another development release of Ceph is ready, v0.53. We are getting
>>>>> pretty
>>>>> close to what will be frozen for the next stable release (bobtail), so
>>>>> if
>>>>> you would like a preview, give this one a go. Notable changes include:
>>>>> 
>>>>>  * librbd: image locking
>>>>>  * rbd: fix list command when more than 1024 (format 2) images
>>>>>  * osd: backfill reservation framework (to avoid flooding new osds with
>>>>>    backfill data)
>>>>>  * osd, mon: honor new 'nobackfill' and 'norecover' osdmap flags
>>>>>  * osd: new 'deep scrub' will compare object content across replicas
>>>>> (once
>>>>>    per week by default)
>>>>>  * osd: crush performance improvements
>>>>>  * osd: some performance improvements related to request queuing
>>>>>  * osd: capability syntax improvements, bug fixes
>>>>>  * osd: misc recovery fixes
>>>>>  * osd: fix memory leak on certain error paths
>>>>>  * osd: default journal size to 1 GB
>>>>>  * crush: default root of tree type is now 'root' instead of 'pool' (to
>>>>>    avoid confusiong wrt rados pools)
>>>>>  * ceph-fuse: fix handling for .. in root directory
>>>>>  * librados: some locking fixes
>>>>>  * mon: some election bug fixes
>>>>>  * mon: some additional on-disk metadata to facilitate future mon
>>>>> changes
>>>>>    (post-bobtail)
>>>>>  * mon: throttle osd flapping based on osd history (limits osdmap
>>>>>    "thrashing" on overloaded or unhappy clusters)
>>>>>  * mon: new 'osd crush create-or-move ...' command
>>>>>  * radosgw: fix copy-object vs attributes
>>>>>  * radosgw: fix bug in bucket stat updates
>>>>>  * mds: fix ino release on abort session close, relative getattr
>>>>> path, mds
>>>>>    shutdown, other misc items
>>>>>  * upstart: stop jobs on shutdown
>>>>>  * common: thread pool sizes can now be adjusted at runtime
>>>>>  * build fixes for Fedora 18, CentOS/RHEL 6
>>>>> 
>>>>> The big items are locking support in RBD, and OSD improvements like deep
>>>>> scrub (which verify object data across replicas) and backfill
>>>>> reservations
>>>>> (which limit load on expanding clusters). And a huge swath of bugfixes
>>>>> and
>>>>> cleanups, many due to feeding the code through scan.coverity.com (they
>>>>> offer free static code analysis for open source projects).
>>>>> 
>>>>> v0.54 is now frozen, and will include many deployment-related fixes
>>>>> (including a new ceph-deploy tool to replace mkcephfs), more bugfixes
>>>>> for
>>>>> libcephfs, ceph-fuse, and the MDS, and the fruits of some performance
>>>>> work
>>>>> on the OSD.
>>>>> 
>>>>> You can get v0.53 from the usual locations:
>>>>> 
>>>>>  * Git at git://github.com/ceph/ceph.git
>>>>>  * Tarball at http://ceph.com/download/ceph-0.53.tar.gz
>>>>>  * For Debian/Ubuntu packages, see
>>>>> http://ceph.com/docs/master/install/debian
>>>>>  * For RPMs, see http://ceph.com/docs/master/install/rpm
>>>>> -- 
>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>>>> the body of a message to majordomo@xxxxxxxxxxxxxxx
>>>>> More majordomo info at http://vger.kernel.org/majordomo-info.html
>>>> 
>>>> 
>>> 
>> 
>> 
>> -- 
>> 
>> Oliver Francke
>> 
>> filoo GmbH
>> Moltkestra?e 25a
>> 33330 G?tersloh
>> HRB4355 AG G?tersloh
>> 
>> Gesch?ftsf?hrer: S.Grewing | J.Rehp?hler | C.Kunz
>> 
>> Folgen Sie uns auf Twitter: http://twitter.com/filoogmbh
>> 
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@xxxxxxxxxxxxxxx
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> 
>> 
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux