Re: MDS Stuck In Replay

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



This looks like a bug we had in our master branch for about 12 hours. Assuming you built from source, pull the latest master and try again. :)

-Greg 

On Jun 21, 2011, at 5:26 PM, Mark Nigh <mnigh@xxxxxxxxxxxxxxx> wrote:

> As you can see, two (2) of the OSDs are down and out. They weren't when I first wrote the post.
> 
> root@ceph001:~# ceph osd dump -o -
> 2011-06-21 19:20:52.025776 mon <- [osd,dump]
> 2011-06-21 19:20:52.026506 mon0 -> 'dumped osdmap epoch 612' (0)
> epoch 612
> fsid f1082578-eb4d-cdd7-4da2-5e6293396315
> created 2011-06-20 16:05:04.711991
> modifed 2011-06-21 18:15:33.498931
> flags
> 
> pg_pool 0 'data' pg_pool(rep pg_size 2 crush_ruleset 0 object_hash rjenkins pg_num 128 pgp_num 128 lpg_num 2 lpgp_num 2 last_change 1 owner 0)
> pg_pool 1 'metadata' pg_pool(rep pg_size 2 crush_ruleset 1 object_hash rjenkins pg_num 128 pgp_num 128 lpg_num 2 lpgp_num 2 last_change 1 owner 0)
> pg_pool 2 'rbd' pg_pool(rep pg_size 2 crush_ruleset 2 object_hash rjenkins pg_num 128 pgp_num 128 lpg_num 2 lpgp_num 2 last_change 1 owner 0)
> 
> max_osd 4
> osd0 up   in  weight 1 up_from 597 up_thru 611 down_at 596 last_clean_interval 2-595 10.6.1.80:6801/1891 10.6.1.80:6802/1891 10.6.1.80:6803/1891
> osd1 up   in  weight 1 up_from 597 up_thru 611 down_at 596 last_clean_interval 3-595 10.6.1.80:6804/1982 10.6.1.80:6805/1982 10.6.1.80:6806/1982
> osd2 down out up_from 596 up_thru 597 down_at 605 last_clean_interval 590-594
> osd3 down out up_from 596 up_thru 597 down_at 605 last_clean_interval 593-594
> 
> They start up fine.
> 
> root@ceph002:~# /etc/init.d/ceph start osd2
> === osd.2 ===
> Mounting Btrfs on ceph002:/mnt/osd2
> Scanning for Btrfs filesystems
> failed to read /dev/sr0
> Starting Ceph osd.2 on ceph002...
> ** WARNING: Ceph is still under development.  Any feedback can be directed  **
> **          at ceph-devel@xxxxxxxxxxxxxxx or http://ceph.newdream.net/.     **
> starting osd2 at 0.0.0.0:6800/8620 osd_data /mnt/osd2 /data/osd2/journal
> root@ceph002:~# /etc/init.d/ceph start osd3
> === osd.3 ===
> Mounting Btrfs on ceph002:/mnt/osd3
> Scanning for Btrfs filesystems
> failed to read /dev/sr0
> Starting Ceph osd.3 on ceph002...
> ** WARNING: Ceph is still under development.  Any feedback can be directed  **
> **          at ceph-devel@xxxxxxxxxxxxxxx or http://ceph.newdream.net/.     **
> starting osd3 at 0.0.0.0:6803/8769 osd_data /mnt/osd3 /data/osd3/journal
> 
> But I do get these messages in the osd logs
> 
> 2011-06-21 19:24:27.937179 7f93df30d700 -- 10.6.1.81:6801/8620 >> 10.6.1.81:6806/8769 pipe(0x30f7a00 sd=19 pgs=0 cs=0 l=0).accept bad authorizer
> 2011-06-21 19:24:27.937254 7f93df30d700 AuthNoneAuthorizeHandle::verify_authorizer() failed to decode
> 2011-06-21 19:24:27.937270 7f93df30d700 -- 10.6.1.81:6801/8620 >> 10.6.1.81:6806/8769 pipe(0x30f7a00 sd=19 pgs=0 cs=0 l=0).accept bad authorizer
> 
> Here are the rule output.
> 
> root@ceph001:~# crushtool -i cm --test --rule 0
> devices weights (hex): [10000,10000,10000,10000]
> rule 0 (data), x = 0..9999
> device 0:      5013
> device 1:      4987
> device 2:      4963
> device 3:      5037
> num results 2: 10000
> 
> root@ceph001:~# crushtool -i cm --test --rule 1
> devices weights (hex): [10000,10000,10000,10000]
> rule 1 (metadata), x = 0..9999
> device 0:      5010
> device 1:      5037
> device 2:      4975
> device 3:      4978
> num results 2: 10000
> 
> 
> Thanks for your assistance.
> 
> Mark Nigh
> Systems Architect
> mnigh@xxxxxxxxxxxxxxx
> (p) 314.392.6926
> 
> 
> 
> -----Original Message-----
> From: Sage Weil [mailto:sage@xxxxxxxxxxxx]
> Sent: Tuesday, June 21, 2011 5:49 PM
> To: Mark Nigh
> Cc: ceph-devel@xxxxxxxxxxxxxxx
> Subject: Re: MDS Stuck In Replay
> 
> On Tue, 21 Jun 2011, Mark Nigh wrote:
>> I am currently testing Ceph on Debian v6.0 (2.6.32) on the following system:
>> 
>> 2 servers each with 2 HDD and 1 osd per HDD for a total of 4 OSD
>> The 1st server has a one (1) mds and mon.
>> 
>> Ceph gets built and functioning correctly but I think it is when I
>> change my crushmap so that no data is stored on a single server that I
>> get my mds in replay. I also get some pgs in peering. See below.
> 
> If there are PGs in peering it will block MDS replay, so that's the core
> issue.
> 
> Can you include the 'ceph osd dump -o -' output so we can verify the
> object pools are using the data and metadata crush rules?
> 
> You can also verify that the individual crush rules are working with
> something like
> 
> $ ceph osd getcrushmap -o cm
> 2011-06-21 15:44:09.434742 mon <- [osd,getcrushmap]
> 2011-06-21 15:44:09.435168 mon0 -> 'got crush map from osdmap epoch 2' (0)
> 2011-06-21 15:44:09.435370 7f8fd5b22720  wrote 320 byte payload to cm
> $ crushtool -i cm --test --rule 0
> devices weights (hex): [10000]
> rule 0 (data), x = 0..9999
> device 0:      10000
> num results 1: 10000
> $ crushtool -i cm --test --rule 1
> devices weights (hex): [10000]
> rule 1 (metadata), x = 0..9999
> device 0:      10000
> num results 1: 10000
> 
> In your case you should see the objects split between 4 OSDs, not the 1 I
> currently have running on my dev box.
> 
> sage
> 
> 
>> 
>> 2011-06-21 16:49:01.551129    pg v1497: 396 pgs: 272 active+clean, 124 peering; 374 MB data, 773 MB used, 11117 GB / 11118 GB avail
>> 2011-06-21 16:49:01.551817   mds e10: 1/1/1 up {0=0=up:replay}
>> 2011-06-21 16:49:01.551836   osd e602: 4 osds: 4 up, 4 in
>> 2011-06-21 16:49:01.551898   log 2011-06-21 16:41:46.966074 mon0 10.6.1.80:6789/0 1 : [INF] mon.0@0 won leader election with quorum 0
>> 2011-06-21 16:49:01.551947   mon e1: 1 mons at {0=10.6.1.80:6789/0}
>> 
>> My crushmap is as follows:
>> 
>> # begin crush map
>> 
>> # devices
>> device 0 device0
>> device 1 device1
>> device 2 device2
>> device 3 device3
>> 
>> # types
>> type 0 device
>> type 1 host
>> type 2 root
>> 
>> # buckets
>> host host0 {
>>        id -1           # do not change unnecessarily
>>        # weight 2.000
>>        alg straw
>>        hash 0  # rjenkins1
>>        item device0 weight 1.000
>>        item device1 weight 1.000
>> }
>> host host1 {
>>        id -2           # do not change unnecessarily
>>        # weight 2.000
>>        alg straw
>>        hash 0  # rjenkins1
>>        item device2 weight 1.000
>>        item device3 weight 1.000
>> }
>> root root {
>>        id -3           # do not change unnecessarily
>>        # weight 2.000
>>        alg straw
>>        hash 0  # rjenkins1
>>        item host0 weight 1.000
>>        item host1 weight 1.000
>> }
>> 
>> # rules
>> rule data {
>>        ruleset 0
>>        type replicated
>>        min_size 1
>>        max_size 10
>>        step take root
>>        step chooseleaf firstn 0 type host
>>        step emit
>> }
>> rule metadata {
>>        ruleset 1
>>        type replicated
>>        min_size 1
>>        max_size 10
>>        step take root
>>        step choose firstn 0 type device
>>        step emit
>> }
>> 
>> I tried to restart the mds daemon with no luck. Here are a few lines of the mds log. Let me know if there is anything else I can provide.
>> 
>> 2011-06-21 16:50:44.047829 7f2fa8483700 -- 10.6.1.80:6800/19997 >> 10.6.1.80:6789/0 pipe(0x281d280 sd=10 pgs=1 cs=1 l=1).reader got message 141 0x283e780 mdsbeacon(4221/0 up:replay seq 141 v10) v2
>> 2011-06-21 16:50:44.047892 7f2fac6fe700 -- 10.6.1.80:6800/19997 >> 10.6.1.80:6789/0 pipe(0x281d280 sd=10 pgs=1 cs=1 l=1).writer: state = 2 policy.server=0
>> 2011-06-21 16:50:44.047918 7f2fac6fe700 -- 10.6.1.80:6800/19997 >> 10.6.1.80:6789/0 pipe(0x281d280 sd=10 pgs=1 cs=1 l=1).write_ack 141
>> 2011-06-21 16:50:44.047944 7f2fac6fe700 -- 10.6.1.80:6800/19997 >> 10.6.1.80:6789/0 pipe(0x281d280 sd=10 pgs=1 cs=1 l=1).writer: state = 2 policy.server=0
>> 2011-06-21 16:50:44.047969 7f2fa9485700 -- 10.6.1.80:6800/19997 <== mon0 10.6.1.80:6789/0 141 ==== mdsbeacon(4221/0 up:replay seq 141 v10) v2 ==== 103+0+0 (2350252520 0 0) 0x283e780 con 0x283d140
>> 2011-06-21 16:50:44.047986 7f2fa9485700 -- 10.6.1.80:6800/19997 dispatch_throttle_release 103 to dispatch throttler 103/104857600
>> 2011-06-21 16:50:44.058654 7f2fa8382700 -- 10.6.1.80:6800/19997 --> osd2 10.6.1.81:6800/1826 -- ping v1 -- ?+0 0x2841300
>> 2011-06-21 16:50:44.058677 7f2fa8382700 -- 10.6.1.80:6800/19997 --> osd0 10.6.1.80:6801/1891 -- ping v1 -- ?+0 0x2841180
>> 2011-06-21 16:50:44.058701 7f2fa6d7a700 -- 10.6.1.80:6800/19997 >> 10.6.1.81:6800/1826 pipe(0x2836a00 sd=9 pgs=7 cs=1 l=1).writer: state = 2 policy.server=0
>> 2011-06-21 16:50:44.058730 7f2fa6d7a700 -- 10.6.1.80:6800/19997 >> 10.6.1.81:6800/1826 pipe(0x2836a00 sd=9 pgs=7 cs=1 l=1).writer: state = 2 policy.server=0
>> 2011-06-21 16:50:44.058765 7f2fa727f700 -- 10.6.1.80:6800/19997 >> 10.6.1.80:6801/1891 pipe(0x281dc80 sd=7 pgs=6 cs=1 l=1).writer: state = 2 policy.server=0
>> 2011-06-21 16:50:44.058810 7f2fa727f700 -- 10.6.1.80:6800/19997 >> 10.6.1.80:6801/1891 pipe(0x281dc80 sd=7 pgs=6 cs=1 l=1).writer: state = 2 policy.server=0
>> 
>> Mark Nigh
>> Systems Architect
>> Netelligent Corporation
>> 
>> 
>> 
>> This transmission and any attached files are privileged, confidential or otherwise the exclusive property of the intended recipient or Netelligent Corporation. If you are not the intended recipient, any disclosure, copying, distribution or use of any of the information contained in or attached to this transmission is strictly prohibited. If you have received this transmission in error, please contact us immediately by responding to this message or by telephone (314-392-6900) and promptly destroy the original transmission and its attachments.
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@xxxxxxxxxxxxxxx
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> 
>> 
> 
> This transmission and any attached files are privileged, confidential or otherwise the exclusive property of the intended recipient or Netelligent Corporation. If you are not the intended recipient, any disclosure, copying, distribution or use of any of the information contained in or attached to this transmission is strictly prohibited. If you have received this transmission in error, please contact us immediately by responding to this message or by telephone (314-392-6900) and promptly destroy the original transmission and its attachments.
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux