Re: Problem with query and any operation on PGs

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello,

> On Wed, 24 May 2017, Łukasz Chrustek wrote:
>> Cześć,
>> 
>> > On Wed, 24 May 2017, Łukasz Chrustek wrote:
>> >> 
>> >> >> And  now  it  is very weird.... I made osd.37 up, and loop
>> >> >> while true;do; ceph tell 1.165 query ;done
>> >> 
>> >> > Here  need  to  explain  more  - all I did was start ceph-osd id=37 on
>> >> > storage node, in ceph osd tree this osd osd is marked as out:
>> >> 
>> >> 
>> >> > -17  21.49995     host stor8
>> >> >  22   1.59999         osd.22            up  1.00000          1.00000 
>> >> >  23   1.59999         osd.23            up  1.00000          1.00000 
>> >> >  36   2.09999         osd.36            up  1.00000          1.00000 
>> >> >  37   2.09999         osd.37            up        0          1.00000 
>> >> >  38   2.50000         osd.38            up  1.00000          1.00000 
>> >> >  39   2.50000         osd.39            up  1.00000          1.00000 
>> >> >  40   2.50000         osd.40            up        0          1.00000 
>> >> >  41   2.50000         osd.41          down        0          1.00000 
>> >> >  42   2.50000         osd.42            up  1.00000          1.00000 
>> >> >  43   1.59999         osd.43            up  1.00000          1.00000
>> >> 
>> >> >  after start of this osd, ceph tell 1.165 query  worked only for one call of this command
>> >> >> catch this:
>> >> 
>> >> >> https://pastebin.com/zKu06fJn
>> >> 
>> >> here is for pg 1.60:
>> >> 
>> >> https://pastebin.com/Xuk5iFXr
>> 
>> > Look at the bottom, after it says
>> 
>> >             "blocked": "peering is blocked due to down osds",
>> 
>> > Did the 1.165 pg recover?
>> 
>> No it didn't:
>> 
>> [root@cc1 ~]# ceph health detail
>> HEALTH_WARN 1 pgs down; 1 pgs incomplete; 1 pgs peering; 2 pgs stuck inactive
>> pg 1.165 is stuck inactive since forever, current state incomplete, last acting [67,88,48]
>> pg 1.60 is stuck inactive since forever, current state down+remapped+peering, last acting [68]
>> pg 1.60 is down+remapped+peering, acting [68]
>> pg 1.165 is incomplete, acting [67,88,48]
>> [root@cc1 ~]#

> Hrm.

>  ceph daemon osd.67 config set debug_osd 20
>  ceph daemon osd.67 config set debug_ms 1
>  ceph osd down 67

> and capture the log resulting log segment, then post it with 
> ceph-post-file.

args: -- /var/log/ceph/ceph-osd.67.log
/usr/bin/ceph-post-file: upload tag 05a02f14-8fd6-43da-9b9c-e42cd1fce560
/usr/bin/ceph-post-file: user: root@stor3
/usr/bin/ceph-post-file: will upload file /var/log/ceph/ceph-osd.67.log
sftp> mkdir post/05a02f14-8fd6-43da-9b9c-e42cd1fce560_root@stor3_8612f2d9-bb31-4d5e-b3e7-3722f8d13314
sftp> cd post/05a02f14-8fd6-43da-9b9c-e42cd1fce560_root@stor3_8612f2d9-bb31-4d5e-b3e7-3722f8d13314
sftp> put /tmp/tmp.rggR3suNMt user
sftp> put /var/log/ceph/ceph-osd.67.log

/usr/bin/ceph-post-file: copy the upload id below to share with a dev:

ceph-post-file: 05a02f14-8fd6-43da-9b9c-e42cd1fce560

-- 
Regards,,
 Łukasz Chrustek

--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux