Re: Problem with query and any operation on PGs

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Cześć,

> On Wed, 24 May 2017, Łukasz Chrustek wrote:
>> Cześć,
>> 
>> > On Tue, 23 May 2017, Łukasz Chrustek wrote:
>> >> Cześć,
>> >> 
>> >> > On Tue, 23 May 2017, Łukasz Chrustek wrote:
>> >> >> I'm  not  sleeping for over 30 hours, and still can't find solution. I
>> >> >> did,      as      You      wrote,     but     turning     off     this
>> >> >> (https://pastebin.com/1npBXeMV) osds didn't resolve issue...
>> >> 
>> >> > The important bit is:
>> >> 
>> >> >             "blocked": "peering is blocked due to down osds",
>> >> >             "down_osds_we_would_probe": [
>> >> >                 6,
>> >> >                 10,
>> >> >                 33,
>> >> >                 37,
>> >> >                 72
>> >> >             ],
>> >> >             "peering_blocked_by": [
>> >> >                 {
>> >> >                     "osd": 6,
>> >> >                     "current_lost_at": 0,
>> >> >                     "comment": "starting or marking this osd lost may let
>> >> > us proceed"
>> >> >                 },
>> >> >                 {
>> >> >                     "osd": 10,
>> >> >                     "current_lost_at": 0,
>> >> >                     "comment": "starting or marking this osd lost may let
>> >> > us proceed"
>> >> >                 },
>> >> >                 {
>> >> >                     "osd": 37,
>> >> >                     "current_lost_at": 0,
>> >> >                     "comment": "starting or marking this osd lost may let
>> >> > us proceed"
>> >> >                 },
>> >> >                 {
>> >> >                     "osd": 72,
>> >> >                     "current_lost_at": 113771,
>> >> >                     "comment": "starting or marking this osd lost may let
>> >> > us proceed"
>> >> >                 }
>> >> >             ]
>> >> >         },
>> >> 
>> >> > Are any of those OSDs startable?
>> >> 
>> >> They were all up and running - but I decided to shut them down and out
>> >> them  from  ceph, now it looks like ceph working ok, but still two PGs
>> >> are in down state, how to get rid of it ?
>> 
>> > If you haven't deleted the data, you should start the OSDs back up.
>> 
>> > If they are partially damanged you can use ceph-objectstore-tool to 
>> > extract just the PGs in question to make sure you haven't lost anything,
>> > inject them on some other OSD(s) and restart those, and *then* mark the
>> > bad OSDs as 'lost'.
>> 
>> > If all else fails, you can just mark those OSDs 'lost', but in doing so
>> > you might be telling the cluster to lose data.
>> 
>> > The best thing to do is definitely to get those OSDs started again.
>> 
>> Now situation looks like this:
>> 
>> [root@cc1 ~]# rbd info volumes/volume-ccc5d976-cecf-4938-a452-1bee6188987b
>> rbd image 'volume-ccc5d976-cecf-4938-a452-1bee6188987b':
>>         size 500 GB in 128000 objects
>>         order 22 (4096 kB objects)
>>         block_name_prefix: rbd_data.ed9d394a851426
>>         format: 2
>>         features: layering
>>         flags:
>> 
>> [root@cc1 ~]# rados -p volumes ls | grep rbd_data.ed9d394a851426
>> (output cutted)
>> rbd_data.ed9d394a851426.000000000000447c
>> rbd_data.ed9d394a851426.0000000000010857
>> rbd_data.ed9d394a851426.000000000000ec8b
>> rbd_data.ed9d394a851426.000000000000fa43
>> rbd_data.ed9d394a851426.000000000001ef2d
>> ^C
>> 
>> it hangs on this object and isn't going further. rbd cp also hangs...
>> rbd map - also...
>> 
>> can  You advice what can be solution for this case ?

> The hang is due to OSD throttling (see my first reply for how to wrok 
> around that and get a pg query).  But you already did that and the cluster
> told you which OSDs it needs to see up in order for it to peer and 
> recover.  If you haven't destroyed those disks, you should start those
> osds and it shoudl be fine.  If you've destroyed the data or the disks are
> truly broken and dead, then you can mark those OSDs lost and the cluster
> *maybe* recover (but hard to say given the information you've shared).

> sage

What information I can bring to You to say it is recoverable ?

here are ceph -s and ceph health detail:

[root@cc1 ~]# ceph -s
    cluster 8cdfbff9-b7be-46de-85bd-9d49866fcf60
     health HEALTH_WARN
            2 pgs down
            2 pgs peering
            2 pgs stuck inactive
     monmap e1: 3 mons at {cc1=192.168.128.1:6789/0,cc2=192.168.128.2:6789/0,cc3=192.168.128.3:6789/0}
            election epoch 872, quorum 0,1,2 cc1,cc2,cc3
     osdmap e115431: 100 osds: 89 up, 86 in; 1 remapped pgs
      pgmap v67641261: 4032 pgs, 18 pools, 26706 GB data, 4855 kobjects
            76705 GB used, 107 TB / 182 TB avail
                4030 active+clean
                   1 down+remapped+peering
                   1 down+peering
  client io 5704 kB/s rd, 24685 kB/s wr, 49 op/s rd, 165 op/s wr
[root@cc1 ~]# ceph health detail
HEALTH_WARN 2 pgs down; 2 pgs peering; 2 pgs stuck inactive
pg 1.165 is stuck inactive since forever, current state down+peering, last acting [67,88,48]
pg 1.60 is stuck inactive since forever, current state down+remapped+peering, last acting [66,40]
pg 1.60 is down+remapped+peering, acting [66,40]
pg 1.165 is down+peering, acting [67,88,48]
[root@cc1 ~]#

-- 
Regards,
 Łukasz Chrustek

--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux