Re: PG active+clean+remapped status

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Maybe try outing the disk that should have a copy of the PG, but doesn't. Then mark it back in. It might check that it has everything properly and pull a copy of the data it's missing. I dunno.


On Sun, Dec 17, 2017, 10:00 PM Karun Josy <karunjosy1@xxxxxxxxx> wrote:
Tried restarting all osds. Still no luck.

Will adding a new disk to any of the server forces a rebalance and fix it?

Karun Josy

On Sun, Dec 17, 2017 at 12:22 PM, Cary <dynamic.cary@xxxxxxxxx> wrote:
Karun,

 Could you paste in the output from "ceph health detail"? Which OSD
was just added?

Cary
-Dynamic

On Sun, Dec 17, 2017 at 4:59 AM, Karun Josy <karunjosy1@xxxxxxxxx> wrote:
> Any help would be appreciated!
>
> Karun Josy
>
> On Sat, Dec 16, 2017 at 11:04 PM, Karun Josy <karunjosy1@xxxxxxxxx> wrote:
>>
>> Hi,
>>
>> Repair didnt fix the issue.
>>
>> In the pg dump details, I notice this None. Seems pg is missing from one
>> of the OSD
>>
>> [0,2,NONE,4,12,10,5,1]
>> [0,2,1,4,12,10,5,1]
>>
>> There is no way Ceph corrects this automatically ? I have to edit/
>> troubleshoot it manually ?
>>
>> Karun
>>
>> On Sat, Dec 16, 2017 at 10:44 PM, Cary <dynamic.cary@xxxxxxxxx> wrote:
>>>
>>> Karun,
>>>
>>>  Running ceph pg repair should not cause any problems. It may not fix
>>> the issue though. If that does not help, there is more information at
>>> the link below.
>>> http://ceph.com/geen-categorie/ceph-manually-repair-object/
>>>
>>> I recommend not rebooting, or restarting while Ceph is repairing or
>>> recovering. If possible, wait until the cluster is in a healthy state
>>> first.
>>>
>>> Cary
>>> -Dynamic
>>>
>>> On Sat, Dec 16, 2017 at 2:05 PM, Karun Josy <karunjosy1@xxxxxxxxx> wrote:
>>> > Hi Cary,
>>> >
>>> > No, I didnt try to repair it.
>>> > I am comparatively new in ceph. Is it okay to try to repair it ?
>>> > Or should I take any precautions while doing it ?
>>> >
>>> > Karun Josy
>>> >
>>> > On Sat, Dec 16, 2017 at 2:08 PM, Cary <dynamic.cary@xxxxxxxxx> wrote:
>>> >>
>>> >> Karun,
>>> >>
>>> >>  Did you attempt a "ceph pg repair <pgid>"? Replace <pgid> with the pg
>>> >> ID that needs repaired, 3.4.
>>> >>
>>> >> Cary
>>> >> -D123
>>> >>
>>> >> On Sat, Dec 16, 2017 at 8:24 AM, Karun Josy <karunjosy1@xxxxxxxxx>
>>> >> wrote:
>>> >> > Hello,
>>> >> >
>>> >> > I added 1 disk to the cluster and after rebalancing, it shows 1 PG
>>> >> > is in
>>> >> > remapped state. How can I correct it ?
>>> >> >
>>> >> > (I had to restart some osds during the rebalancing as there were
>>> >> > some
>>> >> > slow
>>> >> > requests)
>>> >> >
>>> >> > $ ceph pg dump | grep remapped
>>> >> > dumped all
>>> >> > 3.4         981                  0        0         0       0
>>> >> > 2655009792
>>> >> > 1535     1535 active+clean+remapped 2017-12-15 22:07:21.663964
>>> >> > 2824'785115
>>> >> > 2824:2297888 [0,2,NONE,4,12,10,5,1]          0   [0,2,1,4,12,10,5,1]
>>> >> > 0  2288'767367 2017-12-14 11:00:15.576741      417'518549 2017-12-08
>>> >> > 03:56:14.006982
>>> >> >
>>> >> > That PG belongs to an erasure pool with k=5, m =3 profile, failure
>>> >> > domain is
>>> >> > host.
>>> >> >
>>> >> > ===========
>>> >> >
>>> >> > $ ceph osd tree
>>> >> > ID  CLASS WEIGHT   TYPE NAME                STATUS REWEIGHT PRI-AFF
>>> >> >  -1       16.94565 root default
>>> >> >  -3        2.73788     host ceph-a1
>>> >> >   0   ssd  1.86469         osd.0                up  1.00000 1.00000
>>> >> >  14   ssd  0.87320         osd.14               up  1.00000 1.00000
>>> >> >  -5        2.73788     host ceph-a2
>>> >> >   1   ssd  1.86469         osd.1                up  1.00000 1.00000
>>> >> >  15   ssd  0.87320         osd.15               up  1.00000 1.00000
>>> >> >  -7        1.86469     host ceph-a3
>>> >> >   2   ssd  1.86469         osd.2                up  1.00000 1.00000
>>> >> >  -9        1.74640     host ceph-a4
>>> >> >   3   ssd  0.87320         osd.3                up  1.00000 1.00000
>>> >> >   4   ssd  0.87320         osd.4                up  1.00000 1.00000
>>> >> > -11        1.74640     host ceph-a5
>>> >> >   5   ssd  0.87320         osd.5                up  1.00000 1.00000
>>> >> >   6   ssd  0.87320         osd.6                up  1.00000 1.00000
>>> >> > -13        1.74640     host ceph-a6
>>> >> >   7   ssd  0.87320         osd.7                up  1.00000 1.00000
>>> >> >   8   ssd  0.87320         osd.8                up  1.00000 1.00000
>>> >> > -15        1.74640     host ceph-a7
>>> >> >   9   ssd  0.87320         osd.9                up  1.00000 1.00000
>>> >> >  10   ssd  0.87320         osd.10               up  1.00000 1.00000
>>> >> > -17        2.61960     host ceph-a8
>>> >> >  11   ssd  0.87320         osd.11               up  1.00000 1.00000
>>> >> >  12   ssd  0.87320         osd.12               up  1.00000 1.00000
>>> >> >  13   ssd  0.87320         osd.13               up  1.00000 1.00000
>>> >> >
>>> >> >
>>> >> >
>>> >> > Karun
>>> >> >
>>> >> > _______________________________________________
>>> >> > ceph-users mailing list
>>> >> > ceph-users@xxxxxxxxxxxxxx
>>> >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>> >> >
>>> >
>>> >
>>
>>
>

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux