Hi Cary,
No, I didnt try to repair it.
I am comparatively new in ceph. Is it okay to try to repair it ?
Or should I take any precautions while doing it ?
Karun Josy
On Sat, Dec 16, 2017 at 2:08 PM, Cary <dynamic.cary@xxxxxxxxx> wrote:
Karun,
Did you attempt a "ceph pg repair <pgid>"? Replace <pgid> with the pg
ID that needs repaired, 3.4.
Cary
-D123
> ______________________________
On Sat, Dec 16, 2017 at 8:24 AM, Karun Josy <karunjosy1@xxxxxxxxx> wrote:
> Hello,
>
> I added 1 disk to the cluster and after rebalancing, it shows 1 PG is in
> remapped state. How can I correct it ?
>
> (I had to restart some osds during the rebalancing as there were some slow
> requests)
>
> $ ceph pg dump | grep remapped
> dumped all
> 3.4 981 0 0 0 0 2655009792
> 1535 1535 active+clean+remapped 2017-12-15 22:07:21.663964 2824'785115
> 2824:2297888 [0,2,NONE,4,12,10,5,1] 0 [0,2,1,4,12,10,5,1]
> 0 2288'767367 2017-12-14 11:00:15.576741 417'518549 2017-12-08
> 03:56:14.006982
>
> That PG belongs to an erasure pool with k=5, m =3 profile, failure domain is
> host.
>
> ===========
>
> $ ceph osd tree
> ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF
> -1 16.94565 root default
> -3 2.73788 host ceph-a1
> 0 ssd 1.86469 osd.0 up 1.00000 1.00000
> 14 ssd 0.87320 osd.14 up 1.00000 1.00000
> -5 2.73788 host ceph-a2
> 1 ssd 1.86469 osd.1 up 1.00000 1.00000
> 15 ssd 0.87320 osd.15 up 1.00000 1.00000
> -7 1.86469 host ceph-a3
> 2 ssd 1.86469 osd.2 up 1.00000 1.00000
> -9 1.74640 host ceph-a4
> 3 ssd 0.87320 osd.3 up 1.00000 1.00000
> 4 ssd 0.87320 osd.4 up 1.00000 1.00000
> -11 1.74640 host ceph-a5
> 5 ssd 0.87320 osd.5 up 1.00000 1.00000
> 6 ssd 0.87320 osd.6 up 1.00000 1.00000
> -13 1.74640 host ceph-a6
> 7 ssd 0.87320 osd.7 up 1.00000 1.00000
> 8 ssd 0.87320 osd.8 up 1.00000 1.00000
> -15 1.74640 host ceph-a7
> 9 ssd 0.87320 osd.9 up 1.00000 1.00000
> 10 ssd 0.87320 osd.10 up 1.00000 1.00000
> -17 2.61960 host ceph-a8
> 11 ssd 0.87320 osd.11 up 1.00000 1.00000
> 12 ssd 0.87320 osd.12 up 1.00000 1.00000
> 13 ssd 0.87320 osd.13 up 1.00000 1.00000
>
>
>
> Karun
>
_________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph. com
>
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com