Karun, Running ceph pg repair should not cause any problems. It may not fix the issue though. If that does not help, there is more information at the link below. http://ceph.com/geen-categorie/ceph-manually-repair-object/ I recommend not rebooting, or restarting while Ceph is repairing or recovering. If possible, wait until the cluster is in a healthy state first. Cary -Dynamic On Sat, Dec 16, 2017 at 2:05 PM, Karun Josy <karunjosy1@xxxxxxxxx> wrote: > Hi Cary, > > No, I didnt try to repair it. > I am comparatively new in ceph. Is it okay to try to repair it ? > Or should I take any precautions while doing it ? > > Karun Josy > > On Sat, Dec 16, 2017 at 2:08 PM, Cary <dynamic.cary@xxxxxxxxx> wrote: >> >> Karun, >> >> Did you attempt a "ceph pg repair <pgid>"? Replace <pgid> with the pg >> ID that needs repaired, 3.4. >> >> Cary >> -D123 >> >> On Sat, Dec 16, 2017 at 8:24 AM, Karun Josy <karunjosy1@xxxxxxxxx> wrote: >> > Hello, >> > >> > I added 1 disk to the cluster and after rebalancing, it shows 1 PG is in >> > remapped state. How can I correct it ? >> > >> > (I had to restart some osds during the rebalancing as there were some >> > slow >> > requests) >> > >> > $ ceph pg dump | grep remapped >> > dumped all >> > 3.4 981 0 0 0 0 2655009792 >> > 1535 1535 active+clean+remapped 2017-12-15 22:07:21.663964 >> > 2824'785115 >> > 2824:2297888 [0,2,NONE,4,12,10,5,1] 0 [0,2,1,4,12,10,5,1] >> > 0 2288'767367 2017-12-14 11:00:15.576741 417'518549 2017-12-08 >> > 03:56:14.006982 >> > >> > That PG belongs to an erasure pool with k=5, m =3 profile, failure >> > domain is >> > host. >> > >> > =========== >> > >> > $ ceph osd tree >> > ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF >> > -1 16.94565 root default >> > -3 2.73788 host ceph-a1 >> > 0 ssd 1.86469 osd.0 up 1.00000 1.00000 >> > 14 ssd 0.87320 osd.14 up 1.00000 1.00000 >> > -5 2.73788 host ceph-a2 >> > 1 ssd 1.86469 osd.1 up 1.00000 1.00000 >> > 15 ssd 0.87320 osd.15 up 1.00000 1.00000 >> > -7 1.86469 host ceph-a3 >> > 2 ssd 1.86469 osd.2 up 1.00000 1.00000 >> > -9 1.74640 host ceph-a4 >> > 3 ssd 0.87320 osd.3 up 1.00000 1.00000 >> > 4 ssd 0.87320 osd.4 up 1.00000 1.00000 >> > -11 1.74640 host ceph-a5 >> > 5 ssd 0.87320 osd.5 up 1.00000 1.00000 >> > 6 ssd 0.87320 osd.6 up 1.00000 1.00000 >> > -13 1.74640 host ceph-a6 >> > 7 ssd 0.87320 osd.7 up 1.00000 1.00000 >> > 8 ssd 0.87320 osd.8 up 1.00000 1.00000 >> > -15 1.74640 host ceph-a7 >> > 9 ssd 0.87320 osd.9 up 1.00000 1.00000 >> > 10 ssd 0.87320 osd.10 up 1.00000 1.00000 >> > -17 2.61960 host ceph-a8 >> > 11 ssd 0.87320 osd.11 up 1.00000 1.00000 >> > 12 ssd 0.87320 osd.12 up 1.00000 1.00000 >> > 13 ssd 0.87320 osd.13 up 1.00000 1.00000 >> > >> > >> > >> > Karun >> > >> > _______________________________________________ >> > ceph-users mailing list >> > ceph-users@xxxxxxxxxxxxxx >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > > > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com