Karun, Did you attempt a "ceph pg repair <pgid>"? Replace <pgid> with the pg ID that needs repaired, 3.4. Cary -D123 On Sat, Dec 16, 2017 at 8:24 AM, Karun Josy <karunjosy1@xxxxxxxxx> wrote: > Hello, > > I added 1 disk to the cluster and after rebalancing, it shows 1 PG is in > remapped state. How can I correct it ? > > (I had to restart some osds during the rebalancing as there were some slow > requests) > > $ ceph pg dump | grep remapped > dumped all > 3.4 981 0 0 0 0 2655009792 > 1535 1535 active+clean+remapped 2017-12-15 22:07:21.663964 2824'785115 > 2824:2297888 [0,2,NONE,4,12,10,5,1] 0 [0,2,1,4,12,10,5,1] > 0 2288'767367 2017-12-14 11:00:15.576741 417'518549 2017-12-08 > 03:56:14.006982 > > That PG belongs to an erasure pool with k=5, m =3 profile, failure domain is > host. > > =========== > > $ ceph osd tree > ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF > -1 16.94565 root default > -3 2.73788 host ceph-a1 > 0 ssd 1.86469 osd.0 up 1.00000 1.00000 > 14 ssd 0.87320 osd.14 up 1.00000 1.00000 > -5 2.73788 host ceph-a2 > 1 ssd 1.86469 osd.1 up 1.00000 1.00000 > 15 ssd 0.87320 osd.15 up 1.00000 1.00000 > -7 1.86469 host ceph-a3 > 2 ssd 1.86469 osd.2 up 1.00000 1.00000 > -9 1.74640 host ceph-a4 > 3 ssd 0.87320 osd.3 up 1.00000 1.00000 > 4 ssd 0.87320 osd.4 up 1.00000 1.00000 > -11 1.74640 host ceph-a5 > 5 ssd 0.87320 osd.5 up 1.00000 1.00000 > 6 ssd 0.87320 osd.6 up 1.00000 1.00000 > -13 1.74640 host ceph-a6 > 7 ssd 0.87320 osd.7 up 1.00000 1.00000 > 8 ssd 0.87320 osd.8 up 1.00000 1.00000 > -15 1.74640 host ceph-a7 > 9 ssd 0.87320 osd.9 up 1.00000 1.00000 > 10 ssd 0.87320 osd.10 up 1.00000 1.00000 > -17 2.61960 host ceph-a8 > 11 ssd 0.87320 osd.11 up 1.00000 1.00000 > 12 ssd 0.87320 osd.12 up 1.00000 1.00000 > 13 ssd 0.87320 osd.13 up 1.00000 1.00000 > > > > Karun > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com