Hello,
I added 1 disk to the cluster and after rebalancing, it shows 1 PG is in remapped state. How can I correct it ?
(I had to restart some osds during the rebalancing as there were some slow requests)
$ ceph pg dump | grep remapped
dumped all
3.4 981 0 0 0 0 2655009792 1535 1535 active+clean+remapped 2017-12-15 22:07:21.663964 2824'785115 2824:2297888 [0,2,NONE,4,12,10,5,1] 0 [0,2,1,4,12,10,5,1] 0 2288'767367 2017-12-14 11:00:15.576741 417'518549 2017-12-08 03:56:14.006982
That PG belongs to an erasure pool with k=5, m =3 profile, failure domain is host.
===========
$ ceph osd tree
ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF
-1 16.94565 root default
-3 2.73788 host ceph-a1
0 ssd 1.86469 osd.0 up 1.00000 1.00000
14 ssd 0.87320 osd.14 up 1.00000 1.00000
-5 2.73788 host ceph-a2
1 ssd 1.86469 osd.1 up 1.00000 1.00000
15 ssd 0.87320 osd.15 up 1.00000 1.00000
-7 1.86469 host ceph-a3
2 ssd 1.86469 osd.2 up 1.00000 1.00000
-9 1.74640 host ceph-a4
3 ssd 0.87320 osd.3 up 1.00000 1.00000
4 ssd 0.87320 osd.4 up 1.00000 1.00000
-11 1.74640 host ceph-a5
5 ssd 0.87320 osd.5 up 1.00000 1.00000
6 ssd 0.87320 osd.6 up 1.00000 1.00000
-13 1.74640 host ceph-a6
7 ssd 0.87320 osd.7 up 1.00000 1.00000
8 ssd 0.87320 osd.8 up 1.00000 1.00000
-15 1.74640 host ceph-a7
9 ssd 0.87320 osd.9 up 1.00000 1.00000
10 ssd 0.87320 osd.10 up 1.00000 1.00000
-17 2.61960 host ceph-a8
11 ssd 0.87320 osd.11 up 1.00000 1.00000
12 ssd 0.87320 osd.12 up 1.00000 1.00000
13 ssd 0.87320 osd.13 up 1.00000 1.00000
Karun
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com