If you are using the default configuration to create the pool (3 replicas), after losing 1 OSD and having 2 left, CRUSH would not be able to find enough OSDs (at least 3) to map the PG thus it would stuck at unclean. Thanks, Guang ---------------------------------------- > From: chmind@xxxxxxxxx > Date: Wed, 12 Aug 2015 19:46:01 +0300 > To: ceph-users@xxxxxxxxxxxxxx > Subject: osd out > > Hello. > Could you please help me to remove osd from cluster; > > # ceph osd tree > ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY > -1 0.02998 root default > -2 0.00999 host ceph1 > 0 0.00999 osd.0 up 1.00000 1.00000 > -3 0.00999 host ceph2 > 1 0.00999 osd.1 up 1.00000 1.00000 > -4 0.00999 host ceph3 > 2 0.00999 osd.2 up 1.00000 1.00000 > > > # ceph -s > cluster 64f87255-d56e-499d-8ebc-65e0f577e0aa > health HEALTH_OK > monmap e1: 3 mons at {ceph1=10.0.0.101:6789/0,ceph2=10.0.0.102:6789/0,ceph3=10.0.0.103:6789/0} > election epoch 10, quorum 0,1,2 ceph1,ceph2,ceph3 > osdmap e76: 3 osds: 3 up, 3 in > pgmap v328: 128 pgs, 1 pools, 10 bytes data, 1 objects > 120 MB used, 45926 MB / 46046 MB avail > 128 active+clean > > > # ceph osd out 0 > marked out osd.0. > > # ceph -w > cluster 64f87255-d56e-499d-8ebc-65e0f577e0aa > health HEALTH_WARN > 128 pgs stuck unclean > recovery 1/3 objects misplaced (33.333%) > monmap e1: 3 mons at {ceph1=10.0.0.101:6789/0,ceph2=10.0.0.102:6789/0,ceph3=10.0.0.103:6789/0} > election epoch 10, quorum 0,1,2 ceph1,ceph2,ceph3 > osdmap e79: 3 osds: 3 up, 2 in; 128 remapped pgs > pgmap v332: 128 pgs, 1 pools, 10 bytes data, 1 objects > 89120 kB used, 30610 MB / 30697 MB avail > 1/3 objects misplaced (33.333%) > 128 active+remapped > > 2015-08-12 18:43:12.412286 mon.0 [INF] pgmap v332: 128 pgs: 128 active+remapped; 10 bytes data, 89120 kB used, 30610 MB / 30697 MB avail; 1/3 objects misplaced (33.333%) > 2015-08-12 18:43:20.362337 mon.0 [INF] HEALTH_WARN; 128 pgs stuck unclean; recovery 1/3 objects misplaced (33.333%) > 2015-08-12 18:44:15.055825 mon.0 [INF] pgmap v333: 128 pgs: 128 active+remapped; 10 bytes data, 89120 kB used, 30610 MB / 30697 MB avail; 1/3 objects misplaced (33.333%) > > > and it never become active+clean . > What I’m doing wrong ? > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com