After add a new ceph-osd, seems the ceph is back to normal. But there's still a warning message in ceph health detail. During the previous three months, there's a OSD node restart very often due to power supply problem. So I guess maybe it is related to this. But I not quite sure how to fix it, Please help.
HEALTH_WARN 2 pgs backfilling; 2 pgs degraded; 2 pgs stuck unclean; recovery 741/129078 objects degraded (0.574%)
pg 3.1b is stuck unclean for 7081304.433870, current state active+degraded+remapped+backfilling, last acting [0,2]
pg 3.22 is stuck unclean for 7079410.875934, current state active+degraded+remapped+backfilling, last acting [0,2]
pg 3.22 is active+degraded+remapped+backfilling, acting [0,2]
pg 3.1b is active+degraded+remapped+backfilling, acting [0,2]
recovery 741/129078 objects degraded (0.574%)
[root@node-5e40 mail]# ceph pg dump_stuck unclean
ok
pg_stat objects mip degr unf bytes log disklog state state_stamp v reported up up_primary acting acting_primary last_scrub scrub_stamp last_deep_scrub deep_scrub_stamp
3.1b 343 0 401 0 2233819648 10001 10001 active+degraded+remapped+backfilling 2015-04-29 07:04:35.319063 2963'315257 2963:619403 [1,2,0] 1 [0,2] 0 2261'133643 2015-02-06 02:36:17.834174 2261'133007 2015-02-04 02:36:14.332644
3.22 319 0 368 0 2111692288 10001 10001 active+degraded+remapped+backfilling 2015-04-29 07:05:00.954251 2963'320956 2963:414625 [1,0,2] 1 [0,2] 0 2261'176758 2015-02-06 02:09:41.336774 2261'174335 2015-02-04 02:04:40.745000
"last_active": "2015-02-06 09:43:11.587526",
"last_clean": "2015-02-06 09:43:11.587526",
Best Regards
-- Ray
-- Ray
On Tue, Apr 28, 2015 at 10:59 PM, Ray Sun <xiaoquqi@xxxxxxxxx> wrote:
Sébastien,Thanks for your answer, I am a fan of your blog, it really help me a lot.I found there're two ways to do that:The first one is use command line, but after I triedceph pg set_full_ratio 0.98Seems it not worked.Then I tried to modify the ceph.conf and addmon osd full ratio = 0.98But I think this should restart the service, but I really worried about if the service can restart successfully.Please help.Best Regards
-- RayOn Tue, Apr 28, 2015 at 10:54 PM, Sebastien Han <sebastien.han@xxxxxxxxxxxx> wrote:You can try to push the full ratio a bit further and then delete some objects.
> _______________________________________________
> On 28 Apr 2015, at 15:51, Ray Sun <xiaoquqi@xxxxxxxxx> wrote:
>
> More detail about ceph health detail
> [root@controller ~]# ceph health detail
> HEALTH_ERR 20 pgs backfill_toofull; 20 pgs degraded; 20 pgs stuck unclean; recovery 7482/129081 objects degraded (5.796%); 2 full osd(s); 1 near full osd(s)
> pg 3.8 is stuck unclean for 7067109.597691, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.7d is stuck unclean for 1852078.505139, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.21 is stuck unclean for 7072842.637848, current state active+degraded+remapped+backfill_toofull, last acting [0,2]
> pg 3.22 is stuck unclean for 7070880.213397, current state active+degraded+remapped+backfill_toofull, last acting [0,2]
> pg 3.a is stuck unclean for 7067057.863562, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.7f is stuck unclean for 7067122.493746, current state active+degraded+remapped+backfill_toofull, last acting [0,2]
> pg 3.5 is stuck unclean for 7067088.369629, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.1e is stuck unclean for 7073386.246281, current state active+degraded+remapped+backfill_toofull, last acting [0,2]
> pg 3.19 is stuck unclean for 7068035.310269, current state active+degraded+remapped+backfill_toofull, last acting [0,2]
> pg 3.5d is stuck unclean for 1852078.505949, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.1a is stuck unclean for 7067088.429544, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.1b is stuck unclean for 7072773.771385, current state active+degraded+remapped+backfill_toofull, last acting [0,2]
> pg 3.3 is stuck unclean for 7067057.864514, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.15 is stuck unclean for 7067088.825483, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.11 is stuck unclean for 7067057.862408, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.6d is stuck unclean for 7067083.634454, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.6e is stuck unclean for 7067098.452576, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.c is stuck unclean for 5658116.678331, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.e is stuck unclean for 7067078.646953, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.20 is stuck unclean for 7067140.530849, current state active+degraded+remapped+backfill_toofull, last acting [0,2]
> pg 3.7d is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.7f is active+degraded+remapped+backfill_toofull, acting [0,2]
> pg 3.6d is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.6e is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.5d is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.20 is active+degraded+remapped+backfill_toofull, acting [0,2]
> pg 3.21 is active+degraded+remapped+backfill_toofull, acting [0,2]
> pg 3.22 is active+degraded+remapped+backfill_toofull, acting [0,2]
> pg 3.1e is active+degraded+remapped+backfill_toofull, acting [0,2]
> pg 3.19 is active+degraded+remapped+backfill_toofull, acting [0,2]
> pg 3.1a is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.1b is active+degraded+remapped+backfill_toofull, acting [0,2]
> pg 3.15 is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.11 is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.c is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.e is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.8 is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.a is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.5 is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.3 is active+degraded+remapped+backfill_toofull, acting [2,0]
> recovery 7482/129081 objects degraded (5.796%)
> osd.0 is full at 95%
> osd.2 is full at 95%
> osd.1 is near full at 93%
>
> Best Regards
> -- Ray
>
> On Tue, Apr 28, 2015 at 9:43 PM, Ray Sun <xiaoquqi@xxxxxxxxx> wrote:
> Emergency Help!
>
> One of ceph cluster is full, and ceph -s returns:
> [root@controller ~]# ceph -s
> cluster 059f27e8-a23f-4587-9033-3e3679d03b31
> health HEALTH_ERR 20 pgs backfill_toofull; 20 pgs degraded; 20 pgs stuck unclean; recovery 7482/129081 objects degraded (5.796%); 2 full osd(s); 1 near full osd(s)
> monmap e6: 4 mons at {node-5e40.cloud.com=10.10.20.40:6789/0,node-6670.cloud.com=10.10.20.31:6789/0,node-66c4.cloud.com=10.10.20.36:6789/0,node-fb27.cloud.com=10.10.20.41:6789/0}, election epoch 886, quorum 0,1,2,3 node-6670.cloud.com,node-66c4.cloud.com,node-5e40.cloud.com,node-fb27.cloud.com
> osdmap e2743: 3 osds: 3 up, 3 in
> flags full
> pgmap v6564199: 320 pgs, 4 pools, 262 GB data, 43027 objects
> 786 GB used, 47785 MB / 833 GB avail
> 7482/129081 objects degraded (5.796%)
> 300 active+clean
> 20 active+degraded+remapped+backfill_toofull
>
> Then I try to remove some volume, and I got:
> [root@controller ~]# rbd -p volumes rm volume-c55fd052-212d-4107-a2ac-cf53bfc049be
> 2015-04-29 05:31:31.719478 7f5fb82f7760 0 client.4781741.objecter FULL, paused modify 0xe9a9e0 tid 6
>
> Please give me some tip for this, Thanks a lot.
>
>
> Best Regards
> -- Ray
>
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Cheers.
––––
Sébastien Han
Cloud Architect
"Always give 100%. Unless you're giving blood."
Phone: +33 (0)1 49 70 99 72
Mail: sebastien.han@xxxxxxxxxxxx
Address : 11 bis, rue Roquépine - 75008 Paris
Web : www.enovance.com - Twitter : @enovance
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com