Re: Ceph is Full

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



After add a new ceph-osd, seems the ceph is back to normal. But there's still a warning message in ceph health detail. During the previous three months, there's a OSD node restart very often due to power supply problem. So I guess maybe it is related to this. But I not quite sure how to fix it, Please help.

HEALTH_WARN 2 pgs backfilling; 2 pgs degraded; 2 pgs stuck unclean; recovery 741/129078 objects degraded (0.574%)
pg 3.1b is stuck unclean for 7081304.433870, current state active+degraded+remapped+backfilling, last acting [0,2]
pg 3.22 is stuck unclean for 7079410.875934, current state active+degraded+remapped+backfilling, last acting [0,2]
pg 3.22 is active+degraded+remapped+backfilling, acting [0,2]
pg 3.1b is active+degraded+remapped+backfilling, acting [0,2]
recovery 741/129078 objects degraded (0.574%)

[root@node-5e40 mail]# ceph pg dump_stuck unclean
ok
pg_stat objects mip     degr    unf     bytes   log     disklog state   state_stamp     v       reported        up      up_primary      acting  acting_primary  last_scrub   scrub_stamp     last_deep_scrub deep_scrub_stamp
3.1b    343     0       401     0       2233819648      10001   10001   active+degraded+remapped+backfilling    2015-04-29 07:04:35.319063      2963'315257     2963:619403  [1,2,0] 1       [0,2]   0       2261'133643     2015-02-06 02:36:17.834174      2261'133007     2015-02-04 02:36:14.332644
3.22    319     0       368     0       2111692288      10001   10001   active+degraded+remapped+backfilling    2015-04-29 07:05:00.954251      2963'320956     2963:414625  [1,0,2] 1       [0,2]   0       2261'176758     2015-02-06 02:09:41.336774      2261'174335     2015-02-04 02:04:40.745000
"last_active": "2015-02-06 09:43:11.587526",
              "last_clean": "2015-02-06 09:43:11.587526",

Best Regards
-- Ray

On Tue, Apr 28, 2015 at 10:59 PM, Ray Sun <xiaoquqi@xxxxxxxxx> wrote:
Sébastien
​,
​Thanks for your answer, I am a fan of your blog, it really help me a lot.

​I found there're two ways to do that:

The first one is use command line, but after I tried
ceph pg set_full_ratio 0.98​
Seems it not worked.

Then I tried to modify the ceph.conf and add
mon osd full ratio = 0.98
But I think this should restart the service, but I really worried about if the service can restart successfully.

Please help.


Best Regards
-- Ray

On Tue, Apr 28, 2015 at 10:54 PM, Sebastien Han <sebastien.han@xxxxxxxxxxxx> wrote:
You can try to push the full ratio a bit further and then delete some objects.

> On 28 Apr 2015, at 15:51, Ray Sun <xiaoquqi@xxxxxxxxx> wrote:
>
> More detail about ceph health detail
> [root@controller ~]# ceph health detail
> HEALTH_ERR 20 pgs backfill_toofull; 20 pgs degraded; 20 pgs stuck unclean; recovery 7482/129081 objects degraded (5.796%); 2 full osd(s); 1 near full osd(s)
> pg 3.8 is stuck unclean for 7067109.597691, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.7d is stuck unclean for 1852078.505139, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.21 is stuck unclean for 7072842.637848, current state active+degraded+remapped+backfill_toofull, last acting [0,2]
> pg 3.22 is stuck unclean for 7070880.213397, current state active+degraded+remapped+backfill_toofull, last acting [0,2]
> pg 3.a is stuck unclean for 7067057.863562, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.7f is stuck unclean for 7067122.493746, current state active+degraded+remapped+backfill_toofull, last acting [0,2]
> pg 3.5 is stuck unclean for 7067088.369629, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.1e is stuck unclean for 7073386.246281, current state active+degraded+remapped+backfill_toofull, last acting [0,2]
> pg 3.19 is stuck unclean for 7068035.310269, current state active+degraded+remapped+backfill_toofull, last acting [0,2]
> pg 3.5d is stuck unclean for 1852078.505949, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.1a is stuck unclean for 7067088.429544, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.1b is stuck unclean for 7072773.771385, current state active+degraded+remapped+backfill_toofull, last acting [0,2]
> pg 3.3 is stuck unclean for 7067057.864514, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.15 is stuck unclean for 7067088.825483, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.11 is stuck unclean for 7067057.862408, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.6d is stuck unclean for 7067083.634454, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.6e is stuck unclean for 7067098.452576, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.c is stuck unclean for 5658116.678331, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.e is stuck unclean for 7067078.646953, current state active+degraded+remapped+backfill_toofull, last acting [2,0]
> pg 3.20 is stuck unclean for 7067140.530849, current state active+degraded+remapped+backfill_toofull, last acting [0,2]
> pg 3.7d is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.7f is active+degraded+remapped+backfill_toofull, acting [0,2]
> pg 3.6d is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.6e is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.5d is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.20 is active+degraded+remapped+backfill_toofull, acting [0,2]
> pg 3.21 is active+degraded+remapped+backfill_toofull, acting [0,2]
> pg 3.22 is active+degraded+remapped+backfill_toofull, acting [0,2]
> pg 3.1e is active+degraded+remapped+backfill_toofull, acting [0,2]
> pg 3.19 is active+degraded+remapped+backfill_toofull, acting [0,2]
> pg 3.1a is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.1b is active+degraded+remapped+backfill_toofull, acting [0,2]
> pg 3.15 is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.11 is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.c is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.e is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.8 is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.a is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.5 is active+degraded+remapped+backfill_toofull, acting [2,0]
> pg 3.3 is active+degraded+remapped+backfill_toofull, acting [2,0]
> recovery 7482/129081 objects degraded (5.796%)
> osd.0 is full at 95%
> osd.2 is full at 95%
> osd.1 is near full at 93%
>
> Best Regards
> -- Ray
>
> On Tue, Apr 28, 2015 at 9:43 PM, Ray Sun <xiaoquqi@xxxxxxxxx> wrote:
> Emergency Help!
>
> One of ceph cluster is full, and ceph -s returns:
> [root@controller ~]# ceph -s
>     cluster 059f27e8-a23f-4587-9033-3e3679d03b31
>      health HEALTH_ERR 20 pgs backfill_toofull; 20 pgs degraded; 20 pgs stuck unclean; recovery 7482/129081 objects degraded (5.796%); 2 full osd(s); 1 near full osd(s)
>      monmap e6: 4 mons at {node-5e40.cloud.com=10.10.20.40:6789/0,node-6670.cloud.com=10.10.20.31:6789/0,node-66c4.cloud.com=10.10.20.36:6789/0,node-fb27.cloud.com=10.10.20.41:6789/0}, election epoch 886, quorum 0,1,2,3 node-6670.cloud.com,node-66c4.cloud.com,node-5e40.cloud.com,node-fb27.cloud.com
>      osdmap e2743: 3 osds: 3 up, 3 in
>             flags full
>       pgmap v6564199: 320 pgs, 4 pools, 262 GB data, 43027 objects
>             786 GB used, 47785 MB / 833 GB avail
>             7482/129081 objects degraded (5.796%)
>                  300 active+clean
>                   20 active+degraded+remapped+backfill_toofull
>
> Then I try to remove some volume, and I got:
> [root@controller ~]# rbd -p volumes rm volume-c55fd052-212d-4107-a2ac-cf53bfc049be
> 2015-04-29 05:31:31.719478 7f5fb82f7760  0 client.4781741.objecter  FULL, paused modify 0xe9a9e0 tid 6
>
> Please give me some tip for this, Thanks a lot.
>
>
> Best Regards
> -- Ray
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Cheers.
––––
Sébastien Han
Cloud Architect

"Always give 100%. Unless you're giving blood."

Phone: +33 (0)1 49 70 99 72
Mail: sebastien.han@xxxxxxxxxxxx
Address : 11 bis, rue Roquépine - 75008 Paris
Web : www.enovance.com - Twitter : @enovance



_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux