16 osds: 11 up, 16 in

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Craig,

I suspect the disks in question are seeking constantly and the spindle 
contention is causing significant latency. A strategy of throttling 
backfill/recovery and reducing client traffic tends to work for me.

1) You should make sure recovery and backfill are throttled:
ceph tell osd.* injectargs '--osd_max_backfills 1'
ceph tell osd.* injectargs '--osd_recovery_max_active 1'
ceph tell osd.* injectargs '--osd_recovery_op_priority 1'

2) We run a not-particularly critical service with a constant stream of 
95% write/5% read small, random IO. During recovery/backfill, we are 
heavily bound by IOPS. It often times feels like a net win to throttle 
unessential client traffic in an effort to get spindle contention under 
control if Step 1 wasn't enough.

If that all fails, you can try "ceph osd set nodown" which will prevent 
OSDs from being marked down (with or without proper cause), but that 
tends to cause me more trouble than its worth.

Thanks,

Mike Dawson
Co-Founder & Director of Cloud Architecture
Cloudapt LLC
6330 East 75th Street, Suite 170
Indianapolis, IN 46250

On 5/7/2014 1:28 PM, Craig Lewis wrote:
> The 5 OSDs that are down have all been kicked out for being
> unresponsive.  The 5 OSDs are getting kicked faster than they can
> complete the recovery+backfill.  The number of degraded PGs is growing
> over time.
>
> root at ceph0c:~# ceph -w
>      cluster 1604ec7a-6ceb-42fc-8c68-0a7896c4e120
>       health HEALTH_WARN 49 pgs backfill; 926 pgs degraded; 252 pgs
> down; 30 pgs incomplete; 291 pgs peering; 1 pgs recovery_wait; 175 pgs
> stale; 255 pgs stuck inactive; 175 pgs stuck stale; 1234 pgs stuck
> unclean; 66 requests are blocked > 32 sec; recovery 6820014/38055556
> objects degraded (17.921%); 4/16 in osds are down; noout flag(s) set
>       monmap e2: 2 mons at
> {ceph0c=10.193.0.6:6789/0,ceph1c=10.193.0.7:6789/0}, election epoch 238,
> quorum 0,1 ceph0c,ceph1c
>       osdmap e38673: 16 osds: 12 up, 16 in
>              flags noout
>        pgmap v7325233: 2560 pgs, 17 pools, 14090 GB data, 18581 kobjects
>              28456 GB used, 31132 GB / 59588 GB avail
>              6820014/38055556 objects degraded (17.921%)
>                     1 stale+active+clean+scrubbing+deep
>                    15 active
>                  1247 active+clean
>                     1 active+recovery_wait
>                    45 stale+active+clean
>                    39 peering
>                    29 stale+active+degraded+wait_backfill
>                   252 down+peering
>                   827 active+degraded
>                    50 stale+active+degraded
>                    20 stale+active+degraded+remapped+wait_backfill
>                    30 stale+incomplete
>                     4 active+clean+scrubbing+deep
>
> Here's a snippet of ceph.log for one of these OSDs:
> 2014-05-07 09:22:46.747036 mon.0 10.193.0.6:6789/0 39981 : [INF] osd.3
> marked down after no pg stats for 901.212859seconds
> 2014-05-07 09:47:17.930251 mon.0 10.193.0.6:6789/0 40561 : [INF] osd.3
> 10.193.0.6:6812/2830 boot
> 2014-05-07 09:47:16.914519 osd.3 10.193.0.6:6812/2830 823 : [WRN] map
> e38649 wrongly marked me down
>
> root at ceph0c:~# uname -a
> Linux ceph0c 3.5.0-46-generic #70~precise1-Ubuntu SMP Thu Jan 9 23:55:12
> UTC 2014 x86_64 x86_64 x86_64 GNU/Linux
> root at ceph0c:~# lsb_release -a
> No LSB modules are available.
> Distributor ID:    Ubuntu
> Description:    Ubuntu 12.04.4 LTS
> Release:    12.04
> Codename:    precise
> root at ceph0c:~# ceph -v
> ceph version 0.72.2 (a913ded2ff138aefb8cb84d347d72164099cfd60)
>
>
> Any ideas what I can do to make these OSDs stop drying after 15 minutes?
>
>
>
>
> --
>
> *Craig Lewis*
> Senior Systems Engineer
> Office +1.714.602.1309
> Email clewis at centraldesktop.com <mailto:clewis at centraldesktop.com>
>
> *Central Desktop. Work together in ways you never thought possible.*
> Connect with us Website <http://www.centraldesktop.com/>  | Twitter
> <http://www.twitter.com/centraldesktop>  | Facebook
> <http://www.facebook.com/CentralDesktop>  | LinkedIn
> <http://www.linkedin.com/groups?gid=147417>  | Blog
> <http://cdblog.centraldesktop.com/>
>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users at lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux