Martin, Greg is talking about noout. With Ceph, you can specifically preclude OSDs from being marked out when down to prevent rebalancing--e.g., during upgrades, short-term maintenance, etc. http://ceph.com/docs/master/rados/operations/troubleshooting-osd/#stopping-w-out-rebalancing On Thu, Mar 28, 2013 at 11:12 AM, Martin Mailand <martin@xxxxxxxxxxxx> wrote: > Hi Greg, > > setting the osd manually out triggered the recovery. > But now it is the question, why is the osd not marked out after 300 > seconds? That's a default cluster, I use the 0.59 build from your site. > And I didn't change any value, except for the crushmap. > > That's my ceph.conf. > > -martin > > [global] > auth cluster requierd = none > auth service required = none > auth client required = none > # log file = "" > log_max_recent=100 > log_max_new=100 > > [mon] > mon data = /data/mon.$id > [mon.a] > host = store1 > mon addr = 192.168.195.31:6789 > [mon.b] > host = store3 > mon addr = 192.168.195.33:6789 > [mon.c] > host = store5 > mon addr = 192.168.195.35:6789 > [osd] > journal aio = true > osd data = /data/osd.$id > osd mount options btrfs = rw,noatime,nodiratime,autodefrag > osd mkfs options btrfs = -n 32k -l 32k > > [osd.0] > host = store1 > osd journal = /dev/sdg1 > btrfs devs = /dev/sdc > [osd.1] > host = store1 > osd journal = /dev/sdh1 > btrfs devs = /dev/sdd > [osd.2] > host = store1 > osd journal = /dev/sdi1 > btrfs devs = /dev/sde > [osd.3] > host = store1 > osd journal = /dev/sdj1 > btrfs devs = /dev/sdf > [osd.4] > host = store2 > osd journal = /dev/sdg1 > btrfs devs = /dev/sdc > [osd.5] > host = store2 > osd journal = /dev/sdh1 > btrfs devs = /dev/sdd > [osd.6] > host = store2 > osd journal = /dev/sdi1 > btrfs devs = /dev/sde > [osd.7] > host = store2 > osd journal = /dev/sdj1 > btrfs devs = /dev/sdf > [osd.8] > host = store3 > osd journal = /dev/sdg1 > btrfs devs = /dev/sdc > [osd.9] > host = store3 > osd journal = /dev/sdh1 > btrfs devs = /dev/sdd > [osd.10] > host = store3 > osd journal = /dev/sdi1 > btrfs devs = /dev/sde > [osd.11] > host = store3 > osd journal = /dev/sdj1 > btrfs devs = /dev/sdf > [osd.12] > host = store4 > osd journal = /dev/sdg1 > btrfs devs = /dev/sdc > [osd.13] > host = store4 > osd journal = /dev/sdh1 > btrfs devs = /dev/sdd > [osd.14] > host = store4 > osd journal = /dev/sdi1 > btrfs devs = /dev/sde > [osd.15] > host = store4 > osd journal = /dev/sdj1 > btrfs devs = /dev/sdf > [osd.16] > host = store5 > osd journal = /dev/sdg1 > btrfs devs = /dev/sdc > [osd.17] > host = store5 > osd journal = /dev/sdh1 > btrfs devs = /dev/sdd > [osd.18] > host = store5 > osd journal = /dev/sdi1 > btrfs devs = /dev/sde > [osd.19] > host = store5 > osd journal = /dev/sdj1 > btrfs devs = /dev/sdf > [osd.20] > host = store6 > osd journal = /dev/sdg1 > btrfs devs = /dev/sdc > [osd.21] > host = store6 > osd journal = /dev/sdh1 > btrfs devs = /dev/sdd > [osd.22] > host = store6 > osd journal = /dev/sdi1 > btrfs devs = /dev/sde > [osd.23] > host = store6 > osd journal = /dev/sdj1 > btrfs devs = /dev/sdf > > > On 28.03.2013 19:01, Gregory Farnum wrote: >> Your crush map looks fine to me. I'm saying that your ceph -s output >> showed the OSD still hadn't been marked out. No data will be migrated >> until it's marked out. >> After ten minutes it should have been marked out, but that's based on >> a number of factors you have some control over. If you just want a >> quick check of your crush map you can mark it out manually, too. >> -Greg > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- John Wilkins Senior Technical Writer Intank john.wilkins@xxxxxxxxxxx (415) 425-9599 http://inktank.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com