further crush map questions

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



We observe strange behavior with some configurations. PGs stays in degraded state after
a single OSD failure. 

I can also show the behavior using crushtool  with the following map:

----------crush map---------
# begin crush map
tunable choose_local_tries 0
tunable choose_local_fallback_tries 0
tunable choose_total_tries 50
tunable chooseleaf_descend_once 1

# devices
device 0 osd.0
device 1 osd.1
device 2 osd.2
device 3 osd.3
device 4 osd.4
device 5 osd.5
device 6 osd.6
device 7 osd.7
device 8 osd.8
device 9 osd.9
device 10 osd.10
device 11 osd.11

# types
type 0 osd
type 1 host
type 2 rack
type 3 row
type 4 room
type 5 datacenter
type 6 root

# buckets
host prox-ceph-1 {
	id -2		# do not change unnecessarily
	alg straw
	hash 0	# rjenkins1
	item osd.0 weight 1
	item osd.1 weight 1
	item osd.2 weight 1
	item osd.3 weight 1
}
host prox-ceph-2 {
	id -3		# do not change unnecessarily
	# weight 7.260
	alg straw
	hash 0	# rjenkins1
	item osd.4 weight 1
	item osd.5 weight 1
	item osd.6 weight 1
	item osd.7 weight 1
}
host prox-ceph-3 {
	id -4		# do not change unnecessarily
	alg straw
	hash 0	# rjenkins1
	item osd.8 weight 1
	item osd.9 weight 1
	item osd.10 weight 1
	item osd.11 weight 1
}

root default {
	id -1		# do not change unnecessarily
	# weight 21.780
	alg straw
	hash 0	# rjenkins1
	item prox-ceph-1 weight 4
	item prox-ceph-2 weight 4
	item prox-ceph-3 weight 4
}

# rules
rule data {
	ruleset 0
	type replicated
	min_size 1
	max_size 10
	step take default
	step chooseleaf firstn 0 type host
	step emit
}

# end crush map
--------------------------------------

After compiling that map to ''crush-test.map' we run:

# crushtool --test -i 'crush-test.map' --rule 0 --num-rep 3 --weight 11 0 --show-statistics

I set '--weight 11 0' to mark osd.11 as 'out'. The result is:

...
rule 0 (data) num_rep 3 result size == 2:	111/1024
rule 0 (data) num_rep 3 result size == 3:	913/1024

so 111 PG end up in degraded state. I would expect that the data gets 
re-distributed to the remaining OSDs instead.

Can someone explain why that happens?



_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux