Re: all oas crush on start

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Yes, I changed original crushmap, need to take nodes gstore1, gstore2,
cstore5 for new cluster. I have only crushmap from the failed cluster
downloaded immediately after cluster was crushed. It on attachment.


2013/7/17 Gregory Farnum <greg@xxxxxxxxxxx>:
> Have you changed either of these maps since you originally switched to
> use rule 3?
>
> Can you compare them to what you have on your test cluster? In
> particular I see that you have 0 weight for all the buckets in the
> crush pool, which I expect to misbehave but not to cause the OSD to
> crash everywhere.
> Software Engineer #42 @ http://inktank.com | http://ceph.com
>
>
> On Tue, Jul 16, 2013 at 4:00 PM, Vladislav Gorbunov <vadikgo@xxxxxxxxx> wrote:
>> output is in the attached files
>>
>> 2013/7/17 Gregory Farnum <greg@xxxxxxxxxxx>:
>>> The maps in the OSDs only would have gotten there from the monitors.
>>> If a bad map somehow got distributed to the OSDs then cleaning it up
>>> is unfortunately going to take a lot of work without any well-defined
>>> processes.
>>> So if you could just do "ceph osd crush dump" and "ceph osd dump" and
>>> provide the output from those commands, we can look at what the map
>>> actually has and go from there.
>>> -Greg
>>> Software Engineer #42 @ http://inktank.com | http://ceph.com
>>>
>>>
>>> On Tue, Jul 16, 2013 at 3:22 PM, Vladislav Gorbunov <vadikgo@xxxxxxxxx> wrote:
>>>> Gregory, thank for you help!
>>>> After all osd servers downed, i'am back rule set for the iscsi pool
>>>> back to default rule 0:
>>>> ceph osd pool set iscsi crush_ruleset 0
>>>> it does not help, all osd not started, except without data, with weight 0.
>>>> next i remove ruleset iscsi from crush map. It does not help too. And
>>>> after that i post crushmap to this mail list.
>>>> Is any method to extract crush map from downed osd server and inject
>>>> it to the mon server? from /var/lib/ceph/osd/ceph-2/current/omap
>>>> folder?
>>>>
>>>> 2013/7/17 Gregory Farnum <greg@xxxxxxxxxxx>:
>>>>> I notice that your first dump of the crush map didn't include rule #3.
>>>>> Are you sure you've injected it into the cluster? Try extracting it
>>>>> from the monitors and looking at that map directly, instead of a
>>>>> locally cached version.
>>>>> You mentioned some problem with OSDs being positioned wrong too, so
>>>>> you might look at "ceph osd tree" and look at the shape of the map.
>>>>> But it sounds to me like maybe there's a disconnect between what
>>>>> you've put into the cluster, and what you're looking at.
>>>>> -Greg
>>>>> Software Engineer #42 @ http://inktank.com | http://ceph.com
# begin crush map

# devices
device 0 osd.0
device 1 osd.1
device 2 osd.2
device 3 osd.3
device 4 osd.4
device 5 osd.5
device 6 osd.6
device 7 osd.7
device 8 osd.8
device 9 osd.9
device 10 osd.10
device 11 osd.11
device 12 osd.12
device 13 osd.13
device 14 osd.14
device 15 osd.15
device 16 osd.16
device 17 osd.17
device 18 osd.18
device 19 osd.19
device 20 osd.20
device 21 osd.21
device 22 osd.22
device 23 osd.23
device 24 osd.24
device 25 osd.25
device 26 osd.26
device 27 osd.27
device 28 osd.28
device 29 osd.29
device 30 osd.30
device 31 osd.31
device 32 osd.32
device 33 osd.33
device 34 osd.34
device 35 osd.35
device 36 osd.36
device 37 osd.37
device 38 osd.38
device 39 osd.39
device 40 osd.40
device 41 osd.41
device 42 osd.42
device 43 osd.43
device 44 osd.44
device 45 osd.45
device 46 osd.46
device 47 osd.47
device 48 osd.48
device 49 osd.49
device 50 osd.50
device 51 osd.51
device 52 osd.52
device 53 osd.53
device 54 osd.54
device 55 osd.55
device 56 osd.56
device 57 osd.57
device 58 osd.58
device 59 osd.59
device 60 osd.60
device 61 osd.61
device 62 osd.62
device 63 osd.63
device 64 osd.64
device 65 osd.65
device 66 osd.66
device 67 osd.67
device 68 osd.68
device 69 osd.69
device 70 osd.70
device 71 osd.71
device 72 osd.72
device 73 osd.73
device 74 osd.74
device 75 osd.75
device 76 osd.76
device 77 osd.77
device 78 osd.78
device 79 osd.79
device 80 osd.80
device 81 osd.81
device 82 osd.82
device 83 osd.83
device 84 osd.84
device 85 osd.85
device 86 osd.86
device 87 osd.87
device 88 osd.88
device 89 osd.89
device 90 osd.90
device 91 osd.91
device 92 osd.92
device 93 osd.93
device 94 osd.94
device 95 osd.95
device 96 osd.96
device 97 osd.97
device 98 osd.98
device 99 osd.99
device 100 osd.100
device 101 osd.101
device 102 osd.102
device 103 osd.103
device 104 osd.104
device 105 osd.105

# types
type 0 osd
type 1 host
type 2 rack
type 3 row
type 4 room
type 5 datacenter
type 6 pool

# buckets
host gstore5 {
	id -9		# do not change unnecessarily
	# weight 5.000
	alg straw
	hash 0	# rjenkins1
	item osd.2 weight 5.000
}
host cstore1 {
	id -10		# do not change unnecessarily
	# weight 16.000
	alg straw
	hash 0	# rjenkins1
	item osd.5 weight 1.000
	item osd.6 weight 1.000
	item osd.7 weight 1.000
	item osd.8 weight 1.000
	item osd.9 weight 1.000
	item osd.10 weight 1.000
	item osd.11 weight 1.000
	item osd.12 weight 1.000
	item osd.13 weight 1.000
	item osd.14 weight 1.000
	item osd.4 weight 1.000
	item osd.47 weight 1.000
	item osd.48 weight 1.000
	item osd.49 weight 1.000
	item osd.50 weight 1.000
	item osd.51 weight 1.000
}
host cstore2 {
	id -11		# do not change unnecessarily
	# weight 20.000
	alg straw
	hash 0	# rjenkins1
	item osd.15 weight 1.000
	item osd.16 weight 1.000
	item osd.17 weight 1.000
	item osd.18 weight 1.000
	item osd.19 weight 1.000
	item osd.20 weight 1.000
	item osd.21 weight 1.000
	item osd.22 weight 1.000
	item osd.23 weight 1.000
	item osd.24 weight 1.000
	item osd.41 weight 1.000
	item osd.42 weight 1.000
	item osd.43 weight 1.000
	item osd.44 weight 1.000
	item osd.45 weight 1.000
	item osd.46 weight 1.000
	item osd.52 weight 1.000
	item osd.53 weight 1.000
	item osd.54 weight 1.000
	item osd.55 weight 1.000
	item osd.56 weight 0.000
	item osd.57 weight 0.000
}
host cstore3 {
	id -12		# do not change unnecessarily
	# weight 16.000
	alg straw
	hash 0	# rjenkins1
	item osd.25 weight 1.000
	item osd.26 weight 1.000
	item osd.27 weight 1.000
	item osd.28 weight 1.000
	item osd.29 weight 1.000
	item osd.30 weight 1.000
	item osd.31 weight 1.000
	item osd.32 weight 1.000
	item osd.33 weight 1.000
	item osd.34 weight 1.000
	item osd.35 weight 1.000
	item osd.36 weight 1.000
	item osd.37 weight 1.000
	item osd.38 weight 1.000
	item osd.39 weight 1.000
	item osd.40 weight 1.000
}
host cstore4 {
	id -13		# do not change unnecessarily
	# weight 7.640
	alg straw
	hash 0	# rjenkins1
	item osd.62 weight 0.550
	item osd.63 weight 0.550
	item osd.64 weight 0.550
	item osd.65 weight 0.550
	item osd.66 weight 0.550
	item osd.67 weight 0.550
	item osd.68 weight 0.550
	item osd.69 weight 0.550
	item osd.70 weight 0.270
	item osd.71 weight 0.270
	item osd.72 weight 0.270
	item osd.73 weight 0.270
	item osd.74 weight 0.270
	item osd.75 weight 0.270
	item osd.76 weight 0.270
	item osd.77 weight 0.270
	item osd.78 weight 0.270
	item osd.79 weight 0.270
	item osd.80 weight 0.270
	item osd.81 weight 0.270
}
host cstore5 {
	id -14		# do not change unnecessarily
	# weight 3.320
	alg straw
	hash 0	# rjenkins1
	item osd.0 weight 0.400
	item osd.1 weight 0.500
	item osd.3 weight 0.400
	item osd.58 weight 0.400
	item osd.59 weight 0.540
	item osd.60 weight 0.540
	item osd.61 weight 0.540
}
rack unknownrack {
	id -3		# do not change unnecessarily
	# weight 67.960
	alg straw
	hash 0	# rjenkins1
	item gstore5 weight 5.000
	item cstore1 weight 16.000
	item cstore2 weight 20.000
	item cstore3 weight 16.000
	item cstore4 weight 7.640
	item cstore5 weight 3.320
}
pool default {
	id -1		# do not change unnecessarily
	# weight 67.960
	alg straw
	hash 0	# rjenkins1
	item unknownrack weight 67.960
}
host gstore1 {
	id -6		# do not change unnecessarily
	# weight 32.760
	alg straw
	hash 0	# rjenkins1
	item osd.82 weight 2.730
	item osd.83 weight 2.730
	item osd.84 weight 2.730
	item osd.85 weight 2.730
	item osd.86 weight 2.730
	item osd.87 weight 2.730
	item osd.88 weight 2.730
	item osd.89 weight 2.730
	item osd.90 weight 2.730
	item osd.91 weight 2.730
	item osd.92 weight 2.730
	item osd.93 weight 2.730
}
datacenter datacenter-cod {
	id -4		# do not change unnecessarily
	# weight 32.760
	alg straw
	hash 0	# rjenkins1
	item gstore1 weight 32.760
}
host gstore2 {
	id -7		# do not change unnecessarily
	# weight 32.760
	alg straw
	hash 0	# rjenkins1
	item osd.94 weight 2.730
	item osd.95 weight 2.730
	item osd.96 weight 2.730
	item osd.97 weight 2.730
	item osd.98 weight 2.730
	item osd.99 weight 2.730
	item osd.100 weight 2.730
	item osd.101 weight 2.730
	item osd.102 weight 2.730
	item osd.103 weight 2.730
	item osd.104 weight 2.730
	item osd.105 weight 2.730
}
datacenter datacenter-rcod {
	id -5		# do not change unnecessarily
	# weight 32.760
	alg straw
	hash 0	# rjenkins1
	item gstore2 weight 32.760
}
pool iscsi {
	id -2		# do not change unnecessarily
	# weight 65.520
	alg straw
	hash 0	# rjenkins1
	item datacenter-cod weight 32.760
	item datacenter-rcod weight 32.760
}

# rules
rule data {
	ruleset 0
	type replicated
	min_size 1
	max_size 10
	step take default
	step chooseleaf firstn 0 type host
	step emit
}
rule metadata {
	ruleset 1
	type replicated
	min_size 1
	max_size 10
	step take default
	step chooseleaf firstn 0 type host
	step emit
}
rule rbd {
	ruleset 2
	type replicated
	min_size 1
	max_size 10
	step take default
	step chooseleaf firstn 0 type host
	step emit
}
rule iscsi {
	ruleset 3
	type replicated
	min_size 1
	max_size 10
	step take iscsi
	step chooseleaf firstn 0 type datacenter
	step chooseleaf firstn 0 type host
	step emit
}

# end crush map
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux