Re: all oas crush on start

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



ruleset 3 is:
rule iscsi {
        ruleset 3
        type replicated
        min_size 1
        max_size 10
        step take iscsi
        step chooseleaf firstn 0 type datacenter
        step chooseleaf firstn 0 type host
        step emit
}

2013/7/16 Vladislav Gorbunov <vadikgo@xxxxxxxxx>:
> sorry, after i'm try to apply crush ruleset 3 (iscsi) to
> pool iscsi:
> ceph osd pool set iscsi crush_ruleset 3
>
> 2013/7/16 Vladislav Gorbunov <vadikgo@xxxxxxxxx>:
>>>Have you run this crush map through any test mappings yet?
>> Yes, it worked on test cluster, and after apply map to main cluster.
>> OSD servers downed after i'm try to apply crush ruleset 3 (iscsi) to
>> pool iscsi:
>> ceph osd pool set data crush_ruleset 3
>>
>> 2013/7/16 Gregory Farnum <greg@xxxxxxxxxxx>:
>>> It's probably not the same issue as that ticket, which was about the
>>> OSD handling a lack of output incorrectly. (It might be handling the
>>> output incorrectly in some other way, but hopefully not...)
>>>
>>> Have you run this crush map through any test mappings yet?
>>> -Greg
>>> Software Engineer #42 @ http://inktank.com | http://ceph.com
>>>
>>>
>>> On Sun, Jul 14, 2013 at 10:59 PM, Vladislav Gorbunov <vadikgo@xxxxxxxxx> wrote:
>>>> Sympthoms like on http://tracker.ceph.com/issues/4699
>>>>
>>>> all OSDs the process ceph-osd crash with segfault
>>>>
>>>> If i stop MONs daemons then i can start OSDs but if i start MONs back
>>>> then die all OSDs again.
>>>>
>>>> more detailed log:
>>>>      0> 2013-07-15 16:42:05.001242 7ffe5a6fc700 -1 *** Caught signal
>>>> (Segmentation fault) **
>>>>  in thread 7ffe5a6fc700
>>>>
>>>>  ceph version 0.61.4 (1669132fcfc27d0c0b5e5bb93ade59d147e23404)
>>>>  1: /usr/bin/ceph-osd() [0x790e5a]
>>>>  2: (()+0xfcb0) [0x7ffe6b729cb0]
>>>>  3: /usr/bin/ceph-osd() [0x893879]
>>>>  4: (crush_do_rule()+0x1e5) [0x894065]
>>>>  5: (CrushWrapper::do_rule(int, int, std::vector<int,
>>>> std::allocator<int> >&, int, std::vector<unsigned int,
>>>> std::allocator<unsigned int> > const&) const+0x7a) [0x81b2ba]
>>>>  6: (OSDMap::_pg_to_osds(pg_pool_t const&, pg_t, std::vector<int,
>>>> std::allocator<int> >&) const+0x8f) [0x80d7cf]
>>>>  7: (OSDMap::pg_to_up_acting_osds(pg_t, std::vector<int,
>>>> std::allocator<int> >&, std::vector<int, std::allocator<int> >&)
>>>> const+0xa6) [0x80d8a6]
>>>>  8: (OSD::advance_pg(unsigned int, PG*, ThreadPool::TPHandle&,
>>>> PG::RecoveryCtx*, std::set<boost::intrusive_ptr<PG>,
>>>> std::less<boost::intrusive_ptr<PG> >,
>>>> std::allocator<boost::intrusive_ptr<PG> > >*)+0x190) [0x631b70]
>>>>  9: (OSD::process_peering_events(std::list<PG*, std::allocator<PG*> >
>>>> const&, ThreadPool::TPHandle&)+0x244) [0x632254]
>>>>  10: (OSD::PeeringWQ::_process(std::list<PG*, std::allocator<PG*> >
>>>> const&, ThreadPool::TPHandle&)+0x12) [0x66d0f2]
>>>>  11: (ThreadPool::worker(ThreadPool::WorkThread*)+0x4e6) [0x835a66]
>>>>  12: (ThreadPool::WorkThread::entry()+0x10) [0x837890]
>>>>  13: (()+0x7e9a) [0x7ffe6b721e9a]
>>>>  14: (clone()+0x6d) [0x7ffe69f9fccd]
>>>>  NOTE: a copy of the executable, or `objdump -rdS <executable>` is
>>>> needed to interpret this.
>>>>
>>>> --- logging levels ---
>>>>    0/ 5 none
>>>>    0/ 1 lockdep
>>>>    0/ 1 context
>>>>    1/ 1 crush
>>>>    1/ 5 mds
>>>>    1/ 5 mds_balancer
>>>>    1/ 5 mds_locker
>>>>    1/ 5 mds_log
>>>>    1/ 5 mds_log_expire
>>>>    1/ 5 mds_migrator
>>>>    0/ 1 buffer
>>>>    0/ 1 timer
>>>>    0/ 1 filer
>>>>    0/ 1 striper
>>>>    0/ 1 objecter
>>>>    0/ 5 rados
>>>>    0/ 5 rbd
>>>>    0/ 5 journaler
>>>>    0/ 5 objectcacher
>>>>    0/ 5 client
>>>>    0/ 5 osd
>>>>    0/ 5 optracker
>>>>    0/ 5 objclass
>>>>    1/ 3 filestore
>>>>    1/ 3 journal
>>>>    0/ 5 ms
>>>>    1/ 5 mon
>>>>    0/10 monc
>>>>    0/ 5 paxos
>>>>    0/ 5 tp
>>>>    1/ 5 auth
>>>>    1/ 5 crypto
>>>>    1/ 1 finisher
>>>>    1/ 5 heartbeatmap
>>>>    1/ 5 perfcounter
>>>>    1/ 5 rgw
>>>>    1/ 5 hadoop
>>>>    1/ 5 javaclient
>>>>    1/ 5 asok
>>>>    1/ 1 throttle
>>>>   -2/-2 (syslog threshold)
>>>>   -1/-1 (stderr threshold)
>>>>   max_recent     10000
>>>>   max_new         1000
>>>>   log_file /var/log/ceph/ceph-osd.2.log
>>>> --- end dump of recent events ---
>>>>
>>>> 2013/7/14 Vladislav Gorbunov <vadikgo@xxxxxxxxx>:
>>>>> Hello!
>>>>>
>>>>> After change the crush map all osd (ceph version 0.61.4
>>>>> (1669132fcfc27d0c0b5e5bb93ade59d147e23404)) on pool default is crushed
>>>>> with the error:
>>>>> 2013-07-14 17:26:23.755432 7f0c963ad700 -1 *** Caught signal
>>>>> (Segmentation fault) **
>>>>>  in thread 7f0c963ad700
>>>>> ...skipping...
>>>>>  10: (OSD::PeeringWQ::_process(std::list<PG*, std::allocator<PG*> >
>>>>> const&, ThreadPool::TPHandle&)+0x12) [0x66d0f2]
>>>>>  11: (ThreadPool::worker(ThreadPool::WorkThread*)+0x4e6) [0x835a66]
>>>>>  12: (ThreadPool::WorkThread::entry()+0x10) [0x837890]
>>>>>  13: (()+0x7e9a) [0x7fac4e597e9a]
>>>>>  14: (clone()+0x6d) [0x7fac4ce15ccd]
>>>>>  NOTE: a copy of the executable, or `objdump -rdS <executable>` is
>>>>> needed to interpret this.
>>>>>
>>>>> --- logging levels ---
>>>>>    0/ 5 none
>>>>>    0/ 1 lockdep
>>>>>    0/ 1 context
>>>>>    1/ 1 crush
>>>>>    1/ 5 mds
>>>>>    1/ 5 mds_balancer
>>>>>    1/ 5 mds_locker
>>>>>    1/ 5 mds_log
>>>>>    1/ 5 mds_log_expire
>>>>>    1/ 5 mds_migrator
>>>>>    0/ 1 buffer
>>>>>    0/ 1 timer
>>>>>    0/ 1 filer
>>>>>    0/ 1 striper
>>>>>    0/ 1 objecter
>>>>>    0/ 5 rados
>>>>>    0/ 5 rbd
>>>>>    0/ 5 journaler
>>>>>    0/ 5 objectcacher
>>>>>    0/ 5 client
>>>>>    0/ 5 osd
>>>>>    0/ 5 optracker
>>>>>    0/ 5 objclass
>>>>>    1/ 3 filestore
>>>>>    1/ 3 journal
>>>>>    0/ 5 ms
>>>>>    1/ 5 mon
>>>>>    0/10 monc
>>>>>    0/ 5 paxos
>>>>>    0/ 5 tp
>>>>>    1/ 5 auth
>>>>>    1/ 5 crypto
>>>>>    1/ 1 finisher
>>>>>    1/ 5 heartbeatmap
>>>>>    1/ 5 perfcounter
>>>>>    1/ 5 rgw
>>>>>    1/ 5 hadoop
>>>>>    1/ 5 javaclient
>>>>>    1/ 5 asok
>>>>>    1/ 1 throttle
>>>>>   -2/-2 (syslog threshold)
>>>>>   -1/-1 (stderr threshold)
>>>>>   max_recent     10000
>>>>>   max_new         1000
>>>>>   log_file /var/log/ceph/ceph-osd.2.log
>>>>> --- end dump of recent events ---
>>>>>
>>>>> ceph osd complity ignoring osd start/stop an always show this map:
>>>>> # id weight type name up/down reweight
>>>>> -2 65.52 pool iscsi
>>>>> -4 32.76 datacenter datacenter-cod
>>>>> -6 32.76 host gstore1
>>>>> 82 2.73 osd.82 up 1
>>>>> 83 2.73 osd.83 up 1
>>>>> 84 2.73 osd.84 up 1
>>>>> 85 2.73 osd.85 up 1
>>>>> 86 2.73 osd.86 up 1
>>>>> 87 2.73 osd.87 up 1
>>>>> 88 2.73 osd.88 up 1
>>>>> 89 2.73 osd.89 up 1
>>>>> 90 2.73 osd.90 up 1
>>>>> 91 2.73 osd.91 up 1
>>>>> 92 2.73 osd.92 up 1
>>>>> 93 2.73 osd.93 up 1
>>>>> -5 32.76 datacenter datacenter-rcod
>>>>> -7 32.76 host gstore2
>>>>> 94 2.73 osd.94 up 1
>>>>> 95 2.73 osd.95 up 1
>>>>> 96 2.73 osd.96 up 1
>>>>> 97 2.73 osd.97 up 1
>>>>> 98 2.73 osd.98 up 1
>>>>> 99 2.73 osd.99 up 1
>>>>> 100 2.73 osd.100 up 1
>>>>> 101 2.73 osd.101 up 1
>>>>> 102 2.73 osd.102 up 1
>>>>> 103 2.73 osd.103 up 1
>>>>> 104 2.73 osd.104 up 1
>>>>> 105 2.73 osd.105 up 1
>>>>> -1 68.96 pool default
>>>>> -3 68.96 rack unknownrack
>>>>> -9 5 host gstore5
>>>>> 2 5 osd.2 down 1
>>>>> -10 16 host cstore1
>>>>> 5 1 osd.5 down 0
>>>>> 6 1 osd.6 down 0
>>>>> 7 1 osd.7 down 0
>>>>> 8 1 osd.8 down 0
>>>>> 9 1 osd.9 down 0
>>>>> 10 1 osd.10 down 0
>>>>> 11 1 osd.11 down 0
>>>>> 12 1 osd.12 down 0
>>>>> 13 1 osd.13 down 0
>>>>> 14 1 osd.14 down 0
>>>>> 4 1 osd.4 down 0
>>>>> 47 1 osd.47 down 0
>>>>> 48 1 osd.48 down 0
>>>>> 49 1 osd.49 down 0
>>>>> 50 1 osd.50 down 0
>>>>> 51 1 osd.51 down 0
>>>>> -11 21 host cstore2
>>>>> 15 1 osd.15 down 0
>>>>> 16 1 osd.16 down 0
>>>>> 17 1 osd.17 down 0
>>>>> 18 1 osd.18 down 0
>>>>> 19 1 osd.19 down 0
>>>>> 20 1 osd.20 down 0
>>>>> 21 1 osd.21 down 0
>>>>> 22 1 osd.22 down 0
>>>>> 23 1 osd.23 down 0
>>>>> 24 1 osd.24 down 0
>>>>> 41 1 osd.41 down 0
>>>>> 42 1 osd.42 down 0
>>>>> 43 1 osd.43 down 0
>>>>> 44 1 osd.44 down 0
>>>>> 45 1 osd.45 down 0
>>>>> 46 1 osd.46 down 0
>>>>> 52 1 osd.52 down 0
>>>>> 53 1 osd.53 down 0
>>>>> 54 1 osd.54 down 0
>>>>> 55 1 osd.55 down 0
>>>>> 56 1 osd.56 up 1
>>>>> 57 0 osd.57 up 1
>>>>> -12 16 host cstore3
>>>>> 25 1 osd.25 down 0
>>>>> 26 1 osd.26 down 0
>>>>> 27 1 osd.27 down 0
>>>>> 28 1 osd.28 down 0
>>>>> 29 1 osd.29 down 0
>>>>> 30 1 osd.30 down 0
>>>>> 31 1 osd.31 down 0
>>>>> 32 1 osd.32 down 0
>>>>> 33 1 osd.33 down 0
>>>>> 34 1 osd.34 down 0
>>>>> 35 1 osd.35 down 0
>>>>> 36 1 osd.36 down 0
>>>>> 37 1 osd.37 down 0
>>>>> 38 1 osd.38 down 0
>>>>> 39 1 osd.39 down 0
>>>>> 40 1 osd.40 down 0
>>>>> -13 7.64 host cstore4
>>>>> 62 0.55 osd.62 down 0
>>>>> 63 0.55 osd.63 down 0
>>>>> 64 0.55 osd.64 down 0
>>>>> 65 0.55 osd.65 down 0
>>>>> 66 0.55 osd.66 down 0
>>>>> 67 0.55 osd.67 down 0
>>>>> 68 0.55 osd.68 down 0
>>>>> 69 0.55 osd.69 down 0
>>>>> 70 0.27 osd.70 down 0
>>>>> 71 0.27 osd.71 down 0
>>>>> 72 0.27 osd.72 down 0
>>>>> 73 0.27 osd.73 down 0
>>>>> 74 0.27 osd.74 down 0
>>>>> 75 0.27 osd.75 down 0
>>>>> 76 0.27 osd.76 down 0
>>>>> 77 0.27 osd.77 down 0
>>>>> 78 0.27 osd.78 down 0
>>>>> 79 0.27 osd.79 down 0
>>>>> 80 0.27 osd.80 down 0
>>>>> 81 0.27 osd.81 down 0
>>>>> -14 3.32 host cstore5
>>>>> 0 0.4 osd.0 down 0
>>>>> 1 0.5 osd.1 down 0
>>>>> 3 0.4 osd.3 down 0
>>>>> 58 0.4 osd.58 down 1
>>>>> 59 0.54 osd.59 down 1
>>>>> 60 0.54 osd.60 down 1
>>>>> 61 0.54 osd.61 down 1
>>>>>
>>>>> crush map is:
>>>>> # begin crush map
>>>>>
>>>>> # devices
>>>>> device 0 osd.0
>>>>> device 1 osd.1
>>>>> device 2 osd.2
>>>>> device 3 osd.3
>>>>> device 4 osd.4
>>>>> device 5 osd.5
>>>>> device 6 osd.6
>>>>> device 7 osd.7
>>>>> device 8 osd.8
>>>>> device 9 osd.9
>>>>> device 10 osd.10
>>>>> device 11 osd.11
>>>>> device 12 osd.12
>>>>> device 13 osd.13
>>>>> device 14 osd.14
>>>>> device 15 osd.15
>>>>> device 16 osd.16
>>>>> device 17 osd.17
>>>>> device 18 osd.18
>>>>> device 19 osd.19
>>>>> device 20 osd.20
>>>>> device 21 osd.21
>>>>> device 22 osd.22
>>>>> device 23 osd.23
>>>>> device 24 osd.24
>>>>> device 25 osd.25
>>>>> device 26 osd.26
>>>>> device 27 osd.27
>>>>> device 28 osd.28
>>>>> device 29 osd.29
>>>>> device 30 osd.30
>>>>> device 31 osd.31
>>>>> device 32 osd.32
>>>>> device 33 osd.33
>>>>> device 34 osd.34
>>>>> device 35 osd.35
>>>>> device 36 osd.36
>>>>> device 37 osd.37
>>>>> device 38 osd.38
>>>>> device 39 osd.39
>>>>> device 40 osd.40
>>>>> device 41 osd.41
>>>>> device 42 osd.42
>>>>> device 43 osd.43
>>>>> device 44 osd.44
>>>>> device 45 osd.45
>>>>> device 46 osd.46
>>>>> device 47 osd.47
>>>>> device 48 osd.48
>>>>> device 49 osd.49
>>>>> device 50 osd.50
>>>>> device 51 osd.51
>>>>> device 52 osd.52
>>>>> device 53 osd.53
>>>>> device 54 osd.54
>>>>> device 55 osd.55
>>>>> device 56 osd.56
>>>>> device 57 osd.57
>>>>> device 58 osd.58
>>>>> device 59 osd.59
>>>>> device 60 osd.60
>>>>> device 61 osd.61
>>>>> device 62 osd.62
>>>>> device 63 osd.63
>>>>> device 64 osd.64
>>>>> device 65 osd.65
>>>>> device 66 osd.66
>>>>> device 67 osd.67
>>>>> device 68 osd.68
>>>>> device 69 osd.69
>>>>> device 70 osd.70
>>>>> device 71 osd.71
>>>>> device 72 osd.72
>>>>> device 73 osd.73
>>>>> device 74 osd.74
>>>>> device 75 osd.75
>>>>> device 76 osd.76
>>>>> device 77 osd.77
>>>>> device 78 osd.78
>>>>> device 79 osd.79
>>>>> device 80 osd.80
>>>>> device 81 osd.81
>>>>> device 82 osd.82
>>>>> device 83 osd.83
>>>>> device 84 osd.84
>>>>> device 85 osd.85
>>>>> device 86 osd.86
>>>>> device 87 osd.87
>>>>> device 88 osd.88
>>>>> device 89 osd.89
>>>>> device 90 osd.90
>>>>> device 91 osd.91
>>>>> device 92 osd.92
>>>>> device 93 osd.93
>>>>> device 94 osd.94
>>>>> device 95 osd.95
>>>>> device 96 osd.96
>>>>> device 97 osd.97
>>>>> device 98 osd.98
>>>>> device 99 osd.99
>>>>> device 100 osd.100
>>>>> device 101 osd.101
>>>>> device 102 osd.102
>>>>> device 103 osd.103
>>>>> device 104 osd.104
>>>>> device 105 osd.105
>>>>>
>>>>> # types
>>>>> type 0 osd
>>>>> type 1 host
>>>>> type 2 rack
>>>>> type 3 row
>>>>> type 4 room
>>>>> type 5 datacenter
>>>>> type 6 pool
>>>>>
>>>>> # buckets
>>>>> host gstore5 {
>>>>> id -9 # do not change unnecessarily
>>>>> # weight 5.000
>>>>> alg straw
>>>>> hash 0 # rjenkins1
>>>>> item osd.2 weight 5.000
>>>>> }
>>>>> host cstore1 {
>>>>> id -10 # do not change unnecessarily
>>>>> # weight 16.000
>>>>> alg straw
>>>>> hash 0 # rjenkins1
>>>>> item osd.5 weight 1.000
>>>>> item osd.6 weight 1.000
>>>>> item osd.7 weight 1.000
>>>>> item osd.8 weight 1.000
>>>>> item osd.9 weight 1.000
>>>>> item osd.10 weight 1.000
>>>>> item osd.11 weight 1.000
>>>>> item osd.12 weight 1.000
>>>>> item osd.13 weight 1.000
>>>>> item osd.14 weight 1.000
>>>>> item osd.4 weight 1.000
>>>>> item osd.47 weight 1.000
>>>>> item osd.48 weight 1.000
>>>>> item osd.49 weight 1.000
>>>>> item osd.50 weight 1.000
>>>>> item osd.51 weight 1.000
>>>>> }
>>>>> host cstore2 {
>>>>> id -11 # do not change unnecessarily
>>>>> # weight 20.000
>>>>> alg straw
>>>>> hash 0 # rjenkins1
>>>>> item osd.15 weight 1.000
>>>>> item osd.16 weight 1.000
>>>>> item osd.17 weight 1.000
>>>>> item osd.18 weight 1.000
>>>>> item osd.19 weight 1.000
>>>>> item osd.20 weight 1.000
>>>>> item osd.21 weight 1.000
>>>>> item osd.22 weight 1.000
>>>>> item osd.23 weight 1.000
>>>>> item osd.24 weight 1.000
>>>>> item osd.41 weight 1.000
>>>>> item osd.42 weight 1.000
>>>>> item osd.43 weight 1.000
>>>>> item osd.44 weight 1.000
>>>>> item osd.45 weight 1.000
>>>>> item osd.46 weight 1.000
>>>>> item osd.52 weight 1.000
>>>>> item osd.53 weight 1.000
>>>>> item osd.54 weight 1.000
>>>>> item osd.55 weight 1.000
>>>>> item osd.56 weight 0.000
>>>>> item osd.57 weight 0.000
>>>>> }
>>>>> host cstore3 {
>>>>> id -12 # do not change unnecessarily
>>>>> # weight 16.000
>>>>> alg straw
>>>>> hash 0 # rjenkins1
>>>>> item osd.25 weight 1.000
>>>>> item osd.26 weight 1.000
>>>>> item osd.27 weight 1.000
>>>>> item osd.28 weight 1.000
>>>>> item osd.29 weight 1.000
>>>>> item osd.30 weight 1.000
>>>>> item osd.31 weight 1.000
>>>>> item osd.32 weight 1.000
>>>>> item osd.33 weight 1.000
>>>>> item osd.34 weight 1.000
>>>>> item osd.35 weight 1.000
>>>>> item osd.36 weight 1.000
>>>>> item osd.37 weight 1.000
>>>>> item osd.38 weight 1.000
>>>>> item osd.39 weight 1.000
>>>>> item osd.40 weight 1.000
>>>>> }
>>>>> host cstore4 {
>>>>> id -13 # do not change unnecessarily
>>>>> # weight 7.640
>>>>> alg straw
>>>>> hash 0 # rjenkins1
>>>>> item osd.62 weight 0.550
>>>>> item osd.63 weight 0.550
>>>>> item osd.64 weight 0.550
>>>>> item osd.65 weight 0.550
>>>>> item osd.66 weight 0.550
>>>>> item osd.67 weight 0.550
>>>>> item osd.68 weight 0.550
>>>>> item osd.69 weight 0.550
>>>>> item osd.70 weight 0.270
>>>>> item osd.71 weight 0.270
>>>>> item osd.72 weight 0.270
>>>>> item osd.73 weight 0.270
>>>>> item osd.74 weight 0.270
>>>>> item osd.75 weight 0.270
>>>>> item osd.76 weight 0.270
>>>>> item osd.77 weight 0.270
>>>>> item osd.78 weight 0.270
>>>>> item osd.79 weight 0.270
>>>>> item osd.80 weight 0.270
>>>>> item osd.81 weight 0.270
>>>>> }
>>>>> host cstore5 {
>>>>> id -14 # do not change unnecessarily
>>>>> # weight 3.320
>>>>> alg straw
>>>>> hash 0 # rjenkins1
>>>>> item osd.0 weight 0.400
>>>>> item osd.1 weight 0.500
>>>>> item osd.3 weight 0.400
>>>>> item osd.58 weight 0.400
>>>>> item osd.59 weight 0.540
>>>>> item osd.60 weight 0.540
>>>>> item osd.61 weight 0.540
>>>>> }
>>>>> rack unknownrack {
>>>>> id -3 # do not change unnecessarily
>>>>> # weight 67.960
>>>>> alg straw
>>>>> hash 0 # rjenkins1
>>>>> item gstore5 weight 5.000
>>>>> item cstore1 weight 16.000
>>>>> item cstore2 weight 20.000
>>>>> item cstore3 weight 16.000
>>>>> item cstore4 weight 7.640
>>>>> item cstore5 weight 3.320
>>>>> }
>>>>> pool default {
>>>>> id -1 # do not change unnecessarily
>>>>> # weight 67.960
>>>>> alg straw
>>>>> hash 0 # rjenkins1
>>>>> item unknownrack weight 67.960
>>>>> }
>>>>> host gstore1 {
>>>>> id -6 # do not change unnecessarily
>>>>> # weight 32.760
>>>>> alg straw
>>>>> hash 0 # rjenkins1
>>>>> item osd.82 weight 2.730
>>>>> item osd.83 weight 2.730
>>>>> item osd.84 weight 2.730
>>>>> item osd.85 weight 2.730
>>>>> item osd.86 weight 2.730
>>>>> item osd.87 weight 2.730
>>>>> item osd.88 weight 2.730
>>>>> item osd.89 weight 2.730
>>>>> item osd.90 weight 2.730
>>>>> item osd.91 weight 2.730
>>>>> item osd.92 weight 2.730
>>>>> item osd.93 weight 2.730
>>>>> }
>>>>> datacenter datacenter-cod {
>>>>> id -4 # do not change unnecessarily
>>>>> # weight 32.760
>>>>> alg straw
>>>>> hash 0 # rjenkins1
>>>>> item gstore1 weight 32.760
>>>>> }
>>>>> host gstore2 {
>>>>> id -7 # do not change unnecessarily
>>>>> # weight 32.760
>>>>> alg straw
>>>>> hash 0 # rjenkins1
>>>>> item osd.94 weight 2.730
>>>>> item osd.95 weight 2.730
>>>>> item osd.96 weight 2.730
>>>>> item osd.97 weight 2.730
>>>>> item osd.98 weight 2.730
>>>>> item osd.99 weight 2.730
>>>>> item osd.100 weight 2.730
>>>>> item osd.101 weight 2.730
>>>>> item osd.102 weight 2.730
>>>>> item osd.103 weight 2.730
>>>>> item osd.104 weight 2.730
>>>>> item osd.105 weight 2.730
>>>>> }
>>>>> datacenter datacenter-rcod {
>>>>> id -5 # do not change unnecessarily
>>>>> # weight 32.760
>>>>> alg straw
>>>>> hash 0 # rjenkins1
>>>>> item gstore2 weight 32.760
>>>>> }
>>>>> pool iscsi {
>>>>> id -2 # do not change unnecessarily
>>>>> # weight 65.520
>>>>> alg straw
>>>>> hash 0 # rjenkins1
>>>>> item datacenter-cod weight 32.760
>>>>> item datacenter-rcod weight 32.760
>>>>> }
>>>>>
>>>>> # rules
>>>>> rule data {
>>>>> ruleset 0
>>>>> type replicated
>>>>> min_size 1
>>>>> max_size 10
>>>>> step take default
>>>>> step chooseleaf firstn 0 type host
>>>>> step emit
>>>>> }
>>>>> rule metadata {
>>>>> ruleset 1
>>>>> type replicated
>>>>> min_size 1
>>>>> max_size 10
>>>>> step take default
>>>>> step chooseleaf firstn 0 type host
>>>>> step emit
>>>>> }
>>>>> rule rbd {
>>>>> ruleset 2
>>>>> type replicated
>>>>> min_size 1
>>>>> max_size 10
>>>>> step take default
>>>>> step chooseleaf firstn 0 type host
>>>>> step emit
>>>>> }
>>>>>
>>>>> # end crush map
>>>>>
>>>>>
>>>>> Restarting mon don't help.
>>>> _______________________________________________
>>>> ceph-users mailing list
>>>> ceph-users@xxxxxxxxxxxxxx
>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux