Gregory, thank for you help! After all osd servers downed, i'am back rule set for the iscsi pool back to default rule 0: ceph osd pool set iscsi crush_ruleset 0 it does not help, all osd not started, except without data, with weight 0. next i remove ruleset iscsi from crush map. It does not help too. And after that i post crushmap to this mail list. Is any method to extract crush map from downed osd server and inject it to the mon server? from /var/lib/ceph/osd/ceph-2/current/omap folder? 2013/7/17 Gregory Farnum <greg@xxxxxxxxxxx>: > I notice that your first dump of the crush map didn't include rule #3. > Are you sure you've injected it into the cluster? Try extracting it > from the monitors and looking at that map directly, instead of a > locally cached version. > You mentioned some problem with OSDs being positioned wrong too, so > you might look at "ceph osd tree" and look at the shape of the map. > But it sounds to me like maybe there's a disconnect between what > you've put into the cluster, and what you're looking at. > -Greg > Software Engineer #42 @ http://inktank.com | http://ceph.com > > > On Mon, Jul 15, 2013 at 2:11 PM, Vladislav Gorbunov <vadikgo@xxxxxxxxx> wrote: >> ruleset 3 is: >> rule iscsi { >> ruleset 3 >> type replicated >> min_size 1 >> max_size 10 >> step take iscsi >> step chooseleaf firstn 0 type datacenter >> step chooseleaf firstn 0 type host >> step emit >> } >> >> 2013/7/16 Vladislav Gorbunov <vadikgo@xxxxxxxxx>: >>> sorry, after i'm try to apply crush ruleset 3 (iscsi) to >>> pool iscsi: >>> ceph osd pool set iscsi crush_ruleset 3 >>> >>> 2013/7/16 Vladislav Gorbunov <vadikgo@xxxxxxxxx>: >>>>>Have you run this crush map through any test mappings yet? >>>> Yes, it worked on test cluster, and after apply map to main cluster. >>>> OSD servers downed after i'm try to apply crush ruleset 3 (iscsi) to >>>> pool iscsi: >>>> ceph osd pool set data crush_ruleset 3 >>>> >>>> 2013/7/16 Gregory Farnum <greg@xxxxxxxxxxx>: >>>>> It's probably not the same issue as that ticket, which was about the >>>>> OSD handling a lack of output incorrectly. (It might be handling the >>>>> output incorrectly in some other way, but hopefully not...) >>>>> >>>>> Have you run this crush map through any test mappings yet? >>>>> -Greg >>>>> Software Engineer #42 @ http://inktank.com | http://ceph.com >>>>> >>>>> >>>>> On Sun, Jul 14, 2013 at 10:59 PM, Vladislav Gorbunov <vadikgo@xxxxxxxxx> wrote: >>>>>> Sympthoms like on http://tracker.ceph.com/issues/4699 >>>>>> >>>>>> all OSDs the process ceph-osd crash with segfault >>>>>> >>>>>> If i stop MONs daemons then i can start OSDs but if i start MONs back >>>>>> then die all OSDs again. >>>>>> >>>>>> more detailed log: >>>>>> 0> 2013-07-15 16:42:05.001242 7ffe5a6fc700 -1 *** Caught signal >>>>>> (Segmentation fault) ** >>>>>> in thread 7ffe5a6fc700 >>>>>> >>>>>> ceph version 0.61.4 (1669132fcfc27d0c0b5e5bb93ade59d147e23404) >>>>>> 1: /usr/bin/ceph-osd() [0x790e5a] >>>>>> 2: (()+0xfcb0) [0x7ffe6b729cb0] >>>>>> 3: /usr/bin/ceph-osd() [0x893879] >>>>>> 4: (crush_do_rule()+0x1e5) [0x894065] >>>>>> 5: (CrushWrapper::do_rule(int, int, std::vector<int, >>>>>> std::allocator<int> >&, int, std::vector<unsigned int, >>>>>> std::allocator<unsigned int> > const&) const+0x7a) [0x81b2ba] >>>>>> 6: (OSDMap::_pg_to_osds(pg_pool_t const&, pg_t, std::vector<int, >>>>>> std::allocator<int> >&) const+0x8f) [0x80d7cf] >>>>>> 7: (OSDMap::pg_to_up_acting_osds(pg_t, std::vector<int, >>>>>> std::allocator<int> >&, std::vector<int, std::allocator<int> >&) >>>>>> const+0xa6) [0x80d8a6] >>>>>> 8: (OSD::advance_pg(unsigned int, PG*, ThreadPool::TPHandle&, >>>>>> PG::RecoveryCtx*, std::set<boost::intrusive_ptr<PG>, >>>>>> std::less<boost::intrusive_ptr<PG> >, >>>>>> std::allocator<boost::intrusive_ptr<PG> > >*)+0x190) [0x631b70] >>>>>> 9: (OSD::process_peering_events(std::list<PG*, std::allocator<PG*> > >>>>>> const&, ThreadPool::TPHandle&)+0x244) [0x632254] >>>>>> 10: (OSD::PeeringWQ::_process(std::list<PG*, std::allocator<PG*> > >>>>>> const&, ThreadPool::TPHandle&)+0x12) [0x66d0f2] >>>>>> 11: (ThreadPool::worker(ThreadPool::WorkThread*)+0x4e6) [0x835a66] >>>>>> 12: (ThreadPool::WorkThread::entry()+0x10) [0x837890] >>>>>> 13: (()+0x7e9a) [0x7ffe6b721e9a] >>>>>> 14: (clone()+0x6d) [0x7ffe69f9fccd] >>>>>> NOTE: a copy of the executable, or `objdump -rdS <executable>` is >>>>>> needed to interpret this. >>>>>> >>>>>> --- logging levels --- >>>>>> 0/ 5 none >>>>>> 0/ 1 lockdep >>>>>> 0/ 1 context >>>>>> 1/ 1 crush >>>>>> 1/ 5 mds >>>>>> 1/ 5 mds_balancer >>>>>> 1/ 5 mds_locker >>>>>> 1/ 5 mds_log >>>>>> 1/ 5 mds_log_expire >>>>>> 1/ 5 mds_migrator >>>>>> 0/ 1 buffer >>>>>> 0/ 1 timer >>>>>> 0/ 1 filer >>>>>> 0/ 1 striper >>>>>> 0/ 1 objecter >>>>>> 0/ 5 rados >>>>>> 0/ 5 rbd >>>>>> 0/ 5 journaler >>>>>> 0/ 5 objectcacher >>>>>> 0/ 5 client >>>>>> 0/ 5 osd >>>>>> 0/ 5 optracker >>>>>> 0/ 5 objclass >>>>>> 1/ 3 filestore >>>>>> 1/ 3 journal >>>>>> 0/ 5 ms >>>>>> 1/ 5 mon >>>>>> 0/10 monc >>>>>> 0/ 5 paxos >>>>>> 0/ 5 tp >>>>>> 1/ 5 auth >>>>>> 1/ 5 crypto >>>>>> 1/ 1 finisher >>>>>> 1/ 5 heartbeatmap >>>>>> 1/ 5 perfcounter >>>>>> 1/ 5 rgw >>>>>> 1/ 5 hadoop >>>>>> 1/ 5 javaclient >>>>>> 1/ 5 asok >>>>>> 1/ 1 throttle >>>>>> -2/-2 (syslog threshold) >>>>>> -1/-1 (stderr threshold) >>>>>> max_recent 10000 >>>>>> max_new 1000 >>>>>> log_file /var/log/ceph/ceph-osd.2.log >>>>>> --- end dump of recent events --- >>>>>> >>>>>> 2013/7/14 Vladislav Gorbunov <vadikgo@xxxxxxxxx>: >>>>>>> Hello! >>>>>>> >>>>>>> After change the crush map all osd (ceph version 0.61.4 >>>>>>> (1669132fcfc27d0c0b5e5bb93ade59d147e23404)) on pool default is crushed >>>>>>> with the error: >>>>>>> 2013-07-14 17:26:23.755432 7f0c963ad700 -1 *** Caught signal >>>>>>> (Segmentation fault) ** >>>>>>> in thread 7f0c963ad700 >>>>>>> ...skipping... >>>>>>> 10: (OSD::PeeringWQ::_process(std::list<PG*, std::allocator<PG*> > >>>>>>> const&, ThreadPool::TPHandle&)+0x12) [0x66d0f2] >>>>>>> 11: (ThreadPool::worker(ThreadPool::WorkThread*)+0x4e6) [0x835a66] >>>>>>> 12: (ThreadPool::WorkThread::entry()+0x10) [0x837890] >>>>>>> 13: (()+0x7e9a) [0x7fac4e597e9a] >>>>>>> 14: (clone()+0x6d) [0x7fac4ce15ccd] >>>>>>> NOTE: a copy of the executable, or `objdump -rdS <executable>` is >>>>>>> needed to interpret this. >>>>>>> >>>>>>> --- logging levels --- >>>>>>> 0/ 5 none >>>>>>> 0/ 1 lockdep >>>>>>> 0/ 1 context >>>>>>> 1/ 1 crush >>>>>>> 1/ 5 mds >>>>>>> 1/ 5 mds_balancer >>>>>>> 1/ 5 mds_locker >>>>>>> 1/ 5 mds_log >>>>>>> 1/ 5 mds_log_expire >>>>>>> 1/ 5 mds_migrator >>>>>>> 0/ 1 buffer >>>>>>> 0/ 1 timer >>>>>>> 0/ 1 filer >>>>>>> 0/ 1 striper >>>>>>> 0/ 1 objecter >>>>>>> 0/ 5 rados >>>>>>> 0/ 5 rbd >>>>>>> 0/ 5 journaler >>>>>>> 0/ 5 objectcacher >>>>>>> 0/ 5 client >>>>>>> 0/ 5 osd >>>>>>> 0/ 5 optracker >>>>>>> 0/ 5 objclass >>>>>>> 1/ 3 filestore >>>>>>> 1/ 3 journal >>>>>>> 0/ 5 ms >>>>>>> 1/ 5 mon >>>>>>> 0/10 monc >>>>>>> 0/ 5 paxos >>>>>>> 0/ 5 tp >>>>>>> 1/ 5 auth >>>>>>> 1/ 5 crypto >>>>>>> 1/ 1 finisher >>>>>>> 1/ 5 heartbeatmap >>>>>>> 1/ 5 perfcounter >>>>>>> 1/ 5 rgw >>>>>>> 1/ 5 hadoop >>>>>>> 1/ 5 javaclient >>>>>>> 1/ 5 asok >>>>>>> 1/ 1 throttle >>>>>>> -2/-2 (syslog threshold) >>>>>>> -1/-1 (stderr threshold) >>>>>>> max_recent 10000 >>>>>>> max_new 1000 >>>>>>> log_file /var/log/ceph/ceph-osd.2.log >>>>>>> --- end dump of recent events --- >>>>>>> >>>>>>> ceph osd complity ignoring osd start/stop an always show this map: >>>>>>> # id weight type name up/down reweight >>>>>>> -2 65.52 pool iscsi >>>>>>> -4 32.76 datacenter datacenter-cod >>>>>>> -6 32.76 host gstore1 >>>>>>> 82 2.73 osd.82 up 1 >>>>>>> 83 2.73 osd.83 up 1 >>>>>>> 84 2.73 osd.84 up 1 >>>>>>> 85 2.73 osd.85 up 1 >>>>>>> 86 2.73 osd.86 up 1 >>>>>>> 87 2.73 osd.87 up 1 >>>>>>> 88 2.73 osd.88 up 1 >>>>>>> 89 2.73 osd.89 up 1 >>>>>>> 90 2.73 osd.90 up 1 >>>>>>> 91 2.73 osd.91 up 1 >>>>>>> 92 2.73 osd.92 up 1 >>>>>>> 93 2.73 osd.93 up 1 >>>>>>> -5 32.76 datacenter datacenter-rcod >>>>>>> -7 32.76 host gstore2 >>>>>>> 94 2.73 osd.94 up 1 >>>>>>> 95 2.73 osd.95 up 1 >>>>>>> 96 2.73 osd.96 up 1 >>>>>>> 97 2.73 osd.97 up 1 >>>>>>> 98 2.73 osd.98 up 1 >>>>>>> 99 2.73 osd.99 up 1 >>>>>>> 100 2.73 osd.100 up 1 >>>>>>> 101 2.73 osd.101 up 1 >>>>>>> 102 2.73 osd.102 up 1 >>>>>>> 103 2.73 osd.103 up 1 >>>>>>> 104 2.73 osd.104 up 1 >>>>>>> 105 2.73 osd.105 up 1 >>>>>>> -1 68.96 pool default >>>>>>> -3 68.96 rack unknownrack >>>>>>> -9 5 host gstore5 >>>>>>> 2 5 osd.2 down 1 >>>>>>> -10 16 host cstore1 >>>>>>> 5 1 osd.5 down 0 >>>>>>> 6 1 osd.6 down 0 >>>>>>> 7 1 osd.7 down 0 >>>>>>> 8 1 osd.8 down 0 >>>>>>> 9 1 osd.9 down 0 >>>>>>> 10 1 osd.10 down 0 >>>>>>> 11 1 osd.11 down 0 >>>>>>> 12 1 osd.12 down 0 >>>>>>> 13 1 osd.13 down 0 >>>>>>> 14 1 osd.14 down 0 >>>>>>> 4 1 osd.4 down 0 >>>>>>> 47 1 osd.47 down 0 >>>>>>> 48 1 osd.48 down 0 >>>>>>> 49 1 osd.49 down 0 >>>>>>> 50 1 osd.50 down 0 >>>>>>> 51 1 osd.51 down 0 >>>>>>> -11 21 host cstore2 >>>>>>> 15 1 osd.15 down 0 >>>>>>> 16 1 osd.16 down 0 >>>>>>> 17 1 osd.17 down 0 >>>>>>> 18 1 osd.18 down 0 >>>>>>> 19 1 osd.19 down 0 >>>>>>> 20 1 osd.20 down 0 >>>>>>> 21 1 osd.21 down 0 >>>>>>> 22 1 osd.22 down 0 >>>>>>> 23 1 osd.23 down 0 >>>>>>> 24 1 osd.24 down 0 >>>>>>> 41 1 osd.41 down 0 >>>>>>> 42 1 osd.42 down 0 >>>>>>> 43 1 osd.43 down 0 >>>>>>> 44 1 osd.44 down 0 >>>>>>> 45 1 osd.45 down 0 >>>>>>> 46 1 osd.46 down 0 >>>>>>> 52 1 osd.52 down 0 >>>>>>> 53 1 osd.53 down 0 >>>>>>> 54 1 osd.54 down 0 >>>>>>> 55 1 osd.55 down 0 >>>>>>> 56 1 osd.56 up 1 >>>>>>> 57 0 osd.57 up 1 >>>>>>> -12 16 host cstore3 >>>>>>> 25 1 osd.25 down 0 >>>>>>> 26 1 osd.26 down 0 >>>>>>> 27 1 osd.27 down 0 >>>>>>> 28 1 osd.28 down 0 >>>>>>> 29 1 osd.29 down 0 >>>>>>> 30 1 osd.30 down 0 >>>>>>> 31 1 osd.31 down 0 >>>>>>> 32 1 osd.32 down 0 >>>>>>> 33 1 osd.33 down 0 >>>>>>> 34 1 osd.34 down 0 >>>>>>> 35 1 osd.35 down 0 >>>>>>> 36 1 osd.36 down 0 >>>>>>> 37 1 osd.37 down 0 >>>>>>> 38 1 osd.38 down 0 >>>>>>> 39 1 osd.39 down 0 >>>>>>> 40 1 osd.40 down 0 >>>>>>> -13 7.64 host cstore4 >>>>>>> 62 0.55 osd.62 down 0 >>>>>>> 63 0.55 osd.63 down 0 >>>>>>> 64 0.55 osd.64 down 0 >>>>>>> 65 0.55 osd.65 down 0 >>>>>>> 66 0.55 osd.66 down 0 >>>>>>> 67 0.55 osd.67 down 0 >>>>>>> 68 0.55 osd.68 down 0 >>>>>>> 69 0.55 osd.69 down 0 >>>>>>> 70 0.27 osd.70 down 0 >>>>>>> 71 0.27 osd.71 down 0 >>>>>>> 72 0.27 osd.72 down 0 >>>>>>> 73 0.27 osd.73 down 0 >>>>>>> 74 0.27 osd.74 down 0 >>>>>>> 75 0.27 osd.75 down 0 >>>>>>> 76 0.27 osd.76 down 0 >>>>>>> 77 0.27 osd.77 down 0 >>>>>>> 78 0.27 osd.78 down 0 >>>>>>> 79 0.27 osd.79 down 0 >>>>>>> 80 0.27 osd.80 down 0 >>>>>>> 81 0.27 osd.81 down 0 >>>>>>> -14 3.32 host cstore5 >>>>>>> 0 0.4 osd.0 down 0 >>>>>>> 1 0.5 osd.1 down 0 >>>>>>> 3 0.4 osd.3 down 0 >>>>>>> 58 0.4 osd.58 down 1 >>>>>>> 59 0.54 osd.59 down 1 >>>>>>> 60 0.54 osd.60 down 1 >>>>>>> 61 0.54 osd.61 down 1 >>>>>>> >>>>>>> crush map is: >>>>>>> # begin crush map >>>>>>> >>>>>>> # devices >>>>>>> device 0 osd.0 >>>>>>> device 1 osd.1 >>>>>>> device 2 osd.2 >>>>>>> device 3 osd.3 >>>>>>> device 4 osd.4 >>>>>>> device 5 osd.5 >>>>>>> device 6 osd.6 >>>>>>> device 7 osd.7 >>>>>>> device 8 osd.8 >>>>>>> device 9 osd.9 >>>>>>> device 10 osd.10 >>>>>>> device 11 osd.11 >>>>>>> device 12 osd.12 >>>>>>> device 13 osd.13 >>>>>>> device 14 osd.14 >>>>>>> device 15 osd.15 >>>>>>> device 16 osd.16 >>>>>>> device 17 osd.17 >>>>>>> device 18 osd.18 >>>>>>> device 19 osd.19 >>>>>>> device 20 osd.20 >>>>>>> device 21 osd.21 >>>>>>> device 22 osd.22 >>>>>>> device 23 osd.23 >>>>>>> device 24 osd.24 >>>>>>> device 25 osd.25 >>>>>>> device 26 osd.26 >>>>>>> device 27 osd.27 >>>>>>> device 28 osd.28 >>>>>>> device 29 osd.29 >>>>>>> device 30 osd.30 >>>>>>> device 31 osd.31 >>>>>>> device 32 osd.32 >>>>>>> device 33 osd.33 >>>>>>> device 34 osd.34 >>>>>>> device 35 osd.35 >>>>>>> device 36 osd.36 >>>>>>> device 37 osd.37 >>>>>>> device 38 osd.38 >>>>>>> device 39 osd.39 >>>>>>> device 40 osd.40 >>>>>>> device 41 osd.41 >>>>>>> device 42 osd.42 >>>>>>> device 43 osd.43 >>>>>>> device 44 osd.44 >>>>>>> device 45 osd.45 >>>>>>> device 46 osd.46 >>>>>>> device 47 osd.47 >>>>>>> device 48 osd.48 >>>>>>> device 49 osd.49 >>>>>>> device 50 osd.50 >>>>>>> device 51 osd.51 >>>>>>> device 52 osd.52 >>>>>>> device 53 osd.53 >>>>>>> device 54 osd.54 >>>>>>> device 55 osd.55 >>>>>>> device 56 osd.56 >>>>>>> device 57 osd.57 >>>>>>> device 58 osd.58 >>>>>>> device 59 osd.59 >>>>>>> device 60 osd.60 >>>>>>> device 61 osd.61 >>>>>>> device 62 osd.62 >>>>>>> device 63 osd.63 >>>>>>> device 64 osd.64 >>>>>>> device 65 osd.65 >>>>>>> device 66 osd.66 >>>>>>> device 67 osd.67 >>>>>>> device 68 osd.68 >>>>>>> device 69 osd.69 >>>>>>> device 70 osd.70 >>>>>>> device 71 osd.71 >>>>>>> device 72 osd.72 >>>>>>> device 73 osd.73 >>>>>>> device 74 osd.74 >>>>>>> device 75 osd.75 >>>>>>> device 76 osd.76 >>>>>>> device 77 osd.77 >>>>>>> device 78 osd.78 >>>>>>> device 79 osd.79 >>>>>>> device 80 osd.80 >>>>>>> device 81 osd.81 >>>>>>> device 82 osd.82 >>>>>>> device 83 osd.83 >>>>>>> device 84 osd.84 >>>>>>> device 85 osd.85 >>>>>>> device 86 osd.86 >>>>>>> device 87 osd.87 >>>>>>> device 88 osd.88 >>>>>>> device 89 osd.89 >>>>>>> device 90 osd.90 >>>>>>> device 91 osd.91 >>>>>>> device 92 osd.92 >>>>>>> device 93 osd.93 >>>>>>> device 94 osd.94 >>>>>>> device 95 osd.95 >>>>>>> device 96 osd.96 >>>>>>> device 97 osd.97 >>>>>>> device 98 osd.98 >>>>>>> device 99 osd.99 >>>>>>> device 100 osd.100 >>>>>>> device 101 osd.101 >>>>>>> device 102 osd.102 >>>>>>> device 103 osd.103 >>>>>>> device 104 osd.104 >>>>>>> device 105 osd.105 >>>>>>> >>>>>>> # types >>>>>>> type 0 osd >>>>>>> type 1 host >>>>>>> type 2 rack >>>>>>> type 3 row >>>>>>> type 4 room >>>>>>> type 5 datacenter >>>>>>> type 6 pool >>>>>>> >>>>>>> # buckets >>>>>>> host gstore5 { >>>>>>> id -9 # do not change unnecessarily >>>>>>> # weight 5.000 >>>>>>> alg straw >>>>>>> hash 0 # rjenkins1 >>>>>>> item osd.2 weight 5.000 >>>>>>> } >>>>>>> host cstore1 { >>>>>>> id -10 # do not change unnecessarily >>>>>>> # weight 16.000 >>>>>>> alg straw >>>>>>> hash 0 # rjenkins1 >>>>>>> item osd.5 weight 1.000 >>>>>>> item osd.6 weight 1.000 >>>>>>> item osd.7 weight 1.000 >>>>>>> item osd.8 weight 1.000 >>>>>>> item osd.9 weight 1.000 >>>>>>> item osd.10 weight 1.000 >>>>>>> item osd.11 weight 1.000 >>>>>>> item osd.12 weight 1.000 >>>>>>> item osd.13 weight 1.000 >>>>>>> item osd.14 weight 1.000 >>>>>>> item osd.4 weight 1.000 >>>>>>> item osd.47 weight 1.000 >>>>>>> item osd.48 weight 1.000 >>>>>>> item osd.49 weight 1.000 >>>>>>> item osd.50 weight 1.000 >>>>>>> item osd.51 weight 1.000 >>>>>>> } >>>>>>> host cstore2 { >>>>>>> id -11 # do not change unnecessarily >>>>>>> # weight 20.000 >>>>>>> alg straw >>>>>>> hash 0 # rjenkins1 >>>>>>> item osd.15 weight 1.000 >>>>>>> item osd.16 weight 1.000 >>>>>>> item osd.17 weight 1.000 >>>>>>> item osd.18 weight 1.000 >>>>>>> item osd.19 weight 1.000 >>>>>>> item osd.20 weight 1.000 >>>>>>> item osd.21 weight 1.000 >>>>>>> item osd.22 weight 1.000 >>>>>>> item osd.23 weight 1.000 >>>>>>> item osd.24 weight 1.000 >>>>>>> item osd.41 weight 1.000 >>>>>>> item osd.42 weight 1.000 >>>>>>> item osd.43 weight 1.000 >>>>>>> item osd.44 weight 1.000 >>>>>>> item osd.45 weight 1.000 >>>>>>> item osd.46 weight 1.000 >>>>>>> item osd.52 weight 1.000 >>>>>>> item osd.53 weight 1.000 >>>>>>> item osd.54 weight 1.000 >>>>>>> item osd.55 weight 1.000 >>>>>>> item osd.56 weight 0.000 >>>>>>> item osd.57 weight 0.000 >>>>>>> } >>>>>>> host cstore3 { >>>>>>> id -12 # do not change unnecessarily >>>>>>> # weight 16.000 >>>>>>> alg straw >>>>>>> hash 0 # rjenkins1 >>>>>>> item osd.25 weight 1.000 >>>>>>> item osd.26 weight 1.000 >>>>>>> item osd.27 weight 1.000 >>>>>>> item osd.28 weight 1.000 >>>>>>> item osd.29 weight 1.000 >>>>>>> item osd.30 weight 1.000 >>>>>>> item osd.31 weight 1.000 >>>>>>> item osd.32 weight 1.000 >>>>>>> item osd.33 weight 1.000 >>>>>>> item osd.34 weight 1.000 >>>>>>> item osd.35 weight 1.000 >>>>>>> item osd.36 weight 1.000 >>>>>>> item osd.37 weight 1.000 >>>>>>> item osd.38 weight 1.000 >>>>>>> item osd.39 weight 1.000 >>>>>>> item osd.40 weight 1.000 >>>>>>> } >>>>>>> host cstore4 { >>>>>>> id -13 # do not change unnecessarily >>>>>>> # weight 7.640 >>>>>>> alg straw >>>>>>> hash 0 # rjenkins1 >>>>>>> item osd.62 weight 0.550 >>>>>>> item osd.63 weight 0.550 >>>>>>> item osd.64 weight 0.550 >>>>>>> item osd.65 weight 0.550 >>>>>>> item osd.66 weight 0.550 >>>>>>> item osd.67 weight 0.550 >>>>>>> item osd.68 weight 0.550 >>>>>>> item osd.69 weight 0.550 >>>>>>> item osd.70 weight 0.270 >>>>>>> item osd.71 weight 0.270 >>>>>>> item osd.72 weight 0.270 >>>>>>> item osd.73 weight 0.270 >>>>>>> item osd.74 weight 0.270 >>>>>>> item osd.75 weight 0.270 >>>>>>> item osd.76 weight 0.270 >>>>>>> item osd.77 weight 0.270 >>>>>>> item osd.78 weight 0.270 >>>>>>> item osd.79 weight 0.270 >>>>>>> item osd.80 weight 0.270 >>>>>>> item osd.81 weight 0.270 >>>>>>> } >>>>>>> host cstore5 { >>>>>>> id -14 # do not change unnecessarily >>>>>>> # weight 3.320 >>>>>>> alg straw >>>>>>> hash 0 # rjenkins1 >>>>>>> item osd.0 weight 0.400 >>>>>>> item osd.1 weight 0.500 >>>>>>> item osd.3 weight 0.400 >>>>>>> item osd.58 weight 0.400 >>>>>>> item osd.59 weight 0.540 >>>>>>> item osd.60 weight 0.540 >>>>>>> item osd.61 weight 0.540 >>>>>>> } >>>>>>> rack unknownrack { >>>>>>> id -3 # do not change unnecessarily >>>>>>> # weight 67.960 >>>>>>> alg straw >>>>>>> hash 0 # rjenkins1 >>>>>>> item gstore5 weight 5.000 >>>>>>> item cstore1 weight 16.000 >>>>>>> item cstore2 weight 20.000 >>>>>>> item cstore3 weight 16.000 >>>>>>> item cstore4 weight 7.640 >>>>>>> item cstore5 weight 3.320 >>>>>>> } >>>>>>> pool default { >>>>>>> id -1 # do not change unnecessarily >>>>>>> # weight 67.960 >>>>>>> alg straw >>>>>>> hash 0 # rjenkins1 >>>>>>> item unknownrack weight 67.960 >>>>>>> } >>>>>>> host gstore1 { >>>>>>> id -6 # do not change unnecessarily >>>>>>> # weight 32.760 >>>>>>> alg straw >>>>>>> hash 0 # rjenkins1 >>>>>>> item osd.82 weight 2.730 >>>>>>> item osd.83 weight 2.730 >>>>>>> item osd.84 weight 2.730 >>>>>>> item osd.85 weight 2.730 >>>>>>> item osd.86 weight 2.730 >>>>>>> item osd.87 weight 2.730 >>>>>>> item osd.88 weight 2.730 >>>>>>> item osd.89 weight 2.730 >>>>>>> item osd.90 weight 2.730 >>>>>>> item osd.91 weight 2.730 >>>>>>> item osd.92 weight 2.730 >>>>>>> item osd.93 weight 2.730 >>>>>>> } >>>>>>> datacenter datacenter-cod { >>>>>>> id -4 # do not change unnecessarily >>>>>>> # weight 32.760 >>>>>>> alg straw >>>>>>> hash 0 # rjenkins1 >>>>>>> item gstore1 weight 32.760 >>>>>>> } >>>>>>> host gstore2 { >>>>>>> id -7 # do not change unnecessarily >>>>>>> # weight 32.760 >>>>>>> alg straw >>>>>>> hash 0 # rjenkins1 >>>>>>> item osd.94 weight 2.730 >>>>>>> item osd.95 weight 2.730 >>>>>>> item osd.96 weight 2.730 >>>>>>> item osd.97 weight 2.730 >>>>>>> item osd.98 weight 2.730 >>>>>>> item osd.99 weight 2.730 >>>>>>> item osd.100 weight 2.730 >>>>>>> item osd.101 weight 2.730 >>>>>>> item osd.102 weight 2.730 >>>>>>> item osd.103 weight 2.730 >>>>>>> item osd.104 weight 2.730 >>>>>>> item osd.105 weight 2.730 >>>>>>> } >>>>>>> datacenter datacenter-rcod { >>>>>>> id -5 # do not change unnecessarily >>>>>>> # weight 32.760 >>>>>>> alg straw >>>>>>> hash 0 # rjenkins1 >>>>>>> item gstore2 weight 32.760 >>>>>>> } >>>>>>> pool iscsi { >>>>>>> id -2 # do not change unnecessarily >>>>>>> # weight 65.520 >>>>>>> alg straw >>>>>>> hash 0 # rjenkins1 >>>>>>> item datacenter-cod weight 32.760 >>>>>>> item datacenter-rcod weight 32.760 >>>>>>> } >>>>>>> >>>>>>> # rules >>>>>>> rule data { >>>>>>> ruleset 0 >>>>>>> type replicated >>>>>>> min_size 1 >>>>>>> max_size 10 >>>>>>> step take default >>>>>>> step chooseleaf firstn 0 type host >>>>>>> step emit >>>>>>> } >>>>>>> rule metadata { >>>>>>> ruleset 1 >>>>>>> type replicated >>>>>>> min_size 1 >>>>>>> max_size 10 >>>>>>> step take default >>>>>>> step chooseleaf firstn 0 type host >>>>>>> step emit >>>>>>> } >>>>>>> rule rbd { >>>>>>> ruleset 2 >>>>>>> type replicated >>>>>>> min_size 1 >>>>>>> max_size 10 >>>>>>> step take default >>>>>>> step chooseleaf firstn 0 type host >>>>>>> step emit >>>>>>> } >>>>>>> >>>>>>> # end crush map >>>>>>> >>>>>>> >>>>>>> Restarting mon don't help. >>>>>> _______________________________________________ >>>>>> ceph-users mailing list >>>>>> ceph-users@xxxxxxxxxxxxxx >>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com