MON segfaulting when setting a crush ruleset to a pool (firefly 0.80.4)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello,

I'm running a test cluster (mon and osd are debian 7 
with  3.2.57-3+deb7u2 kernel). The client is a debian 7 
with a 3.15.4 kernel that I compiled myself.

The cluster has 3 monitors and 16 osd servers.
I created a pool (periph) and used it a bit and then 
I decided to create some buckets and moved the hosts 
into :

$ ceph osd crush add-bucket rack1-pdu1 rack
$ ceph osd crush add-bucket rack1-pdu2 rack
$ ceph osd crush add-bucket rack1-pdu3 rack
$ ceph osd crush add-bucket rack2-pdu1 rack
$ ceph osd crush add-bucket rack2-pdu2 rack
$ ceph osd crush add-bucket rack2-pdu3 rack
$ ceph osd crush move ceph-osd0 rack=rack1-pdu1 
$ ceph osd crush move ceph-osd1 rack=rack1-pdu1 
$ ceph osd crush move ceph-osd2 rack=rack1-pdu1 
$ ceph osd crush move ceph-osd3 rack=rack1-pdu2 
$ ceph osd crush move ceph-osd4 rack=rack1-pdu2 
$ ceph osd crush move ceph-osd5 rack=rack1-pdu2 
$ ceph osd crush move ceph-osd6 rack=rack1-pdu3 
$ ceph osd crush move ceph-osd7 rack=rack1-pdu3 
$ ceph osd crush move ceph-osd8 rack=rack1-pdu3 
$ ceph osd crush move ceph-osd9 rack=rack2-pdu1 
$ ceph osd crush move ceph-osd10 rack=rack2-pdu1 
$ ceph osd crush move ceph-osd11 rack=rack2-pdu1 
$ ceph osd crush move ceph-osd12 rack=rack2-pdu2 
$ ceph osd crush move ceph-osd13 rack=rack2-pdu2 
$ ceph osd crush move ceph-osd14 rack=rack2-pdu2 
$ ceph osd crush move ceph-osd15 rack=rack2-pdu3 

It did well :

$ ceph osd tree
# id    weight  type name       up/down reweight
-23     0.91    rack rack2-pdu3
-17     0.91            host ceph-osd15
15      0.91                    osd.15  up      1
-22     1.81    rack rack2-pdu2
-14     0.45            host ceph-osd12
12      0.45                    osd.12  up      1
-15     0.45            host ceph-osd13
13      0.45                    osd.13  up      1
-16     0.91            host ceph-osd14
14      0.91                    osd.14  up      1
-21     1.35    rack rack2-pdu1
-11     0.45            host ceph-osd9
9       0.45                    osd.9   up      1
-12     0.45            host ceph-osd10
10      0.45                    osd.10  up      1
-13     0.45            host ceph-osd11
11      0.45                    osd.11  up      1
-20     1.35    rack rack1-pdu3
-8      0.45            host ceph-osd6
6       0.45                    osd.6   up      1
-9      0.45            host ceph-osd7
7       0.45                    osd.7   up      1
-10     0.45            host ceph-osd8
8       0.45                    osd.8   up      1
-19     1.35    rack rack1-pdu2
-5      0.45            host ceph-osd3
3       0.45                    osd.3   up      1
-6      0.45            host ceph-osd4
4       0.45                    osd.4   up      1
-7      0.45            host ceph-osd5
5       0.45                    osd.5   up      1
-18     1.35    rack rack1-pdu1
-2      0.45            host ceph-osd0
0       0.45                    osd.0   up      1
-3      0.45            host ceph-osd1
1       0.45                    osd.1   up      1
-4      0.45            host ceph-osd2
2       0.45                    osd.2   up      1
-1      0       root default


But then, when trying to set the crush_ruleset to the
pool with the command below it crashes two of the three
monitor.

 
$ ceph osd pool set periph crush_ruleset 2
2014-07-23 14:43:38.942811 7fa9696a3700  0 monclient: hunting for new mon

The first monitor come to :

    -4> 2014-07-23 14:43:37.476121 7f52d2f46700  1 -- 10.1.2.1:6789/0 --> 10.1.2.100:0/1027991 -- mon_command_ack([{"prefix": "get_command_descriptions"}]=0  v0) v1 -- ?+29681 0x3b1c780 con 0x2a578c0
    -3> 2014-07-23 14:43:37.598549 7f52d2f46700  1 -- 10.1.2.1:6789/0 <== client.39105 10.1.2.100:0/1027991 8 ==== mon_command({"var": "crush_ruleset", "prefix": "osd pool set", "pool": "periph", "val": "2"} v 0) v1 ==== 122+0+0 (2844980124 0 0) 0x3b1d860 con 0x2a578c0
    -2> 2014-07-23 14:43:37.598602 7f52d2f46700  0 mon.ceph-mon0 at 0(leader) e2 handle_command mon_command({"var": "crush_ruleset", "prefix": "osd pool set", "pool": "periph", "val": "2"} v 0) v1
    -1> 2014-07-23 14:43:37.598705 7f52d2f46700  1 mon.ceph-mon0 at 0(leader).paxos(paxos active c 542663..543338) is_readable now=2014-07-23 14:43:37.598708 lease_expire=2014-07-23 14:43:41.683421 has v0 lc 543338
     0> 2014-07-23 14:43:37.601706 7f52d2f46700 -1 *** Caught signal (Segmentation fault) **
 in thread 7f52d2f46700


Then, after an election try the second monitor goes down also :

     -1> 2014-07-23 14:43:51.772370 7eff4ba15700  1 mon.ceph-mon1 at 1(leader).paxos(paxos active c 542663..543338) is_readable now=2014-07-23 14:43:51.772373 lease_expire=2014-07-23 14:43:56.770906 has v0 lc 543338
     0> 2014-07-23 14:43:51.775817 7eff4ba15700 -1 *** Caught signal (Segmentation fault) **

I can not reactivate the monitors until the command 
"ceph osd pool set periph crush_ruleset 2" is running.

When I kill this command, then the monitors can run
again and retrieve a normal state but it leaves the
cluster with some warnings about data replacements
not achieved I guess (I had some data in the pool).

cluster 4a8669b9-b379-43b2-9488-7fca6e1366bc
     health HEALTH_WARN 152 pgs peering; 166 pgs stuck inactive; 620 pgs stuck unclean; recovery 620/694410 objects degraded (0.089%)
     monmap e2: 3 mons at {ceph-mon0=10.1.2.1:6789/0,ceph-mon1=10.1.2.2:6789/0,ceph-mon2=10.1.2.3:6789/0}, election epoch 50, quorum 0,1,2 ceph-mon0,ceph-mon1,ceph-mon2
     osdmap e1688: 16 osds: 16 up, 16 in
      pgmap v300875: 712 pgs, 5 pools, 1350 GB data, 338 kobjects
            2765 GB used, 5614 GB / 8379 GB avail
            620/694410 objects degraded (0.089%)
                  14 inactive
                 152 peering
                 454 active+remapped
                  92 active+clean

Is there something that I did wrong or forgot to do ?

While writing this mail down I realised that there is
only one host in the rack2-pdu3 rack. May this be a 
cause of the problem ?

Thanks for any hints.

-- 
Olivier DELHOMME

?cole nationale sup?rieure des mines de Paris.
Centre des mat?riaux Pierre-Marie FOURT.

10 rue Henry Desbru?res (Proche SNECMA)
91000 EVRY

Tel : +33 (0)1.60.76.30.48
Fax : +33 (0)1.60.76.31.50

"Quand la v?rit? n'est pas libre, la libert? n'est pas vraie."
					     [Jacques Pr?vert]



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux