Hi I have an OpenStack (pike) environment with a Ceph cluster (ceph version 12.2.5 luminous stable) deployed with kolla-ansible. When I was scaling the Ceph cluster, I found a read/write exception on the OpenStack VM. ( kolla-ansible uses the following script when expanding Ceph https://github.com/openstack/kolla/blob/pike-eol/docker/ceph/ceph-osd/extend_start.sh ) I ran 'ceph mon status' while checking the storage and found that ceph mon was constantly electing. I started to think it was a kolla issue, so I ran the script manually. The problem was reproduced when I executed 'ceph osd crush move 192.168.1.47 root=default '. I don't know if the logs of ceph mon can help to locate the problem. -82> 2022-04-21 23:30:35.329384 7f2f13b27700 5 mon.192.168.1.7@2(leader).paxos(paxos active c 83130286..83130896) is_readable = 1 - now=2022-04-21 23:30:35.329387 lease_expire=2022-04-21 23:30:39.793236 has v0 lc 83130896 -81> 2022-04-21 23:30:35.329438 7f2f13b27700 4 mon.192.168.1.7@2(leader).mgr e1702 beacon from 5534841203 -80> 2022-04-21 23:30:35.329464 7f2f13b27700 4 mon.192.168.1.7@2(leader).mgr e1702 beacon from 5534841203 -79> 2022-04-21 23:30:35.640375 7f2f10b21700 1 -- 192.168.1.7:6789/0 >> - conn(0x5560e28e3800 :6789 s=STATE_ACCEPTING pgs=0 cs=0 l=0)._process_connection sd=219 - -78> 2022-04-21 23:30:35.640745 7f2f10b21700 2 -- 192.168.1.7:6789/0 >> 192.168.1.13:0/1283852369 conn(0x5560e28e3800 :6789 s=STATE_ACCEPTING_WAIT_SEQ pgs=2 cs=1 l=1).handle_connect_msg accept write reply msg done -77> 2022-04-21 23:30:35.640838 7f2f10b21700 2 -- 192.168.1.7:6789/0 >> 192.168.1.13:0/1283852369 conn(0x5560e28e3800 :6789 s=STATE_ACCEPTING_WAIT_SEQ pgs=2 cs=1 l=1)._process_connection accept get newly_acked_seq 0 -76> 2022-04-21 23:30:35.641004 7f2f10b21700 5 -- 192.168.1.7:6789/0 >> 192.168.1.13:0/1283852369 conn(0x5560e28e3800 :6789 s=STATE_OPEN_MESSAGE_READ_FOOTER_AND_DISPATCH pgs=2 cs=1 l=1). rx client.? seq 1 0x5560e3380000 auth(proto 0 31 bytes epoch 0) v1 -75> 2022-04-21 23:30:35.641116 7f2f13b27700 1 -- 192.168.1.7:6789/0 <== client.? 192.168.1.13:0/1283852369 1 ==== auth(proto 0 31 bytes epoch 0) v1 ==== 61+0+0 (535495083 0 0) 0x5560e3380000 con 0x5560e28e3800 -74> 2022-04-21 23:30:35.641207 7f2f13b27700 5 mon.192.168.1.7@2(leader).paxos(paxos active c 83130286..83130896) is_readable = 1 - now=2022-04-21 23:30:35.641211 lease_expire=2022-04-21 23:30:39.793236 has v0 lc 83130896 -73> 2022-04-21 23:30:35.641353 7f2f13b27700 1 -- 192.168.1.7:6789/0 --> 192.168.1.13:0/1283852369 -- mon_map magic: 0 v1 -- 0x5560e30da000 con 0 -72> 2022-04-21 23:30:35.641448 7f2f13b27700 2 mon.192.168.1.7@2(leader) e3 send_reply 0x5560e3bd9320 0x5560e2a27200 auth_reply(proto 2 0 (0) Success) v1 -71> 2022-04-21 23:30:35.641496 7f2f13b27700 1 -- 192.168.1.7:6789/0 --> 192.168.1.13:0/1283852369 -- auth_reply(proto 2 0 (0) Success) v1 -- 0x5560e2a27200 con 0 -70> 2022-04-21 23:30:35.642447 7f2f10b21700 5 -- 192.168.1.7:6789/0 >> 192.168.1.13:0/1283852369 conn(0x5560e28e3800 :6789 s=STATE_OPEN_MESSAGE_READ_FOOTER_AND_DISPATCH pgs=2 cs=1 l=1). rx client.? seq 2 0x5560e2a27200 auth(proto 2 32 bytes epoch 0) v1 -69> 2022-04-21 23:30:35.642557 7f2f13b27700 1 -- 192.168.1.7:6789/0 <== client.? 192.168.1.13:0/1283852369 2 ==== auth(proto 2 32 bytes epoch 0) v1 ==== 62+0+0 (4221912666 0 0) 0x5560e2a27200 con 0x5560e28e3800 -68> 2022-04-21 23:30:35.642630 7f2f13b27700 5 mon.192.168.1.7@2(leader).paxos(paxos active c 83130286..83130896) is_readable = 1 - now=2022-04-21 23:30:35.642633 lease_expire=2022-04-21 23:30:39.793236 has v0 lc 83130896 -67> 2022-04-21 23:30:35.643069 7f2f13b27700 2 mon.192.168.1.7@2(leader) e3 send_reply 0x5560e3bd9320 0x5560e3380000 auth_reply(proto 2 0 (0) Success) v1 -66> 2022-04-21 23:30:35.643145 7f2f13b27700 1 -- 192.168.1.7:6789/0 --> 192.168.1.13:0/1283852369 -- auth_reply(proto 2 0 (0) Success) v1 -- 0x5560e3380000 con 0 -65> 2022-04-21 23:30:35.643677 7f2f10b21700 5 -- 192.168.1.7:6789/0 >> 192.168.1.13:0/1283852369 conn(0x5560e28e3800 :6789 s=STATE_OPEN_MESSAGE_READ_FOOTER_AND_DISPATCH pgs=2 cs=1 l=1). rx client.? seq 3 0x5560e3380000 auth(proto 2 165 bytes epoch 0) v1 -64> 2022-04-21 23:30:35.643771 7f2f13b27700 1 -- 192.168.1.7:6789/0 <== client.? 192.168.1.13:0/1283852369 3 ==== auth(proto 2 165 bytes epoch 0) v1 ==== 195+0+0 (4252124022 0 0) 0x5560e3380000 con 0x5560e28e3800 -63> 2022-04-21 23:30:35.643829 7f2f13b27700 5 mon.192.168.1.7@2(leader).paxos(paxos active c 83130286..83130896) is_readable = 1 - now=2022-04-21 23:30:35.643832 lease_expire=2022-04-21 23:30:39.793236 has v0 lc 83130896 -62> 2022-04-21 23:30:35.644448 7f2f13b27700 2 mon.192.168.1.7@2(leader) e3 send_reply 0x5560e3bd9320 0x5560e2a27200 auth_reply(proto 2 0 (0) Success) v1 -61> 2022-04-21 23:30:35.644499 7f2f13b27700 1 -- 192.168.1.7:6789/0 --> 192.168.1.13:0/1283852369 -- auth_reply(proto 2 0 (0) Success) v1 -- 0x5560e2a27200 con 0 -60> 2022-04-21 23:30:35.645154 7f2f10b21700 1 -- 192.168.1.7:6789/0 >> 192.168.1.13:0/1283852369 conn(0x5560e28e3800 :6789 s=STATE_OPEN pgs=2 cs=1 l=1).read_bulk peer close file descriptor 219 -59> 2022-04-21 23:30:35.645219 7f2f10b21700 1 -- 192.168.1.7:6789/0 >> 192.168.1.13:0/1283852369 conn(0x5560e28e3800 :6789 s=STATE_OPEN pgs=2 cs=1 l=1).read_until read failed -58> 2022-04-21 23:30:35.645242 7f2f10b21700 1 -- 192.168.1.7:6789/0 >> 192.168.1.13:0/1283852369 conn(0x5560e28e3800 :6789 s=STATE_OPEN pgs=2 cs=1 l=1).process read tag failed -57> 2022-04-21 23:30:35.645297 7f2f10b21700 1 -- 192.168.1.7:6789/0 >> 192.168.1.13:0/1283852369 conn(0x5560e28e3800 :6789 s=STATE_OPEN pgs=2 cs=1 l=1).fault on lossy channel, failing -56> 2022-04-21 23:30:35.645348 7f2f10b21700 2 -- 192.168.1.7:6789/0 >> 192.168.1.13:0/1283852369 conn(0x5560e28e3800 :6789 s=STATE_OPEN pgs=2 cs=1 l=1)._stop -55> 2022-04-21 23:30:36.274516 7f2f10320700 5 -- 192.168.1.7:6789/0 >> 192.168.1.11:6789/0 conn(0x5560e1e08800 :-1 s=STATE_OPEN_MESSAGE_READ_FOOTER_AND_DISPATCH pgs=958975211 cs=1 l=0). rx mon.3 seq 38735529 0x5560e329b440 mon_health( service 1 op tell e 0 r 0 ) v1 -54> 2022-04-21 23:30:36.274662 7f2f13b27700 1 -- 192.168.1.7:6789/0 <== mon.3 192.168.1.11:6789/0 38735529 ==== mon_health( service 1 op tell e 0 r 0 ) v1 ==== 108+0+0 (1769890505 0 0) 0x5560e329b440 con 0x5560e1e08800 -53> 2022-04-21 23:30:36.307747 7f2f10b21700 5 -- 192.168.1.7:6789/0 >> 192.168.1.13:6789/0 conn(0x5560e2396000 :-1 s=STATE_OPEN_MESSAGE_READ_FOOTER_AND_DISPATCH pgs=1023323427 cs=1 l=0). rx mon.4 seq 1911532605 0x5560e5a9cf00 forward(mgrbeacon mgr.a5-9c-stor-i620-1(4f7b3bf4-15fa-4ac0-97ab-bae34c68e12d,5535700725, -, 0) v6 caps allow profile mgr tid 29268 con_features 2305244844532236283) v3 -52> 2022-04-21 23:30:36.307882 7f2f13b27700 1 -- 192.168.1.7:6789/0 <== mon.4 192.168.1.13:6789/0 1911532605 ==== forward(mgrbeacon mgr.a5-9c-stor-i620-1(4f7b3bf4-15fa-4ac0-97ab-bae34c68e12d,5535700725, -, 0) v6 caps allow profile mgr tid 29268 con_features 2305244844532236283) v3 ==== 895+0+0 (3218645601 0 0) 0x5560e5a9cf00 con 0x5560e2396000 -51> 2022-04-21 23:30:36.308012 7f2f13b27700 5 mon.192.168.1.7@2(leader).paxos(paxos active c 83130286..83130896) is_readable = 1 - now=2022-04-21 23:30:36.308015 lease_expire=2022-04-21 23:30:39.793236 has v0 lc 83130896 -50> 2022-04-21 23:30:36.308072 7f2f13b27700 1 -- 192.168.1.7:6789/0 --> 192.168.1.13:6789/0 -- route(no-reply tid 29268) v3 -- 0x5560e329b440 con 0 -49> 2022-04-21 23:30:36.308135 7f2f13b27700 4 mon.192.168.1.7@2(leader).mgr e1702 beacon from 5535700725 -48> 2022-04-21 23:30:36.308221 7f2f13b27700 4 mon.192.168.1.7@2(leader).mgr e1702 beacon from 5535700725 -47> 2022-04-21 23:30:36.600706 7f2f10320700 5 -- 192.168.1.7:6789/0 >> 192.168.1.11:6789/0 conn(0x5560e1e08800 :-1 s=STATE_OPEN_MESSAGE_READ_FOOTER_AND_DISPATCH pgs=958975211 cs=1 l=0). rx mon.3 seq 38735530 0x5560e5834780 forward(mon_command({"prefix": "osd crush move", "args": ["root=default"], "name": "192.168.1.47"} v 0) v1 caps allow * tid 124914 con_features 2305244844532236283) v3 -46> 2022-04-21 23:30:36.600851 7f2f13b27700 1 -- 192.168.1.7:6789/0 <== mon.3 192.168.1.11:6789/0 38735530 ==== forward(mon_command({"prefix": "osd crush move", "args": ["root=default"], "name": "192.168.1.47"} v 0) v1 caps allow * tid 124914 con_features 2305244844532236283) v3 ==== 288+0+0 (1352957886 0 0) 0x5560e5834780 con 0x5560e1e08800 -45> 2022-04-21 23:30:36.600974 7f2f10320700 5 -- 192.168.1.7:6789/0 >> 192.168.1.11:6789/0 conn(0x5560e1e08800 :-1 s=STATE_OPEN_MESSAGE_READ_FOOTER_AND_DISPATCH pgs=958975211 cs=1 l=0). rx mon.3 seq 38735531 0x5560e5a91400 forward(log(1 entries from seq 9552 at 2022-04-21 23:30:36.599776) v1 caps allow * tid 124915 con_features 2305244844532236283) v3 -44> 2022-04-21 23:30:36.601144 7f2f13b27700 0 mon.192.168.1.7@2(leader) e3 handle_command mon_command({"prefix": "osd crush move", "args": ["root=default"], "name": "192.168.1.47"} v 0) v1 -43> 2022-04-21 23:30:36.601334 7f2f13b27700 0 log_channel(audit) log [INF] : from='client.5563480993 -' entity='client.admin' cmd=[{"prefix": "osd crush move", "args": ["root=default"], "name": "192.168.1.47"}]: dispatch -42> 2022-04-21 23:30:36.601361 7f2f13b27700 10 log_client _send_to_monlog to self -41> 2022-04-21 23:30:36.601378 7f2f13b27700 10 log_client log_queue is 1 last_log 272 sent 271 num 1 unsent 1 sending 1 -40> 2022-04-21 23:30:36.601390 7f2f13b27700 10 log_client will send 2022-04-21 23:30:36.601358 mon.192.168.1.7 mon.2 192.168.1.7:6789/0 272 : audit [INF] from='client.5563480993 -' entity='client.admin' cmd=[{"prefix": "osd crush move", "args": ["root=default"], "name": "192.168.1.47"}]: dispatch -39> 2022-04-21 23:30:36.601477 7f2f13b27700 1 -- 192.168.1.7:6789/0 --> 192.168.1.7:6789/0 -- log(1 entries from seq 272 at 2022-04-21 23:30:36.601358) v1 -- 0x5560e26f0900 con 0 -38> 2022-04-21 23:30:36.601567 7f2f13b27700 5 mon.192.168.1.7@2(leader).paxos(paxos active c 83130286..83130896) is_readable = 1 - now=2022-04-21 23:30:36.601570 lease_expire=2022-04-21 23:30:39.793236 has v0 lc 83130896 -37> 2022-04-21 23:30:36.623953 7f2f13b27700 0 mon.192.168.1.7@2(leader).osd e47733 moving crush item name '192.168.1.47' to location {root=default} -36> 2022-04-21 23:30:36.627511 7f2f13b27700 5 check_item_loc item -11 loc {root=default} -35> 2022-04-21 23:30:36.627546 7f2f13b27700 2 warning: did not specify location for 'host' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -34> 2022-04-21 23:30:36.627573 7f2f13b27700 2 warning: did not specify location for 'chassis' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -33> 2022-04-21 23:30:36.627593 7f2f13b27700 2 warning: did not specify location for 'rack' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -32> 2022-04-21 23:30:36.627619 7f2f13b27700 2 warning: did not specify location for 'row' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -31> 2022-04-21 23:30:36.627642 7f2f13b27700 2 warning: did not specify location for 'pdu' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -30> 2022-04-21 23:30:36.627665 7f2f13b27700 2 warning: did not specify location for 'pod' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -29> 2022-04-21 23:30:36.627687 7f2f13b27700 2 warning: did not specify location for 'room' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -28> 2022-04-21 23:30:36.627709 7f2f13b27700 2 warning: did not specify location for 'datacenter' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -27> 2022-04-21 23:30:36.627736 7f2f13b27700 2 warning: did not specify location for 'region' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -26> 2022-04-21 23:30:36.627799 7f2f13b27700 5 adjust_item_weight -12 weight 0 -25> 2022-04-21 23:30:36.627816 7f2f13b27700 5 choose_args_adjust_item_weight -11 weight [0] -24> 2022-04-21 23:30:36.627837 7f2f13b27700 5 check_item_loc item -11 loc {root=demo} -23> 2022-04-21 23:30:36.627844 7f2f13b27700 2 warning: did not specify location for 'host' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -22> 2022-04-21 23:30:36.627864 7f2f13b27700 2 warning: did not specify location for 'chassis' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -21> 2022-04-21 23:30:36.627888 7f2f13b27700 2 warning: did not specify location for 'rack' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -20> 2022-04-21 23:30:36.627923 7f2f13b27700 2 warning: did not specify location for 'row' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -19> 2022-04-21 23:30:36.627958 7f2f13b27700 2 warning: did not specify location for 'pdu' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -18> 2022-04-21 23:30:36.627980 7f2f13b27700 2 warning: did not specify location for 'pod' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -17> 2022-04-21 23:30:36.628002 7f2f13b27700 2 warning: did not specify location for 'room' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -16> 2022-04-21 23:30:36.628024 7f2f13b27700 2 warning: did not specify location for 'datacenter' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -15> 2022-04-21 23:30:36.628046 7f2f13b27700 2 warning: did not specify location for 'region' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -14> 2022-04-21 23:30:36.628093 7f2f13b27700 5 insert_item item -11 weight 0 name 192.168.1.47 loc {root=default} -13> 2022-04-21 23:30:36.628131 7f2f13b27700 2 warning: did not specify location for 'host' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -12> 2022-04-21 23:30:36.628155 7f2f13b27700 2 warning: did not specify location for 'chassis' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -11> 2022-04-21 23:30:36.628177 7f2f13b27700 2 warning: did not specify location for 'rack' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -10> 2022-04-21 23:30:36.628200 7f2f13b27700 2 warning: did not specify location for 'row' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -9> 2022-04-21 23:30:36.628222 7f2f13b27700 2 warning: did not specify location for 'pdu' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -8> 2022-04-21 23:30:36.628243 7f2f13b27700 2 warning: did not specify location for 'pod' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -7> 2022-04-21 23:30:36.628266 7f2f13b27700 2 warning: did not specify location for 'room' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -6> 2022-04-21 23:30:36.628289 7f2f13b27700 2 warning: did not specify location for 'datacenter' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -5> 2022-04-21 23:30:36.628319 7f2f13b27700 2 warning: did not specify location for 'region' level (levels are {0=osd,1=host,2=chassis,3=rack,4=row,5=pdu,6=pod,7=room,8=datacenter,9=region,10=root}) -4> 2022-04-21 23:30:36.628339 7f2f13b27700 5 insert_item adding -11 weight 0 to bucket -1 -3> 2022-04-21 23:30:36.628427 7f2f13b27700 5 adjust_item_weight_in_loc -11 weight 0 in {root=default} -2> 2022-04-21 23:30:36.628443 7f2f13b27700 5 adjust_item_weight_in_loc -11 diff 0 in bucket -1 -1> 2022-04-21 23:30:36.628448 7f2f13b27700 5 adjust_item_weight -1 weight 4482660 0> 2022-04-21 23:30:36.647294 7f2f13b27700 -1 *** Caught signal (Segmentation fault) ** in thread 7f2f13b27700 thread_name:ms_dispatch ceph version 12.2.5 (cad919881333ac92274171586c827e01f554a70a) luminous (stable) 1: (()+0x8f4d31) [0x5560d751ed31] 2: (()+0xf6d0) [0x7f2f1ce836d0] 3: (CrushWrapper::device_class_clone(int, int, std::map<int, std::map<int, int, std::less<int>, std::allocator<std::pair<int const, int> > >, std::less<int>, std::allocator<std::pair<int const, std::map<int, int, std::less<int>, std::allocator<std::pair<int const, int> > > > > > const&, std::set<int, std::less<int>, std::allocator<int> > const&, int*, std::map<int, std::map<int, std::vector<int, std::allocator<int> >, std::less<int>, std::allocator<std::pair<int const, std::vector<int, std::allocator<int> > > > >, std::less<int>, std::allocator<std::pair<int const, std::map<int, std::vector<int, std::allocator<int> >, std::less<int>, std::allocator<std::pair<int const, std::vector<int, std::allocator<int> > > > > > > >*)+0xa87) [0x5560d7496eb7] 4: (CrushWrapper::populate_classes(std::map<int, std::map<int, int, std::less<int>, std::allocator<std::pair<int const, int> > >, std::less<int>, std::allocator<std::pair<int const, std::map<int, int, std::less<int>, std::allocator<std::pair<int const, int> > > > > > const&)+0x1cf) [0x5560d74974bf] 5: (CrushWrapper::rebuild_roots_with_classes()+0xfe) [0x5560d749766e] 6: (CrushWrapper::insert_item(CephContext*, int, float, std::string, std::map<std::string, std::string, std::less<std::string>, std::allocator<std::pair<std::string const, std::string> > > const&)+0x78f) [0x5560d74993af] 7: (CrushWrapper::move_bucket(CephContext*, int, std::map<std::string, std::string, std::less<std::string>, std::allocator<std::pair<std::string const, std::string> > > const&)+0xc1) [0x5560d7499eb1] 8: (OSDMonitor::prepare_command_impl(boost::intrusive_ptr<MonOpRequest>, std::map<std::string, boost::variant<std::string, bool, long, double, std::vector<std::string, std::allocator<std::string> >, std::vector<long, std::allocator<long> >, std::vector<double, std::allocator<double> > >, std::less<std::string>, std::allocator<std::pair<std::string const, boost::variant<std::string, bool, long, double, std::vector<std::string, std::allocator<std::string> >, std::vector<long, std::allocator<long> >, std::vector<double, std::allocator<double> > > > > >&)+0x4dd2) [0x5560d7163f82] 9: (OSDMonitor::prepare_command(boost::intrusive_ptr<MonOpRequest>)+0x647) [0x5560d717ed57] 10: (OSDMonitor::prepare_update(boost::intrusive_ptr<MonOpRequest>)+0x39e) [0x5560d717f4be] 11: (PaxosService::dispatch(boost::intrusive_ptr<MonOpRequest>)+0xaf8) [0x5560d710b3e8] 12: (Monitor::handle_command(boost::intrusive_ptr<MonOpRequest>)+0x1d5b) [0x5560d6fe633b] 13: (Monitor::dispatch_op(boost::intrusive_ptr<MonOpRequest>)+0x919) [0x5560d6febee9] 14: (Monitor::_ms_dispatch(Message*)+0x7eb) [0x5560d6fed16b] 15: (Monitor::handle_forward(boost::intrusive_ptr<MonOpRequest>)+0xa8d) [0x5560d6feea7d] 16: (Monitor::dispatch_op(boost::intrusive_ptr<MonOpRequest>)+0xdbd) [0x5560d6fec38d] 17: (Monitor::_ms_dispatch(Message*)+0x7eb) [0x5560d6fed16b] 18: (Monitor::ms_dispatch(Message*)+0x23) [0x5560d70192d3] 19: (DispatchQueue::entry()+0x792) [0x5560d74ca0d2] 20: (DispatchQueue::DispatchThread::entry()+0xd) [0x5560d72c349d] 21: (()+0x7e25) [0x7f2f1ce7be25] 22: (clone()+0x6d) [0x7f2f1a286bad] NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this. Thanks! _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx