Sorry for the late message, I'm back from a short vacation. I would like to try it this weekends. Thanks for your patient :-) On Wed, Sep 3, 2014 at 9:16 PM, Kenneth Waegeman <Kenneth.Waegeman at ugent.be> wrote: > I also can reproduce it on a new slightly different set up (also EC on KV > and Cache) by running ceph pg scrub on a KV pg: this pg will then get the > 'inconsistent' status > > > > ----- Message from Kenneth Waegeman <Kenneth.Waegeman at UGent.be> --------- > Date: Mon, 01 Sep 2014 16:28:31 +0200 > From: Kenneth Waegeman <Kenneth.Waegeman at UGent.be> > Subject: Re: ceph cluster inconsistency keyvaluestore > To: Haomai Wang <haomaiwang at gmail.com> > Cc: ceph-users at lists.ceph.com > > > >> Hi, >> >> >> The cluster got installed with quattor, which uses ceph-deploy for >> installation of daemons, writes the config file and installs the crushmap. >> I have 3 hosts, each 12 disks, having a large KV partition (3.6T) for the >> ECdata pool and a small cache partition (50G) for the cache >> >> I manually did this: >> >> ceph osd pool create cache 1024 1024 >> ceph osd pool set cache size 2 >> ceph osd pool set cache min_size 1 >> ceph osd erasure-code-profile set profile11 k=8 m=3 >> ruleset-failure-domain=osd >> ceph osd pool create ecdata 128 128 erasure profile11 >> ceph osd tier add ecdata cache >> ceph osd tier cache-mode cache writeback >> ceph osd tier set-overlay ecdata cache >> ceph osd pool set cache hit_set_type bloom >> ceph osd pool set cache hit_set_count 1 >> ceph osd pool set cache hit_set_period 3600 >> ceph osd pool set cache target_max_bytes $((280*1024*1024*1024)) >> >> (But the previous time I had the problem already without the cache part) >> >> >> >> Cluster live since 2014-08-29 15:34:16 >> >> Config file on host ceph001: >> >> [global] >> auth_client_required = cephx >> auth_cluster_required = cephx >> auth_service_required = cephx >> cluster_network = 10.143.8.0/24 >> filestore_xattr_use_omap = 1 >> fsid = 82766e04-585b-49a6-a0ac-c13d9ffd0a7d >> mon_cluster_log_to_syslog = 1 >> mon_host = ceph001.cubone.os, ceph002.cubone.os, ceph003.cubone.os >> mon_initial_members = ceph001, ceph002, ceph003 >> osd_crush_update_on_start = 0 >> osd_journal_size = 10240 >> osd_pool_default_min_size = 2 >> osd_pool_default_pg_num = 512 >> osd_pool_default_pgp_num = 512 >> osd_pool_default_size = 3 >> public_network = 10.141.8.0/24 >> >> [osd.11] >> osd_objectstore = keyvaluestore-dev >> >> [osd.13] >> osd_objectstore = keyvaluestore-dev >> >> [osd.15] >> osd_objectstore = keyvaluestore-dev >> >> [osd.17] >> osd_objectstore = keyvaluestore-dev >> >> [osd.19] >> osd_objectstore = keyvaluestore-dev >> >> [osd.21] >> osd_objectstore = keyvaluestore-dev >> >> [osd.23] >> osd_objectstore = keyvaluestore-dev >> >> [osd.25] >> osd_objectstore = keyvaluestore-dev >> >> [osd.3] >> osd_objectstore = keyvaluestore-dev >> >> [osd.5] >> osd_objectstore = keyvaluestore-dev >> >> [osd.7] >> osd_objectstore = keyvaluestore-dev >> >> [osd.9] >> osd_objectstore = keyvaluestore-dev >> >> >> OSDs: >> # id weight type name up/down reweight >> -12 140.6 root default-cache >> -9 46.87 host ceph001-cache >> 2 3.906 osd.2 up 1 >> 4 3.906 osd.4 up 1 >> 6 3.906 osd.6 up 1 >> 8 3.906 osd.8 up 1 >> 10 3.906 osd.10 up 1 >> 12 3.906 osd.12 up 1 >> 14 3.906 osd.14 up 1 >> 16 3.906 osd.16 up 1 >> 18 3.906 osd.18 up 1 >> 20 3.906 osd.20 up 1 >> 22 3.906 osd.22 up 1 >> 24 3.906 osd.24 up 1 >> -10 46.87 host ceph002-cache >> 28 3.906 osd.28 up 1 >> 30 3.906 osd.30 up 1 >> 32 3.906 osd.32 up 1 >> 34 3.906 osd.34 up 1 >> 36 3.906 osd.36 up 1 >> 38 3.906 osd.38 up 1 >> 40 3.906 osd.40 up 1 >> 42 3.906 osd.42 up 1 >> 44 3.906 osd.44 up 1 >> 46 3.906 osd.46 up 1 >> 48 3.906 osd.48 up 1 >> 50 3.906 osd.50 up 1 >> -11 46.87 host ceph003-cache >> 54 3.906 osd.54 up 1 >> 56 3.906 osd.56 up 1 >> 58 3.906 osd.58 up 1 >> 60 3.906 osd.60 up 1 >> 62 3.906 osd.62 up 1 >> 64 3.906 osd.64 up 1 >> 66 3.906 osd.66 up 1 >> 68 3.906 osd.68 up 1 >> 70 3.906 osd.70 up 1 >> 72 3.906 osd.72 up 1 >> 74 3.906 osd.74 up 1 >> 76 3.906 osd.76 up 1 >> -8 140.6 root default-ec >> -5 46.87 host ceph001-ec >> 3 3.906 osd.3 up 1 >> 5 3.906 osd.5 up 1 >> 7 3.906 osd.7 up 1 >> 9 3.906 osd.9 up 1 >> 11 3.906 osd.11 up 1 >> 13 3.906 osd.13 up 1 >> 15 3.906 osd.15 up 1 >> 17 3.906 osd.17 up 1 >> 19 3.906 osd.19 up 1 >> 21 3.906 osd.21 up 1 >> 23 3.906 osd.23 up 1 >> 25 3.906 osd.25 up 1 >> -6 46.87 host ceph002-ec >> 29 3.906 osd.29 up 1 >> 31 3.906 osd.31 up 1 >> 33 3.906 osd.33 up 1 >> 35 3.906 osd.35 up 1 >> 37 3.906 osd.37 up 1 >> 39 3.906 osd.39 up 1 >> 41 3.906 osd.41 up 1 >> 43 3.906 osd.43 up 1 >> 45 3.906 osd.45 up 1 >> 47 3.906 osd.47 up 1 >> 49 3.906 osd.49 up 1 >> 51 3.906 osd.51 up 1 >> -7 46.87 host ceph003-ec >> 55 3.906 osd.55 up 1 >> 57 3.906 osd.57 up 1 >> 59 3.906 osd.59 up 1 >> 61 3.906 osd.61 up 1 >> 63 3.906 osd.63 up 1 >> 65 3.906 osd.65 up 1 >> 67 3.906 osd.67 up 1 >> 69 3.906 osd.69 up 1 >> 71 3.906 osd.71 up 1 >> 73 3.906 osd.73 up 1 >> 75 3.906 osd.75 up 1 >> 77 3.906 osd.77 up 1 >> -4 23.44 root default-ssd >> -1 7.812 host ceph001-ssd >> 0 3.906 osd.0 up 1 >> 1 3.906 osd.1 up 1 >> -2 7.812 host ceph002-ssd >> 26 3.906 osd.26 up 1 >> 27 3.906 osd.27 up 1 >> -3 7.812 host ceph003-ssd >> 52 3.906 osd.52 up 1 >> 53 3.906 osd.53 up 1 >> >> Cache OSDs are each 50G, the EC KV OSDS 3.6T, (ssds not used right now) >> >> Pools: >> pool 0 'rbd' replicated size 3 min_size 2 crush_ruleset 0 object_hash >> rjenkins pg_num 64 pgp_num 64 last_change 1 flags hashpspool stripe_width 0 >> pool 1 'cache' replicated size 2 min_size 1 crush_ruleset 0 object_hash >> rjenkins pg_num 1024 pgp_num 1024 last_change 174 flags >> hashpspool,incomplete_clones tier_of 2 cache_mode writeback target_bytes >> 300647710720 hit_set bloom{false_positive_probability: 0.05, target_size: 0, >> seed: 0} 3600s x1 stripe_width 0 >> pool 2 'ecdata' erasure size 11 min_size 8 crush_ruleset 2 object_hash >> rjenkins pg_num 128 pgp_num 128 last_change 170 lfor 170 flags hashpspool >> tiers 1 read_tier 1 write_tier 1 stripe_width 4096 >> >> >> Crushmap: >> # begin crush map >> tunable choose_local_fallback_tries 0 >> tunable choose_local_tries 0 >> tunable choose_total_tries 50 >> tunable chooseleaf_descend_once 1 >> >> # devices >> device 0 osd.0 >> device 1 osd.1 >> device 2 osd.2 >> device 3 osd.3 >> device 4 osd.4 >> device 5 osd.5 >> device 6 osd.6 >> device 7 osd.7 >> device 8 osd.8 >> device 9 osd.9 >> device 10 osd.10 >> device 11 osd.11 >> device 12 osd.12 >> device 13 osd.13 >> device 14 osd.14 >> device 15 osd.15 >> device 16 osd.16 >> device 17 osd.17 >> device 18 osd.18 >> device 19 osd.19 >> device 20 osd.20 >> device 21 osd.21 >> device 22 osd.22 >> device 23 osd.23 >> device 24 osd.24 >> device 25 osd.25 >> device 26 osd.26 >> device 27 osd.27 >> device 28 osd.28 >> device 29 osd.29 >> device 30 osd.30 >> device 31 osd.31 >> device 32 osd.32 >> device 33 osd.33 >> device 34 osd.34 >> device 35 osd.35 >> device 36 osd.36 >> device 37 osd.37 >> device 38 osd.38 >> device 39 osd.39 >> device 40 osd.40 >> device 41 osd.41 >> device 42 osd.42 >> device 43 osd.43 >> device 44 osd.44 >> device 45 osd.45 >> device 46 osd.46 >> device 47 osd.47 >> device 48 osd.48 >> device 49 osd.49 >> device 50 osd.50 >> device 51 osd.51 >> device 52 osd.52 >> device 53 osd.53 >> device 54 osd.54 >> device 55 osd.55 >> device 56 osd.56 >> device 57 osd.57 >> device 58 osd.58 >> device 59 osd.59 >> device 60 osd.60 >> device 61 osd.61 >> device 62 osd.62 >> device 63 osd.63 >> device 64 osd.64 >> device 65 osd.65 >> device 66 osd.66 >> device 67 osd.67 >> device 68 osd.68 >> device 69 osd.69 >> device 70 osd.70 >> device 71 osd.71 >> device 72 osd.72 >> device 73 osd.73 >> device 74 osd.74 >> device 75 osd.75 >> device 76 osd.76 >> device 77 osd.77 >> >> # types >> type 0 osd >> type 1 host >> type 2 root >> >> # buckets >> host ceph001-ssd { >> id -1 # do not change unnecessarily >> # weight 7.812 >> alg straw >> hash 0 # rjenkins1 >> item osd.0 weight 3.906 >> item osd.1 weight 3.906 >> } >> host ceph002-ssd { >> id -2 # do not change unnecessarily >> # weight 7.812 >> alg straw >> hash 0 # rjenkins1 >> item osd.26 weight 3.906 >> item osd.27 weight 3.906 >> } >> host ceph003-ssd { >> id -3 # do not change unnecessarily >> # weight 7.812 >> alg straw >> hash 0 # rjenkins1 >> item osd.52 weight 3.906 >> item osd.53 weight 3.906 >> } >> root default-ssd { >> id -4 # do not change unnecessarily >> # weight 23.436 >> alg straw >> hash 0 # rjenkins1 >> item ceph001-ssd weight 7.812 >> item ceph002-ssd weight 7.812 >> item ceph003-ssd weight 7.812 >> } >> host ceph001-ec { >> id -5 # do not change unnecessarily >> # weight 46.872 >> alg straw >> hash 0 # rjenkins1 >> item osd.3 weight 3.906 >> item osd.5 weight 3.906 >> item osd.7 weight 3.906 >> item osd.9 weight 3.906 >> item osd.11 weight 3.906 >> item osd.13 weight 3.906 >> item osd.15 weight 3.906 >> item osd.17 weight 3.906 >> item osd.19 weight 3.906 >> item osd.21 weight 3.906 >> item osd.23 weight 3.906 >> item osd.25 weight 3.906 >> } >> host ceph002-ec { >> id -6 # do not change unnecessarily >> # weight 46.872 >> alg straw >> hash 0 # rjenkins1 >> item osd.29 weight 3.906 >> item osd.31 weight 3.906 >> item osd.33 weight 3.906 >> item osd.35 weight 3.906 >> item osd.37 weight 3.906 >> item osd.39 weight 3.906 >> item osd.41 weight 3.906 >> item osd.43 weight 3.906 >> item osd.45 weight 3.906 >> item osd.47 weight 3.906 >> item osd.49 weight 3.906 >> item osd.51 weight 3.906 >> } >> host ceph003-ec { >> id -7 # do not change unnecessarily >> # weight 46.872 >> alg straw >> hash 0 # rjenkins1 >> item osd.55 weight 3.906 >> item osd.57 weight 3.906 >> item osd.59 weight 3.906 >> item osd.61 weight 3.906 >> item osd.63 weight 3.906 >> item osd.65 weight 3.906 >> item osd.67 weight 3.906 >> item osd.69 weight 3.906 >> item osd.71 weight 3.906 >> item osd.73 weight 3.906 >> item osd.75 weight 3.906 >> item osd.77 weight 3.906 >> } >> root default-ec { >> id -8 # do not change unnecessarily >> # weight 140.616 >> alg straw >> hash 0 # rjenkins1 >> item ceph001-ec weight 46.872 >> item ceph002-ec weight 46.872 >> item ceph003-ec weight 46.872 >> } >> host ceph001-cache { >> id -9 # do not change unnecessarily >> # weight 46.872 >> alg straw >> hash 0 # rjenkins1 >> item osd.2 weight 3.906 >> item osd.4 weight 3.906 >> item osd.6 weight 3.906 >> item osd.8 weight 3.906 >> item osd.10 weight 3.906 >> item osd.12 weight 3.906 >> item osd.14 weight 3.906 >> item osd.16 weight 3.906 >> item osd.18 weight 3.906 >> item osd.20 weight 3.906 >> item osd.22 weight 3.906 >> item osd.24 weight 3.906 >> } >> host ceph002-cache { >> id -10 # do not change unnecessarily >> # weight 46.872 >> alg straw >> hash 0 # rjenkins1 >> item osd.28 weight 3.906 >> item osd.30 weight 3.906 >> item osd.32 weight 3.906 >> item osd.34 weight 3.906 >> item osd.36 weight 3.906 >> item osd.38 weight 3.906 >> item osd.40 weight 3.906 >> item osd.42 weight 3.906 >> item osd.44 weight 3.906 >> item osd.46 weight 3.906 >> item osd.48 weight 3.906 >> item osd.50 weight 3.906 >> } >> host ceph003-cache { >> id -11 # do not change unnecessarily >> # weight 46.872 >> alg straw >> hash 0 # rjenkins1 >> item osd.54 weight 3.906 >> item osd.56 weight 3.906 >> item osd.58 weight 3.906 >> item osd.60 weight 3.906 >> item osd.62 weight 3.906 >> item osd.64 weight 3.906 >> item osd.66 weight 3.906 >> item osd.68 weight 3.906 >> item osd.70 weight 3.906 >> item osd.72 weight 3.906 >> item osd.74 weight 3.906 >> item osd.76 weight 3.906 >> } >> root default-cache { >> id -12 # do not change unnecessarily >> # weight 140.616 >> alg straw >> hash 0 # rjenkins1 >> item ceph001-cache weight 46.872 >> item ceph002-cache weight 46.872 >> item ceph003-cache weight 46.872 >> } >> >> # rules >> rule cache { >> ruleset 0 >> type replicated >> min_size 1 >> max_size 10 >> step take default-cache >> step chooseleaf firstn 0 type host >> step emit >> } >> rule metadata { >> ruleset 1 >> type replicated >> min_size 1 >> max_size 10 >> step take default-ssd >> step chooseleaf firstn 0 type host >> step emit >> } >> rule ecdata { >> ruleset 2 >> type erasure >> min_size 3 >> max_size 20 >> step set_chooseleaf_tries 5 >> step take default-ec >> step choose indep 0 type osd >> step emit >> } >> >> # end crush map >> >> The benchmarks I then did: >> >> ./benchrw 50000 >> >> benchrw: >> /usr/bin/rados -p ecdata bench $1 write --no-cleanup >> /usr/bin/rados -p ecdata bench $1 seq >> /usr/bin/rados -p ecdata bench $1 seq & >> /usr/bin/rados -p ecdata bench $1 write --no-cleanup >> >> >> Srubbing errors started soon after that: 2014-08-31 10:59:14 >> >> >> Please let me know if you need more information, and thanks ! >> >> Kenneth >> >> ----- Message from Haomai Wang <haomaiwang at gmail.com> --------- >> Date: Mon, 1 Sep 2014 21:30:16 +0800 >> From: Haomai Wang <haomaiwang at gmail.com> >> Subject: Re: ceph cluster inconsistency keyvaluestore >> To: Kenneth Waegeman <Kenneth.Waegeman at ugent.be> >> Cc: ceph-users at lists.ceph.com >> >> >>> Hmm, could you please list your instructions including cluster existing >>> time and all relevant ops? I want to reproduce it. >>> >>> >>> On Mon, Sep 1, 2014 at 4:45 PM, Kenneth Waegeman >>> <Kenneth.Waegeman at ugent.be> >>> wrote: >>> >>>> Hi, >>>> >>>> I reinstalled the cluster with 0.84, and tried again running rados bench >>>> on a EC coded pool on keyvaluestore. >>>> Nothing crashed this time, but when I check the status: >>>> >>>> health HEALTH_ERR 128 pgs inconsistent; 128 scrub errors; too few >>>> pgs >>>> per osd (15 < min 20) >>>> monmap e1: 3 mons at {ceph001=10.141.8.180:6789/0, >>>> ceph002=10.141.8.181:6789/0,ceph003=10.141.8.182:6789/0}, election epoch >>>> 8, quorum 0,1,2 ceph001,ceph002,ceph003 >>>> osdmap e174: 78 osds: 78 up, 78 in >>>> pgmap v147680: 1216 pgs, 3 pools, 14758 GB data, 3690 kobjects >>>> 1753 GB used, 129 TB / 131 TB avail >>>> 1088 active+clean >>>> 128 active+clean+inconsistent >>>> >>>> the 128 inconsistent pgs are ALL the pgs of the EC KV store ( the others >>>> are on Filestore) >>>> >>>> The only thing I can see in the logs is that after the rados tests, it >>>> start scrubbing, and for each KV pg I get something like this: >>>> >>>> 2014-08-31 11:14:09.050747 osd.11 10.141.8.180:6833/61098 4 : [ERR] >>>> 2.3s0 >>>> scrub stat mismatch, got 28164/29291 objects, 0/0 clones, 28164/29291 >>>> dirty, 0/0 omap, 0/0 hit_set_archive, 0/0 whiteouts, >>>> 118128377856/122855358464 bytes. >>>> >>>> What could here be the problem? >>>> Thanks again!! >>>> >>>> Kenneth >>>> >>>> >>>> ----- Message from Haomai Wang <haomaiwang at gmail.com> --------- >>>> Date: Tue, 26 Aug 2014 17:11:43 +0800 >>>> From: Haomai Wang <haomaiwang at gmail.com> >>>> Subject: Re: ceph cluster inconsistency? >>>> To: Kenneth Waegeman <Kenneth.Waegeman at ugent.be> >>>> Cc: ceph-users at lists.ceph.com >>>> >>>> >>>> Hmm, it looks like you hit this >>>> bug(http://tracker.ceph.com/issues/9223). >>>>> >>>>> >>>>> Sorry for the late message, I forget that this fix is merged into 0.84. >>>>> >>>>> Thanks for your patient :-) >>>>> >>>>> On Tue, Aug 26, 2014 at 4:39 PM, Kenneth Waegeman >>>>> <Kenneth.Waegeman at ugent.be> wrote: >>>>> >>>>>> >>>>>> Hi, >>>>>> >>>>>> In the meantime I already tried with upgrading the cluster to 0.84, to >>>>>> see >>>>>> if that made a difference, and it seems it does. >>>>>> I can't reproduce the crashing osds by doing a 'rados -p ecdata ls' >>>>>> anymore. >>>>>> >>>>>> But now the cluster detect it is inconsistent: >>>>>> >>>>>> cluster 82766e04-585b-49a6-a0ac-c13d9ffd0a7d >>>>>> health HEALTH_ERR 40 pgs inconsistent; 40 scrub errors; too few >>>>>> pgs >>>>>> per osd (4 < min 20); mon.ceph002 low disk space >>>>>> monmap e3: 3 mons at >>>>>> {ceph001=10.141.8.180:6789/0,ceph002=10.141.8.181:6789/0, >>>>>> ceph003=10.141.8.182:6789/0}, >>>>>> election epoch 30, quorum 0,1,2 ceph001,ceph002,ceph003 >>>>>> mdsmap e78951: 1/1/1 up {0=ceph003.cubone.os=up:active}, 3 >>>>>> up:standby >>>>>> osdmap e145384: 78 osds: 78 up, 78 in >>>>>> pgmap v247095: 320 pgs, 4 pools, 15366 GB data, 3841 kobjects >>>>>> 1502 GB used, 129 TB / 131 TB avail >>>>>> 279 active+clean >>>>>> 40 active+clean+inconsistent >>>>>> 1 active+clean+scrubbing+deep >>>>>> >>>>>> >>>>>> I tried to do ceph pg repair for all the inconsistent pgs: >>>>>> >>>>>> cluster 82766e04-585b-49a6-a0ac-c13d9ffd0a7d >>>>>> health HEALTH_ERR 40 pgs inconsistent; 1 pgs repair; 40 scrub >>>>>> errors; >>>>>> too few pgs per osd (4 < min 20); mon.ceph002 low disk space >>>>>> monmap e3: 3 mons at >>>>>> {ceph001=10.141.8.180:6789/0,ceph002=10.141.8.181:6789/0, >>>>>> ceph003=10.141.8.182:6789/0}, >>>>>> election epoch 30, quorum 0,1,2 ceph001,ceph002,ceph003 >>>>>> mdsmap e79486: 1/1/1 up {0=ceph003.cubone.os=up:active}, 3 >>>>>> up:standby >>>>>> osdmap e146452: 78 osds: 78 up, 78 in >>>>>> pgmap v248520: 320 pgs, 4 pools, 15366 GB data, 3841 kobjects >>>>>> 1503 GB used, 129 TB / 131 TB avail >>>>>> 279 active+clean >>>>>> 39 active+clean+inconsistent >>>>>> 1 active+clean+scrubbing+deep >>>>>> 1 active+clean+scrubbing+deep+inconsistent+repair >>>>>> >>>>>> I let it recovering through the night, but this morning the mons were >>>>>> all >>>>>> gone, nothing to see in the log files.. The osds were all still up! >>>>>> >>>>>> cluster 82766e04-585b-49a6-a0ac-c13d9ffd0a7d >>>>>> health HEALTH_ERR 36 pgs inconsistent; 1 pgs repair; 36 scrub >>>>>> errors; >>>>>> too few pgs per osd (4 < min 20) >>>>>> monmap e7: 3 mons at >>>>>> {ceph001=10.141.8.180:6789/0,ceph002=10.141.8.181:6789/0, >>>>>> ceph003=10.141.8.182:6789/0}, >>>>>> election epoch 44, quorum 0,1,2 ceph001,ceph002,ceph003 >>>>>> mdsmap e109481: 1/1/1 up {0=ceph003.cubone.os=up:active}, 3 >>>>>> up:standby >>>>>> osdmap e203410: 78 osds: 78 up, 78 in >>>>>> pgmap v331747: 320 pgs, 4 pools, 15251 GB data, 3812 kobjects >>>>>> 1547 GB used, 129 TB / 131 TB avail >>>>>> 1 active+clean+scrubbing+deep+inconsistent+repair >>>>>> 284 active+clean >>>>>> 35 active+clean+inconsistent >>>>>> >>>>>> I restarted the monitors now, I will let you know when I see something >>>>>> more.. >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> ----- Message from Haomai Wang <haomaiwang at gmail.com> --------- >>>>>> Date: Sun, 24 Aug 2014 12:51:41 +0800 >>>>>> >>>>>> From: Haomai Wang <haomaiwang at gmail.com> >>>>>> Subject: Re: ceph cluster inconsistency? >>>>>> To: Kenneth Waegeman <Kenneth.Waegeman at ugent.be>, >>>>>> ceph-users at lists.ceph.com >>>>>> >>>>>> >>>>>> It's really strange! I write a test program according the key ordering >>>>>>> >>>>>>> you provided and parse the corresponding value. It's true! >>>>>>> >>>>>>> I have no idea now. If free, could you add this debug code to >>>>>>> "src/os/GenericObjectMap.cc" and insert *before* "assert(start <= >>>>>>> header.oid);": >>>>>>> >>>>>>> dout(0) << "start: " << start << "header.oid: " << header.oid << >>>>>>> dendl; >>>>>>> >>>>>>> Then you need to recompile ceph-osd and run it again. The output log >>>>>>> can help it! >>>>>>> >>>>>>> On Tue, Aug 19, 2014 at 10:19 PM, Haomai Wang <haomaiwang at gmail.com> >>>>>>> wrote: >>>>>>> >>>>>>>> >>>>>>>> I feel a little embarrassed, 1024 rows still true for me. >>>>>>>> >>>>>>>> I was wondering if you could give your all keys via >>>>>>>> ""ceph-kvstore-tool /var/lib/ceph/osd/ceph-67/current/ list >>>>>>>> _GHOBJTOSEQ_ > keys.log?. >>>>>>>> >>>>>>>> thanks! >>>>>>>> >>>>>>>> On Tue, Aug 19, 2014 at 4:58 PM, Kenneth Waegeman >>>>>>>> <Kenneth.Waegeman at ugent.be> wrote: >>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> ----- Message from Haomai Wang <haomaiwang at gmail.com> --------- >>>>>>>>> Date: Tue, 19 Aug 2014 12:28:27 +0800 >>>>>>>>> >>>>>>>>> From: Haomai Wang <haomaiwang at gmail.com> >>>>>>>>> Subject: Re: ceph cluster inconsistency? >>>>>>>>> To: Kenneth Waegeman <Kenneth.Waegeman at ugent.be> >>>>>>>>> Cc: Sage Weil <sweil at redhat.com>, ceph-users at lists.ceph.com >>>>>>>>> >>>>>>>>> >>>>>>>>> On Mon, Aug 18, 2014 at 7:32 PM, Kenneth Waegeman >>>>>>>>>> >>>>>>>>>> <Kenneth.Waegeman at ugent.be> wrote: >>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> ----- Message from Haomai Wang <haomaiwang at gmail.com> --------- >>>>>>>>>>> Date: Mon, 18 Aug 2014 18:34:11 +0800 >>>>>>>>>>> >>>>>>>>>>> From: Haomai Wang <haomaiwang at gmail.com> >>>>>>>>>>> Subject: Re: ceph cluster inconsistency? >>>>>>>>>>> To: Kenneth Waegeman <Kenneth.Waegeman at ugent.be> >>>>>>>>>>> Cc: Sage Weil <sweil at redhat.com>, ceph-users at lists.ceph.com >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Mon, Aug 18, 2014 at 5:38 PM, Kenneth Waegeman >>>>>>>>>>>> >>>>>>>>>>>> <Kenneth.Waegeman at ugent.be> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> Hi, >>>>>>>>>>>>> >>>>>>>>>>>>> I tried this after restarting the osd, but I guess that was not >>>>>>>>>>>>> the >>>>>>>>>>>>> aim >>>>>>>>>>>>> ( >>>>>>>>>>>>> # ceph-kvstore-tool /var/lib/ceph/osd/ceph-67/current/ list >>>>>>>>>>>>> _GHOBJTOSEQ_| >>>>>>>>>>>>> grep 6adb1100 -A 100 >>>>>>>>>>>>> IO error: lock /var/lib/ceph/osd/ceph-67/current//LOCK: >>>>>>>>>>>>> Resource >>>>>>>>>>>>> temporarily >>>>>>>>>>>>> unavailable >>>>>>>>>>>>> tools/ceph_kvstore_tool.cc: In function >>>>>>>>>>>>> 'StoreTool::StoreTool(const >>>>>>>>>>>>> string&)' thread 7f8fecf7d780 time 2014-08-18 11:12:29.551780 >>>>>>>>>>>>> tools/ceph_kvstore_tool.cc: 38: FAILED >>>>>>>>>>>>> assert(!db_ptr->open(std::cerr)) >>>>>>>>>>>>> .. >>>>>>>>>>>>> ) >>>>>>>>>>>>> >>>>>>>>>>>>> When I run it after bringing the osd down, it takes a while, >>>>>>>>>>>>> but >>>>>>>>>>>>> it >>>>>>>>>>>>> has >>>>>>>>>>>>> no >>>>>>>>>>>>> output.. (When running it without the grep, I'm getting a huge >>>>>>>>>>>>> list >>>>>>>>>>>>> ) >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Oh, sorry for it! I made a mistake, the hash value(6adb1100) >>>>>>>>>>>> will >>>>>>>>>>>> be >>>>>>>>>>>> reversed into leveldb. >>>>>>>>>>>> So grep "benchmark_data_ceph001.cubone.os_5560_object789734" >>>>>>>>>>>> should >>>>>>>>>>>> be >>>>>>>>>>>> help it. >>>>>>>>>>>> >>>>>>>>>>>> this gives: >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> [root at ceph003 ~]# ceph-kvstore-tool /var/lib/ceph/osd/ceph-67/ >>>>>>>>>>> current/ >>>>>>>>>>> list >>>>>>>>>>> _GHOBJTOSEQ_ | grep 5560_object789734 -A 100 >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0011BDA6!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object789734!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0011C027!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object1330170!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0011C6FD!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_4919_object227366!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0011CB03!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object1363631!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0011CDF0!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object1573957!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0011D02C!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object1019282!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0011E2B5!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object1283563!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0011E511!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_4919_object273736!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0011E547!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object1170628!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0011EAAB!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_4919_object256335!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0011F446!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object1484196!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0011FC59!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object884178!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!001203F3!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object853746!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!001208E3!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object36633!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00120B37!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object1235337!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!001210B6!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object1661351!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!001210CB!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object238126!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012184C!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object339943!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00121916!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object1047094!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!001219C1!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object520642!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!001222BB!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object639565!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!001223AA!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_4919_object231080!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012243C!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object858050!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012289C!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object241796!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00122D28!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_4919_object7462!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00122DFE!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object243798!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00122EFC!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_8961_object109512!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!001232D7!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object653973!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!001234A3!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object1378169!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00123714!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object512925!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!001237D9!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_4919_object23289!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00123854!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object1108852!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00123971!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object704026!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00123F75!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_8961_object250441!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00124083!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object706178!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!001240FA!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object316952!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012447D!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object538734!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!001244D9!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object789215!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!001247CD!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_8961_object265993!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00124897!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object610597!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00124BE4!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object691723!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00124C9B!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object1306135!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00124E1D!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object520580!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012534C!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object659767!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00125A81!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object184060!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00125E77!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object1292867!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00126562!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object1201410!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00126B34!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object1657326!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00127383!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object1269787!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00127396!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object500115!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!001277F8!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object394932!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!001279DD!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_4919_object252963!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00127B40!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object936811!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00127BAC!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object1481773!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012894E!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object999885!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00128D05!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object943667!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012908A!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object212990!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00129519!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object437596!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00129716!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object1585330!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00129798!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object603505!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!001299C9!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object808800!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00129B7A!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object23193!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00129B9A!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object1158397!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012A932!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object542450!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012B77A!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_8961_object195480!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012BE8C!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_4919_object312911!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012BF74!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object1563783!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012C65C!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object1123980!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012C6FE!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_3411_object913!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012CCAD!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object400863!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012CDBB!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object789667!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012D14B!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object1020723!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012D95B!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_8961_object106293!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012E3C8!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object1355526!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012E5B3!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object1491348!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012F2BB!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_8961_object338872!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012F374!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object1337264!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012FBE5!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object1512395!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012FCE3!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_8961_object298610!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0012FEB6!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_4919_object120824!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!001301CA!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object816326!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00130263!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object777163!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00130529!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object1413173!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!001317D9!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object809510!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0013204F!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object471416!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00132400!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object695087!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00132A19!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object591945!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00132BF8!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object302000!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00132F5B!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object1645443!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00133B8B!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object761911!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!0013433E!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object1467727!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00134446!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object791960!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00134678!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object677078!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00134A96!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object254923!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!001355D0!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_31461_object321528!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00135690!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_4919_object36935!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00135B62!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object1228272!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00135C72!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_4812_object2180!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00135DEE!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object425705!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00136366!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object141569!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _GHOBJTOSEQ_:3%e0s0_head!00136371!!3!!benchmark_data_ >>>>>>>>>>> ceph001%ecubone%eos_5560_object564213!head >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> 100 rows seemed true for me. I found the min list objects is 1024. >>>>>>>>>> Please could you run >>>>>>>>>> "ceph-kvstore-tool /var/lib/ceph/osd/ceph-67/current/ list >>>>>>>>>> _GHOBJTOSEQ_| grep 6adb1100 -A 1024" >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> I got the output in attachment >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>>> >>>>>>>>>>>>> Or should I run this immediately after the osd is crashed, >>>>>>>>>>>>> (because >>>>>>>>>>>>> it >>>>>>>>>>>>> maybe >>>>>>>>>>>>> rebalanced? I did already restarted the cluster) >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> I don't know if it is related, but before I could all do that, >>>>>>>>>>>>> I >>>>>>>>>>>>> had >>>>>>>>>>>>> to >>>>>>>>>>>>> fix >>>>>>>>>>>>> something else: A monitor did run out if disk space, using 8GB >>>>>>>>>>>>> for >>>>>>>>>>>>> his >>>>>>>>>>>>> store.db folder (lot of sst files). Other monitors are also >>>>>>>>>>>>> near >>>>>>>>>>>>> that >>>>>>>>>>>>> level. >>>>>>>>>>>>> Never had that problem on previous setups before. I recreated a >>>>>>>>>>>>> monitor >>>>>>>>>>>>> and >>>>>>>>>>>>> now it uses 3.8GB. >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> It exists some duplicate data which needed to be compacted. >>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> Another idea, maybe you can make KeyValueStore's stripe size >>>>>>>>>>>> align >>>>>>>>>>>> with EC stripe size. >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> How can I do that? Is there some documentation about that? >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> ceph --show-config | grep keyvaluestore >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> debug_keyvaluestore = 0/0 >>>>>>>>>> keyvaluestore_queue_max_ops = 50 >>>>>>>>>> keyvaluestore_queue_max_bytes = 104857600 >>>>>>>>>> keyvaluestore_debug_check_backend = false >>>>>>>>>> keyvaluestore_op_threads = 2 >>>>>>>>>> keyvaluestore_op_thread_timeout = 60 >>>>>>>>>> keyvaluestore_op_thread_suicide_timeout = 180 >>>>>>>>>> keyvaluestore_default_strip_size = 4096 >>>>>>>>>> keyvaluestore_max_expected_write_size = 16777216 >>>>>>>>>> keyvaluestore_header_cache_size = 4096 >>>>>>>>>> keyvaluestore_backend = leveldb >>>>>>>>>> >>>>>>>>>> keyvaluestore_default_strip_size is the wanted >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> I haven't think deeply and maybe I will try it later. >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Thanks! >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> Kenneth >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> ----- Message from Sage Weil <sweil at redhat.com> --------- >>>>>>>>>>>>> Date: Fri, 15 Aug 2014 06:10:34 -0700 (PDT) >>>>>>>>>>>>> From: Sage Weil <sweil at redhat.com> >>>>>>>>>>>>> >>>>>>>>>>>>> Subject: Re: ceph cluster inconsistency? >>>>>>>>>>>>> To: Haomai Wang <haomaiwang at gmail.com> >>>>>>>>>>>>> Cc: Kenneth Waegeman <Kenneth.Waegeman at ugent.be>, >>>>>>>>>>>>> ceph-users at lists.ceph.com >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On Fri, 15 Aug 2014, Haomai Wang wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Hi Kenneth, >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> I don't find valuable info in your logs, it lack of the >>>>>>>>>>>>>>> necessary >>>>>>>>>>>>>>> debug output when accessing crash code. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> But I scan the encode/decode implementation in >>>>>>>>>>>>>>> GenericObjectMap >>>>>>>>>>>>>>> and >>>>>>>>>>>>>>> find something bad. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> For example, two oid has same hash and their name is: >>>>>>>>>>>>>>> A: "rb.data.123" >>>>>>>>>>>>>>> B: "rb-123" >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> In ghobject_t compare level, A < B. But GenericObjectMap >>>>>>>>>>>>>>> encode >>>>>>>>>>>>>>> "." >>>>>>>>>>>>>>> to >>>>>>>>>>>>>>> "%e", so the key in DB is: >>>>>>>>>>>>>>> A: _GHOBJTOSEQ_:blah!51615000!!none!!rb%edata%e123!head >>>>>>>>>>>>>>> B: _GHOBJTOSEQ_:blah!51615000!!none!!rb-123!head >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> A > B >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> And it seemed that the escape function is useless and should >>>>>>>>>>>>>>> be >>>>>>>>>>>>>>> disabled. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> I'm not sure whether Kenneth's problem is touching this bug. >>>>>>>>>>>>>>> Because >>>>>>>>>>>>>>> this scene only occur when the object set is very large and >>>>>>>>>>>>>>> make >>>>>>>>>>>>>>> the >>>>>>>>>>>>>>> two object has same hash value. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Kenneth, could you have time to run "ceph-kv-store >>>>>>>>>>>>>>> [path-to-osd] >>>>>>>>>>>>>>> list >>>>>>>>>>>>>>> _GHOBJTOSEQ_| grep 6adb1100 -A 100". ceph-kv-store is a debug >>>>>>>>>>>>>>> tool >>>>>>>>>>>>>>> which can be compiled from source. You can clone ceph repo >>>>>>>>>>>>>>> and >>>>>>>>>>>>>>> run >>>>>>>>>>>>>>> "./authongen.sh; ./configure; cd src; make >>>>>>>>>>>>>>> ceph-kvstore-tool". >>>>>>>>>>>>>>> "path-to-osd" should be "/var/lib/ceph/osd-[id]/current/". >>>>>>>>>>>>>>> "6adb1100" >>>>>>>>>>>>>>> is from your verbose log and the next 100 rows should know >>>>>>>>>>>>>>> necessary >>>>>>>>>>>>>>> infos. >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> You can also get ceph-kvstore-tool from the 'ceph-tests' >>>>>>>>>>>>>> package. >>>>>>>>>>>>>> >>>>>>>>>>>>>> Hi sage, do you think we need to provided with upgrade >>>>>>>>>>>>>> function >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> to >>>>>>>>>>>>>>> fix >>>>>>>>>>>>>>> it? >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> Hmm, we might. This only affects the key/value encoding >>>>>>>>>>>>>> right? >>>>>>>>>>>>>> The >>>>>>>>>>>>>> FileStore is using its own function to map these to file >>>>>>>>>>>>>> names? >>>>>>>>>>>>>> >>>>>>>>>>>>>> Can you open a ticket in the tracker for this? >>>>>>>>>>>>>> >>>>>>>>>>>>>> Thanks! >>>>>>>>>>>>>> sage >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Thu, Aug 14, 2014 at 7:36 PM, Kenneth Waegeman >>>>>>>>>>>>>>> <Kenneth.Waegeman at ugent.be> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> ----- Message from Haomai Wang <haomaiwang at gmail.com> >>>>>>>>>>>>>>>> --------- >>>>>>>>>>>>>>>> Date: Thu, 14 Aug 2014 19:11:55 +0800 >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> From: Haomai Wang <haomaiwang at gmail.com> >>>>>>>>>>>>>>>> Subject: Re: ceph cluster inconsistency? >>>>>>>>>>>>>>>> To: Kenneth Waegeman <Kenneth.Waegeman at ugent.be> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Could you add config "debug_keyvaluestore = 20/20" to the >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> crashed >>>>>>>>>>>>>>>>> osd >>>>>>>>>>>>>>>>> and replay the command causing crash? >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> I would like to get more debug infos! Thanks. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> I included the log in attachment! >>>>>>>>>>>>>>>> Thanks! >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> On Thu, Aug 14, 2014 at 4:41 PM, Kenneth Waegeman >>>>>>>>>>>>>>>>> <Kenneth.Waegeman at ugent.be> wrote: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> I have: >>>>>>>>>>>>>>>>>> osd_objectstore = keyvaluestore-dev >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> in the global section of my ceph.conf >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> [root at ceph002 ~]# ceph osd erasure-code-profile get >>>>>>>>>>>>>>>>>> profile11 >>>>>>>>>>>>>>>>>> directory=/usr/lib64/ceph/erasure-code >>>>>>>>>>>>>>>>>> k=8 >>>>>>>>>>>>>>>>>> m=3 >>>>>>>>>>>>>>>>>> plugin=jerasure >>>>>>>>>>>>>>>>>> ruleset-failure-domain=osd >>>>>>>>>>>>>>>>>> technique=reed_sol_van >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> the ecdata pool has this as profile >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> pool 3 'ecdata' erasure size 11 min_size 8 crush_ruleset 2 >>>>>>>>>>>>>>>>>> object_hash >>>>>>>>>>>>>>>>>> rjenkins pg_num 128 pgp_num 128 last_change 161 flags >>>>>>>>>>>>>>>>>> hashpspool >>>>>>>>>>>>>>>>>> stripe_width 4096 >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> ECrule in crushmap >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> rule ecdata { >>>>>>>>>>>>>>>>>> ruleset 2 >>>>>>>>>>>>>>>>>> type erasure >>>>>>>>>>>>>>>>>> min_size 3 >>>>>>>>>>>>>>>>>> max_size 20 >>>>>>>>>>>>>>>>>> step set_chooseleaf_tries 5 >>>>>>>>>>>>>>>>>> step take default-ec >>>>>>>>>>>>>>>>>> step choose indep 0 type osd >>>>>>>>>>>>>>>>>> step emit >>>>>>>>>>>>>>>>>> } >>>>>>>>>>>>>>>>>> root default-ec { >>>>>>>>>>>>>>>>>> id -8 # do not change unnecessarily >>>>>>>>>>>>>>>>>> # weight 140.616 >>>>>>>>>>>>>>>>>> alg straw >>>>>>>>>>>>>>>>>> hash 0 # rjenkins1 >>>>>>>>>>>>>>>>>> item ceph001-ec weight 46.872 >>>>>>>>>>>>>>>>>> item ceph002-ec weight 46.872 >>>>>>>>>>>>>>>>>> item ceph003-ec weight 46.872 >>>>>>>>>>>>>>>>>> ... >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Cheers! >>>>>>>>>>>>>>>>>> Kenneth >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> ----- Message from Haomai Wang <haomaiwang at gmail.com> >>>>>>>>>>>>>>>>>> --------- >>>>>>>>>>>>>>>>>> Date: Thu, 14 Aug 2014 10:07:50 +0800 >>>>>>>>>>>>>>>>>> From: Haomai Wang <haomaiwang at gmail.com> >>>>>>>>>>>>>>>>>> Subject: Re: ceph cluster inconsistency? >>>>>>>>>>>>>>>>>> To: Kenneth Waegeman <Kenneth.Waegeman at ugent.be> >>>>>>>>>>>>>>>>>> Cc: ceph-users <ceph-users at lists.ceph.com> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Hi Kenneth, >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Could you give your configuration related to EC and >>>>>>>>>>>>>>>>>>> KeyValueStore? >>>>>>>>>>>>>>>>>>> Not sure whether it's bug on KeyValueStore >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> On Thu, Aug 14, 2014 at 12:06 AM, Kenneth Waegeman >>>>>>>>>>>>>>>>>>> <Kenneth.Waegeman at ugent.be> wrote: >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> I was doing some tests with rados bench on a Erasure >>>>>>>>>>>>>>>>>>>> Coded >>>>>>>>>>>>>>>>>>>> pool >>>>>>>>>>>>>>>>>>>> (using >>>>>>>>>>>>>>>>>>>> keyvaluestore-dev objectstore) on 0.83, and I see some >>>>>>>>>>>>>>>>>>>> strangs >>>>>>>>>>>>>>>>>>>> things: >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> [root at ceph001 ~]# ceph status >>>>>>>>>>>>>>>>>>>> cluster 82766e04-585b-49a6-a0ac-c13d9ffd0a7d >>>>>>>>>>>>>>>>>>>> health HEALTH_WARN too few pgs per osd (4 < min 20) >>>>>>>>>>>>>>>>>>>> monmap e1: 3 mons at >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> {ceph001=10.141.8.180:6789/0,ceph002=10.141.8.181:6789/0, >>>>>>>>>>>>>>>>>>>> ceph003=10.141.8.182:6789/0}, >>>>>>>>>>>>>>>>>>>> election epoch 6, quorum 0,1,2 ceph001,ceph002,ceph003 >>>>>>>>>>>>>>>>>>>> mdsmap e116: 1/1/1 up >>>>>>>>>>>>>>>>>>>> {0=ceph001.cubone.os=up:active}, >>>>>>>>>>>>>>>>>>>> 2 >>>>>>>>>>>>>>>>>>>> up:standby >>>>>>>>>>>>>>>>>>>> osdmap e292: 78 osds: 78 up, 78 in >>>>>>>>>>>>>>>>>>>> pgmap v48873: 320 pgs, 4 pools, 15366 GB data, 3841 >>>>>>>>>>>>>>>>>>>> kobjects >>>>>>>>>>>>>>>>>>>> 1381 GB used, 129 TB / 131 TB avail >>>>>>>>>>>>>>>>>>>> 320 active+clean >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> There is around 15T of data, but only 1.3 T usage. >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> This is also visible in rados: >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> [root at ceph001 ~]# rados df >>>>>>>>>>>>>>>>>>>> pool name category KB objects >>>>>>>>>>>>>>>>>>>> clones >>>>>>>>>>>>>>>>>>>> degraded unfound rd rd KB >>>>>>>>>>>>>>>>>>>> wr >>>>>>>>>>>>>>>>>>>> wr >>>>>>>>>>>>>>>>>>>> KB >>>>>>>>>>>>>>>>>>>> data - 0 >>>>>>>>>>>>>>>>>>>> 0 >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> 0 0 0 0 0 >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> ecdata - 16113451009 >>>>>>>>>>>>>>>>>>>> 3933959 >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> 0 0 1 1 3935632 >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> 16116850711 >>>>>>>>>>>>>>>>>>>> metadata - 2 >>>>>>>>>>>>>>>>>>>> 20 >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> 0 0 33 36 21 >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> 8 >>>>>>>>>>>>>>>>>>>> rbd - 0 >>>>>>>>>>>>>>>>>>>> 0 >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> 0 0 0 0 0 >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> total used 1448266016 3933979 >>>>>>>>>>>>>>>>>>>> total avail 139400181016 >>>>>>>>>>>>>>>>>>>> total space 140848447032 >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Another (related?) thing: if I do rados -p ecdata ls, I >>>>>>>>>>>>>>>>>>>> trigger >>>>>>>>>>>>>>>>>>>> osd >>>>>>>>>>>>>>>>>>>> shutdowns (each time): >>>>>>>>>>>>>>>>>>>> I get a list followed by an error: >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> ... >>>>>>>>>>>>>>>>>>>> benchmark_data_ceph001.cubone.os_8961_object243839 >>>>>>>>>>>>>>>>>>>> benchmark_data_ceph001.cubone.os_5560_object801983 >>>>>>>>>>>>>>>>>>>> benchmark_data_ceph001.cubone.os_31461_object856489 >>>>>>>>>>>>>>>>>>>> benchmark_data_ceph001.cubone.os_8961_object202232 >>>>>>>>>>>>>>>>>>>> benchmark_data_ceph001.cubone.os_4919_object33199 >>>>>>>>>>>>>>>>>>>> benchmark_data_ceph001.cubone.os_5560_object807797 >>>>>>>>>>>>>>>>>>>> benchmark_data_ceph001.cubone.os_4919_object74729 >>>>>>>>>>>>>>>>>>>> benchmark_data_ceph001.cubone.os_31461_object1264121 >>>>>>>>>>>>>>>>>>>> benchmark_data_ceph001.cubone.os_5560_object1318513 >>>>>>>>>>>>>>>>>>>> benchmark_data_ceph001.cubone.os_5560_object1202111 >>>>>>>>>>>>>>>>>>>> benchmark_data_ceph001.cubone.os_31461_object939107 >>>>>>>>>>>>>>>>>>>> benchmark_data_ceph001.cubone.os_31461_object729682 >>>>>>>>>>>>>>>>>>>> benchmark_data_ceph001.cubone.os_5560_object122915 >>>>>>>>>>>>>>>>>>>> benchmark_data_ceph001.cubone.os_5560_object76521 >>>>>>>>>>>>>>>>>>>> benchmark_data_ceph001.cubone.os_5560_object113261 >>>>>>>>>>>>>>>>>>>> benchmark_data_ceph001.cubone.os_31461_object575079 >>>>>>>>>>>>>>>>>>>> benchmark_data_ceph001.cubone.os_5560_object671042 >>>>>>>>>>>>>>>>>>>> benchmark_data_ceph001.cubone.os_5560_object381146 >>>>>>>>>>>>>>>>>>>> 2014-08-13 17:57:48.736150 7f65047b5700 0 -- >>>>>>>>>>>>>>>>>>>> 10.141.8.180:0/1023295 >> >>>>>>>>>>>>>>>>>>>> 10.141.8.182:6839/4471 pipe(0x7f64fc019b20 sd=5 :0 s=1 >>>>>>>>>>>>>>>>>>>> pgs=0 >>>>>>>>>>>>>>>>>>>> cs=0 >>>>>>>>>>>>>>>>>>>> l=1 >>>>>>>>>>>>>>>>>>>> c=0x7f64fc019db0).fault >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> And I can see this in the log files: >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> -25> 2014-08-13 17:52:56.323908 7f8a97fa4700 1 -- >>>>>>>>>>>>>>>>>>>> 10.143.8.182:6827/64670 <== osd.57 10.141.8.182:0/15796 >>>>>>>>>>>>>>>>>>>> 51 >>>>>>>>>>>>>>>>>>>> ==== >>>>>>>>>>>>>>>>>>>> osd_ping(ping e220 stamp 2014-08-13 17:52:56.323092) v2 >>>>>>>>>>>>>>>>>>>> ==== >>>>>>>>>>>>>>>>>>>> 47+0+0 >>>>>>>>>>>>>>>>>>>> (3227325175 0 0) 0xf475940 con 0xee89fa0 >>>>>>>>>>>>>>>>>>>> -24> 2014-08-13 17:52:56.323938 7f8a97fa4700 1 -- >>>>>>>>>>>>>>>>>>>> 10.143.8.182:6827/64670 --> 10.141.8.182:0/15796 -- >>>>>>>>>>>>>>>>>>>> osd_ping(ping_reply >>>>>>>>>>>>>>>>>>>> e220 >>>>>>>>>>>>>>>>>>>> stamp 2014-08-13 17:52:56.323092) v2 -- ?+0 0xf815b00 >>>>>>>>>>>>>>>>>>>> con >>>>>>>>>>>>>>>>>>>> 0xee89fa0 >>>>>>>>>>>>>>>>>>>> -23> 2014-08-13 17:52:56.324078 7f8a997a7700 1 -- >>>>>>>>>>>>>>>>>>>> 10.141.8.182:6840/64670 <== osd.57 10.141.8.182:0/15796 >>>>>>>>>>>>>>>>>>>> 51 >>>>>>>>>>>>>>>>>>>> ==== >>>>>>>>>>>>>>>>>>>> osd_ping(ping e220 stamp 2014-08-13 17:52:56.323092) v2 >>>>>>>>>>>>>>>>>>>> ==== >>>>>>>>>>>>>>>>>>>> 47+0+0 >>>>>>>>>>>>>>>>>>>> (3227325175 0 0) 0xf132bc0 con 0xee8a680 >>>>>>>>>>>>>>>>>>>> -22> 2014-08-13 17:52:56.324111 7f8a997a7700 1 -- >>>>>>>>>>>>>>>>>>>> 10.141.8.182:6840/64670 --> 10.141.8.182:0/15796 -- >>>>>>>>>>>>>>>>>>>> osd_ping(ping_reply >>>>>>>>>>>>>>>>>>>> e220 >>>>>>>>>>>>>>>>>>>> stamp 2014-08-13 17:52:56.323092) v2 -- ?+0 0xf811a40 >>>>>>>>>>>>>>>>>>>> con >>>>>>>>>>>>>>>>>>>> 0xee8a680 >>>>>>>>>>>>>>>>>>>> -21> 2014-08-13 17:52:56.584461 7f8a997a7700 1 -- >>>>>>>>>>>>>>>>>>>> 10.141.8.182:6840/64670 <== osd.29 10.143.8.181:0/12142 >>>>>>>>>>>>>>>>>>>> 47 >>>>>>>>>>>>>>>>>>>> ==== >>>>>>>>>>>>>>>>>>>> osd_ping(ping e220 stamp 2014-08-13 17:52:56.583010) v2 >>>>>>>>>>>>>>>>>>>> ==== >>>>>>>>>>>>>>>>>>>> 47+0+0 >>>>>>>>>>>>>>>>>>>> (3355887204 0 0) 0xf655940 con 0xee88b00 >>>>>>>>>>>>>>>>>>>> -20> 2014-08-13 17:52:56.584486 7f8a997a7700 1 -- >>>>>>>>>>>>>>>>>>>> 10.141.8.182:6840/64670 --> 10.143.8.181:0/12142 -- >>>>>>>>>>>>>>>>>>>> osd_ping(ping_reply >>>>>>>>>>>>>>>>>>>> e220 >>>>>>>>>>>>>>>>>>>> stamp 2014-08-13 17:52:56.583010) v2 -- ?+0 0xf132bc0 >>>>>>>>>>>>>>>>>>>> con >>>>>>>>>>>>>>>>>>>> 0xee88b00 >>>>>>>>>>>>>>>>>>>> -19> 2014-08-13 17:52:56.584498 7f8a97fa4700 1 -- >>>>>>>>>>>>>>>>>>>> 10.143.8.182:6827/64670 <== osd.29 10.143.8.181:0/12142 >>>>>>>>>>>>>>>>>>>> 47 >>>>>>>>>>>>>>>>>>>> ==== >>>>>>>>>>>>>>>>>>>> osd_ping(ping e220 stamp 2014-08-13 17:52:56.583010) v2 >>>>>>>>>>>>>>>>>>>> ==== >>>>>>>>>>>>>>>>>>>> 47+0+0 >>>>>>>>>>>>>>>>>>>> (3355887204 0 0) 0xf20e040 con 0xee886e0 >>>>>>>>>>>>>>>>>>>> -18> 2014-08-13 17:52:56.584526 7f8a97fa4700 1 -- >>>>>>>>>>>>>>>>>>>> 10.143.8.182:6827/64670 --> 10.143.8.181:0/12142 -- >>>>>>>>>>>>>>>>>>>> osd_ping(ping_reply >>>>>>>>>>>>>>>>>>>> e220 >>>>>>>>>>>>>>>>>>>> stamp 2014-08-13 17:52:56.583010) v2 -- ?+0 0xf475940 >>>>>>>>>>>>>>>>>>>> con >>>>>>>>>>>>>>>>>>>> 0xee886e0 >>>>>>>>>>>>>>>>>>>> -17> 2014-08-13 17:52:56.594448 7f8a798c7700 1 -- >>>>>>>>>>>>>>>>>>>> 10.141.8.182:6839/64670 >> :/0 pipe(0xec15f00 sd=74 >>>>>>>>>>>>>>>>>>>> :6839 >>>>>>>>>>>>>>>>>>>> s=0 >>>>>>>>>>>>>>>>>>>> pgs=0 >>>>>>>>>>>>>>>>>>>> cs=0 >>>>>>>>>>>>>>>>>>>> l=0 >>>>>>>>>>>>>>>>>>>> c=0xee856a0).accept sd=74 10.141.8.180:47641/0 >>>>>>>>>>>>>>>>>>>> -16> 2014-08-13 17:52:56.594921 7f8a798c7700 1 -- >>>>>>>>>>>>>>>>>>>> 10.141.8.182:6839/64670 <== client.7512 >>>>>>>>>>>>>>>>>>>> 10.141.8.180:0/1018433 >>>>>>>>>>>>>>>>>>>> 1 >>>>>>>>>>>>>>>>>>>> ==== >>>>>>>>>>>>>>>>>>>> osd_op(client.7512.0:1 [pgls start_epoch 0] 3.0 >>>>>>>>>>>>>>>>>>>> ack+read+known_if_redirected e220) v4 ==== 151+0+39 >>>>>>>>>>>>>>>>>>>> (1972163119 >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> 4174233976) 0xf3bca40 con 0xee856a0 >>>>>>>>>>>>>>>>>>>> -15> 2014-08-13 17:52:56.594957 7f8a798c7700 5 -- op >>>>>>>>>>>>>>>>>>>> tracker >>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> , >>>>>>>>>>>>>>>>>>>> seq: >>>>>>>>>>>>>>>>>>>> 299, time: 2014-08-13 17:52:56.594874, event: >>>>>>>>>>>>>>>>>>>> header_read, >>>>>>>>>>>>>>>>>>>> op: >>>>>>>>>>>>>>>>>>>> osd_op(client.7512.0:1 [pgls start_epoch 0] 3.0 >>>>>>>>>>>>>>>>>>>> ack+read+known_if_redirected e220) >>>>>>>>>>>>>>>>>>>> -14> 2014-08-13 17:52:56.594970 7f8a798c7700 5 -- op >>>>>>>>>>>>>>>>>>>> tracker >>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> , >>>>>>>>>>>>>>>>>>>> seq: >>>>>>>>>>>>>>>>>>>> 299, time: 2014-08-13 17:52:56.594880, event: throttled, >>>>>>>>>>>>>>>>>>>> op: >>>>>>>>>>>>>>>>>>>> osd_op(client.7512.0:1 [pgls start_epoch 0] 3.0 >>>>>>>>>>>>>>>>>>>> ack+read+known_if_redirected e220) >>>>>>>>>>>>>>>>>>>> -13> 2014-08-13 17:52:56.594978 7f8a798c7700 5 -- op >>>>>>>>>>>>>>>>>>>> tracker >>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> , >>>>>>>>>>>>>>>>>>>> seq: >>>>>>>>>>>>>>>>>>>> 299, time: 2014-08-13 17:52:56.594917, event: all_read, >>>>>>>>>>>>>>>>>>>> op: >>>>>>>>>>>>>>>>>>>> osd_op(client.7512.0:1 [pgls start_epoch 0] 3.0 >>>>>>>>>>>>>>>>>>>> ack+read+known_if_redirected e220) >>>>>>>>>>>>>>>>>>>> -12> 2014-08-13 17:52:56.594986 7f8a798c7700 5 -- op >>>>>>>>>>>>>>>>>>>> tracker >>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> , >>>>>>>>>>>>>>>>>>>> seq: >>>>>>>>>>>>>>>>>>>> 299, time: 0.000000, event: dispatched, op: >>>>>>>>>>>>>>>>>>>> osd_op(client.7512.0:1 >>>>>>>>>>>>>>>>>>>> [pgls >>>>>>>>>>>>>>>>>>>> start_epoch 0] 3.0 ack+read+known_if_redirected e220) >>>>>>>>>>>>>>>>>>>> -11> 2014-08-13 17:52:56.595127 7f8a90795700 5 -- op >>>>>>>>>>>>>>>>>>>> tracker >>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> , >>>>>>>>>>>>>>>>>>>> seq: >>>>>>>>>>>>>>>>>>>> 299, time: 2014-08-13 17:52:56.595104, event: >>>>>>>>>>>>>>>>>>>> reached_pg, >>>>>>>>>>>>>>>>>>>> op: >>>>>>>>>>>>>>>>>>>> osd_op(client.7512.0:1 [pgls start_epoch 0] 3.0 >>>>>>>>>>>>>>>>>>>> ack+read+known_if_redirected e220) >>>>>>>>>>>>>>>>>>>> -10> 2014-08-13 17:52:56.595159 7f8a90795700 5 -- op >>>>>>>>>>>>>>>>>>>> tracker >>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> , >>>>>>>>>>>>>>>>>>>> seq: >>>>>>>>>>>>>>>>>>>> 299, time: 2014-08-13 17:52:56.595153, event: started, >>>>>>>>>>>>>>>>>>>> op: >>>>>>>>>>>>>>>>>>>> osd_op(client.7512.0:1 [pgls start_epoch 0] 3.0 >>>>>>>>>>>>>>>>>>>> ack+read+known_if_redirected e220) >>>>>>>>>>>>>>>>>>>> -9> 2014-08-13 17:52:56.602179 7f8a90795700 1 -- >>>>>>>>>>>>>>>>>>>> 10.141.8.182:6839/64670 --> 10.141.8.180:0/1018433 -- >>>>>>>>>>>>>>>>>>>> osd_op_reply(1 >>>>>>>>>>>>>>>>>>>> [pgls >>>>>>>>>>>>>>>>>>>> start_epoch 0] v164'30654 uv30654 ondisk = 0) v6 -- ?+0 >>>>>>>>>>>>>>>>>>>> 0xec16180 >>>>>>>>>>>>>>>>>>>> con >>>>>>>>>>>>>>>>>>>> 0xee856a0 >>>>>>>>>>>>>>>>>>>> -8> 2014-08-13 17:52:56.602211 7f8a90795700 5 -- op >>>>>>>>>>>>>>>>>>>> tracker >>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> , >>>>>>>>>>>>>>>>>>>> seq: >>>>>>>>>>>>>>>>>>>> 299, time: 2014-08-13 17:52:56.602205, event: done, op: >>>>>>>>>>>>>>>>>>>> osd_op(client.7512.0:1 [pgls start_epoch 0] 3.0 >>>>>>>>>>>>>>>>>>>> ack+read+known_if_redirected e220) >>>>>>>>>>>>>>>>>>>> -7> 2014-08-13 17:52:56.614839 7f8a798c7700 1 -- >>>>>>>>>>>>>>>>>>>> 10.141.8.182:6839/64670 <== client.7512 >>>>>>>>>>>>>>>>>>>> 10.141.8.180:0/1018433 >>>>>>>>>>>>>>>>>>>> 2 >>>>>>>>>>>>>>>>>>>> ==== >>>>>>>>>>>>>>>>>>>> osd_op(client.7512.0:2 [pgls start_epoch 220] 3.0 >>>>>>>>>>>>>>>>>>>> ack+read+known_if_redirected e220) v4 ==== 151+0+89 >>>>>>>>>>>>>>>>>>>> (3460833343 >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> 2600845095) 0xf3bcec0 con 0xee856a0 >>>>>>>>>>>>>>>>>>>> -6> 2014-08-13 17:52:56.614864 7f8a798c7700 5 -- op >>>>>>>>>>>>>>>>>>>> tracker >>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> , >>>>>>>>>>>>>>>>>>>> seq: >>>>>>>>>>>>>>>>>>>> 300, time: 2014-08-13 17:52:56.614789, event: >>>>>>>>>>>>>>>>>>>> header_read, >>>>>>>>>>>>>>>>>>>> op: >>>>>>>>>>>>>>>>>>>> osd_op(client.7512.0:2 [pgls start_epoch 220] 3.0 >>>>>>>>>>>>>>>>>>>> ack+read+known_if_redirected e220) >>>>>>>>>>>>>>>>>>>> -5> 2014-08-13 17:52:56.614874 7f8a798c7700 5 -- op >>>>>>>>>>>>>>>>>>>> tracker >>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> , >>>>>>>>>>>>>>>>>>>> seq: >>>>>>>>>>>>>>>>>>>> 300, time: 2014-08-13 17:52:56.614792, event: throttled, >>>>>>>>>>>>>>>>>>>> op: >>>>>>>>>>>>>>>>>>>> osd_op(client.7512.0:2 [pgls start_epoch 220] 3.0 >>>>>>>>>>>>>>>>>>>> ack+read+known_if_redirected e220) >>>>>>>>>>>>>>>>>>>> -4> 2014-08-13 17:52:56.614884 7f8a798c7700 5 -- op >>>>>>>>>>>>>>>>>>>> tracker >>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> , >>>>>>>>>>>>>>>>>>>> seq: >>>>>>>>>>>>>>>>>>>> 300, time: 2014-08-13 17:52:56.614835, event: all_read, >>>>>>>>>>>>>>>>>>>> op: >>>>>>>>>>>>>>>>>>>> osd_op(client.7512.0:2 [pgls start_epoch 220] 3.0 >>>>>>>>>>>>>>>>>>>> ack+read+known_if_redirected e220) >>>>>>>>>>>>>>>>>>>> -3> 2014-08-13 17:52:56.614891 7f8a798c7700 5 -- op >>>>>>>>>>>>>>>>>>>> tracker >>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> , >>>>>>>>>>>>>>>>>>>> seq: >>>>>>>>>>>>>>>>>>>> 300, time: 0.000000, event: dispatched, op: >>>>>>>>>>>>>>>>>>>> osd_op(client.7512.0:2 >>>>>>>>>>>>>>>>>>>> [pgls >>>>>>>>>>>>>>>>>>>> start_epoch 220] 3.0 ack+read+known_if_redirected e220) >>>>>>>>>>>>>>>>>>>> -2> 2014-08-13 17:52:56.614972 7f8a92f9a700 5 -- op >>>>>>>>>>>>>>>>>>>> tracker >>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> , >>>>>>>>>>>>>>>>>>>> seq: >>>>>>>>>>>>>>>>>>>> 300, time: 2014-08-13 17:52:56.614958, event: >>>>>>>>>>>>>>>>>>>> reached_pg, >>>>>>>>>>>>>>>>>>>> op: >>>>>>>>>>>>>>>>>>>> osd_op(client.7512.0:2 [pgls start_epoch 220] 3.0 >>>>>>>>>>>>>>>>>>>> ack+read+known_if_redirected e220) >>>>>>>>>>>>>>>>>>>> -1> 2014-08-13 17:52:56.614993 7f8a92f9a700 5 -- op >>>>>>>>>>>>>>>>>>>> tracker >>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> , >>>>>>>>>>>>>>>>>>>> seq: >>>>>>>>>>>>>>>>>>>> 300, time: 2014-08-13 17:52:56.614986, event: started, >>>>>>>>>>>>>>>>>>>> op: >>>>>>>>>>>>>>>>>>>> osd_op(client.7512.0:2 [pgls start_epoch 220] 3.0 >>>>>>>>>>>>>>>>>>>> ack+read+known_if_redirected e220) >>>>>>>>>>>>>>>>>>>> 0> 2014-08-13 17:52:56.617087 7f8a92f9a700 -1 >>>>>>>>>>>>>>>>>>>> os/GenericObjectMap.cc: >>>>>>>>>>>>>>>>>>>> In function 'int GenericObjectMap::list_objects(const >>>>>>>>>>>>>>>>>>>> coll_t&, >>>>>>>>>>>>>>>>>>>> ghobject_t, >>>>>>>>>>>>>>>>>>>> int, std::vector<ghobject_t>*, ghobject_t*)' thread >>>>>>>>>>>>>>>>>>>> 7f8a92f9a700 >>>>>>>>>>>>>>>>>>>> time >>>>>>>>>>>>>>>>>>>> 2014-08-13 17:52:56.615073 >>>>>>>>>>>>>>>>>>>> os/GenericObjectMap.cc: 1118: FAILED assert(start <= >>>>>>>>>>>>>>>>>>>> header.oid) >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> ceph version 0.83 (78ff1f0a5dfd3c5850805b40217385 >>>>>>>>>>>>>>>>>>>> 64c36c92b8) >>>>>>>>>>>>>>>>>>>> 1: (GenericObjectMap::list_objects(coll_t const&, >>>>>>>>>>>>>>>>>>>> ghobject_t, >>>>>>>>>>>>>>>>>>>> int, >>>>>>>>>>>>>>>>>>>> std::vector<ghobject_t, std::allocator<ghobject_t> >*, >>>>>>>>>>>>>>>>>>>> ghobject_t*)+0x474) >>>>>>>>>>>>>>>>>>>> [0x98f774] >>>>>>>>>>>>>>>>>>>> 2: (KeyValueStore::collection_list_partial(coll_t, >>>>>>>>>>>>>>>>>>>> ghobject_t, >>>>>>>>>>>>>>>>>>>> int, >>>>>>>>>>>>>>>>>>>> int, >>>>>>>>>>>>>>>>>>>> snapid_t, std::vector<ghobject_t, >>>>>>>>>>>>>>>>>>>> std::allocator<ghobject_t> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> *, >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> ghobject_t*)+0x274) [0x8c5b54] >>>>>>>>>>>>>>>>>>>> 3: (PGBackend::objects_list_partial(hobject_t const&, >>>>>>>>>>>>>>>>>>>> int, >>>>>>>>>>>>>>>>>>>> int, >>>>>>>>>>>>>>>>>>>> snapid_t, >>>>>>>>>>>>>>>>>>>> std::vector<hobject_t, std::allocator<hobject_t> >*, >>>>>>>>>>>>>>>>>>>> hobject_t*)+0x1c9) >>>>>>>>>>>>>>>>>>>> [0x862de9] >>>>>>>>>>>>>>>>>>>> 4: >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> (ReplicatedPG::do_pg_op(std::tr1::shared_ptr<OpRequest>)+ >>>>>>>>>>>>>>>>>>>> 0xea5) >>>>>>>>>>>>>>>>>>>> [0x7f67f5] >>>>>>>>>>>>>>>>>>>> 5: >>>>>>>>>>>>>>>>>>>> (ReplicatedPG::do_op(std::tr1: >>>>>>>>>>>>>>>>>>>> :shared_ptr<OpRequest>)+0x1f3) >>>>>>>>>>>>>>>>>>>> [0x8177b3] >>>>>>>>>>>>>>>>>>>> 6: (ReplicatedPG::do_request(std: >>>>>>>>>>>>>>>>>>>> :tr1::shared_ptr<OpRequest>, >>>>>>>>>>>>>>>>>>>> ThreadPool::TPHandle&)+0x5d5) [0x7b8045] >>>>>>>>>>>>>>>>>>>> 7: (OSD::dequeue_op(boost::intrusive_ptr<PG>, >>>>>>>>>>>>>>>>>>>> std::tr1::shared_ptr<OpRequest>, >>>>>>>>>>>>>>>>>>>> ThreadPool::TPHandle&)+0x47d) >>>>>>>>>>>>>>>>>>>> [0x62bf8d] >>>>>>>>>>>>>>>>>>>> 8: (OSD::ShardedOpWQ::_process(unsigned int, >>>>>>>>>>>>>>>>>>>> ceph::heartbeat_handle_d*)+0x35c) [0x62c56c] >>>>>>>>>>>>>>>>>>>> 9: (ShardedThreadPool::shardedthreadpool_worker(unsigned >>>>>>>>>>>>>>>>>>>> int)+0x8cd) >>>>>>>>>>>>>>>>>>>> [0xa776fd] >>>>>>>>>>>>>>>>>>>> 10: (ShardedThreadPool::WorkThreadSharded::entry()+0x10) >>>>>>>>>>>>>>>>>>>> [0xa79980] >>>>>>>>>>>>>>>>>>>> 11: (()+0x7df3) [0x7f8aac71fdf3] >>>>>>>>>>>>>>>>>>>> 12: (clone()+0x6d) [0x7f8aab1963dd] >>>>>>>>>>>>>>>>>>>> NOTE: a copy of the executable, or `objdump -rdS >>>>>>>>>>>>>>>>>>>> <executable>` >>>>>>>>>>>>>>>>>>>> is >>>>>>>>>>>>>>>>>>>> needed >>>>>>>>>>>>>>>>>>>> to >>>>>>>>>>>>>>>>>>>> interpret this. >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> ceph version 0.83 (78ff1f0a5dfd3c5850805b40217385 >>>>>>>>>>>>>>>>>>>> 64c36c92b8) >>>>>>>>>>>>>>>>>>>> 1: /usr/bin/ceph-osd() [0x99b466] >>>>>>>>>>>>>>>>>>>> 2: (()+0xf130) [0x7f8aac727130] >>>>>>>>>>>>>>>>>>>> 3: (gsignal()+0x39) [0x7f8aab0d5989] >>>>>>>>>>>>>>>>>>>> 4: (abort()+0x148) [0x7f8aab0d7098] >>>>>>>>>>>>>>>>>>>> 5: (__gnu_cxx::__verbose_terminate_handler()+0x165) >>>>>>>>>>>>>>>>>>>> [0x7f8aab9e89d5] >>>>>>>>>>>>>>>>>>>> 6: (()+0x5e946) [0x7f8aab9e6946] >>>>>>>>>>>>>>>>>>>> 7: (()+0x5e973) [0x7f8aab9e6973] >>>>>>>>>>>>>>>>>>>> 8: (()+0x5eb9f) [0x7f8aab9e6b9f] >>>>>>>>>>>>>>>>>>>> 9: (ceph::__ceph_assert_fail(char const*, char const*, >>>>>>>>>>>>>>>>>>>> int, >>>>>>>>>>>>>>>>>>>> char >>>>>>>>>>>>>>>>>>>> const*)+0x1ef) [0xa8805f] >>>>>>>>>>>>>>>>>>>> 10: (GenericObjectMap::list_objects(coll_t const&, >>>>>>>>>>>>>>>>>>>> ghobject_t, >>>>>>>>>>>>>>>>>>>> int, >>>>>>>>>>>>>>>>>>>> std::vector<ghobject_t, std::allocator<ghobject_t> >*, >>>>>>>>>>>>>>>>>>>> ghobject_t*)+0x474) >>>>>>>>>>>>>>>>>>>> [0x98f774] >>>>>>>>>>>>>>>>>>>> 11: (KeyValueStore::collection_list_partial(coll_t, >>>>>>>>>>>>>>>>>>>> ghobject_t, >>>>>>>>>>>>>>>>>>>> int, >>>>>>>>>>>>>>>>>>>> int, >>>>>>>>>>>>>>>>>>>> snapid_t, std::vector<ghobject_t, >>>>>>>>>>>>>>>>>>>> std::allocator<ghobject_t> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> *, >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> ghobject_t*)+0x274) [0x8c5b54] >>>>>>>>>>>>>>>>>>>> 12: (PGBackend::objects_list_partial(hobject_t const&, >>>>>>>>>>>>>>>>>>>> int, >>>>>>>>>>>>>>>>>>>> int, >>>>>>>>>>>>>>>>>>>> snapid_t, >>>>>>>>>>>>>>>>>>>> std::vector<hobject_t, std::allocator<hobject_t> >*, >>>>>>>>>>>>>>>>>>>> hobject_t*)+0x1c9) >>>>>>>>>>>>>>>>>>>> [0x862de9] >>>>>>>>>>>>>>>>>>>> 13: >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> (ReplicatedPG::do_pg_op(std::tr1::shared_ptr<OpRequest>)+ >>>>>>>>>>>>>>>>>>>> 0xea5) >>>>>>>>>>>>>>>>>>>> [0x7f67f5] >>>>>>>>>>>>>>>>>>>> 14: >>>>>>>>>>>>>>>>>>>> (ReplicatedPG::do_op(std::tr1: >>>>>>>>>>>>>>>>>>>> :shared_ptr<OpRequest>)+0x1f3) >>>>>>>>>>>>>>>>>>>> [0x8177b3] >>>>>>>>>>>>>>>>>>>> 15: >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> (ReplicatedPG::do_request(std::tr1::shared_ptr<OpRequest>, >>>>>>>>>>>>>>>>>>>> ThreadPool::TPHandle&)+0x5d5) [0x7b8045] >>>>>>>>>>>>>>>>>>>> 16: (OSD::dequeue_op(boost::intrusive_ptr<PG>, >>>>>>>>>>>>>>>>>>>> std::tr1::shared_ptr<OpRequest>, >>>>>>>>>>>>>>>>>>>> ThreadPool::TPHandle&)+0x47d) >>>>>>>>>>>>>>>>>>>> [0x62bf8d] >>>>>>>>>>>>>>>>>>>> 17: (OSD::ShardedOpWQ::_process(unsigned int, >>>>>>>>>>>>>>>>>>>> ceph::heartbeat_handle_d*)+0x35c) [0x62c56c] >>>>>>>>>>>>>>>>>>>> 18: >>>>>>>>>>>>>>>>>>>> (ShardedThreadPool::shardedthreadpool_worker(unsigned >>>>>>>>>>>>>>>>>>>> int)+0x8cd) >>>>>>>>>>>>>>>>>>>> [0xa776fd] >>>>>>>>>>>>>>>>>>>> 19: (ShardedThreadPool::WorkThreadSharded::entry()+0x10) >>>>>>>>>>>>>>>>>>>> [0xa79980] >>>>>>>>>>>>>>>>>>>> 20: (()+0x7df3) [0x7f8aac71fdf3] >>>>>>>>>>>>>>>>>>>> 21: (clone()+0x6d) [0x7f8aab1963dd] >>>>>>>>>>>>>>>>>>>> NOTE: a copy of the executable, or `objdump -rdS >>>>>>>>>>>>>>>>>>>> <executable>` >>>>>>>>>>>>>>>>>>>> is >>>>>>>>>>>>>>>>>>>> needed >>>>>>>>>>>>>>>>>>>> to >>>>>>>>>>>>>>>>>>>> interpret this. >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> --- begin dump of recent events --- >>>>>>>>>>>>>>>>>>>> 0> 2014-08-13 17:52:56.714214 7f8a92f9a700 -1 *** >>>>>>>>>>>>>>>>>>>> Caught >>>>>>>>>>>>>>>>>>>> signal >>>>>>>>>>>>>>>>>>>> (Aborted) ** >>>>>>>>>>>>>>>>>>>> in thread 7f8a92f9a700 >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> ceph version 0.83 (78ff1f0a5dfd3c5850805b40217385 >>>>>>>>>>>>>>>>>>>> 64c36c92b8) >>>>>>>>>>>>>>>>>>>> 1: /usr/bin/ceph-osd() [0x99b466] >>>>>>>>>>>>>>>>>>>> 2: (()+0xf130) [0x7f8aac727130] >>>>>>>>>>>>>>>>>>>> 3: (gsignal()+0x39) [0x7f8aab0d5989] >>>>>>>>>>>>>>>>>>>> 4: (abort()+0x148) [0x7f8aab0d7098] >>>>>>>>>>>>>>>>>>>> 5: (__gnu_cxx::__verbose_terminate_handler()+0x165) >>>>>>>>>>>>>>>>>>>> [0x7f8aab9e89d5] >>>>>>>>>>>>>>>>>>>> 6: (()+0x5e946) [0x7f8aab9e6946] >>>>>>>>>>>>>>>>>>>> 7: (()+0x5e973) [0x7f8aab9e6973] >>>>>>>>>>>>>>>>>>>> 8: (()+0x5eb9f) [0x7f8aab9e6b9f] >>>>>>>>>>>>>>>>>>>> 9: (ceph::__ceph_assert_fail(char const*, char const*, >>>>>>>>>>>>>>>>>>>> int, >>>>>>>>>>>>>>>>>>>> char >>>>>>>>>>>>>>>>>>>> const*)+0x1ef) [0xa8805f] >>>>>>>>>>>>>>>>>>>> 10: (GenericObjectMap::list_objects(coll_t const&, >>>>>>>>>>>>>>>>>>>> ghobject_t, >>>>>>>>>>>>>>>>>>>> int, >>>>>>>>>>>>>>>>>>>> std::vector<ghobject_t, std::allocator<ghobject_t> >*, >>>>>>>>>>>>>>>>>>>> ghobject_t*)+0x474) >>>>>>>>>>>>>>>>>>>> [0x98f774] >>>>>>>>>>>>>>>>>>>> 11: (KeyValueStore::collection_list_partial(coll_t, >>>>>>>>>>>>>>>>>>>> ghobject_t, >>>>>>>>>>>>>>>>>>>> int, >>>>>>>>>>>>>>>>>>>> int, >>>>>>>>>>>>>>>>>>>> snapid_t, std::vector<ghobject_t, >>>>>>>>>>>>>>>>>>>> std::allocator<ghobject_t> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> *, >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> ghobject_t*)+0x274) [0x8c5b54] >>>>>>>>>>>>>>>>>>>> 12: (PGBackend::objects_list_partial(hobject_t const&, >>>>>>>>>>>>>>>>>>>> int, >>>>>>>>>>>>>>>>>>>> int, >>>>>>>>>>>>>>>>>>>> snapid_t, >>>>>>>>>>>>>>>>>>>> std::vector<hobject_t, std::allocator<hobject_t> >*, >>>>>>>>>>>>>>>>>>>> hobject_t*)+0x1c9) >>>>>>>>>>>>>>>>>>>> [0x862de9] >>>>>>>>>>>>>>>>>>>> 13: >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> (ReplicatedPG::do_pg_op(std::tr1::shared_ptr<OpRequest>)+ >>>>>>>>>>>>>>>>>>>> 0xea5) >>>>>>>>>>>>>>>>>>>> [0x7f67f5] >>>>>>>>>>>>>>>>>>>> 14: >>>>>>>>>>>>>>>>>>>> (ReplicatedPG::do_op(std::tr1: >>>>>>>>>>>>>>>>>>>> :shared_ptr<OpRequest>)+0x1f3) >>>>>>>>>>>>>>>>>>>> [0x8177b3] >>>>>>>>>>>>>>>>>>>> 15: >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> (ReplicatedPG::do_request(std::tr1::shared_ptr<OpRequest>, >>>>>>>>>>>>>>>>>>>> ThreadPool::TPHandle&)+0x5d5) [0x7b8045] >>>>>>>>>>>>>>>>>>>> 16: (OSD::dequeue_op(boost::intrusive_ptr<PG>, >>>>>>>>>>>>>>>>>>>> std::tr1::shared_ptr<OpRequest>, >>>>>>>>>>>>>>>>>>>> ThreadPool::TPHandle&)+0x47d) >>>>>>>>>>>>>>>>>>>> [0x62bf8d] >>>>>>>>>>>>>>>>>>>> 17: (OSD::ShardedOpWQ::_process(unsigned int, >>>>>>>>>>>>>>>>>>>> ceph::heartbeat_handle_d*)+0x35c) [0x62c56c] >>>>>>>>>>>>>>>>>>>> 18: >>>>>>>>>>>>>>>>>>>> (ShardedThreadPool::shardedthreadpool_worker(unsigned >>>>>>>>>>>>>>>>>>>> int)+0x8cd) >>>>>>>>>>>>>>>>>>>> [0xa776fd] >>>>>>>>>>>>>>>>>>>> 19: (ShardedThreadPool::WorkThreadSharded::entry()+0x10) >>>>>>>>>>>>>>>>>>>> [0xa79980] >>>>>>>>>>>>>>>>>>>> 20: (()+0x7df3) [0x7f8aac71fdf3] >>>>>>>>>>>>>>>>>>>> 21: (clone()+0x6d) [0x7f8aab1963dd] >>>>>>>>>>>>>>>>>>>> NOTE: a copy of the executable, or `objdump -rdS >>>>>>>>>>>>>>>>>>>> <executable>` >>>>>>>>>>>>>>>>>>>> is >>>>>>>>>>>>>>>>>>>> needed >>>>>>>>>>>>>>>>>>>> to >>>>>>>>>>>>>>>>>>>> interpret this. >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> I guess this has something to do with using the dev >>>>>>>>>>>>>>>>>>>> Keyvaluestore? >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Thanks! >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Kenneth >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>>>>>> ceph-users mailing list >>>>>>>>>>>>>>>>>>>> ceph-users at lists.ceph.com >>>>>>>>>>>>>>>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>> Best Regards, >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Wheat >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> ----- End message from Haomai Wang <haomaiwang at gmail.com> >>>>>>>>>>>>>>>>>> ----- >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Met vriendelijke groeten, >>>>>>>>>>>>>>>>>> Kenneth Waegeman >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>> Best Regards, >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Wheat >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> ----- End message from Haomai Wang <haomaiwang at gmail.com> >>>>>>>>>>>>>>>> ----- >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Met vriendelijke groeten, >>>>>>>>>>>>>>>> Kenneth Waegeman >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>> Best Regards, >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Wheat >>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>> ceph-users mailing list >>>>>>>>>>>>>>> ceph-users at lists.ceph.com >>>>>>>>>>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> ----- End message from Sage Weil <sweil at redhat.com> ----- >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> >>>>>>>>>>>>> Met vriendelijke groeten, > > >>>>>>>>>>>>> Kenneth Waegeman >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> Best Regards, >>>>>>>>>>>> >>>>>>>>>>>> Wheat >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> ----- End message from Haomai Wang <haomaiwang at gmail.com> ----- >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> >>>>>>>>>>> Met vriendelijke groeten, >>>>>>>>>>> Kenneth Waegeman >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> Best Regards, >>>>>>>>>> >>>>>>>>>> Wheat >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> ----- End message from Haomai Wang <haomaiwang at gmail.com> ----- >>>>>>>>> >>>>>>>>> -- >>>>>>>>> >>>>>>>>> Met vriendelijke groeten, >>>>>>>>> Kenneth Waegeman >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Best Regards, >>>>>>>> >>>>>>>> Wheat >>>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Best Regards, >>>>>>> >>>>>>> Wheat >>>>>>> >>>>>> >>>>>> >>>>>> >>>>>> ----- End message from Haomai Wang <haomaiwang at gmail.com> ----- >>>>>> >>>>>> -- >>>>>> >>>>>> Met vriendelijke groeten, >>>>>> Kenneth Waegeman >>>>>> >>>>>> >>>>>> >>>>>> >>>>> >>>>> >>>>> -- >>>>> Best Regards, >>>>> >>>>> Wheat >>>>> >>>> >>>> >>>> ----- End message from Haomai Wang <haomaiwang at gmail.com> ----- >>>> >>>> -- >>>> >>>> Met vriendelijke groeten, >>>> Kenneth Waegeman >>>> >>>> >>>> >>> >>> >>> -- >>> >>> Best Regards, >>> >>> Wheat >> >> >> >> ----- End message from Haomai Wang <haomaiwang at gmail.com> ----- >> >> -- >> >> Met vriendelijke groeten, >> Kenneth Waegeman > > > > ----- End message from Kenneth Waegeman <Kenneth.Waegeman at UGent.be> ----- > > > -- > > Met vriendelijke groeten, > Kenneth Waegeman > -- Best Regards, Wheat