Hi Gregory, I checked again and again each line change about spill out codes, still failed to find anything wrong. I ran "ceph_test_rados" then activate scrub process several times locally, nothing unusual. I'm familiar with teuthoghy jobs, maybe we can find the common thing among fail jobs. On Sat, May 31, 2014 at 2:06 AM, Gregory Farnum <greg@xxxxxxxxxxx> wrote: > On Fri, May 30, 2014 at 2:18 AM, Haomai Wang <haomaiwang@xxxxxxxxx> wrote: >> Hi Gregory, >> >> I try to reproduce the bug in my local machine but failed. >> >> My test cmdline: >> ./ceph_test_rados --op read 100 --op write 100 --op delete 50 >> --max-ops 400000 --objects 1024 --max-in-flight 64 --size 4000000 >> --min-stride-size 400000 --max-stride-size 800000 --max-seconds 600 >> --op copy_from 50 --op snap_create 50 --op snap_remove 50 --op >> rollback 50 --op setattr 25 --op rmattr 25 --pool unique_pool_0 >> >> Is there any tip to reproduce it? > > Hmm. I've been directly running the teuthology tests that failed; that > is the command line it's running though so I think the only difference > would be that I've got OSD thrashing (ie, recovery) happening while > the test is running. > Most of the other failures were scrub turning up inconsistencies in > the xattrs at each replica/shard of an object. I didn't see any > obvious mechanism by which storing values in xattrs versus leveldb > would impact these higher-level primitives, but maybe you have some > idea? > -Greg -- Best Regards, Wheat -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html