/familiar/not familiar/ On Tue, Jun 3, 2014 at 10:33 PM, Haomai Wang <haomaiwang@xxxxxxxxx> wrote: > Hi Gregory, > > I checked again and again each line change about spill out codes, > still failed to find anything wrong. > > I ran "ceph_test_rados" then activate scrub process several times > locally, nothing unusual. I'm familiar with teuthoghy jobs, maybe we > can find the common thing among fail jobs. > > > On Sat, May 31, 2014 at 2:06 AM, Gregory Farnum <greg@xxxxxxxxxxx> wrote: >> On Fri, May 30, 2014 at 2:18 AM, Haomai Wang <haomaiwang@xxxxxxxxx> wrote: >>> Hi Gregory, >>> >>> I try to reproduce the bug in my local machine but failed. >>> >>> My test cmdline: >>> ./ceph_test_rados --op read 100 --op write 100 --op delete 50 >>> --max-ops 400000 --objects 1024 --max-in-flight 64 --size 4000000 >>> --min-stride-size 400000 --max-stride-size 800000 --max-seconds 600 >>> --op copy_from 50 --op snap_create 50 --op snap_remove 50 --op >>> rollback 50 --op setattr 25 --op rmattr 25 --pool unique_pool_0 >>> >>> Is there any tip to reproduce it? >> >> Hmm. I've been directly running the teuthology tests that failed; that >> is the command line it's running though so I think the only difference >> would be that I've got OSD thrashing (ie, recovery) happening while >> the test is running. >> Most of the other failures were scrub turning up inconsistencies in >> the xattrs at each replica/shard of an object. I didn't see any >> obvious mechanism by which storing values in xattrs versus leveldb >> would impact these higher-level primitives, but maybe you have some >> idea? >> -Greg > > > > -- > Best Regards, > > Wheat -- Best Regards, Wheat -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html