Hi John, http://pulpito.ceph.com/loic-2016-08-15_07:35:11-fs-jewel-backports-distro-basic-smithi/364579/ has the following error: 2016-08-15T08:13:22.919 INFO:teuthology.orchestra.run.smithi052.stderr:create_volume: /volumes/grpid/volid 2016-08-15T08:13:22.919 INFO:teuthology.orchestra.run.smithi052.stderr:create_volume: grpid/volid, create pool fsvolume_volid as data_isolated =True. 2016-08-15T08:13:22.919 INFO:teuthology.orchestra.run.smithi052.stderr:Traceback (most recent call last): 2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr: File "<string>", line 11, in <module> 2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr: File "/usr/lib/python2.7/dist-packages/ceph_volume_client.py", line 632, in create_volume 2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr: self.fs.setxattr(path, 'ceph.dir.layout.pool', pool_name, 0) 2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr: File "cephfs.pyx", line 779, in cephfs.LibCephFS.setxattr (/srv/autobuild-ceph/gitbuilder.git/build/out~/ceph-10.2.2-351-g431d02a/src/build/cephfs.c:10542) 2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr:cephfs.InvalidValue: error in setxattr This is jewel with a number of CephFS related backports https://github.com/ceph/ceph/tree/jewel-backports but I can't see which one could cause that kind of error. There are a few differences between the jewel and master branch of ceph-qa-suite, but it does not seem to be the cause; git log --no-merges --oneline --cherry-mark --left-right ceph/jewel...ceph/master -- suites/fs > ed1e7f1 suites/fs: fix log whitelist for inotable repair > 41e51eb suites: fix asok_dump_tree.yaml > c669f1e cephfs: add test for dump tree admin socket command > 60dc968 suites/fs: log whitelist for inotable repair = 1558a48 fs: add snapshot tests to mds thrashing = b9b18c7 fs: add snapshot tests to mds thrashing > dc165e6 cephfs: test fragment size limit > 4179c85 suites/fs: use simple messenger some places > 367973b cephfs: test readahead is working > 795d586 suites/fs/permission: run qa/workunits/fs/misc/{acl.sh,chmod.sh} > 45b8e9c suites/fs: fix config for enabling libcephfs posix ACL = fe74a2c suites: allow four remote clients for fs/recovery = b970f97 suites: allow four remote clients for fs/recovery If that rings a bell, let me know. Otherwise I'll keep digging to narrow it down. Cheers -- Loïc Dachary, Artisan Logiciel Libre -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html