On Tue, Aug 16, 2016 at 12:47 AM, Loic Dachary <loic@xxxxxxxxxxx> wrote: > Hi John, > > http://pulpito.ceph.com/loic-2016-08-15_07:35:11-fs-jewel-backports-distro-basic-smithi/364579/ has the following error: > > 2016-08-15T08:13:22.919 INFO:teuthology.orchestra.run.smithi052.stderr:create_volume: /volumes/grpid/volid > 2016-08-15T08:13:22.919 INFO:teuthology.orchestra.run.smithi052.stderr:create_volume: grpid/volid, create pool fsvolume_volid as data_isolated =True. > 2016-08-15T08:13:22.919 INFO:teuthology.orchestra.run.smithi052.stderr:Traceback (most recent call last): > 2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr: File "<string>", line 11, in <module> > 2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr: File "/usr/lib/python2.7/dist-packages/ceph_volume_client.py", line 632, in create_volume > 2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr: self.fs.setxattr(path, 'ceph.dir.layout.pool', pool_name, 0) > 2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr: File "cephfs.pyx", line 779, in cephfs.LibCephFS.setxattr (/srv/autobuild-ceph/gitbuilder.git/build/out~/ceph-10.2.2-351-g431d02a/src/build/cephfs.c:10542) > 2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr:cephfs.InvalidValue: error in setxattr > The error is because MDS had outdated osdmap and thought the newly creately pool does not exist. (MDS has code that makes sure its osdmap is the same as or newer than fs client's osdmap) For this case, It seems both mds and fs client had outdated osdmap. Pool creation was through self.rados. self.rados had the newest olsdmap, but self.fs might have outdated osdmap. Regards Yan, Zheng > This is jewel with a number of CephFS related backports > > https://github.com/ceph/ceph/tree/jewel-backports but I can't see which one could cause that kind of error. > > There are a few differences between the jewel and master branch of ceph-qa-suite, but it does not seem to be the cause; > > git log --no-merges --oneline --cherry-mark --left-right ceph/jewel...ceph/master -- suites/fs > >> ed1e7f1 suites/fs: fix log whitelist for inotable repair >> 41e51eb suites: fix asok_dump_tree.yaml >> c669f1e cephfs: add test for dump tree admin socket command >> 60dc968 suites/fs: log whitelist for inotable repair > = 1558a48 fs: add snapshot tests to mds thrashing > = b9b18c7 fs: add snapshot tests to mds thrashing >> dc165e6 cephfs: test fragment size limit >> 4179c85 suites/fs: use simple messenger some places >> 367973b cephfs: test readahead is working >> 795d586 suites/fs/permission: run qa/workunits/fs/misc/{acl.sh,chmod.sh} >> 45b8e9c suites/fs: fix config for enabling libcephfs posix ACL > = fe74a2c suites: allow four remote clients for fs/recovery > = b970f97 suites: allow four remote clients for fs/recovery > > If that rings a bell, let me know. Otherwise I'll keep digging to narrow it down. > > Cheers > > -- > Loïc Dachary, Artisan Logiciel Libre > -- > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html