Hi John, On 29/08/2016 18:53, John Spray wrote: > On Mon, Aug 22, 2016 at 7:16 PM, Gregory Farnum <gfarnum@xxxxxxxxxx> wrote: >> On Tue, Aug 16, 2016 at 1:44 AM, Loic Dachary <loic@xxxxxxxxxxx> wrote: >>> Hi Yan, >>> >>> On 16/08/2016 04:16, Yan, Zheng wrote: >>>> On Tue, Aug 16, 2016 at 12:47 AM, Loic Dachary <loic@xxxxxxxxxxx> wrote: >>>>> Hi John, >>>>> >>>>> http://pulpito.ceph.com/loic-2016-08-15_07:35:11-fs-jewel-backports-distro-basic-smithi/364579/ has the following error: >>>>> >>>>> 2016-08-15T08:13:22.919 INFO:teuthology.orchestra.run.smithi052.stderr:create_volume: /volumes/grpid/volid >>>>> 2016-08-15T08:13:22.919 INFO:teuthology.orchestra.run.smithi052.stderr:create_volume: grpid/volid, create pool fsvolume_volid as data_isolated =True. >>>>> 2016-08-15T08:13:22.919 INFO:teuthology.orchestra.run.smithi052.stderr:Traceback (most recent call last): >>>>> 2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr: File "<string>", line 11, in <module> >>>>> 2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr: File "/usr/lib/python2.7/dist-packages/ceph_volume_client.py", line 632, in create_volume >>>>> 2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr: self.fs.setxattr(path, 'ceph.dir.layout.pool', pool_name, 0) >>>>> 2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr: File "cephfs.pyx", line 779, in cephfs.LibCephFS.setxattr (/srv/autobuild-ceph/gitbuilder.git/build/out~/ceph-10.2.2-351-g431d02a/src/build/cephfs.c:10542) >>>>> 2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr:cephfs.InvalidValue: error in setxattr >>>>> >>>> >>>> The error is because MDS had outdated osdmap and thought the newly >>>> creately pool does not exist. (MDS has code that makes sure its osdmap >>>> is the same as or newer than fs client's osdmap) For this case, It >>>> seems both mds and fs client had outdated osdmap. Pool creation was >>>> through self.rados. self.rados had the newest olsdmap, but self.fs >>>> might have outdated osdmap. >>> >>> Interesting. Do you know why this happens ? Is there a specific pull request that causes this ? >>> >>> Thanks a lot for your help ! >> >> Not sure about the specific PR, but in general when running commands >> referencing pools, you need a new enough OSDMap to see the pool >> everywhere it's used. We have a lot of logic and extra data passing in >> the FS layers to make sure those OSDMaps appear transparently, but if >> you create the pool through RADOS the FS clients have no idea of its >> existence and the caller needs to wait themselves. > > Loic, was this failure reproducible or a one off? It was a one off. See http://tracker.ceph.com/issues/16344#note-21 for two other runs of the same job, in an attempt to reproduce it. > > What's supposed to happen here is that Client::ll_setxattr calls > wait_for_latest_osdmap when it sees a set to ceph.dir.layout.pool, and > thereby picks up the pool that was just created. It shouldn't be racy > :-/ > > There is only the MDS log from this failure, in which the EINVAL is > being generated on the server side. Hmm. > > John > -- > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html > -- Loïc Dachary, Artisan Logiciel Libre -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html