Re: jewel backports: cephfs.InvalidValue: error in setxattr

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi John,

On 29/08/2016 18:53, John Spray wrote:
> On Mon, Aug 22, 2016 at 7:16 PM, Gregory Farnum <gfarnum@xxxxxxxxxx> wrote:
>> On Tue, Aug 16, 2016 at 1:44 AM, Loic Dachary <loic@xxxxxxxxxxx> wrote:
>>> Hi Yan,
>>>
>>> On 16/08/2016 04:16, Yan, Zheng wrote:
>>>> On Tue, Aug 16, 2016 at 12:47 AM, Loic Dachary <loic@xxxxxxxxxxx> wrote:
>>>>> Hi John,
>>>>>
>>>>> http://pulpito.ceph.com/loic-2016-08-15_07:35:11-fs-jewel-backports-distro-basic-smithi/364579/ has the following error:
>>>>>
>>>>> 2016-08-15T08:13:22.919 INFO:teuthology.orchestra.run.smithi052.stderr:create_volume: /volumes/grpid/volid
>>>>> 2016-08-15T08:13:22.919 INFO:teuthology.orchestra.run.smithi052.stderr:create_volume: grpid/volid, create pool fsvolume_volid as data_isolated =True.
>>>>> 2016-08-15T08:13:22.919 INFO:teuthology.orchestra.run.smithi052.stderr:Traceback (most recent call last):
>>>>> 2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr:  File "<string>", line 11, in <module>
>>>>> 2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr:  File "/usr/lib/python2.7/dist-packages/ceph_volume_client.py", line 632, in create_volume
>>>>> 2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr:    self.fs.setxattr(path, 'ceph.dir.layout.pool', pool_name, 0)
>>>>> 2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr:  File "cephfs.pyx", line 779, in cephfs.LibCephFS.setxattr (/srv/autobuild-ceph/gitbuilder.git/build/out~/ceph-10.2.2-351-g431d02a/src/build/cephfs.c:10542)
>>>>> 2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr:cephfs.InvalidValue: error in setxattr
>>>>>
>>>>
>>>> The error is because MDS had outdated osdmap and thought the newly
>>>> creately pool does not exist. (MDS has code that makes sure its osdmap
>>>> is the same as or newer than fs client's osdmap)  For this case, It
>>>> seems both mds and fs client had outdated osdmap.  Pool creation was
>>>> through self.rados. self.rados had the newest olsdmap, but self.fs
>>>> might have outdated osdmap.
>>>
>>> Interesting. Do you know why this happens ? Is there a specific pull request that causes this ?
>>>
>>> Thanks a lot for your help !
>>
>> Not sure about the specific PR, but in general when running commands
>> referencing pools, you need a new enough OSDMap to see the pool
>> everywhere it's used. We have a lot of logic and extra data passing in
>> the FS layers to make sure those OSDMaps appear transparently, but if
>> you create the pool through RADOS the FS clients have no idea of its
>> existence and the caller needs to wait themselves.
> 
> Loic, was this failure reproducible or a one off?

It was a one off. See http://tracker.ceph.com/issues/16344#note-21 for two other runs of the same job, in an attempt to reproduce it.

> 
> What's supposed to happen here is that Client::ll_setxattr calls
> wait_for_latest_osdmap when it sees a set to ceph.dir.layout.pool, and
> thereby picks up the pool that was just created.  It shouldn't be racy
> :-/
> 
> There is only the MDS log from this failure, in which the EINVAL is
> being generated on the server side.  Hmm.
> 
> John
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 

-- 
Loïc Dachary, Artisan Logiciel Libre
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux