jewel backports: cephfs.InvalidValue: error in setxattr

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi John,

http://pulpito.ceph.com/loic-2016-08-15_07:35:11-fs-jewel-backports-distro-basic-smithi/364579/ has the following error:

2016-08-15T08:13:22.919 INFO:teuthology.orchestra.run.smithi052.stderr:create_volume: /volumes/grpid/volid
2016-08-15T08:13:22.919 INFO:teuthology.orchestra.run.smithi052.stderr:create_volume: grpid/volid, create pool fsvolume_volid as data_isolated =True.
2016-08-15T08:13:22.919 INFO:teuthology.orchestra.run.smithi052.stderr:Traceback (most recent call last):
2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr:  File "<string>", line 11, in <module>
2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr:  File "/usr/lib/python2.7/dist-packages/ceph_volume_client.py", line 632, in create_volume
2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr:    self.fs.setxattr(path, 'ceph.dir.layout.pool', pool_name, 0)
2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr:  File "cephfs.pyx", line 779, in cephfs.LibCephFS.setxattr (/srv/autobuild-ceph/gitbuilder.git/build/out~/ceph-10.2.2-351-g431d02a/src/build/cephfs.c:10542)
2016-08-15T08:13:22.920 INFO:teuthology.orchestra.run.smithi052.stderr:cephfs.InvalidValue: error in setxattr

This is jewel with a number of CephFS related backports 

https://github.com/ceph/ceph/tree/jewel-backports but I can't see which one could cause that kind of error. 

There are a few differences between the jewel and master branch of ceph-qa-suite, but it does not seem to be the cause;

git log --no-merges --oneline --cherry-mark --left-right ceph/jewel...ceph/master -- suites/fs

> ed1e7f1 suites/fs: fix log whitelist for inotable repair
> 41e51eb suites: fix asok_dump_tree.yaml
> c669f1e cephfs: add test for dump tree admin socket command
> 60dc968 suites/fs: log whitelist for inotable repair
= 1558a48 fs: add snapshot tests to mds thrashing
= b9b18c7 fs: add snapshot tests to mds thrashing
> dc165e6 cephfs: test fragment size limit
> 4179c85 suites/fs: use simple messenger some places
> 367973b cephfs: test readahead is working
> 795d586 suites/fs/permission: run qa/workunits/fs/misc/{acl.sh,chmod.sh}
> 45b8e9c suites/fs: fix config for enabling libcephfs posix ACL
= fe74a2c suites: allow four remote clients for fs/recovery
= b970f97 suites: allow four remote clients for fs/recovery

If that rings a bell, let me know. Otherwise I'll keep digging to narrow it down.

Cheers

-- 
Loïc Dachary, Artisan Logiciel Libre
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux