On 2020/2/19 19:27, Ilya Dryomov wrote:
On Wed, Feb 19, 2020 at 12:01 PM Xiubo Li <xiubli@xxxxxxxxxx> wrote:
On 2020/2/19 18:53, Ilya Dryomov wrote:
On Wed, Feb 19, 2020 at 10:39 AM Xiubo Li <xiubli@xxxxxxxxxx> wrote:
Hi Jeff, Ilya and all
I hit this call traces by running some test cases when unmounting the fs
mount points.
It seems there still have some inodes or dentries are not destroyed.
Will this be a problem ? Any idea ?
Hi Xiubo,
Of course it is a problem ;)
These are all in ceph_inode_info and ceph_dentry_info caches, but
I see traces of rbd mappings as well. Could you please share your
test cases? How are you unloading modules?
I am not sure exactly in which one, mostly I was running the following
commands.
1, ./bin/rbd map share -o mount_timeout=30
2, ./bin/rbd unmap share
3, ./bin/mount.ceph :/ /mnt/cephfs/
4, `for i in {0..1000}; do mkdir /mnt/cephfs/dir$0; done` and `for i in
{0..1000}; do rm -rf /mnt/cephfs/dir$0; done`
5, umount /mnt/cephfs/
6, rmmod ceph; rmmod rbd; rmmod libceph
This it seems none business with the rbd mappings.
Is this on more or less plain upstream or with async unlink and
possibly other filesystem patches applied?
Using the latest test branch:
https://github.com/ceph/ceph-client/tree/testing.
thanks
Thanks,
Ilya