Re: BUG: ceph_inode_cachep and ceph_dentry_cachep caches are not clean when destroying

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 2020/2/19 18:53, Ilya Dryomov wrote:
On Wed, Feb 19, 2020 at 10:39 AM Xiubo Li <xiubli@xxxxxxxxxx> wrote:
Hi Jeff, Ilya and all

I hit this call traces by running some test cases when unmounting the fs
mount points.

It seems there still have some inodes or dentries are not destroyed.

Will this be a problem ? Any idea ?
Hi Xiubo,

Of course it is a problem ;)

These are all in ceph_inode_info and ceph_dentry_info caches, but
I see traces of rbd mappings as well.  Could you please share your
test cases?  How are you unloading modules?

I am not sure exactly in which one, mostly I was running the following commands.

1, ./bin/rbd map share -o mount_timeout=30

2, ./bin/rbd unmap share

3, ./bin/mount.ceph :/ /mnt/cephfs/

4, `for i in {0..1000}; do mkdir /mnt/cephfs/dir$0; done` and `for i in {0..1000}; do rm -rf /mnt/cephfs/dir$0; done`

5, umount /mnt/cephfs/

6, rmmod ceph; rmmod rbd; rmmod libceph

This it seems none business with the rbd mappings.

Thanks.

BRs

Xiubo


Thanks,

                 Ilya





[Index of Archives]     [CEPH Users]     [Ceph Large]     [Ceph Dev]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux