On 2020/2/19 18:53, Ilya Dryomov wrote:
On Wed, Feb 19, 2020 at 10:39 AM Xiubo Li <xiubli@xxxxxxxxxx> wrote:
Hi Jeff, Ilya and all
I hit this call traces by running some test cases when unmounting the fs
mount points.
It seems there still have some inodes or dentries are not destroyed.
Will this be a problem ? Any idea ?
Hi Xiubo,
Of course it is a problem ;)
These are all in ceph_inode_info and ceph_dentry_info caches, but
I see traces of rbd mappings as well. Could you please share your
test cases? How are you unloading modules?
I am not sure exactly in which one, mostly I was running the following
commands.
1, ./bin/rbd map share -o mount_timeout=30
2, ./bin/rbd unmap share
3, ./bin/mount.ceph :/ /mnt/cephfs/
4, `for i in {0..1000}; do mkdir /mnt/cephfs/dir$0; done` and `for i in
{0..1000}; do rm -rf /mnt/cephfs/dir$0; done`
5, umount /mnt/cephfs/
6, rmmod ceph; rmmod rbd; rmmod libceph
This it seems none business with the rbd mappings.
Thanks.
BRs
Xiubo
Thanks,
Ilya