On Wed, Feb 19, 2020 at 12:01 PM Xiubo Li <xiubli@xxxxxxxxxx> wrote: > > On 2020/2/19 18:53, Ilya Dryomov wrote: > > On Wed, Feb 19, 2020 at 10:39 AM Xiubo Li <xiubli@xxxxxxxxxx> wrote: > >> Hi Jeff, Ilya and all > >> > >> I hit this call traces by running some test cases when unmounting the fs > >> mount points. > >> > >> It seems there still have some inodes or dentries are not destroyed. > >> > >> Will this be a problem ? Any idea ? > > Hi Xiubo, > > > > Of course it is a problem ;) > > > > These are all in ceph_inode_info and ceph_dentry_info caches, but > > I see traces of rbd mappings as well. Could you please share your > > test cases? How are you unloading modules? > > I am not sure exactly in which one, mostly I was running the following > commands. > > 1, ./bin/rbd map share -o mount_timeout=30 > > 2, ./bin/rbd unmap share > > 3, ./bin/mount.ceph :/ /mnt/cephfs/ > > 4, `for i in {0..1000}; do mkdir /mnt/cephfs/dir$0; done` and `for i in > {0..1000}; do rm -rf /mnt/cephfs/dir$0; done` > > 5, umount /mnt/cephfs/ > > 6, rmmod ceph; rmmod rbd; rmmod libceph > > This it seems none business with the rbd mappings. Is this on more or less plain upstream or with async unlink and possibly other filesystem patches applied? Thanks, Ilya