Re: BUG: ceph_inode_cachep and ceph_dentry_cachep caches are not clean when destroying

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, Feb 19, 2020 at 12:01 PM Xiubo Li <xiubli@xxxxxxxxxx> wrote:
>
> On 2020/2/19 18:53, Ilya Dryomov wrote:
> > On Wed, Feb 19, 2020 at 10:39 AM Xiubo Li <xiubli@xxxxxxxxxx> wrote:
> >> Hi Jeff, Ilya and all
> >>
> >> I hit this call traces by running some test cases when unmounting the fs
> >> mount points.
> >>
> >> It seems there still have some inodes or dentries are not destroyed.
> >>
> >> Will this be a problem ? Any idea ?
> > Hi Xiubo,
> >
> > Of course it is a problem ;)
> >
> > These are all in ceph_inode_info and ceph_dentry_info caches, but
> > I see traces of rbd mappings as well.  Could you please share your
> > test cases?  How are you unloading modules?
>
> I am not sure exactly in which one, mostly I was running the following
> commands.
>
> 1, ./bin/rbd map share -o mount_timeout=30
>
> 2, ./bin/rbd unmap share
>
> 3, ./bin/mount.ceph :/ /mnt/cephfs/
>
> 4, `for i in {0..1000}; do mkdir /mnt/cephfs/dir$0; done` and `for i in
> {0..1000}; do rm -rf /mnt/cephfs/dir$0; done`
>
> 5, umount /mnt/cephfs/
>
> 6, rmmod ceph; rmmod rbd; rmmod libceph
>
> This it seems none business with the rbd mappings.

Is this on more or less plain upstream or with async unlink and
possibly other filesystem patches applied?

Thanks,

                Ilya



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Ceph Dev]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux