Re: cephfs file layouts, empty objects in first data pool

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Mon, Feb 10, 2020 at 12:29 AM Håkan T Johansson <f96hajo@xxxxxxxxxxx> wrote:
>
>
> On Mon, 10 Feb 2020, Gregory Farnum wrote:
>
> > On Sun, Feb 9, 2020 at 3:24 PM Håkan T Johansson <f96hajo@xxxxxxxxxxx> wrote:
> >
> >       Hi,
> >
> >       running 14.2.6, debian buster (backports).
> >
> >       Have set up a cephfs with 3 data pools and one metadata pool:
> >       myfs_data, myfs_data_hdd, myfs_data_ssd, and myfs_metadata.
> >
> >       The data of all files are with the use of ceph.dir.layout.pool either
> >       stored in the pools myfs_data_hdd or myfs_data_ssd.  This has also been
> >       checked by dumping the ceph.file.layout.pool attributes of all files.
> >
> >       The filesystem has 1617949 files and 36042 directories.
> >
> >       There are however approximately as many objects in the first pool created
> >       for the cephfs, myfs_data, as there are files.  They also becomes more or
> >       fewer as files are created or deleted (so cannot be some leftover from
> >       earlier exercises).  Note how the USED size is reported as 0 bytes,
> >       correctly reflecting that no file data is stored in them.
> >
> >       POOL_NAME        USED OBJECTS CLONES  COPIES MISSING_ON_PRIMARY UNFOUND DEGRADED  RD_OPS      RD   WR_OPS      WR USED COMPR UNDER COMPR
> >       myfs_data         0 B 1618229      0 4854687                  0       0        0 2263590 129 GiB 23312479 124 GiB        0 B         0 B
> >       myfs_data_hdd 831 GiB  136309      0  408927                  0       0        0  106046 200 GiB   269084 277 GiB        0 B         0 B
> >       myfs_data_ssd  43 GiB 1552412      0 4657236                  0       0        0  181468 2.3 GiB  4661935  12 GiB        0 B         0 B
> >       myfs_metadata 1.2 GiB   36096      0  108288                  0       0        0 4828623  82 GiB  1355102 143 GiB        0 B         0 B
> >
> >       Is this expected?
> >
> >       I was assuming that in this scenario, all objects, both their data and any
> >       keys would be either in the metadata pool, or the two pools where the
> >       objects are stored.
> >
> >       Is it some additional metadata keys that are stored in the first
> >       created data pool for cephfs?  This would not be so nice in case the osd
> >       selection rules for it are using worse disks than the data itself...
> >
> >
> > https://docs.ceph.com/docs/master/cephfs/file-layouts/#adding-a-data-pool-to-the-mds notes there is “a small amount of metadata” kept in the primary pool.
>
> Thanks!  This I managed to miss, probably as it was at the bottom of the
> page.  In case one wants to use layouts to separate fast (likely many)
> from slow (likely large) files, it then sounds as the primary pool should
> the fast kind too, due to the large amount of objects.  Thus this needs to
> be highlighted early in that documentation.
>
> > That’s not terribly clear; what is actually stored is a per-file location backtrace (its location in the directory tree) used for hardlink lookups and disaster recovery
> > scenarios.
>
> This info would be nice to add to the manual page.  It is nice to know
> what kind of information is stored there.

Yeah, PRs welcome. :p
Just to be clear though, that shouldn't be performance-critical. It's
lazily updated by the MDS when the directory location changes, but not
otherwise.

>
> Again thanks for the clarification!
>
> >       Btw: is there any tool to see the amount of key value data size associated
> >       with a pool?  'ceph osd df' gives omap and meta for osds, but not broken
> >       down per pool.
> >
> >
> > I think this is in the newest master code, but I’m not certain which release it’s in...
>
> Would it then (when available) also be in the 'rados df' command?

I really don't remember how everything is shared out but I think so?

>
> Best regards,
> Håkan
>
>
> > -Greg
> >
> >
> >
> >       Best regards,
> >       Håkan
> >       _______________________________________________
> >       ceph-users mailing list -- ceph-users@xxxxxxx
> >       To unsubscribe send an email to ceph-users-leave@xxxxxxx
> >
> >
> >
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux