Re: s3 requires twice the space it should use

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Cheers,

[root@s3db1 ~]#  ceph daemon osd.23 perf dump | grep numpg
        "numpg": 187,
        "numpg_primary": 64,
        "numpg_replica": 121,
        "numpg_stray": 2,
        "numpg_removing": 0,


Am Do., 15. Apr. 2021 um 18:18 Uhr schrieb 胡 玮文 <huww98@xxxxxxxxxxx>:

> Hi Boris,
>
> Could you check something like
>
> ceph daemon osd.23 perf dump | grep numpg
>
> to see if there are some stray or removing PG?
>
> Weiwen Hu
>
> > 在 2021年4月15日,22:53,Boris Behrens <bb@xxxxxxxxx> 写道:
> >
> > Ah you are right.
> > [root@s3db1 ~]# ceph daemon osd.23 config get
> bluestore_min_alloc_size_hdd
> > {
> >    "bluestore_min_alloc_size_hdd": "65536"
> > }
> > But I also checked how many objects our s3 hold and the numbers just do
> not
> > add up.
> > There are only 26509200 objects, which would result in around 1TB "waste"
> > if every object would be empty.
> >
> > I think the problem began when I updated the PG count from 1024 to 2048.
> > Could there be an issue where the data is written twice?
> >
> >
> >> Am Do., 15. Apr. 2021 um 16:48 Uhr schrieb Amit Ghadge <
> amitg.b14@xxxxxxxxx
> >>> :
> >>
> >> verify those two parameter values ,bluestore_min_alloc_size_hdd &
> >> bluestore_min_alloc_size_sdd, If you are using hdd disk then
> >> bluestore_min_alloc_size_hdd are applicable.
> >>
> >>> On Thu, Apr 15, 2021 at 8:06 PM Boris Behrens <bb@xxxxxxxxx> wrote:
> >>>
> >>> So, I need to live with it? A value of zero leads to use the default?
> >>> [root@s3db1 ~]# ceph daemon osd.23 config get bluestore_min_alloc_size
> >>> {
> >>>    "bluestore_min_alloc_size": "0"
> >>> }
> >>>
> >>> I also checked the fragmentation on the bluestore OSDs and it is around
> >>> 0.80 - 0.89 on most OSDs. yikes.
> >>> [root@s3db1 ~]# ceph daemon osd.23 bluestore allocator score block
> >>> {
> >>>    "fragmentation_rating": 0.85906054329923576
> >>> }
> >>>
> >>> The problem I currently have is, that I barely keep up with adding OSD
> >>> disks.
> >>>
> >>> Am Do., 15. Apr. 2021 um 16:18 Uhr schrieb Amit Ghadge <
> >>> amitg.b14@xxxxxxxxx>:
> >>>
> >>>> size_kb_actual are actually bucket object size but on OSD level the
> >>>> bluestore_min_alloc_size default 64KB and SSD are 16KB
> >>>>
> >>>>
> >>>>
> https://apac01.safelinks.protection.outlook.com/?url=https%3A%2F%2Faccess.redhat.com%2Fdocumentation%2Fen-us%2Fred_hat_ceph_storage%2F3%2Fhtml%2Fadministration_guide%2Fosd-bluestore&amp;data=04%7C01%7C%7Cba98c0dff13941ea96ff08d9001e3759%7C84df9e7fe9f640afb435aaaaaaaaaaaa%7C1%7C0%7C637540952043049058%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&amp;sdata=wfSqqyiDHRXp4ypOGTxx4p%2Buy902OGPEmGkNfJ2BF6I%3D&amp;reserved=0
> >>>>
> >>>> -AmitG
> >>>>
> >>>> On Thu, Apr 15, 2021 at 7:29 PM Boris Behrens <bb@xxxxxxxxx> wrote:
> >>>>
> >>>>> Hi,
> >>>>>
> >>>>> maybe it is just a problem in my understanding, but it looks like
> our s3
> >>>>> requires twice the space it should use.
> >>>>>
> >>>>> I ran "radosgw-admin bucket stats", and added all "size_kb_actual"
> >>>>> values
> >>>>> up and divided to TB (/1024/1024/1024).
> >>>>> The resulting space is 135,1636733 TB. When I tripple it because of
> >>>>> replication I end up with around 405TB which is nearly half the
> space of
> >>>>> what ceph df tells me.
> >>>>>
> >>>>> Hope someone can help me.
> >>>>>
> >>>>> ceph df shows
> >>>>> RAW STORAGE:
> >>>>>    CLASS     SIZE         AVAIL       USED        RAW USED     %RAW
> >>>>> USED
> >>>>>    hdd       1009 TiB     189 TiB     820 TiB      820 TiB
> >>>>> 81.26
> >>>>>    TOTAL     1009 TiB     189 TiB     820 TiB      820 TiB
> >>>>> 81.26
> >>>>>
> >>>>> POOLS:
> >>>>>    POOL                                ID     PGS      STORED
> >>>>> OBJECTS
> >>>>>    USED        %USED     MAX AVAIL
> >>>>>    rbd                                  0       64         0 B
> >>>>>   0
> >>>>>        0 B         0        18 TiB
> >>>>>    .rgw.root                            1       64      99 KiB
> >>>>> 119
> >>>>>     99 KiB         0        18 TiB
> >>>>>    eu-central-1.rgw.control             2       64         0 B
> >>>>>   8
> >>>>>        0 B         0        18 TiB
> >>>>>    eu-central-1.rgw.data.root           3       64     1.0 MiB
> >>>>> 3.15k
> >>>>>    1.0 MiB         0        18 TiB
> >>>>>    eu-central-1.rgw.gc                  4       64      71 MiB
> >>>>>  32
> >>>>>     71 MiB         0        18 TiB
> >>>>>    eu-central-1.rgw.log                 5       64     267 MiB
> >>>>> 564
> >>>>>    267 MiB         0        18 TiB
> >>>>>    eu-central-1.rgw.users.uid           6       64     2.8 MiB
> >>>>> 6.91k
> >>>>>    2.8 MiB         0        18 TiB
> >>>>>    eu-central-1.rgw.users.keys          7       64     263 KiB
> >>>>> 6.73k
> >>>>>    263 KiB         0        18 TiB
> >>>>>    eu-central-1.rgw.meta                8       64     384 KiB
> >>>>>  1k
> >>>>>    384 KiB         0        18 TiB
> >>>>>    eu-central-1.rgw.users.email         9       64        40 B
> >>>>>   1
> >>>>>       40 B         0        18 TiB
> >>>>>    eu-central-1.rgw.buckets.index      10       64      10 GiB
> >>>>> 67.61k
> >>>>>     10 GiB      0.02        18 TiB
> >>>>>    eu-central-1.rgw.buckets.data       11     2048     264 TiB
> >>>>> 138.31M
> >>>>>    264 TiB     83.37        18 TiB
> >>>>>    eu-central-1.rgw.buckets.non-ec     12       64     297 MiB
> >>>>> 11.32k
> >>>>>    297 MiB         0        18 TiB
> >>>>>    eu-central-1.rgw.usage              13       64     536 MiB
> >>>>>  32
> >>>>>    536 MiB         0        18 TiB
> >>>>>    eu-msg-1.rgw.control                56       64         0 B
> >>>>>   8
> >>>>>        0 B         0        18 TiB
> >>>>>    eu-msg-1.rgw.data.root              57       64      72 KiB
> >>>>> 227
> >>>>>     72 KiB         0        18 TiB
> >>>>>    eu-msg-1.rgw.gc                     58       64     300 KiB
> >>>>>  32
> >>>>>    300 KiB         0        18 TiB
> >>>>>    eu-msg-1.rgw.log                    59       64     835 KiB
> >>>>> 242
> >>>>>    835 KiB         0        18 TiB
> >>>>>    eu-msg-1.rgw.users.uid              60       64      56 KiB
> >>>>> 104
> >>>>>     56 KiB         0        18 TiB
> >>>>>    eu-msg-1.rgw.usage                  61       64      37 MiB
> >>>>>  25
> >>>>>     37 MiB         0        18 TiB
> >>>>>    eu-msg-1.rgw.users.keys             62       64     3.8 KiB
> >>>>>  97
> >>>>>    3.8 KiB         0        18 TiB
> >>>>>    eu-msg-1.rgw.meta                   63       64     607 KiB
> >>>>> 1.60k
> >>>>>    607 KiB         0        18 TiB
> >>>>>    eu-msg-1.rgw.buckets.index          64       64      71 MiB
> >>>>> 119
> >>>>>     71 MiB         0        18 TiB
> >>>>>    eu-msg-1.rgw.users.email            65       64         0 B
> >>>>>   0
> >>>>>        0 B         0        18 TiB
> >>>>>    eu-msg-1.rgw.buckets.data           66       64     2.9 TiB
> >>>>> 1.16M
> >>>>>    2.9 TiB      5.30        18 TiB
> >>>>>    eu-msg-1.rgw.buckets.non-ec         67       64     2.2 MiB
> >>>>> 354
> >>>>>    2.2 MiB         0        18 TiB
> >>>>>    default.rgw.control                 69       32         0 B
> >>>>>   8
> >>>>>        0 B         0        18 TiB
> >>>>>    default.rgw.data.root               70       32         0 B
> >>>>>   0
> >>>>>        0 B         0        18 TiB
> >>>>>    default.rgw.gc                      71       32         0 B
> >>>>>   0
> >>>>>        0 B         0        18 TiB
> >>>>>    default.rgw.log                     72       32         0 B
> >>>>>   0
> >>>>>        0 B         0        18 TiB
> >>>>>    default.rgw.users.uid               73       32         0 B
> >>>>>   0
> >>>>>        0 B         0        18 TiB
> >>>>>    fra-1.rgw.control                   74       32         0 B
> >>>>>   8
> >>>>>        0 B         0        18 TiB
> >>>>>    fra-1.rgw.meta                      75       32         0 B
> >>>>>   0
> >>>>>        0 B         0        18 TiB
> >>>>>    fra-1.rgw.log                       76       32        50 B
> >>>>>  28
> >>>>>       50 B         0        18 TiB
> >>>>>
> >>>>>
> >>>>> --
> >>>>> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal
> abweichend im
> >>>>> groüen Saal.
> >>>>> _______________________________________________
> >>>>> ceph-users mailing list -- ceph-users@xxxxxxx
> >>>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
> >>>>>
> >>>>
> >>>
> >>> --
> >>> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend
> im
> >>> groüen Saal.
> >>>
> >>
> >
> > --
> > Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im
> > groüen Saal.
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@xxxxxxx
> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
>


-- 
Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im
groüen Saal.
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux