Re: s3 requires twice the space it should use

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Ah you are right.
[root@s3db1 ~]# ceph daemon osd.23 config get bluestore_min_alloc_size_hdd
{
    "bluestore_min_alloc_size_hdd": "65536"
}
But I also checked how many objects our s3 hold and the numbers just do not
add up.
There are only 26509200 objects, which would result in around 1TB "waste"
if every object would be empty.

I think the problem began when I updated the PG count from 1024 to 2048.
Could there be an issue where the data is written twice?


Am Do., 15. Apr. 2021 um 16:48 Uhr schrieb Amit Ghadge <amitg.b14@xxxxxxxxx
>:

> verify those two parameter values ,bluestore_min_alloc_size_hdd &
> bluestore_min_alloc_size_sdd, If you are using hdd disk then
> bluestore_min_alloc_size_hdd are applicable.
>
> On Thu, Apr 15, 2021 at 8:06 PM Boris Behrens <bb@xxxxxxxxx> wrote:
>
>> So, I need to live with it? A value of zero leads to use the default?
>> [root@s3db1 ~]# ceph daemon osd.23 config get bluestore_min_alloc_size
>> {
>>     "bluestore_min_alloc_size": "0"
>> }
>>
>> I also checked the fragmentation on the bluestore OSDs and it is around
>> 0.80 - 0.89 on most OSDs. yikes.
>> [root@s3db1 ~]# ceph daemon osd.23 bluestore allocator score block
>> {
>>     "fragmentation_rating": 0.85906054329923576
>> }
>>
>> The problem I currently have is, that I barely keep up with adding OSD
>> disks.
>>
>> Am Do., 15. Apr. 2021 um 16:18 Uhr schrieb Amit Ghadge <
>> amitg.b14@xxxxxxxxx>:
>>
>>> size_kb_actual are actually bucket object size but on OSD level the
>>> bluestore_min_alloc_size default 64KB and SSD are 16KB
>>>
>>>
>>> https://access.redhat.com/documentation/en-us/red_hat_ceph_storage/3/html/administration_guide/osd-bluestore
>>>
>>> -AmitG
>>>
>>> On Thu, Apr 15, 2021 at 7:29 PM Boris Behrens <bb@xxxxxxxxx> wrote:
>>>
>>>> Hi,
>>>>
>>>> maybe it is just a problem in my understanding, but it looks like our s3
>>>> requires twice the space it should use.
>>>>
>>>> I ran "radosgw-admin bucket stats", and added all "size_kb_actual"
>>>> values
>>>> up and divided to TB (/1024/1024/1024).
>>>> The resulting space is 135,1636733 TB. When I tripple it because of
>>>> replication I end up with around 405TB which is nearly half the space of
>>>> what ceph df tells me.
>>>>
>>>> Hope someone can help me.
>>>>
>>>> ceph df shows
>>>> RAW STORAGE:
>>>>     CLASS     SIZE         AVAIL       USED        RAW USED     %RAW
>>>> USED
>>>>     hdd       1009 TiB     189 TiB     820 TiB      820 TiB
>>>>  81.26
>>>>     TOTAL     1009 TiB     189 TiB     820 TiB      820 TiB
>>>>  81.26
>>>>
>>>> POOLS:
>>>>     POOL                                ID     PGS      STORED
>>>> OBJECTS
>>>>     USED        %USED     MAX AVAIL
>>>>     rbd                                  0       64         0 B
>>>>    0
>>>>         0 B         0        18 TiB
>>>>     .rgw.root                            1       64      99 KiB
>>>>  119
>>>>      99 KiB         0        18 TiB
>>>>     eu-central-1.rgw.control             2       64         0 B
>>>>    8
>>>>         0 B         0        18 TiB
>>>>     eu-central-1.rgw.data.root           3       64     1.0 MiB
>>>>  3.15k
>>>>     1.0 MiB         0        18 TiB
>>>>     eu-central-1.rgw.gc                  4       64      71 MiB
>>>>   32
>>>>      71 MiB         0        18 TiB
>>>>     eu-central-1.rgw.log                 5       64     267 MiB
>>>>  564
>>>>     267 MiB         0        18 TiB
>>>>     eu-central-1.rgw.users.uid           6       64     2.8 MiB
>>>>  6.91k
>>>>     2.8 MiB         0        18 TiB
>>>>     eu-central-1.rgw.users.keys          7       64     263 KiB
>>>>  6.73k
>>>>     263 KiB         0        18 TiB
>>>>     eu-central-1.rgw.meta                8       64     384 KiB
>>>>   1k
>>>>     384 KiB         0        18 TiB
>>>>     eu-central-1.rgw.users.email         9       64        40 B
>>>>    1
>>>>        40 B         0        18 TiB
>>>>     eu-central-1.rgw.buckets.index      10       64      10 GiB
>>>> 67.61k
>>>>      10 GiB      0.02        18 TiB
>>>>     eu-central-1.rgw.buckets.data       11     2048     264 TiB
>>>>  138.31M
>>>>     264 TiB     83.37        18 TiB
>>>>     eu-central-1.rgw.buckets.non-ec     12       64     297 MiB
>>>> 11.32k
>>>>     297 MiB         0        18 TiB
>>>>     eu-central-1.rgw.usage              13       64     536 MiB
>>>>   32
>>>>     536 MiB         0        18 TiB
>>>>     eu-msg-1.rgw.control                56       64         0 B
>>>>    8
>>>>         0 B         0        18 TiB
>>>>     eu-msg-1.rgw.data.root              57       64      72 KiB
>>>>  227
>>>>      72 KiB         0        18 TiB
>>>>     eu-msg-1.rgw.gc                     58       64     300 KiB
>>>>   32
>>>>     300 KiB         0        18 TiB
>>>>     eu-msg-1.rgw.log                    59       64     835 KiB
>>>>  242
>>>>     835 KiB         0        18 TiB
>>>>     eu-msg-1.rgw.users.uid              60       64      56 KiB
>>>>  104
>>>>      56 KiB         0        18 TiB
>>>>     eu-msg-1.rgw.usage                  61       64      37 MiB
>>>>   25
>>>>      37 MiB         0        18 TiB
>>>>     eu-msg-1.rgw.users.keys             62       64     3.8 KiB
>>>>   97
>>>>     3.8 KiB         0        18 TiB
>>>>     eu-msg-1.rgw.meta                   63       64     607 KiB
>>>>  1.60k
>>>>     607 KiB         0        18 TiB
>>>>     eu-msg-1.rgw.buckets.index          64       64      71 MiB
>>>>  119
>>>>      71 MiB         0        18 TiB
>>>>     eu-msg-1.rgw.users.email            65       64         0 B
>>>>    0
>>>>         0 B         0        18 TiB
>>>>     eu-msg-1.rgw.buckets.data           66       64     2.9 TiB
>>>>  1.16M
>>>>     2.9 TiB      5.30        18 TiB
>>>>     eu-msg-1.rgw.buckets.non-ec         67       64     2.2 MiB
>>>>  354
>>>>     2.2 MiB         0        18 TiB
>>>>     default.rgw.control                 69       32         0 B
>>>>    8
>>>>         0 B         0        18 TiB
>>>>     default.rgw.data.root               70       32         0 B
>>>>    0
>>>>         0 B         0        18 TiB
>>>>     default.rgw.gc                      71       32         0 B
>>>>    0
>>>>         0 B         0        18 TiB
>>>>     default.rgw.log                     72       32         0 B
>>>>    0
>>>>         0 B         0        18 TiB
>>>>     default.rgw.users.uid               73       32         0 B
>>>>    0
>>>>         0 B         0        18 TiB
>>>>     fra-1.rgw.control                   74       32         0 B
>>>>    8
>>>>         0 B         0        18 TiB
>>>>     fra-1.rgw.meta                      75       32         0 B
>>>>    0
>>>>         0 B         0        18 TiB
>>>>     fra-1.rgw.log                       76       32        50 B
>>>>   28
>>>>        50 B         0        18 TiB
>>>>
>>>>
>>>> --
>>>> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im
>>>> groüen Saal.
>>>> _______________________________________________
>>>> ceph-users mailing list -- ceph-users@xxxxxxx
>>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>>>
>>>
>>
>> --
>> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im
>> groüen Saal.
>>
>

-- 
Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im
groüen Saal.
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux