Re: s3 requires twice the space it should use

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



So, I need to live with it? A value of zero leads to use the default?
[root@s3db1 ~]# ceph daemon osd.23 config get bluestore_min_alloc_size
{
    "bluestore_min_alloc_size": "0"
}

I also checked the fragmentation on the bluestore OSDs and it is around
0.80 - 0.89 on most OSDs. yikes.
[root@s3db1 ~]# ceph daemon osd.23 bluestore allocator score block
{
    "fragmentation_rating": 0.85906054329923576
}

The problem I currently have is, that I barely keep up with adding OSD
disks.

Am Do., 15. Apr. 2021 um 16:18 Uhr schrieb Amit Ghadge <amitg.b14@xxxxxxxxx
>:

> size_kb_actual are actually bucket object size but on OSD level the
> bluestore_min_alloc_size default 64KB and SSD are 16KB
>
>
> https://access.redhat.com/documentation/en-us/red_hat_ceph_storage/3/html/administration_guide/osd-bluestore
>
> -AmitG
>
> On Thu, Apr 15, 2021 at 7:29 PM Boris Behrens <bb@xxxxxxxxx> wrote:
>
>> Hi,
>>
>> maybe it is just a problem in my understanding, but it looks like our s3
>> requires twice the space it should use.
>>
>> I ran "radosgw-admin bucket stats", and added all "size_kb_actual" values
>> up and divided to TB (/1024/1024/1024).
>> The resulting space is 135,1636733 TB. When I tripple it because of
>> replication I end up with around 405TB which is nearly half the space of
>> what ceph df tells me.
>>
>> Hope someone can help me.
>>
>> ceph df shows
>> RAW STORAGE:
>>     CLASS     SIZE         AVAIL       USED        RAW USED     %RAW USED
>>     hdd       1009 TiB     189 TiB     820 TiB      820 TiB         81.26
>>     TOTAL     1009 TiB     189 TiB     820 TiB      820 TiB         81.26
>>
>> POOLS:
>>     POOL                                ID     PGS      STORED
>> OBJECTS
>>     USED        %USED     MAX AVAIL
>>     rbd                                  0       64         0 B
>>  0
>>         0 B         0        18 TiB
>>     .rgw.root                            1       64      99 KiB
>>  119
>>      99 KiB         0        18 TiB
>>     eu-central-1.rgw.control             2       64         0 B
>>  8
>>         0 B         0        18 TiB
>>     eu-central-1.rgw.data.root           3       64     1.0 MiB
>>  3.15k
>>     1.0 MiB         0        18 TiB
>>     eu-central-1.rgw.gc                  4       64      71 MiB
>> 32
>>      71 MiB         0        18 TiB
>>     eu-central-1.rgw.log                 5       64     267 MiB
>>  564
>>     267 MiB         0        18 TiB
>>     eu-central-1.rgw.users.uid           6       64     2.8 MiB
>>  6.91k
>>     2.8 MiB         0        18 TiB
>>     eu-central-1.rgw.users.keys          7       64     263 KiB
>>  6.73k
>>     263 KiB         0        18 TiB
>>     eu-central-1.rgw.meta                8       64     384 KiB
>> 1k
>>     384 KiB         0        18 TiB
>>     eu-central-1.rgw.users.email         9       64        40 B
>>  1
>>        40 B         0        18 TiB
>>     eu-central-1.rgw.buckets.index      10       64      10 GiB
>> 67.61k
>>      10 GiB      0.02        18 TiB
>>     eu-central-1.rgw.buckets.data       11     2048     264 TiB
>>  138.31M
>>     264 TiB     83.37        18 TiB
>>     eu-central-1.rgw.buckets.non-ec     12       64     297 MiB
>> 11.32k
>>     297 MiB         0        18 TiB
>>     eu-central-1.rgw.usage              13       64     536 MiB
>> 32
>>     536 MiB         0        18 TiB
>>     eu-msg-1.rgw.control                56       64         0 B
>>  8
>>         0 B         0        18 TiB
>>     eu-msg-1.rgw.data.root              57       64      72 KiB
>>  227
>>      72 KiB         0        18 TiB
>>     eu-msg-1.rgw.gc                     58       64     300 KiB
>> 32
>>     300 KiB         0        18 TiB
>>     eu-msg-1.rgw.log                    59       64     835 KiB
>>  242
>>     835 KiB         0        18 TiB
>>     eu-msg-1.rgw.users.uid              60       64      56 KiB
>>  104
>>      56 KiB         0        18 TiB
>>     eu-msg-1.rgw.usage                  61       64      37 MiB
>> 25
>>      37 MiB         0        18 TiB
>>     eu-msg-1.rgw.users.keys             62       64     3.8 KiB
>> 97
>>     3.8 KiB         0        18 TiB
>>     eu-msg-1.rgw.meta                   63       64     607 KiB
>>  1.60k
>>     607 KiB         0        18 TiB
>>     eu-msg-1.rgw.buckets.index          64       64      71 MiB
>>  119
>>      71 MiB         0        18 TiB
>>     eu-msg-1.rgw.users.email            65       64         0 B
>>  0
>>         0 B         0        18 TiB
>>     eu-msg-1.rgw.buckets.data           66       64     2.9 TiB
>>  1.16M
>>     2.9 TiB      5.30        18 TiB
>>     eu-msg-1.rgw.buckets.non-ec         67       64     2.2 MiB
>>  354
>>     2.2 MiB         0        18 TiB
>>     default.rgw.control                 69       32         0 B
>>  8
>>         0 B         0        18 TiB
>>     default.rgw.data.root               70       32         0 B
>>  0
>>         0 B         0        18 TiB
>>     default.rgw.gc                      71       32         0 B
>>  0
>>         0 B         0        18 TiB
>>     default.rgw.log                     72       32         0 B
>>  0
>>         0 B         0        18 TiB
>>     default.rgw.users.uid               73       32         0 B
>>  0
>>         0 B         0        18 TiB
>>     fra-1.rgw.control                   74       32         0 B
>>  8
>>         0 B         0        18 TiB
>>     fra-1.rgw.meta                      75       32         0 B
>>  0
>>         0 B         0        18 TiB
>>     fra-1.rgw.log                       76       32        50 B
>> 28
>>        50 B         0        18 TiB
>>
>>
>> --
>> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im
>> groüen Saal.
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@xxxxxxx
>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>
>

-- 
Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im
groüen Saal.
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux