Re: Recently deployed cluster showing 9Tb of raw usage without any load deployed

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



To add more information, in case that helps:
```
# ceph -s
  cluster:
    id:     <Pool_UUID>
    health: HEALTH_OK

 ....

  task status:

  data:
    pools:   6 pools, 161 pgs
    objects: 223 objects, 7.0 KiB
    usage:   9.3 TiB used, 364 TiB / 373 TiB avail
    pgs:     161 active+clean

# ceph df
--- RAW STORAGE ---
CLASS  SIZE     AVAIL    USED     RAW USED  %RAW USED
hdd    373 TiB  364 TiB  9.3 TiB   9.3 TiB       2.50
TOTAL  373 TiB  364 TiB  9.3 TiB   9.3 TiB       2.50

--- POOLS ---
POOL                   ID  PGS  STORED   OBJECTS  USED     %USED  MAX AVAIL
device_health_metrics   1    1      0 B        0      0 B      0    115 TiB
.rgw.root               2   32  3.6 KiB        8  1.5 MiB      0    115 TiB
default.rgw.log         3   32  3.4 KiB      207    6 MiB      0    115 TiB
default.rgw.control     4   32      0 B        8      0 B      0    115 TiB
default.rgw.meta        5   32      0 B        0      0 B      0    115 TiB
rbd                     6   32      0 B        0      0 B      0    115 TiB
```

On Mon, Apr 3, 2023 at 10:25 PM Work Ceph <work.ceph.user.mailing@xxxxxxxxx>
wrote:

> Hello guys!
>
>
> We noticed an unexpected situation. In a recently deployed Ceph cluster we
> are seeing a raw usage, that is a bit odd. We have the following setup:
>
>
> We have a new cluster with 5 nodes with the following setup:
>
>    - 128 GB of RAM
>    - 2 cpus Intel(R) Intel Xeon Silver 4210R
>    - 1 NVME of 2 TB for the rocks DB caching
>    - 5 HDDs of 14TB
>    - 1 NIC dual port of 25GiB in BOND mode.
>
>
> Right after deploying the Ceph cluster, we see a raw usage of about 9TiB.
> However, no load has been applied onto the cluster. Have you guys seen such
> a situation? Or, can you guys help understand it?
>
>
> We are using Ceph Octopus, and we have set the following configurations:
>
> ```
>
> ceph_conf_overrides:
>
>   global:
>
>     osd pool default size: 3
>
>     osd pool default min size: 1
>
>     osd pool default pg autoscale mode: "warn"
>
>     perf: true
>
>     rocksdb perf: true
>
>   mon:
>
>     mon osd down out interval: 120
>
>   osd:
>
>     bluestore min alloc size hdd: 65536
>
>
> ```
>
>
> Any tip or help on how to explain this situation is welcome!
>
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux