Re: Please help collecting stats of Ceph monitor disk writes

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Thank you, Frank.

Tbh, I think it doesn't matter if the number of manual compactions is for
24h or for a smaller period, as long as it's over a reasonable period of
time, so that an average number of compactions per hour can be calculated.

/Z

On Fri, 13 Oct 2023 at 16:01, Frank Schilder <frans@xxxxxx> wrote:

> Hi Zakhar,
>
> I'm pretty sure you wanted the #manual compactions for an entire day, not
> from whenever the log starts to current time, which is most often not
> 23:59. You need to get the date from the previous day and make sure the log
> contains a full 00:00-23:59 window.
>
> 1) iotop results:
>     TID  PRIO  USER     DISK READ  DISK WRITE  SWAPIN      IO    COMMAND
>    2256 be/4 ceph          0.00 B     17.48 M  0.00 %  0.80 % ceph-mon
> --cluster ceph --setuser ceph --setgroup ceph --foreground -i ceph-01
> --mon-data /var/lib/ceph/mon/ceph-ceph-01 --public-addr 192.168.32.65
> [safe_timer]
>    2230 be/4 ceph          0.00 B   1514.19 M  0.00 %  0.37 % ceph-mon
> --cluster ceph --setuser ceph --setgroup ceph --foreground -i ceph-01
> --mon-data /var/lib/ceph/mon/ceph-ceph-01 --public-addr 192.168.32.65
> [rocksdb:low0]
>    2250 be/4 ceph          0.00 B     36.23 M  0.00 %  0.15 % ceph-mon
> --cluster ceph --setuser ceph --setgroup ceph --foreground -i ceph-01
> --mon-data /var/lib/ceph/mon/ceph-ceph-01 --public-addr 192.168.32.65
> [fn_monstore]
>    2231 be/4 ceph          0.00 B     50.52 M  0.00 %  0.02 % ceph-mon
> --cluster ceph --setuser ceph --setgroup ceph --foreground -i ceph-01
> --mon-data /var/lib/ceph/mon/ceph-ceph-01 --public-addr 192.168.32.65
> [rocksdb:high0]
>    2225 be/4 ceph          0.00 B    120.00 K  0.00 %  0.00 % ceph-mon
> --cluster ceph --setuser ceph --setgroup ceph --foreground -i ceph-01
> --mon-data /var/lib/ceph/mon/ceph-ceph-01 --public-addr 192.168.32.65 [log]
>
> 2) manual compactions (over a full 24h window): 1882
>
> 3) monitor store.db size: 616M
>
> 4) cluster version and status:
>
> ceph version 15.2.17 (8a82819d84cf884bd39c17e3236e0632ac146dc4) octopus
> (stable)
>
>   cluster:
>     id:     xxx
>     health: HEALTH_WARN
>             1 large omap objects
>
>   services:
>     mon: 5 daemons, quorum ceph-01,ceph-02,ceph-03,ceph-25,ceph-26 (age 7w)
>     mgr: ceph-25(active, since 4w), standbys: ceph-26, ceph-01, ceph-03,
> ceph-02
>     mds: con-fs2:8 4 up:standby 8 up:active
>     osd: 1284 osds: 1282 up (since 27h), 1282 in (since 3w)
>
>   task status:
>
>   data:
>     pools:   14 pools, 25065 pgs
>     objects: 2.18G objects, 3.9 PiB
>     usage:   4.8 PiB used, 8.3 PiB / 13 PiB avail
>     pgs:     25037 active+clean
>              26    active+clean+scrubbing+deep
>              2     active+clean+scrubbing
>
>   io:
>     client:   1.7 GiB/s rd, 1013 MiB/s wr, 3.02k op/s rd, 1.78k op/s wr
>
> Best regards,
> =================
> Frank Schilder
> AIT Risø Campus
> Bygning 109, rum S14
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux