Thank you, Frank. Tbh, I think it doesn't matter if the number of manual compactions is for 24h or for a smaller period, as long as it's over a reasonable period of time, so that an average number of compactions per hour can be calculated. /Z On Fri, 13 Oct 2023 at 16:01, Frank Schilder <frans@xxxxxx> wrote: > Hi Zakhar, > > I'm pretty sure you wanted the #manual compactions for an entire day, not > from whenever the log starts to current time, which is most often not > 23:59. You need to get the date from the previous day and make sure the log > contains a full 00:00-23:59 window. > > 1) iotop results: > TID PRIO USER DISK READ DISK WRITE SWAPIN IO COMMAND > 2256 be/4 ceph 0.00 B 17.48 M 0.00 % 0.80 % ceph-mon > --cluster ceph --setuser ceph --setgroup ceph --foreground -i ceph-01 > --mon-data /var/lib/ceph/mon/ceph-ceph-01 --public-addr 192.168.32.65 > [safe_timer] > 2230 be/4 ceph 0.00 B 1514.19 M 0.00 % 0.37 % ceph-mon > --cluster ceph --setuser ceph --setgroup ceph --foreground -i ceph-01 > --mon-data /var/lib/ceph/mon/ceph-ceph-01 --public-addr 192.168.32.65 > [rocksdb:low0] > 2250 be/4 ceph 0.00 B 36.23 M 0.00 % 0.15 % ceph-mon > --cluster ceph --setuser ceph --setgroup ceph --foreground -i ceph-01 > --mon-data /var/lib/ceph/mon/ceph-ceph-01 --public-addr 192.168.32.65 > [fn_monstore] > 2231 be/4 ceph 0.00 B 50.52 M 0.00 % 0.02 % ceph-mon > --cluster ceph --setuser ceph --setgroup ceph --foreground -i ceph-01 > --mon-data /var/lib/ceph/mon/ceph-ceph-01 --public-addr 192.168.32.65 > [rocksdb:high0] > 2225 be/4 ceph 0.00 B 120.00 K 0.00 % 0.00 % ceph-mon > --cluster ceph --setuser ceph --setgroup ceph --foreground -i ceph-01 > --mon-data /var/lib/ceph/mon/ceph-ceph-01 --public-addr 192.168.32.65 [log] > > 2) manual compactions (over a full 24h window): 1882 > > 3) monitor store.db size: 616M > > 4) cluster version and status: > > ceph version 15.2.17 (8a82819d84cf884bd39c17e3236e0632ac146dc4) octopus > (stable) > > cluster: > id: xxx > health: HEALTH_WARN > 1 large omap objects > > services: > mon: 5 daemons, quorum ceph-01,ceph-02,ceph-03,ceph-25,ceph-26 (age 7w) > mgr: ceph-25(active, since 4w), standbys: ceph-26, ceph-01, ceph-03, > ceph-02 > mds: con-fs2:8 4 up:standby 8 up:active > osd: 1284 osds: 1282 up (since 27h), 1282 in (since 3w) > > task status: > > data: > pools: 14 pools, 25065 pgs > objects: 2.18G objects, 3.9 PiB > usage: 4.8 PiB used, 8.3 PiB / 13 PiB avail > pgs: 25037 active+clean > 26 active+clean+scrubbing+deep > 2 active+clean+scrubbing > > io: > client: 1.7 GiB/s rd, 1013 MiB/s wr, 3.02k op/s rd, 1.78k op/s wr > > Best regards, > ================= > Frank Schilder > AIT Risø Campus > Bygning 109, rum S14 > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx