Re: CephFS constant high write I/O to the metadata pool

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I tried my luck and upgraded to 17.2.4 but unfortunately that didn't make
any difference here either.

I also looked more again at all kinds of client op and request stats and
wotnot which only made me even more certain that this io is not caused by
any clients.

What internal mds operation or mechanism could cause such high idle write
io? I've tried to fiddle a bit with some of the mds cache trim and memory
settings but I haven't noticed any effect there. Any pointers appreciated.

Cheers,
---------------------------
Olli Rajala - Lead TD
Anima Vitae Ltd.
www.anima.fi
---------------------------


On Mon, Oct 17, 2022 at 10:28 AM Olli Rajala <olli.rajala@xxxxxxxx> wrote:

> Hi Patrick,
>
> With "objecter_ops" did you mean "ceph tell mds.pve-core-1 ops" and/or
> "ceph tell mds.pve-core-1 objecter_requests"? Both these show very few
> requests/ops - many times just returning empty lists. I'm pretty sure that
> this I/O isn't generated by any clients - I've earlier tried to isolate
> this by shutting down all cephfs clients and this didn't have any
> noticeable effect.
>
> I tried to watch what is going on with that "perf dump" but to be honest
> all I can see is some numbers going up in the different sections :)
> ...don't have a clue what to focus on and how to interpret that.
>
> Here's a perf dump if you or anyone could make something out of that:
> https://gist.github.com/olliRJL/43c10173aafd82be22c080a9cd28e673
>
> Tnx!
> o.
>
> ---------------------------
> Olli Rajala - Lead TD
> Anima Vitae Ltd.
> www.anima.fi
> ---------------------------
>
>
> On Fri, Oct 14, 2022 at 8:32 PM Patrick Donnelly <pdonnell@xxxxxxxxxx>
> wrote:
>
>> Hello Olli,
>>
>> On Thu, Oct 13, 2022 at 5:01 AM Olli Rajala <olli.rajala@xxxxxxxx> wrote:
>> >
>> > Hi,
>> >
>> > I'm seeing constant 25-50MB/s writes to the metadata pool even when all
>> > clients and the cluster is idling and in clean state. This surely can't
>> be
>> > normal?
>> >
>> > There's no apparent issues with the performance of the cluster but this
>> > write rate seems excessive and I don't know where to look for the
>> culprit.
>> >
>> > The setup is Ceph 16.2.9 running in hyperconverged 3 node core cluster
>> and
>> > 6 hdd osd nodes.
>> >
>> > Here's typical status when pretty much all clients are idling. Most of
>> that
>> > write bandwidth and maybe fifth of the write iops is hitting the
>> > metadata pool.
>> >
>> >
>> ---------------------------------------------------------------------------------------------------
>> > root@pve-core-1:~# ceph -s
>> >   cluster:
>> >     id:     2088b4b1-8de1-44d4-956e-aa3d3afff77f
>> >     health: HEALTH_OK
>> >
>> >   services:
>> >     mon: 3 daemons, quorum pve-core-1,pve-core-2,pve-core-3 (age 2w)
>> >     mgr: pve-core-1(active, since 4w), standbys: pve-core-2, pve-core-3
>> >     mds: 1/1 daemons up, 2 standby
>> >     osd: 48 osds: 48 up (since 5h), 48 in (since 4M)
>> >
>> >   data:
>> >     volumes: 1/1 healthy
>> >     pools:   10 pools, 625 pgs
>> >     objects: 70.06M objects, 46 TiB
>> >     usage:   95 TiB used, 182 TiB / 278 TiB avail
>> >     pgs:     625 active+clean
>> >
>> >   io:
>> >     client:   45 KiB/s rd, 38 MiB/s wr, 6 op/s rd, 287 op/s wr
>> >
>> ---------------------------------------------------------------------------------------------------
>> >
>> > Here's some daemonperf dump:
>> >
>> >
>> ---------------------------------------------------------------------------------------------------
>> > root@pve-core-1:~# ceph daemonperf mds.`hostname -s`
>> >
>> ----------------------------------------mds-----------------------------------------
>> > --mds_cache--- ------mds_log------ -mds_mem- -------mds_server-------
>> mds_
>> > -----objecter------ purg
>> > req  rlat fwd  inos caps exi  imi  hifc crev cgra ctru cfsa cfa  hcc
>> hccd
>> > hccr prcr|stry recy recd|subm evts segs repl|ino  dn  |hcr  hcs  hsr
>> cre
>> >  cat |sess|actv rd   wr   rdwr|purg|
>> >  40    0    0  767k  78k   0    0    0    1    6    1    0    0    5
>> 5
>> >  3    7 |1.1k   0    0 | 17  3.7k 134    0 |767k 767k| 40    5    0    0
>> >  0 |110 |  4    2   21    0 |  2
>> >  57    2    0  767k  78k   0    0    0    3   16    3    0    0   11
>>  11
>> >  0   17 |1.1k   0    0 | 45  3.7k 137    0 |767k 767k| 57    8    0    0
>> >  0 |110 |  0    2   28    0 |  4
>> >  57    4    0  767k  78k   0    0    0    4   34    4    0    0   34
>>  33
>> >  2   26 |1.0k   0    0 |134  3.9k 139    0 |767k 767k| 57   13    0    0
>> >  0 |110 |  0    2  112    0 | 19
>> >  67    3    0  767k  78k   0    0    0    6   32    6    0    0   22
>>  22
>> >  0   32 |1.1k   0    0 | 78  3.9k 141    0 |767k 768k| 67    4    0    0
>> >  0 |110 |  0    2   56    0 |  2
>> >
>> ---------------------------------------------------------------------------------------------------
>> > Any ideas where to look at?
>>
>> Check the perf dump output of the mds:
>>
>> ceph tell mds.<fs_name>:0 perf dump
>>
>> over a period of time to identify what's going on. You can also look
>> at the objecter_ops (another tell command) for the MDS.
>>
>> --
>> Patrick Donnelly, Ph.D.
>> He / Him / His
>> Principal Software Engineer
>> Red Hat, Inc.
>> GPG: 19F28A586F808C2402351B93C3301A3E258DD79D
>>
>>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux