Re: Write i/o in CephFS metadata pool

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Sammy;

I had a thought; since you say the FS has high read activity, but you're seeing large write I/O... is it possible that this is related to  atime (Linux last access time)?  If I remember my Linux FS basics, atime is stored in the file entry for the file in the directory, and I believe directory information is stored in the metadata pool (dentries?).

As a test; you might try mounting the CephFS with the noatime flag.  Then see if the write I/O is reduced.

I honestly don't know if CephFS supports atime, but I would expect it would.

Thank you,

Dominic L. Hilsbos, MBA 
Director - Information Technology 
Perform Air International Inc.
DHilsbos@xxxxxxxxxxxxxx 
www.PerformAir.com



-----Original Message-----
From: Samy Ascha [mailto:samy@xxxxxx] 
Sent: Wednesday, January 29, 2020 2:25 AM
To: ceph-users@xxxxxxx
Subject:  Write i/o in CephFS metadata pool

Hi!

I've been running CephFS for a while now and ever since setting it up, I've seen unexpectedly large write i/o on the CephFS metadata pool.

The filesystem is otherwise stable and I'm seeing no usage issues.

I'm in a read-intensive environment, from the clients' perspective and throughput for the metadata pool is consistently larger than that of the data pool.

For example:

# ceph osd pool stats
pool cephfs_data id 1
  client io 7.6 MiB/s rd, 19 KiB/s wr, 404 op/s rd, 1 op/s wr

pool cephfs_metadata id 2
  client io 338 KiB/s rd, 43 MiB/s wr, 84 op/s rd, 26 op/s wr

I realise, of course, that this is a momentary display of statistics, but I see this unbalanced r/w activity consistently when monitoring it live.

I would like some insight into what may be causing this large imbalance in r/w, especially since I'm in a read-intensive (web hosting) environment.

Some of it may be expected in when considering details of my environment and CephFS implementation specifics, so please ask away if more details are needed.

With my experience using NFS, I would start by looking at client io stats, like `nfsstat` and tuning e.g. mount options, but I haven't been able to find such statistics for CephFS clients.

Is there anything of the sort for CephFS? Are similar stats obtainable in some other way?

This might be a somewhat broad question and shallow description, so yeah, let me know if there's anything you would like more details on.

Thanks a lot,
Samy
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux