Re: Write i/o in CephFS metadata pool

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Dominic,

I should have mentioned that I've set noatime already.

I have not found any obvious other mount options that would contribute to 'write on read' behaviour..

Thx

Samy

> On 29 Jan 2020, at 15:43, DHilsbos@xxxxxxxxxxxxxx wrote:
> 
> Sammy;
> 
> I had a thought; since you say the FS has high read activity, but you're seeing large write I/O... is it possible that this is related to  atime (Linux last access time)?  If I remember my Linux FS basics, atime is stored in the file entry for the file in the directory, and I believe directory information is stored in the metadata pool (dentries?).
> 
> As a test; you might try mounting the CephFS with the noatime flag.  Then see if the write I/O is reduced.
> 
> I honestly don't know if CephFS supports atime, but I would expect it would.
> 
> Thank you,
> 
> Dominic L. Hilsbos, MBA 
> Director - Information Technology 
> Perform Air International Inc.
> DHilsbos@xxxxxxxxxxxxxx 
> www.PerformAir.com
> 
> 
> 
> -----Original Message-----
> From: Samy Ascha [mailto:samy@xxxxxx] 
> Sent: Wednesday, January 29, 2020 2:25 AM
> To: ceph-users@xxxxxxx
> Subject:  Write i/o in CephFS metadata pool
> 
> Hi!
> 
> I've been running CephFS for a while now and ever since setting it up, I've seen unexpectedly large write i/o on the CephFS metadata pool.
> 
> The filesystem is otherwise stable and I'm seeing no usage issues.
> 
> I'm in a read-intensive environment, from the clients' perspective and throughput for the metadata pool is consistently larger than that of the data pool.
> 
> For example:
> 
> # ceph osd pool stats
> pool cephfs_data id 1
> client io 7.6 MiB/s rd, 19 KiB/s wr, 404 op/s rd, 1 op/s wr
> 
> pool cephfs_metadata id 2
> client io 338 KiB/s rd, 43 MiB/s wr, 84 op/s rd, 26 op/s wr
> 
> I realise, of course, that this is a momentary display of statistics, but I see this unbalanced r/w activity consistently when monitoring it live.
> 
> I would like some insight into what may be causing this large imbalance in r/w, especially since I'm in a read-intensive (web hosting) environment.
> 
> Some of it may be expected in when considering details of my environment and CephFS implementation specifics, so please ask away if more details are needed.
> 
> With my experience using NFS, I would start by looking at client io stats, like `nfsstat` and tuning e.g. mount options, but I haven't been able to find such statistics for CephFS clients.
> 
> Is there anything of the sort for CephFS? Are similar stats obtainable in some other way?
> 
> This might be a somewhat broad question and shallow description, so yeah, let me know if there's anything you would like more details on.
> 
> Thanks a lot,
> Samy
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux