Re: mds0: Client failing to respond to cache pressure

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi John,

I am seeing this problem with Ceph v9.0.1 with the v4.1 kernel on all
nodes.  This system is using 4 Ceph FS client systems. They all have
the kernel driver version of CephFS loaded, but none are mounting the
file system. All 4 clients are using the libcephfs VFS interface to
Ganesha NFS (V2.2.0-2) and Samba (Version 4.3.0pre1-GIT-0791bb0) to
share out the Ceph file system.

# ceph -s
    cluster 6d8aae1e-1125-11e5-a708-001b78e265be
     health HEALTH_WARN
            4 near full osd(s)
            mds0: Client ede-c2-gw01 failing to respond to cache pressure
            mds0: Client ede-c2-gw02:cephfs failing to respond to cache pressure
            mds0: Client ede-c2-gw03:cephfs failing to respond to cache pressure
     monmap e1: 3 mons at
{ede-c2-mon01=10.15.2.121:6789/0,ede-c2-mon02=10.15.2.122:6789/0,ede-c2-mon03=10.15.2.123:6789/0}
            election epoch 8, quorum 0,1,2
ede-c2-mon01,ede-c2-mon02,ede-c2-mon03
     mdsmap e912: 1/1/1 up {0=ede-c2-mds03=up:active}, 2 up:standby
     osdmap e272: 8 osds: 8 up, 8 in
      pgmap v225264: 832 pgs, 4 pools, 188 GB data, 5173 kobjects
            212 GB used, 48715 MB / 263 GB avail
                 832 active+clean
  client io 1379 kB/s rd, 20653 B/s wr, 98 op/s

# cat /proc/version
Linux version 4.1.0-040100-generic (kernel@gomeisa) (gcc version 4.6.3
(Ubuntu/Linaro 4.6.3-1ubuntu5) ) #201506220235 SMP Mon Jun 22 06:36:19
UTC 2015

# ceph -v
ceph version 9.0.1 (997b3f998d565a744bfefaaf34b08b891f8dbf64)

The systems are all running Ubuntu Trusty that has been upgraded to
the 4.1 kernel. This is all physical machines and no VMs.  The test
run that caused the problem was create and verifying 5 million small
files.

We have some tools that flag when Ceph is in a WARN state so it would
be nice to get rid of this warning.

Please let me know what additional information you need.

Thanks,

Eric

On Fri, Jul 10, 2015 at 4:19 AM, 谷枫 <feicheche@xxxxxxxxx> wrote:
> Thank you John,
> All my server is ubuntu14.04 with 3.16 kernel.
> Not all of clients appear this problem, the cluster seems functioning well
> now.
> As you say,i will change the mds_cache_size to 500000 from 100000 to take a
> test, thanks again!
>
> 2015-07-10 17:00 GMT+08:00 John Spray <john.spray@xxxxxxxxxx>:
>>
>>
>> This is usually caused by use of older kernel clients.  I don't remember
>> exactly what version it was fixed in, but iirc we've seen the problem with
>> 3.14 and seen it go away with 3.18.
>>
>> If your system is otherwise functioning well, this is not a critical error
>> -- it just means that the MDS might not be able to fully control its memory
>> usage (i.e. it can exceed mds_cache_size).
>>
>> John
>>
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux