On Thu, Nov 23, 2017 at 9:17 PM, David C <dcsysengineer@xxxxxxxxx> wrote: > Hi All > > I upgraded my 12.2.0 cluster to 12.2.1 a month or two back. I've noticed > that the number of inodes held in cache is only approx 1/5th of my > inode_max. This is a surprise to me as with 12.2.0, and before that Jewel, > after starting an MDS server, the cache would typically fill to the max > within 24 hours. I have ~8 million entries my file system, and most of this > is fairly hot data. I'm seeing quite frequent "failing to respond to cache > pressure" messages, I just have two Kernel clients accessing the filesystem. > > Are there some new defaults I need to change perhaps? Or potentially a bug? > we introduced config option 'mds_cache_memory_limit'. Regards Yan, Zheng > Output of perf dump mds: > >>> "mds": { >>> >>> "request": 184132091, >>> >>> "reply": 184132064, >>> >>> "reply_latency": { >>> >>> "avgcount": 184132064, >>> >>> "sum": 125364.905594355, >>> >>> "avgtime": 0.000680842 >>> >>> }, >>> >>> "forward": 0, >>> >>> "dir_fetch": 9846671, >>> >>> "dir_commit": 562495, >>> >>> "dir_split": 0, >>> >>> "dir_merge": 0, >>> >>> "inode_max": 2500000, >>> >>> "inodes": 444642, >>> >>> "inodes_top": 185845, >>> >>> "inodes_bottom": 127878, >>> >>> "inodes_pin_tail": 130919, >>> >>> "inodes_pinned": 179149, >>> >>> "inodes_expired": 135604208, >>> >>> "inodes_with_caps": 165900, >>> >>> "caps": 165948, >>> >>> "subtrees": 2, >>> >>> "traverse": 187280168, >>> >>> "traverse_hit": 185739606, >>> >>> "traverse_forward": 0, >>> >>> "traverse_discover": 0, >>> >>> "traverse_dir_fetch": 118150, >>> >>> "traverse_remote_ino": 8, >>> >>> "traverse_lock": 60256, >>> >>> "load_cent": 18413221445, >>> >>> "q": 0, >>> >>> "exported": 0, >>> >>> "exported_inodes": 0, >>> >>> "imported": 0, >>> >>> "imported_inodes": 0 >>> >>> } >> >> > > A few extra details: > > Running two MDS servers, one is standby. > Both have mds_cache_size = 2500000. > CentOS 7.3 servers. > Kernel clients are CentOS 7.3 (3.10.0-514.2.2.el7.x86_64) > > Thanks, > David > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com