> simple thing to do in any case. ÂYou can watch the entries in slabinfo > and see if any of the ones with sizes over 4096 bytes are getting used > often. ÂYou can also watch /proc/buddyinfo and see how often columns > other than the first couple are moving around. I collected some information from /proc/{buddyinfo,meminfo,slabinfo,vmstat} and let it sit, polling approximately once per minute. I have some results correlated with another page eviction in graphs. The graph is here: http://files.spotify.com/memcut/memgraph-20101124.png The last sudden eviction there occurred somewhere between 22:30 and 22:45. Some URL:s that can be compared for those periods: Before: http://files.spotify.com/memcut/memstat-20101124/2010-11-24T22:39:30/vmstat http://files.spotify.com/memcut/memstat-20101124/2010-11-24T22:39:30/buddyinfo http://files.spotify.com/memcut/memstat-20101124/2010-11-24T22:39:30/meminfo http://files.spotify.com/memcut/memstat-20101124/2010-11-24T22:39:30/slabinfo After: http://files.spotify.com/memcut/memstat-20101124/2010-11-24T22:45:31/vmstat http://files.spotify.com/memcut/memstat-20101124/2010-11-24T22:45:31/buddyinfo http://files.spotify.com/memcut/memstat-20101124/2010-11-24T22:45:31/meminfo http://files.spotify.com/memcut/memstat-20101124/2010-11-24T22:45:31/slabinfo A more complete set of files for several minutes before/after/during is in this tarball: http://files.spotify.com/memcut/memgraph-20101124.tar.gz Diffing the two slabinfos above yields among other things the following expected or at least plausible decreases since they correlate with the symptomes: -ext3_inode_cache 254638 693552 760 43 8 : tunables 0 0 0 : slabdata 16130 16130 0 +ext3_inode_cache 81698 468958 760 43 8 : tunables 0 0 0 : slabdata 10906 10906 0 -dentry 96504 344232 192 42 2 : tunables 0 0 0 : slabdata 8196 8196 0 +dentry 55628 243810 192 42 2 : tunables 0 0 0 : slabdata 5805 5805 0 And if my understanding is correct, buffer_head would be meta-data associated with cached pages and thus be expected to drop in correlation with less data cached: -buffer_head 2109250 4979052 104 39 1 : tunables 0 0 0 : slabdata 127668 127668 0 +buffer_head 838859 4059822 104 39 1 : tunables 0 0 0 : slabdata 104098 104098 0 My knowledge of the implementation is lacking far too much to know where best to look for the likely culprit in terms of the root cause of the eviction. The one thing I thought looked suspicious was the kmalloc increases: -kmalloc-4096 301 328 4096 8 8 : tunables 0 0 0 : slabdata 41 41 0 +kmalloc-4096 637 680 4096 8 8 : tunables 0 0 0 : slabdata 85 85 0 -kmalloc-2048 18215 19696 2048 16 8 : tunables 0 0 0 : slabdata 1231 1231 0 +kmalloc-2048 41908 51792 2048 16 8 : tunables 0 0 0 : slabdata 3237 3237 0 -kmalloc-1024 85444 97280 1024 32 8 : tunables 0 0 0 : slabdata 3040 3040 0 +kmalloc-1024 267031 327104 1024 32 8 : tunables 0 0 0 : slabdata 10222 10222 0 -kmalloc-512 1988 2176 512 32 4 : tunables 0 0 0 : slabdata 68 68 0 +kmalloc-512 1692 2080 512 32 4 : tunables 0 0 0 : slabdata 65 65 0 -kmalloc-256 102588 119776 256 32 2 : tunables 0 0 0 : slabdata 3743 3743 0 +kmalloc-256 308470 370720 256 32 2 : tunables 0 0 0 : slabdata 11585 11585 0 -kmalloc-128 8435 9760 128 32 1 : tunables 0 0 0 : slabdata 305 305 0 +kmalloc-128 8524 9760 128 32 1 : tunables 0 0 0 : slabdata 305 305 0 -kmalloc-64 96176 405440 64 64 1 : tunables 0 0 0 : slabdata 6335 6335 0 +kmalloc-64 50001 352448 64 64 1 : tunables 0 0 0 : slabdata 5507 5507 0 If my interpretation and understanding is correct, this indicates that for example, ~3000 to ~10000 3-order allocations resulting from 1 kb kmalloc():s. Meaning about 0.2 gig ( 7000*4*8*1024/1024/1024). Add the other ones and we get some more, but only a few hundred megs in total. Going by the hypothesis that we are seeing the same thing as reported by Simon Kirby (I'll respond to that E-Mail separately), the total amount is (as far as I understand) not the important part, but the fact that we saw a non-trivial increase in 3-order allocations would perhaps be a consistent observation in that frequent 3-order allocations might be more likely to trigger the behavior Simon reports. I can do additional post-processing on the data that was dumped (such as graphing it), but I'm not sure which parts would be the most interesting for figuring out what is going on. Is there something else I should be collecting for that matter, than the vmstat/slabinfo/buddyinfo/meminfo? -- / Peter Schuller aka scode -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxxx For more info on Linux MM, see: http://www.linux-mm.org/ . Fight unfair telecom policy in Canada: sign http://dissolvethecrtc.ca/ Don't email: <a href