Re: PROBLEM: Memory leaking when running kubernetes cronjobs

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Roman,

unfortunately Kubernetes seems to be using version 1 cgroups, so I can't 
see that stat - I'll investigate if there's a way to get Kube to use V2 so 
we can check this..

There wasn't memory pressure, I just run it in a pretty controlled way 
when running the test - so initially it sounds like what I saw was 
expected. I then ran stress --vm 16 --vm-bytes 2147483648 which did create 
some memory pressure and I saw oom killer getting invoked - it seemed 
pretty similar behaviour to before where only a small amount of the "lost" 
memory was reclaimed... Maybe I was being too severe with stress and the 
memory would be reclaimed at a slower rate under more reasonable memory 
pressure?

thanks,

Dan McGinnes

IBM Cloud - Containers performance

Int Tel: 247359        Ext Tel: 01962 817359

Notes: Daniel McGinnes/UK/IBM
Email: MCGINNES@xxxxxxxxxx

IBM (UK) Ltd, Hursley Park,Winchester,Hampshire, SO21 2JN



From:   Roman Gushchin <guro@xxxxxx>
To:     Daniel McGinnes <MCGINNES@xxxxxxxxxx>
Cc:     "cgroups@xxxxxxxxxxxxxxx" <cgroups@xxxxxxxxxxxxxxx>, Nathaniel 
Rockwell <nrockwell@xxxxxxxxxx>
Date:   19/09/2018 17:01
Subject:        Re: PROBLEM: Memory leaking when running kubernetes 
cronjobs



Hi Daniel!

The best thing to look at is nr_dying_descendants in 
<cgroupfs>/cgroup.stat.
This is exactly the number of deleted but not yet reclaimed cgroups in the
subtree. But you need to be on cgroup v2 in order to have it.

The second question is: is there any memory pressure on your host?
If not, it's pretty much expected that dying cgroups will be not reclaimed
(assuming new cgroups are periodically created).

MemAvailable will decrease in this case, because all per-cpu memory will 
be
treated as unreclaimable. It's a known accounting issue. MemAvailable was
never precise.

Thanks!


On Wed, Sep 19, 2018 at 10:37:27AM +0000, Daniel McGinnes wrote:
> Hi Roman,
> 
> many thanks for the response, very promising news. I just wondered 
exactly 
> what difference in behaviour I should expect with the patches? Would you 

> expect to see MemAvailable not decreasing at all, or do they mean that 
the 
> memory gets reclaimed better?
> 
> I applied those patches to 4.19-rc3 and after the first hour it looks to 

> have leaked at a similar rate (I haven't checked if it will be reclaimed 

> under pressure yet)
> 
> This is my first time applying patches to a kernel build so it is 
possible 
> I made a mistake, but just wanted to double check the expected behaviour 

> whilst I double check that.
> 
> thanks,
> 
> Dan McGinnes
> 
> IBM Cloud - Containers performance
> 
> Int Tel: 247359        Ext Tel: 01962 817359
> 
> Notes: Daniel McGinnes/UK/IBM
> Email: MCGINNES@xxxxxxxxxx
> 
> IBM (UK) Ltd, Hursley Park,Winchester,Hampshire, SO21 2JN
> 
> 
> 
> From:   Roman Gushchin <guro@xxxxxx>
> To:     Daniel McGinnes <MCGINNES@xxxxxxxxxx>
> Cc:     "cgroups@xxxxxxxxxxxxxxx" <cgroups@xxxxxxxxxxxxxxx>
> Date:   18/09/2018 16:54
> Subject:        Re: PROBLEM: Memory leaking when running kubernetes 
> cronjobs
> 
> 
> 
> Hi Daniel!
> 
> Thank you for the very detailed and full report!
> We've experienced the same (or very similar problem), when memory 
cgroups
> were staying in the dying state for a long time, so that the number of
> dying cgroups grew steadily with time.
> 
> I've investigated the issue and found several problems in the memory
> reclaim and accounting code.
> 
> The following commits from the next tree are solving the problem in our 
> case:
> 
> 010cb21d4ede math64: prevent double calculation of DIV64_U64_ROUND_UP() 
> arguments
> f77d7a05670d mm: don't miss the last page because of round-off error
> d18bf0af683e mm: drain memcg stocks on css offlining
> 71cd51b2e1ca mm: rework memcg kernel stack accounting
> f3a2fccbce15 mm: slowly shrink slabs with a relatively small number of 
> objects
> 
> It should be relatively easy to backport them, please, let me know,
> if I can help somehow.
> 
> Thank you!
> 
> Roman
> 
> 
> On Tue, Sep 18, 2018 at 01:33:20PM +0100, Daniel McGinnes wrote:
> > (Resending to mailing list as first attempt got bounced for having 
html)
> > Hi, 
> > 
> > we are hitting the following kernel issue, which we think might be 
> caused 
> > by memory cgroups.: 
> > 
> > [1.] One line summary of the problem: Memory leaking when running 
> > kubernetes cronjobs 
> > 
> > [2.] Full description of the problem/report: 
> > We are using Kubernetes V1.8.15 with docker 18.03.1-ce. 
> > We schedule 50 Kubernetes cronjobs to run every 5 minutes. Each 
cronjob 
> > will create a simple busybox container, echo hello world, then 
> terminate. 
> > In the first set of data I have I let this run for 1 hour, and in this 

> > time the Available memory had reduced from 31256704 kB to 30461224 kB 
- 
> so 
> > a loss of 776 MB. 
> > There doesn't appear to be any processes left behind, or any growth in 

> any 
> > other processes to explain where the memory has gone. 
> > echo 3 > /proc/sys/vm/drop_caches causes some of the memory to be 
> > returned, but the majority remains leaked, and the only way to free it 

> > appears to be to reboot the system. 
> > 
> > We are currently running Ubuntu 4.15.0-32.35-generic 4.15.18 and have 
> > previously observed similar issues on Ubuntu 16.04 with Kernel 
> > 4.4.0-89-generic #112-Ubuntu SMP Mon Jul 31 19:38:41 UTC 2017 and 
Debian 
> 
> > 9.4 running 4.9.0-6-amd64 #1 SMP Debian 4.9.82-1+deb9u3 (2018-03-02), 
> and 
> > I have just recreated on Linux version 4.19.0-041900rc3-generic 
> > 
> > The leak was more severe on the Debian system, and investigations 
there 
> > showed leaks in pcpu_get_vm_areas and were related to memory cgroups. 
> > Running with Kernel 4.17 on debian showed a leak at a similar rate to 
> what 
> > we now observe on 4.15 & 4.19 Kernels with Ubuntu. This leak causes us 

> > issues as we need to run cronjobs regularly for backups and want the 
> > systems to remain up for months. 
> > 
> > Kubernetes will create a new cgroup each time the cronjob runs, but 
> these 
> > are removed when the job completes (which takes a few seconds). If I 
use 
> 
> > systemd-cgtop I don't see any increase in cgroups over time - but if I 

> > monitor /proc/cgroups over time I can see num_cgroups for memory 
> increases 
> > for about the first 18 hours, and then stabilises at around 4300. 
> > 
> > For the duration of the test I collected slabinfo, meminfo, 
vmallocinfo 
> & 
> > cgroups - which I can provide. 
> > 
> > I stopped the cronjobs after 1 hour, and I then left the system idle 
for 
> 
> > 10 minutes. I then ran echo 3 > /proc/sys/vm/drop_caches . This seemed 

> to 
> > free ~240MB - but this still leaves ~500MB lost. I then left the 
system 
> > idle for a further 20 minutes, and MemoryAvailable didn't seem to be 
> > increasing significantly. 
> > 
> > I have performed another run for a longer period, and after 89 hours, 
> the 
> > MemoryAvailable has reduced by ~5.5GB - the rate of decrease seems 
less 
> > severe after the first 4-5 hours - but clearly continues to decrease 
at 
> a 
> > rate of ~1 GB per day for the duration of the test. 
> > 
> > After ~110 hours I ran `stress --vm 16 --vm-bytes 2147483648` to 
> generate 
> > some memory pressure to see how much would be reclaimed. 
> > Before running the stress utility  MemAvailable was 24585576 
> > After running the stress utility MemAvailable was 27266788 
> > 
> > I ran ps aux | awk '{sum+=$6} END {print sum / 1024}' 
> > which showed: 1220.44 
> > 
> > Since the start of the test it looks like we have leaked 3.8GB which 
is 
> > not reclaimable under memory pressure. 
> > 
> > This was originally raised as this Ubuntu bug: 
> > 
> 
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1792349 

> , which 
> > requested a kernel bug be opened. 
> > 
> > [3.] Keywords (i.e., modules, networking, kernel): 
> > [4.] Kernel version (from /proc/version):Kernel version: Linux version 

> > 4.19.0-041900rc3-generic (kernel@gloin) (gcc version 8.2.0 (Ubuntu 
> > 8.2.0-6ubuntu1)) #201809120832 SMP Wed Sep 12 12:35:08 UTC 2018 
> > [5.] Output of Oops.. message (if applicable) with symbolic 
information 
> > resolved (see Documentation/oops-tracing.txt): n/a 
> > [6.] A small shell script or example program which triggers the 
problem 
> > (if possible): Tricky to provide as Kubernetes is fairly complex to 
> > setup/configure 
> > [7.] Environment 
> >         Description:        Ubuntu 18.04.1 LTS 
> >         Release:        18.04 
> > [7.1.] Software (add the output of the ver_linux script here): 
> > Linux stage-dal09-carrier1-worker-37.alchemy.ibm.com 
> > 4.19.0-041900rc3-generic #201809120832 SMP Wed Sep 12 12:35:08 UTC 
2018 
> > x86_64 x86_64 x86_64 GNU/Linux 
> > 
> > GNU Make                    4.1 
> > Binutils                    2.30 
> > Util-linux                  2.31.1 
> > Mount                       2.31.1 
> > Module-init-tools           24 
> > E2fsprogs                   1.44.1 
> > Xfsprogs                    4.9.0 
> > Nfs-utils                   1.3.3 
> > Linux C Library             2.27 
> > Dynamic linker (ldd)        2.27 
> > Linux C++ Library           6.0.25 
> > Procps                      3.3.12 
> > Net-tools                   2.10 
> > Kbd                         2.0.4 
> > Console-tools               2.0.4 
> > Sh-utils                    8.28 
> > Udev                        237 
> > Modules Loaded              async_memcpy async_pq async_raid6_recov 
> > async_tx async_xor autofs4 binfmt_misc bpfilter bridge btrfs cirrus 
> > crc32_pclmul crct10dif_pclmul cryptd drm drm_kms_helper fb_sys_fops 
> floppy 
> > ghash_clmulni_intel hid hid_generic i2c_piix4 ib_cm ib_core ib_iser 
> > ide_core ide_pci_generic input_leds intel_rapl intel_rapl_perf 
> ip6_tables 
> > ip6t_REJECT ip6t_rt ip6table_filter ip_set ip_set_hash_ip 
> ip_set_hash_net 
> > ip_tables ip_tunnel ip_vs ipip ipt_MASQUERADE ipt_REJECT 
iptable_filter 
> > iptable_mangle iptable_nat iptable_raw iscsi_tcp iw_cm joydev 
libcrc32c 
> > libiscsi libiscsi_tcp linear llc mac_hid multipath nf_conntrack 
> > nf_conntrack_broadcast nf_conntrack_ftp nf_conntrack_netbios_ns 
> > nf_conntrack_netlink nf_defrag_ipv4 nf_defrag_ipv6 nf_log_common 
> > nf_log_ipv4 nf_log_ipv6 nf_nat nf_nat_ftp nf_nat_ipv4 nf_reject_ipv4 
> > nf_reject_ipv6 nfnetlink overlay parport parport_pc pata_acpi piix 
ppdev 
> 
> > psmouse raid0 raid1 raid10 raid456 raid6_pq rdma_cm sb_edac 
sch_fq_codel 
> 
> > scsi_transport_iscsi serio_raw stp sunrpc syscopyarea sysfillrect 
> > sysimgblt ttm tunnel4 usbhid veth x_tables xen_privcmd xenfs xfrm_algo 

> > xfrm_user xor xt_LOG xt_addrtype xt_comment xt_conntrack xt_hl 
xt_limit 
> > xt_mark xt_multiport xt_nat xt_recent xt_set xt_statistic xt_tcpudp 
> > zstd_compress 
> > 
> > [7.2.] Processor information (from /proc/cpuinfo): 
> > 
> > (There are 16 processors - just including first one for brevity) 
> > 
> > processor        : 0 
> > vendor_id        : GenuineIntel 
> > cpu family        : 6 
> > model                : 63 
> > model name        : Intel(R) Xeon(R) CPU E5-2683 v3 @ 2.00GHz 
> > stepping        : 2 
> > microcode        : 0x3d 
> > cpu MHz                : 2000.018 
> > cache size        : 35840 KB 
> > physical id        : 0 
> > siblings        : 16 
> > core id                : 0 
> > cpu cores        : 16 
> > apicid                : 0 
> > initial apicid        : 0 
> > fpu                : yes 
> > fpu_exception        : yes 
> > cpuid level        : 13 
> > wp                : yes 
> > flags                : fpu vme de pse tsc msr pae mce cx8 apic sep 
mtrr 
> > pge mca cmov pat pse36 clflush acpi mmx fxsr sse sse2 ht syscall 
pdpe1gb 
> 
> > rdtscp lm constant_tsc rep_good nopl cpuid pni pclmulqdq ssse3 fma 
cx16 
> > pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer xsave avx 
f16c 
> 
> > rdrand hypervisor lahf_lm abm cpuid_fault invpcid_single pti 
intel_ppin 
> > ssbd ibrs ibpb stibp fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms 
> invpcid 
> > xsaveopt flush_l1d 
> > bugs                : cpu_meltdown spectre_v1 spectre_v2 
> spec_store_bypass 
> > l1tf 
> > bogomips        : 4000.08 
> > clflush size        : 64 
> > cache_alignment        : 64 
> > address sizes        : 46 bits physical, 48 bits virtual 
> > power management: 
> > 
> > [7.3.] Module information (from /proc/modules): 
> > binfmt_misc 20480 1 - Live 0xffffffffc0a7c000 
> > veth 24576 0 - Live 0xffffffffc0a75000 
> > xt_set 20480 4 - Live 0xffffffffc0a6f000 
> > xt_multiport 16384 11 - Live 0xffffffffc0a65000 
> > iptable_raw 16384 1 - Live 0xffffffffc0a60000 
> > iptable_mangle 16384 1 - Live 0xffffffffc0a45000 
> > ip_set_hash_ip 32768 1 - Live 0xffffffffc0a57000 
> > ip_set_hash_net 32768 2 - Live 0xffffffffc0a4a000 
> > ip_set 40960 3 xt_set,ip_set_hash_ip,ip_set_hash_net, Live 
> > 0xffffffffc0a3a000 
> > ipip 16384 0 - Live 0xffffffffc0a35000 
> > tunnel4 16384 1 ipip, Live 0xffffffffc0a30000 
> > ip_tunnel 24576 1 ipip, Live 0xffffffffc0a29000 
> > xt_statistic 16384 5172 - Live 0xffffffffc0a24000 
> > xt_nat 16384 8652 - Live 0xffffffffc0a1f000 
> > xt_recent 20480 2 - Live 0xffffffffc0a15000 
> > ipt_MASQUERADE 16384 3 - Live 0xffffffffc0a10000 
> > xt_mark 16384 568 - Live 0xffffffffc096c000 
> > bridge 159744 0 - Live 0xffffffffc09e8000 
> > stp 16384 1 bridge, Live 0xffffffffc09e1000 
> > llc 16384 2 bridge,stp, Live 0xffffffffc09d9000 
> > xfrm_user 36864 1 - Live 0xffffffffc09cb000 
> > xfrm_algo 16384 1 xfrm_user, Live 0xffffffffc09c6000 
> > nf_conntrack_netlink 40960 0 - Live 0xffffffffc09bb000 
> > nfnetlink 16384 3 ip_set,nf_conntrack_netlink, Live 0xffffffffc0967000 

> > xt_comment 16384 39264 - Live 0xffffffffc095c000 
> > overlay 106496 8 - Live 0xffffffffc09a0000 
> > iptable_nat 16384 1 - Live 0xffffffffc094e000 
> > nf_nat_ipv4 16384 2 ipt_MASQUERADE,iptable_nat, Live 
0xffffffffc0957000 
> > ip_vs 147456 0 - Live 0xffffffffc097b000 
> > ppdev 20480 0 - Live 0xffffffffc0896000 
> > intel_rapl 20480 0 - Live 0xffffffffc0975000 
> > sb_edac 24576 0 - Live 0xffffffffc0947000 
> > joydev 20480 0 - Live 0xffffffffc0961000 
> > intel_rapl_perf 16384 0 - Live 0xffffffffc0891000 
> > input_leds 16384 0 - Live 0xffffffffc088c000 
> > serio_raw 16384 0 - Live 0xffffffffc0887000 
> > parport_pc 36864 0 - Live 0xffffffffc08d4000 
> > parport 49152 2 ppdev,parport_pc, Live 0xffffffffc08c2000 
> > mac_hid 16384 0 - Live 0xffffffffc0882000 
> > ip6t_REJECT 16384 1 - Live 0xffffffffc087a000 
> > nf_reject_ipv6 16384 1 ip6t_REJECT, Live 0xffffffffc0875000 
> > nf_log_ipv6 16384 3 - Live 0xffffffffc0870000 
> > xt_hl 16384 22 - Live 0xffffffffc086b000 
> > ip6t_rt 16384 3 - Live 0xffffffffc0863000 
> > ipt_REJECT 16384 23 - Live 0xffffffffc085b000 
> > nf_reject_ipv4 16384 1 ipt_REJECT, Live 0xffffffffc0856000 
> > nf_log_ipv4 16384 3 - Live 0xffffffffc084e000 
> > nf_log_common 16384 2 nf_log_ipv6,nf_log_ipv4, Live 0xffffffffc0846000 

> > xt_LOG 16384 6 - Live 0xffffffffc083e000 
> > xt_limit 16384 9 - Live 0xffffffffc082f000 
> > xt_tcpudp 16384 20850 - Live 0xffffffffc082a000 
> > xt_addrtype 16384 18 - Live 0xffffffffc0795000 
> > sch_fq_codel 20480 17 - Live 0xffffffffc078f000 
> > xt_conntrack 16384 155 - Live 0xffffffffc0822000 
> > ip6table_filter 16384 1 - Live 0xffffffffc07c4000 
> > ip6_tables 28672 53 ip6table_filter, Live 0xffffffffc0816000 
> > nf_conntrack_netbios_ns 16384 0 - Live 0xffffffffc080e000 
> > nf_conntrack_broadcast 16384 1 nf_conntrack_netbios_ns, Live 
> > 0xffffffffc073d000 
> > nf_nat_ftp 16384 0 - Live 0xffffffffc06cc000 
> > nf_nat 32768 3 xt_nat,nf_nat_ipv4,nf_nat_ftp, Live 0xffffffffc079f000 
> > nf_conntrack_ftp 20480 1 nf_nat_ftp, Live 0xffffffffc0804000 
> > nf_conntrack 143360 11 
> > 
> 
xt_nat,ipt_MASQUERADE,nf_conntrack_netlink,nf_nat_ipv4,ip_vs,xt_conntrack,nf_conntrack_netbios_ns,nf_conntrack_broadcast,nf_nat_ftp,nf_nat,nf_conntrack_ftp, 

> 
> > Live 0xffffffffc07d4000 
> > nf_defrag_ipv6 20480 1 nf_conntrack, Live 0xffffffffc07ca000 
> > ib_iser 53248 0 - Live 0xffffffffc07b6000 
> > nf_defrag_ipv4 16384 1 nf_conntrack, Live 0xffffffffc0743000 
> > rdma_cm 57344 1 ib_iser, Live 0xffffffffc072e000 
> > iptable_filter 16384 1 - Live 0xffffffffc0729000 
> > iw_cm 45056 1 rdma_cm, Live 0xffffffffc06b1000 
> > ib_cm 53248 1 rdma_cm, Live 0xffffffffc07a8000 
> > bpfilter 16384 0 - Live 0xffffffffc053a000 
> > ib_core 237568 4 ib_iser,rdma_cm,iw_cm,ib_cm, Live 0xffffffffc0754000 
> > iscsi_tcp 20480 0 - Live 0xffffffffc074e000 
> > libiscsi_tcp 20480 1 iscsi_tcp, Live 0xffffffffc0748000 
> > sunrpc 356352 1 - Live 0xffffffffc06d1000 
> > libiscsi 57344 3 ib_iser,iscsi_tcp,libiscsi_tcp, Live 
0xffffffffc06bd000 
> 
> > xenfs 16384 1 - Live 0xffffffffc0523000 
> > xen_privcmd 20480 1 xenfs, Live 0xffffffffc0330000 
> > scsi_transport_iscsi 98304 4 ib_iser,iscsi_tcp,libiscsi, Live 
> > 0xffffffffc0698000 
> > ip_tables 24576 14 
> iptable_raw,iptable_mangle,iptable_nat,iptable_filter, 
> > Live 0xffffffffc045f000 
> > x_tables 36864 23 
> > 
> 
xt_set,xt_multiport,iptable_raw,iptable_mangle,xt_statistic,xt_nat,xt_recent,ipt_MASQUERADE,xt_mark,xt_comment,ip6t_REJECT,xt_hl,ip6t_rt,ipt_REJECT,xt_LOG,xt_limit,xt_tcpudp,xt_addrtype,xt_conntrack,ip6table_filter,ip6_tables,iptable_filter,ip_tables, 

> 
> > Live 0xffffffffc068e000 
> > autofs4 40960 2 - Live 0xffffffffc0683000 
> > btrfs 1159168 0 - Live 0xffffffffc0567000 
> > zstd_compress 159744 1 btrfs, Live 0xffffffffc053f000 
> > raid10 53248 0 - Live 0xffffffffc052c000 
> > raid456 151552 0 - Live 0xffffffffc04fd000 
> > async_raid6_recov 20480 1 raid456, Live 0xffffffffc04f7000 
> > async_memcpy 16384 2 raid456,async_raid6_recov, Live 
0xffffffffc04f2000 
> > async_pq 16384 2 raid456,async_raid6_recov, Live 0xffffffffc04ed000 
> > async_xor 16384 3 raid456,async_raid6_recov,async_pq, Live 
> > 0xffffffffc04b5000 
> > async_tx 16384 5 
> > raid456,async_raid6_recov,async_memcpy,async_pq,async_xor, Live 
> > 0xffffffffc04b0000 
> > xor 24576 2 btrfs,async_xor, Live 0xffffffffc0483000 
> > raid6_pq 114688 4 btrfs,raid456,async_raid6_recov,async_pq, Live 
> > 0xffffffffc0493000 
> > libcrc32c 16384 5 ip_vs,nf_nat,nf_conntrack,btrfs,raid456, Live 
> > 0xffffffffc047e000 
> > raid1 40960 0 - Live 0xffffffffc046b000 
> > raid0 20480 0 - Live 0xffffffffc0422000 
> > multipath 16384 0 - Live 0xffffffffc02fe000 
> > linear 16384 0 - Live 0xffffffffc0312000 
> > hid_generic 16384 0 - Live 0xffffffffc0304000 
> > usbhid 49152 0 - Live 0xffffffffc02f1000 
> > hid 126976 2 hid_generic,usbhid, Live 0xffffffffc042e000 
> > cirrus 24576 1 - Live 0xffffffffc02e4000 
> > ttm 106496 1 cirrus, Live 0xffffffffc03eb000 
> > drm_kms_helper 167936 1 cirrus, Live 0xffffffffc04c3000 
> > syscopyarea 16384 1 drm_kms_helper, Live 0xffffffffc02ec000 
> > crct10dif_pclmul 16384 0 - Live 0xffffffffc04bb000 
> > sysfillrect 16384 1 drm_kms_helper, Live 0xffffffffc048e000 
> > crc32_pclmul 16384 0 - Live 0xffffffffc0429000 
> > ide_pci_generic 16384 0 - Live 0xffffffffc045a000 
> > sysimgblt 16384 1 drm_kms_helper, Live 0xffffffffc0317000 
> > ghash_clmulni_intel 16384 0 - Live 0xffffffffc0479000 
> > fb_sys_fops 16384 1 drm_kms_helper, Live 0xffffffffc0466000 
> > piix 16384 0 - Live 0xffffffffc0452000 
> > cryptd 24576 1 ghash_clmulni_intel, Live 0xffffffffc0347000 
> > ide_core 106496 2 ide_pci_generic,piix, Live 0xffffffffc0407000 
> > drm 466944 4 cirrus,ttm,drm_kms_helper, Live 0xffffffffc0352000 
> > psmouse 151552 0 - Live 0xffffffffc03c5000 
> > i2c_piix4 24576 0 - Live 0xffffffffc033c000 
> > pata_acpi 16384 0 - Live 0xffffffffc0337000 
> > floppy 77824 0 - Live 0xffffffffc031c000 
> > 
> > 
> > [7.4.] Loaded driver and hardware information (/proc/ioports, 
> /proc/iomem) 
> > 
> > cat /proc/ioports 
> > 0000-0cf7 : PCI Bus 0000:00 
> >   0000-001f : dma1 
> >   0020-0021 : pic1 
> >   0040-0043 : timer0 
> >   0050-0053 : timer1 
> >   0060-0060 : keyboard 
> >   0061-0061 : PNP0800:00 
> >   0064-0064 : keyboard 
> >   0070-0071 : rtc0 
> >   0080-008f : dma page reg 
> >   00a0-00a1 : pic2 
> >   00c0-00df : dma2 
> >   00f0-00ff : fpu 
> >   0170-0177 : 0000:00:01.1 
> >     0170-0177 : ata_piix 
> >   01f0-01f7 : 0000:00:01.1 
> >     01f0-01f7 : ata_piix 
> >   0376-0376 : 0000:00:01.1 
> >     0376-0376 : ata_piix 
> >   0378-037a : parport0 
> >   03c0-03df : vga+ 
> >   03f2-03f2 : floppy 
> >   03f4-03f5 : floppy 
> >   03f6-03f6 : 0000:00:01.1 
> >     03f6-03f6 : ata_piix 
> >   03f7-03f7 : floppy 
> >   03f8-03ff : serial 
> >   04d0-04d1 : pnp 00:01 
> >   08a0-08a3 : pnp 00:01 
> >   0cc0-0ccf : pnp 00:01 
> > 0cf8-0cff : PCI conf1 
> > 0d00-ffff : PCI Bus 0000:00 
> >   10c0-1141 : pnp 00:08 
> >   afe0-afe3 : ACPI GPE0_BLK 
> >   b000-b03f : 0000:00:01.3 
> >     b000-b003 : ACPI PM1a_EVT_BLK 
> >     b004-b005 : ACPI PM1a_CNT_BLK 
> >     b008-b00b : ACPI PM_TMR 
> >   b044-b047 : pnp 00:08 
> >   c000-c0ff : 0000:00:03.0 
> >     c000-c0ff : xen-platform-pci 
> >   c300-c31f : 0000:00:01.2 
> >     c300-c31f : uhci_hcd 
> >   c320-c32f : 0000:00:01.1 
> >     c320-c32f : ata_piix 
> > 
> > cat /proc/iomem 
> > 00000000-00000fff : Reserved 
> > 00001000-0009dfff : System RAM 
> > 0009e000-0009ffff : Reserved 
> > 000a0000-000bffff : PCI Bus 0000:00 
> > 000c0000-000c8bff : Video ROM 
> > 000c9000-000c99ff : Adapter ROM 
> > 000e0000-000fffff : Reserved 
> >   000f0000-000fffff : System ROM 
> > 00100000-efffffff : System RAM 
> > f0000000-fbffffff : PCI Bus 0000:00 
> >   f0000000-f1ffffff : 0000:00:02.0 
> >     f0000000-f1ffffff : cirrusdrmfb_vram 
> >   f2000000-f2ffffff : 0000:00:03.0 
> >     f2000000-f2ffffff : xen-platform-pci 
> >   f3000000-f33fffff : 0000:00:06.0 
> >   f3400000-f3400fff : 0000:00:02.0 
> >     f3400000-f3400fff : cirrusdrmfb_mmio 
> > fc000000-ffffffff : Reserved 
> >   fec00000-fec003ff : IOAPIC 0 
> >   fed00000-fed003ff : HPET 0 
> >     fed00000-fed003ff : PNP0103:00 
> >   fee00000-fee00fff : Local APIC 
> > 100000000-80fbfffff : System RAM 
> >   30b400000-30c2031d0 : Kernel code 
> >   30c2031d1-30cc9fa3f : Kernel data 
> >   30cf22000-30d16dfff : Kernel bss 
> > 80fc00000-80fffffff : RAM buffer 
> > 
> > [7.5.] PCI information ('lspci -vvv' as root) 
> > 
> > sudo lspci -vvv 
> > 00:00.0 Host bridge: Intel Corporation 440FX - 82441FX PMC [Natoma] 
(rev 
> 
> > 02) 
> >         Subsystem: Red Hat, Inc Qemu virtual machine 
> >         Physical Slot: 0 
> >         Control: I/O- Mem- BusMaster+ SpecCycle- MemWINV- VGASnoop- 
> > ParErr- Stepping- SERR- FastB2B- DisINTx- 
> >         Status: Cap- 66MHz- UDF- FastB2B- ParErr- DEVSEL=fast >TAbort- 

> > <TAbort- <MAbort- >SERR- <PERR- INTx- 
> >         Latency: 0 
> > 
> > 00:01.0 ISA bridge: Intel Corporation 82371SB PIIX3 ISA [Natoma/Triton 

> II] 
> > 
> >         Subsystem: Red Hat, Inc Qemu virtual machine 
> >         Physical Slot: 1 
> >         Control: I/O+ Mem+ BusMaster+ SpecCycle- MemWINV- VGASnoop- 
> > ParErr- Stepping- SERR- FastB2B- DisINTx- 
> >         Status: Cap- 66MHz- UDF- FastB2B- ParErr- DEVSEL=medium 
>TAbort- 
> 
> > <TAbort- <MAbort- >SERR- <PERR- INTx- 
> >         Latency: 0 
> > 
> > 00:01.1 IDE interface: Intel Corporation 82371SB PIIX3 IDE 
> [Natoma/Triton 
> > II] (prog-if 80 [Master]) 
> >         Subsystem: XenSource, Inc. 82371SB PIIX3 IDE [Natoma/Triton 
II] 
> >         Physical Slot: 1 
> >         Control: I/O+ Mem- BusMaster+ SpecCycle- MemWINV- VGASnoop- 
> > ParErr- Stepping- SERR- FastB2B- DisINTx- 
> >         Status: Cap- 66MHz- UDF- FastB2B+ ParErr- DEVSEL=medium 
>TAbort- 
> 
> > <TAbort- <MAbort- >SERR- <PERR- INTx- 
> >         Latency: 64 
> >         Region 0: [virtual] Memory at 000001f0 (32-bit, 
> non-prefetchable) 
> > [size=8] 
> >         Region 1: [virtual] Memory at 000003f0 (type 3, 
> non-prefetchable) 
> >         Region 2: [virtual] Memory at 00000170 (32-bit, 
> non-prefetchable) 
> > [size=8] 
> >         Region 3: [virtual] Memory at 00000370 (type 3, 
> non-prefetchable) 
> >         Region 4: I/O ports at c320 [size=16] 
> >         Kernel driver in use: ata_piix 
> >         Kernel modules: pata_acpi, piix, ide_pci_generic 
> > 
> > 00:01.2 USB controller: Intel Corporation 82371SB PIIX3 USB 
> [Natoma/Triton 
> > II] (rev 01) (prog-if 00 [UHCI]) 
> >         Subsystem: XenSource, Inc. 82371SB PIIX3 USB [Natoma/Triton 
II] 
> >         Physical Slot: 1 
> >         Control: I/O+ Mem- BusMaster+ SpecCycle- MemWINV- VGASnoop- 
> > ParErr- Stepping- SERR- FastB2B- DisINTx- 
> >         Status: Cap- 66MHz- UDF- FastB2B- ParErr- DEVSEL=fast >TAbort- 

> > <TAbort- <MAbort- >SERR- <PERR- INTx- 
> >         Latency: 64 
> >         Interrupt: pin D routed to IRQ 23 
> >         Region 4: I/O ports at c300 [size=32] 
> >         Kernel driver in use: uhci_hcd 
> > 
> > 00:01.3 Bridge: Intel Corporation 82371AB/EB/MB PIIX4 ACPI (rev 01) 
> >         Subsystem: Red Hat, Inc Qemu virtual machine 
> >         Physical Slot: 1 
> >         Control: I/O- Mem- BusMaster+ SpecCycle- MemWINV- VGASnoop- 
> > ParErr- Stepping- SERR- FastB2B- DisINTx- 
> >         Status: Cap- 66MHz- UDF- FastB2B- ParErr- DEVSEL=fast >TAbort- 

> > <TAbort- <MAbort- >SERR- <PERR- INTx- 
> >         Latency: 0 
> >         Interrupt: pin A routed to IRQ 9 
> >         Kernel modules: i2c_piix4 
> > 
> > 00:02.0 VGA compatible controller: Cirrus Logic GD 5446 (prog-if 00 
[VGA 
> 
> > controller]) 
> >         Subsystem: XenSource, Inc. GD 5446 
> >         Physical Slot: 2 
> >         Control: I/O+ Mem+ BusMaster+ SpecCycle- MemWINV- VGASnoop- 
> > ParErr- Stepping- SERR- FastB2B- DisINTx- 
> >         Status: Cap- 66MHz- UDF- FastB2B- ParErr- DEVSEL=fast >TAbort- 

> > <TAbort- <MAbort- >SERR- <PERR- INTx- 
> >         Latency: 0 
> >         Interrupt: pin A routed to IRQ 24 
> >         Region 0: Memory at f0000000 (32-bit, prefetchable) [size=32M] 

> >         Region 1: Memory at f3400000 (32-bit, non-prefetchable) 
> [size=4K] 
> >         [virtual] Expansion ROM at 000c0000 [disabled] [size=128K] 
> >         Kernel driver in use: cirrus 
> >         Kernel modules: cirrusfb, cirrus 
> > 
> > 00:03.0 SCSI storage controller: XenSource, Inc. Xen Platform Device 
> (rev 
> > 01) 
> >         Subsystem: XenSource, Inc. Xen Platform Device 
> >         Physical Slot: 3 
> >         Control: I/O+ Mem+ BusMaster+ SpecCycle- MemWINV- VGASnoop- 
> > ParErr- Stepping- SERR- FastB2B- DisINTx- 
> >         Status: Cap- 66MHz- UDF- FastB2B- ParErr- DEVSEL=fast >TAbort- 

> > <TAbort- <MAbort- >SERR- <PERR- INTx- 
> >         Latency: 0 
> >         Interrupt: pin C routed to IRQ 30 
> >         Region 0: I/O ports at c000 [size=256] 
> >         Region 1: Memory at f2000000 (32-bit, prefetchable) [size=16M] 

> >         Kernel driver in use: xen-platform-pci 
> > 
> > 00:06.0 System peripheral: XenSource, Inc. Citrix XenServer PCI Device 

> for 
> > Windows Update (rev 01) 
> >         Subsystem: XenSource, Inc. Citrix XenServer PCI Device for 
> Windows 
> > Update 
> >         Physical Slot: 6 
> >         Control: I/O- Mem+ BusMaster+ SpecCycle- MemWINV- VGASnoop- 
> > ParErr- Stepping- SERR- FastB2B- DisINTx- 
> >         Status: Cap- 66MHz- UDF- FastB2B- ParErr- DEVSEL=fast >TAbort- 

> > <TAbort- <MAbort- >SERR- <PERR- INTx- 
> >         Latency: 0 
> >         Interrupt: pin C routed to IRQ 5 
> >         Region 1: Memory at f3000000 (32-bit, prefetchable) [size=4M] 
> > 
> > [7.6.] SCSI information (from /proc/scsi/scsi) 
> > Attached devices: 
> > 
> > [7.7.] Other information that might be relevant to the problem 
> >        (please look in /proc and include all information that you 
> >        think to be relevant): 
> > 
> > cat slabinfo 
> > 
> > slabinfo - version: 2.1 
> > # name            <active_objs> <num_objs> <objsize> <objperslab> 
> > <pagesperslab> : tunables <limit> <batchcount> <sharedfactor> : 
slabdata 
> 
> > <active_slabs> <num_slabs> <sharedavail> 
> > ovl_inode           9408   9408    680   48    8 : tunables    0    0 
0 
> > : slabdata    196    196      0 
> > ext4_groupinfo_1k     60     60    136   60    2 : tunables    0    0 
0 
> > : slabdata      1      1      0 
> > nf_conntrack_expect      0      0    216   37    2 : tunables    0 0 
> 0 
> > : slabdata      0      0      0 
> > nf_conntrack        6783   6783    320   51    4 : tunables    0    0 
0 
> > : slabdata    133    133      0 
> > rpc_inode_cache      102    102    640   51    8 : tunables    0    0 
0 
> > : slabdata      2      2      0 
> > ext4_groupinfo_4k   3248   3248    144   56    2 : tunables    0    0 
0 
> > : slabdata     58     58      0 
> > btrfs_delayed_ref_head      0      0    152   53    2 : tunables    0 
0 
> >    0 : slabdata      0      0      0 
> > btrfs_delayed_node      0      0    296   55    4 : tunables    0    0 
0 
> : 
> > slabdata      0      0      0 
> > btrfs_ordered_extent      0      0    416   39    4 : tunables    0 0 
> 
> > 0 : slabdata      0      0      0 
> > btrfs_extent_map       0      0    144   56    2 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > btrfs_extent_buffer      0      0    280   58    4 : tunables    0 0 
> 0 
> > : slabdata      0      0      0 
> > btrfs_path             0      0    112   36    1 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > btrfs_inode            0      0   1136   28    8 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > PINGv6                 0      0   1152   28    8 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > RAWv6               2184   2408   1152   28    8 : tunables    0    0 
0 
> > : slabdata     86     86      0 
> > UDPv6               4037   4175   1280   25    8 : tunables    0    0 
0 
> > : slabdata    167    167      0 
> > tw_sock_TCPv6       2176   2176    240   68    4 : tunables    0    0 
0 
> > : slabdata     32     32      0 
> > request_sock_TCPv6      0      0    304   53    4 : tunables    0    0 
0 
> : 
> > slabdata      0      0      0 
> > TCPv6               2716   2716   2304   14    8 : tunables    0    0 
0 
> > : slabdata    194    194      0 
> > kcopyd_job             0      0   3312    9    8 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > dm_uevent              0      0   2632   12    8 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > scsi_sense_cache       0      0    128   64    2 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > cfq_io_cq              0      0    120   68    2 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > mqueue_inode_cache    544    544    960   34    8 : tunables    0    0 
0 
> : 
> > slabdata     16     16      0 
> > fuse_request         640    640    400   40    4 : tunables    0    0 
0 
> > : slabdata     16     16      0 
> > fuse_inode          7566   7566    832   39    8 : tunables    0    0 
0 
> > : slabdata    194    194      0 
> > ecryptfs_key_record_cache      0      0    576   56    8 : tunables 0 
> 
> > 0    0 : slabdata      0      0      0 
> > ecryptfs_headers       8      8   4096    8    8 : tunables    0    0 
0 
> > : slabdata      1      1      0 
> > ecryptfs_inode_cache      0      0    960   34    8 : tunables    0 0 
> 
> > 0 : slabdata      0      0      0 
> > ecryptfs_dentry_info_cache    128    128     32  128    1 : tunables 0 

> >  0    0 : slabdata      1      1      0 
> > ecryptfs_file_cache      0      0     16  256    1 : tunables    0 0 
> 0 
> > : slabdata      0      0      0 
> > ecryptfs_auth_tok_list_item      0      0    832   39    8 : tunables 
0 
> >    0    0 : slabdata      0      0      0 
> > fat_inode_cache        0      0    736   44    8 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > fat_cache              0      0     40  102    1 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > squashfs_inode_cache     92     92    704   46    8 : tunables    0 0 
> 
> > 0 : slabdata      2      2      0 
> > jbd2_journal_head  12376  12376    120   68    2 : tunables    0    0 
0 
> > : slabdata    182    182      0 
> > jbd2_revoke_table_s    768    768     16  256    1 : tunables    0 0 
> 0 
> > : slabdata      3      3      0 
> > ext4_inode_cache   73909  82020   1080   30    8 : tunables    0    0 
0 
> > : slabdata   2734   2734      0 
> > ext4_allocation_context   1024   1024    128   64    2 : tunables    0 
0 
> 
> > 0 : slabdata     16     16      0 
> > ext4_io_end         1984   1984     64   64    1 : tunables    0    0 
0 
> > : slabdata     31     31      0 
> > ext4_extent_status  82192  83130     40  102    1 : tunables    0    0 
0 
> : 
> > slabdata    815    815      0 
> > mbcache             5548   5548     56   73    1 : tunables    0    0 
0 
> > : slabdata     76     76      0 
> > fscrypt_info        7680   7680     32  128    1 : tunables    0    0 
0 
> > : slabdata     60     60      0 
> > fscrypt_ctx         1360   1360     48   85    1 : tunables    0    0 
0 
> > : slabdata     16     16      0 
> > userfaultfd_ctx_cache      0      0    192   42    2 : tunables    0 0 

> >  0 : slabdata      0      0      0 
> > dnotify_struct         0      0     32  128    1 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > posix_timers_cache    544    544    240   68    4 : tunables    0    0 
0 
> : 
> > slabdata      8      8      0 
> > UNIX                8544   8544   1024   32    8 : tunables    0    0 
0 
> > : slabdata    267    267      0 
> > ip4-frags            624    624    208   39    2 : tunables    0    0 
0 
> > : slabdata     16     16      0 
> > secpath_cache       1024   1024    128   64    2 : tunables    0    0 
0 
> > : slabdata     16     16      0 
> > xfrm_dst_cache         0      0    320   51    4 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > xfrm_state             0      0    768   42    8 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > PING                   0      0    960   34    8 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > RAW                 6086   6392    960   34    8 : tunables    0    0 
0 
> > : slabdata    188    188      0 
> > tw_sock_TCP         2176   2176    240   68    4 : tunables    0    0 
0 
> > : slabdata     32     32      0 
> > request_sock_TCP     901    901    304   53    4 : tunables    0    0 
0 
> > : slabdata     17     17      0 
> > TCP                 1605   1605   2176   15    8 : tunables    0    0 
0 
> > : slabdata    107    107      0 
> > hugetlbfs_inode_cache    104    104    624   52    8 : tunables    0 0 

> >  0 : slabdata      2      2      0 
> > dquot               1024   1024    256   64    4 : tunables    0    0 
0 
> > : slabdata     16     16      0 
> > eventpoll_pwq      25984  25984     72   56    1 : tunables    0    0 
0 
> > : slabdata    464    464      0 
> > inotify_inode_mark   3978   3978     80   51    1 : tunables    0    0 
0 
> : 
> > slabdata     78     78      0 
> > dax_cache             42     42    768   42    8 : tunables    0    0 
0 
> > : slabdata      1      1      0 
> > request_queue         39     39   2480   13    8 : tunables    0    0 
0 
> > : slabdata      3      3      0 
> > blkdev_requests       52     52    312   52    4 : tunables    0    0 
0 
> > : slabdata      1      1      0 
> > biovec-max           300    304   8192    4    8 : tunables    0    0 
0 
> > : slabdata     76     76      0 
> > biovec-128            32     32   2048   16    8 : tunables    0    0 
0 
> > : slabdata      2      2      0 
> > biovec-64            512    512   1024   32    8 : tunables    0    0 
0 
> > : slabdata     16     16      0 
> > dmaengine-unmap-256     15     15   2112   15    8 : tunables    0 0 
> 0 
> > : slabdata      1      1      0 
> > dmaengine-unmap-128     30     30   1088   30    8 : tunables    0 0 
> 0 
> > : slabdata      1      1      0 
> > dmaengine-unmap-16  13860  13860    192   42    2 : tunables    0    0 
0 
> : 
> > slabdata    330    330      0 
> > dmaengine-unmap-2   3072   3072     64   64    1 : tunables    0    0 
0 
> > : slabdata     48     48      0 
> > sock_inode_cache   26914  27232    704   46    8 : tunables    0    0 
0 
> > : slabdata    592    592      0 
> > skbuff_fclone_cache   1216   1216    512   64    8 : tunables    0 0 
> 0 
> > : slabdata     19     19      0 
> > skbuff_head_cache  33488  34368    256   64    4 : tunables    0    0 
0 
> > : slabdata    537    537      0 
> > file_lock_cache      640    640    200   40    2 : tunables    0    0 
0 
> > : slabdata     16     16      0 
> > fsnotify_mark_connector   2720   2720     24  170    1 : tunables    0 
0 
> 
> > 0 : slabdata     16     16      0 
> > net_namespace        180    185   6336    5    8 : tunables    0    0 
0 
> > : slabdata     37     37      0 
> > shmem_inode_cache  17822  18400    704   46    8 : tunables    0    0 
0 
> > : slabdata    400    400      0 
> > taskstats            784    784    328   49    4 : tunables    0    0 
0 
> > : slabdata     16     16      0 
> > proc_dir_entry      8603   8694    192   42    2 : tunables    0    0 
0 
> > : slabdata    207    207      0 
> > pde_opener         22848  22848     40  102    1 : tunables    0    0 
0 
> > : slabdata    224    224      0 
> > proc_inode_cache   49963  51888    672   48    8 : tunables    0    0 
0 
> > : slabdata   1081   1081      0 
> > sigqueue             816    816    160   51    2 : tunables    0    0 
0 
> > : slabdata     16     16      0 
> > bdev_cache           312    312    832   39    8 : tunables    0    0 
0 
> > : slabdata      8      8      0 
> > kernfs_node_cache 7775161 7775400    136   60    2 : tunables    0 0 
> 0 
> > : slabdata 129590 129590      0 
> > mnt_cache           5215   5250    384   42    4 : tunables    0    0 
0 
> > : slabdata    125    125      0 
> > filp              167618 171136    256   64    4 : tunables    0    0 
0 
> > : slabdata   2674   2674      0 
> > inode_cache       5517103 5523174    600   54    8 : tunables    0 0 
> 0 
> > : slabdata 102281 102281      0 
> > dentry            14733403 14827218    192   42    2 : tunables    0 0 

> >  0 : slabdata 353029 353029      0 
> > names_cache          296    296   4096    8    8 : tunables    0    0 
0 
> > : slabdata     37     37      0 
> > iint_cache             0      0    120   68    2 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > buffer_head       2184093 2604927    104   39    1 : tunables    0 0 
> 0 
> > : slabdata  66793  66793      0 
> > uts_namespace        592    592    440   37    4 : tunables    0    0 
0 
> > : slabdata     16     16      0 
> > nsproxy             1752   1752     56   73    1 : tunables    0    0 
0 
> > : slabdata     24     24      0 
> > vm_area_struct    108257 109080    200   40    2 : tunables    0    0 
0 
> > : slabdata   2727   2727      0 
> > mm_struct           9330   9330   1088   30    8 : tunables    0    0 
0 
> > : slabdata    311    311      0 
> > files_cache        11040  11040    704   46    8 : tunables    0    0 
0 
> > : slabdata    240    240      0 
> > signal_cache       12270  12270   1088   30    8 : tunables    0    0 
0 
> > : slabdata    409    409      0 
> > sighand_cache       4780   4800   2112   15    8 : tunables    0    0 
0 
> > : slabdata    320    320      0 
> > task_struct         3625   3815   5888    5    8 : tunables    0    0 
0 
> > : slabdata    763    763      0 
> > cred_jar           56532  56532    192   42    2 : tunables    0    0 
0 
> > : slabdata   1346   1346      0 
> > anon_vma_chain    130241 132096     64   64    1 : tunables    0    0 
0 
> > : slabdata   2064   2064      0 
> > anon_vma           83964  84364     88   46    1 : tunables    0    0 
0 
> > : slabdata   1834   1834      0 
> > pid                41603  41728    128   64    2 : tunables    0    0 
0 
> > : slabdata    652    652      0 
> > Acpi-Operand        4256   4256     72   56    1 : tunables    0    0 
0 
> > : slabdata     76     76      0 
> > Acpi-ParseExt        624    624    104   39    1 : tunables    0    0 
0 
> > : slabdata     16     16      0 
> > Acpi-State           459    459     80   51    1 : tunables    0    0 
0 
> > : slabdata      9      9      0 
> > Acpi-Namespace      4284   4284     40  102    1 : tunables    0    0 
0 
> > : slabdata     42     42      0 
> > numa_policy          186    186    264   62    4 : tunables    0    0 
0 
> > : slabdata      3      3      0 
> > trace_event_file    2116   2116     88   46    1 : tunables    0    0 
0 
> > : slabdata     46     46      0 
> > ftrace_event_field  24294  25670     48   85    1 : tunables    0    0 
0 
> : 
> > slabdata    302    302      0 
> > pool_workqueue      4736   4736    256   64    4 : tunables    0    0 
0 
> > : slabdata     74     74      0 
> > radix_tree_node   2679706 2689960    584   56    8 : tunables    0 0 
> 0 
> > : slabdata  48035  48035      0 
> > task_group           816    816    640   51    8 : tunables    0    0 
0 
> > : slabdata     16     16      0 
> > dma-kmalloc-8192       0      0   8192    4    8 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > dma-kmalloc-4096       0      0   4096    8    8 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > dma-kmalloc-2048       0      0   2048   16    8 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > dma-kmalloc-1024       0      0   1024   32    8 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > dma-kmalloc-512        0      0    512   64    8 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > dma-kmalloc-256        0      0    256   64    4 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > dma-kmalloc-128        0      0    128   64    2 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > dma-kmalloc-64         0      0     64   64    1 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > dma-kmalloc-32         0      0     32  128    1 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > dma-kmalloc-16         0      0     16  256    1 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > dma-kmalloc-8          0      0      8  512    1 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > dma-kmalloc-192        0      0    192   42    2 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > dma-kmalloc-96         0      0     96   42    1 : tunables    0    0 
0 
> > : slabdata      0      0      0 
> > kmalloc-8192         827    844   8192    4    8 : tunables    0    0 
0 
> > : slabdata    211    211      0 
> > kmalloc-4096        7301   8208   4096    8    8 : tunables    0    0 
0 
> > : slabdata   1026   1026      0 
> > kmalloc-2048        9739  13440   2048   16    8 : tunables    0    0 
0 
> > : slabdata    840    840      0 
> > kmalloc-1024       20177  22272   1024   32    8 : tunables    0    0 
0 
> > : slabdata    696    696      0 
> > kmalloc-512        62892  94464    512   64    8 : tunables    0    0 
0 
> > : slabdata   1476   1476      0 
> > kmalloc-256        22049  25024    256   64    4 : tunables    0    0 
0 
> > : slabdata    391    391      0 
> > kmalloc-192       1296144 1296918    192   42    2 : tunables    0 0 
> 0 
> > : slabdata  30879  30879      0 
> > kmalloc-128       7230544 7231680    128   64    2 : tunables    0 0 
> 0 
> > : slabdata 112995 112995      0 
> > kmalloc-96        8202597 8204952     96   42    1 : tunables    0 0 
> 0 
> > : slabdata 195356 195356      0 
> > kmalloc-64        2169882 2190016     64   64    1 : tunables    0 0 
> 0 
> > : slabdata  34219  34219      0 
> > kmalloc-32        147668 303872     32  128    1 : tunables    0    0 
0 
> > : slabdata   2374   2374      0 
> > kmalloc-16        156395 170496     16  256    1 : tunables    0    0 
0 
> > : slabdata    666    666      0 
> > kmalloc-8         131584 131584      8  512    1 : tunables    0    0 
0 
> > : slabdata    257    257      0 
> > kmem_cache_node    74067  74880     64   64    1 : tunables    0    0 
0 
> > : slabdata   1170   1170      0 
> > kmem_cache         73491  74004    384   42    4 : tunables    0    0 
0 
> > : slabdata   1762   1762      0 
> > 
> > cat meminfo 
> > MemTotal:       32910232 kB 
> > MemFree:         6742744 kB 
> > MemAvailable:   25470032 kB 
> > Buffers:          947924 kB 
> > Cached:          8611528 kB 
> > SwapCached:            0 kB 
> > Active:          5668624 kB 
> > Inactive:        5374688 kB 
> > Active(anon):    1488612 kB 
> > Inactive(anon):      620 kB 
> > Active(file):    4180012 kB 
> > Inactive(file):  5374068 kB 
> > Unevictable:        5408 kB 
> > Mlocked:            5408 kB 
> > SwapTotal:             0 kB 
> > SwapFree:              0 kB 
> > Dirty:               520 kB 
> > Writeback:             0 kB 
> > AnonPages:       1478048 kB 
> > Mapped:           305152 kB 
> > Shmem:              1164 kB 
> > Slab:           11596304 kB 
> > SReclaimable:    8061752 kB 
> > SUnreclaim:      3534552 kB 
> > KernelStack:       14800 kB 
> > PageTables:        13676 kB 
> > NFS_Unstable:          0 kB 
> > Bounce:                0 kB 
> > WritebackTmp:          0 kB 
> > CommitLimit:    16455116 kB 
> > Committed_AS:    3378396 kB 
> > VmallocTotal:   34359738367 kB 
> > VmallocUsed:           0 kB 
> > VmallocChunk:          0 kB 
> > Percpu:          3316224 kB 
> > HardwareCorrupted:     0 kB 
> > AnonHugePages:     10240 kB 
> > ShmemHugePages:        0 kB 
> > ShmemPmdMapped:        0 kB 
> > CmaTotal:              0 kB 
> > CmaFree:               0 kB 
> > HugePages_Total:       0 
> > HugePages_Free:        0 
> > HugePages_Rsvd:        0 
> > HugePages_Surp:        0 
> > Hugepagesize:       2048 kB 
> > Hugetlb:               0 kB 
> > DirectMap4k:      387072 kB 
> > DirectMap2M:    27920384 kB 
> > DirectMap1G:     6291456 kB 
> > 
> > 
> > 
> > Dan McGinnes
> > 
> > IBM Cloud - Containers performance
> > 
> > Int Tel: 247359        Ext Tel: 01962 817359
> > 
> > Notes: Daniel McGinnes/UK/IBM
> > Email: MCGINNES@xxxxxxxxxx
> > 
> > IBM (UK) Ltd, Hursley Park,Winchester,Hampshire, SO21 2JN
> > Unless stated otherwise above:
> > IBM United Kingdom Limited - Registered in England and Wales with 
number 
> 
> > 741598. 
> > Registered office: PO Box 41, North Harbour, Portsmouth, Hampshire PO6 

> 3AU
> > 
> 
> 
> 
> 
> Unless stated otherwise above:
> IBM United Kingdom Limited - Registered in England and Wales with number 

> 741598. 
> Registered office: PO Box 41, North Harbour, Portsmouth, Hampshire PO6 
3AU
> 




Unless stated otherwise above:
IBM United Kingdom Limited - Registered in England and Wales with number 
741598. 
Registered office: PO Box 41, North Harbour, Portsmouth, Hampshire PO6 3AU




[Index of Archives]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Security]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]     [Monitors]

  Powered by Linux