[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [Xen-users] unexpected Out Of Memory (OOM)
Le jeudi 08 aoÃt 2013 Ã 14:25 +0100, Wei Liu a Ãcrit : > On Thu, Aug 08, 2013 at 01:43:08PM +0200, Olivier Bonvalet wrote: > > > > > > Le jeudi 08 aoÃt 2013 Ã 11:18 +0100, Ian Campbell a Ãcrit : > > > On Thu, 2013-08-08 at 12:10 +0200, Olivier Bonvalet wrote: > > > > > > > > Le jeudi 08 aoÃt 2013 Ã 09:58 +0100, Ian Campbell a Ãcrit : > > > > > On Wed, 2013-08-07 at 23:37 +0200, Olivier Bonvalet wrote: > > > > > > So I recompiled a kernel with the kmemleak feature. I obtain that > > > > > > kind > > > > > > of list, but not sure that it's usefull : > > > > > > > > > > These look to me like valid things to be allocating at boot time, and > > > > > even if they are leaked there isn't enough here to exhaust 8GB by a > > > > > long > > > > > way. > > > > > > > > > > It'd be worth monitoring to see if it grows at all or if anything > > > > > interesting shows up after running for a while with the leak. > > > > > > > > > > Likewise it'd be worth keeping an eye on the process list and slabtop > > > > > and seeing if anything appears to be growing without bound. > > > > > > > > > > Other than that I'm afraid I don't have many smart ideas. > > > > > > > > > > Ian. > > > > > > > > > > > > > > > > > > Ok, then I will become crazy : when I start the kernel with kmemleak=on > > > > in fact I haven't memory leak. The memory usage stay near 300MB. > > > > > > > > Then I restart on the same kernel, without kmemleak=on, the memory usage > > > > jump to 600MB and still grow. > > > > > > > > Olivier > > > > > > > > PS : I retry several time, to confirm that. > > > > > > *boggles* > > > > > > Ian. > > > > > > > > > > > > So, I retried the slabtop test, with more leaked memory, to have better > > visibility : > > > > --- a 2013-08-08 12:29:48.437966407 +0200 > > +++ c 2013-08-08 13:33:41.213711305 +0200 > > @@ -1,23 +1,23 @@ > > - Active / Total Objects (% used) : 186382 / 189232 (98.5%) > > - Active / Total Slabs (% used) : 6600 / 6600 (100.0%) > > - Active / Total Caches (% used) : 100 / 151 (66.2%) > > - Active / Total Size (% used) : 111474.55K / 113631.58K (98.1%) > > - Minimum / Average / Maximum Object : 0.33K / 0.60K / 8.32K > > + Active / Total Objects (% used) : 2033635 / 2037851 (99.8%) > > + Active / Total Slabs (% used) : 70560 / 70560 (100.0%) > > + Active / Total Caches (% used) : 101 / 151 (66.9%) > > + Active / Total Size (% used) : 1289959.44K / 1292725.98K (99.8%) > > + Minimum / Average / Maximum Object : 0.33K / 0.63K / 8.32K > > > > OBJS ACTIVE USE OBJ SIZE SLABS OBJ/SLAB CACHE SIZE NAME > > > > - 55048 55038 99% 0.56K 1966 28 31456K filp > > > > - 29536 29528 99% 0.50K 923 32 14768K cred_jar > > > > - 22909 22909 100% 0.51K 739 31 11824K dentry > > > > +831572 831552 99% 0.56K 29699 28 475184K filp > > > > +501664 501635 99% 0.50K 15677 32 250832K cred_jar > > > > +172453 172432 99% 0.51K 5563 31 89008K dentry > > > > +150920 150906 99% 0.91K 4312 35 137984K proc_inode_cache > > > > + 54686 54652 99% 0.43K 1478 37 23648K task_delay_info > > > > + 54656 54651 99% 1.98K 3416 16 109312K task_struct > > > > + 54652 54651 99% 1.19K 2102 26 67264K task_xstate > > > > + 54648 54644 99% 0.44K 1518 36 24288K pid > > > > + 54648 54645 99% 1.38K 2376 23 76032K signal_cache > > > > + 38200 38188 99% 0.38K 1910 20 15280K kmalloc-64 > > > > 11803 11774 99% 0.43K 319 37 5104K sysfs_dir_cache > > > > - 7350 7327 99% 0.91K 210 35 6720K proc_inode_cache > > > > - 5520 5465 99% 0.38K 276 20 2208K anon_vma_chain > > > > - 5216 5137 98% 0.50K 163 32 2608K vm_area_struct > > > > - 3984 3978 99% 0.33K 166 24 1328K kmalloc-8 > > > > - 3811 3798 99% 0.84K 103 37 3296K inode_cache > > > > - 3384 3359 99% 0.44K 94 36 1504K pid > > > > - 3381 3362 99% 1.38K 147 23 4704K signal_cache > > > > - 3380 3366 99% 1.19K 130 26 4160K task_xstate > > > > - 3376 3366 99% 1.98K 211 16 6752K task_struct > > > > - 3367 3367 100% 0.43K 91 37 1456K task_delay_info > > > > - 2886 2864 99% 0.42K 78 37 1248K buffer_head > > > > - 2720 2714 99% 0.93K 80 34 2560K shmem_inode_cache > > > > + 7920 7676 96% 0.38K 396 20 3168K anon_vma_chain > > > > + 7808 7227 92% 0.50K 244 32 3904K vm_area_struct > > > > + 5624 5581 99% 0.42K 152 37 2432K buffer_head > > > > + 4316 4308 99% 1.22K 166 26 5312K ext4_inode_cache > > > > + 3984 3977 99% 0.33K 166 24 1328K kmalloc-8 > > > > > > So in 1 hour, "filp" and "cred_jar" eat a lot of memory. > > > > filp should be the pointer to struct file. cred_jar is slab cache to > store credentials. These two are not Xen-specific. > > Dentry grows too. But that's not Xen-specific either. > > > Wei. > Great. So I stop here, I will try with kernel developers. Thanks for the help ! _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |