[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] Load increase after memory upgrade (part2)
On Tue, Jan 24, 2012 at 08:58:22AM +0000, Jan Beulich wrote: > >>> On 23.01.12 at 23:32, Konrad Rzeszutek Wilk <konrad.wilk@xxxxxxxxxx> > >>> wrote: > > On Wed, Jan 18, 2012 at 10:29:23AM -0400, Konrad Rzeszutek Wilk wrote: > >> On Wed, Jan 18, 2012 at 11:35:35AM +0000, Jan Beulich wrote: > >> > >>> On 17.01.12 at 22:02, Konrad Rzeszutek Wilk <konrad.wilk@xxxxxxxxxx> > >> > >>> wrote: > >> > > The issue as I understand is that the DVB drivers allocate their > >> > > buffers > >> > > from 0->4GB most (all the time?) so they never have to do > >> > > bounce-buffering. > >> > > > >> > > While the pv-ops one ends up quite frequently doing the > >> > > bounce-buffering, > >> > > which > >> > > implies that the DVB drivers end up allocating their buffers above the > > 4GB. > >> > > This means we end up spending some CPU time (in the guest) copying the > >> > > memory > >> > > from >4GB to 0-4GB region (And vice-versa). > >> > > >> > This reminds me of something (not sure what XenoLinux you use for > >> > comparison) - how are they allocating that memory? Not vmalloc_32() > >> > >> I was using the 2.6.18, then the one I saw on Google for Gentoo, and now > >> I am going to look at the 2.6.38 from OpenSuSE. > >> > >> > by chance (I remember having seen numerous uses under - iirc - > >> > drivers/media/)? > >> > > >> > Obviously, vmalloc_32() and any GFP_DMA32 allocations do *not* do > >> > what their (driver) callers might expect in a PV guest (including the > >> > contiguity assumption for the latter, recalling that you earlier said > >> > you were able to see the problem after several guest starts), and I > >> > had put into our kernels an adjustment to make vmalloc_32() actually > >> > behave as expected. > >> > >> Aaah.. The plot thickens! Let me look in the sources! Thanks for the > >> pointer. > > > > Jan hints lead me to the videobuf-dma-sg.c which does indeed to vmalloc_32 > > and then performs PCI DMA operations on the allocted vmalloc_32 > > area. > > > > So I cobbled up the attached patch (hadn't actually tested it and sadly > > won't until next week) which removes the call to vmalloc_32 and instead > > sets up DMA allocated set of pages. > > What a big patch (which would need re-doing for every vmalloc_32() > caller)! Fixing vmalloc_32() would be much less intrusive (reproducing > our 3.2 version of the affected function below, but clearly that's not > pv-ops ready). I just want to get to the bottom of this before attempting a proper fix. > > Jan > > static void *__vmalloc_area_node(struct vm_struct *area, gfp_t gfp_mask, > pgprot_t prot, int node, void *caller) > { > const int order = 0; > struct page **pages; > unsigned int nr_pages, array_size, i; > gfp_t nested_gfp = (gfp_mask & GFP_RECLAIM_MASK) | __GFP_ZERO; > #ifdef CONFIG_XEN > gfp_t dma_mask = gfp_mask & (__GFP_DMA | __GFP_DMA32); > > BUILD_BUG_ON((__GFP_DMA | __GFP_DMA32) != (__GFP_DMA + __GFP_DMA32)); > if (dma_mask == (__GFP_DMA | __GFP_DMA32)) > gfp_mask &= ~(__GFP_DMA | __GFP_DMA32); > #endif > > nr_pages = (area->size - PAGE_SIZE) >> PAGE_SHIFT; > array_size = (nr_pages * sizeof(struct page *)); > > area->nr_pages = nr_pages; > /* Please note that the recursion is strictly bounded. */ > if (array_size > PAGE_SIZE) { > pages = __vmalloc_node(array_size, 1, nested_gfp|__GFP_HIGHMEM, > PAGE_KERNEL, node, caller); > area->flags |= VM_VPAGES; > } else { > pages = kmalloc_node(array_size, nested_gfp, node); > } > area->pages = pages; > area->caller = caller; > if (!area->pages) { > remove_vm_area(area->addr); > kfree(area); > return NULL; > } > > for (i = 0; i < area->nr_pages; i++) { > struct page *page; > gfp_t tmp_mask = gfp_mask | __GFP_NOWARN; > > if (node < 0) > page = alloc_page(tmp_mask); > else > page = alloc_pages_node(node, tmp_mask, order); > > if (unlikely(!page)) { > /* Successfully allocated i pages, free them in > __vunmap() */ > area->nr_pages = i; > goto fail; > } > area->pages[i] = page; > #ifdef CONFIG_XEN > if (dma_mask) { > if (xen_limit_pages_to_max_mfn(page, 0, 32)) { > area->nr_pages = i + 1; > goto fail; > } > if (gfp_mask & __GFP_ZERO) > clear_highpage(page); > } > #endif > } > > if (map_vm_area(area, prot, &pages)) > goto fail; > return area->addr; > > fail: > warn_alloc_failed(gfp_mask, order, > "vmalloc: allocation failure, allocated %ld of %ld > bytes\n", > (area->nr_pages*PAGE_SIZE), area->size); > vfree(area->addr); > return NULL; > } > > ... > > #if defined(CONFIG_64BIT) && defined(CONFIG_ZONE_DMA32) > #define GFP_VMALLOC32 GFP_DMA32 | GFP_KERNEL > #elif defined(CONFIG_64BIT) && defined(CONFIG_ZONE_DMA) > #define GFP_VMALLOC32 GFP_DMA | GFP_KERNEL > #elif defined(CONFIG_XEN) > #define GFP_VMALLOC32 __GFP_DMA | __GFP_DMA32 | GFP_KERNEL > #else > #define GFP_VMALLOC32 GFP_KERNEL > #endif _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |