[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH RESEND RFC 5/8] mm: Do not discard already-scrubbed pages softirqs are pending
Signed-off-by: Boris Ostrovsky <boris.ostrovsky@xxxxxxxxxx> --- xen/common/page_alloc.c | 65 +++++++++++++++++++++++++++++++++++++++++----- 1 files changed, 58 insertions(+), 7 deletions(-) diff --git a/xen/common/page_alloc.c b/xen/common/page_alloc.c index ac15406..3469185 100644 --- a/xen/common/page_alloc.c +++ b/xen/common/page_alloc.c @@ -1032,11 +1032,14 @@ static void merge_chunks(struct page_info *pg, unsigned int node, page_list_add(pg, &heap(node, zone, order)); } +#define SCRUB_CHUNK_ORDER 8 bool_t scrub_free_pages(unsigned int node) { struct page_info *pg; unsigned int i, zone, cpu; int order; + unsigned int num_scrubbed, scrub_order, start, end; + bool_t preempt; static unsigned node_scrubbing; if ( !node_need_scrub[node] ) @@ -1046,6 +1049,7 @@ bool_t scrub_free_pages(unsigned int node) return 0; cpu = smp_processor_id(); + preempt = 0; spin_lock(&heap_lock); @@ -1060,16 +1064,63 @@ bool_t scrub_free_pages(unsigned int node) if ( !test_bit(_PGC_need_scrub, &pg[0].count_info) ) break; - for ( i = 0; i < (1 << order); i++) + page_list_del(pg, &heap(node, zone, order)); + + scrub_order = (order > SCRUB_CHUNK_ORDER) ? SCRUB_CHUNK_ORDER : order; + num_scrubbed = 0; + while ( num_scrubbed < (1 << order) ) { - scrub_one_page(&pg[i]); + for ( i = 0; i < (1 << scrub_order); i++ ) + scrub_one_page(&pg[num_scrubbed + i]); + + num_scrubbed += (1 << scrub_order); if ( softirq_pending(cpu) ) - goto out; + { + preempt = 1; + break; + } + } + + start = 0; + end = num_scrubbed; + + /* Merge clean pages */ + while ( start < end ) + { + /* + * Largest power-of-two chunk starting @start, + * not greater than @end + */ + unsigned chunk_order = flsl(end - start) - 1; + struct page_info *ppg = &pg[start]; + + for ( i = 0; i < (1 << chunk_order); i++ ) + ppg[i].count_info &= ~PGC_need_scrub; + + node_need_scrub[node] -= (1 << chunk_order); + + PFN_ORDER(ppg) = chunk_order; + merge_chunks(ppg, node, zone, chunk_order, 0); + start += (1 << chunk_order); + } + + /* Merge unscrubbed pages */ + while ( end < (1 << order) ) + { + /* + * Largest power-of-two chunk starting @end, not crossing + * next power-of-two boundary + */ + unsigned chunk_order = ffsl(end) - 1; + struct page_info *ppg = &pg[end]; + + PFN_ORDER(ppg) = chunk_order; + merge_chunks(ppg, node, zone, chunk_order, 1); + end += (1 << chunk_order); } - - node_need_scrub[node] -= (1 << order); - for ( i = 0; i < (1 << order); i++) - pg[i].count_info &= ~PGC_need_scrub; + + if ( preempt ) + goto out; } } } -- 1.7.1 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx https://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |