[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH V2 11/13] xen: introduce xen_event_channel_register_3level
Signed-off-by: Wei Liu <wei.liu2@xxxxxxxxxx> --- drivers/xen/events.c | 94 ++++++++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 94 insertions(+) diff --git a/drivers/xen/events.c b/drivers/xen/events.c index f3f0fdc..6e0283c 100644 --- a/drivers/xen/events.c +++ b/drivers/xen/events.c @@ -53,6 +53,9 @@ /* Helper macro(s) */ #define LONG_BITORDER (BITS_PER_LONG == 64 ? 6 : 5) +/* event bitmap size: 1 page for 32 bit and 8 pages for 64 bit */ +#define BITMAP_PG_ORDER (BITS_PER_LONG == 64 ? 3 : 1) +#define BITMAP_NR_PAGES (BITMAP_PG_ORDER == 3 ? 8 : 1) /* N-level event channel, starting from 2 */ unsigned int evtchn_level = 2; @@ -2123,6 +2126,97 @@ void xen_callback_vector(void) void xen_callback_vector(void) {} #endif +static int xen_event_channel_register_3level(void) +{ + evtchn_register_nlevel_t reg; + int i, cpu; + unsigned long *_evtchn_pending = NULL; + unsigned long *_evtchn_mask = NULL; + unsigned long *l2sel_mfns = NULL; + unsigned long *l2sel_offsets = NULL; + int rc; + + /* If we come from restore path, we don't need to allocate + * pages. + */ + if (!evtchn_pending && !evtchn_mask) { + evtchn_pending = + (unsigned long *)__get_free_pages(GFP_KERNEL, + BITMAP_PG_ORDER); + evtchn_mask = + (unsigned long *)__get_free_pages(GFP_KERNEL, + BITMAP_PG_ORDER); + if (!evtchn_pending || !evtchn_mask) { + free_pages((unsigned long)evtchn_pending, BITMAP_PG_ORDER); + free_pages((unsigned long)evtchn_mask, BITMAP_PG_ORDER); + evtchn_pending = NULL; + evtchn_mask = NULL; + rc = -ENOMEM; + goto err; + } + } + + rc = -ENOMEM; /* Common error code for following operations */ +#define __ALLOC_ARRAY(_ptr, _nr) \ + do { \ + (_ptr) = kzalloc(sizeof(unsigned long) * (_nr), \ + GFP_KERNEL); \ + if (!(_ptr)) \ + goto out; \ + } while (0) + + __ALLOC_ARRAY(_evtchn_pending, BITMAP_NR_PAGES); + __ALLOC_ARRAY(_evtchn_mask, BITMAP_NR_PAGES); + __ALLOC_ARRAY(l2sel_mfns, nr_cpu_ids); + __ALLOC_ARRAY(l2sel_offsets, nr_cpu_ids); +#undef __ALLOC_ARRAY + + memset(®, 0, sizeof(reg)); + + for (i = 0; i < BITMAP_NR_PAGES; i++) { + unsigned long offset = PAGE_SIZE * i; + _evtchn_pending[i] = + arbitrary_virt_to_mfn( + (void *)((unsigned long)evtchn_pending+offset)); + _evtchn_mask[i] = + arbitrary_virt_to_mfn( + (void *)((unsigned long)evtchn_mask+offset)); + } + + for_each_possible_cpu(cpu) { + l2sel_mfns[cpu] = + arbitrary_virt_to_mfn(&per_cpu(evtchn_sel_l2, cpu)); + l2sel_offsets[cpu] = + offset_in_page(&per_cpu(evtchn_sel_l2, cpu)); + } + + reg.u.l3.nr_pages = BITMAP_NR_PAGES; + reg.u.l3.evtchn_pending = _evtchn_pending; + reg.u.l3.evtchn_mask = _evtchn_mask; + + reg.u.l3.nr_vcpus = nr_cpu_ids; + reg.u.l3.l2sel_mfns = l2sel_mfns; + reg.u.l3.l2sel_offsets = l2sel_offsets; + + reg.level = 3; + + rc = HYPERVISOR_event_channel_op(EVTCHNOP_register_nlevel, ®); + if (rc) { + free_pages((unsigned long)evtchn_pending, BITMAP_PG_ORDER); + free_pages((unsigned long)evtchn_mask, BITMAP_PG_ORDER); + evtchn_pending = NULL; + evtchn_mask = NULL; + } + +out: + kfree(_evtchn_pending); + kfree(_evtchn_mask); + kfree(l2sel_mfns); + kfree(l2sel_offsets); +err: + return rc; +} + void __init xen_init_IRQ(void) { int i, rc; -- 1.7.10.4 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |