[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-changelog] [xen-unstable] [IA64] fix coding style in vmx_init.c
# HG changeset patch # User awilliam@xxxxxxxxxxx # Node ID 8ee7bb2af497d2e2ee1ff26ddbee4aede61c8e68 # Parent 1eb880e9ff944de77d534517492d0a583fc9587e [IA64] fix coding style in vmx_init.c Signed-off-by: Akio Takebe <takebe_akio@xxxxxxxxxxxxxx> --- xen/arch/ia64/vmx/vmx_init.c | 73 +++++++++++++++++++++---------------------- 1 files changed, 36 insertions(+), 37 deletions(-) diff -r 1eb880e9ff94 -r 8ee7bb2af497 xen/arch/ia64/vmx/vmx_init.c --- a/xen/arch/ia64/vmx/vmx_init.c Tue Oct 31 22:25:25 2006 -0700 +++ b/xen/arch/ia64/vmx/vmx_init.c Tue Oct 31 22:25:31 2006 -0700 @@ -183,7 +183,6 @@ static vpd_t *alloc_vpd(void) mregs->vac.a_cover = 1; mregs->vac.a_bsw = 1; mregs->vac.a_int = 1; - mregs->vdc.d_vmsw = 1; return vpd; @@ -276,7 +275,7 @@ static void vmx_create_event_channels(st for_each_vcpu(v->domain, o) { p = get_vio(v->domain, o->vcpu_id); o->arch.arch_vmx.xen_port = p->vp_eport = - alloc_unbound_xen_event_channel(o, 0); + alloc_unbound_xen_event_channel(o, 0); DPRINTK("Allocated port %d for hvm.\n", o->arch.arch_vmx.xen_port); } @@ -306,8 +305,8 @@ vmx_final_setup_guest(struct vcpu *v) /* Per-domain vTLB and vhpt implementation. Now vmx domain will stick * to this solution. Maybe it can be deferred until we know created * one as vmx domain */ -#ifndef HASH_VHPT - init_domain_tlb(v); +#ifndef HASH_VHPT + init_domain_tlb(v); #endif vmx_create_event_channels(v); @@ -378,44 +377,44 @@ static void vmx_build_physmap_table(stru /* Mark I/O ranges */ for (i = 0; i < (sizeof(io_ranges) / sizeof(io_range_t)); i++) { - for (j = io_ranges[i].start; - j < io_ranges[i].start + io_ranges[i].size; - j += PAGE_SIZE) - (void)__assign_domain_page(d, j, io_ranges[i].type, - ASSIGN_writable); + for (j = io_ranges[i].start; + j < io_ranges[i].start + io_ranges[i].size; j += PAGE_SIZE) + (void)__assign_domain_page(d, j, io_ranges[i].type, + ASSIGN_writable); } /* Map normal memory below 3G */ end = VMX_CONFIG_PAGES(d) << PAGE_SHIFT; tmp = end < MMIO_START ? end : MMIO_START; for (i = 0; (i < tmp) && (list_ent != &d->page_list); i += PAGE_SIZE) { - mfn = page_to_mfn(list_entry(list_ent, struct page_info, list)); - list_ent = mfn_to_page(mfn)->list.next; - if (VGA_IO_START <= i && i < VGA_IO_START + VGA_IO_SIZE) - continue; - assign_domain_page(d, i, mfn << PAGE_SHIFT); + mfn = page_to_mfn(list_entry(list_ent, struct page_info, list)); + list_ent = mfn_to_page(mfn)->list.next; + if (VGA_IO_START <= i && i < VGA_IO_START + VGA_IO_SIZE) + continue; + assign_domain_page(d, i, mfn << PAGE_SHIFT); } ASSERT(list_ent != &d->page_list); /* Map normal memory beyond 4G */ if (unlikely(end > MMIO_START)) { - start = 4 * MEM_G; - end = start + (end - 3 * MEM_G); - for (i = start; - (i < end) && (list_ent != &d->page_list); i += PAGE_SIZE) { + start = 4 * MEM_G; + end = start + (end - 3 * MEM_G); + for (i = start; + (i < end) && (list_ent != &d->page_list); i += PAGE_SIZE) { + mfn = page_to_mfn(list_entry(list_ent, + struct page_info, list)); + assign_domain_page(d, i, mfn << PAGE_SHIFT); + list_ent = mfn_to_page(mfn)->list.next; + } + ASSERT(list_ent != &d->page_list); + } + + /* Map guest firmware */ + for (i = GFW_START; (i < GFW_START + GFW_SIZE) && + (list_ent != &d->page_list); i += PAGE_SIZE) { mfn = page_to_mfn(list_entry(list_ent, struct page_info, list)); assign_domain_page(d, i, mfn << PAGE_SHIFT); list_ent = mfn_to_page(mfn)->list.next; - } - ASSERT(list_ent != &d->page_list); - } - - /* Map guest firmware */ - for (i = GFW_START; (i < GFW_START + GFW_SIZE) && - (list_ent != &d->page_list); i += PAGE_SIZE) { - mfn = page_to_mfn(list_entry(list_ent, struct page_info, list)); - assign_domain_page(d, i, mfn << PAGE_SHIFT); - list_ent = mfn_to_page(mfn)->list.next; } ASSERT(list_ent != &d->page_list); @@ -429,10 +428,10 @@ static void vmx_build_physmap_table(stru assign_domain_page(d, STORE_PAGE_START, mfn << PAGE_SHIFT); list_ent = mfn_to_page(mfn)->list.next; ASSERT(list_ent != &d->page_list); - - mfn = page_to_mfn(list_entry(list_ent, struct page_info, list)); - assign_domain_page(d, BUFFER_IO_PAGE_START, mfn << PAGE_SHIFT); - list_ent = mfn_to_page(mfn)->list.next; + + mfn = page_to_mfn(list_entry(list_ent, struct page_info, list)); + assign_domain_page(d, BUFFER_IO_PAGE_START, mfn << PAGE_SHIFT); + list_ent = mfn_to_page(mfn)->list.next; ASSERT(list_ent == &d->page_list); } @@ -444,10 +443,10 @@ void vmx_setup_platform(struct domain *d d->arch.vmx_platform.shared_page_va = (unsigned long)__va(__gpa_to_mpa(d, IO_PAGE_START)); - //For buffered IO requests. - spin_lock_init(&d->arch.hvm_domain.buffered_io_lock); - d->arch.hvm_domain.buffered_io_va = - (unsigned long)__va(__gpa_to_mpa(d, BUFFER_IO_PAGE_START)); + /* For buffered IO requests. */ + spin_lock_init(&d->arch.hvm_domain.buffered_io_lock); + d->arch.hvm_domain.buffered_io_va = + (unsigned long)__va(__gpa_to_mpa(d, BUFFER_IO_PAGE_START)); /* TEMP */ d->arch.vmx_platform.pib_base = 0xfee00000UL; @@ -455,7 +454,7 @@ void vmx_setup_platform(struct domain *d /* Only open one port for I/O and interrupt emulation */ memset(&d->shared_info->evtchn_mask[0], 0xff, - sizeof(d->shared_info->evtchn_mask)); + sizeof(d->shared_info->evtchn_mask)); /* initiate spinlock for pass virq */ spin_lock_init(&d->arch.arch_vmx.virq_assist_lock); _______________________________________________ Xen-changelog mailing list Xen-changelog@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-changelog
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |