[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-changelog] [xen-unstable] [IA64] Cleanup in vmx
# HG changeset patch # User awilliam@xxxxxxxxxxxx # Date 1171839984 25200 # Node ID 54247a642f5e0eeeae1bcb12202fbbab946b58a7 # Parent 5abf33a383cf651acd17156effd807bc5f44ffd2 [IA64] Cleanup in vmx Cleanup: static added, unused code #if'ed, typos. Signed-off-by: Tristan Gingold <tgingold@xxxxxxx> --- xen/arch/ia64/vmx/vmx_interrupt.c | 15 ++++-- xen/arch/ia64/vmx/vmx_support.c | 9 --- xen/arch/ia64/vmx/vmx_vcpu.c | 24 ---------- xen/arch/ia64/vmx/vmx_virt.c | 89 +++++++++++++++++++------------------- 4 files changed, 58 insertions(+), 79 deletions(-) diff -r 5abf33a383cf -r 54247a642f5e xen/arch/ia64/vmx/vmx_interrupt.c --- a/xen/arch/ia64/vmx/vmx_interrupt.c Sun Feb 18 16:00:52 2007 -0700 +++ b/xen/arch/ia64/vmx/vmx_interrupt.c Sun Feb 18 16:06:24 2007 -0700 @@ -20,15 +20,15 @@ * Xiaoyan Feng (Fleming Feng) <fleming.feng@xxxxxxxxx> * Xuefei Xu (Anthony Xu) (Anthony.xu@xxxxxxxxx) */ - - #include <xen/types.h> #include <asm/vmx_vcpu.h> #include <asm/vmx_mm_def.h> #include <asm/vmx_pal_vsa.h> + /* SDM vol2 5.5 - IVA based interruption handling */ #define INITIAL_PSR_VALUE_AT_INTERRUPTION 0x0000001808028034 -void + +static void collect_interruption(VCPU *vcpu) { u64 ipsr; @@ -92,14 +92,19 @@ inject_guest_interruption(VCPU *vcpu, u6 u64 viva; REGS *regs; ISR pt_isr; + perfc_incra(vmx_inject_guest_interruption, vec >> 8); - regs=vcpu_regs(vcpu); - // clear cr.isr.ri + + regs = vcpu_regs(vcpu); + + // clear cr.isr.ir (incomplete register frame) pt_isr.val = VMX(vcpu,cr_isr); pt_isr.ir = 0; VMX(vcpu,cr_isr) = pt_isr.val; + collect_interruption(vcpu); vmx_ia64_set_dcr(vcpu); + vmx_vcpu_get_iva(vcpu,&viva); regs->cr_iip = viva + vec; } diff -r 5abf33a383cf -r 54247a642f5e xen/arch/ia64/vmx/vmx_support.c --- a/xen/arch/ia64/vmx/vmx_support.c Sun Feb 18 16:00:52 2007 -0700 +++ b/xen/arch/ia64/vmx/vmx_support.c Sun Feb 18 16:06:24 2007 -0700 @@ -44,7 +44,8 @@ void vmx_io_assist(struct vcpu *v) */ vio = get_vio(v->domain, v->vcpu_id); if (!vio) - panic_domain(vcpu_regs(v),"Corruption: bad shared page: %lx\n", (unsigned long)vio); + panic_domain(vcpu_regs(v),"Corruption: bad shared page: %lx\n", + (unsigned long)vio); p = &vio->vp_ioreq; @@ -98,9 +99,3 @@ void vmx_send_assist_req(struct vcpu *v) /* the code under this line is completer phase... */ vmx_io_assist(v); } - -/* Wake up a vcpu whihc is waiting for interrupts to come in */ -void vmx_prod_vcpu(struct vcpu *v) -{ - vcpu_unblock(v); -} diff -r 5abf33a383cf -r 54247a642f5e xen/arch/ia64/vmx/vmx_vcpu.c --- a/xen/arch/ia64/vmx/vmx_vcpu.c Sun Feb 18 16:00:52 2007 -0700 +++ b/xen/arch/ia64/vmx/vmx_vcpu.c Sun Feb 18 16:06:24 2007 -0700 @@ -22,7 +22,6 @@ * Yaozu Dong (Eddie Dong) (Eddie.dong@xxxxxxxxx) * Xuefei Xu (Anthony Xu) (Anthony.xu@xxxxxxxxx) */ - #include <xen/sched.h> #include <public/xen.h> #include <asm/ia64_int.h> @@ -36,29 +35,6 @@ #include <asm/vmx_mm_def.h> #include <asm/vmx.h> #include <asm/vmx_phy_mode.h> -//u64 fire_itc; -//u64 fire_itc2; -//u64 fire_itm; -//u64 fire_itm2; -/* - * Copyright (c) 2005 Intel Corporation. - * Anthony Xu (anthony.xu@xxxxxxxxx) - * Yaozu Dong (Eddie Dong) (Eddie.dong@xxxxxxxxx) - * - * This program is free software; you can redistribute it and/or modify it - * under the terms and conditions of the GNU General Public License, - * version 2, as published by the Free Software Foundation. - * - * This program is distributed in the hope it will be useful, but WITHOUT - * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or - * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for - * more details. - * - * You should have received a copy of the GNU General Public License along with - * this program; if not, write to the Free Software Foundation, Inc., 59 Temple - * Place - Suite 330, Boston, MA 02111-1307 USA. - * - */ /************************************************************************** VCPU general register access routines diff -r 5abf33a383cf -r 54247a642f5e xen/arch/ia64/vmx/vmx_virt.c --- a/xen/arch/ia64/vmx/vmx_virt.c Sun Feb 18 16:00:52 2007 -0700 +++ b/xen/arch/ia64/vmx/vmx_virt.c Sun Feb 18 16:06:24 2007 -0700 @@ -31,7 +31,8 @@ #include <asm/virt_event.h> #include <asm/vmx_phy_mode.h> -void +#ifdef BYPASS_VMAL_OPCODE +static void ia64_priv_decoder(IA64_SLOT_TYPE slot_type, INST64 inst, u64 * cause) { *cause=0; @@ -141,20 +142,21 @@ ia64_priv_decoder(IA64_SLOT_TYPE slot_ty break; } } - -IA64FAULT vmx_emul_rsm(VCPU *vcpu, INST64 inst) +#endif + +static IA64FAULT vmx_emul_rsm(VCPU *vcpu, INST64 inst) { u64 imm24 = (inst.M44.i << 23) | (inst.M44.i2 << 21) | inst.M44.imm; return vmx_vcpu_reset_psr_sm(vcpu,imm24); } -IA64FAULT vmx_emul_ssm(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_ssm(VCPU *vcpu, INST64 inst) { u64 imm24 = (inst.M44.i << 23) | (inst.M44.i2 << 21) | inst.M44.imm; return vmx_vcpu_set_psr_sm(vcpu,imm24); } -IA64FAULT vmx_emul_mov_from_psr(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_mov_from_psr(VCPU *vcpu, INST64 inst) { u64 tgt = inst.M33.r1; u64 val; @@ -172,7 +174,7 @@ IA64FAULT vmx_emul_mov_from_psr(VCPU *vc /** * @todo Check for reserved bits and return IA64_RSVDREG_FAULT. */ -IA64FAULT vmx_emul_mov_to_psr(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_mov_to_psr(VCPU *vcpu, INST64 inst) { u64 val; @@ -187,7 +189,7 @@ Privileged operation emulation routines Privileged operation emulation routines **************************************************************************/ -IA64FAULT vmx_emul_rfi(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_rfi(VCPU *vcpu, INST64 inst) { IA64_PSR vpsr; REGS *regs; @@ -209,7 +211,7 @@ IA64FAULT vmx_emul_rfi(VCPU *vcpu, INST6 return vmx_vcpu_rfi(vcpu); } -IA64FAULT vmx_emul_bsw0(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_bsw0(VCPU *vcpu, INST64 inst) { #ifdef CHECK_FAULT IA64_PSR vpsr; @@ -224,7 +226,7 @@ IA64FAULT vmx_emul_bsw0(VCPU *vcpu, INST return vcpu_bsw0(vcpu); } -IA64FAULT vmx_emul_bsw1(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_bsw1(VCPU *vcpu, INST64 inst) { #ifdef CHECK_FAULT IA64_PSR vpsr; @@ -239,12 +241,12 @@ IA64FAULT vmx_emul_bsw1(VCPU *vcpu, INST return vcpu_bsw1(vcpu); } -IA64FAULT vmx_emul_cover(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_cover(VCPU *vcpu, INST64 inst) { return vmx_vcpu_cover(vcpu); } -IA64FAULT vmx_emul_ptc_l(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_ptc_l(VCPU *vcpu, INST64 inst) { u64 r2,r3; #ifdef VMAL_NO_FAULT_CHECK @@ -278,7 +280,7 @@ IA64FAULT vmx_emul_ptc_l(VCPU *vcpu, INS return vmx_vcpu_ptc_l(vcpu,r3,bits(r2,2,7)); } -IA64FAULT vmx_emul_ptc_e(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_ptc_e(VCPU *vcpu, INST64 inst) { u64 r3; #ifdef VMAL_NO_FAULT_CHECK @@ -303,7 +305,7 @@ IA64FAULT vmx_emul_ptc_e(VCPU *vcpu, INS return vmx_vcpu_ptc_e(vcpu,r3); } -IA64FAULT vmx_emul_ptc_g(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_ptc_g(VCPU *vcpu, INST64 inst) { u64 r2,r3; #ifdef VMAL_NO_FAULT_CHECK @@ -336,7 +338,7 @@ IA64FAULT vmx_emul_ptc_g(VCPU *vcpu, INS return vmx_vcpu_ptc_g(vcpu,r3,bits(r2,2,7)); } -IA64FAULT vmx_emul_ptc_ga(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_ptc_ga(VCPU *vcpu, INST64 inst) { u64 r2,r3; #ifdef VMAL_NO_FAULT_CHECK @@ -369,7 +371,7 @@ IA64FAULT vmx_emul_ptc_ga(VCPU *vcpu, IN return vmx_vcpu_ptc_ga(vcpu,r3,bits(r2,2,7)); } -IA64FAULT ptr_fault_check(VCPU *vcpu, INST64 inst, u64 *pr2, u64 *pr3) +static IA64FAULT ptr_fault_check(VCPU *vcpu, INST64 inst, u64 *pr2, u64 *pr3) { IA64FAULT ret1, ret2; @@ -403,7 +405,7 @@ IA64FAULT ptr_fault_check(VCPU *vcpu, IN return IA64_NO_FAULT; } -IA64FAULT vmx_emul_ptr_d(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_ptr_d(VCPU *vcpu, INST64 inst) { u64 r2,r3; if ( ptr_fault_check(vcpu, inst, &r2, &r3 ) == IA64_FAULT ) @@ -411,7 +413,7 @@ IA64FAULT vmx_emul_ptr_d(VCPU *vcpu, INS return vmx_vcpu_ptr_d(vcpu,r3,bits(r2,2,7)); } -IA64FAULT vmx_emul_ptr_i(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_ptr_i(VCPU *vcpu, INST64 inst) { u64 r2,r3; if ( ptr_fault_check(vcpu, inst, &r2, &r3 ) == IA64_FAULT ) @@ -420,7 +422,7 @@ IA64FAULT vmx_emul_ptr_i(VCPU *vcpu, INS } -IA64FAULT vmx_emul_thash(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_thash(VCPU *vcpu, INST64 inst) { u64 r1,r3; #ifdef CHECK_FAULT @@ -450,7 +452,7 @@ IA64FAULT vmx_emul_thash(VCPU *vcpu, INS } -IA64FAULT vmx_emul_ttag(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_ttag(VCPU *vcpu, INST64 inst) { u64 r1,r3; #ifdef CHECK_FAULT @@ -482,7 +484,7 @@ IA64FAULT vmx_emul_ttag(VCPU *vcpu, INST } -IA64FAULT vmx_emul_tpa(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_tpa(VCPU *vcpu, INST64 inst) { u64 r1,r3; #ifdef CHECK_FAULT @@ -526,7 +528,7 @@ IA64FAULT vmx_emul_tpa(VCPU *vcpu, INST6 return(IA64_NO_FAULT); } -IA64FAULT vmx_emul_tak(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_tak(VCPU *vcpu, INST64 inst) { u64 r1,r3; #ifdef CHECK_FAULT @@ -564,7 +566,7 @@ IA64FAULT vmx_emul_tak(VCPU *vcpu, INST6 * Insert translation register/cache ************************************/ -IA64FAULT vmx_emul_itr_d(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_itr_d(VCPU *vcpu, INST64 inst) { u64 itir, ifa, pte, slot; #ifdef VMAL_NO_FAULT_CHECK @@ -621,7 +623,7 @@ IA64FAULT vmx_emul_itr_d(VCPU *vcpu, INS return (vmx_vcpu_itr_d(vcpu,slot,pte,itir,ifa)); } -IA64FAULT vmx_emul_itr_i(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_itr_i(VCPU *vcpu, INST64 inst) { u64 itir, ifa, pte, slot; #ifdef VMAL_NO_FAULT_CHECK @@ -678,7 +680,8 @@ IA64FAULT vmx_emul_itr_i(VCPU *vcpu, INS return (vmx_vcpu_itr_i(vcpu,slot,pte,itir,ifa)); } -IA64FAULT itc_fault_check(VCPU *vcpu, INST64 inst, u64 *itir, u64 *ifa,u64 *pte) +static IA64FAULT itc_fault_check(VCPU *vcpu, INST64 inst, + u64 *itir, u64 *ifa, u64 *pte) { IA64FAULT ret1; @@ -727,7 +730,7 @@ IA64FAULT itc_fault_check(VCPU *vcpu, IN return IA64_NO_FAULT; } -IA64FAULT vmx_emul_itc_d(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_itc_d(VCPU *vcpu, INST64 inst) { u64 itir, ifa, pte; @@ -738,7 +741,7 @@ IA64FAULT vmx_emul_itc_d(VCPU *vcpu, INS return (vmx_vcpu_itc_d(vcpu,pte,itir,ifa)); } -IA64FAULT vmx_emul_itc_i(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_itc_i(VCPU *vcpu, INST64 inst) { u64 itir, ifa, pte; @@ -754,7 +757,7 @@ IA64FAULT vmx_emul_itc_i(VCPU *vcpu, INS * Moves to semi-privileged registers *************************************/ -IA64FAULT vmx_emul_mov_to_ar_imm(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_mov_to_ar_imm(VCPU *vcpu, INST64 inst) { // I27 and M30 are identical for these fields u64 imm; @@ -780,7 +783,7 @@ IA64FAULT vmx_emul_mov_to_ar_imm(VCPU *v return (vmx_vcpu_set_itc(vcpu, imm)); } -IA64FAULT vmx_emul_mov_to_ar_reg(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_mov_to_ar_reg(VCPU *vcpu, INST64 inst) { // I26 and M29 are identical for these fields u64 r2; @@ -808,7 +811,7 @@ IA64FAULT vmx_emul_mov_to_ar_reg(VCPU *v } -IA64FAULT vmx_emul_mov_from_ar_reg(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_mov_from_ar_reg(VCPU *vcpu, INST64 inst) { // I27 and M30 are identical for these fields u64 r1; @@ -840,7 +843,7 @@ IA64FAULT vmx_emul_mov_from_ar_reg(VCPU * Moves to privileged registers ********************************/ -IA64FAULT vmx_emul_mov_to_pkr(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_mov_to_pkr(VCPU *vcpu, INST64 inst) { u64 r3,r2; #ifdef CHECK_FAULT @@ -863,7 +866,7 @@ IA64FAULT vmx_emul_mov_to_pkr(VCPU *vcpu return (vmx_vcpu_set_pkr(vcpu,r3,r2)); } -IA64FAULT vmx_emul_mov_to_rr(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_mov_to_rr(VCPU *vcpu, INST64 inst) { u64 r3,r2; #ifdef CHECK_FAULT @@ -886,7 +889,7 @@ IA64FAULT vmx_emul_mov_to_rr(VCPU *vcpu, return (vmx_vcpu_set_rr(vcpu,r3,r2)); } -IA64FAULT vmx_emul_mov_to_dbr(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_mov_to_dbr(VCPU *vcpu, INST64 inst) { u64 r3,r2; return IA64_NO_FAULT; @@ -910,7 +913,7 @@ IA64FAULT vmx_emul_mov_to_dbr(VCPU *vcpu return (vmx_vcpu_set_dbr(vcpu,r3,r2)); } -IA64FAULT vmx_emul_mov_to_ibr(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_mov_to_ibr(VCPU *vcpu, INST64 inst) { u64 r3,r2; return IA64_NO_FAULT; @@ -934,7 +937,7 @@ IA64FAULT vmx_emul_mov_to_ibr(VCPU *vcpu return (vmx_vcpu_set_ibr(vcpu,r3,r2)); } -IA64FAULT vmx_emul_mov_to_pmc(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_mov_to_pmc(VCPU *vcpu, INST64 inst) { u64 r3,r2; #ifdef CHECK_FAULT @@ -957,7 +960,7 @@ IA64FAULT vmx_emul_mov_to_pmc(VCPU *vcpu return (vmx_vcpu_set_pmc(vcpu,r3,r2)); } -IA64FAULT vmx_emul_mov_to_pmd(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_mov_to_pmd(VCPU *vcpu, INST64 inst) { u64 r3,r2; #ifdef CHECK_FAULT @@ -985,7 +988,7 @@ IA64FAULT vmx_emul_mov_to_pmd(VCPU *vcpu * Moves from privileged registers **********************************/ -IA64FAULT vmx_emul_mov_from_rr(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_mov_from_rr(VCPU *vcpu, INST64 inst) { u64 r3,r1; #ifdef CHECK_FAULT @@ -1021,7 +1024,7 @@ IA64FAULT vmx_emul_mov_from_rr(VCPU *vcp return vcpu_set_gr(vcpu, inst.M43.r1, r1,0); } -IA64FAULT vmx_emul_mov_from_pkr(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_mov_from_pkr(VCPU *vcpu, INST64 inst) { u64 r3,r1; #ifdef CHECK_FAULT @@ -1058,7 +1061,7 @@ IA64FAULT vmx_emul_mov_from_pkr(VCPU *vc return vcpu_set_gr(vcpu, inst.M43.r1, r1,0); } -IA64FAULT vmx_emul_mov_from_dbr(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_mov_from_dbr(VCPU *vcpu, INST64 inst) { u64 r3,r1; #ifdef CHECK_FAULT @@ -1095,7 +1098,7 @@ IA64FAULT vmx_emul_mov_from_dbr(VCPU *vc return vcpu_set_gr(vcpu, inst.M43.r1, r1,0); } -IA64FAULT vmx_emul_mov_from_ibr(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_mov_from_ibr(VCPU *vcpu, INST64 inst) { u64 r3,r1; #ifdef CHECK_FAULT @@ -1132,7 +1135,7 @@ IA64FAULT vmx_emul_mov_from_ibr(VCPU *vc return vcpu_set_gr(vcpu, inst.M43.r1, r1,0); } -IA64FAULT vmx_emul_mov_from_pmc(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_mov_from_pmc(VCPU *vcpu, INST64 inst) { u64 r3,r1; #ifdef CHECK_FAULT @@ -1169,7 +1172,7 @@ IA64FAULT vmx_emul_mov_from_pmc(VCPU *vc return vcpu_set_gr(vcpu, inst.M43.r1, r1,0); } -IA64FAULT vmx_emul_mov_from_cpuid(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_mov_from_cpuid(VCPU *vcpu, INST64 inst) { u64 r3,r1; #ifdef CHECK_FAULT @@ -1197,7 +1200,7 @@ IA64FAULT vmx_emul_mov_from_cpuid(VCPU * return vcpu_set_gr(vcpu, inst.M43.r1, r1,0); } -IA64FAULT vmx_emul_mov_to_cr(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_mov_to_cr(VCPU *vcpu, INST64 inst) { u64 r2; extern u64 cr_igfld_mask(int index, u64 value); @@ -1275,7 +1278,7 @@ IA64FAULT vmx_emul_mov_to_cr(VCPU *vcpu, ((fault=vmx_vcpu_get_##cr(vcpu,&val))==IA64_NO_FAULT)?\ vcpu_set_gr(vcpu, tgt, val,0):fault; -IA64FAULT vmx_emul_mov_from_cr(VCPU *vcpu, INST64 inst) +static IA64FAULT vmx_emul_mov_from_cr(VCPU *vcpu, INST64 inst) { u64 tgt = inst.M33.r1; u64 val; _______________________________________________ Xen-changelog mailing list Xen-changelog@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-changelog
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |