[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH 1/6] x86/vmx: Fix handing of MSR_DEBUGCTL on VMExit
Currently, whenever the guest writes a nonzero value to MSR_DEBUGCTL, Xen updates a host MSR load list entry with the current hardware value of MSR_DEBUGCTL. This is wrong. On VMExit, hardware automatically resets MSR_DEBUGCTL to 0. The only case where different behaviour is needed is if Xen is debugging itself, and this needs setting up unconditionally for the lifetime of the VM, rather than based on guest actions. This could be fixed by using a host MSR load list entry set up during construct_vmcs(). However, a more efficient option is to use an alternative block in the VMExit path, keyed on whether hypervisor debugging has been enabled. In order to set this up, drop the per cpu ler_msr variable (as there is no point having it per cpu when it will be the same everywhere), and use a single read_mostly variable instead. Split calc_ler_msr() out of percpu_traps_init() for clarity, and only perform the calculation on the first call. Finally, clean up do_debug(). Rearrange it to have a common tail, and call out the LBR behaviour specifically. Signed-off-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx> --- CC: Jan Beulich <JBeulich@xxxxxxxx> CC: Jun Nakajima <jun.nakajima@xxxxxxxxx> CC: Kevin Tian <kevin.tian@xxxxxxxxx> CC: Wei Liu <wei.liu2@xxxxxxxxxx> CC: Roger Pau Monné <roger.pau@xxxxxxxxxx> Initially, I tried to have a common xen_msr_debugctl variable, but rip-relative addresses don't resolve correctly in alternative blocks. LBR-only has been fine for ages, and I don't see that changing any time soon. --- xen/arch/x86/hvm/vmx/entry.S | 9 ++++++ xen/arch/x86/hvm/vmx/vmx.c | 3 +- xen/arch/x86/traps.c | 58 +++++++++++++++++++-------------------- xen/arch/x86/x86_64/traps.c | 7 +++-- xen/include/asm-x86/cpufeature.h | 1 + xen/include/asm-x86/cpufeatures.h | 1 + xen/include/asm-x86/msr.h | 2 +- 7 files changed, 45 insertions(+), 36 deletions(-) diff --git a/xen/arch/x86/hvm/vmx/entry.S b/xen/arch/x86/hvm/vmx/entry.S index aa2f103..afd552f 100644 --- a/xen/arch/x86/hvm/vmx/entry.S +++ b/xen/arch/x86/hvm/vmx/entry.S @@ -41,6 +41,15 @@ ENTRY(vmx_asm_vmexit_handler) SPEC_CTRL_ENTRY_FROM_HVM /* Req: b=curr %rsp=regs/cpuinfo, Clob: acd */ /* WARNING! `ret`, `call *`, `jmp *` not safe before this point. */ + /* Hardware clears MSR_DEBUGCTL on VMExit. Reinstate it if debugging Xen. */ + .macro restore_lbr + mov $IA32_DEBUGCTLMSR_LBR, %eax + mov $MSR_IA32_DEBUGCTLMSR, %ecx + xor %edx, %edx + wrmsr + .endm + ALTERNATIVE "", restore_lbr, X86_FEATURE_XEN_LBR + mov %rsp,%rdi call vmx_vmexit_handler diff --git a/xen/arch/x86/hvm/vmx/vmx.c b/xen/arch/x86/hvm/vmx/vmx.c index 9707514..33d39f6 100644 --- a/xen/arch/x86/hvm/vmx/vmx.c +++ b/xen/arch/x86/hvm/vmx/vmx.c @@ -3120,8 +3120,7 @@ static int vmx_msr_write_intercept(unsigned int msr, uint64_t msr_content) } } - if ( (rc < 0) || - (msr_content && (vmx_add_host_load_msr(msr) < 0)) ) + if ( rc < 0 ) hvm_inject_hw_exception(TRAP_machine_check, X86_EVENT_NO_EC); else __vmwrite(GUEST_IA32_DEBUGCTL, msr_content); diff --git a/xen/arch/x86/traps.c b/xen/arch/x86/traps.c index 8a99174..74784a2 100644 --- a/xen/arch/x86/traps.c +++ b/xen/arch/x86/traps.c @@ -96,8 +96,6 @@ string_param("nmi", opt_nmi); DEFINE_PER_CPU(uint64_t, efer); static DEFINE_PER_CPU(unsigned long, last_extable_addr); -DEFINE_PER_CPU_READ_MOSTLY(u32, ler_msr); - DEFINE_PER_CPU_READ_MOSTLY(struct desc_struct *, gdt_table); DEFINE_PER_CPU_READ_MOSTLY(struct desc_struct *, compat_gdt_table); @@ -117,6 +115,9 @@ integer_param("debug_stack_lines", debug_stack_lines); static bool opt_ler; boolean_param("ler", opt_ler); +/* LastExceptionFromIP on this hardware. Zero if LER is not in use. */ +uint32_t __read_mostly ler_msr; + #define stack_words_per_line 4 #define ESP_BEFORE_EXCEPTION(regs) ((unsigned long *)regs->rsp) @@ -1764,17 +1765,6 @@ void do_device_not_available(struct cpu_user_regs *regs) return; } -static void ler_enable(void) -{ - u64 debugctl; - - if ( !this_cpu(ler_msr) ) - return; - - rdmsrl(MSR_IA32_DEBUGCTLMSR, debugctl); - wrmsrl(MSR_IA32_DEBUGCTLMSR, debugctl | IA32_DEBUGCTLMSR_LBR); -} - void do_debug(struct cpu_user_regs *regs) { unsigned long dr6; @@ -1870,13 +1860,13 @@ void do_debug(struct cpu_user_regs *regs) v->arch.debugreg[6] |= (dr6 & ~X86_DR6_DEFAULT); v->arch.debugreg[6] &= (dr6 | ~X86_DR6_DEFAULT); - ler_enable(); pv_inject_hw_exception(TRAP_debug, X86_EVENT_NO_EC); - return; out: - ler_enable(); - return; + + /* #DB automatically disabled LBR. Reinstate it if debugging Xen. */ + if ( cpu_has_xen_lbr ) + wrmsrl(MSR_IA32_DEBUGCTLMSR, IA32_DEBUGCTLMSR_LBR); } static void __init noinline __set_intr_gate(unsigned int n, @@ -1920,38 +1910,46 @@ void load_TR(void) : "=m" (old_gdt) : "rm" (TSS_ENTRY << 3), "m" (tss_gdt) : "memory" ); } -void percpu_traps_init(void) +static uint32_t calc_ler_msr(void) { - subarch_percpu_traps_init(); - - if ( !opt_ler ) - return; - switch ( boot_cpu_data.x86_vendor ) { case X86_VENDOR_INTEL: switch ( boot_cpu_data.x86 ) { case 6: - this_cpu(ler_msr) = MSR_IA32_LASTINTFROMIP; - break; + return MSR_IA32_LASTINTFROMIP; + case 15: - this_cpu(ler_msr) = MSR_P4_LER_FROM_LIP; - break; + return MSR_P4_LER_FROM_LIP; } break; + case X86_VENDOR_AMD: switch ( boot_cpu_data.x86 ) { case 6: case 0xf ... 0x17: - this_cpu(ler_msr) = MSR_IA32_LASTINTFROMIP; - break; + return MSR_IA32_LASTINTFROMIP; } break; } - ler_enable(); + return 0; +} + +void percpu_traps_init(void) +{ + subarch_percpu_traps_init(); + + if ( !opt_ler ) + return; + + if ( !ler_msr && (ler_msr = calc_ler_msr()) ) + setup_force_cpu_cap(X86_FEATURE_XEN_LBR); + + if ( cpu_has_xen_lbr ) + wrmsrl(MSR_IA32_DEBUGCTLMSR, IA32_DEBUGCTLMSR_LBR); } void __init init_idt_traps(void) diff --git a/xen/arch/x86/x86_64/traps.c b/xen/arch/x86/x86_64/traps.c index f7f6928..b040185 100644 --- a/xen/arch/x86/x86_64/traps.c +++ b/xen/arch/x86/x86_64/traps.c @@ -144,11 +144,12 @@ void show_registers(const struct cpu_user_regs *regs) printk("CPU: %d\n", smp_processor_id()); _show_registers(&fault_regs, fault_crs, context, v); - if ( this_cpu(ler_msr) && !guest_mode(regs) ) + if ( ler_msr && !guest_mode(regs) ) { u64 from, to; - rdmsrl(this_cpu(ler_msr), from); - rdmsrl(this_cpu(ler_msr) + 1, to); + + rdmsrl(ler_msr, from); + rdmsrl(ler_msr + 1, to); printk("ler: %016lx -> %016lx\n", from, to); } } diff --git a/xen/include/asm-x86/cpufeature.h b/xen/include/asm-x86/cpufeature.h index 2cf8f7e..b237da1 100644 --- a/xen/include/asm-x86/cpufeature.h +++ b/xen/include/asm-x86/cpufeature.h @@ -113,6 +113,7 @@ #define cpu_has_aperfmperf boot_cpu_has(X86_FEATURE_APERFMPERF) #define cpu_has_lfence_dispatch boot_cpu_has(X86_FEATURE_LFENCE_DISPATCH) #define cpu_has_no_xpti boot_cpu_has(X86_FEATURE_NO_XPTI) +#define cpu_has_xen_lbr boot_cpu_has(X86_FEATURE_XEN_LBR) enum _cache_type { CACHE_TYPE_NULL = 0, diff --git a/xen/include/asm-x86/cpufeatures.h b/xen/include/asm-x86/cpufeatures.h index b90aa2d..8e5cc53 100644 --- a/xen/include/asm-x86/cpufeatures.h +++ b/xen/include/asm-x86/cpufeatures.h @@ -32,3 +32,4 @@ XEN_CPUFEATURE(SC_RSB_PV, (FSCAPINTS+0)*32+18) /* RSB overwrite needed for XEN_CPUFEATURE(SC_RSB_HVM, (FSCAPINTS+0)*32+19) /* RSB overwrite needed for HVM */ XEN_CPUFEATURE(NO_XPTI, (FSCAPINTS+0)*32+20) /* XPTI mitigation not in use */ XEN_CPUFEATURE(SC_MSR_IDLE, (FSCAPINTS+0)*32+21) /* (SC_MSR_PV || SC_MSR_HVM) && default_xen_spec_ctrl */ +XEN_CPUFEATURE(XEN_LBR, (FSCAPINTS+0)*32+22) /* Xen uses MSR_DEBUGCTL.LBR */ diff --git a/xen/include/asm-x86/msr.h b/xen/include/asm-x86/msr.h index f14f265..9f6d3b2 100644 --- a/xen/include/asm-x86/msr.h +++ b/xen/include/asm-x86/msr.h @@ -241,7 +241,7 @@ static inline void write_efer(uint64_t val) wrmsrl(MSR_EFER, val); } -DECLARE_PER_CPU(u32, ler_msr); +extern uint32_t ler_msr; DECLARE_PER_CPU(uint32_t, tsc_aux); -- 2.1.4 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxxx https://lists.xenproject.org/mailman/listinfo/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |