[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH RESEND v1 5/7] x86: Implement Intel Processor Trace context switch
On Tue, Jan 16, 2018 at 02:12:31AM +0800, Luwei Kang wrote: > Load/Store Intel processor trace register in context switch. > MSR IA32_RTIT_CTL is loaded/stored automatically from VMCS. > When Intel PT is supported in guest, we need load/restore > PT MSRs only when PT is enabled in guest. > > Signed-off-by: Luwei Kang <luwei.kang@xxxxxxxxx> Is there a reason to not use xsaves/xrstors when they are available? > --- > xen/arch/x86/cpu/intel_pt.c | 69 > ++++++++++++++++++++++++++++++++++++++ > xen/arch/x86/hvm/vmx/vmx.c | 4 +++ > xen/include/asm-x86/hvm/vmx/vmcs.h | 2 ++ > xen/include/asm-x86/intel_pt.h | 4 +++ > 4 files changed, 79 insertions(+) > > diff --git a/xen/arch/x86/cpu/intel_pt.c b/xen/arch/x86/cpu/intel_pt.c > index 520e0ca..c0e9e68 100644 > --- a/xen/arch/x86/cpu/intel_pt.c > +++ b/xen/arch/x86/cpu/intel_pt.c > @@ -21,7 +21,76 @@ > #include <xen/types.h> > #include <xen/cache.h> > #include <xen/init.h> > +#include <asm/hvm/vmx/vmx.h> > +#include <asm/intel_pt.h> > > /* intel_pt: Flag to enable Intel Processor Trace (default on). */ > bool_t __read_mostly opt_intel_pt = 1; > boolean_param("intel_pt", opt_intel_pt); > + > +static inline void pt_load_msr(struct pt_ctx *ctx, u32 addr_num) Use unsigned int for addr_num. > +{ > + u32 i; Use unsigned int and add a blank line. > + wrmsrl(MSR_IA32_RTIT_STATUS, ctx->status); > + wrmsrl(MSR_IA32_RTIT_OUTPUT_BASE, ctx->output_base); > + wrmsrl(MSR_IA32_RTIT_OUTPUT_MASK, ctx->output_mask); > + wrmsrl(MSR_IA32_RTIT_CR3_MATCH, ctx->cr3_match); > + for ( i = 0; i < addr_num; i++ ) > + wrmsrl(MSR_IA32_RTIT_ADDR0_A + i, ctx->addr[i]); > +} > + > +static inline void pt_save_msr(struct pt_ctx *ctx, u32 addr_num) > +{ > + u32 i; Ditto. > + rdmsrl(MSR_IA32_RTIT_STATUS, ctx->status); > + rdmsrl(MSR_IA32_RTIT_OUTPUT_BASE, ctx->output_base); > + rdmsrl(MSR_IA32_RTIT_OUTPUT_MASK, ctx->output_mask); > + rdmsrl(MSR_IA32_RTIT_CR3_MATCH, ctx->cr3_match); > + for ( i = 0; i < addr_num; i++ ) > + rdmsrl(MSR_IA32_RTIT_ADDR0_A + i, ctx->addr[i]); > +} > + > +void pt_guest_enter(struct vcpu *v) > +{ > + struct pt_desc *pt = &v->arch.hvm_vmx.pt_desc; > + > + if ( pt->intel_pt_enabled && > + (pt->guest_pt_ctx.ctl & MSR_IA32_RTIT_CTL_TRACEEN) ) > + pt_load_msr(&pt->guest_pt_ctx, pt->addr_num); > +} > + > +void pt_guest_exit(struct vcpu *v) > +{ > + struct pt_desc *pt = &v->arch.hvm_vmx.pt_desc; > + > + if ( pt->intel_pt_enabled && > + (pt->guest_pt_ctx.ctl & MSR_IA32_RTIT_CTL_TRACEEN) ) > + pt_save_msr(&pt->guest_pt_ctx, pt->addr_num); > +} > + > +void pt_vcpu_init(struct vcpu *v) > +{ > + struct pt_desc *pt = &v->arch.hvm_vmx.pt_desc; > + unsigned int eax, ebx, ecx, edx; > + > + memset(pt, 0, sizeof(struct pt_desc)); > + pt->intel_pt_enabled = false; > + > + if ( !cpu_has_intel_pt || !opt_intel_pt || > + !(v->arch.hvm_vmx.secondary_exec_control & > SECONDARY_EXEC_PT_USE_GPA) ) > + return; > + > + /* get the number of address ranges */ > + if ( cpuid_eax(0x14) == 1 ) > + cpuid_count(0x14, 1, &eax, &ebx, &ecx, &edx); > + else > + return; > + if ( cpuid_eax(0x14) != 1 ) return; cpuid_count(....); This reduces the indentation needed. > + pt->addr_num = eax & 0x7; > + pt->guest_pt_ctx.output_mask = 0x7F; > + pt->intel_pt_enabled = true; > + > + vmx_vmcs_enter(v); > + __vmwrite(GUEST_IA32_RTIT_CTL, 0); > + vmx_vmcs_exit(v); > +} > diff --git a/xen/arch/x86/hvm/vmx/vmx.c b/xen/arch/x86/hvm/vmx/vmx.c > index e036303..f386933 100644 > --- a/xen/arch/x86/hvm/vmx/vmx.c > +++ b/xen/arch/x86/hvm/vmx/vmx.c > @@ -467,6 +467,8 @@ static int vmx_vcpu_initialise(struct vcpu *v) > if ( v->vcpu_id == 0 ) > v->arch.user_regs.rax = 1; > > + pt_vcpu_init(v); > + > return 0; > } > > @@ -3513,6 +3515,7 @@ void vmx_vmexit_handler(struct cpu_user_regs *regs) > __vmread(GUEST_RSP, ®s->rsp); > __vmread(GUEST_RFLAGS, ®s->rflags); > > + pt_guest_exit(v); > hvm_invalidate_regs_fields(regs); > > if ( paging_mode_hap(v->domain) ) > @@ -4281,6 +4284,7 @@ bool vmx_vmenter_helper(const struct cpu_user_regs > *regs) > } > } > > + pt_guest_enter(curr); > out: > if ( unlikely(curr->arch.hvm_vmx.lbr_fixup_enabled) ) > lbr_fixup(); > diff --git a/xen/include/asm-x86/hvm/vmx/vmcs.h > b/xen/include/asm-x86/hvm/vmx/vmcs.h > index 33ec3e6..46c386f 100644 > --- a/xen/include/asm-x86/hvm/vmx/vmcs.h > +++ b/xen/include/asm-x86/hvm/vmx/vmcs.h > @@ -421,6 +421,8 @@ enum vmcs_field { > GUEST_PDPTE0 = 0x0000280a, > #define GUEST_PDPTE(n) (GUEST_PDPTE0 + (n) * 2) /* n = 0...3 */ > GUEST_BNDCFGS = 0x00002812, > + GUEST_IA32_RTIT_CTL = 0x00002814, > + GUEST_IA32_RTIT_CTL_HIGH = 0x00002815, > HOST_PAT = 0x00002c00, > HOST_EFER = 0x00002c02, > HOST_PERF_GLOBAL_CTRL = 0x00002c04, > diff --git a/xen/include/asm-x86/intel_pt.h b/xen/include/asm-x86/intel_pt.h > index 909e22f..9505c8f 100644 > --- a/xen/include/asm-x86/intel_pt.h > +++ b/xen/include/asm-x86/intel_pt.h > @@ -40,4 +40,8 @@ struct pt_desc { > > extern bool_t opt_intel_pt; > > +void pt_vcpu_init(struct vcpu *v); > +void pt_guest_enter(struct vcpu *v); > +void pt_guest_exit(struct vcpu *v); > + > #endif /* __ASM_X86_HVM_INTEL_PT_H_ */ > -- > 1.8.3.1 > _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxxx https://lists.xenproject.org/mailman/listinfo/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |