[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-changelog] [xen master] x86/SVM: Fix intercepted {RD, WR}MSR for the SYS{CALL, ENTER} MSRs
commit c04c1866e5131e450ddcd114e32401477c60b816 Author: Andrew Cooper <andrew.cooper3@xxxxxxxxxx> AuthorDate: Tue Apr 24 17:08:18 2018 +0100 Commit: Andrew Cooper <andrew.cooper3@xxxxxxxxxx> CommitDate: Wed Apr 25 13:08:13 2018 +0100 x86/SVM: Fix intercepted {RD,WR}MSR for the SYS{CALL,ENTER} MSRs By default, the SYSCALL MSRs are not intercepted, and accesses are completed by hardware. The SYSENTER MSRs are intercepted for cross-vendor purposes (albeit needlessly in the common case), and are fully emulated. However, {RD,WR}MSR instructions which happen to be emulated (FEP, introspection, or older versions of Xen which intercepted #UD), or when the MSRs are explicitly intercepted (introspection), will be completed incorrectly. svm_msr_read_intercept() appears to return the correct values, but only because of the default read-everything case (which is going to disappear), and that in vcpu context, hardware should have the guest values in context. Update the read path to explicitly sync the VMCB and complete the accesses, rather than falling all the way through to the default case. svm_msr_write_intercept() silently discard all updates. Synchronise the VMCB for all applicable MSRs, and implement suitable checks. The actual behaviour of AMD hardware is to truncate the SYSENTER and SFMASK MSRs at 32 bits, but this isn't implemented yet to remain compatible with the cross-vendor case. Drop one bit of trailing whitespace while modifing this area of the code. Reported-by: Razvan Cojocaru <rcojocaru@xxxxxxxxxxxxxxx> Signed-off-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx> Reviewed-by: Boris Ostrovsky <boris.ostrovsky@xxxxxxxxxx> Release-acked-by: Juergen Gross <jgross@xxxxxxxx> --- xen/arch/x86/hvm/svm/svm.c | 114 +++++++++++++++++++++++++++++++++++++++++---- 1 file changed, 105 insertions(+), 9 deletions(-) diff --git a/xen/arch/x86/hvm/svm/svm.c b/xen/arch/x86/hvm/svm/svm.c index c7616456dd..88938e6ae6 100644 --- a/xen/arch/x86/hvm/svm/svm.c +++ b/xen/arch/x86/hvm/svm/svm.c @@ -1882,6 +1882,25 @@ static int svm_msr_read_intercept(unsigned int msr, uint64_t *msr_content) switch ( msr ) { + /* + * Sync not needed while the cross-vendor logic is in unilateral effect. + case MSR_IA32_SYSENTER_CS: + case MSR_IA32_SYSENTER_ESP: + case MSR_IA32_SYSENTER_EIP: + */ + case MSR_STAR: + case MSR_LSTAR: + case MSR_CSTAR: + case MSR_SYSCALL_MASK: + case MSR_FS_BASE: + case MSR_GS_BASE: + case MSR_SHADOW_GS_BASE: + svm_sync_vmcb(v); + break; + } + + switch ( msr ) + { case MSR_IA32_SYSENTER_CS: *msr_content = v->arch.hvm_svm.guest_sysenter_cs; break; @@ -1892,6 +1911,34 @@ static int svm_msr_read_intercept(unsigned int msr, uint64_t *msr_content) *msr_content = v->arch.hvm_svm.guest_sysenter_eip; break; + case MSR_STAR: + *msr_content = vmcb->star; + break; + + case MSR_LSTAR: + *msr_content = vmcb->lstar; + break; + + case MSR_CSTAR: + *msr_content = vmcb->cstar; + break; + + case MSR_SYSCALL_MASK: + *msr_content = vmcb->sfmask; + break; + + case MSR_FS_BASE: + *msr_content = vmcb->fs.base; + break; + + case MSR_GS_BASE: + *msr_content = vmcb->gs.base; + break; + + case MSR_SHADOW_GS_BASE: + *msr_content = vmcb->kerngsbase; + break; + case MSR_IA32_MCx_MISC(4): /* Threshold register */ case MSR_F10_MC4_MISC1 ... MSR_F10_MC4_MISC3: /* @@ -2020,32 +2067,81 @@ static int svm_msr_write_intercept(unsigned int msr, uint64_t msr_content) int ret, result = X86EMUL_OKAY; struct vcpu *v = current; struct vmcb_struct *vmcb = v->arch.hvm_svm.vmcb; - int sync = 0; + bool sync = false; switch ( msr ) { case MSR_IA32_SYSENTER_CS: case MSR_IA32_SYSENTER_ESP: case MSR_IA32_SYSENTER_EIP: - sync = 1; - break; - default: + case MSR_STAR: + case MSR_LSTAR: + case MSR_CSTAR: + case MSR_SYSCALL_MASK: + case MSR_FS_BASE: + case MSR_GS_BASE: + case MSR_SHADOW_GS_BASE: + sync = true; break; } if ( sync ) - svm_sync_vmcb(v); + svm_sync_vmcb(v); switch ( msr ) { + case MSR_IA32_SYSENTER_ESP: + case MSR_IA32_SYSENTER_EIP: + case MSR_LSTAR: + case MSR_CSTAR: + case MSR_FS_BASE: + case MSR_GS_BASE: + case MSR_SHADOW_GS_BASE: + if ( !is_canonical_address(msr_content) ) + goto gpf; + + switch ( msr ) + { + case MSR_IA32_SYSENTER_ESP: + vmcb->sysenter_esp = v->arch.hvm_svm.guest_sysenter_esp = msr_content; + break; + + case MSR_IA32_SYSENTER_EIP: + vmcb->sysenter_eip = v->arch.hvm_svm.guest_sysenter_eip = msr_content; + break; + + case MSR_LSTAR: + vmcb->lstar = msr_content; + break; + + case MSR_CSTAR: + vmcb->cstar = msr_content; + break; + + case MSR_FS_BASE: + vmcb->fs.base = msr_content; + break; + + case MSR_GS_BASE: + vmcb->gs.base = msr_content; + break; + + case MSR_SHADOW_GS_BASE: + vmcb->kerngsbase = msr_content; + break; + } + break; + case MSR_IA32_SYSENTER_CS: vmcb->sysenter_cs = v->arch.hvm_svm.guest_sysenter_cs = msr_content; break; - case MSR_IA32_SYSENTER_ESP: - vmcb->sysenter_esp = v->arch.hvm_svm.guest_sysenter_esp = msr_content; + + case MSR_STAR: + vmcb->star = msr_content; break; - case MSR_IA32_SYSENTER_EIP: - vmcb->sysenter_eip = v->arch.hvm_svm.guest_sysenter_eip = msr_content; + + case MSR_SYSCALL_MASK: + vmcb->sfmask = msr_content; break; case MSR_IA32_DEBUGCTLMSR: -- generated by git-patchbot for /home/xen/git/xen.git#master _______________________________________________ Xen-changelog mailing list Xen-changelog@xxxxxxxxxxxxxxxxxxxx https://lists.xenproject.org/xen-changelog
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |