[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

[Xen-changelog] [xen-unstable] x86 hvm: Make a couple of hypercall state flags per-vcpu



# HG changeset patch
# User Keir Fraser <keir@xxxxxxx>
# Date 1289911355 0
# Node ID 07bbfe6dd27d51f2ace6dc1e127f899eee672ba6
# Parent  379123d8dbab42351f9ebf41bcb0a21ef8440c24
x86 hvm: Make a couple of hypercall state flags per-vcpu

This is a prerequisite for allowing guest descheduling within a
hypercall.

Signed-off-by: Keir Fraser <keir@xxxxxxx>
---
 xen/arch/x86/domain.c                  |   12 ++++--------
 xen/arch/x86/hvm/hvm.c                 |   26 ++++++++++++--------------
 xen/include/asm-x86/hvm/guest_access.h |    3 ---
 xen/include/asm-x86/hvm/vcpu.h         |    3 +++
 xen/include/asm-x86/hypercall.h        |    9 ---------
 5 files changed, 19 insertions(+), 34 deletions(-)

diff -r 379123d8dbab -r 07bbfe6dd27d xen/arch/x86/domain.c
--- a/xen/arch/x86/domain.c     Tue Nov 16 11:44:09 2010 +0000
+++ b/xen/arch/x86/domain.c     Tue Nov 16 12:42:35 2010 +0000
@@ -1553,8 +1553,6 @@ void sync_vcpu_execstate(struct vcpu *v)
     __arg;                                                                  \
 })
 
-DEFINE_PER_CPU(char, hc_preempted);
-
 unsigned long hypercall_create_continuation(
     unsigned int op, const char *format, ...)
 {
@@ -1583,12 +1581,12 @@ unsigned long hypercall_create_continuat
     {
         regs       = guest_cpu_user_regs();
         regs->eax  = op;
-        /*
-         * For PV guest, we update EIP to re-execute 'syscall' / 'int 0x82';
-         * HVM does not need this since 'vmcall' / 'vmmcall' is fault-like.
-         */
+
+        /* Ensure the hypercall trap instruction is re-executed. */
         if ( !is_hvm_vcpu(current) )
             regs->eip -= 2;  /* re-execute 'syscall' / 'int 0x82' */
+        else
+            current->arch.hvm_vcpu.hcall_preempted = 1;
 
 #ifdef __x86_64__
         if ( !is_hvm_vcpu(current) ?
@@ -1629,8 +1627,6 @@ unsigned long hypercall_create_continuat
                 }
             }
         }
-
-        this_cpu(hc_preempted) = 1;
     }
 
     va_end(args);
diff -r 379123d8dbab -r 07bbfe6dd27d xen/arch/x86/hvm/hvm.c
--- a/xen/arch/x86/hvm/hvm.c    Tue Nov 16 11:44:09 2010 +0000
+++ b/xen/arch/x86/hvm/hvm.c    Tue Nov 16 12:42:35 2010 +0000
@@ -2028,16 +2028,13 @@ enum hvm_copy_result hvm_fetch_from_gues
                       PFEC_page_present | pfec);
 }
 
+unsigned long copy_to_user_hvm(void *to, const void *from, unsigned int len)
+{
+    int rc;
+
 #ifdef __x86_64__
-DEFINE_PER_CPU(bool_t, hvm_64bit_hcall);
-#endif
-
-unsigned long copy_to_user_hvm(void *to, const void *from, unsigned int len)
-{
-    int rc;
-
-#ifdef __x86_64__
-    if ( !this_cpu(hvm_64bit_hcall) && is_compat_arg_xlat_range(to, len) )
+    if ( !current->arch.hvm_vcpu.hcall_64bit &&
+         is_compat_arg_xlat_range(to, len) )
     {
         memcpy(to, from, len);
         return 0;
@@ -2054,7 +2051,8 @@ unsigned long copy_from_user_hvm(void *t
     int rc;
 
 #ifdef __x86_64__
-    if ( !this_cpu(hvm_64bit_hcall) && is_compat_arg_xlat_range(from, len) )
+    if ( !current->arch.hvm_vcpu.hcall_64bit &&
+         is_compat_arg_xlat_range(from, len) )
     {
         memcpy(to, from, len);
         return 0;
@@ -2567,7 +2565,7 @@ int hvm_do_hypercall(struct cpu_user_reg
         return HVM_HCALL_completed;
     }
 
-    this_cpu(hc_preempted) = 0;
+    curr->arch.hvm_vcpu.hcall_preempted = 0;
 
 #ifdef __x86_64__
     if ( mode == 8 )
@@ -2575,13 +2573,13 @@ int hvm_do_hypercall(struct cpu_user_reg
         HVM_DBG_LOG(DBG_LEVEL_HCALL, "hcall%u(%lx, %lx, %lx, %lx, %lx)", eax,
                     regs->rdi, regs->rsi, regs->rdx, regs->r10, regs->r8);
 
-        this_cpu(hvm_64bit_hcall) = 1;
+        curr->arch.hvm_vcpu.hcall_64bit = 1;
         regs->rax = hvm_hypercall64_table[eax](regs->rdi,
                                                regs->rsi,
                                                regs->rdx,
                                                regs->r10,
                                                regs->r8); 
-        this_cpu(hvm_64bit_hcall) = 0;
+        curr->arch.hvm_vcpu.hcall_64bit = 0;
     }
     else
 #endif
@@ -2601,7 +2599,7 @@ int hvm_do_hypercall(struct cpu_user_reg
     HVM_DBG_LOG(DBG_LEVEL_HCALL, "hcall%u -> %lx",
                 eax, (unsigned long)regs->eax);
 
-    if ( this_cpu(hc_preempted) )
+    if ( curr->arch.hvm_vcpu.hcall_preempted )
         return HVM_HCALL_preempted;
 
     if ( unlikely(curr->domain->arch.hvm_domain.qemu_mapcache_invalidate) &&
diff -r 379123d8dbab -r 07bbfe6dd27d xen/include/asm-x86/hvm/guest_access.h
--- a/xen/include/asm-x86/hvm/guest_access.h    Tue Nov 16 11:44:09 2010 +0000
+++ b/xen/include/asm-x86/hvm/guest_access.h    Tue Nov 16 12:42:35 2010 +0000
@@ -1,8 +1,5 @@
 #ifndef __ASM_X86_HVM_GUEST_ACCESS_H__
 #define __ASM_X86_HVM_GUEST_ACCESS_H__
-
-#include <xen/percpu.h>
-DECLARE_PER_CPU(bool_t, hvm_64bit_hcall);
 
 unsigned long copy_to_user_hvm(void *to, const void *from, unsigned len);
 unsigned long copy_from_user_hvm(void *to, const void *from, unsigned len);
diff -r 379123d8dbab -r 07bbfe6dd27d xen/include/asm-x86/hvm/vcpu.h
--- a/xen/include/asm-x86/hvm/vcpu.h    Tue Nov 16 11:44:09 2010 +0000
+++ b/xen/include/asm-x86/hvm/vcpu.h    Tue Nov 16 12:42:35 2010 +0000
@@ -63,6 +63,9 @@ struct hvm_vcpu {
     bool_t              debug_state_latch;
     bool_t              single_step;
 
+    bool_t              hcall_preempted;
+    bool_t              hcall_64bit;
+
     u64                 asid_generation;
     u32                 asid;
 
diff -r 379123d8dbab -r 07bbfe6dd27d xen/include/asm-x86/hypercall.h
--- a/xen/include/asm-x86/hypercall.h   Tue Nov 16 11:44:09 2010 +0000
+++ b/xen/include/asm-x86/hypercall.h   Tue Nov 16 12:42:35 2010 +0000
@@ -16,15 +16,6 @@
  * invocation of do_mmu_update() is resuming a previously preempted call.
  */
 #define MMU_UPDATE_PREEMPTED          (~(~0U>>1))
-
-/*
- * This gets set to a non-zero value whenever hypercall_create_continuation()
- * is used (outside of multicall context; in multicall context the second call
- * from do_multicall() itself will have this effect). Internal callers of
- * hypercall handlers interested in this condition must clear the flag prior
- * to invoking the respective handler(s).
- */
-DECLARE_PER_CPU(char, hc_preempted);
 
 extern long
 do_event_channel_op_compat(

_______________________________________________
Xen-changelog mailing list
Xen-changelog@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-changelog


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.