[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

[Xen-devel] [PATCH RFC v1 58/74] xen/pvshim: add migration support



From: Roger Pau Monne <roger.pau@xxxxxxxxxx>

Signed-off-by: Roger Pau Monné <roger.pau@xxxxxxxxxx>
---
 xen/arch/x86/guest/xen.c          |  29 +++++++
 xen/arch/x86/mm.c                 |   3 +-
 xen/arch/x86/pv/shim.c            | 163 +++++++++++++++++++++++++++++++++++++-
 xen/common/domain.c               |  11 ++-
 xen/common/schedule.c             |   3 +-
 xen/drivers/char/xen_pv_console.c |   2 +-
 xen/include/asm-x86/guest/xen.h   |   5 ++
 xen/include/asm-x86/pv/shim.h     |   5 +-
 xen/include/xen/sched.h           |   2 +-
 9 files changed, 206 insertions(+), 17 deletions(-)

diff --git a/xen/arch/x86/guest/xen.c b/xen/arch/x86/guest/xen.c
index 653a7366ab..54c997b9e0 100644
--- a/xen/arch/x86/guest/xen.c
+++ b/xen/arch/x86/guest/xen.c
@@ -361,6 +361,35 @@ uint32_t hypervisor_cpuid_base(void)
     return xen_cpuid_base;
 }
 
+void ap_resume(void *unused)
+{
+    map_vcpuinfo();
+    init_evtchn();
+}
+
+void hypervisor_resume(void)
+{
+    /* Reset shared info page. */
+    map_shared_info();
+
+    /*
+     * Reset vcpu_info. Just clean the mapped bitmap and try to map the vcpu
+     * area again. On failure to map (when it was previously mapped) panic
+     * since it's impossible to safely shut down running guest vCPUs in order
+     * to meet the new XEN_LEGACY_MAX_VCPUS requirement.
+     */
+    memset(vcpu_info_mapped, 0, sizeof(vcpu_info_mapped));
+    if ( map_vcpuinfo() && nr_cpu_ids > XEN_LEGACY_MAX_VCPUS )
+        panic("unable to remap vCPU info and vCPUs > legacy limit");
+
+    /* Setup event channel upcall vector. */
+    init_evtchn();
+    smp_call_function(ap_resume, NULL, 1);
+
+    if ( pv_console )
+        pv_console_init();
+}
+
 /*
  * Local variables:
  * mode: C
diff --git a/xen/arch/x86/mm.c b/xen/arch/x86/mm.c
index 4332d3bb39..e726c62064 100644
--- a/xen/arch/x86/mm.c
+++ b/xen/arch/x86/mm.c
@@ -466,8 +466,7 @@ void share_xen_page_with_guest(
     spin_unlock(&d->page_alloc_lock);
 }
 
-int __init unshare_xen_page_with_guest(struct page_info *page,
-                                       struct domain *d)
+int unshare_xen_page_with_guest(struct page_info *page, struct domain *d)
 {
     if ( page_get_owner(page) != d || !is_xen_heap_page(page) )
         return -EINVAL;
diff --git a/xen/arch/x86/pv/shim.c b/xen/arch/x86/pv/shim.c
index 74b24bf950..56ecaea2d2 100644
--- a/xen/arch/x86/pv/shim.c
+++ b/xen/arch/x86/pv/shim.c
@@ -151,10 +151,167 @@ void __init pv_shim_setup_dom(struct domain *d, 
l4_pgentry_t *l4start,
     }
 }
 
-void pv_shim_shutdown(uint8_t reason)
+static void write_start_info(struct domain *d)
 {
-    /* XXX: handle suspend */
-    xen_hypercall_shutdown(reason);
+    struct cpu_user_regs *regs = guest_cpu_user_regs();
+    start_info_t *si = map_domain_page(_mfn(is_pv_32bit_domain(d) ? regs->edx
+                                                                  : 
regs->rdx));
+    uint64_t param;
+
+    BUG_ON(!si);
+
+    snprintf(si->magic, sizeof(si->magic), "xen-3.0-x86_%s",
+             is_pv_32bit_domain(d) ? "32p" : "64");
+    si->nr_pages = d->tot_pages;
+    si->shared_info = virt_to_maddr(d->shared_info);
+    si->flags = (xen_processor_pmbits << 8) & SIF_PM_MASK;
+    BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_STORE_PFN, &si->store_mfn));
+    BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_STORE_EVTCHN, &param));
+    si->store_evtchn = param;
+    BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_CONSOLE_EVTCHN, &param));
+    si->console.domU.evtchn = param;
+    if ( !pv_console )
+        BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_CONSOLE_PFN,
+                                           &si->console.domU.mfn));
+    else
+        si->console.domU.mfn = virt_to_mfn(consoled_get_ring_addr());
+
+    if ( is_pv_32bit_domain(d) )
+        xlat_start_info(si, XLAT_start_info_console_domU);
+
+    unmap_domain_page(si);
+}
+
+int pv_shim_shutdown(uint8_t reason)
+{
+    long rc;
+
+    if ( reason == SHUTDOWN_suspend )
+    {
+        struct domain *d = current->domain;
+        struct vcpu *v;
+        unsigned int i;
+        uint64_t old_store_pfn, old_console_pfn = 0, store_pfn, console_pfn;
+        uint64_t store_evtchn, console_evtchn;
+
+        BUG_ON(current->vcpu_id != 0);
+
+        BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_STORE_PFN,
+                                           &old_store_pfn));
+        if ( !pv_console )
+            BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_CONSOLE_PFN,
+                                               &old_console_pfn));
+
+        /* Pause the other vcpus before starting the migration. */
+        for_each_vcpu(d, v)
+            if ( v != current )
+                vcpu_pause_by_systemcontroller(v);
+
+        rc = xen_hypercall_shutdown(SHUTDOWN_suspend);
+        if ( rc )
+        {
+            for_each_vcpu(d, v)
+                if ( v != current )
+                    vcpu_unpause_by_systemcontroller(v);
+
+            return rc;
+        }
+
+        /* Resume the shim itself first. */
+        hypervisor_resume();
+
+        /*
+         * ATM there's nothing Xen can do if the console/store pfn changes,
+         * because Xen won't have a page_info struct for it.
+         */
+        BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_STORE_PFN,
+                                           &store_pfn));
+        BUG_ON(old_store_pfn != store_pfn);
+        if ( !pv_console )
+        {
+            BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_CONSOLE_PFN,
+                                               &console_pfn));
+            BUG_ON(old_console_pfn != console_pfn);
+        }
+
+        /* Update domain id. */
+        d->domain_id = get_dom0_domid();
+
+        /* Clean the iomem range. */
+        BUG_ON(iomem_deny_access(d, 0, ~0UL));
+
+        /* Clean grant frames. */
+        xfree(grant_frames);
+        grant_frames = NULL;
+        nr_grant_list = 0;
+
+        /* Clean event channels. */
+        for ( i = 0; i < EVTCHN_2L_NR_CHANNELS; i++ )
+        {
+            if ( !port_is_valid(d, i) )
+                continue;
+
+            if ( evtchn_handled(d, i) )
+                evtchn_close(d, i, false);
+            else
+                evtchn_free(d, evtchn_from_port(d, i));
+        }
+
+        /* Reserve store/console event channel. */
+        BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_STORE_EVTCHN,
+                                           &store_evtchn));
+        BUG_ON(evtchn_allocate_port(d, store_evtchn));
+        evtchn_reserve(d, store_evtchn);
+        BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_CONSOLE_EVTCHN,
+                                           &console_evtchn));
+        BUG_ON(evtchn_allocate_port(d, console_evtchn));
+        evtchn_reserve(d, console_evtchn);
+
+        /* Clean watchdogs. */
+        watchdog_domain_destroy(d);
+        watchdog_domain_init(d);
+
+        /* Clean the PIRQ EOI page. */
+        if ( d->arch.pirq_eoi_map != NULL )
+        {
+            unmap_domain_page_global(d->arch.pirq_eoi_map);
+            put_page_and_type(mfn_to_page(d->arch.pirq_eoi_map_mfn));
+            d->arch.pirq_eoi_map = NULL;
+            d->arch.pirq_eoi_map_mfn = 0;
+            d->arch.auto_unmask = 0;
+        }
+
+        /*
+         * NB: there's no need to fixup the p2m, since the mfns assigned
+         * to the PV guest have not changed at all. Just re-write the
+         * start_info fields with the appropriate value.
+         */
+        write_start_info(d);
+
+        for_each_vcpu(d, v)
+        {
+            /* Unmap guest vcpu_info pages. */
+            unmap_vcpu_info(v);
+
+            /* Reset the periodic timer to the default value. */
+            v->periodic_period = MILLISECS(10);
+            /* Stop the singleshot timer. */
+            stop_timer(&v->singleshot_timer);
+
+            if ( test_bit(_VPF_down, &v->pause_flags) )
+                BUG_ON(vcpu_reset(v));
+
+            if ( v != current )
+                vcpu_unpause_by_systemcontroller(v);
+            else
+                vcpu_force_reschedule(v);
+        }
+    }
+    else
+        /* Forward to L0. */
+        rc = xen_hypercall_shutdown(reason);
+
+    return rc;
 }
 
 long pv_shim_event_channel_op(int cmd, XEN_GUEST_HANDLE_PARAM(void) arg)
diff --git a/xen/common/domain.c b/xen/common/domain.c
index d653a0b0bb..bc2ceb2d36 100644
--- a/xen/common/domain.c
+++ b/xen/common/domain.c
@@ -701,15 +701,12 @@ void __domain_crash_synchronous(void)
 }
 
 
-void domain_shutdown(struct domain *d, u8 reason)
+int domain_shutdown(struct domain *d, u8 reason)
 {
     struct vcpu *v;
 
     if ( pv_shim )
-    {
-        pv_shim_shutdown(reason);
-        return;
-    }
+        return pv_shim_shutdown(reason);
 
     spin_lock(&d->shutdown_lock);
 
@@ -723,7 +720,7 @@ void domain_shutdown(struct domain *d, u8 reason)
     if ( d->is_shutting_down )
     {
         spin_unlock(&d->shutdown_lock);
-        return;
+        return 0;
     }
 
     d->is_shutting_down = 1;
@@ -745,6 +742,8 @@ void domain_shutdown(struct domain *d, u8 reason)
     __domain_finalise_shutdown(d);
 
     spin_unlock(&d->shutdown_lock);
+
+    return 0;
 }
 
 void domain_resume(struct domain *d)
diff --git a/xen/common/schedule.c b/xen/common/schedule.c
index 88279213e8..b7884263f2 100644
--- a/xen/common/schedule.c
+++ b/xen/common/schedule.c
@@ -1149,11 +1149,10 @@ ret_t do_sched_op(int cmd, XEN_GUEST_HANDLE_PARAM(void) 
arg)
         if ( copy_from_guest(&sched_shutdown, arg, 1) )
             break;
 
-        ret = 0;
         TRACE_3D(TRC_SCHED_SHUTDOWN,
                  current->domain->domain_id, current->vcpu_id,
                  sched_shutdown.reason);
-        domain_shutdown(current->domain, (u8)sched_shutdown.reason);
+        ret = domain_shutdown(current->domain, (u8)sched_shutdown.reason);
 
         break;
     }
diff --git a/xen/drivers/char/xen_pv_console.c 
b/xen/drivers/char/xen_pv_console.c
index fb5a7893be..d2a28478ac 100644
--- a/xen/drivers/char/xen_pv_console.c
+++ b/xen/drivers/char/xen_pv_console.c
@@ -37,7 +37,7 @@ static DEFINE_SPINLOCK(tx_lock);
 
 bool pv_console;
 
-void __init pv_console_init(void)
+void pv_console_init(void)
 {
     struct evtchn_unmask unmask;
     long r;
diff --git a/xen/include/asm-x86/guest/xen.h b/xen/include/asm-x86/guest/xen.h
index 94f781c30f..c244569e83 100644
--- a/xen/include/asm-x86/guest/xen.h
+++ b/xen/include/asm-x86/guest/xen.h
@@ -40,6 +40,7 @@ void hypervisor_fixup_e820(struct e820map *e820);
 void hypervisor_init_memory(void);
 const unsigned long *hypervisor_reserved_pages(unsigned int *size);
 uint32_t hypervisor_cpuid_base(void);
+void hypervisor_resume(void);
 
 #else
 
@@ -91,6 +92,10 @@ static inline uint32_t hypervisor_cpuid_base(void)
     ASSERT_UNREACHABLE();
     return 0;
 };
+static inline void hypervisor_resume(void)
+{
+    ASSERT_UNREACHABLE();
+};
 
 #endif /* CONFIG_XEN_GUEST */
 #endif /* __X86_GUEST_XEN_H__ */
diff --git a/xen/include/asm-x86/pv/shim.h b/xen/include/asm-x86/pv/shim.h
index 47bc4267af..0207348a85 100644
--- a/xen/include/asm-x86/pv/shim.h
+++ b/xen/include/asm-x86/pv/shim.h
@@ -35,7 +35,7 @@ void pv_shim_setup_dom(struct domain *d, l4_pgentry_t 
*l4start,
                        unsigned long va_start, unsigned long store_va,
                        unsigned long console_va, unsigned long vphysmap,
                        start_info_t *si);
-void pv_shim_shutdown(uint8_t reason);
+int pv_shim_shutdown(uint8_t reason);
 long pv_shim_event_channel_op(int cmd, XEN_GUEST_HANDLE_PARAM(void) arg);
 void pv_shim_inject_evtchn(unsigned int port);
 long pv_shim_grant_table_op(unsigned int cmd, XEN_GUEST_HANDLE_PARAM(void) uop,
@@ -53,9 +53,10 @@ static inline void pv_shim_setup_dom(struct domain *d, 
l4_pgentry_t *l4start,
 {
     ASSERT_UNREACHABLE();
 }
-static inline void pv_shim_shutdown(uint8_t reason)
+static inline int pv_shim_shutdown(uint8_t reason)
 {
     ASSERT_UNREACHABLE();
+    return 0;
 }
 static inline long pv_shim_event_channel_op(int cmd,
                                             XEN_GUEST_HANDLE_PARAM(void) arg)
diff --git a/xen/include/xen/sched.h b/xen/include/xen/sched.h
index ac65d0c0df..70db377eae 100644
--- a/xen/include/xen/sched.h
+++ b/xen/include/xen/sched.h
@@ -605,7 +605,7 @@ static inline struct domain *rcu_lock_current_domain(void)
 struct domain *get_domain_by_id(domid_t dom);
 void domain_destroy(struct domain *d);
 int domain_kill(struct domain *d);
-void domain_shutdown(struct domain *d, u8 reason);
+int domain_shutdown(struct domain *d, u8 reason);
 void domain_resume(struct domain *d);
 void domain_pause_for_debugger(void);
 
-- 
2.11.0


_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxxx
https://lists.xenproject.org/mailman/listinfo/xen-devel

 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.