|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH V4] X86/vMCE: handle broken page with regard to migration
Sorry all, please ignore last 2 email, I mis-operate hg email :(
I'm check the reason and will send out later.
Thanks,
Jinsong
Liu Jinsong wrote:
> Generally, there are 2 cases:
> 1. broken page occurs before migration
> 2. broken page occurs during migration
>
> We had submitted 2 versions of patches. Their difference is how to
> handle case 2:
> V1 patch aborted migration for the case 'broken page occurs during
> migration';
> V2 patch marked broken page to dirty bitmap then got handled next
> round and no abort;
>
> This is V3 patch, adding handle for vMCE occur at last iteration of
> migration:
> At the sender
> For case 1, the broken page will be mapped but not copied to target
> (otherwise it may trigger more serious error, say, SRAR error).
> While its pfn_type and pfn number will be transferred to target
> so that target take appropriate action.
>
> For case 2, mce handler marks the broken page to dirty bitmap, so
> that at copypages stage of migration, its pfn_type and pfn number
> will be transferred to target and then take appropriate action.
>
> At the target
> When migration target populates pages for guest. As for the case
> of broken page wrt migration, we prefer to keep the type of the
> page, for the sake of seamless migration.
>
> At target it will set p2m as p2m_ram_broken for broken page. Guest
> MCE may have good chance to handle its broken page, while if guest
> access the broken page again it will kill itself as expected.
>
> If vMCE occur at the last memory copy iteration of migration, we do
> one more
> iteration so that the broken page's pfn_type and pfn number has
> chance to
> be transferred to target.
>
> Suggested-by: George Dunlap <george.dunlap@xxxxxxxxxxxxx>
> Signed-off-by: Liu Jinsong <jinsong.liu@xxxxxxxxx>
> Acked-by: George Dunlap <george.dunlap@xxxxxxxxxxxxx>
>
> - for tools part of the patch
> Acked-by: Ian Jackson <Ian.Jackson@xxxxxxxxxxxxx>
>
> - for hypervisor part of the patch
> Acked-by: Jan Beulich <JBeulich@xxxxxxxx>
>
> diff -r 8b93ac0c93f3 -r 446f6b9bfc89 tools/libxc/xc_domain.c
> --- a/tools/libxc/xc_domain.c Tue Nov 13 11:19:17 2012 +0000
> +++ b/tools/libxc/xc_domain.c Sat Nov 17 09:46:05 2012 +0800
> @@ -283,6 +283,22 @@
> return ret;
> }
>
> +/* set broken page p2m */
> +int xc_set_broken_page_p2m(xc_interface *xch,
> + uint32_t domid,
> + unsigned long pfn)
> +{
> + int ret;
> + DECLARE_DOMCTL;
> +
> + domctl.cmd = XEN_DOMCTL_set_broken_page_p2m;
> + domctl.domain = (domid_t)domid;
> + domctl.u.set_broken_page_p2m.pfn = pfn;
> + ret = do_domctl(xch, &domctl);
> +
> + return ret ? -1 : 0;
> +}
> +
> /* get info from hvm guest for save */
> int xc_domain_hvm_getcontext(xc_interface *xch,
> uint32_t domid,
> diff -r 8b93ac0c93f3 -r 446f6b9bfc89 tools/libxc/xc_domain_restore.c
> --- a/tools/libxc/xc_domain_restore.c Tue Nov 13 11:19:17 2012 +0000
> +++ b/tools/libxc/xc_domain_restore.c Sat Nov 17 09:46:05 2012 +0800
> @@ -1023,9 +1023,15 @@
>
> countpages = count;
> for (i = oldcount; i < buf->nr_pages; ++i)
> - if ((buf->pfn_types[i] & XEN_DOMCTL_PFINFO_LTAB_MASK) ==
> XEN_DOMCTL_PFINFO_XTAB
> - ||(buf->pfn_types[i] & XEN_DOMCTL_PFINFO_LTAB_MASK) ==
> XEN_DOMCTL_PFINFO_XALLOC) + {
> + unsigned long pagetype;
> +
> + pagetype = buf->pfn_types[i] & XEN_DOMCTL_PFINFO_LTAB_MASK;
> + if ( pagetype == XEN_DOMCTL_PFINFO_XTAB ||
> + pagetype == XEN_DOMCTL_PFINFO_BROKEN ||
> + pagetype == XEN_DOMCTL_PFINFO_XALLOC )
> --countpages;
> + }
>
> if (!countpages)
> return count;
> @@ -1267,6 +1273,17 @@
> /* a bogus/unmapped/allocate-only page: skip it */
> continue;
>
> + if ( pagetype == XEN_DOMCTL_PFINFO_BROKEN )
> + {
> + if ( xc_set_broken_page_p2m(xch, dom, pfn) )
> + {
> + ERROR("Set p2m for broken page failed, "
> + "dom=%d, pfn=%lx\n", dom, pfn);
> + goto err_mapped;
> + }
> + continue;
> + }
> +
> if (pfn_err[i])
> {
> ERROR("unexpected PFN mapping failure pfn %lx map_mfn
> %lx p2m_mfn %lx",
> diff -r 8b93ac0c93f3 -r 446f6b9bfc89 tools/libxc/xc_domain_save.c
> --- a/tools/libxc/xc_domain_save.c Tue Nov 13 11:19:17 2012 +0000
> +++ b/tools/libxc/xc_domain_save.c Sat Nov 17 09:46:05 2012 +0800
> @@ -1118,7 +1118,7 @@
> /* Now write out each data page, canonicalising page tables as
> we go... */ for ( ; ; )
> {
> - unsigned int N, batch, run;
> + unsigned int N, batch, run, broken_page_num1,
> broken_page_num2; char reportbuf[80];
>
> snprintf(reportbuf, sizeof(reportbuf),
> @@ -1270,6 +1270,7 @@
> goto out;
> }
>
> + broken_page_num1 = 0;
> for ( run = j = 0; j < batch; j++ )
> {
> unsigned long gmfn = pfn_batch[j];
> @@ -1277,6 +1278,14 @@
> if ( !hvm )
> gmfn = pfn_to_mfn(gmfn);
>
> + if ( pfn_type[j] == XEN_DOMCTL_PFINFO_BROKEN )
> + {
> + pfn_type[j] |= pfn_batch[j];
> + ++broken_page_num1;
> + ++run;
> + continue;
> + }
> +
> if ( pfn_err[j] )
> {
> if ( pfn_type[j] == XEN_DOMCTL_PFINFO_XTAB )
> @@ -1371,8 +1380,12 @@
> }
> }
>
> - /* skip pages that aren't present or are alloc-only
> */ + /*
> + * skip pages that aren't present,
> + * or are broken, or are alloc-only
> + */
> if ( pagetype == XEN_DOMCTL_PFINFO_XTAB
> + || pagetype == XEN_DOMCTL_PFINFO_BROKEN
> || pagetype == XEN_DOMCTL_PFINFO_XALLOC )
> continue;
>
> @@ -1484,6 +1497,35 @@
>
> munmap(region_base, batch*PAGE_SIZE);
>
> + /*
> + * if vMCE occur at last iter, do one more iter so that
> it get + * chance to transfer broken page's pfn_type and
> pfn number to + * target and then take appropriate action
> + */
> + if ( last_iter )
> + {
> + for ( j = 0; j < batch; j++ )
> + {
> + if ( hvm )
> + pfn_type[j] = pfn_batch[j];
> + else
> + pfn_type[j] = pfn_to_mfn(pfn_batch[j]);
> + }
> +
> + if ( xc_get_pfn_type_batch(xch, dom, batch,
> pfn_type) ) + {
> + PERROR("get_pfn_type_batch failed");
> + goto out;
> + }
> +
> + broken_page_num2 = 0;
> + for ( j = 0; j < batch; j++ )
> + if ( pfn_type[j] == XEN_DOMCTL_PFINFO_BROKEN )
> + broken_page_num2++;
> +
> + if ( broken_page_num1 < broken_page_num2 )
> + last_iter = 0;
> + }
> } /* end of this while loop for this iteration */
>
> skip:
> @@ -1550,23 +1592,22 @@
> PERROR("Error when writing to state file (tsc)");
> goto out;
> }
> + }
> + }
> + else
> + last_iter = 1;
>
> + if ( xc_shadow_control(xch, dom,
> + XEN_DOMCTL_SHADOW_OP_CLEAN,
> HYPERCALL_BUFFER(to_send), +
> dinfo->p2m_size, NULL, 0, &shadow_stats) != dinfo->p2m_size ) +
> { + PERROR("Error flushing shadow PT");
> + goto out;
> + }
>
> - }
> + sent_last_iter = sent_this_iter;
>
> - if ( xc_shadow_control(xch, dom,
> - XEN_DOMCTL_SHADOW_OP_CLEAN,
> HYPERCALL_BUFFER(to_send),
> - dinfo->p2m_size, NULL, 0,
> &shadow_stats) != dinfo->p2m_size )
> - {
> - PERROR("Error flushing shadow PT");
> - goto out;
> - }
> -
> - sent_last_iter = sent_this_iter;
> -
> - print_stats(xch, dom, sent_this_iter, &time_stats,
> &shadow_stats, 1); -
> - }
> + print_stats(xch, dom, sent_this_iter, &time_stats,
> &shadow_stats, 1); } /* end of infinite for loop */
>
> DPRINTF("All memory is saved\n");
> diff -r 8b93ac0c93f3 -r 446f6b9bfc89 tools/libxc/xenctrl.h
> --- a/tools/libxc/xenctrl.h Tue Nov 13 11:19:17 2012 +0000
> +++ b/tools/libxc/xenctrl.h Sat Nov 17 09:46:05 2012 +0800
> @@ -575,6 +575,17 @@
> xc_domaininfo_t *info);
>
> /**
> + * This function set p2m for broken page
> + * &parm xch a handle to an open hypervisor interface
> + * @parm domid the domain id which broken page belong to
> + * @parm pfn the pfn number of the broken page
> + * @return 0 on success, -1 on failure
> + */
> +int xc_set_broken_page_p2m(xc_interface *xch,
> + uint32_t domid,
> + unsigned long pfn);
> +
> +/**
> * This function returns information about the context of a hvm
> domain
> * @parm xch a handle to an open hypervisor interface
> * @parm domid the domain to get information from
> diff -r 8b93ac0c93f3 -r 446f6b9bfc89
> xen/arch/x86/cpu/mcheck/mcaction.c ---
> a/xen/arch/x86/cpu/mcheck/mcaction.c Tue Nov 13 11:19:17 2012 +0000
> +++ b/xen/arch/x86/cpu/mcheck/mcaction.c Sat Nov 17 09:46:05 2012
> +0800 @@ -1,5 +1,6 @@ #include <xen/types.h>
> #include <xen/sched.h>
> +#include <asm/p2m.h>
> #include "mcaction.h"
> #include "vmce.h"
> #include "mce.h"
> @@ -91,6 +92,24 @@
> goto vmce_failed;
> }
>
> + if ( is_hvm_domain(d) &&
> !d->arch.hvm_domain.dirty_vram && +
> paging_mode_log_dirty(d) ) + {
> + /*
> + * vMCE occur during migration
> + *
> + * At sender, it marks broken page to dirty
> bitmap, + * so that at copypages stage of
> migration, broken + * page's pfn_type and pfn
> number would be transferred + * to target and
> then take appropriate action. + *
> + * At target, it would set p2m as
> p2m_ram_broken for + * broken page, so that if
> guest access the broken page + * again, it
> would kill itself as expected. + */
> + paging_mark_dirty(d, mfn);
> + }
> +
> if ( unmmap_broken_page(d, _mfn(mfn), gfn) )
> {
> printk("Unmap broken memory %lx for DOM%d
> failed\n",
> diff -r 8b93ac0c93f3 -r 446f6b9bfc89 xen/arch/x86/domctl.c
> --- a/xen/arch/x86/domctl.c Tue Nov 13 11:19:17 2012 +0000
> +++ b/xen/arch/x86/domctl.c Sat Nov 17 09:46:05 2012 +0800
> @@ -209,12 +209,18 @@
> for ( j = 0; j < k; j++ )
> {
> unsigned long type = 0;
> + p2m_type_t t;
>
> - page = get_page_from_gfn(d, arr[j], NULL,
> P2M_ALLOC); + page = get_page_from_gfn(d, arr[j],
> &t, P2M_ALLOC);
>
> if ( unlikely(!page) ||
> unlikely(is_xen_heap_page(page)) )
> - type = XEN_DOMCTL_PFINFO_XTAB;
> + {
> + if ( p2m_is_broken(t) )
> + type = XEN_DOMCTL_PFINFO_BROKEN;
> + else
> + type = XEN_DOMCTL_PFINFO_XTAB;
> + }
> else
> {
> switch( page->u.inuse.type_info &
> PGT_type_mask ) @@ -235,6 +241,9 @@
>
> if ( page->u.inuse.type_info & PGT_pinned )
> type |= XEN_DOMCTL_PFINFO_LPINTAB;
> +
> + if ( page->count_info & PGC_broken )
> + type = XEN_DOMCTL_PFINFO_BROKEN;
> }
>
> if ( page )
> @@ -1568,6 +1577,29 @@
> }
> break;
>
> + case XEN_DOMCTL_set_broken_page_p2m:
> + {
> + struct domain *d;
> +
> + d = rcu_lock_domain_by_id(domctl->domain);
> + if ( d != NULL )
> + {
> + p2m_type_t pt;
> + unsigned long pfn = domctl->u.set_broken_page_p2m.pfn;
> + mfn_t mfn = get_gfn_query(d, pfn, &pt);
> +
> + if ( unlikely(!mfn_valid(mfn_x(mfn)) || !p2m_is_ram(pt)
> || + (p2m_change_type(d, pfn, pt,
> p2m_ram_broken) != pt)) ) + ret = -EINVAL;
> +
> + put_gfn(d, pfn);
> + rcu_unlock_domain(d);
> + }
> + else
> + ret = -ESRCH;
> + }
> + break;
> +
> default:
> ret = iommu_do_domctl(domctl, u_domctl);
> break;
> diff -r 8b93ac0c93f3 -r 446f6b9bfc89 xen/include/public/domctl.h
> --- a/xen/include/public/domctl.h Tue Nov 13 11:19:17 2012 +0000
> +++ b/xen/include/public/domctl.h Sat Nov 17 09:46:05 2012 +0800
> @@ -136,6 +136,7 @@
> #define XEN_DOMCTL_PFINFO_LPINTAB (0x1U<<31)
> #define XEN_DOMCTL_PFINFO_XTAB (0xfU<<28) /* invalid page */
> #define XEN_DOMCTL_PFINFO_XALLOC (0xeU<<28) /* allocate-only page */
> +#define XEN_DOMCTL_PFINFO_BROKEN (0xdU<<28) /* broken page */
> #define XEN_DOMCTL_PFINFO_LTAB_MASK (0xfU<<28)
>
> struct xen_domctl_getpageframeinfo {
> @@ -834,6 +835,12 @@
> typedef struct xen_domctl_set_access_required
> xen_domctl_set_access_required_t;
> DEFINE_XEN_GUEST_HANDLE(xen_domctl_set_access_required_t);
>
> +struct xen_domctl_set_broken_page_p2m {
> + uint64_aligned_t pfn;
> +};
> +typedef struct xen_domctl_set_broken_page_p2m
> xen_domctl_set_broken_page_p2m_t;
> +DEFINE_XEN_GUEST_HANDLE(xen_domctl_set_broken_page_p2m_t); +
> struct xen_domctl {
> uint32_t cmd;
> #define XEN_DOMCTL_createdomain 1
> @@ -899,6 +906,7 @@
> #define XEN_DOMCTL_set_access_required 64
> #define XEN_DOMCTL_audit_p2m 65
> #define XEN_DOMCTL_set_virq_handler 66
> +#define XEN_DOMCTL_set_broken_page_p2m 67
> #define XEN_DOMCTL_gdbsx_guestmemio 1000
> #define XEN_DOMCTL_gdbsx_pausevcpu 1001
> #define XEN_DOMCTL_gdbsx_unpausevcpu 1002
> @@ -954,6 +962,7 @@
> struct xen_domctl_audit_p2m audit_p2m;
> struct xen_domctl_set_virq_handler set_virq_handler;
> struct xen_domctl_gdbsx_memio gdbsx_guest_memio;
> + struct xen_domctl_set_broken_page_p2m set_broken_page_p2m;
> struct xen_domctl_gdbsx_pauseunp_vcpu gdbsx_pauseunp_vcpu;
> struct xen_domctl_gdbsx_domstatus gdbsx_domstatus;
> uint8_t pad[128];
>
> _______________________________________________
> Xen-devel mailing list
> Xen-devel@xxxxxxxxxxxxx
> http://lists.xen.org/xen-devel
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |