[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [PATCH] x86/pv: Flush TLB in response to paging structure changes
On 20/10/2020 16:24, Andrew Cooper wrote: > With MMU_UPDATE, a PV guest can make changes to higher level pagetables. This > is from Xen's point of view (as the update only affects guest mappings), and > the guest is required to flush suitably after making updates. > > However, Xen's use of linear pagetables (UPDATE_VA_MAPPING, GNTTABOP_map, > writeable pagetables, etc.) is an implementation detail outside of the > API/ABI. > > Changes in the paging structure require invalidations in the linear pagetable > range for subsequent accesses into the linear pagetables to access non-stale > mappings. Xen must provide suitable flushing to prevent intermixed guest > actions from accidentally accessing/modifying the wrong pagetable. > > For all L2 and higher modifications, flush the full TLB. (This could in > principle be an order 39 flush starting at LINEAR_PT_VIRT_START, but no such > mechanism exists in practice.) > > As this combines with sync_guest for XPTI L4 "shadowing", replace the > sync_guest boolean with flush_flags and accumulate flags. The sync_guest case > now always needs to flush, there is no point trying to exclude the current CPU > from the flush mask. Use pt_owner->dirty_cpumask directly. > > This is XSA-286. > > Signed-off-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx> > --- > CC: Jan Beulich <JBeulich@xxxxxxxx> > CC: Roger Pau Monné <roger.pau@xxxxxxxxxx> > CC: Wei Liu <wl@xxxxxxx> > > A couple of minor points. > > * PV guests can create global mappings. I can't reason any safe way to relax > FLUSH_TLB_GLOBAL to just FLUSH_TLB. Sorry - forgot one of the points here. We could in principle relax the flush entirely if we know that we're editing from a not-present to present entry, but plumbing this up through mod_l?_entry() isn't trivial, and its also not not obvious how much of an optimisation it would be in practice. ~Andrew > * Performance tests are still ongoing, but so far is fairing better than the > embargoed alternative. > --- > xen/arch/x86/mm.c | 31 +++++++++++++++---------------- > 1 file changed, 15 insertions(+), 16 deletions(-) > > diff --git a/xen/arch/x86/mm.c b/xen/arch/x86/mm.c > index 918ee2bbe3..a6a7fcb56c 100644 > --- a/xen/arch/x86/mm.c > +++ b/xen/arch/x86/mm.c > @@ -3883,11 +3883,10 @@ long do_mmu_update( > void *va = NULL; > unsigned long gpfn, gmfn; > struct page_info *page; > - unsigned int cmd, i = 0, done = 0, pt_dom; > + unsigned int cmd, i = 0, done = 0, pt_dom, flush_flags = 0; > struct vcpu *curr = current, *v = curr; > struct domain *d = v->domain, *pt_owner = d, *pg_owner; > mfn_t map_mfn = INVALID_MFN, mfn; > - bool sync_guest = false; > uint32_t xsm_needed = 0; > uint32_t xsm_checked = 0; > int rc = put_old_guest_table(curr); > @@ -4037,6 +4036,8 @@ long do_mmu_update( > break; > rc = mod_l2_entry(va, l2e_from_intpte(req.val), mfn, > cmd == MMU_PT_UPDATE_PRESERVE_AD, v); > + if ( !rc ) > + flush_flags |= FLUSH_TLB_GLOBAL; > break; > > case PGT_l3_page_table: > @@ -4044,6 +4045,8 @@ long do_mmu_update( > break; > rc = mod_l3_entry(va, l3e_from_intpte(req.val), mfn, > cmd == MMU_PT_UPDATE_PRESERVE_AD, v); > + if ( !rc ) > + flush_flags |= FLUSH_TLB_GLOBAL; > break; > > case PGT_l4_page_table: > @@ -4051,6 +4054,8 @@ long do_mmu_update( > break; > rc = mod_l4_entry(va, l4e_from_intpte(req.val), mfn, > cmd == MMU_PT_UPDATE_PRESERVE_AD, v); > + if ( !rc ) > + flush_flags |= FLUSH_TLB_GLOBAL; > if ( !rc && pt_owner->arch.pv.xpti ) > { > bool local_in_use = false; > @@ -4071,7 +4076,7 @@ long do_mmu_update( > (1 + !!(page->u.inuse.type_info & PGT_pinned) + > > mfn_eq(pagetable_get_mfn(curr->arch.guest_table_user), > mfn) + local_in_use) ) > - sync_guest = true; > + flush_flags |= FLUSH_ROOT_PGTBL; > } > break; > > @@ -4173,19 +4178,13 @@ long do_mmu_update( > if ( va ) > unmap_domain_page(va); > > - if ( sync_guest ) > - { > - /* > - * Force other vCPU-s of the affected guest to pick up L4 entry > - * changes (if any). > - */ > - unsigned int cpu = smp_processor_id(); > - cpumask_t *mask = per_cpu(scratch_cpumask, cpu); > - > - cpumask_andnot(mask, pt_owner->dirty_cpumask, cpumask_of(cpu)); > - if ( !cpumask_empty(mask) ) > - flush_mask(mask, FLUSH_TLB_GLOBAL | FLUSH_ROOT_PGTBL); > - } > + /* > + * Flush TLBs if an L2 or higher was changed (invalidates the structure > of > + * the linear pagetables), or an L4 in use by other CPUs was made (needs > + * to resync the XPTI copy of the table). > + */ > + if ( flush_flags ) > + flush_mask(pt_owner->dirty_cpumask, flush_flags); > > perfc_add(num_page_updates, i); >
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |