[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH RFC 11/44] x86/pt-shadow: Always set _PAGE_ACCESSED on L4e updates
Signed-off-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx> Reviewed-by: Jan Beulich <jbeulich@xxxxxxxx> --- xen/arch/x86/pv/mm.h | 19 ++++++++++++++++--- 1 file changed, 16 insertions(+), 3 deletions(-) diff --git a/xen/arch/x86/pv/mm.h b/xen/arch/x86/pv/mm.h index 7502d53..a10b09a 100644 --- a/xen/arch/x86/pv/mm.h +++ b/xen/arch/x86/pv/mm.h @@ -144,9 +144,22 @@ static inline l3_pgentry_t unadjust_guest_l3e(l3_pgentry_t l3e, static inline l4_pgentry_t adjust_guest_l4e(l4_pgentry_t l4e, const struct domain *d) { - if ( likely(l4e_get_flags(l4e) & _PAGE_PRESENT) && - likely(!is_pv_32bit_domain(d)) ) - l4e_add_flags(l4e, _PAGE_USER); + /* + * When shadowing an L4 for per-pcpu purposes, we cannot efficiently sync + * access bit updates from hardware (on the shadow tables) back into the + * guest view. We therefore always set _PAGE_ACCESSED even in the guests + * view. + * + * This will appear to the guest as a CPU which proactively pulls all + * valid L4e's into its TLB, which is compatible with the x86 ABI. + * + * Furthermore, at the time of writing, all PV guests I can locate choose + * to set the access bit anyway, so this is no actual change in their + * behaviour. + */ + if ( likely(l4e_get_flags(l4e) & _PAGE_PRESENT) ) + l4e_add_flags(l4e, (_PAGE_ACCESSED | + (is_pv_32bit_domain(d) ? 0 : _PAGE_USER))); return l4e; } -- 2.1.4 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxxx https://lists.xenproject.org/mailman/listinfo/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |