[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [xen master] x86/shadow: only 4-level guest code needs building when !HVM
commit aff8bf94ce652bab55d4e0ec8832445bcc1c071f Author: Jan Beulich <jbeulich@xxxxxxxx> AuthorDate: Mon Apr 12 12:34:04 2021 +0200 Commit: Jan Beulich <jbeulich@xxxxxxxx> CommitDate: Mon Apr 12 12:34:04 2021 +0200 x86/shadow: only 4-level guest code needs building when !HVM In order to limit #ifdef-ary, provide "stub" #define-s for SH_type_{l1,fl1,l2}_{32,pae}_shadow and SHF_{L1,FL1,L2}_{32,PAE}. The change in shadow_vcpu_init() is necessary to cover for "x86: correct is_pv_domain() when !CONFIG_PV" (or any other change along those lines) - we should only rely on is_hvm_*() to become a build time constant. Signed-off-by: Jan Beulich <jbeulich@xxxxxxxx> Reviewed-by: Tim Deegan <tim@xxxxxxx> --- xen/arch/x86/mm/Makefile | 2 +- xen/arch/x86/mm/shadow/Makefile | 6 +++--- xen/arch/x86/mm/shadow/common.c | 30 +++++++++++++++++++++++++++--- xen/arch/x86/mm/shadow/private.h | 36 ++++++++++++++++++++++++++++++++++++ 4 files changed, 67 insertions(+), 7 deletions(-) diff --git a/xen/arch/x86/mm/Makefile b/xen/arch/x86/mm/Makefile index a2431fde6b..24f4ec40e0 100644 --- a/xen/arch/x86/mm/Makefile +++ b/xen/arch/x86/mm/Makefile @@ -3,7 +3,7 @@ obj-$(CONFIG_HVM) += hap/ obj-$(CONFIG_HVM) += altp2m.o obj-$(CONFIG_HVM) += guest_walk_2.o guest_walk_3.o guest_walk_4.o -obj-$(CONFIG_SHADOW_PAGING) += guest_walk_2.o guest_walk_3.o guest_walk_4.o +obj-$(CONFIG_SHADOW_PAGING) += guest_walk_4.o obj-$(CONFIG_MEM_ACCESS) += mem_access.o obj-y += mem_paging.o obj-$(CONFIG_MEM_SHARING) += mem_sharing.o diff --git a/xen/arch/x86/mm/shadow/Makefile b/xen/arch/x86/mm/shadow/Makefile index 05ffc164eb..770213fe9d 100644 --- a/xen/arch/x86/mm/shadow/Makefile +++ b/xen/arch/x86/mm/shadow/Makefile @@ -1,7 +1,7 @@ ifeq ($(CONFIG_SHADOW_PAGING),y) -obj-y += common.o guest_2.o guest_3.o guest_4.o set.o -obj-$(CONFIG_HVM) += hvm.o -obj-$(CONFIG_PV) += pv.o +obj-y += common.o set.o +obj-$(CONFIG_HVM) += hvm.o guest_2.o guest_3.o guest_4.o +obj-$(CONFIG_PV) += pv.o guest_4.o else obj-y += none.o endif diff --git a/xen/arch/x86/mm/shadow/common.c b/xen/arch/x86/mm/shadow/common.c index 2b75ea6856..c0c4ce46e4 100644 --- a/xen/arch/x86/mm/shadow/common.c +++ b/xen/arch/x86/mm/shadow/common.c @@ -90,9 +90,9 @@ void shadow_vcpu_init(struct vcpu *v) } #endif - v->arch.paging.mode = is_pv_vcpu(v) ? - &SHADOW_INTERNAL_NAME(sh_paging_mode, 4) : - &SHADOW_INTERNAL_NAME(sh_paging_mode, 3); + v->arch.paging.mode = is_hvm_vcpu(v) ? + &SHADOW_INTERNAL_NAME(sh_paging_mode, 3) : + &SHADOW_INTERNAL_NAME(sh_paging_mode, 4); } #if SHADOW_AUDIT @@ -257,6 +257,7 @@ static int sh_remove_write_access_from_sl1p(struct domain *d, mfn_t gmfn, switch ( mfn_to_page(smfn)->u.sh.type ) { +#ifdef CONFIG_HVM case SH_type_l1_32_shadow: case SH_type_fl1_32_shadow: return SHADOW_INTERNAL_NAME(sh_rm_write_access_from_sl1p, 2) @@ -266,6 +267,7 @@ static int sh_remove_write_access_from_sl1p(struct domain *d, mfn_t gmfn, case SH_type_fl1_pae_shadow: return SHADOW_INTERNAL_NAME(sh_rm_write_access_from_sl1p, 3) (d, gmfn, smfn, off); +#endif case SH_type_l1_64_shadow: case SH_type_fl1_64_shadow: @@ -848,6 +850,7 @@ sh_validate_guest_entry(struct vcpu *v, mfn_t gmfn, void *entry, u32 size) * the free pool. */ +#ifdef CONFIG_HVM const u8 sh_type_to_size[] = { 1, /* SH_type_none */ 2, /* SH_type_l1_32_shadow */ @@ -866,6 +869,7 @@ const u8 sh_type_to_size[] = { 1, /* SH_type_monitor_table */ 1 /* SH_type_oos_snapshot */ }; +#endif /* * Figure out the least acceptable quantity of shadow memory. @@ -892,12 +896,14 @@ void shadow_unhook_mappings(struct domain *d, mfn_t smfn, int user_only) struct page_info *sp = mfn_to_page(smfn); switch ( sp->u.sh.type ) { +#ifdef CONFIG_HVM case SH_type_l2_32_shadow: SHADOW_INTERNAL_NAME(sh_unhook_32b_mappings, 2)(d, smfn, user_only); break; case SH_type_l2_pae_shadow: SHADOW_INTERNAL_NAME(sh_unhook_pae_mappings, 3)(d, smfn, user_only); break; +#endif case SH_type_l4_64_shadow: SHADOW_INTERNAL_NAME(sh_unhook_64b_mappings, 4)(d, smfn, user_only); break; @@ -1104,8 +1110,10 @@ mfn_t shadow_alloc(struct domain *d, /* Backpointers that are MFNs need to be packed into PDXs (PFNs don't) */ switch (shadow_type) { +#ifdef CONFIG_HVM case SH_type_fl1_32_shadow: case SH_type_fl1_pae_shadow: +#endif case SH_type_fl1_64_shadow: break; default: @@ -1746,6 +1754,7 @@ void sh_destroy_shadow(struct domain *d, mfn_t smfn) * small numbers that the compiler will enjoy */ switch ( t ) { +#ifdef CONFIG_HVM case SH_type_l1_32_shadow: case SH_type_fl1_32_shadow: SHADOW_INTERNAL_NAME(sh_destroy_l1_shadow, 2)(d, smfn); @@ -1761,6 +1770,7 @@ void sh_destroy_shadow(struct domain *d, mfn_t smfn) case SH_type_l2_pae_shadow: SHADOW_INTERNAL_NAME(sh_destroy_l2_shadow, 3)(d, smfn); break; +#endif case SH_type_l1_64_shadow: case SH_type_fl1_64_shadow: @@ -1811,12 +1821,14 @@ int sh_remove_write_access(struct domain *d, mfn_t gmfn, /* Dispatch table for getting per-type functions */ static const hash_domain_callback_t callbacks[SH_type_unused] = { NULL, /* none */ +#ifdef CONFIG_HVM SHADOW_INTERNAL_NAME(sh_rm_write_access_from_l1, 2), /* l1_32 */ SHADOW_INTERNAL_NAME(sh_rm_write_access_from_l1, 2), /* fl1_32 */ NULL, /* l2_32 */ SHADOW_INTERNAL_NAME(sh_rm_write_access_from_l1, 3), /* l1_pae */ SHADOW_INTERNAL_NAME(sh_rm_write_access_from_l1, 3), /* fl1_pae */ NULL, /* l2_pae */ +#endif SHADOW_INTERNAL_NAME(sh_rm_write_access_from_l1, 4), /* l1_64 */ SHADOW_INTERNAL_NAME(sh_rm_write_access_from_l1, 4), /* fl1_64 */ NULL, /* l2_64 */ @@ -2034,12 +2046,14 @@ int sh_remove_all_mappings(struct domain *d, mfn_t gmfn, gfn_t gfn) /* Dispatch table for getting per-type functions */ static const hash_domain_callback_t callbacks[SH_type_unused] = { NULL, /* none */ +#ifdef CONFIG_HVM SHADOW_INTERNAL_NAME(sh_rm_mappings_from_l1, 2), /* l1_32 */ SHADOW_INTERNAL_NAME(sh_rm_mappings_from_l1, 2), /* fl1_32 */ NULL, /* l2_32 */ SHADOW_INTERNAL_NAME(sh_rm_mappings_from_l1, 3), /* l1_pae */ SHADOW_INTERNAL_NAME(sh_rm_mappings_from_l1, 3), /* fl1_pae */ NULL, /* l2_pae */ +#endif SHADOW_INTERNAL_NAME(sh_rm_mappings_from_l1, 4), /* l1_64 */ SHADOW_INTERNAL_NAME(sh_rm_mappings_from_l1, 4), /* fl1_64 */ NULL, /* l2_64 */ @@ -2131,6 +2145,7 @@ static int sh_remove_shadow_via_pointer(struct domain *d, mfn_t smfn) /* Blank the offending entry */ switch (sp->u.sh.type) { +#ifdef CONFIG_HVM case SH_type_l1_32_shadow: case SH_type_l2_32_shadow: SHADOW_INTERNAL_NAME(sh_clear_shadow_entry, 2)(d, vaddr, pmfn); @@ -2139,6 +2154,7 @@ static int sh_remove_shadow_via_pointer(struct domain *d, mfn_t smfn) case SH_type_l2_pae_shadow: SHADOW_INTERNAL_NAME(sh_clear_shadow_entry, 3)(d, vaddr, pmfn); break; +#endif case SH_type_l1_64_shadow: case SH_type_l2_64_shadow: case SH_type_l2h_64_shadow: @@ -2175,12 +2191,14 @@ void sh_remove_shadows(struct domain *d, mfn_t gmfn, int fast, int all) * be called with the function to remove a lower-level shadow. */ static const hash_domain_callback_t callbacks[SH_type_unused] = { NULL, /* none */ +#ifdef CONFIG_HVM NULL, /* l1_32 */ NULL, /* fl1_32 */ SHADOW_INTERNAL_NAME(sh_remove_l1_shadow, 2), /* l2_32 */ NULL, /* l1_pae */ NULL, /* fl1_pae */ SHADOW_INTERNAL_NAME(sh_remove_l1_shadow, 3), /* l2_pae */ +#endif NULL, /* l1_64 */ NULL, /* fl1_64 */ SHADOW_INTERNAL_NAME(sh_remove_l1_shadow, 4), /* l2_64 */ @@ -2194,12 +2212,14 @@ void sh_remove_shadows(struct domain *d, mfn_t gmfn, int fast, int all) /* Another lookup table, for choosing which mask to use */ static const unsigned int masks[SH_type_unused] = { 0, /* none */ +#ifdef CONFIG_HVM SHF_L2_32, /* l1_32 */ 0, /* fl1_32 */ 0, /* l2_32 */ SHF_L2_PAE, /* l1_pae */ 0, /* fl1_pae */ 0, /* l2_pae */ +#endif SHF_L2H_64 | SHF_L2_64, /* l1_64 */ 0, /* fl1_64 */ SHF_L3_64, /* l2_64 */ @@ -2338,12 +2358,14 @@ void sh_reset_l3_up_pointers(struct vcpu *v) { static const hash_vcpu_callback_t callbacks[SH_type_unused] = { NULL, /* none */ +#ifdef CONFIG_HVM NULL, /* l1_32 */ NULL, /* fl1_32 */ NULL, /* l2_32 */ NULL, /* l1_pae */ NULL, /* fl1_pae */ NULL, /* l2_pae */ +#endif NULL, /* l1_64 */ NULL, /* fl1_64 */ NULL, /* l2_64 */ @@ -3362,12 +3384,14 @@ void shadow_audit_tables(struct vcpu *v) static const hash_vcpu_callback_t callbacks[SH_type_unused] = { NULL, /* none */ #if SHADOW_AUDIT & (SHADOW_AUDIT_ENTRIES | SHADOW_AUDIT_ENTRIES_FULL) +# ifdef CONFIG_HVM SHADOW_INTERNAL_NAME(sh_audit_l1_table, 2), /* l1_32 */ SHADOW_INTERNAL_NAME(sh_audit_fl1_table, 2), /* fl1_32 */ SHADOW_INTERNAL_NAME(sh_audit_l2_table, 2), /* l2_32 */ SHADOW_INTERNAL_NAME(sh_audit_l1_table, 3), /* l1_pae */ SHADOW_INTERNAL_NAME(sh_audit_fl1_table, 3), /* fl1_pae */ SHADOW_INTERNAL_NAME(sh_audit_l2_table, 3), /* l2_pae */ +# endif SHADOW_INTERNAL_NAME(sh_audit_l1_table, 4), /* l1_64 */ SHADOW_INTERNAL_NAME(sh_audit_fl1_table, 4), /* fl1_64 */ SHADOW_INTERNAL_NAME(sh_audit_l2_table, 4), /* l2_64 */ diff --git a/xen/arch/x86/mm/shadow/private.h b/xen/arch/x86/mm/shadow/private.h index 5b899399b6..35efb1b984 100644 --- a/xen/arch/x86/mm/shadow/private.h +++ b/xen/arch/x86/mm/shadow/private.h @@ -171,6 +171,7 @@ extern void shadow_audit_tables(struct vcpu *v); /* Shadow type codes */ #define SH_type_none (0U) /* on the shadow free list */ #define SH_type_min_shadow (1U) +#ifdef CONFIG_HVM #define SH_type_l1_32_shadow (1U) /* shadowing a 32-bit L1 guest page */ #define SH_type_fl1_32_shadow (2U) /* L1 shadow for a 32b 4M superpage */ #define SH_type_l2_32_shadow (3U) /* shadowing a 32-bit L2 guest page */ @@ -188,6 +189,25 @@ extern void shadow_audit_tables(struct vcpu *v); #define SH_type_monitor_table (14U) /* in use as a monitor table */ #define SH_type_oos_snapshot (15U) /* in use as OOS snapshot */ #define SH_type_unused (16U) +#else +#define SH_type_l1_32_shadow SH_type_unused +#define SH_type_fl1_32_shadow SH_type_unused +#define SH_type_l2_32_shadow SH_type_unused +#define SH_type_l1_pae_shadow SH_type_unused +#define SH_type_fl1_pae_shadow SH_type_unused +#define SH_type_l2_pae_shadow SH_type_unused +#define SH_type_l1_64_shadow 1U /* shadowing a 64-bit L1 page */ +#define SH_type_fl1_64_shadow 2U /* L1 shadow for 64-bit 2M superpg */ +#define SH_type_l2_64_shadow 3U /* shadowing a 64-bit L2 page */ +#define SH_type_l2h_64_shadow 4U /* shadowing a compat PAE L2 high page */ +#define SH_type_l3_64_shadow 5U /* shadowing a 64-bit L3 page */ +#define SH_type_l4_64_shadow 6U /* shadowing a 64-bit L4 page */ +#define SH_type_max_shadow 6U +#define SH_type_p2m_table 7U /* in use as the p2m table */ +#define SH_type_monitor_table 8U /* in use as a monitor table */ +#define SH_type_oos_snapshot 9U /* in use as OOS snapshot */ +#define SH_type_unused 10U +#endif /* * What counts as a pinnable shadow? @@ -248,12 +268,21 @@ static inline void sh_terminate_list(struct page_list_head *tmp_list) (((1u << (SH_type_max_shadow + 1u)) - 1u) - \ ((1u << SH_type_min_shadow) - 1u)) +#ifdef CONFIG_HVM #define SHF_L1_32 (1u << SH_type_l1_32_shadow) #define SHF_FL1_32 (1u << SH_type_fl1_32_shadow) #define SHF_L2_32 (1u << SH_type_l2_32_shadow) #define SHF_L1_PAE (1u << SH_type_l1_pae_shadow) #define SHF_FL1_PAE (1u << SH_type_fl1_pae_shadow) #define SHF_L2_PAE (1u << SH_type_l2_pae_shadow) +#else +#define SHF_L1_32 0 +#define SHF_FL1_32 0 +#define SHF_L2_32 0 +#define SHF_L1_PAE 0 +#define SHF_FL1_PAE 0 +#define SHF_L2_PAE 0 +#endif #define SHF_L1_64 (1u << SH_type_l1_64_shadow) #define SHF_FL1_64 (1u << SH_type_fl1_64_shadow) #define SHF_L2_64 (1u << SH_type_l2_64_shadow) @@ -329,8 +358,13 @@ extern const u8 sh_type_to_size[SH_type_unused]; static inline unsigned int shadow_size(unsigned int shadow_type) { +#ifdef CONFIG_HVM ASSERT(shadow_type < ARRAY_SIZE(sh_type_to_size)); return sh_type_to_size[shadow_type]; +#else + ASSERT(shadow_type < SH_type_unused); + return 1; +#endif } /****************************************************************************** @@ -488,8 +522,10 @@ static inline unsigned long __backpointer(const struct page_info *sp) { switch (sp->u.sh.type) { +#ifdef CONFIG_HVM case SH_type_fl1_32_shadow: case SH_type_fl1_pae_shadow: +#endif case SH_type_fl1_64_shadow: return sp->v.sh.back; } -- generated by git-patchbot for /home/xen/git/xen.git#master
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |