[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v2 06/30] x86/paging: introduce paging_set_allocation
On 27/09/16 16:57, Roger Pau Monne wrote: > ... and remove hap_set_alloc_for_pvh_dom0. > > Signed-off-by: Roger Pau Monné <roger.pau@xxxxxxxxxx> > Acked-by: Tim Deegan <tim@xxxxxxx> Acked-by: George Dunlap <george.dunlap@xxxxxxxxxx> > --- > Cc: Jan Beulich <jbeulich@xxxxxxxx> > Cc: Andrew Cooper <andrew.cooper3@xxxxxxxxxx> > Cc: George Dunlap <george.dunlap@xxxxxxxxxxxxx> > Cc: Tim Deegan <tim@xxxxxxx> > --- > Changes since RFC: > - Make paging_set_allocation preemtable. > - Move comments. > --- > xen/arch/x86/domain_build.c | 17 +++++++++++++---- > xen/arch/x86/mm/hap/hap.c | 14 +------------- > xen/arch/x86/mm/paging.c | 16 ++++++++++++++++ > xen/arch/x86/mm/shadow/common.c | 7 +------ > xen/include/asm-x86/hap.h | 2 +- > xen/include/asm-x86/paging.h | 7 +++++++ > xen/include/asm-x86/shadow.h | 8 ++++++++ > 7 files changed, 47 insertions(+), 24 deletions(-) > > diff --git a/xen/arch/x86/domain_build.c b/xen/arch/x86/domain_build.c > index 0a02d65..04d6cb0 100644 > --- a/xen/arch/x86/domain_build.c > +++ b/xen/arch/x86/domain_build.c > @@ -35,7 +35,6 @@ > #include <asm/setup.h> > #include <asm/bzimage.h> /* for bzimage_parse */ > #include <asm/io_apic.h> > -#include <asm/hap.h> > #include <asm/hpet.h> > > #include <public/version.h> > @@ -1383,15 +1382,25 @@ int __init construct_dom0( > nr_pages); > } > > - if ( is_pvh_domain(d) ) > - hap_set_alloc_for_pvh_dom0(d, dom0_paging_pages(d, nr_pages)); > - > /* > * We enable paging mode again so guest_physmap_add_page will do the > * right thing for us. > */ > d->arch.paging.mode = save_pvh_pg_mode; > > + if ( is_pvh_domain(d) ) > + { > + int preempted; > + > + do { > + preempted = 0; > + paging_set_allocation(d, dom0_paging_pages(d, nr_pages), > + &preempted); > + process_pending_softirqs(); > + } while ( preempted ); > + } > + > + > /* Write the phys->machine and machine->phys table entries. */ > for ( pfn = 0; pfn < count; pfn++ ) > { > diff --git a/xen/arch/x86/mm/hap/hap.c b/xen/arch/x86/mm/hap/hap.c > index 2dc82f5..4420e4e 100644 > --- a/xen/arch/x86/mm/hap/hap.c > +++ b/xen/arch/x86/mm/hap/hap.c > @@ -334,7 +334,7 @@ hap_get_allocation(struct domain *d) > > /* Set the pool of pages to the required number of pages. > * Returns 0 for success, non-zero for failure. */ > -static int > +int > hap_set_allocation(struct domain *d, unsigned long pages, int *preempted) > { > struct page_info *pg; > @@ -640,18 +640,6 @@ int hap_domctl(struct domain *d, xen_domctl_shadow_op_t > *sc, > } > } > > -void __init hap_set_alloc_for_pvh_dom0(struct domain *d, > - unsigned long hap_pages) > -{ > - int rc; > - > - paging_lock(d); > - rc = hap_set_allocation(d, hap_pages, NULL); > - paging_unlock(d); > - > - BUG_ON(rc); > -} > - > static const struct paging_mode hap_paging_real_mode; > static const struct paging_mode hap_paging_protected_mode; > static const struct paging_mode hap_paging_pae_mode; > diff --git a/xen/arch/x86/mm/paging.c b/xen/arch/x86/mm/paging.c > index cc44682..2717bd3 100644 > --- a/xen/arch/x86/mm/paging.c > +++ b/xen/arch/x86/mm/paging.c > @@ -954,6 +954,22 @@ void paging_write_p2m_entry(struct p2m_domain *p2m, > unsigned long gfn, > safe_write_pte(p, new); > } > > +int paging_set_allocation(struct domain *d, unsigned long pages, int > *preempted) > +{ > + int rc; > + > + ASSERT(paging_mode_enabled(d)); > + > + paging_lock(d); > + if ( hap_enabled(d) ) > + rc = hap_set_allocation(d, pages, preempted); > + else > + rc = sh_set_allocation(d, pages, preempted); > + paging_unlock(d); > + > + return rc; > +} > + > /* > * Local variables: > * mode: C > diff --git a/xen/arch/x86/mm/shadow/common.c b/xen/arch/x86/mm/shadow/common.c > index d3cc2cc..53ffe1a 100644 > --- a/xen/arch/x86/mm/shadow/common.c > +++ b/xen/arch/x86/mm/shadow/common.c > @@ -1609,12 +1609,7 @@ shadow_free_p2m_page(struct domain *d, struct > page_info *pg) > paging_unlock(d); > } > > -/* Set the pool of shadow pages to the required number of pages. > - * Input will be rounded up to at least shadow_min_acceptable_pages(), > - * plus space for the p2m table. > - * Returns 0 for success, non-zero for failure. */ > -static int sh_set_allocation(struct domain *d, unsigned long pages, > - int *preempted) > +int sh_set_allocation(struct domain *d, unsigned long pages, int *preempted) > { > struct page_info *sp; > unsigned int lower_bound; > diff --git a/xen/include/asm-x86/hap.h b/xen/include/asm-x86/hap.h > index c613836..9d59430 100644 > --- a/xen/include/asm-x86/hap.h > +++ b/xen/include/asm-x86/hap.h > @@ -46,7 +46,7 @@ int hap_track_dirty_vram(struct domain *d, > XEN_GUEST_HANDLE_64(uint8) dirty_bitmap); > > extern const struct paging_mode *hap_paging_get_mode(struct vcpu *); > -void hap_set_alloc_for_pvh_dom0(struct domain *d, unsigned long num_pages); > +int hap_set_allocation(struct domain *d, unsigned long pages, int > *preempted); > > #endif /* XEN_HAP_H */ > > diff --git a/xen/include/asm-x86/paging.h b/xen/include/asm-x86/paging.h > index 56eef6b..c2d60d3 100644 > --- a/xen/include/asm-x86/paging.h > +++ b/xen/include/asm-x86/paging.h > @@ -347,6 +347,13 @@ void pagetable_dying(struct domain *d, paddr_t gpa); > void paging_dump_domain_info(struct domain *d); > void paging_dump_vcpu_info(struct vcpu *v); > > +/* Set the pool of shadow pages to the required number of pages. > + * Input might be rounded up to at minimum amount of pages, plus > + * space for the p2m table. > + * Returns 0 for success, non-zero for failure. */ > +int paging_set_allocation(struct domain *d, unsigned long pages, > + int *preempted); > + > #endif /* XEN_PAGING_H */ > > /* > diff --git a/xen/include/asm-x86/shadow.h b/xen/include/asm-x86/shadow.h > index 6d0aefb..f0e2227 100644 > --- a/xen/include/asm-x86/shadow.h > +++ b/xen/include/asm-x86/shadow.h > @@ -83,6 +83,12 @@ void sh_remove_shadows(struct domain *d, mfn_t gmfn, int > fast, int all); > /* Discard _all_ mappings from the domain's shadows. */ > void shadow_blow_tables_per_domain(struct domain *d); > > +/* Set the pool of shadow pages to the required number of pages. > + * Input will be rounded up to at least shadow_min_acceptable_pages(), > + * plus space for the p2m table. > + * Returns 0 for success, non-zero for failure. */ > +int sh_set_allocation(struct domain *d, unsigned long pages, int *preempted); > + > #else /* !CONFIG_SHADOW_PAGING */ > > #define shadow_teardown(d, p) ASSERT(is_pv_domain(d)) > @@ -91,6 +97,8 @@ void shadow_blow_tables_per_domain(struct domain *d); > ({ ASSERT(is_pv_domain(d)); -EOPNOTSUPP; }) > #define shadow_track_dirty_vram(d, begin_pfn, nr, bitmap) \ > ({ ASSERT_UNREACHABLE(); -EOPNOTSUPP; }) > +#define sh_set_allocation(d, pages, preempted) \ > + ({ ASSERT_UNREACHABLE(); -EOPNOTSUPP; }) > > static inline void sh_remove_shadows(struct domain *d, mfn_t gmfn, > bool_t fast, bool_t all) {} > _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx https://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |