|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v2 06/30] x86/paging: introduce paging_set_allocation
On 27/09/16 16:57, Roger Pau Monne wrote:
> ... and remove hap_set_alloc_for_pvh_dom0.
>
> Signed-off-by: Roger Pau Monné <roger.pau@xxxxxxxxxx>
> Acked-by: Tim Deegan <tim@xxxxxxx>
Acked-by: George Dunlap <george.dunlap@xxxxxxxxxx>
> ---
> Cc: Jan Beulich <jbeulich@xxxxxxxx>
> Cc: Andrew Cooper <andrew.cooper3@xxxxxxxxxx>
> Cc: George Dunlap <george.dunlap@xxxxxxxxxxxxx>
> Cc: Tim Deegan <tim@xxxxxxx>
> ---
> Changes since RFC:
> - Make paging_set_allocation preemtable.
> - Move comments.
> ---
> xen/arch/x86/domain_build.c | 17 +++++++++++++----
> xen/arch/x86/mm/hap/hap.c | 14 +-------------
> xen/arch/x86/mm/paging.c | 16 ++++++++++++++++
> xen/arch/x86/mm/shadow/common.c | 7 +------
> xen/include/asm-x86/hap.h | 2 +-
> xen/include/asm-x86/paging.h | 7 +++++++
> xen/include/asm-x86/shadow.h | 8 ++++++++
> 7 files changed, 47 insertions(+), 24 deletions(-)
>
> diff --git a/xen/arch/x86/domain_build.c b/xen/arch/x86/domain_build.c
> index 0a02d65..04d6cb0 100644
> --- a/xen/arch/x86/domain_build.c
> +++ b/xen/arch/x86/domain_build.c
> @@ -35,7 +35,6 @@
> #include <asm/setup.h>
> #include <asm/bzimage.h> /* for bzimage_parse */
> #include <asm/io_apic.h>
> -#include <asm/hap.h>
> #include <asm/hpet.h>
>
> #include <public/version.h>
> @@ -1383,15 +1382,25 @@ int __init construct_dom0(
> nr_pages);
> }
>
> - if ( is_pvh_domain(d) )
> - hap_set_alloc_for_pvh_dom0(d, dom0_paging_pages(d, nr_pages));
> -
> /*
> * We enable paging mode again so guest_physmap_add_page will do the
> * right thing for us.
> */
> d->arch.paging.mode = save_pvh_pg_mode;
>
> + if ( is_pvh_domain(d) )
> + {
> + int preempted;
> +
> + do {
> + preempted = 0;
> + paging_set_allocation(d, dom0_paging_pages(d, nr_pages),
> + &preempted);
> + process_pending_softirqs();
> + } while ( preempted );
> + }
> +
> +
> /* Write the phys->machine and machine->phys table entries. */
> for ( pfn = 0; pfn < count; pfn++ )
> {
> diff --git a/xen/arch/x86/mm/hap/hap.c b/xen/arch/x86/mm/hap/hap.c
> index 2dc82f5..4420e4e 100644
> --- a/xen/arch/x86/mm/hap/hap.c
> +++ b/xen/arch/x86/mm/hap/hap.c
> @@ -334,7 +334,7 @@ hap_get_allocation(struct domain *d)
>
> /* Set the pool of pages to the required number of pages.
> * Returns 0 for success, non-zero for failure. */
> -static int
> +int
> hap_set_allocation(struct domain *d, unsigned long pages, int *preempted)
> {
> struct page_info *pg;
> @@ -640,18 +640,6 @@ int hap_domctl(struct domain *d, xen_domctl_shadow_op_t
> *sc,
> }
> }
>
> -void __init hap_set_alloc_for_pvh_dom0(struct domain *d,
> - unsigned long hap_pages)
> -{
> - int rc;
> -
> - paging_lock(d);
> - rc = hap_set_allocation(d, hap_pages, NULL);
> - paging_unlock(d);
> -
> - BUG_ON(rc);
> -}
> -
> static const struct paging_mode hap_paging_real_mode;
> static const struct paging_mode hap_paging_protected_mode;
> static const struct paging_mode hap_paging_pae_mode;
> diff --git a/xen/arch/x86/mm/paging.c b/xen/arch/x86/mm/paging.c
> index cc44682..2717bd3 100644
> --- a/xen/arch/x86/mm/paging.c
> +++ b/xen/arch/x86/mm/paging.c
> @@ -954,6 +954,22 @@ void paging_write_p2m_entry(struct p2m_domain *p2m,
> unsigned long gfn,
> safe_write_pte(p, new);
> }
>
> +int paging_set_allocation(struct domain *d, unsigned long pages, int
> *preempted)
> +{
> + int rc;
> +
> + ASSERT(paging_mode_enabled(d));
> +
> + paging_lock(d);
> + if ( hap_enabled(d) )
> + rc = hap_set_allocation(d, pages, preempted);
> + else
> + rc = sh_set_allocation(d, pages, preempted);
> + paging_unlock(d);
> +
> + return rc;
> +}
> +
> /*
> * Local variables:
> * mode: C
> diff --git a/xen/arch/x86/mm/shadow/common.c b/xen/arch/x86/mm/shadow/common.c
> index d3cc2cc..53ffe1a 100644
> --- a/xen/arch/x86/mm/shadow/common.c
> +++ b/xen/arch/x86/mm/shadow/common.c
> @@ -1609,12 +1609,7 @@ shadow_free_p2m_page(struct domain *d, struct
> page_info *pg)
> paging_unlock(d);
> }
>
> -/* Set the pool of shadow pages to the required number of pages.
> - * Input will be rounded up to at least shadow_min_acceptable_pages(),
> - * plus space for the p2m table.
> - * Returns 0 for success, non-zero for failure. */
> -static int sh_set_allocation(struct domain *d, unsigned long pages,
> - int *preempted)
> +int sh_set_allocation(struct domain *d, unsigned long pages, int *preempted)
> {
> struct page_info *sp;
> unsigned int lower_bound;
> diff --git a/xen/include/asm-x86/hap.h b/xen/include/asm-x86/hap.h
> index c613836..9d59430 100644
> --- a/xen/include/asm-x86/hap.h
> +++ b/xen/include/asm-x86/hap.h
> @@ -46,7 +46,7 @@ int hap_track_dirty_vram(struct domain *d,
> XEN_GUEST_HANDLE_64(uint8) dirty_bitmap);
>
> extern const struct paging_mode *hap_paging_get_mode(struct vcpu *);
> -void hap_set_alloc_for_pvh_dom0(struct domain *d, unsigned long num_pages);
> +int hap_set_allocation(struct domain *d, unsigned long pages, int
> *preempted);
>
> #endif /* XEN_HAP_H */
>
> diff --git a/xen/include/asm-x86/paging.h b/xen/include/asm-x86/paging.h
> index 56eef6b..c2d60d3 100644
> --- a/xen/include/asm-x86/paging.h
> +++ b/xen/include/asm-x86/paging.h
> @@ -347,6 +347,13 @@ void pagetable_dying(struct domain *d, paddr_t gpa);
> void paging_dump_domain_info(struct domain *d);
> void paging_dump_vcpu_info(struct vcpu *v);
>
> +/* Set the pool of shadow pages to the required number of pages.
> + * Input might be rounded up to at minimum amount of pages, plus
> + * space for the p2m table.
> + * Returns 0 for success, non-zero for failure. */
> +int paging_set_allocation(struct domain *d, unsigned long pages,
> + int *preempted);
> +
> #endif /* XEN_PAGING_H */
>
> /*
> diff --git a/xen/include/asm-x86/shadow.h b/xen/include/asm-x86/shadow.h
> index 6d0aefb..f0e2227 100644
> --- a/xen/include/asm-x86/shadow.h
> +++ b/xen/include/asm-x86/shadow.h
> @@ -83,6 +83,12 @@ void sh_remove_shadows(struct domain *d, mfn_t gmfn, int
> fast, int all);
> /* Discard _all_ mappings from the domain's shadows. */
> void shadow_blow_tables_per_domain(struct domain *d);
>
> +/* Set the pool of shadow pages to the required number of pages.
> + * Input will be rounded up to at least shadow_min_acceptable_pages(),
> + * plus space for the p2m table.
> + * Returns 0 for success, non-zero for failure. */
> +int sh_set_allocation(struct domain *d, unsigned long pages, int *preempted);
> +
> #else /* !CONFIG_SHADOW_PAGING */
>
> #define shadow_teardown(d, p) ASSERT(is_pv_domain(d))
> @@ -91,6 +97,8 @@ void shadow_blow_tables_per_domain(struct domain *d);
> ({ ASSERT(is_pv_domain(d)); -EOPNOTSUPP; })
> #define shadow_track_dirty_vram(d, begin_pfn, nr, bitmap) \
> ({ ASSERT_UNREACHABLE(); -EOPNOTSUPP; })
> +#define sh_set_allocation(d, pages, preempted) \
> + ({ ASSERT_UNREACHABLE(); -EOPNOTSUPP; })
>
> static inline void sh_remove_shadows(struct domain *d, mfn_t gmfn,
> bool_t fast, bool_t all) {}
>
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
https://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |