[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH v6 3/5] xen/arm: Implement virtual-linear page table for guest
Implement vlpt (virtual-linear page table) for fast accessing of 3rd PTE of guest p2m. When creating a mapping for vlpt, just copy the 1st level PTE of guest p2m to the xen's 2nd level PTE. Then the mapping becomes the following: xen's 1st PTE --> xen's 2nd PTE (which is the same as 1st PTE of guest p2m) --> guest p2m's 2nd PTE --> guest p2m's 3rd PTE (the memory contents where the vlpt points) For more info about vlpt, see: http://www.technovelty.org/linux/virtual-linear-page-table.html This function is used in dirty-page tracing: when domU write-fault is trapped by xen, xen can immediately locate the 3rd PTE of guest p2m. The following link shows the performance comparison for handling a dirty-page between vlpt and typical page table walking. http://lists.xen.org/archives/html/xen-devel/2013-08/msg01503.html Signed-off-by: Jaeyong Yoo <jaeyong.yoo@xxxxxxxxxxx> Signed-off-by: Junghyun Yoo <yjhyun.yoo@xxxxxxxxxxx> --- xen/arch/arm/domain.c | 14 +++++++ xen/arch/arm/mm.c | 83 ++++++++++++++++++++++++++++++++++++++++ xen/include/asm-arm/arm32/page.h | 36 +++++++++++++---- xen/include/asm-arm/config.h | 18 +++++++-- xen/include/asm-arm/domain.h | 9 +++++ xen/include/asm-arm/mm.h | 24 ++++++++++++ 6 files changed, 173 insertions(+), 11 deletions(-) diff --git a/xen/arch/arm/domain.c b/xen/arch/arm/domain.c index 2ae6941..4978765 100644 --- a/xen/arch/arm/domain.c +++ b/xen/arch/arm/domain.c @@ -208,6 +208,12 @@ static void ctxt_switch_to(struct vcpu *n) isb(); + /* for dirty-page tracing + * XXX: how do we consider SMP case? + */ + if ( n->domain->arch.dirty.mode ) + restore_vlpt(n->domain); + /* This is could trigger an hardware interrupt from the virtual * timer. The interrupt needs to be injected into the guest. */ WRITE_SYSREG32(n->arch.cntkctl, CNTKCTL_EL1); @@ -504,6 +510,14 @@ int arch_domain_create(struct domain *d, unsigned int domcr_flags) /* Default the virtual ID to match the physical */ d->arch.vpidr = boot_cpu_data.midr.bits; + /* init for dirty-page tracing */ + d->arch.dirty.mode = 0; + + d->arch.dirty.p2m_start_idx = 0; + d->arch.dirty.p2m_end_idx = 0; + d->arch.dirty.p2m_first[0] = NULL; + d->arch.dirty.p2m_first[1] = NULL; + clear_page(d->shared_info); share_xen_page_with_guest( virt_to_page(d->shared_info), d, XENSHARE_writable); diff --git a/xen/arch/arm/mm.c b/xen/arch/arm/mm.c index 7e8e06a..0fc9d9a 100644 --- a/xen/arch/arm/mm.c +++ b/xen/arch/arm/mm.c @@ -1245,6 +1245,89 @@ void clear_and_clean_page(struct page_info *page) unmap_domain_page(p); } +paddr_t get_gma_end(struct domain *d) +{ + return GUEST_RAM_BASE + ((paddr_t) d->max_pages << PAGE_SHIFT); +} + +/* restore the xen page table for vlpt mapping for domain d */ +void restore_vlpt(struct domain *d) +{ + int vlpt_base_idx; + int i; + lpae_t *map = this_cpu(xen_dommap); + + vlpt_base_idx = VLPT_VA_TO_IDX(VIRT_LIN_P2M_START); + dsb(sy); + for ( i = d->arch.dirty.p2m_start_idx; i < d->arch.dirty.p2m_end_idx; ++i ) + { + uint32_t in = INNER_IDX(i, LPAE_ENTRIES); + uint32_t out = OUTER_IDX(i, LPAE_ENTRIES); + if ( map[vlpt_base_idx+i].bits != + d->arch.dirty.p2m_first[out][in].bits ) + { + write_pte(&map[vlpt_base_idx+i], + d->arch.dirty.p2m_first[out][in]); + __flush_xen_data_tlb_range_va(VLPT_IDX_TO_VA(vlpt_base_idx+i), + VLPT_SIZE); + } + } + dsb(sy); + isb(); +} + +/* setting up the xen page table for vlpt mapping for domain d */ +int prepare_vlpt(struct domain *d) +{ + int vlpt_base_idx; + int p2m_start_index, p2m_end_index; + struct p2m_domain *p2m = &d->arch.p2m; + paddr_t gma_start = GUEST_RAM_BASE; + paddr_t gma_end = 0; + uint64_t required, avail = VIRT_LIN_P2M_END - VIRT_LIN_P2M_START; + + gma_end = get_gma_end(d); + required = ((gma_end - gma_start) >> PAGE_SHIFT) * (sizeof(lpae_t)); + + vlpt_base_idx = VLPT_VA_TO_IDX(VIRT_LIN_P2M_START); + + if ( required > avail ) + { + dprintk(XENLOG_ERR, "Available VLPT is small for domU guest" + "(avail: %llx, required: %llx)\n", + avail, required); + return -ENOMEM; + } + + p2m_start_index = gma_start >> FIRST_SHIFT; + p2m_end_index = (gma_end >> FIRST_SHIFT) + 1; + + if ( vlpt_base_idx + p2m_end_index >= LPAE_ENTRIES * 2 ) + { + dprintk(XENLOG_ERR, "xen second page is small for VLPT for domU"); + return -ENOMEM; + } + + /* First level p2m is 2 consecutive pages */ + d->arch.dirty.p2m_first[0] = map_domain_page_global( + page_to_mfn(p2m->first_level) ); + d->arch.dirty.p2m_first[1] = map_domain_page_global( + page_to_mfn(p2m->first_level+1) ); + + /* storing the start and end index of guest p2m first*/ + d->arch.dirty.p2m_start_idx = p2m_start_index; + d->arch.dirty.p2m_end_idx = p2m_end_index; + + return 0; +} + +void cleanup_vlpt(struct domain *d) +{ + /* First level p2m is 2 consecutive pages */ + unmap_domain_page_global(d->arch.dirty.p2m_first[0]); + unmap_domain_page_global(d->arch.dirty.p2m_first[1]); +} + /* * Local variables: * mode: C diff --git a/xen/include/asm-arm/arm32/page.h b/xen/include/asm-arm/arm32/page.h index 4abb281..7377557 100644 --- a/xen/include/asm-arm/arm32/page.h +++ b/xen/include/asm-arm/arm32/page.h @@ -3,22 +3,31 @@ #ifndef __ASSEMBLY__ -/* Write a pagetable entry. +/* + * Write a pagetable entry. * - * If the table entry is changing a text mapping, it is responsibility - * of the caller to issue an ISB after write_pte. + * All necessary barriers are responsibility of the caller */ -static inline void write_pte(lpae_t *p, lpae_t pte) +static inline void __write_pte(lpae_t *p, lpae_t pte) { asm volatile ( - /* Ensure any writes have completed with the old mappings. */ - "dsb;" /* Safely write the entry (STRD is atomic on CPUs that support LPAE) */ "strd %0, %H0, [%1];" - "dsb;" : : "r" (pte.bits), "r" (p) : "memory"); } +/* Write a pagetable entry. + * + * If the table entry is changing a text mapping, it is responsibility + * of the caller to issue an ISB after write_pte. + */ +static inline void write_pte(lpae_t *p, lpae_t pte) +{ + dsb(); + __write_pte(p, pte); + dsb(); +} + /* Inline ASM to flush dcache on register R (may be an inline asm operand) */ #define __clean_xen_dcache_one(R) STORE_CP32(R, DCCMVAC) @@ -74,6 +83,19 @@ static inline void __flush_xen_data_tlb_one(vaddr_t va) asm volatile(STORE_CP32(0, TLBIMVAHIS) : : "r" (va) : "memory"); } +/* + * Flush a range of VA's hypervisor mappings from the data TLB. + * All necessary barriers are responsibility of the caller + */ +static inline void __flush_xen_data_tlb_range_va(unsigned long va, unsigned long size) +{ + unsigned long end = va + size; + while ( va < end ) { + __flush_xen_data_tlb_one(va); + va += PAGE_SIZE; + } +} + /* Ask the MMU to translate a VA for us */ static inline uint64_t __va_to_par(vaddr_t va) { diff --git a/xen/include/asm-arm/config.h b/xen/include/asm-arm/config.h index 1c3abcf..337bb7c 100644 --- a/xen/include/asm-arm/config.h +++ b/xen/include/asm-arm/config.h @@ -93,7 +93,8 @@ * space * * 1G - 2G Xenheap: always-mapped memory - * 2G - 4G Domheap: on-demand-mapped + * 2G - 3G 896M Domheap: on-demand-mapped + * 3G 896M - 4G Virtual-linear mapping to P2M table (128MB size) * * ARM64 layout: * 0x0000000000000000 - 0x0000007fffffffff (512GB, L0 slot [0]) @@ -130,14 +131,21 @@ #define XENHEAP_VIRT_START _AT(vaddr_t,0x40000000) #define XENHEAP_VIRT_END _AT(vaddr_t,0x7fffffff) #define DOMHEAP_VIRT_START _AT(vaddr_t,0x80000000) -#define DOMHEAP_VIRT_END _AT(vaddr_t,0xffffffff) +#define DOMHEAP_VIRT_END _AT(vaddr_t,0xe7ffffff) +#define VIRT_LIN_P2M_START _AT(vaddr_t,0xe8000000) +#define VIRT_LIN_P2M_END _AT(vaddr_t,0xffffffff) #define VMAP_VIRT_END XENHEAP_VIRT_START -#define DOMHEAP_ENTRIES 1024 /* 1024 2MB mapping slots */ +/* 1024 2MB mapping slots */ +#define DOMHEAP_ENTRIES ((DOMHEAP_VIRT_END - DOMHEAP_VIRT_START+1) >> SECOND_SHIFT) /* Number of domheap pagetable pages required at the second level (2MB mappings) */ -#define DOMHEAP_SECOND_PAGES ((DOMHEAP_VIRT_END - DOMHEAP_VIRT_START + 1) >> FIRST_SHIFT) +/* We piggy-back VLPT areas in the back of domain heap, so the one that actually + * allocated for domheap table should be domheap size + vlpt size. */ +#define __DOMHEAP_SIZE (DOMHEAP_VIRT_END - DOMHEAP_VIRT_START + 1) +#define __VLPT_SIZE (VIRT_LIN_P2M_END - VIRT_LIN_P2M_START + 1) +#define DOMHEAP_SECOND_PAGES ((__DOMHEAP_SIZE + __VLPT_SIZE) >> FIRST_SHIFT) #else /* ARM_64 */ @@ -159,6 +167,8 @@ #define HYPERVISOR_VIRT_END DIRECTMAP_VIRT_END +/*TODO (ARM_64): define VIRT_LIN_P2M_START VIRT_LIN_P2M_END */ + #endif /* Fixmap slots */ diff --git a/xen/include/asm-arm/domain.h b/xen/include/asm-arm/domain.h index b296923..9674175 100644 --- a/xen/include/asm-arm/domain.h +++ b/xen/include/asm-arm/domain.h @@ -164,6 +164,15 @@ struct arch_domain } vuart; unsigned int evtchn_irq; + + /* dirty-page tracing */ + struct { + int mode; /* 1 if dirty pages tracing enabled */ + lpae_t *p2m_first[2]; /* copy of guest p2m's first */ + int p2m_start_idx; /* start index of p2m_first */ + int p2m_end_idx; /* end index of p2m_first */ + } dirty; + } __cacheline_aligned; struct arch_vcpu diff --git a/xen/include/asm-arm/mm.h b/xen/include/asm-arm/mm.h index 3bef93f..7ceb568 100644 --- a/xen/include/asm-arm/mm.h +++ b/xen/include/asm-arm/mm.h @@ -344,6 +344,30 @@ static inline void put_page_and_type(struct page_info *page) void clear_and_clean_page(struct page_info *page); +/* routine for dirty-page tracing */ +#define VLPT_SIZE (1 << SECOND_SHIFT) +#define VLPT_VA_TO_IDX(va) ((va - DOMHEAP_VIRT_START) >> SECOND_SHIFT) +#define VLPT_IDX_TO_VA(idx) (DOMHEAP_VIRT_START + ((idx) << SECOND_SHIFT)) +#define INNER_IDX(i, s) ((i) % (s)) +#define OUTER_IDX(i, s) ((uint32_t)(i) / (uint32_t)(s)) + +paddr_t get_gma_end(struct domain *d); +int prepare_vlpt(struct domain *d); +void cleanup_vlpt(struct domain *d); +void restore_vlpt(struct domain *d); + +/* calculate the xen's virtual address for accessing the leaf PTE of + * a given address (GPA) */ +static inline lpae_t * get_vlpt_3lvl_pte(paddr_t addr) +{ + lpae_t *table = (lpae_t *)VIRT_LIN_P2M_START; + + /* Since we slotted the guest's first p2m page table to xen's + * second page table, one shift is enough for calculating the + * index of guest p2m table entry */ + return &table[addr >> PAGE_SHIFT]; +} + #endif /* __ARCH_ARM_MM__ */ /* * Local variables: -- 1.8.1.2 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |