[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] [PATCHv4 18/33] x86/xen: convert __xen_pgd_walk() and xen_cleanmfnmap() to support p4d



> +static int xen_p4d_walk(struct mm_struct *mm, p4d_t *p4d,
> +             int (*func)(struct mm_struct *mm, struct page *, enum pt_level),
> +             bool last, unsigned long limit)
> +{
> +     int i, nr, flush = 0;
> +
> +     nr = last ? p4d_index(limit) + 1 : PTRS_PER_P4D;
> +     for (i = 0; i < nr; i++) {
> +             pud_t *pud;
> +
> +             if (p4d_none(p4d[i]))
> +                     continue;
> +
> +             pud = pud_offset(&p4d[i], 0);
> +             if (PTRS_PER_PUD > 1)
> +                     flush |= (*func)(mm, virt_to_page(pud), PT_PUD);
> +             xen_pud_walk(mm, pud, func, last && i == nr - 1, limit);
> +     }
> +     return flush;
> +}

..

> +             p4d = p4d_offset(&pgd[i], 0);
> +             if (PTRS_PER_P4D > 1)
> +                     flush |= (*func)(mm, virt_to_page(p4d), PT_P4D);
> +             xen_p4d_walk(mm, p4d, func, i == nr - 1, limit);


We are losing flush status at all levels so we need something like

flush |= xen_XXX_walk(...)



>       }
>  
> -out:
>       /* Do the top level last, so that the callbacks can use it as
>          a cue to do final things like tlb flushes. */
>       flush |= (*func)(mm, virt_to_page(pgd), PT_PGD);
> @@ -1150,57 +1161,97 @@ static void __init xen_cleanmfnmap_free_pgtbl(void 
> *pgtbl, bool unpin)
>       xen_free_ro_pages(pa, PAGE_SIZE);
>  }
>  
> +static void __init xen_cleanmfnmap_pmd(pmd_t *pmd, bool unpin)
> +{
> +     unsigned long pa;
> +     pte_t *pte_tbl;
> +     int i;
> +
> +     if (pmd_large(*pmd)) {
> +             pa = pmd_val(*pmd) & PHYSICAL_PAGE_MASK;
> +             xen_free_ro_pages(pa, PMD_SIZE);
> +             return;
> +     }
> +
> +     pte_tbl = pte_offset_kernel(pmd, 0);
> +     for (i = 0; i < PTRS_PER_PTE; i++) {
> +             if (pte_none(pte_tbl[i]))
> +                     continue;
> +             pa = pte_pfn(pte_tbl[i]) << PAGE_SHIFT;
> +             xen_free_ro_pages(pa, PAGE_SIZE);
> +     }
> +     set_pmd(pmd, __pmd(0));
> +     xen_cleanmfnmap_free_pgtbl(pte_tbl, unpin);
> +}
> +
> +static void __init xen_cleanmfnmap_pud(pud_t *pud, bool unpin)
> +{
> +     unsigned long pa;
> +     pmd_t *pmd_tbl;
> +     int i;
> +
> +     if (pud_large(*pud)) {
> +             pa = pud_val(*pud) & PHYSICAL_PAGE_MASK;
> +             xen_free_ro_pages(pa, PUD_SIZE);
> +             return;
> +     }
> +
> +     pmd_tbl = pmd_offset(pud, 0);
> +     for (i = 0; i < PTRS_PER_PMD; i++) {
> +             if (pmd_none(pmd_tbl[i]))
> +                     continue;
> +             xen_cleanmfnmap_pmd(pmd_tbl + i, unpin);
> +     }
> +     set_pud(pud, __pud(0));
> +     xen_cleanmfnmap_free_pgtbl(pmd_tbl, unpin);
> +}
> +
> +static void __init xen_cleanmfnmap_p4d(p4d_t *p4d, bool unpin)
> +{
> +     unsigned long pa;
> +     pud_t *pud_tbl;
> +     int i;
> +
> +     if (p4d_large(*p4d)) {
> +             pa = p4d_val(*p4d) & PHYSICAL_PAGE_MASK;
> +             xen_free_ro_pages(pa, P4D_SIZE);
> +             return;
> +     }
> +
> +     pud_tbl = pud_offset(p4d, 0);
> +     for (i = 0; i < PTRS_PER_PUD; i++) {
> +             if (pud_none(pud_tbl[i]))
> +                     continue;
> +             xen_cleanmfnmap_pud(pud_tbl + i, unpin);
> +     }
> +     set_p4d(p4d, __p4d(0));
> +     xen_cleanmfnmap_free_pgtbl(pud_tbl, unpin);
> +}
> +
>  /*
>   * Since it is well isolated we can (and since it is perhaps large we should)
>   * also free the page tables mapping the initial P->M table.
>   */
>  static void __init xen_cleanmfnmap(unsigned long vaddr)
>  {
> -     unsigned long va = vaddr & PMD_MASK;
> -     unsigned long pa;
> -     pgd_t *pgd = pgd_offset_k(va);
> -     pud_t *pud_page = pud_offset(pgd, 0);
> -     pud_t *pud;
> -     pmd_t *pmd;
> -     pte_t *pte;
> +     pgd_t *pgd;
> +     p4d_t *p4d;
>       unsigned int i;
>       bool unpin;
>  
>       unpin = (vaddr == 2 * PGDIR_SIZE);
> -     set_pgd(pgd, __pgd(0));
> -     do {
> -             pud = pud_page + pud_index(va);
> -             if (pud_none(*pud)) {
> -                     va += PUD_SIZE;
> -             } else if (pud_large(*pud)) {
> -                     pa = pud_val(*pud) & PHYSICAL_PAGE_MASK;
> -                     xen_free_ro_pages(pa, PUD_SIZE);
> -                     va += PUD_SIZE;
> -             } else {
> -                     pmd = pmd_offset(pud, va);
> -                     if (pmd_large(*pmd)) {
> -                             pa = pmd_val(*pmd) & PHYSICAL_PAGE_MASK;
> -                             xen_free_ro_pages(pa, PMD_SIZE);
> -                     } else if (!pmd_none(*pmd)) {
> -                             pte = pte_offset_kernel(pmd, va);
> -                             set_pmd(pmd, __pmd(0));
> -                             for (i = 0; i < PTRS_PER_PTE; ++i) {
> -                                     if (pte_none(pte[i]))
> -                                             break;
> -                                     pa = pte_pfn(pte[i]) << PAGE_SHIFT;
> -                                     xen_free_ro_pages(pa, PAGE_SIZE);
> -                             }
> -                             xen_cleanmfnmap_free_pgtbl(pte, unpin);
> -                     }
> -                     va += PMD_SIZE;
> -                     if (pmd_index(va))
> -                             continue;
> -                     set_pud(pud, __pud(0));
> -                     xen_cleanmfnmap_free_pgtbl(pmd, unpin);
> -             }
> -
> -     } while (pud_index(va) || pmd_index(va));
> -     xen_cleanmfnmap_free_pgtbl(pud_page, unpin);
> +     vaddr &= PMD_MASK;
> +     pgd = pgd_offset_k(vaddr);
> +     p4d = p4d_offset(pgd, 0);
> +     for (i = 0; i < PTRS_PER_P4D; i++) {
> +             if (p4d_none(p4d[i]))
> +                     continue;
> +             xen_cleanmfnmap_p4d(p4d + i, unpin);
> +     }

Don't we need to pass vaddr down to all routines so that they select
appropriate tables? You seem to always be choosing the first one.

-boris

> +     if (IS_ENABLED(CONFIG_X86_5LEVEL)) {
> +             set_pgd(pgd, __pgd(0));
> +             xen_cleanmfnmap_free_pgtbl(p4d, unpin);
> +     }
>  }
>  
>  static void __init xen_pagetable_p2m_free(void)
> diff --git a/arch/x86/xen/mmu.h b/arch/x86/xen/mmu.h
> index 73809bb951b4..3fe2b3292915 100644
> --- a/arch/x86/xen/mmu.h
> +++ b/arch/x86/xen/mmu.h
> @@ -5,6 +5,7 @@
>  
>  enum pt_level {
>       PT_PGD,
> +     PT_P4D,
>       PT_PUD,
>       PT_PMD,
>       PT_PTE



_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
https://lists.xen.org/xen-devel

 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.