[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v5 2/8] xen/arm: rename *_xen_dcache_* operations to *_dcache_*
Hi Stefano, On 10/13/2014 03:58 PM, Stefano Stabellini wrote: > Signed-off-by: Stefano Stabellini <stefano.stabellini@xxxxxxxxxxxxx> I guess this patch is just mechanical changes and has been build tested on both arm32 and arm64, so: Reviewed-by: Julien Grall <julien.grall@xxxxxxxxxx> Regards, > --- > xen/arch/arm/guestcopy.c | 2 +- > xen/arch/arm/kernel.c | 2 +- > xen/arch/arm/mm.c | 24 ++++++++++++------------ > xen/arch/arm/p2m.c | 4 ++-- > xen/arch/arm/smpboot.c | 2 +- > xen/include/asm-arm/arm32/page.h | 4 ++-- > xen/include/asm-arm/arm64/page.h | 4 ++-- > xen/include/asm-arm/page.h | 20 ++++++++++---------- > 8 files changed, 31 insertions(+), 31 deletions(-) > > diff --git a/xen/arch/arm/guestcopy.c b/xen/arch/arm/guestcopy.c > index 0173597..7dbaeca 100644 > --- a/xen/arch/arm/guestcopy.c > +++ b/xen/arch/arm/guestcopy.c > @@ -27,7 +27,7 @@ static unsigned long raw_copy_to_guest_helper(void *to, > const void *from, > p += offset; > memcpy(p, from, size); > if ( flush_dcache ) > - clean_xen_dcache_va_range(p, size); > + clean_dcache_va_range(p, size); > > unmap_domain_page(p - offset); > put_page(page); > diff --git a/xen/arch/arm/kernel.c b/xen/arch/arm/kernel.c > index 1b8ac9a..209c3dd 100644 > --- a/xen/arch/arm/kernel.c > +++ b/xen/arch/arm/kernel.c > @@ -57,7 +57,7 @@ void copy_from_paddr(void *dst, paddr_t paddr, unsigned > long len) > > set_fixmap(FIXMAP_MISC, p, BUFFERABLE); > memcpy(dst, src + s, l); > - clean_xen_dcache_va_range(dst, l); > + clean_dcache_va_range(dst, l); > > paddr += l; > dst += l; > diff --git a/xen/arch/arm/mm.c b/xen/arch/arm/mm.c > index 675a51b..996687b 100644 > --- a/xen/arch/arm/mm.c > +++ b/xen/arch/arm/mm.c > @@ -381,7 +381,7 @@ void flush_page_to_ram(unsigned long mfn) > { > void *v = map_domain_page(mfn); > > - clean_and_invalidate_xen_dcache_va_range(v, PAGE_SIZE); > + clean_and_invalidate_dcache_va_range(v, PAGE_SIZE); > unmap_domain_page(v); > } > > @@ -504,17 +504,17 @@ void __init setup_pagetables(unsigned long > boot_phys_offset, paddr_t xen_paddr) > /* Clear the copy of the boot pagetables. Each secondary CPU > * rebuilds these itself (see head.S) */ > memset(boot_pgtable, 0x0, PAGE_SIZE); > - clean_and_invalidate_xen_dcache(boot_pgtable); > + clean_and_invalidate_dcache(boot_pgtable); > #ifdef CONFIG_ARM_64 > memset(boot_first, 0x0, PAGE_SIZE); > - clean_and_invalidate_xen_dcache(boot_first); > + clean_and_invalidate_dcache(boot_first); > memset(boot_first_id, 0x0, PAGE_SIZE); > - clean_and_invalidate_xen_dcache(boot_first_id); > + clean_and_invalidate_dcache(boot_first_id); > #endif > memset(boot_second, 0x0, PAGE_SIZE); > - clean_and_invalidate_xen_dcache(boot_second); > + clean_and_invalidate_dcache(boot_second); > memset(boot_third, 0x0, PAGE_SIZE); > - clean_and_invalidate_xen_dcache(boot_third); > + clean_and_invalidate_dcache(boot_third); > > /* Break up the Xen mapping into 4k pages and protect them separately. */ > for ( i = 0; i < LPAE_ENTRIES; i++ ) > @@ -552,7 +552,7 @@ void __init setup_pagetables(unsigned long > boot_phys_offset, paddr_t xen_paddr) > > /* Make sure it is clear */ > memset(this_cpu(xen_dommap), 0, DOMHEAP_SECOND_PAGES*PAGE_SIZE); > - clean_xen_dcache_va_range(this_cpu(xen_dommap), > + clean_dcache_va_range(this_cpu(xen_dommap), > DOMHEAP_SECOND_PAGES*PAGE_SIZE); > #endif > } > @@ -563,7 +563,7 @@ int init_secondary_pagetables(int cpu) > /* Set init_ttbr for this CPU coming up. All CPus share a single setof > * pagetables, but rewrite it each time for consistency with 32 bit. */ > init_ttbr = (uintptr_t) xen_pgtable + phys_offset; > - clean_xen_dcache(init_ttbr); > + clean_dcache(init_ttbr); > return 0; > } > #else > @@ -598,15 +598,15 @@ int init_secondary_pagetables(int cpu) > > write_pte(&first[first_table_offset(DOMHEAP_VIRT_START+i*FIRST_SIZE)], pte); > } > > - clean_xen_dcache_va_range(first, PAGE_SIZE); > - clean_xen_dcache_va_range(domheap, DOMHEAP_SECOND_PAGES*PAGE_SIZE); > + clean_dcache_va_range(first, PAGE_SIZE); > + clean_dcache_va_range(domheap, DOMHEAP_SECOND_PAGES*PAGE_SIZE); > > per_cpu(xen_pgtable, cpu) = first; > per_cpu(xen_dommap, cpu) = domheap; > > /* Set init_ttbr for this CPU coming up */ > init_ttbr = __pa(first); > - clean_xen_dcache(init_ttbr); > + clean_dcache(init_ttbr); > > return 0; > } > @@ -1273,7 +1273,7 @@ void clear_and_clean_page(struct page_info *page) > void *p = __map_domain_page(page); > > clear_page(p); > - clean_xen_dcache_va_range(p, PAGE_SIZE); > + clean_dcache_va_range(p, PAGE_SIZE); > unmap_domain_page(p); > } > > diff --git a/xen/arch/arm/p2m.c b/xen/arch/arm/p2m.c > index 70929fc..6ff8c17 100644 > --- a/xen/arch/arm/p2m.c > +++ b/xen/arch/arm/p2m.c > @@ -302,7 +302,7 @@ static inline void p2m_write_pte(lpae_t *p, lpae_t pte, > bool_t flush_cache) > { > write_pte(p, pte); > if ( flush_cache ) > - clean_xen_dcache(*p); > + clean_dcache(*p); > } > > /* > @@ -362,7 +362,7 @@ static int p2m_create_table(struct domain *d, lpae_t > *entry, > clear_page(p); > > if ( flush_cache ) > - clean_xen_dcache_va_range(p, PAGE_SIZE); > + clean_dcache_va_range(p, PAGE_SIZE); > > unmap_domain_page(p); > > diff --git a/xen/arch/arm/smpboot.c b/xen/arch/arm/smpboot.c > index ee395a1..14054ae 100644 > --- a/xen/arch/arm/smpboot.c > +++ b/xen/arch/arm/smpboot.c > @@ -374,7 +374,7 @@ int __cpu_up(unsigned int cpu) > > /* Open the gate for this CPU */ > smp_up_cpu = cpu_logical_map(cpu); > - clean_xen_dcache(smp_up_cpu); > + clean_dcache(smp_up_cpu); > > rc = arch_cpu_up(cpu); > > diff --git a/xen/include/asm-arm/arm32/page.h > b/xen/include/asm-arm/arm32/page.h > index 9740672..20a6a7f 100644 > --- a/xen/include/asm-arm/arm32/page.h > +++ b/xen/include/asm-arm/arm32/page.h > @@ -20,11 +20,11 @@ static inline void write_pte(lpae_t *p, lpae_t pte) > } > > /* Inline ASM to flush dcache on register R (may be an inline asm operand) */ > -#define __clean_xen_dcache_one(R) STORE_CP32(R, DCCMVAC) > +#define __clean_dcache_one(R) STORE_CP32(R, DCCMVAC) > > /* Inline ASM to clean and invalidate dcache on register R (may be an > * inline asm operand) */ > -#define __clean_and_invalidate_xen_dcache_one(R) STORE_CP32(R, DCCIMVAC) > +#define __clean_and_invalidate_dcache_one(R) STORE_CP32(R, DCCIMVAC) > > /* > * Flush all hypervisor mappings from the TLB and branch predictor of > diff --git a/xen/include/asm-arm/arm64/page.h > b/xen/include/asm-arm/arm64/page.h > index bb10164..101d7a8 100644 > --- a/xen/include/asm-arm/arm64/page.h > +++ b/xen/include/asm-arm/arm64/page.h > @@ -15,11 +15,11 @@ static inline void write_pte(lpae_t *p, lpae_t pte) > } > > /* Inline ASM to flush dcache on register R (may be an inline asm operand) */ > -#define __clean_xen_dcache_one(R) "dc cvac, %" #R ";" > +#define __clean_dcache_one(R) "dc cvac, %" #R ";" > > /* Inline ASM to clean and invalidate dcache on register R (may be an > * inline asm operand) */ > -#define __clean_and_invalidate_xen_dcache_one(R) "dc civac, %" #R ";" > +#define __clean_and_invalidate_dcache_one(R) "dc civac, %" #R ";" > > /* > * Flush all hypervisor mappings from the TLB of the local processor. > diff --git a/xen/include/asm-arm/page.h b/xen/include/asm-arm/page.h > index d758b61..1327b00 100644 > --- a/xen/include/asm-arm/page.h > +++ b/xen/include/asm-arm/page.h > @@ -268,48 +268,48 @@ extern size_t cacheline_bytes; > /* Functions for flushing medium-sized areas. > * if 'range' is large enough we might want to use model-specific > * full-cache flushes. */ > -static inline void clean_xen_dcache_va_range(const void *p, unsigned long > size) > +static inline void clean_dcache_va_range(const void *p, unsigned long size) > { > const void *end; > dsb(sy); /* So the CPU issues all writes to the range */ > for ( end = p + size; p < end; p += cacheline_bytes ) > - asm volatile (__clean_xen_dcache_one(0) : : "r" (p)); > + asm volatile (__clean_dcache_one(0) : : "r" (p)); > dsb(sy); /* So we know the flushes happen before continuing */ > } > > -static inline void clean_and_invalidate_xen_dcache_va_range > +static inline void clean_and_invalidate_dcache_va_range > (const void *p, unsigned long size) > { > const void *end; > dsb(sy); /* So the CPU issues all writes to the range */ > for ( end = p + size; p < end; p += cacheline_bytes ) > - asm volatile (__clean_and_invalidate_xen_dcache_one(0) : : "r" (p)); > + asm volatile (__clean_and_invalidate_dcache_one(0) : : "r" (p)); > dsb(sy); /* So we know the flushes happen before continuing */ > } > > /* Macros for flushing a single small item. The predicate is always > * compile-time constant so this will compile down to 3 instructions in > * the common case. */ > -#define clean_xen_dcache(x) do { \ > +#define clean_dcache(x) do { \ > typeof(x) *_p = &(x); \ > if ( sizeof(x) > MIN_CACHELINE_BYTES || sizeof(x) > alignof(x) ) \ > - clean_xen_dcache_va_range(_p, sizeof(x)); \ > + clean_dcache_va_range(_p, sizeof(x)); \ > else \ > asm volatile ( \ > "dsb sy;" /* Finish all earlier writes */ \ > - __clean_xen_dcache_one(0) \ > + __clean_dcache_one(0) \ > "dsb sy;" /* Finish flush before continuing */ \ > : : "r" (_p), "m" (*_p)); \ > } while (0) > > -#define clean_and_invalidate_xen_dcache(x) do { \ > +#define clean_and_invalidate_dcache(x) do { \ > typeof(x) *_p = &(x); \ > if ( sizeof(x) > MIN_CACHELINE_BYTES || sizeof(x) > alignof(x) ) \ > - clean_and_invalidate_xen_dcache_va_range(_p, sizeof(x)); \ > + clean_and_invalidate_dcache_va_range(_p, sizeof(x)); \ > else \ > asm volatile ( \ > "dsb sy;" /* Finish all earlier writes */ \ > - __clean_and_invalidate_xen_dcache_one(0) \ > + __clean_and_invalidate_dcache_one(0) \ > "dsb sy;" /* Finish flush before continuing */ \ > : : "r" (_p), "m" (*_p)); \ > } while (0) > -- Julien Grall _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |