|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH] xen: arm: bitops take unsigned int (Was: Re: [PATCH] xen/arm64: disable alignment check)
On Fri, 9 May 2014, Ian Campbell wrote:
> (Just adding the other ARM guys...)
>
> On Fri, 2014-05-09 at 14:24 +0100, Ian Campbell wrote:
> > On Tue, 2014-04-29 at 09:58 +0100, Ian Campbell wrote:
> > > On Tue, 2014-04-29 at 08:38 +0100, Vladimir Murzin wrote:
> > > > On Mon, Apr 28, 2014 at 11:43 AM, Ian Campbell
> > > > <Ian.Campbell@xxxxxxxxxx> wrote:
> > > > > But I also wanted confirmation that the problematic instruction was
> > > > > generated by gcc and not by some handcoded asm somewhere which we
> > > > > hadn't
> > > > > properly fixed.
> > >
> > > > I believe it comes form test_bit (xen/include/asm-arm/bitops.h).
> > >
> > > Ah, then I think this code needs fixing too. Probably switching to
> > > unsigned int * throughout would work, what do you think?
> >
> > I finally managed to upgrade to a new enough kernel to trigger this.
> >
> > This Works For Me(tm), along with the Linux patch "xen/events/fifo:
> > correctly align bitops" which is queued for 3.15 Linus (but not sent
> > yet?)
The change looks good to me.
> > 8<-------------------
> >
> > From aa6afe6520ea22241fb0ce430ef315c49a73867f Mon Sep 17 00:00:00 2001
> > From: Ian Campbell <ian.campbell@xxxxxxxxxx>
> > Date: Thu, 8 May 2014 16:13:55 +0100
> > Subject: [PATCH] xen: arm: bitops take unsigned int
> >
> > Xen bitmaps can be 4 rather than 8 byte aligned, so use the appropriate
> > type.
> > Otherwise the compiler can generate unaligned 8 byte accesses and cause
> > traps.
> >
> > Signed-off-by: Ian Campbell <ian.campbell@xxxxxxxxxx>
> > ---
> > xen/include/asm-arm/bitops.h | 37 +++++++++++++++++++------------------
> > 1 file changed, 19 insertions(+), 18 deletions(-)
> >
> > diff --git a/xen/include/asm-arm/bitops.h b/xen/include/asm-arm/bitops.h
> > index 0a7caee..25f96c8 100644
> > --- a/xen/include/asm-arm/bitops.h
> > +++ b/xen/include/asm-arm/bitops.h
> > @@ -18,13 +18,14 @@
> > #define __set_bit(n,p) set_bit(n,p)
> > #define __clear_bit(n,p) clear_bit(n,p)
> >
> > +#define BITS_PER_WORD 32
> > #define BIT(nr) (1UL << (nr))
> > -#define BIT_MASK(nr) (1UL << ((nr) % BITS_PER_LONG))
> > -#define BIT_WORD(nr) ((nr) / BITS_PER_LONG)
> > +#define BIT_MASK(nr) (1UL << ((nr) % BITS_PER_WORD))
> > +#define BIT_WORD(nr) ((nr) / BITS_PER_WORD)
> > #define BITS_PER_BYTE 8
> >
> > -#define ADDR (*(volatile long *) addr)
> > -#define CONST_ADDR (*(const volatile long *) addr)
> > +#define ADDR (*(volatile int *) addr)
> > +#define CONST_ADDR (*(const volatile int *) addr)
> >
> > #if defined(CONFIG_ARM_32)
> > # include <asm/arm32/bitops.h>
> > @@ -45,10 +46,10 @@
> > */
> > static inline int __test_and_set_bit(int nr, volatile void *addr)
> > {
> > - unsigned long mask = BIT_MASK(nr);
> > - volatile unsigned long *p =
> > - ((volatile unsigned long *)addr) + BIT_WORD(nr);
> > - unsigned long old = *p;
> > + unsigned int mask = BIT_MASK(nr);
> > + volatile unsigned int *p =
> > + ((volatile unsigned int *)addr) + BIT_WORD(nr);
> > + unsigned int old = *p;
> >
> > *p = old | mask;
> > return (old & mask) != 0;
> > @@ -65,10 +66,10 @@ static inline int __test_and_set_bit(int nr, volatile
> > void *addr)
> > */
> > static inline int __test_and_clear_bit(int nr, volatile void *addr)
> > {
> > - unsigned long mask = BIT_MASK(nr);
> > - volatile unsigned long *p =
> > - ((volatile unsigned long *)addr) + BIT_WORD(nr);
> > - unsigned long old = *p;
> > + unsigned int mask = BIT_MASK(nr);
> > + volatile unsigned int *p =
> > + ((volatile unsigned int *)addr) + BIT_WORD(nr);
> > + unsigned int old = *p;
> >
> > *p = old & ~mask;
> > return (old & mask) != 0;
> > @@ -78,10 +79,10 @@ static inline int __test_and_clear_bit(int nr, volatile
> > void *addr)
> > static inline int __test_and_change_bit(int nr,
> > volatile void *addr)
> > {
> > - unsigned long mask = BIT_MASK(nr);
> > - volatile unsigned long *p =
> > - ((volatile unsigned long *)addr) + BIT_WORD(nr);
> > - unsigned long old = *p;
> > + unsigned int mask = BIT_MASK(nr);
> > + volatile unsigned int *p =
> > + ((volatile unsigned int *)addr) + BIT_WORD(nr);
> > + unsigned int old = *p;
> >
> > *p = old ^ mask;
> > return (old & mask) != 0;
> > @@ -94,8 +95,8 @@ static inline int __test_and_change_bit(int nr,
> > */
> > static inline int test_bit(int nr, const volatile void *addr)
> > {
> > - const volatile unsigned long *p = (const volatile unsigned long
> > *)addr;
> > - return 1UL & (p[BIT_WORD(nr)] >> (nr & (BITS_PER_LONG-1)));
> > + const volatile unsigned int *p = (const volatile unsigned int
> > *)addr;
> > + return 1UL & (p[BIT_WORD(nr)] >> (nr & (BITS_PER_WORD-1)));
> > }
> >
> > static inline int constant_fls(int x)
>
>
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |