[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-changelog] [xen master] arm: provide add_sized()
commit 890674d13feb4a270aa112ca452dcf62fdd53f34 Author: David Vrabel <david.vrabel@xxxxxxxxxx> AuthorDate: Wed May 13 15:01:25 2015 +0200 Commit: Jan Beulich <jbeulich@xxxxxxxx> CommitDate: Wed May 13 15:01:25 2015 +0200 arm: provide add_sized() add_sized(ptr, inc) adds inc to the value at ptr using only the correct size of loads and stores for the type of *ptr. The add is /not/ atomic. This is needed for ticket locks to ensure the increment of the head ticket does not affect the tail ticket. Signed-off-by: David Vrabel <david.vrabel@xxxxxxxxxx> Acked-by: Ian Campbell <ian.campbell@xxxxxxxxxx> --- xen/include/asm-arm/atomic.h | 26 ++++++++++++++++++++++++++ 1 files changed, 26 insertions(+), 0 deletions(-) diff --git a/xen/include/asm-arm/atomic.h b/xen/include/asm-arm/atomic.h index 7d15fb0..5a38c67 100644 --- a/xen/include/asm-arm/atomic.h +++ b/xen/include/asm-arm/atomic.h @@ -23,6 +23,17 @@ static inline void name(volatile type *addr, type val) \ : reg (val)); \ } +#define build_add_sized(name, size, width, type, reg) \ +static inline void name(volatile type *addr, type val) \ +{ \ + type t; \ + asm volatile("ldr" size " %"width"1,%0\n" \ + "add %"width"1,%"width"1,%"width"2\n" \ + "str" size " %"width"1,%0" \ + : "=m" (*(volatile type *)addr), "=r" (t) \ + : reg (val)); \ +} + #if defined (CONFIG_ARM_32) #define BYTE "" #define WORD "" @@ -46,6 +57,10 @@ build_atomic_read(read_u64_atomic, "x", uint64_t, "=r") build_atomic_write(write_u64_atomic, "x", uint64_t, "r") #endif +build_add_sized(add_u8_sized, "b", BYTE, uint8_t, "ri") +build_add_sized(add_u16_sized, "h", WORD, uint16_t, "ri") +build_add_sized(add_u32_sized, "", WORD, uint32_t, "ri") + void __bad_atomic_size(void); #define read_atomic(p) ({ \ @@ -70,6 +85,17 @@ void __bad_atomic_size(void); __x; \ }) +#define add_sized(p, x) ({ \ + typeof(*(p)) __x = (x); \ + switch ( sizeof(*(p)) ) \ + { \ + case 1: add_u8_sized((uint8_t *)(p), __x); break; \ + case 2: add_u16_sized((uint16_t *)(p), __x); break; \ + case 4: add_u32_sized((uint32_t *)(p), __x); break; \ + default: __bad_atomic_size(); break; \ + } \ +}) + /* * NB. I've pushed the volatile qualifier into the operations. This allows * fast accessors such as _atomic_read() and _atomic_set() which don't give -- generated by git-patchbot for /home/xen/git/xen.git#master _______________________________________________ Xen-changelog mailing list Xen-changelog@xxxxxxxxxxxxx http://lists.xensource.com/xen-changelog
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |