[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH RFC 10/12] x86/pvticketlock: keep count of blocked cpus
When a CPU blocks by calling into __ticket_lock_spinning, keep a count in the spinlock. This allows __ticket_lock_kick to more accurately tell whether it has any work to do (in many cases, a spinlock may be contended, but none of the waiters have gone into blocking). This adds two locked instructions to the spinlock slow path (once the lock has already spun for SPIN_THRESHOLD iterations), and adds another one or two bytes to struct arch_spinlock. We need to make sure we increment the waiting counter before doing the last-chance check of the lock to see if we picked it up in the meantime. If we don't then there's a potential deadlock: lock holder lock waiter clear event channel check lock for pickup (did not) release lock check waiting counter (=0, no kick) add waiting counter block (=deadlock) Moving the "add waiting counter earler" avoids the deadlock: lock holder lock waiter clear event channel add waiting counter check lock for pickup (did not) release lock check waiting counter (=1, kick) block (and immediately wake) Signed-off-by: Jeremy Fitzhardinge <jeremy.fitzhardinge@xxxxxxxxxx> --- arch/x86/include/asm/spinlock.h | 27 ++++++++++++++++++++++++++- arch/x86/include/asm/spinlock_types.h | 3 +++ arch/x86/xen/spinlock.c | 4 ++++ 3 files changed, 33 insertions(+), 1 deletions(-) diff --git a/arch/x86/include/asm/spinlock.h b/arch/x86/include/asm/spinlock.h index a79dfee..3deabca 100644 --- a/arch/x86/include/asm/spinlock.h +++ b/arch/x86/include/asm/spinlock.h @@ -65,6 +65,31 @@ static __always_inline void ____ticket_unlock_kick(struct arch_spinlock *lock, u { } +static __always_inline bool __ticket_lock_waiters(const struct arch_spinlock *lock) +{ + return false; +} +#else +static inline void __ticket_add_waiting(struct arch_spinlock *lock) +{ + if (sizeof(lock->waiting) == sizeof(u8)) + asm (LOCK_PREFIX "addb $1, %0" : "+m" (lock->waiting) : : "memory"); + else + asm (LOCK_PREFIX "addw $1, %0" : "+m" (lock->waiting) : : "memory"); +} + +static inline void __ticket_sub_waiting(struct arch_spinlock *lock) +{ + if (sizeof(lock->waiting) == sizeof(u8)) + asm (LOCK_PREFIX "subb $1, %0" : "+m" (lock->waiting) : : "memory"); + else + asm (LOCK_PREFIX "subw $1, %0" : "+m" (lock->waiting) : : "memory"); +} + +static __always_inline bool __ticket_lock_waiters(const struct arch_spinlock *lock) +{ + return ACCESS_ONCE(lock->waiting) != 0; +} #endif /* CONFIG_PARAVIRT_SPINLOCKS */ /* @@ -106,7 +131,7 @@ static __always_inline struct __raw_tickets __ticket_spin_claim(struct arch_spin */ static __always_inline void __ticket_unlock_kick(struct arch_spinlock *lock, __ticket_t next) { - if (unlikely(lock->tickets.tail != next)) + if (unlikely(__ticket_lock_waiters(lock))) ____ticket_unlock_kick(lock, next); } diff --git a/arch/x86/include/asm/spinlock_types.h b/arch/x86/include/asm/spinlock_types.h index 48dafc3..b396ed5 100644 --- a/arch/x86/include/asm/spinlock_types.h +++ b/arch/x86/include/asm/spinlock_types.h @@ -26,6 +26,9 @@ typedef struct arch_spinlock { __ticket_t head, tail; } tickets; }; +#ifdef CONFIG_PARAVIRT_SPINLOCKS + __ticket_t waiting; +#endif } arch_spinlock_t; #define __ARCH_SPIN_LOCK_UNLOCKED { { .slock = 0 } } diff --git a/arch/x86/xen/spinlock.c b/arch/x86/xen/spinlock.c index e60d5f1..2f81d5e 100644 --- a/arch/x86/xen/spinlock.c +++ b/arch/x86/xen/spinlock.c @@ -118,6 +118,8 @@ static void xen_lock_spinning(struct arch_spinlock *lock, unsigned want) /* Only check lock once pending cleared */ barrier(); + __ticket_add_waiting(lock); + /* check again make sure it didn't become free while we weren't looking */ if (ACCESS_ONCE(lock->tickets.head) == want) { @@ -132,6 +134,8 @@ static void xen_lock_spinning(struct arch_spinlock *lock, unsigned want) kstat_incr_irqs_this_cpu(irq, irq_to_desc(irq)); out: + __ticket_sub_waiting(lock); + cpumask_clear_cpu(cpu, &waiting_cpus); w->lock = NULL; spin_time_accum_blocked(start); -- 1.7.1.1 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |