|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH V5] x86 spinlock: Fix memory corruption on completing completions
Well, I regret I mentioned the lack of barrier after enter_slowpath ;)
On 02/15, Raghavendra K T wrote:
>
> @@ -46,7 +46,8 @@ static __always_inline bool static_key_false(struct
> static_key *key);
>
> static inline void __ticket_enter_slowpath(arch_spinlock_t *lock)
> {
> - set_bit(0, (volatile unsigned long *)&lock->tickets.tail);
> + set_bit(0, (volatile unsigned long *)&lock->tickets.head);
> + barrier();
> }
Because this barrier() looks really confusing.
Firsty, it is equally unneeded on x86. At the same time, it can not help.
We need a memory barrier() between set_bit(SLOWPATH) and READ_ONCE(head)
to avoid the race with spin_unlock().
So I think you should replace it with smp_mb__after_atomic() or remove it.
Other than that I believe this version is correct. So I won't insist, this
is cosmetic after all.
Oleg.
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |