|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] RE: [PATCH 11/12] evtchn: convert vIRQ lock to an r/w one
> -----Original Message-----
> From: Xen-devel <xen-devel-bounces@xxxxxxxxxxxxxxxxxxxx> On Behalf Of Jan
> Beulich
> Sent: 28 September 2020 12:02
> To: xen-devel@xxxxxxxxxxxxxxxxxxxx
> Cc: Andrew Cooper <andrew.cooper3@xxxxxxxxxx>; George Dunlap
> <George.Dunlap@xxxxxxxxxxxxx>; Ian
> Jackson <iwj@xxxxxxxxxxxxxx>; Julien Grall <julien@xxxxxxx>; Wei Liu
> <wl@xxxxxxx>; Stefano Stabellini
> <sstabellini@xxxxxxxxxx>
> Subject: [PATCH 11/12] evtchn: convert vIRQ lock to an r/w one
>
> There's no need to serialize all sending of vIRQ-s; all that's needed
> is serialization against the closing of the respective event channels
> (by means of a barrier). To facilitate the conversion, introduce a new
> rw_barrier().
>
> Signed-off-by: Jan Beulich <jbeulich@xxxxxxxx>
>
> --- a/xen/common/domain.c
> +++ b/xen/common/domain.c
> @@ -160,7 +160,7 @@ struct vcpu *vcpu_create(struct domain *
> v->vcpu_id = vcpu_id;
> v->dirty_cpu = VCPU_CPU_CLEAN;
>
> - spin_lock_init(&v->virq_lock);
> + rwlock_init(&v->virq_lock);
>
> tasklet_init(&v->continue_hypercall_tasklet, NULL, NULL);
>
> --- a/xen/common/event_channel.c
> +++ b/xen/common/event_channel.c
> @@ -640,7 +640,7 @@ int evtchn_close(struct domain *d1, int
> if ( v->virq_to_evtchn[chn1->u.virq] != port1 )
> continue;
> v->virq_to_evtchn[chn1->u.virq] = 0;
> - spin_barrier(&v->virq_lock);
> + rw_barrier(&v->virq_lock);
> }
> break;
>
> @@ -794,7 +794,7 @@ void send_guest_vcpu_virq(struct vcpu *v
>
> ASSERT(!virq_is_global(virq));
>
> - spin_lock_irqsave(&v->virq_lock, flags);
> + read_lock_irqsave(&v->virq_lock, flags);
>
> port = v->virq_to_evtchn[virq];
> if ( unlikely(port == 0) )
> @@ -807,7 +807,7 @@ void send_guest_vcpu_virq(struct vcpu *v
> spin_unlock(&chn->lock);
>
> out:
> - spin_unlock_irqrestore(&v->virq_lock, flags);
> + read_unlock_irqrestore(&v->virq_lock, flags);
> }
>
> void send_guest_global_virq(struct domain *d, uint32_t virq)
> @@ -826,7 +826,7 @@ void send_guest_global_virq(struct domai
> if ( unlikely(v == NULL) )
> return;
>
> - spin_lock_irqsave(&v->virq_lock, flags);
> + read_lock_irqsave(&v->virq_lock, flags);
>
> port = v->virq_to_evtchn[virq];
> if ( unlikely(port == 0) )
> @@ -838,7 +838,7 @@ void send_guest_global_virq(struct domai
> spin_unlock(&chn->lock);
>
> out:
> - spin_unlock_irqrestore(&v->virq_lock, flags);
> + read_unlock_irqrestore(&v->virq_lock, flags);
> }
>
> void send_guest_pirq(struct domain *d, const struct pirq *pirq)
> --- a/xen/common/spinlock.c
> +++ b/xen/common/spinlock.c
> @@ -2,7 +2,7 @@
> #include <xen/irq.h>
> #include <xen/smp.h>
> #include <xen/time.h>
> -#include <xen/spinlock.h>
> +#include <xen/rwlock.h>
> #include <xen/guest_access.h>
> #include <xen/preempt.h>
> #include <public/sysctl.h>
> @@ -334,6 +334,12 @@ void _spin_unlock_recursive(spinlock_t *
> }
> }
>
> +void _rw_barrier(rwlock_t *lock)
> +{
> + check_barrier(&lock->lock.debug);
> + do { smp_mb(); } while ( _rw_is_locked(lock) );
> +}
Should you not have a cpu_relax() somewhere in here?
TBH though, the fact this lock is never taken as a writer makes me wonder
whether there needs to be a lock at all.
Paul
> +
> #ifdef CONFIG_DEBUG_LOCK_PROFILE
>
> struct lock_profile_anc {
> --- a/xen/include/xen/rwlock.h
> +++ b/xen/include/xen/rwlock.h
> @@ -237,6 +237,8 @@ static inline int _rw_is_write_locked(rw
> return (atomic_read(&lock->cnts) & _QW_WMASK) == _QW_LOCKED;
> }
>
> +void _rw_barrier(rwlock_t *lock);
> +
> #define read_lock(l) _read_lock(l)
> #define read_lock_irq(l) _read_lock_irq(l)
> #define read_lock_irqsave(l, f) \
> @@ -266,6 +268,7 @@ static inline int _rw_is_write_locked(rw
> #define rw_is_locked(l) _rw_is_locked(l)
> #define rw_is_write_locked(l) _rw_is_write_locked(l)
>
> +#define rw_barrier(l) _rw_barrier(l)
>
> typedef struct percpu_rwlock percpu_rwlock_t;
>
> --- a/xen/include/xen/sched.h
> +++ b/xen/include/xen/sched.h
> @@ -235,7 +235,7 @@ struct vcpu
>
> /* IRQ-safe virq_lock protects against delivering VIRQ to stale evtchn.
> */
> evtchn_port_t virq_to_evtchn[NR_VIRQS];
> - spinlock_t virq_lock;
> + rwlock_t virq_lock;
>
> /* Tasklet for continue_hypercall_on_cpu(). */
> struct tasklet continue_hypercall_tasklet;
>
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |