[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH] domctl: improve locking during domain destruction
On 06/12/17 15:31, Jan Beulich wrote: > There is no need to hold the global domctl lock across domain_kill() - > the domain lock is fully sufficient here, and parallel cleanup after > multiple domains performs quite a bit better this way. > > Signed-off-by: Jan Beulich <jbeulich@xxxxxxxx> This is clearer. Reviewed-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx> > --- > Changes since RFC: Comment added. > --- > Obviously other domctl-s could benefit from similar adjustments, so > this is meant to be just a start. > > --- a/xen/common/domain.c > +++ b/xen/common/domain.c > @@ -615,13 +615,21 @@ int domain_kill(struct domain *d) > if ( d == current->domain ) > return -EINVAL; > > - /* Protected by domctl_lock. */ > + /* Protected by d->domain_lock. */ > switch ( d->is_dying ) > { > case DOMDYING_alive: > + domain_unlock(d); > domain_pause(d); > + domain_lock(d); > + /* > + * With the domain lock dropped, d->is_dying may have changed. Call > + * ourselves recursively if so, which is safe as then we won't come > + * back here. > + */ > + if ( d->is_dying != DOMDYING_alive ) > + return domain_kill(d); > d->is_dying = DOMDYING_dying; > - spin_barrier(&d->domain_lock); > evtchn_destroy(d); > gnttab_release_mappings(d); > tmem_destroy(d->tmem_client); > --- a/xen/common/domctl.c > +++ b/xen/common/domctl.c > @@ -665,11 +665,14 @@ long do_domctl(XEN_GUEST_HANDLE_PARAM(xe > break; > > case XEN_DOMCTL_destroydomain: > + domctl_lock_release(); > + domain_lock(d); > ret = domain_kill(d); > + domain_unlock(d); > if ( ret == -ERESTART ) > ret = hypercall_create_continuation( > __HYPERVISOR_domctl, "h", u_domctl); > - break; > + goto domctl_out_unlock_domonly; > > case XEN_DOMCTL_setnodeaffinity: > { > > > _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxxx https://lists.xenproject.org/mailman/listinfo/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |