[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-changelog] [xen-unstable] Inroduce new vcpu_lock_affinity() and vcpu_unlock_affinity() helper
# HG changeset patch # User Keir Fraser <keir.fraser@xxxxxxxxxx> # Date 1201346748 0 # Node ID 7327e1c2a42ced6b1b473e0e738b12e141fb0584 # Parent 4fd33f77be6b116538c8cc286c26434267a4135f Inroduce new vcpu_lock_affinity() and vcpu_unlock_affinity() helper functions for use by x86's continue_hypercall_on_cpu(). This has two advantages: 1. We can lock out ordinary vcpu_set_affinity() commands from dom0. 2. We avoid the (in this case bogus) check for dom0_vcpus_pin. Signed-off-by: Keir Fraser <keir.fraser@xxxxxxxxxx> --- xen/arch/x86/domain.c | 9 ++++----- xen/common/schedule.c | 47 +++++++++++++++++++++++++++++++++++++++++------ xen/include/xen/sched.h | 4 ++++ 3 files changed, 49 insertions(+), 11 deletions(-) diff -r 4fd33f77be6b -r 7327e1c2a42c xen/arch/x86/domain.c --- a/xen/arch/x86/domain.c Sat Jan 26 10:39:58 2008 +0000 +++ b/xen/arch/x86/domain.c Sat Jan 26 11:25:48 2008 +0000 @@ -1416,16 +1416,16 @@ static void continue_hypercall_on_cpu_he { struct cpu_user_regs *regs = guest_cpu_user_regs(); struct migrate_info *info = v->arch.continue_info; + cpumask_t mask = info->saved_affinity; regs->eax = info->func(info->data); v->arch.schedule_tail = info->saved_schedule_tail; - v->cpu_affinity = info->saved_affinity; v->arch.continue_info = NULL; xfree(info); - vcpu_set_affinity(v, &v->cpu_affinity); + vcpu_unlock_affinity(v, &mask); schedule_tail(v); } @@ -1433,7 +1433,6 @@ int continue_hypercall_on_cpu(int cpu, l { struct vcpu *v = current; struct migrate_info *info; - cpumask_t mask = cpumask_of_cpu(cpu); int rc; if ( cpu == smp_processor_id() ) @@ -1446,12 +1445,12 @@ int continue_hypercall_on_cpu(int cpu, l info->func = func; info->data = data; info->saved_schedule_tail = v->arch.schedule_tail; - info->saved_affinity = v->cpu_affinity; + info->saved_affinity = cpumask_of_cpu(cpu); v->arch.schedule_tail = continue_hypercall_on_cpu_helper; v->arch.continue_info = info; - rc = vcpu_set_affinity(v, &mask); + rc = vcpu_lock_affinity(v, &info->saved_affinity); if ( rc ) { v->arch.schedule_tail = info->saved_schedule_tail; diff -r 4fd33f77be6b -r 7327e1c2a42c xen/common/schedule.c --- a/xen/common/schedule.c Sat Jan 26 10:39:58 2008 +0000 +++ b/xen/common/schedule.c Sat Jan 26 11:25:48 2008 +0000 @@ -262,12 +262,11 @@ void vcpu_force_reschedule(struct vcpu * } } -int vcpu_set_affinity(struct vcpu *v, cpumask_t *affinity) -{ - cpumask_t online_affinity; - - if ( (v->domain->domain_id == 0) && opt_dom0_vcpus_pin ) - return -EINVAL; +static int __vcpu_set_affinity( + struct vcpu *v, cpumask_t *affinity, + bool_t old_lock_status, bool_t new_lock_status) +{ + cpumask_t online_affinity, old_affinity; cpus_and(online_affinity, *affinity, cpu_online_map); if ( cpus_empty(online_affinity) ) @@ -275,7 +274,18 @@ int vcpu_set_affinity(struct vcpu *v, cp vcpu_schedule_lock_irq(v); + if ( v->affinity_locked != old_lock_status ) + { + BUG_ON(!v->affinity_locked); + vcpu_schedule_unlock_irq(v); + return -EBUSY; + } + + v->affinity_locked = new_lock_status; + + old_affinity = v->cpu_affinity; v->cpu_affinity = *affinity; + *affinity = old_affinity; if ( !cpu_isset(v->processor, v->cpu_affinity) ) set_bit(_VPF_migrating, &v->pause_flags); @@ -288,6 +298,31 @@ int vcpu_set_affinity(struct vcpu *v, cp } return 0; +} + +int vcpu_set_affinity(struct vcpu *v, cpumask_t *affinity) +{ + if ( (v->domain->domain_id == 0) && opt_dom0_vcpus_pin ) + return -EINVAL; + return __vcpu_set_affinity(v, affinity, 0, 0); +} + +int vcpu_lock_affinity(struct vcpu *v, cpumask_t *affinity) +{ + return __vcpu_set_affinity(v, affinity, 0, 1); +} + +void vcpu_unlock_affinity(struct vcpu *v, cpumask_t *affinity) +{ + cpumask_t online_affinity; + + /* Do not fail if no CPU in old affinity mask is online. */ + cpus_and(online_affinity, *affinity, cpu_online_map); + if ( cpus_empty(online_affinity) ) + *affinity = cpu_online_map; + + if ( __vcpu_set_affinity(v, affinity, 1, 0) != 0 ) + BUG(); } /* Block the currently-executing domain until a pertinent event occurs. */ diff -r 4fd33f77be6b -r 7327e1c2a42c xen/include/xen/sched.h --- a/xen/include/xen/sched.h Sat Jan 26 10:39:58 2008 +0000 +++ b/xen/include/xen/sched.h Sat Jan 26 11:25:48 2008 +0000 @@ -122,6 +122,8 @@ struct vcpu bool_t defer_shutdown; /* VCPU is paused following shutdown request (d->is_shutting_down)? */ bool_t paused_for_shutdown; + /* VCPU affinity is temporarily locked from controller changes? */ + bool_t affinity_locked; unsigned long pause_flags; atomic_t pause_count; @@ -485,6 +487,8 @@ void cpu_init(void); void vcpu_force_reschedule(struct vcpu *v); int vcpu_set_affinity(struct vcpu *v, cpumask_t *affinity); +int vcpu_lock_affinity(struct vcpu *v, cpumask_t *affinity); +void vcpu_unlock_affinity(struct vcpu *v, cpumask_t *affinity); void vcpu_runstate_get(struct vcpu *v, struct vcpu_runstate_info *runstate); _______________________________________________ Xen-changelog mailing list Xen-changelog@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-changelog
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |