[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH 47/60] xen/sched: move per-cpu variable scheduler to struct sched_resource
Having a pointer to struct scheduler in struct sched_resource instead of per cpu is enough. Signed-off-by: Juergen Gross <jgross@xxxxxxxx> --- V1: new patch --- xen/common/sched_credit.c | 18 +++++++++++------- xen/common/sched_credit2.c | 3 ++- xen/common/schedule.c | 21 ++++++++++----------- xen/include/xen/sched-if.h | 2 +- 4 files changed, 24 insertions(+), 20 deletions(-) diff --git a/xen/common/sched_credit.c b/xen/common/sched_credit.c index 15339e6fae..5e60788112 100644 --- a/xen/common/sched_credit.c +++ b/xen/common/sched_credit.c @@ -350,9 +350,10 @@ DEFINE_PER_CPU(unsigned int, last_tickle_cpu); static inline void __runq_tickle(struct csched_unit *new) { unsigned int cpu = sched_unit_cpu(new->unit); + struct sched_resource *sd = get_sched_res(cpu); struct sched_unit *unit = new->unit; struct csched_unit * const cur = CSCHED_UNIT(curr_on_cpu(cpu)); - struct csched_private *prv = CSCHED_PRIV(per_cpu(scheduler, cpu)); + struct csched_private *prv = CSCHED_PRIV(sd->scheduler); cpumask_t mask, idle_mask, *online; int balance_step, idlers_empty; @@ -937,7 +938,8 @@ csched_unit_acct(struct csched_private *prv, unsigned int cpu) { struct sched_unit *currunit = current->sched_unit; struct csched_unit * const svc = CSCHED_UNIT(currunit); - const struct scheduler *ops = per_cpu(scheduler, cpu); + struct sched_resource *sd = get_sched_res(cpu); + const struct scheduler *ops = sd->scheduler; ASSERT( sched_unit_cpu(currunit) == cpu ); ASSERT( svc->sdom != NULL ); @@ -993,8 +995,7 @@ csched_unit_acct(struct csched_private *prv, unsigned int cpu) * idlers. But, if we are here, it means there is someone running * on it, and hence the bit must be zero already. */ - ASSERT(!cpumask_test_cpu(cpu, - CSCHED_PRIV(per_cpu(scheduler, cpu))->idlers)); + ASSERT(!cpumask_test_cpu(cpu, CSCHED_PRIV(sd->scheduler)->idlers)); cpu_raise_softirq(cpu, SCHEDULE_SOFTIRQ); } } @@ -1089,6 +1090,7 @@ csched_unit_sleep(const struct scheduler *ops, struct sched_unit *unit) { struct csched_unit * const svc = CSCHED_UNIT(unit); unsigned int cpu = sched_unit_cpu(unit); + struct sched_resource *sd = get_sched_res(cpu); SCHED_STAT_CRANK(unit_sleep); @@ -1101,7 +1103,7 @@ csched_unit_sleep(const struct scheduler *ops, struct sched_unit *unit) * But, we are here because unit is going to sleep while running on cpu, * so the bit must be zero already. */ - ASSERT(!cpumask_test_cpu(cpu, CSCHED_PRIV(per_cpu(scheduler, cpu))->idlers)); + ASSERT(!cpumask_test_cpu(cpu, CSCHED_PRIV(sd->scheduler)->idlers)); cpu_raise_softirq(cpu, SCHEDULE_SOFTIRQ); } else if ( __unit_on_runq(svc) ) @@ -1581,8 +1583,9 @@ static void csched_tick(void *_cpu) { unsigned int cpu = (unsigned long)_cpu; + struct sched_resource *sd = get_sched_res(cpu); struct csched_pcpu *spc = CSCHED_PCPU(cpu); - struct csched_private *prv = CSCHED_PRIV(per_cpu(scheduler, cpu)); + struct csched_private *prv = CSCHED_PRIV(sd->scheduler); spc->tick++; @@ -1607,7 +1610,8 @@ csched_tick(void *_cpu) static struct csched_unit * csched_runq_steal(int peer_cpu, int cpu, int pri, int balance_step) { - const struct csched_private * const prv = CSCHED_PRIV(per_cpu(scheduler, cpu)); + struct sched_resource *sd = get_sched_res(cpu); + const struct csched_private * const prv = CSCHED_PRIV(sd->scheduler); const struct csched_pcpu * const peer_pcpu = CSCHED_PCPU(peer_cpu); struct csched_unit *speer; struct list_head *iter; diff --git a/xen/common/sched_credit2.c b/xen/common/sched_credit2.c index 3bfceefa46..1764aa704e 100644 --- a/xen/common/sched_credit2.c +++ b/xen/common/sched_credit2.c @@ -3265,8 +3265,9 @@ runq_candidate(struct csched2_runqueue_data *rqd, unsigned int *skipped) { struct list_head *iter, *temp; + struct sched_resource *sd = get_sched_res(cpu); struct csched2_unit *snext = NULL; - struct csched2_private *prv = csched2_priv(per_cpu(scheduler, cpu)); + struct csched2_private *prv = csched2_priv(sd->scheduler); bool yield = false, soft_aff_preempt = false; *skipped = 0; diff --git a/xen/common/schedule.c b/xen/common/schedule.c index 9bff4dc183..34c95d1dc6 100644 --- a/xen/common/schedule.c +++ b/xen/common/schedule.c @@ -66,7 +66,6 @@ static void vcpu_singleshot_timer_fn(void *data); static void poll_timer_fn(void *data); /* This is global for now so that private implementations can reach it */ -DEFINE_PER_CPU(struct scheduler *, scheduler); DEFINE_PER_CPU(struct sched_resource *, sched_res); static DEFINE_PER_CPU(unsigned int, sched_res_idx); @@ -133,7 +132,7 @@ static inline struct scheduler *vcpu_scheduler(const struct vcpu *v) * for idle vCPUs, it is safe to use it, with no locks, to figure that out. */ ASSERT(is_idle_domain(d)); - return per_cpu(scheduler, v->processor); + return get_sched_res(v->processor)->scheduler; } #define VCPU2ONLINE(_v) cpupool_domain_cpumask((_v)->domain) @@ -1767,8 +1766,8 @@ static bool sched_tasklet_check(unsigned int cpu) static struct sched_unit *do_schedule(struct sched_unit *prev, s_time_t now, unsigned int cpu) { - struct scheduler *sched = per_cpu(scheduler, cpu); struct sched_resource *sd = get_sched_res(cpu); + struct scheduler *sched = sd->scheduler; struct sched_unit *next; /* get policy-specific decision on scheduling... */ @@ -2144,7 +2143,7 @@ static int cpu_schedule_up(unsigned int cpu) sd->cpus = cpumask_of(cpu); set_sched_res(cpu, sd); - per_cpu(scheduler, cpu) = &ops; + sd->scheduler = &ops; spin_lock_init(&sd->_lock); sd->schedule_lock = &sd->_lock; init_timer(&sd->s_timer, s_timer_fn, NULL, cpu); @@ -2203,7 +2202,7 @@ static int cpu_schedule_up(unsigned int cpu) static void cpu_schedule_down(unsigned int cpu) { struct sched_resource *sd = get_sched_res(cpu); - struct scheduler *sched = per_cpu(scheduler, cpu); + struct scheduler *sched = sd->scheduler; sched_free_pdata(sched, sd->sched_priv, cpu); sched_free_vdata(sched, idle_vcpu[cpu]->sched_unit->priv); @@ -2219,8 +2218,8 @@ static void cpu_schedule_down(unsigned int cpu) void scheduler_percpu_init(unsigned int cpu) { - struct scheduler *sched = per_cpu(scheduler, cpu); struct sched_resource *sd = get_sched_res(cpu); + struct scheduler *sched = sd->scheduler; if ( system_state != SYS_STATE_resume ) sched_init_pdata(sched, sd->sched_priv, cpu); @@ -2230,8 +2229,8 @@ static int cpu_schedule_callback( struct notifier_block *nfb, unsigned long action, void *hcpu) { unsigned int cpu = (unsigned long)hcpu; - struct scheduler *sched = per_cpu(scheduler, cpu); struct sched_resource *sd = get_sched_res(cpu); + struct scheduler *sched = sd->scheduler; int rc = 0; /* @@ -2407,7 +2406,7 @@ int schedule_cpu_switch(unsigned int cpu, struct cpupool *c) { struct vcpu *idle; void *ppriv, *ppriv_old, *vpriv, *vpriv_old; - struct scheduler *old_ops = per_cpu(scheduler, cpu); + struct scheduler *old_ops = get_sched_res(cpu)->scheduler; struct scheduler *new_ops = (c == NULL) ? &ops : c->sched; struct cpupool *old_pool = per_cpu(cpupool, cpu); struct sched_resource *sd = get_sched_res(cpu); @@ -2473,7 +2472,7 @@ int schedule_cpu_switch(unsigned int cpu, struct cpupool *c) ppriv_old = sd->sched_priv; new_lock = sched_switch_sched(new_ops, cpu, ppriv, vpriv); - per_cpu(scheduler, cpu) = new_ops; + sd->scheduler = new_ops; sd->sched_priv = ppriv; /* @@ -2574,7 +2573,7 @@ void sched_tick_suspend(void) struct scheduler *sched; unsigned int cpu = smp_processor_id(); - sched = per_cpu(scheduler, cpu); + sched = get_sched_res(cpu)->scheduler; sched_do_tick_suspend(sched, cpu); rcu_idle_enter(cpu); rcu_idle_timer_start(); @@ -2587,7 +2586,7 @@ void sched_tick_resume(void) rcu_idle_timer_stop(); rcu_idle_exit(cpu); - sched = per_cpu(scheduler, cpu); + sched = get_sched_res(cpu)->scheduler; sched_do_tick_resume(sched, cpu); } diff --git a/xen/include/xen/sched-if.h b/xen/include/xen/sched-if.h index f5962cbcfb..ad6cf43425 100644 --- a/xen/include/xen/sched-if.h +++ b/xen/include/xen/sched-if.h @@ -36,6 +36,7 @@ extern const cpumask_t *sched_res_mask; * as the rest of the struct. Just have the scheduler point to the * one it wants (This may be the one right in front of it).*/ struct sched_resource { + struct scheduler *scheduler; spinlock_t *schedule_lock, _lock; struct sched_unit *curr; /* current task */ @@ -50,7 +51,6 @@ struct sched_resource { #define curr_on_cpu(c) (get_sched_res(c)->curr) -DECLARE_PER_CPU(struct scheduler *, scheduler); DECLARE_PER_CPU(struct cpupool *, cpupool); DECLARE_PER_CPU(struct sched_resource *, sched_res); -- 2.16.4 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxxx https://lists.xenproject.org/mailman/listinfo/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |