|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH] xen/sched_credit: Use delay to control scheduling frequency
>>> On 26.12.11 at 09:46, Hui Lv <hui.lv@xxxxxxxxx> wrote:
> Some modifications for this patch.
> 1.Based on George's proposal, added a ratelimit_us element to csched_priv to
> constrain prv->ratelimit_us <= 1000* prv->tslice_ms in csched_init.
I do not see why you need a per-scheduler-instance variable for
issuing the warning and correcting the value - one warning (during
the first initialization) is entirely sufficient. (Otoh the already existing
tslice_ms field is pointless currently too, as it never gets changed
after being initialized from sched_credit_tslice_ms - George?)
Jan
> 2.Move the definition of sched_ratelimit_us to xen/sched.h
> 3.Remove the redundant "else" structure
>
> See if you have any comment for such changes.
>
>
> This patch can improve Xen performance:
> 1. Basically, the "delay method" can achieve 11% overall performance boost
> for SPECvirt than original credit scheduler.
> 2. We have tried 1ms delay and 10ms delay, there is no big difference
> between these two configurations. (1ms is enough to achieve a good
> performance)
> 3. We have compared different load level response time/latency (low, high,
> peak), "delay method" didn't bring very much response time increase.
> 4. 1ms delay can reduce 30% context switch at peak performance, where
> produces the benefits. (int sched_ratelimit_us = 1000 is the recommended
> setting)
>
> Signed-off-by: Hui Lv <hui.lv@xxxxxxxxx>
> Signed-off-by: George Dunlap <George.Dunlap@xxxxxxxxxxxxx>
>
> diff -r a4bff36780a3 -r 18f40e1d0274 xen/common/sched_credit.c
> --- a/xen/common/sched_credit.c Fri Dec 16 18:46:27 2011 +0000
> +++ b/xen/common/sched_credit.c Mon Dec 26 03:44:39 2011 -0500
> @@ -172,6 +172,7 @@ struct csched_private {
> uint32_t credit;
> int credit_balance;
> uint32_t runq_sort;
> + unsigned ratelimit_us;
> /* Period of master and tick in milliseconds */
> unsigned tslice_ms, tick_period_us, ticks_per_tslice;
> unsigned credits_per_tslice;
> @@ -1297,10 +1298,15 @@ csched_schedule(
> struct csched_private *prv = CSCHED_PRIV(ops);
> struct csched_vcpu *snext;
> struct task_slice ret;
> + s_time_t runtime, tslice;
>
> CSCHED_STAT_CRANK(schedule);
> CSCHED_VCPU_CHECK(current);
>
> + runtime = now - current->runstate.state_entry_time;
> + if ( runtime < 0 ) /* Does this ever happen? */
> + runtime = 0;
> +
> if ( !is_idle_vcpu(scurr->vcpu) )
> {
> /* Update credits of a non-idle VCPU. */
> @@ -1313,6 +1319,35 @@ csched_schedule(
> scurr->pri = CSCHED_PRI_IDLE;
> }
>
> + /* Choices, choices:
> + * - If we have a tasklet, we need to run the idle vcpu no matter what.
> + * - If sched rate limiting is in effect, and the current vcpu has
> + * run for less than that amount of time, continue the current one,
> + * but with a shorter timeslice and return it immediately
> + * - Otherwise, chose the one with the highest priority (which may
> + * be the one currently running)
> + * - If the currently running one is TS_OVER, see if there
> + * is a higher priority one waiting on the runqueue of another
> + * cpu and steal it.
> + */
> +
> + /* If we have schedule rate limiting enabled, check to see
> + * how long we've run for. */
> + if ( !tasklet_work_scheduled
> + && prv->ratelimit_us
> + && vcpu_runnable(current)
> + && !is_idle_vcpu(current)
> + && runtime < MICROSECS(prv->ratelimit_us) )
> + {
> + snext = scurr;
> + snext->start_time += now;
> + perfc_incr(delay_ms);
> + tslice = MICROSECS(prv->ratelimit_us);
> + ret.migrated = 0;
> + goto out;
> + }
> + tslice = MILLISECS(prv->tslice_ms);
> +
> /*
> * Select next runnable local VCPU (ie top of local runq)
> */
> @@ -1367,11 +1402,12 @@ csched_schedule(
> if ( !is_idle_vcpu(snext->vcpu) )
> snext->start_time += now;
>
> +out:
> /*
> * Return task to run next...
> */
> ret.time = (is_idle_vcpu(snext->vcpu) ?
> - -1 : MILLISECS(prv->tslice_ms));
> + -1 : tslice);
> ret.task = snext->vcpu;
>
> CSCHED_VCPU_CHECK(ret.task);
> @@ -1533,6 +1569,15 @@ csched_init(struct scheduler *ops)
> prv->tick_period_us = prv->tslice_ms * 1000 / prv->ticks_per_tslice;
> prv->credits_per_tslice = CSCHED_CREDITS_PER_MSEC * prv->tslice_ms;
>
> + if ( MICROSECS(sched_ratelimit_us) > MILLISECS(sched_credit_tslice_ms) )
> + {
> + printk("WARNING: sched_ratelimit_us >"
> + "sched_credit_tslice_ms is undefined\n"
> + "ratelimit_us is set to 1000 * tslice_ms forcely\n");
> + prv->ratelimit_us = 1000 * prv->tslice_ms;
> + }
> + else
> + prv->ratelimit_us = sched_ratelimit_us;
> return 0;
> }
>
> diff -r a4bff36780a3 -r 18f40e1d0274 xen/common/schedule.c
> --- a/xen/common/schedule.c Fri Dec 16 18:46:27 2011 +0000
> +++ b/xen/common/schedule.c Mon Dec 26 03:44:39 2011 -0500
> @@ -47,6 +47,11 @@ string_param("sched", opt_sched);
> bool_t sched_smt_power_savings = 0;
> boolean_param("sched_smt_power_savings", sched_smt_power_savings);
>
> +/* Default scheduling rate limit: 1ms
> + * The behavior when sched_ratelimit_us is greater than
> sched_credit_tslice_ms is undefined
> + * */
> +int sched_ratelimit_us = 1000;
> +integer_param("sched_ratelimit_us", sched_ratelimit_us);
> /* Various timer handlers. */
> static void s_timer_fn(void *unused);
> static void vcpu_periodic_timer_fn(void *data);
> diff -r a4bff36780a3 -r 18f40e1d0274 xen/include/xen/perfc_defn.h
> --- a/xen/include/xen/perfc_defn.h Fri Dec 16 18:46:27 2011 +0000
> +++ b/xen/include/xen/perfc_defn.h Mon Dec 26 03:44:39 2011 -0500
> @@ -16,6 +16,7 @@ PERFCOUNTER(sched_irq, "sch
> PERFCOUNTER(sched_run, "sched: runs through scheduler")
> PERFCOUNTER(sched_ctx, "sched: context switches")
>
> +PERFCOUNTER(delay_ms, "csched: delay")
> PERFCOUNTER(vcpu_check, "csched: vcpu_check")
> PERFCOUNTER(schedule, "csched: schedule")
> PERFCOUNTER(acct_run, "csched: acct_run")
> diff -r a4bff36780a3 -r 18f40e1d0274 xen/include/xen/sched-if.h
> --- a/xen/include/xen/sched-if.h Fri Dec 16 18:46:27 2011 +0000
> +++ b/xen/include/xen/sched-if.h Mon Dec 26 03:44:39 2011 -0500
> @@ -16,6 +16,11 @@ extern struct cpupool *cpupool0;
> /* cpus currently in no cpupool */
> extern cpumask_t cpupool_free_cpus;
>
> +/* Scheduler generic parameters
> + * */
> +extern int sched_ratelimit_us;
> +
> +
> /*
> * In order to allow a scheduler to remap the lock->cpu mapping,
> * we have a per-cpu pointer, along with a pre-allocated set of
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |