[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-changelog] [xen-unstable] xen,credit1: Add variable timeslice
# HG changeset patch # User George Dunlap <george.dunlap@xxxxxxxxxxxxx> # Date 1315907023 -3600 # Node ID 1d40b27937230ba849c6e4743fc17eaf3727e35f # Parent ffe8e65f6687cde49b2cb2bf991bd3f5dff9c189 xen,credit1: Add variable timeslice Add a xen command-line parameter, sched_credit_tslice_ms, to set the timeslice of the credit1 scheduler. Signed-off-by: George Dunlap <george.dunlap@xxxxxxxxxxxxx> --- diff -r ffe8e65f6687 -r 1d40b2793723 xen/common/sched_credit.c --- a/xen/common/sched_credit.c Tue Sep 13 10:33:10 2011 +0100 +++ b/xen/common/sched_credit.c Tue Sep 13 10:43:43 2011 +0100 @@ -41,15 +41,9 @@ */ #define CSCHED_DEFAULT_WEIGHT 256 #define CSCHED_TICKS_PER_TSLICE 3 -#define CSCHED_TICKS_PER_ACCT 3 -#define CSCHED_MSECS_PER_TICK 10 -#define CSCHED_MSECS_PER_TSLICE \ - (CSCHED_MSECS_PER_TICK * CSCHED_TICKS_PER_TSLICE) +/* Default timeslice: 30ms */ +#define CSCHED_DEFAULT_TSLICE_MS 30 #define CSCHED_CREDITS_PER_MSEC 10 -#define CSCHED_CREDITS_PER_TSLICE \ - (CSCHED_CREDITS_PER_MSEC * CSCHED_MSECS_PER_TSLICE) -#define CSCHED_CREDITS_PER_ACCT \ - (CSCHED_CREDITS_PER_MSEC * CSCHED_MSECS_PER_TICK * CSCHED_TICKS_PER_ACCT) /* @@ -113,6 +107,8 @@ */ static bool_t __read_mostly sched_credit_default_yield; boolean_param("sched_credit_default_yield", sched_credit_default_yield); +static int __read_mostly sched_credit_tslice_ms = CSCHED_DEFAULT_TSLICE_MS; +integer_param("sched_credit_tslice_ms", sched_credit_tslice_ms); /* * Physical CPU @@ -176,6 +172,9 @@ uint32_t credit; int credit_balance; uint32_t runq_sort; + /* Period of master and tick in milliseconds */ + unsigned tslice_ms, tick_period_us, ticks_per_tslice; + unsigned credits_per_tslice; }; static void csched_tick(void *_cpu); @@ -326,7 +325,7 @@ spin_lock_irqsave(&prv->lock, flags); - prv->credit -= CSCHED_CREDITS_PER_ACCT; + prv->credit -= prv->credits_per_tslice; prv->ncpus--; cpu_clear(cpu, prv->idlers); cpu_clear(cpu, prv->cpus); @@ -360,19 +359,19 @@ spin_lock_irqsave(&prv->lock, flags); /* Initialize/update system-wide config */ - prv->credit += CSCHED_CREDITS_PER_ACCT; + prv->credit += prv->credits_per_tslice; prv->ncpus++; cpu_set(cpu, prv->cpus); if ( prv->ncpus == 1 ) { prv->master = cpu; init_timer(&prv->master_ticker, csched_acct, prv, cpu); - set_timer(&prv->master_ticker, NOW() + - MILLISECS(CSCHED_MSECS_PER_TICK) * CSCHED_TICKS_PER_ACCT); + set_timer(&prv->master_ticker, + NOW() + MILLISECS(prv->tslice_ms)); } init_timer(&spc->ticker, csched_tick, (void *)(unsigned long)cpu, cpu); - set_timer(&spc->ticker, NOW() + MILLISECS(CSCHED_MSECS_PER_TICK)); + set_timer(&spc->ticker, NOW() + MICROSECS(prv->tick_period_us) ); INIT_LIST_HEAD(&spc->runq); spc->runq_sort_last = prv->runq_sort; @@ -1002,7 +1001,7 @@ * for one full accounting period. We allow a domain to earn more * only when the system-wide credit balance is negative. */ - credit_peak = sdom->active_vcpu_count * CSCHED_CREDITS_PER_ACCT; + credit_peak = sdom->active_vcpu_count * prv->credits_per_tslice; if ( prv->credit_balance < 0 ) { credit_peak += ( ( -prv->credit_balance @@ -1014,7 +1013,7 @@ if ( sdom->cap != 0U ) { - credit_cap = ((sdom->cap * CSCHED_CREDITS_PER_ACCT) + 99) / 100; + credit_cap = ((sdom->cap * prv->credits_per_tslice) + 99) / 100; if ( credit_cap < credit_peak ) credit_peak = credit_cap; @@ -1092,10 +1091,10 @@ } /* Lower bound on credits */ - if ( credit < -CSCHED_CREDITS_PER_TSLICE ) + if ( credit < -prv->credits_per_tslice ) { CSCHED_STAT_CRANK(acct_min_credit); - credit = -CSCHED_CREDITS_PER_TSLICE; + credit = -prv->credits_per_tslice; atomic_set(&svc->credit, credit); } } @@ -1117,7 +1116,7 @@ } /* Upper bound on credits means VCPU stops earning */ - if ( credit > CSCHED_CREDITS_PER_TSLICE ) + if ( credit > prv->credits_per_tslice ) { __csched_vcpu_acct_stop_locked(prv, svc); /* Divide credits in half, so that when it starts @@ -1141,8 +1140,8 @@ prv->runq_sort++; out: - set_timer( &prv->master_ticker, NOW() + - MILLISECS(CSCHED_MSECS_PER_TICK) * CSCHED_TICKS_PER_ACCT ); + set_timer( &prv->master_ticker, + NOW() + MILLISECS(prv->tslice_ms)); } static void @@ -1169,7 +1168,7 @@ */ csched_runq_sort(prv, cpu); - set_timer(&spc->ticker, NOW() + MILLISECS(CSCHED_MSECS_PER_TICK)); + set_timer(&spc->ticker, NOW() + MICROSECS(prv->tick_period_us) ); } static struct csched_vcpu * @@ -1375,7 +1374,7 @@ * Return task to run next... */ ret.time = (is_idle_vcpu(snext->vcpu) ? - -1 : MILLISECS(CSCHED_MSECS_PER_TSLICE)); + -1 : MILLISECS(prv->tslice_ms)); ret.task = snext->vcpu; CSCHED_VCPU_CHECK(ret.task); @@ -1469,10 +1468,9 @@ "\tweight = %u\n" "\trunq_sort = %u\n" "\tdefault-weight = %d\n" - "\tmsecs per tick = %dms\n" + "\ttslice = %dms\n" "\tcredits per msec = %d\n" "\tticks per tslice = %d\n" - "\tticks per acct = %d\n" "\tmigration delay = %uus\n", prv->ncpus, prv->master, @@ -1481,10 +1479,9 @@ prv->weight, prv->runq_sort, CSCHED_DEFAULT_WEIGHT, - CSCHED_MSECS_PER_TICK, + prv->tslice_ms, CSCHED_CREDITS_PER_MSEC, - CSCHED_TICKS_PER_TSLICE, - CSCHED_TICKS_PER_ACCT, + prv->ticks_per_tslice, vcpu_migration_delay); cpumask_scnprintf(idlers_buf, sizeof(idlers_buf), prv->idlers); @@ -1526,6 +1523,13 @@ INIT_LIST_HEAD(&prv->active_sdom); prv->master = UINT_MAX; + prv->tslice_ms = sched_credit_tslice_ms; + prv->ticks_per_tslice = CSCHED_TICKS_PER_TSLICE; + if ( prv->tslice_ms < prv->ticks_per_tslice ) + prv->ticks_per_tslice = 1; + prv->tick_period_us = prv->tslice_ms * 1000 / prv->ticks_per_tslice; + prv->credits_per_tslice = CSCHED_CREDITS_PER_MSEC * prv->tslice_ms; + return 0; } @@ -1550,13 +1554,16 @@ static void csched_tick_resume(const struct scheduler *ops, unsigned int cpu) { + struct csched_private *prv; struct csched_pcpu *spc; uint64_t now = NOW(); spc = CSCHED_PCPU(cpu); - set_timer(&spc->ticker, now + MILLISECS(CSCHED_MSECS_PER_TICK) - - now % MILLISECS(CSCHED_MSECS_PER_TICK) ); + prv = CSCHED_PRIV(ops); + + set_timer(&spc->ticker, now + MICROSECS(prv->tick_period_us) + - now % MICROSECS(prv->tick_period_us) ); } static struct csched_private _csched_priv; _______________________________________________ Xen-changelog mailing list Xen-changelog@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-changelog
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |