[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH RFC V2 45/45] xen/sched: add scheduling granularity enum
On 10/05/2019 10:53, Jan Beulich wrote: >>>> On 08.05.19 at 16:36, <jgross@xxxxxxxx> wrote: >> On 06/05/2019 12:01, Jan Beulich wrote: >>>>>> On 06.05.19 at 11:23, <jgross@xxxxxxxx> wrote: >>>> And that was mentioned in the cover letter: cpu hotplug is not yet >>>> handled (hence the RFC status of the series). >>>> >>>> When cpu hotplug is being added it might be appropriate to switch the >>>> scheme as you suggested. Right now the current solution is much more >>>> simple. >>> >>> I see (I did notice the cover letter remark, but managed to not >>> honor it when writing the reply), but I'm unconvinced if incurring >>> more code churn by not dealing with things the "dynamic" way >>> right away is indeed the "more simple" (overall) solution. >> >> I have started to address cpu on/offlining now. >> >> There are multiple design decisions to take. >> >> 1. Interaction between sched-gran and smt boot parameters >> 2. Interaction between sched-gran and xen-hptool smt switching >> 3. Interaction between sched-gran and single cpu on/offlining >> >> Right now any guest won't see a difference regarding sched-gran >> selection. This means we don't have to think about potential migration >> restrictions. This might change in future when we want to enable the >> guest to e.g. use core scheduling themselves in order to mitigate >> against side channel attacks within the guest. >> >> The most simple solution would be (and I'd like to send out V1 of my >> series with that implemented): >> >> sched-gran=core and sched-gran=socket don't allow dynamical switching >> of smt via xen-hptool. >> >> With sched-gran=core or sched-gran=socket offlining a single cpu results >> in moving the complete core or socket to cpupool_free_cpus and then >> offlining from there. Only complete cores/sockets can be moved to any >> cpupool. When onlining a cpu it is added to cpupool_free_cpus and if >> the core/socket is completely online it will automatically be added to >> Pool-0 (as today any single onlined cpu). > > Well, this is in line with what was discussed on the call yesterday, so > I think it's an acceptable initial state to end up in. Albeit, just for > completeness, I'm not convinced there's no use for "smt-{dis,en}able" > anymore with core-aware scheduling implemented just in Xen - it > may still be considered useful as long as we don't expose proper > topology to guests, for them to be able to do something similar. As the extra complexity for supporting that is significant I'd like to at least postpone it. And with the (later) introduction of per-cpupool smt on/off I guess this would be even less important. Juergen _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxxx https://lists.xenproject.org/mailman/listinfo/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |