|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [PATCH 1/3] xen/sched: introduce cpupool_update_node_affinity()
On 03.08.2022 10:01, Juergen Gross wrote:
> On 03.08.22 09:50, Jan Beulich wrote:
>> On 02.08.2022 15:27, Juergen Gross wrote:
>>> --- a/xen/common/sched/core.c
>>> +++ b/xen/common/sched/core.c
>>> @@ -1790,28 +1790,14 @@ int vcpu_affinity_domctl(struct domain *d, uint32_t
>>> cmd,
>>> return ret;
>>> }
>>>
>>> -void domain_update_node_affinity(struct domain *d)
>>> +void domain_update_node_affinity_noalloc(struct domain *d,
>>> + const cpumask_t *online,
>>> + struct affinity_masks *affinity)
>>> {
>>> - cpumask_var_t dom_cpumask, dom_cpumask_soft;
>>> cpumask_t *dom_affinity;
>>> - const cpumask_t *online;
>>> struct sched_unit *unit;
>>> unsigned int cpu;
>>>
>>> - /* Do we have vcpus already? If not, no need to update node-affinity.
>>> */
>>> - if ( !d->vcpu || !d->vcpu[0] )
>>> - return;
>>> -
>>> - if ( !zalloc_cpumask_var(&dom_cpumask) )
>>> - return;
>>> - if ( !zalloc_cpumask_var(&dom_cpumask_soft) )
>>> - {
>>> - free_cpumask_var(dom_cpumask);
>>> - return;
>>> - }
>>
>> Instead of splitting the function, did you consider using
>> cond_zalloc_cpumask_var() here, thus allowing (but not requiring)
>> callers to pre-allocate the masks? Would imo be quite a bit less
>> code churn (I think).
>
> This would require to change all callers of domain_update_node_affinity()
> to add the additional mask parameter. The now common/sched local struct
> affinity_masks would then need to made globally visible.
>
> I'm not sure this is a good idea.
Hmm, I see there are quite a few callers (so there would be code churn
elsewhere). But I don't think the struct would need making globally
visible - the majority of callers could simply pass NULL, making the
function use a local instance of the struct instead. Personally I think
that would still be neater than having a _noalloc-suffixed variant of a
function (and specifically in this case also with an already long name).
But I guess this is then up to you / the scheduler maintainers.
>>> --- a/xen/common/sched/cpupool.c
>>> +++ b/xen/common/sched/cpupool.c
>>> @@ -410,6 +410,48 @@ int cpupool_move_domain(struct domain *d, struct
>>> cpupool *c)
>>> return ret;
>>> }
>>>
>>> +/* Update affinities of all domains in a cpupool. */
>>> +static int cpupool_alloc_affin_masks(struct affinity_masks *masks)
>>> +{
>>> + if ( !alloc_cpumask_var(&masks->hard) )
>>> + return -ENOMEM;
>>> + if ( alloc_cpumask_var(&masks->soft) )
>>> + return 0;
>>> +
>>> + free_cpumask_var(masks->hard);
>>> + return -ENOMEM;
>>> +}
>>
>> Wouldn't this be a nice general helper function, also usable from
>> outside of this CU?
>
> I considered that, but wasn't sure this is really helpful. The only
> potential other user would be domain_update_node_affinity(), requiring
> to use the zalloc variant of the allocation in the helper (not that this
> would be a major problem, though).
I was actually thinking the other way around - the clearing of the masks
might better move into what is domain_update_node_affinity_noalloc() in
this version of the patch, so the helper could continue to use the non-
clearing allocations.
Jan
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |