[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [PATCH v2] xen/arm: implement GICD_I[S/C]ACTIVER reads


  • To: Julien Grall <julien@xxxxxxx>
  • From: Bertrand Marquis <Bertrand.Marquis@xxxxxxx>
  • Date: Thu, 2 Apr 2020 08:17:45 +0000
  • Accept-language: en-GB, en-US
  • Arc-authentication-results: i=1; mx.microsoft.com 1; spf=pass smtp.mailfrom=arm.com; dmarc=pass action=none header.from=arm.com; dkim=pass header.d=arm.com; arc=none
  • Arc-message-signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=microsoft.com; s=arcselector9901; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-SenderADCheck; bh=49QIskKHJQLPIVNSMR8LT2XHuYRAqL92AqAUdN6HKhc=; b=CrQ9FQak13jfU0etHoCTdQ6BQXCjlFggQSjmocVqEHwUtcXgWeWTbsaf3gtMfPRTPuyrlicLYevMYqXGZJUKH/XvO8yXsQsBBUPQ7tqgFhI5f3TDvqxKG5Yz3LcO232g9uDSynO7Fs2hxUv4pfsvOi9NIrlL55k2Ck41udjvXdzoHQCJV9atBiBRCOeJmfwOPgIKLNXw6NzOkNHlANPPqnkZPB1OlMscCkrUOzVVZAEeRsvTovoU/tZ8Bo7ZfSekPr+hszv/Ulrupys/5zczR5Nk+YdA/OMqHEZtBnqShLvXMgK+Uh6STH9zBKdoUxshhRV8v0O2NbHwdBEGfNdf7g==
  • Arc-seal: i=1; a=rsa-sha256; s=arcselector9901; d=microsoft.com; cv=none; b=cgquZ9O/YP/s+uhrBR/I10AXRwsPBYxikxlqyE6w12NFV9gtfYJaLHGE2u0lgcHk12a5ba3n2/5nVrsUcqvA9fY4Z29KwUqaaDqrq7gAk9RGqnvXRDQGmiTAHRift0slY4IsYL3CpTCQY1wmZ4/WD19wkjN/oW19NUkG5nS0wH1Ahwlr+y+rtjYlCi3h+HFR5dBrG6x5nqW5dQ+IaoEwZjrcJUbH44F9mnceCsqHxUcfRXgmgGbVqibVhjtuk4R51zHJmuPgpj7pSRGQP4OAtCTuU5oIgst1b/EEVSh5O/dfUG/iXw/ZU9nsxAwUAEiNFBtLQRLOZi1sxL12Q6UUSQ==
  • Authentication-results: spf=pass (sender IP is 63.35.35.123) smtp.mailfrom=arm.com; lists.xenproject.org; dkim=pass (signature was verified) header.d=armh.onmicrosoft.com;lists.xenproject.org; dmarc=bestguesspass action=none header.from=arm.com;
  • Authentication-results-original: spf=none (sender IP is ) smtp.mailfrom=Bertrand.Marquis@xxxxxxx;
  • Cc: Peng Fan <peng.fan@xxxxxxx>, Stefano Stabellini <sstabellini@xxxxxxxxxx>, Stefano Stabellini <stefano.stabellini@xxxxxxxxxx>, "George.Dunlap@xxxxxxxxxxxxx" <George.Dunlap@xxxxxxxxxxxxx>, Wei Xu <xuwei5@xxxxxxxxxxxxx>, "xen-devel@xxxxxxxxxxxxxxxxxxxx" <xen-devel@xxxxxxxxxxxxxxxxxxxx>, nd <nd@xxxxxxx>
  • Delivery-date: Thu, 02 Apr 2020 08:18:15 +0000
  • List-id: Xen developer discussion <xen-devel.lists.xenproject.org>
  • Nodisclaimer: true
  • Original-authentication-results: spf=none (sender IP is ) smtp.mailfrom=Bertrand.Marquis@xxxxxxx;
  • Thread-index: AQHWB8CUPtV45V7pfkaBDuUahGESj6hj8BaAgAAXawCAAH2GAIAA+deA
  • Thread-topic: [PATCH v2] xen/arm: implement GICD_I[S/C]ACTIVER reads


> On 1 Apr 2020, at 18:23, Julien Grall <julien@xxxxxxx> wrote:
> 
> 
> 
> On 01/04/2020 10:54, Bertrand Marquis wrote:
>>> On 1 Apr 2020, at 09:30, Julien Grall <julien@xxxxxxx 
>>> <mailto:julien@xxxxxxx>> wrote:
>>> 
>>> 
>>> 
>>> On 01/04/2020 01:57, Stefano Stabellini wrote:
>>>> On Mon, 30 Mar 2020, Julien Grall wrote:
>>>>> Hi Stefano,
>>>>> 
>>>>> On 30/03/2020 17:35, Stefano Stabellini wrote:
>>>>>> On Sat, 28 Mar 2020, Julien Grall wrote:
>>>>>>> qHi Stefano,
>>>>>>> 
>>>>>>> On 27/03/2020 02:34, Stefano Stabellini wrote:
>>>>>>>> This is a simple implementation of GICD_ICACTIVER / GICD_ISACTIVER
>>>>>>>> reads. It doesn't take into account the latest state of interrupts on
>>>>>>>> other vCPUs. Only the current vCPU is up-to-date. A full solution is
>>>>>>>> not possible because it would require synchronization among all vCPUs,
>>>>>>>> which would be very expensive in terms or latency.
>>>>>>> 
>>>>>>> Your sentence suggests you have number showing that correctly emulating
>>>>>>> the
>>>>>>> registers would be too slow. Mind sharing them?
>>>>>> 
>>>>>> No, I don't have any numbers. Would you prefer a different wording or a
>>>>>> better explanation? I also realized there is a typo in there (or/of).
>>>>> Let me start with I think correctness is more important than speed.
>>>>> So I would have expected your commit message to contain some fact why
>>>>> synchronization is going to be slow and why this is a problem.
>>>>> 
>>>>> To give you a concrete example, the implementation of set/way 
>>>>> instructions are
>>>>> really slow (it could take a few seconds depending on the setup). However,
>>>>> this was fine because not implementing them correctly would have a greater
>>>>> impact on the guest (corruption) and they are not used often.
>>>>> 
>>>>> I don't think the performance in our case will be in same order 
>>>>> magnitude. It
>>>>> is most likely to be in the range of milliseconds (if not less) which I 
>>>>> think
>>>>> is acceptable for emulation (particularly for the vGIC) and the current 
>>>>> uses.
>>>> Writing on the mailing list some of our discussions today.
>>>> Correctness is not just in terms of compliance to a specification but it
>>>> is also about not breaking guests. Introducing latency in the range of
>>>> milliseconds, or hundreds of microseconds, would break any latency
>>>> sensitive workloads. We don't have numbers so we don't know for certain
>>>> the effect that your suggestion would have.
>>> 
>>> You missed part of the discussion. I don't disagree that latency is 
>>> important. However, if an implementation is only 95% reliable, then it 
>>> means 5% of the time your guest may break (corruption, crash, deadlock...). 
>>> At which point the latency is the last of your concern.
>>> 
>>>> It would be interesting to have those numbers, and I'll add to my TODO
>>>> list to run the experiments you suggested, but I'll put it on the
>>>> back-burner (from a Xilinx perspective it is low priority as no
>>>> customers are affected.)
>>> 
>>> How about we get a correct implementation merge first and then discuss 
>>> about optimization? This would allow the community to check whether there 
>>> are actually noticeable latency in their workload.
>> Hi,
> 
> Hi,

Hi,

> 
>> I am not sure that pushing something with a performance impact to later fix 
>> it is the right approach here.
>> The patch is an improvement compared to the current code and it can be 
>> further improved later to handle more cases (other cores).
> 
> If you read my other answer on this patch you will see that this is going to 
> introduce a deadlock in the guest using multiple vCPUs. How is it an 
> improvement compare to today?

I fully agree with the fact that a deadlock even with low probability is a 
no-go.
Current implementation returning always 0 is not creating this deadlock case 
but can mislead in the other way letting on CPU think that the interrupt has 
been handled already when it has possibly not been.

> 
>> If we really have to sync all vCPUs here, this will cost a lot and the 
>> result will still be the status in the past in fact because nothing will 
>> make sure that at the point the guest gets back the value it is still valid.
> 
> I hope you are aware the problem is exactly the same in the hardware. As soon 
> as you read the ISACTIVER, then the value may not be correct anymore. So I 
> don't see the issue to have an outdated value here.

Main difference being that in the hardware you can still poll for it and be 
sure that it won’t end up in your deadlock.
So I agree, we need to find a clean solution here.

> 
> As I said to Stefano yesterday, I would be happy to compromise as long as the 
> implementation does not introduce an outright deadlock in the guest.
> 
> At the moment, I don't have a better approach than kick all the vCPUs. Feel 
> free to suggest a better one.

Only one that I see is to enforce a service interrupt when interrupts have been 
handled to make sure that we clean up the internal active status as soon as 
interrupt has actually been handle.
This would ensure that the polling is behaving properly from the guest point of 
view but would prevent the cross cpu sync (status will be right as soon as the 
interrupt will have been serviced).
But the trade off will be a bigger overhead by enforcing a return back to xen 
each time and interrupt is handled by a guest.

I will try to spend more time on this and discuss with the GIC engineers at Arm 
to dig deeper in this case.

Cheers
Bertrand

> 
> Cheers,
> 
> -- 
> Julien Grall


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.