[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] [PATCH for-next 07/16] xen/arm: Introduce copy_to_guest_phys_flush_dcache



On Fri, 8 Dec 2017, Julien Grall wrote:
> On 8 Dec 2017 22:26, "Stefano Stabellini" <sstabellini@xxxxxxxxxx> wrote:
>       On Fri, 8 Dec 2017, Julien Grall wrote:
>       > On 06/12/17 12:27, Julien Grall wrote:
>       > > On 12/06/2017 01:26 AM, Stefano Stabellini wrote:
>       > > > On Thu, 23 Nov 2017, Julien Grall wrote:
>       > > > > Hi Andrew,
>       > > > >
>       > > > > On 23/11/17 18:49, Andrew Cooper wrote:
>       > > > > > On 23/11/17 18:32, Julien Grall wrote:
>       > > > > > > This new function will be used in a follow-up patch to copy 
> data to
>       > > > > > > the
>       > > > > > > guest
>       > > > > > > using the IPA (aka guest physical address) and then clean 
> the cache.
>       > > > > > >
>       > > > > > > Signed-off-by: Julien Grall <julien.grall@xxxxxxxxxx>
>       > > > > > > ---
>       > > > > > >    xen/arch/arm/guestcopy.c           | 10 ++++++++++
>       > > > > > >    xen/include/asm-arm/guest_access.h |  6 ++++++
>       > > > > > >    2 files changed, 16 insertions(+)
>       > > > > > >
>       > > > > > > diff --git a/xen/arch/arm/guestcopy.c 
> b/xen/arch/arm/guestcopy.c
>       > > > > > > index be53bee559..7958663970 100644
>       > > > > > > --- a/xen/arch/arm/guestcopy.c
>       > > > > > > +++ b/xen/arch/arm/guestcopy.c
>       > > > > > > @@ -110,6 +110,16 @@ unsigned long raw_copy_from_guest(void 
> *to,
>       > > > > > > const
>       > > > > > > void __user *from, unsigned le
>       > > > > > >                          COPY_from_guest | COPY_linear);
>       > > > > > >    }
>       > > > > > >    +unsigned long copy_to_guest_phys_flush_dcache(struct 
> domain *d,
>       > > > > > > +                                              paddr_t gpa,
>       > > > > > > +                                              void *buf,
>       > > > > > > +                                              unsigned int 
> len)
>       > > > > > > +{
>       > > > > > > +    /* P2M is shared between all vCPUs, so the vCPU used 
> does not
>       > > > > > > matter.
>       > > > > > > */
>       > > > > >
>       > > > > > Be very careful with this line of thinking.  It is only works 
> after
>       > > > > > DOMCTL_max_vcpus has succeeded, and before that point, it is 
> a latent
>       > > > > > NULL pointer dereference.
>       > > > >
>       > > > > I really don't expect that function been used before 
> DOMCT_max_vcpus is
>       > > > > set.
>       > > > > It is only used for hardware emulation or Xen loading image 
> into the
>       > > > > hardware
>       > > > > domain memory. I could add a check d->vcpus to be safe.
>       > > > >
>       > > > > >
>       > > > > > Also, what about vcpus configured with alternative views?
>       > > > >
>       > > > > It is not important because the underlying call is 
> get_page_from_gfn
>       > > > > that does
>       > > > > not care about the alternative view (that function take a 
> domain in
>       > > > > parameter). I can update the comment.
>       > > > Since this is a new function, would it make sense to take a struct
>       > > > vcpu* as parameter, instead of a struct domain* ?
>       > >
>       > > Well, I suggested this patch this way because likely everyone will 
> use with
>       > > d->vcpus[0]. And then you would have to wonder why d->vcpus[0] and 
> not
>       > > d->vcpus[1]...
>       >
>       > Thinking a bit more to this, it might be better/safer to pass either 
> a domain
>       > or a vCPU to copy_guest. I can see 2 solutions:
>       >       1# Introduce a union that use the same parameter:
>       >               union
>       >               {
>       >                       struct
>       >                       {
>       >                               struct domain *d;
>       >                       } ipa;
>       >                       struct
>       >                       {
>       >                               struct vcpu *v;
>       >                       } gva;
>       >               }
>       >         The structure here would be to ensure that it is clear that 
> only
>       > domain (resp. vcpu) should be used with ipa (resp. gva).
>       >
>       >       2# Have 2 parameters, vcpu and domain.
>       >
>       > Any opinions?
> 
> I think that would be clearer. You could also add a paddr_t/vaddr_t
> correspondingly inside the union maybe.
> 
> 
> Well, you will have nameclash happening I think.
> 
> 
> And vaddr_t and paddr_t are confusing because you don't know which address 
> you speak about (guest vs hypervisor). At least ipa/gpa, gva are known naming.

That's not what I meant. ipa and gva are good names.

I was suggesting to put an additional address field inside the union to
avoid the issue with paddr_t and vaddr_t discussed elsewhere
(alpine.DEB.2.10.1712081313070.8052@sstabellini-ThinkPad-X260).

I am happy either way, it was just a suggestion.
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxxx
https://lists.xenproject.org/mailman/listinfo/xen-devel

 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.