[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [RFC v1 08/15] Update IRTE according to guest interrupt config changes
> -----Original Message----- > From: Andrew Cooper [mailto:andrew.cooper3@xxxxxxxxxx] > Sent: Friday, March 27, 2015 3:59 AM > To: Wu, Feng; xen-devel@xxxxxxxxxxxxx > Cc: Zhang, Yang Z; Tian, Kevin; keir@xxxxxxx; JBeulich@xxxxxxxx > Subject: Re: [Xen-devel] [RFC v1 08/15] Update IRTE according to guest > interrupt config changes > > On 25/03/15 12:31, Feng Wu wrote: > > When guest changes its interrupt configuration (such as, vector, etc.) > > for direct-assigned devices, we need to update the associated IRTE > > with the new guest vector, so external interrupts from the assigned > > devices can be injected to guests without VM-Exit. > > > > For lowest-priority interrupts, we use vector-hashing mechamisn to find > > the destination vCPU. This follows the hardware behavior, since modern > > Intel CPUs use vector hashing to handle the lowest-priority interrupt. > > > > For multicase/broadcast vCPU, we cannot handle it via interrupt posting, > > still use interrupt remapping. > > > > Signed-off-by: Feng Wu <feng.wu@xxxxxxxxx> > > --- > > xen/drivers/passthrough/io.c | 77 > +++++++++++++++++++++++++++++++++++++++++++- > > 1 file changed, 76 insertions(+), 1 deletion(-) > > > > diff --git a/xen/drivers/passthrough/io.c b/xen/drivers/passthrough/io.c > > index ae050df..1d9a132 100644 > > --- a/xen/drivers/passthrough/io.c > > +++ b/xen/drivers/passthrough/io.c > > @@ -26,6 +26,7 @@ > > #include <asm/hvm/iommu.h> > > #include <asm/hvm/support.h> > > #include <xen/hvm/irq.h> > > +#include <asm/io_apic.h> > > > > static DEFINE_PER_CPU(struct list_head, dpci_list); > > > > @@ -199,6 +200,61 @@ void free_hvm_irq_dpci(struct hvm_irq_dpci *dpci) > > xfree(dpci); > > } > > > > +/* > > + * Here we handle the following cases: > > + * - For lowest-priority interrupts, we find the destination vCPU from the > > + * guest vector using vector-hashing mechamisn and return true. This > follows > > + * the hardware behavior, since modern Intel CPUs use vector hashing to > > + * handle the lowest-priority interrupt. > > What is the hashing algorithm, or can I have some hint as to where to > find it in a manual? I asked hardware guys about this, there is no document about how hardware implements the hashing algorithm. > > > + * - Otherwise, for single destination interrupt, it is straightforward to > > + * find the destination vCPU and return true. > > + * - For multicase/broadcast vCPU, we cannot handle it via interrupt > > posting, > > + * so return false. > > + */ > > +static bool_t pi_find_dest_vcpu(struct domain *d, uint8_t dest_id, > > + uint8_t dest_mode, uint8_t > deliver_mode, > > + uint32_t gvec, struct vcpu > **dest_vcpu) > > +{ > > + struct vcpu *v, **dest_vcpu_array; > > + unsigned int dest_vcpu_num = 0; > > + int ret; > > + > > + if ( deliver_mode == dest_LowestPrio ) > > + dest_vcpu_array = xzalloc_array(struct vcpu *, d->max_vcpus); > > This allocation can fail, but you really should see about avoiding it > entirely, if possible. > > > + > > + for_each_vcpu ( d, v ) > > + { > > + if ( !vlapic_match_dest(vcpu_vlapic(v), NULL, 0, > > + dest_id, dest_mode) ) > > + continue; > > + > > + dest_vcpu_num++; > > + > > + if ( deliver_mode == dest_LowestPrio ) > > + dest_vcpu_array[dest_vcpu_num] = v; > > + else > > + *dest_vcpu = v; > > + } > > + > > + if ( deliver_mode == dest_LowestPrio ) > > + { > > + if ( dest_vcpu_num != 0 ) > > + { > > + *dest_vcpu = dest_vcpu_array[gvec % dest_vcpu_num]; > > + ret = 1; > > + } > > + else > > + ret = 0; > > + > > + xfree(dest_vcpu_array); > > + return ret; > > + } > > + else if ( dest_vcpu_num == 1 ) > > + return 1; > > + else > > + return 0; > > +} > > + > > int pt_irq_create_bind( > > struct domain *d, xen_domctl_bind_pt_irq_t *pt_irq_bind) > > { > > @@ -257,7 +313,7 @@ int pt_irq_create_bind( > > { > > case PT_IRQ_TYPE_MSI: > > { > > - uint8_t dest, dest_mode; > > + uint8_t dest, dest_mode, deliver_mode; > > int dest_vcpu_id; > > > > if ( !(pirq_dpci->flags & HVM_IRQ_DPCI_MAPPED) ) > > @@ -330,11 +386,30 @@ int pt_irq_create_bind( > > /* Calculate dest_vcpu_id for MSI-type pirq migration. */ > > dest = pirq_dpci->gmsi.gflags & VMSI_DEST_ID_MASK; > > dest_mode = !!(pirq_dpci->gmsi.gflags & VMSI_DM_MASK); > > + deliver_mode = (pirq_dpci->gmsi.gflags >> > GFLAGS_SHIFT_DELIV_MODE) & > > + VMSI_DELIV_MASK; > > s/deliver/delivery/ > > Also, you should be able to use MASK_EXTR() rather than manual shifts > and masks. > > > dest_vcpu_id = hvm_girq_dest_2_vcpu_id(d, dest, dest_mode); > > pirq_dpci->gmsi.dest_vcpu_id = dest_vcpu_id; > > spin_unlock(&d->event_lock); > > if ( dest_vcpu_id >= 0 ) > > hvm_migrate_pirqs(d->vcpu[dest_vcpu_id]); > > + > > + /* Use interrupt posting if it is supported */ > > + if ( iommu_intpost ) > > + { > > + struct vcpu *vcpu = NULL; > > + > > + if ( !pi_find_dest_vcpu(d, dest, dest_mode, deliver_mode, > > + pirq_dpci->gmsi.gvec, &vcpu) ) > > + break; > > + > > + if ( pi_update_irte( vcpu, info, pirq_dpci->gmsi.gvec ) != 0 ) > > + { > > + dprintk(XENLOG_G_INFO, "failed to update PI IRTE\n"); > > Please put far more information this error message. > > > + return -EBUSY; > > Under what circumstances can this happen. I don't think it is valid to > fail the userspace bind hypercall in this case. Please see the reply to Konrad, BTW, do you have any sugguestion about this point? Thanks, Feng > > ~Andrew > > > + } > > + } > > + > > break; > > } > > _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |