[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [RFC v1 08/15] Update IRTE according to guest interrupt config changes
On 25/03/15 12:31, Feng Wu wrote: When guest changes its interrupt configuration (such as, vector, etc.) for direct-assigned devices, we need to update the associated IRTE with the new guest vector, so external interrupts from the assigned devices can be injected to guests without VM-Exit. For lowest-priority interrupts, we use vector-hashing mechamisn to find the destination vCPU. This follows the hardware behavior, since modern Intel CPUs use vector hashing to handle the lowest-priority interrupt. For multicase/broadcast vCPU, we cannot handle it via interrupt posting, still use interrupt remapping. Signed-off-by: Feng Wu <feng.wu@xxxxxxxxx> --- xen/drivers/passthrough/io.c | 77 +++++++++++++++++++++++++++++++++++++++++++- 1 file changed, 76 insertions(+), 1 deletion(-) diff --git a/xen/drivers/passthrough/io.c b/xen/drivers/passthrough/io.c index ae050df..1d9a132 100644 --- a/xen/drivers/passthrough/io.c +++ b/xen/drivers/passthrough/io.c @@ -26,6 +26,7 @@ #include <asm/hvm/iommu.h> #include <asm/hvm/support.h> #include <xen/hvm/irq.h> +#include <asm/io_apic.h>static DEFINE_PER_CPU(struct list_head, dpci_list); @@ -199,6 +200,61 @@ void free_hvm_irq_dpci(struct hvm_irq_dpci *dpci)xfree(dpci); }+/*+ * Here we handle the following cases: + * - For lowest-priority interrupts, we find the destination vCPU from the + * guest vector using vector-hashing mechamisn and return true. This follows + * the hardware behavior, since modern Intel CPUs use vector hashing to + * handle the lowest-priority interrupt. What is the hashing algorithm, or can I have some hint as to where to find it in a manual? + * - Otherwise, for single destination interrupt, it is straightforward to + * find the destination vCPU and return true. + * - For multicase/broadcast vCPU, we cannot handle it via interrupt posting, + * so return false. + */ +static bool_t pi_find_dest_vcpu(struct domain *d, uint8_t dest_id, + uint8_t dest_mode, uint8_t deliver_mode, + uint32_t gvec, struct vcpu **dest_vcpu) +{ + struct vcpu *v, **dest_vcpu_array; + unsigned int dest_vcpu_num = 0; + int ret; + + if ( deliver_mode == dest_LowestPrio ) + dest_vcpu_array = xzalloc_array(struct vcpu *, d->max_vcpus); This allocation can fail, but you really should see about avoiding it entirely, if possible. + + for_each_vcpu ( d, v ) + { + if ( !vlapic_match_dest(vcpu_vlapic(v), NULL, 0, + dest_id, dest_mode) ) + continue; + + dest_vcpu_num++; + + if ( deliver_mode == dest_LowestPrio ) + dest_vcpu_array[dest_vcpu_num] = v; + else + *dest_vcpu = v; + } + + if ( deliver_mode == dest_LowestPrio ) + { + if ( dest_vcpu_num != 0 ) + { + *dest_vcpu = dest_vcpu_array[gvec % dest_vcpu_num]; + ret = 1; + } + else + ret = 0; + + xfree(dest_vcpu_array); + return ret; + } + else if ( dest_vcpu_num == 1 ) + return 1; + else + return 0; +} + int pt_irq_create_bind( struct domain *d, xen_domctl_bind_pt_irq_t *pt_irq_bind) { @@ -257,7 +313,7 @@ int pt_irq_create_bind( { case PT_IRQ_TYPE_MSI: { - uint8_t dest, dest_mode; + uint8_t dest, dest_mode, deliver_mode; int dest_vcpu_id;if ( !(pirq_dpci->flags & HVM_IRQ_DPCI_MAPPED) )@@ -330,11 +386,30 @@ int pt_irq_create_bind( /* Calculate dest_vcpu_id for MSI-type pirq migration. */ dest = pirq_dpci->gmsi.gflags & VMSI_DEST_ID_MASK; dest_mode = !!(pirq_dpci->gmsi.gflags & VMSI_DM_MASK); + deliver_mode = (pirq_dpci->gmsi.gflags >> GFLAGS_SHIFT_DELIV_MODE) & + VMSI_DELIV_MASK; s/deliver/delivery/Also, you should be able to use MASK_EXTR() rather than manual shifts and masks. dest_vcpu_id = hvm_girq_dest_2_vcpu_id(d, dest, dest_mode); pirq_dpci->gmsi.dest_vcpu_id = dest_vcpu_id; spin_unlock(&d->event_lock); if ( dest_vcpu_id >= 0 ) hvm_migrate_pirqs(d->vcpu[dest_vcpu_id]); + + /* Use interrupt posting if it is supported */ + if ( iommu_intpost ) + { + struct vcpu *vcpu = NULL; + + if ( !pi_find_dest_vcpu(d, dest, dest_mode, deliver_mode, + pirq_dpci->gmsi.gvec, &vcpu) ) + break; + + if ( pi_update_irte( vcpu, info, pirq_dpci->gmsi.gvec ) != 0 ) + { + dprintk(XENLOG_G_INFO, "failed to update PI IRTE\n"); Please put far more information this error message. + return -EBUSY; Under what circumstances can this happen. I don't think it is valid to fail the userspace bind hypercall in this case. ~Andrew + } + } + break; } _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |