[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH 1/3] VMX: Properly adjust the status of pi descriptor
> -----Original Message----- > From: Jan Beulich [mailto:JBeulich@xxxxxxxx] > Sent: Monday, May 23, 2016 8:31 PM > To: Wu, Feng <feng.wu@xxxxxxxxx> > Cc: andrew.cooper3@xxxxxxxxxx; dario.faggioli@xxxxxxxxxx; > george.dunlap@xxxxxxxxxxxxx; Tian, Kevin <kevin.tian@xxxxxxxxx>; xen- > devel@xxxxxxxxxxxxx; konrad.wilk@xxxxxxxxxx; keir@xxxxxxx > Subject: Re: [PATCH 1/3] VMX: Properly adjust the status of pi descriptor > > >>> On 20.05.16 at 10:53, <feng.wu@xxxxxxxxx> wrote: > > --- a/xen/arch/x86/hvm/vmx/vmx.c > > +++ b/xen/arch/x86/hvm/vmx/vmx.c > > @@ -107,12 +107,22 @@ void vmx_pi_per_cpu_init(unsigned int cpu) > > static void vmx_vcpu_block(struct vcpu *v) > > { > > unsigned long flags; > > - unsigned int dest; > > + unsigned int dest = cpu_physical_id(v->processor); > > spinlock_t *old_lock; > > spinlock_t *pi_blocking_list_lock = > > &per_cpu(vmx_pi_blocking, v->processor).lock; > > struct pi_desc *pi_desc = &v->arch.hvm_vmx.pi_desc; > > > > + if (v->arch.hvm_vmx.pi_back_from_hotplug == 1) > > Coding style (missing blanks). Also the flag is of boolean nature, so > it should be declared bool_t and you should drop the "== 1" here. > > > + { > > + write_atomic(&pi_desc->ndst, > > + x2apic_enabled ? dest : MASK_INSR(dest, > > PI_xAPIC_NDST_MASK)); > > + write_atomic(&pi_desc->nv, posted_intr_vector); > > + pi_clear_sn(pi_desc); > > + > > + v->arch.hvm_vmx.pi_back_from_hotplug = 0; > > + } > > + > > spin_lock_irqsave(pi_blocking_list_lock, flags); > > old_lock = cmpxchg(&v->arch.hvm_vmx.pi_blocking.lock, NULL, > > pi_blocking_list_lock); > > @@ -130,8 +140,6 @@ static void vmx_vcpu_block(struct vcpu *v) > > > > ASSERT(!pi_test_sn(pi_desc)); > > > > - dest = cpu_physical_id(v->processor); > > - > > ASSERT(pi_desc->ndst == > > (x2apic_enabled ? dest : MASK_INSR(dest, PI_xAPIC_NDST_MASK))); > > > > @@ -164,6 +172,16 @@ static void vmx_pi_do_resume(struct vcpu *v) > > unsigned long flags; > > spinlock_t *pi_blocking_list_lock; > > struct pi_desc *pi_desc = &v->arch.hvm_vmx.pi_desc; > > + unsigned int dest = cpu_physical_id(v->processor); > > + > > + if (v->arch.hvm_vmx.pi_back_from_hotplug == 1) > > + { > > + write_atomic(&pi_desc->ndst, > > + x2apic_enabled ? dest : MASK_INSR(dest, > > PI_xAPIC_NDST_MASK)); > > + pi_clear_sn(pi_desc); > > + > > + v->arch.hvm_vmx.pi_back_from_hotplug = 0; > > + } > > Compared with the adjustment above you don't write ->nv here, as > that happens ... > > > ASSERT(!test_bit(_VPF_blocked, &v->pause_flags)); > > after this ASSERT() (which suggests that your code addition would > better go here). That raises the question of moving the entire body > of the first hunk's if() into a helper function, which then would also > be used here. That would then also address my question about > ordering: Shouldn't pi_clear_sn() always happen last? > > Furthermore, are both of these additions really eliminating the > issue rather than just reducing the likelihood for it to occur? I.e. > what if the new flag gets set immediately after either of the if()-s > above? My original idea is: when we are setting the new flags in vmx_pi_hooks_assign(), the hooks are NULL pointer, so the new flag is always set before the checking point. However, seems there are still some race condition, such as: vmx_pi_hooks_assign() gets called right after vmx_pi_hooks_dessign(), so theoretically there might be some in-flight calls to the hooks when vmx_pi_hooks_assign() is being called. Hence the situation you mentioned above might happen.... > > > @@ -202,9 +220,14 @@ static void vmx_pi_do_resume(struct vcpu *v) > > /* This function is called when pcidevs_lock is held */ > > void vmx_pi_hooks_assign(struct domain *d) > > { > > + struct vcpu *v; > > + > > if ( !iommu_intpost || !has_hvm_container_domain(d) ) > > return; > > > > + for_each_vcpu ( d, v ) > > + v->arch.hvm_vmx.pi_back_from_hotplug = 1; > > Thinking of possible alternatives: What about doing the necessary > adjustments right here, instead of setting the new flag? > > Of course it may still be the case that the whole issue means we > indeed shouldn't zap those hooks upon device removal. Yes, even we do the adjustments here, we may still suffer from the race condition because as mentioned above, theoretically there might be in-fight calls to the hooks in vmx_pi_hooks_assign(). May need more thinking about the concern cases. Thanks, Feng > > Jan _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |