[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-changelog] [xen-unstable] Rename evtchn_lock to event_lock, since it protects more than just
# HG changeset patch # User Keir Fraser <keir.fraser@xxxxxxxxxx> # Date 1223547471 -3600 # Node ID 0033c944318f266a0e367678bf9f46042ae03397 # Parent a11ad61bdb5b188a8116b533c87c31d6e9bd62d4 Rename evtchn_lock to event_lock, since it protects more than just event-channel state now. Signed-off-by: Keir Fraser <keir.fraser@xxxxxxxxxx> --- xen/arch/x86/hvm/svm/intr.c | 6 - xen/arch/x86/hvm/vmx/intr.c | 6 - xen/arch/x86/irq.c | 16 ++-- xen/arch/x86/physdev.c | 12 +-- xen/common/event_channel.c | 88 ++++++++++++------------ xen/drivers/passthrough/io.c | 36 ++++----- xen/drivers/passthrough/pci.c | 4 - xen/drivers/passthrough/vtd/x86/vtd.c | 6 - xen/include/asm-x86/domain.h | 2 xen/include/asm-x86/hvm/irq.h | 2 xen/include/xen/sched.h | 2 xen/xsm/acm/acm_simple_type_enforcement_hooks.c | 8 +- 12 files changed, 94 insertions(+), 94 deletions(-) diff -r a11ad61bdb5b -r 0033c944318f xen/arch/x86/hvm/svm/intr.c --- a/xen/arch/x86/hvm/svm/intr.c Thu Oct 09 11:14:52 2008 +0100 +++ b/xen/arch/x86/hvm/svm/intr.c Thu Oct 09 11:17:51 2008 +0100 @@ -124,11 +124,11 @@ static void svm_dirq_assist(struct vcpu if ( !test_and_clear_bit(irq, &hvm_irq_dpci->dirq_mask) ) continue; - spin_lock(&d->evtchn_lock); + spin_lock(&d->event_lock); if ( test_bit(_HVM_IRQ_DPCI_MSI, &hvm_irq_dpci->mirq[irq].flags) ) { hvm_pci_msi_assert(d, irq); - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); continue; } @@ -151,7 +151,7 @@ static void svm_dirq_assist(struct vcpu */ set_timer(&hvm_irq_dpci->hvm_timer[domain_irq_to_vector(d, irq)], NOW() + PT_IRQ_TIME_OUT); - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); } } diff -r a11ad61bdb5b -r 0033c944318f xen/arch/x86/hvm/vmx/intr.c --- a/xen/arch/x86/hvm/vmx/intr.c Thu Oct 09 11:14:52 2008 +0100 +++ b/xen/arch/x86/hvm/vmx/intr.c Thu Oct 09 11:17:51 2008 +0100 @@ -127,11 +127,11 @@ static void vmx_dirq_assist(struct vcpu if ( !test_and_clear_bit(irq, &hvm_irq_dpci->dirq_mask) ) continue; - spin_lock(&d->evtchn_lock); + spin_lock(&d->event_lock); if ( test_bit(_HVM_IRQ_DPCI_MSI, &hvm_irq_dpci->mirq[irq].flags) ) { hvm_pci_msi_assert(d, irq); - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); continue; } @@ -154,7 +154,7 @@ static void vmx_dirq_assist(struct vcpu */ set_timer(&hvm_irq_dpci->hvm_timer[domain_irq_to_vector(d, irq)], NOW() + PT_IRQ_TIME_OUT); - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); } } diff -r a11ad61bdb5b -r 0033c944318f xen/arch/x86/irq.c --- a/xen/arch/x86/irq.c Thu Oct 09 11:14:52 2008 +0100 +++ b/xen/arch/x86/irq.c Thu Oct 09 11:17:51 2008 +0100 @@ -514,7 +514,7 @@ int pirq_guest_bind(struct vcpu *v, int int rc = 0; cpumask_t cpumask = CPU_MASK_NONE; - WARN_ON(!spin_is_locked(&v->domain->evtchn_lock)); + WARN_ON(!spin_is_locked(&v->domain->event_lock)); BUG_ON(!local_irq_is_enabled()); retry: @@ -684,7 +684,7 @@ void pirq_guest_unbind(struct domain *d, irq_desc_t *desc; int vector; - WARN_ON(!spin_is_locked(&d->evtchn_lock)); + WARN_ON(!spin_is_locked(&d->event_lock)); BUG_ON(!local_irq_is_enabled()); desc = domain_spin_lock_irq_desc(d, irq, NULL); @@ -711,7 +711,7 @@ int pirq_guest_force_unbind(struct domai irq_guest_action_t *action; int i, bound = 0; - WARN_ON(!spin_is_locked(&d->evtchn_lock)); + WARN_ON(!spin_is_locked(&d->event_lock)); BUG_ON(!local_irq_is_enabled()); desc = domain_spin_lock_irq_desc(d, irq, NULL); @@ -738,7 +738,7 @@ int get_free_pirq(struct domain *d, int { int i; - ASSERT(spin_is_locked(&d->evtchn_lock)); + ASSERT(spin_is_locked(&d->event_lock)); if ( type == MAP_PIRQ_TYPE_GSI ) { @@ -768,7 +768,7 @@ int map_domain_pirq( irq_desc_t *desc; unsigned long flags; - ASSERT(spin_is_locked(&d->evtchn_lock)); + ASSERT(spin_is_locked(&d->event_lock)); if ( !IS_PRIV(current->domain) ) return -EPERM; @@ -836,7 +836,7 @@ int unmap_domain_pirq(struct domain *d, if ( !IS_PRIV(current->domain) ) return -EINVAL; - ASSERT(spin_is_locked(&d->evtchn_lock)); + ASSERT(spin_is_locked(&d->event_lock)); vector = d->arch.pirq_vector[pirq]; if ( vector <= 0 ) @@ -892,13 +892,13 @@ void free_domain_pirqs(struct domain *d) { int i; - spin_lock(&d->evtchn_lock); + spin_lock(&d->event_lock); for ( i = 0; i < NR_PIRQS; i++ ) if ( d->arch.pirq_vector[i] > 0 ) unmap_domain_pirq(d, i); - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); } extern void dump_ioapic_irq_info(void); diff -r a11ad61bdb5b -r 0033c944318f xen/arch/x86/physdev.c --- a/xen/arch/x86/physdev.c Thu Oct 09 11:14:52 2008 +0100 +++ b/xen/arch/x86/physdev.c Thu Oct 09 11:17:51 2008 +0100 @@ -100,7 +100,7 @@ static int physdev_map_pirq(struct physd } /* Verify or get pirq. */ - spin_lock(&d->evtchn_lock); + spin_lock(&d->event_lock); if ( map->pirq < 0 ) { if ( d->arch.vector_pirq[vector] ) @@ -145,7 +145,7 @@ static int physdev_map_pirq(struct physd map->pirq = pirq; done: - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); if ( (ret != 0) && (map->type == MAP_PIRQ_TYPE_MSI) && (map->index == -1) ) free_irq_vector(vector); free_domain: @@ -169,9 +169,9 @@ static int physdev_unmap_pirq(struct phy if ( d == NULL ) return -ESRCH; - spin_lock(&d->evtchn_lock); + spin_lock(&d->event_lock); ret = unmap_domain_pirq(d, unmap->pirq); - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); rcu_unlock_domain(d); @@ -298,10 +298,10 @@ ret_t do_physdev_op(int cmd, XEN_GUEST_H irq_op.vector = assign_irq_vector(irq); - spin_lock(&dom0->evtchn_lock); + spin_lock(&dom0->event_lock); ret = map_domain_pirq(dom0, irq_op.irq, irq_op.vector, MAP_PIRQ_TYPE_GSI, NULL); - spin_unlock(&dom0->evtchn_lock); + spin_unlock(&dom0->event_lock); if ( copy_to_guest(arg, &irq_op, 1) != 0 ) ret = -EFAULT; diff -r a11ad61bdb5b -r 0033c944318f xen/common/event_channel.c --- a/xen/common/event_channel.c Thu Oct 09 11:14:52 2008 +0100 +++ b/xen/common/event_channel.c Thu Oct 09 11:17:51 2008 +0100 @@ -133,7 +133,7 @@ static long evtchn_alloc_unbound(evtchn_ if ( rc ) return rc; - spin_lock(&d->evtchn_lock); + spin_lock(&d->event_lock); if ( (port = get_free_port(d)) < 0 ) ERROR_EXIT_DOM(port, d); @@ -150,7 +150,7 @@ static long evtchn_alloc_unbound(evtchn_ alloc->port = port; out: - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); rcu_unlock_domain(d); return rc; @@ -174,14 +174,14 @@ static long evtchn_bind_interdomain(evtc /* Avoid deadlock by first acquiring lock of domain with smaller id. */ if ( ld < rd ) { - spin_lock(&ld->evtchn_lock); - spin_lock(&rd->evtchn_lock); + spin_lock(&ld->event_lock); + spin_lock(&rd->event_lock); } else { if ( ld != rd ) - spin_lock(&rd->evtchn_lock); - spin_lock(&ld->evtchn_lock); + spin_lock(&rd->event_lock); + spin_lock(&ld->event_lock); } if ( (lport = get_free_port(ld)) < 0 ) @@ -216,9 +216,9 @@ static long evtchn_bind_interdomain(evtc bind->local_port = lport; out: - spin_unlock(&ld->evtchn_lock); + spin_unlock(&ld->event_lock); if ( ld != rd ) - spin_unlock(&rd->evtchn_lock); + spin_unlock(&rd->event_lock); rcu_unlock_domain(rd); @@ -244,7 +244,7 @@ static long evtchn_bind_virq(evtchn_bind ((v = d->vcpu[vcpu]) == NULL) ) return -ENOENT; - spin_lock(&d->evtchn_lock); + spin_lock(&d->event_lock); if ( v->virq_to_evtchn[virq] != 0 ) ERROR_EXIT(-EEXIST); @@ -260,7 +260,7 @@ static long evtchn_bind_virq(evtchn_bind v->virq_to_evtchn[virq] = bind->port = port; out: - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); return rc; } @@ -277,7 +277,7 @@ static long evtchn_bind_ipi(evtchn_bind_ (d->vcpu[vcpu] == NULL) ) return -ENOENT; - spin_lock(&d->evtchn_lock); + spin_lock(&d->event_lock); if ( (port = get_free_port(d)) < 0 ) ERROR_EXIT(port); @@ -289,7 +289,7 @@ static long evtchn_bind_ipi(evtchn_bind_ bind->port = port; out: - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); return rc; } @@ -308,7 +308,7 @@ static long evtchn_bind_pirq(evtchn_bind if ( !irq_access_permitted(d, pirq) ) return -EPERM; - spin_lock(&d->evtchn_lock); + spin_lock(&d->event_lock); if ( d->pirq_to_evtchn[pirq] != 0 ) ERROR_EXIT(-EEXIST); @@ -333,7 +333,7 @@ static long evtchn_bind_pirq(evtchn_bind bind->port = port; out: - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); return rc; } @@ -348,7 +348,7 @@ static long __evtchn_close(struct domain long rc = 0; again: - spin_lock(&d1->evtchn_lock); + spin_lock(&d1->event_lock); if ( !port_is_valid(d1, port1) ) { @@ -404,12 +404,12 @@ static long __evtchn_close(struct domain if ( d1 < d2 ) { - spin_lock(&d2->evtchn_lock); + spin_lock(&d2->event_lock); } else if ( d1 != d2 ) { - spin_unlock(&d1->evtchn_lock); - spin_lock(&d2->evtchn_lock); + spin_unlock(&d1->event_lock); + spin_lock(&d2->event_lock); goto again; } } @@ -454,11 +454,11 @@ static long __evtchn_close(struct domain if ( d2 != NULL ) { if ( d1 != d2 ) - spin_unlock(&d2->evtchn_lock); + spin_unlock(&d2->event_lock); put_domain(d2); } - spin_unlock(&d1->evtchn_lock); + spin_unlock(&d1->event_lock); return rc; } @@ -476,11 +476,11 @@ int evtchn_send(struct domain *d, unsign struct vcpu *rvcpu; int rport, ret = 0; - spin_lock(&ld->evtchn_lock); + spin_lock(&ld->event_lock); if ( unlikely(!port_is_valid(ld, lport)) ) { - spin_unlock(&ld->evtchn_lock); + spin_unlock(&ld->event_lock); return -EINVAL; } @@ -489,7 +489,7 @@ int evtchn_send(struct domain *d, unsign /* Guest cannot send via a Xen-attached event channel. */ if ( unlikely(lchn->consumer_is_xen) ) { - spin_unlock(&ld->evtchn_lock); + spin_unlock(&ld->event_lock); return -EINVAL; } @@ -527,7 +527,7 @@ int evtchn_send(struct domain *d, unsign } out: - spin_unlock(&ld->evtchn_lock); + spin_unlock(&ld->event_lock); return ret; } @@ -656,7 +656,7 @@ static long evtchn_status(evtchn_status_ if ( rc ) return rc; - spin_lock(&d->evtchn_lock); + spin_lock(&d->event_lock); if ( !port_is_valid(d, port) ) { @@ -704,7 +704,7 @@ static long evtchn_status(evtchn_status_ status->vcpu = chn->notify_vcpu_id; out: - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); rcu_unlock_domain(d); return rc; @@ -720,7 +720,7 @@ long evtchn_bind_vcpu(unsigned int port, if ( (vcpu_id >= ARRAY_SIZE(d->vcpu)) || (d->vcpu[vcpu_id] == NULL) ) return -ENOENT; - spin_lock(&d->evtchn_lock); + spin_lock(&d->event_lock); if ( !port_is_valid(d, port) ) { @@ -756,7 +756,7 @@ long evtchn_bind_vcpu(unsigned int port, } out: - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); return rc; } @@ -768,11 +768,11 @@ static long evtchn_unmask(evtchn_unmask_ int port = unmask->port; struct vcpu *v; - spin_lock(&d->evtchn_lock); + spin_lock(&d->event_lock); if ( unlikely(!port_is_valid(d, port)) ) { - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); return -EINVAL; } @@ -790,7 +790,7 @@ static long evtchn_unmask(evtchn_unmask_ vcpu_mark_events_pending(v); } - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); return 0; } @@ -944,7 +944,7 @@ int alloc_unbound_xen_event_channel( struct domain *d = local_vcpu->domain; int port; - spin_lock(&d->evtchn_lock); + spin_lock(&d->event_lock); if ( (port = get_free_port(d)) < 0 ) goto out; @@ -956,7 +956,7 @@ int alloc_unbound_xen_event_channel( chn->u.unbound.remote_domid = remote_domid; out: - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); return port; } @@ -968,11 +968,11 @@ void free_xen_event_channel( struct evtchn *chn; struct domain *d = local_vcpu->domain; - spin_lock(&d->evtchn_lock); + spin_lock(&d->event_lock); if ( unlikely(d->is_dying) ) { - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); return; } @@ -981,7 +981,7 @@ void free_xen_event_channel( BUG_ON(!chn->consumer_is_xen); chn->consumer_is_xen = 0; - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); (void)__evtchn_close(d, port); } @@ -993,7 +993,7 @@ void notify_via_xen_event_channel(int lp struct domain *ld = current->domain, *rd; int rport; - spin_lock(&ld->evtchn_lock); + spin_lock(&ld->event_lock); ASSERT(port_is_valid(ld, lport)); lchn = evtchn_from_port(ld, lport); @@ -1007,13 +1007,13 @@ void notify_via_xen_event_channel(int lp evtchn_set_pending(rd->vcpu[rchn->notify_vcpu_id], rport); } - spin_unlock(&ld->evtchn_lock); + spin_unlock(&ld->event_lock); } int evtchn_init(struct domain *d) { - spin_lock_init(&d->evtchn_lock); + spin_lock_init(&d->event_lock); if ( get_free_port(d) != 0 ) return -EINVAL; evtchn_from_port(d, 0)->state = ECS_RESERVED; @@ -1027,7 +1027,7 @@ void evtchn_destroy(struct domain *d) /* After this barrier no new event-channel allocations can occur. */ BUG_ON(!d->is_dying); - spin_barrier(&d->evtchn_lock); + spin_barrier(&d->event_lock); /* Close all existing event channels. */ for ( i = 0; port_is_valid(d, i); i++ ) @@ -1037,14 +1037,14 @@ void evtchn_destroy(struct domain *d) } /* Free all event-channel buckets. */ - spin_lock(&d->evtchn_lock); + spin_lock(&d->event_lock); for ( i = 0; i < NR_EVTCHN_BUCKETS; i++ ) { xsm_free_security_evtchn(d->evtchn[i]); xfree(d->evtchn[i]); d->evtchn[i] = NULL; } - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); } static void domain_dump_evtchn_info(struct domain *d) @@ -1053,7 +1053,7 @@ static void domain_dump_evtchn_info(stru printk("Domain %d polling vCPUs: %08lx\n", d->domain_id, d->poll_mask[0]); - if ( !spin_trylock(&d->evtchn_lock) ) + if ( !spin_trylock(&d->event_lock) ) return; printk("Event channel information for domain %d:\n", @@ -1094,7 +1094,7 @@ static void domain_dump_evtchn_info(stru printk(" x=%d\n", chn->consumer_is_xen); } - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); } static void dump_evtchn_info(unsigned char key) diff -r a11ad61bdb5b -r 0033c944318f xen/drivers/passthrough/io.c --- a/xen/drivers/passthrough/io.c Thu Oct 09 11:14:52 2008 +0100 +++ b/xen/drivers/passthrough/io.c Thu Oct 09 11:17:51 2008 +0100 @@ -30,7 +30,7 @@ static void pt_irq_time_out(void *data) struct dev_intx_gsi_link *digl; uint32_t device, intx; - spin_lock(&irq_map->dom->evtchn_lock); + spin_lock(&irq_map->dom->event_lock); dpci = domain_get_irq_dpci(irq_map->dom); ASSERT(dpci); @@ -46,7 +46,7 @@ static void pt_irq_time_out(void *data) clear_bit(machine_gsi, dpci->dirq_mask); vector = domain_irq_to_vector(irq_map->dom, machine_gsi); dpci->mirq[machine_gsi].pending = 0; - spin_unlock(&irq_map->dom->evtchn_lock); + spin_unlock(&irq_map->dom->event_lock); pirq_guest_eoi(irq_map->dom, machine_gsi); } @@ -62,7 +62,7 @@ int pt_irq_create_bind_vtd( if ( pirq < 0 || pirq >= NR_PIRQS ) return -EINVAL; - spin_lock(&d->evtchn_lock); + spin_lock(&d->event_lock); hvm_irq_dpci = domain_get_irq_dpci(d); if ( hvm_irq_dpci == NULL ) @@ -70,7 +70,7 @@ int pt_irq_create_bind_vtd( hvm_irq_dpci = xmalloc(struct hvm_irq_dpci); if ( hvm_irq_dpci == NULL ) { - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); return -ENOMEM; } memset(hvm_irq_dpci, 0, sizeof(*hvm_irq_dpci)); @@ -81,7 +81,7 @@ int pt_irq_create_bind_vtd( if ( domain_set_irq_dpci(d, hvm_irq_dpci) == 0 ) { xfree(hvm_irq_dpci); - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); return -EINVAL; } @@ -101,7 +101,7 @@ int pt_irq_create_bind_vtd( ||hvm_irq_dpci->msi_gvec_pirq[pt_irq_bind->u.msi.gvec] != pirq) { - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); return -EBUSY; } } @@ -117,7 +117,7 @@ int pt_irq_create_bind_vtd( digl = xmalloc(struct dev_intx_gsi_link); if ( !digl ) { - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); return -ENOMEM; } @@ -149,7 +149,7 @@ int pt_irq_create_bind_vtd( "VT-d irq bind: m_irq = %x device = %x intx = %x\n", machine_gsi, device, intx); } - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); return 0; } @@ -172,13 +172,13 @@ int pt_irq_destroy_bind_vtd( "pt_irq_destroy_bind_vtd: machine_gsi=%d " "guest_gsi=%d, device=%d, intx=%d.\n", machine_gsi, guest_gsi, device, intx); - spin_lock(&d->evtchn_lock); + spin_lock(&d->event_lock); hvm_irq_dpci = domain_get_irq_dpci(d); if ( hvm_irq_dpci == NULL ) { - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); return -EINVAL; } @@ -213,7 +213,7 @@ int pt_irq_destroy_bind_vtd( clear_bit(machine_gsi, hvm_irq_dpci->mapping); } } - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); gdprintk(XENLOG_INFO, "XEN_DOMCTL_irq_unmapping: m_irq = %x device = %x intx = %x\n", machine_gsi, device, intx); @@ -254,7 +254,7 @@ void hvm_dpci_msi_eoi(struct domain *d, if ( !iommu_enabled || (hvm_irq_dpci == NULL) ) return; - spin_lock(&d->evtchn_lock); + spin_lock(&d->event_lock); pirq = hvm_irq_dpci->msi_gvec_pirq[vector]; if ( ( pirq >= 0 ) && (pirq < NR_PIRQS) && @@ -265,7 +265,7 @@ void hvm_dpci_msi_eoi(struct domain *d, desc = domain_spin_lock_irq_desc(d, pirq, NULL); if (!desc) { - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); return; } @@ -275,7 +275,7 @@ void hvm_dpci_msi_eoi(struct domain *d, pirq_guest_eoi(d, pirq); } - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); } void hvm_dpci_eoi(struct domain *d, unsigned int guest_gsi, @@ -293,14 +293,14 @@ void hvm_dpci_eoi(struct domain *d, unsi return; } - spin_lock(&d->evtchn_lock); + spin_lock(&d->event_lock); hvm_irq_dpci = domain_get_irq_dpci(d); if((hvm_irq_dpci == NULL) || (guest_gsi >= NR_ISAIRQS && !hvm_irq_dpci->girq[guest_gsi].valid) ) { - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); return; } @@ -322,5 +322,5 @@ void hvm_dpci_eoi(struct domain *d, unsi pirq_guest_eoi(d, machine_gsi); } } - spin_unlock(&d->evtchn_lock); -} + spin_unlock(&d->event_lock); +} diff -r a11ad61bdb5b -r 0033c944318f xen/drivers/passthrough/pci.c --- a/xen/drivers/passthrough/pci.c Thu Oct 09 11:14:52 2008 +0100 +++ b/xen/drivers/passthrough/pci.c Thu Oct 09 11:17:51 2008 +0100 @@ -165,7 +165,7 @@ static void pci_clean_dpci_irqs(struct d if ( !is_hvm_domain(d) && !need_iommu(d) ) return; - spin_lock(&d->evtchn_lock); + spin_lock(&d->event_lock); hvm_irq_dpci = domain_get_irq_dpci(d); if ( hvm_irq_dpci != NULL ) { @@ -189,7 +189,7 @@ static void pci_clean_dpci_irqs(struct d d->arch.hvm_domain.irq.dpci = NULL; xfree(hvm_irq_dpci); } - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); } void pci_release_devices(struct domain *d) diff -r a11ad61bdb5b -r 0033c944318f xen/drivers/passthrough/vtd/x86/vtd.c --- a/xen/drivers/passthrough/vtd/x86/vtd.c Thu Oct 09 11:14:52 2008 +0100 +++ b/xen/drivers/passthrough/vtd/x86/vtd.c Thu Oct 09 11:17:51 2008 +0100 @@ -93,13 +93,13 @@ void hvm_dpci_isairq_eoi(struct domain * if ( !vtd_enabled) return; - spin_lock(&d->evtchn_lock); + spin_lock(&d->event_lock); dpci = domain_get_irq_dpci(d); if ( !dpci || !test_bit(isairq, dpci->isairq_map) ) { - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); return; } /* Multiple mirq may be mapped to one isa irq */ @@ -121,5 +121,5 @@ void hvm_dpci_isairq_eoi(struct domain * } } } - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); } diff -r a11ad61bdb5b -r 0033c944318f xen/include/asm-x86/domain.h --- a/xen/include/asm-x86/domain.h Thu Oct 09 11:14:52 2008 +0100 +++ b/xen/include/asm-x86/domain.h Thu Oct 09 11:17:51 2008 +0100 @@ -235,7 +235,7 @@ struct arch_domain /* Shadow translated domain: P2M mapping */ pagetable_t phys_table; - /* NB. protected by d->evtchn_lock and by irq_desc[vector].lock */ + /* NB. protected by d->event_lock and by irq_desc[vector].lock */ int vector_pirq[NR_VECTORS]; int pirq_vector[NR_PIRQS]; diff -r a11ad61bdb5b -r 0033c944318f xen/include/asm-x86/hvm/irq.h --- a/xen/include/asm-x86/hvm/irq.h Thu Oct 09 11:14:52 2008 +0100 +++ b/xen/include/asm-x86/hvm/irq.h Thu Oct 09 11:17:51 2008 +0100 @@ -63,7 +63,7 @@ struct hvm_girq_dpci_mapping { #define NR_ISAIRQS 16 #define NR_LINK 4 -/* Protected by domain's evtchn_lock */ +/* Protected by domain's event_lock */ struct hvm_irq_dpci { /* Machine IRQ to guest device/intx mapping. */ DECLARE_BITMAP(mapping, NR_PIRQS); diff -r a11ad61bdb5b -r 0033c944318f xen/include/xen/sched.h --- a/xen/include/xen/sched.h Thu Oct 09 11:14:52 2008 +0100 +++ b/xen/include/xen/sched.h Thu Oct 09 11:17:51 2008 +0100 @@ -188,7 +188,7 @@ struct domain /* Event channel information. */ struct evtchn *evtchn[NR_EVTCHN_BUCKETS]; - spinlock_t evtchn_lock; + spinlock_t event_lock; struct grant_table *grant_table; diff -r a11ad61bdb5b -r 0033c944318f xen/xsm/acm/acm_simple_type_enforcement_hooks.c --- a/xen/xsm/acm/acm_simple_type_enforcement_hooks.c Thu Oct 09 11:14:52 2008 +0100 +++ b/xen/xsm/acm/acm_simple_type_enforcement_hooks.c Thu Oct 09 11:17:51 2008 +0100 @@ -248,11 +248,11 @@ ste_init_state(struct acm_sized_buffer * /* a) check for event channel conflicts */ for ( bucket = 0; bucket < NR_EVTCHN_BUCKETS; bucket++ ) { - spin_lock(&d->evtchn_lock); + spin_lock(&d->event_lock); ports = d->evtchn[bucket]; if ( ports == NULL) { - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); break; } @@ -280,7 +280,7 @@ ste_init_state(struct acm_sized_buffer * printkd("%s: Policy violation in event channel domain " "%x -> domain %x.\n", __func__, d->domain_id, rdomid); - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); acm_array_append_tuple(errors, ACM_EVTCHN_SHARING_VIOLATION, @@ -288,7 +288,7 @@ ste_init_state(struct acm_sized_buffer * goto out; } } - spin_unlock(&d->evtchn_lock); + spin_unlock(&d->event_lock); } _______________________________________________ Xen-changelog mailing list Xen-changelog@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-changelog
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |