[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-changelog] [xen master] ioreq-server: centralize access to ioreq structures
commit dea8c75c82a8c24c9554fd1f31a0c4866aeae12d Author: Paul Durrant <paul.durrant@xxxxxxxxxx> AuthorDate: Mon May 12 12:01:43 2014 +0200 Commit: Jan Beulich <jbeulich@xxxxxxxx> CommitDate: Mon May 12 12:01:43 2014 +0200 ioreq-server: centralize access to ioreq structures To simplify creation of the ioreq server abstraction in a subsequent patch, this patch centralizes all use of the shared ioreq structure and the buffered ioreq ring to the source module xen/arch/x86/hvm/hvm.c. The patch moves an rmb() from inside hvm_io_assist() to hvm_do_resume() because the former may now be passed a data structure on stack, in which case the barrier is unnecessary. Signed-off-by: Paul Durrant <paul.durrant@xxxxxxxxxx> Acked-by: Jan Beulich <jbeulich@xxxxxxxx> Reviewed-by: Kevin Tian <kevin.tian@xxxxxxxxx> --- xen/arch/x86/hvm/emulate.c | 66 ++++++++++------------ xen/arch/x86/hvm/hvm.c | 111 ++++++++++++++++++++++++++++++++++++- xen/arch/x86/hvm/io.c | 104 ++-------------------------------- xen/arch/x86/hvm/vmx/vvmx.c | 13 ++++- xen/include/asm-x86/hvm/hvm.h | 15 +++++- xen/include/asm-x86/hvm/support.h | 21 +++---- 6 files changed, 181 insertions(+), 149 deletions(-) diff --git a/xen/arch/x86/hvm/emulate.c b/xen/arch/x86/hvm/emulate.c index 6d3522a..904c71a 100644 --- a/xen/arch/x86/hvm/emulate.c +++ b/xen/arch/x86/hvm/emulate.c @@ -57,24 +57,11 @@ static int hvmemul_do_io( int value_is_ptr = (p_data == NULL); struct vcpu *curr = current; struct hvm_vcpu_io *vio; - ioreq_t *p = get_ioreq(curr); - ioreq_t _ioreq; + ioreq_t p; unsigned long ram_gfn = paddr_to_pfn(ram_gpa); p2m_type_t p2mt; struct page_info *ram_page; int rc; - bool_t has_dm = 1; - - /* - * Domains without a backing DM, don't have an ioreq page. Just - * point to a struct on the stack, initialising the state as needed. - */ - if ( !p ) - { - has_dm = 0; - p = &_ioreq; - p->state = STATE_IOREQ_NONE; - } /* Check for paged out page */ ram_page = get_page_from_gfn(curr->domain, ram_gfn, &p2mt, P2M_UNSHARE); @@ -173,10 +160,9 @@ static int hvmemul_do_io( return X86EMUL_UNHANDLEABLE; } - if ( p->state != STATE_IOREQ_NONE ) + if ( hvm_io_pending(curr) ) { - gdprintk(XENLOG_WARNING, "WARNING: io already pending (%d)?\n", - p->state); + gdprintk(XENLOG_WARNING, "WARNING: io already pending?\n"); if ( ram_page ) put_page(ram_page); return X86EMUL_UNHANDLEABLE; @@ -193,38 +179,38 @@ static int hvmemul_do_io( if ( vio->mmio_retrying ) *reps = 1; - p->dir = dir; - p->data_is_ptr = value_is_ptr; - p->type = is_mmio ? IOREQ_TYPE_COPY : IOREQ_TYPE_PIO; - p->size = size; - p->addr = addr; - p->count = *reps; - p->df = df; - p->data = value; + p.dir = dir; + p.data_is_ptr = value_is_ptr; + p.type = is_mmio ? IOREQ_TYPE_COPY : IOREQ_TYPE_PIO; + p.size = size; + p.addr = addr; + p.count = *reps; + p.df = df; + p.data = value; if ( dir == IOREQ_WRITE ) - hvmtrace_io_assist(is_mmio, p); + hvmtrace_io_assist(is_mmio, &p); if ( is_mmio ) { - rc = hvm_mmio_intercept(p); + rc = hvm_mmio_intercept(&p); if ( rc == X86EMUL_UNHANDLEABLE ) - rc = hvm_buffered_io_intercept(p); + rc = hvm_buffered_io_intercept(&p); } else { - rc = hvm_portio_intercept(p); + rc = hvm_portio_intercept(&p); } switch ( rc ) { case X86EMUL_OKAY: case X86EMUL_RETRY: - *reps = p->count; - p->state = STATE_IORESP_READY; + *reps = p.count; + p.state = STATE_IORESP_READY; if ( !vio->mmio_retry ) { - hvm_io_assist(p); + hvm_io_assist(&p); vio->io_state = HVMIO_none; } else @@ -233,7 +219,7 @@ static int hvmemul_do_io( break; case X86EMUL_UNHANDLEABLE: /* If there is no backing DM, just ignore accesses */ - if ( !has_dm ) + if ( !hvm_has_dm(curr->domain) ) { rc = X86EMUL_OKAY; vio->io_state = HVMIO_none; @@ -241,7 +227,7 @@ static int hvmemul_do_io( else { rc = X86EMUL_RETRY; - if ( !hvm_send_assist_req() ) + if ( !hvm_send_assist_req(&p) ) vio->io_state = HVMIO_none; else if ( p_data == NULL ) rc = X86EMUL_OKAY; @@ -260,7 +246,7 @@ static int hvmemul_do_io( finish_access: if ( dir == IOREQ_READ ) - hvmtrace_io_assist(is_mmio, p); + hvmtrace_io_assist(is_mmio, &p); if ( p_data != NULL ) memcpy(p_data, &vio->io_data, size); @@ -1292,3 +1278,13 @@ struct segment_register *hvmemul_get_seg_reg( hvm_get_segment_register(current, seg, &hvmemul_ctxt->seg_reg[seg]); return &hvmemul_ctxt->seg_reg[seg]; } + +/* + * Local variables: + * mode: C + * c-file-style: "BSD" + * c-basic-offset: 4 + * tab-width: 4 + * indent-tabs-mode: nil + * End: + */ diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c index 43880c8..5e07bd2 100644 --- a/xen/arch/x86/hvm/hvm.c +++ b/xen/arch/x86/hvm/hvm.c @@ -363,6 +363,26 @@ void hvm_migrate_pirqs(struct vcpu *v) spin_unlock(&d->event_lock); } +static ioreq_t *get_ioreq(struct vcpu *v) +{ + struct domain *d = v->domain; + shared_iopage_t *p = d->arch.hvm_domain.ioreq.va; + + ASSERT((v == current) || spin_is_locked(&d->arch.hvm_domain.ioreq.lock)); + + return p ? &p->vcpu_ioreq[v->vcpu_id] : NULL; +} + +bool_t hvm_io_pending(struct vcpu *v) +{ + ioreq_t *p = get_ioreq(v); + + if ( !p ) + return 0; + + return p->state != STATE_IOREQ_NONE; +} + void hvm_do_resume(struct vcpu *v) { ioreq_t *p = get_ioreq(v); @@ -380,6 +400,7 @@ void hvm_do_resume(struct vcpu *v) switch ( p->state ) { case STATE_IORESP_READY: /* IORESP_READY -> NONE */ + rmb(); /* see IORESP_READY /then/ read contents of ioreq */ hvm_io_assist(p); break; case STATE_IOREQ_READY: /* IOREQ_{READY,INPROCESS} -> IORESP_READY */ @@ -1424,7 +1445,87 @@ void hvm_vcpu_down(struct vcpu *v) } } -bool_t hvm_send_assist_req(void) +int hvm_buffered_io_send(ioreq_t *p) +{ + struct vcpu *v = current; + struct domain *d = v->domain; + struct hvm_ioreq_page *iorp = &d->arch.hvm_domain.buf_ioreq; + buffered_iopage_t *pg = iorp->va; + buf_ioreq_t bp = { .data = p->data, + .addr = p->addr, + .type = p->type, + .dir = p->dir }; + /* Timeoffset sends 64b data, but no address. Use two consecutive slots. */ + int qw = 0; + + /* Ensure buffered_iopage fits in a page */ + BUILD_BUG_ON(sizeof(buffered_iopage_t) > PAGE_SIZE); + + /* + * Return 0 for the cases we can't deal with: + * - 'addr' is only a 20-bit field, so we cannot address beyond 1MB + * - we cannot buffer accesses to guest memory buffers, as the guest + * may expect the memory buffer to be synchronously accessed + * - the count field is usually used with data_is_ptr and since we don't + * support data_is_ptr we do not waste space for the count field either + */ + if ( (p->addr > 0xffffful) || p->data_is_ptr || (p->count != 1) ) + return 0; + + switch ( p->size ) + { + case 1: + bp.size = 0; + break; + case 2: + bp.size = 1; + break; + case 4: + bp.size = 2; + break; + case 8: + bp.size = 3; + qw = 1; + break; + default: + gdprintk(XENLOG_WARNING, "unexpected ioreq size: %u\n", p->size); + return 0; + } + + spin_lock(&iorp->lock); + + if ( (pg->write_pointer - pg->read_pointer) >= + (IOREQ_BUFFER_SLOT_NUM - qw) ) + { + /* The queue is full: send the iopacket through the normal path. */ + spin_unlock(&iorp->lock); + return 0; + } + + pg->buf_ioreq[pg->write_pointer % IOREQ_BUFFER_SLOT_NUM] = bp; + + if ( qw ) + { + bp.data = p->data >> 32; + pg->buf_ioreq[(pg->write_pointer+1) % IOREQ_BUFFER_SLOT_NUM] = bp; + } + + /* Make the ioreq_t visible /before/ write_pointer. */ + wmb(); + pg->write_pointer += qw ? 2 : 1; + + notify_via_xen_event_channel(d, d->arch.hvm_domain.params[HVM_PARAM_BUFIOREQ_EVTCHN]); + spin_unlock(&iorp->lock); + + return 1; +} + +bool_t hvm_has_dm(struct domain *d) +{ + return !!d->arch.hvm_domain.ioreq.va; +} + +bool_t hvm_send_assist_req(ioreq_t *proto_p) { struct vcpu *v = current; ioreq_t *p = get_ioreq(v); @@ -1443,14 +1544,18 @@ bool_t hvm_send_assist_req(void) return 0; } - prepare_wait_on_xen_event_channel(v->arch.hvm_vcpu.xen_port); + proto_p->state = STATE_IOREQ_NONE; + proto_p->vp_eport = p->vp_eport; + *p = *proto_p; + + prepare_wait_on_xen_event_channel(p->vp_eport); /* * Following happens /after/ blocking and setting up ioreq contents. * prepare_wait_on_xen_event_channel() is an implicit barrier. */ p->state = STATE_IOREQ_READY; - notify_via_xen_event_channel(v->domain, v->arch.hvm_vcpu.xen_port); + notify_via_xen_event_channel(v->domain, p->vp_eport); return 1; } diff --git a/xen/arch/x86/hvm/io.c b/xen/arch/x86/hvm/io.c index 44b4e20..f5ad9be 100644 --- a/xen/arch/x86/hvm/io.c +++ b/xen/arch/x86/hvm/io.c @@ -46,82 +46,6 @@ #include <xen/iocap.h> #include <public/hvm/ioreq.h> -int hvm_buffered_io_send(ioreq_t *p) -{ - struct vcpu *v = current; - struct domain *d = v->domain; - struct hvm_ioreq_page *iorp = &d->arch.hvm_domain.buf_ioreq; - buffered_iopage_t *pg = iorp->va; - buf_ioreq_t bp = { .data = p->data, - .addr = p->addr, - .type = p->type, - .dir = p->dir }; - /* Timeoffset sends 64b data, but no address. Use two consecutive slots. */ - int qw = 0; - - /* Ensure buffered_iopage fits in a page */ - BUILD_BUG_ON(sizeof(buffered_iopage_t) > PAGE_SIZE); - - /* - * Return 0 for the cases we can't deal with: - * - 'addr' is only a 20-bit field, so we cannot address beyond 1MB - * - we cannot buffer accesses to guest memory buffers, as the guest - * may expect the memory buffer to be synchronously accessed - * - the count field is usually used with data_is_ptr and since we don't - * support data_is_ptr we do not waste space for the count field either - */ - if ( (p->addr > 0xffffful) || p->data_is_ptr || (p->count != 1) ) - return 0; - - switch ( p->size ) - { - case 1: - bp.size = 0; - break; - case 2: - bp.size = 1; - break; - case 4: - bp.size = 2; - break; - case 8: - bp.size = 3; - qw = 1; - break; - default: - gdprintk(XENLOG_WARNING, "unexpected ioreq size: %u\n", p->size); - return 0; - } - - spin_lock(&iorp->lock); - - if ( (pg->write_pointer - pg->read_pointer) >= - (IOREQ_BUFFER_SLOT_NUM - qw) ) - { - /* The queue is full: send the iopacket through the normal path. */ - spin_unlock(&iorp->lock); - return 0; - } - - pg->buf_ioreq[pg->write_pointer % IOREQ_BUFFER_SLOT_NUM] = bp; - - if ( qw ) - { - bp.data = p->data >> 32; - pg->buf_ioreq[(pg->write_pointer+1) % IOREQ_BUFFER_SLOT_NUM] = bp; - } - - /* Make the ioreq_t visible /before/ write_pointer. */ - wmb(); - pg->write_pointer += qw ? 2 : 1; - - notify_via_xen_event_channel(d, - d->arch.hvm_domain.params[HVM_PARAM_BUFIOREQ_EVTCHN]); - spin_unlock(&iorp->lock); - - return 1; -} - void send_timeoffset_req(unsigned long timeoff) { ioreq_t p = { @@ -143,26 +67,14 @@ void send_timeoffset_req(unsigned long timeoff) /* Ask ioemu mapcache to invalidate mappings. */ void send_invalidate_req(void) { - struct vcpu *v = current; - ioreq_t *p = get_ioreq(v); - - if ( !p ) - return; - - if ( p->state != STATE_IOREQ_NONE ) - { - gdprintk(XENLOG_ERR, "WARNING: send invalidate req with something " - "already pending (%d)?\n", p->state); - domain_crash(v->domain); - return; - } - - p->type = IOREQ_TYPE_INVALIDATE; - p->size = 4; - p->dir = IOREQ_WRITE; - p->data = ~0UL; /* flush all */ + ioreq_t p = { + .type = IOREQ_TYPE_INVALIDATE, + .size = 4, + .dir = IOREQ_WRITE, + .data = ~0UL, /* flush all */ + }; - (void)hvm_send_assist_req(); + (void)hvm_send_assist_req(&p); } int handle_mmio(void) @@ -265,8 +177,6 @@ void hvm_io_assist(ioreq_t *p) struct hvm_vcpu_io *vio = &curr->arch.hvm_vcpu.hvm_io; enum hvm_io_state io_state; - rmb(); /* see IORESP_READY /then/ read contents of ioreq */ - p->state = STATE_IOREQ_NONE; io_state = vio->io_state; diff --git a/xen/arch/x86/hvm/vmx/vvmx.c b/xen/arch/x86/hvm/vmx/vvmx.c index e263376..9ccc03f 100644 --- a/xen/arch/x86/hvm/vmx/vvmx.c +++ b/xen/arch/x86/hvm/vmx/vvmx.c @@ -1394,7 +1394,6 @@ void nvmx_switch_guest(void) struct vcpu *v = current; struct nestedvcpu *nvcpu = &vcpu_nestedhvm(v); struct cpu_user_regs *regs = guest_cpu_user_regs(); - const ioreq_t *ioreq = get_ioreq(v); /* * A pending IO emulation may still be not finished. In this case, no @@ -1404,7 +1403,7 @@ void nvmx_switch_guest(void) * don't want to continue as this setup is not implemented nor supported * as of right now. */ - if ( !ioreq || ioreq->state != STATE_IOREQ_NONE ) + if ( hvm_io_pending(v) ) return; /* * a softirq may interrupt us between a virtual vmentry is @@ -2522,3 +2521,13 @@ void nvmx_set_cr_read_shadow(struct vcpu *v, unsigned int cr) /* nvcpu.guest_cr is what L2 write to cr actually. */ __vmwrite(read_shadow_field, v->arch.hvm_vcpu.nvcpu.guest_cr[cr]); } + +/* + * Local variables: + * mode: C + * c-file-style: "BSD" + * c-basic-offset: 4 + * tab-width: 4 + * indent-tabs-mode: nil + * End: + */ diff --git a/xen/include/asm-x86/hvm/hvm.h b/xen/include/asm-x86/hvm/hvm.h index 4fb7e22..251625d 100644 --- a/xen/include/asm-x86/hvm/hvm.h +++ b/xen/include/asm-x86/hvm/hvm.h @@ -26,6 +26,7 @@ #include <asm/hvm/asid.h> #include <public/domctl.h> #include <public/hvm/save.h> +#include <public/hvm/ioreq.h> #include <asm/mm.h> /* Interrupt acknowledgement sources. */ @@ -231,7 +232,7 @@ int prepare_ring_for_helper(struct domain *d, unsigned long gmfn, struct page_info **_page, void **_va); void destroy_ring_for_helper(void **_va, struct page_info *page); -bool_t hvm_send_assist_req(void); +bool_t hvm_send_assist_req(ioreq_t *p); void hvm_get_guest_pat(struct vcpu *v, u64 *guest_pat); int hvm_set_guest_pat(struct vcpu *v, u64 guest_pat); @@ -349,6 +350,8 @@ void hvm_hypervisor_cpuid_leaf(uint32_t sub_idx, void hvm_cpuid(unsigned int input, unsigned int *eax, unsigned int *ebx, unsigned int *ecx, unsigned int *edx); void hvm_migrate_timers(struct vcpu *v); +bool_t hvm_has_dm(struct domain *d); +bool_t hvm_io_pending(struct vcpu *v); void hvm_do_resume(struct vcpu *v); void hvm_migrate_pirqs(struct vcpu *v); @@ -545,3 +548,13 @@ bool_t nhvm_vmcx_hap_enabled(struct vcpu *v); enum hvm_intblk nhvm_interrupt_blocked(struct vcpu *v); #endif /* __ASM_X86_HVM_HVM_H__ */ + +/* + * Local variables: + * mode: C + * c-file-style: "BSD" + * c-basic-offset: 4 + * tab-width: 4 + * indent-tabs-mode: nil + * End: + */ diff --git a/xen/include/asm-x86/hvm/support.h b/xen/include/asm-x86/hvm/support.h index 1dc2f2d..05ef5c5 100644 --- a/xen/include/asm-x86/hvm/support.h +++ b/xen/include/asm-x86/hvm/support.h @@ -22,21 +22,10 @@ #define __ASM_X86_HVM_SUPPORT_H__ #include <xen/types.h> -#include <public/hvm/ioreq.h> #include <xen/sched.h> #include <xen/hvm/save.h> #include <asm/processor.h> -static inline ioreq_t *get_ioreq(struct vcpu *v) -{ - struct domain *d = v->domain; - shared_iopage_t *p = d->arch.hvm_domain.ioreq.va; - - ASSERT((v == current) || spin_is_locked(&d->arch.hvm_domain.ioreq.lock)); - - return p ? &p->vcpu_ioreq[v->vcpu_id] : NULL; -} - #define HVM_DELIVER_NO_ERROR_CODE -1 #ifndef NDEBUG @@ -144,3 +133,13 @@ int hvm_mov_to_cr(unsigned int cr, unsigned int gpr); int hvm_mov_from_cr(unsigned int cr, unsigned int gpr); #endif /* __ASM_X86_HVM_SUPPORT_H__ */ + +/* + * Local variables: + * mode: C + * c-file-style: "BSD" + * c-basic-offset: 4 + * tab-width: 4 + * indent-tabs-mode: nil + * End: + */ -- generated by git-patchbot for /home/xen/git/xen.git#master _______________________________________________ Xen-changelog mailing list Xen-changelog@xxxxxxxxxxxxx http://lists.xensource.com/xen-changelog
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |