[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH v5 6/9] ioreq-server: add support for multiple servers
The previous single ioreq server that was created on demand now becomes the default server and an API is created to allow secondary servers, which handle specific IO ranges or PCI devices, to be added. When the guest issues an IO the list of secondary servers is checked for a matching IO range or PCI device. If none is found then the IO is passed to the default server. NOTE: To prevent emulators running in non-privileged guests from potentially allocating very large amounts of xen heap, the core rangeset code has been modified to introduce a hard limit of 256 ranges per set. This patch also introduces an implementation of asprintf() for Xen to allow meaningful names to be supplied for the secondary server rangesets. Secondary servers use guest pages to communicate with emulators, in the same way as the default server. These pages need to be in the guest physmap otherwise there is no suitable reference that can be queried by an emulator in order to map them. Therefore a pool of pages in the current E820 reserved region, just below the special pages is used. Secondary servers allocate from and free to this pool as they are created and destroyed. The size of the pool is currently hardcoded in the domain build at a value of 8. This should be sufficient for now and both the location and size of the pool can be modified in future without any need to change the API. Signed-off-by: Paul Durrant <paul.durrant@xxxxxxxxxx> Cc: Ian Jackson <ian.jackson@xxxxxxxxxxxxx> Cc: Stefano Stabellini <stefano.stabellini@xxxxxxxxxxxxx> Cc: Ian Campbell <ian.campbell@xxxxxxxxxx> Cc: Jan Beulich <jbeulich@xxxxxxxx> --- tools/libxc/xc_domain.c | 219 +++++++++ tools/libxc/xc_domain_restore.c | 45 ++ tools/libxc/xc_domain_save.c | 24 + tools/libxc/xc_hvm_build_x86.c | 30 +- tools/libxc/xenctrl.h | 133 ++++++ tools/libxc/xg_save_restore.h | 2 + xen/arch/x86/hvm/hvm.c | 933 ++++++++++++++++++++++++++++++++++---- xen/arch/x86/hvm/io.c | 2 +- xen/common/rangeset.c | 52 ++- xen/common/vsprintf.c | 56 +++ xen/include/asm-x86/hvm/domain.h | 22 +- xen/include/asm-x86/hvm/hvm.h | 1 + xen/include/public/hvm/hvm_op.h | 118 +++++ xen/include/public/hvm/ioreq.h | 15 +- xen/include/public/hvm/params.h | 5 +- xen/include/xen/lib.h | 4 + xen/include/xen/stdarg.h | 1 + 17 files changed, 1557 insertions(+), 105 deletions(-) diff --git a/tools/libxc/xc_domain.c b/tools/libxc/xc_domain.c index 369c3f3..b3ed029 100644 --- a/tools/libxc/xc_domain.c +++ b/tools/libxc/xc_domain.c @@ -1284,6 +1284,225 @@ int xc_get_hvm_param(xc_interface *handle, domid_t dom, int param, unsigned long return rc; } +int xc_hvm_create_ioreq_server(xc_interface *xch, + domid_t domid, + ioservid_t *id) +{ + DECLARE_HYPERCALL; + DECLARE_HYPERCALL_BUFFER(xen_hvm_create_ioreq_server_t, arg); + int rc; + + arg = xc_hypercall_buffer_alloc(xch, arg, sizeof(*arg)); + if ( arg == NULL ) + return -1; + + hypercall.op = __HYPERVISOR_hvm_op; + hypercall.arg[0] = HVMOP_create_ioreq_server; + hypercall.arg[1] = HYPERCALL_BUFFER_AS_ARG(arg); + + arg->domid = domid; + + rc = do_xen_hypercall(xch, &hypercall); + + *id = arg->id; + + xc_hypercall_buffer_free(xch, arg); + return rc; +} + +int xc_hvm_get_ioreq_server_info(xc_interface *xch, + domid_t domid, + ioservid_t id, + xen_pfn_t *ioreq_pfn, + xen_pfn_t *bufioreq_pfn, + evtchn_port_t *bufioreq_port) +{ + DECLARE_HYPERCALL; + DECLARE_HYPERCALL_BUFFER(xen_hvm_get_ioreq_server_info_t, arg); + int rc; + + arg = xc_hypercall_buffer_alloc(xch, arg, sizeof(*arg)); + if ( arg == NULL ) + return -1; + + hypercall.op = __HYPERVISOR_hvm_op; + hypercall.arg[0] = HVMOP_get_ioreq_server_info; + hypercall.arg[1] = HYPERCALL_BUFFER_AS_ARG(arg); + + arg->domid = domid; + arg->id = id; + + rc = do_xen_hypercall(xch, &hypercall); + if ( rc != 0 ) + goto done; + + if ( ioreq_pfn ) + *ioreq_pfn = arg->ioreq_pfn; + + if ( bufioreq_pfn ) + *bufioreq_pfn = arg->bufioreq_pfn; + + if ( bufioreq_port ) + *bufioreq_port = arg->bufioreq_port; + +done: + xc_hypercall_buffer_free(xch, arg); + return rc; +} + +int xc_hvm_map_io_range_to_ioreq_server(xc_interface *xch, domid_t domid, + ioservid_t id, int is_mmio, + uint64_t start, uint64_t end) +{ + DECLARE_HYPERCALL; + DECLARE_HYPERCALL_BUFFER(xen_hvm_io_range_t, arg); + int rc; + + arg = xc_hypercall_buffer_alloc(xch, arg, sizeof(*arg)); + if ( arg == NULL ) + return -1; + + hypercall.op = __HYPERVISOR_hvm_op; + hypercall.arg[0] = HVMOP_map_io_range_to_ioreq_server; + hypercall.arg[1] = HYPERCALL_BUFFER_AS_ARG(arg); + + arg->domid = domid; + arg->id = id; + arg->type = is_mmio ? HVMOP_IO_RANGE_MEMORY : HVMOP_IO_RANGE_PORT; + arg->start = start; + arg->end = end; + + rc = do_xen_hypercall(xch, &hypercall); + + xc_hypercall_buffer_free(xch, arg); + return rc; +} + +int xc_hvm_unmap_io_range_from_ioreq_server(xc_interface *xch, domid_t domid, + ioservid_t id, int is_mmio, + uint64_t start, uint64_t end) +{ + DECLARE_HYPERCALL; + DECLARE_HYPERCALL_BUFFER(xen_hvm_io_range_t, arg); + int rc; + + arg = xc_hypercall_buffer_alloc(xch, arg, sizeof(*arg)); + if ( arg == NULL ) + return -1; + + hypercall.op = __HYPERVISOR_hvm_op; + hypercall.arg[0] = HVMOP_unmap_io_range_from_ioreq_server; + hypercall.arg[1] = HYPERCALL_BUFFER_AS_ARG(arg); + + arg->domid = domid; + arg->id = id; + arg->type = is_mmio ? HVMOP_IO_RANGE_MEMORY : HVMOP_IO_RANGE_PORT; + arg->start = start; + arg->end = end; + + rc = do_xen_hypercall(xch, &hypercall); + + xc_hypercall_buffer_free(xch, arg); + return rc; +} + +int xc_hvm_map_pcidev_to_ioreq_server(xc_interface *xch, domid_t domid, + ioservid_t id, uint16_t segment, + uint8_t bus, uint8_t device, + uint8_t function) +{ + DECLARE_HYPERCALL; + DECLARE_HYPERCALL_BUFFER(xen_hvm_io_range_t, arg); + int rc; + + if (device > 0x1f || function > 0x7) { + errno = EINVAL; + return -1; + } + + arg = xc_hypercall_buffer_alloc(xch, arg, sizeof(*arg)); + if ( arg == NULL ) + return -1; + + hypercall.op = __HYPERVISOR_hvm_op; + hypercall.arg[0] = HVMOP_map_io_range_to_ioreq_server; + hypercall.arg[1] = HYPERCALL_BUFFER_AS_ARG(arg); + + arg->domid = domid; + arg->id = id; + arg->type = HVMOP_IO_RANGE_PCI; + arg->start = arg->end = HVMOP_PCI_SBDF((uint64_t)segment, + (uint64_t)bus, + (uint64_t)device, + (uint64_t)function); + + rc = do_xen_hypercall(xch, &hypercall); + + xc_hypercall_buffer_free(xch, arg); + return rc; +} + +int xc_hvm_unmap_pcidev_from_ioreq_server(xc_interface *xch, domid_t domid, + ioservid_t id, uint16_t segment, + uint8_t bus, uint8_t device, + uint8_t function) +{ + DECLARE_HYPERCALL; + DECLARE_HYPERCALL_BUFFER(xen_hvm_io_range_t, arg); + int rc; + + if (device > 0x1f || function > 0x7) { + errno = EINVAL; + return -1; + } + + arg = xc_hypercall_buffer_alloc(xch, arg, sizeof(*arg)); + if ( arg == NULL ) + return -1; + + hypercall.op = __HYPERVISOR_hvm_op; + hypercall.arg[0] = HVMOP_unmap_io_range_from_ioreq_server; + hypercall.arg[1] = HYPERCALL_BUFFER_AS_ARG(arg); + + arg->domid = domid; + arg->id = id; + arg->type = HVMOP_IO_RANGE_PCI; + arg->start = arg->end = HVMOP_PCI_SBDF((uint64_t)segment, + (uint64_t)bus, + (uint64_t)device, + (uint64_t)function); + + rc = do_xen_hypercall(xch, &hypercall); + + xc_hypercall_buffer_free(xch, arg); + return rc; +} + +int xc_hvm_destroy_ioreq_server(xc_interface *xch, + domid_t domid, + ioservid_t id) +{ + DECLARE_HYPERCALL; + DECLARE_HYPERCALL_BUFFER(xen_hvm_destroy_ioreq_server_t, arg); + int rc; + + arg = xc_hypercall_buffer_alloc(xch, arg, sizeof(*arg)); + if ( arg == NULL ) + return -1; + + hypercall.op = __HYPERVISOR_hvm_op; + hypercall.arg[0] = HVMOP_destroy_ioreq_server; + hypercall.arg[1] = HYPERCALL_BUFFER_AS_ARG(arg); + + arg->domid = domid; + arg->id = id; + + rc = do_xen_hypercall(xch, &hypercall); + + xc_hypercall_buffer_free(xch, arg); + return rc; +} + int xc_domain_setdebugging(xc_interface *xch, uint32_t domid, unsigned int enable) diff --git a/tools/libxc/xc_domain_restore.c b/tools/libxc/xc_domain_restore.c index bcb0ae0..af2bf3a 100644 --- a/tools/libxc/xc_domain_restore.c +++ b/tools/libxc/xc_domain_restore.c @@ -740,6 +740,8 @@ typedef struct { uint64_t acpi_ioport_location; uint64_t viridian; uint64_t vm_generationid_addr; + uint64_t ioreq_server_pfn; + uint64_t nr_ioreq_server_pages; struct toolstack_data_t tdata; } pagebuf_t; @@ -990,6 +992,26 @@ static int pagebuf_get_one(xc_interface *xch, struct restore_ctx *ctx, DPRINTF("read generation id buffer address"); return pagebuf_get_one(xch, ctx, buf, fd, dom); + case XC_SAVE_ID_HVM_IOREQ_SERVER_PFN: + /* Skip padding 4 bytes then read the ioreq server gmfn base. */ + if ( RDEXACT(fd, &buf->ioreq_server_pfn, sizeof(uint32_t)) || + RDEXACT(fd, &buf->ioreq_server_pfn, sizeof(uint64_t)) ) + { + PERROR("error read the ioreq server gmfn base"); + return -1; + } + return pagebuf_get_one(xch, ctx, buf, fd, dom); + + case XC_SAVE_ID_HVM_NR_IOREQ_SERVER_PAGES: + /* Skip padding 4 bytes then read the ioreq server gmfn count. */ + if ( RDEXACT(fd, &buf->nr_ioreq_server_pages, sizeof(uint32_t)) || + RDEXACT(fd, &buf->nr_ioreq_server_pages, sizeof(uint64_t)) ) + { + PERROR("error read the ioreq server gmfn count"); + return -1; + } + return pagebuf_get_one(xch, ctx, buf, fd, dom); + default: if ( (count > MAX_BATCH_SIZE) || (count < 0) ) { ERROR("Max batch size exceeded (%d). Giving up.", count); @@ -1748,6 +1770,29 @@ int xc_domain_restore(xc_interface *xch, int io_fd, uint32_t dom, if (pagebuf.viridian != 0) xc_set_hvm_param(xch, dom, HVM_PARAM_VIRIDIAN, 1); + /* + * If we are migrating in from a host that does not support + * secondary emulators then nr_ioreq_server_pages will be 0, since + * there will be no XC_SAVE_ID_HVM_NR_IOREQ_SERVER_PAGES chunk in + * the image. + * If we are migrating from a host that does support secondary + * emulators then the XC_SAVE_ID_HVM_NR_IOREQ_SERVER_PAGES chunk + * will exist and is guaranteed to have a non-zero value. The + * existence of that chunk also implies the existence of the + * XC_SAVE_ID_HVM_IOREQ_SERVER_PFN chunk, which is also guaranteed + * to have a non-zero value. + */ + if (pagebuf.nr_ioreq_server_pages != 0) { + if (pagebuf.ioreq_server_pfn != 0) { + xc_set_hvm_param(xch, dom, HVM_PARAM_NR_IOREQ_SERVER_PAGES, + pagebuf.nr_ioreq_server_pages); + xc_set_hvm_param(xch, dom, HVM_PARAM_IOREQ_SERVER_PFN, + pagebuf.ioreq_server_pfn); + } else { + ERROR("ioreq_server_pfn is invalid"); + } + } + if (pagebuf.acpi_ioport_location == 1) { DBGPRINTF("Use new firmware ioport from the checkpoint\n"); xc_set_hvm_param(xch, dom, HVM_PARAM_ACPI_IOPORTS_LOCATION, 1); diff --git a/tools/libxc/xc_domain_save.c b/tools/libxc/xc_domain_save.c index 71f9b59..acf3685 100644 --- a/tools/libxc/xc_domain_save.c +++ b/tools/libxc/xc_domain_save.c @@ -1737,6 +1737,30 @@ int xc_domain_save(xc_interface *xch, int io_fd, uint32_t dom, uint32_t max_iter PERROR("Error when writing the viridian flag"); goto out; } + + chunk.id = XC_SAVE_ID_HVM_IOREQ_SERVER_PFN; + chunk.data = 0; + xc_get_hvm_param(xch, dom, HVM_PARAM_IOREQ_SERVER_PFN, + (unsigned long *)&chunk.data); + + if ( (chunk.data != 0) && + wrexact(io_fd, &chunk, sizeof(chunk)) ) + { + PERROR("Error when writing the ioreq server gmfn base"); + goto out; + } + + chunk.id = XC_SAVE_ID_HVM_NR_IOREQ_SERVER_PAGES; + chunk.data = 0; + xc_get_hvm_param(xch, dom, HVM_PARAM_NR_IOREQ_SERVER_PAGES, + (unsigned long *)&chunk.data); + + if ( (chunk.data != 0) && + wrexact(io_fd, &chunk, sizeof(chunk)) ) + { + PERROR("Error when writing the ioreq server gmfn count"); + goto out; + } } if ( callbacks != NULL && callbacks->toolstack_save != NULL ) diff --git a/tools/libxc/xc_hvm_build_x86.c b/tools/libxc/xc_hvm_build_x86.c index dd3b522..3564e8b 100644 --- a/tools/libxc/xc_hvm_build_x86.c +++ b/tools/libxc/xc_hvm_build_x86.c @@ -49,6 +49,9 @@ #define NR_SPECIAL_PAGES 8 #define special_pfn(x) (0xff000u - NR_SPECIAL_PAGES + (x)) +#define NR_IOREQ_SERVER_PAGES 8 +#define ioreq_server_pfn(x) (special_pfn(0) - NR_IOREQ_SERVER_PAGES + (x)) + #define VGA_HOLE_SIZE (0x20) static int modules_init(struct xc_hvm_build_args *args, @@ -114,7 +117,7 @@ static void build_hvm_info(void *hvm_info_page, uint64_t mem_size, /* Memory parameters. */ hvm_info->low_mem_pgend = lowmem_end >> PAGE_SHIFT; hvm_info->high_mem_pgend = highmem_end >> PAGE_SHIFT; - hvm_info->reserved_mem_pgstart = special_pfn(0); + hvm_info->reserved_mem_pgstart = ioreq_server_pfn(0); /* Finish with the checksum. */ for ( i = 0, sum = 0; i < hvm_info->length; i++ ) @@ -502,6 +505,31 @@ static int setup_guest(xc_interface *xch, special_pfn(SPECIALPAGE_SHARING)); /* + * Allocate and clear additional ioreq server pages. The default + * server will use the IOREQ and BUFIOREQ special pages above. + */ + for ( i = 0; i < NR_IOREQ_SERVER_PAGES; i++ ) + { + xen_pfn_t pfn = ioreq_server_pfn(i); + + rc = xc_domain_populate_physmap_exact(xch, dom, 1, 0, 0, &pfn); + if ( rc != 0 ) + { + PERROR("Could not allocate %d'th ioreq server page.", i); + goto error_out; + } + + if ( xc_clear_domain_page(xch, dom, pfn) ) + goto error_out; + } + + /* Tell the domain where the pages are and how many there are */ + xc_set_hvm_param(xch, dom, HVM_PARAM_IOREQ_SERVER_PFN, + ioreq_server_pfn(0)); + xc_set_hvm_param(xch, dom, HVM_PARAM_NR_IOREQ_SERVER_PAGES, + NR_IOREQ_SERVER_PAGES); + + /* * Identity-map page table is required for running with CR0.PG=0 when * using Intel EPT. Create a 32-bit non-PAE page directory of superpages. */ diff --git a/tools/libxc/xenctrl.h b/tools/libxc/xenctrl.h index 02129f7..74fb738 100644 --- a/tools/libxc/xenctrl.h +++ b/tools/libxc/xenctrl.h @@ -1787,6 +1787,129 @@ void xc_clear_last_error(xc_interface *xch); int xc_set_hvm_param(xc_interface *handle, domid_t dom, int param, unsigned long value); int xc_get_hvm_param(xc_interface *handle, domid_t dom, int param, unsigned long *value); +/* + * IOREQ Server API. (See section on IOREQ Servers in public/hvm_op.h). + */ + +/** + * This function instantiates an IOREQ Server. + * + * @parm xch a handle to an open hypervisor interface. + * @parm domid the domain id to be serviced + * @parm id pointer to an ioservid_t to receive the IOREQ Server id. + * @return 0 on success, -1 on failure. + */ +int xc_hvm_create_ioreq_server(xc_interface *xch, + domid_t domid, + ioservid_t *id); + +/** + * This function retrieves the necessary information to allow an + * emulator to use an IOREQ Server. + * + * @parm xch a handle to an open hypervisor interface. + * @parm domid the domain id to be serviced + * @parm id the IOREQ Server id. + * @parm ioreq_pfn pointer to a xen_pfn_t to receive the synchronous ioreq gmfn + * @parm bufioreq_pfn pointer to a xen_pfn_t to receive the buffered ioreq gmfn + * @parm bufioreq_port pointer to a evtchn_port_t to receive the buffered ioreq event channel + * @return 0 on success, -1 on failure. + */ +int xc_hvm_get_ioreq_server_info(xc_interface *xch, + domid_t domid, + ioservid_t id, + xen_pfn_t *ioreq_pfn, + xen_pfn_t *bufioreq_pfn, + evtchn_port_t *bufioreq_port); + +/** + * This function registers a range of memory or I/O ports for emulation. + * + * @parm xch a handle to an open hypervisor interface. + * @parm domid the domain id to be serviced + * @parm id the IOREQ Server id. + * @parm is_mmio is this a range of ports or memory + * @parm start start of range + * @parm end end of range (inclusive). + * @return 0 on success, -1 on failure. + */ +int xc_hvm_map_io_range_to_ioreq_server(xc_interface *xch, + domid_t domid, + ioservid_t id, + int is_mmio, + uint64_t start, + uint64_t end); + +/** + * This function deregisters a range of memory or I/O ports for emulation. + * + * @parm xch a handle to an open hypervisor interface. + * @parm domid the domain id to be serviced + * @parm id the IOREQ Server id. + * @parm is_mmio is this a range of ports or memory + * @parm start start of range + * @parm end end of range (inclusive). + * @return 0 on success, -1 on failure. + */ +int xc_hvm_unmap_io_range_from_ioreq_server(xc_interface *xch, + domid_t domid, + ioservid_t id, + int is_mmio, + uint64_t start, + uint64_t end); + +/** + * This function registers a PCI device for config space emulation. + * + * @parm xch a handle to an open hypervisor interface. + * @parm domid the domain id to be serviced + * @parm id the IOREQ Server id. + * @parm segment the PCI segment of the device + * @parm bus the PCI bus of the device + * @parm device the 'slot' number of the device + * @parm function the function number of the device + * @return 0 on success, -1 on failure. + */ +int xc_hvm_map_pcidev_to_ioreq_server(xc_interface *xch, + domid_t domid, + ioservid_t id, + uint16_t segment, + uint8_t bus, + uint8_t device, + uint8_t function); + +/** + * This function deregisters a PCI device for config space emulation. + * + * @parm xch a handle to an open hypervisor interface. + * @parm domid the domain id to be serviced + * @parm id the IOREQ Server id. + * @parm segment the PCI segment of the device + * @parm bus the PCI bus of the device + * @parm device the 'slot' number of the device + * @parm function the function number of the device + * @return 0 on success, -1 on failure. + */ +int xc_hvm_unmap_pcidev_from_ioreq_server(xc_interface *xch, + domid_t domid, + ioservid_t id, + uint16_t segment, + uint8_t bus, + uint8_t device, + uint8_t function); + +/** + * This function destroys an IOREQ Server. + * + * @parm xch a handle to an open hypervisor interface. + * @parm domid the domain id to be serviced + * @parm id the IOREQ Server id. + * @return 0 on success, -1 on failure. + */ +int xc_hvm_destroy_ioreq_server(xc_interface *xch, + domid_t domid, + ioservid_t id); + /* HVM guest pass-through */ int xc_assign_device(xc_interface *xch, uint32_t domid, @@ -2425,3 +2548,13 @@ int xc_kexec_load(xc_interface *xch, uint8_t type, uint16_t arch, int xc_kexec_unload(xc_interface *xch, int type); #endif /* XENCTRL_H */ + +/* + * Local variables: + * mode: C + * c-file-style: "BSD" + * c-basic-offset: 4 + * tab-width: 4 + * indent-tabs-mode: nil + * End: + */ diff --git a/tools/libxc/xg_save_restore.h b/tools/libxc/xg_save_restore.h index f859621..f1ec7f5 100644 --- a/tools/libxc/xg_save_restore.h +++ b/tools/libxc/xg_save_restore.h @@ -259,6 +259,8 @@ #define XC_SAVE_ID_HVM_ACCESS_RING_PFN -16 #define XC_SAVE_ID_HVM_SHARING_RING_PFN -17 #define XC_SAVE_ID_TOOLSTACK -18 /* Optional toolstack specific info */ +#define XC_SAVE_ID_HVM_IOREQ_SERVER_PFN -19 +#define XC_SAVE_ID_HVM_NR_IOREQ_SERVER_PAGES -20 /* ** We process save/restore/migrate in batches of pages; the below diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c index bc073b5..5ac2d93 100644 --- a/xen/arch/x86/hvm/hvm.c +++ b/xen/arch/x86/hvm/hvm.c @@ -66,6 +66,7 @@ #include <asm/mem_event.h> #include <asm/mem_access.h> #include <public/mem_event.h> +#include <xen/rangeset.h> bool_t __read_mostly hvm_enabled; @@ -375,27 +376,36 @@ static ioreq_t *get_ioreq(struct hvm_ioreq_server *s, struct vcpu *v) bool_t hvm_io_pending(struct vcpu *v) { - struct hvm_ioreq_server *s = v->domain->arch.hvm_domain.ioreq_server; - ioreq_t *p; - - if ( !s ) - return 0; + struct domain *d = v->domain; + struct hvm_ioreq_server *s; + + list_for_each_entry ( s, + &d->arch.hvm_domain.ioreq_server_list, + list_entry ) + { + ioreq_t *p = get_ioreq(s, v); + + p = get_ioreq(s, v); + if ( p->state != STATE_IOREQ_NONE ) + return 1; + } - p = get_ioreq(s, v); - return p->state != STATE_IOREQ_NONE; + return 0; } void hvm_do_resume(struct vcpu *v) { struct domain *d = v->domain; - struct hvm_ioreq_server *s = d->arch.hvm_domain.ioreq_server; + struct hvm_ioreq_server *s; check_wakeup_from_wait(); if ( is_hvm_vcpu(v) ) pt_restore_timer(v); - if ( s ) + list_for_each_entry ( s, + &d->arch.hvm_domain.ioreq_server_list, + list_entry ) { ioreq_t *p = get_ioreq(s, v); @@ -429,6 +439,31 @@ void hvm_do_resume(struct vcpu *v) } } +static int hvm_alloc_ioreq_gmfn(struct domain *d, unsigned long *gmfn) +{ + unsigned int i; + int rc; + + rc = -ENOMEM; + for ( i = 0; i < d->arch.hvm_domain.ioreq_gmfn_count; i++ ) + { + if ( !test_and_set_bit(i, &d->arch.hvm_domain.ioreq_gmfn_mask) ) { + *gmfn = d->arch.hvm_domain.ioreq_gmfn_base + i; + rc = 0; + break; + } + } + + return rc; +} + +static void hvm_free_ioreq_gmfn(struct domain *d, unsigned long gmfn) +{ + unsigned int i = gmfn - d->arch.hvm_domain.ioreq_gmfn_base; + + clear_bit(i, &d->arch.hvm_domain.ioreq_gmfn_mask); +} + void destroy_ring_for_helper( void **_va, struct page_info *page) { @@ -513,6 +548,7 @@ static int hvm_map_ioreq_page( iorp->va = va; iorp->page = page; + iorp->gmfn = gmfn; return 0; } @@ -543,6 +579,127 @@ static int hvm_print_line( return X86EMUL_OKAY; } +static int hvm_access_cf8( + int dir, uint32_t port, uint32_t bytes, uint32_t *val) +{ + struct vcpu *curr = current; + struct hvm_domain *hd = &curr->domain->arch.hvm_domain; + int rc; + + BUG_ON(port < 0xcf8); + port -= 0xcf8; + + spin_lock(&hd->pci_lock); + + if ( dir == IOREQ_WRITE ) + { + switch ( bytes ) + { + case 4: + hd->pci_cf8 = *val; + break; + + case 2: + { + uint32_t mask = 0xffff << (port * 8); + uint32_t subval = *val << (port * 8); + + hd->pci_cf8 = (hd->pci_cf8 & ~mask) | + (subval & mask); + break; + } + + case 1: + { + uint32_t mask = 0xff << (port * 8); + uint32_t subval = *val << (port * 8); + + hd->pci_cf8 = (hd->pci_cf8 & ~mask) | + (subval & mask); + break; + } + + default: + break; + } + + /* We always need to fall through to the catch all emulator */ + rc = X86EMUL_UNHANDLEABLE; + } + else + { + switch ( bytes ) + { + case 4: + *val = hd->pci_cf8; + break; + + case 2: + *val = (hd->pci_cf8 >> (port * 8)) & 0xffff; + break; + + case 1: + *val = (hd->pci_cf8 >> (port * 8)) & 0xff; + break; + + default: + break; + } + + rc = X86EMUL_OKAY; + } + + spin_unlock(&hd->pci_lock); + + return rc; +} + +static int hvm_access_cfc( + int dir, uint32_t port, uint32_t bytes, uint32_t *val) +{ + struct vcpu *curr = current; + struct hvm_domain *hd = &curr->domain->arch.hvm_domain; + int rc; + + BUG_ON(port < 0xcfc); + port -= 0xcfc; + + spin_lock(&hd->pci_lock); + + if ( hd->pci_cf8 & (1 << 31) ) { + /* Fall through to an emulator */ + rc = X86EMUL_UNHANDLEABLE; + } else { + /* Config access disabled */ + if ( dir == IOREQ_READ ) + { + switch ( bytes ) + { + case 4: + *val = 0xffffffff; + break; + + case 2: + *val = 0xffff; + break; + + case 1: + *val = 0xff; + break; + + default: + break; + } + } + + rc = X86EMUL_OKAY; + } + + spin_unlock(&hd->pci_lock); + + return rc; +} + static int handle_pvh_io( int dir, uint32_t port, uint32_t bytes, uint32_t *val) { @@ -571,7 +728,7 @@ static void hvm_update_ioreq_evtchn(struct hvm_ioreq_server *s, } static int hvm_ioreq_server_add_vcpu(struct hvm_ioreq_server *s, - struct vcpu *v) + bool_t is_default, struct vcpu *v) { struct hvm_ioreq_vcpu *sv; int rc; @@ -599,8 +756,9 @@ static int hvm_ioreq_server_add_vcpu(struct hvm_ioreq_server *s, goto fail3; s->bufioreq_evtchn = rc; - d->arch.hvm_domain.params[HVM_PARAM_BUFIOREQ_EVTCHN] = - s->bufioreq_evtchn; + if ( is_default ) + d->arch.hvm_domain.params[HVM_PARAM_BUFIOREQ_EVTCHN] = + s->bufioreq_evtchn; } sv->vcpu = v; @@ -677,45 +835,128 @@ static void hvm_ioreq_server_remove_all_vcpus(struct hvm_ioreq_server *s) spin_unlock(&s->lock); } -static int hvm_ioreq_server_map_pages(struct hvm_ioreq_server *s) +static int hvm_ioreq_server_map_pages(struct hvm_ioreq_server *s, + bool_t is_default) { struct domain *d = s->domain; - unsigned long pfn; + unsigned long ioreq_pfn, bufioreq_pfn; int rc; - pfn = d->arch.hvm_domain.params[HVM_PARAM_IOREQ_PFN]; - rc = hvm_map_ioreq_page(s, 0, pfn); + if ( is_default ) { + ioreq_pfn = d->arch.hvm_domain.params[HVM_PARAM_IOREQ_PFN]; + bufioreq_pfn = d->arch.hvm_domain.params[HVM_PARAM_BUFIOREQ_PFN]; + } else { + rc = hvm_alloc_ioreq_gmfn(d, &ioreq_pfn); + if ( rc ) + goto fail1; + + rc = hvm_alloc_ioreq_gmfn(d, &bufioreq_pfn); + if ( rc ) + goto fail2; + } + + rc = hvm_map_ioreq_page(s, 0, ioreq_pfn); if ( rc ) - goto fail1; + goto fail3; - pfn = d->arch.hvm_domain.params[HVM_PARAM_BUFIOREQ_PFN]; - rc = hvm_map_ioreq_page(s, 1, pfn); + rc = hvm_map_ioreq_page(s, 1, bufioreq_pfn); if ( rc ) - goto fail2; + goto fail4; return 0; -fail2: +fail4: hvm_unmap_ioreq_page(s, 0); +fail3: + if ( !is_default ) + hvm_free_ioreq_gmfn(d, bufioreq_pfn); + +fail2: + if ( !is_default ) + hvm_free_ioreq_gmfn(d, ioreq_pfn); + fail1: return rc; } -static void hvm_ioreq_server_unmap_pages(struct hvm_ioreq_server *s) +static void hvm_ioreq_server_unmap_pages(struct hvm_ioreq_server *s, + bool_t is_default) { + struct domain *d = s->domain; + hvm_unmap_ioreq_page(s, 1); hvm_unmap_ioreq_page(s, 0); + + if ( !is_default ) { + hvm_free_ioreq_gmfn(d, s->bufioreq.gmfn); + hvm_free_ioreq_gmfn(d, s->ioreq.gmfn); + } +} + +static int hvm_ioreq_server_alloc_rangesets(struct hvm_ioreq_server *s, + bool_t is_default) +{ + int i; + int rc; + + if ( is_default ) + goto done; + + for ( i = 0; i < MAX_IO_RANGE_TYPE; i++ ) { + char *name; + + rc = asprintf(&name, "ioreq_server %d:%d %s", s->domid, s->id, + (i == HVMOP_IO_RANGE_PORT) ? "port" : + (i == HVMOP_IO_RANGE_MEMORY) ? "memory" : + (i == HVMOP_IO_RANGE_PCI) ? "pci" : + ""); + if ( rc ) + goto fail; + + s->range[i] = rangeset_new(s->domain, name, + RANGESETF_prettyprint_hex); + + xfree(name); + + rc = -ENOMEM; + if ( !s->range[i] ) + goto fail; + } + + done: + return 0; + + fail: + while ( --i >= 0 ) + rangeset_destroy(s->range[i]); + + return rc; +} + +static void hvm_ioreq_server_free_rangesets(struct hvm_ioreq_server *s, + bool_t is_default) +{ + int i; + + if ( is_default ) + return; + + for ( i = 0; i < MAX_IO_RANGE_TYPE; i++ ) + rangeset_destroy(s->range[i]); } static int hvm_ioreq_server_init(struct hvm_ioreq_server *s, struct domain *d, - domid_t domid) + domid_t domid, bool_t is_default, + ioservid_t id) { struct vcpu *v; int rc; - gdprintk(XENLOG_DEBUG, "%s %d\n", __func__, domid); + gdprintk(XENLOG_DEBUG, "%s %d:%d %s\n", __func__, domid, id, + is_default ? "[DEFAULT]" : ""); + s->id = id; s->domain = d; s->domid = domid; @@ -723,35 +964,49 @@ static int hvm_ioreq_server_init(struct hvm_ioreq_server *s, struct domain *d, INIT_LIST_HEAD(&s->ioreq_vcpu_list); spin_lock_init(&s->bufioreq_lock); - rc = hvm_ioreq_server_map_pages(s); + rc = hvm_ioreq_server_alloc_rangesets(s, is_default); if ( rc ) - return rc; + goto fail1; + + rc = hvm_ioreq_server_map_pages(s, is_default); + if ( rc ) + goto fail2; for_each_vcpu ( d, v ) { - rc = hvm_ioreq_server_add_vcpu(s, v); + rc = hvm_ioreq_server_add_vcpu(s, is_default, v); if ( rc ) - goto fail; + goto fail3; } return 0; - fail: + fail3: hvm_ioreq_server_remove_all_vcpus(s); - hvm_ioreq_server_unmap_pages(s); + hvm_ioreq_server_unmap_pages(s, is_default); + fail2: + hvm_ioreq_server_free_rangesets(s, is_default); + + fail1: + spin_unlock(&d->arch.hvm_domain.ioreq_server_lock); return rc; } -static void hvm_ioreq_server_deinit(struct hvm_ioreq_server *s) +static void hvm_ioreq_server_deinit(struct hvm_ioreq_server *s, + bool_t is_default) { - gdprintk(XENLOG_DEBUG, "%s %d\n", __func__, s->domid); + + gdprintk(XENLOG_DEBUG, "%s %d:%d %s\n", __func__, s->domid, s->id, + is_default ? "[DEFAULT]" : ""); hvm_ioreq_server_remove_all_vcpus(s); - hvm_ioreq_server_unmap_pages(s); + hvm_ioreq_server_unmap_pages(s, is_default); + hvm_ioreq_server_free_rangesets(s, is_default); } -static int hvm_create_ioreq_server(struct domain *d, domid_t domid) +static int hvm_create_ioreq_server(struct domain *d, domid_t domid, + bool_t is_default, ioservid_t *id) { struct hvm_ioreq_server *s; int rc; @@ -764,52 +1019,270 @@ static int hvm_create_ioreq_server(struct domain *d, domid_t domid) domain_pause(d); spin_lock(&d->arch.hvm_domain.ioreq_server_lock); - rc = -EEXIST; - if ( d->arch.hvm_domain.ioreq_server != NULL ) - goto fail2; + rc = -EEXIST; + if ( is_default && d->arch.hvm_domain.default_ioreq_server != NULL ) + goto fail2; + + rc = hvm_ioreq_server_init(s, d, domid, is_default, + d->arch.hvm_domain.ioreq_server_id++); + if ( rc ) + goto fail3; + + list_add(&s->list_entry, + &d->arch.hvm_domain.ioreq_server_list); + d->arch.hvm_domain.ioreq_server_count++; + + if ( is_default ) + d->arch.hvm_domain.default_ioreq_server = s; + + if (id != NULL) + *id = s->id; + + spin_unlock(&d->arch.hvm_domain.ioreq_server_lock); + domain_unpause(d); + + return 0; + + fail3: + fail2: + spin_unlock(&d->arch.hvm_domain.ioreq_server_lock); + domain_unpause(d); + + xfree(s); + fail1: + return rc; +} + +static void hvm_destroy_ioreq_server(struct domain *d, ioservid_t id) +{ + struct hvm_ioreq_server *s; + + spin_lock(&d->arch.hvm_domain.ioreq_server_lock); + + list_for_each_entry ( s, + &d->arch.hvm_domain.ioreq_server_list, + list_entry ) + { + bool_t is_default = ( s == d->arch.hvm_domain.default_ioreq_server); + + if ( s->id != id ) + continue; + + domain_pause(d); + + if ( is_default ) + d->arch.hvm_domain.default_ioreq_server = NULL; + + --d->arch.hvm_domain.ioreq_server_count; + list_del_init(&s->list_entry); + + hvm_ioreq_server_deinit(s, is_default); + + domain_unpause(d); + + xfree(s); + + break; + } + + spin_unlock(&d->arch.hvm_domain.ioreq_server_lock); +} + +static int hvm_get_ioreq_server_info(struct domain *d, ioservid_t id, + unsigned long *ioreq_pfn, + unsigned long *bufioreq_pfn, + evtchn_port_t *bufioreq_port) +{ + struct hvm_ioreq_server *s; + int rc; + + spin_lock(&d->arch.hvm_domain.ioreq_server_lock); + + rc = -ENOENT; + list_for_each_entry ( s, + &d->arch.hvm_domain.ioreq_server_list, + list_entry ) + { + if ( s->id != id ) + continue; + + *ioreq_pfn = s->ioreq.gmfn; + *bufioreq_pfn = s->bufioreq.gmfn; + *bufioreq_port = s->bufioreq_evtchn; + + rc = 0; + break; + } + + spin_unlock(&d->arch.hvm_domain.ioreq_server_lock); + + return rc; +} + +static int hvm_map_io_range_to_ioreq_server(struct domain *d, ioservid_t id, + uint32_t type, uint64_t start, uint64_t end) +{ + struct hvm_ioreq_server *s; + int rc; + + spin_lock(&d->arch.hvm_domain.ioreq_server_lock); + + rc = -ENOENT; + list_for_each_entry ( s, + &d->arch.hvm_domain.ioreq_server_list, + list_entry ) + { + if ( s->id == id ) + { + struct rangeset *r; + + switch ( type ) + { + case HVMOP_IO_RANGE_PORT: + case HVMOP_IO_RANGE_MEMORY: + case HVMOP_IO_RANGE_PCI: + r = s->range[type]; + break; + + default: + r = NULL; + break; + } + + rc = -EINVAL; + if ( !r ) + break; + + rc = rangeset_add_range(r, start, end); + break; + } + } + + spin_unlock(&d->arch.hvm_domain.ioreq_server_lock); + + return rc; +} + +static int hvm_unmap_io_range_from_ioreq_server(struct domain *d, ioservid_t id, + uint32_t type, uint64_t start, uint64_t end) +{ + struct hvm_ioreq_server *s; + int rc; + + spin_lock(&d->arch.hvm_domain.ioreq_server_lock); + + rc = -ENOENT; + list_for_each_entry ( s, + &d->arch.hvm_domain.ioreq_server_list, + list_entry ) + { + if ( s->id == id ) + { + struct rangeset *r; + + switch ( type ) + { + case HVMOP_IO_RANGE_PORT: + case HVMOP_IO_RANGE_MEMORY: + case HVMOP_IO_RANGE_PCI: + r = s->range[type]; + break; + + default: + r = NULL; + break; + } + + rc = -EINVAL; + if ( !r ) + break; + + rc = rangeset_remove_range(r, start, end); + break; + } + } + + spin_unlock(&d->arch.hvm_domain.ioreq_server_lock); + + return rc; +} + +static int hvm_all_ioreq_servers_add_vcpu(struct domain *d, struct vcpu *v) +{ + struct hvm_ioreq_server *s; + int rc; + + spin_lock(&d->arch.hvm_domain.ioreq_server_lock); - rc = hvm_ioreq_server_init(s, d, domid); - if ( rc ) - goto fail3; + list_for_each_entry ( s, + &d->arch.hvm_domain.ioreq_server_list, + list_entry ) + { + bool_t is_default = ( s == d->arch.hvm_domain.default_ioreq_server); - d->arch.hvm_domain.ioreq_server = s; + rc = hvm_ioreq_server_add_vcpu(s, is_default, v); + if ( rc ) + goto fail; + } spin_unlock(&d->arch.hvm_domain.ioreq_server_lock); - domain_unpause(d); return 0; - fail3: - fail2: + fail: + list_for_each_entry ( s, + &d->arch.hvm_domain.ioreq_server_list, + list_entry ) + hvm_ioreq_server_remove_vcpu(s, v); + spin_unlock(&d->arch.hvm_domain.ioreq_server_lock); - domain_unpause(d); - xfree(s); - fail1: return rc; } -static void hvm_destroy_ioreq_server(struct domain *d) +static void hvm_all_ioreq_servers_remove_vcpu(struct domain *d, struct vcpu *v) { struct hvm_ioreq_server *s; - domain_pause(d); spin_lock(&d->arch.hvm_domain.ioreq_server_lock); - s = d->arch.hvm_domain.ioreq_server; - if ( !s ) - goto done; + list_for_each_entry ( s, + &d->arch.hvm_domain.ioreq_server_list, + list_entry ) + hvm_ioreq_server_remove_vcpu(s, v); - d->arch.hvm_domain.ioreq_server = NULL; + spin_unlock(&d->arch.hvm_domain.ioreq_server_lock); +} - hvm_ioreq_server_deinit(s); +static void hvm_destroy_all_ioreq_servers(struct domain *d) +{ + struct hvm_ioreq_server *s, *next; - done: - spin_unlock(&d->arch.hvm_domain.ioreq_server_lock); - domain_unpause(d); + spin_lock(&d->arch.hvm_domain.ioreq_server_lock); + + list_for_each_entry_safe ( s, + next, + &d->arch.hvm_domain.ioreq_server_list, + list_entry ) + { + bool_t is_default = ( s == d->arch.hvm_domain.default_ioreq_server); + + domain_pause(d); + + if ( is_default ) + d->arch.hvm_domain.default_ioreq_server = NULL; + + --d->arch.hvm_domain.ioreq_server_count; + list_del_init(&s->list_entry); + + hvm_ioreq_server_deinit(s, is_default); + + domain_unpause(d); - if ( s ) xfree(s); + } + + spin_unlock(&d->arch.hvm_domain.ioreq_server_lock); } static int hvm_replace_event_channel(struct vcpu *v, domid_t remote_domid, @@ -834,7 +1307,7 @@ static int hvm_set_dm_domain(struct domain *d, domid_t domid) spin_lock(&d->arch.hvm_domain.ioreq_server_lock); - s = d->arch.hvm_domain.ioreq_server; + s = d->arch.hvm_domain.default_ioreq_server; if ( !s ) goto done; @@ -906,6 +1379,8 @@ int hvm_domain_initialise(struct domain *d) } spin_lock_init(&d->arch.hvm_domain.ioreq_server_lock); + INIT_LIST_HEAD(&d->arch.hvm_domain.ioreq_server_list); + spin_lock_init(&d->arch.hvm_domain.pci_lock); spin_lock_init(&d->arch.hvm_domain.irq_lock); spin_lock_init(&d->arch.hvm_domain.uc_lock); @@ -947,6 +1422,8 @@ int hvm_domain_initialise(struct domain *d) rtc_init(d); register_portio_handler(d, 0xe9, 1, hvm_print_line); + register_portio_handler(d, 0xcf8, 4, hvm_access_cf8); + register_portio_handler(d, 0xcfc, 4, hvm_access_cfc); rc = hvm_funcs.domain_initialise(d); if ( rc != 0 ) @@ -977,7 +1454,7 @@ void hvm_domain_relinquish_resources(struct domain *d) if ( hvm_funcs.nhvm_domain_relinquish_resources ) hvm_funcs.nhvm_domain_relinquish_resources(d); - hvm_destroy_ioreq_server(d); + hvm_destroy_all_ioreq_servers(d); msixtbl_pt_cleanup(d); @@ -1610,7 +2087,6 @@ int hvm_vcpu_initialise(struct vcpu *v) { int rc; struct domain *d = v->domain; - struct hvm_ioreq_server *s; hvm_asid_flush_vcpu(v); @@ -1653,14 +2129,7 @@ int hvm_vcpu_initialise(struct vcpu *v) && (rc = nestedhvm_vcpu_initialise(v)) < 0 ) /* teardown: nestedhvm_vcpu_destroy */ goto fail5; - spin_lock(&d->arch.hvm_domain.ioreq_server_lock); - - s = d->arch.hvm_domain.ioreq_server; - if ( s ) - rc = hvm_ioreq_server_add_vcpu(s, v); - - spin_unlock(&d->arch.hvm_domain.ioreq_server_lock); - + rc = hvm_all_ioreq_servers_add_vcpu(d, v); if ( rc != 0 ) goto fail6; @@ -1697,15 +2166,8 @@ int hvm_vcpu_initialise(struct vcpu *v) void hvm_vcpu_destroy(struct vcpu *v) { struct domain *d = v->domain; - struct hvm_ioreq_server *s; - - spin_lock(&d->arch.hvm_domain.ioreq_server_lock); - - s = d->arch.hvm_domain.ioreq_server; - if ( s ) - hvm_ioreq_server_remove_vcpu(s, v); - spin_unlock(&d->arch.hvm_domain.ioreq_server_lock); + hvm_all_ioreq_servers_remove_vcpu(d, v); nestedhvm_vcpu_destroy(v); @@ -1744,11 +2206,94 @@ void hvm_vcpu_down(struct vcpu *v) } } +static struct hvm_ioreq_server *hvm_select_ioreq_server(struct domain *d, + ioreq_t *p) +{ +#define CF8_BDF(cf8) (((cf8) & 0x00ffff00) >> 8) +#define CF8_ADDR(cf8) ((cf8) & 0x000000ff) + + struct hvm_ioreq_server *s; + uint8_t type; + uint64_t addr; + + if ( d->arch.hvm_domain.ioreq_server_count == 1 && + d->arch.hvm_domain.default_ioreq_server ) + goto done; + + if ( p->type == IOREQ_TYPE_PIO && + (p->addr & ~3) == 0xcfc ) + { + uint32_t cf8; + uint32_t sbdf; + + /* PCI config data cycle */ + type = IOREQ_TYPE_PCI_CONFIG; + + spin_lock(&d->arch.hvm_domain.pci_lock); + cf8 = d->arch.hvm_domain.pci_cf8; + sbdf = HVMOP_PCI_SBDF(0, + PCI_BUS(CF8_BDF(cf8)), + PCI_SLOT(CF8_BDF(cf8)), + PCI_FUNC(CF8_BDF(cf8))); + addr = ((uint64_t)sbdf << 32) | (CF8_ADDR(cf8) + (p->addr & 3)); + spin_unlock(&d->arch.hvm_domain.pci_lock); + } + else + { + type = p->type; + addr = p->addr; + } + + switch ( type ) + { + case IOREQ_TYPE_COPY: + case IOREQ_TYPE_PIO: + case IOREQ_TYPE_PCI_CONFIG: + break; + default: + goto done; + } + + list_for_each_entry ( s, + &d->arch.hvm_domain.ioreq_server_list, + list_entry ) + { + if ( s == d->arch.hvm_domain.default_ioreq_server ) + continue; + + switch ( type ) + { + case IOREQ_TYPE_COPY: + case IOREQ_TYPE_PIO: + if ( rangeset_contains_singleton(s->range[type], addr) ) + goto found; + + break; + case IOREQ_TYPE_PCI_CONFIG: + if ( rangeset_contains_singleton(s->range[type], addr >> 32) ) { + p->type = type; + p->addr = addr; + goto found; + } + + break; + } + } + + done: + s = d->arch.hvm_domain.default_ioreq_server; + + found: + return s; + +#undef CF8_ADDR +#undef CF8_BDF +} + int hvm_buffered_io_send(ioreq_t *p) { - struct vcpu *v = current; - struct domain *d = v->domain; - struct hvm_ioreq_server *s = d->arch.hvm_domain.ioreq_server; + struct domain *d = current->domain; + struct hvm_ioreq_server *s = hvm_select_ioreq_server(d, p); struct hvm_ioreq_page *iorp; buffered_iopage_t *pg; buf_ioreq_t bp; @@ -1830,22 +2375,19 @@ int hvm_buffered_io_send(ioreq_t *p) bool_t hvm_has_dm(struct domain *d) { - return !!d->arch.hvm_domain.ioreq_server; + return !list_empty(&d->arch.hvm_domain.ioreq_server_list); } -bool_t hvm_send_assist_req(ioreq_t *proto_p) +bool_t hvm_send_assist_req_to_ioreq_server(struct hvm_ioreq_server *s, + struct vcpu *v, + ioreq_t *proto_p) { - struct vcpu *v = current; struct domain *d = v->domain; - struct hvm_ioreq_server *s = d->arch.hvm_domain.ioreq_server; ioreq_t *p; if ( unlikely(!vcpu_start_shutdown_deferral(v)) ) return 0; /* implicitly bins the i/o operation */ - if ( !s ) - return 0; - p = get_ioreq(s, v); if ( unlikely(p->state != STATE_IOREQ_NONE) ) @@ -1867,11 +2409,35 @@ bool_t hvm_send_assist_req(ioreq_t *proto_p) * prepare_wait_on_xen_event_channel() is an implicit barrier. */ p->state = STATE_IOREQ_READY; - notify_via_xen_event_channel(v->domain, p->vp_eport); + notify_via_xen_event_channel(d, p->vp_eport); return 1; } +bool_t hvm_send_assist_req(ioreq_t *p) +{ + struct vcpu *v = current; + struct domain *d = v->domain; + struct hvm_ioreq_server *s = hvm_select_ioreq_server(d, p); + + if ( !s ) + return 0; + + return hvm_send_assist_req_to_ioreq_server(s, v, p); +} + +void hvm_broadcast_assist_req(ioreq_t *p) +{ + struct vcpu *v = current; + struct domain *d = v->domain; + struct hvm_ioreq_server *s; + + list_for_each_entry ( s, + &d->arch.hvm_domain.ioreq_server_list, + list_entry ) + (void) hvm_send_assist_req_to_ioreq_server(s, v, p); +} + void hvm_hlt(unsigned long rflags) { struct vcpu *curr = current; @@ -4473,6 +5039,145 @@ static int hvmop_flush_tlb_all(void) return 0; } +static int hvmop_create_ioreq_server( + XEN_GUEST_HANDLE_PARAM(xen_hvm_create_ioreq_server_t) uop) +{ + struct domain *curr_d = current->domain; + xen_hvm_create_ioreq_server_t op; + struct domain *d; + int rc; + + if ( copy_from_guest(&op, uop, 1) ) + return -EFAULT; + + rc = rcu_lock_remote_domain_by_id(op.domid, &d); + if ( rc != 0 ) + return rc; + + rc = -EINVAL; + if ( !is_hvm_domain(d) ) + goto out; + + rc = hvm_create_ioreq_server(d, curr_d->domain_id, 0, &op.id); + if ( rc != 0 ) + goto out; + + rc = copy_to_guest(uop, &op, 1) ? -EFAULT : 0; + + out: + rcu_unlock_domain(d); + return rc; +} + +static int hvmop_get_ioreq_server_info( + XEN_GUEST_HANDLE_PARAM(xen_hvm_get_ioreq_server_info_t) uop) +{ + xen_hvm_get_ioreq_server_info_t op; + struct domain *d; + int rc; + + if ( copy_from_guest(&op, uop, 1) ) + return -EFAULT; + + rc = rcu_lock_remote_domain_by_id(op.domid, &d); + if ( rc != 0 ) + return rc; + + rc = -EINVAL; + if ( !is_hvm_domain(d) ) + goto out; + + if ( (rc = hvm_get_ioreq_server_info(d, op.id, + &op.ioreq_pfn, + &op.bufioreq_pfn, + &op.bufioreq_port)) < 0 ) + goto out; + + rc = copy_to_guest(uop, &op, 1) ? -EFAULT : 0; + + out: + rcu_unlock_domain(d); + return rc; +} + +static int hvmop_map_io_range_to_ioreq_server( + XEN_GUEST_HANDLE_PARAM(xen_hvm_io_range_t) uop) +{ + xen_hvm_io_range_t op; + struct domain *d; + int rc; + + if ( copy_from_guest(&op, uop, 1) ) + return -EFAULT; + + rc = rcu_lock_remote_domain_by_id(op.domid, &d); + if ( rc != 0 ) + return rc; + + rc = -EINVAL; + if ( !is_hvm_domain(d) ) + goto out; + + rc = hvm_map_io_range_to_ioreq_server(d, op.id, op.type, + op.start, op.end); + + out: + rcu_unlock_domain(d); + return rc; +} + +static int hvmop_unmap_io_range_from_ioreq_server( + XEN_GUEST_HANDLE_PARAM(xen_hvm_io_range_t) uop) +{ + xen_hvm_io_range_t op; + struct domain *d; + int rc; + + if ( copy_from_guest(&op, uop, 1) ) + return -EFAULT; + + rc = rcu_lock_remote_domain_by_id(op.domid, &d); + if ( rc != 0 ) + return rc; + + rc = -EINVAL; + if ( !is_hvm_domain(d) ) + goto out; + + rc = hvm_unmap_io_range_from_ioreq_server(d, op.id, op.type, + op.start, op.end); + + out: + rcu_unlock_domain(d); + return rc; +} + +static int hvmop_destroy_ioreq_server( + XEN_GUEST_HANDLE_PARAM(xen_hvm_destroy_ioreq_server_t) uop) +{ + xen_hvm_destroy_ioreq_server_t op; + struct domain *d; + int rc; + + if ( copy_from_guest(&op, uop, 1) ) + return -EFAULT; + + rc = rcu_lock_remote_domain_by_id(op.domid, &d); + if ( rc != 0 ) + return rc; + + rc = -EINVAL; + if ( !is_hvm_domain(d) ) + goto out; + + hvm_destroy_ioreq_server(d, op.id); + rc = 0; + + out: + rcu_unlock_domain(d); + return rc; +} + #define HVMOP_op_mask 0xff long do_hvm_op(unsigned long op, XEN_GUEST_HANDLE_PARAM(void) arg) @@ -4484,6 +5189,31 @@ long do_hvm_op(unsigned long op, XEN_GUEST_HANDLE_PARAM(void) arg) switch ( op &= HVMOP_op_mask ) { + case HVMOP_create_ioreq_server: + rc = hvmop_create_ioreq_server( + guest_handle_cast(arg, xen_hvm_create_ioreq_server_t)); + break; + + case HVMOP_get_ioreq_server_info: + rc = hvmop_get_ioreq_server_info( + guest_handle_cast(arg, xen_hvm_get_ioreq_server_info_t)); + break; + + case HVMOP_map_io_range_to_ioreq_server: + rc = hvmop_map_io_range_to_ioreq_server( + guest_handle_cast(arg, xen_hvm_io_range_t)); + break; + + case HVMOP_unmap_io_range_from_ioreq_server: + rc = hvmop_unmap_io_range_from_ioreq_server( + guest_handle_cast(arg, xen_hvm_io_range_t)); + break; + + case HVMOP_destroy_ioreq_server: + rc = hvmop_destroy_ioreq_server( + guest_handle_cast(arg, xen_hvm_destroy_ioreq_server_t)); + break; + case HVMOP_set_param: case HVMOP_get_param: { @@ -4637,6 +5367,25 @@ long do_hvm_op(unsigned long op, XEN_GUEST_HANDLE_PARAM(void) arg) if ( a.value > SHUTDOWN_MAX ) rc = -EINVAL; break; + case HVM_PARAM_IOREQ_SERVER_PFN: + if ( d == current->domain ) { + rc = -EPERM; + break; + } + d->arch.hvm_domain.ioreq_gmfn_base = a.value; + break; + case HVM_PARAM_NR_IOREQ_SERVER_PAGES: + if ( d == current->domain ) { + rc = -EPERM; + break; + } + if ( a.value == 0 || + a.value > sizeof(unsigned long) * 8 ) { + rc = -EINVAL; + break; + } + d->arch.hvm_domain.ioreq_gmfn_count = a.value; + break; } if ( rc == 0 ) @@ -4670,6 +5419,12 @@ long do_hvm_op(unsigned long op, XEN_GUEST_HANDLE_PARAM(void) arg) case HVM_PARAM_ACPI_S_STATE: a.value = d->arch.hvm_domain.is_s3_suspended ? 3 : 0; break; + case HVM_PARAM_IOREQ_SERVER_PFN: + case HVM_PARAM_NR_IOREQ_SERVER_PAGES: + if ( d == current->domain ) { + rc = -EPERM; + break; + } case HVM_PARAM_IOREQ_PFN: case HVM_PARAM_BUFIOREQ_PFN: case HVM_PARAM_BUFIOREQ_EVTCHN: { @@ -4677,7 +5432,7 @@ long do_hvm_op(unsigned long op, XEN_GUEST_HANDLE_PARAM(void) arg) /* May need to create server */ domid = d->arch.hvm_domain.params[HVM_PARAM_DM_DOMAIN]; - rc = hvm_create_ioreq_server(d, domid); + rc = hvm_create_ioreq_server(d, domid, 1, NULL); if ( rc != 0 && rc != -EEXIST ) goto param_fail; /*FALLTHRU*/ diff --git a/xen/arch/x86/hvm/io.c b/xen/arch/x86/hvm/io.c index f5ad9be..bf68837 100644 --- a/xen/arch/x86/hvm/io.c +++ b/xen/arch/x86/hvm/io.c @@ -74,7 +74,7 @@ void send_invalidate_req(void) .data = ~0UL, /* flush all */ }; - (void)hvm_send_assist_req(&p); + hvm_broadcast_assist_req(&p); } int handle_mmio(void) diff --git a/xen/common/rangeset.c b/xen/common/rangeset.c index 2b986fb..ccda2f4 100644 --- a/xen/common/rangeset.c +++ b/xen/common/rangeset.c @@ -25,6 +25,9 @@ struct rangeset { /* Ordered list of ranges contained in this set, and protecting lock. */ struct list_head range_list; + unsigned int range_count; +# define MAX_RANGE_COUNT 256 + spinlock_t lock; /* Pretty-printing name. */ @@ -81,12 +84,32 @@ static void insert_range( /* Remove a range from its list and free it. */ static void destroy_range( - struct range *x) + struct rangeset *r, struct range *x) { + ASSERT(r->range_count != 0); + r->range_count--; + list_del(&x->list); xfree(x); } +/* Allocate a new range */ +static struct range *alloc_range( + struct rangeset *r) +{ + struct range *x; + + ASSERT(r->range_count <= MAX_RANGE_COUNT); + if ( r->range_count == MAX_RANGE_COUNT) + return NULL; + + x = xmalloc(struct range); + if ( x ) + r->range_count++; + + return x; +} + /***************************** * Core public functions */ @@ -108,7 +131,7 @@ int rangeset_add_range( { if ( (x == NULL) || ((x->e < s) && ((x->e + 1) != s)) ) { - x = xmalloc(struct range); + x = alloc_range(r); if ( x == NULL ) { rc = -ENOMEM; @@ -143,7 +166,7 @@ int rangeset_add_range( y = next_range(r, x); if ( (y == NULL) || (y->e > x->e) ) break; - destroy_range(y); + destroy_range(r, y); } } @@ -151,7 +174,7 @@ int rangeset_add_range( if ( (y != NULL) && ((x->e + 1) == y->s) ) { x->e = y->e; - destroy_range(y); + destroy_range(r, y); } out: @@ -179,7 +202,7 @@ int rangeset_remove_range( if ( (x->s < s) && (x->e > e) ) { - y = xmalloc(struct range); + y = alloc_range(r); if ( y == NULL ) { rc = -ENOMEM; @@ -193,7 +216,7 @@ int rangeset_remove_range( insert_range(r, x, y); } else if ( (x->s == s) && (x->e <= e) ) - destroy_range(x); + destroy_range(r, x); else if ( x->s == s ) x->s = e + 1; else if ( x->e <= e ) @@ -214,12 +237,12 @@ int rangeset_remove_range( { t = x; x = next_range(r, x); - destroy_range(t); + destroy_range(r, t); } x->s = e + 1; if ( x->s > x->e ) - destroy_range(x); + destroy_range(r, x); } out: @@ -312,6 +335,7 @@ struct rangeset *rangeset_new( spin_lock_init(&r->lock); INIT_LIST_HEAD(&r->range_list); + r->range_count = 0; BUG_ON(flags & ~RANGESETF_prettyprint_hex); r->flags = flags; @@ -351,7 +375,7 @@ void rangeset_destroy( } while ( (x = first_range(r)) != NULL ) - destroy_range(x); + destroy_range(r, x); xfree(r); } @@ -461,3 +485,13 @@ void rangeset_domain_printk( spin_unlock(&d->rangesets_lock); } + +/* + * Local variables: + * mode: C + * c-file-style: "BSD" + * c-basic-offset: 4 + * tab-width: 4 + * indent-tabs-mode: nil + * End: + */ diff --git a/xen/common/vsprintf.c b/xen/common/vsprintf.c index 8c43282..5e5816d 100644 --- a/xen/common/vsprintf.c +++ b/xen/common/vsprintf.c @@ -631,6 +631,62 @@ int scnprintf(char * buf, size_t size, const char *fmt, ...) } EXPORT_SYMBOL(scnprintf); +/** + * vasprintf - Format a string and allocate a buffer to place it in + * + * @bufp: Pointer to a pointer to receive the allocated buffer + * @fmt: The format string to use + * @args: Arguments for the format string + * + * -ENOMEM is returned on failure and @bufp is not touched. + * On success, 0 is returned. The buffer passed back is + * guaranteed to be null terminated. The memory is allocated + * from xenheap, so the buffer should be freed with xfree(). + */ +int vasprintf(char **bufp, const char *fmt, va_list args) +{ + va_list args_copy; + size_t size; + char *buf, dummy[1]; + + va_copy(args_copy, args); + size = vsnprintf(dummy, 0, fmt, args_copy); + va_end(args_copy); + + buf = _xmalloc(++size, 1); + if ( !buf ) + return -ENOMEM; + + (void) vsnprintf(buf, size, fmt, args); + + *bufp = buf; + return 0; +} +EXPORT_SYMBOL(vasprintf); + +/** + * asprintf - Format a string and place it in a buffer + * @bufp: Pointer to a pointer to receive the allocated buffer + * @fmt: The format string to use + * @...: Arguments for the format string + * + * -ENOMEM is returned on failure and @bufp is not touched. + * On success, 0 is returned. The buffer passed back is + * guaranteed to be null terminated. The memory is allocated + * from xenheap, so the buffer should be freed with xfree(). + */ +int asprintf(char **bufp, const char *fmt, ...) +{ + va_list args; + int i; + + va_start(args, fmt); + i=vasprintf(bufp,fmt,args); + va_end(args); + return i; +} +EXPORT_SYMBOL(asprintf); + /* * Local variables: * mode: C diff --git a/xen/include/asm-x86/hvm/domain.h b/xen/include/asm-x86/hvm/domain.h index cd885de..522e2c7 100644 --- a/xen/include/asm-x86/hvm/domain.h +++ b/xen/include/asm-x86/hvm/domain.h @@ -34,8 +34,10 @@ #include <public/grant_table.h> #include <public/hvm/params.h> #include <public/hvm/save.h> +#include <public/hvm/hvm_op.h> struct hvm_ioreq_page { + unsigned long gmfn; struct page_info *page; void *va; }; @@ -46,7 +48,10 @@ struct hvm_ioreq_vcpu { evtchn_port_t ioreq_evtchn; }; +#define MAX_IO_RANGE_TYPE (HVMOP_IO_RANGE_PCI + 1) + struct hvm_ioreq_server { + struct list_head list_entry; struct domain *domain; /* Lock to serialize toolstack modifications */ @@ -54,6 +59,7 @@ struct hvm_ioreq_server { /* Domain id of emulating domain */ domid_t domid; + ioservid_t id; struct hvm_ioreq_page ioreq; struct list_head ioreq_vcpu_list; struct hvm_ioreq_page bufioreq; @@ -61,11 +67,25 @@ struct hvm_ioreq_server { /* Lock to serialize access to buffered ioreq ring */ spinlock_t bufioreq_lock; evtchn_port_t bufioreq_evtchn; + struct rangeset *range[MAX_IO_RANGE_TYPE]; }; struct hvm_domain { - struct hvm_ioreq_server *ioreq_server; + /* Guest page range used for non-default ioreq servers */ + unsigned long ioreq_gmfn_base; + unsigned long ioreq_gmfn_mask; + unsigned int ioreq_gmfn_count; + + /* Lock protects all other values in the following block */ spinlock_t ioreq_server_lock; + ioservid_t ioreq_server_id; + unsigned int ioreq_server_count; + struct list_head ioreq_server_list; + struct hvm_ioreq_server *default_ioreq_server; + + /* Cached CF8 for guest PCI config cycles */ + uint32_t pci_cf8; + spinlock_t pci_lock; struct pl_time pl_time; diff --git a/xen/include/asm-x86/hvm/hvm.h b/xen/include/asm-x86/hvm/hvm.h index b0f7be5..28c2bd9 100644 --- a/xen/include/asm-x86/hvm/hvm.h +++ b/xen/include/asm-x86/hvm/hvm.h @@ -229,6 +229,7 @@ int prepare_ring_for_helper(struct domain *d, unsigned long gmfn, void destroy_ring_for_helper(void **_va, struct page_info *page); bool_t hvm_send_assist_req(ioreq_t *p); +void hvm_broadcast_assist_req(ioreq_t *p); void hvm_get_guest_pat(struct vcpu *v, u64 *guest_pat); int hvm_set_guest_pat(struct vcpu *v, u64 guest_pat); diff --git a/xen/include/public/hvm/hvm_op.h b/xen/include/public/hvm/hvm_op.h index f00f6d2..f29f2b6 100644 --- a/xen/include/public/hvm/hvm_op.h +++ b/xen/include/public/hvm/hvm_op.h @@ -23,6 +23,7 @@ #include "../xen.h" #include "../trace.h" +#include "../event_channel.h" /* Get/set subcommands: extra argument == pointer to xen_hvm_param struct. */ #define HVMOP_set_param 0 @@ -232,6 +233,123 @@ struct xen_hvm_inject_msi { typedef struct xen_hvm_inject_msi xen_hvm_inject_msi_t; DEFINE_XEN_GUEST_HANDLE(xen_hvm_inject_msi_t); +/* + * IOREQ Servers + * + * The interface between an I/O emulator an Xen is called an IOREQ Server. + * A domain supports a single 'legacy' IOREQ Server which is instantiated if + * parameter... + * + * HVM_PARAM_IOREQ_PFN is read (to get the gmfn containing the synchronous + * ioreq structures), or... + * HVM_PARAM_BUFIOREQ_PFN is read (to get the gmfn containing the buffered + * ioreq ring), or... + * HVM_PARAM_BUFIOREQ_EVTCHN is read (to get the event channel that Xen uses + * to request buffered I/O emulation). + * + * The following hypercalls facilitate the creation of IOREQ Servers for + * 'secondary' emulators which are invoked to implement port I/O, memory, or + * PCI config space ranges which they explicitly register. + */ + +typedef uint16_t ioservid_t; +DEFINE_XEN_GUEST_HANDLE(ioservid_t); + +/* + * HVMOP_create_ioreq_server: Instantiate a new IOREQ Server for a secondary + * emulator servicing domain <domid>. + * + * The <id> handed back is unique for <domid>. + */ +#define HVMOP_create_ioreq_server 17 +struct xen_hvm_create_ioreq_server { + domid_t domid; /* IN - domain to be serviced */ + ioservid_t id; /* OUT - server id */ +}; +typedef struct xen_hvm_create_ioreq_server xen_hvm_create_ioreq_server_t; +DEFINE_XEN_GUEST_HANDLE(xen_hvm_create_ioreq_server_t); + +/* + * HVMOP_get_ioreq_server_info: Get all the information necessary to access + * IOREQ Server <id>. + * + * The emulator needs to map the synchronous ioreq structures and buffered + * ioreq ring that Xen uses to request emulation. These are hosted in domain + * <domid>'s gmfns <ioreq_pfn> and <bufioreq_pfn> respectively. In addition the + * emulator needs to bind to event channel <bufioreq_port> to listen for + * buffered emulation requests. (The event channels used for synchronous + * emulation requests are specified in the per-CPU ioreq structures in + * <ioreq_pfn>). + */ +#define HVMOP_get_ioreq_server_info 18 +struct xen_hvm_get_ioreq_server_info { + domid_t domid; /* IN - domain to be serviced */ + ioservid_t id; /* IN - server id */ + xen_pfn_t ioreq_pfn; /* OUT - sync ioreq pfn */ + xen_pfn_t bufioreq_pfn; /* OUT - buffered ioreq pfn */ + evtchn_port_t bufioreq_port; /* OUT - buffered ioreq port */ +}; +typedef struct xen_hvm_get_ioreq_server_info xen_hvm_get_ioreq_server_info_t; +DEFINE_XEN_GUEST_HANDLE(xen_hvm_get_ioreq_server_info_t); + +/* + * HVM_map_io_range_to_ioreq_server: Register an I/O range of domain <domid> + * for emulation by the client of IOREQ + * Server <id> + * HVM_unmap_io_range_from_ioreq_server: Deregister an I/O range of <domid> + * for emulation by the client of IOREQ + * Server <id> + * + * There are three types of I/O that can be emulated: port I/O, memory accesses + * and PCI config space accesses. The <type> field denotes which type of range + * the <start> and <end> (inclusive) fields are specifying. + * PCI config space ranges are specified by segment/bus/device/function values + * which should be encoded using the HVMOP_PCI_SBDF helper macro below. + */ +#define HVMOP_map_io_range_to_ioreq_server 19 +#define HVMOP_unmap_io_range_from_ioreq_server 20 +struct xen_hvm_io_range { + domid_t domid; /* IN - domain to be serviced */ + ioservid_t id; /* IN - server id */ + uint32_t type; /* IN - type of range */ +# define HVMOP_IO_RANGE_PORT 0 /* I/O port range */ +# define HVMOP_IO_RANGE_MEMORY 1 /* MMIO range */ +# define HVMOP_IO_RANGE_PCI 2 /* PCI segment/bus/dev/func range */ + uint64_aligned_t start, end; /* IN - inclusive start and end of range */ +}; +typedef struct xen_hvm_io_range xen_hvm_io_range_t; +DEFINE_XEN_GUEST_HANDLE(xen_hvm_io_range_t); + +#define HVMOP_PCI_SBDF(s,b,d,f) \ + ((((s) & 0xffff) << 16) | \ + (((b) & 0xff) << 8) | \ + (((d) & 0x1f) << 3) | \ + ((f) & 0x07)) + +/* + * HVMOP_destroy_ioreq_server: Destroy the IOREQ Server <id> servicing domain + * <domid>. + * + * Any registered I/O ranges will be automatically deregistered. + */ +#define HVMOP_destroy_ioreq_server 21 +struct xen_hvm_destroy_ioreq_server { + domid_t domid; /* IN - domain to be serviced */ + ioservid_t id; /* IN - server id */ +}; +typedef struct xen_hvm_destroy_ioreq_server xen_hvm_destroy_ioreq_server_t; +DEFINE_XEN_GUEST_HANDLE(xen_hvm_destroy_ioreq_server_t); + #endif /* defined(__XEN__) || defined(__XEN_TOOLS__) */ #endif /* __XEN_PUBLIC_HVM_HVM_OP_H__ */ + +/* + * Local variables: + * mode: C + * c-file-style: "BSD" + * c-basic-offset: 4 + * tab-width: 4 + * indent-tabs-mode: nil + * End: + */ diff --git a/xen/include/public/hvm/ioreq.h b/xen/include/public/hvm/ioreq.h index f05d130..533089d 100644 --- a/xen/include/public/hvm/ioreq.h +++ b/xen/include/public/hvm/ioreq.h @@ -24,6 +24,8 @@ #ifndef _IOREQ_H_ #define _IOREQ_H_ +#include "hvm_op.h" + #define IOREQ_READ 1 #define IOREQ_WRITE 0 @@ -32,15 +34,22 @@ #define STATE_IOREQ_INPROCESS 2 #define STATE_IORESP_READY 3 -#define IOREQ_TYPE_PIO 0 /* pio */ -#define IOREQ_TYPE_COPY 1 /* mmio ops */ +#define IOREQ_TYPE_PIO (HVMOP_IO_RANGE_PORT) +#define IOREQ_TYPE_COPY (HVMOP_IO_RANGE_MEMORY) +#define IOREQ_TYPE_PCI_CONFIG (HVMOP_IO_RANGE_PCI) #define IOREQ_TYPE_TIMEOFFSET 7 #define IOREQ_TYPE_INVALIDATE 8 /* mapcache */ /* * VMExit dispatcher should cooperate with instruction decoder to * prepare this structure and notify service OS and DM by sending - * virq + * virq. + * + * For I/O type IOREQ_TYPE_PCI_CONFIG, the physical address is formatted + * as follows: + * + * 63....48|47..40|39..35|34..32|31........0 + * SEGMENT |BUS |DEV |FN |OFFSET */ struct ioreq { uint64_t addr; /* physical address */ diff --git a/xen/include/public/hvm/params.h b/xen/include/public/hvm/params.h index 517a184..f830bdd 100644 --- a/xen/include/public/hvm/params.h +++ b/xen/include/public/hvm/params.h @@ -145,6 +145,9 @@ /* SHUTDOWN_* action in case of a triple fault */ #define HVM_PARAM_TRIPLE_FAULT_REASON 31 -#define HVM_NR_PARAMS 32 +#define HVM_PARAM_IOREQ_SERVER_PFN 32 +#define HVM_PARAM_NR_IOREQ_SERVER_PAGES 33 + +#define HVM_NR_PARAMS 34 #endif /* __XEN_PUBLIC_HVM_PARAMS_H__ */ diff --git a/xen/include/xen/lib.h b/xen/include/xen/lib.h index 1369b2b..e81b80e 100644 --- a/xen/include/xen/lib.h +++ b/xen/include/xen/lib.h @@ -104,6 +104,10 @@ extern int scnprintf(char * buf, size_t size, const char * fmt, ...) __attribute__ ((format (printf, 3, 4))); extern int vscnprintf(char *buf, size_t size, const char *fmt, va_list args) __attribute__ ((format (printf, 3, 0))); +extern int asprintf(char ** bufp, const char * fmt, ...) + __attribute__ ((format (printf, 2, 3))); +extern int vasprintf(char ** bufp, const char * fmt, va_list args) + __attribute__ ((format (printf, 2, 0))); long simple_strtol( const char *cp,const char **endp, unsigned int base); diff --git a/xen/include/xen/stdarg.h b/xen/include/xen/stdarg.h index 216fe6d..29249a1 100644 --- a/xen/include/xen/stdarg.h +++ b/xen/include/xen/stdarg.h @@ -2,6 +2,7 @@ #define __XEN_STDARG_H__ typedef __builtin_va_list va_list; +#define va_copy(dest, src) __builtin_va_copy((dest), (src)) #define va_start(ap, last) __builtin_va_start((ap), (last)) #define va_end(ap) __builtin_va_end(ap) #define va_arg __builtin_va_arg -- 1.7.10.4 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |