[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH 1/2] xen: Implement ioctl to restrict privcmd to a specific domain
On Thu, Jul 31, 2014 at 02:16:44PM +0100, Frediano Ziglio wrote: > Add a RESTRICT ioctl to /dev/xen/privcmd, which allows privileged commands > file descriptor to be restricted to only working with a particular domain. I feel I am missing an justification or use case here. > > Signed-off-by: Frediano Ziglio <frediano.ziglio@xxxxxxxxxx> > --- > drivers/xen/privcmd.c | 209 ++++++- > include/uapi/xen/privcmd.h | 6 + > include/xen/interface/domctl.h | 1090 > ++++++++++++++++++++++++++++++++++++ > include/xen/interface/hvm/hvm_op.h | 66 +++ > include/xen/interface/memory.h | 8 + > include/xen/interface/xen.h | 1 + > 6 files changed, 1373 insertions(+), 7 deletions(-) > create mode 100644 include/xen/interface/domctl.h > > diff --git a/drivers/xen/privcmd.c b/drivers/xen/privcmd.c > index 569a13b..c177850 100644 > --- a/drivers/xen/privcmd.c > +++ b/drivers/xen/privcmd.c > @@ -32,6 +32,10 @@ > #include <xen/xen.h> > #include <xen/privcmd.h> > #include <xen/interface/xen.h> > +#include <xen/interface/sched.h> > +#include <xen/interface/memory.h> > +#include <xen/interface/domctl.h> > +#include <xen/interface/hvm/hvm_op.h> > #include <xen/features.h> > #include <xen/page.h> > #include <xen/xen-ops.h> > @@ -43,24 +47,173 @@ MODULE_LICENSE("GPL"); > > #define PRIV_VMA_LOCKED ((void *)1) > > +#define UNRESTRICTED_DOMID ((domid_t)-1) > + > static int privcmd_vma_range_is_mapped( > struct vm_area_struct *vma, > unsigned long addr, > unsigned long nr_pages); > > -static long privcmd_ioctl_hypercall(void __user *udata) > +struct privcmd_check_buf { > + unsigned copy_back; > + void __user *copy_ptr; > + union { > + struct sched_remote_shutdown remote_shutdown; > + struct xen_memory_exchange mem_exchange; > + struct xen_domctl domctl; > + unsigned char buf[1]; > + } u; > +}; > + > +static long privcmd_check_hypercall(struct privcmd_hypercall *hypercall, > + struct privcmd_check *check, > + domid_t restrict_domid) > +{ > +#define DOMID_AT(type, field) do { \ > + BUILD_BUG_ON(sizeof(type) > sizeof(check->u)); \ > + BUILD_BUG_ON(sizeof(((type *) 0)->field) != sizeof(domid_t)); \ > + check->copy_back = sizeof(type); \ > + domid_offset = offsetof(type, field); \ > + } while (0) > + > +/* we copy from userspace and replace arguments to avoid unsafe data */ > +#define FETCH_ARG(dest, arg_num, size) do { \ > + check->copy_ptr = (void *) (long) hypercall->arg[arg_num]; \ > + if (copy_from_user(dest, check->copy_ptr, size)) \ > + return -EFAULT; \ > + hypercall->arg[arg_num] = (long) dest; \ > + } while (0) > + > + unsigned domid_offset; > + > + /* default to invalid so on cases not handled we fail */ > + domid_t domid = UNRESTRICTED_DOMID; > + > + switch (hypercall->op) { > + case __HYPERVISOR_sched_op: > + if (hypercall->arg[0] == SCHEDOP_remote_shutdown) { > + FETCH_ARG(&check->u.remote_shutdown, 1, > + sizeof(check->u.remote_shutdown)); > + domid = check->u.remote_shutdown.domain_id; > + } > + break; > + > + case __HYPERVISOR_domctl: > + FETCH_ARG(&check->u.domctl, 0, sizeof(check->u.domctl)); > + check->copy_back = sizeof(check->u.domctl); > + /* avoid to create a domain */ > + if (check->u.domctl.cmd == XEN_DOMCTL_createdomain) > + return -EACCES; > + /* limit versions to avoid possible future bigger buffer */ > + if (check->u.domctl.interface_version > > XEN_DOMCTL_INTERFACE_VERSION) > + return -EACCES; > + domid = check->u.domctl.domain; > + break; > + > + case __HYPERVISOR_memory_op: > + switch (hypercall->arg[0]) { > + case XENMEM_increase_reservation: > + case XENMEM_decrease_reservation: > + case XENMEM_populate_physmap: > + DOMID_AT(struct xen_memory_reservation, domid); > + break; > + case XENMEM_exchange: > + DOMID_AT(struct xen_memory_exchange, in.domid); > + break; > + case XENMEM_current_reservation: > + case XENMEM_maximum_reservation: > + case XENMEM_maximum_gpfn: > + check->copy_back = sizeof(domid); > + domid_offset = 0; > + break; > + case XENMEM_add_to_physmap: > + DOMID_AT(struct xen_add_to_physmap, domid); > + break; > + case XENMEM_set_memory_map: > + DOMID_AT(struct xen_foreign_memory_map, domid); > + break; > + default: > + return -EACCES; > + } > + FETCH_ARG(&check->u, 1, check->copy_back); > + domid = *((domid_t *) &check->u.buf[domid_offset]); > + > + /* extra check for XENMEM_exchange, exchange in the same > + * domain */ > + if (hypercall->arg[0] == XENMEM_exchange && > + check->u.mem_exchange.in.domid != > check->u.mem_exchange.out.domid) > + return -EACCES; > + break; > + > + case __HYPERVISOR_hvm_op: > + switch (hypercall->arg[0]) { > + case HVMOP_set_param: > + case HVMOP_get_param: > + DOMID_AT(struct xen_hvm_param, domid); > + break; > + case HVMOP_set_pci_intx_level: > + DOMID_AT(struct xen_hvm_set_pci_intx_level, domid); > + break; > + case HVMOP_set_isa_irq_level: > + DOMID_AT(struct xen_hvm_set_isa_irq_level, domid); > + break; > + case HVMOP_set_pci_link_route: > + DOMID_AT(struct xen_hvm_set_pci_link_route, domid); > + break; > + case HVMOP_modified_memory: > + DOMID_AT(struct xen_hvm_modified_memory, domid); > + break; > + case HVMOP_set_mem_type: > + DOMID_AT(struct xen_hvm_set_mem_type, domid); > + break; > + case HVMOP_track_dirty_vram: > + DOMID_AT(struct xen_hvm_track_dirty_vram, domid); > + break; > + default: > + return -EACCES; > + } > + FETCH_ARG(&check->u, 1, check->copy_back); > + domid = *((domid_t *) &check->u.buf[domid_offset]); > + break; > + } > + > + if (domid != restrict_domid) > + return -EACCES; > + > + return 0; > +} > + > +static long privcmd_ioctl_hypercall(void __user *udata, > + domid_t restrict_domid) > { > struct privcmd_hypercall hypercall; > + struct privcmd_check_buf check_buf; > long ret; > > + check_buf.copy_back = 0; > + check_buf.copy_ptr = NULL; > + > if (copy_from_user(&hypercall, udata, sizeof(hypercall))) > return -EFAULT; > > + /* we must check domain we are using */ > + if (restrict_domid != UNRESTRICTED_DOMID) { > + ret = privcmd_check_hypercall(&hypercall, &check_buf, > + restrict_domid); > + if (ret) > + return ret; > + } > + > ret = privcmd_call(hypercall.op, > hypercall.arg[0], hypercall.arg[1], > hypercall.arg[2], hypercall.arg[3], > hypercall.arg[4]); > > + if (check_buf.copy_back && check_buf.copy_ptr && ret >= 0) > + if (copy_to_user(check_buf.copy_ptr, &check_buf.u, > + check_buf.copy_back)) > + ret = -EFAULT; > + > return ret; > } > > @@ -193,7 +346,7 @@ static int mmap_mfn_range(void *data, void *state) > return 0; > } > > -static long privcmd_ioctl_mmap(void __user *udata) > +static long privcmd_ioctl_mmap(void __user *udata, domid_t restrict_domid) > { > struct privcmd_mmap mmapcmd; > struct mm_struct *mm = current->mm; > @@ -209,6 +362,10 @@ static long privcmd_ioctl_mmap(void __user *udata) > if (copy_from_user(&mmapcmd, udata, sizeof(mmapcmd))) > return -EFAULT; > > + if (restrict_domid != UNRESTRICTED_DOMID && > + restrict_domid != mmapcmd.dom) > + return -EACCES; > + > rc = gather_array(&pagelist, > mmapcmd.num, sizeof(struct privcmd_mmap_entry), > mmapcmd.entry); > @@ -367,7 +524,8 @@ static int alloc_empty_pages(struct vm_area_struct *vma, > int numpgs) > > static struct vm_operations_struct privcmd_vm_ops; > > -static long privcmd_ioctl_mmap_batch(void __user *udata, int version) > +static long privcmd_ioctl_mmap_batch(void __user *udata, int version, > + domid_t restrict_domid) > { > int ret; > struct privcmd_mmapbatch_v2 m; > @@ -397,6 +555,10 @@ static long privcmd_ioctl_mmap_batch(void __user *udata, > int version) > return -EINVAL; > } > > + if (restrict_domid != UNRESTRICTED_DOMID && > + restrict_domid != m.dom) > + return -EACCES; > + > nr_pages = m.num; > if ((m.num <= 0) || (nr_pages > (LONG_MAX >> PAGE_SHIFT))) > return -EINVAL; > @@ -498,27 +660,53 @@ out_unlock: > goto out; > } > > +static inline domid_t privcmd_get_restrict_domid(const struct file *file) > +{ > + return (domid_t) (long) file->private_data; > +} > + > +static inline void privcmd_set_restrict_domid(struct file *file, > + domid_t domid) > +{ > + file->private_data = (void *) (long) domid; > +} > + > static long privcmd_ioctl(struct file *file, > unsigned int cmd, unsigned long data) > { > int ret = -ENOSYS; > void __user *udata = (void __user *) data; > + domid_t restrict_domid = privcmd_get_restrict_domid(file); > > switch (cmd) { > case IOCTL_PRIVCMD_HYPERCALL: > - ret = privcmd_ioctl_hypercall(udata); > + ret = privcmd_ioctl_hypercall(udata, restrict_domid); > break; > > case IOCTL_PRIVCMD_MMAP: > - ret = privcmd_ioctl_mmap(udata); > + ret = privcmd_ioctl_mmap(udata, restrict_domid); > break; > > case IOCTL_PRIVCMD_MMAPBATCH: > - ret = privcmd_ioctl_mmap_batch(udata, 1); > + ret = privcmd_ioctl_mmap_batch(udata, 1, restrict_domid); > break; > > case IOCTL_PRIVCMD_MMAPBATCH_V2: > - ret = privcmd_ioctl_mmap_batch(udata, 2); > + ret = privcmd_ioctl_mmap_batch(udata, 2, restrict_domid); > + break; > + > + case IOCTL_PRIVCMD_RESTRICT_DOMID: { > + struct privcmd_restrict_domid prd; > + > + if (restrict_domid != UNRESTRICTED_DOMID) > + return -EACCES; > + if (copy_from_user(&prd, udata, sizeof(prd))) > + return -EFAULT; > + if (prd.domid >= DOMID_FIRST_RESERVED) > + return -EINVAL; > + privcmd_set_restrict_domid(file, prd.domid); > + ret = 0; > + } > break; > > default: > @@ -593,10 +781,17 @@ static int privcmd_vma_range_is_mapped( > is_mapped_fn, NULL) != 0; > } > > +static int privcmd_open(struct inode *ino, struct file *filp) > +{ > + privcmd_set_restrict_domid(filp, UNRESTRICTED_DOMID); > + return 0; > +} > + > const struct file_operations xen_privcmd_fops = { > .owner = THIS_MODULE, > .unlocked_ioctl = privcmd_ioctl, > .mmap = privcmd_mmap, > + .open = privcmd_open, > }; > EXPORT_SYMBOL_GPL(xen_privcmd_fops); > > diff --git a/include/uapi/xen/privcmd.h b/include/uapi/xen/privcmd.h > index a853168..461a999 100644 > --- a/include/uapi/xen/privcmd.h > +++ b/include/uapi/xen/privcmd.h > @@ -73,6 +73,10 @@ struct privcmd_mmapbatch_v2 { > int __user *err; /* array of error codes */ > }; > > +struct privcmd_restrict_domid { > + domid_t domid; > +}; > + > /* > * @cmd: IOCTL_PRIVCMD_HYPERCALL > * @arg: &privcmd_hypercall_t > @@ -94,5 +98,7 @@ struct privcmd_mmapbatch_v2 { > _IOC(_IOC_NONE, 'P', 3, sizeof(struct privcmd_mmapbatch)) > #define IOCTL_PRIVCMD_MMAPBATCH_V2 \ > _IOC(_IOC_NONE, 'P', 4, sizeof(struct privcmd_mmapbatch_v2)) > +#define IOCTL_PRIVCMD_RESTRICT_DOMID \ > + _IOC(_IOC_NONE, 'P', 5, sizeof(struct privcmd_restrict_domid)) > > #endif /* __LINUX_PUBLIC_PRIVCMD_H__ */ > diff --git a/include/xen/interface/domctl.h b/include/xen/interface/domctl.h > new file mode 100644 > index 0000000..0668fed > --- /dev/null > +++ b/include/xen/interface/domctl.h > @@ -0,0 +1,1090 @@ > +/****************************************************************************** > + * domctl.h > + * > + * Domain management operations. For use by node control stack. > + * > + * Permission is hereby granted, free of charge, to any person obtaining a > copy > + * of this software and associated documentation files (the "Software"), to > + * deal in the Software without restriction, including without limitation the > + * rights to use, copy, modify, merge, publish, distribute, sublicense, > and/or > + * sell copies of the Software, and to permit persons to whom the Software is > + * furnished to do so, subject to the following conditions: > + * > + * The above copyright notice and this permission notice shall be included in > + * all copies or substantial portions of the Software. > + * > + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR > + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, > + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL > THE > + * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER > + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING > + * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER > + * DEALINGS IN THE SOFTWARE. > + * > + * Copyright (c) 2002-2003, B Dragovic > + * Copyright (c) 2002-2006, K Fraser > + */ > + > +#ifndef __XEN_PUBLIC_DOMCTL_H__ > +#define __XEN_PUBLIC_DOMCTL_H__ > + > +#include "xen.h" > +#include "grant_table.h" > + > +#define XEN_DOMCTL_INTERFACE_VERSION 0x0000000a > + > +#if 1 > +/* > + * NB. xen_domctl.domain is an IN/OUT parameter for this operation. > + * If it is specified as zero, an id is auto-allocated and returned. > + */ > +/* XEN_DOMCTL_createdomain */ > +struct xen_domctl_createdomain { > + /* IN parameters */ > + uint32_t ssidref; > + xen_domain_handle_t handle; > + /* Is this an HVM guest (as opposed to a PVH or PV guest)? */ > +#define _XEN_DOMCTL_CDF_hvm_guest 0 > +#define XEN_DOMCTL_CDF_hvm_guest (1U<<_XEN_DOMCTL_CDF_hvm_guest) > + /* Use hardware-assisted paging if available? */ > +#define _XEN_DOMCTL_CDF_hap 1 > +#define XEN_DOMCTL_CDF_hap (1U<<_XEN_DOMCTL_CDF_hap) > + /* Should domain memory integrity be verifed by tboot during Sx? */ > +#define _XEN_DOMCTL_CDF_s3_integrity 2 > +#define XEN_DOMCTL_CDF_s3_integrity (1U<<_XEN_DOMCTL_CDF_s3_integrity) > + /* Disable out-of-sync shadow page tables? */ > +#define _XEN_DOMCTL_CDF_oos_off 3 > +#define XEN_DOMCTL_CDF_oos_off (1U<<_XEN_DOMCTL_CDF_oos_off) > + /* Is this a PVH guest (as opposed to an HVM or PV guest)? */ > +#define _XEN_DOMCTL_CDF_pvh_guest 4 > +#define XEN_DOMCTL_CDF_pvh_guest (1U<<_XEN_DOMCTL_CDF_pvh_guest) > + uint32_t flags; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_createdomain); > + > +/* XEN_DOMCTL_getdomaininfo */ > +struct xen_domctl_getdomaininfo { > + /* OUT variables. */ > + domid_t domain; /* Also echoed in domctl.domain */ > + /* Domain is scheduled to die. */ > +#define _XEN_DOMINF_dying 0 > +#define XEN_DOMINF_dying (1U<<_XEN_DOMINF_dying) > + /* Domain is an HVM guest (as opposed to a PV guest). */ > +#define _XEN_DOMINF_hvm_guest 1 > +#define XEN_DOMINF_hvm_guest (1U<<_XEN_DOMINF_hvm_guest) > + /* The guest OS has shut down. */ > +#define _XEN_DOMINF_shutdown 2 > +#define XEN_DOMINF_shutdown (1U<<_XEN_DOMINF_shutdown) > + /* Currently paused by control software. */ > +#define _XEN_DOMINF_paused 3 > +#define XEN_DOMINF_paused (1U<<_XEN_DOMINF_paused) > + /* Currently blocked pending an event. */ > +#define _XEN_DOMINF_blocked 4 > +#define XEN_DOMINF_blocked (1U<<_XEN_DOMINF_blocked) > + /* Domain is currently running. */ > +#define _XEN_DOMINF_running 5 > +#define XEN_DOMINF_running (1U<<_XEN_DOMINF_running) > + /* Being debugged. */ > +#define _XEN_DOMINF_debugged 6 > +#define XEN_DOMINF_debugged (1U<<_XEN_DOMINF_debugged) > +/* domain is PVH */ > +#define _XEN_DOMINF_pvh_guest 7 > +#define XEN_DOMINF_pvh_guest (1U<<_XEN_DOMINF_pvh_guest) > + /* XEN_DOMINF_shutdown guest-supplied code. */ > +#define XEN_DOMINF_shutdownmask 255 > +#define XEN_DOMINF_shutdownshift 16 > + uint32_t flags; /* XEN_DOMINF_* */ > + aligned_u64 tot_pages; > + aligned_u64 max_pages; > + aligned_u64 outstanding_pages; > + aligned_u64 shr_pages; > + aligned_u64 paged_pages; > + aligned_u64 shared_info_frame; /* GMFN of shared_info struct */ > + aligned_u64 cpu_time; > + uint32_t nr_online_vcpus; /* Number of VCPUs currently online. */ > + uint32_t max_vcpu_id; /* Maximum VCPUID in use by this domain. */ > + uint32_t ssidref; > + xen_domain_handle_t handle; > + uint32_t cpupool; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_getdomaininfo); > + > + > +/* XEN_DOMCTL_getmemlist */ > +struct xen_domctl_getmemlist { > + /* IN variables. */ > + /* Max entries to write to output buffer. */ > + aligned_u64 max_pfns; > + /* Start index in guest's page list. */ > + aligned_u64 start_pfn; > + GUEST_HANDLE(uint64_t) buffer; > + /* OUT variables. */ > + aligned_u64 num_pfns; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_getmemlist); > + > + > +/* XEN_DOMCTL_getpageframeinfo */ > + > +#define XEN_DOMCTL_PFINFO_LTAB_SHIFT 28 > +#define XEN_DOMCTL_PFINFO_NOTAB (0x0U<<28) > +#define XEN_DOMCTL_PFINFO_L1TAB (0x1U<<28) > +#define XEN_DOMCTL_PFINFO_L2TAB (0x2U<<28) > +#define XEN_DOMCTL_PFINFO_L3TAB (0x3U<<28) > +#define XEN_DOMCTL_PFINFO_L4TAB (0x4U<<28) > +#define XEN_DOMCTL_PFINFO_LTABTYPE_MASK (0x7U<<28) > +#define XEN_DOMCTL_PFINFO_LPINTAB (0x1U<<31) > +#define XEN_DOMCTL_PFINFO_XTAB (0xfU<<28) /* invalid page */ > +#define XEN_DOMCTL_PFINFO_XALLOC (0xeU<<28) /* allocate-only page */ > +#define XEN_DOMCTL_PFINFO_BROKEN (0xdU<<28) /* broken page */ > +#define XEN_DOMCTL_PFINFO_LTAB_MASK (0xfU<<28) > + > +struct xen_domctl_getpageframeinfo { > + /* IN variables. */ > + aligned_u64 gmfn; /* GMFN to query */ > + /* OUT variables. */ > + /* Is the page PINNED to a type? */ > + uint32_t type; /* see above type defs */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_getpageframeinfo); > + > + > +/* XEN_DOMCTL_getpageframeinfo2 */ > +struct xen_domctl_getpageframeinfo2 { > + /* IN variables. */ > + aligned_u64 num; > + /* IN/OUT variables. */ > + GUEST_HANDLE(uint32_t) array; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_getpageframeinfo2); > + > +/* XEN_DOMCTL_getpageframeinfo3 */ > +struct xen_domctl_getpageframeinfo3 { > + /* IN variables. */ > + aligned_u64 num; > + /* IN/OUT variables. */ > + GUEST_HANDLE(xen_pfn_t) array; > +}; > + > + > +/* > + * Control shadow pagetables operation > + */ > +/* XEN_DOMCTL_shadow_op */ > + > +/* Disable shadow mode. */ > +#define XEN_DOMCTL_SHADOW_OP_OFF 0 > + > +/* Enable shadow mode (mode contains ORed XEN_DOMCTL_SHADOW_ENABLE_* flags). > */ > +#define XEN_DOMCTL_SHADOW_OP_ENABLE 32 > + > +/* Log-dirty bitmap operations. */ > + /* Return the bitmap and clean internal copy for next round. */ > +#define XEN_DOMCTL_SHADOW_OP_CLEAN 11 > + /* Return the bitmap but do not modify internal copy. */ > +#define XEN_DOMCTL_SHADOW_OP_PEEK 12 > + > +/* Memory allocation accessors. */ > +#define XEN_DOMCTL_SHADOW_OP_GET_ALLOCATION 30 > +#define XEN_DOMCTL_SHADOW_OP_SET_ALLOCATION 31 > + > +/* Legacy enable operations. */ > + /* Equiv. to ENABLE with no mode flags. */ > +#define XEN_DOMCTL_SHADOW_OP_ENABLE_TEST 1 > + /* Equiv. to ENABLE with mode flag ENABLE_LOG_DIRTY. */ > +#define XEN_DOMCTL_SHADOW_OP_ENABLE_LOGDIRTY 2 > + /* Equiv. to ENABLE with mode flags ENABLE_REFCOUNT and ENABLE_TRANSLATE. */ > +#define XEN_DOMCTL_SHADOW_OP_ENABLE_TRANSLATE 3 > + > +/* Mode flags for XEN_DOMCTL_SHADOW_OP_ENABLE. */ > + /* > + * Shadow pagetables are refcounted: guest does not use explicit mmu > + * operations nor write-protect its pagetables. > + */ > +#define XEN_DOMCTL_SHADOW_ENABLE_REFCOUNT (1 << 1) > + /* > + * Log pages in a bitmap as they are dirtied. > + * Used for live relocation to determine which pages must be re-sent. > + */ > +#define XEN_DOMCTL_SHADOW_ENABLE_LOG_DIRTY (1 << 2) > + /* > + * Automatically translate GPFNs into MFNs. > + */ > +#define XEN_DOMCTL_SHADOW_ENABLE_TRANSLATE (1 << 3) > + /* > + * Xen does not steal virtual address space from the guest. > + * Requires HVM support. > + */ > +#define XEN_DOMCTL_SHADOW_ENABLE_EXTERNAL (1 << 4) > + > +struct xen_domctl_shadow_op_stats { > + uint32_t fault_count; > + uint32_t dirty_count; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_shadow_op_stats); > + > +struct xen_domctl_shadow_op { > + /* IN variables. */ > + uint32_t op; /* XEN_DOMCTL_SHADOW_OP_* */ > + > + /* OP_ENABLE */ > + uint32_t mode; /* XEN_DOMCTL_SHADOW_ENABLE_* */ > + > + /* OP_GET_ALLOCATION / OP_SET_ALLOCATION */ > + uint32_t mb; /* Shadow memory allocation in MB */ > + > + /* OP_PEEK / OP_CLEAN */ > + GUEST_HANDLE(uchar) dirty_bitmap; > + aligned_u64 pages; /* Size of buffer. Updated with actual size. */ > + struct xen_domctl_shadow_op_stats stats; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_shadow_op); > + > + > +/* XEN_DOMCTL_max_mem */ > +struct xen_domctl_max_mem { > + /* IN variables. */ > + aligned_u64 max_memkb; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_max_mem); > + > + > +/* XEN_DOMCTL_setvcpucontext */ > +/* XEN_DOMCTL_getvcpucontext */ > +struct xen_domctl_vcpucontext { > + uint32_t vcpu; /* IN */ > + GUEST_HANDLE(vcpu_guest_context) ctxt; /* IN/OUT */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_vcpucontext); > + > + > +/* XEN_DOMCTL_getvcpuinfo */ > +struct xen_domctl_getvcpuinfo { > + /* IN variables. */ > + uint32_t vcpu; > + /* OUT variables. */ > + uint8_t online; /* currently online (not hotplugged)? */ > + uint8_t blocked; /* blocked waiting for an event? */ > + uint8_t running; /* currently scheduled on its CPU? */ > + aligned_u64 cpu_time; /* total cpu time consumed (ns) */ > + uint32_t cpu; /* current mapping */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_getvcpuinfo); > + > +#if 0 > +/* Get/set the NUMA node(s) with which the guest has affinity with. */ > +/* XEN_DOMCTL_setnodeaffinity */ > +/* XEN_DOMCTL_getnodeaffinity */ > +struct xen_domctl_nodeaffinity { > + struct xenctl_bitmap nodemap;/* IN */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_nodeaffinity); > + > + > +/* Get/set which physical cpus a vcpu can execute on. */ > +/* XEN_DOMCTL_setvcpuaffinity */ > +/* XEN_DOMCTL_getvcpuaffinity */ > +struct xen_domctl_vcpuaffinity { > + uint32_t vcpu; /* IN */ > + struct xenctl_bitmap cpumap; /* IN/OUT */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_vcpuaffinity); > +#endif > + > + > +/* XEN_DOMCTL_max_vcpus */ > +struct xen_domctl_max_vcpus { > + uint32_t max; /* maximum number of vcpus */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_max_vcpus); > + > + > +/* XEN_DOMCTL_scheduler_op */ > +/* Scheduler types. */ > +#define XEN_SCHEDULER_SEDF 4 > +#define XEN_SCHEDULER_CREDIT 5 > +#define XEN_SCHEDULER_CREDIT2 6 > +#define XEN_SCHEDULER_ARINC653 7 > +/* Set or get info? */ > +#define XEN_DOMCTL_SCHEDOP_putinfo 0 > +#define XEN_DOMCTL_SCHEDOP_getinfo 1 > +struct xen_domctl_scheduler_op { > + uint32_t sched_id; /* XEN_SCHEDULER_* */ > + uint32_t cmd; /* XEN_DOMCTL_SCHEDOP_* */ > + union { > + struct xen_domctl_sched_sedf { > + aligned_u64 period; > + aligned_u64 slice; > + aligned_u64 latency; > + uint32_t extratime; > + uint32_t weight; > + } sedf; > + struct xen_domctl_sched_credit { > + uint16_t weight; > + uint16_t cap; > + } credit; > + struct xen_domctl_sched_credit2 { > + uint16_t weight; > + } credit2; > + } u; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_scheduler_op); > + > + > +/* XEN_DOMCTL_setdomainhandle */ > +struct xen_domctl_setdomainhandle { > + xen_domain_handle_t handle; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_setdomainhandle); > + > + > +/* XEN_DOMCTL_setdebugging */ > +struct xen_domctl_setdebugging { > + uint8_t enable; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_setdebugging); > + > + > +/* XEN_DOMCTL_irq_permission */ > +struct xen_domctl_irq_permission { > + uint8_t pirq; > + uint8_t allow_access; /* flag to specify enable/disable of IRQ access */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_irq_permission); > + > + > +/* XEN_DOMCTL_iomem_permission */ > +struct xen_domctl_iomem_permission { > + aligned_u64 first_mfn;/* first page (physical page number) in range */ > + aligned_u64 nr_mfns; /* number of pages in range (>0) */ > + uint8_t allow_access; /* allow (!0) or deny (0) access to range? */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_iomem_permission); > + > + > +/* XEN_DOMCTL_ioport_permission */ > +struct xen_domctl_ioport_permission { > + uint32_t first_port; /* first port int range */ > + uint32_t nr_ports; /* size of port range */ > + uint8_t allow_access; /* allow or deny access to range? */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_ioport_permission); > + > + > +/* XEN_DOMCTL_hypercall_init */ > +struct xen_domctl_hypercall_init { > + aligned_u64 gmfn; /* GMFN to be initialised */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_hypercall_init); > + > + > +/* XEN_DOMCTL_arch_setup */ > +#define _XEN_DOMAINSETUP_hvm_guest 0 > +#define XEN_DOMAINSETUP_hvm_guest (1UL<<_XEN_DOMAINSETUP_hvm_guest) > +#define _XEN_DOMAINSETUP_query 1 /* Get parameters (for save) */ > +#define XEN_DOMAINSETUP_query (1UL<<_XEN_DOMAINSETUP_query) > +#define _XEN_DOMAINSETUP_sioemu_guest 2 > +#define XEN_DOMAINSETUP_sioemu_guest (1UL<<_XEN_DOMAINSETUP_sioemu_guest) > +struct xen_domctl_arch_setup { > + aligned_u64 flags; /* XEN_DOMAINSETUP_* */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_arch_setup); > + > + > +/* XEN_DOMCTL_settimeoffset */ > +struct xen_domctl_settimeoffset { > + int32_t time_offset_seconds; /* applied to domain wallclock time */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_settimeoffset); > + > +/* XEN_DOMCTL_gethvmcontext */ > +/* XEN_DOMCTL_sethvmcontext */ > +struct xen_domctl_hvmcontext { > + uint32_t size; /* IN/OUT: size of buffer / bytes filled */ > + GUEST_HANDLE(uchar) buffer; /* IN/OUT: data, or call > + * gethvmcontext with NULL > + * buffer to get size req'd */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_hvmcontext); > + > + > +/* XEN_DOMCTL_set_address_size */ > +/* XEN_DOMCTL_get_address_size */ > +struct xen_domctl_address_size { > + uint32_t size; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_address_size); > + > + > +/* XEN_DOMCTL_real_mode_area */ > +struct xen_domctl_real_mode_area { > + uint32_t log; /* log2 of Real Mode Area size */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_real_mode_area); > + > + > +/* XEN_DOMCTL_sendtrigger */ > +#define XEN_DOMCTL_SENDTRIGGER_NMI 0 > +#define XEN_DOMCTL_SENDTRIGGER_RESET 1 > +#define XEN_DOMCTL_SENDTRIGGER_INIT 2 > +#define XEN_DOMCTL_SENDTRIGGER_POWER 3 > +#define XEN_DOMCTL_SENDTRIGGER_SLEEP 4 > +struct xen_domctl_sendtrigger { > + uint32_t trigger; /* IN */ > + uint32_t vcpu; /* IN */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_sendtrigger); > + > + > +/* Assign PCI device to HVM guest. Sets up IOMMU structures. */ > +/* XEN_DOMCTL_assign_device */ > +/* XEN_DOMCTL_test_assign_device */ > +/* XEN_DOMCTL_deassign_device */ > +struct xen_domctl_assign_device { > + uint32_t machine_sbdf; /* machine PCI ID of assigned device */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_assign_device); > + > +/* Retrieve sibling devices infomation of machine_sbdf */ > +/* XEN_DOMCTL_get_device_group */ > +struct xen_domctl_get_device_group { > + uint32_t machine_sbdf; /* IN */ > + uint32_t max_sdevs; /* IN */ > + uint32_t num_sdevs; /* OUT */ > + GUEST_HANDLE(uint32_t) sdev_array; /* OUT */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_get_device_group); > + > +/* Pass-through interrupts: bind real irq -> hvm devfn. */ > +/* XEN_DOMCTL_bind_pt_irq */ > +/* XEN_DOMCTL_unbind_pt_irq */ > +enum pt_irq_type_e { > + PT_IRQ_TYPE_PCI, > + PT_IRQ_TYPE_ISA, > + PT_IRQ_TYPE_MSI, > + PT_IRQ_TYPE_MSI_TRANSLATE, > +}; > +struct xen_domctl_bind_pt_irq { > + uint32_t machine_irq; > + enum pt_irq_type_e irq_type; > + uint32_t hvm_domid; > + > + union { > + struct { > + uint8_t isa_irq; > + } isa; > + struct { > + uint8_t bus; > + uint8_t device; > + uint8_t intx; > + } pci; > + struct { > + uint8_t gvec; > + uint32_t gflags; > + aligned_u64 gtable; > + } msi; > + } u; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_bind_pt_irq); > + > + > +/* Bind machine I/O address range -> HVM address range. */ > +/* XEN_DOMCTL_memory_mapping */ > +#define DPCI_ADD_MAPPING 1 > +#define DPCI_REMOVE_MAPPING 0 > +struct xen_domctl_memory_mapping { > + aligned_u64 first_gfn; /* first page (hvm guest phys page) in range */ > + aligned_u64 first_mfn; /* first page (machine page) in range */ > + aligned_u64 nr_mfns; /* number of pages in range (>0) */ > + uint32_t add_mapping; /* add or remove mapping */ > + uint32_t padding; /* padding for 64-bit aligned structure */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_memory_mapping); > + > + > +/* Bind machine I/O port range -> HVM I/O port range. */ > +/* XEN_DOMCTL_ioport_mapping */ > +struct xen_domctl_ioport_mapping { > + uint32_t first_gport; /* first guest IO port*/ > + uint32_t first_mport; /* first machine IO port */ > + uint32_t nr_ports; /* size of port range */ > + uint32_t add_mapping; /* add or remove mapping */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_ioport_mapping); > + > + > +/* > + * Pin caching type of RAM space for x86 HVM domU. > + */ > +/* XEN_DOMCTL_pin_mem_cacheattr */ > +/* Caching types: these happen to be the same as x86 MTRR/PAT type codes. */ > +#define XEN_DOMCTL_MEM_CACHEATTR_UC 0 > +#define XEN_DOMCTL_MEM_CACHEATTR_WC 1 > +#define XEN_DOMCTL_MEM_CACHEATTR_WT 4 > +#define XEN_DOMCTL_MEM_CACHEATTR_WP 5 > +#define XEN_DOMCTL_MEM_CACHEATTR_WB 6 > +#define XEN_DOMCTL_MEM_CACHEATTR_UCM 7 > +struct xen_domctl_pin_mem_cacheattr { > + aligned_u64 start, end; > + uint32_t type; /* XEN_DOMCTL_MEM_CACHEATTR_* */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_pin_mem_cacheattr); > + > + > +#if 0 > +/* XEN_DOMCTL_set_ext_vcpucontext */ > +/* XEN_DOMCTL_get_ext_vcpucontext */ > +struct xen_domctl_ext_vcpucontext { > + /* IN: VCPU that this call applies to. */ > + uint32_t vcpu; > + /* > + * SET: Size of struct (IN) > + * GET: Size of struct (OUT, up to 128 bytes) > + */ > + uint32_t size; > +#if defined(__i386__) || defined(__x86_64__) > + /* SYSCALL from 32-bit mode and SYSENTER callback information. */ > + /* NB. SYSCALL from 64-bit mode is contained in vcpu_guest_context_t */ > + aligned_u64 syscall32_callback_eip; > + aligned_u64 sysenter_callback_eip; > + uint16_t syscall32_callback_cs; > + uint16_t sysenter_callback_cs; > + uint8_t syscall32_disables_events; > + uint8_t sysenter_disables_events; > +#if defined(__GNUC__) > + union { > + aligned_u64 mcg_cap; > + struct hvm_vmce_vcpu vmce; > + }; > +#else > + struct hvm_vmce_vcpu vmce; > +#endif > +#endif > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_ext_vcpucontext); > +#endif > + > +/* > + * Set the target domain for a domain > + */ > +/* XEN_DOMCTL_set_target */ > +struct xen_domctl_set_target { > + domid_t target; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_set_target); > + > +#if defined(__i386__) || defined(__x86_64__) > +# define XEN_CPUID_INPUT_UNUSED 0xFFFFFFFF > +/* XEN_DOMCTL_set_cpuid */ > +struct xen_domctl_cpuid { > + uint32_t input[2]; > + uint32_t eax; > + uint32_t ebx; > + uint32_t ecx; > + uint32_t edx; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_cpuid); > +#endif > + > +/* > + * Arranges that if the domain suspends (specifically, if it shuts > + * down with code SHUTDOWN_suspend), this event channel will be > + * notified. > + * > + * This is _instead of_ the usual notification to the global > + * VIRQ_DOM_EXC. (In most systems that pirq is owned by xenstored.) > + * > + * Only one subscription per domain is possible. Last subscriber > + * wins; others are silently displaced. > + * > + * NB that contrary to the rather general name, it only applies to > + * domain shutdown with code suspend. Shutdown for other reasons > + * (including crash), and domain death, are notified to VIRQ_DOM_EXC > + * regardless. > + */ > +/* XEN_DOMCTL_subscribe */ > +struct xen_domctl_subscribe { > + uint32_t port; /* IN */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_subscribe); > + > +/* > + * Define the maximum machine address size which should be allocated > + * to a guest. > + */ > +/* XEN_DOMCTL_set_machine_address_size */ > +/* XEN_DOMCTL_get_machine_address_size */ > + > +/* > + * Do not inject spurious page faults into this domain. > + */ > +/* XEN_DOMCTL_suppress_spurious_page_faults */ > + > +/* XEN_DOMCTL_debug_op */ > +#define XEN_DOMCTL_DEBUG_OP_SINGLE_STEP_OFF 0 > +#define XEN_DOMCTL_DEBUG_OP_SINGLE_STEP_ON 1 > +struct xen_domctl_debug_op { > + uint32_t op; /* IN */ > + uint32_t vcpu; /* IN */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_debug_op); > + > +/* > + * Request a particular record from the HVM context > + */ > +/* XEN_DOMCTL_gethvmcontext_partial */ > +struct xen_domctl_hvmcontext_partial { > + uint32_t type; /* IN: Type of record required */ > + uint32_t instance; /* IN: Instance of that type */ > + GUEST_HANDLE(uchar) buffer; /* OUT: buffer to write record into */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_hvmcontext_partial); > + > +/* XEN_DOMCTL_disable_migrate */ > +struct xen_domctl_disable_migrate { > + uint32_t disable; /* IN: 1: disable migration and restore */ > +}; > + > + > +/* XEN_DOMCTL_gettscinfo */ > +/* XEN_DOMCTL_settscinfo */ > +struct xen_guest_tsc_info { > + uint32_t tsc_mode; > + uint32_t gtsc_khz; > + uint32_t incarnation; > + uint32_t pad; > + aligned_u64 elapsed_nsec; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_guest_tsc_info); > + > +struct xen_domctl_tsc_info { > + GUEST_HANDLE(xen_guest_tsc_info) out_info; /* OUT */ > + struct xen_guest_tsc_info info; /* IN */ > +}; > + > +/* XEN_DOMCTL_gdbsx_guestmemio guest mem io */ > +struct xen_domctl_gdbsx_memio { > + /* IN */ > + aligned_u64 pgd3val;/* optional: init_mm.pgd[3] value */ > + aligned_u64 gva; /* guest virtual address */ > + aligned_u64 uva; /* user buffer virtual address */ > + uint32_t len; /* number of bytes to read/write */ > + uint8_t gwr; /* 0 = read from guest. 1 = write to guest */ > + /* OUT */ > + uint32_t remain; /* bytes remaining to be copied */ > +}; > + > +/* XEN_DOMCTL_gdbsx_pausevcpu */ > +/* XEN_DOMCTL_gdbsx_unpausevcpu */ > +struct xen_domctl_gdbsx_pauseunp_vcpu { /* pause/unpause a vcpu */ > + uint32_t vcpu; /* which vcpu */ > +}; > + > +/* XEN_DOMCTL_gdbsx_domstatus */ > +struct xen_domctl_gdbsx_domstatus { > + /* OUT */ > + uint8_t paused; /* is the domain paused */ > + uint32_t vcpu_id; /* any vcpu in an event? */ > + uint32_t vcpu_ev; /* if yes, what event? */ > +}; > + > +/* > + * Memory event operations > + */ > + > +/* XEN_DOMCTL_mem_event_op */ > + > +/* > + * Domain memory paging > + * Page memory in and out. > + * Domctl interface to set up and tear down the > + * pager<->hypervisor interface. Use XENMEM_paging_op* > + * to perform per-page operations. > + * > + * The XEN_DOMCTL_MEM_EVENT_OP_PAGING_ENABLE domctl returns several > + * non-standard error codes to indicate why paging could not be enabled: > + * ENODEV - host lacks HAP support (EPT/NPT) or HAP is disabled in guest > + * EMLINK - guest has iommu passthrough enabled > + * EXDEV - guest has PoD enabled > + * EBUSY - guest has or had paging enabled, ring buffer still active > + */ > +#define XEN_DOMCTL_MEM_EVENT_OP_PAGING 1 > + > +#define XEN_DOMCTL_MEM_EVENT_OP_PAGING_ENABLE 0 > +#define XEN_DOMCTL_MEM_EVENT_OP_PAGING_DISABLE 1 > + > +/* > + * Access permissions. > + * > + * As with paging, use the domctl for teardown/setup of the > + * helper<->hypervisor interface. > + * > + * There are HVM hypercalls to set the per-page access permissions of every > + * page in a domain. When one of these permissions--independent, read, > + * write, and execute--is violated, the VCPU is paused and a memory event > + * is sent with what happened. (See public/mem_event.h) . > + * > + * The memory event handler can then resume the VCPU and redo the access > + * with a XENMEM_access_op_resume hypercall. > + * > + * The XEN_DOMCTL_MEM_EVENT_OP_ACCESS_ENABLE domctl returns several > + * non-standard error codes to indicate why access could not be enabled: > + * ENODEV - host lacks HAP support (EPT/NPT) or HAP is disabled in guest > + * EBUSY - guest has or had access enabled, ring buffer still active > + */ > +#define XEN_DOMCTL_MEM_EVENT_OP_ACCESS 2 > + > +#define XEN_DOMCTL_MEM_EVENT_OP_ACCESS_ENABLE 0 > +#define XEN_DOMCTL_MEM_EVENT_OP_ACCESS_DISABLE 1 > + > +/* > + * Sharing ENOMEM helper. > + * > + * As with paging, use the domctl for teardown/setup of the > + * helper<->hypervisor interface. > + * > + * If setup, this ring is used to communicate failed allocations > + * in the unshare path. XENMEM_sharing_op_resume is used to wake up > + * vcpus that could not unshare. > + * > + * Note that shring can be turned on (as per the domctl below) > + * *without* this ring being setup. > + */ > +#define XEN_DOMCTL_MEM_EVENT_OP_SHARING 3 > + > +#define XEN_DOMCTL_MEM_EVENT_OP_SHARING_ENABLE 0 > +#define XEN_DOMCTL_MEM_EVENT_OP_SHARING_DISABLE 1 > + > +/* Use for teardown/setup of helper<->hypervisor interface for paging, > + * access and sharing.*/ > +struct xen_domctl_mem_event_op { > + uint32_t op; /* XEN_DOMCTL_MEM_EVENT_OP_*_* */ > + uint32_t mode; /* XEN_DOMCTL_MEM_EVENT_OP_* */ > + > + uint32_t port; /* OUT: event channel for ring */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_mem_event_op); > + > +/* > + * Memory sharing operations > + */ > +/* XEN_DOMCTL_mem_sharing_op. > + * The CONTROL sub-domctl is used for bringup/teardown. */ > +#define XEN_DOMCTL_MEM_SHARING_CONTROL 0 > + > +struct xen_domctl_mem_sharing_op { > + uint8_t op; /* XEN_DOMCTL_MEM_SHARING_* */ > + > + union { > + uint8_t enable; /* CONTROL */ > + } u; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_mem_sharing_op); > + > +struct xen_domctl_audit_p2m { > + /* OUT error counts */ > + uint64_t orphans; > + uint64_t m2p_bad; > + uint64_t p2m_bad; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_audit_p2m); > + > +struct xen_domctl_set_virq_handler { > + uint32_t virq; /* IN */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_set_virq_handler); > + > +#if defined(__i386__) || defined(__x86_64__) > +/* XEN_DOMCTL_setvcpuextstate */ > +/* XEN_DOMCTL_getvcpuextstate */ > +struct xen_domctl_vcpuextstate { > + /* IN: VCPU that this call applies to. */ > + uint32_t vcpu; > + /* > + * SET: Ignored. > + * GET: xfeature support mask of struct (IN/OUT) > + * xfeature mask is served as identifications of the saving format > + * so that compatible CPUs can have a check on format to decide > + * whether it can restore. > + */ > + aligned_u64 xfeature_mask; > + /* > + * SET: Size of struct (IN) > + * GET: Size of struct (IN/OUT) > + */ > + aligned_u64 size; > + GUEST_HANDLE(uint64_t) buffer; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_vcpuextstate); > +#endif > + > +/* XEN_DOMCTL_set_access_required: sets whether a memory event listener > + * must be present to handle page access events: if false, the page > + * access will revert to full permissions if no one is listening; > + * */ > +struct xen_domctl_set_access_required { > + uint8_t access_required; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_set_access_required); > + > +struct xen_domctl_set_broken_page_p2m { > + aligned_u64 pfn; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_set_broken_page_p2m); > + > +/* > + * XEN_DOMCTL_set_max_evtchn: sets the maximum event channel port > + * number the guest may use. Use this limit the amount of resources > + * (global mapping space, xenheap) a guest may use for event channels. > + */ > +struct xen_domctl_set_max_evtchn { > + uint32_t max_port; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_set_max_evtchn); > + > +/* > + * ARM: Clean and invalidate caches associated with given region of > + * guest memory. > + */ > +struct xen_domctl_cacheflush { > + /* IN: page range to flush. */ > + xen_pfn_t start_pfn, nr_pfns; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_cacheflush); > + > +#if defined(__i386__) || defined(__x86_64__) > +struct xen_domctl_vcpu_msr { > + uint32_t index; > + uint32_t reserved; > + aligned_u64 value; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_vcpu_msr); > + > +/* > + * XEN_DOMCTL_set_vcpu_msrs / XEN_DOMCTL_get_vcpu_msrs. > + * > + * Input: > + * - A NULL 'msrs' guest handle is a request for the maximum 'msr_count'. > + * - Otherwise, 'msr_count' is the number of entries in 'msrs'. > + * > + * Output for get: > + * - If 'msr_count' is less than the number Xen needs to write, -ENOBUFS > shall > + * be returned and 'msr_count' updated to reflect the intended number. > + * - On success, 'msr_count' shall indicate the number of MSRs written, which > + * may be less than the maximum if some are not currently used by the vcpu. > + * > + * Output for set: > + * - If Xen encounters an error with a specific MSR, -EINVAL shall be > returned > + * and 'msr_count' shall be set to the offending index, to aid debugging. > + */ > +struct xen_domctl_vcpu_msrs { > + uint32_t vcpu; /* IN */ > + uint32_t msr_count; /* IN/OUT */ > + GUEST_HANDLE(xen_domctl_vcpu_msr) msrs; /* IN/OUT */ > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_vcpu_msrs); > +#endif > + > +/* > + * Return information about the state and running time of a domain. > + * The "domain runstate" is based on the runstates of all the vcpus of the > + * domain (see below). > + * @extra_arg == pointer to domain_runstate_info structure. > + */ > +struct xen_domctl_runstate_info { > + /* VCPU's current state (RUNSTATE_*). */ > + uint32_t state; > + uint32_t missed_changes; > + /* Number of times we missed an update due to contention */ > + /* When was current state entered (system time, ns)? */ > + uint64_t state_entry_time; > + /* > + * Time spent in each RUNSTATE_* (ns). The sum of these times is > + * NOT guaranteed not to drift from system time. > + */ > + uint64_t time[6]; > +}; > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_runstate_info); > + > +/* All vcpus are running */ > +#define DOMAIN_RUNSTATE_full_run 0 > + > +/* All vcpus are runnable (i.e., waiting for cpu) */ > +#define DOMAIN_RUNSTATE_full_contention 1 > + > +/* Some vcpus are running, some are runnable */ > +#define DOMAIN_RUNSTATE_concurrency_hazard 2 > + > +/* All vcpus are blocked / offline */ > +#define DOMAIN_RUNSTATE_blocked 3 > + > +/* Some vpcus are running, some are blocked */ > +#define DOMAIN_RUNSTATE_partial_run 4 > + > +/* Some vcpus are runnable, some are blocked */ > +#define DOMAIN_RUNSTATE_partial_contention 5 > + > +struct xen_domctl_corespersocket { > + uint32_t cores_per_socket; > +}; > + > +DEFINE_GUEST_HANDLE_STRUCT(xen_domctl_corespersocket); > +#endif > + > +struct xen_domctl { > + uint32_t cmd; > +#define XEN_DOMCTL_createdomain 1 > +#define XEN_DOMCTL_destroydomain 2 > +#define XEN_DOMCTL_pausedomain 3 > +#define XEN_DOMCTL_unpausedomain 4 > +#define XEN_DOMCTL_getdomaininfo 5 > +#define XEN_DOMCTL_getmemlist 6 > +#define XEN_DOMCTL_getpageframeinfo 7 > +#define XEN_DOMCTL_getpageframeinfo2 8 > +#define XEN_DOMCTL_setvcpuaffinity 9 > +#define XEN_DOMCTL_shadow_op 10 > +#define XEN_DOMCTL_max_mem 11 > +#define XEN_DOMCTL_setvcpucontext 12 > +#define XEN_DOMCTL_getvcpucontext 13 > +#define XEN_DOMCTL_getvcpuinfo 14 > +#define XEN_DOMCTL_max_vcpus 15 > +#define XEN_DOMCTL_scheduler_op 16 > +#define XEN_DOMCTL_setdomainhandle 17 > +#define XEN_DOMCTL_setdebugging 18 > +#define XEN_DOMCTL_irq_permission 19 > +#define XEN_DOMCTL_iomem_permission 20 > +#define XEN_DOMCTL_ioport_permission 21 > +#define XEN_DOMCTL_hypercall_init 22 > +#define XEN_DOMCTL_arch_setup 23 > +#define XEN_DOMCTL_settimeoffset 24 > +#define XEN_DOMCTL_getvcpuaffinity 25 > +#define XEN_DOMCTL_real_mode_area 26 > +#define XEN_DOMCTL_resumedomain 27 > +#define XEN_DOMCTL_sendtrigger 28 > +#define XEN_DOMCTL_subscribe 29 > +#define XEN_DOMCTL_gethvmcontext 33 > +#define XEN_DOMCTL_sethvmcontext 34 > +#define XEN_DOMCTL_set_address_size 35 > +#define XEN_DOMCTL_get_address_size 36 > +#define XEN_DOMCTL_assign_device 37 > +#define XEN_DOMCTL_bind_pt_irq 38 > +#define XEN_DOMCTL_memory_mapping 39 > +#define XEN_DOMCTL_ioport_mapping 40 > +#define XEN_DOMCTL_pin_mem_cacheattr 41 > +#define XEN_DOMCTL_set_ext_vcpucontext 42 > +#define XEN_DOMCTL_get_ext_vcpucontext 43 > +#define XEN_DOMCTL_set_opt_feature 44 /* Obsolete IA64 only */ > +#define XEN_DOMCTL_test_assign_device 45 > +#define XEN_DOMCTL_set_target 46 > +#define XEN_DOMCTL_deassign_device 47 > +#define XEN_DOMCTL_unbind_pt_irq 48 > +#define XEN_DOMCTL_set_cpuid 49 > +#define XEN_DOMCTL_get_device_group 50 > +#define XEN_DOMCTL_set_machine_address_size 51 > +#define XEN_DOMCTL_get_machine_address_size 52 > +#define XEN_DOMCTL_suppress_spurious_page_faults 53 > +#define XEN_DOMCTL_debug_op 54 > +#define XEN_DOMCTL_gethvmcontext_partial 55 > +#define XEN_DOMCTL_mem_event_op 56 > +#define XEN_DOMCTL_mem_sharing_op 57 > +#define XEN_DOMCTL_disable_migrate 58 > +#define XEN_DOMCTL_gettscinfo 59 > +#define XEN_DOMCTL_settscinfo 60 > +#define XEN_DOMCTL_getpageframeinfo3 61 > +#define XEN_DOMCTL_setvcpuextstate 62 > +#define XEN_DOMCTL_getvcpuextstate 63 > +#define XEN_DOMCTL_set_access_required 64 > +#define XEN_DOMCTL_audit_p2m 65 > +#define XEN_DOMCTL_set_virq_handler 66 > +#define XEN_DOMCTL_set_broken_page_p2m 67 > +#define XEN_DOMCTL_setnodeaffinity 68 > +#define XEN_DOMCTL_getnodeaffinity 69 > +#define XEN_DOMCTL_set_max_evtchn 70 > +#define XEN_DOMCTL_cacheflush 71 > +#define XEN_DOMCTL_get_vcpu_msrs 72 > +#define XEN_DOMCTL_set_vcpu_msrs 73 > +#define XEN_DOMCTL_get_runstate_info 98 > +#define XEN_DOMCTL_gdbsx_guestmemio 1000 > +#define XEN_DOMCTL_gdbsx_pausevcpu 1001 > +#define XEN_DOMCTL_gdbsx_unpausevcpu 1002 > +#define XEN_DOMCTL_gdbsx_domstatus 1003 > +#define XEN_DOMCTL_setcorespersocket 4001 > + uint32_t interface_version; /* XEN_DOMCTL_INTERFACE_VERSION */ > + domid_t domain; > + union { > + struct xen_domctl_createdomain createdomain; > + struct xen_domctl_getdomaininfo getdomaininfo; > + struct xen_domctl_getmemlist getmemlist; > + struct xen_domctl_getpageframeinfo getpageframeinfo; > + struct xen_domctl_getpageframeinfo2 getpageframeinfo2; > + struct xen_domctl_getpageframeinfo3 getpageframeinfo3; > +#if 0 > + struct xen_domctl_nodeaffinity nodeaffinity; > + struct xen_domctl_vcpuaffinity vcpuaffinity; > +#endif > + struct xen_domctl_shadow_op shadow_op; > + struct xen_domctl_max_mem max_mem; > + struct xen_domctl_vcpucontext vcpucontext; > + struct xen_domctl_getvcpuinfo getvcpuinfo; > + struct xen_domctl_max_vcpus max_vcpus; > + struct xen_domctl_scheduler_op scheduler_op; > + struct xen_domctl_setdomainhandle setdomainhandle; > + struct xen_domctl_setdebugging setdebugging; > + struct xen_domctl_irq_permission irq_permission; > + struct xen_domctl_iomem_permission iomem_permission; > + struct xen_domctl_ioport_permission ioport_permission; > + struct xen_domctl_hypercall_init hypercall_init; > + struct xen_domctl_arch_setup arch_setup; > + struct xen_domctl_settimeoffset settimeoffset; > + struct xen_domctl_disable_migrate disable_migrate; > + struct xen_domctl_tsc_info tsc_info; > + struct xen_domctl_real_mode_area real_mode_area; > + struct xen_domctl_hvmcontext hvmcontext; > + struct xen_domctl_hvmcontext_partial hvmcontext_partial; > + struct xen_domctl_address_size address_size; > + struct xen_domctl_sendtrigger sendtrigger; > + struct xen_domctl_get_device_group get_device_group; > + struct xen_domctl_assign_device assign_device; > + struct xen_domctl_bind_pt_irq bind_pt_irq; > + struct xen_domctl_memory_mapping memory_mapping; > + struct xen_domctl_ioport_mapping ioport_mapping; > + struct xen_domctl_pin_mem_cacheattr pin_mem_cacheattr; > +#if 0 > + struct xen_domctl_ext_vcpucontext ext_vcpucontext; > +#endif > + struct xen_domctl_set_target set_target; > + struct xen_domctl_subscribe subscribe; > + struct xen_domctl_debug_op debug_op; > + struct xen_domctl_mem_event_op mem_event_op; > + struct xen_domctl_mem_sharing_op mem_sharing_op; > +#if defined(__i386__) || defined(__x86_64__) > + struct xen_domctl_cpuid cpuid; > + struct xen_domctl_vcpuextstate vcpuextstate; > + struct xen_domctl_vcpu_msrs vcpu_msrs; > +#endif > + struct xen_domctl_set_access_required access_required; > + struct xen_domctl_audit_p2m audit_p2m; > + struct xen_domctl_set_virq_handler set_virq_handler; > + struct xen_domctl_set_max_evtchn set_max_evtchn; > + struct xen_domctl_runstate_info domain_runstate; > + struct xen_domctl_corespersocket corespersocket; > + struct xen_domctl_gdbsx_memio gdbsx_guest_memio; > + struct xen_domctl_set_broken_page_p2m set_broken_page_p2m; > + struct xen_domctl_cacheflush cacheflush; > + struct xen_domctl_gdbsx_pauseunp_vcpu gdbsx_pauseunp_vcpu; > + struct xen_domctl_gdbsx_domstatus gdbsx_domstatus; > + uint8_t pad[128]; > + } u __aligned(8); > +}; > + > +#endif /* __XEN_PUBLIC_DOMCTL_H__ */ > + > +/* > + * Local variables: > + * mode: C > + * c-file-style: "BSD" > + * c-basic-offset: 4 > + * tab-width: 4 > + * indent-tabs-mode: nil > + * End: > + */ > diff --git a/include/xen/interface/hvm/hvm_op.h > b/include/xen/interface/hvm/hvm_op.h > index 956a046..5fb5260 100644 > --- a/include/xen/interface/hvm/hvm_op.h > +++ b/include/xen/interface/hvm/hvm_op.h > @@ -32,6 +32,72 @@ struct xen_hvm_param { > }; > DEFINE_GUEST_HANDLE_STRUCT(xen_hvm_param); > > +#define HVMOP_set_pci_intx_level 2 > +struct xen_hvm_set_pci_intx_level { > + /* Domain to be updated. */ > + domid_t domid; > + /* PCI INTx identification in PCI topology (domain:bus:device:intx). */ > + uint8_t domain, bus, device, intx; > + /* Assertion level (0 = unasserted, 1 = asserted). */ > + uint8_t level; > +}; > + > +#define HVMOP_set_isa_irq_level 3 > +struct xen_hvm_set_isa_irq_level { > + /* Domain to be updated. */ > + domid_t domid; > + /* ISA device identification, by ISA IRQ (0-15). */ > + uint8_t isa_irq; > + /* Assertion level (0 = unasserted, 1 = asserted). */ > + uint8_t level; > +}; > + > +#define HVMOP_set_pci_link_route 4 > +struct xen_hvm_set_pci_link_route { > + /* Domain to be updated. */ > + domid_t domid; > + /* PCI link identifier (0-3). */ > + uint8_t link; > + /* ISA IRQ (1-15), or 0 (disable link). */ > + uint8_t isa_irq; > +}; > + > +#define HVMOP_track_dirty_vram 6 > +struct xen_hvm_track_dirty_vram { > + /* Domain to be tracked. */ > + domid_t domid; > + /* First pfn to track. */ > + aligned_u64 first_pfn; > + /* Number of pages to track. */ > + aligned_u64 nr; > + /* OUT variable. */ > + /* Dirty bitmap buffer. */ > + aligned_u64 dirty_bitmap; > +}; > + > +#define HVMOP_modified_memory 7 > +struct xen_hvm_modified_memory { > + /* Domain to be updated. */ > + domid_t domid; > + /* First pfn. */ > + aligned_u64 first_pfn; > + /* Number of pages. */ > + aligned_u64 nr; > +}; > + > +#define HVMOP_set_mem_type 8 > +/* Notify that a region of memory is to be treated in a specific way. */ > +struct xen_hvm_set_mem_type { > + /* Domain to be updated. */ > + domid_t domid; > + /* Memory type */ > + uint16_t hvmmem_type; > + /* Number of pages. */ > + uint32_t nr; > + /* First pfn. */ > + aligned_u64 first_pfn; > +}; > + > /* Hint from PV drivers for pagetable destruction. */ > #define HVMOP_pagetable_dying 9 > struct xen_hvm_pagetable_dying { > diff --git a/include/xen/interface/memory.h b/include/xen/interface/memory.h > index 2ecfe4f..a5fd2e6 100644 > --- a/include/xen/interface/memory.h > +++ b/include/xen/interface/memory.h > @@ -248,6 +248,14 @@ DEFINE_GUEST_HANDLE_STRUCT(xen_memory_map); > */ > extern spinlock_t xen_reservation_lock; > > +#define XENMEM_set_memory_map 13 > +struct xen_foreign_memory_map { > + domid_t domid; > + struct xen_memory_map map; > +}; > + > +#define XENMEM_maximum_gpfn 14 > + > /* > * Unmaps the page appearing at a particular GPFN from the specified guest's > * pseudophysical address space. > diff --git a/include/xen/interface/xen.h b/include/xen/interface/xen.h > index de08213..075cb6f 100644 > --- a/include/xen/interface/xen.h > +++ b/include/xen/interface/xen.h > @@ -57,6 +57,7 @@ > #define __HYPERVISOR_event_channel_op 32 > #define __HYPERVISOR_physdev_op 33 > #define __HYPERVISOR_hvm_op 34 > +#define __HYPERVISOR_domctl 36 > #define __HYPERVISOR_tmem_op 38 > > /* Architecture-specific hypercall definitions. */ > -- > 1.9.1 > > _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |