[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] [PATCH 6/7] xen: Allow hardare domain != dom0



>>> On 18.03.14 at 22:34, Daniel De Graaf <dgdegra@xxxxxxxxxxxxx> wrote:
> This adds a hypervisor command line option "hardware_dom=" which takes a
> domain ID.  When the domain with this ID is created, it will be used
> as the hardware domain.
> 
> This is intended to be used when domain 0 is a dedicated stub domain for
> domain building, allowing the hardware domain to be de-privileged and
> act only as a driver domain.
> 
> Signed-off-by: Daniel De Graaf <dgdegra@xxxxxxxxxxxxx>
> Cc: Keir Fraser <keir@xxxxxxx>

Reviewed-by: Jan Beulich <jbeulich@xxxxxxxx>

(assuming the LATE_HWDOM_ENABLE -> CONFIG_LATE_HWDOM
will get done)

> ---
>  xen/arch/x86/domain_build.c         |  4 +++-
>  xen/arch/x86/setup.c                |  2 ++
>  xen/common/domain.c                 |  3 ++-
>  xen/common/domctl.c                 | 40 
> +++++++++++++++++++++++++++++++++++++
>  xen/common/rangeset.c               | 40 
> +++++++++++++++++++++++++++++++++++++
>  xen/include/xen/rangeset.h          |  3 +++
>  xen/include/xen/sched.h             |  3 ++-
>  xen/include/xsm/dummy.h             |  6 ++++++
>  xen/include/xsm/xsm.h               |  6 ++++++
>  xen/xsm/dummy.c                     |  2 ++
>  xen/xsm/flask/hooks.c               |  6 ++++++
>  xen/xsm/flask/policy/access_vectors |  2 ++
>  12 files changed, 114 insertions(+), 3 deletions(-)
> 
> diff --git a/xen/arch/x86/domain_build.c b/xen/arch/x86/domain_build.c
> index f75f6e7..a554d3b 100644
> --- a/xen/arch/x86/domain_build.c
> +++ b/xen/arch/x86/domain_build.c
> @@ -1150,7 +1150,9 @@ int __init construct_dom0(
>          printk(" Xen warning: dom0 kernel broken ELF: %s\n",
>                 elf_check_broken(&elf));
>  
> -    iommu_hwdom_init(hardware_domain);
> +    if ( is_hardware_domain(d) )
> +        iommu_hwdom_init(d);
> +
>      return 0;
>  
>  out:
> diff --git a/xen/arch/x86/setup.c b/xen/arch/x86/setup.c
> index 75cf212..f246ac3 100644
> --- a/xen/arch/x86/setup.c
> +++ b/xen/arch/x86/setup.c
> @@ -84,6 +84,8 @@ unsigned long __initdata highmem_start;
>  size_param("highmem-start", highmem_start);
>  #endif
>  
> +integer_param("hardware_dom", hardware_domid);
> +
>  cpumask_t __read_mostly cpu_present_map;
>  
>  unsigned long __read_mostly xen_phys_start;
> diff --git a/xen/common/domain.c b/xen/common/domain.c
> index c8414ed..a77f8af 100644
> --- a/xen/common/domain.c
> +++ b/xen/common/domain.c
> @@ -59,7 +59,8 @@ DEFINE_RCU_READ_LOCK(domlist_read_lock);
>  static struct domain *domain_hash[DOMAIN_HASH_SIZE];
>  struct domain *domain_list;
>  
> -struct domain *hardware_domain;
> +struct domain *hardware_domain __read_mostly;
> +domid_t hardware_domid __read_mostly;
>  
>  struct vcpu *idle_vcpu[NR_CPUS] __read_mostly;
>  
> diff --git a/xen/common/domctl.c b/xen/common/domctl.c
> index eebeee7..9af24bf 100644
> --- a/xen/common/domctl.c
> +++ b/xen/common/domctl.c
> @@ -472,6 +472,46 @@ long do_domctl(XEN_GUEST_HANDLE_PARAM(xen_domctl_t) 
> u_domctl)
>              break;
>          }
>  
> +        /*
> +         * Late initialization of the hardware domain is only supported on 
> x86,
> +         * so only check for it there.
> +         */
> +#ifdef LATE_HWDOM_ENABLE
> +        if ( is_hardware_domain(d) )
> +        {
> +            struct domain *dom0 = hardware_domain;
> +            ASSERT(dom0->domain_id == 0);
> +
> +            ret = xsm_init_hardware_domain(XSM_HOOK, d);
> +            if ( ret )
> +            {
> +                domain_kill(d);
> +                d = NULL;
> +                break;
> +            }
> +
> +            printk("Initialising hardware domain %d\n", hardware_domid);
> +            hardware_domain = d;
> +
> +            /* Hardware resource ranges for domain 0 have been set up from
> +             * various sources intended to restrict the hardware domain's
> +             * access. Apply these ranges to the actual hardware domain.
> +             *
> +             * Because the lists are being swapped, a side effect of this
> +             * operation is that Domain 0's rangesets are cleared. Since 
> domain
> +             * 0 should not be accessing the hardware when it constructs a
> +             * hardware domain, this should not be a problem. Both lists 
> should
> +             * be modified after this hypercall returns if a more complex
> +             * device model is desired.
> +             */
> +            rangeset_swap(d->irq_caps, dom0->irq_caps);
> +            rangeset_swap(d->iomem_caps, dom0->iomem_caps);
> +            rangeset_swap(d->arch.ioport_caps, dom0->arch.ioport_caps);
> +
> +            iommu_hwdom_init(d);
> +        }
> +#endif
> +
>          ret = 0;
>  
>          memcpy(d->handle, op->u.createdomain.handle,
> diff --git a/xen/common/rangeset.c b/xen/common/rangeset.c
> index f09c0c4..52fae1f 100644
> --- a/xen/common/rangeset.c
> +++ b/xen/common/rangeset.c
> @@ -438,3 +438,43 @@ void rangeset_domain_printk(
>  
>      spin_unlock(&d->rangesets_lock);
>  }
> +
> +void rangeset_swap(struct rangeset *a, struct rangeset *b)
> +{
> +    struct list_head tmp;
> +    if ( &a < &b )
> +    {
> +        spin_lock(&a->lock);
> +        spin_lock(&b->lock);
> +    }
> +    else
> +    {
> +        spin_lock(&b->lock);
> +        spin_lock(&a->lock);
> +    }
> +    memcpy(&tmp, &a->range_list, sizeof(tmp));
> +    memcpy(&a->range_list, &b->range_list, sizeof(tmp));
> +    memcpy(&b->range_list, &tmp, sizeof(tmp));
> +    if ( a->range_list.next == &b->range_list )
> +    {
> +        a->range_list.next = &a->range_list;
> +        a->range_list.prev = &a->range_list;
> +    }
> +    else
> +    {
> +        a->range_list.next->prev = &a->range_list;
> +        a->range_list.prev->next = &a->range_list;
> +    }
> +    if ( b->range_list.next == &a->range_list )
> +    {
> +        b->range_list.next = &b->range_list;
> +        b->range_list.prev = &b->range_list;
> +    }
> +    else
> +    {
> +        b->range_list.next->prev = &b->range_list;
> +        b->range_list.prev->next = &b->range_list;
> +    }
> +    spin_unlock(&a->lock);
> +    spin_unlock(&b->lock);
> +}
> diff --git a/xen/include/xen/rangeset.h b/xen/include/xen/rangeset.h
> index 1e16a6b..805ebde 100644
> --- a/xen/include/xen/rangeset.h
> +++ b/xen/include/xen/rangeset.h
> @@ -73,4 +73,7 @@ void rangeset_printk(
>  void rangeset_domain_printk(
>      struct domain *d);
>  
> +/* swap contents */
> +void rangeset_swap(struct rangeset *a, struct rangeset *b);
> +
>  #endif /* __XEN_RANGESET_H__ */
> diff --git a/xen/include/xen/sched.h b/xen/include/xen/sched.h
> index 146514d..ce37c77 100644
> --- a/xen/include/xen/sched.h
> +++ b/xen/include/xen/sched.h
> @@ -45,6 +45,7 @@ DEFINE_XEN_GUEST_HANDLE(vcpu_runstate_info_compat_t);
>  
>  /* A global pointer to the hardware domain (usually DOM0). */
>  extern struct domain *hardware_domain;
> +extern domid_t hardware_domid;
>  
>  #ifndef CONFIG_COMPAT
>  #define BITS_PER_EVTCHN_WORD(d) BITS_PER_XEN_ULONG
> @@ -778,7 +779,7 @@ void watchdog_domain_destroy(struct domain *d);
>   *    (that is, this would not be suitable for a driver domain)
>   *  - There is never a reason to deny dom0 access to this
>   */
> -#define is_hardware_domain(_d) ((_d)->domain_id == 0)
> +#define is_hardware_domain(d)  ((d)->domain_id == hardware_domid)
>  
>  /* This check is for functionality specific to a control domain */
>  #define is_control_domain(_d) ((_d)->is_privileged)
> diff --git a/xen/include/xsm/dummy.h b/xen/include/xsm/dummy.h
> index 3bcd941..180cc88 100644
> --- a/xen/include/xsm/dummy.h
> +++ b/xen/include/xsm/dummy.h
> @@ -305,6 +305,12 @@ static XSM_INLINE char *xsm_show_security_evtchn(struct 
> domain *d, const struct
>      return NULL;
>  }
>  
> +static XSM_INLINE int xsm_init_hardware_domain(XSM_DEFAULT_ARG struct 
> domain *d)
> +{
> +    XSM_ASSERT_ACTION(XSM_HOOK);
> +    return xsm_default_action(action, current->domain, d);
> +}
> +
>  static XSM_INLINE int xsm_get_pod_target(XSM_DEFAULT_ARG struct domain *d)
>  {
>      XSM_ASSERT_ACTION(XSM_PRIV);
> diff --git a/xen/include/xsm/xsm.h b/xen/include/xsm/xsm.h
> index de9cf86..6ab9ed1 100644
> --- a/xen/include/xsm/xsm.h
> +++ b/xen/include/xsm/xsm.h
> @@ -83,6 +83,7 @@ struct xsm_operations {
>      int (*alloc_security_evtchn) (struct evtchn *chn);
>      void (*free_security_evtchn) (struct evtchn *chn);
>      char *(*show_security_evtchn) (struct domain *d, const struct evtchn 
> *chn);
> +    int (*init_hardware_domain) (struct domain *d);
>  
>      int (*get_pod_target) (struct domain *d);
>      int (*set_pod_target) (struct domain *d);
> @@ -314,6 +315,11 @@ static inline char *xsm_show_security_evtchn (struct 
> domain *d, const struct evt
>      return xsm_ops->show_security_evtchn(d, chn);
>  }
>  
> +static inline int xsm_init_hardware_domain (xsm_default_t def, struct 
> domain *d)
> +{
> +    return xsm_ops->init_hardware_domain(d);
> +}
> +
>  static inline int xsm_get_pod_target (xsm_default_t def, struct domain *d)
>  {
>      return xsm_ops->get_pod_target(d);
> diff --git a/xen/xsm/dummy.c b/xen/xsm/dummy.c
> index 3fe4c59..689af3d 100644
> --- a/xen/xsm/dummy.c
> +++ b/xen/xsm/dummy.c
> @@ -59,6 +59,8 @@ void xsm_fixup_ops (struct xsm_operations *ops)
>      set_to_dummy_if_null(ops, alloc_security_evtchn);
>      set_to_dummy_if_null(ops, free_security_evtchn);
>      set_to_dummy_if_null(ops, show_security_evtchn);
> +    set_to_dummy_if_null(ops, init_hardware_domain);
> +
>      set_to_dummy_if_null(ops, get_pod_target);
>      set_to_dummy_if_null(ops, set_pod_target);
>  
> diff --git a/xen/xsm/flask/hooks.c b/xen/xsm/flask/hooks.c
> index 96276ac..5b906d6 100644
> --- a/xen/xsm/flask/hooks.c
> +++ b/xen/xsm/flask/hooks.c
> @@ -344,6 +344,11 @@ static char *flask_show_security_evtchn(struct domain 
> *d, const struct evtchn *c
>      return ctx;
>  }
>  
> +static int flask_init_hardware_domain(struct domain *d)
> +{
> +    return current_has_perm(d, SECCLASS_DOMAIN2, 
> DOMAIN2__CREATE_HARDWARE_DOMAIN);
> +}
> +
>  static int flask_grant_mapref(struct domain *d1, struct domain *d2, 
>                                uint32_t flags)
>  {
> @@ -1491,6 +1496,7 @@ static struct xsm_operations flask_ops = {
>      .alloc_security_evtchn = flask_alloc_security_evtchn,
>      .free_security_evtchn = flask_free_security_evtchn,
>      .show_security_evtchn = flask_show_security_evtchn,
> +    .init_hardware_domain = flask_init_hardware_domain,
>  
>      .get_pod_target = flask_get_pod_target,
>      .set_pod_target = flask_set_pod_target,
> diff --git a/xen/xsm/flask/policy/access_vectors 
> b/xen/xsm/flask/policy/access_vectors
> index a0ed13d..32371a9 100644
> --- a/xen/xsm/flask/policy/access_vectors
> +++ b/xen/xsm/flask/policy/access_vectors
> @@ -198,6 +198,8 @@ class domain2
>      set_max_evtchn
>  # XEN_DOMCTL_cacheflush
>      cacheflush
> +# Creation of the hardware domain when it is not dom0
> +    create_hardware_domain
>  }
>  
>  # Similar to class domain, but primarily contains domctls related to HVM 
> domains
> -- 
> 1.8.5.3



_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.