|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [PATCH for-4.19???] x86/physdev: replace physdev_{,un}map_pirq() checking against DOMID_SELF
It's hardly ever correct to check for just DOMID_SELF, as guests have
ways to figure out their domain IDs and hence could instead use those as
inputs to respective hypercalls. Note, however, that for ordinary DomU-s
the adjustment is relaxing things rather than tightening them, since
- as a result of XSA-237 - the respective XSM checks would have rejected
self (un)mapping attempts for other than the control domain.
Since in physdev_map_pirq() handling overall is a little easier this
way, move obtaining of the domain pointer into the caller. Doing the
same for physdev_unmap_pirq() is just to keep both consistent in this
regard. For both this has the advantage that it is now provable (by the
build not failing) that there are no DOMID_SELF checks left (and none
could easily be re-added).
Fixes: 0b469cd68708 ("Interrupt remapping to PIRQs in HVM guests")
Fixes: 9e1a3415b773 ("x86: fixes after emuirq changes")
Signed-off-by: Jan Beulich <jbeulich@xxxxxxxx>
---
Note that the moving of rcu_lock_domain_by_any_id() is also going to
help
https://lists.xen.org/archives/html/xen-devel/2024-06/msg00206.html.
--- a/xen/arch/x86/physdev.c
+++ b/xen/arch/x86/physdev.c
@@ -18,9 +18,9 @@
#include <xsm/xsm.h>
#include <asm/p2m.h>
-int physdev_map_pirq(domid_t domid, int type, int *index, int *pirq_p,
+int physdev_map_pirq(struct domain *d, int type, int *index, int *pirq_p,
struct msi_info *msi);
-int physdev_unmap_pirq(domid_t domid, int pirq);
+int physdev_unmap_pirq(struct domain *d, int pirq);
#include "x86_64/mmconfig.h"
@@ -88,13 +88,12 @@ static int physdev_hvm_map_pirq(
return ret;
}
-int physdev_map_pirq(domid_t domid, int type, int *index, int *pirq_p,
+int physdev_map_pirq(struct domain *d, int type, int *index, int *pirq_p,
struct msi_info *msi)
{
- struct domain *d = current->domain;
int ret;
- if ( domid == DOMID_SELF && is_hvm_domain(d) && has_pirq(d) )
+ if ( d == current->domain && is_hvm_domain(d) && has_pirq(d) )
{
/*
* Only makes sense for vector-based callback, else HVM-IRQ logic
@@ -106,13 +105,9 @@ int physdev_map_pirq(domid_t domid, int
return physdev_hvm_map_pirq(d, type, index, pirq_p);
}
- d = rcu_lock_domain_by_any_id(domid);
- if ( d == NULL )
- return -ESRCH;
-
ret = xsm_map_domain_pirq(XSM_DM_PRIV, d);
if ( ret )
- goto free_domain;
+ return ret;
/* Verify or get irq. */
switch ( type )
@@ -135,24 +130,17 @@ int physdev_map_pirq(domid_t domid, int
break;
}
- free_domain:
- rcu_unlock_domain(d);
return ret;
}
-int physdev_unmap_pirq(domid_t domid, int pirq)
+int physdev_unmap_pirq(struct domain *d, int pirq)
{
- struct domain *d;
int ret = 0;
- d = rcu_lock_domain_by_any_id(domid);
- if ( d == NULL )
- return -ESRCH;
-
- if ( domid != DOMID_SELF || !is_hvm_domain(d) || !has_pirq(d) )
+ if ( d != current->domain || !is_hvm_domain(d) || !has_pirq(d) )
ret = xsm_unmap_domain_pirq(XSM_DM_PRIV, d);
if ( ret )
- goto free_domain;
+ return ret;
if ( is_hvm_domain(d) && has_pirq(d) )
{
@@ -160,8 +148,8 @@ int physdev_unmap_pirq(domid_t domid, in
if ( domain_pirq_to_emuirq(d, pirq) != IRQ_UNBOUND )
ret = unmap_domain_pirq_emuirq(d, pirq);
write_unlock(&d->event_lock);
- if ( domid == DOMID_SELF || ret )
- goto free_domain;
+ if ( d == current->domain || ret )
+ return ret;
}
pcidevs_lock();
@@ -170,8 +158,6 @@ int physdev_unmap_pirq(domid_t domid, in
write_unlock(&d->event_lock);
pcidevs_unlock();
- free_domain:
- rcu_unlock_domain(d);
return ret;
}
#endif /* COMPAT */
@@ -184,6 +170,8 @@ ret_t do_physdev_op(int cmd, XEN_GUEST_H
switch ( cmd )
{
+ struct domain *d;
+
case PHYSDEVOP_eoi: {
struct physdev_eoi eoi;
struct pirq *pirq;
@@ -331,8 +319,15 @@ ret_t do_physdev_op(int cmd, XEN_GUEST_H
msi.sbdf.devfn = map.devfn;
msi.entry_nr = map.entry_nr;
msi.table_base = map.table_base;
- ret = physdev_map_pirq(map.domid, map.type, &map.index, &map.pirq,
- &msi);
+
+ d = rcu_lock_domain_by_any_id(map.domid);
+ ret = -ESRCH;
+ if ( !d )
+ break;
+
+ ret = physdev_map_pirq(d, map.type, &map.index, &map.pirq, &msi);
+
+ rcu_unlock_domain(d);
if ( map.type == MAP_PIRQ_TYPE_MULTI_MSI )
map.entry_nr = msi.entry_nr;
@@ -348,7 +343,15 @@ ret_t do_physdev_op(int cmd, XEN_GUEST_H
if ( copy_from_guest(&unmap, arg, 1) != 0 )
break;
- ret = physdev_unmap_pirq(unmap.domid, unmap.pirq);
+ d = rcu_lock_domain_by_any_id(unmap.domid);
+ ret = -ESRCH;
+ if ( !d )
+ break;
+
+ ret = physdev_unmap_pirq(d, unmap.pirq);
+
+ rcu_unlock_domain(d);
+
break;
}
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |