[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH 1/1] xen/blkback: rework connect_ring() to avoid inconsistent xenstore 'ring-page-order' set by malicious blkfront
> -----Original Message----- > From: Dongli Zhang [mailto:dongli.zhang@xxxxxxxxxx] > Sent: 07 December 2018 15:10 > To: Paul Durrant <Paul.Durrant@xxxxxxxxxx>; linux-kernel@xxxxxxxxxxxxxxx; > xen-devel@xxxxxxxxxxxxxxxxxxxx; linux-block@xxxxxxxxxxxxxxx > Cc: axboe@xxxxxxxxx; Roger Pau Monne <roger.pau@xxxxxxxxxx>; > konrad.wilk@xxxxxxxxxx > Subject: Re: [Xen-devel] [PATCH 1/1] xen/blkback: rework connect_ring() to > avoid inconsistent xenstore 'ring-page-order' set by malicious blkfront > > Hi Paul, > > On 12/07/2018 05:39 PM, Paul Durrant wrote: > >> -----Original Message----- > >> From: Xen-devel [mailto:xen-devel-bounces@xxxxxxxxxxxxxxxxxxxx] On > Behalf > >> Of Dongli Zhang > >> Sent: 07 December 2018 04:18 > >> To: linux-kernel@xxxxxxxxxxxxxxx; xen-devel@xxxxxxxxxxxxxxxxxxxx; > linux- > >> block@xxxxxxxxxxxxxxx > >> Cc: axboe@xxxxxxxxx; Roger Pau Monne <roger.pau@xxxxxxxxxx>; > >> konrad.wilk@xxxxxxxxxx > >> Subject: [Xen-devel] [PATCH 1/1] xen/blkback: rework connect_ring() to > >> avoid inconsistent xenstore 'ring-page-order' set by malicious blkfront > >> > >> The xenstore 'ring-page-order' is used globally for each blkback queue > and > >> therefore should be read from xenstore only once. However, it is > obtained > >> in read_per_ring_refs() which might be called multiple times during the > >> initialization of each blkback queue. > > > > That is certainly sub-optimal. > > > >> > >> If the blkfront is malicious and the 'ring-page-order' is set in > different > >> value by blkfront every time before blkback reads it, this may end up > at > >> the "WARN_ON(i != (XEN_BLKIF_REQS_PER_PAGE * blkif->nr_ring_pages));" > in > >> xen_blkif_disconnect() when frontend is destroyed. > > > > I can't actually see what useful function blkif->nr_ring_pages actually > performs any more. Perhaps you could actually get rid of it? > > How about we keep it? Other than reading from xenstore, it is the only > place for > us to know the value from 'ring-page-order'. > > This helps calculate the initialized number of elements on all > xen_blkif_ring->pending_free lists. That's how "WARN_ON(i != > (XEN_BLKIF_REQS_PER_PAGE * blkif->nr_ring_pages));" is used to double > check if > there is no leak of elements reclaimed from all xen_blkif_ring- > >pending_free. > > It helps vmcore analysis as well. Given blkif->nr_ring_pages, we would be > able > to double check if the number of ring buffer slots are correct. > > I could not see any drawback leaving blkif->nr_ring_pagesin the code. No, there's no drawback apart from space, but apart from that cross-check and, as you say, core analysis it seems to have little value. Paul > > Dongli Zhang > > > > >> > >> This patch reworks connect_ring() to read xenstore 'ring-page-order' > only > >> once. > > > > That is certainly a good thing :-) > > > > Paul > > > >> > >> Signed-off-by: Dongli Zhang <dongli.zhang@xxxxxxxxxx> > >> --- > >> drivers/block/xen-blkback/xenbus.c | 49 ++++++++++++++++++++++++------ > --- > >> ----- > >> 1 file changed, 31 insertions(+), 18 deletions(-) > >> > >> diff --git a/drivers/block/xen-blkback/xenbus.c b/drivers/block/xen- > >> blkback/xenbus.c > >> index a4bc74e..4a8ce20 100644 > >> --- a/drivers/block/xen-blkback/xenbus.c > >> +++ b/drivers/block/xen-blkback/xenbus.c > >> @@ -919,14 +919,15 @@ static void connect(struct backend_info *be) > >> /* > >> * Each ring may have multi pages, depends on "ring-page-order". > >> */ > >> -static int read_per_ring_refs(struct xen_blkif_ring *ring, const char > >> *dir) > >> +static int read_per_ring_refs(struct xen_blkif_ring *ring, const char > >> *dir, > >> + bool use_ring_page_order) > >> { > >> unsigned int ring_ref[XENBUS_MAX_RING_GRANTS]; > >> struct pending_req *req, *n; > >> int err, i, j; > >> struct xen_blkif *blkif = ring->blkif; > >> struct xenbus_device *dev = blkif->be->dev; > >> - unsigned int ring_page_order, nr_grefs, evtchn; > >> + unsigned int nr_grefs, evtchn; > >> > >> err = xenbus_scanf(XBT_NIL, dir, "event-channel", "%u", > >> &evtchn); > >> @@ -936,28 +937,18 @@ static int read_per_ring_refs(struct > xen_blkif_ring > >> *ring, const char *dir) > >> return err; > >> } > >> > >> - err = xenbus_scanf(XBT_NIL, dev->otherend, "ring-page-order", "%u", > >> - &ring_page_order); > >> - if (err != 1) { > >> + nr_grefs = blkif->nr_ring_pages; > >> + > >> + if (!use_ring_page_order) { > >> err = xenbus_scanf(XBT_NIL, dir, "ring-ref", "%u", > >> &ring_ref[0]); > >> if (err != 1) { > >> err = -EINVAL; > >> xenbus_dev_fatal(dev, err, "reading %s/ring-ref", dir); > >> return err; > >> } > >> - nr_grefs = 1; > >> } else { > >> unsigned int i; > >> > >> - if (ring_page_order > xen_blkif_max_ring_order) { > >> - err = -EINVAL; > >> - xenbus_dev_fatal(dev, err, "%s/request %d ring page > >> order exceed max:%d", > >> - dir, ring_page_order, > >> - xen_blkif_max_ring_order); > >> - return err; > >> - } > >> - > >> - nr_grefs = 1 << ring_page_order; > >> for (i = 0; i < nr_grefs; i++) { > >> char ring_ref_name[RINGREF_NAME_LEN]; > >> > >> @@ -972,7 +963,6 @@ static int read_per_ring_refs(struct xen_blkif_ring > >> *ring, const char *dir) > >> } > >> } > >> } > >> - blkif->nr_ring_pages = nr_grefs; > >> > >> for (i = 0; i < nr_grefs * XEN_BLKIF_REQS_PER_PAGE; i++) { > >> req = kzalloc(sizeof(*req), GFP_KERNEL); > >> @@ -1030,6 +1020,8 @@ static int connect_ring(struct backend_info *be) > >> size_t xspathsize; > >> const size_t xenstore_path_ext_size = 11; /* sufficient for "/queue- > >> NNN" */ > >> unsigned int requested_num_queues = 0; > >> + bool use_ring_page_order = false; > >> + unsigned int ring_page_order; > >> > >> pr_debug("%s %s\n", __func__, dev->otherend); > >> > >> @@ -1075,8 +1067,28 @@ static int connect_ring(struct backend_info *be) > >> be->blkif->nr_rings, be->blkif->blk_protocol, protocol, > >> pers_grants ? "persistent grants" : ""); > >> > >> + err = xenbus_scanf(XBT_NIL, dev->otherend, "ring-page-order", "%u", > >> + &ring_page_order); > >> + > >> + if (err != 1) { > >> + be->blkif->nr_ring_pages = 1; > >> + } else { > >> + if (ring_page_order > xen_blkif_max_ring_order) { > >> + err = -EINVAL; > >> + xenbus_dev_fatal(dev, err, > >> + "requested ring page order %d exceed > >> max:%d", > >> + ring_page_order, > >> + xen_blkif_max_ring_order); > >> + return err; > >> + } > >> + > >> + use_ring_page_order = true; > >> + be->blkif->nr_ring_pages = 1 << ring_page_order; > >> + } > >> + > >> if (be->blkif->nr_rings == 1) > >> - return read_per_ring_refs(&be->blkif->rings[0], dev- > >>> otherend); > >> + return read_per_ring_refs(&be->blkif->rings[0], dev->otherend, > >> + use_ring_page_order); > >> else { > >> xspathsize = strlen(dev->otherend) + xenstore_path_ext_size; > >> xspath = kmalloc(xspathsize, GFP_KERNEL); > >> @@ -1088,7 +1100,8 @@ static int connect_ring(struct backend_info *be) > >> for (i = 0; i < be->blkif->nr_rings; i++) { > >> memset(xspath, 0, xspathsize); > >> snprintf(xspath, xspathsize, "%s/queue-%u", dev- > >>> otherend, i); > >> - err = read_per_ring_refs(&be->blkif->rings[i], xspath); > >> + err = read_per_ring_refs(&be->blkif->rings[i], xspath, > >> + use_ring_page_order); > >> if (err) { > >> kfree(xspath); > >> return err; > >> -- > >> 2.7.4 > >> > >> > >> _______________________________________________ > >> Xen-devel mailing list > >> Xen-devel@xxxxxxxxxxxxxxxxxxxx > >> https://lists.xenproject.org/mailman/listinfo/xen-devel _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxxx https://lists.xenproject.org/mailman/listinfo/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |