[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

[Xen-devel] [RFC PATCH 06/13] xen-netback: copy buffer on xenvif_start_xmit()



By introducing persistent grants we speed up the RX thread with the
decreased copy cost, that leads to a throughput decrease of 20%.
It is observed that the rx_queue stays mostly at 10% of its capacity,
as opposed to full capacity when using grant copy. And a finer measure
with lock_stat (below with pkt_size 64, burst 1) shows much higher wait
queue contention on queue->wq, which hints that the RX kthread is
waits/wake_up more often, that is actually doing work.

Without persistent grants:

class name    con-bounces    contentions   waittime-min   waittime-max
waittime-total   waittime-avg    acq-bounces   acquisitions   holdtime-min
holdtime-max holdtime-total   holdtime-avg
--------------------------------------------------------------------------
&queue->wq:   792            792           0.36          24.36
1140.30           1.44           4208        1002671           0.00
46.75      538164.02           0.54
----------
&queue->wq    326          [<ffffffff8115949f>] __wake_up+0x2f/0x80
&queue->wq    410          [<ffffffff811592bf>] finish_wait+0x4f/0xa0
&queue->wq     56          [<ffffffff811593eb>] prepare_to_wait+0x2b/0xb0
----------
&queue->wq    202          [<ffffffff811593eb>] prepare_to_wait+0x2b/0xb0
&queue->wq    467          [<ffffffff8115949f>] __wake_up+0x2f/0x80
&queue->wq    123          [<ffffffff811592bf>] finish_wait+0x4f/0xa0

With persistent grants:

&queue->wq:   61834          61836           0.32          30.12
99710.27           1.61         241400        1125308           0.00
75.61     1106578.82           0.98
----------
&queue->wq     5079        [<ffffffff8115949f>] __wake_up+0x2f/0x80
&queue->wq    56280        [<ffffffff811592bf>] finish_wait+0x4f/0xa0
&queue->wq      479        [<ffffffff811593eb>] prepare_to_wait+0x2b/0xb0
----------
&queue->wq     1005        [<ffffffff811592bf>] finish_wait+0x4f/0xa0
&queue->wq    56761        [<ffffffff8115949f>] __wake_up+0x2f/0x80
&queue->wq     4072        [<ffffffff811593eb>] prepare_to_wait+0x2b/0xb0

Also, with persistent grants, we don't require batching grant copy ops
(besides the initial copy+map) which makes me believe that deferring
the skb to the RX kthread just adds up unnecessary overhead (for this
particular case). This patch proposes copying the buffer on
xenvif_start_xmit(), which lets us both remove the contention on
queue->wq and lock on rx_queue. Here, an alternative to
xenvif_rx_action routine is added namely xenvif_rx_map() that maps
and copies the buffer to the guest. This is only used when persistent
grants are used, since it would otherwise mean an hypercall per
packet.

Improvements are up to a factor of 2.14 with a single queue getting us
from 1.04 Mpps to 1.7 Mpps (burst 1, pkt_size 64) and 1.5 to 2.6 Mpps
(burst 2, pkt_size 64) compared to using the kthread. Maximum with grant
copy is 1.2 Mpps, irrespective of the burst. All of this, measured on
an Intel(R) Xeon(R) CPU E5-1650 v2 @ 3.50GHz.

Signed-off-by: Joao Martins <joao.martins@xxxxxxxxx>
---
 drivers/net/xen-netback/common.h    |  2 ++
 drivers/net/xen-netback/interface.c | 11 +++++---
 drivers/net/xen-netback/netback.c   | 52 +++++++++++++++++++++++++++++--------
 3 files changed, 51 insertions(+), 14 deletions(-)

diff --git a/drivers/net/xen-netback/common.h b/drivers/net/xen-netback/common.h
index 23deb6a..f3ece12 100644
--- a/drivers/net/xen-netback/common.h
+++ b/drivers/net/xen-netback/common.h
@@ -363,6 +363,8 @@ void xenvif_kick_thread(struct xenvif_queue *queue);
 
 int xenvif_dealloc_kthread(void *data);
 
+int xenvif_rx_map(struct xenvif_queue *queue, struct sk_buff *skb);
+
 void xenvif_rx_queue_tail(struct xenvif_queue *queue, struct sk_buff *skb);
 
 /* Determine whether the needed number of slots (req) are available,
diff --git a/drivers/net/xen-netback/interface.c 
b/drivers/net/xen-netback/interface.c
index 1103568..dfe2b7b 100644
--- a/drivers/net/xen-netback/interface.c
+++ b/drivers/net/xen-netback/interface.c
@@ -109,7 +109,8 @@ static irqreturn_t xenvif_rx_interrupt(int irq, void 
*dev_id)
 {
        struct xenvif_queue *queue = dev_id;
 
-       xenvif_kick_thread(queue);
+       if (!queue->vif->persistent_grants)
+               xenvif_kick_thread(queue);
 
        return IRQ_HANDLED;
 }
@@ -168,8 +169,12 @@ static int xenvif_start_xmit(struct sk_buff *skb, struct 
net_device *dev)
        cb = XENVIF_RX_CB(skb);
        cb->expires = jiffies + vif->drain_timeout;
 
-       xenvif_rx_queue_tail(queue, skb);
-       xenvif_kick_thread(queue);
+       if (!queue->vif->persistent_grants) {
+               xenvif_rx_queue_tail(queue, skb);
+               xenvif_kick_thread(queue);
+       } else if (xenvif_rx_map(queue, skb)) {
+               return NETDEV_TX_BUSY;
+       }
 
        return NETDEV_TX_OK;
 
diff --git a/drivers/net/xen-netback/netback.c 
b/drivers/net/xen-netback/netback.c
index c4f57d7..228df92 100644
--- a/drivers/net/xen-netback/netback.c
+++ b/drivers/net/xen-netback/netback.c
@@ -883,9 +883,48 @@ static bool xenvif_rx_submit(struct xenvif_queue *queue,
        return !!ret;
 }
 
+int xenvif_rx_map(struct xenvif_queue *queue, struct sk_buff *skb)
+{
+       int ret = -EBUSY;
+       struct netrx_pending_operations npo = {
+               .copy  = queue->grant_copy_op,
+               .meta  = queue->meta
+       };
+
+       if (!xenvif_rx_ring_slots_available(queue, XEN_NETBK_LEGACY_SLOTS_MAX))
+               goto done;
+
+       xenvif_rx_build_gops(queue, &npo, skb);
+
+       BUG_ON(npo.meta_prod > ARRAY_SIZE(queue->meta));
+       if (!npo.copy_done && !npo.copy_prod)
+               goto done;
+
+       BUG_ON(npo.map_prod > MAX_GRANT_COPY_OPS);
+       if (npo.map_prod) {
+               ret = gnttab_map_refs(queue->rx_map_ops,
+                                     NULL,
+                                     queue->rx_pages_to_map,
+                                     npo.map_prod);
+               BUG_ON(ret);
+       }
+
+       BUG_ON(npo.copy_prod > MAX_GRANT_COPY_OPS);
+       if (npo.copy_prod)
+               gnttab_batch_copy(npo.copy, npo.copy_prod);
+
+       if (xenvif_rx_submit(queue, &npo, skb))
+               notify_remote_via_irq(queue->rx_irq);
+
+       ret = 0; /* clear error */
+done:
+       if (xenvif_queue_stopped(queue))
+               xenvif_wake_queue(queue);
+       return ret;
+}
+
 static void xenvif_rx_action(struct xenvif_queue *queue)
 {
-       int ret;
        struct sk_buff *skb;
        struct sk_buff_head rxq;
        bool need_to_notify = false;
@@ -905,22 +944,13 @@ static void xenvif_rx_action(struct xenvif_queue *queue)
        }
 
        BUG_ON(npo.meta_prod > XEN_NETIF_RX_RING_SIZE);
-       if (!npo.copy_done && !npo.copy_prod)
+       if (!npo.copy_prod)
                return;
 
        BUG_ON(npo.copy_prod > MAX_GRANT_COPY_OPS);
        if (npo.copy_prod)
                gnttab_batch_copy(npo.copy, npo.copy_prod);
 
-       BUG_ON(npo.map_prod > MAX_GRANT_COPY_OPS);
-       if (npo.map_prod) {
-               ret = gnttab_map_refs(queue->rx_map_ops,
-                                     NULL,
-                                     queue->rx_pages_to_map,
-                                     npo.map_prod);
-               BUG_ON(ret);
-       }
-
        while ((skb = __skb_dequeue(&rxq)) != NULL)
                need_to_notify |= xenvif_rx_submit(queue, &npo, skb);
 
-- 
2.1.3


_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.