[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH net v2 1/3] xen-netback: remove pointless clause from if statement
> -----Original Message----- > From: Sander Eikelenboom [mailto:linux@xxxxxxxxxxxxxx] > Sent: 27 March 2014 16:46 > To: Paul Durrant > Cc: xen-devel@xxxxxxxxxxxxx; netdev@xxxxxxxxxxxxxxx; Ian Campbell; Wei Liu > Subject: Re: [PATCH net v2 1/3] xen-netback: remove pointless clause from if > statement > > > Thursday, March 27, 2014, 3:09:32 PM, you wrote: > > >> -----Original Message----- > >> From: Sander Eikelenboom [mailto:linux@xxxxxxxxxxxxxx] > >> Sent: 27 March 2014 14:03 > >> To: Paul Durrant > >> Cc: xen-devel@xxxxxxxxxxxxx; netdev@xxxxxxxxxxxxxxx; Ian Campbell; Wei > Liu > >> Subject: Re: [PATCH net v2 1/3] xen-netback: remove pointless clause > from if > >> statement > >> > >> > >> Thursday, March 27, 2014, 2:54:46 PM, you wrote: > >> > >> >> -----Original Message----- > >> >> From: Sander Eikelenboom [mailto:linux@xxxxxxxxxxxxxx] > >> >> Sent: 27 March 2014 13:46 > >> >> To: Paul Durrant > >> >> Cc: xen-devel@xxxxxxxxxxxxx; netdev@xxxxxxxxxxxxxxx; Ian Campbell; > Wei > >> Liu > >> >> Subject: Re: [PATCH net v2 1/3] xen-netback: remove pointless clause > >> from if > >> >> statement > >> >> > >> >> > >> >> Thursday, March 27, 2014, 1:56:11 PM, you wrote: > >> >> > >> >> > This patch removes a test in start_new_rx_buffer() that checks > whether > >> >> > a copy operation is less than MAX_BUFFER_OFFSET in length, since > >> >> > MAX_BUFFER_OFFSET is defined to be PAGE_SIZE and the only caller > of > >> >> > start_new_rx_buffer() already limits copy operations to PAGE_SIZE > or > >> less. > >> >> > >> >> > Signed-off-by: Paul Durrant <paul.durrant@xxxxxxxxxx> > >> >> > Cc: Ian Campbell <ian.campbell@xxxxxxxxxx> > >> >> > Cc: Wei Liu <wei.liu2@xxxxxxxxxx> > >> >> > Cc: Sander Eikelenboom <linux@xxxxxxxxxxxxxx> > >> >> > --- > >> >> > >> >> > v2: > >> >> > - Add BUG_ON() as suggested by Ian Campbell > >> >> > >> >> > drivers/net/xen-netback/netback.c | 4 ++-- > >> >> > 1 file changed, 2 insertions(+), 2 deletions(-) > >> >> > >> >> > diff --git a/drivers/net/xen-netback/netback.c b/drivers/net/xen- > >> >> netback/netback.c > >> >> > index 438d0c0..72314c7 100644 > >> >> > --- a/drivers/net/xen-netback/netback.c > >> >> > +++ b/drivers/net/xen-netback/netback.c > >> >> > @@ -192,8 +192,8 @@ static bool start_new_rx_buffer(int offset, > >> >> unsigned long size, int head) > >> >> > * into multiple copies tend to give large frags their > >> >> > * own buffers as before. > >> >> > */ > >> >> > - if ((offset + size > MAX_BUFFER_OFFSET) && > >> >> > - (size <= MAX_BUFFER_OFFSET) && offset && !head) > >> >> > + BUG_ON(size > MAX_BUFFER_OFFSET); > >> >> > + if ((offset + size > MAX_BUFFER_OFFSET) && offset && !head) > >> >> > return true; > >> >> > > >> >> > return false; > >> >> > >> >> Hi Paul, > >> >> > >> >> Unfortunately .. no good .. > >> >> > >> >> With these patches (v2) applied to 3.14-rc8 it all seems to work well, > >> >> until i do my test case .. it still chokes and now effectively > >> >> permanently > >> stalls > >> >> network traffic to that guest. > >> >> > >> >> No error messages or anything in either xl dmesg or dmesg on the host > .. > >> and > >> >> nothing in dmesg in the guest either. > >> >> > >> >> But in the guest the TX bytes ifconfig reports for eth0 still increase > >> >> but > RX > >> >> bytes does nothing, so it seems only the RX path is effected) > >> >> > >> > >> > But you're not getting ring overflow, right? So that suggests this > >> > series is > >> working and you're now hitting another problem? I don't see how these > >> patches could directly cause the new behaviour you're seeing. > >> > >> Don't know .. how ever .. i previously tested: > >> - unconditionally doing "max_slots_needed + 1" in > >> "net_rx_action()", > >> and that circumvented the problem reliably without causing anything else > >> - reverting the calculation of "max_slots_needed + 1" in > >> "net_rx_action()" to what it was before : > >> int max = DIV_ROUND_UP(vif->dev->mtu, PAGE_SIZE); > >> if (vif->can_sg || vif->gso_mask || vif->gso_prefix_mask) > >> max += MAX_SKB_FRAGS + 1; /* extra_info + frags */ > >> > > > So, it may be that the worse-case estimate is now too bad. In the case > where it's failing for you it would be nice to know what the estimate was. > Looking at netfront, we could be in trouble if it ever goes above 64. > > With your patches + some extra printk's > > Ok you seem to be right .. > > [ 967.957014] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 26 > [ 968.164711] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 26 > [ 968.310899] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 26 > [ 968.674412] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 26 > [ 968.928398] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 26 > [ 969.105993] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 26 > [ 969.434961] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 26 > [ 969.719368] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 26 > [ 969.729606] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 26 > [ 970.195451] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 26 > [ 970.493106] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 26 > [ 970.581056] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 26 > [ 970.594934] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 26 > [ 970.754355] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 26 > [ 970.991755] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 26 > [ 976.978261] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 26 frag:8 size:1460 offset:15478 > [ 976.980183] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 31 frag:9 size:313 offset:17398 > [ 976.982154] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 32 > [ 976.984078] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 27 frag:3 size:1460 offset:25846 > [ 976.986466] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 35 frag:4 size:1460 offset:27766 > [ 976.988540] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 43 frag:5 size:1460 offset:29686 > [ 976.990809] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 44 frag:6 size:1460 offset:118 > [ 976.993038] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 45 frag:7 size:1460 offset:2038 > [ 976.994966] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 47 frag:8 size:1460 offset:3958 > [ 976.996987] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 49 frag:9 size:1460 offset:5878 > [ 976.998947] vif vif-7-0 vif7.0: ?!?!? xenvif_rx_action Insane numer of > slots > needed: 52 > [ 982.624852] net_ratelimit: 97 callbacks suppressed > [ 982.627279] vif vif-7-0 vif7.0: ?!?!? rx_work_todo waiting for > rx_last_skb_slots: 52 > [ 982.637833] vif vif-7-0 vif7.0: ?!?!? rx_work_todo waiting for > rx_last_skb_slots: 52 > [ 983.432482] vif vif-7-0 vif7.0: ?!?!? rx_work_todo waiting for > rx_last_skb_slots: 52 > [ 983.640017] vif vif-7-0 vif7.0: ?!?!? rx_work_todo waiting for > rx_last_skb_slots: 52 > [ 983.642974] vif vif-7-0 vif7.0: ?!?!? rx_work_todo waiting for > rx_last_skb_slots: 52 > [ 983.656408] vif vif-7-0 vif7.0: ?!?!? rx_work_todo waiting for > rx_last_skb_slots: 52 > [ 983.779142] vif vif-7-0 vif7.0: ?!?!? rx_work_todo waiting for > rx_last_skb_slots: 52 > [ 984.644546] vif vif-7-0 vif7.0: ?!?!? rx_work_todo waiting for > rx_last_skb_slots: 52 > [ 984.657728] vif vif-7-0 vif7.0: ?!?!? rx_work_todo waiting for > rx_last_skb_slots: 52 > [ 985.459147] vif vif-7-0 vif7.0: ?!?!? rx_work_todo waiting for > rx_last_skb_slots: 52 > [ 987.668407] net_ratelimit: 8 callbacks suppressed > [ 987.671661] vif vif-7-0 vif7.0: ?!?!? rx_work_todo waiting for > rx_last_skb_slots: 52 > [ 987.678483] vif vif-7-0 vif7.0: ?!?!? rx_work_todo waiting for > rx_last_skb_slots: 52 > [ 988.671510] vif vif-7-0 vif7.0: ?!?!? rx_work_todo waiting for > rx_last_skb_slots: 52 > [ 988.681210] vif vif-7-0 vif7.0: ?!?!? rx_work_todo waiting for > rx_last_skb_slots: 52 > [ 989.472372] vif vif-7-0 vif7.0: ?!?!? rx_work_todo waiting for > rx_last_skb_slots: 52 > [ 989.685166] vif vif-7-0 vif7.0: ?!?!? rx_work_todo waiting for > rx_last_skb_slots: 52 > [ 989.700220] vif vif-7-0 vif7.0: ?!?!? rx_work_todo waiting for > rx_last_skb_slots: 52 > [ 990.058987] vif vif-7-0 vif7.0: ?!?!? rx_work_todo waiting for > rx_last_skb_slots: 52 > [ 990.192480] vif vif-7-0 vif7.0: ?!?!? rx_work_todo waiting for > rx_last_skb_slots: 52 > [ 990.687626] vif vif-7-0 vif7.0: ?!?!? rx_work_todo waiting for > rx_last_skb_slots: 52 > [ 992.707878] net_ratelimit: 5 callbacks suppressed > > But it's done at 52 instead of 64. > > And this worst case is a *lot* larger (and probably needless .. since i > previously could do with only one slot extra). > Ok, so we cannot be too pessimistic. In that case I don't see there's a lot of choice but to stick with the existing DIV_ROUND_UP (i.e. don't assume start_new_rx_buffer() returns true every time) and just add the extra 1. Paul > > > > > Paul > > >> So that leads me to think it's something caused by this patch set. > >> Patch1 could be a candidate .. perhaps that check was needed for some > >> reason .. will see what not applying that one does > >> > >> -- > >> Sander > >> > >> > >> > Paul > >> > >> >> So it now seems i now have the situation which you described in the > >> commit > >> >> message from "ca2f09f2b2c6c25047cfc545d057c4edfcfe561c", > >> >> "Without this patch I can trivially stall netback permanently by just > doing a > >> >> large guest to guest file copy between two Windows Server 2008R2 > VMs > >> on a > >> >> single host." > >> >> > >> >> -- > >> >> Sander > >> > >> > > _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |