[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-users] Re: XEN 3.1: critical bug: vif init failure after creating 15-17 VMs (XENBUS: Timeout connecting to device: device/vif)

What dom0 kernel image are you running? It looks like vmalloc_sync_all(), called from alloc_vm_area() has not caused the pte that will map the rx ring to be made present in the currently-running page tables. The code looks okay on inspection though.

 -- Keir

On 14/7/07 03:32, "Eric Tessler" <maiden1134@xxxxxxxxx> wrote:

Also, by suggestion of Keir, I tried the XEN 3.0.4 kernel in my 16th VM (, it failed the same way. The only difference is that instead of extending the grant table from 4 to 5 frames, it was extended from 4 to 16 frames:
   (XEN) grant_table.c:557:d18 Expanding dom (18) grant table from (4) to (16) frames.
   (XEN) mm.c:2605:d0 Could not find L1 PTE for address d1400000
I believe the following stack trace represents the trace of the failure (starting from within XenBus, traced by hand):
connect_rings                       linux-2.6-xen-sparse/drivers/xen/netback/xenbus.c
netif_map                           linux-2.6-xen-sparse/drivers/xen/netback/interface.c
map_frontend_pages                  linux-2.6-xen-sparse/drivers/xen/netback/interface.c
__gnttab_map_grant_ref (hypercall)             xen/common/grant_table.c
create_grant_host_mapping           xen/arch/x86/mm.c
create_grant_va_mapping             xen/arch/x86/mm.c
   guest_map_l1e                    xen/arch/x86/mm.c
     (this is the function that is ultimately failing)
Any clue as to what is causing this failure or how to fix it? Is there any other debug info I can provide here that would be of any help in resolving this issue? I have some free time tomorrow to debug this issue, but need some direction; this is in an area of XEN I don't understand very well.

Xen-users mailing list



Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.