[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

RE: [Xen-users] Error starting stubdom HVM on Xen-3.4.3-rc4-pre



I found this link:

http://lists.xensource.com/archives/html/xen-devel/2009-05/msg00698.html

Was Xen 3.4.x ever fixed to used phy:/ disk entries?

And also:

http://lists.xensource.com/archives/html/xen-devel/2009-10/msg00169.html

Is Xen still prone to crashing when using stubdoms?


I have yet to get stubdom HVM guest running? Could someone please post a working hvm-stubdom example... I have been using /etc/xen/xmexample.hvm-stubdom with no luck.

 
Thanks.
 

-M




From: viaum@xxxxxxxxxxxxxxx
To: lists@xxxxxxxxx
Subject: RE: [Xen-users] Error starting stubdom HVM on Xen-3.4.3-rc4-pre
Date: Wed, 10 Mar 2010 10:57:10 -0500
CC: xen-users@xxxxxxxxxxxxxxxxxxx

> Subject: Re: [Xen-users] Error starting stubdom HVM on Xen-3.4.3-rc4-pre
> From: lists@xxxxxxxxx
> To: viaum@xxxxxxxxxxxxxxx
> Date: Wed, 10 Mar 2010 16:29:35 +0100
> CC: xen-users@xxxxxxxxxxxxxxxxxxx
>
> Hi Mike,
>
> Am Mittwoch, den 10.03.2010, 09:45 -0500 schrieb Mike Viau:
> > Hi there,
> >
> > Last night I was trying to start a HVM domU via stubdom-dm device
> > model. Initially I did not receive any error to stdout when I did so
> > with Xen-3.4.2.
> >
> > My Xen-3.4.2 installation works fine with qemu-dm (or regular HVM
> > guests). The stubdom-dm guest I was trying to create did not really
> > operate as I was unable to connect to the VNC console. The output of
> > xm list showed the DomU was there, but it did not show it was actively
> > running, almost like it was pause at 0:00:00.
> >
> > This morning I installed Xen-3.4.3-rc4-pre with the same domU HVM
> > stubdom configuration and got this message.
> >
> > localhost:~/XENSTORE/router#xm create router
> >
> > Using config file "./router".
> > Error: Device 768 (vbd) could not be connected.
> > File /root/XENSTORE/router/hd.img is loopback-mounted
> > through /dev/loop1,
> > which is mounted in a guest domain,
> > and so cannot be mounted now.
>
> So, your domU has this disk mounted or used as loop-device, which
>
> losetup -a
>
> should show. The Disk can be detached through
>
> losetup -d
>
> This should help getting your VM start.
>
> # man losetup
>
> hth,
>
> thomas
>
>

I was unaware or the losetup but unfortunately this did not solve the problem. It appears that the loop device is freed up after the error appears.

localhost:~/XENSTORE/router# xm create router
Using config file "./router".
Error: Device 768 (vbd) could not be connected.
File /root/XENSTORE/router/hd.img is loopback-mounted through /dev/loop1,
which is mounted in a guest domain,
and so cannot be mounted now.
localhost:~/XENSTORE/router# losetup -a
localhost:~/XENSTORE/router#


 
> >
> > AND
> >
> > localhost:/var/log/xen# cat qemu-dm-router-dm.log
> >
> > domid: 2
> > Warning: vlan 0 is not connected to host network
> > Watching /local/domain/0/device-model/2/logdirty/next-active
> > Watching /local/domain/0/device-model/2/command
> > char device redirected to /dev/pts/2
> > /usr/local/src/xen-3.4-testing.hg/tools/ioemu-dir/hw/xen_blktap.c:628:
> > Init blktap pipes
> > bind() failed <<-- This was also the last line in the log when
> > using Xen-3.4.2
> >
> >
> > AND
> >
> >
> > localhost:~/XENSTORE/router# xm info
> > host : localhost.site
> > release : 2.6.33-xen
> > version : #2 SMP Sat Mar 6 15:26:37 EST 2010
> > machine : x86_64
> > nr_cpus : 8
> > nr_nodes : 1
> > cores_per_socket : 4
> > threads_per_core : 2
> > cpu_mhz : 2800
> > hw_caps :
> > bfebfbff:28100800:00000000:00000340:0098e3fd:00000000:00000001:00000000
> > virt_caps : hvm
> > total_memory : 8182
> > free_memory : 7061
> > node_to_cpu : node0:0-7
> > node_to_memory : node0:7061
> > xen_major : 3
> > xen_minor : 4
> > xen_extra : .3-rc4-pre
> > xen_caps : xen-3.0-x86_64 xen-3.0-x86_32p hvm-3.0-x86_32
> > hvm-3.0-x86_32p hvm-3.0-x86_64
> > xen_scheduler : credit
> > xen_pagesize : 4096
> > platform_params : virt_start=0xffff800000000000
> > xen_changeset : Tue Mar 09 18:15:52 2010 +0000
> > 19924:6e72f8c3c017
> > cc_compiler : gcc version 4.3.2 (Debian 4.3.2-1.1)
> > cc_compile_by : root
> > cc_compile_domain : site
> > cc_compile_date : Wed Mar 10 08:46:50 EST 2010
> > xend_config_format : 4
> >
> >
> > AND the configuration file
> >
> >
> > kernel = "/usr/lib/xen/boot/hvmloader"
> > builder='hvm'
> > memory = 1024
> > name = "router"
> > vcpus=1
> > pae=1
> > acpi=1
> > apic=1
> > vif = [ 'type=ioemu, mac=00:16:3e:11:11:01, bridge=br100',
> > 'type=ioemu, mac=00:16:3e:11:11:02, bridge=br101',
> > 'type=ioemu, mac=00:16:3e:11:11:03, bridge=br200',
> > 'type=ioemu, mac=00:16:3e:11:11:04, bridge=br300',
> > 'type=ioemu, mac=00:16:3e:11:11:05, bridge=br400' ]
> > disk = [ 'file:/root/XENSTORE/router/hd.img,hda,w',
> >
> > 'file:/root/XENSTORE/router/pfSense-1.2.3-RELEASE-LiveCD-Installer.iso,
> > hdc:cdrom,r' ]
> > device_model = '/usr/lib/xen/bin/stubdom-dm'
> > boot="cda"
> > sdl=0
> > opengl=1
> > vnc=1
> > vnclisten="10.254.2.81"
> > vncdisplay=0
> > vncunused=0
> > vncpasswd='router'
> > stdvga=0
> >
> >
> > LASTLY the router-dm file
> >
> > #This file is autogenerated, edit router instead!
> > kernel = '/usr/lib/xen/boot/ioemu-stubdom.gz'
> > vfb = ['sdl=0, opengl=1, xauthority=/root/.Xauthority, vnc=1,
> > vncdisplay=0, vnclisten=10.254.2.81, vncunused=0, vncpasswd=router']
> > disk =
> > [ 'file:/root/XENSTORE/router/hd.img,hda:disk,w','file:/root/XENSTORE/router/pfSense-1.2.3-RELEASE-LiveCD-Installer.iso,hdc:cdrom,r' ]
> > vif =
> > [ 'mac=00:16:3e:11:11:01,bridge=br100','mac=00:16:3e:11:11:02,bridge=br101','mac=00:16:3e:11:11:03,bridge=br200','mac=00:16:3e:11:11:04,bridge=br300','mac=00:16:3e:11:11:05,bridge=br400' ]
> >
> >
> >
> > Can someone suggest what I might have done wrong, thanks!
> >
> >
> > -M


Take your contacts everywhere. Try Messenger for mobile
_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users

 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.