[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-users] PVHVM example for xen 4.3.1
> > Hello, > > > I'm looking for an example configuration for a PVHVM domU that works with > 4.3.1. The one at http://wiki.xen.org/wiki/Xen_Linux_PV_on_HVM_drivers > causes 3 warnings and some errors: > > WARNING: ignoring "kernel" directive for HVM guest. Use > "firmware_override" instead if you really want a non-default firmware > WARNING: ignoring device_model directive. > WARNING: Use "device_model_override" instead if you really want a > non-default device_model > xc: info: VIRTUAL MEMORY ARRANGEMENT: > Loader: 0000000000100000->000000000019ed04 > Modules: 0000000000000000->0000000000000000 > TOTAL: 0000000000000000->000000000f800000 > ENTRY ADDRESS: 0000000000100000 > xc: info: PHYSICAL MEMORY ALLOCATION: > 4KB PAGES: 0x0000000000000200 > 2MB PAGES: 0x000000000000007b > 1GB PAGES: 0x0000000000000000 > libxl: error: libxl_dm.c:1275:device_model_spawn_outcome: domain 2 device > model: spawn failed (rc=-3) > libxl: error: libxl_create.c:1075:domcreate_devmodel_started: device model > did not start: -3 > libxl: error: libxl_dm.c:1306:libxl__destroy_device_model: Device Model > already exited > > If I follow the suggestions in the warning messages and specify > firmware_override and device_model_override like so: > > firmware_override = "/usr/lib64/xen/boot/hvmloader" > device_model_override = '/usr/lib64/xen/bin/qemu-dm' > > > it only removes some of the warnings but not the errors: > > WARNING: device model override given without specific DM version > same errors as above > > > Does anybody have a config that works or any idea what's wrong with my > config? > > One more noteworthy thing: > > # cat /var/log/xen/qemu-dm-auth.log > /usr/lib64/xen/bin/qemu-dm: invalid option -- '-xen-domid' > > > My config: > > #kernel = "hvmloader" > #kernel = "/usr/lib64/xen/boot/hvmloader" > firmware_override = "/usr/lib64/xen/boot/hvmloader" > builder = "hvm" > #device_model = 'qemu-dm' > device_model_override = '/usr/lib64/xen/bin/qemu-dm' > name = "auth" > memory = 256 > vcpus=1 > pae=1 > acpi=1 > apic=1 > vif = [ 'mac=00:16:3e:10:20:30, bridge=virbr0, model=e1000' ] > disk = [ '/mnt/iso/grml96-full_2013.09.iso,raw,xvda:cdrom,r' ] > boot="cd" > #root = "/dev/sda1 ro" > xen_platform_pci=1 > on_poweroff = 'destroy' > on_reboot = 'restart' > on_crash = 'restart' > sdl=0 > vnc=1 > vncpasswd='' > stdvga=0 > serial='pty' > keymap='en' > > > - John > > > _______________________________________________ > Xen-users mailing list > Xen-users@xxxxxxxxxxxxx > http://lists.xen.org/xen-users > My examples are for xl toolstack. ----------------------------------------------- A PV guest that use an external kernel ----------------------------------------------- builder="generic" vcpus=4 memory=4096 name="127.0.0.1" kernel="/home/servers/xen/kernels/vmlinuz-2.6.32.61-INTEL-SMP-grsec" #ramdisk="" extra="root=/dev/xvda1 xencons=tty1 earlyprintk=xen" vif = ["ip=127.0.0.1,mac=xx:xx:xx:xx:xx:xx,bridge=br0,rate=100Mb/s"] disk = [ "raw:/xen/images/127.0.0.1.img,ioemu:xvda,w" ] vfb = [ "type=vnc,vnclisten=127.0.0.1,vncdisplay=14,vncpasswd=password" ] serial="pty" boot = "c" sdl=0 monitor=1 localtime=0 xen_platform_pci=1 stdvga=1 videoram=16 gfx_passthru=0 on_poweroff="destroy" on_reboot="restart" on_crash="restart" ----------------------------------------------- A HVM guest ----------------------------------------------- builder="hvm" vcpus=4 memory=4096 name="127.0.0.1" bootloader="pygrub" vif = ["ip=127.0.0.1,mac=xx:xx:xx:xx:xx:xx,bridge=br0,rate=100Mb/s"] disk = [ "raw:/xen/images/127.0.0.1.img,ioemu:xvda,w" ] vfb = [ "type=vnc,vnclisten=127.0.0.1,vncdisplay=14,vncpasswd=password" ] serial="pty" boot = "c" sdl=0 monitor=1 pae=1 acpi=1 apic=1 localtime=0 xen_platform_pci=1 stdvga=1 videoram=16 gfx_passthru=0 on_poweroff="destroy" on_reboot="restart" on_crash="restart" _______________________________________________ Xen-users mailing list Xen-users@xxxxxxxxxxxxx http://lists.xen.org/xen-users
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |