[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

RE: [Xen-users] drbd8 - lvm - xen


  • To: xen-users@xxxxxxxxxxxxxxxxxxx
  • From: Sebastian Igerl <sig@xxxxxxxxx>
  • Date: Thu, 11 Sep 2008 11:45:02 +0200
  • Delivery-date: Thu, 11 Sep 2008 02:45:25 -0700
  • List-id: Xen user discussion <xen-users.lists.xensource.com>

oups.. forgot the text :-)


i did.. and e.g. /dev/sda5 without drbd gives the same error...
but just with 
xen-3.3.0 !

On Thu, 2008-09-11 at 11:08 +0200, Sebastian Igerl wrote:
> On Thu, 2008-09-11 at 09:42 +0100, Robert Dunkley wrote:
> > The DRBD device is in use by Dom0 or another DomU. Try one DRBD device
> > per VM, if you have already done this then make sure the DRBD device is
> > NOT mounted in Dom0.
> > 
> > Rob
> > 
> > -----Original Message-----
> > From: Sebastian Igerl [mailto:sig@xxxxxxxxx] 
> > Sent: 11 September 2008 09:36
> > To: Robert Dunkley
> > Cc: xen-users@xxxxxxxxxxxxxxxxxxx
> > Subject: RE: [Xen-users] drbd8 - lvm - xen
> > 
> > Hi,
> > 
> > Found out that it didn't had anything to to with drbd or lvm... 
> > 
> > I upgraded from 3.2.1  to xen-3.3.0...
> > 
> > No i can't start any domu's... regardless if they are on lvm drbd oder
> > local partiton...
> > 
> > Does anyone has any idea ?
> > 
> > xen4 conf # xm create test1.cfg 
> > > Using config file "./test1.cfg".
> > > Error: Device 768 (vbd) could not be connected.
> > > Device /dev/xen3and4/hd23 is mounted in the privileged domain,
> > 
> > 
> > 
> > xen4 conf # xm create test1.cfg 
> > > Using config file "./test1.cfg".
> > > Error: Device 768 (vbd) could not be connected.
> > > Device /dev/sda5 is mounted in the privileged domain,
> > 
> > 
> > 
> > 
> > another thing is... : is ist possible to make 20 different drbd devices
> > on one host ? drbd0 - drbd20 ? aren't there limits ? 
> > 
> > 
> > On Wed, 2008-09-10 at 15:27 +0100, Robert Dunkley wrote:
> > > Hi Sebastian,
> > > 
> > > I think you should split the volume into LVM partitions and then run a
> > > DRBD device for each partition. I run this config, RaidArray -> LVM ->
> > > DRBD8 -> Xen and it works well. You need the independence of one drbd
> > > device per VM otherwise things are difficult. You may also want to ask
> > > xen-users -at- antibozo.net for their DRBD script, on my Centos Xen
> > 3.30
> > > setup it works much better, it also avoids the need for IMHO dangerous
> > > dual primary config.
> > > 
> > > I synch multiple arrays with multiple DRBD devices at once. DRBD
> > equally
> > > shares the array write bandwidth between DRBD devices. My Raid1 arrays
> > > manage 80Mb/second each write speed, 2 DRBD devices synching at once
> > on
> > > each of the 3 arrays yields 40mb.sec per DRBD device which totals a
> > > perfect 240Mb/sec synch rate (I used Infiniband to link the two DRBD
> > > servers together so am not bound by the bandwidth constraints of GB
> > > Ethernet). 
> > > 
> > > Rob
> > > 
> > > -----Original Message-----
> > > From: xen-users-bounces@xxxxxxxxxxxxxxxxxxx
> > > [mailto:xen-users-bounces@xxxxxxxxxxxxxxxxxxx] On Behalf Of Sebastian
> > > Igerl
> > > Sent: 10 September 2008 14:54
> > > To: xen-users@xxxxxxxxxxxxxxxxxxx
> > > Subject: [Xen-users] drbd8 - lvm - xen
> > > 
> > > Hi,
> > > 
> > > What I'm trying to to is to set up 1 drbd active/active device then
> > > split this with lvm creating about 20 volumes and then run xen on top
> > of
> > > that lvm volums !
> > > 
> > > I tried for days now and it isn't working... anyhow, do you see any
> > > problems why this shouldn't work..
> > > 
> > > Should I create 20 drbd devices and use the drbd helper script for xen
> > ?
> > > isn't using so many drbd syncs a lot of overhead ? never tried it...
> > > 
> > > 
> > > xen4 conf # xm create test1.cfg 
> > > Using config file "./test1.cfg".
> > > Error: Device 768 (vbd) could not be connected.
> > > Device /dev/xen3and4/hd23 is mounted in the privileged domain,
> > > 
> > > 
> > > This is what i'm getting right now.. 
> > > /dev/xen3and4/hd23 isn't mounted...
> > > 
> > > 
> > > 
> > > xen4 conf # cat test1.cfg 
> > > 
> > > kernel = "/usr/lib64/xen/boot/hvmloader"
> > > builder = "hvm"
> > > device_model = "/usr/lib64/xen/bin/qemu-dm"
> > > memory = "1024"
> > > maxmem = 2048
> > > maxmemory = 2048
> > > name = "test1"
> > > 
> > > dhcp = "off"
> > > vif = ["type=ioemu, mac=00:16:3e:00:00:20, bridge=xenbr0"]
> > > #cpus = ""
> > > #vcpus = 1 
> > > disk = ["phy:/dev/xen3and4/hd23,ioemu:hda,w" ]
> > > 
> > > serial="pty"
> > > boot="c"
> > > 
> > > vnc=1
> > > vncviewer=0
> > > keymap="de"
> > > vfb = [ "type=vnc, vncpasswd=ssss, vnclisten=0.0.0.0, vncdisplay=23" ]
> > > sdl=0
> > > 
> > > 
> > > 
> > > since i'm using lvm on top of drbd, xen should'nt be aware of drbd..
> > > using only lvm it workd...
> > > 
> > > mouting   drbd-lvm2-device works too.... ( only on one site because i
> > > haven't set up any cluster filesystems.. )
> > > 
> > > 
> > > 
> > > Thanks,
> > > 
> > > Sebastian
> > > 
> > > 
> > > _______________________________________________
> > > Xen-users mailing list
> > > Xen-users@xxxxxxxxxxxxxxxxxxx
> > > http://lists.xensource.com/xen-users
> > > 
> > > The SAQ Group
> > > Registered Office: 18 Chapel Street, Petersfield, Hampshire. GU32 3DZ
> > > SEMTEC Limited trading as SAQ is Registered in England & Wales
> > > Company Number: 06481952
> > > 
> > > http://www.saqnet.co.uk AS29219
> > > SAQ Group Delivers high quality, honestly priced communication and
> > I.T. services to UK Business.
> > > DSL : Domains : Email : Hosting : CoLo : Servers : Racks : Transit :
> > Backups : Managed Networks : Remote Support.
> > > 
> > > Find us in http://www.thebestof.co.uk/petersfield
> > > 
> > 
> 
> 
> _______________________________________________
> Xen-users mailing list
> Xen-users@xxxxxxxxxxxxxxxxxxx
> http://lists.xensource.com/xen-users


_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.