[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

RE: [Xen-users] drbd8 - lvm - xen


  • To: "Sebastian Igerl" <sig@xxxxxxxxx>,<xen-users@xxxxxxxxxxxxxxxxxxx>
  • From: "Robert Dunkley" <Robert@xxxxxxxxx>
  • Date: Wed, 10 Sep 2008 15:27:59 +0100
  • Cc:
  • Delivery-date: Wed, 10 Sep 2008 07:30:04 -0700
  • List-id: Xen user discussion <xen-users.lists.xensource.com>
  • Thread-index: AckTTKx7a0G8CBWbRnm4fgPKKRf+iAAA3k4Q
  • Thread-topic: [Xen-users] drbd8 - lvm - xen

Hi Sebastian,

I think you should split the volume into LVM partitions and then run a
DRBD device for each partition. I run this config, RaidArray -> LVM ->
DRBD8 -> Xen and it works well. You need the independence of one drbd
device per VM otherwise things are difficult. You may also want to ask
xen-users -at- antibozo.net for their DRBD script, on my Centos Xen 3.30
setup it works much better, it also avoids the need for IMHO dangerous
dual primary config.

I synch multiple arrays with multiple DRBD devices at once. DRBD equally
shares the array write bandwidth between DRBD devices. My Raid1 arrays
manage 80Mb/second each write speed, 2 DRBD devices synching at once on
each of the 3 arrays yields 40mb.sec per DRBD device which totals a
perfect 240Mb/sec synch rate (I used Infiniband to link the two DRBD
servers together so am not bound by the bandwidth constraints of GB
Ethernet). 

Rob

-----Original Message-----
From: xen-users-bounces@xxxxxxxxxxxxxxxxxxx
[mailto:xen-users-bounces@xxxxxxxxxxxxxxxxxxx] On Behalf Of Sebastian
Igerl
Sent: 10 September 2008 14:54
To: xen-users@xxxxxxxxxxxxxxxxxxx
Subject: [Xen-users] drbd8 - lvm - xen

Hi,

What I'm trying to to is to set up 1 drbd active/active device then
split this with lvm creating about 20 volumes and then run xen on top of
that lvm volums !

I tried for days now and it isn't working... anyhow, do you see any
problems why this shouldn't work..

Should I create 20 drbd devices and use the drbd helper script for xen ?
isn't using so many drbd syncs a lot of overhead ? never tried it...


xen4 conf # xm create test1.cfg 
Using config file "./test1.cfg".
Error: Device 768 (vbd) could not be connected.
Device /dev/xen3and4/hd23 is mounted in the privileged domain,


This is what i'm getting right now.. 
/dev/xen3and4/hd23 isn't mounted...



xen4 conf # cat test1.cfg 

kernel = "/usr/lib64/xen/boot/hvmloader"
builder = "hvm"
device_model = "/usr/lib64/xen/bin/qemu-dm"
memory = "1024"
maxmem = 2048
maxmemory = 2048
name = "test1"

dhcp = "off"
vif = ["type=ioemu, mac=00:16:3e:00:00:20, bridge=xenbr0"]
#cpus = ""
#vcpus = 1 
disk = ["phy:/dev/xen3and4/hd23,ioemu:hda,w" ]

serial="pty"
boot="c"

vnc=1
vncviewer=0
keymap="de"
vfb = [ "type=vnc, vncpasswd=ssss, vnclisten=0.0.0.0, vncdisplay=23" ]
sdl=0



since i'm using lvm on top of drbd, xen should'nt be aware of drbd..
using only lvm it workd...

mouting   drbd-lvm2-device works too.... ( only on one site because i
haven't set up any cluster filesystems.. )



Thanks,

Sebastian


_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users

The SAQ Group
Registered Office: 18 Chapel Street, Petersfield, Hampshire. GU32 3DZ
SEMTEC Limited trading as SAQ is Registered in England & Wales
Company Number: 06481952

http://www.saqnet.co.uk AS29219
SAQ Group Delivers high quality, honestly priced communication and I.T. 
services to UK Business.
DSL : Domains : Email : Hosting : CoLo : Servers : Racks : Transit : Backups : 
Managed Networks : Remote Support.

Find us in http://www.thebestof.co.uk/petersfield


_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.