[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-users] guest boots sucessfully only one time then lvm complains no uuid found
Usually, when there is mention of LVM and Xen, it refers to domU:s stored in LV:s on the dom0. I see no reference to LVM in your domU config and conclude that you have created a VG *in your domU*, which you want to boot from? Can't help you there, but you may want to check The Linux Documentation Project (TLDP). They have a howto on LVM which I think includes how to boot from LVM. If I remember correctly, there are a few steps you have to take to make it work. Cheers, /Chris > christoffer@xxxxxxxxx wrote: >> It should be possible - I do precisely that: LVM on software raid1. I >> think more detail is needed. What are your device names and what do you >> export to your domU? >> >> Cheers, >> /Chris >> >>> I set up a guest ( under RHEL5 dom0) that has a root volume group >>> extends onto /dev/md which is a raid 1. >>> >>> The first time after setting it up, that guest boots just fine but the >>> next reboot will panic with the root lvm cannot the uuid of the >>> /dev/md0. >>> >>> Because I have a backup of that image, so I can restore and restart it >>> again but then even without doing anything, the next reboot will panic. >>> > > I do not have problem when my main VG ( the VG that has the LV > containing the / partition ) is built directly on the /dev/md0. I have > another guest with that configuration and it runs just fine. > > This guest with problem has the main VG built just on one drive and it > boots normally. Then I build the /dev/md0 on two other drives, and > *extend* the main VG into this /dev/md0 's PV : just extend the VG, not > yet resize the LV inside it. I did run mdadm --detail --scan > > /etc/mdadm.conf. > > The first time I boot, the LVM can see /dev/md0 and its PV, so the VG is > found ok. The 2nd time I boot, it complains no UUID ( for /dev/md0 )is > found. All of the below configuation output is from the first boot: > > Below is my guest config: > > [root@xen2 ~]# more /etc/xen/g1 > name = "g1" > uuid = "2731f3ac-3bde-7105-46b9-794004cb3894" > maxmem = 1024 > memory = 1024 > vcpus = 1 > bootloader = "/usr/bin/pygrub" > on_poweroff = "destroy" > on_reboot = "restart" > on_crash = "restart" > vfb = [ ] > disk = [ "tap:aio:/data/xen/images/g1.img,xvda,w", > "tap:aio:/data/xen/images/g1_ > b.img,xvdb,w","tap:aio:/data/xen/images/g1_c.img,xvdc,w" ] > vif = [ "mac=00:16:3e:6e:35:77,bridge=xenbr0" ] > > [root@g1 ~]# vgdisplay -v > Finding all volume groups > Finding volume group "g1volgroup00" > --- Volume group --- > VG Name g1volgroup00 > System ID > Format lvm2 > Metadata Areas 2 > Metadata Sequence No 25 > VG Access read/write > VG Status resizable > MAX LV 0 > Cur LV 2 > Open LV 2 > Max PV 0 > Cur PV 2 > Act PV 2 > VG Size 10.04 GB > PE Size 8.00 MB > Total PE 1285 > Alloc PE / Size 1266 / 9.89 GB > Free PE / Size 19 / 152.00 MB > VG UUID EFxLNW-31H9-lyLD-oprg-UvLB-lVNF-3Qv6zG > > --- Logical volume --- > LV Name /dev/g1volgroup00/lvroot > VG Name g1volgroup00 > LV UUID y8C8j8-Ar8C-py5x-OL2E-K5zS-yqft-4vXDWU > LV Write Access read/write > LV Status available > # open 1 > LV Size 8.89 GB > Current LE 1138 > Segments 1 > Allocation inherit > Read ahead sectors auto > - currently set to 256 > Block device 253:0 > > --- Logical volume --- > LV Name /dev/g1volgroup00/lvswap > VG Name g1volgroup00 > LV UUID UTPDI8-LAGf-Tvp6-lLfD-vi6p-0RUv-S0rdKo > LV Write Access read/write > LV Status available > # open 1 > LV Size 1.00 GB > Current LE 128 > Segments 1 > Allocation inherit > Read ahead sectors auto > - currently set to 256 > Block device 253:1 > > --- Physical volumes --- > PV Name /dev/xvda2 > PV UUID H0owWs-k8Li-w0Mb-3CgN-5gsM-haJy-BYYkrh > PV Status allocatable > Total PE / Free PE 1266 / 0 > > PV Name /dev/md0 > PV UUID 4m2GFx-QDv0-Gb8q-qxAG-QMIc-If9i-b1ge12 > PV Status allocatable > Total PE / Free PE 19 / 19 > > [root@g1 ~]# pvscan > PV /dev/xvda2 VG g1volgroup00 lvm2 [9.89 GB / 0 free] > PV /dev/md0 VG g1volgroup00 lvm2 [152.00 MB / 152.00 MB free] > Total: 2 [10.04 GB] / in use: 2 [10.04 GB] / in no VG: 0 [0 ] > > [root@g1 ~]# pvdisplay > --- Physical volume --- > PV Name /dev/xvda2 > VG Name g1volgroup00 > PV Size 9.90 GB / not usable 6.76 MB > Allocatable yes (but full) > PE Size (KByte) 8192 > Total PE 1266 > Free PE 0 > Allocated PE 1266 > PV UUID H0owWs-k8Li-w0Mb-3CgN-5gsM-haJy-BYYkrh > > --- Physical volume --- > PV Name /dev/md0 > VG Name g1volgroup00 > PV Size 156.19 MB / not usable 4.19 MB > Allocatable yes > PE Size (KByte) 8192 > Total PE 19 > Free PE 19 > Allocated PE 0 > PV UUID 4m2GFx-QDv0-Gb8q-qxAG-QMIc-If9i-b1ge12 > > [root@g1 ~]# mdadm --detail /dev/md0 > /dev/md0: > Version : 00.90.03 > Creation Time : Sat Jan 17 11:11:37 2009 > Raid Level : raid1 > Array Size : 159936 (156.21 MiB 163.77 MB) > Used Dev Size : 159936 (156.21 MiB 163.77 MB) > Raid Devices : 2 > Total Devices : 2 > Preferred Minor : 0 > Persistence : Superblock is persistent > > Intent Bitmap : Internal > > Update Time : Sun Jan 18 13:10:00 2009 > State : active > Active Devices : 2 > Working Devices : 2 > Failed Devices : 0 > Spare Devices : 0 > > UUID : f38cb367:f475b06c:d2f4e2c1:3c452b16 > Events : 0.6 > > Number Major Minor RaidDevice State > 0 202 16 0 active sync /dev/xvdb > 1 202 32 1 active sync /dev/xvdc > [root@g1 ~]# > > > > > Thanks, > > Vu > > _______________________________________________ > Xen-users mailing list > Xen-users@xxxxxxxxxxxxxxxxxxx > http://lists.xensource.com/xen-users > _______________________________________________ Xen-users mailing list Xen-users@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-users
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |