[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

[Xen-devel] Problem with starting guest domain using root nfs


I've been trying to start a guest domain using root nfs but haven't been successful. I have a linux os mounted exported on (I've made sure that I can access and mount this from a different host) and I edited the fstab under to be the following:           /       nfs     defaults        0 0
#LABEL=/                 /                       ext3    defaults        1 1
none                    /dev/pts                devpts  gid=5,mode=620  0 0
none                    /dev/shm                tmpfs   defaults        0 0
none                    /proc                   proc    defaults        0 0
none                    /sys                    sysfs   defaults        0 0

the xen configuration file on the xen host machine is as follows:

kernel = "/boot/vmlinuz-2.6-xenU"
memory = 192
name = "vServer2"
cpu = -1   # leave to Xen to pick
hostname= "vServer2"
root = "/dev/nfs"
# The nfs server.
nfs_server = 'pplasma2'
# Root directory on the nfs server.
nfs_root   = '/export/linux_os'
extra = "4"

I got some rpc error when trying to start this guest domain:

[root@n1-demo-5 xen]# xm create ./*nfs -c
Using config file "./vServer2.nfs".
Started domain vServer2, console on port 9633
************ REMOTE CONSOLE: CTRL-] TO QUIT ********
Linux version (root@xxxxxxxxxxxxxxxxxxxxxxx) (gcc version 3.4.3 20050227 (Red Hat 3.4.3-22.1)) #1 Sat Aug 6 16:02:40 PDT 2005
BIOS-provided physical RAM map:
Xen: 0000000000000000 - 000000000c000000 (usable)
192MB LOWMEM available.
DMI not present.
IRQ lockup detection disabled
Allocating PCI resources starting at 0c000000 (gap: 0c000000:f4000000)
Built 1 zonelists
Kernel command line:  root=/dev/nfs nfsroot= 4
Initializing CPU#0
PID hash table entries: 1024 (order: 10, 16384 bytes)
Xen reported: 1794.928 MHz processor.
Using tsc for high-res timesource
Dentry cache hash table entries: 32768 (order: 5, 131072 bytes)
Inode-cache hash table entries: 16384 (order: 4, 65536 bytes)
vmalloc area: cc800000-fbff9000, maxmem 34000000
Memory: 192128k/196608k available (1558k kernel code, 4352k reserved, 408k data, 104k init, 0k highmem)
Checking if this processor honours the WP bit even in supervisor mode... Ok.
Mount-cache hash table entries: 512 (order: 0, 4096 bytes)
CPU: L1 I Cache: 64K (64 bytes/line), D cache 64K (64 bytes/line)
CPU: L2 Cache: 1024K (64 bytes/line)
CPU: AMD Opteron(tm) Processor 244 stepping 0a
Enabling fast FPU save and restore... done.
Enabling unmasked SIMD FPU exception support... done.
Checking 'hlt' instruction... disabled
NET: Registered protocol family 16
xen_mem: Initialising balloon driver.
Initializing Cryptographic API
io scheduler noop registered
io scheduler anticipatory registered
io scheduler deadline registered
io scheduler cfq registered
RAMDISK driver initialized: 16 RAM disks of 4096K size 1024 blocksize
Xen virtual console successfully installed as tty
Event-channel device installed.
xen_blk: Initialising virtual block device driver
xen_blk: Could not probe disks (0)
xen_blk: failed to probe
xen_net: Initialising virtual ethernet driver.
NET: Registered protocol family 2
IP: routing cache hash table of 2048 buckets, 16Kbytes
TCP established hash table entries: 8192 (order: 4, 65536 bytes)
TCP bind hash table entries: 8192 (order: 3, 32768 bytes)
TCP: Hash tables configured (established 8192 bind 8192)
NET: Registered protocol family 1
NET: Registered protocol family 17
Looking up port of RPC 100003/2 on
RPC: sendmsg returned error 101
portmap: RPC call returned error 101
Root-NFS: Unable to get nfsd port number from server, using default
Looking up port of RPC 100005/1 on
RPC: sendmsg returned error 101
portmap: RPC call returned error 101
Root-NFS: Unable to get mountd port number from server, using default
RPC: sendmsg returned error 101
mount: RPC call returned error 101

Did I miss anything?


Xen-devel mailing list



Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.