[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-API] The vdi is not available



[root@nj-xen-01 ~]# xe pbd-list sr-uuid=9f9aa794-86c0-9c36-a99d-1e5fdc14a206
uuid ( RO)                  : c53d12f6-c3a6-0ae2-75fb-c67c761b2716
             host-uuid ( RO): b8ca0c69-6023-48c5-9b61-bd5871093f4e
               sr-uuid ( RO): 9f9aa794-86c0-9c36-a99d-1e5fdc14a206
         device-config (MRO): serverpath: /xen; options: ; server: 10.254.253.9
    currently-attached ( RO): true


uuid ( RO)                  : a0739a97-408b-afed-7ac2-fe76ffec3ee7
             host-uuid ( RO): a464b853-47d7-4756-b9ab-49cb00c5aebb
               sr-uuid ( RO): 9f9aa794-86c0-9c36-a99d-1e5fdc14a206
         device-config (MRO): serverpath: /xen; options: ; server: 10.254.253.9
    currently-attached ( RO): true


uuid ( RO)                  : 6f2c0e7d-fdda-e406-c2e1-d4ef81552b17
             host-uuid ( RO): dab9cd1a-7ca8-4441-a78f-445580d851d2
               sr-uuid ( RO): 9f9aa794-86c0-9c36-a99d-1e5fdc14a206
         device-config (MRO): serverpath: /xen; options: ; server: 10.254.253.9
    currently-attached ( RO): true

[root@nj-xen-01 ~]# xe host-list
uuid ( RO)                : a464b853-47d7-4756-b9ab-49cb00c5aebb
          name-label ( RW): nj-xen-03
    name-description ( RW): Default install of XenServer


uuid ( RO)                : dab9cd1a-7ca8-4441-a78f-445580d851d2
          name-label ( RW): nj-xen-04
    name-description ( RW): Default install of XenServer


uuid ( RO)                : b8ca0c69-6023-48c5-9b61-bd5871093f4e
          name-label ( RW): nj-xen-01
    name-description ( RW): Default install of XenServer



----- Original Message -----
From: "SÃÂbastien RICCIO" <sr@xxxxxxxxxxxxxxx>
To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx>, xen-api@xxxxxxxxxxxxx
Sent: Thursday, July 25, 2013 11:09:21 AM
Subject: Re: [Xen-API] The vdi is not available

Actually it is right to have:

10.254.253.9:/xen/9f9aa794-86c0-9c36-a99d-1e5fdc14a206 instead of 
10.254.253.9:/xen

That is why on this non-working server your file resides in 
/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/9f9aa794-86c0-9c36-a99d-1e5fdc14a206

When you create a NFS SR in XCP and specify for example 
10.254.253.9:/xen as share to use, it will first create a directory on 
the share with the id of the SR (9f9aa794-86c0-9c36-a99d-1e5fdc14a206) 
and then it remounts 10.254.253.9:/xen/9f9aa794-86c0-9c36-a99d-1e5fdc14a206.

What is strange is that if your servers are in a pool they should share 
the same mount path. Are they all in the same pool ?

Can you please post the results of a :

xe pbd-list sr-uuid=9f9aa794-86c0-9c36-a99d-1e5fdc14a206

and a

xe host-list

thanks


On 25.07.2013 16:51, Andres E. Moya wrote:
> the mounts are not the same, but what is odd is that the servers that have it 
> working correctly, actually seem to be mounting incorrectly?
> please see below
>
> the servers that are working correctly have
>
> Filesystem            Size  Used Avail Use% Mounted on
> /dev/sda1             4.0G  2.1G  1.7G  56% /
> none                  373M   20K  373M   1% /dev/shm
> 10.254.253.9:/xen/9f9aa794-86c0-9c36-a99d-1e5fdc14a206
>                         25T  127G   25T   1% 
> /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206
> 10.254.253.9:/iso      25T  127G   25T   1% 
> /var/run/sr-mount/fbfbf5b3-a37a-288a-86aa-d8d168173f98
> //10.254.254.30/share
>                        196G   26G  160G  14% 
> /var/run/sr-mount/fc63fc27-89ca-dbc8-228d-27e3c74779bb
>
> and the one that doesnt work has it in
>
> Filesystem            Size  Used Avail Use% Mounted on
> /dev/sda1             4.0G  2.0G  1.8G  54% /
> none                  373M   24K  373M   1% /dev/shm
> //10.254.254.30/share
>                        196G   26G  160G  14% 
> /var/run/sr-mount/fc63fc27-89ca-dbc8-228d-27e3c74779bb
> 10.254.253.9:/xen      25T  126G   25T   1% 
> /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206
> 10.254.253.9:/iso      25T  126G   25T   1% 
> /var/run/sr-mount/fbfbf5b3-a37a-288a-86aa-d8d168173f98
>
>
> ----- Original Message -----
> From: "SÃÆÃÂbastien RICCIO" <sr@xxxxxxxxxxxxxxx>
> To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx>, xen-api@xxxxxxxxxxxxx
> Sent: Thursday, July 25, 2013 10:38:02 AM
> Subject: Re: [Xen-API] The vdi is not available
>
> Okay I think you got something here.
>
> do a df -h on each server to check the mount path for the SR on them.
>
> Looks like one or more of your servers mounted it wrong.
>
> /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/
> this is not supposed to happen :(
>
>
>
> On 25.07.2013 16:31, Andres E. Moya wrote:
>> I actually just took a look and in the the servers where everything is 
>> working correctly everything is under
>> /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/
>>
>> and on the one that complains that it cant find the file, the file is 
>> actually located in
>> /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/
>>
>> it's as if it is mounting the storage repository within itself.
>>
>>
>> how can i check if thin provisioning is enabled?
>>
>>
>> ----- Original Message -----
>> From: "SÃÆÃâÃâÃÂbastien RICCIO" <sr@xxxxxxxxxxxxxxx>
>> To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx>
>> Cc: "xen-api" <xen-api@xxxxxxxxxxxxx>, "Alberto Castrillo" 
>> <castrillo@xxxxxxxxxx>
>> Sent: Thursday, July 25, 2013 10:21:44 AM
>> Subject: Re: [Xen-API] The vdi is not available
>>
>> According to:
>>
>> [25610] 2013-07-25 09:51:46.036917      ***** generic exception: vdi_attach: 
>> EXCEPTION SR.SROSError, The VDI is not available 
>> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw
>>  not found]
>>
>> [16462] 2013-07-25 10:02:49.485672 ['/usr/sbin/td-util', 'query', 'vhd',
>> '-vpf',
>> '/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd']
>>
>> there is something wrong. It looks it tries to open a .raw file instead
>> of .vhd.
>>
>> Maybe one of your server has been installed selecting the "thin
>> provisionning" feature and the others servers not ?
>>
>> As far as I know thin provisioning uses vhd, non thin provisioning uses raw.
>> So if you have mixed installations that will not work when using a
>> shared storage between them.
>>
>> My guess is that if you create VM on one server it will create a .vhd
>> image, and on the other a .raw image.
>> I can't be 100% certain as I've always used thin provisionning.
>>
>> You maybe can check if you have mixed raw/vhd in
>> /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/
>>
>>
>>
>>
>> On 25.07.2013 16:04, Andres E. Moya wrote:
>>> this was trying to start up the vm
>>>
>>> [25610] 2013-07-25 09:51:45.997895      lock: acquired 
>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>>> [25610] 2013-07-25 09:51:46.035698      Raising exception [46, The VDI is 
>>> not available 
>>> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw
>>>  not found]]
>>> [25610] 2013-07-25 09:51:46.035831      lock: released 
>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>>> [25610] 2013-07-25 09:51:46.036917      ***** generic exception: 
>>> vdi_attach: EXCEPTION SR.SROSError, The VDI is not available 
>>> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw
>>>  not found]
>>>      File "/opt/xensource/sm/SRCommand.py", line 96, in run
>>>        return self._run_locked(sr)
>>>      File "/opt/xensource/sm/SRCommand.py", line 137, in _run_locked
>>>        target = sr.vdi(self.vdi_uuid)
>>>      File "/opt/xensource/sm/NFSSR", line 213, in vdi
>>>        return NFSFileVDI(self, uuid)
>>>      File "/opt/xensource/sm/VDI.py", line 102, in __init__
>>>        self.load(uuid)
>>>      File "/opt/xensource/sm/FileSR.py", line 370, in load
>>>        opterr="%s not found" % self.path)
>>>      File "/opt/xensource/sm/xs_errors.py", line 49, in __init__
>>>        raise SR.SROSError(errorcode, errormessage)
>>>
>>> [25610] 2013-07-25 09:51:46.037204      lock: closed 
>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>>>
>>>
>>> and this is on a migrate(destination)
>>>
>>> [29480] 2013-07-25 09:53:18.859918      lock: acquired 
>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>>> [29480] 2013-07-25 09:53:18.897479      Raising exception [46, The VDI is 
>>> not available 
>>> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw
>>>  not found]]
>>> [29480] 2013-07-25 09:53:18.897609      lock: released 
>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>>> [29480] 2013-07-25 09:53:18.898701      ***** generic exception: 
>>> vdi_attach: EXCEPTION SR.SROSError, The VDI is not available 
>>> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw
>>>  not found]
>>>      File "/opt/xensource/sm/SRCommand.py", line 96, in run
>>>        return self._run_locked(sr)
>>>      File "/opt/xensource/sm/SRCommand.py", line 137, in _run_locked
>>>        target = sr.vdi(self.vdi_uuid)
>>>      File "/opt/xensource/sm/NFSSR", line 213, in vdi
>>>        return NFSFileVDI(self, uuid)
>>>      File "/opt/xensource/sm/VDI.py", line 102, in __init__
>>>        self.load(uuid)
>>>      File "/opt/xensource/sm/FileSR.py", line 370, in load
>>>        opterr="%s not found" % self.path)
>>>      File "/opt/xensource/sm/xs_errors.py", line 49, in __init__
>>>        raise SR.SROSError(errorcode, errormessage)
>>>
>>> [29480] 2013-07-25 09:53:18.898972      lock: closed 
>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>>>
>>> this is on migrate (source)
>>>
>>> [16462] 2013-07-25 10:02:48.800862      blktap2.deactivate
>>> [16462] 2013-07-25 10:02:48.800965      lock: acquired 
>>> /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi
>>> [16462] 2013-07-25 10:02:48.819441      ['/usr/sbin/tap-ctl', 'close', 
>>> '-p', '5578', '-m', '7']
>>> [16462] 2013-07-25 10:02:49.295250       = 0
>>> [16462] 2013-07-25 10:02:49.295467      ['/usr/sbin/tap-ctl', 'detach', 
>>> '-p', '5578', '-m', '7']
>>> [16462] 2013-07-25 10:02:49.299579       = 0
>>> [16462] 2013-07-25 10:02:49.299794      ['/usr/sbin/tap-ctl', 'free', '-m', 
>>> '7']
>>> [16462] 2013-07-25 10:02:49.303645       = 0
>>> [16462] 2013-07-25 10:02:49.303902      tap.deactivate: Shut down 
>>> Tapdisk(vhd:/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd,
>>>  pid=5578, minor=7, state=R)
>>> [16462] 2013-07-25 10:02:49.485672      ['/usr/sbin/td-util', 'query', 
>>> 'vhd', '-vpf', 
>>> '/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd']
>>> [16462] 2013-07-25 10:02:49.510929        pread SUCCESS
>>> [16462] 2013-07-25 10:02:49.537296      Removed host key 
>>> host_OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080 for 
>>> 72ad514a-f1f8-4a34-9907-9c6a3506520b
>>> [16462] 2013-07-25 10:02:49.537451      lock: released 
>>> /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi
>>> [16462] 2013-07-25 10:02:49.537540      lock: closed 
>>> /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi
>>> [16462] 2013-07-25 10:02:49.537641      lock: closed 
>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>>> [16462] 2013-07-25 10:02:49.537862      lock: closed 
>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>>> [16636] 2013-07-25 10:02:50.103352      lock: acquired 
>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>>> [16636] 2013-07-25 10:02:50.117961      ['/usr/sbin/td-util', 'query', 
>>> 'vhd', '-vpf', 
>>> '/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd']
>>> [16636] 2013-07-25 10:02:50.137963        pread SUCCESS
>>> [16636] 2013-07-25 10:02:50.139106      vdi_detach {'sr_uuid': 
>>> '9f9aa794-86c0-9c36-a99d-1e5fdc14a206', 'subtask_of': 
>>> 'DummyRef:|ebe0d00f-b082-77ba-b209-095e71a0c1c7|VDI.detach', 'vdi_ref': 
>>> 'OpaqueRef:31009428-3c98-c005-67ed-ddcc5e432e03', 'vdi_on_boot': 'persist', 
>>> 'args': [], 'vdi_location': '72ad514a-f1f8-4a34-9907-9c6a3506520b', 
>>> 'host_ref': 'OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080', 
>>> 'session_ref': 'OpaqueRef:f4170801-402a-0935-a759-19a46e700a87', 
>>> 'device_config': {'server': '10.254.253.9', 'SRmaster': 'true', 
>>> 'serverpath': '/xen', 'options': ''}, 'command': 'vdi_detach', 
>>> 'vdi_allow_caching': 'false', 'sr_ref': 
>>> 'OpaqueRef:fefba283-7462-1f5a-b4e2-d58169c4b318', 'vdi_uuid': 
>>> '72ad514a-f1f8-4a34-9907-9c6a3506520b'}
>>> [16636] 2013-07-25 10:02:50.139415      lock: closed 
>>> /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi
>>> [16636] 2013-07-25 10:02:50.139520      lock: released 
>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>>> [16636] 2013-07-25 10:02:50.139779      lock: closed 
>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>>> [17886] 2013-07-25 10:03:16.326423      sr_scan {'sr_uuid': 
>>> 'fc63fc27-89ca-dbc8-228d-27e3c74779bb', 'subtask_of': 
>>> 'DummyRef:|2f34582a-2b3b-82be-b1f6-7f374565c8e8|SR.scan', 'args': [], 
>>> 'host_ref': 'OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080', 
>>> 'session_ref': 'OpaqueRef:bfffd224-6edc-1cb4-9145-c0c95cbb063b', 
>>> 'device_config': {'iso_path': '/iso', 'type': 'cifs', 'SRmaster': 'true', 
>>> 'location': '//10.254.254.30/share'}, 'command': 'sr_scan', 'sr_ref': 
>>> 'OpaqueRef:9c7f5cd0-fd88-16e2-2426-6e066a1183ab'}
>>>
>>>
>>> ----- Original Message -----
>>> From: "SÃÆÃâÃâÃââÃÆÃâÅÃâÃÂbastien RICCIO" <sr@xxxxxxxxxxxxxxx>
>>> To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx>, "Alberto Castrillo" 
>>> <castrillo@xxxxxxxxxx>
>>> Cc: "xen-api" <xen-api@xxxxxxxxxxxxx>
>>> Sent: Wednesday, July 24, 2013 10:55:40 PM
>>> Subject: Re: [Xen-API] The vdi is not available
>>>
>>> Hi,
>>>
>>> When this happens, what does /var/log/SMlog says ?
>>>
>>> Can you please tail -f /var/log/SMlog on both source and destination,
>>> try to migrate the VM and paste the results?
>>>
>>> Cheers,
>>> SÃÆÃâÃâÃââÃÆÃâÅÃâÃÂbastien
>>>
>>> On 24.07.2013 23:09, Andres E. Moya wrote:
>>>> I also just tried creating a new storage repository moving the vdi to the 
>>>> new storage repository is successful, i then try to migrate it to server C 
>>>> and still have the same issue
>>>>
>>>> Moya Solutions, Inc.
>>>> amoya@xxxxxxxxxxxxxxxxx
>>>> 0 | 646-918-5238 x 102
>>>> F | 646-390-1806
>>>>
>>>> ----- Original Message -----
>>>> From: "Alberto Castrillo" <castrillo@xxxxxxxxxx>
>>>> To: "xen-api" <xen-api@xxxxxxxxxxxxx>
>>>> Sent: Wednesday, July 24, 2013 4:12:13 PM
>>>> Subject: Re: [Xen-API] The vdi is not available
>>>>
>>>>
>>>>
>>>> We use NFS as shared storage, and have faced some "VDI not available" 
>>>> issues with our VMs. I haven't been able to start a VM with the method of 
>>>> that URL in XCP 1.6 (in 1.1 and 1.5 beta worked). What worked for me:
>>>>
>>>>
>>>> - Detach the VDI from the VM
>>>> - Detach and forget the SR where the VDI is stored
>>>> - Reattach the forgotten SR (create new SR, give the same info that the 
>>>> detached SR, re-use the SR-UUID, ...)
>>>> - Reattach the VDI to the VM
>>>>
>>>>
>>>>
>>>>
>>>> El 24/07/2013, a las 21:10, hook escribiÃÆÃâÃâÃââÃÆÃâÅÃâÃÂ:
>>>>
>>>>
>>>>
>>>> Past weekend (as usual O_o) we have experienced the issue in our XCP 1.6 
>>>> production pool.
>>>> Shared iSCSI storage was shutted down due to misconfigured UPS settings 
>>>> while XCP servers continued to work.
>>>>
>>>>
>>>> When storage was returned to working state and reconnected to pool most VM 
>>>> did not boot with the same message - VDI is not available.
>>>> Googling give me mentioned above method - forgot and reconnect VDI.
>>>> Result was even worser - the whole SR become unusable.
>>>> Storage rescan gazered lot of errors like bad header on LVM and many other.
>>>>
>>>>
>>>> Finally i've disconnect failed SR from pool, connect it back and SR become 
>>>> healthy (it looks so). But anyone VM was not start with disk from this SR 
>>>> and freeze during startup.
>>>> I did not find solution and restored most VMs from backup (long live VMPP!)
>>>>
>>>>
>>>> So, i just wanna say - be highly careful with VDI on shared storage 
>>>> repository in production environment)
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> 2013/7/24 Brian Menges < bmenges@xxxxxxxxxx >
>>>>
>>>>
>>>> Have you tried the following?: 
>>>> http://community.spiceworks.com/how_to/show/14199-xcp-xen-cloud-platform-xenserver-the-vdi-is-not-available
>>>>
>>>> - Brian Menges
>>>> Principal Engineer, DevOps
>>>> GoGrid | ServePath | ColoServe | UpStream Networks
>>>>
>>>>
>>>> -----Original Message-----
>>>> From: xen-api-bounces@xxxxxxxxxxxxx [mailto: xen-api-bounces@xxxxxxxxxxxxx 
>>>> ] On Behalf Of Andres E. Moya
>>>> Sent: Wednesday, July 24, 2013 09:32
>>>> To: xen-api@xxxxxxxxxxxxx
>>>> Subject: [Xen-API] The vdi is not available
>>>>
>>>> Guys need help trouble shooting this issue
>>>>
>>>> I have an xcp 1.6 pool with 3 machines A,B, and C
>>>>
>>>> I can migrate from A to B and B to A
>>>>
>>>> WE cannot migrate from A or B to C, we also cannot shutdown a vm and start 
>>>> it up on C, when we do that we get the message The vdi is not available.
>>>>
>>>> We have tried removing machine C from the pool and re joining and still 
>>>> have the issue.
>>>>
>>>> when we first add host C to the pool it cannot load the nfs storage 
>>>> repository because we need to create a management interface from a bonded 
>>>> vlan that gets created after joining the pool. After we create the 
>>>> interface and run a re plug on the storage repository it says its 
>>>> connected / re plugged.
>>>>
>>>> Thanks for any help in advance
>>>>
>>>>
>>>> _______________________________________________
>>>> Xen-api mailing list
>>>> Xen-api@xxxxxxxxxxxxx
>>>> http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api
>>>>
>>>> ________________________________
>>>>
>>>> The information contained in this message, and any attachments, may 
>>>> contain confidential and legally privileged material. It is solely for the 
>>>> use of the person or entity to which it is addressed. Any review, 
>>>> retransmission, dissemination, or action taken in reliance upon this 
>>>> information by persons or entities other than the intended recipient is 
>>>> prohibited. If you receive this in error, please contact the sender and 
>>>> delete the material from any computer.
>>>>
>>>> _______________________________________________
>>>> Xen-api mailing list
>>>> Xen-api@xxxxxxxxxxxxx
>>>> http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api
>>>>
>>>>
>>>>
>>>>
>




_______________________________________________
Xen-api mailing list
Xen-api@xxxxxxxxxxxxx
http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api

 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.