[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-users] Block level domU backup


  • To: xen-users@xxxxxxxxxxxxxxxxxxx
  • From: "Fajar A. Nugraha" <fajar@xxxxxxxxx>
  • Date: Wed, 29 Oct 2008 13:49:53 +0700
  • Delivery-date: Tue, 28 Oct 2008 23:50:44 -0700
  • List-id: Xen user discussion <xen-users.lists.xensource.com>

Luke S Crawford wrote:
> "Fajar A. Nugraha" <fajar@xxxxxxxxx> writes:
>   
>> With that in mind, it should be easier to simply use snapshot without
>> the need of xm save/restore. It will save some domU "downtime" (the time
>> needed to save and restore domU).
>>     
>
> the idea is that 'xm save' saves your ram, including any write-back
> disk cash that has not been flushed.  So if I do a 'xm save' and save the 
> savefile, and take a bit-for-bit copy of the backing device while the 
> domain is still frozen, I should be able to restore the bit-for-bit copy 
> of the backing device at some point in the future, and then 'xm restore' the
> savefile I saved, and end up exactly where I was, with no inconsistancies 
> or corruptions, as all disk writes that had not been flushed to disk are 
> still in ram.  
>
> (I can reduce downtime by only taking a snapshot while the domU is down, then
> doing the bit-for-bit copy off the snapshot.) 
>
>   
Even with snapshot, there's still the time required to "xm save" and "xm
restore". I guess it's more about choice, really.

If I snapshot without xm save-restore, I get a "dirty" filesystem
backup, but services would run as usual. If I do xm save-restore, I get
a "clean" backup, but that also means all services on that domU would be
unavailable for (at least) the duration of xm save-restore. I choose the
first one.

> you want to test 
> restoring your backup to another server.   A backup that isn't tested is 
> no backup at all.  
>
>   

Good point on that.
>> Another thing to consider, when the question "how to backup domU" arised
>> on this list in the past (and it comes up quite often, search the list
>> archive) I'd generally reply "try using zfs snapshot". Which means :
>> - for backup in domU, you either need an opensolaris or zfs-fuse/linux
>> running on domU
>>     
>
> Yeah, that's great if you are using opensolaris in the DomU (or something
> else that supports zfs well)  but from what I understand, the linux zfs-fuse
> stuff is pretty slow.  
>
>   

Not really. zfs-fuse is slow if you let it handle raid (about half
lvm/md throughput). Since I mostly need the snapshot feature, I use
zfs-fuse on top of lvm. Performance-wise, depending on how you use it,
it's similar to ext3. Best case scenario, if you :
- disable checksum
- enable compression
- set application block size to match zfs block size (or vice versa)
you can actually get better read i/o performance (with cpu usage tradeoff).

>> - for backup in dom0, you need opensolaris dom0 (using zfs volume),
>> whatever the OS/fs running on domU.
>>     
>
> This does sound interesting, though I haven't tried it.  
>
>   

I'm using opensolaris snv_98 dom0, and it works fine for the most part.
There are differences from linux dom0 though, like the fact that (for
now) you can't bridge a vlan interface to dom0 (you can only bridge
physical interfaces).

>> Another alternative is to have an opensolaris server exporting zfs
>> volumes via iscsi, have dom0/domU import it, and do all backups on the
>> storage server.
>>     
>
> this is also interesting.  Software ISCSI is obvously going to be slower
> than native disk, but how much slower?  it is an interesting question.
>
>   

This thread might give some info
http://mail.opensolaris.org/pipermail/zfs-discuss/2008-October/051749.html

> Right now, all my storage is local to the Dom0, and many hosts have 
> excess disk.  I've been thinking about exporting the excess disk via iscsi
> or NFS so that customers who want to buy more storage can do so without me
> worrying about balancing the local storage on various Dom0 hosts.  
>
> and it would be easy enough to do that from within a OpenSolaris DomU.
>
> the big question in my mind is 'how much of the zfs benifits do I retain
> if I export over iscsi and format the block device ext3?'
>
>   
You can get :
- zfs checksum and raidz, which would ensure data integrity (up to the
exported block-level anyway)
- transparent compression. Having compressed ext3 volumes is nice for
certain usage.
- snapshot and clone. Similar to qcow, but with block-device benefits.


> If the dd is causing performance problems, use ionice, and set it to the
> 'idle' class.  your backup will be really, really slow but will not
> interfere with other I/O.  (I have tested that, and it does seem to work 
> as advertised.)  Now, you're right about lvm snapshots being slow, so the 
> domain being backed up is going to be slow until the backup finishes and the 
> snapshot is cleared,  but ionice makes a huge difference for the other 
> domains on the box.  
>
>   

Good hint on ionice. At least it can isolate the performance penalty.

Regards,

Fajar

Attachment: smime.p7s
Description: S/MIME Cryptographic Signature

_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users

 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.