[Gluster-users] Shared VM disk/image on gluster for redundancy?

Nikolai K. Bochev n.bochev at grandstarco.com
Fri Jul 2 07:39:42 UTC 2010


I'd recommend KVM, as it is being pushed up pretty nice lately. It does have a live migration and if you use libvirt it makes management way easier. 


From: "Emmanuel Noobadmin" <centos.admin at gmail.com> 
To: "Gluster General Discussion List" <gluster-users at gluster.org> 
Sent: Tuesday, June 29, 2010 7:10:50 PM 
Subject: Re: [Gluster-users] Shared VM disk/image on gluster for redundancy? 

Thanks for the help! Now that I'm certain I'm not trying something 
undoable, I'll probably test drive on VMs first before ordering the 
physical machines. But before that, I'll still have to decide between 
ESXi/Xen/KVM. 



On 6/29/10, Jeff Darcy <jdarcy at redhat.com> wrote: 
> On 06/29/2010 10:04 AM, Emmanuel Noobadmin wrote: 
>> So most likely I would run two or more physical machines with VM to 
>> failover to each other to catch situations of a single machine 
>> failure. Along with that a pair of storage server. In the case of a 
>> total failure where both the primary & secondary VM dies physically, 
>> roll in a new machine to load up the VM images still safe on the 
>> gluster data servers. 
>> 
>> So in this case would I be correct that my configuration, assuming a 
>> basic 2 physical VM host server and 2 storage server would probably 
>> look something like 
>> 
>> volume rep0 
>> type cluster/replicate 
>> option read-subvolume vmsrv0vol0 
>> subvolumes vmsrv0vol0 datasrv0vol0 datasrv1vol0 
>> end-volume 
>> 
>> 
>> volume rep1 
>> type cluster/replicate 
>> option read-subvolume vmsrv1vol0 
>> subvolumes vmsrv1vol0 datasrv0vol0 datasrv1vol0 
>> end-volume 
>> 
>> volume my_nufa 
>> type cluster/nufa 
>> option local-volume-name rep0 
>> subvolumes rep0 rep1 
>> end-volume 
>> 
>> Or did I lose my way somewhere? :) 
> 
> That looks reasonable to me, except that the last stanza would only 
> apply on vmsrv0. For vmsrv1, you'd want this instead: 
> 
> volume my_nufa 
> type cluster/nufa 
> option local-volume-name rep1 # this is the only difference 
> subvolumes rep0 rep1 
> end-volume 
> 
> It's a little unfortunate that you can't do this with a single volfile, 
> perhaps with $-variable substitutions or some such, but that's the way 
> it is AFAIK. 
> 
> 
>> Does it make any sense to replicate across all 3 or should I simply 
>> spec the VM servers with tiny drives and put everything on the gluster 
>> storage which I suppose would impact performance severely? 
> 
> That's a pretty murky area. With a fast interconnect it's tempting to 
> say the "storage of record" should be only on the data nodes and the app 
> nodes should only do caching. It would certainly be simpler, though 
> with more than two data nodes you'd have to do essentially the same 
> layering of distribute on top of replicate (nufa wouldn't be 
> particularly useful in that configuration). If you wanted to stick with 
> something more like the above, you'd just need to pair each app node 
> with a data node, so e.g. rep0=vmsrv0vol0+datasrv0vol0 and 
> rep1=vmsrv1vol0+datasrv1vol1. You would probably also want to "cross" 
> the read-subvolume assignments, so for example vol0 would go first to 
> datasrv1vol0 instead of vmsrv1vol0 for rep1. This avoids having the app 
> nodes talk to each other when they could be talking to data nodes instead. 
> _______________________________________________ 
> Gluster-users mailing list 
> Gluster-users at gluster.org 
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users 
> 
_______________________________________________ 
Gluster-users mailing list 
Gluster-users at gluster.org 
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users 


-- 


Nikolai K. Bochev 
System Administrator 





More information about the Gluster-users mailing list