[Gluster-users] GlusterFS on a two-node setup

Ramon Diaz-Uriarte rdiaz02 at gmail.com
Mon May 21 17:50:42 UTC 2012




On Mon, 21 May 2012 13:47:52 +0000,John Jolet <jjolet at drillinginfo.com> wrote:
> i suspect that an rsync with the proper argument will be fine before starting glusterd on the recovered node….

Thanks.


R.

> On May 21, 2012, at 7:58 AM, Ramon Diaz-Uriarte wrote:

> > 
> > 
> > 
> > On Mon, 21 May 2012 00:46:33 +0000,John Jolet <jjolet at drillinginfo.com> wrote:
> > 
> >> On May 20, 2012, at 4:55 PM, Ramon Diaz-Uriarte wrote:
> > 
> >>> 
> >>> 
> >>> 
> >>> On Sun, 20 May 2012 20:38:02 +0100,Brian Candler <B.Candler at pobox.com> wrote:
> >>>> On Sun, May 20, 2012 at 01:26:51AM +0200, Ramon Diaz-Uriarte wrote:
> >>>>> Questions:
> >>>>> ==========
> >>>>> 
> >>>>> 1. Is using GlusterFS an overkill? (I guess the alternative would be to use
> >>>>> NFS from one of the nodes to the other)
> >>> 
> >>>> In my opinion, the other main option you should be looking at is DRBD
> >>>> (www.drbd.org).  This works at the block level, unlike glusterfs which works
> >>>> at the file level.  Using this you can mirror your disk remotely.
> >>> 
> >>> 
> >>> Brian, thanks for your reply. 
> >>> 
> >>> 
> >>> I might have to look at DRBD more carefully, but I do not think it fits my
> >>> needs: I need both nodes to be working (and thus doing I/O) at the same
> >>> time. These are basically number crunching nodes and data needs to be
> >>> accessible from both nodes (e.g., some jobs will use MPI over the
> >>> CPUs/cores of both nodes ---assuming both nodes are up, of course ;-).
> >>> 
> >>> 
> >>> 
> >>> 
> >>>> If you are doing virtualisation then look at Ganeti: this is an environment
> >>>> which combines LVM plus DRBD and allows you to run VMs on either node and
> >>>> live-migrate them from one to the other.
> >>>> http://docs.ganeti.org/ganeti/current/html/
> >>> 
> >>> I am not doing virtualisation. I should have said that explicitly. 
> >>> 
> >>> 
> >>>> If a node fails, you just restart the VMs on the other node and away you go.
> >>> 
> >>>>> 2. I plan on using a dedicated partition from each node as a brick. Should
> >>>>> I use replicated or distributed volumes?
> >>> 
> >>>> A distributed volume will only increase the size of storage available (e.g. 
> >>>> combining two 600GB drives into one 1.2GB volume - as long as no single file
> >>>> is too large).  If this is all you need, you'd probably be better off buying
> >>>> bigger disks in the first place.
> >>> 
> >>>> A replicated volume allows you to have a copy of every file on both nodes
> >>>> simultaneously, kept in sync in real time, and gives you resilience against
> >>>> one of the nodes failing.
> >>> 
> >>> 
> >>> But from the docs and the mailing list I get the impression that
> >>> replication has severe performance penalties when writing and some
> >>> penalties when reading. And with a two-node setup, it is unclear to me
> >>> that, even with replication, if one node fails, gluster will continue to
> >>> work (i.e., the other node will continue to work). I've not been able to
> >>> find what is the recommended procedure to continue working, with
> >>> replicated volumes, when one of the two nodes fails. So that is why I am
> >>> wondering what would replication really give me in this case.
> >>> 
> >>> 
> >> replicated volumes have a performance penalty on the client.  for
> >> instance, i have a replicated volume, with one replica on each of two
> >> nodes.  I'm front ending this with an ubuntu box running samba for cifs
> >> sharing.  if my windows client sends 100MB to the cifs server, the cifs
> >> server will send 100MB to each node in the replica set.  As for what you
> >> have to do to continue working if a node went down, i have tested this.
> >> Not on purpose, but one of my nodes was accidentally downed.  my client
> >> saw no difference.  however, running 3.2.x, in order to get the client
> >> to use the downed node after it was brought back up, i had to remount
> >> the share on the cifs server.  this is supposedly fixed in 3.3.
> > 
> > OK, great. Thanks for the info. It is clear, then, that several of you
> > report that this will work just fine.
> > 
> > 
> >> It's important to note that self-healing will create files created while
> >> the node was offline, but does not DELETE files deleted while the node
> >> was offline.  not sure what the official line is there, but my use is
> >> archival, so it doesn't matter enough to me to run down (if they'd
> >> delete files, i wouldn't need gluster..)
> > 
> > 
> > That is good to know, but is not something I'd want. Is there any way to
> > get files to be deleted? Maybe rsync'ing or similar before self-healing
> > starts? Or will that lead to chaos?
> > 
> > 
> > 
> > Best,
> > 
> > R.
> > 
> > 
> >>> Best,
> >>> 
> >>> R.
> >>> 
> >>> 
> >>> 
> >>> 
> >>>> Regards,
> >>> 
> >>>> Brian.
> >>> -- 
> >>> Ramon Diaz-Uriarte
> >>> Department of Biochemistry, Lab B-25
> >>> Facultad de Medicina 
> >>> Universidad Autónoma de Madrid 
> >>> Arzobispo Morcillo, 4
> >>> 28029 Madrid
> >>> Spain
> >>> 
> >>> Phone: +34-91-497-2412
> >>> 
> >>> Email: rdiaz02 at gmail.com
> >>>      ramon.diaz at iib.uam.es
> >>> 
> >>> http://ligarto.org/rdiaz
> >>> 
> >>> _______________________________________________
> >>> Gluster-users mailing list
> >>> Gluster-users at gluster.org
> >>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
> > 
> > -- 
> > Ramon Diaz-Uriarte
> > Department of Biochemistry, Lab B-25
> > Facultad de Medicina 
> > Universidad Autónoma de Madrid 
> > Arzobispo Morcillo, 4
> > 28029 Madrid
> > Spain
> > 
> > Phone: +34-91-497-2412
> > 
> > Email: rdiaz02 at gmail.com
> >       ramon.diaz at iib.uam.es
> > 
> > http://ligarto.org/rdiaz
> > 

-- 
Ramon Diaz-Uriarte
Department of Biochemistry, Lab B-25
Facultad de Medicina 
Universidad Autónoma de Madrid 
Arzobispo Morcillo, 4
28029 Madrid
Spain

Phone: +34-91-497-2412

Email: rdiaz02 at gmail.com
       ramon.diaz at iib.uam.es

http://ligarto.org/rdiaz




More information about the Gluster-users mailing list