[Gluster-users] how well will this work

Miles Fidelman mfidelman at meetinghouse.net
Thu Dec 27 23:00:18 UTC 2012


Ok... now that's diametrically the opposite response from Dan Cyr's of a 
few minutes ago.

Can you say just a bit more about your configuration - how many nodes, 
do you have storage and processing combined or separated, how do you 
have your drives partitioned, and so forth?

Thanks,

Miles


Joe Julian wrote:
> Trying to return to the actual question, the way I handle those is to 
> mount gluster volumes that host the data for those tasks from within 
> the vm. I've done that successfully since 2.0 with all of those services.
>
> The limitations that others are expressing have as much to do with 
> limitations placed on their own designs as with their hardware. Sure, 
> there are other less stable and/or scalable systems that are faster, 
> but with proper engineering you should be able to build a system that 
> meets those design requirements.
>
> The one piece that wasn't there before but now is in 3.3 is the 
> "locking and performance problems during disk rebuilds" which is now 
> done at a much more granular level and I have successfully self-healed 
> several vm images simultaneously while doing it on all of them without 
> any measurable delays.
>
> Miles Fidelman <mfidelman at meetinghouse.net> wrote:
>
>     Joe Julian wrote:
>
>         It would probably be better to ask this with end-goal
>         questions instead of with a unspecified "critical feature"
>         list and "performance problems".
>
>
>     Ok... I'm running a 2-node cluster that's essentially a mini cloud stack
>     - with storage and processing combined on the same boxes.  I'm running a
>     production VM that hosts a mail server, list server, web server, and
>     database; another production VM providing a backup server for the
>     cluster and for a bunch of desktop machines; and several VMs used for a
>     variety of development and testing purposes. It's all backed by a
>     storage stack consisting of linux raid10 -> lvm -> drbd, and uses
>     pacemaker for high-availability failover of the
>     production VMs.  It all
>     performs reasonably well under moderate load (mail flows, web servers
>     respond, database transactions complete, without notable user-level
>     delays; queues don't back up; cpu and io loads stay within reasonable
>     bounds).
>
>     The goals are to:
>     - add storage and processing capacity by adding two more nodes - each
>     consisting of several CPU cores and 4 disks each
>     - maintain the flexibility to create/delete/migrate/failover virtual
>     machines - across 4 nodes instead of 2
>     - avoid having to play games with pairwise DRBD configurations by moving
>     to a clustered filesystem
>     - in essence, I'm looking to do what Sheepdog purports to do, except in
>     a Xen environment
>
>     Earlier versions of gluster had reported problems with:
>     - supporting databases
>     - supporting VMs
>     - locking and performance problems during disk rebuilds
>     - and... most of the gluster documentation implies that it's
>     preferable
>     to separate storage nodes from processing nodes
>
>     It looks like Gluster 3.2 and 3.3 have addressed some of these issues,
>     and I'm trying to get a general read on whether it's worth putting in
>     the effort of moving forward with some experimentation, or whether this
>     is a non-starter.  Is there anyone out there who's tried to run this
>     kind of mini-cloud with gluster?  What kind of results have you had?
>
>
>
>         On 12/26/2012 08:24 PM, Miles Fidelman wrote:
>
>             Hi Folks, I find myself trying to expand a 2-node
>             high-availability cluster from to a 4-node cluster. I'm
>             running Xen virtualization, and currently using DRBD to
>             mirror data, and pacemaker to failover cleanly. The thing
>             is, I'm trying to add 2 nodes to the cluster, and DRBD
>             doesn't scale. Also, as a function of rackspace limits,
>             and the hardware at hand, I can't separate storage nodes
>             from compute nodes - instead, I have to live with 4 nodes,
>             each with 4 large drives (but also w/ 4 gigE ports per
>             server). The obvious thought is to use Gluster to assemble
>             all the drives into one large storage pool, with
>             replication. But.. last time I looked at this (6 months or
>             so back), it looked like some of the critical features
>             were brand new, and performance seemed to be a problem in
>             the configuration I'm thinking of. Which leads me to my
>             question: Has the situation improved to the point that I
>             can use Gluster this way? Thanks very much, Miles Fidelman
>
>         ------------------------------------------------------------------------
>         Gluster-users mailing list Gluster-users at gluster.org
>         http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
>


-- 
In theory, there is no difference between theory and practice.
In practice, there is.   .... Yogi Berra




More information about the Gluster-users mailing list