[Gluster-users] RAID on GLUSTER node

Pranith Kumar Karampuri pkarampu at redhat.com
Wed Jan 13 04:07:32 UTC 2016



On 01/13/2016 02:21 AM, Pawan Devaiah wrote:
> Thanks for the response Pranith
>
> If we take EC out of the equation and say I go with RAID on the 
> physical disk, do you think GlusterFS is good for the 2 workloads that 
> I mentioned before.
>
> Basically it is going to be a NFS storage for VM and data but with 
> different RAIDs, 10 for VM and 6 for data.
What will be the kind of volume you will be using with these disks?

Pranith
> Thanks
> Dev
>
> On Tue, Jan 12, 2016 at 9:46 PM, Pranith Kumar Karampuri 
> <pkarampu at redhat.com <mailto:pkarampu at redhat.com>> wrote:
>
>
>
>     On 01/12/2016 01:26 PM, Pawan Devaiah wrote:
>>     Thanks for your response Pranith and Mathieu,
>>
>>     Pranith: To answer your question, I am planning to use this
>>     storage for two main workloads.
>>
>>     1. As a shared storage for VMs.
>     EC as it is today is not good for this.
>>     2. As a NFS Storage for files.
>     If the above is for storing archive data. EC is nice here.
>
>     Pranith
>
>>
>>     We are a online backup company so we store few hundred Terra
>>     bytes of data.
>>
>>
>>     Mathieu: I appreciate your concern, however as a system admins
>>     sometimes we get paranoid and try to control everything under the
>>     Sun.
>>     I know I can only control what I can.
>>
>>     Having said that, No, I have pair of servers to start with so at
>>     the moment I am just evaluating and preparing for proof of
>>     concept, after which I am going to propose to my management, if
>>     they are happy then we will proceed further.
>>
>>     Regards,
>>     Dev
>>
>>     On Tue, Jan 12, 2016 at 7:30 PM, Mathieu Chateau
>>     <mathieu.chateau at lotp.fr <mailto:mathieu.chateau at lotp.fr>> wrote:
>>
>>         Hello,
>>
>>         For any system, 36 disks raise disk failure probability. Do
>>         you plan GlusterFS with only one server?
>>
>>         You should think about failure at each level and be prepared
>>         for it:
>>
>>           * Motherboard failure (full server down)
>>           * Disks failure
>>           * Network cable failure
>>           * File system corruption (time needed for fsck)
>>           * File/folder removed by mistake (backup)
>>
>>         Using or not raid depend on your answer on these questions
>>         and performance needed.
>>         It also depend how "good" is raid controller in your server,
>>         like if it has battery and 1GB of cache.
>>
>>         When many disks are bought at same time (1 order, serial
>>         number close to each other), they may fail in near time to
>>         each other (if something bad happened in manufactory).
>>         I already saw like 3 disks failing in few days.
>>
>>         just my 2 cents,
>>
>>
>>
>>         Cordialement,
>>         Mathieu CHATEAU
>>         http://www.lotp.fr
>>
>>         2016-01-12 4:36 GMT+01:00 Pranith Kumar Karampuri
>>         <pkarampu at redhat.com <mailto:pkarampu at redhat.com>>:
>>
>>
>>
>>             On 01/12/2016 04:34 AM, Pawan Devaiah wrote:
>>>             Hi All,
>>>
>>>             We have a fairly powerful server sitting at office with
>>>             128 Gig RAM and 36 X 4 TB drives. I am planning to
>>>             utilize this server as a backend storage with GlusterFS
>>>             on it.
>>>             I have been doing lot of reading on Glusterfs, but I do
>>>             not see any definite recommendation on having RAID on
>>>             GLUSTER nodes.
>>>             Is it recommended to have RAID on GLUSTER nodes
>>>             specially for the bricks?
>>>             If Yes, is it not contrary to the latest Erasure code
>>>             implemented in Gluster or is it still not ready for
>>>             production environment?
>>>             I am happy to implement RAID but my two main concern are
>>>             1. I want to make most of the disk space available.
>>>             2. I am also concerned about the rebuild time after disk
>>>             failure on the RAID.
>>             What is the workload you have?
>>
>>             We found in our testing that random read/write workload
>>             with Erasure coded volumes is not as good as we get with
>>             replication. There are enhancements in progress at the
>>             moment to address these things which we are yet to merge
>>             and re-test.
>>
>>             Pranith
>>>
>>>             Thanks
>>>             Dev
>>>
>>>
>>>
>>>
>>>             _______________________________________________
>>>             Gluster-users mailing list
>>>             Gluster-users at gluster.org  <mailto:Gluster-users at gluster.org>
>>>             http://www.gluster.org/mailman/listinfo/gluster-users
>>
>>
>>             _______________________________________________
>>             Gluster-users mailing list
>>             Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
>>             http://www.gluster.org/mailman/listinfo/gluster-users
>>
>>
>>
>
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160113/572bc4c4/attachment.html>


More information about the Gluster-users mailing list