[Gluster-users] how many hosts could be down in a 12x(4+2) distributed dispersed volume?

Mauro Tridici mauro.tridici at cmcc.it
Wed Sep 20 15:51:30 UTC 2017


Dear Serkan,

thank you very much for your support and explanation.
I really appreciated the information you provided.

Regards,
Mauro

> Il giorno 20 set 2017, alle ore 08:26, Serkan Çoban <cobanserkan at gmail.com> ha scritto:
> 
> If you add bricks to existing volume one host could be down in each
> three host group, If you recreate the volume with one brick on each
> host, then two random hosts can be tolerated.
> Assume s1,s2,s3 are current servers and you add s4,s5,s6 and extend
> volume. If any two servers in each group goes down you loose data. If
> you chose random two host the probability you loose data will be %20
> in this case.
> If you recreate volume with s1,s2,s3,s4,s5,s6 with one brick on each
> host any random two servers can go down. If you chose random two host
> the probability you loose data will be %0 in this case.
> 
> On Mon, Sep 18, 2017 at 10:39 PM, Mauro Tridici <mauro.tridici at cmcc.it> wrote:
>> Dear All,
>> 
>> I just implemented a (6x(4+2)) DISTRIBUTED DISPERSED gluster (v.3.10) volume based on the following hardware:
>> 
>> - 3 gluster servers (each server with 2 CPU 10 cores, 64GB RAM, 12 hard disk SAS 12Gb/s, 10GbE storage network)
>> 
>> Now, we need to add 3 new servers with the same hardware configuration respecting the current volume topology.
>> If I'm right, we will obtain a DITRIBUTED DISPERSED gluster volume with 12 subvolumes, each volume will contain (4+2) bricks, that is a [12x(4+2)] volume.
>> 
>> My question is: in the current volume configuration, only 2 bricks per subvolume or one host could be down without losing data. What it will happen in the next configuration? How many hosts could be down without losing data?
>> 
>> Thank you very much.
>> Mauro Tridici
>> 
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://lists.gluster.org/mailman/listinfo/gluster-users



More information about the Gluster-users mailing list