[Gluster-users] Two VMS as arbiter...

Gilberto Nunes gilberto.nunes32 at gmail.com
Wed Aug 5 20:41:57 UTC 2020


I'm in trouble here.
When I shutdown the pve01 server, the shared folder over glusterfs is EMPTY!
It's supposed to be a qcow2 file inside it.
The content is show right, just after I power on pve01 backup...

Some advice?


Thanks

---
Gilberto Nunes Ferreira

(47) 3025-5907
(47) 99676-7530 - Whatsapp / Telegram

Skype: gilberto.nunes36





Em qua., 5 de ago. de 2020 às 11:07, Gilberto Nunes <
gilberto.nunes32 at gmail.com> escreveu:

> Well...
> I do the follow:
>
> gluster vol create VMS replica 3 arbiter 1 pve01:/DATA/brick1
> pve02:/DATA/brick1.5 pve01:/DATA/arbiter1.5 pve02:/DATA/brick2 pv
> e01:/DATA/brick2.5 pve02:/DATA/arbiter2.5 force
>
> And now I have:
> gluster vol info
>
> Volume Name: VMS
> Type: Distributed-Replicate
> Volume ID: 1bd712f5-ccb9-4322-8275-abe363d1ffdd
> Status: Started
> Snapshot Count: 0
> Number of Bricks: 2 x (2 + 1) = 6
> Transport-type: tcp
> Bricks:
> Brick1: pve01:/DATA/brick1
> Brick2: pve02:/DATA/brick1.5
> Brick3: pve01:/DATA/arbiter1.5 (arbiter)
> Brick4: pve02:/DATA/brick2
> Brick5: pve01:/DATA/brick2.5
> Brick6: pve02:/DATA/arbiter2.5 (arbiter)
> Options Reconfigured:
> cluster.quorum-count: 1
> cluster.quorum-reads: false
> cluster.self-heal-daemon: enable
> cluster.heal-timeout: 10
> storage.fips-mode-rchecksum: on
> transport.address-family: inet
> nfs.disable: on
> performance.client-io-threads: off
>
> This values I have put it myself, in order to see if could improve the
> time to make the volume available, when pve01 goes down with ifupdown
> cluster.quorum-count: 1
> cluster.quorum-reads: false
> cluster.self-heal-daemon: enable
> cluster.heal-timeout: 10
>
> Nevertheless, it took more than 1 minutes to the volume VMS available in
> the other host (pve02).
> Is there any trick to reduce this time ?
>
> Thanks
>
> ---
> Gilberto Nunes Ferreira
>
>
>
>
>
>
> Em qua., 5 de ago. de 2020 às 08:57, Gilberto Nunes <
> gilberto.nunes32 at gmail.com> escreveu:
>
>> hum I see... like this:
>> [image: image.png]
>> ---
>> Gilberto Nunes Ferreira
>>
>> (47) 3025-5907
>> (47) 99676-7530 - Whatsapp / Telegram
>>
>> Skype: gilberto.nunes36
>>
>>
>>
>>
>>
>> Em qua., 5 de ago. de 2020 às 02:14, Computerisms Corporation <
>> bob at computerisms.ca> escreveu:
>>
>>> check the example of the chained configuration on this page:
>>>
>>>
>>> https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.3/html/administration_guide/creating_arbitrated_replicated_volumes
>>>
>>> and apply it to two servers...
>>>
>>> On 2020-08-04 8:25 p.m., Gilberto Nunes wrote:
>>> > Hi Bob!
>>> >
>>> > Could you, please, send me more detail about this configuration?
>>> > I will appreciate that!
>>> >
>>> > Thank you
>>> > ---
>>> > Gilberto Nunes Ferreira
>>> >
>>> > (47) 3025-5907
>>> > **
>>> > (47) 99676-7530 - Whatsapp / Telegram
>>> >
>>> > Skype: gilberto.nunes36
>>> >
>>> >
>>> >
>>> >
>>> >
>>> > Em ter., 4 de ago. de 2020 às 23:47, Computerisms Corporation
>>> > <bob at computerisms.ca <mailto:bob at computerisms.ca>> escreveu:
>>> >
>>> >     Hi Gilberto,
>>> >
>>> >     My understanding is there can only be one arbiter per replicated
>>> >     set.  I
>>> >     don't have a lot of practice with gluster, so this could be bad
>>> advice,
>>> >     but the way I dealt with it on my two servers was to use 6 bricks
>>> as
>>> >     distributed-replicated (this is also relatively easy to migrate to
>>> 3
>>> >     servers if that happens for you in the future):
>>> >
>>> >     Server1     Server2
>>> >     brick1      brick1.5
>>> >     arbiter1.5  brick2
>>> >     brick2.5    arbiter2.5
>>> >
>>> >     On 2020-08-04 7:00 p.m., Gilberto Nunes wrote:
>>> >      > Hi there.
>>> >      > I have two physical servers deployed as replica 2 and,
>>> obviously,
>>> >     I got
>>> >      > a split-brain.
>>> >      > So I am thinking in use two virtual machines,each one in
>>> physical
>>> >      > servers....
>>> >      > Then this two VMS act as a artiber of gluster set....
>>> >      >
>>> >      > Is this doable?
>>> >      >
>>> >      > Thanks
>>> >      >
>>> >      > ________
>>> >      >
>>> >      >
>>> >      >
>>> >      > Community Meeting Calendar:
>>> >      >
>>> >      > Schedule -
>>> >      > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
>>> >      > Bridge: https://bluejeans.com/441850968
>>> >      >
>>> >      > Gluster-users mailing list
>>> >      > Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
>>> >      > https://lists.gluster.org/mailman/listinfo/gluster-users
>>> >      >
>>> >     ________
>>> >
>>> >
>>> >
>>> >     Community Meeting Calendar:
>>> >
>>> >     Schedule -
>>> >     Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
>>> >     Bridge: https://bluejeans.com/441850968
>>> >
>>> >     Gluster-users mailing list
>>> >     Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
>>> >     https://lists.gluster.org/mailman/listinfo/gluster-users
>>> >
>>>
>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20200805/746ac4bd/attachment.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: image.png
Type: image/png
Size: 54749 bytes
Desc: not available
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20200805/746ac4bd/attachment.png>


More information about the Gluster-users mailing list