[Gluster-users] Two VMS as arbiter...

Gilberto Nunes gilberto.nunes32 at gmail.com
Wed Aug 5 22:56:58 UTC 2020


Ok...Thanks a lot Strahil

This gluster volume set VMS cluster.favorite-child-policy size do the trick
to me here!

Cheers
---
Gilberto Nunes Ferreira

(47) 3025-5907
(47) 99676-7530 - Whatsapp / Telegram

Skype: gilberto.nunes36





Em qua., 5 de ago. de 2020 às 18:15, Strahil Nikolov <hunter86_bg at yahoo.com>
escreveu:

> This could happen if you have pending heals. Did you reboot that node
> recently ?
> Did you set automatic unsplit-brain ?
>
> Check for pending heals and files in splitbrain.
>
> If not, you can check
> https://docs.gluster.org/en/latest/Troubleshooting/resolving-splitbrain/
> (look at point 5).
>
> Best Regards,
> Strahil Nikolov
>
> На 5 август 2020 г. 23:41:57 GMT+03:00, Gilberto Nunes <
> gilberto.nunes32 at gmail.com> написа:
> >I'm in trouble here.
> >When I shutdown the pve01 server, the shared folder over glusterfs is
> >EMPTY!
> >It's supposed to be a qcow2 file inside it.
> >The content is show right, just after I power on pve01 backup...
> >
> >Some advice?
> >
> >
> >Thanks
> >
> >---
> >Gilberto Nunes Ferreira
> >
> >(47) 3025-5907
> >(47) 99676-7530 - Whatsapp / Telegram
> >
> >Skype: gilberto.nunes36
> >
> >
> >
> >
> >
> >Em qua., 5 de ago. de 2020 às 11:07, Gilberto Nunes <
> >gilberto.nunes32 at gmail.com> escreveu:
> >
> >> Well...
> >> I do the follow:
> >>
> >> gluster vol create VMS replica 3 arbiter 1 pve01:/DATA/brick1
> >> pve02:/DATA/brick1.5 pve01:/DATA/arbiter1.5 pve02:/DATA/brick2 pv
> >> e01:/DATA/brick2.5 pve02:/DATA/arbiter2.5 force
> >>
> >> And now I have:
> >> gluster vol info
> >>
> >> Volume Name: VMS
> >> Type: Distributed-Replicate
> >> Volume ID: 1bd712f5-ccb9-4322-8275-abe363d1ffdd
> >> Status: Started
> >> Snapshot Count: 0
> >> Number of Bricks: 2 x (2 + 1) = 6
> >> Transport-type: tcp
> >> Bricks:
> >> Brick1: pve01:/DATA/brick1
> >> Brick2: pve02:/DATA/brick1.5
> >> Brick3: pve01:/DATA/arbiter1.5 (arbiter)
> >> Brick4: pve02:/DATA/brick2
> >> Brick5: pve01:/DATA/brick2.5
> >> Brick6: pve02:/DATA/arbiter2.5 (arbiter)
> >> Options Reconfigured:
> >> cluster.quorum-count: 1
> >> cluster.quorum-reads: false
> >> cluster.self-heal-daemon: enable
> >> cluster.heal-timeout: 10
> >> storage.fips-mode-rchecksum: on
> >> transport.address-family: inet
> >> nfs.disable: on
> >> performance.client-io-threads: off
> >>
> >> This values I have put it myself, in order to see if could improve
> >the
> >> time to make the volume available, when pve01 goes down with ifupdown
> >> cluster.quorum-count: 1
> >> cluster.quorum-reads: false
> >> cluster.self-heal-daemon: enable
> >> cluster.heal-timeout: 10
> >>
> >> Nevertheless, it took more than 1 minutes to the volume VMS available
> >in
> >> the other host (pve02).
> >> Is there any trick to reduce this time ?
> >>
> >> Thanks
> >>
> >> ---
> >> Gilberto Nunes Ferreira
> >>
> >>
> >>
> >>
> >>
> >>
> >> Em qua., 5 de ago. de 2020 às 08:57, Gilberto Nunes <
> >> gilberto.nunes32 at gmail.com> escreveu:
> >>
> >>> hum I see... like this:
> >>> [image: image.png]
> >>> ---
> >>> Gilberto Nunes Ferreira
> >>>
> >>> (47) 3025-5907
> >>> (47) 99676-7530 - Whatsapp / Telegram
> >>>
> >>> Skype: gilberto.nunes36
> >>>
> >>>
> >>>
> >>>
> >>>
> >>> Em qua., 5 de ago. de 2020 às 02:14, Computerisms Corporation <
> >>> bob at computerisms.ca> escreveu:
> >>>
> >>>> check the example of the chained configuration on this page:
> >>>>
> >>>>
> >>>>
> >
> https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.3/html/administration_guide/creating_arbitrated_replicated_volumes
> >>>>
> >>>> and apply it to two servers...
> >>>>
> >>>> On 2020-08-04 8:25 p.m., Gilberto Nunes wrote:
> >>>> > Hi Bob!
> >>>> >
> >>>> > Could you, please, send me more detail about this configuration?
> >>>> > I will appreciate that!
> >>>> >
> >>>> > Thank you
> >>>> > ---
> >>>> > Gilberto Nunes Ferreira
> >>>> >
> >>>> > (47) 3025-5907
> >>>> > **
> >>>> > (47) 99676-7530 - Whatsapp / Telegram
> >>>> >
> >>>> > Skype: gilberto.nunes36
> >>>> >
> >>>> >
> >>>> >
> >>>> >
> >>>> >
> >>>> > Em ter., 4 de ago. de 2020 às 23:47, Computerisms Corporation
> >>>> > <bob at computerisms.ca <mailto:bob at computerisms.ca>> escreveu:
> >>>> >
> >>>> >     Hi Gilberto,
> >>>> >
> >>>> >     My understanding is there can only be one arbiter per
> >replicated
> >>>> >     set.  I
> >>>> >     don't have a lot of practice with gluster, so this could be
> >bad
> >>>> advice,
> >>>> >     but the way I dealt with it on my two servers was to use 6
> >bricks
> >>>> as
> >>>> >     distributed-replicated (this is also relatively easy to
> >migrate to
> >>>> 3
> >>>> >     servers if that happens for you in the future):
> >>>> >
> >>>> >     Server1     Server2
> >>>> >     brick1      brick1.5
> >>>> >     arbiter1.5  brick2
> >>>> >     brick2.5    arbiter2.5
> >>>> >
> >>>> >     On 2020-08-04 7:00 p.m., Gilberto Nunes wrote:
> >>>> >      > Hi there.
> >>>> >      > I have two physical servers deployed as replica 2 and,
> >>>> obviously,
> >>>> >     I got
> >>>> >      > a split-brain.
> >>>> >      > So I am thinking in use two virtual machines,each one in
> >>>> physical
> >>>> >      > servers....
> >>>> >      > Then this two VMS act as a artiber of gluster set....
> >>>> >      >
> >>>> >      > Is this doable?
> >>>> >      >
> >>>> >      > Thanks
> >>>> >      >
> >>>> >      > ________
> >>>> >      >
> >>>> >      >
> >>>> >      >
> >>>> >      > Community Meeting Calendar:
> >>>> >      >
> >>>> >      > Schedule -
> >>>> >      > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
> >>>> >      > Bridge: https://bluejeans.com/441850968
> >>>> >      >
> >>>> >      > Gluster-users mailing list
> >>>> >      > Gluster-users at gluster.org
> ><mailto:Gluster-users at gluster.org>
> >>>> >      > https://lists.gluster.org/mailman/listinfo/gluster-users
> >>>> >      >
> >>>> >     ________
> >>>> >
> >>>> >
> >>>> >
> >>>> >     Community Meeting Calendar:
> >>>> >
> >>>> >     Schedule -
> >>>> >     Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
> >>>> >     Bridge: https://bluejeans.com/441850968
> >>>> >
> >>>> >     Gluster-users mailing list
> >>>> >     Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
> >>>> >     https://lists.gluster.org/mailman/listinfo/gluster-users
> >>>> >
> >>>>
> >>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20200805/ff3596ed/attachment.html>


More information about the Gluster-users mailing list