[Gluster-users] Two VMS as arbiter...

Gilberto Nunes gilberto.nunes32 at gmail.com
Thu Aug 6 13:37:07 UTC 2020


Oh I see... I was confused because the terms... Now I read this and
everything becomes clear...

https://staged-gluster-docs.readthedocs.io/en/release3.7.0beta1/Features/shard/

https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.3/html/configuring_red_hat_virtualization_with_red_hat_gluster_storage/chap-hosting_virtual_machine_images_on_red_hat_storage_volumes


Should I use cluster.granular-entrey-heal-enable too, since I am working
with big files?

Thanks

---
Gilberto Nunes Ferreira

(47) 3025-5907
(47) 99676-7530 - Whatsapp / Telegram

Skype: gilberto.nunes36





Em qui., 6 de ago. de 2020 às 09:32, Gilberto Nunes <
gilberto.nunes32 at gmail.com> escreveu:

> What do you mean "sharding"? Do you mean sharing folders between two
> servers to host qcow2 or raw vm images?
> Here I am using Proxmox which uses qemu but not virsh.
>
> Thanks
> ---
> Gilberto Nunes Ferreira
>
> (47) 3025-5907
> (47) 99676-7530 - Whatsapp / Telegram
>
> Skype: gilberto.nunes36
>
>
>
>
>
> Em qui., 6 de ago. de 2020 às 01:09, Strahil Nikolov <
> hunter86_bg at yahoo.com> escreveu:
>
>> As  you mentioned qcow2  files,  check the virt group
>> (/var/lib/glusterfs/group or something like that). It has optimal setttins
>> for VMs and is used by oVirt.
>>
>> WARNING: If you decide to enable the group, which will also enable
>> sharding, NEVER EVER DISABLE SHARDING -> ONCE ENABLED STAYS ENABLED !!!
>> Sharding helps reduce loocking during replica heals.
>>
>> WARNING2: As virt group uses sharding (fixes the size of file into shard
>> size),  you should consider cluster.favorite-child-policy with value
>> ctime/mtime.
>>
>> Best Regards,
>> Strahil Nikolov
>>
>> На 6 август 2020 г. 1:56:58 GMT+03:00, Gilberto Nunes <
>> gilberto.nunes32 at gmail.com> написа:
>> >Ok...Thanks a lot Strahil
>> >
>> >This gluster volume set VMS cluster.favorite-child-policy size do the
>> >trick
>> >to me here!
>> >
>> >Cheers
>> >---
>> >Gilberto Nunes Ferreira
>> >
>> >(47) 3025-5907
>> >(47) 99676-7530 - Whatsapp / Telegram
>> >
>> >Skype: gilberto.nunes36
>> >
>> >
>> >
>> >
>> >
>> >Em qua., 5 de ago. de 2020 às 18:15, Strahil Nikolov
>> ><hunter86_bg at yahoo.com>
>> >escreveu:
>> >
>> >> This could happen if you have pending heals. Did you reboot that node
>> >> recently ?
>> >> Did you set automatic unsplit-brain ?
>> >>
>> >> Check for pending heals and files in splitbrain.
>> >>
>> >> If not, you can check
>> >>
>> >https://docs.gluster.org/en/latest/Troubleshooting/resolving-splitbrain/
>> >> (look at point 5).
>> >>
>> >> Best Regards,
>> >> Strahil Nikolov
>> >>
>> >> На 5 август 2020 г. 23:41:57 GMT+03:00, Gilberto Nunes <
>> >> gilberto.nunes32 at gmail.com> написа:
>> >> >I'm in trouble here.
>> >> >When I shutdown the pve01 server, the shared folder over glusterfs
>> >is
>> >> >EMPTY!
>> >> >It's supposed to be a qcow2 file inside it.
>> >> >The content is show right, just after I power on pve01 backup...
>> >> >
>> >> >Some advice?
>> >> >
>> >> >
>> >> >Thanks
>> >> >
>> >> >---
>> >> >Gilberto Nunes Ferreira
>> >> >
>> >> >(47) 3025-5907
>> >> >(47) 99676-7530 - Whatsapp / Telegram
>> >> >
>> >> >Skype: gilberto.nunes36
>> >> >
>> >> >
>> >> >
>> >> >
>> >> >
>> >> >Em qua., 5 de ago. de 2020 às 11:07, Gilberto Nunes <
>> >> >gilberto.nunes32 at gmail.com> escreveu:
>> >> >
>> >> >> Well...
>> >> >> I do the follow:
>> >> >>
>> >> >> gluster vol create VMS replica 3 arbiter 1 pve01:/DATA/brick1
>> >> >> pve02:/DATA/brick1.5 pve01:/DATA/arbiter1.5 pve02:/DATA/brick2 pv
>> >> >> e01:/DATA/brick2.5 pve02:/DATA/arbiter2.5 force
>> >> >>
>> >> >> And now I have:
>> >> >> gluster vol info
>> >> >>
>> >> >> Volume Name: VMS
>> >> >> Type: Distributed-Replicate
>> >> >> Volume ID: 1bd712f5-ccb9-4322-8275-abe363d1ffdd
>> >> >> Status: Started
>> >> >> Snapshot Count: 0
>> >> >> Number of Bricks: 2 x (2 + 1) = 6
>> >> >> Transport-type: tcp
>> >> >> Bricks:
>> >> >> Brick1: pve01:/DATA/brick1
>> >> >> Brick2: pve02:/DATA/brick1.5
>> >> >> Brick3: pve01:/DATA/arbiter1.5 (arbiter)
>> >> >> Brick4: pve02:/DATA/brick2
>> >> >> Brick5: pve01:/DATA/brick2.5
>> >> >> Brick6: pve02:/DATA/arbiter2.5 (arbiter)
>> >> >> Options Reconfigured:
>> >> >> cluster.quorum-count: 1
>> >> >> cluster.quorum-reads: false
>> >> >> cluster.self-heal-daemon: enable
>> >> >> cluster.heal-timeout: 10
>> >> >> storage.fips-mode-rchecksum: on
>> >> >> transport.address-family: inet
>> >> >> nfs.disable: on
>> >> >> performance.client-io-threads: off
>> >> >>
>> >> >> This values I have put it myself, in order to see if could improve
>> >> >the
>> >> >> time to make the volume available, when pve01 goes down with
>> >ifupdown
>> >> >> cluster.quorum-count: 1
>> >> >> cluster.quorum-reads: false
>> >> >> cluster.self-heal-daemon: enable
>> >> >> cluster.heal-timeout: 10
>> >> >>
>> >> >> Nevertheless, it took more than 1 minutes to the volume VMS
>> >available
>> >> >in
>> >> >> the other host (pve02).
>> >> >> Is there any trick to reduce this time ?
>> >> >>
>> >> >> Thanks
>> >> >>
>> >> >> ---
>> >> >> Gilberto Nunes Ferreira
>> >> >>
>> >> >>
>> >> >>
>> >> >>
>> >> >>
>> >> >>
>> >> >> Em qua., 5 de ago. de 2020 às 08:57, Gilberto Nunes <
>> >> >> gilberto.nunes32 at gmail.com> escreveu:
>> >> >>
>> >> >>> hum I see... like this:
>> >> >>> [image: image.png]
>> >> >>> ---
>> >> >>> Gilberto Nunes Ferreira
>> >> >>>
>> >> >>> (47) 3025-5907
>> >> >>> (47) 99676-7530 - Whatsapp / Telegram
>> >> >>>
>> >> >>> Skype: gilberto.nunes36
>> >> >>>
>> >> >>>
>> >> >>>
>> >> >>>
>> >> >>>
>> >> >>> Em qua., 5 de ago. de 2020 às 02:14, Computerisms Corporation <
>> >> >>> bob at computerisms.ca> escreveu:
>> >> >>>
>> >> >>>> check the example of the chained configuration on this page:
>> >> >>>>
>> >> >>>>
>> >> >>>>
>> >> >
>> >>
>> >
>> https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.3/html/administration_guide/creating_arbitrated_replicated_volumes
>> >> >>>>
>> >> >>>> and apply it to two servers...
>> >> >>>>
>> >> >>>> On 2020-08-04 8:25 p.m., Gilberto Nunes wrote:
>> >> >>>> > Hi Bob!
>> >> >>>> >
>> >> >>>> > Could you, please, send me more detail about this
>> >configuration?
>> >> >>>> > I will appreciate that!
>> >> >>>> >
>> >> >>>> > Thank you
>> >> >>>> > ---
>> >> >>>> > Gilberto Nunes Ferreira
>> >> >>>> >
>> >> >>>> > (47) 3025-5907
>> >> >>>> > **
>> >> >>>> > (47) 99676-7530 - Whatsapp / Telegram
>> >> >>>> >
>> >> >>>> > Skype: gilberto.nunes36
>> >> >>>> >
>> >> >>>> >
>> >> >>>> >
>> >> >>>> >
>> >> >>>> >
>> >> >>>> > Em ter., 4 de ago. de 2020 às 23:47, Computerisms Corporation
>> >> >>>> > <bob at computerisms.ca <mailto:bob at computerisms.ca>> escreveu:
>> >> >>>> >
>> >> >>>> >     Hi Gilberto,
>> >> >>>> >
>> >> >>>> >     My understanding is there can only be one arbiter per
>> >> >replicated
>> >> >>>> >     set.  I
>> >> >>>> >     don't have a lot of practice with gluster, so this could
>> >be
>> >> >bad
>> >> >>>> advice,
>> >> >>>> >     but the way I dealt with it on my two servers was to use 6
>> >> >bricks
>> >> >>>> as
>> >> >>>> >     distributed-replicated (this is also relatively easy to
>> >> >migrate to
>> >> >>>> 3
>> >> >>>> >     servers if that happens for you in the future):
>> >> >>>> >
>> >> >>>> >     Server1     Server2
>> >> >>>> >     brick1      brick1.5
>> >> >>>> >     arbiter1.5  brick2
>> >> >>>> >     brick2.5    arbiter2.5
>> >> >>>> >
>> >> >>>> >     On 2020-08-04 7:00 p.m., Gilberto Nunes wrote:
>> >> >>>> >      > Hi there.
>> >> >>>> >      > I have two physical servers deployed as replica 2 and,
>> >> >>>> obviously,
>> >> >>>> >     I got
>> >> >>>> >      > a split-brain.
>> >> >>>> >      > So I am thinking in use two virtual machines,each one
>> >in
>> >> >>>> physical
>> >> >>>> >      > servers....
>> >> >>>> >      > Then this two VMS act as a artiber of gluster set....
>> >> >>>> >      >
>> >> >>>> >      > Is this doable?
>> >> >>>> >      >
>> >> >>>> >      > Thanks
>> >> >>>> >      >
>> >> >>>> >      > ________
>> >> >>>> >      >
>> >> >>>> >      >
>> >> >>>> >      >
>> >> >>>> >      > Community Meeting Calendar:
>> >> >>>> >      >
>> >> >>>> >      > Schedule -
>> >> >>>> >      > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
>> >> >>>> >      > Bridge: https://bluejeans.com/441850968
>> >> >>>> >      >
>> >> >>>> >      > Gluster-users mailing list
>> >> >>>> >      > Gluster-users at gluster.org
>> >> ><mailto:Gluster-users at gluster.org>
>> >> >>>> >      >
>> >https://lists.gluster.org/mailman/listinfo/gluster-users
>> >> >>>> >      >
>> >> >>>> >     ________
>> >> >>>> >
>> >> >>>> >
>> >> >>>> >
>> >> >>>> >     Community Meeting Calendar:
>> >> >>>> >
>> >> >>>> >     Schedule -
>> >> >>>> >     Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
>> >> >>>> >     Bridge: https://bluejeans.com/441850968
>> >> >>>> >
>> >> >>>> >     Gluster-users mailing list
>> >> >>>> >     Gluster-users at gluster.org
>> ><mailto:Gluster-users at gluster.org>
>> >> >>>> >     https://lists.gluster.org/mailman/listinfo/gluster-users
>> >> >>>> >
>> >> >>>>
>> >> >>>
>> >>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20200806/8ae6bbe4/attachment.html>


More information about the Gluster-users mailing list