[Gluster-users] Experimenting with thin-arbiter

Diego Zuccato diego.zuccato at unibo.it
Wed Feb 16 05:44:54 UTC 2022


No such process is defined. Just the standard glusterd.service and 
glustereventsd.service . Using Debian stable.

Il 15/02/2022 15:41, Strahil Nikolov ha scritto:
> Any errors in gluster-ta.service on the arbiter node ?
> 
> Best Regards,
> Strahil Nikolov
> 
>     On Tue, Feb 15, 2022 at 14:28, Diego Zuccato
>     <diego.zuccato at unibo.it> wrote:
>     Hello all.
> 
>     I'm experimenting with thin-arbiter and getting disappointing results.
> 
>     I have 3 hosts in the trusted pool:
>     root at nas1 <mailto:root at nas1>:~# gluster --version
>     glusterfs 9.2
>     [...]
>     root at nas1 <mailto:root at nas1>:~# gluster pool list
>     UUID                                    Hostname        State
>     d4791fed-3e6d-4f8f-bdb6-4e0043610ead    nas3            Connected
>     bff398f0-9d1d-4bd0-8a47-0bf481d1d593    nas2            Connected
>     4607034c-919d-4675-b5fc-14e1cad90214    localhost      Connected
> 
>     When I try to create a new volume, the first initialization succeeds:
>     root at nas1 <mailto:root at nas1>:~# gluster v create Bck replica 2
>     thin-arbiter 1
>     nas{1,3}:/bricks/00/Bck nas2:/bricks/arbiter/Bck
>     volume create: Bck: success: please start the volume to access data
> 
>     But adding a second brick segfaults the daemon:
>     root at nas1 <mailto:root at nas1>:~# gluster v add-brick Bck
>     nas{1,3}:/bricks/01/Bck
>     Connection failed. Please check if gluster daemon is operational.
> 
>     After erroring out, systemctl status glusterd reports daemon in
>     "restarting" state and it eventually restarts. But the new brick is not
>     added to the volume, even if trying to re-add it yelds a "brick is
>     already part of a volume" error. Seems glusterd crashes between marking
>     brick dir as used and recording its data in the config.
> 
>     If I try to add all the bricks during the creation, glusterd does not
>     die but the volume doesn't get created:
>     root at nas1 <mailto:root at nas1>:~# rm -rf /bricks/{00..07}/Bck && mkdir
>     /bricks/{00..07}/Bck
>     root at nas1 <mailto:root at nas1>:~# gluster v create Bck replica 2
>     thin-arbiter 1
>     nas{1,3}:/bricks/00/Bck nas{1,3}:/bricks/01/Bck nas{1,3}:/bricks/02/Bck
>     nas{1,3}:/bricks/03/Bck nas{1,3}:/bricks/04/Bck nas{1,3}:/bricks/05/Bck
>     nas{1,3}:/bricks/06/Bck nas{1,3}:/bricks/07/Bck nas2:/bricks/arbiter/Bck
>     volume create: Bck: failed: Commit failed on localhost. Please check
>     the
>     log file for more details.
> 
>     Couldn't find anything useful in the logs :(
> 
>     If I create a "replica 3 arbiter 1" over the same brick directories
>     (just adding some directories to keep arbiters separated), it succeeds:
>     root at nas1 <mailto:root at nas1>:~# gluster v create Bck replica 3
>     arbiter 1
>     nas{1,3}:/bricks/00/Bck nas2:/bricks/arbiter/Bck/00
>     volume create: Bck: success: please start the volume to access data
>     root at nas1 <mailto:root at nas1>:~# for T in {01..07}; do gluster v
>     add-brick Bck
>     nas{1,3}:/bricks/$T/Bck nas2:/bricks/arbiter/Bck/$T ; done
>     volume add-brick: success
>     volume add-brick: success
>     volume add-brick: success
>     volume add-brick: success
>     volume add-brick: success
>     volume add-brick: success
>     volume add-brick: success
>     root at nas1 <mailto:root at nas1>:~# gluster v start Bck
>     volume start: Bck: success
>     root at nas1 <mailto:root at nas1>:~# gluster v info Bck
> 
>     Volume Name: Bck
>     Type: Distributed-Replicate
>     Volume ID: 4786e747-8203-42bf-abe8-107a50b238ee
>     Status: Started
>     Snapshot Count: 0
>     Number of Bricks: 8 x (2 + 1) = 24
>     Transport-type: tcp
>     Bricks:
>     Brick1: nas1:/bricks/00/Bck
>     Brick2: nas3:/bricks/00/Bck
>     Brick3: nas2:/bricks/arbiter/Bck/00 (arbiter)
>     Brick4: nas1:/bricks/01/Bck
>     Brick5: nas3:/bricks/01/Bck
>     Brick6: nas2:/bricks/arbiter/Bck/01 (arbiter)
>     Brick7: nas1:/bricks/02/Bck
>     Brick8: nas3:/bricks/02/Bck
>     Brick9: nas2:/bricks/arbiter/Bck/02 (arbiter)
>     Brick10: nas1:/bricks/03/Bck
>     Brick11: nas3:/bricks/03/Bck
>     Brick12: nas2:/bricks/arbiter/Bck/03 (arbiter)
>     Brick13: nas1:/bricks/04/Bck
>     Brick14: nas3:/bricks/04/Bck
>     Brick15: nas2:/bricks/arbiter/Bck/04 (arbiter)
>     Brick16: nas1:/bricks/05/Bck
>     Brick17: nas3:/bricks/05/Bck
>     Brick18: nas2:/bricks/arbiter/Bck/05 (arbiter)
>     Brick19: nas1:/bricks/06/Bck
>     Brick20: nas3:/bricks/06/Bck
>     Brick21: nas2:/bricks/arbiter/Bck/06 (arbiter)
>     Brick22: nas1:/bricks/07/Bck
>     Brick23: nas3:/bricks/07/Bck
>     Brick24: nas2:/bricks/arbiter/Bck/07 (arbiter)
>     Options Reconfigured:
>     cluster.granular-entry-heal: on
>     storage.fips-mode-rchecksum: on
>     transport.address-family: inet
>     nfs.disable: on
>     performance.client-io-threads: off
> 
>     Does thin arbiter support just one replica of bricks?
> 
>     -- 
>     Diego Zuccato
>     DIFA - Dip. di Fisica e Astronomia
>     Servizi Informatici
>     Alma Mater Studiorum - Università di Bologna
>     V.le Berti-Pichat 6/2 - 40127 Bologna - Italy
>     tel.: +39 051 20 95786
>     ________
> 
> 
> 
>     Community Meeting Calendar:
> 
>     Schedule -
>     Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
>     Bridge: https://meet.google.com/cpu-eiue-hvk
>     <https://meet.google.com/cpu-eiue-hvk>
>     Gluster-users mailing list
>     Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
>     https://lists.gluster.org/mailman/listinfo/gluster-users
>     <https://lists.gluster.org/mailman/listinfo/gluster-users>
> 

-- 
Diego Zuccato
DIFA - Dip. di Fisica e Astronomia
Servizi Informatici
Alma Mater Studiorum - Università di Bologna
V.le Berti-Pichat 6/2 - 40127 Bologna - Italy
tel.: +39 051 20 95786


More information about the Gluster-users mailing list