[Gluster-users] Add an arbiter when have multiple bricks at    same server.

Gilberto Ferreira gilberto.nunes32 at gmail.com
Tue Nov 5 19:17:01 UTC 2024


Still getting error
pve01:~# gluster vol info

Volume Name: VMS
Type: Distributed-Replicate
Volume ID: e1a4f787-3f62-441e-a7ce-c0ae6b111ebf
Status: Started
Snapshot Count: 0
Number of Bricks: 3 x 2 = 6
Transport-type: tcp
Bricks:
Brick1: gluster1:/disco2TB-0/vms
Brick2: gluster2:/disco2TB-0/vms
Brick3: gluster1:/disco1TB-0/vms
Brick4: gluster2:/disco1TB-0/vms
Brick5: gluster1:/disco1TB-1/vms
Brick6: gluster2:/disco1TB-1/vms
Options Reconfigured:
cluster.self-heal-daemon: off
cluster.entry-self-heal: off
cluster.metadata-self-heal: off
cluster.data-self-heal: off
cluster.granular-entry-heal: on
storage.fips-mode-rchecksum: on
transport.address-family: inet
performance.client-io-threads: off
pve01:~# gluster volume add-brick VMS replica 3 arbiter 1
gluster1:/disco2TB-0/vms gluster2:/disco2TB-0/vms arbiter:/arbiter1 force
volume add-brick: failed: Brick: gluster1:/disco2TB-0/vms not available.
Brick may be containing or be contained by an existing brick.
pve01:~#
---


Gilberto Nunes Ferreira







Em ter., 5 de nov. de 2024 às 14:33, Gilberto Ferreira <
gilberto.nunes32 at gmail.com> escreveu:

> Yes but I want to add.
> Is it the same logic?
> ---
> Gilberto Nunes Ferreira
> +55 (47) 99676-7530
> Proxmox VE
> VinChin Backup & Restore
>
> Em ter., 5 de nov. de 2024, 14:09, Aravinda <aravinda at kadalu.tech>
> escreveu:
>
>> Hello Gilberto,
>>
>> You can create a Arbiter volume using three bricks. Two of them will be
>> data bricks and one will be Arbiter brick.
>>
>> gluster volume create VMS replica 3 arbiter 1 gluster1:/disco2TB-0/vms
>> gluster2:/disco2TB-0/vms arbiter:/arbiter1
>>
>> To make this Volume as distributed Arbiter, add more bricks (multiple of
>> 3, two data bricks and one arbiter brick) similar to above.
>>
>> --
>> Aravinda
>>
>>
>> ---- On Tue, 05 Nov 2024 22:24:38 +0530 *Gilberto Ferreira
>> <gilberto.nunes32 at gmail.com <gilberto.nunes32 at gmail.com>>* wrote ---
>>
>> Clearly I am doing something wrong
>>
>> pve01:~# gluster vol info
>>
>> Volume Name: VMS
>> Type: Distributed-Replicate
>> Volume ID: e1a4f787-3f62-441e-a7ce-c0ae6b111ebf
>> Status: Started
>> Snapshot Count: 0
>> Number of Bricks: 3 x 2 = 6
>> Transport-type: tcp
>> Bricks:
>> Brick1: gluster1:/disco2TB-0/vms
>> Brick2: gluster2:/disco2TB-0/vms
>> Brick3: gluster1:/disco1TB-0/vms
>> Brick4: gluster2:/disco1TB-0/vms
>> Brick5: gluster1:/disco1TB-1/vms
>> Brick6: gluster2:/disco1TB-1/vms
>> Options Reconfigured:
>> cluster.self-heal-daemon: off
>> cluster.entry-self-heal: off
>> cluster.metadata-self-heal: off
>> cluster.data-self-heal: off
>> cluster.granular-entry-heal: on
>> storage.fips-mode-rchecksum: on
>> transport.address-family: inet
>> performance.client-io-threads: off
>> pve01:~# gluster vol add-brick VMS replica 3 arbiter 1
>> gluster1:/disco2TB-0/vms gluster2:/disco2TB-0/vms gluster1:/disco1TB-0/vms
>> gluster2:/disco1TB-0/vms gluster1:/disco1TB-1/vms gluster2:
>> /disco1TB-1/vms arbiter:/arbiter1
>> volume add-brick: failed: Operation failed
>> pve01:~# gluster vol add-brick VMS replica 3 arbiter 1
>> gluster1:/disco2TB-0/vms gluster2:/disco2TB-0/vms gluster1:/disco1TB-0/vms
>> gluster2:/disco1TB-0/vms gluster1:/disco1TB-1/vms gluster2:
>> /disco1TB-1/vms arbiter:/arbiter1 arbiter:/arbiter2 arbiter:/arbiter3
>> arbiter:/arbiter4
>> volume add-brick: failed: Operation failed
>>
>> ---
>>
>>
>> Gilberto Nunes Ferreira
>>
>>
>>
>>
>>
>>
>>
>> Em ter., 5 de nov. de 2024 às 13:39, Andreas Schwibbe <a.schwibbe at gmx.net>
>> escreveu:
>>
>>
>> If you create a volume with replica 2 arbiter 1
>>
>> you create 2 data bricks that are mirrored (makes 2 file copies)
>> +
>> you create 1 arbiter that holds metadata of all files on these bricks.
>>
>> You "can" create all on the same server, but this makes no sense, because
>> when the server goes down, no files on these disks are accessible anymore,
>> hence why bestpractice is to spread out over 3 servers, so when one
>> server (or disk) goes down, you will still have 1 file copy and 1 arbiter
>> with metadata online.
>> Which is also very handy when the down server comes up again, because
>> then you prevent splitbrain as you have matching file copy + metadata
>> showing which version of each file is newest, thus self-heal can jump in to
>> get you back to 2 file copies.
>>
>> when you want to add further bricks, you must add pairs i.e.
>> you will add again 2 bricks 1 arbiter and these bricks and arbiter belong
>> together and share the same files and metadata.
>>
>> Hth.
>> A.
>>
>>
>> Am Dienstag, dem 05.11.2024 um 13:28 -0300 schrieb Gilberto Ferreira:
>>
>> Ok.
>> I got confused here!
>> For each brick I will need one arbiter brick, in a different
>> partition/folder?
>> And what if in some point in the future I decide to add a new brick in
>> the main servers?
>> Do I need to provide another partition/folder in the arbiter and then
>> adjust the arbiter brick counter?
>>
>> ---
>>
>>
>> Gilberto Nunes Ferreira
>>
>>
>>
>>
>>
>>
>>
>> Em ter., 5 de nov. de 2024 às 13:22, Andreas Schwibbe <a.schwibbe at gmx.net>
>> escreveu:
>>
>> Your add-brick command adds 2 bricks 1 arbiter (even though you name them
>> all arbiter!)
>>
>> The sequence is important:
>>
>> gluster v add-brick VMS replica 2 arbiter 1 gluster1:/gv0 gluster2:/gv0
>> arbiter1:/arb1
>>
>> adds two data bricks and a corresponding arbiter from 3 different servers
>> and 3 different disks,
>> thus you can loose any one server OR any one disk and stay up and
>> consistent.
>>
>> adding more bricks to the volume you can follow the pattern.
>>
>> A.
>>
>> Am Dienstag, dem 05.11.2024 um 12:51 -0300 schrieb Gilberto Ferreira:
>>
>> Hi there.
>>
>> In previous emails, I comment with you guys, about 2 node gluster server,
>> where the bricks lay down in different size and folders in the same server,
>> like
>>
>> gluster vol create VMS replica 2 gluster1:/disco2TB-0/vms
>> gluster2:/disco2TB-0/vms gluster1:/disco1TB-0/vms gluster2:/disco1TB-0/vms
>> gluster1:/disco1TB-1/vms gluster2:/disco1TB-1/vms
>>
>> So I went ahead and installed a Debian 12 and installed the same gluster
>> version that the other servers, which is now 11.1 or something like that.
>> In this new server, I have a small disk like 480G in size.
>> And I created 3 partitions formatted with XFS using imaxpct=75, as
>> suggested in previous emails.
>>
>> And than in the gluster nodes, I tried to add the brick
>> gluster vol add-brick VMS replica 3 arbiter 1 arbiter:/arbiter1/arbiter1
>> arbiter:/arbiter2/arbiter2 arbiter:/arbiter3/arbiter3
>>
>> But to my surprise (or not!) I got this message:
>> volume add-brick: failed: Multiple bricks of a replicate volume are
>> present on the same server. This setup is not optimal. Bricks should be on
>> different nodes to have best fault tolerant co
>> nfiguration. Use 'force' at the end of the command if you want to
>> override this behavior.
>>
>> Why is that?
>>
>>
>>
>>
>>
>>
>> ---
>>
>>
>> Gilberto Nunes Ferreira
>>
>>
>>
>>
>>
>> ________
>>
>>
>>
>> Community Meeting Calendar:
>>
>> Schedule -
>> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
>> Bridge: https://meet.google.com/cpu-eiue-hvk
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> https://lists.gluster.org/mailman/listinfo/gluster-users
>>
>>
>>
>>
>>
>> ________
>>
>>
>>
>> Community Meeting Calendar:
>>
>> Schedule -
>> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
>> Bridge: https://meet.google.com/cpu-eiue-hvk
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> https://lists.gluster.org/mailman/listinfo/gluster-users
>>
>> ________
>>
>>
>>
>> Community Meeting Calendar:
>>
>> Schedule -
>> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
>> Bridge: https://meet.google.com/cpu-eiue-hvk
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> https://lists.gluster.org/mailman/listinfo/gluster-users
>>
>>
>>
>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20241105/36206b82/attachment.html>


More information about the Gluster-users mailing list