[Gluster-users] proper way to temporarily remove brick server from replica cluster to avoid kvm guest disruption

Todd Pfaff pfaff at rhpcs.mcmaster.ca
Sun Mar 6 02:06:36 UTC 2022


[root at vh1 ~]# gluster volume info vol1

Volume Name: vol1
Type: Replicate
Volume ID: dfd681bb-5b68-4831-9863-e13f9f027620
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 4 = 4
Transport-type: tcp
Bricks:
Brick1: vh1:/pool/gluster/brick1/data
Brick2: vh2:/pool/gluster/brick1/data
Brick3: vh3:/pool/gluster/brick1/data
Brick4: vh4:/pool/gluster/brick1/data
Options Reconfigured:
transport.address-family: inet
nfs.disable: on
performance.client-io-threads: off


[root at vh1 ~]# gluster pool list
UUID                                    Hostname        State
75fc4258-fabd-47c9-8198-bbe6e6a906fb    vh2             Connected
00697e28-96c0-4534-a314-e878070b653d    vh3             Connected
2a9b891b-35d0-496c-bb06-f5dab4feb6bf    vh4             Connected
8ba6fb80-3b13-4379-94cf-22662cbb48a2    vh5             Disconnected
1298d334-3500-4b40-a8bd-cc781f7349d0    vh6             Connected
79a533ac-3d89-44b9-b0ce-823cfec8cf75    vh7             Connected
4141cd74-9c13-404c-a02c-f553fa19bc22    vh8             Connected


On Sat, 5 Mar 2022, Strahil Nikolov wrote:

> Hey Todd,
> 
> can you provide 'gluster volume info <VOLUME>' ?
> 
> Best Regards,
> Strahil Nikolov
>
>       On Sat, Mar 5, 2022 at 18:17, Todd Pfaff
> <pfaff at rhpcs.mcmaster.ca> wrote:
> I have a replica volume created as:
> 
> gluster volume create vol1 replica 4 \
>   host{1,2,3,4}:/mnt/gluster/brick1/data \
>   force
> 
> 
> All hosts host{1,2,3,4} mount this volume as:
> 
> localhost:/vol1 /mnt/gluster/vol1 glusterfs defaults
> 
> 
> Some other hosts are trusted peers but do not contribute bricks, and
> they
> also mount vol1 in the same way:
> 
> localhost:/vol1 /mnt/gluster/vol1 glusterfs defaults
> 
> 
> All hosts run CentOS 7.9, and all are running glusterfs 9.4 or 9.5 from
> centos-release-gluster9-1.0-1.el7.noarch.
> 
> 
> All hosts run kvm guests that use qcow2 files for root filesystems that
> are stored on gluster volume vol1.
> 
> 
> This is all working well, as long as none of host{1,2,3,4} go offline.
> 
> 
> I want to take one of host{1,2,3,4} offline temporarily for
> maintenance.
> I'll refer to this as hostX.
> 
> I understand that hostX will need to be healed when it comes back
> online.
> 
> I would, of course, migrate guests from hostX to another host, in which
> case hostX would then only be participating as a gluster replica brick
> provider and serving gluster client requests.
> 
> What I've experienced is that if I take one of host{1,2,3,4} offline,
> this
> can disrupt some of the VM guests on various other hosts such that
> their
> root filesystems go to read-only.
> 
> What I'm looking for here are suggestions as to how to properly take
> one
> of host{1,2,3,4} offline to avoid such disruption or how to tune the
> libvirt kvm hosts and guests to be sufficiently resilient in the face
> of
> taking one gluster replica node offline.
> 
> Thanks,
> Todd
> ________
> 
> 
> 
> Community Meeting Calendar:
> 
> Schedule -
> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
> Bridge: https://meet.google.com/cpu-eiue-hvk
> Gluster-users mailing list
> Gluster-users at gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users
> 
> 
>


More information about the Gluster-users mailing list