[Gluster-users] healing of big files
Strahil
hunter86_bg at yahoo.com
Tue Jan 14 15:21:11 UTC 2020
Are you using the 'virt' group for setting all optimal settings for virtualization ?
The groups' settings are in /var/lib/gluster/groups on each node.
Best Regards,
Strahil NikolovOn Jan 14, 2020 10:57, vlad f halilov <vfh at swemel.ru> wrote:
>
> Hi there. I'm trying to deploy replicated 2+1 cluster with gluster 5.5.1
> (debian 10) as vm image storage for libvirt, but have an issue with
> healing disconnected node. Example:
>
> 1) node4 + node5 - storages with one brick, node2 - arbiter.
> 2) Creating volume with defaults, and making a file (5Gb) in mounted fs,
> running 'dd bs=1024k count=5000 if=/dev/zero of=/mnt/zero'
> 3) Now restarting node5, and executing 'echo >>/mnt/zero' on node2. So
> file /mnt/zero become little changed and need to heal.
>
> # gluster volume heal vmpool info
>
> Brick node5:/data/glusterfs/vmpool/brick1/brick
> Status: enpoint not connected
> Number of entries: -
>
> Brick node4:/data/glusterfs/vmpool/brick1/brick
> /zero
> Status: Connected
> Number of entries: 1
>
> Brick node2:/data/glusterfs/vmpool/brick1/brick
> /zero
> Status: Connected
> Number of entries: 1
>
>
> 4) After node5 has reconnected, healing procedure has begun, but looks
> that about to entire file have transferred from synced node to unsynced,
> instead just changed blocks. (ifconfig on node5 showed ' RX
> packets 269790 bytes 5293724127 (4.9 GiB)) - so little less than 5Gb
> file' I tried 'difff' mode (gluster volume set vmpool
> cluster.data-self-heal-algorithm diff) but with no effect. Any idea,
> whats wrong?
>
> --
> vlad f halilov
> swemel
>
> ________
>
> Community Meeting Calendar:
>
> APAC Schedule -
> Every 2nd and 4th Tuesday at 11:30 AM IST
> Bridge: https://bluejeans.com/441850968
>
> NA/EMEA Schedule -
> Every 1st and 3rd Tuesday at 01:00 PM EDT
> Bridge: https://bluejeans.com/441850968
>
> Gluster-users mailing list
> Gluster-users at gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users
More information about the Gluster-users
mailing list