[Gluster-users] Slow performance over rsync in Replicated-Distributed Setup

Felix Kölzow felix.koelzow at gmx.de
Sat Mar 6 17:54:26 UTC 2021


Dear Shubhank,

this small file performance appears to be slow on glusterfs usually.

Can you provide more details according to your setup? (zfs settings,
bonding, tuned-adm profile, etc, ...)


 From a gluster point of view, setting performance.write-behind-window
to 128MB increases performance.

I was able to hit the cpu limit using smallfile benchmark tool
(available on github) and native glusterfs-client with that knob.


Furthermore, throughput increases if you increase the number of rsync
processes ( -> github -> msrsync works well here).


Regards,

Felix


On 06/03/2021 15:27, Shubhank Gaur wrote:
> Hello users,
>
> I have started using gluster just a few weeks ago and I am rocking a
> Replicated-Distributed setup with arbiters (A) and SATA Volumes (V). I
> have around 6 volumes and 3 arbiters in this setup:
> V+V+A | V+V+A | V+V+A
>
> All these volumes are spread across 3 different nodes, all of them
> being 1Gbit. Due to hardware limitations, SSD or 10Gbit network is not
> available.
>
> But even then, testing via iperf and normal rsync of files between
> servers, I am easily able to achieve 700Mbps~
> [ ID] Interval           Transfer     Bandwidth       Retr  Cwnd
> [  4]   0.00-1.00   sec  49.9 MBytes   419 Mbits/sec   21  132 KBytes
> [  4]   1.00-2.00   sec  80.0 MBytes   671 Mbits/sec    0  214 KBytes
> [  4]   2.00-3.00   sec  87.0 MBytes   730 Mbits/sec    3  228 KBytes
> [  4]   3.00-4.00   sec  91.6 MBytes   769 Mbits/sec   15  215 KBytes
>
> But when rsyncing data from same server to another node with mounted
> glusterVolume, I am getting measly 50Mbps (7MBps).
>
> All servers have 64GB Ram and their memory usage is around 50% and CPU
> usage less than 10%.
> All bricks are zfs volumes, no Raid setup or anything.  All volumes
> are direct hard disks formatted as ZFS (JBOD setup).
>
>
> My Gluster Vol Info
>
> gluster vol info
>
> Volume Name: glusterStore
> Type: Distributed-Replicate
> Volume ID: c7ac8094-f379-45fc-8cfd-f2937355e03d
> Status: Started
> Snapshot Count: 0
> Number of Bricks: 3 x (2 + 1) = 9
> Transport-type: tcp
> Bricks:
> Brick1: 62.0.0.1:/zpool1/proxmox
> Brick2: 5.0.0.1:/zpool1/proxmox
> Brick3: 62.0.0.1 :/home/glusterArbiter (arbiter)
> Brick4: 62.0.0.1 :/zpool2/proxmox
> Brick5: 5.0.0.1 :/zpool2/proxmox
> Brick6: 62.0.0.2:/home/glusterArbiter2 (arbiter)
> Brick7: 62.0.0.2:/zpool/proxmox
> Brick8: 5.0.0.1 :/zpool3/proxmox
> Brick9: 62.0.0.2:/home/glusterArbiter (arbiter)
> Options Reconfigured:
> performance.readdir-ahead: enable
> cluster.rsync-hash-regex: none
> client.event-threads: 16
> server.event-threads: 16
> network.ping-timeout: 5
> performance.normal-prio-threads: 64
> performance.high-prio-threads: 64
> performance.io-thread-count: 64
> performance.cache-size: 1GB
> performance.read-ahead: off
> performance.io-cache: off
> performance.flush-behind: off
> performance.quick-read: on
> network.frame-timeout: 60
> storage.batch-fsync-delay-usec: 0
> server.allow-insecure: on
> performance.stat-prefetch: off
> cluster.lookup-optimize: on
> performance.write-behind: on
> cluster.granular-entry-heal: on
> storage.fips-mode-rchecksum: on
> transport.address-family: inet
> nfs.disable: on
> performance.client-io-threads: off
>
>
> Regards
>
> ________
>
>
>
> Community Meeting Calendar:
>
> Schedule -
> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
> Bridge: https://meet.google.com/cpu-eiue-hvk
> Gluster-users mailing list
> Gluster-users at gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20210306/895ecb67/attachment.html>


More information about the Gluster-users mailing list