[Gluster-users] GlusterFS Scale

Lindolfo Meira meira at cesup.ufrgs.br
Mon Feb 18 17:53:19 UTC 2019

We're running some benchmarks on a striped glusterfs volume.

We have 6 identical servers acting as bricks. Measured link speed between 
these servers is 3.36GB/s. Link speed between clients of the parallel file 
system and its servers is also 3.36GB/s. So we're expecting this system to 
have a write performance of around 20.16GB/s (6 times 3.36GB/s) minus some 
write overhead.

If we write to the system from a single client, we manage to write at 
around 3.36GB/s. That's okay, because we're limited by the max throughput 
of that client's network adapter. But when we account for that and write 
from 6 or more clients, we can never get past 11GB/s. Is that right? Is 
this really the overhead to be expected? We'd appreciate any inputs.

Output of gluster volume info:

Volume Name: gfs0
Type: Stripe
Volume ID: 2ca3dd45-6209-43ff-a164-7f2694097c64
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 6 = 6
Transport-type: tcp
Brick1: pfs01-ib:/mnt/data
Brick2: pfs02-ib:/mnt/data
Brick3: pfs03-ib:/mnt/data
Brick4: pfs04-ib:/mnt/data
Brick5: pfs05-ib:/mnt/data
Brick6: pfs06-ib:/mnt/data
Options Reconfigured:
cluster.stripe-block-size: 128KB
performance.cache-size: 32MB
performance.write-behind-window-size: 1MB
performance.strict-write-ordering: off
performance.strict-o-direct: off
performance.stat-prefetch: off
server.event-threads: 4
client.event-threads: 2
performance.io-thread-count: 16
transport.address-family: inet
nfs.disable: on
cluster.localtime-logging: enable


Lindolfo Meira, MSc
Diretor Geral, Centro Nacional de Supercomputação
Universidade Federal do Rio Grande do Sul
+55 (51) 3308-3139

More information about the Gluster-users mailing list