[Gluster-devel] Weird performance behavior
Xavi Hernandez
xhernandez at redhat.com
Thu May 2 20:58:12 UTC 2019
Hi,
doing some tests to compare performance I've found some weird results. I've
seen this in different tests, but probably the more clear an easier to
reproduce is to use smallfile tool to create files.
The test command is:
# python smallfile_cli.py --operation create --files-per-dir 100
--file-size 32768 --threads 16 --files 256 --top <mountpoint> --stonewall no
I've run this test 5 times sequentially using the same initial conditions
(at least this is what I think): bricks cleared, all gluster processes
stopped, volume destroyed and recreated, caches emptied.
This is the data I've obtained for each execution:
Time us sy ni id wa hi si st read
write use
435 1.80 3.70 0.00 81.62 11.06 0.00 0.00 0.00 32.931
608715.575 97.632
450 1.67 3.62 0.00 80.67 12.19 0.00 0.00 0.00 30.989
589078.308 97.714
425 1.74 3.75 0.00 81.85 10.76 0.00 0.00 0.00 37.588
622034.812 97.706
320 2.47 5.06 0.00 82.84 7.75 0.00 0.00 0.00 46.406
828637.359 96.891
365 2.19 4.44 0.00 84.45 7.12 0.00 0.00 0.00 45.822
734566.685 97.466
Time is in seconds. us, sy, ni, id, wa, hi, si and st are the CPU times, as
reported by top. read and write are the disk throughput in KiB/s. use is
the disk usage percentage.
Based on this we can see that there's a big difference between the best and
the worst cases. But it seems more relevant that when it performed better,
in fact disk utilization and CPU wait time were a bit lower.
Disk is a NVMe and I used a recent commit from master (2b86da69). Volume
type is a replica 3 with 3 bricks.
I'm not sure what can be causing this. Any idea ? can anyone try to
reproduce it to see if it's a problem in my environment or it's a common
problem ?
Thanks,
Xavi
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-devel/attachments/20190502/07036d6a/attachment.html>
More information about the Gluster-devel
mailing list