[Gluster-users] self healing with sharding
Lindsay Mathieson
lindsay.mathieson at gmail.com
Sat Jul 9 02:46:46 UTC 2016
On 8/07/2016 9:40 PM, Gandalf Corvotempesta wrote:
> How did you mesure the performance? I would like to test in the same
> way, so that results are comparable.
Not particularity scientific. I have four main tests I run
1. CrystalDiskMark in a Windows VM. This lets me see IOPS as
experienced by the VM. I'm suspicious of std disk becnhmarks though,
they don't really reflect day-day usage.
2. The build server for our enterprise product, a fairly large cmd
line build, a real world usage that exercises random read/writes fairly
well.
3. Starting up and running std applications - eclipse, Office 365,
outlook etc. More subjective, which does matter.
4. Multiple simultaneous VM starts, a good stress test.
> Which network/hardware/servers topology are you using ?
3 Compute Servers - Combined VM hosts and gluster nodes, for a replica 3
gluster volume
VNA:
- Dual Xeon E5-2660 2.2Ghz
- 64GB EEC Ram
- 2 * 1Gb Bond
- 4x3TB WD red in ZFS RAID10
VNB, VNG :
- Xeon E5-2620 2.0 Ghz
- 64GB Ram
- 3 * 1Gb Bond
- 4x3TB WD red in ZFS RAID10
All Bonds are LACP Balance-tcp with a dedicated Switch. VNA is supposed
to have 3*1Gb as well but we had driver problems with the 3rd card and I
haven't got round to fixing it :(
Internal & external traffic share the bond. External traffic is minimal.
--
Lindsay Mathieson
More information about the Gluster-users
mailing list