[Gluster-users] infiniband speed

Christopher Hawkins chawkins at bplinux.com
Tue Jan 11 00:43:18 UTC 2011


I am testing Infiniband for the first time. It seems that I should be able to get a lot more speed than I am with some pretty basic tests. Maybe someone running Infiniband can confirm that what I am seeing is way out of line, and/or help diagnose? 

I have two systems connected using 3.1.2qa3. With 3.1.1 infiniband wouldn't even start, it gave an error about unable to intialize rdma. But with the latest version and an upgrade to OFED 1.5.2, everything starts up with no errors and I can create a volume and mount it. 

The underlying Infiniband seems ok, and a basic ibv_rc_pingpong test shows I can move data pretty fast:
81920000 bytes in 0.23 seconds = 2858.45 Mbit/sec
10000 iters in 0.23 seconds = 22.93 usec/iter

So now I have two volumes created, one that uses tcp over a gig-e link and one that uses rdma. I mount them and do some file copy tests... And they are almost exactly the same? What? 

gluster volume info

Volume Name: test2_volume
Type: Replicate
Status: Started
Number of Bricks: 2
Transport-type: tcp
Bricks:
Brick1: bravo:/cluster/shadow/test2
Brick2: backup:/cluster/shadow/test2

Volume Name: test_volume
Type: Replicate
Status: Started
Number of Bricks: 2
Transport-type: rdma
Bricks:
Brick1: bravo:/cluster/shadow/test
Brick2: backup:/cluster/shadow/test

mount:
glusterfs#localhost:/test_volume on /mnt/test type fuse (rw,allow_other,default_permissions,max_read=131072)
glusterfs#localhost:/test2_volume on /mnt/test2 type fuse (rw,allow_other,default_permissions,max_read=131072)


time cp files.tar /mnt/test2/

real    0m11.159s
user    0m0.123s
sys     0m1.244s

files.tar is single file, 390MB, so this about 35MB/s. Fine for gig-e. 
----------------------------

time cp files.tar /mnt/test/

real    0m5.656s
user    0m0.116s
sys     0m0.962s

69MB/s... ehhh. Faster at least. On a few runs, this was not any faster at all. Maybe a cache effect? 
----------------------------

time cp -av /usr/src/kernels /mnt/test2/
real    0m49.605s
user    0m0.681s
sys     0m2.593s

kernels dir is 34MB of small files. The low latency of IB should really show an improvement here I thought. 
-----------------------------

time cp -av /usr/src/kernels /mnt/test/

real    0m56.046s
user    0m0.625s
sys     0m2.675s

It took LONGER? That can't be right. 
------------------------------

And finally, this error is appearing in the rdma mount log every 3 seconds on both nodes:

[2011-01-10 19:46:56.728127] E [rdma.c:4428:tcp_connect_finish] test_volume-client-1: tcp connect to  failed (Connection refused)
[2011-01-10 19:46:59.738291] E [rdma.c:4428:tcp_connect_finish] test_volume-client-1: tcp connect to  failed (Connection refused)
[2011-01-10 19:47:02.748260] E [rdma.c:4428:tcp_connect_finish] test_volume-client-1: tcp connect to  failed (Connection refused)
[2011-01-10 19:47:05.758256] E [rdma.c:4428:tcp_connect_finish] test_volume-client-1: tcp connect to  failed (Connection refused)
[2011-01-10 19:47:08.768299] E [rdma.c:4428:tcp_connect_finish] test_volume-client-1: tcp connect to  failed (Connection refused)
[2011-01-10 19:47:11.778308] E [rdma.c:4428:tcp_connect_finish] test_volume-client-1: tcp connect to  failed (Connection refused)
[2011-01-10 19:47:14.788356] E [rdma.c:4428:tcp_connect_finish] test_volume-client-1: tcp connect to  failed (Connection refused)
[2011-01-10 19:47:17.798381] E [rdma.c:4428:tcp_connect_finish] test_volume-client-1: tcp connect to  failed (Connection refused)
[2011-01-10 19:47:20.808413] E [rdma.c:4428:tcp_connect_finish] test_volume-client-1: tcp connect to  failed (Connection refused)

But there are no restrictions in the config. Everything is allow *. So my questions are, can anyone else tell me what kind of basic file copy performance they see using IB? And what can I do to troubleshoot?

Thanks List and Devs, 

Chris



More information about the Gluster-users mailing list