[Gluster-users] problems running a vol over IPoIB, and qemu off it?

lejeczek peljasz at yahoo.co.uk
Mon Oct 23 15:59:05 UTC 2017


hi people

I wonder if anybody experience any problems with vols in 
replica mode that run across IPoIB links and libvirt stores 
qcow image on such a volume?

I wonder if maybe devel could confirm it should just work, 
and then hardware/Infiniband I should blame.

I have a direct IPoIB link between two hosts, gluster 
replica volume, libvirt store disk images there.

I start a guest on hostA and I get below on hostB(which is 
IB subnet manager):

[Mon Oct 23 16:43:32 2017] Workqueue: ipoib_wq 
ipoib_cm_tx_start [ib_ipoib]
[Mon Oct 23 16:43:32 2017]  0000000000008010 
00000000553c90b1 ffff880c1c6eb818 ffffffff816a3db1
[Mon Oct 23 16:43:32 2017]  ffff880c1c6eb8a8 
ffffffff81188810 0000000000000000 ffff88042ffdb000
[Mon Oct 23 16:43:32 2017]  0000000000000004 
0000000000008010 ffff880c1c6eb8a8 00000000553c90b1
[Mon Oct 23 16:43:32 2017] Call Trace:
[Mon Oct 23 16:43:32 2017]  [<ffffffff816a3db1>] 
dump_stack+0x19/0x1b
[Mon Oct 23 16:43:32 2017]  [<ffffffff81188810>] 
warn_alloc_failed+0x110/0x180
[Mon Oct 23 16:43:32 2017]  [<ffffffff8169fd8a>] 
__alloc_pages_slowpath+0x6b6/0x724
[Mon Oct 23 16:43:32 2017]  [<ffffffff8118cd85>] 
__alloc_pages_nodemask+0x405/0x420
[Mon Oct 23 16:43:32 2017]  [<ffffffff81030f8f>] 
dma_generic_alloc_coherent+0x8f/0x140
[Mon Oct 23 16:43:32 2017]  [<ffffffff81065c0d>] 
gart_alloc_coherent+0x2d/0x40
[Mon Oct 23 16:43:32 2017]  [<ffffffffc012e4d3>] 
mlx4_buf_direct_alloc.isra.6+0xd3/0x1a0 [mlx4_core]
[Mon Oct 23 16:43:32 2017]  [<ffffffffc012e76b>] 
mlx4_buf_alloc+0x1cb/0x240 [mlx4_core]
[Mon Oct 23 16:43:32 2017]  [<ffffffffc04dd85e>] 
create_qp_common.isra.31+0x62e/0x10d0 [mlx4_ib]
[Mon Oct 23 16:43:32 2017]  [<ffffffffc04de44e>] 
mlx4_ib_create_qp+0x14e/0x480 [mlx4_ib]
[Mon Oct 23 16:43:32 2017]  [<ffffffffc06df20c>] ? 
ipoib_cm_tx_init+0x5c/0x400 [ib_ipoib]
[Mon Oct 23 16:43:32 2017]  [<ffffffffc0639c3a>] 
ib_create_qp+0x7a/0x2f0 [ib_core]
[Mon Oct 23 16:43:32 2017]  [<ffffffffc06df2b3>] 
ipoib_cm_tx_init+0x103/0x400 [ib_ipoib]
[Mon Oct 23 16:43:32 2017]  [<ffffffffc06e1608>] 
ipoib_cm_tx_start+0x268/0x3f0 [ib_ipoib]
[Mon Oct 23 16:43:32 2017]  [<ffffffff810a881a>] 
process_one_work+0x17a/0x440
[Mon Oct 23 16:43:32 2017]  [<ffffffff810a94e6>] 
worker_thread+0x126/0x3c0
[Mon Oct 23 16:43:32 2017]  [<ffffffff810a93c0>] ? 
manage_workers.isra.24+0x2a0/0x2a0
[Mon Oct 23 16:43:32 2017]  [<ffffffff810b098f>] 
kthread+0xcf/0xe0
[Mon Oct 23 16:43:32 2017]  [<ffffffff810b08c0>] ? 
insert_kthread_work+0x40/0x40
[Mon Oct 23 16:43:32 2017]  [<ffffffff816b4f58>] 
ret_from_fork+0x58/0x90
[Mon Oct 23 16:43:32 2017]  [<ffffffff810b08c0>] ? 
insert_kthread_work+0x40/0x40
[Mon Oct 23 16:43:32 2017] Mem-Info:
[Mon Oct 23 16:43:32 2017] active_anon:2389656 
inactive_anon:17792 isolated_anon:0
  active_file:14294829 inactive_file:14609973 isolated_file:0
  unevictable:24185 dirty:11846 writeback:9907 unstable:0
  slab_reclaimable:1024309 slab_unreclaimable:127961
  mapped:74895 shmem:28096 pagetables:30088 bounce:0
  free:142329 free_pcp:249 free_cma:0
[Mon Oct 23 16:43:32 2017] Node 0 DMA free:15320kB min:24kB 
low:28kB high:36kB active_anon:0kB inactive_anon:0kB 
active_file:0kB inactive_file:0kB unevictable:0kB 
isolated(anon):0kB isolated(file):0kB present:15984kB 
managed:15900kB mlocked:0kB dirty:0kB writeback:0kB 
mapped:0kB shmem:0kB slab_reclaimable:0kB 
slab_unreclaimable:64kB kernel_stack:0kB pagetables:0kB 
unstable:0kB bounce:0kB free_pcp:0kB local_pcp:0kB 
free_cma:0kB writeback_tmp:0kB pages_scanned:0 
all_unreclaimable? yes


To clarify - other volumes which use that IPoIB link do not 
seem to case that, or any other problem.


More information about the Gluster-users mailing list