The backtrace you have provided here suggests that the issue could be with mellanox driver, though the question still valid to users of the IPoIB. Regards Rafi KC On 10/23/2017 09:29 PM, lejeczek wrote: > hi people > > I wonder if anybody experience any problems with vols in replica mode > that run across IPoIB links and libvirt stores qcow image on such a > volume? > > I wonder if maybe devel could confirm it should just work, and then > hardware/Infiniband I should blame. > > I have a direct IPoIB link between two hosts, gluster replica volume, > libvirt store disk images there. > > I start a guest on hostA and I get below on hostB(which is IB subnet > manager): > > [Mon Oct 23 16:43:32 2017] Workqueue: ipoib_wq ipoib_cm_tx_start > [ib_ipoib] > [Mon Oct 23 16:43:32 2017] 0000000000008010 00000000553c90b1 > ffff880c1c6eb818 ffffffff816a3db1 > [Mon Oct 23 16:43:32 2017] ffff880c1c6eb8a8 ffffffff81188810 > 0000000000000000 ffff88042ffdb000 > [Mon Oct 23 16:43:32 2017] 0000000000000004 0000000000008010 > ffff880c1c6eb8a8 00000000553c90b1 > [Mon Oct 23 16:43:32 2017] Call Trace: > [Mon Oct 23 16:43:32 2017] [<ffffffff816a3db1>] dump_stack+0x19/0x1b > [Mon Oct 23 16:43:32 2017] [<ffffffff81188810>] > warn_alloc_failed+0x110/0x180 > [Mon Oct 23 16:43:32 2017] [<ffffffff8169fd8a>] > __alloc_pages_slowpath+0x6b6/0x724 > [Mon Oct 23 16:43:32 2017] [<ffffffff8118cd85>] > __alloc_pages_nodemask+0x405/0x420 > [Mon Oct 23 16:43:32 2017] [<ffffffff81030f8f>] > dma_generic_alloc_coherent+0x8f/0x140 > [Mon Oct 23 16:43:32 2017] [<ffffffff81065c0d>] > gart_alloc_coherent+0x2d/0x40 > [Mon Oct 23 16:43:32 2017] [<ffffffffc012e4d3>] > mlx4_buf_direct_alloc.isra.6+0xd3/0x1a0 [mlx4_core] > [Mon Oct 23 16:43:32 2017] [<ffffffffc012e76b>] > mlx4_buf_alloc+0x1cb/0x240 [mlx4_core] > [Mon Oct 23 16:43:32 2017] [<ffffffffc04dd85e>] > create_qp_common.isra.31+0x62e/0x10d0 [mlx4_ib] > [Mon Oct 23 16:43:32 2017] [<ffffffffc04de44e>] > mlx4_ib_create_qp+0x14e/0x480 [mlx4_ib] > [Mon Oct 23 16:43:32 2017] [<ffffffffc06df20c>] ? > ipoib_cm_tx_init+0x5c/0x400 [ib_ipoib] > [Mon Oct 23 16:43:32 2017] [<ffffffffc0639c3a>] > ib_create_qp+0x7a/0x2f0 [ib_core] > [Mon Oct 23 16:43:32 2017] [<ffffffffc06df2b3>] > ipoib_cm_tx_init+0x103/0x400 [ib_ipoib] > [Mon Oct 23 16:43:32 2017] [<ffffffffc06e1608>] > ipoib_cm_tx_start+0x268/0x3f0 [ib_ipoib] > [Mon Oct 23 16:43:32 2017] [<ffffffff810a881a>] > process_one_work+0x17a/0x440 > [Mon Oct 23 16:43:32 2017] [<ffffffff810a94e6>] > worker_thread+0x126/0x3c0 > [Mon Oct 23 16:43:32 2017] [<ffffffff810a93c0>] ? > manage_workers.isra.24+0x2a0/0x2a0 > [Mon Oct 23 16:43:32 2017] [<ffffffff810b098f>] kthread+0xcf/0xe0 > [Mon Oct 23 16:43:32 2017] [<ffffffff810b08c0>] ? > insert_kthread_work+0x40/0x40 > [Mon Oct 23 16:43:32 2017] [<ffffffff816b4f58>] ret_from_fork+0x58/0x90 > [Mon Oct 23 16:43:32 2017] [<ffffffff810b08c0>] ? > insert_kthread_work+0x40/0x40 > [Mon Oct 23 16:43:32 2017] Mem-Info: > [Mon Oct 23 16:43:32 2017] active_anon:2389656 inactive_anon:17792 > isolated_anon:0 > active_file:14294829 inactive_file:14609973 isolated_file:0 > unevictable:24185 dirty:11846 writeback:9907 unstable:0 > slab_reclaimable:1024309 slab_unreclaimable:127961 > mapped:74895 shmem:28096 pagetables:30088 bounce:0 > free:142329 free_pcp:249 free_cma:0 > [Mon Oct 23 16:43:32 2017] Node 0 DMA free:15320kB min:24kB low:28kB > high:36kB active_anon:0kB inactive_anon:0kB active_file:0kB > inactive_file:0kB unevictable:0kB isolated(anon):0kB > isolated(file):0kB present:15984kB managed:15900kB mlocked:0kB > dirty:0kB writeback:0kB mapped:0kB shmem:0kB slab_reclaimable:0kB > slab_unreclaimable:64kB kernel_stack:0kB pagetables:0kB unstable:0kB > bounce:0kB free_pcp:0kB local_pcp:0kB free_cma:0kB writeback_tmp:0kB > pages_scanned:0 all_unreclaimable? yes > > > To clarify - other volumes which use that IPoIB link do not seem to > case that, or any other problem. > _______________________________________________ > Gluster-users mailing list > Gluster-users@xxxxxxxxxxx > http://lists.gluster.org/mailman/listinfo/gluster-users _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://lists.gluster.org/mailman/listinfo/gluster-users