Hello The below WARNING was triggered on the host side after a connection on 5.17.0-rc3, here is the full log, let me if you need any tests for it, thanks HW: # lspci | grep -i mel 04:00.0 Infiniband controller: Mellanox Technologies MT27700 Family [ConnectX-4] 04:00.1 Infiniband controller: Mellanox Technologies MT27700 Family [ConnectX-4] dmesg: [ 507.229417] ------------[ cut here ]------------ [ 507.235233] DMA-API: mlx5_core 0000:04:00.0: cacheline tracking EEXIST, overlapping mappings aren't supported [ 507.246475] WARNING: CPU: 10 PID: 855 at kernel/dma/debug.c:571 add_dma_entry+0x36a/0x4c0 [ 507.255751] Modules linked in: nvme_rdma nvme_fabrics nvme_core 8021q garp mrp bonding bridge stp llc rfkill rpcrdma sunrpc rdma_ucm ib_srpt ib_isert iscsi_target_mod target_core_mod ib_iser libiscsi ib_umad scsi_transport_iscsi rdma_cm ib_ipoib iw_cm ib_cm intel_rapl_msr intel_rapl_common sb_edac x86_pkg_temp_thermal intel_powerclamp coretemp kvm_intel kvm mlx5_ib iTCO_wdt iTCO_vendor_support irqbypass ib_uverbs crct10dif_pclmul crc32_pclmul ib_core ghash_clmulni_intel rapl intel_cstate intel_uncore pcspkr lpc_ich mei_me mxm_wmi ipmi_ssif mei ipmi_si ipmi_devintf ipmi_msghandler acpi_power_meter xfs libcrc32c mgag200 i2c_algo_bit drm_shmem_helper sd_mod drm_kms_helper t10_pi sg syscopyarea sysfillrect sysimgblt fb_sys_fops mlx5_core drm ahci mlxfw libahci crc32c_intel pci_hyperv_intf tls tg3 libata psample wmi dm_mirror dm_region_hash dm_log dm_mod [ 507.341004] CPU: 10 PID: 855 Comm: kworker/u80:3 Tainted: G S 5.17.0-rc3.rc3.debug #3 [ 507.351351] Hardware name: Dell Inc. PowerEdge R430/03XKDV, BIOS 2.9.1 12/07/2018 [ 507.361252] Workqueue: ib_addr process_one_req [ib_core] [ 507.368756] RIP: 0010:add_dma_entry+0x36a/0x4c0 [ 507.375291] Code: 01 00 00 4c 8b 6d 50 4d 85 ed 0f 84 9d 00 00 00 48 89 ef e8 18 e5 0e 01 48 89 c6 4c 89 ea 48 c7 c7 80 a0 8c 8d e8 a6 d6 da 01 <0f> 0b 48 85 db 0f 85 20 c9 dc 01 8b 05 75 56 24 04 85 c0 0f 85 c2 [ 507.399392] RSP: 0018:ffff8890b6aaf7a0 EFLAGS: 00010282 [ 507.406756] RAX: 0000000000000000 RBX: ffff8885d1d6e600 RCX: 0000000000000000 [ 507.416302] RDX: 0000000000000001 RSI: ffffffff8da92ac0 RDI: ffffffff90ff9880 [ 507.425860] RBP: ffff8882580640d0 R08: ffffed11854bfb39 R09: ffffed11854bfb39 [ 507.435450] R10: ffff888c2a5fd9c7 R11: ffffed11854bfb38 R12: 1ffff11216d55ef6 [ 507.445054] R13: ffff889086697b20 R14: 00000000ffffffef R15: 0000000000000202 [ 507.454683] FS: 0000000000000000(0000) GS:ffff888c2a400000(0000) knlGS:0000000000000000 [ 507.465423] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 [ 507.473553] CR2: 00007f080b8fe000 CR3: 00000017fd62c001 CR4: 00000000001706e0 [ 507.483282] Call Trace: [ 507.487748] <TASK> [ 507.491829] ? dma_debug_init+0x130/0x130 [ 507.498064] ? debug_dma_map_page+0x278/0x320 [ 507.504722] dma_map_page_attrs+0x3bb/0xa90 [ 507.511191] ? dma_need_sync+0x100/0x100 [ 507.517369] ? kmem_cache_alloc_trace+0x13b/0x220 [ 507.524457] nvme_rdma_create_queue_ib+0x8a6/0x12c0 [nvme_rdma] [ 507.532949] ? nvme_rdma_reconnect_ctrl_work+0x60/0x60 [nvme_rdma] [ 507.541744] ? rdma_restrack_add+0xfb/0x4a0 [ib_core] [ 507.549302] ? rcu_read_lock_sched_held+0xaf/0xe0 [ 507.556469] nvme_rdma_cm_handler+0x71e/0xa20 [nvme_rdma] [ 507.564436] ? nvme_rdma_create_queue_ib+0x12c0/0x12c0 [nvme_rdma] [ 507.573293] ? lock_release+0x42f/0xc90 [ 507.579540] ? lock_downgrade+0x6b0/0x6b0 [ 507.585978] ? lock_is_held_type+0xd9/0x130 [ 507.592632] ? rcu_read_lock_sched_held+0xaf/0xe0 [ 507.599879] ? rcu_read_lock_bh_held+0xc0/0xc0 [ 507.606842] ? lock_is_held_type+0xd9/0x130 [ 507.613510] cma_cm_event_handler+0xf2/0x530 [rdma_cm] [ 507.621292] addr_handler+0x189/0x2a0 [rdma_cm] [ 507.628405] ? cma_work_handler+0x220/0x220 [rdma_cm] [ 507.636179] ? cma_work_handler+0x220/0x220 [rdma_cm] [ 507.643834] process_one_req+0xdd/0x590 [ib_core] [ 507.651094] process_one_work+0x97e/0x1700 [ 507.657607] ? pwq_dec_nr_in_flight+0x270/0x270 [ 507.664571] worker_thread+0x87/0xb30 [ 507.670501] ? process_one_work+0x1700/0x1700 [ 507.677145] kthread+0x29d/0x340 [ 507.682465] ? kthread_complete_and_exit+0x20/0x20 [ 507.689503] ret_from_fork+0x22/0x30 [ 507.695143] </TASK> [ 507.699133] irq event stamp: 473271 [ 507.704543] hardirqs last enabled at (473281): [<ffffffff8b39a9b2>] __up_console_sem+0x52/0x60 [ 507.715840] hardirqs last disabled at (473290): [<ffffffff8b39a997>] __up_console_sem+0x37/0x60 [ 507.727102] softirqs last enabled at (473218): [<ffffffff8d60064a>] __do_softirq+0x64a/0xa4c [ 507.738185] softirqs last disabled at (473309): [<ffffffff8b21561e>] irq_exit_rcu+0x1ce/0x240 [ 507.749206] ---[ end trace 0000000000000000 ]--- [ 507.755780] DMA-API: Mapped at: [ 507.760660] debug_dma_map_page+0x86/0x320 [ 507.766603] dma_map_page_attrs+0x3bb/0xa90 [ 507.772628] nvme_rdma_create_queue_ib+0x8a6/0x12c0 [nvme_rdma] [ 507.780613] nvme_rdma_cm_handler+0x71e/0xa20 [nvme_rdma] [ 507.788009] cma_cm_event_handler+0xf2/0x530 [rdma_cm] [ 508.648545] mlx5_ib0.8008: multicast join failed for ff12:401b:8008:0000:0000:0000:ffff:ffff, status -22 [ 509.655833] nvme nvme0: creating 40 I/O queues. [ 509.672625] mlx5_ib0.8016: multicast join failed for ff12:401b:8016:0000:0000:0000:ffff:ffff, status -22 [ 511.267023] DMA-API: dma_debug_entry pool grown to 2162688 (3300%) [ 520.572519] DMA-API: dma_debug_entry pool grown to 2228224 (3400%) [ 523.451478] nvme nvme0: mapped 40/0/0 default/read/poll queues. [ 523.631836] nvme nvme0: new ctrl: NQN "testnqn", addr 172.31.0.202:4420 -- Best Regards, Yi Zhang