: Multipath Setup for NVMe over Fabric

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi All, 

I am working on NVMe over fabric and want to experiment the Multipathing support for the same.

Setup Info:
RHEL 7.2 with Kernel 4.9.3


[root@localhost ~]# nvme list
Node             SN                   Model                                           Namespace Usage                      Format           FW Rev
---------------- -------------------- ---------------------------------------- --------- -------------------------- ---------------- --------
/dev/nvme0n1     30501b622ed15184     Linux                                    10        268.44  GB / 268.44  GB    512   B +  0 B   4.9.3
/dev/nvme1n1     ef730272d9be107c      Linux                                    10        268.44  GB / 268.44  GB    512   B +  0 B   4.9.3


[root@localhost ~]# ps ax | grep multipath
1272 ?        SLl    0:00 /sbin/multipathd


I have connected my Initiator to both the ports of Ethernet Adapter(Target) to get 2 IO Paths, from the above data "/dev/nvme0n1" is path 1 and "/dev/nvme1n1" is path 2 for the same namespace.

Note: I am using Null Block device on the Target Side.

But still the multipath is showing an error ie no path to Host for All the NVMe Drives mapped on the Initiator. Does multipathd supports NVMe over Fabric ??
Or what I am missing from configuration side ??

Thanks in advance!!


BR~
Ankur





--
dm-devel mailing list
dm-devel@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/dm-devel

[Index of Archives]     [DM Crypt]     [Fedora Desktop]     [ATA RAID]     [Fedora Marketing]     [Fedora Packaging]     [Fedora SELinux]     [Yosemite Discussion]     [KDE Users]     [Fedora Docs]

  Powered by Linux