Duplicate multipathd process

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi ALL.

I'm having some problems and could not find solution with technical support.

My server is two multipathd running processes. I noticed that when there is a failure to communicate with the SAN device looks like this:

volclu2 (36000144000000010700aa26e0e57b5b9) dm-36 EMC Invista
size = 1.0G features = '0' hwhandler = '0' wp = rw
`- + - Policy = 'queue-length 0' prio = 0 status = enabled
| - 0: 0: 0: 20 65:64 sdu failed running faulty
| - 0: 0: 1: 20 sdaq 66: 160 failed running faulty
| - 1: 0: 0: 20 sddf 70: 208 failed running faulty
`- 1: 0: 1: 20 sdeb 128: 48 failed running faulty

And then the communication is re-established looks like this:

volclu2 (36000144000000010700aa26e0e57b5b9) dm-36 EMC Invista
size = 1.0G features = '0' hwhandler = '0' wp = rw
`- + - Policy = 'queue-length 0' prio = 1 Status = enabled
| - 0: 0: 0: 20 65:64 sdu failed running ready
| - 0: 0: 1: 20 sdaq 66: 160 failed running ready
| - 1: 0: 0: 20 sddf 70: 208 failed running ready
`- 1: 0: 1: 20 sdeb 128: 48  failed running ready

The failback is not being automatic. The status of the device is "enable" rather than "active".

The status only changes to "active" when I run "multipath -v3".

- What can cause the creation of two daemon multipathd?

Regards,
    Juliano.

--
dm-devel mailing list
dm-devel@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/dm-devel

[Index of Archives]     [DM Crypt]     [Fedora Desktop]     [ATA RAID]     [Fedora Marketing]     [Fedora Packaging]     [Fedora SELinux]     [Yosemite Discussion]     [KDE Users]     [Fedora Docs]

  Powered by Linux