Re: tcmu-runner: "Acquired exclusive lock" every 21s

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 





Am 05.08.19 um 18:31 schrieb Mike Christie:
On 08/05/2019 05:58 AM, Matthias Leopold wrote:
Hi,

I'm still testing my 2 node (dedicated) iSCSI gateway with ceph 12.2.12
before I dare to put it into production. I installed latest tcmu-runner
release (1.5.1) and (like before) I'm seeing that both nodes switch
exclusive locks for the disk images every 21 seconds. tcmu-runner logs
look like this:

2019-08-05 12:53:04.184 13742 [WARN] tcmu_notify_lock_lost:222
rbd/iscsi.test03: Async lock drop. Old state 1
2019-08-05 12:53:04.714 13742 [WARN] tcmu_rbd_lock:762 rbd/iscsi.test03:
Acquired exclusive lock.
2019-08-05 12:53:25.186 13742 [WARN] tcmu_notify_lock_lost:222
rbd/iscsi.test03: Async lock drop. Old state 1
2019-08-05 12:53:25.773 13742 [WARN] tcmu_rbd_lock:762 rbd/iscsi.test03:
Acquired exclusive lock.

Old state can sometimes be 0 or 2.
Is this expected behaviour?

What initiator OS are you using?


I'm using CentOS 7 initiators and I somehow missed to configure multipathd on them correctly (device { vendor "LIO.ORG" ... }). After fixing that the above problem disappeared and the output of "multipath -ll" finally looks correct. Thanks for pointing me to this.

Nevertheless there's now another problem visible in the logs. As soon as an initiator logs in tcmu-runner on the gateway node that doesn't own the image being accessed logs

[ERROR] tcmu_rbd_has_lock:516 rbd/iscsi.test02: Could not check lock ownership. Error: Cannot send after transport endpoint shutdown.

This disappears after the osd blacklist entries for the node expire (visible with "ceph osd blacklist ls"). I haven't yet understood how this is supposed to work, right now I restarted from scratch (logged out, waited till all blacklist entries disappeared, logged in) and I'm again seeing several blacklist entries for both gateway nodes (and the above error message in tcmu-runner.log). This doesn't seem to interfere with the iSCSI service, but I want this explained/resolved before I can start using the gateways.

To make things easier I'm using "cluster_client_name = client.admin" in iscsi-gateway.cfg again.

thanks
Matthias
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux