Re: [RFC v4 10/19] lpfc: Remove host_lock unlock() + lock() from lpfc_queuecommand()

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 09/27/2010 09:06 PM, Nicholas A. Bellinger wrote:
From: Nicholas Bellinger<nab@xxxxxxxxxxxxxxx>

This patch removes the now legacy host_lock unlock() + lock() optimization
from lpfc_scsi.c:lpfc_queuecommand().  This also includes setting the
SHT->unlocked_qcmd=1 for host_lock less lpfc lpfc_queuecommand() operation.

Signed-off-by: Nicholas A. Bellinger<nab@xxxxxxxxxxxxxxx>
---
  drivers/scsi/lpfc/lpfc_scsi.c |    4 ++--
  1 files changed, 2 insertions(+), 2 deletions(-)

diff --git a/drivers/scsi/lpfc/lpfc_scsi.c b/drivers/scsi/lpfc/lpfc_scsi.c
index 2e51aa6..69fe31e 100644
--- a/drivers/scsi/lpfc/lpfc_scsi.c
+++ b/drivers/scsi/lpfc/lpfc_scsi.c
@@ -3023,11 +3023,9 @@ lpfc_queuecommand(struct scsi_cmnd *cmnd, void (*done) (struct scsi_cmnd *))
  		goto out_host_busy_free_buf;
  	}
  	if (phba->cfg_poll&  ENABLE_FCP_RING_POLLING) {
-		spin_unlock(shost->host_lock);
  		lpfc_sli_handle_fast_ring_event(phba,
  			&phba->sli.ring[LPFC_FCP_RING], HA_R0RE_REQ);

-		spin_lock(shost->host_lock);
  		if (phba->cfg_poll&  DISABLE_FCP_RING_INT)
  			lpfc_poll_rearm_timer(phba);
  	}
@@ -3723,6 +3721,7 @@ struct scsi_host_template lpfc_template = {
  	.slave_destroy		= lpfc_slave_destroy,
  	.scan_finished		= lpfc_scan_finished,
  	.this_id		= -1,
+	.unlocked_qcmd		= 1,
  	.sg_tablesize		= LPFC_DEFAULT_SG_SEG_CNT,
  	.cmd_per_lun		= LPFC_CMD_PER_LUN,
  	.use_clustering		= ENABLE_CLUSTERING,
@@ -3746,6 +3745,7 @@ struct scsi_host_template lpfc_vport_template = {
  	.slave_destroy		= lpfc_slave_destroy,
  	.scan_finished		= lpfc_scan_finished,
  	.this_id		= -1,
+	.unlocked_qcmd		= 1,
  	.sg_tablesize		= LPFC_DEFAULT_SG_SEG_CNT,
  	.cmd_per_lun		= LPFC_CMD_PER_LUN,
  	.use_clustering		= ENABLE_CLUSTERING,

The FC class sets the rport state and bits with the host lock held. Drivers were then calling fc_remote_port_chkready from the queuecommand with the host lock held. If we remove the host lock from queuecommand is it possible that the on one proc the fc class calls fc_remote_port_add to re-add a rport, this sets the rport state to online, it unblocks the devices, but then on some other processor we start calling queuecommand and see that the rport is not online (maybe blocked with FC_RPORT_FAST_FAIL_TIMEDOUT set) and so we end up failing the IO?
--
To unsubscribe from this list: send the line "unsubscribe linux-scsi" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[Index of Archives]     [SCSI Target Devel]     [Linux SCSI Target Infrastructure]     [Kernel Newbies]     [IDE]     [Security]     [Git]     [Netfilter]     [Bugtraq]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux RAID]     [Linux ATA RAID]     [Linux IIO]     [Samba]     [Device Mapper]
  Powered by Linux