mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-28 15:13:55 +08:00
43fac4d97a
Reverted back a change in qla*_intr_handler code that caused an increase in cpu cycles by allowing interrupts to occur while the instance hardware lock was being held. Fix by taking the lock in irqsave mode. Reported-and-tested-by: Douglas W. Styner <douglas.w.styner@intel.com> Signed-off-by: Anirban Chakraborty <anirban.chakraborty@qlogic.com> Signed-off-by: James Bottomley <James.Bottomley@HansenPartnership.com>
2088 lines
54 KiB
C
2088 lines
54 KiB
C
/*
|
|
* QLogic Fibre Channel HBA Driver
|
|
* Copyright (c) 2003-2008 QLogic Corporation
|
|
*
|
|
* See LICENSE.qla2xxx for copyright and licensing details.
|
|
*/
|
|
#include "qla_def.h"
|
|
|
|
#include <linux/delay.h>
|
|
#include <scsi/scsi_tcq.h>
|
|
|
|
static void qla2x00_mbx_completion(scsi_qla_host_t *, uint16_t);
|
|
static void qla2x00_process_completed_request(struct scsi_qla_host *,
|
|
struct req_que *, uint32_t);
|
|
static void qla2x00_status_entry(scsi_qla_host_t *, struct rsp_que *, void *);
|
|
static void qla2x00_status_cont_entry(struct rsp_que *, sts_cont_entry_t *);
|
|
static void qla2x00_error_entry(scsi_qla_host_t *, struct rsp_que *,
|
|
sts_entry_t *);
|
|
|
|
/**
|
|
* qla2100_intr_handler() - Process interrupts for the ISP2100 and ISP2200.
|
|
* @irq:
|
|
* @dev_id: SCSI driver HA context
|
|
*
|
|
* Called by system whenever the host adapter generates an interrupt.
|
|
*
|
|
* Returns handled flag.
|
|
*/
|
|
irqreturn_t
|
|
qla2100_intr_handler(int irq, void *dev_id)
|
|
{
|
|
scsi_qla_host_t *vha;
|
|
struct qla_hw_data *ha;
|
|
struct device_reg_2xxx __iomem *reg;
|
|
int status;
|
|
unsigned long iter;
|
|
uint16_t hccr;
|
|
uint16_t mb[4];
|
|
struct rsp_que *rsp;
|
|
unsigned long flags;
|
|
|
|
rsp = (struct rsp_que *) dev_id;
|
|
if (!rsp) {
|
|
printk(KERN_INFO
|
|
"%s(): NULL response queue pointer\n", __func__);
|
|
return (IRQ_NONE);
|
|
}
|
|
|
|
ha = rsp->hw;
|
|
reg = &ha->iobase->isp;
|
|
status = 0;
|
|
|
|
spin_lock_irqsave(&ha->hardware_lock, flags);
|
|
vha = pci_get_drvdata(ha->pdev);
|
|
for (iter = 50; iter--; ) {
|
|
hccr = RD_REG_WORD(®->hccr);
|
|
if (hccr & HCCR_RISC_PAUSE) {
|
|
if (pci_channel_offline(ha->pdev))
|
|
break;
|
|
|
|
/*
|
|
* Issue a "HARD" reset in order for the RISC interrupt
|
|
* bit to be cleared. Schedule a big hammmer to get
|
|
* out of the RISC PAUSED state.
|
|
*/
|
|
WRT_REG_WORD(®->hccr, HCCR_RESET_RISC);
|
|
RD_REG_WORD(®->hccr);
|
|
|
|
ha->isp_ops->fw_dump(vha, 1);
|
|
set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
|
|
break;
|
|
} else if ((RD_REG_WORD(®->istatus) & ISR_RISC_INT) == 0)
|
|
break;
|
|
|
|
if (RD_REG_WORD(®->semaphore) & BIT_0) {
|
|
WRT_REG_WORD(®->hccr, HCCR_CLR_RISC_INT);
|
|
RD_REG_WORD(®->hccr);
|
|
|
|
/* Get mailbox data. */
|
|
mb[0] = RD_MAILBOX_REG(ha, reg, 0);
|
|
if (mb[0] > 0x3fff && mb[0] < 0x8000) {
|
|
qla2x00_mbx_completion(vha, mb[0]);
|
|
status |= MBX_INTERRUPT;
|
|
} else if (mb[0] > 0x7fff && mb[0] < 0xc000) {
|
|
mb[1] = RD_MAILBOX_REG(ha, reg, 1);
|
|
mb[2] = RD_MAILBOX_REG(ha, reg, 2);
|
|
mb[3] = RD_MAILBOX_REG(ha, reg, 3);
|
|
qla2x00_async_event(vha, rsp, mb);
|
|
} else {
|
|
/*EMPTY*/
|
|
DEBUG2(printk("scsi(%ld): Unrecognized "
|
|
"interrupt type (%d).\n",
|
|
vha->host_no, mb[0]));
|
|
}
|
|
/* Release mailbox registers. */
|
|
WRT_REG_WORD(®->semaphore, 0);
|
|
RD_REG_WORD(®->semaphore);
|
|
} else {
|
|
qla2x00_process_response_queue(rsp);
|
|
|
|
WRT_REG_WORD(®->hccr, HCCR_CLR_RISC_INT);
|
|
RD_REG_WORD(®->hccr);
|
|
}
|
|
}
|
|
spin_unlock_irqrestore(&ha->hardware_lock, flags);
|
|
|
|
if (test_bit(MBX_INTR_WAIT, &ha->mbx_cmd_flags) &&
|
|
(status & MBX_INTERRUPT) && ha->flags.mbox_int) {
|
|
set_bit(MBX_INTERRUPT, &ha->mbx_cmd_flags);
|
|
complete(&ha->mbx_intr_comp);
|
|
}
|
|
|
|
return (IRQ_HANDLED);
|
|
}
|
|
|
|
/**
|
|
* qla2300_intr_handler() - Process interrupts for the ISP23xx and ISP63xx.
|
|
* @irq:
|
|
* @dev_id: SCSI driver HA context
|
|
*
|
|
* Called by system whenever the host adapter generates an interrupt.
|
|
*
|
|
* Returns handled flag.
|
|
*/
|
|
irqreturn_t
|
|
qla2300_intr_handler(int irq, void *dev_id)
|
|
{
|
|
scsi_qla_host_t *vha;
|
|
struct device_reg_2xxx __iomem *reg;
|
|
int status;
|
|
unsigned long iter;
|
|
uint32_t stat;
|
|
uint16_t hccr;
|
|
uint16_t mb[4];
|
|
struct rsp_que *rsp;
|
|
struct qla_hw_data *ha;
|
|
unsigned long flags;
|
|
|
|
rsp = (struct rsp_que *) dev_id;
|
|
if (!rsp) {
|
|
printk(KERN_INFO
|
|
"%s(): NULL response queue pointer\n", __func__);
|
|
return (IRQ_NONE);
|
|
}
|
|
|
|
ha = rsp->hw;
|
|
reg = &ha->iobase->isp;
|
|
status = 0;
|
|
|
|
spin_lock_irqsave(&ha->hardware_lock, flags);
|
|
vha = pci_get_drvdata(ha->pdev);
|
|
for (iter = 50; iter--; ) {
|
|
stat = RD_REG_DWORD(®->u.isp2300.host_status);
|
|
if (stat & HSR_RISC_PAUSED) {
|
|
if (pci_channel_offline(ha->pdev))
|
|
break;
|
|
|
|
hccr = RD_REG_WORD(®->hccr);
|
|
if (hccr & (BIT_15 | BIT_13 | BIT_11 | BIT_8))
|
|
qla_printk(KERN_INFO, ha, "Parity error -- "
|
|
"HCCR=%x, Dumping firmware!\n", hccr);
|
|
else
|
|
qla_printk(KERN_INFO, ha, "RISC paused -- "
|
|
"HCCR=%x, Dumping firmware!\n", hccr);
|
|
|
|
/*
|
|
* Issue a "HARD" reset in order for the RISC
|
|
* interrupt bit to be cleared. Schedule a big
|
|
* hammmer to get out of the RISC PAUSED state.
|
|
*/
|
|
WRT_REG_WORD(®->hccr, HCCR_RESET_RISC);
|
|
RD_REG_WORD(®->hccr);
|
|
|
|
ha->isp_ops->fw_dump(vha, 1);
|
|
set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
|
|
break;
|
|
} else if ((stat & HSR_RISC_INT) == 0)
|
|
break;
|
|
|
|
switch (stat & 0xff) {
|
|
case 0x1:
|
|
case 0x2:
|
|
case 0x10:
|
|
case 0x11:
|
|
qla2x00_mbx_completion(vha, MSW(stat));
|
|
status |= MBX_INTERRUPT;
|
|
|
|
/* Release mailbox registers. */
|
|
WRT_REG_WORD(®->semaphore, 0);
|
|
break;
|
|
case 0x12:
|
|
mb[0] = MSW(stat);
|
|
mb[1] = RD_MAILBOX_REG(ha, reg, 1);
|
|
mb[2] = RD_MAILBOX_REG(ha, reg, 2);
|
|
mb[3] = RD_MAILBOX_REG(ha, reg, 3);
|
|
qla2x00_async_event(vha, rsp, mb);
|
|
break;
|
|
case 0x13:
|
|
qla2x00_process_response_queue(rsp);
|
|
break;
|
|
case 0x15:
|
|
mb[0] = MBA_CMPLT_1_16BIT;
|
|
mb[1] = MSW(stat);
|
|
qla2x00_async_event(vha, rsp, mb);
|
|
break;
|
|
case 0x16:
|
|
mb[0] = MBA_SCSI_COMPLETION;
|
|
mb[1] = MSW(stat);
|
|
mb[2] = RD_MAILBOX_REG(ha, reg, 2);
|
|
qla2x00_async_event(vha, rsp, mb);
|
|
break;
|
|
default:
|
|
DEBUG2(printk("scsi(%ld): Unrecognized interrupt type "
|
|
"(%d).\n",
|
|
vha->host_no, stat & 0xff));
|
|
break;
|
|
}
|
|
WRT_REG_WORD(®->hccr, HCCR_CLR_RISC_INT);
|
|
RD_REG_WORD_RELAXED(®->hccr);
|
|
}
|
|
spin_unlock_irqrestore(&ha->hardware_lock, flags);
|
|
|
|
if (test_bit(MBX_INTR_WAIT, &ha->mbx_cmd_flags) &&
|
|
(status & MBX_INTERRUPT) && ha->flags.mbox_int) {
|
|
set_bit(MBX_INTERRUPT, &ha->mbx_cmd_flags);
|
|
complete(&ha->mbx_intr_comp);
|
|
}
|
|
|
|
return (IRQ_HANDLED);
|
|
}
|
|
|
|
/**
|
|
* qla2x00_mbx_completion() - Process mailbox command completions.
|
|
* @ha: SCSI driver HA context
|
|
* @mb0: Mailbox0 register
|
|
*/
|
|
static void
|
|
qla2x00_mbx_completion(scsi_qla_host_t *vha, uint16_t mb0)
|
|
{
|
|
uint16_t cnt;
|
|
uint16_t __iomem *wptr;
|
|
struct qla_hw_data *ha = vha->hw;
|
|
struct device_reg_2xxx __iomem *reg = &ha->iobase->isp;
|
|
|
|
/* Load return mailbox registers. */
|
|
ha->flags.mbox_int = 1;
|
|
ha->mailbox_out[0] = mb0;
|
|
wptr = (uint16_t __iomem *)MAILBOX_REG(ha, reg, 1);
|
|
|
|
for (cnt = 1; cnt < ha->mbx_count; cnt++) {
|
|
if (IS_QLA2200(ha) && cnt == 8)
|
|
wptr = (uint16_t __iomem *)MAILBOX_REG(ha, reg, 8);
|
|
if (cnt == 4 || cnt == 5)
|
|
ha->mailbox_out[cnt] = qla2x00_debounce_register(wptr);
|
|
else
|
|
ha->mailbox_out[cnt] = RD_REG_WORD(wptr);
|
|
|
|
wptr++;
|
|
}
|
|
|
|
if (ha->mcp) {
|
|
DEBUG3(printk("%s(%ld): Got mailbox completion. cmd=%x.\n",
|
|
__func__, vha->host_no, ha->mcp->mb[0]));
|
|
} else {
|
|
DEBUG2_3(printk("%s(%ld): MBX pointer ERROR!\n",
|
|
__func__, vha->host_no));
|
|
}
|
|
}
|
|
|
|
static void
|
|
qla81xx_idc_event(scsi_qla_host_t *vha, uint16_t aen, uint16_t descr)
|
|
{
|
|
static char *event[] =
|
|
{ "Complete", "Request Notification", "Time Extension" };
|
|
int rval;
|
|
struct device_reg_24xx __iomem *reg24 = &vha->hw->iobase->isp24;
|
|
uint16_t __iomem *wptr;
|
|
uint16_t cnt, timeout, mb[QLA_IDC_ACK_REGS];
|
|
|
|
/* Seed data -- mailbox1 -> mailbox7. */
|
|
wptr = (uint16_t __iomem *)®24->mailbox1;
|
|
for (cnt = 0; cnt < QLA_IDC_ACK_REGS; cnt++, wptr++)
|
|
mb[cnt] = RD_REG_WORD(wptr);
|
|
|
|
DEBUG2(printk("scsi(%ld): Inter-Driver Commucation %s -- "
|
|
"%04x %04x %04x %04x %04x %04x %04x.\n", vha->host_no,
|
|
event[aen & 0xff],
|
|
mb[0], mb[1], mb[2], mb[3], mb[4], mb[5], mb[6]));
|
|
|
|
/* Acknowledgement needed? [Notify && non-zero timeout]. */
|
|
timeout = (descr >> 8) & 0xf;
|
|
if (aen != MBA_IDC_NOTIFY || !timeout)
|
|
return;
|
|
|
|
DEBUG2(printk("scsi(%ld): Inter-Driver Commucation %s -- "
|
|
"ACK timeout=%d.\n", vha->host_no, event[aen & 0xff], timeout));
|
|
|
|
rval = qla2x00_post_idc_ack_work(vha, mb);
|
|
if (rval != QLA_SUCCESS)
|
|
qla_printk(KERN_WARNING, vha->hw,
|
|
"IDC failed to post ACK.\n");
|
|
}
|
|
|
|
/**
|
|
* qla2x00_async_event() - Process aynchronous events.
|
|
* @ha: SCSI driver HA context
|
|
* @mb: Mailbox registers (0 - 3)
|
|
*/
|
|
void
|
|
qla2x00_async_event(scsi_qla_host_t *vha, struct rsp_que *rsp, uint16_t *mb)
|
|
{
|
|
#define LS_UNKNOWN 2
|
|
static char *link_speeds[] = { "1", "2", "?", "4", "8", "10" };
|
|
char *link_speed;
|
|
uint16_t handle_cnt;
|
|
uint16_t cnt;
|
|
uint32_t handles[5];
|
|
struct qla_hw_data *ha = vha->hw;
|
|
struct device_reg_2xxx __iomem *reg = &ha->iobase->isp;
|
|
uint32_t rscn_entry, host_pid;
|
|
uint8_t rscn_queue_index;
|
|
unsigned long flags;
|
|
|
|
/* Setup to process RIO completion. */
|
|
handle_cnt = 0;
|
|
if (IS_QLA81XX(ha))
|
|
goto skip_rio;
|
|
switch (mb[0]) {
|
|
case MBA_SCSI_COMPLETION:
|
|
handles[0] = le32_to_cpu((uint32_t)((mb[2] << 16) | mb[1]));
|
|
handle_cnt = 1;
|
|
break;
|
|
case MBA_CMPLT_1_16BIT:
|
|
handles[0] = mb[1];
|
|
handle_cnt = 1;
|
|
mb[0] = MBA_SCSI_COMPLETION;
|
|
break;
|
|
case MBA_CMPLT_2_16BIT:
|
|
handles[0] = mb[1];
|
|
handles[1] = mb[2];
|
|
handle_cnt = 2;
|
|
mb[0] = MBA_SCSI_COMPLETION;
|
|
break;
|
|
case MBA_CMPLT_3_16BIT:
|
|
handles[0] = mb[1];
|
|
handles[1] = mb[2];
|
|
handles[2] = mb[3];
|
|
handle_cnt = 3;
|
|
mb[0] = MBA_SCSI_COMPLETION;
|
|
break;
|
|
case MBA_CMPLT_4_16BIT:
|
|
handles[0] = mb[1];
|
|
handles[1] = mb[2];
|
|
handles[2] = mb[3];
|
|
handles[3] = (uint32_t)RD_MAILBOX_REG(ha, reg, 6);
|
|
handle_cnt = 4;
|
|
mb[0] = MBA_SCSI_COMPLETION;
|
|
break;
|
|
case MBA_CMPLT_5_16BIT:
|
|
handles[0] = mb[1];
|
|
handles[1] = mb[2];
|
|
handles[2] = mb[3];
|
|
handles[3] = (uint32_t)RD_MAILBOX_REG(ha, reg, 6);
|
|
handles[4] = (uint32_t)RD_MAILBOX_REG(ha, reg, 7);
|
|
handle_cnt = 5;
|
|
mb[0] = MBA_SCSI_COMPLETION;
|
|
break;
|
|
case MBA_CMPLT_2_32BIT:
|
|
handles[0] = le32_to_cpu((uint32_t)((mb[2] << 16) | mb[1]));
|
|
handles[1] = le32_to_cpu(
|
|
((uint32_t)(RD_MAILBOX_REG(ha, reg, 7) << 16)) |
|
|
RD_MAILBOX_REG(ha, reg, 6));
|
|
handle_cnt = 2;
|
|
mb[0] = MBA_SCSI_COMPLETION;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
skip_rio:
|
|
switch (mb[0]) {
|
|
case MBA_SCSI_COMPLETION: /* Fast Post */
|
|
if (!vha->flags.online)
|
|
break;
|
|
|
|
for (cnt = 0; cnt < handle_cnt; cnt++)
|
|
qla2x00_process_completed_request(vha, rsp->req,
|
|
handles[cnt]);
|
|
break;
|
|
|
|
case MBA_RESET: /* Reset */
|
|
DEBUG2(printk("scsi(%ld): Asynchronous RESET.\n",
|
|
vha->host_no));
|
|
|
|
set_bit(RESET_MARKER_NEEDED, &vha->dpc_flags);
|
|
break;
|
|
|
|
case MBA_SYSTEM_ERR: /* System Error */
|
|
qla_printk(KERN_INFO, ha,
|
|
"ISP System Error - mbx1=%xh mbx2=%xh mbx3=%xh.\n",
|
|
mb[1], mb[2], mb[3]);
|
|
|
|
ha->isp_ops->fw_dump(vha, 1);
|
|
|
|
if (IS_FWI2_CAPABLE(ha)) {
|
|
if (mb[1] == 0 && mb[2] == 0) {
|
|
qla_printk(KERN_ERR, ha,
|
|
"Unrecoverable Hardware Error: adapter "
|
|
"marked OFFLINE!\n");
|
|
vha->flags.online = 0;
|
|
} else
|
|
set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
|
|
} else if (mb[1] == 0) {
|
|
qla_printk(KERN_INFO, ha,
|
|
"Unrecoverable Hardware Error: adapter marked "
|
|
"OFFLINE!\n");
|
|
vha->flags.online = 0;
|
|
} else
|
|
set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
|
|
break;
|
|
|
|
case MBA_REQ_TRANSFER_ERR: /* Request Transfer Error */
|
|
DEBUG2(printk("scsi(%ld): ISP Request Transfer Error.\n",
|
|
vha->host_no));
|
|
qla_printk(KERN_WARNING, ha, "ISP Request Transfer Error.\n");
|
|
|
|
set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
|
|
break;
|
|
|
|
case MBA_RSP_TRANSFER_ERR: /* Response Transfer Error */
|
|
DEBUG2(printk("scsi(%ld): ISP Response Transfer Error.\n",
|
|
vha->host_no));
|
|
qla_printk(KERN_WARNING, ha, "ISP Response Transfer Error.\n");
|
|
|
|
set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
|
|
break;
|
|
|
|
case MBA_WAKEUP_THRES: /* Request Queue Wake-up */
|
|
DEBUG2(printk("scsi(%ld): Asynchronous WAKEUP_THRES.\n",
|
|
vha->host_no));
|
|
break;
|
|
|
|
case MBA_LIP_OCCURRED: /* Loop Initialization Procedure */
|
|
DEBUG2(printk("scsi(%ld): LIP occurred (%x).\n", vha->host_no,
|
|
mb[1]));
|
|
qla_printk(KERN_INFO, ha, "LIP occurred (%x).\n", mb[1]);
|
|
|
|
if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
|
|
atomic_set(&vha->loop_state, LOOP_DOWN);
|
|
atomic_set(&vha->loop_down_timer, LOOP_DOWN_TIME);
|
|
qla2x00_mark_all_devices_lost(vha, 1);
|
|
}
|
|
|
|
if (vha->vp_idx) {
|
|
atomic_set(&vha->vp_state, VP_FAILED);
|
|
fc_vport_set_state(vha->fc_vport, FC_VPORT_FAILED);
|
|
}
|
|
|
|
set_bit(REGISTER_FC4_NEEDED, &vha->dpc_flags);
|
|
set_bit(REGISTER_FDMI_NEEDED, &vha->dpc_flags);
|
|
|
|
vha->flags.management_server_logged_in = 0;
|
|
qla2x00_post_aen_work(vha, FCH_EVT_LIP, mb[1]);
|
|
break;
|
|
|
|
case MBA_LOOP_UP: /* Loop Up Event */
|
|
if (IS_QLA2100(ha) || IS_QLA2200(ha)) {
|
|
link_speed = link_speeds[0];
|
|
ha->link_data_rate = PORT_SPEED_1GB;
|
|
} else {
|
|
link_speed = link_speeds[LS_UNKNOWN];
|
|
if (mb[1] < 5)
|
|
link_speed = link_speeds[mb[1]];
|
|
else if (mb[1] == 0x13)
|
|
link_speed = link_speeds[5];
|
|
ha->link_data_rate = mb[1];
|
|
}
|
|
|
|
DEBUG2(printk("scsi(%ld): Asynchronous LOOP UP (%s Gbps).\n",
|
|
vha->host_no, link_speed));
|
|
qla_printk(KERN_INFO, ha, "LOOP UP detected (%s Gbps).\n",
|
|
link_speed);
|
|
|
|
vha->flags.management_server_logged_in = 0;
|
|
qla2x00_post_aen_work(vha, FCH_EVT_LINKUP, ha->link_data_rate);
|
|
break;
|
|
|
|
case MBA_LOOP_DOWN: /* Loop Down Event */
|
|
DEBUG2(printk("scsi(%ld): Asynchronous LOOP DOWN "
|
|
"(%x %x %x).\n", vha->host_no, mb[1], mb[2], mb[3]));
|
|
qla_printk(KERN_INFO, ha, "LOOP DOWN detected (%x %x %x).\n",
|
|
mb[1], mb[2], mb[3]);
|
|
|
|
if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
|
|
atomic_set(&vha->loop_state, LOOP_DOWN);
|
|
atomic_set(&vha->loop_down_timer, LOOP_DOWN_TIME);
|
|
vha->device_flags |= DFLG_NO_CABLE;
|
|
qla2x00_mark_all_devices_lost(vha, 1);
|
|
}
|
|
|
|
if (vha->vp_idx) {
|
|
atomic_set(&vha->vp_state, VP_FAILED);
|
|
fc_vport_set_state(vha->fc_vport, FC_VPORT_FAILED);
|
|
}
|
|
|
|
vha->flags.management_server_logged_in = 0;
|
|
ha->link_data_rate = PORT_SPEED_UNKNOWN;
|
|
qla2x00_post_aen_work(vha, FCH_EVT_LINKDOWN, 0);
|
|
break;
|
|
|
|
case MBA_LIP_RESET: /* LIP reset occurred */
|
|
DEBUG2(printk("scsi(%ld): Asynchronous LIP RESET (%x).\n",
|
|
vha->host_no, mb[1]));
|
|
qla_printk(KERN_INFO, ha,
|
|
"LIP reset occurred (%x).\n", mb[1]);
|
|
|
|
if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
|
|
atomic_set(&vha->loop_state, LOOP_DOWN);
|
|
atomic_set(&vha->loop_down_timer, LOOP_DOWN_TIME);
|
|
qla2x00_mark_all_devices_lost(vha, 1);
|
|
}
|
|
|
|
if (vha->vp_idx) {
|
|
atomic_set(&vha->vp_state, VP_FAILED);
|
|
fc_vport_set_state(vha->fc_vport, FC_VPORT_FAILED);
|
|
}
|
|
|
|
set_bit(RESET_MARKER_NEEDED, &vha->dpc_flags);
|
|
|
|
ha->operating_mode = LOOP;
|
|
vha->flags.management_server_logged_in = 0;
|
|
qla2x00_post_aen_work(vha, FCH_EVT_LIPRESET, mb[1]);
|
|
break;
|
|
|
|
/* case MBA_DCBX_COMPLETE: */
|
|
case MBA_POINT_TO_POINT: /* Point-to-Point */
|
|
if (IS_QLA2100(ha))
|
|
break;
|
|
|
|
if (IS_QLA81XX(ha))
|
|
DEBUG2(printk("scsi(%ld): DCBX Completed -- %04x %04x "
|
|
"%04x\n", vha->host_no, mb[1], mb[2], mb[3]));
|
|
else
|
|
DEBUG2(printk("scsi(%ld): Asynchronous P2P MODE "
|
|
"received.\n", vha->host_no));
|
|
|
|
/*
|
|
* Until there's a transition from loop down to loop up, treat
|
|
* this as loop down only.
|
|
*/
|
|
if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
|
|
atomic_set(&vha->loop_state, LOOP_DOWN);
|
|
if (!atomic_read(&vha->loop_down_timer))
|
|
atomic_set(&vha->loop_down_timer,
|
|
LOOP_DOWN_TIME);
|
|
qla2x00_mark_all_devices_lost(vha, 1);
|
|
}
|
|
|
|
if (vha->vp_idx) {
|
|
atomic_set(&vha->vp_state, VP_FAILED);
|
|
fc_vport_set_state(vha->fc_vport, FC_VPORT_FAILED);
|
|
}
|
|
|
|
if (!(test_bit(ABORT_ISP_ACTIVE, &vha->dpc_flags)))
|
|
set_bit(RESET_MARKER_NEEDED, &vha->dpc_flags);
|
|
|
|
set_bit(REGISTER_FC4_NEEDED, &vha->dpc_flags);
|
|
set_bit(REGISTER_FDMI_NEEDED, &vha->dpc_flags);
|
|
|
|
ha->flags.gpsc_supported = 1;
|
|
vha->flags.management_server_logged_in = 0;
|
|
break;
|
|
|
|
case MBA_CHG_IN_CONNECTION: /* Change in connection mode */
|
|
if (IS_QLA2100(ha))
|
|
break;
|
|
|
|
DEBUG2(printk("scsi(%ld): Asynchronous Change In Connection "
|
|
"received.\n",
|
|
vha->host_no));
|
|
qla_printk(KERN_INFO, ha,
|
|
"Configuration change detected: value=%x.\n", mb[1]);
|
|
|
|
if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
|
|
atomic_set(&vha->loop_state, LOOP_DOWN);
|
|
if (!atomic_read(&vha->loop_down_timer))
|
|
atomic_set(&vha->loop_down_timer,
|
|
LOOP_DOWN_TIME);
|
|
qla2x00_mark_all_devices_lost(vha, 1);
|
|
}
|
|
|
|
if (vha->vp_idx) {
|
|
atomic_set(&vha->vp_state, VP_FAILED);
|
|
fc_vport_set_state(vha->fc_vport, FC_VPORT_FAILED);
|
|
}
|
|
|
|
set_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags);
|
|
set_bit(LOCAL_LOOP_UPDATE, &vha->dpc_flags);
|
|
break;
|
|
|
|
case MBA_PORT_UPDATE: /* Port database update */
|
|
/* Only handle SCNs for our Vport index. */
|
|
if (vha->vp_idx && vha->vp_idx != (mb[3] & 0xff))
|
|
break;
|
|
|
|
/*
|
|
* If PORT UPDATE is global (received LIP_OCCURRED/LIP_RESET
|
|
* event etc. earlier indicating loop is down) then process
|
|
* it. Otherwise ignore it and Wait for RSCN to come in.
|
|
*/
|
|
atomic_set(&vha->loop_down_timer, 0);
|
|
if (atomic_read(&vha->loop_state) != LOOP_DOWN &&
|
|
atomic_read(&vha->loop_state) != LOOP_DEAD) {
|
|
DEBUG2(printk("scsi(%ld): Asynchronous PORT UPDATE "
|
|
"ignored %04x/%04x/%04x.\n", vha->host_no, mb[1],
|
|
mb[2], mb[3]));
|
|
break;
|
|
}
|
|
|
|
DEBUG2(printk("scsi(%ld): Asynchronous PORT UPDATE.\n",
|
|
vha->host_no));
|
|
DEBUG(printk(KERN_INFO
|
|
"scsi(%ld): Port database changed %04x %04x %04x.\n",
|
|
vha->host_no, mb[1], mb[2], mb[3]));
|
|
|
|
/*
|
|
* Mark all devices as missing so we will login again.
|
|
*/
|
|
atomic_set(&vha->loop_state, LOOP_UP);
|
|
|
|
qla2x00_mark_all_devices_lost(vha, 1);
|
|
|
|
vha->flags.rscn_queue_overflow = 1;
|
|
|
|
set_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags);
|
|
set_bit(LOCAL_LOOP_UPDATE, &vha->dpc_flags);
|
|
break;
|
|
|
|
case MBA_RSCN_UPDATE: /* State Change Registration */
|
|
/* Check if the Vport has issued a SCR */
|
|
if (vha->vp_idx && test_bit(VP_SCR_NEEDED, &vha->vp_flags))
|
|
break;
|
|
/* Only handle SCNs for our Vport index. */
|
|
if (vha->vp_idx && vha->vp_idx != (mb[3] & 0xff))
|
|
break;
|
|
DEBUG2(printk("scsi(%ld): Asynchronous RSCR UPDATE.\n",
|
|
vha->host_no));
|
|
DEBUG(printk(KERN_INFO
|
|
"scsi(%ld): RSCN database changed -- %04x %04x %04x.\n",
|
|
vha->host_no, mb[1], mb[2], mb[3]));
|
|
|
|
rscn_entry = ((mb[1] & 0xff) << 16) | mb[2];
|
|
host_pid = (vha->d_id.b.domain << 16) | (vha->d_id.b.area << 8)
|
|
| vha->d_id.b.al_pa;
|
|
if (rscn_entry == host_pid) {
|
|
DEBUG(printk(KERN_INFO
|
|
"scsi(%ld): Ignoring RSCN update to local host "
|
|
"port ID (%06x)\n",
|
|
vha->host_no, host_pid));
|
|
break;
|
|
}
|
|
|
|
/* Ignore reserved bits from RSCN-payload. */
|
|
rscn_entry = ((mb[1] & 0x3ff) << 16) | mb[2];
|
|
rscn_queue_index = vha->rscn_in_ptr + 1;
|
|
if (rscn_queue_index == MAX_RSCN_COUNT)
|
|
rscn_queue_index = 0;
|
|
if (rscn_queue_index != vha->rscn_out_ptr) {
|
|
vha->rscn_queue[vha->rscn_in_ptr] = rscn_entry;
|
|
vha->rscn_in_ptr = rscn_queue_index;
|
|
} else {
|
|
vha->flags.rscn_queue_overflow = 1;
|
|
}
|
|
|
|
atomic_set(&vha->loop_state, LOOP_UPDATE);
|
|
atomic_set(&vha->loop_down_timer, 0);
|
|
vha->flags.management_server_logged_in = 0;
|
|
|
|
set_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags);
|
|
set_bit(RSCN_UPDATE, &vha->dpc_flags);
|
|
qla2x00_post_aen_work(vha, FCH_EVT_RSCN, rscn_entry);
|
|
break;
|
|
|
|
/* case MBA_RIO_RESPONSE: */
|
|
case MBA_ZIO_RESPONSE:
|
|
DEBUG3(printk("scsi(%ld): [R|Z]IO update completion.\n",
|
|
vha->host_no));
|
|
|
|
if (IS_FWI2_CAPABLE(ha))
|
|
qla24xx_process_response_queue(vha, rsp);
|
|
else
|
|
qla2x00_process_response_queue(rsp);
|
|
break;
|
|
|
|
case MBA_DISCARD_RND_FRAME:
|
|
DEBUG2(printk("scsi(%ld): Discard RND Frame -- %04x %04x "
|
|
"%04x.\n", vha->host_no, mb[1], mb[2], mb[3]));
|
|
break;
|
|
|
|
case MBA_TRACE_NOTIFICATION:
|
|
DEBUG2(printk("scsi(%ld): Trace Notification -- %04x %04x.\n",
|
|
vha->host_no, mb[1], mb[2]));
|
|
break;
|
|
|
|
case MBA_ISP84XX_ALERT:
|
|
DEBUG2(printk("scsi(%ld): ISP84XX Alert Notification -- "
|
|
"%04x %04x %04x\n", vha->host_no, mb[1], mb[2], mb[3]));
|
|
|
|
spin_lock_irqsave(&ha->cs84xx->access_lock, flags);
|
|
switch (mb[1]) {
|
|
case A84_PANIC_RECOVERY:
|
|
qla_printk(KERN_INFO, ha, "Alert 84XX: panic recovery "
|
|
"%04x %04x\n", mb[2], mb[3]);
|
|
break;
|
|
case A84_OP_LOGIN_COMPLETE:
|
|
ha->cs84xx->op_fw_version = mb[3] << 16 | mb[2];
|
|
DEBUG2(qla_printk(KERN_INFO, ha, "Alert 84XX:"
|
|
"firmware version %x\n", ha->cs84xx->op_fw_version));
|
|
break;
|
|
case A84_DIAG_LOGIN_COMPLETE:
|
|
ha->cs84xx->diag_fw_version = mb[3] << 16 | mb[2];
|
|
DEBUG2(qla_printk(KERN_INFO, ha, "Alert 84XX:"
|
|
"diagnostic firmware version %x\n",
|
|
ha->cs84xx->diag_fw_version));
|
|
break;
|
|
case A84_GOLD_LOGIN_COMPLETE:
|
|
ha->cs84xx->diag_fw_version = mb[3] << 16 | mb[2];
|
|
ha->cs84xx->fw_update = 1;
|
|
DEBUG2(qla_printk(KERN_INFO, ha, "Alert 84XX: gold "
|
|
"firmware version %x\n",
|
|
ha->cs84xx->gold_fw_version));
|
|
break;
|
|
default:
|
|
qla_printk(KERN_ERR, ha,
|
|
"Alert 84xx: Invalid Alert %04x %04x %04x\n",
|
|
mb[1], mb[2], mb[3]);
|
|
}
|
|
spin_unlock_irqrestore(&ha->cs84xx->access_lock, flags);
|
|
break;
|
|
case MBA_DCBX_START:
|
|
DEBUG2(printk("scsi(%ld): DCBX Started -- %04x %04x %04x\n",
|
|
vha->host_no, mb[1], mb[2], mb[3]));
|
|
break;
|
|
case MBA_DCBX_PARAM_UPDATE:
|
|
DEBUG2(printk("scsi(%ld): DCBX Parameters Updated -- "
|
|
"%04x %04x %04x\n", vha->host_no, mb[1], mb[2], mb[3]));
|
|
break;
|
|
case MBA_FCF_CONF_ERR:
|
|
DEBUG2(printk("scsi(%ld): FCF Configuration Error -- "
|
|
"%04x %04x %04x\n", vha->host_no, mb[1], mb[2], mb[3]));
|
|
break;
|
|
case MBA_IDC_COMPLETE:
|
|
case MBA_IDC_NOTIFY:
|
|
case MBA_IDC_TIME_EXT:
|
|
qla81xx_idc_event(vha, mb[0], mb[1]);
|
|
break;
|
|
}
|
|
|
|
if (!vha->vp_idx && ha->num_vhosts)
|
|
qla2x00_alert_all_vps(rsp, mb);
|
|
}
|
|
|
|
static void
|
|
qla2x00_adjust_sdev_qdepth_up(struct scsi_device *sdev, void *data)
|
|
{
|
|
fc_port_t *fcport = data;
|
|
struct scsi_qla_host *vha = fcport->vha;
|
|
struct qla_hw_data *ha = vha->hw;
|
|
struct req_que *req = NULL;
|
|
|
|
if (!ql2xqfulltracking)
|
|
return;
|
|
|
|
req = vha->req;
|
|
if (!req)
|
|
return;
|
|
if (req->max_q_depth <= sdev->queue_depth)
|
|
return;
|
|
|
|
if (sdev->ordered_tags)
|
|
scsi_adjust_queue_depth(sdev, MSG_ORDERED_TAG,
|
|
sdev->queue_depth + 1);
|
|
else
|
|
scsi_adjust_queue_depth(sdev, MSG_SIMPLE_TAG,
|
|
sdev->queue_depth + 1);
|
|
|
|
fcport->last_ramp_up = jiffies;
|
|
|
|
DEBUG2(qla_printk(KERN_INFO, ha,
|
|
"scsi(%ld:%d:%d:%d): Queue depth adjusted-up to %d.\n",
|
|
fcport->vha->host_no, sdev->channel, sdev->id, sdev->lun,
|
|
sdev->queue_depth));
|
|
}
|
|
|
|
static void
|
|
qla2x00_adjust_sdev_qdepth_down(struct scsi_device *sdev, void *data)
|
|
{
|
|
fc_port_t *fcport = data;
|
|
|
|
if (!scsi_track_queue_full(sdev, sdev->queue_depth - 1))
|
|
return;
|
|
|
|
DEBUG2(qla_printk(KERN_INFO, fcport->vha->hw,
|
|
"scsi(%ld:%d:%d:%d): Queue depth adjusted-down to %d.\n",
|
|
fcport->vha->host_no, sdev->channel, sdev->id, sdev->lun,
|
|
sdev->queue_depth));
|
|
}
|
|
|
|
static inline void
|
|
qla2x00_ramp_up_queue_depth(scsi_qla_host_t *vha, struct req_que *req,
|
|
srb_t *sp)
|
|
{
|
|
fc_port_t *fcport;
|
|
struct scsi_device *sdev;
|
|
|
|
if (!ql2xqfulltracking)
|
|
return;
|
|
|
|
sdev = sp->cmd->device;
|
|
if (sdev->queue_depth >= req->max_q_depth)
|
|
return;
|
|
|
|
fcport = sp->fcport;
|
|
if (time_before(jiffies,
|
|
fcport->last_ramp_up + ql2xqfullrampup * HZ))
|
|
return;
|
|
if (time_before(jiffies,
|
|
fcport->last_queue_full + ql2xqfullrampup * HZ))
|
|
return;
|
|
|
|
starget_for_each_device(sdev->sdev_target, fcport,
|
|
qla2x00_adjust_sdev_qdepth_up);
|
|
}
|
|
|
|
/**
|
|
* qla2x00_process_completed_request() - Process a Fast Post response.
|
|
* @ha: SCSI driver HA context
|
|
* @index: SRB index
|
|
*/
|
|
static void
|
|
qla2x00_process_completed_request(struct scsi_qla_host *vha,
|
|
struct req_que *req, uint32_t index)
|
|
{
|
|
srb_t *sp;
|
|
struct qla_hw_data *ha = vha->hw;
|
|
|
|
/* Validate handle. */
|
|
if (index >= MAX_OUTSTANDING_COMMANDS) {
|
|
DEBUG2(printk("scsi(%ld): Invalid SCSI completion handle %d.\n",
|
|
vha->host_no, index));
|
|
qla_printk(KERN_WARNING, ha,
|
|
"Invalid SCSI completion handle %d.\n", index);
|
|
|
|
set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
|
|
return;
|
|
}
|
|
|
|
sp = req->outstanding_cmds[index];
|
|
if (sp) {
|
|
/* Free outstanding command slot. */
|
|
req->outstanding_cmds[index] = NULL;
|
|
|
|
/* Save ISP completion status */
|
|
sp->cmd->result = DID_OK << 16;
|
|
|
|
qla2x00_ramp_up_queue_depth(vha, req, sp);
|
|
qla2x00_sp_compl(ha, sp);
|
|
} else {
|
|
DEBUG2(printk("scsi(%ld) Req:%d: Invalid ISP SCSI completion"
|
|
" handle(%d)\n", vha->host_no, req->id, index));
|
|
qla_printk(KERN_WARNING, ha,
|
|
"Invalid ISP SCSI completion handle\n");
|
|
|
|
set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* qla2x00_process_response_queue() - Process response queue entries.
|
|
* @ha: SCSI driver HA context
|
|
*/
|
|
void
|
|
qla2x00_process_response_queue(struct rsp_que *rsp)
|
|
{
|
|
struct scsi_qla_host *vha;
|
|
struct qla_hw_data *ha = rsp->hw;
|
|
struct device_reg_2xxx __iomem *reg = &ha->iobase->isp;
|
|
sts_entry_t *pkt;
|
|
uint16_t handle_cnt;
|
|
uint16_t cnt;
|
|
|
|
vha = pci_get_drvdata(ha->pdev);
|
|
|
|
if (!vha->flags.online)
|
|
return;
|
|
|
|
while (rsp->ring_ptr->signature != RESPONSE_PROCESSED) {
|
|
pkt = (sts_entry_t *)rsp->ring_ptr;
|
|
|
|
rsp->ring_index++;
|
|
if (rsp->ring_index == rsp->length) {
|
|
rsp->ring_index = 0;
|
|
rsp->ring_ptr = rsp->ring;
|
|
} else {
|
|
rsp->ring_ptr++;
|
|
}
|
|
|
|
if (pkt->entry_status != 0) {
|
|
DEBUG3(printk(KERN_INFO
|
|
"scsi(%ld): Process error entry.\n", vha->host_no));
|
|
|
|
qla2x00_error_entry(vha, rsp, pkt);
|
|
((response_t *)pkt)->signature = RESPONSE_PROCESSED;
|
|
wmb();
|
|
continue;
|
|
}
|
|
|
|
switch (pkt->entry_type) {
|
|
case STATUS_TYPE:
|
|
qla2x00_status_entry(vha, rsp, pkt);
|
|
break;
|
|
case STATUS_TYPE_21:
|
|
handle_cnt = ((sts21_entry_t *)pkt)->handle_count;
|
|
for (cnt = 0; cnt < handle_cnt; cnt++) {
|
|
qla2x00_process_completed_request(vha, rsp->req,
|
|
((sts21_entry_t *)pkt)->handle[cnt]);
|
|
}
|
|
break;
|
|
case STATUS_TYPE_22:
|
|
handle_cnt = ((sts22_entry_t *)pkt)->handle_count;
|
|
for (cnt = 0; cnt < handle_cnt; cnt++) {
|
|
qla2x00_process_completed_request(vha, rsp->req,
|
|
((sts22_entry_t *)pkt)->handle[cnt]);
|
|
}
|
|
break;
|
|
case STATUS_CONT_TYPE:
|
|
qla2x00_status_cont_entry(rsp, (sts_cont_entry_t *)pkt);
|
|
break;
|
|
default:
|
|
/* Type Not Supported. */
|
|
DEBUG4(printk(KERN_WARNING
|
|
"scsi(%ld): Received unknown response pkt type %x "
|
|
"entry status=%x.\n",
|
|
vha->host_no, pkt->entry_type, pkt->entry_status));
|
|
break;
|
|
}
|
|
((response_t *)pkt)->signature = RESPONSE_PROCESSED;
|
|
wmb();
|
|
}
|
|
|
|
/* Adjust ring index */
|
|
WRT_REG_WORD(ISP_RSP_Q_OUT(ha, reg), rsp->ring_index);
|
|
}
|
|
|
|
static inline void
|
|
qla2x00_handle_sense(srb_t *sp, uint8_t *sense_data, uint32_t sense_len,
|
|
struct rsp_que *rsp)
|
|
{
|
|
struct scsi_cmnd *cp = sp->cmd;
|
|
|
|
if (sense_len >= SCSI_SENSE_BUFFERSIZE)
|
|
sense_len = SCSI_SENSE_BUFFERSIZE;
|
|
|
|
sp->request_sense_length = sense_len;
|
|
sp->request_sense_ptr = cp->sense_buffer;
|
|
if (sp->request_sense_length > 32)
|
|
sense_len = 32;
|
|
|
|
memcpy(cp->sense_buffer, sense_data, sense_len);
|
|
|
|
sp->request_sense_ptr += sense_len;
|
|
sp->request_sense_length -= sense_len;
|
|
if (sp->request_sense_length != 0)
|
|
rsp->status_srb = sp;
|
|
|
|
DEBUG5(printk("%s(): Check condition Sense data, scsi(%ld:%d:%d:%d) "
|
|
"cmd=%p pid=%ld\n", __func__, sp->fcport->vha->host_no,
|
|
cp->device->channel, cp->device->id, cp->device->lun, cp,
|
|
cp->serial_number));
|
|
if (sense_len)
|
|
DEBUG5(qla2x00_dump_buffer(cp->sense_buffer, sense_len));
|
|
}
|
|
|
|
/**
|
|
* qla2x00_status_entry() - Process a Status IOCB entry.
|
|
* @ha: SCSI driver HA context
|
|
* @pkt: Entry pointer
|
|
*/
|
|
static void
|
|
qla2x00_status_entry(scsi_qla_host_t *vha, struct rsp_que *rsp, void *pkt)
|
|
{
|
|
srb_t *sp;
|
|
fc_port_t *fcport;
|
|
struct scsi_cmnd *cp;
|
|
sts_entry_t *sts;
|
|
struct sts_entry_24xx *sts24;
|
|
uint16_t comp_status;
|
|
uint16_t scsi_status;
|
|
uint8_t lscsi_status;
|
|
int32_t resid;
|
|
uint32_t sense_len, rsp_info_len, resid_len, fw_resid_len;
|
|
uint8_t *rsp_info, *sense_data;
|
|
struct qla_hw_data *ha = vha->hw;
|
|
uint32_t handle;
|
|
uint16_t que;
|
|
struct req_que *req;
|
|
|
|
sts = (sts_entry_t *) pkt;
|
|
sts24 = (struct sts_entry_24xx *) pkt;
|
|
if (IS_FWI2_CAPABLE(ha)) {
|
|
comp_status = le16_to_cpu(sts24->comp_status);
|
|
scsi_status = le16_to_cpu(sts24->scsi_status) & SS_MASK;
|
|
} else {
|
|
comp_status = le16_to_cpu(sts->comp_status);
|
|
scsi_status = le16_to_cpu(sts->scsi_status) & SS_MASK;
|
|
}
|
|
handle = (uint32_t) LSW(sts->handle);
|
|
que = MSW(sts->handle);
|
|
req = ha->req_q_map[que];
|
|
/* Fast path completion. */
|
|
if (comp_status == CS_COMPLETE && scsi_status == 0) {
|
|
qla2x00_process_completed_request(vha, req, handle);
|
|
|
|
return;
|
|
}
|
|
|
|
/* Validate handle. */
|
|
if (handle < MAX_OUTSTANDING_COMMANDS) {
|
|
sp = req->outstanding_cmds[handle];
|
|
req->outstanding_cmds[handle] = NULL;
|
|
} else
|
|
sp = NULL;
|
|
|
|
if (sp == NULL) {
|
|
DEBUG2(printk("scsi(%ld): Status Entry invalid handle.\n",
|
|
vha->host_no));
|
|
qla_printk(KERN_WARNING, ha, "Status Entry invalid handle.\n");
|
|
|
|
set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
|
|
qla2xxx_wake_dpc(vha);
|
|
return;
|
|
}
|
|
cp = sp->cmd;
|
|
if (cp == NULL) {
|
|
DEBUG2(printk("scsi(%ld): Command already returned back to OS "
|
|
"pkt->handle=%d sp=%p.\n", vha->host_no, handle, sp));
|
|
qla_printk(KERN_WARNING, ha,
|
|
"Command is NULL: already returned to OS (sp=%p)\n", sp);
|
|
|
|
return;
|
|
}
|
|
|
|
lscsi_status = scsi_status & STATUS_MASK;
|
|
|
|
fcport = sp->fcport;
|
|
|
|
sense_len = rsp_info_len = resid_len = fw_resid_len = 0;
|
|
if (IS_FWI2_CAPABLE(ha)) {
|
|
sense_len = le32_to_cpu(sts24->sense_len);
|
|
rsp_info_len = le32_to_cpu(sts24->rsp_data_len);
|
|
resid_len = le32_to_cpu(sts24->rsp_residual_count);
|
|
fw_resid_len = le32_to_cpu(sts24->residual_len);
|
|
rsp_info = sts24->data;
|
|
sense_data = sts24->data;
|
|
host_to_fcp_swap(sts24->data, sizeof(sts24->data));
|
|
} else {
|
|
sense_len = le16_to_cpu(sts->req_sense_length);
|
|
rsp_info_len = le16_to_cpu(sts->rsp_info_len);
|
|
resid_len = le32_to_cpu(sts->residual_length);
|
|
rsp_info = sts->rsp_info;
|
|
sense_data = sts->req_sense_data;
|
|
}
|
|
|
|
/* Check for any FCP transport errors. */
|
|
if (scsi_status & SS_RESPONSE_INFO_LEN_VALID) {
|
|
/* Sense data lies beyond any FCP RESPONSE data. */
|
|
if (IS_FWI2_CAPABLE(ha))
|
|
sense_data += rsp_info_len;
|
|
if (rsp_info_len > 3 && rsp_info[3]) {
|
|
DEBUG2(printk("scsi(%ld:%d:%d:%d) FCP I/O protocol "
|
|
"failure (%x/%02x%02x%02x%02x%02x%02x%02x%02x)..."
|
|
"retrying command\n", vha->host_no,
|
|
cp->device->channel, cp->device->id,
|
|
cp->device->lun, rsp_info_len, rsp_info[0],
|
|
rsp_info[1], rsp_info[2], rsp_info[3], rsp_info[4],
|
|
rsp_info[5], rsp_info[6], rsp_info[7]));
|
|
|
|
cp->result = DID_BUS_BUSY << 16;
|
|
qla2x00_sp_compl(ha, sp);
|
|
return;
|
|
}
|
|
}
|
|
|
|
/* Check for overrun. */
|
|
if (IS_FWI2_CAPABLE(ha) && comp_status == CS_COMPLETE &&
|
|
scsi_status & SS_RESIDUAL_OVER)
|
|
comp_status = CS_DATA_OVERRUN;
|
|
|
|
/*
|
|
* Based on Host and scsi status generate status code for Linux
|
|
*/
|
|
switch (comp_status) {
|
|
case CS_COMPLETE:
|
|
case CS_QUEUE_FULL:
|
|
if (scsi_status == 0) {
|
|
cp->result = DID_OK << 16;
|
|
break;
|
|
}
|
|
if (scsi_status & (SS_RESIDUAL_UNDER | SS_RESIDUAL_OVER)) {
|
|
resid = resid_len;
|
|
scsi_set_resid(cp, resid);
|
|
|
|
if (!lscsi_status &&
|
|
((unsigned)(scsi_bufflen(cp) - resid) <
|
|
cp->underflow)) {
|
|
qla_printk(KERN_INFO, ha,
|
|
"scsi(%ld:%d:%d:%d): Mid-layer underflow "
|
|
"detected (%x of %x bytes)...returning "
|
|
"error status.\n", vha->host_no,
|
|
cp->device->channel, cp->device->id,
|
|
cp->device->lun, resid,
|
|
scsi_bufflen(cp));
|
|
|
|
cp->result = DID_ERROR << 16;
|
|
break;
|
|
}
|
|
}
|
|
cp->result = DID_OK << 16 | lscsi_status;
|
|
|
|
if (lscsi_status == SAM_STAT_TASK_SET_FULL) {
|
|
DEBUG2(printk(KERN_INFO
|
|
"scsi(%ld): QUEUE FULL status detected "
|
|
"0x%x-0x%x.\n", vha->host_no, comp_status,
|
|
scsi_status));
|
|
|
|
/* Adjust queue depth for all luns on the port. */
|
|
if (!ql2xqfulltracking)
|
|
break;
|
|
fcport->last_queue_full = jiffies;
|
|
starget_for_each_device(cp->device->sdev_target,
|
|
fcport, qla2x00_adjust_sdev_qdepth_down);
|
|
break;
|
|
}
|
|
if (lscsi_status != SS_CHECK_CONDITION)
|
|
break;
|
|
|
|
memset(cp->sense_buffer, 0, SCSI_SENSE_BUFFERSIZE);
|
|
if (!(scsi_status & SS_SENSE_LEN_VALID))
|
|
break;
|
|
|
|
qla2x00_handle_sense(sp, sense_data, sense_len, rsp);
|
|
break;
|
|
|
|
case CS_DATA_UNDERRUN:
|
|
resid = resid_len;
|
|
/* Use F/W calculated residual length. */
|
|
if (IS_FWI2_CAPABLE(ha)) {
|
|
if (!(scsi_status & SS_RESIDUAL_UNDER)) {
|
|
lscsi_status = 0;
|
|
} else if (resid != fw_resid_len) {
|
|
scsi_status &= ~SS_RESIDUAL_UNDER;
|
|
lscsi_status = 0;
|
|
}
|
|
resid = fw_resid_len;
|
|
}
|
|
|
|
if (scsi_status & SS_RESIDUAL_UNDER) {
|
|
scsi_set_resid(cp, resid);
|
|
} else {
|
|
DEBUG2(printk(KERN_INFO
|
|
"scsi(%ld:%d:%d) UNDERRUN status detected "
|
|
"0x%x-0x%x. resid=0x%x fw_resid=0x%x cdb=0x%x "
|
|
"os_underflow=0x%x\n", vha->host_no,
|
|
cp->device->id, cp->device->lun, comp_status,
|
|
scsi_status, resid_len, resid, cp->cmnd[0],
|
|
cp->underflow));
|
|
|
|
}
|
|
|
|
/*
|
|
* Check to see if SCSI Status is non zero. If so report SCSI
|
|
* Status.
|
|
*/
|
|
if (lscsi_status != 0) {
|
|
cp->result = DID_OK << 16 | lscsi_status;
|
|
|
|
if (lscsi_status == SAM_STAT_TASK_SET_FULL) {
|
|
DEBUG2(printk(KERN_INFO
|
|
"scsi(%ld): QUEUE FULL status detected "
|
|
"0x%x-0x%x.\n", vha->host_no, comp_status,
|
|
scsi_status));
|
|
|
|
/*
|
|
* Adjust queue depth for all luns on the
|
|
* port.
|
|
*/
|
|
if (!ql2xqfulltracking)
|
|
break;
|
|
fcport->last_queue_full = jiffies;
|
|
starget_for_each_device(
|
|
cp->device->sdev_target, fcport,
|
|
qla2x00_adjust_sdev_qdepth_down);
|
|
break;
|
|
}
|
|
if (lscsi_status != SS_CHECK_CONDITION)
|
|
break;
|
|
|
|
memset(cp->sense_buffer, 0, SCSI_SENSE_BUFFERSIZE);
|
|
if (!(scsi_status & SS_SENSE_LEN_VALID))
|
|
break;
|
|
|
|
qla2x00_handle_sense(sp, sense_data, sense_len, rsp);
|
|
} else {
|
|
/*
|
|
* If RISC reports underrun and target does not report
|
|
* it then we must have a lost frame, so tell upper
|
|
* layer to retry it by reporting an error.
|
|
*/
|
|
if (!(scsi_status & SS_RESIDUAL_UNDER)) {
|
|
DEBUG2(printk("scsi(%ld:%d:%d:%d) Dropped "
|
|
"frame(s) detected (%x of %x bytes)..."
|
|
"retrying command.\n",
|
|
vha->host_no, cp->device->channel,
|
|
cp->device->id, cp->device->lun, resid,
|
|
scsi_bufflen(cp)));
|
|
|
|
cp->result = DID_ERROR << 16;
|
|
break;
|
|
}
|
|
|
|
/* Handle mid-layer underflow */
|
|
if ((unsigned)(scsi_bufflen(cp) - resid) <
|
|
cp->underflow) {
|
|
qla_printk(KERN_INFO, ha,
|
|
"scsi(%ld:%d:%d:%d): Mid-layer underflow "
|
|
"detected (%x of %x bytes)...returning "
|
|
"error status.\n", vha->host_no,
|
|
cp->device->channel, cp->device->id,
|
|
cp->device->lun, resid,
|
|
scsi_bufflen(cp));
|
|
|
|
cp->result = DID_ERROR << 16;
|
|
break;
|
|
}
|
|
|
|
/* Everybody online, looking good... */
|
|
cp->result = DID_OK << 16;
|
|
}
|
|
break;
|
|
|
|
case CS_DATA_OVERRUN:
|
|
DEBUG2(printk(KERN_INFO
|
|
"scsi(%ld:%d:%d): OVERRUN status detected 0x%x-0x%x\n",
|
|
vha->host_no, cp->device->id, cp->device->lun, comp_status,
|
|
scsi_status));
|
|
DEBUG2(printk(KERN_INFO
|
|
"CDB: 0x%x 0x%x 0x%x 0x%x 0x%x 0x%x\n",
|
|
cp->cmnd[0], cp->cmnd[1], cp->cmnd[2], cp->cmnd[3],
|
|
cp->cmnd[4], cp->cmnd[5]));
|
|
DEBUG2(printk(KERN_INFO
|
|
"PID=0x%lx req=0x%x xtra=0x%x -- returning DID_ERROR "
|
|
"status!\n",
|
|
cp->serial_number, scsi_bufflen(cp), resid_len));
|
|
|
|
cp->result = DID_ERROR << 16;
|
|
break;
|
|
|
|
case CS_PORT_LOGGED_OUT:
|
|
case CS_PORT_CONFIG_CHG:
|
|
case CS_PORT_BUSY:
|
|
case CS_INCOMPLETE:
|
|
case CS_PORT_UNAVAILABLE:
|
|
/*
|
|
* If the port is in Target Down state, return all IOs for this
|
|
* Target with DID_NO_CONNECT ELSE Queue the IOs in the
|
|
* retry_queue.
|
|
*/
|
|
DEBUG2(printk("scsi(%ld:%d:%d): status_entry: Port Down "
|
|
"pid=%ld, compl status=0x%x, port state=0x%x\n",
|
|
vha->host_no, cp->device->id, cp->device->lun,
|
|
cp->serial_number, comp_status,
|
|
atomic_read(&fcport->state)));
|
|
|
|
/*
|
|
* We are going to have the fc class block the rport
|
|
* while we try to recover so instruct the mid layer
|
|
* to requeue until the class decides how to handle this.
|
|
*/
|
|
cp->result = DID_TRANSPORT_DISRUPTED << 16;
|
|
if (atomic_read(&fcport->state) == FCS_ONLINE)
|
|
qla2x00_mark_device_lost(fcport->vha, fcport, 1, 1);
|
|
break;
|
|
|
|
case CS_RESET:
|
|
DEBUG2(printk(KERN_INFO
|
|
"scsi(%ld): RESET status detected 0x%x-0x%x.\n",
|
|
vha->host_no, comp_status, scsi_status));
|
|
|
|
cp->result = DID_RESET << 16;
|
|
break;
|
|
|
|
case CS_ABORTED:
|
|
/*
|
|
* hv2.19.12 - DID_ABORT does not retry the request if we
|
|
* aborted this request then abort otherwise it must be a
|
|
* reset.
|
|
*/
|
|
DEBUG2(printk(KERN_INFO
|
|
"scsi(%ld): ABORT status detected 0x%x-0x%x.\n",
|
|
vha->host_no, comp_status, scsi_status));
|
|
|
|
cp->result = DID_RESET << 16;
|
|
break;
|
|
|
|
case CS_TIMEOUT:
|
|
/*
|
|
* We are going to have the fc class block the rport
|
|
* while we try to recover so instruct the mid layer
|
|
* to requeue until the class decides how to handle this.
|
|
*/
|
|
cp->result = DID_TRANSPORT_DISRUPTED << 16;
|
|
|
|
if (IS_FWI2_CAPABLE(ha)) {
|
|
DEBUG2(printk(KERN_INFO
|
|
"scsi(%ld:%d:%d:%d): TIMEOUT status detected "
|
|
"0x%x-0x%x\n", vha->host_no, cp->device->channel,
|
|
cp->device->id, cp->device->lun, comp_status,
|
|
scsi_status));
|
|
break;
|
|
}
|
|
DEBUG2(printk(KERN_INFO
|
|
"scsi(%ld:%d:%d:%d): TIMEOUT status detected 0x%x-0x%x "
|
|
"sflags=%x.\n", vha->host_no, cp->device->channel,
|
|
cp->device->id, cp->device->lun, comp_status, scsi_status,
|
|
le16_to_cpu(sts->status_flags)));
|
|
|
|
/* Check to see if logout occurred. */
|
|
if ((le16_to_cpu(sts->status_flags) & SF_LOGOUT_SENT))
|
|
qla2x00_mark_device_lost(fcport->vha, fcport, 1, 1);
|
|
break;
|
|
|
|
default:
|
|
DEBUG3(printk("scsi(%ld): Error detected (unknown status) "
|
|
"0x%x-0x%x.\n", vha->host_no, comp_status, scsi_status));
|
|
qla_printk(KERN_INFO, ha,
|
|
"Unknown status detected 0x%x-0x%x.\n",
|
|
comp_status, scsi_status);
|
|
|
|
cp->result = DID_ERROR << 16;
|
|
break;
|
|
}
|
|
|
|
/* Place command on done queue. */
|
|
if (rsp->status_srb == NULL)
|
|
qla2x00_sp_compl(ha, sp);
|
|
}
|
|
|
|
/**
|
|
* qla2x00_status_cont_entry() - Process a Status Continuations entry.
|
|
* @ha: SCSI driver HA context
|
|
* @pkt: Entry pointer
|
|
*
|
|
* Extended sense data.
|
|
*/
|
|
static void
|
|
qla2x00_status_cont_entry(struct rsp_que *rsp, sts_cont_entry_t *pkt)
|
|
{
|
|
uint8_t sense_sz = 0;
|
|
struct qla_hw_data *ha = rsp->hw;
|
|
srb_t *sp = rsp->status_srb;
|
|
struct scsi_cmnd *cp;
|
|
|
|
if (sp != NULL && sp->request_sense_length != 0) {
|
|
cp = sp->cmd;
|
|
if (cp == NULL) {
|
|
DEBUG2(printk("%s(): Cmd already returned back to OS "
|
|
"sp=%p.\n", __func__, sp));
|
|
qla_printk(KERN_INFO, ha,
|
|
"cmd is NULL: already returned to OS (sp=%p)\n",
|
|
sp);
|
|
|
|
rsp->status_srb = NULL;
|
|
return;
|
|
}
|
|
|
|
if (sp->request_sense_length > sizeof(pkt->data)) {
|
|
sense_sz = sizeof(pkt->data);
|
|
} else {
|
|
sense_sz = sp->request_sense_length;
|
|
}
|
|
|
|
/* Move sense data. */
|
|
if (IS_FWI2_CAPABLE(ha))
|
|
host_to_fcp_swap(pkt->data, sizeof(pkt->data));
|
|
memcpy(sp->request_sense_ptr, pkt->data, sense_sz);
|
|
DEBUG5(qla2x00_dump_buffer(sp->request_sense_ptr, sense_sz));
|
|
|
|
sp->request_sense_ptr += sense_sz;
|
|
sp->request_sense_length -= sense_sz;
|
|
|
|
/* Place command on done queue. */
|
|
if (sp->request_sense_length == 0) {
|
|
rsp->status_srb = NULL;
|
|
qla2x00_sp_compl(ha, sp);
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* qla2x00_error_entry() - Process an error entry.
|
|
* @ha: SCSI driver HA context
|
|
* @pkt: Entry pointer
|
|
*/
|
|
static void
|
|
qla2x00_error_entry(scsi_qla_host_t *vha, struct rsp_que *rsp, sts_entry_t *pkt)
|
|
{
|
|
srb_t *sp;
|
|
struct qla_hw_data *ha = vha->hw;
|
|
uint32_t handle = LSW(pkt->handle);
|
|
uint16_t que = MSW(pkt->handle);
|
|
struct req_que *req = ha->req_q_map[que];
|
|
#if defined(QL_DEBUG_LEVEL_2)
|
|
if (pkt->entry_status & RF_INV_E_ORDER)
|
|
qla_printk(KERN_ERR, ha, "%s: Invalid Entry Order\n", __func__);
|
|
else if (pkt->entry_status & RF_INV_E_COUNT)
|
|
qla_printk(KERN_ERR, ha, "%s: Invalid Entry Count\n", __func__);
|
|
else if (pkt->entry_status & RF_INV_E_PARAM)
|
|
qla_printk(KERN_ERR, ha,
|
|
"%s: Invalid Entry Parameter\n", __func__);
|
|
else if (pkt->entry_status & RF_INV_E_TYPE)
|
|
qla_printk(KERN_ERR, ha, "%s: Invalid Entry Type\n", __func__);
|
|
else if (pkt->entry_status & RF_BUSY)
|
|
qla_printk(KERN_ERR, ha, "%s: Busy\n", __func__);
|
|
else
|
|
qla_printk(KERN_ERR, ha, "%s: UNKNOWN flag error\n", __func__);
|
|
#endif
|
|
|
|
/* Validate handle. */
|
|
if (handle < MAX_OUTSTANDING_COMMANDS)
|
|
sp = req->outstanding_cmds[handle];
|
|
else
|
|
sp = NULL;
|
|
|
|
if (sp) {
|
|
/* Free outstanding command slot. */
|
|
req->outstanding_cmds[handle] = NULL;
|
|
|
|
/* Bad payload or header */
|
|
if (pkt->entry_status &
|
|
(RF_INV_E_ORDER | RF_INV_E_COUNT |
|
|
RF_INV_E_PARAM | RF_INV_E_TYPE)) {
|
|
sp->cmd->result = DID_ERROR << 16;
|
|
} else if (pkt->entry_status & RF_BUSY) {
|
|
sp->cmd->result = DID_BUS_BUSY << 16;
|
|
} else {
|
|
sp->cmd->result = DID_ERROR << 16;
|
|
}
|
|
qla2x00_sp_compl(ha, sp);
|
|
|
|
} else if (pkt->entry_type == COMMAND_A64_TYPE || pkt->entry_type ==
|
|
COMMAND_TYPE || pkt->entry_type == COMMAND_TYPE_7) {
|
|
DEBUG2(printk("scsi(%ld): Error entry - invalid handle\n",
|
|
vha->host_no));
|
|
qla_printk(KERN_WARNING, ha,
|
|
"Error entry - invalid handle\n");
|
|
|
|
set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
|
|
qla2xxx_wake_dpc(vha);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* qla24xx_mbx_completion() - Process mailbox command completions.
|
|
* @ha: SCSI driver HA context
|
|
* @mb0: Mailbox0 register
|
|
*/
|
|
static void
|
|
qla24xx_mbx_completion(scsi_qla_host_t *vha, uint16_t mb0)
|
|
{
|
|
uint16_t cnt;
|
|
uint16_t __iomem *wptr;
|
|
struct qla_hw_data *ha = vha->hw;
|
|
struct device_reg_24xx __iomem *reg = &ha->iobase->isp24;
|
|
|
|
/* Load return mailbox registers. */
|
|
ha->flags.mbox_int = 1;
|
|
ha->mailbox_out[0] = mb0;
|
|
wptr = (uint16_t __iomem *)®->mailbox1;
|
|
|
|
for (cnt = 1; cnt < ha->mbx_count; cnt++) {
|
|
ha->mailbox_out[cnt] = RD_REG_WORD(wptr);
|
|
wptr++;
|
|
}
|
|
|
|
if (ha->mcp) {
|
|
DEBUG3(printk("%s(%ld): Got mailbox completion. cmd=%x.\n",
|
|
__func__, vha->host_no, ha->mcp->mb[0]));
|
|
} else {
|
|
DEBUG2_3(printk("%s(%ld): MBX pointer ERROR!\n",
|
|
__func__, vha->host_no));
|
|
}
|
|
}
|
|
|
|
/**
|
|
* qla24xx_process_response_queue() - Process response queue entries.
|
|
* @ha: SCSI driver HA context
|
|
*/
|
|
void qla24xx_process_response_queue(struct scsi_qla_host *vha,
|
|
struct rsp_que *rsp)
|
|
{
|
|
struct sts_entry_24xx *pkt;
|
|
|
|
if (!vha->flags.online)
|
|
return;
|
|
|
|
while (rsp->ring_ptr->signature != RESPONSE_PROCESSED) {
|
|
pkt = (struct sts_entry_24xx *)rsp->ring_ptr;
|
|
|
|
rsp->ring_index++;
|
|
if (rsp->ring_index == rsp->length) {
|
|
rsp->ring_index = 0;
|
|
rsp->ring_ptr = rsp->ring;
|
|
} else {
|
|
rsp->ring_ptr++;
|
|
}
|
|
|
|
if (pkt->entry_status != 0) {
|
|
DEBUG3(printk(KERN_INFO
|
|
"scsi(%ld): Process error entry.\n", vha->host_no));
|
|
|
|
qla2x00_error_entry(vha, rsp, (sts_entry_t *) pkt);
|
|
((response_t *)pkt)->signature = RESPONSE_PROCESSED;
|
|
wmb();
|
|
continue;
|
|
}
|
|
|
|
switch (pkt->entry_type) {
|
|
case STATUS_TYPE:
|
|
qla2x00_status_entry(vha, rsp, pkt);
|
|
break;
|
|
case STATUS_CONT_TYPE:
|
|
qla2x00_status_cont_entry(rsp, (sts_cont_entry_t *)pkt);
|
|
break;
|
|
case VP_RPT_ID_IOCB_TYPE:
|
|
qla24xx_report_id_acquisition(vha,
|
|
(struct vp_rpt_id_entry_24xx *)pkt);
|
|
break;
|
|
default:
|
|
/* Type Not Supported. */
|
|
DEBUG4(printk(KERN_WARNING
|
|
"scsi(%ld): Received unknown response pkt type %x "
|
|
"entry status=%x.\n",
|
|
vha->host_no, pkt->entry_type, pkt->entry_status));
|
|
break;
|
|
}
|
|
((response_t *)pkt)->signature = RESPONSE_PROCESSED;
|
|
wmb();
|
|
}
|
|
|
|
/* Adjust ring index */
|
|
WRT_REG_DWORD(rsp->rsp_q_out, rsp->ring_index);
|
|
}
|
|
|
|
static void
|
|
qla2xxx_check_risc_status(scsi_qla_host_t *vha)
|
|
{
|
|
int rval;
|
|
uint32_t cnt;
|
|
struct qla_hw_data *ha = vha->hw;
|
|
struct device_reg_24xx __iomem *reg = &ha->iobase->isp24;
|
|
|
|
if (!IS_QLA25XX(ha) && !IS_QLA81XX(ha))
|
|
return;
|
|
|
|
rval = QLA_SUCCESS;
|
|
WRT_REG_DWORD(®->iobase_addr, 0x7C00);
|
|
RD_REG_DWORD(®->iobase_addr);
|
|
WRT_REG_DWORD(®->iobase_window, 0x0001);
|
|
for (cnt = 10000; (RD_REG_DWORD(®->iobase_window) & BIT_0) == 0 &&
|
|
rval == QLA_SUCCESS; cnt--) {
|
|
if (cnt) {
|
|
WRT_REG_DWORD(®->iobase_window, 0x0001);
|
|
udelay(10);
|
|
} else
|
|
rval = QLA_FUNCTION_TIMEOUT;
|
|
}
|
|
if (rval == QLA_SUCCESS)
|
|
goto next_test;
|
|
|
|
WRT_REG_DWORD(®->iobase_window, 0x0003);
|
|
for (cnt = 100; (RD_REG_DWORD(®->iobase_window) & BIT_0) == 0 &&
|
|
rval == QLA_SUCCESS; cnt--) {
|
|
if (cnt) {
|
|
WRT_REG_DWORD(®->iobase_window, 0x0003);
|
|
udelay(10);
|
|
} else
|
|
rval = QLA_FUNCTION_TIMEOUT;
|
|
}
|
|
if (rval != QLA_SUCCESS)
|
|
goto done;
|
|
|
|
next_test:
|
|
if (RD_REG_DWORD(®->iobase_c8) & BIT_3)
|
|
qla_printk(KERN_INFO, ha, "Additional code -- 0x55AA.\n");
|
|
|
|
done:
|
|
WRT_REG_DWORD(®->iobase_window, 0x0000);
|
|
RD_REG_DWORD(®->iobase_window);
|
|
}
|
|
|
|
/**
|
|
* qla24xx_intr_handler() - Process interrupts for the ISP23xx and ISP63xx.
|
|
* @irq:
|
|
* @dev_id: SCSI driver HA context
|
|
*
|
|
* Called by system whenever the host adapter generates an interrupt.
|
|
*
|
|
* Returns handled flag.
|
|
*/
|
|
irqreturn_t
|
|
qla24xx_intr_handler(int irq, void *dev_id)
|
|
{
|
|
scsi_qla_host_t *vha;
|
|
struct qla_hw_data *ha;
|
|
struct device_reg_24xx __iomem *reg;
|
|
int status;
|
|
unsigned long iter;
|
|
uint32_t stat;
|
|
uint32_t hccr;
|
|
uint16_t mb[4];
|
|
struct rsp_que *rsp;
|
|
unsigned long flags;
|
|
|
|
rsp = (struct rsp_que *) dev_id;
|
|
if (!rsp) {
|
|
printk(KERN_INFO
|
|
"%s(): NULL response queue pointer\n", __func__);
|
|
return IRQ_NONE;
|
|
}
|
|
|
|
ha = rsp->hw;
|
|
reg = &ha->iobase->isp24;
|
|
status = 0;
|
|
|
|
spin_lock_irqsave(&ha->hardware_lock, flags);
|
|
vha = pci_get_drvdata(ha->pdev);
|
|
for (iter = 50; iter--; ) {
|
|
stat = RD_REG_DWORD(®->host_status);
|
|
if (stat & HSRX_RISC_PAUSED) {
|
|
if (pci_channel_offline(ha->pdev))
|
|
break;
|
|
|
|
hccr = RD_REG_DWORD(®->hccr);
|
|
|
|
qla_printk(KERN_INFO, ha, "RISC paused -- HCCR=%x, "
|
|
"Dumping firmware!\n", hccr);
|
|
|
|
qla2xxx_check_risc_status(vha);
|
|
|
|
ha->isp_ops->fw_dump(vha, 1);
|
|
set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
|
|
break;
|
|
} else if ((stat & HSRX_RISC_INT) == 0)
|
|
break;
|
|
|
|
switch (stat & 0xff) {
|
|
case 0x1:
|
|
case 0x2:
|
|
case 0x10:
|
|
case 0x11:
|
|
qla24xx_mbx_completion(vha, MSW(stat));
|
|
status |= MBX_INTERRUPT;
|
|
|
|
break;
|
|
case 0x12:
|
|
mb[0] = MSW(stat);
|
|
mb[1] = RD_REG_WORD(®->mailbox1);
|
|
mb[2] = RD_REG_WORD(®->mailbox2);
|
|
mb[3] = RD_REG_WORD(®->mailbox3);
|
|
qla2x00_async_event(vha, rsp, mb);
|
|
break;
|
|
case 0x13:
|
|
case 0x14:
|
|
qla24xx_process_response_queue(vha, rsp);
|
|
break;
|
|
default:
|
|
DEBUG2(printk("scsi(%ld): Unrecognized interrupt type "
|
|
"(%d).\n",
|
|
vha->host_no, stat & 0xff));
|
|
break;
|
|
}
|
|
WRT_REG_DWORD(®->hccr, HCCRX_CLR_RISC_INT);
|
|
RD_REG_DWORD_RELAXED(®->hccr);
|
|
}
|
|
spin_unlock_irqrestore(&ha->hardware_lock, flags);
|
|
|
|
if (test_bit(MBX_INTR_WAIT, &ha->mbx_cmd_flags) &&
|
|
(status & MBX_INTERRUPT) && ha->flags.mbox_int) {
|
|
set_bit(MBX_INTERRUPT, &ha->mbx_cmd_flags);
|
|
complete(&ha->mbx_intr_comp);
|
|
}
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static irqreturn_t
|
|
qla24xx_msix_rsp_q(int irq, void *dev_id)
|
|
{
|
|
struct qla_hw_data *ha;
|
|
struct rsp_que *rsp;
|
|
struct device_reg_24xx __iomem *reg;
|
|
struct scsi_qla_host *vha;
|
|
|
|
rsp = (struct rsp_que *) dev_id;
|
|
if (!rsp) {
|
|
printk(KERN_INFO
|
|
"%s(): NULL response queue pointer\n", __func__);
|
|
return IRQ_NONE;
|
|
}
|
|
ha = rsp->hw;
|
|
reg = &ha->iobase->isp24;
|
|
|
|
spin_lock_irq(&ha->hardware_lock);
|
|
|
|
vha = qla25xx_get_host(rsp);
|
|
qla24xx_process_response_queue(vha, rsp);
|
|
WRT_REG_DWORD(®->hccr, HCCRX_CLR_RISC_INT);
|
|
|
|
spin_unlock_irq(&ha->hardware_lock);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static irqreturn_t
|
|
qla25xx_msix_rsp_q(int irq, void *dev_id)
|
|
{
|
|
struct qla_hw_data *ha;
|
|
struct rsp_que *rsp;
|
|
|
|
rsp = (struct rsp_que *) dev_id;
|
|
if (!rsp) {
|
|
printk(KERN_INFO
|
|
"%s(): NULL response queue pointer\n", __func__);
|
|
return IRQ_NONE;
|
|
}
|
|
ha = rsp->hw;
|
|
|
|
queue_work_on((int) (rsp->id - 1), ha->wq, &rsp->q_work);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static irqreturn_t
|
|
qla24xx_msix_default(int irq, void *dev_id)
|
|
{
|
|
scsi_qla_host_t *vha;
|
|
struct qla_hw_data *ha;
|
|
struct rsp_que *rsp;
|
|
struct device_reg_24xx __iomem *reg;
|
|
int status;
|
|
uint32_t stat;
|
|
uint32_t hccr;
|
|
uint16_t mb[4];
|
|
|
|
rsp = (struct rsp_que *) dev_id;
|
|
if (!rsp) {
|
|
DEBUG(printk(
|
|
"%s(): NULL response queue pointer\n", __func__));
|
|
return IRQ_NONE;
|
|
}
|
|
ha = rsp->hw;
|
|
reg = &ha->iobase->isp24;
|
|
status = 0;
|
|
|
|
spin_lock_irq(&ha->hardware_lock);
|
|
vha = pci_get_drvdata(ha->pdev);
|
|
do {
|
|
stat = RD_REG_DWORD(®->host_status);
|
|
if (stat & HSRX_RISC_PAUSED) {
|
|
if (pci_channel_offline(ha->pdev))
|
|
break;
|
|
|
|
hccr = RD_REG_DWORD(®->hccr);
|
|
|
|
qla_printk(KERN_INFO, ha, "RISC paused -- HCCR=%x, "
|
|
"Dumping firmware!\n", hccr);
|
|
|
|
qla2xxx_check_risc_status(vha);
|
|
|
|
ha->isp_ops->fw_dump(vha, 1);
|
|
set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
|
|
break;
|
|
} else if ((stat & HSRX_RISC_INT) == 0)
|
|
break;
|
|
|
|
switch (stat & 0xff) {
|
|
case 0x1:
|
|
case 0x2:
|
|
case 0x10:
|
|
case 0x11:
|
|
qla24xx_mbx_completion(vha, MSW(stat));
|
|
status |= MBX_INTERRUPT;
|
|
|
|
break;
|
|
case 0x12:
|
|
mb[0] = MSW(stat);
|
|
mb[1] = RD_REG_WORD(®->mailbox1);
|
|
mb[2] = RD_REG_WORD(®->mailbox2);
|
|
mb[3] = RD_REG_WORD(®->mailbox3);
|
|
qla2x00_async_event(vha, rsp, mb);
|
|
break;
|
|
case 0x13:
|
|
case 0x14:
|
|
qla24xx_process_response_queue(vha, rsp);
|
|
break;
|
|
default:
|
|
DEBUG2(printk("scsi(%ld): Unrecognized interrupt type "
|
|
"(%d).\n",
|
|
vha->host_no, stat & 0xff));
|
|
break;
|
|
}
|
|
WRT_REG_DWORD(®->hccr, HCCRX_CLR_RISC_INT);
|
|
} while (0);
|
|
spin_unlock_irq(&ha->hardware_lock);
|
|
|
|
if (test_bit(MBX_INTR_WAIT, &ha->mbx_cmd_flags) &&
|
|
(status & MBX_INTERRUPT) && ha->flags.mbox_int) {
|
|
set_bit(MBX_INTERRUPT, &ha->mbx_cmd_flags);
|
|
complete(&ha->mbx_intr_comp);
|
|
}
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
/* Interrupt handling helpers. */
|
|
|
|
struct qla_init_msix_entry {
|
|
const char *name;
|
|
irq_handler_t handler;
|
|
};
|
|
|
|
static struct qla_init_msix_entry msix_entries[3] = {
|
|
{ "qla2xxx (default)", qla24xx_msix_default },
|
|
{ "qla2xxx (rsp_q)", qla24xx_msix_rsp_q },
|
|
{ "qla2xxx (multiq)", qla25xx_msix_rsp_q },
|
|
};
|
|
|
|
static void
|
|
qla24xx_disable_msix(struct qla_hw_data *ha)
|
|
{
|
|
int i;
|
|
struct qla_msix_entry *qentry;
|
|
|
|
for (i = 0; i < ha->msix_count; i++) {
|
|
qentry = &ha->msix_entries[i];
|
|
if (qentry->have_irq)
|
|
free_irq(qentry->vector, qentry->rsp);
|
|
}
|
|
pci_disable_msix(ha->pdev);
|
|
kfree(ha->msix_entries);
|
|
ha->msix_entries = NULL;
|
|
ha->flags.msix_enabled = 0;
|
|
}
|
|
|
|
static int
|
|
qla24xx_enable_msix(struct qla_hw_data *ha, struct rsp_que *rsp)
|
|
{
|
|
#define MIN_MSIX_COUNT 2
|
|
int i, ret;
|
|
struct msix_entry *entries;
|
|
struct qla_msix_entry *qentry;
|
|
|
|
entries = kzalloc(sizeof(struct msix_entry) * ha->msix_count,
|
|
GFP_KERNEL);
|
|
if (!entries)
|
|
return -ENOMEM;
|
|
|
|
for (i = 0; i < ha->msix_count; i++)
|
|
entries[i].entry = i;
|
|
|
|
ret = pci_enable_msix(ha->pdev, entries, ha->msix_count);
|
|
if (ret) {
|
|
if (ret < MIN_MSIX_COUNT)
|
|
goto msix_failed;
|
|
|
|
qla_printk(KERN_WARNING, ha,
|
|
"MSI-X: Failed to enable support -- %d/%d\n"
|
|
" Retry with %d vectors\n", ha->msix_count, ret, ret);
|
|
ha->msix_count = ret;
|
|
ret = pci_enable_msix(ha->pdev, entries, ha->msix_count);
|
|
if (ret) {
|
|
msix_failed:
|
|
qla_printk(KERN_WARNING, ha, "MSI-X: Failed to enable"
|
|
" support, giving up -- %d/%d\n",
|
|
ha->msix_count, ret);
|
|
goto msix_out;
|
|
}
|
|
ha->max_rsp_queues = ha->msix_count - 1;
|
|
}
|
|
ha->msix_entries = kzalloc(sizeof(struct qla_msix_entry) *
|
|
ha->msix_count, GFP_KERNEL);
|
|
if (!ha->msix_entries) {
|
|
ret = -ENOMEM;
|
|
goto msix_out;
|
|
}
|
|
ha->flags.msix_enabled = 1;
|
|
|
|
for (i = 0; i < ha->msix_count; i++) {
|
|
qentry = &ha->msix_entries[i];
|
|
qentry->vector = entries[i].vector;
|
|
qentry->entry = entries[i].entry;
|
|
qentry->have_irq = 0;
|
|
qentry->rsp = NULL;
|
|
}
|
|
|
|
/* Enable MSI-X vectors for the base queue */
|
|
for (i = 0; i < 2; i++) {
|
|
qentry = &ha->msix_entries[i];
|
|
ret = request_irq(qentry->vector, msix_entries[i].handler,
|
|
0, msix_entries[i].name, rsp);
|
|
if (ret) {
|
|
qla_printk(KERN_WARNING, ha,
|
|
"MSI-X: Unable to register handler -- %x/%d.\n",
|
|
qentry->vector, ret);
|
|
qla24xx_disable_msix(ha);
|
|
ha->mqenable = 0;
|
|
goto msix_out;
|
|
}
|
|
qentry->have_irq = 1;
|
|
qentry->rsp = rsp;
|
|
rsp->msix = qentry;
|
|
}
|
|
|
|
/* Enable MSI-X vector for response queue update for queue 0 */
|
|
if (ha->mqiobase && (ha->max_rsp_queues > 1 || ha->max_req_queues > 1))
|
|
ha->mqenable = 1;
|
|
|
|
msix_out:
|
|
kfree(entries);
|
|
return ret;
|
|
}
|
|
|
|
int
|
|
qla2x00_request_irqs(struct qla_hw_data *ha, struct rsp_que *rsp)
|
|
{
|
|
int ret;
|
|
device_reg_t __iomem *reg = ha->iobase;
|
|
|
|
/* If possible, enable MSI-X. */
|
|
if (!IS_QLA2432(ha) && !IS_QLA2532(ha) &&
|
|
!IS_QLA8432(ha) && !IS_QLA8001(ha))
|
|
goto skip_msix;
|
|
|
|
if (IS_QLA2432(ha) && (ha->pdev->revision < QLA_MSIX_CHIP_REV_24XX ||
|
|
!QLA_MSIX_FW_MODE_1(ha->fw_attributes))) {
|
|
DEBUG2(qla_printk(KERN_WARNING, ha,
|
|
"MSI-X: Unsupported ISP2432 (0x%X, 0x%X).\n",
|
|
ha->pdev->revision, ha->fw_attributes));
|
|
|
|
goto skip_msix;
|
|
}
|
|
|
|
if (ha->pdev->subsystem_vendor == PCI_VENDOR_ID_HP &&
|
|
(ha->pdev->subsystem_device == 0x7040 ||
|
|
ha->pdev->subsystem_device == 0x7041 ||
|
|
ha->pdev->subsystem_device == 0x1705)) {
|
|
DEBUG2(qla_printk(KERN_WARNING, ha,
|
|
"MSI-X: Unsupported ISP2432 SSVID/SSDID (0x%X, 0x%X).\n",
|
|
ha->pdev->subsystem_vendor,
|
|
ha->pdev->subsystem_device));
|
|
|
|
goto skip_msi;
|
|
}
|
|
|
|
ret = qla24xx_enable_msix(ha, rsp);
|
|
if (!ret) {
|
|
DEBUG2(qla_printk(KERN_INFO, ha,
|
|
"MSI-X: Enabled (0x%X, 0x%X).\n", ha->chip_revision,
|
|
ha->fw_attributes));
|
|
goto clear_risc_ints;
|
|
}
|
|
qla_printk(KERN_WARNING, ha,
|
|
"MSI-X: Falling back-to INTa mode -- %d.\n", ret);
|
|
skip_msix:
|
|
|
|
if (!IS_QLA24XX(ha) && !IS_QLA2532(ha) && !IS_QLA8432(ha) &&
|
|
!IS_QLA8001(ha))
|
|
goto skip_msi;
|
|
|
|
ret = pci_enable_msi(ha->pdev);
|
|
if (!ret) {
|
|
DEBUG2(qla_printk(KERN_INFO, ha, "MSI: Enabled.\n"));
|
|
ha->flags.msi_enabled = 1;
|
|
}
|
|
skip_msi:
|
|
|
|
ret = request_irq(ha->pdev->irq, ha->isp_ops->intr_handler,
|
|
IRQF_SHARED, QLA2XXX_DRIVER_NAME, rsp);
|
|
if (ret) {
|
|
qla_printk(KERN_WARNING, ha,
|
|
"Failed to reserve interrupt %d already in use.\n",
|
|
ha->pdev->irq);
|
|
goto fail;
|
|
}
|
|
ha->flags.inta_enabled = 1;
|
|
clear_risc_ints:
|
|
|
|
/*
|
|
* FIXME: Noted that 8014s were being dropped during NK testing.
|
|
* Timing deltas during MSI-X/INTa transitions?
|
|
*/
|
|
if (IS_QLA81XX(ha))
|
|
goto fail;
|
|
spin_lock_irq(&ha->hardware_lock);
|
|
if (IS_FWI2_CAPABLE(ha)) {
|
|
WRT_REG_DWORD(®->isp24.hccr, HCCRX_CLR_HOST_INT);
|
|
WRT_REG_DWORD(®->isp24.hccr, HCCRX_CLR_RISC_INT);
|
|
} else {
|
|
WRT_REG_WORD(®->isp.semaphore, 0);
|
|
WRT_REG_WORD(®->isp.hccr, HCCR_CLR_RISC_INT);
|
|
WRT_REG_WORD(®->isp.hccr, HCCR_CLR_HOST_INT);
|
|
}
|
|
spin_unlock_irq(&ha->hardware_lock);
|
|
|
|
fail:
|
|
return ret;
|
|
}
|
|
|
|
void
|
|
qla2x00_free_irqs(scsi_qla_host_t *vha)
|
|
{
|
|
struct qla_hw_data *ha = vha->hw;
|
|
struct rsp_que *rsp = ha->rsp_q_map[0];
|
|
|
|
if (ha->flags.msix_enabled)
|
|
qla24xx_disable_msix(ha);
|
|
else if (ha->flags.inta_enabled) {
|
|
free_irq(ha->pdev->irq, rsp);
|
|
pci_disable_msi(ha->pdev);
|
|
}
|
|
}
|
|
|
|
|
|
int qla25xx_request_irq(struct rsp_que *rsp)
|
|
{
|
|
struct qla_hw_data *ha = rsp->hw;
|
|
struct qla_init_msix_entry *intr = &msix_entries[2];
|
|
struct qla_msix_entry *msix = rsp->msix;
|
|
int ret;
|
|
|
|
ret = request_irq(msix->vector, intr->handler, 0, intr->name, rsp);
|
|
if (ret) {
|
|
qla_printk(KERN_WARNING, ha,
|
|
"MSI-X: Unable to register handler -- %x/%d.\n",
|
|
msix->vector, ret);
|
|
return ret;
|
|
}
|
|
msix->have_irq = 1;
|
|
msix->rsp = rsp;
|
|
return ret;
|
|
}
|
|
|
|
struct scsi_qla_host *
|
|
qla25xx_get_host(struct rsp_que *rsp)
|
|
{
|
|
srb_t *sp;
|
|
struct qla_hw_data *ha = rsp->hw;
|
|
struct scsi_qla_host *vha = NULL;
|
|
struct sts_entry_24xx *pkt;
|
|
struct req_que *req;
|
|
uint16_t que;
|
|
uint32_t handle;
|
|
|
|
pkt = (struct sts_entry_24xx *) rsp->ring_ptr;
|
|
que = MSW(pkt->handle);
|
|
handle = (uint32_t) LSW(pkt->handle);
|
|
req = ha->req_q_map[que];
|
|
if (handle < MAX_OUTSTANDING_COMMANDS) {
|
|
sp = req->outstanding_cmds[handle];
|
|
if (sp)
|
|
return sp->fcport->vha;
|
|
else
|
|
goto base_que;
|
|
}
|
|
base_que:
|
|
vha = pci_get_drvdata(ha->pdev);
|
|
return vha;
|
|
}
|