2
0
mirror of https://github.com/edk2-porting/linux-next.git synced 2024-12-17 09:43:59 +08:00
linux-next/drivers/scsi/qedi/qedi_fw.c
Linus Torvalds b2fe5fa686 Merge git://git.kernel.org/pub/scm/linux/kernel/git/davem/net-next
Pull networking updates from David Miller:

 1) Significantly shrink the core networking routing structures. Result
    of http://vger.kernel.org/~davem/seoul2017_netdev_keynote.pdf

 2) Add netdevsim driver for testing various offloads, from Jakub
    Kicinski.

 3) Support cross-chip FDB operations in DSA, from Vivien Didelot.

 4) Add a 2nd listener hash table for TCP, similar to what was done for
    UDP. From Martin KaFai Lau.

 5) Add eBPF based queue selection to tun, from Jason Wang.

 6) Lockless qdisc support, from John Fastabend.

 7) SCTP stream interleave support, from Xin Long.

 8) Smoother TCP receive autotuning, from Eric Dumazet.

 9) Lots of erspan tunneling enhancements, from William Tu.

10) Add true function call support to BPF, from Alexei Starovoitov.

11) Add explicit support for GRO HW offloading, from Michael Chan.

12) Support extack generation in more netlink subsystems. From Alexander
    Aring, Quentin Monnet, and Jakub Kicinski.

13) Add 1000BaseX, flow control, and EEE support to mvneta driver. From
    Russell King.

14) Add flow table abstraction to netfilter, from Pablo Neira Ayuso.

15) Many improvements and simplifications to the NFP driver bpf JIT,
    from Jakub Kicinski.

16) Support for ipv6 non-equal cost multipath routing, from Ido
    Schimmel.

17) Add resource abstration to devlink, from Arkadi Sharshevsky.

18) Packet scheduler classifier shared filter block support, from Jiri
    Pirko.

19) Avoid locking in act_csum, from Davide Caratti.

20) devinet_ioctl() simplifications from Al viro.

21) More TCP bpf improvements from Lawrence Brakmo.

22) Add support for onlink ipv6 route flag, similar to ipv4, from David
    Ahern.

* git://git.kernel.org/pub/scm/linux/kernel/git/davem/net-next: (1925 commits)
  tls: Add support for encryption using async offload accelerator
  ip6mr: fix stale iterator
  net/sched: kconfig: Remove blank help texts
  openvswitch: meter: Use 64-bit arithmetic instead of 32-bit
  tcp_nv: fix potential integer overflow in tcpnv_acked
  r8169: fix RTL8168EP take too long to complete driver initialization.
  qmi_wwan: Add support for Quectel EP06
  rtnetlink: enable IFLA_IF_NETNSID for RTM_NEWLINK
  ipmr: Fix ptrdiff_t print formatting
  ibmvnic: Wait for device response when changing MAC
  qlcnic: fix deadlock bug
  tcp: release sk_frag.page in tcp_disconnect
  ipv4: Get the address of interface correctly.
  net_sched: gen_estimator: fix lockdep splat
  net: macb: Handle HRESP error
  net/mlx5e: IPoIB, Fix copy-paste bug in flow steering refactoring
  ipv6: addrconf: break critical section in addrconf_verify_rtnl()
  ipv6: change route cache aging logic
  i40e/i40evf: Update DESC_NEEDED value to reflect larger value
  bnxt_en: cleanup DIM work on device shutdown
  ...
2018-01-31 14:31:10 -08:00

2229 lines
67 KiB
C

/*
* QLogic iSCSI Offload Driver
* Copyright (c) 2016 Cavium Inc.
*
* This software is available under the terms of the GNU General Public License
* (GPL) Version 2, available from the file COPYING in the main directory of
* this source tree.
*/
#include <linux/blkdev.h>
#include <scsi/scsi_tcq.h>
#include <linux/delay.h>
#include "qedi.h"
#include "qedi_iscsi.h"
#include "qedi_gbl.h"
#include "qedi_fw_iscsi.h"
#include "qedi_fw_scsi.h"
static int qedi_send_iscsi_tmf(struct qedi_conn *qedi_conn,
struct iscsi_task *mtask);
void qedi_iscsi_unmap_sg_list(struct qedi_cmd *cmd)
{
struct scsi_cmnd *sc = cmd->scsi_cmd;
if (cmd->io_tbl.sge_valid && sc) {
cmd->io_tbl.sge_valid = 0;
scsi_dma_unmap(sc);
}
}
static void qedi_process_logout_resp(struct qedi_ctx *qedi,
union iscsi_cqe *cqe,
struct iscsi_task *task,
struct qedi_conn *qedi_conn)
{
struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
struct iscsi_logout_rsp *resp_hdr;
struct iscsi_session *session = conn->session;
struct iscsi_logout_response_hdr *cqe_logout_response;
struct qedi_cmd *cmd;
cmd = (struct qedi_cmd *)task->dd_data;
cqe_logout_response = &cqe->cqe_common.iscsi_hdr.logout_response;
spin_lock(&session->back_lock);
resp_hdr = (struct iscsi_logout_rsp *)&qedi_conn->gen_pdu.resp_hdr;
memset(resp_hdr, 0, sizeof(struct iscsi_hdr));
resp_hdr->opcode = cqe_logout_response->opcode;
resp_hdr->flags = cqe_logout_response->flags;
resp_hdr->hlength = 0;
resp_hdr->itt = build_itt(cqe->cqe_solicited.itid, conn->session->age);
resp_hdr->statsn = cpu_to_be32(cqe_logout_response->stat_sn);
resp_hdr->exp_cmdsn = cpu_to_be32(cqe_logout_response->exp_cmd_sn);
resp_hdr->max_cmdsn = cpu_to_be32(cqe_logout_response->max_cmd_sn);
resp_hdr->t2wait = cpu_to_be32(cqe_logout_response->time_2_wait);
resp_hdr->t2retain = cpu_to_be32(cqe_logout_response->time_2_retain);
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_TID,
"Freeing tid=0x%x for cid=0x%x\n",
cmd->task_id, qedi_conn->iscsi_conn_id);
if (likely(cmd->io_cmd_in_list)) {
cmd->io_cmd_in_list = false;
list_del_init(&cmd->io_cmd);
qedi_conn->active_cmd_count--;
} else {
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_INFO,
"Active cmd list node already deleted, tid=0x%x, cid=0x%x, io_cmd_node=%p\n",
cmd->task_id, qedi_conn->iscsi_conn_id,
&cmd->io_cmd);
}
cmd->state = RESPONSE_RECEIVED;
qedi_clear_task_idx(qedi, cmd->task_id);
__iscsi_complete_pdu(conn, (struct iscsi_hdr *)resp_hdr, NULL, 0);
spin_unlock(&session->back_lock);
}
static void qedi_process_text_resp(struct qedi_ctx *qedi,
union iscsi_cqe *cqe,
struct iscsi_task *task,
struct qedi_conn *qedi_conn)
{
struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
struct iscsi_session *session = conn->session;
struct e4_iscsi_task_context *task_ctx;
struct iscsi_text_rsp *resp_hdr_ptr;
struct iscsi_text_response_hdr *cqe_text_response;
struct qedi_cmd *cmd;
int pld_len;
cmd = (struct qedi_cmd *)task->dd_data;
task_ctx = qedi_get_task_mem(&qedi->tasks, cmd->task_id);
cqe_text_response = &cqe->cqe_common.iscsi_hdr.text_response;
spin_lock(&session->back_lock);
resp_hdr_ptr = (struct iscsi_text_rsp *)&qedi_conn->gen_pdu.resp_hdr;
memset(resp_hdr_ptr, 0, sizeof(struct iscsi_hdr));
resp_hdr_ptr->opcode = cqe_text_response->opcode;
resp_hdr_ptr->flags = cqe_text_response->flags;
resp_hdr_ptr->hlength = 0;
hton24(resp_hdr_ptr->dlength,
(cqe_text_response->hdr_second_dword &
ISCSI_TEXT_RESPONSE_HDR_DATA_SEG_LEN_MASK));
resp_hdr_ptr->itt = build_itt(cqe->cqe_solicited.itid,
conn->session->age);
resp_hdr_ptr->ttt = cqe_text_response->ttt;
resp_hdr_ptr->statsn = cpu_to_be32(cqe_text_response->stat_sn);
resp_hdr_ptr->exp_cmdsn = cpu_to_be32(cqe_text_response->exp_cmd_sn);
resp_hdr_ptr->max_cmdsn = cpu_to_be32(cqe_text_response->max_cmd_sn);
pld_len = cqe_text_response->hdr_second_dword &
ISCSI_TEXT_RESPONSE_HDR_DATA_SEG_LEN_MASK;
qedi_conn->gen_pdu.resp_wr_ptr = qedi_conn->gen_pdu.resp_buf + pld_len;
memset(task_ctx, '\0', sizeof(*task_ctx));
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_TID,
"Freeing tid=0x%x for cid=0x%x\n",
cmd->task_id, qedi_conn->iscsi_conn_id);
if (likely(cmd->io_cmd_in_list)) {
cmd->io_cmd_in_list = false;
list_del_init(&cmd->io_cmd);
qedi_conn->active_cmd_count--;
} else {
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_INFO,
"Active cmd list node already deleted, tid=0x%x, cid=0x%x, io_cmd_node=%p\n",
cmd->task_id, qedi_conn->iscsi_conn_id,
&cmd->io_cmd);
}
cmd->state = RESPONSE_RECEIVED;
qedi_clear_task_idx(qedi, cmd->task_id);
__iscsi_complete_pdu(conn, (struct iscsi_hdr *)resp_hdr_ptr,
qedi_conn->gen_pdu.resp_buf,
(qedi_conn->gen_pdu.resp_wr_ptr -
qedi_conn->gen_pdu.resp_buf));
spin_unlock(&session->back_lock);
}
static void qedi_tmf_resp_work(struct work_struct *work)
{
struct qedi_cmd *qedi_cmd =
container_of(work, struct qedi_cmd, tmf_work);
struct qedi_conn *qedi_conn = qedi_cmd->conn;
struct qedi_ctx *qedi = qedi_conn->qedi;
struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
struct iscsi_session *session = conn->session;
struct iscsi_tm_rsp *resp_hdr_ptr;
struct iscsi_cls_session *cls_sess;
int rval = 0;
set_bit(QEDI_CONN_FW_CLEANUP, &qedi_conn->flags);
resp_hdr_ptr = (struct iscsi_tm_rsp *)qedi_cmd->tmf_resp_buf;
cls_sess = iscsi_conn_to_session(qedi_conn->cls_conn);
iscsi_block_session(session->cls_session);
rval = qedi_cleanup_all_io(qedi, qedi_conn, qedi_cmd->task, true);
if (rval) {
qedi_clear_task_idx(qedi, qedi_cmd->task_id);
iscsi_unblock_session(session->cls_session);
goto exit_tmf_resp;
}
iscsi_unblock_session(session->cls_session);
qedi_clear_task_idx(qedi, qedi_cmd->task_id);
spin_lock(&session->back_lock);
__iscsi_complete_pdu(conn, (struct iscsi_hdr *)resp_hdr_ptr, NULL, 0);
spin_unlock(&session->back_lock);
exit_tmf_resp:
kfree(resp_hdr_ptr);
clear_bit(QEDI_CONN_FW_CLEANUP, &qedi_conn->flags);
}
static void qedi_process_tmf_resp(struct qedi_ctx *qedi,
union iscsi_cqe *cqe,
struct iscsi_task *task,
struct qedi_conn *qedi_conn)
{
struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
struct iscsi_session *session = conn->session;
struct iscsi_tmf_response_hdr *cqe_tmp_response;
struct iscsi_tm_rsp *resp_hdr_ptr;
struct iscsi_tm *tmf_hdr;
struct qedi_cmd *qedi_cmd = NULL;
cqe_tmp_response = &cqe->cqe_common.iscsi_hdr.tmf_response;
qedi_cmd = task->dd_data;
qedi_cmd->tmf_resp_buf = kzalloc(sizeof(*resp_hdr_ptr), GFP_ATOMIC);
if (!qedi_cmd->tmf_resp_buf) {
QEDI_ERR(&qedi->dbg_ctx,
"Failed to allocate resp buf, cid=0x%x\n",
qedi_conn->iscsi_conn_id);
return;
}
spin_lock(&session->back_lock);
resp_hdr_ptr = (struct iscsi_tm_rsp *)qedi_cmd->tmf_resp_buf;
memset(resp_hdr_ptr, 0, sizeof(struct iscsi_tm_rsp));
/* Fill up the header */
resp_hdr_ptr->opcode = cqe_tmp_response->opcode;
resp_hdr_ptr->flags = cqe_tmp_response->hdr_flags;
resp_hdr_ptr->response = cqe_tmp_response->hdr_response;
resp_hdr_ptr->hlength = 0;
hton24(resp_hdr_ptr->dlength,
(cqe_tmp_response->hdr_second_dword &
ISCSI_TMF_RESPONSE_HDR_DATA_SEG_LEN_MASK));
resp_hdr_ptr->itt = build_itt(cqe->cqe_solicited.itid,
conn->session->age);
resp_hdr_ptr->statsn = cpu_to_be32(cqe_tmp_response->stat_sn);
resp_hdr_ptr->exp_cmdsn = cpu_to_be32(cqe_tmp_response->exp_cmd_sn);
resp_hdr_ptr->max_cmdsn = cpu_to_be32(cqe_tmp_response->max_cmd_sn);
tmf_hdr = (struct iscsi_tm *)qedi_cmd->task->hdr;
if (likely(qedi_cmd->io_cmd_in_list)) {
qedi_cmd->io_cmd_in_list = false;
list_del_init(&qedi_cmd->io_cmd);
qedi_conn->active_cmd_count--;
}
if (((tmf_hdr->flags & ISCSI_FLAG_TM_FUNC_MASK) ==
ISCSI_TM_FUNC_LOGICAL_UNIT_RESET) ||
((tmf_hdr->flags & ISCSI_FLAG_TM_FUNC_MASK) ==
ISCSI_TM_FUNC_TARGET_WARM_RESET) ||
((tmf_hdr->flags & ISCSI_FLAG_TM_FUNC_MASK) ==
ISCSI_TM_FUNC_TARGET_COLD_RESET)) {
INIT_WORK(&qedi_cmd->tmf_work, qedi_tmf_resp_work);
queue_work(qedi->tmf_thread, &qedi_cmd->tmf_work);
goto unblock_sess;
}
qedi_clear_task_idx(qedi, qedi_cmd->task_id);
__iscsi_complete_pdu(conn, (struct iscsi_hdr *)resp_hdr_ptr, NULL, 0);
kfree(resp_hdr_ptr);
unblock_sess:
spin_unlock(&session->back_lock);
}
static void qedi_process_login_resp(struct qedi_ctx *qedi,
union iscsi_cqe *cqe,
struct iscsi_task *task,
struct qedi_conn *qedi_conn)
{
struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
struct iscsi_session *session = conn->session;
struct e4_iscsi_task_context *task_ctx;
struct iscsi_login_rsp *resp_hdr_ptr;
struct iscsi_login_response_hdr *cqe_login_response;
struct qedi_cmd *cmd;
int pld_len;
cmd = (struct qedi_cmd *)task->dd_data;
cqe_login_response = &cqe->cqe_common.iscsi_hdr.login_response;
task_ctx = qedi_get_task_mem(&qedi->tasks, cmd->task_id);
spin_lock(&session->back_lock);
resp_hdr_ptr = (struct iscsi_login_rsp *)&qedi_conn->gen_pdu.resp_hdr;
memset(resp_hdr_ptr, 0, sizeof(struct iscsi_login_rsp));
resp_hdr_ptr->opcode = cqe_login_response->opcode;
resp_hdr_ptr->flags = cqe_login_response->flags_attr;
resp_hdr_ptr->hlength = 0;
hton24(resp_hdr_ptr->dlength,
(cqe_login_response->hdr_second_dword &
ISCSI_LOGIN_RESPONSE_HDR_DATA_SEG_LEN_MASK));
resp_hdr_ptr->itt = build_itt(cqe->cqe_solicited.itid,
conn->session->age);
resp_hdr_ptr->tsih = cqe_login_response->tsih;
resp_hdr_ptr->statsn = cpu_to_be32(cqe_login_response->stat_sn);
resp_hdr_ptr->exp_cmdsn = cpu_to_be32(cqe_login_response->exp_cmd_sn);
resp_hdr_ptr->max_cmdsn = cpu_to_be32(cqe_login_response->max_cmd_sn);
resp_hdr_ptr->status_class = cqe_login_response->status_class;
resp_hdr_ptr->status_detail = cqe_login_response->status_detail;
pld_len = cqe_login_response->hdr_second_dword &
ISCSI_LOGIN_RESPONSE_HDR_DATA_SEG_LEN_MASK;
qedi_conn->gen_pdu.resp_wr_ptr = qedi_conn->gen_pdu.resp_buf + pld_len;
if (likely(cmd->io_cmd_in_list)) {
cmd->io_cmd_in_list = false;
list_del_init(&cmd->io_cmd);
qedi_conn->active_cmd_count--;
}
memset(task_ctx, '\0', sizeof(*task_ctx));
__iscsi_complete_pdu(conn, (struct iscsi_hdr *)resp_hdr_ptr,
qedi_conn->gen_pdu.resp_buf,
(qedi_conn->gen_pdu.resp_wr_ptr -
qedi_conn->gen_pdu.resp_buf));
spin_unlock(&session->back_lock);
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_TID,
"Freeing tid=0x%x for cid=0x%x\n",
cmd->task_id, qedi_conn->iscsi_conn_id);
cmd->state = RESPONSE_RECEIVED;
qedi_clear_task_idx(qedi, cmd->task_id);
}
static void qedi_get_rq_bdq_buf(struct qedi_ctx *qedi,
struct iscsi_cqe_unsolicited *cqe,
char *ptr, int len)
{
u16 idx = 0;
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
"pld_len [%d], bdq_prod_idx [%d], idx [%d]\n",
len, qedi->bdq_prod_idx,
(qedi->bdq_prod_idx % qedi->rq_num_entries));
/* Obtain buffer address from rqe_opaque */
idx = cqe->rqe_opaque;
if (idx > (QEDI_BDQ_NUM - 1)) {
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
"wrong idx %d returned by FW, dropping the unsolicited pkt\n",
idx);
return;
}
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
"rqe_opaque [0x%p], idx [%d]\n", cqe->rqe_opaque, idx);
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
"unsol_cqe_type = %d\n", cqe->unsol_cqe_type);
switch (cqe->unsol_cqe_type) {
case ISCSI_CQE_UNSOLICITED_SINGLE:
case ISCSI_CQE_UNSOLICITED_FIRST:
if (len)
memcpy(ptr, (void *)qedi->bdq[idx].buf_addr, len);
break;
case ISCSI_CQE_UNSOLICITED_MIDDLE:
case ISCSI_CQE_UNSOLICITED_LAST:
break;
default:
break;
}
}
static void qedi_put_rq_bdq_buf(struct qedi_ctx *qedi,
struct iscsi_cqe_unsolicited *cqe,
int count)
{
u16 tmp;
u16 idx = 0;
struct scsi_bd *pbl;
/* Obtain buffer address from rqe_opaque */
idx = cqe->rqe_opaque;
if (idx > (QEDI_BDQ_NUM - 1)) {
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
"wrong idx %d returned by FW, dropping the unsolicited pkt\n",
idx);
return;
}
pbl = (struct scsi_bd *)qedi->bdq_pbl;
pbl += (qedi->bdq_prod_idx % qedi->rq_num_entries);
pbl->address.hi = cpu_to_le32(QEDI_U64_HI(qedi->bdq[idx].buf_dma));
pbl->address.lo = cpu_to_le32(QEDI_U64_LO(qedi->bdq[idx].buf_dma));
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
"pbl [0x%p] pbl->address hi [0x%llx] lo [0x%llx] idx [%d]\n",
pbl, pbl->address.hi, pbl->address.lo, idx);
pbl->opaque.iscsi_opaque.reserved_zero[0] = 0;
pbl->opaque.iscsi_opaque.reserved_zero[1] = 0;
pbl->opaque.iscsi_opaque.reserved_zero[2] = 0;
pbl->opaque.iscsi_opaque.opaque = cpu_to_le32(idx);
/* Increment producer to let f/w know we've handled the frame */
qedi->bdq_prod_idx += count;
writew(qedi->bdq_prod_idx, qedi->bdq_primary_prod);
tmp = readw(qedi->bdq_primary_prod);
writew(qedi->bdq_prod_idx, qedi->bdq_secondary_prod);
tmp = readw(qedi->bdq_secondary_prod);
}
static void qedi_unsol_pdu_adjust_bdq(struct qedi_ctx *qedi,
struct iscsi_cqe_unsolicited *cqe,
u32 pdu_len, u32 num_bdqs,
char *bdq_data)
{
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
"num_bdqs [%d]\n", num_bdqs);
qedi_get_rq_bdq_buf(qedi, cqe, bdq_data, pdu_len);
qedi_put_rq_bdq_buf(qedi, cqe, (num_bdqs + 1));
}
static int qedi_process_nopin_mesg(struct qedi_ctx *qedi,
union iscsi_cqe *cqe,
struct iscsi_task *task,
struct qedi_conn *qedi_conn, u16 que_idx)
{
struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
struct iscsi_session *session = conn->session;
struct iscsi_nop_in_hdr *cqe_nop_in;
struct iscsi_nopin *hdr;
struct qedi_cmd *cmd;
int tgt_async_nop = 0;
u32 lun[2];
u32 pdu_len, num_bdqs;
char bdq_data[QEDI_BDQ_BUF_SIZE];
unsigned long flags;
spin_lock_bh(&session->back_lock);
cqe_nop_in = &cqe->cqe_common.iscsi_hdr.nop_in;
pdu_len = cqe_nop_in->hdr_second_dword &
ISCSI_NOP_IN_HDR_DATA_SEG_LEN_MASK;
num_bdqs = pdu_len / QEDI_BDQ_BUF_SIZE;
hdr = (struct iscsi_nopin *)&qedi_conn->gen_pdu.resp_hdr;
memset(hdr, 0, sizeof(struct iscsi_hdr));
hdr->opcode = cqe_nop_in->opcode;
hdr->max_cmdsn = cpu_to_be32(cqe_nop_in->max_cmd_sn);
hdr->exp_cmdsn = cpu_to_be32(cqe_nop_in->exp_cmd_sn);
hdr->statsn = cpu_to_be32(cqe_nop_in->stat_sn);
hdr->ttt = cpu_to_be32(cqe_nop_in->ttt);
if (cqe->cqe_common.cqe_type == ISCSI_CQE_TYPE_UNSOLICITED) {
spin_lock_irqsave(&qedi->hba_lock, flags);
qedi_unsol_pdu_adjust_bdq(qedi, &cqe->cqe_unsolicited,
pdu_len, num_bdqs, bdq_data);
hdr->itt = RESERVED_ITT;
tgt_async_nop = 1;
spin_unlock_irqrestore(&qedi->hba_lock, flags);
goto done;
}
/* Response to one of our nop-outs */
if (task) {
cmd = task->dd_data;
hdr->flags = ISCSI_FLAG_CMD_FINAL;
hdr->itt = build_itt(cqe->cqe_solicited.itid,
conn->session->age);
lun[0] = 0xffffffff;
lun[1] = 0xffffffff;
memcpy(&hdr->lun, lun, sizeof(struct scsi_lun));
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_TID,
"Freeing tid=0x%x for cid=0x%x\n",
cmd->task_id, qedi_conn->iscsi_conn_id);
cmd->state = RESPONSE_RECEIVED;
spin_lock(&qedi_conn->list_lock);
if (likely(cmd->io_cmd_in_list)) {
cmd->io_cmd_in_list = false;
list_del_init(&cmd->io_cmd);
qedi_conn->active_cmd_count--;
}
spin_unlock(&qedi_conn->list_lock);
qedi_clear_task_idx(qedi, cmd->task_id);
}
done:
__iscsi_complete_pdu(conn, (struct iscsi_hdr *)hdr, bdq_data, pdu_len);
spin_unlock_bh(&session->back_lock);
return tgt_async_nop;
}
static void qedi_process_async_mesg(struct qedi_ctx *qedi,
union iscsi_cqe *cqe,
struct iscsi_task *task,
struct qedi_conn *qedi_conn,
u16 que_idx)
{
struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
struct iscsi_session *session = conn->session;
struct iscsi_async_msg_hdr *cqe_async_msg;
struct iscsi_async *resp_hdr;
u32 lun[2];
u32 pdu_len, num_bdqs;
char bdq_data[QEDI_BDQ_BUF_SIZE];
unsigned long flags;
spin_lock_bh(&session->back_lock);
cqe_async_msg = &cqe->cqe_common.iscsi_hdr.async_msg;
pdu_len = cqe_async_msg->hdr_second_dword &
ISCSI_ASYNC_MSG_HDR_DATA_SEG_LEN_MASK;
num_bdqs = pdu_len / QEDI_BDQ_BUF_SIZE;
if (cqe->cqe_common.cqe_type == ISCSI_CQE_TYPE_UNSOLICITED) {
spin_lock_irqsave(&qedi->hba_lock, flags);
qedi_unsol_pdu_adjust_bdq(qedi, &cqe->cqe_unsolicited,
pdu_len, num_bdqs, bdq_data);
spin_unlock_irqrestore(&qedi->hba_lock, flags);
}
resp_hdr = (struct iscsi_async *)&qedi_conn->gen_pdu.resp_hdr;
memset(resp_hdr, 0, sizeof(struct iscsi_hdr));
resp_hdr->opcode = cqe_async_msg->opcode;
resp_hdr->flags = 0x80;
lun[0] = cpu_to_be32(cqe_async_msg->lun.lo);
lun[1] = cpu_to_be32(cqe_async_msg->lun.hi);
memcpy(&resp_hdr->lun, lun, sizeof(struct scsi_lun));
resp_hdr->exp_cmdsn = cpu_to_be32(cqe_async_msg->exp_cmd_sn);
resp_hdr->max_cmdsn = cpu_to_be32(cqe_async_msg->max_cmd_sn);
resp_hdr->statsn = cpu_to_be32(cqe_async_msg->stat_sn);
resp_hdr->async_event = cqe_async_msg->async_event;
resp_hdr->async_vcode = cqe_async_msg->async_vcode;
resp_hdr->param1 = cpu_to_be16(cqe_async_msg->param1_rsrv);
resp_hdr->param2 = cpu_to_be16(cqe_async_msg->param2_rsrv);
resp_hdr->param3 = cpu_to_be16(cqe_async_msg->param3_rsrv);
__iscsi_complete_pdu(conn, (struct iscsi_hdr *)resp_hdr, bdq_data,
pdu_len);
spin_unlock_bh(&session->back_lock);
}
static void qedi_process_reject_mesg(struct qedi_ctx *qedi,
union iscsi_cqe *cqe,
struct iscsi_task *task,
struct qedi_conn *qedi_conn,
uint16_t que_idx)
{
struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
struct iscsi_session *session = conn->session;
struct iscsi_reject_hdr *cqe_reject;
struct iscsi_reject *hdr;
u32 pld_len, num_bdqs;
unsigned long flags;
spin_lock_bh(&session->back_lock);
cqe_reject = &cqe->cqe_common.iscsi_hdr.reject;
pld_len = cqe_reject->hdr_second_dword &
ISCSI_REJECT_HDR_DATA_SEG_LEN_MASK;
num_bdqs = pld_len / QEDI_BDQ_BUF_SIZE;
if (cqe->cqe_common.cqe_type == ISCSI_CQE_TYPE_UNSOLICITED) {
spin_lock_irqsave(&qedi->hba_lock, flags);
qedi_unsol_pdu_adjust_bdq(qedi, &cqe->cqe_unsolicited,
pld_len, num_bdqs, conn->data);
spin_unlock_irqrestore(&qedi->hba_lock, flags);
}
hdr = (struct iscsi_reject *)&qedi_conn->gen_pdu.resp_hdr;
memset(hdr, 0, sizeof(struct iscsi_hdr));
hdr->opcode = cqe_reject->opcode;
hdr->reason = cqe_reject->hdr_reason;
hdr->flags = cqe_reject->hdr_flags;
hton24(hdr->dlength, (cqe_reject->hdr_second_dword &
ISCSI_REJECT_HDR_DATA_SEG_LEN_MASK));
hdr->max_cmdsn = cpu_to_be32(cqe_reject->max_cmd_sn);
hdr->exp_cmdsn = cpu_to_be32(cqe_reject->exp_cmd_sn);
hdr->statsn = cpu_to_be32(cqe_reject->stat_sn);
hdr->ffffffff = cpu_to_be32(0xffffffff);
__iscsi_complete_pdu(conn, (struct iscsi_hdr *)hdr,
conn->data, pld_len);
spin_unlock_bh(&session->back_lock);
}
static void qedi_scsi_completion(struct qedi_ctx *qedi,
union iscsi_cqe *cqe,
struct iscsi_task *task,
struct iscsi_conn *conn)
{
struct scsi_cmnd *sc_cmd;
struct qedi_cmd *cmd = task->dd_data;
struct iscsi_session *session = conn->session;
struct iscsi_scsi_rsp *hdr;
struct iscsi_data_in_hdr *cqe_data_in;
int datalen = 0;
struct qedi_conn *qedi_conn;
u32 iscsi_cid;
u8 cqe_err_bits = 0;
iscsi_cid = cqe->cqe_common.conn_id;
qedi_conn = qedi->cid_que.conn_cid_tbl[iscsi_cid];
cqe_data_in = &cqe->cqe_common.iscsi_hdr.data_in;
cqe_err_bits =
cqe->cqe_common.error_bitmap.error_bits.cqe_error_status_bits;
spin_lock_bh(&session->back_lock);
/* get the scsi command */
sc_cmd = cmd->scsi_cmd;
if (!sc_cmd) {
QEDI_WARN(&qedi->dbg_ctx, "sc_cmd is NULL!\n");
goto error;
}
if (!sc_cmd->SCp.ptr) {
QEDI_WARN(&qedi->dbg_ctx,
"SCp.ptr is NULL, returned in another context.\n");
goto error;
}
if (!sc_cmd->request) {
QEDI_WARN(&qedi->dbg_ctx,
"sc_cmd->request is NULL, sc_cmd=%p.\n",
sc_cmd);
goto error;
}
if (!sc_cmd->request->special) {
QEDI_WARN(&qedi->dbg_ctx,
"request->special is NULL so request not valid, sc_cmd=%p.\n",
sc_cmd);
goto error;
}
if (!sc_cmd->request->q) {
QEDI_WARN(&qedi->dbg_ctx,
"request->q is NULL so request is not valid, sc_cmd=%p.\n",
sc_cmd);
goto error;
}
qedi_iscsi_unmap_sg_list(cmd);
hdr = (struct iscsi_scsi_rsp *)task->hdr;
hdr->opcode = cqe_data_in->opcode;
hdr->max_cmdsn = cpu_to_be32(cqe_data_in->max_cmd_sn);
hdr->exp_cmdsn = cpu_to_be32(cqe_data_in->exp_cmd_sn);
hdr->itt = build_itt(cqe->cqe_solicited.itid, conn->session->age);
hdr->response = cqe_data_in->reserved1;
hdr->cmd_status = cqe_data_in->status_rsvd;
hdr->flags = cqe_data_in->flags;
hdr->residual_count = cpu_to_be32(cqe_data_in->residual_count);
if (hdr->cmd_status == SAM_STAT_CHECK_CONDITION) {
datalen = cqe_data_in->reserved2 &
ISCSI_COMMON_HDR_DATA_SEG_LEN_MASK;
memcpy((char *)conn->data, (char *)cmd->sense_buffer, datalen);
}
/* If f/w reports data underrun err then set residual to IO transfer
* length, set Underrun flag and clear Overrun flag explicitly
*/
if (unlikely(cqe_err_bits &&
GET_FIELD(cqe_err_bits, CQE_ERROR_BITMAP_UNDER_RUN_ERR))) {
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_INFO,
"Under flow itt=0x%x proto flags=0x%x tid=0x%x cid 0x%x fw resid 0x%x sc dlen 0x%x\n",
hdr->itt, cqe_data_in->flags, cmd->task_id,
qedi_conn->iscsi_conn_id, hdr->residual_count,
scsi_bufflen(sc_cmd));
hdr->residual_count = cpu_to_be32(scsi_bufflen(sc_cmd));
hdr->flags |= ISCSI_FLAG_CMD_UNDERFLOW;
hdr->flags &= (~ISCSI_FLAG_CMD_OVERFLOW);
}
spin_lock(&qedi_conn->list_lock);
if (likely(cmd->io_cmd_in_list)) {
cmd->io_cmd_in_list = false;
list_del_init(&cmd->io_cmd);
qedi_conn->active_cmd_count--;
}
spin_unlock(&qedi_conn->list_lock);
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_TID,
"Freeing tid=0x%x for cid=0x%x\n",
cmd->task_id, qedi_conn->iscsi_conn_id);
cmd->state = RESPONSE_RECEIVED;
if (qedi_io_tracing)
qedi_trace_io(qedi, task, cmd->task_id, QEDI_IO_TRACE_RSP);
qedi_clear_task_idx(qedi, cmd->task_id);
__iscsi_complete_pdu(conn, (struct iscsi_hdr *)hdr,
conn->data, datalen);
error:
spin_unlock_bh(&session->back_lock);
}
static void qedi_mtask_completion(struct qedi_ctx *qedi,
union iscsi_cqe *cqe,
struct iscsi_task *task,
struct qedi_conn *conn, uint16_t que_idx)
{
struct iscsi_conn *iscsi_conn;
u32 hdr_opcode;
hdr_opcode = cqe->cqe_common.iscsi_hdr.common.hdr_first_byte;
iscsi_conn = conn->cls_conn->dd_data;
switch (hdr_opcode) {
case ISCSI_OPCODE_SCSI_RESPONSE:
case ISCSI_OPCODE_DATA_IN:
qedi_scsi_completion(qedi, cqe, task, iscsi_conn);
break;
case ISCSI_OPCODE_LOGIN_RESPONSE:
qedi_process_login_resp(qedi, cqe, task, conn);
break;
case ISCSI_OPCODE_TMF_RESPONSE:
qedi_process_tmf_resp(qedi, cqe, task, conn);
break;
case ISCSI_OPCODE_TEXT_RESPONSE:
qedi_process_text_resp(qedi, cqe, task, conn);
break;
case ISCSI_OPCODE_LOGOUT_RESPONSE:
qedi_process_logout_resp(qedi, cqe, task, conn);
break;
case ISCSI_OPCODE_NOP_IN:
qedi_process_nopin_mesg(qedi, cqe, task, conn, que_idx);
break;
default:
QEDI_ERR(&qedi->dbg_ctx, "unknown opcode\n");
}
}
static void qedi_process_nopin_local_cmpl(struct qedi_ctx *qedi,
struct iscsi_cqe_solicited *cqe,
struct iscsi_task *task,
struct qedi_conn *qedi_conn)
{
struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
struct iscsi_session *session = conn->session;
struct qedi_cmd *cmd = task->dd_data;
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_UNSOL,
"itid=0x%x, cmd task id=0x%x\n",
cqe->itid, cmd->task_id);
cmd->state = RESPONSE_RECEIVED;
qedi_clear_task_idx(qedi, cmd->task_id);
spin_lock_bh(&session->back_lock);
__iscsi_put_task(task);
spin_unlock_bh(&session->back_lock);
}
static void qedi_process_cmd_cleanup_resp(struct qedi_ctx *qedi,
struct iscsi_cqe_solicited *cqe,
struct iscsi_task *task,
struct iscsi_conn *conn)
{
struct qedi_work_map *work, *work_tmp;
u32 proto_itt = cqe->itid;
u32 ptmp_itt = 0;
itt_t protoitt = 0;
int found = 0;
struct qedi_cmd *qedi_cmd = NULL;
u32 rtid = 0;
u32 iscsi_cid;
struct qedi_conn *qedi_conn;
struct qedi_cmd *dbg_cmd;
struct iscsi_task *mtask;
struct iscsi_tm *tmf_hdr = NULL;
iscsi_cid = cqe->conn_id;
qedi_conn = qedi->cid_que.conn_cid_tbl[iscsi_cid];
/* Based on this itt get the corresponding qedi_cmd */
spin_lock_bh(&qedi_conn->tmf_work_lock);
list_for_each_entry_safe(work, work_tmp, &qedi_conn->tmf_work_list,
list) {
if (work->rtid == proto_itt) {
/* We found the command */
qedi_cmd = work->qedi_cmd;
if (!qedi_cmd->list_tmf_work) {
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"TMF work not found, cqe->tid=0x%x, cid=0x%x\n",
proto_itt, qedi_conn->iscsi_conn_id);
WARN_ON(1);
}
found = 1;
mtask = qedi_cmd->task;
tmf_hdr = (struct iscsi_tm *)mtask->hdr;
rtid = work->rtid;
list_del_init(&work->list);
kfree(work);
qedi_cmd->list_tmf_work = NULL;
}
}
spin_unlock_bh(&qedi_conn->tmf_work_lock);
if (found) {
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"TMF work, cqe->tid=0x%x, tmf flags=0x%x, cid=0x%x\n",
proto_itt, tmf_hdr->flags, qedi_conn->iscsi_conn_id);
if ((tmf_hdr->flags & ISCSI_FLAG_TM_FUNC_MASK) ==
ISCSI_TM_FUNC_ABORT_TASK) {
spin_lock_bh(&conn->session->back_lock);
protoitt = build_itt(get_itt(tmf_hdr->rtt),
conn->session->age);
task = iscsi_itt_to_task(conn, protoitt);
spin_unlock_bh(&conn->session->back_lock);
if (!task) {
QEDI_NOTICE(&qedi->dbg_ctx,
"IO task completed, tmf rtt=0x%x, cid=0x%x\n",
get_itt(tmf_hdr->rtt),
qedi_conn->iscsi_conn_id);
return;
}
dbg_cmd = task->dd_data;
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"Abort tmf rtt=0x%x, i/o itt=0x%x, i/o tid=0x%x, cid=0x%x\n",
get_itt(tmf_hdr->rtt), get_itt(task->itt),
dbg_cmd->task_id, qedi_conn->iscsi_conn_id);
if (qedi_cmd->state == CLEANUP_WAIT_FAILED)
qedi_cmd->state = CLEANUP_RECV;
qedi_clear_task_idx(qedi_conn->qedi, rtid);
spin_lock(&qedi_conn->list_lock);
list_del_init(&dbg_cmd->io_cmd);
qedi_conn->active_cmd_count--;
spin_unlock(&qedi_conn->list_lock);
qedi_cmd->state = CLEANUP_RECV;
wake_up_interruptible(&qedi_conn->wait_queue);
}
} else if (qedi_conn->cmd_cleanup_req > 0) {
spin_lock_bh(&conn->session->back_lock);
qedi_get_proto_itt(qedi, cqe->itid, &ptmp_itt);
protoitt = build_itt(ptmp_itt, conn->session->age);
task = iscsi_itt_to_task(conn, protoitt);
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"cleanup io itid=0x%x, protoitt=0x%x, cmd_cleanup_cmpl=%d, cid=0x%x\n",
cqe->itid, protoitt, qedi_conn->cmd_cleanup_cmpl,
qedi_conn->iscsi_conn_id);
spin_unlock_bh(&conn->session->back_lock);
if (!task) {
QEDI_NOTICE(&qedi->dbg_ctx,
"task is null, itid=0x%x, cid=0x%x\n",
cqe->itid, qedi_conn->iscsi_conn_id);
return;
}
qedi_conn->cmd_cleanup_cmpl++;
wake_up(&qedi_conn->wait_queue);
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_TID,
"Freeing tid=0x%x for cid=0x%x\n",
cqe->itid, qedi_conn->iscsi_conn_id);
qedi_clear_task_idx(qedi_conn->qedi, cqe->itid);
} else {
qedi_get_proto_itt(qedi, cqe->itid, &ptmp_itt);
protoitt = build_itt(ptmp_itt, conn->session->age);
task = iscsi_itt_to_task(conn, protoitt);
QEDI_ERR(&qedi->dbg_ctx,
"Delayed or untracked cleanup response, itt=0x%x, tid=0x%x, cid=0x%x, task=%p\n",
protoitt, cqe->itid, qedi_conn->iscsi_conn_id, task);
}
}
void qedi_fp_process_cqes(struct qedi_work *work)
{
struct qedi_ctx *qedi = work->qedi;
union iscsi_cqe *cqe = &work->cqe;
struct iscsi_task *task = NULL;
struct iscsi_nopout *nopout_hdr;
struct qedi_conn *q_conn;
struct iscsi_conn *conn;
struct qedi_cmd *qedi_cmd;
u32 comp_type;
u32 iscsi_cid;
u32 hdr_opcode;
u16 que_idx = work->que_idx;
u8 cqe_err_bits = 0;
comp_type = cqe->cqe_common.cqe_type;
hdr_opcode = cqe->cqe_common.iscsi_hdr.common.hdr_first_byte;
cqe_err_bits =
cqe->cqe_common.error_bitmap.error_bits.cqe_error_status_bits;
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
"fw_cid=0x%x, cqe type=0x%x, opcode=0x%x\n",
cqe->cqe_common.conn_id, comp_type, hdr_opcode);
if (comp_type >= MAX_ISCSI_CQES_TYPE) {
QEDI_WARN(&qedi->dbg_ctx, "Invalid CqE type\n");
return;
}
iscsi_cid = cqe->cqe_common.conn_id;
q_conn = qedi->cid_que.conn_cid_tbl[iscsi_cid];
if (!q_conn) {
QEDI_WARN(&qedi->dbg_ctx,
"Session no longer exists for cid=0x%x!!\n",
iscsi_cid);
return;
}
conn = q_conn->cls_conn->dd_data;
if (unlikely(cqe_err_bits &&
GET_FIELD(cqe_err_bits,
CQE_ERROR_BITMAP_DATA_DIGEST_ERR))) {
iscsi_conn_failure(conn, ISCSI_ERR_DATA_DGST);
return;
}
switch (comp_type) {
case ISCSI_CQE_TYPE_SOLICITED:
case ISCSI_CQE_TYPE_SOLICITED_WITH_SENSE:
qedi_cmd = container_of(work, struct qedi_cmd, cqe_work);
task = qedi_cmd->task;
if (!task) {
QEDI_WARN(&qedi->dbg_ctx, "task is NULL\n");
return;
}
/* Process NOPIN local completion */
nopout_hdr = (struct iscsi_nopout *)task->hdr;
if ((nopout_hdr->itt == RESERVED_ITT) &&
(cqe->cqe_solicited.itid != (u16)RESERVED_ITT)) {
qedi_process_nopin_local_cmpl(qedi, &cqe->cqe_solicited,
task, q_conn);
} else {
cqe->cqe_solicited.itid =
qedi_get_itt(cqe->cqe_solicited);
/* Process other solicited responses */
qedi_mtask_completion(qedi, cqe, task, q_conn, que_idx);
}
break;
case ISCSI_CQE_TYPE_UNSOLICITED:
switch (hdr_opcode) {
case ISCSI_OPCODE_NOP_IN:
qedi_process_nopin_mesg(qedi, cqe, task, q_conn,
que_idx);
break;
case ISCSI_OPCODE_ASYNC_MSG:
qedi_process_async_mesg(qedi, cqe, task, q_conn,
que_idx);
break;
case ISCSI_OPCODE_REJECT:
qedi_process_reject_mesg(qedi, cqe, task, q_conn,
que_idx);
break;
}
goto exit_fp_process;
case ISCSI_CQE_TYPE_DUMMY:
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM, "Dummy CqE\n");
goto exit_fp_process;
case ISCSI_CQE_TYPE_TASK_CLEANUP:
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM, "CleanUp CqE\n");
qedi_process_cmd_cleanup_resp(qedi, &cqe->cqe_solicited, task,
conn);
goto exit_fp_process;
default:
QEDI_ERR(&qedi->dbg_ctx, "Error cqe.\n");
break;
}
exit_fp_process:
return;
}
static void qedi_ring_doorbell(struct qedi_conn *qedi_conn)
{
struct iscsi_db_data dbell = { 0 };
dbell.agg_flags = 0;
dbell.params |= DB_DEST_XCM << ISCSI_DB_DATA_DEST_SHIFT;
dbell.params |= DB_AGG_CMD_SET << ISCSI_DB_DATA_AGG_CMD_SHIFT;
dbell.params |=
DQ_XCM_ISCSI_SQ_PROD_CMD << ISCSI_DB_DATA_AGG_VAL_SEL_SHIFT;
dbell.sq_prod = qedi_conn->ep->fw_sq_prod_idx;
writel(*(u32 *)&dbell, qedi_conn->ep->p_doorbell);
/* Make sure fw write idx is coherent, and include both memory barriers
* as a failsafe as for some architectures the call is the same but on
* others they are two different assembly operations.
*/
wmb();
mmiowb();
QEDI_INFO(&qedi_conn->qedi->dbg_ctx, QEDI_LOG_MP_REQ,
"prod_idx=0x%x, fw_prod_idx=0x%x, cid=0x%x\n",
qedi_conn->ep->sq_prod_idx, qedi_conn->ep->fw_sq_prod_idx,
qedi_conn->iscsi_conn_id);
}
static u16 qedi_get_wqe_idx(struct qedi_conn *qedi_conn)
{
struct qedi_endpoint *ep;
u16 rval;
ep = qedi_conn->ep;
rval = ep->sq_prod_idx;
/* Increament SQ index */
ep->sq_prod_idx++;
ep->fw_sq_prod_idx++;
if (ep->sq_prod_idx == QEDI_SQ_SIZE)
ep->sq_prod_idx = 0;
return rval;
}
int qedi_send_iscsi_login(struct qedi_conn *qedi_conn,
struct iscsi_task *task)
{
struct iscsi_login_req_hdr login_req_pdu_header;
struct scsi_sgl_task_params tx_sgl_task_params;
struct scsi_sgl_task_params rx_sgl_task_params;
struct iscsi_task_params task_params;
struct e4_iscsi_task_context *fw_task_ctx;
struct qedi_ctx *qedi = qedi_conn->qedi;
struct iscsi_login_req *login_hdr;
struct scsi_sge *resp_sge = NULL;
struct qedi_cmd *qedi_cmd;
struct qedi_endpoint *ep;
s16 tid = 0;
u16 sq_idx = 0;
int rval = 0;
resp_sge = (struct scsi_sge *)qedi_conn->gen_pdu.resp_bd_tbl;
qedi_cmd = (struct qedi_cmd *)task->dd_data;
ep = qedi_conn->ep;
login_hdr = (struct iscsi_login_req *)task->hdr;
tid = qedi_get_task_idx(qedi);
if (tid == -1)
return -ENOMEM;
fw_task_ctx =
(struct e4_iscsi_task_context *)qedi_get_task_mem(&qedi->tasks,
tid);
memset(fw_task_ctx, 0, sizeof(struct e4_iscsi_task_context));
qedi_cmd->task_id = tid;
memset(&task_params, 0, sizeof(task_params));
memset(&login_req_pdu_header, 0, sizeof(login_req_pdu_header));
memset(&tx_sgl_task_params, 0, sizeof(tx_sgl_task_params));
memset(&rx_sgl_task_params, 0, sizeof(rx_sgl_task_params));
/* Update header info */
login_req_pdu_header.opcode = login_hdr->opcode;
login_req_pdu_header.version_min = login_hdr->min_version;
login_req_pdu_header.version_max = login_hdr->max_version;
login_req_pdu_header.flags_attr = login_hdr->flags;
login_req_pdu_header.isid_tabc = swab32p((u32 *)login_hdr->isid);
login_req_pdu_header.isid_d = swab16p((u16 *)&login_hdr->isid[4]);
login_req_pdu_header.tsih = login_hdr->tsih;
login_req_pdu_header.hdr_second_dword = ntoh24(login_hdr->dlength);
qedi_update_itt_map(qedi, tid, task->itt, qedi_cmd);
login_req_pdu_header.itt = qedi_set_itt(tid, get_itt(task->itt));
login_req_pdu_header.cid = qedi_conn->iscsi_conn_id;
login_req_pdu_header.cmd_sn = be32_to_cpu(login_hdr->cmdsn);
login_req_pdu_header.exp_stat_sn = be32_to_cpu(login_hdr->exp_statsn);
login_req_pdu_header.exp_stat_sn = 0;
/* Fill tx AHS and rx buffer */
tx_sgl_task_params.sgl =
(struct scsi_sge *)qedi_conn->gen_pdu.req_bd_tbl;
tx_sgl_task_params.sgl_phys_addr.lo =
(u32)(qedi_conn->gen_pdu.req_dma_addr);
tx_sgl_task_params.sgl_phys_addr.hi =
(u32)((u64)qedi_conn->gen_pdu.req_dma_addr >> 32);
tx_sgl_task_params.total_buffer_size = ntoh24(login_hdr->dlength);
tx_sgl_task_params.num_sges = 1;
rx_sgl_task_params.sgl =
(struct scsi_sge *)qedi_conn->gen_pdu.resp_bd_tbl;
rx_sgl_task_params.sgl_phys_addr.lo =
(u32)(qedi_conn->gen_pdu.resp_dma_addr);
rx_sgl_task_params.sgl_phys_addr.hi =
(u32)((u64)qedi_conn->gen_pdu.resp_dma_addr >> 32);
rx_sgl_task_params.total_buffer_size = resp_sge->sge_len;
rx_sgl_task_params.num_sges = 1;
/* Fill fw input params */
task_params.context = fw_task_ctx;
task_params.conn_icid = (u16)qedi_conn->iscsi_conn_id;
task_params.itid = tid;
task_params.cq_rss_number = 0;
task_params.tx_io_size = ntoh24(login_hdr->dlength);
task_params.rx_io_size = resp_sge->sge_len;
sq_idx = qedi_get_wqe_idx(qedi_conn);
task_params.sqe = &ep->sq[sq_idx];
memset(task_params.sqe, 0, sizeof(struct iscsi_wqe));
rval = init_initiator_login_request_task(&task_params,
&login_req_pdu_header,
&tx_sgl_task_params,
&rx_sgl_task_params);
if (rval)
return -1;
spin_lock(&qedi_conn->list_lock);
list_add_tail(&qedi_cmd->io_cmd, &qedi_conn->active_cmd_list);
qedi_cmd->io_cmd_in_list = true;
qedi_conn->active_cmd_count++;
spin_unlock(&qedi_conn->list_lock);
qedi_ring_doorbell(qedi_conn);
return 0;
}
int qedi_send_iscsi_logout(struct qedi_conn *qedi_conn,
struct iscsi_task *task)
{
struct iscsi_logout_req_hdr logout_pdu_header;
struct scsi_sgl_task_params tx_sgl_task_params;
struct scsi_sgl_task_params rx_sgl_task_params;
struct iscsi_task_params task_params;
struct e4_iscsi_task_context *fw_task_ctx;
struct iscsi_logout *logout_hdr = NULL;
struct qedi_ctx *qedi = qedi_conn->qedi;
struct qedi_cmd *qedi_cmd;
struct qedi_endpoint *ep;
s16 tid = 0;
u16 sq_idx = 0;
int rval = 0;
qedi_cmd = (struct qedi_cmd *)task->dd_data;
logout_hdr = (struct iscsi_logout *)task->hdr;
ep = qedi_conn->ep;
tid = qedi_get_task_idx(qedi);
if (tid == -1)
return -ENOMEM;
fw_task_ctx =
(struct e4_iscsi_task_context *)qedi_get_task_mem(&qedi->tasks,
tid);
memset(fw_task_ctx, 0, sizeof(struct e4_iscsi_task_context));
qedi_cmd->task_id = tid;
memset(&task_params, 0, sizeof(task_params));
memset(&logout_pdu_header, 0, sizeof(logout_pdu_header));
memset(&tx_sgl_task_params, 0, sizeof(tx_sgl_task_params));
memset(&rx_sgl_task_params, 0, sizeof(rx_sgl_task_params));
/* Update header info */
logout_pdu_header.opcode = logout_hdr->opcode;
logout_pdu_header.reason_code = 0x80 | logout_hdr->flags;
qedi_update_itt_map(qedi, tid, task->itt, qedi_cmd);
logout_pdu_header.itt = qedi_set_itt(tid, get_itt(task->itt));
logout_pdu_header.exp_stat_sn = be32_to_cpu(logout_hdr->exp_statsn);
logout_pdu_header.cmd_sn = be32_to_cpu(logout_hdr->cmdsn);
logout_pdu_header.cid = qedi_conn->iscsi_conn_id;
/* Fill fw input params */
task_params.context = fw_task_ctx;
task_params.conn_icid = (u16)qedi_conn->iscsi_conn_id;
task_params.itid = tid;
task_params.cq_rss_number = 0;
task_params.tx_io_size = 0;
task_params.rx_io_size = 0;
sq_idx = qedi_get_wqe_idx(qedi_conn);
task_params.sqe = &ep->sq[sq_idx];
memset(task_params.sqe, 0, sizeof(struct iscsi_wqe));
rval = init_initiator_logout_request_task(&task_params,
&logout_pdu_header,
NULL, NULL);
if (rval)
return -1;
spin_lock(&qedi_conn->list_lock);
list_add_tail(&qedi_cmd->io_cmd, &qedi_conn->active_cmd_list);
qedi_cmd->io_cmd_in_list = true;
qedi_conn->active_cmd_count++;
spin_unlock(&qedi_conn->list_lock);
qedi_ring_doorbell(qedi_conn);
return 0;
}
int qedi_cleanup_all_io(struct qedi_ctx *qedi, struct qedi_conn *qedi_conn,
struct iscsi_task *task, bool in_recovery)
{
int rval;
struct iscsi_task *ctask;
struct qedi_cmd *cmd, *cmd_tmp;
struct iscsi_tm *tmf_hdr;
unsigned int lun = 0;
bool lun_reset = false;
struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
struct iscsi_session *session = conn->session;
/* From recovery, task is NULL or from tmf resp valid task */
if (task) {
tmf_hdr = (struct iscsi_tm *)task->hdr;
if ((tmf_hdr->flags & ISCSI_FLAG_TM_FUNC_MASK) ==
ISCSI_TM_FUNC_LOGICAL_UNIT_RESET) {
lun_reset = true;
lun = scsilun_to_int(&tmf_hdr->lun);
}
}
qedi_conn->cmd_cleanup_req = 0;
qedi_conn->cmd_cleanup_cmpl = 0;
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"active_cmd_count=%d, cid=0x%x, in_recovery=%d, lun_reset=%d\n",
qedi_conn->active_cmd_count, qedi_conn->iscsi_conn_id,
in_recovery, lun_reset);
if (lun_reset)
spin_lock_bh(&session->back_lock);
spin_lock(&qedi_conn->list_lock);
list_for_each_entry_safe(cmd, cmd_tmp, &qedi_conn->active_cmd_list,
io_cmd) {
ctask = cmd->task;
if (ctask == task)
continue;
if (lun_reset) {
if (cmd->scsi_cmd && cmd->scsi_cmd->device) {
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"tid=0x%x itt=0x%x scsi_cmd_ptr=%p device=%p task_state=%d cmd_state=0%x cid=0x%x\n",
cmd->task_id, get_itt(ctask->itt),
cmd->scsi_cmd, cmd->scsi_cmd->device,
ctask->state, cmd->state,
qedi_conn->iscsi_conn_id);
if (cmd->scsi_cmd->device->lun != lun)
continue;
}
}
qedi_conn->cmd_cleanup_req++;
qedi_iscsi_cleanup_task(ctask, true);
list_del_init(&cmd->io_cmd);
qedi_conn->active_cmd_count--;
QEDI_WARN(&qedi->dbg_ctx,
"Deleted active cmd list node io_cmd=%p, cid=0x%x\n",
&cmd->io_cmd, qedi_conn->iscsi_conn_id);
}
spin_unlock(&qedi_conn->list_lock);
if (lun_reset)
spin_unlock_bh(&session->back_lock);
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"cmd_cleanup_req=%d, cid=0x%x\n",
qedi_conn->cmd_cleanup_req,
qedi_conn->iscsi_conn_id);
rval = wait_event_interruptible_timeout(qedi_conn->wait_queue,
((qedi_conn->cmd_cleanup_req ==
qedi_conn->cmd_cleanup_cmpl) ||
qedi_conn->ep),
5 * HZ);
if (rval) {
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"i/o cmd_cleanup_req=%d, equal to cmd_cleanup_cmpl=%d, cid=0x%x\n",
qedi_conn->cmd_cleanup_req,
qedi_conn->cmd_cleanup_cmpl,
qedi_conn->iscsi_conn_id);
return 0;
}
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"i/o cmd_cleanup_req=%d, not equal to cmd_cleanup_cmpl=%d, cid=0x%x\n",
qedi_conn->cmd_cleanup_req,
qedi_conn->cmd_cleanup_cmpl,
qedi_conn->iscsi_conn_id);
iscsi_host_for_each_session(qedi->shost,
qedi_mark_device_missing);
qedi_ops->common->drain(qedi->cdev);
/* Enable IOs for all other sessions except current.*/
if (!wait_event_interruptible_timeout(qedi_conn->wait_queue,
(qedi_conn->cmd_cleanup_req ==
qedi_conn->cmd_cleanup_cmpl),
5 * HZ)) {
iscsi_host_for_each_session(qedi->shost,
qedi_mark_device_available);
return -1;
}
iscsi_host_for_each_session(qedi->shost,
qedi_mark_device_available);
return 0;
}
void qedi_clearsq(struct qedi_ctx *qedi, struct qedi_conn *qedi_conn,
struct iscsi_task *task)
{
struct qedi_endpoint *qedi_ep;
int rval;
qedi_ep = qedi_conn->ep;
qedi_conn->cmd_cleanup_req = 0;
qedi_conn->cmd_cleanup_cmpl = 0;
if (!qedi_ep) {
QEDI_WARN(&qedi->dbg_ctx,
"Cannot proceed, ep already disconnected, cid=0x%x\n",
qedi_conn->iscsi_conn_id);
return;
}
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_INFO,
"Clearing SQ for cid=0x%x, conn=%p, ep=%p\n",
qedi_conn->iscsi_conn_id, qedi_conn, qedi_ep);
qedi_ops->clear_sq(qedi->cdev, qedi_ep->handle);
rval = qedi_cleanup_all_io(qedi, qedi_conn, task, true);
if (rval) {
QEDI_ERR(&qedi->dbg_ctx,
"fatal error, need hard reset, cid=0x%x\n",
qedi_conn->iscsi_conn_id);
WARN_ON(1);
}
}
static int qedi_wait_for_cleanup_request(struct qedi_ctx *qedi,
struct qedi_conn *qedi_conn,
struct iscsi_task *task,
struct qedi_cmd *qedi_cmd,
struct qedi_work_map *list_work)
{
struct qedi_cmd *cmd = (struct qedi_cmd *)task->dd_data;
int wait;
wait = wait_event_interruptible_timeout(qedi_conn->wait_queue,
((qedi_cmd->state ==
CLEANUP_RECV) ||
((qedi_cmd->type == TYPEIO) &&
(cmd->state ==
RESPONSE_RECEIVED))),
5 * HZ);
if (!wait) {
qedi_cmd->state = CLEANUP_WAIT_FAILED;
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"Cleanup timedout tid=0x%x, issue connection recovery, cid=0x%x\n",
cmd->task_id, qedi_conn->iscsi_conn_id);
return -1;
}
return 0;
}
static void qedi_tmf_work(struct work_struct *work)
{
struct qedi_cmd *qedi_cmd =
container_of(work, struct qedi_cmd, tmf_work);
struct qedi_conn *qedi_conn = qedi_cmd->conn;
struct qedi_ctx *qedi = qedi_conn->qedi;
struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
struct iscsi_cls_session *cls_sess;
struct qedi_work_map *list_work = NULL;
struct iscsi_task *mtask;
struct qedi_cmd *cmd;
struct iscsi_task *ctask;
struct iscsi_tm *tmf_hdr;
s16 rval = 0;
s16 tid = 0;
mtask = qedi_cmd->task;
tmf_hdr = (struct iscsi_tm *)mtask->hdr;
cls_sess = iscsi_conn_to_session(qedi_conn->cls_conn);
set_bit(QEDI_CONN_FW_CLEANUP, &qedi_conn->flags);
ctask = iscsi_itt_to_task(conn, tmf_hdr->rtt);
if (!ctask || !ctask->sc) {
QEDI_ERR(&qedi->dbg_ctx, "Task already completed\n");
goto abort_ret;
}
cmd = (struct qedi_cmd *)ctask->dd_data;
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_INFO,
"Abort tmf rtt=0x%x, cmd itt=0x%x, cmd tid=0x%x, cid=0x%x\n",
get_itt(tmf_hdr->rtt), get_itt(ctask->itt), cmd->task_id,
qedi_conn->iscsi_conn_id);
if (qedi_do_not_recover) {
QEDI_ERR(&qedi->dbg_ctx, "DONT SEND CLEANUP/ABORT %d\n",
qedi_do_not_recover);
goto abort_ret;
}
list_work = kzalloc(sizeof(*list_work), GFP_ATOMIC);
if (!list_work) {
QEDI_ERR(&qedi->dbg_ctx, "Memory allocation failed\n");
goto abort_ret;
}
qedi_cmd->type = TYPEIO;
list_work->qedi_cmd = qedi_cmd;
list_work->rtid = cmd->task_id;
list_work->state = QEDI_WORK_SCHEDULED;
qedi_cmd->list_tmf_work = list_work;
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"Queue tmf work=%p, list node=%p, cid=0x%x, tmf flags=0x%x\n",
list_work->ptr_tmf_work, list_work, qedi_conn->iscsi_conn_id,
tmf_hdr->flags);
spin_lock_bh(&qedi_conn->tmf_work_lock);
list_add_tail(&list_work->list, &qedi_conn->tmf_work_list);
spin_unlock_bh(&qedi_conn->tmf_work_lock);
qedi_iscsi_cleanup_task(ctask, false);
rval = qedi_wait_for_cleanup_request(qedi, qedi_conn, ctask, qedi_cmd,
list_work);
if (rval == -1) {
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_INFO,
"FW cleanup got escalated, cid=0x%x\n",
qedi_conn->iscsi_conn_id);
goto ldel_exit;
}
tid = qedi_get_task_idx(qedi);
if (tid == -1) {
QEDI_ERR(&qedi->dbg_ctx, "Invalid tid, cid=0x%x\n",
qedi_conn->iscsi_conn_id);
goto ldel_exit;
}
qedi_cmd->task_id = tid;
qedi_send_iscsi_tmf(qedi_conn, qedi_cmd->task);
abort_ret:
clear_bit(QEDI_CONN_FW_CLEANUP, &qedi_conn->flags);
return;
ldel_exit:
spin_lock_bh(&qedi_conn->tmf_work_lock);
if (!qedi_cmd->list_tmf_work) {
list_del_init(&list_work->list);
qedi_cmd->list_tmf_work = NULL;
kfree(list_work);
}
spin_unlock_bh(&qedi_conn->tmf_work_lock);
spin_lock(&qedi_conn->list_lock);
list_del_init(&cmd->io_cmd);
qedi_conn->active_cmd_count--;
spin_unlock(&qedi_conn->list_lock);
clear_bit(QEDI_CONN_FW_CLEANUP, &qedi_conn->flags);
}
static int qedi_send_iscsi_tmf(struct qedi_conn *qedi_conn,
struct iscsi_task *mtask)
{
struct iscsi_tmf_request_hdr tmf_pdu_header;
struct iscsi_task_params task_params;
struct qedi_ctx *qedi = qedi_conn->qedi;
struct e4_iscsi_task_context *fw_task_ctx;
struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
struct iscsi_task *ctask;
struct iscsi_tm *tmf_hdr;
struct qedi_cmd *qedi_cmd;
struct qedi_cmd *cmd;
struct qedi_endpoint *ep;
u32 scsi_lun[2];
s16 tid = 0;
u16 sq_idx = 0;
int rval = 0;
tmf_hdr = (struct iscsi_tm *)mtask->hdr;
qedi_cmd = (struct qedi_cmd *)mtask->dd_data;
ep = qedi_conn->ep;
if (!ep)
return -ENODEV;
tid = qedi_get_task_idx(qedi);
if (tid == -1)
return -ENOMEM;
fw_task_ctx =
(struct e4_iscsi_task_context *)qedi_get_task_mem(&qedi->tasks,
tid);
memset(fw_task_ctx, 0, sizeof(struct e4_iscsi_task_context));
qedi_cmd->task_id = tid;
memset(&task_params, 0, sizeof(task_params));
memset(&tmf_pdu_header, 0, sizeof(tmf_pdu_header));
/* Update header info */
qedi_update_itt_map(qedi, tid, mtask->itt, qedi_cmd);
tmf_pdu_header.itt = qedi_set_itt(tid, get_itt(mtask->itt));
tmf_pdu_header.cmd_sn = be32_to_cpu(tmf_hdr->cmdsn);
memcpy(scsi_lun, &tmf_hdr->lun, sizeof(struct scsi_lun));
tmf_pdu_header.lun.lo = be32_to_cpu(scsi_lun[0]);
tmf_pdu_header.lun.hi = be32_to_cpu(scsi_lun[1]);
if ((tmf_hdr->flags & ISCSI_FLAG_TM_FUNC_MASK) ==
ISCSI_TM_FUNC_ABORT_TASK) {
ctask = iscsi_itt_to_task(conn, tmf_hdr->rtt);
if (!ctask || !ctask->sc) {
QEDI_ERR(&qedi->dbg_ctx,
"Could not get reference task\n");
return 0;
}
cmd = (struct qedi_cmd *)ctask->dd_data;
tmf_pdu_header.rtt =
qedi_set_itt(cmd->task_id,
get_itt(tmf_hdr->rtt));
} else {
tmf_pdu_header.rtt = ISCSI_RESERVED_TAG;
}
tmf_pdu_header.opcode = tmf_hdr->opcode;
tmf_pdu_header.function = tmf_hdr->flags;
tmf_pdu_header.hdr_second_dword = ntoh24(tmf_hdr->dlength);
tmf_pdu_header.ref_cmd_sn = be32_to_cpu(tmf_hdr->refcmdsn);
/* Fill fw input params */
task_params.context = fw_task_ctx;
task_params.conn_icid = (u16)qedi_conn->iscsi_conn_id;
task_params.itid = tid;
task_params.cq_rss_number = 0;
task_params.tx_io_size = 0;
task_params.rx_io_size = 0;
sq_idx = qedi_get_wqe_idx(qedi_conn);
task_params.sqe = &ep->sq[sq_idx];
memset(task_params.sqe, 0, sizeof(struct iscsi_wqe));
rval = init_initiator_tmf_request_task(&task_params,
&tmf_pdu_header);
if (rval)
return -1;
spin_lock(&qedi_conn->list_lock);
list_add_tail(&qedi_cmd->io_cmd, &qedi_conn->active_cmd_list);
qedi_cmd->io_cmd_in_list = true;
qedi_conn->active_cmd_count++;
spin_unlock(&qedi_conn->list_lock);
qedi_ring_doorbell(qedi_conn);
return 0;
}
int qedi_iscsi_abort_work(struct qedi_conn *qedi_conn,
struct iscsi_task *mtask)
{
struct qedi_ctx *qedi = qedi_conn->qedi;
struct iscsi_tm *tmf_hdr;
struct qedi_cmd *qedi_cmd = (struct qedi_cmd *)mtask->dd_data;
s16 tid = 0;
tmf_hdr = (struct iscsi_tm *)mtask->hdr;
qedi_cmd->task = mtask;
/* If abort task then schedule the work and return */
if ((tmf_hdr->flags & ISCSI_FLAG_TM_FUNC_MASK) ==
ISCSI_TM_FUNC_ABORT_TASK) {
qedi_cmd->state = CLEANUP_WAIT;
INIT_WORK(&qedi_cmd->tmf_work, qedi_tmf_work);
queue_work(qedi->tmf_thread, &qedi_cmd->tmf_work);
} else if (((tmf_hdr->flags & ISCSI_FLAG_TM_FUNC_MASK) ==
ISCSI_TM_FUNC_LOGICAL_UNIT_RESET) ||
((tmf_hdr->flags & ISCSI_FLAG_TM_FUNC_MASK) ==
ISCSI_TM_FUNC_TARGET_WARM_RESET) ||
((tmf_hdr->flags & ISCSI_FLAG_TM_FUNC_MASK) ==
ISCSI_TM_FUNC_TARGET_COLD_RESET)) {
tid = qedi_get_task_idx(qedi);
if (tid == -1) {
QEDI_ERR(&qedi->dbg_ctx, "Invalid tid, cid=0x%x\n",
qedi_conn->iscsi_conn_id);
return -1;
}
qedi_cmd->task_id = tid;
qedi_send_iscsi_tmf(qedi_conn, qedi_cmd->task);
} else {
QEDI_ERR(&qedi->dbg_ctx, "Invalid tmf, cid=0x%x\n",
qedi_conn->iscsi_conn_id);
return -1;
}
return 0;
}
int qedi_send_iscsi_text(struct qedi_conn *qedi_conn,
struct iscsi_task *task)
{
struct iscsi_text_request_hdr text_request_pdu_header;
struct scsi_sgl_task_params tx_sgl_task_params;
struct scsi_sgl_task_params rx_sgl_task_params;
struct iscsi_task_params task_params;
struct e4_iscsi_task_context *fw_task_ctx;
struct qedi_ctx *qedi = qedi_conn->qedi;
struct iscsi_text *text_hdr;
struct scsi_sge *req_sge = NULL;
struct scsi_sge *resp_sge = NULL;
struct qedi_cmd *qedi_cmd;
struct qedi_endpoint *ep;
s16 tid = 0;
u16 sq_idx = 0;
int rval = 0;
req_sge = (struct scsi_sge *)qedi_conn->gen_pdu.req_bd_tbl;
resp_sge = (struct scsi_sge *)qedi_conn->gen_pdu.resp_bd_tbl;
qedi_cmd = (struct qedi_cmd *)task->dd_data;
text_hdr = (struct iscsi_text *)task->hdr;
ep = qedi_conn->ep;
tid = qedi_get_task_idx(qedi);
if (tid == -1)
return -ENOMEM;
fw_task_ctx =
(struct e4_iscsi_task_context *)qedi_get_task_mem(&qedi->tasks,
tid);
memset(fw_task_ctx, 0, sizeof(struct e4_iscsi_task_context));
qedi_cmd->task_id = tid;
memset(&task_params, 0, sizeof(task_params));
memset(&text_request_pdu_header, 0, sizeof(text_request_pdu_header));
memset(&tx_sgl_task_params, 0, sizeof(tx_sgl_task_params));
memset(&rx_sgl_task_params, 0, sizeof(rx_sgl_task_params));
/* Update header info */
text_request_pdu_header.opcode = text_hdr->opcode;
text_request_pdu_header.flags_attr = text_hdr->flags;
qedi_update_itt_map(qedi, tid, task->itt, qedi_cmd);
text_request_pdu_header.itt = qedi_set_itt(tid, get_itt(task->itt));
text_request_pdu_header.ttt = text_hdr->ttt;
text_request_pdu_header.cmd_sn = be32_to_cpu(text_hdr->cmdsn);
text_request_pdu_header.exp_stat_sn = be32_to_cpu(text_hdr->exp_statsn);
text_request_pdu_header.hdr_second_dword = ntoh24(text_hdr->dlength);
/* Fill tx AHS and rx buffer */
tx_sgl_task_params.sgl =
(struct scsi_sge *)qedi_conn->gen_pdu.req_bd_tbl;
tx_sgl_task_params.sgl_phys_addr.lo =
(u32)(qedi_conn->gen_pdu.req_dma_addr);
tx_sgl_task_params.sgl_phys_addr.hi =
(u32)((u64)qedi_conn->gen_pdu.req_dma_addr >> 32);
tx_sgl_task_params.total_buffer_size = req_sge->sge_len;
tx_sgl_task_params.num_sges = 1;
rx_sgl_task_params.sgl =
(struct scsi_sge *)qedi_conn->gen_pdu.resp_bd_tbl;
rx_sgl_task_params.sgl_phys_addr.lo =
(u32)(qedi_conn->gen_pdu.resp_dma_addr);
rx_sgl_task_params.sgl_phys_addr.hi =
(u32)((u64)qedi_conn->gen_pdu.resp_dma_addr >> 32);
rx_sgl_task_params.total_buffer_size = resp_sge->sge_len;
rx_sgl_task_params.num_sges = 1;
/* Fill fw input params */
task_params.context = fw_task_ctx;
task_params.conn_icid = (u16)qedi_conn->iscsi_conn_id;
task_params.itid = tid;
task_params.cq_rss_number = 0;
task_params.tx_io_size = ntoh24(text_hdr->dlength);
task_params.rx_io_size = resp_sge->sge_len;
sq_idx = qedi_get_wqe_idx(qedi_conn);
task_params.sqe = &ep->sq[sq_idx];
memset(task_params.sqe, 0, sizeof(struct iscsi_wqe));
rval = init_initiator_text_request_task(&task_params,
&text_request_pdu_header,
&tx_sgl_task_params,
&rx_sgl_task_params);
if (rval)
return -1;
spin_lock(&qedi_conn->list_lock);
list_add_tail(&qedi_cmd->io_cmd, &qedi_conn->active_cmd_list);
qedi_cmd->io_cmd_in_list = true;
qedi_conn->active_cmd_count++;
spin_unlock(&qedi_conn->list_lock);
qedi_ring_doorbell(qedi_conn);
return 0;
}
int qedi_send_iscsi_nopout(struct qedi_conn *qedi_conn,
struct iscsi_task *task,
char *datap, int data_len, int unsol)
{
struct iscsi_nop_out_hdr nop_out_pdu_header;
struct scsi_sgl_task_params tx_sgl_task_params;
struct scsi_sgl_task_params rx_sgl_task_params;
struct iscsi_task_params task_params;
struct qedi_ctx *qedi = qedi_conn->qedi;
struct e4_iscsi_task_context *fw_task_ctx;
struct iscsi_nopout *nopout_hdr;
struct scsi_sge *resp_sge = NULL;
struct qedi_cmd *qedi_cmd;
struct qedi_endpoint *ep;
u32 scsi_lun[2];
s16 tid = 0;
u16 sq_idx = 0;
int rval = 0;
resp_sge = (struct scsi_sge *)qedi_conn->gen_pdu.resp_bd_tbl;
qedi_cmd = (struct qedi_cmd *)task->dd_data;
nopout_hdr = (struct iscsi_nopout *)task->hdr;
ep = qedi_conn->ep;
tid = qedi_get_task_idx(qedi);
if (tid == -1)
return -ENOMEM;
fw_task_ctx =
(struct e4_iscsi_task_context *)qedi_get_task_mem(&qedi->tasks,
tid);
memset(fw_task_ctx, 0, sizeof(struct e4_iscsi_task_context));
qedi_cmd->task_id = tid;
memset(&task_params, 0, sizeof(task_params));
memset(&nop_out_pdu_header, 0, sizeof(nop_out_pdu_header));
memset(&tx_sgl_task_params, 0, sizeof(tx_sgl_task_params));
memset(&rx_sgl_task_params, 0, sizeof(rx_sgl_task_params));
/* Update header info */
nop_out_pdu_header.opcode = nopout_hdr->opcode;
SET_FIELD(nop_out_pdu_header.flags_attr, ISCSI_NOP_OUT_HDR_CONST1, 1);
SET_FIELD(nop_out_pdu_header.flags_attr, ISCSI_NOP_OUT_HDR_RSRV, 0);
memcpy(scsi_lun, &nopout_hdr->lun, sizeof(struct scsi_lun));
nop_out_pdu_header.lun.lo = be32_to_cpu(scsi_lun[0]);
nop_out_pdu_header.lun.hi = be32_to_cpu(scsi_lun[1]);
nop_out_pdu_header.cmd_sn = be32_to_cpu(nopout_hdr->cmdsn);
nop_out_pdu_header.exp_stat_sn = be32_to_cpu(nopout_hdr->exp_statsn);
qedi_update_itt_map(qedi, tid, task->itt, qedi_cmd);
if (nopout_hdr->ttt != ISCSI_TTT_ALL_ONES) {
nop_out_pdu_header.itt = be32_to_cpu(nopout_hdr->itt);
nop_out_pdu_header.ttt = be32_to_cpu(nopout_hdr->ttt);
} else {
nop_out_pdu_header.itt = qedi_set_itt(tid, get_itt(task->itt));
nop_out_pdu_header.ttt = ISCSI_TTT_ALL_ONES;
spin_lock(&qedi_conn->list_lock);
list_add_tail(&qedi_cmd->io_cmd, &qedi_conn->active_cmd_list);
qedi_cmd->io_cmd_in_list = true;
qedi_conn->active_cmd_count++;
spin_unlock(&qedi_conn->list_lock);
}
/* Fill tx AHS and rx buffer */
if (data_len) {
tx_sgl_task_params.sgl =
(struct scsi_sge *)qedi_conn->gen_pdu.req_bd_tbl;
tx_sgl_task_params.sgl_phys_addr.lo =
(u32)(qedi_conn->gen_pdu.req_dma_addr);
tx_sgl_task_params.sgl_phys_addr.hi =
(u32)((u64)qedi_conn->gen_pdu.req_dma_addr >> 32);
tx_sgl_task_params.total_buffer_size = data_len;
tx_sgl_task_params.num_sges = 1;
rx_sgl_task_params.sgl =
(struct scsi_sge *)qedi_conn->gen_pdu.resp_bd_tbl;
rx_sgl_task_params.sgl_phys_addr.lo =
(u32)(qedi_conn->gen_pdu.resp_dma_addr);
rx_sgl_task_params.sgl_phys_addr.hi =
(u32)((u64)qedi_conn->gen_pdu.resp_dma_addr >> 32);
rx_sgl_task_params.total_buffer_size = resp_sge->sge_len;
rx_sgl_task_params.num_sges = 1;
}
/* Fill fw input params */
task_params.context = fw_task_ctx;
task_params.conn_icid = (u16)qedi_conn->iscsi_conn_id;
task_params.itid = tid;
task_params.cq_rss_number = 0;
task_params.tx_io_size = data_len;
task_params.rx_io_size = resp_sge->sge_len;
sq_idx = qedi_get_wqe_idx(qedi_conn);
task_params.sqe = &ep->sq[sq_idx];
memset(task_params.sqe, 0, sizeof(struct iscsi_wqe));
rval = init_initiator_nop_out_task(&task_params,
&nop_out_pdu_header,
&tx_sgl_task_params,
&rx_sgl_task_params);
if (rval)
return -1;
qedi_ring_doorbell(qedi_conn);
return 0;
}
static int qedi_split_bd(struct qedi_cmd *cmd, u64 addr, int sg_len,
int bd_index)
{
struct scsi_sge *bd = cmd->io_tbl.sge_tbl;
int frag_size, sg_frags;
sg_frags = 0;
while (sg_len) {
if (addr % QEDI_PAGE_SIZE)
frag_size =
(QEDI_PAGE_SIZE - (addr % QEDI_PAGE_SIZE));
else
frag_size = (sg_len > QEDI_BD_SPLIT_SZ) ? 0 :
(sg_len % QEDI_BD_SPLIT_SZ);
if (frag_size == 0)
frag_size = QEDI_BD_SPLIT_SZ;
bd[bd_index + sg_frags].sge_addr.lo = (addr & 0xffffffff);
bd[bd_index + sg_frags].sge_addr.hi = (addr >> 32);
bd[bd_index + sg_frags].sge_len = (u16)frag_size;
QEDI_INFO(&cmd->conn->qedi->dbg_ctx, QEDI_LOG_IO,
"split sge %d: addr=%llx, len=%x",
(bd_index + sg_frags), addr, frag_size);
addr += (u64)frag_size;
sg_frags++;
sg_len -= frag_size;
}
return sg_frags;
}
static int qedi_map_scsi_sg(struct qedi_ctx *qedi, struct qedi_cmd *cmd)
{
struct scsi_cmnd *sc = cmd->scsi_cmd;
struct scsi_sge *bd = cmd->io_tbl.sge_tbl;
struct scatterlist *sg;
int byte_count = 0;
int bd_count = 0;
int sg_count;
int sg_len;
int sg_frags;
u64 addr, end_addr;
int i;
WARN_ON(scsi_sg_count(sc) > QEDI_ISCSI_MAX_BDS_PER_CMD);
sg_count = dma_map_sg(&qedi->pdev->dev, scsi_sglist(sc),
scsi_sg_count(sc), sc->sc_data_direction);
/*
* New condition to send single SGE as cached-SGL.
* Single SGE with length less than 64K.
*/
sg = scsi_sglist(sc);
if ((sg_count == 1) && (sg_dma_len(sg) <= MAX_SGLEN_FOR_CACHESGL)) {
sg_len = sg_dma_len(sg);
addr = (u64)sg_dma_address(sg);
bd[bd_count].sge_addr.lo = (addr & 0xffffffff);
bd[bd_count].sge_addr.hi = (addr >> 32);
bd[bd_count].sge_len = (u16)sg_len;
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_IO,
"single-cashed-sgl: bd_count:%d addr=%llx, len=%x",
sg_count, addr, sg_len);
return ++bd_count;
}
scsi_for_each_sg(sc, sg, sg_count, i) {
sg_len = sg_dma_len(sg);
addr = (u64)sg_dma_address(sg);
end_addr = (addr + sg_len);
/*
* first sg elem in the 'list',
* check if end addr is page-aligned.
*/
if ((i == 0) && (sg_count > 1) && (end_addr % QEDI_PAGE_SIZE))
cmd->use_slowpath = true;
/*
* last sg elem in the 'list',
* check if start addr is page-aligned.
*/
else if ((i == (sg_count - 1)) &&
(sg_count > 1) && (addr % QEDI_PAGE_SIZE))
cmd->use_slowpath = true;
/*
* middle sg elements in list,
* check if start and end addr is page-aligned
*/
else if ((i != 0) && (i != (sg_count - 1)) &&
((addr % QEDI_PAGE_SIZE) ||
(end_addr % QEDI_PAGE_SIZE)))
cmd->use_slowpath = true;
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_IO, "sg[%d] size=0x%x",
i, sg_len);
if (sg_len > QEDI_BD_SPLIT_SZ) {
sg_frags = qedi_split_bd(cmd, addr, sg_len, bd_count);
} else {
sg_frags = 1;
bd[bd_count].sge_addr.lo = addr & 0xffffffff;
bd[bd_count].sge_addr.hi = addr >> 32;
bd[bd_count].sge_len = sg_len;
}
byte_count += sg_len;
bd_count += sg_frags;
}
if (byte_count != scsi_bufflen(sc))
QEDI_ERR(&qedi->dbg_ctx,
"byte_count = %d != scsi_bufflen = %d\n", byte_count,
scsi_bufflen(sc));
else
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_IO, "byte_count = %d\n",
byte_count);
WARN_ON(byte_count != scsi_bufflen(sc));
return bd_count;
}
static void qedi_iscsi_map_sg_list(struct qedi_cmd *cmd)
{
int bd_count;
struct scsi_cmnd *sc = cmd->scsi_cmd;
if (scsi_sg_count(sc)) {
bd_count = qedi_map_scsi_sg(cmd->conn->qedi, cmd);
if (bd_count == 0)
return;
} else {
struct scsi_sge *bd = cmd->io_tbl.sge_tbl;
bd[0].sge_addr.lo = 0;
bd[0].sge_addr.hi = 0;
bd[0].sge_len = 0;
bd_count = 0;
}
cmd->io_tbl.sge_valid = bd_count;
}
static void qedi_cpy_scsi_cdb(struct scsi_cmnd *sc, u32 *dstp)
{
u32 dword;
int lpcnt;
u8 *srcp;
lpcnt = sc->cmd_len / sizeof(dword);
srcp = (u8 *)sc->cmnd;
while (lpcnt--) {
memcpy(&dword, (const void *)srcp, 4);
*dstp = cpu_to_be32(dword);
srcp += 4;
dstp++;
}
if (sc->cmd_len & 0x3) {
dword = (u32)srcp[0] | ((u32)srcp[1] << 8);
*dstp = cpu_to_be32(dword);
}
}
void qedi_trace_io(struct qedi_ctx *qedi, struct iscsi_task *task,
u16 tid, int8_t direction)
{
struct qedi_io_log *io_log;
struct iscsi_conn *conn = task->conn;
struct qedi_conn *qedi_conn = conn->dd_data;
struct scsi_cmnd *sc_cmd = task->sc;
unsigned long flags;
spin_lock_irqsave(&qedi->io_trace_lock, flags);
io_log = &qedi->io_trace_buf[qedi->io_trace_idx];
io_log->direction = direction;
io_log->task_id = tid;
io_log->cid = qedi_conn->iscsi_conn_id;
io_log->lun = sc_cmd->device->lun;
io_log->op = sc_cmd->cmnd[0];
io_log->lba[0] = sc_cmd->cmnd[2];
io_log->lba[1] = sc_cmd->cmnd[3];
io_log->lba[2] = sc_cmd->cmnd[4];
io_log->lba[3] = sc_cmd->cmnd[5];
io_log->bufflen = scsi_bufflen(sc_cmd);
io_log->sg_count = scsi_sg_count(sc_cmd);
io_log->fast_sgs = qedi->fast_sgls;
io_log->cached_sgs = qedi->cached_sgls;
io_log->slow_sgs = qedi->slow_sgls;
io_log->cached_sge = qedi->use_cached_sge;
io_log->slow_sge = qedi->use_slow_sge;
io_log->fast_sge = qedi->use_fast_sge;
io_log->result = sc_cmd->result;
io_log->jiffies = jiffies;
io_log->blk_req_cpu = smp_processor_id();
if (direction == QEDI_IO_TRACE_REQ) {
/* For requests we only care about the submission CPU */
io_log->req_cpu = smp_processor_id() % qedi->num_queues;
io_log->intr_cpu = 0;
io_log->blk_rsp_cpu = 0;
} else if (direction == QEDI_IO_TRACE_RSP) {
io_log->req_cpu = smp_processor_id() % qedi->num_queues;
io_log->intr_cpu = qedi->intr_cpu;
io_log->blk_rsp_cpu = smp_processor_id();
}
qedi->io_trace_idx++;
if (qedi->io_trace_idx == QEDI_IO_TRACE_SIZE)
qedi->io_trace_idx = 0;
qedi->use_cached_sge = false;
qedi->use_slow_sge = false;
qedi->use_fast_sge = false;
spin_unlock_irqrestore(&qedi->io_trace_lock, flags);
}
int qedi_iscsi_send_ioreq(struct iscsi_task *task)
{
struct iscsi_conn *conn = task->conn;
struct iscsi_session *session = conn->session;
struct Scsi_Host *shost = iscsi_session_to_shost(session->cls_session);
struct qedi_ctx *qedi = iscsi_host_priv(shost);
struct qedi_conn *qedi_conn = conn->dd_data;
struct qedi_cmd *cmd = task->dd_data;
struct scsi_cmnd *sc = task->sc;
struct iscsi_cmd_hdr cmd_pdu_header;
struct scsi_sgl_task_params tx_sgl_task_params;
struct scsi_sgl_task_params rx_sgl_task_params;
struct scsi_sgl_task_params *prx_sgl = NULL;
struct scsi_sgl_task_params *ptx_sgl = NULL;
struct iscsi_task_params task_params;
struct iscsi_conn_params conn_params;
struct scsi_initiator_cmd_params cmd_params;
struct e4_iscsi_task_context *fw_task_ctx;
struct iscsi_cls_conn *cls_conn;
struct iscsi_scsi_req *hdr = (struct iscsi_scsi_req *)task->hdr;
enum iscsi_task_type task_type = MAX_ISCSI_TASK_TYPE;
struct qedi_endpoint *ep;
u32 scsi_lun[2];
s16 tid = 0;
u16 sq_idx = 0;
u16 cq_idx;
int rval = 0;
ep = qedi_conn->ep;
cls_conn = qedi_conn->cls_conn;
conn = cls_conn->dd_data;
qedi_iscsi_map_sg_list(cmd);
int_to_scsilun(sc->device->lun, (struct scsi_lun *)scsi_lun);
tid = qedi_get_task_idx(qedi);
if (tid == -1)
return -ENOMEM;
fw_task_ctx =
(struct e4_iscsi_task_context *)qedi_get_task_mem(&qedi->tasks,
tid);
memset(fw_task_ctx, 0, sizeof(struct e4_iscsi_task_context));
cmd->task_id = tid;
memset(&task_params, 0, sizeof(task_params));
memset(&cmd_pdu_header, 0, sizeof(cmd_pdu_header));
memset(&tx_sgl_task_params, 0, sizeof(tx_sgl_task_params));
memset(&rx_sgl_task_params, 0, sizeof(rx_sgl_task_params));
memset(&conn_params, 0, sizeof(conn_params));
memset(&cmd_params, 0, sizeof(cmd_params));
cq_idx = smp_processor_id() % qedi->num_queues;
/* Update header info */
SET_FIELD(cmd_pdu_header.flags_attr, ISCSI_CMD_HDR_ATTR,
ISCSI_ATTR_SIMPLE);
if (hdr->cdb[0] != TEST_UNIT_READY) {
if (sc->sc_data_direction == DMA_TO_DEVICE) {
SET_FIELD(cmd_pdu_header.flags_attr,
ISCSI_CMD_HDR_WRITE, 1);
task_type = ISCSI_TASK_TYPE_INITIATOR_WRITE;
} else {
SET_FIELD(cmd_pdu_header.flags_attr,
ISCSI_CMD_HDR_READ, 1);
task_type = ISCSI_TASK_TYPE_INITIATOR_READ;
}
}
cmd_pdu_header.lun.lo = be32_to_cpu(scsi_lun[0]);
cmd_pdu_header.lun.hi = be32_to_cpu(scsi_lun[1]);
qedi_update_itt_map(qedi, tid, task->itt, cmd);
cmd_pdu_header.itt = qedi_set_itt(tid, get_itt(task->itt));
cmd_pdu_header.expected_transfer_length = cpu_to_be32(hdr->data_length);
cmd_pdu_header.hdr_second_dword = ntoh24(hdr->dlength);
cmd_pdu_header.cmd_sn = be32_to_cpu(hdr->cmdsn);
cmd_pdu_header.hdr_first_byte = hdr->opcode;
qedi_cpy_scsi_cdb(sc, (u32 *)cmd_pdu_header.cdb);
/* Fill tx AHS and rx buffer */
if (task_type == ISCSI_TASK_TYPE_INITIATOR_WRITE) {
tx_sgl_task_params.sgl = cmd->io_tbl.sge_tbl;
tx_sgl_task_params.sgl_phys_addr.lo =
(u32)(cmd->io_tbl.sge_tbl_dma);
tx_sgl_task_params.sgl_phys_addr.hi =
(u32)((u64)cmd->io_tbl.sge_tbl_dma >> 32);
tx_sgl_task_params.total_buffer_size = scsi_bufflen(sc);
tx_sgl_task_params.num_sges = cmd->io_tbl.sge_valid;
if (cmd->use_slowpath)
tx_sgl_task_params.small_mid_sge = true;
} else if (task_type == ISCSI_TASK_TYPE_INITIATOR_READ) {
rx_sgl_task_params.sgl = cmd->io_tbl.sge_tbl;
rx_sgl_task_params.sgl_phys_addr.lo =
(u32)(cmd->io_tbl.sge_tbl_dma);
rx_sgl_task_params.sgl_phys_addr.hi =
(u32)((u64)cmd->io_tbl.sge_tbl_dma >> 32);
rx_sgl_task_params.total_buffer_size = scsi_bufflen(sc);
rx_sgl_task_params.num_sges = cmd->io_tbl.sge_valid;
}
/* Add conn param */
conn_params.first_burst_length = conn->session->first_burst;
conn_params.max_send_pdu_length = conn->max_xmit_dlength;
conn_params.max_burst_length = conn->session->max_burst;
if (conn->session->initial_r2t_en)
conn_params.initial_r2t = true;
if (conn->session->imm_data_en)
conn_params.immediate_data = true;
/* Add cmd params */
cmd_params.sense_data_buffer_phys_addr.lo = (u32)cmd->sense_buffer_dma;
cmd_params.sense_data_buffer_phys_addr.hi =
(u32)((u64)cmd->sense_buffer_dma >> 32);
/* Fill fw input params */
task_params.context = fw_task_ctx;
task_params.conn_icid = (u16)qedi_conn->iscsi_conn_id;
task_params.itid = tid;
task_params.cq_rss_number = cq_idx;
if (task_type == ISCSI_TASK_TYPE_INITIATOR_WRITE)
task_params.tx_io_size = scsi_bufflen(sc);
else if (task_type == ISCSI_TASK_TYPE_INITIATOR_READ)
task_params.rx_io_size = scsi_bufflen(sc);
sq_idx = qedi_get_wqe_idx(qedi_conn);
task_params.sqe = &ep->sq[sq_idx];
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_IO,
"%s: %s-SGL: sg_len=0x%x num_sges=0x%x first-sge-lo=0x%x first-sge-hi=0x%x\n",
(task_type == ISCSI_TASK_TYPE_INITIATOR_WRITE) ?
"Write " : "Read ", (cmd->io_tbl.sge_valid == 1) ?
"Single" : (cmd->use_slowpath ? "SLOW" : "FAST"),
(u16)cmd->io_tbl.sge_valid, scsi_bufflen(sc),
(u32)(cmd->io_tbl.sge_tbl_dma),
(u32)((u64)cmd->io_tbl.sge_tbl_dma >> 32));
memset(task_params.sqe, 0, sizeof(struct iscsi_wqe));
if (task_params.tx_io_size != 0)
ptx_sgl = &tx_sgl_task_params;
if (task_params.rx_io_size != 0)
prx_sgl = &rx_sgl_task_params;
rval = init_initiator_rw_iscsi_task(&task_params, &conn_params,
&cmd_params, &cmd_pdu_header,
ptx_sgl, prx_sgl,
NULL);
if (rval)
return -1;
spin_lock(&qedi_conn->list_lock);
list_add_tail(&cmd->io_cmd, &qedi_conn->active_cmd_list);
cmd->io_cmd_in_list = true;
qedi_conn->active_cmd_count++;
spin_unlock(&qedi_conn->list_lock);
qedi_ring_doorbell(qedi_conn);
return 0;
}
int qedi_iscsi_cleanup_task(struct iscsi_task *task, bool mark_cmd_node_deleted)
{
struct iscsi_task_params task_params;
struct qedi_endpoint *ep;
struct iscsi_conn *conn = task->conn;
struct qedi_conn *qedi_conn = conn->dd_data;
struct qedi_cmd *cmd = task->dd_data;
u16 sq_idx = 0;
int rval = 0;
QEDI_INFO(&qedi_conn->qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"issue cleanup tid=0x%x itt=0x%x task_state=%d cmd_state=0%x cid=0x%x\n",
cmd->task_id, get_itt(task->itt), task->state,
cmd->state, qedi_conn->iscsi_conn_id);
memset(&task_params, 0, sizeof(task_params));
ep = qedi_conn->ep;
sq_idx = qedi_get_wqe_idx(qedi_conn);
task_params.sqe = &ep->sq[sq_idx];
memset(task_params.sqe, 0, sizeof(struct iscsi_wqe));
task_params.itid = cmd->task_id;
rval = init_cleanup_task(&task_params);
if (rval)
return rval;
qedi_ring_doorbell(qedi_conn);
return 0;
}