mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-15 08:44:14 +08:00
Merge branch '5.14/scsi-fixes' into 5.15/scsi-staging
Resolve mpt3sas conflict between 5.14/scsi-fixes and 5.15/scsi-staging reported by sfr. Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
This commit is contained in:
commit
315480209b
@ -453,8 +453,8 @@ static int initialize_controller(struct scsi_device *sdev,
|
||||
if (!h->ctlr)
|
||||
err = SCSI_DH_RES_TEMP_UNAVAIL;
|
||||
else {
|
||||
list_add_rcu(&h->node, &h->ctlr->dh_list);
|
||||
h->sdev = sdev;
|
||||
list_add_rcu(&h->node, &h->ctlr->dh_list);
|
||||
}
|
||||
spin_unlock(&list_lock);
|
||||
err = SCSI_DH_OK;
|
||||
@ -778,11 +778,11 @@ static void rdac_bus_detach( struct scsi_device *sdev )
|
||||
spin_lock(&list_lock);
|
||||
if (h->ctlr) {
|
||||
list_del_rcu(&h->node);
|
||||
h->sdev = NULL;
|
||||
kref_put(&h->ctlr->kref, release_controller);
|
||||
}
|
||||
spin_unlock(&list_lock);
|
||||
sdev->handler_data = NULL;
|
||||
synchronize_rcu();
|
||||
kfree(h);
|
||||
}
|
||||
|
||||
|
@ -807,6 +807,13 @@ static int ibmvfc_init_event_pool(struct ibmvfc_host *vhost,
|
||||
for (i = 0; i < size; ++i) {
|
||||
struct ibmvfc_event *evt = &pool->events[i];
|
||||
|
||||
/*
|
||||
* evt->active states
|
||||
* 1 = in flight
|
||||
* 0 = being completed
|
||||
* -1 = free/freed
|
||||
*/
|
||||
atomic_set(&evt->active, -1);
|
||||
atomic_set(&evt->free, 1);
|
||||
evt->crq.valid = 0x80;
|
||||
evt->crq.ioba = cpu_to_be64(pool->iu_token + (sizeof(*evt->xfer_iu) * i));
|
||||
@ -1017,6 +1024,7 @@ static void ibmvfc_free_event(struct ibmvfc_event *evt)
|
||||
|
||||
BUG_ON(!ibmvfc_valid_event(pool, evt));
|
||||
BUG_ON(atomic_inc_return(&evt->free) != 1);
|
||||
BUG_ON(atomic_dec_and_test(&evt->active));
|
||||
|
||||
spin_lock_irqsave(&evt->queue->l_lock, flags);
|
||||
list_add_tail(&evt->queue_list, &evt->queue->free);
|
||||
@ -1072,6 +1080,12 @@ static void ibmvfc_complete_purge(struct list_head *purge_list)
|
||||
**/
|
||||
static void ibmvfc_fail_request(struct ibmvfc_event *evt, int error_code)
|
||||
{
|
||||
/*
|
||||
* Anything we are failing should still be active. Otherwise, it
|
||||
* implies we already got a response for the command and are doing
|
||||
* something bad like double completing it.
|
||||
*/
|
||||
BUG_ON(!atomic_dec_and_test(&evt->active));
|
||||
if (evt->cmnd) {
|
||||
evt->cmnd->result = (error_code << 16);
|
||||
evt->done = ibmvfc_scsi_eh_done;
|
||||
@ -1723,6 +1737,7 @@ static int ibmvfc_send_event(struct ibmvfc_event *evt,
|
||||
|
||||
evt->done(evt);
|
||||
} else {
|
||||
atomic_set(&evt->active, 1);
|
||||
spin_unlock_irqrestore(&evt->queue->l_lock, flags);
|
||||
ibmvfc_trc_start(evt);
|
||||
}
|
||||
@ -3251,7 +3266,7 @@ static void ibmvfc_handle_crq(struct ibmvfc_crq *crq, struct ibmvfc_host *vhost,
|
||||
return;
|
||||
}
|
||||
|
||||
if (unlikely(atomic_read(&evt->free))) {
|
||||
if (unlikely(atomic_dec_if_positive(&evt->active))) {
|
||||
dev_err(vhost->dev, "Received duplicate correlation_token 0x%08llx!\n",
|
||||
crq->ioba);
|
||||
return;
|
||||
@ -3778,7 +3793,7 @@ static void ibmvfc_handle_scrq(struct ibmvfc_crq *crq, struct ibmvfc_host *vhost
|
||||
return;
|
||||
}
|
||||
|
||||
if (unlikely(atomic_read(&evt->free))) {
|
||||
if (unlikely(atomic_dec_if_positive(&evt->active))) {
|
||||
dev_err(vhost->dev, "Received duplicate correlation_token 0x%08llx!\n",
|
||||
crq->ioba);
|
||||
return;
|
||||
|
@ -745,6 +745,7 @@ struct ibmvfc_event {
|
||||
struct ibmvfc_target *tgt;
|
||||
struct scsi_cmnd *cmnd;
|
||||
atomic_t free;
|
||||
atomic_t active;
|
||||
union ibmvfc_iu *xfer_iu;
|
||||
void (*done)(struct ibmvfc_event *evt);
|
||||
void (*_done)(struct ibmvfc_event *evt);
|
||||
|
@ -238,7 +238,7 @@ mraid_mm_get_adapter(mimd_t __user *umimd, int *rval)
|
||||
mimd_t mimd;
|
||||
uint32_t adapno;
|
||||
int iterator;
|
||||
|
||||
bool is_found;
|
||||
|
||||
if (copy_from_user(&mimd, umimd, sizeof(mimd_t))) {
|
||||
*rval = -EFAULT;
|
||||
@ -254,12 +254,16 @@ mraid_mm_get_adapter(mimd_t __user *umimd, int *rval)
|
||||
|
||||
adapter = NULL;
|
||||
iterator = 0;
|
||||
is_found = false;
|
||||
|
||||
list_for_each_entry(adapter, &adapters_list_g, list) {
|
||||
if (iterator++ == adapno) break;
|
||||
if (iterator++ == adapno) {
|
||||
is_found = true;
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
if (!adapter) {
|
||||
if (!is_found) {
|
||||
*rval = -ENODEV;
|
||||
return NULL;
|
||||
}
|
||||
@ -725,6 +729,7 @@ ioctl_done(uioc_t *kioc)
|
||||
uint32_t adapno;
|
||||
int iterator;
|
||||
mraid_mmadp_t* adapter;
|
||||
bool is_found;
|
||||
|
||||
/*
|
||||
* When the kioc returns from driver, make sure it still doesn't
|
||||
@ -747,19 +752,23 @@ ioctl_done(uioc_t *kioc)
|
||||
iterator = 0;
|
||||
adapter = NULL;
|
||||
adapno = kioc->adapno;
|
||||
is_found = false;
|
||||
|
||||
con_log(CL_ANN, ( KERN_WARNING "megaraid cmm: completed "
|
||||
"ioctl that was timedout before\n"));
|
||||
|
||||
list_for_each_entry(adapter, &adapters_list_g, list) {
|
||||
if (iterator++ == adapno) break;
|
||||
if (iterator++ == adapno) {
|
||||
is_found = true;
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
kioc->timedout = 0;
|
||||
|
||||
if (adapter) {
|
||||
if (is_found)
|
||||
mraid_mm_dealloc_kioc( adapter, kioc );
|
||||
}
|
||||
|
||||
}
|
||||
else {
|
||||
wake_up(&wait_q);
|
||||
|
@ -3076,13 +3076,13 @@ _base_check_enable_msix(struct MPT3SAS_ADAPTER *ioc)
|
||||
}
|
||||
|
||||
/**
|
||||
* _base_free_irq - free irq
|
||||
* mpt3sas_base_free_irq - free irq
|
||||
* @ioc: per adapter object
|
||||
*
|
||||
* Freeing respective reply_queue from the list.
|
||||
*/
|
||||
static void
|
||||
_base_free_irq(struct MPT3SAS_ADAPTER *ioc)
|
||||
void
|
||||
mpt3sas_base_free_irq(struct MPT3SAS_ADAPTER *ioc)
|
||||
{
|
||||
struct adapter_reply_queue *reply_q, *next;
|
||||
|
||||
@ -3308,12 +3308,12 @@ _base_check_and_enable_high_iops_queues(struct MPT3SAS_ADAPTER *ioc,
|
||||
}
|
||||
|
||||
/**
|
||||
* _base_disable_msix - disables msix
|
||||
* mpt3sas_base_disable_msix - disables msix
|
||||
* @ioc: per adapter object
|
||||
*
|
||||
*/
|
||||
static void
|
||||
_base_disable_msix(struct MPT3SAS_ADAPTER *ioc)
|
||||
void
|
||||
mpt3sas_base_disable_msix(struct MPT3SAS_ADAPTER *ioc)
|
||||
{
|
||||
if (!ioc->msix_enable)
|
||||
return;
|
||||
@ -3486,8 +3486,8 @@ _base_enable_msix(struct MPT3SAS_ADAPTER *ioc)
|
||||
for (i = 0; i < ioc->reply_queue_count; i++) {
|
||||
r = _base_request_irq(ioc, i);
|
||||
if (r) {
|
||||
_base_free_irq(ioc);
|
||||
_base_disable_msix(ioc);
|
||||
mpt3sas_base_free_irq(ioc);
|
||||
mpt3sas_base_disable_msix(ioc);
|
||||
goto try_ioapic;
|
||||
}
|
||||
}
|
||||
@ -3525,8 +3525,8 @@ mpt3sas_base_unmap_resources(struct MPT3SAS_ADAPTER *ioc)
|
||||
|
||||
dexitprintk(ioc, ioc_info(ioc, "%s\n", __func__));
|
||||
|
||||
_base_free_irq(ioc);
|
||||
_base_disable_msix(ioc);
|
||||
mpt3sas_base_free_irq(ioc);
|
||||
mpt3sas_base_disable_msix(ioc);
|
||||
|
||||
kfree(ioc->replyPostRegisterIndex);
|
||||
ioc->replyPostRegisterIndex = NULL;
|
||||
@ -7880,14 +7880,14 @@ _base_diag_reset(struct MPT3SAS_ADAPTER *ioc)
|
||||
}
|
||||
|
||||
/**
|
||||
* _base_make_ioc_ready - put controller in READY state
|
||||
* mpt3sas_base_make_ioc_ready - put controller in READY state
|
||||
* @ioc: per adapter object
|
||||
* @type: FORCE_BIG_HAMMER or SOFT_RESET
|
||||
*
|
||||
* Return: 0 for success, non-zero for failure.
|
||||
*/
|
||||
static int
|
||||
_base_make_ioc_ready(struct MPT3SAS_ADAPTER *ioc, enum reset_type type)
|
||||
int
|
||||
mpt3sas_base_make_ioc_ready(struct MPT3SAS_ADAPTER *ioc, enum reset_type type)
|
||||
{
|
||||
u32 ioc_state;
|
||||
int rc;
|
||||
@ -8164,7 +8164,7 @@ mpt3sas_base_free_resources(struct MPT3SAS_ADAPTER *ioc)
|
||||
if (ioc->chip_phys && ioc->chip) {
|
||||
mpt3sas_base_mask_interrupts(ioc);
|
||||
ioc->shost_recovery = 1;
|
||||
_base_make_ioc_ready(ioc, SOFT_RESET);
|
||||
mpt3sas_base_make_ioc_ready(ioc, SOFT_RESET);
|
||||
ioc->shost_recovery = 0;
|
||||
}
|
||||
|
||||
@ -8284,7 +8284,7 @@ mpt3sas_base_attach(struct MPT3SAS_ADAPTER *ioc)
|
||||
ioc->build_sg_mpi = &_base_build_sg;
|
||||
ioc->build_zero_len_sge_mpi = &_base_build_zero_len_sge;
|
||||
|
||||
r = _base_make_ioc_ready(ioc, SOFT_RESET);
|
||||
r = mpt3sas_base_make_ioc_ready(ioc, SOFT_RESET);
|
||||
if (r)
|
||||
goto out_free_resources;
|
||||
|
||||
@ -8739,7 +8739,7 @@ mpt3sas_base_hard_reset_handler(struct MPT3SAS_ADAPTER *ioc,
|
||||
mpt3sas_wait_for_commands_to_complete(ioc);
|
||||
mpt3sas_base_mask_interrupts(ioc);
|
||||
mpt3sas_base_pause_mq_polling(ioc);
|
||||
r = _base_make_ioc_ready(ioc, type);
|
||||
r = mpt3sas_base_make_ioc_ready(ioc, type);
|
||||
if (r)
|
||||
goto out;
|
||||
_base_clear_outstanding_commands(ioc);
|
||||
|
@ -1755,6 +1755,9 @@ do { ioc_err(ioc, "In func: %s\n", __func__); \
|
||||
status, mpi_request, sz); } while (0)
|
||||
|
||||
int mpt3sas_wait_for_ioc(struct MPT3SAS_ADAPTER *ioc, int wait_count);
|
||||
int mpt3sas_base_make_ioc_ready(struct MPT3SAS_ADAPTER *ioc, enum reset_type type);
|
||||
void mpt3sas_base_free_irq(struct MPT3SAS_ADAPTER *ioc);
|
||||
void mpt3sas_base_disable_msix(struct MPT3SAS_ADAPTER *ioc);
|
||||
int mpt3sas_blk_mq_poll(struct Scsi_Host *shost, unsigned int queue_num);
|
||||
void mpt3sas_base_pause_mq_polling(struct MPT3SAS_ADAPTER *ioc);
|
||||
void mpt3sas_base_resume_mq_polling(struct MPT3SAS_ADAPTER *ioc);
|
||||
|
@ -11342,7 +11342,12 @@ scsih_shutdown(struct pci_dev *pdev)
|
||||
|
||||
_scsih_ir_shutdown(ioc);
|
||||
_scsih_nvme_shutdown(ioc);
|
||||
mpt3sas_base_detach(ioc);
|
||||
mpt3sas_base_mask_interrupts(ioc);
|
||||
ioc->shost_recovery = 1;
|
||||
mpt3sas_base_make_ioc_ready(ioc, SOFT_RESET);
|
||||
ioc->shost_recovery = 0;
|
||||
mpt3sas_base_free_irq(ioc);
|
||||
mpt3sas_base_disable_msix(ioc);
|
||||
}
|
||||
|
||||
|
||||
|
@ -684,8 +684,7 @@ int pm8001_dev_found(struct domain_device *dev)
|
||||
|
||||
void pm8001_task_done(struct sas_task *task)
|
||||
{
|
||||
if (!del_timer(&task->slow_task->timer))
|
||||
return;
|
||||
del_timer(&task->slow_task->timer);
|
||||
complete(&task->slow_task->completion);
|
||||
}
|
||||
|
||||
@ -693,9 +692,14 @@ static void pm8001_tmf_timedout(struct timer_list *t)
|
||||
{
|
||||
struct sas_task_slow *slow = from_timer(slow, t, timer);
|
||||
struct sas_task *task = slow->task;
|
||||
unsigned long flags;
|
||||
|
||||
task->task_state_flags |= SAS_TASK_STATE_ABORTED;
|
||||
complete(&task->slow_task->completion);
|
||||
spin_lock_irqsave(&task->task_state_lock, flags);
|
||||
if (!(task->task_state_flags & SAS_TASK_STATE_DONE)) {
|
||||
task->task_state_flags |= SAS_TASK_STATE_ABORTED;
|
||||
complete(&task->slow_task->completion);
|
||||
}
|
||||
spin_unlock_irqrestore(&task->task_state_lock, flags);
|
||||
}
|
||||
|
||||
#define PM8001_TASK_TIMEOUT 20
|
||||
@ -748,13 +752,10 @@ static int pm8001_exec_internal_tmf_task(struct domain_device *dev,
|
||||
}
|
||||
res = -TMF_RESP_FUNC_FAILED;
|
||||
/* Even TMF timed out, return direct. */
|
||||
if ((task->task_state_flags & SAS_TASK_STATE_ABORTED)) {
|
||||
if (!(task->task_state_flags & SAS_TASK_STATE_DONE)) {
|
||||
pm8001_dbg(pm8001_ha, FAIL,
|
||||
"TMF task[%x]timeout.\n",
|
||||
tmf->tmf);
|
||||
goto ex_err;
|
||||
}
|
||||
if (task->task_state_flags & SAS_TASK_STATE_ABORTED) {
|
||||
pm8001_dbg(pm8001_ha, FAIL, "TMF task[%x]timeout.\n",
|
||||
tmf->tmf);
|
||||
goto ex_err;
|
||||
}
|
||||
|
||||
if (task->task_status.resp == SAS_TASK_COMPLETE &&
|
||||
@ -834,12 +835,9 @@ pm8001_exec_internal_task_abort(struct pm8001_hba_info *pm8001_ha,
|
||||
wait_for_completion(&task->slow_task->completion);
|
||||
res = TMF_RESP_FUNC_FAILED;
|
||||
/* Even TMF timed out, return direct. */
|
||||
if ((task->task_state_flags & SAS_TASK_STATE_ABORTED)) {
|
||||
if (!(task->task_state_flags & SAS_TASK_STATE_DONE)) {
|
||||
pm8001_dbg(pm8001_ha, FAIL,
|
||||
"TMF task timeout.\n");
|
||||
goto ex_err;
|
||||
}
|
||||
if (task->task_state_flags & SAS_TASK_STATE_ABORTED) {
|
||||
pm8001_dbg(pm8001_ha, FAIL, "TMF task timeout.\n");
|
||||
goto ex_err;
|
||||
}
|
||||
|
||||
if (task->task_status.resp == SAS_TASK_COMPLETE &&
|
||||
|
@ -477,7 +477,8 @@ static struct scsi_target *scsi_alloc_target(struct device *parent,
|
||||
error = shost->hostt->target_alloc(starget);
|
||||
|
||||
if(error) {
|
||||
dev_printk(KERN_ERR, dev, "target allocation failed, error %d\n", error);
|
||||
if (error != -ENXIO)
|
||||
dev_err(dev, "target allocation failed, error %d\n", error);
|
||||
/* don't want scsi_target_reap to do the final
|
||||
* put because it will be under the host lock */
|
||||
scsi_target_destroy(starget);
|
||||
|
@ -808,11 +808,14 @@ store_state_field(struct device *dev, struct device_attribute *attr,
|
||||
mutex_lock(&sdev->state_mutex);
|
||||
ret = scsi_device_set_state(sdev, state);
|
||||
/*
|
||||
* If the device state changes to SDEV_RUNNING, we need to run
|
||||
* the queue to avoid I/O hang.
|
||||
* If the device state changes to SDEV_RUNNING, we need to
|
||||
* rescan the device to revalidate it, and run the queue to
|
||||
* avoid I/O hang.
|
||||
*/
|
||||
if (ret == 0 && state == SDEV_RUNNING)
|
||||
if (ret == 0 && state == SDEV_RUNNING) {
|
||||
scsi_rescan_device(dev);
|
||||
blk_mq_run_hw_queues(sdev->request_queue, true);
|
||||
}
|
||||
mutex_unlock(&sdev->state_mutex);
|
||||
|
||||
return ret == 0 ? count : -EINVAL;
|
||||
|
@ -439,39 +439,10 @@ static umode_t iscsi_iface_attr_is_visible(struct kobject *kobj,
|
||||
struct device *dev = container_of(kobj, struct device, kobj);
|
||||
struct iscsi_iface *iface = iscsi_dev_to_iface(dev);
|
||||
struct iscsi_transport *t = iface->transport;
|
||||
int param;
|
||||
int param_type;
|
||||
int param = -1;
|
||||
|
||||
if (attr == &dev_attr_iface_enabled.attr)
|
||||
param = ISCSI_NET_PARAM_IFACE_ENABLE;
|
||||
else if (attr == &dev_attr_iface_vlan_id.attr)
|
||||
param = ISCSI_NET_PARAM_VLAN_ID;
|
||||
else if (attr == &dev_attr_iface_vlan_priority.attr)
|
||||
param = ISCSI_NET_PARAM_VLAN_PRIORITY;
|
||||
else if (attr == &dev_attr_iface_vlan_enabled.attr)
|
||||
param = ISCSI_NET_PARAM_VLAN_ENABLED;
|
||||
else if (attr == &dev_attr_iface_mtu.attr)
|
||||
param = ISCSI_NET_PARAM_MTU;
|
||||
else if (attr == &dev_attr_iface_port.attr)
|
||||
param = ISCSI_NET_PARAM_PORT;
|
||||
else if (attr == &dev_attr_iface_ipaddress_state.attr)
|
||||
param = ISCSI_NET_PARAM_IPADDR_STATE;
|
||||
else if (attr == &dev_attr_iface_delayed_ack_en.attr)
|
||||
param = ISCSI_NET_PARAM_DELAYED_ACK_EN;
|
||||
else if (attr == &dev_attr_iface_tcp_nagle_disable.attr)
|
||||
param = ISCSI_NET_PARAM_TCP_NAGLE_DISABLE;
|
||||
else if (attr == &dev_attr_iface_tcp_wsf_disable.attr)
|
||||
param = ISCSI_NET_PARAM_TCP_WSF_DISABLE;
|
||||
else if (attr == &dev_attr_iface_tcp_wsf.attr)
|
||||
param = ISCSI_NET_PARAM_TCP_WSF;
|
||||
else if (attr == &dev_attr_iface_tcp_timer_scale.attr)
|
||||
param = ISCSI_NET_PARAM_TCP_TIMER_SCALE;
|
||||
else if (attr == &dev_attr_iface_tcp_timestamp_en.attr)
|
||||
param = ISCSI_NET_PARAM_TCP_TIMESTAMP_EN;
|
||||
else if (attr == &dev_attr_iface_cache_id.attr)
|
||||
param = ISCSI_NET_PARAM_CACHE_ID;
|
||||
else if (attr == &dev_attr_iface_redirect_en.attr)
|
||||
param = ISCSI_NET_PARAM_REDIRECT_EN;
|
||||
else if (attr == &dev_attr_iface_def_taskmgmt_tmo.attr)
|
||||
param = ISCSI_IFACE_PARAM_DEF_TASKMGMT_TMO;
|
||||
else if (attr == &dev_attr_iface_header_digest.attr)
|
||||
@ -508,6 +479,38 @@ static umode_t iscsi_iface_attr_is_visible(struct kobject *kobj,
|
||||
param = ISCSI_IFACE_PARAM_STRICT_LOGIN_COMP_EN;
|
||||
else if (attr == &dev_attr_iface_initiator_name.attr)
|
||||
param = ISCSI_IFACE_PARAM_INITIATOR_NAME;
|
||||
|
||||
if (param != -1)
|
||||
return t->attr_is_visible(ISCSI_IFACE_PARAM, param);
|
||||
|
||||
if (attr == &dev_attr_iface_vlan_id.attr)
|
||||
param = ISCSI_NET_PARAM_VLAN_ID;
|
||||
else if (attr == &dev_attr_iface_vlan_priority.attr)
|
||||
param = ISCSI_NET_PARAM_VLAN_PRIORITY;
|
||||
else if (attr == &dev_attr_iface_vlan_enabled.attr)
|
||||
param = ISCSI_NET_PARAM_VLAN_ENABLED;
|
||||
else if (attr == &dev_attr_iface_mtu.attr)
|
||||
param = ISCSI_NET_PARAM_MTU;
|
||||
else if (attr == &dev_attr_iface_port.attr)
|
||||
param = ISCSI_NET_PARAM_PORT;
|
||||
else if (attr == &dev_attr_iface_ipaddress_state.attr)
|
||||
param = ISCSI_NET_PARAM_IPADDR_STATE;
|
||||
else if (attr == &dev_attr_iface_delayed_ack_en.attr)
|
||||
param = ISCSI_NET_PARAM_DELAYED_ACK_EN;
|
||||
else if (attr == &dev_attr_iface_tcp_nagle_disable.attr)
|
||||
param = ISCSI_NET_PARAM_TCP_NAGLE_DISABLE;
|
||||
else if (attr == &dev_attr_iface_tcp_wsf_disable.attr)
|
||||
param = ISCSI_NET_PARAM_TCP_WSF_DISABLE;
|
||||
else if (attr == &dev_attr_iface_tcp_wsf.attr)
|
||||
param = ISCSI_NET_PARAM_TCP_WSF;
|
||||
else if (attr == &dev_attr_iface_tcp_timer_scale.attr)
|
||||
param = ISCSI_NET_PARAM_TCP_TIMER_SCALE;
|
||||
else if (attr == &dev_attr_iface_tcp_timestamp_en.attr)
|
||||
param = ISCSI_NET_PARAM_TCP_TIMESTAMP_EN;
|
||||
else if (attr == &dev_attr_iface_cache_id.attr)
|
||||
param = ISCSI_NET_PARAM_CACHE_ID;
|
||||
else if (attr == &dev_attr_iface_redirect_en.attr)
|
||||
param = ISCSI_NET_PARAM_REDIRECT_EN;
|
||||
else if (iface->iface_type == ISCSI_IFACE_TYPE_IPV4) {
|
||||
if (attr == &dev_attr_ipv4_iface_ipaddress.attr)
|
||||
param = ISCSI_NET_PARAM_IPV4_ADDR;
|
||||
@ -598,32 +601,7 @@ static umode_t iscsi_iface_attr_is_visible(struct kobject *kobj,
|
||||
return 0;
|
||||
}
|
||||
|
||||
switch (param) {
|
||||
case ISCSI_IFACE_PARAM_DEF_TASKMGMT_TMO:
|
||||
case ISCSI_IFACE_PARAM_HDRDGST_EN:
|
||||
case ISCSI_IFACE_PARAM_DATADGST_EN:
|
||||
case ISCSI_IFACE_PARAM_IMM_DATA_EN:
|
||||
case ISCSI_IFACE_PARAM_INITIAL_R2T_EN:
|
||||
case ISCSI_IFACE_PARAM_DATASEQ_INORDER_EN:
|
||||
case ISCSI_IFACE_PARAM_PDU_INORDER_EN:
|
||||
case ISCSI_IFACE_PARAM_ERL:
|
||||
case ISCSI_IFACE_PARAM_MAX_RECV_DLENGTH:
|
||||
case ISCSI_IFACE_PARAM_FIRST_BURST:
|
||||
case ISCSI_IFACE_PARAM_MAX_R2T:
|
||||
case ISCSI_IFACE_PARAM_MAX_BURST:
|
||||
case ISCSI_IFACE_PARAM_CHAP_AUTH_EN:
|
||||
case ISCSI_IFACE_PARAM_BIDI_CHAP_EN:
|
||||
case ISCSI_IFACE_PARAM_DISCOVERY_AUTH_OPTIONAL:
|
||||
case ISCSI_IFACE_PARAM_DISCOVERY_LOGOUT_EN:
|
||||
case ISCSI_IFACE_PARAM_STRICT_LOGIN_COMP_EN:
|
||||
case ISCSI_IFACE_PARAM_INITIATOR_NAME:
|
||||
param_type = ISCSI_IFACE_PARAM;
|
||||
break;
|
||||
default:
|
||||
param_type = ISCSI_NET_PARAM;
|
||||
}
|
||||
|
||||
return t->attr_is_visible(param_type, param);
|
||||
return t->attr_is_visible(ISCSI_NET_PARAM, param);
|
||||
}
|
||||
|
||||
static struct attribute *iscsi_iface_attrs[] = {
|
||||
|
@ -224,7 +224,7 @@ static unsigned int sr_get_events(struct scsi_device *sdev)
|
||||
else if (med->media_event_code == 2)
|
||||
return DISK_EVENT_MEDIA_CHANGE;
|
||||
else if (med->media_event_code == 3)
|
||||
return DISK_EVENT_EJECT_REQUEST;
|
||||
return DISK_EVENT_MEDIA_CHANGE;
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
@ -25,7 +25,7 @@
|
||||
#include "target_core_alua.h"
|
||||
|
||||
static sense_reason_t
|
||||
sbc_check_prot(struct se_device *, struct se_cmd *, unsigned char *, u32, bool);
|
||||
sbc_check_prot(struct se_device *, struct se_cmd *, unsigned char, u32, bool);
|
||||
static sense_reason_t sbc_execute_unmap(struct se_cmd *cmd);
|
||||
|
||||
static sense_reason_t
|
||||
@ -279,14 +279,14 @@ static inline unsigned long long transport_lba_64_ext(unsigned char *cdb)
|
||||
}
|
||||
|
||||
static sense_reason_t
|
||||
sbc_setup_write_same(struct se_cmd *cmd, unsigned char *flags, struct sbc_ops *ops)
|
||||
sbc_setup_write_same(struct se_cmd *cmd, unsigned char flags, struct sbc_ops *ops)
|
||||
{
|
||||
struct se_device *dev = cmd->se_dev;
|
||||
sector_t end_lba = dev->transport->get_blocks(dev) + 1;
|
||||
unsigned int sectors = sbc_get_write_same_sectors(cmd);
|
||||
sense_reason_t ret;
|
||||
|
||||
if ((flags[0] & 0x04) || (flags[0] & 0x02)) {
|
||||
if ((flags & 0x04) || (flags & 0x02)) {
|
||||
pr_err("WRITE_SAME PBDATA and LBDATA"
|
||||
" bits not supported for Block Discard"
|
||||
" Emulation\n");
|
||||
@ -308,7 +308,7 @@ sbc_setup_write_same(struct se_cmd *cmd, unsigned char *flags, struct sbc_ops *o
|
||||
}
|
||||
|
||||
/* We always have ANC_SUP == 0 so setting ANCHOR is always an error */
|
||||
if (flags[0] & 0x10) {
|
||||
if (flags & 0x10) {
|
||||
pr_warn("WRITE SAME with ANCHOR not supported\n");
|
||||
return TCM_INVALID_CDB_FIELD;
|
||||
}
|
||||
@ -316,7 +316,7 @@ sbc_setup_write_same(struct se_cmd *cmd, unsigned char *flags, struct sbc_ops *o
|
||||
* Special case for WRITE_SAME w/ UNMAP=1 that ends up getting
|
||||
* translated into block discard requests within backend code.
|
||||
*/
|
||||
if (flags[0] & 0x08) {
|
||||
if (flags & 0x08) {
|
||||
if (!ops->execute_unmap)
|
||||
return TCM_UNSUPPORTED_SCSI_OPCODE;
|
||||
|
||||
@ -331,7 +331,7 @@ sbc_setup_write_same(struct se_cmd *cmd, unsigned char *flags, struct sbc_ops *o
|
||||
if (!ops->execute_write_same)
|
||||
return TCM_UNSUPPORTED_SCSI_OPCODE;
|
||||
|
||||
ret = sbc_check_prot(dev, cmd, &cmd->t_task_cdb[0], sectors, true);
|
||||
ret = sbc_check_prot(dev, cmd, flags >> 5, sectors, true);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
@ -717,10 +717,9 @@ sbc_set_prot_op_checks(u8 protect, bool fabric_prot, enum target_prot_type prot_
|
||||
}
|
||||
|
||||
static sense_reason_t
|
||||
sbc_check_prot(struct se_device *dev, struct se_cmd *cmd, unsigned char *cdb,
|
||||
sbc_check_prot(struct se_device *dev, struct se_cmd *cmd, unsigned char protect,
|
||||
u32 sectors, bool is_write)
|
||||
{
|
||||
u8 protect = cdb[1] >> 5;
|
||||
int sp_ops = cmd->se_sess->sup_prot_ops;
|
||||
int pi_prot_type = dev->dev_attrib.pi_prot_type;
|
||||
bool fabric_prot = false;
|
||||
@ -768,7 +767,7 @@ sbc_check_prot(struct se_device *dev, struct se_cmd *cmd, unsigned char *cdb,
|
||||
fallthrough;
|
||||
default:
|
||||
pr_err("Unable to determine pi_prot_type for CDB: 0x%02x "
|
||||
"PROTECT: 0x%02x\n", cdb[0], protect);
|
||||
"PROTECT: 0x%02x\n", cmd->t_task_cdb[0], protect);
|
||||
return TCM_INVALID_CDB_FIELD;
|
||||
}
|
||||
|
||||
@ -843,7 +842,7 @@ sbc_parse_cdb(struct se_cmd *cmd, struct sbc_ops *ops)
|
||||
if (sbc_check_dpofua(dev, cmd, cdb))
|
||||
return TCM_INVALID_CDB_FIELD;
|
||||
|
||||
ret = sbc_check_prot(dev, cmd, cdb, sectors, false);
|
||||
ret = sbc_check_prot(dev, cmd, cdb[1] >> 5, sectors, false);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
@ -857,7 +856,7 @@ sbc_parse_cdb(struct se_cmd *cmd, struct sbc_ops *ops)
|
||||
if (sbc_check_dpofua(dev, cmd, cdb))
|
||||
return TCM_INVALID_CDB_FIELD;
|
||||
|
||||
ret = sbc_check_prot(dev, cmd, cdb, sectors, false);
|
||||
ret = sbc_check_prot(dev, cmd, cdb[1] >> 5, sectors, false);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
@ -871,7 +870,7 @@ sbc_parse_cdb(struct se_cmd *cmd, struct sbc_ops *ops)
|
||||
if (sbc_check_dpofua(dev, cmd, cdb))
|
||||
return TCM_INVALID_CDB_FIELD;
|
||||
|
||||
ret = sbc_check_prot(dev, cmd, cdb, sectors, false);
|
||||
ret = sbc_check_prot(dev, cmd, cdb[1] >> 5, sectors, false);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
@ -892,7 +891,7 @@ sbc_parse_cdb(struct se_cmd *cmd, struct sbc_ops *ops)
|
||||
if (sbc_check_dpofua(dev, cmd, cdb))
|
||||
return TCM_INVALID_CDB_FIELD;
|
||||
|
||||
ret = sbc_check_prot(dev, cmd, cdb, sectors, true);
|
||||
ret = sbc_check_prot(dev, cmd, cdb[1] >> 5, sectors, true);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
@ -906,7 +905,7 @@ sbc_parse_cdb(struct se_cmd *cmd, struct sbc_ops *ops)
|
||||
if (sbc_check_dpofua(dev, cmd, cdb))
|
||||
return TCM_INVALID_CDB_FIELD;
|
||||
|
||||
ret = sbc_check_prot(dev, cmd, cdb, sectors, true);
|
||||
ret = sbc_check_prot(dev, cmd, cdb[1] >> 5, sectors, true);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
@ -921,7 +920,7 @@ sbc_parse_cdb(struct se_cmd *cmd, struct sbc_ops *ops)
|
||||
if (sbc_check_dpofua(dev, cmd, cdb))
|
||||
return TCM_INVALID_CDB_FIELD;
|
||||
|
||||
ret = sbc_check_prot(dev, cmd, cdb, sectors, true);
|
||||
ret = sbc_check_prot(dev, cmd, cdb[1] >> 5, sectors, true);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
@ -980,7 +979,7 @@ sbc_parse_cdb(struct se_cmd *cmd, struct sbc_ops *ops)
|
||||
size = sbc_get_size(cmd, 1);
|
||||
cmd->t_task_lba = get_unaligned_be64(&cdb[12]);
|
||||
|
||||
ret = sbc_setup_write_same(cmd, &cdb[10], ops);
|
||||
ret = sbc_setup_write_same(cmd, cdb[10], ops);
|
||||
if (ret)
|
||||
return ret;
|
||||
break;
|
||||
@ -1079,7 +1078,7 @@ sbc_parse_cdb(struct se_cmd *cmd, struct sbc_ops *ops)
|
||||
size = sbc_get_size(cmd, 1);
|
||||
cmd->t_task_lba = get_unaligned_be64(&cdb[2]);
|
||||
|
||||
ret = sbc_setup_write_same(cmd, &cdb[1], ops);
|
||||
ret = sbc_setup_write_same(cmd, cdb[1], ops);
|
||||
if (ret)
|
||||
return ret;
|
||||
break;
|
||||
@ -1097,7 +1096,7 @@ sbc_parse_cdb(struct se_cmd *cmd, struct sbc_ops *ops)
|
||||
* Follow sbcr26 with WRITE_SAME (10) and check for the existence
|
||||
* of byte 1 bit 3 UNMAP instead of original reserved field
|
||||
*/
|
||||
ret = sbc_setup_write_same(cmd, &cdb[1], ops);
|
||||
ret = sbc_setup_write_same(cmd, cdb[1], ops);
|
||||
if (ret)
|
||||
return ret;
|
||||
break;
|
||||
|
@ -886,7 +886,7 @@ void target_complete_cmd(struct se_cmd *cmd, u8 scsi_status)
|
||||
INIT_WORK(&cmd->work, success ? target_complete_ok_work :
|
||||
target_complete_failure_work);
|
||||
|
||||
if (wwn->cmd_compl_affinity == SE_COMPL_AFFINITY_CPUID)
|
||||
if (!wwn || wwn->cmd_compl_affinity == SE_COMPL_AFFINITY_CPUID)
|
||||
cpu = cmd->cpuid;
|
||||
else
|
||||
cpu = wwn->cmd_compl_affinity;
|
||||
|
Loading…
Reference in New Issue
Block a user