mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-17 01:34:00 +08:00
4dd3cc5caf
The current ccw_device_set_options() sets a specified mask of options and clears those not specified, but there is no way to find out which options have already been set. In order to fix this up, introduce the following interface changes: ccw_device_set_options() now only sets the specified bits, but does not clear those that are not specified. ccw_device_clear_options() clears the specified bits. ccw_device_set_options_mask() provides the old semantics (setting only the specified bits and clearing the others). Device drivers now work as expected. qdio has been adapted. Signed-off-by: Cornelia Huck <cornelia.huck@de.ibm.com> Signed-off-by: Martin Schwidefsky <schwidefsky@de.ibm.com>
648 lines
16 KiB
C
648 lines
16 KiB
C
/*
|
|
* drivers/s390/cio/device_ops.c
|
|
*
|
|
* Copyright (C) 2002 IBM Deutschland Entwicklung GmbH,
|
|
* IBM Corporation
|
|
* Author(s): Martin Schwidefsky (schwidefsky@de.ibm.com)
|
|
* Cornelia Huck (cornelia.huck@de.ibm.com)
|
|
*/
|
|
#include <linux/module.h>
|
|
#include <linux/init.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/list.h>
|
|
#include <linux/device.h>
|
|
#include <linux/delay.h>
|
|
|
|
#include <asm/ccwdev.h>
|
|
#include <asm/idals.h>
|
|
|
|
#include "cio.h"
|
|
#include "cio_debug.h"
|
|
#include "css.h"
|
|
#include "chsc.h"
|
|
#include "device.h"
|
|
|
|
int ccw_device_set_options_mask(struct ccw_device *cdev, unsigned long flags)
|
|
{
|
|
/*
|
|
* The flag usage is mutal exclusive ...
|
|
*/
|
|
if ((flags & CCWDEV_EARLY_NOTIFICATION) &&
|
|
(flags & CCWDEV_REPORT_ALL))
|
|
return -EINVAL;
|
|
cdev->private->options.fast = (flags & CCWDEV_EARLY_NOTIFICATION) != 0;
|
|
cdev->private->options.repall = (flags & CCWDEV_REPORT_ALL) != 0;
|
|
cdev->private->options.pgroup = (flags & CCWDEV_DO_PATHGROUP) != 0;
|
|
cdev->private->options.force = (flags & CCWDEV_ALLOW_FORCE) != 0;
|
|
return 0;
|
|
}
|
|
|
|
int ccw_device_set_options(struct ccw_device *cdev, unsigned long flags)
|
|
{
|
|
/*
|
|
* The flag usage is mutal exclusive ...
|
|
*/
|
|
if (((flags & CCWDEV_EARLY_NOTIFICATION) &&
|
|
(flags & CCWDEV_REPORT_ALL)) ||
|
|
((flags & CCWDEV_EARLY_NOTIFICATION) &&
|
|
cdev->private->options.repall) ||
|
|
((flags & CCWDEV_REPORT_ALL) &&
|
|
cdev->private->options.fast))
|
|
return -EINVAL;
|
|
cdev->private->options.fast |= (flags & CCWDEV_EARLY_NOTIFICATION) != 0;
|
|
cdev->private->options.repall |= (flags & CCWDEV_REPORT_ALL) != 0;
|
|
cdev->private->options.pgroup |= (flags & CCWDEV_DO_PATHGROUP) != 0;
|
|
cdev->private->options.force |= (flags & CCWDEV_ALLOW_FORCE) != 0;
|
|
return 0;
|
|
}
|
|
|
|
void ccw_device_clear_options(struct ccw_device *cdev, unsigned long flags)
|
|
{
|
|
cdev->private->options.fast &= (flags & CCWDEV_EARLY_NOTIFICATION) == 0;
|
|
cdev->private->options.repall &= (flags & CCWDEV_REPORT_ALL) == 0;
|
|
cdev->private->options.pgroup &= (flags & CCWDEV_DO_PATHGROUP) == 0;
|
|
cdev->private->options.force &= (flags & CCWDEV_ALLOW_FORCE) == 0;
|
|
}
|
|
|
|
int
|
|
ccw_device_clear(struct ccw_device *cdev, unsigned long intparm)
|
|
{
|
|
struct subchannel *sch;
|
|
int ret;
|
|
|
|
if (!cdev)
|
|
return -ENODEV;
|
|
if (cdev->private->state == DEV_STATE_NOT_OPER)
|
|
return -ENODEV;
|
|
if (cdev->private->state != DEV_STATE_ONLINE &&
|
|
cdev->private->state != DEV_STATE_W4SENSE)
|
|
return -EINVAL;
|
|
sch = to_subchannel(cdev->dev.parent);
|
|
if (!sch)
|
|
return -ENODEV;
|
|
ret = cio_clear(sch);
|
|
if (ret == 0)
|
|
cdev->private->intparm = intparm;
|
|
return ret;
|
|
}
|
|
|
|
int
|
|
ccw_device_start_key(struct ccw_device *cdev, struct ccw1 *cpa,
|
|
unsigned long intparm, __u8 lpm, __u8 key,
|
|
unsigned long flags)
|
|
{
|
|
struct subchannel *sch;
|
|
int ret;
|
|
|
|
if (!cdev)
|
|
return -ENODEV;
|
|
sch = to_subchannel(cdev->dev.parent);
|
|
if (!sch)
|
|
return -ENODEV;
|
|
if (cdev->private->state == DEV_STATE_NOT_OPER)
|
|
return -ENODEV;
|
|
if (cdev->private->state == DEV_STATE_VERIFY ||
|
|
cdev->private->state == DEV_STATE_CLEAR_VERIFY) {
|
|
/* Remember to fake irb when finished. */
|
|
if (!cdev->private->flags.fake_irb) {
|
|
cdev->private->flags.fake_irb = 1;
|
|
cdev->private->intparm = intparm;
|
|
return 0;
|
|
} else
|
|
/* There's already a fake I/O around. */
|
|
return -EBUSY;
|
|
}
|
|
if (cdev->private->state != DEV_STATE_ONLINE ||
|
|
((sch->schib.scsw.stctl & SCSW_STCTL_PRIM_STATUS) &&
|
|
!(sch->schib.scsw.stctl & SCSW_STCTL_SEC_STATUS)) ||
|
|
cdev->private->flags.doverify)
|
|
return -EBUSY;
|
|
ret = cio_set_options (sch, flags);
|
|
if (ret)
|
|
return ret;
|
|
/* Adjust requested path mask to excluded varied off paths. */
|
|
if (lpm) {
|
|
lpm &= sch->opm;
|
|
if (lpm == 0)
|
|
return -EACCES;
|
|
}
|
|
ret = cio_start_key (sch, cpa, lpm, key);
|
|
if (ret == 0)
|
|
cdev->private->intparm = intparm;
|
|
return ret;
|
|
}
|
|
|
|
|
|
int
|
|
ccw_device_start_timeout_key(struct ccw_device *cdev, struct ccw1 *cpa,
|
|
unsigned long intparm, __u8 lpm, __u8 key,
|
|
unsigned long flags, int expires)
|
|
{
|
|
int ret;
|
|
|
|
if (!cdev)
|
|
return -ENODEV;
|
|
ccw_device_set_timeout(cdev, expires);
|
|
ret = ccw_device_start_key(cdev, cpa, intparm, lpm, key, flags);
|
|
if (ret != 0)
|
|
ccw_device_set_timeout(cdev, 0);
|
|
return ret;
|
|
}
|
|
|
|
int
|
|
ccw_device_start(struct ccw_device *cdev, struct ccw1 *cpa,
|
|
unsigned long intparm, __u8 lpm, unsigned long flags)
|
|
{
|
|
return ccw_device_start_key(cdev, cpa, intparm, lpm,
|
|
PAGE_DEFAULT_KEY, flags);
|
|
}
|
|
|
|
int
|
|
ccw_device_start_timeout(struct ccw_device *cdev, struct ccw1 *cpa,
|
|
unsigned long intparm, __u8 lpm, unsigned long flags,
|
|
int expires)
|
|
{
|
|
return ccw_device_start_timeout_key(cdev, cpa, intparm, lpm,
|
|
PAGE_DEFAULT_KEY, flags,
|
|
expires);
|
|
}
|
|
|
|
|
|
int
|
|
ccw_device_halt(struct ccw_device *cdev, unsigned long intparm)
|
|
{
|
|
struct subchannel *sch;
|
|
int ret;
|
|
|
|
if (!cdev)
|
|
return -ENODEV;
|
|
if (cdev->private->state == DEV_STATE_NOT_OPER)
|
|
return -ENODEV;
|
|
if (cdev->private->state != DEV_STATE_ONLINE &&
|
|
cdev->private->state != DEV_STATE_W4SENSE)
|
|
return -EINVAL;
|
|
sch = to_subchannel(cdev->dev.parent);
|
|
if (!sch)
|
|
return -ENODEV;
|
|
ret = cio_halt(sch);
|
|
if (ret == 0)
|
|
cdev->private->intparm = intparm;
|
|
return ret;
|
|
}
|
|
|
|
int
|
|
ccw_device_resume(struct ccw_device *cdev)
|
|
{
|
|
struct subchannel *sch;
|
|
|
|
if (!cdev)
|
|
return -ENODEV;
|
|
sch = to_subchannel(cdev->dev.parent);
|
|
if (!sch)
|
|
return -ENODEV;
|
|
if (cdev->private->state == DEV_STATE_NOT_OPER)
|
|
return -ENODEV;
|
|
if (cdev->private->state != DEV_STATE_ONLINE ||
|
|
!(sch->schib.scsw.actl & SCSW_ACTL_SUSPENDED))
|
|
return -EINVAL;
|
|
return cio_resume(sch);
|
|
}
|
|
|
|
/*
|
|
* Pass interrupt to device driver.
|
|
*/
|
|
int
|
|
ccw_device_call_handler(struct ccw_device *cdev)
|
|
{
|
|
struct subchannel *sch;
|
|
unsigned int stctl;
|
|
int ending_status;
|
|
|
|
sch = to_subchannel(cdev->dev.parent);
|
|
|
|
/*
|
|
* we allow for the device action handler if .
|
|
* - we received ending status
|
|
* - the action handler requested to see all interrupts
|
|
* - we received an intermediate status
|
|
* - fast notification was requested (primary status)
|
|
* - unsolicited interrupts
|
|
*/
|
|
stctl = cdev->private->irb.scsw.stctl;
|
|
ending_status = (stctl & SCSW_STCTL_SEC_STATUS) ||
|
|
(stctl == (SCSW_STCTL_ALERT_STATUS | SCSW_STCTL_STATUS_PEND)) ||
|
|
(stctl == SCSW_STCTL_STATUS_PEND);
|
|
if (!ending_status &&
|
|
!cdev->private->options.repall &&
|
|
!(stctl & SCSW_STCTL_INTER_STATUS) &&
|
|
!(cdev->private->options.fast &&
|
|
(stctl & SCSW_STCTL_PRIM_STATUS)))
|
|
return 0;
|
|
|
|
/* Clear pending timers for device driver initiated I/O. */
|
|
if (ending_status)
|
|
ccw_device_set_timeout(cdev, 0);
|
|
/*
|
|
* Now we are ready to call the device driver interrupt handler.
|
|
*/
|
|
if (cdev->handler)
|
|
cdev->handler(cdev, cdev->private->intparm,
|
|
&cdev->private->irb);
|
|
|
|
/*
|
|
* Clear the old and now useless interrupt response block.
|
|
*/
|
|
memset(&cdev->private->irb, 0, sizeof(struct irb));
|
|
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* Search for CIW command in extended sense data.
|
|
*/
|
|
struct ciw *
|
|
ccw_device_get_ciw(struct ccw_device *cdev, __u32 ct)
|
|
{
|
|
int ciw_cnt;
|
|
|
|
if (cdev->private->flags.esid == 0)
|
|
return NULL;
|
|
for (ciw_cnt = 0; ciw_cnt < MAX_CIWS; ciw_cnt++)
|
|
if (cdev->private->senseid.ciw[ciw_cnt].ct == ct)
|
|
return cdev->private->senseid.ciw + ciw_cnt;
|
|
return NULL;
|
|
}
|
|
|
|
__u8
|
|
ccw_device_get_path_mask(struct ccw_device *cdev)
|
|
{
|
|
struct subchannel *sch;
|
|
|
|
sch = to_subchannel(cdev->dev.parent);
|
|
if (!sch)
|
|
return 0;
|
|
else
|
|
return sch->lpm;
|
|
}
|
|
|
|
static void
|
|
ccw_device_wake_up(struct ccw_device *cdev, unsigned long ip, struct irb *irb)
|
|
{
|
|
if (!ip)
|
|
/* unsolicited interrupt */
|
|
return;
|
|
|
|
/* Abuse intparm for error reporting. */
|
|
if (IS_ERR(irb))
|
|
cdev->private->intparm = -EIO;
|
|
else if (irb->scsw.cc == 1)
|
|
/* Retry for deferred condition code. */
|
|
cdev->private->intparm = -EAGAIN;
|
|
else if ((irb->scsw.dstat !=
|
|
(DEV_STAT_CHN_END|DEV_STAT_DEV_END)) ||
|
|
(irb->scsw.cstat != 0)) {
|
|
/*
|
|
* We didn't get channel end / device end. Check if path
|
|
* verification has been started; we can retry after it has
|
|
* finished. We also retry unit checks except for command reject
|
|
* or intervention required. Also check for long busy
|
|
* conditions.
|
|
*/
|
|
if (cdev->private->flags.doverify ||
|
|
cdev->private->state == DEV_STATE_VERIFY)
|
|
cdev->private->intparm = -EAGAIN;
|
|
else if ((irb->scsw.dstat & DEV_STAT_UNIT_CHECK) &&
|
|
!(irb->ecw[0] &
|
|
(SNS0_CMD_REJECT | SNS0_INTERVENTION_REQ)))
|
|
cdev->private->intparm = -EAGAIN;
|
|
else if ((irb->scsw.dstat & DEV_STAT_ATTENTION) &&
|
|
(irb->scsw.dstat & DEV_STAT_DEV_END) &&
|
|
(irb->scsw.dstat & DEV_STAT_UNIT_EXCEP))
|
|
cdev->private->intparm = -EAGAIN;
|
|
else
|
|
cdev->private->intparm = -EIO;
|
|
|
|
} else
|
|
cdev->private->intparm = 0;
|
|
wake_up(&cdev->private->wait_q);
|
|
}
|
|
|
|
static int
|
|
__ccw_device_retry_loop(struct ccw_device *cdev, struct ccw1 *ccw, long magic, __u8 lpm)
|
|
{
|
|
int ret;
|
|
struct subchannel *sch;
|
|
|
|
sch = to_subchannel(cdev->dev.parent);
|
|
do {
|
|
ccw_device_set_timeout(cdev, 60 * HZ);
|
|
ret = cio_start (sch, ccw, lpm);
|
|
if (ret != 0)
|
|
ccw_device_set_timeout(cdev, 0);
|
|
if (ret == -EBUSY) {
|
|
/* Try again later. */
|
|
spin_unlock_irq(sch->lock);
|
|
msleep(10);
|
|
spin_lock_irq(sch->lock);
|
|
continue;
|
|
}
|
|
if (ret != 0)
|
|
/* Non-retryable error. */
|
|
break;
|
|
/* Wait for end of request. */
|
|
cdev->private->intparm = magic;
|
|
spin_unlock_irq(sch->lock);
|
|
wait_event(cdev->private->wait_q,
|
|
(cdev->private->intparm == -EIO) ||
|
|
(cdev->private->intparm == -EAGAIN) ||
|
|
(cdev->private->intparm == 0));
|
|
spin_lock_irq(sch->lock);
|
|
/* Check at least for channel end / device end */
|
|
if (cdev->private->intparm == -EIO) {
|
|
/* Non-retryable error. */
|
|
ret = -EIO;
|
|
break;
|
|
}
|
|
if (cdev->private->intparm == 0)
|
|
/* Success. */
|
|
break;
|
|
/* Try again later. */
|
|
spin_unlock_irq(sch->lock);
|
|
msleep(10);
|
|
spin_lock_irq(sch->lock);
|
|
} while (1);
|
|
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* read_dev_chars() - read device characteristics
|
|
* @param cdev target ccw device
|
|
* @param buffer pointer to buffer for rdc data
|
|
* @param length size of rdc data
|
|
* @returns 0 for success, negative error value on failure
|
|
*
|
|
* Context:
|
|
* called for online device, lock not held
|
|
**/
|
|
int
|
|
read_dev_chars (struct ccw_device *cdev, void **buffer, int length)
|
|
{
|
|
void (*handler)(struct ccw_device *, unsigned long, struct irb *);
|
|
struct subchannel *sch;
|
|
int ret;
|
|
struct ccw1 *rdc_ccw;
|
|
|
|
if (!cdev)
|
|
return -ENODEV;
|
|
if (!buffer || !length)
|
|
return -EINVAL;
|
|
sch = to_subchannel(cdev->dev.parent);
|
|
|
|
CIO_TRACE_EVENT (4, "rddevch");
|
|
CIO_TRACE_EVENT (4, sch->dev.bus_id);
|
|
|
|
rdc_ccw = kzalloc(sizeof(struct ccw1), GFP_KERNEL | GFP_DMA);
|
|
if (!rdc_ccw)
|
|
return -ENOMEM;
|
|
rdc_ccw->cmd_code = CCW_CMD_RDC;
|
|
rdc_ccw->count = length;
|
|
rdc_ccw->flags = CCW_FLAG_SLI;
|
|
ret = set_normalized_cda (rdc_ccw, (*buffer));
|
|
if (ret != 0) {
|
|
kfree(rdc_ccw);
|
|
return ret;
|
|
}
|
|
|
|
spin_lock_irq(sch->lock);
|
|
/* Save interrupt handler. */
|
|
handler = cdev->handler;
|
|
/* Temporarily install own handler. */
|
|
cdev->handler = ccw_device_wake_up;
|
|
if (cdev->private->state != DEV_STATE_ONLINE)
|
|
ret = -ENODEV;
|
|
else if (((sch->schib.scsw.stctl & SCSW_STCTL_PRIM_STATUS) &&
|
|
!(sch->schib.scsw.stctl & SCSW_STCTL_SEC_STATUS)) ||
|
|
cdev->private->flags.doverify)
|
|
ret = -EBUSY;
|
|
else
|
|
/* 0x00D9C4C3 == ebcdic "RDC" */
|
|
ret = __ccw_device_retry_loop(cdev, rdc_ccw, 0x00D9C4C3, 0);
|
|
|
|
/* Restore interrupt handler. */
|
|
cdev->handler = handler;
|
|
spin_unlock_irq(sch->lock);
|
|
|
|
clear_normalized_cda (rdc_ccw);
|
|
kfree(rdc_ccw);
|
|
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* Read Configuration data using path mask
|
|
*/
|
|
int
|
|
read_conf_data_lpm (struct ccw_device *cdev, void **buffer, int *length, __u8 lpm)
|
|
{
|
|
void (*handler)(struct ccw_device *, unsigned long, struct irb *);
|
|
struct subchannel *sch;
|
|
struct ciw *ciw;
|
|
char *rcd_buf;
|
|
int ret;
|
|
struct ccw1 *rcd_ccw;
|
|
|
|
if (!cdev)
|
|
return -ENODEV;
|
|
if (!buffer || !length)
|
|
return -EINVAL;
|
|
sch = to_subchannel(cdev->dev.parent);
|
|
|
|
CIO_TRACE_EVENT (4, "rdconf");
|
|
CIO_TRACE_EVENT (4, sch->dev.bus_id);
|
|
|
|
/*
|
|
* scan for RCD command in extended SenseID data
|
|
*/
|
|
ciw = ccw_device_get_ciw(cdev, CIW_TYPE_RCD);
|
|
if (!ciw || ciw->cmd == 0)
|
|
return -EOPNOTSUPP;
|
|
|
|
/* Adjust requested path mask to excluded varied off paths. */
|
|
if (lpm) {
|
|
lpm &= sch->opm;
|
|
if (lpm == 0)
|
|
return -EACCES;
|
|
}
|
|
|
|
rcd_ccw = kzalloc(sizeof(struct ccw1), GFP_KERNEL | GFP_DMA);
|
|
if (!rcd_ccw)
|
|
return -ENOMEM;
|
|
rcd_buf = kzalloc(ciw->count, GFP_KERNEL | GFP_DMA);
|
|
if (!rcd_buf) {
|
|
kfree(rcd_ccw);
|
|
return -ENOMEM;
|
|
}
|
|
rcd_ccw->cmd_code = ciw->cmd;
|
|
rcd_ccw->cda = (__u32) __pa (rcd_buf);
|
|
rcd_ccw->count = ciw->count;
|
|
rcd_ccw->flags = CCW_FLAG_SLI;
|
|
|
|
spin_lock_irq(sch->lock);
|
|
/* Save interrupt handler. */
|
|
handler = cdev->handler;
|
|
/* Temporarily install own handler. */
|
|
cdev->handler = ccw_device_wake_up;
|
|
if (cdev->private->state != DEV_STATE_ONLINE)
|
|
ret = -ENODEV;
|
|
else if (((sch->schib.scsw.stctl & SCSW_STCTL_PRIM_STATUS) &&
|
|
!(sch->schib.scsw.stctl & SCSW_STCTL_SEC_STATUS)) ||
|
|
cdev->private->flags.doverify)
|
|
ret = -EBUSY;
|
|
else
|
|
/* 0x00D9C3C4 == ebcdic "RCD" */
|
|
ret = __ccw_device_retry_loop(cdev, rcd_ccw, 0x00D9C3C4, lpm);
|
|
|
|
/* Restore interrupt handler. */
|
|
cdev->handler = handler;
|
|
spin_unlock_irq(sch->lock);
|
|
|
|
/*
|
|
* on success we update the user input parms
|
|
*/
|
|
if (ret) {
|
|
kfree (rcd_buf);
|
|
*buffer = NULL;
|
|
*length = 0;
|
|
} else {
|
|
*length = ciw->count;
|
|
*buffer = rcd_buf;
|
|
}
|
|
kfree(rcd_ccw);
|
|
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* Read Configuration data
|
|
*/
|
|
int
|
|
read_conf_data (struct ccw_device *cdev, void **buffer, int *length)
|
|
{
|
|
return read_conf_data_lpm (cdev, buffer, length, 0);
|
|
}
|
|
|
|
/*
|
|
* Try to break the lock on a boxed device.
|
|
*/
|
|
int
|
|
ccw_device_stlck(struct ccw_device *cdev)
|
|
{
|
|
void *buf, *buf2;
|
|
unsigned long flags;
|
|
struct subchannel *sch;
|
|
int ret;
|
|
|
|
if (!cdev)
|
|
return -ENODEV;
|
|
|
|
if (cdev->drv && !cdev->private->options.force)
|
|
return -EINVAL;
|
|
|
|
sch = to_subchannel(cdev->dev.parent);
|
|
|
|
CIO_TRACE_EVENT(2, "stl lock");
|
|
CIO_TRACE_EVENT(2, cdev->dev.bus_id);
|
|
|
|
buf = kmalloc(32*sizeof(char), GFP_DMA|GFP_KERNEL);
|
|
if (!buf)
|
|
return -ENOMEM;
|
|
buf2 = kmalloc(32*sizeof(char), GFP_DMA|GFP_KERNEL);
|
|
if (!buf2) {
|
|
kfree(buf);
|
|
return -ENOMEM;
|
|
}
|
|
spin_lock_irqsave(sch->lock, flags);
|
|
ret = cio_enable_subchannel(sch, 3);
|
|
if (ret)
|
|
goto out_unlock;
|
|
/*
|
|
* Setup ccw. We chain an unconditional reserve and a release so we
|
|
* only break the lock.
|
|
*/
|
|
cdev->private->iccws[0].cmd_code = CCW_CMD_STLCK;
|
|
cdev->private->iccws[0].cda = (__u32) __pa(buf);
|
|
cdev->private->iccws[0].count = 32;
|
|
cdev->private->iccws[0].flags = CCW_FLAG_CC;
|
|
cdev->private->iccws[1].cmd_code = CCW_CMD_RELEASE;
|
|
cdev->private->iccws[1].cda = (__u32) __pa(buf2);
|
|
cdev->private->iccws[1].count = 32;
|
|
cdev->private->iccws[1].flags = 0;
|
|
ret = cio_start(sch, cdev->private->iccws, 0);
|
|
if (ret) {
|
|
cio_disable_subchannel(sch); //FIXME: return code?
|
|
goto out_unlock;
|
|
}
|
|
cdev->private->irb.scsw.actl |= SCSW_ACTL_START_PEND;
|
|
spin_unlock_irqrestore(sch->lock, flags);
|
|
wait_event(cdev->private->wait_q, cdev->private->irb.scsw.actl == 0);
|
|
spin_lock_irqsave(sch->lock, flags);
|
|
cio_disable_subchannel(sch); //FIXME: return code?
|
|
if ((cdev->private->irb.scsw.dstat !=
|
|
(DEV_STAT_CHN_END|DEV_STAT_DEV_END)) ||
|
|
(cdev->private->irb.scsw.cstat != 0))
|
|
ret = -EIO;
|
|
/* Clear irb. */
|
|
memset(&cdev->private->irb, 0, sizeof(struct irb));
|
|
out_unlock:
|
|
kfree(buf);
|
|
kfree(buf2);
|
|
spin_unlock_irqrestore(sch->lock, flags);
|
|
return ret;
|
|
}
|
|
|
|
void *
|
|
ccw_device_get_chp_desc(struct ccw_device *cdev, int chp_no)
|
|
{
|
|
struct subchannel *sch;
|
|
|
|
sch = to_subchannel(cdev->dev.parent);
|
|
return chsc_get_chp_desc(sch, chp_no);
|
|
}
|
|
|
|
// FIXME: these have to go:
|
|
|
|
int
|
|
_ccw_device_get_subchannel_number(struct ccw_device *cdev)
|
|
{
|
|
return cdev->private->schid.sch_no;
|
|
}
|
|
|
|
int
|
|
_ccw_device_get_device_number(struct ccw_device *cdev)
|
|
{
|
|
return cdev->private->dev_id.devno;
|
|
}
|
|
|
|
|
|
MODULE_LICENSE("GPL");
|
|
EXPORT_SYMBOL(ccw_device_set_options_mask);
|
|
EXPORT_SYMBOL(ccw_device_set_options);
|
|
EXPORT_SYMBOL(ccw_device_clear_options);
|
|
EXPORT_SYMBOL(ccw_device_clear);
|
|
EXPORT_SYMBOL(ccw_device_halt);
|
|
EXPORT_SYMBOL(ccw_device_resume);
|
|
EXPORT_SYMBOL(ccw_device_start_timeout);
|
|
EXPORT_SYMBOL(ccw_device_start);
|
|
EXPORT_SYMBOL(ccw_device_start_timeout_key);
|
|
EXPORT_SYMBOL(ccw_device_start_key);
|
|
EXPORT_SYMBOL(ccw_device_get_ciw);
|
|
EXPORT_SYMBOL(ccw_device_get_path_mask);
|
|
EXPORT_SYMBOL(read_conf_data);
|
|
EXPORT_SYMBOL(read_dev_chars);
|
|
EXPORT_SYMBOL(_ccw_device_get_subchannel_number);
|
|
EXPORT_SYMBOL(_ccw_device_get_device_number);
|
|
EXPORT_SYMBOL_GPL(ccw_device_get_chp_desc);
|
|
EXPORT_SYMBOL_GPL(read_conf_data_lpm);
|