mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-17 01:34:00 +08:00
cciss: factor out cciss_big_passthru
Signed-off-by: Stephen M. Cameron <scameron@beardog.cce.hp.com> Signed-off-by: Jens Axboe <jaxboe@fusionio.com>
This commit is contained in:
parent
f32f125b1c
commit
0c9f5ba7cb
@ -1498,6 +1498,155 @@ static int cciss_passthru(ctlr_info_t *h, void __user *argp)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int cciss_bigpassthru(ctlr_info_t *h, void __user *argp)
|
||||
{
|
||||
BIG_IOCTL_Command_struct *ioc;
|
||||
CommandList_struct *c;
|
||||
unsigned char **buff = NULL;
|
||||
int *buff_size = NULL;
|
||||
u64bit temp64;
|
||||
BYTE sg_used = 0;
|
||||
int status = 0;
|
||||
int i;
|
||||
DECLARE_COMPLETION_ONSTACK(wait);
|
||||
__u32 left;
|
||||
__u32 sz;
|
||||
BYTE __user *data_ptr;
|
||||
|
||||
if (!argp)
|
||||
return -EINVAL;
|
||||
if (!capable(CAP_SYS_RAWIO))
|
||||
return -EPERM;
|
||||
ioc = (BIG_IOCTL_Command_struct *)
|
||||
kmalloc(sizeof(*ioc), GFP_KERNEL);
|
||||
if (!ioc) {
|
||||
status = -ENOMEM;
|
||||
goto cleanup1;
|
||||
}
|
||||
if (copy_from_user(ioc, argp, sizeof(*ioc))) {
|
||||
status = -EFAULT;
|
||||
goto cleanup1;
|
||||
}
|
||||
if ((ioc->buf_size < 1) &&
|
||||
(ioc->Request.Type.Direction != XFER_NONE)) {
|
||||
status = -EINVAL;
|
||||
goto cleanup1;
|
||||
}
|
||||
/* Check kmalloc limits using all SGs */
|
||||
if (ioc->malloc_size > MAX_KMALLOC_SIZE) {
|
||||
status = -EINVAL;
|
||||
goto cleanup1;
|
||||
}
|
||||
if (ioc->buf_size > ioc->malloc_size * MAXSGENTRIES) {
|
||||
status = -EINVAL;
|
||||
goto cleanup1;
|
||||
}
|
||||
buff = kzalloc(MAXSGENTRIES * sizeof(char *), GFP_KERNEL);
|
||||
if (!buff) {
|
||||
status = -ENOMEM;
|
||||
goto cleanup1;
|
||||
}
|
||||
buff_size = kmalloc(MAXSGENTRIES * sizeof(int), GFP_KERNEL);
|
||||
if (!buff_size) {
|
||||
status = -ENOMEM;
|
||||
goto cleanup1;
|
||||
}
|
||||
left = ioc->buf_size;
|
||||
data_ptr = ioc->buf;
|
||||
while (left) {
|
||||
sz = (left > ioc->malloc_size) ? ioc->malloc_size : left;
|
||||
buff_size[sg_used] = sz;
|
||||
buff[sg_used] = kmalloc(sz, GFP_KERNEL);
|
||||
if (buff[sg_used] == NULL) {
|
||||
status = -ENOMEM;
|
||||
goto cleanup1;
|
||||
}
|
||||
if (ioc->Request.Type.Direction == XFER_WRITE) {
|
||||
if (copy_from_user(buff[sg_used], data_ptr, sz)) {
|
||||
status = -EFAULT;
|
||||
goto cleanup1;
|
||||
}
|
||||
} else {
|
||||
memset(buff[sg_used], 0, sz);
|
||||
}
|
||||
left -= sz;
|
||||
data_ptr += sz;
|
||||
sg_used++;
|
||||
}
|
||||
c = cmd_special_alloc(h);
|
||||
if (!c) {
|
||||
status = -ENOMEM;
|
||||
goto cleanup1;
|
||||
}
|
||||
c->cmd_type = CMD_IOCTL_PEND;
|
||||
c->Header.ReplyQueue = 0;
|
||||
|
||||
if (ioc->buf_size > 0) {
|
||||
c->Header.SGList = sg_used;
|
||||
c->Header.SGTotal = sg_used;
|
||||
} else {
|
||||
c->Header.SGList = 0;
|
||||
c->Header.SGTotal = 0;
|
||||
}
|
||||
c->Header.LUN = ioc->LUN_info;
|
||||
c->Header.Tag.lower = c->busaddr;
|
||||
|
||||
c->Request = ioc->Request;
|
||||
if (ioc->buf_size > 0) {
|
||||
for (i = 0; i < sg_used; i++) {
|
||||
temp64.val =
|
||||
pci_map_single(h->pdev, buff[i], buff_size[i],
|
||||
PCI_DMA_BIDIRECTIONAL);
|
||||
c->SG[i].Addr.lower = temp64.val32.lower;
|
||||
c->SG[i].Addr.upper = temp64.val32.upper;
|
||||
c->SG[i].Len = buff_size[i];
|
||||
c->SG[i].Ext = 0; /* we are not chaining */
|
||||
}
|
||||
}
|
||||
c->waiting = &wait;
|
||||
enqueue_cmd_and_start_io(h, c);
|
||||
wait_for_completion(&wait);
|
||||
/* unlock the buffers from DMA */
|
||||
for (i = 0; i < sg_used; i++) {
|
||||
temp64.val32.lower = c->SG[i].Addr.lower;
|
||||
temp64.val32.upper = c->SG[i].Addr.upper;
|
||||
pci_unmap_single(h->pdev,
|
||||
(dma_addr_t) temp64.val, buff_size[i],
|
||||
PCI_DMA_BIDIRECTIONAL);
|
||||
}
|
||||
check_ioctl_unit_attention(h, c);
|
||||
/* Copy the error information out */
|
||||
ioc->error_info = *(c->err_info);
|
||||
if (copy_to_user(argp, ioc, sizeof(*ioc))) {
|
||||
cmd_special_free(h, c);
|
||||
status = -EFAULT;
|
||||
goto cleanup1;
|
||||
}
|
||||
if (ioc->Request.Type.Direction == XFER_READ) {
|
||||
/* Copy the data out of the buffer we created */
|
||||
BYTE __user *ptr = ioc->buf;
|
||||
for (i = 0; i < sg_used; i++) {
|
||||
if (copy_to_user(ptr, buff[i], buff_size[i])) {
|
||||
cmd_special_free(h, c);
|
||||
status = -EFAULT;
|
||||
goto cleanup1;
|
||||
}
|
||||
ptr += buff_size[i];
|
||||
}
|
||||
}
|
||||
cmd_special_free(h, c);
|
||||
status = 0;
|
||||
cleanup1:
|
||||
if (buff) {
|
||||
for (i = 0; i < sg_used; i++)
|
||||
kfree(buff[i]);
|
||||
kfree(buff);
|
||||
}
|
||||
kfree(buff_size);
|
||||
kfree(ioc);
|
||||
return status;
|
||||
}
|
||||
|
||||
static int cciss_ioctl(struct block_device *bdev, fmode_t mode,
|
||||
unsigned int cmd, unsigned long arg)
|
||||
{
|
||||
@ -1534,162 +1683,8 @@ static int cciss_ioctl(struct block_device *bdev, fmode_t mode,
|
||||
return cciss_getluninfo(h, disk, argp);
|
||||
case CCISS_PASSTHRU:
|
||||
return cciss_passthru(h, argp);
|
||||
case CCISS_BIG_PASSTHRU:{
|
||||
BIG_IOCTL_Command_struct *ioc;
|
||||
CommandList_struct *c;
|
||||
unsigned char **buff = NULL;
|
||||
int *buff_size = NULL;
|
||||
u64bit temp64;
|
||||
BYTE sg_used = 0;
|
||||
int status = 0;
|
||||
int i;
|
||||
DECLARE_COMPLETION_ONSTACK(wait);
|
||||
__u32 left;
|
||||
__u32 sz;
|
||||
BYTE __user *data_ptr;
|
||||
|
||||
if (!arg)
|
||||
return -EINVAL;
|
||||
if (!capable(CAP_SYS_RAWIO))
|
||||
return -EPERM;
|
||||
ioc = (BIG_IOCTL_Command_struct *)
|
||||
kmalloc(sizeof(*ioc), GFP_KERNEL);
|
||||
if (!ioc) {
|
||||
status = -ENOMEM;
|
||||
goto cleanup1;
|
||||
}
|
||||
if (copy_from_user(ioc, argp, sizeof(*ioc))) {
|
||||
status = -EFAULT;
|
||||
goto cleanup1;
|
||||
}
|
||||
if ((ioc->buf_size < 1) &&
|
||||
(ioc->Request.Type.Direction != XFER_NONE)) {
|
||||
status = -EINVAL;
|
||||
goto cleanup1;
|
||||
}
|
||||
/* Check kmalloc limits using all SGs */
|
||||
if (ioc->malloc_size > MAX_KMALLOC_SIZE) {
|
||||
status = -EINVAL;
|
||||
goto cleanup1;
|
||||
}
|
||||
if (ioc->buf_size > ioc->malloc_size * MAXSGENTRIES) {
|
||||
status = -EINVAL;
|
||||
goto cleanup1;
|
||||
}
|
||||
buff =
|
||||
kzalloc(MAXSGENTRIES * sizeof(char *), GFP_KERNEL);
|
||||
if (!buff) {
|
||||
status = -ENOMEM;
|
||||
goto cleanup1;
|
||||
}
|
||||
buff_size = kmalloc(MAXSGENTRIES * sizeof(int),
|
||||
GFP_KERNEL);
|
||||
if (!buff_size) {
|
||||
status = -ENOMEM;
|
||||
goto cleanup1;
|
||||
}
|
||||
left = ioc->buf_size;
|
||||
data_ptr = ioc->buf;
|
||||
while (left) {
|
||||
sz = (left >
|
||||
ioc->malloc_size) ? ioc->
|
||||
malloc_size : left;
|
||||
buff_size[sg_used] = sz;
|
||||
buff[sg_used] = kmalloc(sz, GFP_KERNEL);
|
||||
if (buff[sg_used] == NULL) {
|
||||
status = -ENOMEM;
|
||||
goto cleanup1;
|
||||
}
|
||||
if (ioc->Request.Type.Direction == XFER_WRITE) {
|
||||
if (copy_from_user
|
||||
(buff[sg_used], data_ptr, sz)) {
|
||||
status = -EFAULT;
|
||||
goto cleanup1;
|
||||
}
|
||||
} else {
|
||||
memset(buff[sg_used], 0, sz);
|
||||
}
|
||||
left -= sz;
|
||||
data_ptr += sz;
|
||||
sg_used++;
|
||||
}
|
||||
c = cmd_special_alloc(h);
|
||||
if (!c) {
|
||||
status = -ENOMEM;
|
||||
goto cleanup1;
|
||||
}
|
||||
c->cmd_type = CMD_IOCTL_PEND;
|
||||
c->Header.ReplyQueue = 0;
|
||||
|
||||
if (ioc->buf_size > 0) {
|
||||
c->Header.SGList = sg_used;
|
||||
c->Header.SGTotal = sg_used;
|
||||
} else {
|
||||
c->Header.SGList = 0;
|
||||
c->Header.SGTotal = 0;
|
||||
}
|
||||
c->Header.LUN = ioc->LUN_info;
|
||||
c->Header.Tag.lower = c->busaddr;
|
||||
|
||||
c->Request = ioc->Request;
|
||||
if (ioc->buf_size > 0) {
|
||||
for (i = 0; i < sg_used; i++) {
|
||||
temp64.val =
|
||||
pci_map_single(h->pdev, buff[i],
|
||||
buff_size[i],
|
||||
PCI_DMA_BIDIRECTIONAL);
|
||||
c->SG[i].Addr.lower =
|
||||
temp64.val32.lower;
|
||||
c->SG[i].Addr.upper =
|
||||
temp64.val32.upper;
|
||||
c->SG[i].Len = buff_size[i];
|
||||
c->SG[i].Ext = 0; /* we are not chaining */
|
||||
}
|
||||
}
|
||||
c->waiting = &wait;
|
||||
enqueue_cmd_and_start_io(h, c);
|
||||
wait_for_completion(&wait);
|
||||
/* unlock the buffers from DMA */
|
||||
for (i = 0; i < sg_used; i++) {
|
||||
temp64.val32.lower = c->SG[i].Addr.lower;
|
||||
temp64.val32.upper = c->SG[i].Addr.upper;
|
||||
pci_unmap_single(h->pdev,
|
||||
(dma_addr_t) temp64.val, buff_size[i],
|
||||
PCI_DMA_BIDIRECTIONAL);
|
||||
}
|
||||
check_ioctl_unit_attention(h, c);
|
||||
/* Copy the error information out */
|
||||
ioc->error_info = *(c->err_info);
|
||||
if (copy_to_user(argp, ioc, sizeof(*ioc))) {
|
||||
cmd_special_free(h, c);
|
||||
status = -EFAULT;
|
||||
goto cleanup1;
|
||||
}
|
||||
if (ioc->Request.Type.Direction == XFER_READ) {
|
||||
/* Copy the data out of the buffer we created */
|
||||
BYTE __user *ptr = ioc->buf;
|
||||
for (i = 0; i < sg_used; i++) {
|
||||
if (copy_to_user
|
||||
(ptr, buff[i], buff_size[i])) {
|
||||
cmd_special_free(h, c);
|
||||
status = -EFAULT;
|
||||
goto cleanup1;
|
||||
}
|
||||
ptr += buff_size[i];
|
||||
}
|
||||
}
|
||||
cmd_special_free(h, c);
|
||||
status = 0;
|
||||
cleanup1:
|
||||
if (buff) {
|
||||
for (i = 0; i < sg_used; i++)
|
||||
kfree(buff[i]);
|
||||
kfree(buff);
|
||||
}
|
||||
kfree(buff_size);
|
||||
kfree(ioc);
|
||||
return status;
|
||||
}
|
||||
case CCISS_BIG_PASSTHRU:
|
||||
return cciss_bigpassthru(h, argp);
|
||||
|
||||
/* scsi_cmd_ioctl handles these, below, though some are not */
|
||||
/* very meaningful for cciss. SG_IO is the main one people want. */
|
||||
|
Loading…
Reference in New Issue
Block a user