mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-19 18:53:52 +08:00
nvme/pci: Switch to pci_request_irq()
Signed-off-by: Christoph Hellwig <hch@lst.de> Signed-off-by: Bjorn Helgaas <bhelgaas@google.com> Reviewed-by: Thomas Gleixner <tglx@linutronix.de> Reviewed-by: Keith Busch <keith.busch@intel.com>
This commit is contained in:
parent
704e8953d3
commit
0ff199cb48
@ -117,7 +117,6 @@ static inline struct nvme_dev *to_nvme_dev(struct nvme_ctrl *ctrl)
|
||||
struct nvme_queue {
|
||||
struct device *q_dmadev;
|
||||
struct nvme_dev *dev;
|
||||
char irqname[24]; /* nvme4294967295-65535\0 */
|
||||
spinlock_t q_lock;
|
||||
struct nvme_command *sq_cmds;
|
||||
struct nvme_command __iomem *sq_cmds_io;
|
||||
@ -204,11 +203,6 @@ static unsigned int nvme_cmd_size(struct nvme_dev *dev)
|
||||
nvme_iod_alloc_size(dev, NVME_INT_BYTES(dev), NVME_INT_PAGES);
|
||||
}
|
||||
|
||||
static int nvmeq_irq(struct nvme_queue *nvmeq)
|
||||
{
|
||||
return pci_irq_vector(to_pci_dev(nvmeq->dev->dev), nvmeq->cq_vector);
|
||||
}
|
||||
|
||||
static int nvme_admin_init_hctx(struct blk_mq_hw_ctx *hctx, void *data,
|
||||
unsigned int hctx_idx)
|
||||
{
|
||||
@ -962,7 +956,7 @@ static int nvme_suspend_queue(struct nvme_queue *nvmeq)
|
||||
spin_unlock_irq(&nvmeq->q_lock);
|
||||
return 1;
|
||||
}
|
||||
vector = nvmeq_irq(nvmeq);
|
||||
vector = nvmeq->cq_vector;
|
||||
nvmeq->dev->online_queues--;
|
||||
nvmeq->cq_vector = -1;
|
||||
spin_unlock_irq(&nvmeq->q_lock);
|
||||
@ -970,7 +964,7 @@ static int nvme_suspend_queue(struct nvme_queue *nvmeq)
|
||||
if (!nvmeq->qid && nvmeq->dev->ctrl.admin_q)
|
||||
blk_mq_stop_hw_queues(nvmeq->dev->ctrl.admin_q);
|
||||
|
||||
free_irq(vector, nvmeq);
|
||||
pci_free_irq(to_pci_dev(nvmeq->dev->dev), vector, nvmeq);
|
||||
|
||||
return 0;
|
||||
}
|
||||
@ -1055,8 +1049,6 @@ static struct nvme_queue *nvme_alloc_queue(struct nvme_dev *dev, int qid,
|
||||
|
||||
nvmeq->q_dmadev = dev->dev;
|
||||
nvmeq->dev = dev;
|
||||
snprintf(nvmeq->irqname, sizeof(nvmeq->irqname), "nvme%dq%d",
|
||||
dev->ctrl.instance, qid);
|
||||
spin_lock_init(&nvmeq->q_lock);
|
||||
nvmeq->cq_head = 0;
|
||||
nvmeq->cq_phase = 1;
|
||||
@ -1079,12 +1071,16 @@ static struct nvme_queue *nvme_alloc_queue(struct nvme_dev *dev, int qid,
|
||||
|
||||
static int queue_request_irq(struct nvme_queue *nvmeq)
|
||||
{
|
||||
if (use_threaded_interrupts)
|
||||
return request_threaded_irq(nvmeq_irq(nvmeq), nvme_irq_check,
|
||||
nvme_irq, IRQF_SHARED, nvmeq->irqname, nvmeq);
|
||||
else
|
||||
return request_irq(nvmeq_irq(nvmeq), nvme_irq, IRQF_SHARED,
|
||||
nvmeq->irqname, nvmeq);
|
||||
struct pci_dev *pdev = to_pci_dev(nvmeq->dev->dev);
|
||||
int nr = nvmeq->dev->ctrl.instance;
|
||||
|
||||
if (use_threaded_interrupts) {
|
||||
return pci_request_irq(pdev, nvmeq->cq_vector, nvme_irq_check,
|
||||
nvme_irq, nvmeq, "nvme%dq%d", nr, nvmeq->qid);
|
||||
} else {
|
||||
return pci_request_irq(pdev, nvmeq->cq_vector, nvme_irq,
|
||||
NULL, nvmeq, "nvme%dq%d", nr, nvmeq->qid);
|
||||
}
|
||||
}
|
||||
|
||||
static void nvme_init_queue(struct nvme_queue *nvmeq, u16 qid)
|
||||
@ -1440,7 +1436,7 @@ static int nvme_setup_io_queues(struct nvme_dev *dev)
|
||||
}
|
||||
|
||||
/* Deregister the admin queue's interrupt */
|
||||
free_irq(pci_irq_vector(pdev, 0), adminq);
|
||||
pci_free_irq(pdev, 0, adminq);
|
||||
|
||||
/*
|
||||
* If we enable msix early due to not intx, disable it again before
|
||||
|
Loading…
Reference in New Issue
Block a user