nvme-pci: consolidate code for polling non-dedicated queues

We have three places that can poll for I/O completions on a normal
interrupt-enabled queue.  All of them are in slow path code, so
consolidate them to a single helper that uses spin_lock_irqsave and
removes the fast path cqe_pending check.

Reviewed-by: Sagi Grimberg <sagi@grimberg.me>
Signed-off-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Jens Axboe <axboe@kernel.dk>
This commit is contained in:
Christoph Hellwig 2018-12-02 17:46:20 +01:00 committed by Jens Axboe
parent c6d962aeba
commit 0b2a8a9f4b

View File

@ -1072,17 +1072,19 @@ static irqreturn_t nvme_irq_check(int irq, void *data)
return IRQ_NONE; return IRQ_NONE;
} }
static int __nvme_poll(struct nvme_queue *nvmeq, unsigned int tag) /*
* Poll for completions any queue, including those not dedicated to polling.
* Can be called from any context.
*/
static int nvme_poll_irqdisable(struct nvme_queue *nvmeq, unsigned int tag)
{ {
unsigned long flags;
u16 start, end; u16 start, end;
int found; int found;
if (!nvme_cqe_pending(nvmeq)) spin_lock_irqsave(&nvmeq->cq_lock, flags);
return 0;
spin_lock_irq(&nvmeq->cq_lock);
found = nvme_process_cq(nvmeq, &start, &end, tag); found = nvme_process_cq(nvmeq, &start, &end, tag);
spin_unlock_irq(&nvmeq->cq_lock); spin_unlock_irqrestore(&nvmeq->cq_lock, flags);
nvme_complete_cqes(nvmeq, start, end); nvme_complete_cqes(nvmeq, start, end);
return found; return found;
@ -1279,7 +1281,7 @@ static enum blk_eh_timer_return nvme_timeout(struct request *req, bool reserved)
/* /*
* Did we miss an interrupt? * Did we miss an interrupt?
*/ */
if (__nvme_poll(nvmeq, req->tag)) { if (nvme_poll_irqdisable(nvmeq, req->tag)) {
dev_warn(dev->ctrl.device, dev_warn(dev->ctrl.device,
"I/O %d QID %d timeout, completion polled\n", "I/O %d QID %d timeout, completion polled\n",
req->tag, nvmeq->qid); req->tag, nvmeq->qid);
@ -1406,18 +1408,13 @@ static int nvme_suspend_queue(struct nvme_queue *nvmeq)
static void nvme_disable_admin_queue(struct nvme_dev *dev, bool shutdown) static void nvme_disable_admin_queue(struct nvme_dev *dev, bool shutdown)
{ {
struct nvme_queue *nvmeq = &dev->queues[0]; struct nvme_queue *nvmeq = &dev->queues[0];
u16 start, end;
if (shutdown) if (shutdown)
nvme_shutdown_ctrl(&dev->ctrl); nvme_shutdown_ctrl(&dev->ctrl);
else else
nvme_disable_ctrl(&dev->ctrl, dev->ctrl.cap); nvme_disable_ctrl(&dev->ctrl, dev->ctrl.cap);
spin_lock_irq(&nvmeq->cq_lock); nvme_poll_irqdisable(nvmeq, -1);
nvme_process_cq(nvmeq, &start, &end, -1);
spin_unlock_irq(&nvmeq->cq_lock);
nvme_complete_cqes(nvmeq, start, end);
} }
static int nvme_cmb_qdepth(struct nvme_dev *dev, int nr_io_queues, static int nvme_cmb_qdepth(struct nvme_dev *dev, int nr_io_queues,
@ -2217,17 +2214,9 @@ static void nvme_del_queue_end(struct request *req, blk_status_t error)
static void nvme_del_cq_end(struct request *req, blk_status_t error) static void nvme_del_cq_end(struct request *req, blk_status_t error)
{ {
struct nvme_queue *nvmeq = req->end_io_data; struct nvme_queue *nvmeq = req->end_io_data;
u16 start, end;
if (!error) { if (!error)
unsigned long flags; nvme_poll_irqdisable(nvmeq, -1);
spin_lock_irqsave(&nvmeq->cq_lock, flags);
nvme_process_cq(nvmeq, &start, &end, -1);
spin_unlock_irqrestore(&nvmeq->cq_lock, flags);
nvme_complete_cqes(nvmeq, start, end);
}
nvme_del_queue_end(req, error); nvme_del_queue_end(req, error);
} }