nvme: switch abort to blk_execute_rq_nowait
And remove the now unused nvme_submit_cmd helper. Signed-off-by: Christoph Hellwig <hch@lst.de> Acked-by: Keith Busch <keith.busch@intel.com> Signed-off-by: Jens Axboe <axboe@fb.com>
This commit is contained in:
Родитель
d8f32166a9
Коммит
e7a2a87d59
|
@ -375,20 +375,6 @@ static void async_req_completion(struct nvme_queue *nvmeq, void *ctx,
|
|||
}
|
||||
}
|
||||
|
||||
static void abort_completion(struct nvme_queue *nvmeq, void *ctx,
|
||||
struct nvme_completion *cqe)
|
||||
{
|
||||
struct request *req = ctx;
|
||||
|
||||
u16 status = le16_to_cpup(&cqe->status) >> 1;
|
||||
u32 result = le32_to_cpup(&cqe->result);
|
||||
|
||||
blk_mq_free_request(req);
|
||||
|
||||
dev_warn(nvmeq->q_dmadev, "Abort status:%x result:%x", status, result);
|
||||
atomic_inc(&nvmeq->dev->ctrl.abort_limit);
|
||||
}
|
||||
|
||||
static inline struct nvme_cmd_info *get_cmd_from_tag(struct nvme_queue *nvmeq,
|
||||
unsigned int tag)
|
||||
{
|
||||
|
@ -440,14 +426,6 @@ static void __nvme_submit_cmd(struct nvme_queue *nvmeq,
|
|||
nvmeq->sq_tail = tail;
|
||||
}
|
||||
|
||||
static void nvme_submit_cmd(struct nvme_queue *nvmeq, struct nvme_command *cmd)
|
||||
{
|
||||
unsigned long flags;
|
||||
spin_lock_irqsave(&nvmeq->q_lock, flags);
|
||||
__nvme_submit_cmd(nvmeq, cmd);
|
||||
spin_unlock_irqrestore(&nvmeq->q_lock, flags);
|
||||
}
|
||||
|
||||
static __le64 **iod_list(struct nvme_iod *iod)
|
||||
{
|
||||
return ((void *)iod) + iod->offset;
|
||||
|
@ -1045,13 +1023,25 @@ static int adapter_delete_sq(struct nvme_dev *dev, u16 sqid)
|
|||
return adapter_delete_queue(dev, nvme_admin_delete_sq, sqid);
|
||||
}
|
||||
|
||||
static void abort_endio(struct request *req, int error)
|
||||
{
|
||||
struct nvme_cmd_info *cmd = blk_mq_rq_to_pdu(req);
|
||||
struct nvme_queue *nvmeq = cmd->nvmeq;
|
||||
u32 result = (u32)(uintptr_t)req->special;
|
||||
u16 status = req->errors;
|
||||
|
||||
dev_warn(nvmeq->q_dmadev, "Abort status:%x result:%x", status, result);
|
||||
atomic_inc(&nvmeq->dev->ctrl.abort_limit);
|
||||
|
||||
blk_mq_free_request(req);
|
||||
}
|
||||
|
||||
static enum blk_eh_timer_return nvme_timeout(struct request *req, bool reserved)
|
||||
{
|
||||
struct nvme_cmd_info *cmd_rq = blk_mq_rq_to_pdu(req);
|
||||
struct nvme_queue *nvmeq = cmd_rq->nvmeq;
|
||||
struct nvme_dev *dev = nvmeq->dev;
|
||||
struct request *abort_req;
|
||||
struct nvme_cmd_info *abort_cmd;
|
||||
struct nvme_command cmd;
|
||||
|
||||
/*
|
||||
|
@ -1089,30 +1079,31 @@ static enum blk_eh_timer_return nvme_timeout(struct request *req, bool reserved)
|
|||
return BLK_EH_HANDLED;
|
||||
}
|
||||
|
||||
if (atomic_dec_and_test(&dev->ctrl.abort_limit))
|
||||
return BLK_EH_RESET_TIMER;
|
||||
cmd_rq->aborted = 1;
|
||||
|
||||
abort_req = blk_mq_alloc_request(dev->ctrl.admin_q, WRITE,
|
||||
if (atomic_dec_return(&dev->ctrl.abort_limit) < 0) {
|
||||
atomic_inc(&dev->ctrl.abort_limit);
|
||||
return BLK_EH_RESET_TIMER;
|
||||
}
|
||||
|
||||
memset(&cmd, 0, sizeof(cmd));
|
||||
cmd.abort.opcode = nvme_admin_abort_cmd;
|
||||
cmd.abort.cid = req->tag;
|
||||
cmd.abort.sqid = cpu_to_le16(nvmeq->qid);
|
||||
|
||||
dev_warn(nvmeq->q_dmadev, "I/O %d QID %d timeout, aborting\n",
|
||||
req->tag, nvmeq->qid);
|
||||
|
||||
abort_req = nvme_alloc_request(dev->ctrl.admin_q, &cmd,
|
||||
BLK_MQ_REQ_NOWAIT);
|
||||
if (IS_ERR(abort_req)) {
|
||||
atomic_inc(&dev->ctrl.abort_limit);
|
||||
return BLK_EH_RESET_TIMER;
|
||||
}
|
||||
|
||||
abort_cmd = blk_mq_rq_to_pdu(abort_req);
|
||||
nvme_set_info(abort_cmd, abort_req, abort_completion);
|
||||
|
||||
memset(&cmd, 0, sizeof(cmd));
|
||||
cmd.abort.opcode = nvme_admin_abort_cmd;
|
||||
cmd.abort.cid = req->tag;
|
||||
cmd.abort.sqid = cpu_to_le16(nvmeq->qid);
|
||||
cmd.abort.command_id = abort_req->tag;
|
||||
|
||||
cmd_rq->aborted = 1;
|
||||
|
||||
dev_warn(nvmeq->q_dmadev, "I/O %d QID %d timeout, aborting\n",
|
||||
req->tag, nvmeq->qid);
|
||||
nvme_submit_cmd(dev->queues[0], &cmd);
|
||||
abort_req->timeout = ADMIN_TIMEOUT;
|
||||
abort_req->end_io_data = NULL;
|
||||
blk_execute_rq_nowait(abort_req->q, NULL, abort_req, 0, abort_endio);
|
||||
|
||||
/*
|
||||
* The aborted req will be completed on receiving the abort req.
|
||||
|
|
Загрузка…
Ссылка в новой задаче