]> git.proxmox.com Git - mirror_ubuntu-zesty-kernel.git/blobdiff - drivers/nvme/host/core.c
nvme: introduce struct nvme_request
[mirror_ubuntu-zesty-kernel.git] / drivers / nvme / host / core.c
index 4669c052239ed21beb09028736e15824dbb04681..2fd632bcd97501b3f1076bb3dfac31c0567e7203 100644 (file)
@@ -201,13 +201,7 @@ fail:
 
 void nvme_requeue_req(struct request *req)
 {
-       unsigned long flags;
-
-       blk_mq_requeue_request(req);
-       spin_lock_irqsave(req->q->queue_lock, flags);
-       if (!blk_queue_stopped(req->q))
-               blk_mq_kick_requeue_list(req->q);
-       spin_unlock_irqrestore(req->q->queue_lock, flags);
+       blk_mq_requeue_request(req, !blk_mq_queue_stopped(req->q));
 }
 EXPORT_SYMBOL_GPL(nvme_requeue_req);
 
@@ -227,8 +221,7 @@ struct request *nvme_alloc_request(struct request_queue *q,
 
        req->cmd_type = REQ_TYPE_DRV_PRIV;
        req->cmd_flags |= REQ_FAILFAST_DRIVER;
-       req->cmd = (unsigned char *)cmd;
-       req->cmd_len = sizeof(struct nvme_command);
+       nvme_req(req)->cmd = cmd;
 
        return req;
 }
@@ -327,7 +320,7 @@ int nvme_setup_cmd(struct nvme_ns *ns, struct request *req,
        int ret = 0;
 
        if (req->cmd_type == REQ_TYPE_DRV_PRIV)
-               memcpy(cmd, req->cmd, sizeof(*cmd));
+               memcpy(cmd, nvme_req(req)->cmd, sizeof(*cmd));
        else if (req_op(req) == REQ_OP_FLUSH)
                nvme_setup_flush(ns, cmd);
        else if (req_op(req) == REQ_OP_DISCARD)
@@ -344,7 +337,7 @@ EXPORT_SYMBOL_GPL(nvme_setup_cmd);
  * if the result is positive, it's an NVM Express status code
  */
 int __nvme_submit_sync_cmd(struct request_queue *q, struct nvme_command *cmd,
-               struct nvme_completion *cqe, void *buffer, unsigned bufflen,
+               union nvme_result *result, void *buffer, unsigned bufflen,
                unsigned timeout, int qid, int at_head, int flags)
 {
        struct request *req;
@@ -355,7 +348,6 @@ int __nvme_submit_sync_cmd(struct request_queue *q, struct nvme_command *cmd,
                return PTR_ERR(req);
 
        req->timeout = timeout ? timeout : ADMIN_TIMEOUT;
-       req->special = cqe;
 
        if (buffer && bufflen) {
                ret = blk_rq_map_kern(q, req, buffer, bufflen, GFP_KERNEL);
@@ -364,6 +356,8 @@ int __nvme_submit_sync_cmd(struct request_queue *q, struct nvme_command *cmd,
        }
 
        blk_execute_rq(req->q, NULL, req, at_head);
+       if (result)
+               *result = nvme_req(req)->result;
        ret = req->errors;
  out:
        blk_mq_free_request(req);
@@ -385,7 +379,6 @@ int __nvme_submit_user_cmd(struct request_queue *q, struct nvme_command *cmd,
                u32 *result, unsigned timeout)
 {
        bool write = nvme_is_write(cmd);
-       struct nvme_completion cqe;
        struct nvme_ns *ns = q->queuedata;
        struct gendisk *disk = ns ? ns->disk : NULL;
        struct request *req;
@@ -398,7 +391,6 @@ int __nvme_submit_user_cmd(struct request_queue *q, struct nvme_command *cmd,
                return PTR_ERR(req);
 
        req->timeout = timeout ? timeout : ADMIN_TIMEOUT;
-       req->special = &cqe;
 
        if (ubuffer && bufflen) {
                ret = blk_rq_map_user(q, req, NULL, ubuffer, bufflen,
@@ -453,7 +445,7 @@ int __nvme_submit_user_cmd(struct request_queue *q, struct nvme_command *cmd,
        blk_execute_rq(req->q, disk, req, 0);
        ret = req->errors;
        if (result)
-               *result = le32_to_cpu(cqe.result);
+               *result = le32_to_cpu(nvme_req(req)->result.u32);
        if (meta && !ret && !write) {
                if (copy_to_user(meta_buffer, meta, meta_len))
                        ret = -EFAULT;
@@ -602,7 +594,7 @@ int nvme_get_features(struct nvme_ctrl *dev, unsigned fid, unsigned nsid,
                      void *buffer, size_t buflen, u32 *result)
 {
        struct nvme_command c;
-       struct nvme_completion cqe;
+       union nvme_result res;
        int ret;
 
        memset(&c, 0, sizeof(c));
@@ -610,10 +602,10 @@ int nvme_get_features(struct nvme_ctrl *dev, unsigned fid, unsigned nsid,
        c.features.nsid = cpu_to_le32(nsid);
        c.features.fid = cpu_to_le32(fid);
 
-       ret = __nvme_submit_sync_cmd(dev->admin_q, &c, &cqe, buffer, buflen, 0,
+       ret = __nvme_submit_sync_cmd(dev->admin_q, &c, &res, buffer, buflen, 0,
                        NVME_QID_ANY, 0, 0);
        if (ret >= 0 && result)
-               *result = le32_to_cpu(cqe.result);
+               *result = le32_to_cpu(res.u32);
        return ret;
 }
 
@@ -621,7 +613,7 @@ int nvme_set_features(struct nvme_ctrl *dev, unsigned fid, unsigned dword11,
                      void *buffer, size_t buflen, u32 *result)
 {
        struct nvme_command c;
-       struct nvme_completion cqe;
+       union nvme_result res;
        int ret;
 
        memset(&c, 0, sizeof(c));
@@ -629,10 +621,10 @@ int nvme_set_features(struct nvme_ctrl *dev, unsigned fid, unsigned dword11,
        c.features.fid = cpu_to_le32(fid);
        c.features.dword11 = cpu_to_le32(dword11);
 
-       ret = __nvme_submit_sync_cmd(dev->admin_q, &c, &cqe,
+       ret = __nvme_submit_sync_cmd(dev->admin_q, &c, &res,
                        buffer, buflen, 0, NVME_QID_ANY, 0, 0);
        if (ret >= 0 && result)
-               *result = le32_to_cpu(cqe.result);
+               *result = le32_to_cpu(res.u32);
        return ret;
 }
 
@@ -1851,9 +1843,6 @@ static void nvme_scan_work(struct work_struct *work)
        list_sort(NULL, &ctrl->namespaces, ns_cmp);
        mutex_unlock(&ctrl->namespaces_mutex);
        kfree(id);
-
-       if (ctrl->ops->post_scan)
-               ctrl->ops->post_scan(ctrl);
 }
 
 void nvme_queue_scan(struct nvme_ctrl *ctrl)
@@ -1910,7 +1899,7 @@ void nvme_complete_async_event(struct nvme_ctrl *ctrl,
                struct nvme_completion *cqe)
 {
        u16 status = le16_to_cpu(cqe->status) >> 1;
-       u32 result = le32_to_cpu(cqe->result);
+       u32 result = le32_to_cpu(cqe->result.u32);
 
        if (status == NVME_SC_SUCCESS || status == NVME_SC_ABORT_REQ) {
                ++ctrl->event_limit;
@@ -2079,14 +2068,8 @@ void nvme_stop_queues(struct nvme_ctrl *ctrl)
        struct nvme_ns *ns;
 
        mutex_lock(&ctrl->namespaces_mutex);
-       list_for_each_entry(ns, &ctrl->namespaces, list) {
-               spin_lock_irq(ns->queue->queue_lock);
-               queue_flag_set(QUEUE_FLAG_STOPPED, ns->queue);
-               spin_unlock_irq(ns->queue->queue_lock);
-
-               blk_mq_cancel_requeue_work(ns->queue);
-               blk_mq_stop_hw_queues(ns->queue);
-       }
+       list_for_each_entry(ns, &ctrl->namespaces, list)
+               blk_mq_quiesce_queue(ns->queue);
        mutex_unlock(&ctrl->namespaces_mutex);
 }
 EXPORT_SYMBOL_GPL(nvme_stop_queues);
@@ -2097,7 +2080,6 @@ void nvme_start_queues(struct nvme_ctrl *ctrl)
 
        mutex_lock(&ctrl->namespaces_mutex);
        list_for_each_entry(ns, &ctrl->namespaces, list) {
-               queue_flag_clear_unlocked(QUEUE_FLAG_STOPPED, ns->queue);
                blk_mq_start_stopped_hw_queues(ns->queue, true);
                blk_mq_kick_requeue_list(ns->queue);
        }