}
EXPORT_SYMBOL(__blk_mq_alloc_disk);
-/*
- * Helper for setting up a queue with mq ops, given queue depth, and
- * the passed in mq ops flags.
- */
-struct request_queue *blk_mq_init_sq_queue(struct blk_mq_tag_set *set,
- const struct blk_mq_ops *ops,
- unsigned int queue_depth,
- unsigned int set_flags)
-{
- struct request_queue *q;
- int ret;
-
- ret = blk_mq_alloc_sq_tag_set(set, ops, queue_depth, set_flags);
- if (ret)
- return ERR_PTR(ret);
- q = blk_mq_init_queue(set);
- if (IS_ERR(q))
- blk_mq_free_tag_set(set);
- return q;
-}
-EXPORT_SYMBOL(blk_mq_init_sq_queue);
-
static struct blk_mq_hw_ctx *blk_mq_alloc_and_init_hctx(
struct blk_mq_tag_set *set, struct request_queue *q,
int hctx_idx, int node)
void *queuedata);
int blk_mq_init_allocated_queue(struct blk_mq_tag_set *set,
struct request_queue *q);
-struct request_queue *blk_mq_init_sq_queue(struct blk_mq_tag_set *set,
- const struct blk_mq_ops *ops,
- unsigned int queue_depth,
- unsigned int set_flags);
void blk_mq_unregister_dev(struct device *, struct request_queue *);
int blk_mq_alloc_tag_set(struct blk_mq_tag_set *set);