block: pass a queue_limits argument to blk_alloc_queue
Pass a queue_limits to blk_alloc_queue and apply it after validating and capping the values using blk_validate_limits. This will allow allocating queues with valid queue limits instead of setting the values one at a time later. Signed-off-by: Christoph Hellwig <hch@lst.de> Reviewed-by: Damien Le Moal <dlemoal@kernel.org> Reviewed-by: Hannes Reinecke <hare@suse.de> Link: https://lore.kernel.org/r/20240213073425.1621680-9-hch@lst.de Signed-off-by: Jens Axboe <axboe@kernel.dk>
This commit is contained in:
parent
ff956a3be9
commit
ad751ba1f8
|
@ -394,24 +394,34 @@ static void blk_timeout_work(struct work_struct *work)
|
||||||
{
|
{
|
||||||
}
|
}
|
||||||
|
|
||||||
struct request_queue *blk_alloc_queue(int node_id)
|
struct request_queue *blk_alloc_queue(struct queue_limits *lim, int node_id)
|
||||||
{
|
{
|
||||||
struct request_queue *q;
|
struct request_queue *q;
|
||||||
|
int error;
|
||||||
|
|
||||||
q = kmem_cache_alloc_node(blk_requestq_cachep, GFP_KERNEL | __GFP_ZERO,
|
q = kmem_cache_alloc_node(blk_requestq_cachep, GFP_KERNEL | __GFP_ZERO,
|
||||||
node_id);
|
node_id);
|
||||||
if (!q)
|
if (!q)
|
||||||
return NULL;
|
return ERR_PTR(-ENOMEM);
|
||||||
|
|
||||||
q->last_merge = NULL;
|
q->last_merge = NULL;
|
||||||
|
|
||||||
q->id = ida_alloc(&blk_queue_ida, GFP_KERNEL);
|
q->id = ida_alloc(&blk_queue_ida, GFP_KERNEL);
|
||||||
if (q->id < 0)
|
if (q->id < 0) {
|
||||||
|
error = q->id;
|
||||||
goto fail_q;
|
goto fail_q;
|
||||||
|
}
|
||||||
|
|
||||||
q->stats = blk_alloc_queue_stats();
|
q->stats = blk_alloc_queue_stats();
|
||||||
if (!q->stats)
|
if (!q->stats) {
|
||||||
|
error = -ENOMEM;
|
||||||
goto fail_id;
|
goto fail_id;
|
||||||
|
}
|
||||||
|
|
||||||
|
error = blk_set_default_limits(lim);
|
||||||
|
if (error)
|
||||||
|
goto fail_stats;
|
||||||
|
q->limits = *lim;
|
||||||
|
|
||||||
q->node = node_id;
|
q->node = node_id;
|
||||||
|
|
||||||
|
@ -436,12 +446,12 @@ struct request_queue *blk_alloc_queue(int node_id)
|
||||||
* Init percpu_ref in atomic mode so that it's faster to shutdown.
|
* Init percpu_ref in atomic mode so that it's faster to shutdown.
|
||||||
* See blk_register_queue() for details.
|
* See blk_register_queue() for details.
|
||||||
*/
|
*/
|
||||||
if (percpu_ref_init(&q->q_usage_counter,
|
error = percpu_ref_init(&q->q_usage_counter,
|
||||||
blk_queue_usage_counter_release,
|
blk_queue_usage_counter_release,
|
||||||
PERCPU_REF_INIT_ATOMIC, GFP_KERNEL))
|
PERCPU_REF_INIT_ATOMIC, GFP_KERNEL);
|
||||||
|
if (error)
|
||||||
goto fail_stats;
|
goto fail_stats;
|
||||||
|
|
||||||
blk_set_default_limits(&q->limits);
|
|
||||||
q->nr_requests = BLKDEV_DEFAULT_RQ;
|
q->nr_requests = BLKDEV_DEFAULT_RQ;
|
||||||
|
|
||||||
return q;
|
return q;
|
||||||
|
@ -452,7 +462,7 @@ fail_id:
|
||||||
ida_free(&blk_queue_ida, q->id);
|
ida_free(&blk_queue_ida, q->id);
|
||||||
fail_q:
|
fail_q:
|
||||||
kmem_cache_free(blk_requestq_cachep, q);
|
kmem_cache_free(blk_requestq_cachep, q);
|
||||||
return NULL;
|
return ERR_PTR(error);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
|
@ -4086,12 +4086,13 @@ void blk_mq_release(struct request_queue *q)
|
||||||
static struct request_queue *blk_mq_init_queue_data(struct blk_mq_tag_set *set,
|
static struct request_queue *blk_mq_init_queue_data(struct blk_mq_tag_set *set,
|
||||||
void *queuedata)
|
void *queuedata)
|
||||||
{
|
{
|
||||||
|
struct queue_limits lim = { };
|
||||||
struct request_queue *q;
|
struct request_queue *q;
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
q = blk_alloc_queue(set->numa_node);
|
q = blk_alloc_queue(&lim, set->numa_node);
|
||||||
if (!q)
|
if (IS_ERR(q))
|
||||||
return ERR_PTR(-ENOMEM);
|
return q;
|
||||||
q->queuedata = queuedata;
|
q->queuedata = queuedata;
|
||||||
ret = blk_mq_init_allocated_queue(set, q);
|
ret = blk_mq_init_allocated_queue(set, q);
|
||||||
if (ret) {
|
if (ret) {
|
||||||
|
|
|
@ -448,7 +448,7 @@ static inline void bio_release_page(struct bio *bio, struct page *page)
|
||||||
unpin_user_page(page);
|
unpin_user_page(page);
|
||||||
}
|
}
|
||||||
|
|
||||||
struct request_queue *blk_alloc_queue(int node_id);
|
struct request_queue *blk_alloc_queue(struct queue_limits *lim, int node_id);
|
||||||
|
|
||||||
int disk_scan_partitions(struct gendisk *disk, blk_mode_t mode);
|
int disk_scan_partitions(struct gendisk *disk, blk_mode_t mode);
|
||||||
|
|
||||||
|
|
|
@ -1393,11 +1393,12 @@ out_free_disk:
|
||||||
|
|
||||||
struct gendisk *__blk_alloc_disk(int node, struct lock_class_key *lkclass)
|
struct gendisk *__blk_alloc_disk(int node, struct lock_class_key *lkclass)
|
||||||
{
|
{
|
||||||
|
struct queue_limits lim = { };
|
||||||
struct request_queue *q;
|
struct request_queue *q;
|
||||||
struct gendisk *disk;
|
struct gendisk *disk;
|
||||||
|
|
||||||
q = blk_alloc_queue(node);
|
q = blk_alloc_queue(&lim, node);
|
||||||
if (!q)
|
if (IS_ERR(q))
|
||||||
return NULL;
|
return NULL;
|
||||||
|
|
||||||
disk = __alloc_disk_node(q, node, lkclass);
|
disk = __alloc_disk_node(q, node, lkclass);
|
||||||
|
|
Loading…
Reference in New Issue