summaryrefslogtreecommitdiffstats
path: root/block/blk-mq-tag.h
diff options
context:
space:
mode:
authorChristoph Hellwig <hch@lst.de>2014-05-27 20:59:48 +0200
committerJens Axboe <axboe@fb.com>2014-05-28 09:49:23 -0600
commita3bd77567cae6af700dcd245148befc73fc89a50 (patch)
tree465934c4f1fb0687da5d862e5221d54c0d34eb2f /block/blk-mq-tag.h
parent5dee857720db15e2c8ef0c03f7eeac00c4c63cb2 (diff)
downloadlinux-0-day-a3bd77567cae6af700dcd245148befc73fc89a50.tar.gz
linux-0-day-a3bd77567cae6af700dcd245148befc73fc89a50.tar.xz
blk-mq: remove blk_mq_wait_for_tags
The current logic for blocking tag allocation is rather confusing, as we first allocated and then free again a tag in blk_mq_wait_for_tags, just to attempt a non-blocking allocation and then repeat if someone else managed to grab the tag before us. Instead change blk_mq_alloc_request_pinned to simply do a blocking tag allocation itself and use the request we get back from it. Signed-off-by: Christoph Hellwig <hch@lst.de> Signed-off-by: Jens Axboe <axboe@fb.com>
Diffstat (limited to 'block/blk-mq-tag.h')
-rw-r--r--block/blk-mq-tag.h1
1 files changed, 0 insertions, 1 deletions
diff --git a/block/blk-mq-tag.h b/block/blk-mq-tag.h
index 2e5e6872d0892..c959de58d2a55 100644
--- a/block/blk-mq-tag.h
+++ b/block/blk-mq-tag.h
@@ -49,7 +49,6 @@ extern struct blk_mq_tags *blk_mq_init_tags(unsigned int nr_tags, unsigned int r
extern void blk_mq_free_tags(struct blk_mq_tags *tags);
extern unsigned int blk_mq_get_tag(struct blk_mq_hw_ctx *hctx, unsigned int *last_tag, gfp_t gfp, bool reserved);
-extern void blk_mq_wait_for_tags(struct blk_mq_hw_ctx *hctx, bool reserved);
extern void blk_mq_put_tag(struct blk_mq_hw_ctx *hctx, unsigned int tag, unsigned int *last_tag);
extern bool blk_mq_has_free_tags(struct blk_mq_tags *tags);
extern ssize_t blk_mq_tag_sysfs_show(struct blk_mq_tags *tags, char *page);