summaryrefslogtreecommitdiffstats
path: root/block/blk-mq.c
diff options
context:
space:
mode:
authorChristoph Hellwig <hch@lst.de>2014-05-27 20:59:48 +0200
committerJens Axboe <axboe@fb.com>2014-05-28 09:49:23 -0600
commita3bd77567cae6af700dcd245148befc73fc89a50 (patch)
tree465934c4f1fb0687da5d862e5221d54c0d34eb2f /block/blk-mq.c
parent5dee857720db15e2c8ef0c03f7eeac00c4c63cb2 (diff)
downloadlinux-a3bd77567cae6af700dcd245148befc73fc89a50.tar.bz2
blk-mq: remove blk_mq_wait_for_tags
The current logic for blocking tag allocation is rather confusing, as we first allocated and then free again a tag in blk_mq_wait_for_tags, just to attempt a non-blocking allocation and then repeat if someone else managed to grab the tag before us. Instead change blk_mq_alloc_request_pinned to simply do a blocking tag allocation itself and use the request we get back from it. Signed-off-by: Christoph Hellwig <hch@lst.de> Signed-off-by: Jens Axboe <axboe@fb.com>
Diffstat (limited to 'block/blk-mq.c')
-rw-r--r--block/blk-mq.c13
1 files changed, 6 insertions, 7 deletions
diff --git a/block/blk-mq.c b/block/blk-mq.c
index 04ef7ecb3c7f..3224888d329a 100644
--- a/block/blk-mq.c
+++ b/block/blk-mq.c
@@ -264,31 +264,30 @@ __blk_mq_alloc_request(struct request_queue *q, struct blk_mq_hw_ctx *hctx,
return NULL;
}
-
static struct request *blk_mq_alloc_request_pinned(struct request_queue *q,
int rw, gfp_t gfp,
bool reserved)
{
+ bool gfp_mask = gfp & ~__GFP_WAIT;
struct request *rq;
do {
struct blk_mq_ctx *ctx = blk_mq_get_ctx(q);
struct blk_mq_hw_ctx *hctx = q->mq_ops->map_queue(q, ctx->cpu);
- rq = __blk_mq_alloc_request(q, hctx, ctx, rw, gfp & ~__GFP_WAIT,
+ rq = __blk_mq_alloc_request(q, hctx, ctx, rw, gfp_mask,
reserved);
if (rq)
break;
- if (gfp & __GFP_WAIT) {
- __blk_mq_run_hw_queue(hctx);
- blk_mq_put_ctx(ctx);
- } else {
+ if (!(gfp & __GFP_WAIT)) {
blk_mq_put_ctx(ctx);
break;
}
- blk_mq_wait_for_tags(hctx, reserved);
+ __blk_mq_run_hw_queue(hctx);
+ blk_mq_put_ctx(ctx);
+ gfp_mask = gfp;
} while (1);
return rq;