diff options
author | Jeff Moyer <jmoyer@redhat.com> | 2015-10-20 23:13:51 +0800 |
---|---|---|
committer | Jens Axboe <axboe@fb.com> | 2015-10-21 15:00:48 -0600 |
commit | 0809e3ac62319dc7534b64f95ac37e230d740e8a (patch) | |
tree | 70814dd32f930745834e4186445acf8dcbd1e3c1 /block/blk-core.c | |
parent | 3380f4589f6d9725e275525fd0580c8ee2b5cbbc (diff) | |
download | linux-0809e3ac62319dc7534b64f95ac37e230d740e8a.tar.bz2 |
block: fix plug list flushing for nomerge queues
Request queues with merging disabled will not flush the plug list after
BLK_MAX_REQUEST_COUNT requests have been queued, since the code relies
on blk_attempt_plug_merge to compute the request_count. Fix this by
computing the number of queued requests even for nomerge queues.
Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
Signed-off-by: Jens Axboe <axboe@fb.com>
Diffstat (limited to 'block/blk-core.c')
-rw-r--r-- | block/blk-core.c | 32 |
1 files changed, 29 insertions, 3 deletions
diff --git a/block/blk-core.c b/block/blk-core.c index 2eb722d48773..f0ae087ead06 100644 --- a/block/blk-core.c +++ b/block/blk-core.c @@ -1594,6 +1594,30 @@ out: return ret; } +unsigned int blk_plug_queued_count(struct request_queue *q) +{ + struct blk_plug *plug; + struct request *rq; + struct list_head *plug_list; + unsigned int ret = 0; + + plug = current->plug; + if (!plug) + goto out; + + if (q->mq_ops) + plug_list = &plug->mq_list; + else + plug_list = &plug->list; + + list_for_each_entry(rq, plug_list, queuelist) { + if (rq->q == q) + ret++; + } +out: + return ret; +} + void init_request_from_bio(struct request *req, struct bio *bio) { req->cmd_type = REQ_TYPE_FS; @@ -1641,9 +1665,11 @@ static void blk_queue_bio(struct request_queue *q, struct bio *bio) * Check if we can merge with the plugged list before grabbing * any locks. */ - if (!blk_queue_nomerges(q) && - blk_attempt_plug_merge(q, bio, &request_count, NULL)) - return; + if (!blk_queue_nomerges(q)) { + if (blk_attempt_plug_merge(q, bio, &request_count, NULL)) + return; + } else + request_count = blk_plug_queued_count(q); spin_lock_irq(q->queue_lock); |