summaryrefslogtreecommitdiff
path: root/block
diff options
context:
space:
mode:
authorJan Kara <jack@suse.cz>2022-04-01 12:27:46 +0200
committerGreg Kroah-Hartman <gregkh@linuxfoundation.org>2022-06-09 10:23:19 +0200
commita107df383c163fdf9fdf748bc8f1a5e8f00d29c1 (patch)
tree4458020fff12ac3ea9533a3423920ccb1878cf2d /block
parente8821f45612f2e6d9adb9c6ba0fb4184f57692aa (diff)
bfq: Drop pointless unlock-lock pair
commit fc84e1f941b91221092da5b3102ec82da24c5673 upstream. In bfq_insert_request() we unlock bfqd->lock only to call trace_block_rq_insert() and then lock bfqd->lock again. This is really pointless since tracing is disabled if we really care about performance and even if the tracepoint is enabled, it is a quick call. CC: stable@vger.kernel.org Tested-by: "yukuai (C)" <yukuai3@huawei.com> Signed-off-by: Jan Kara <jack@suse.cz> Reviewed-by: Christoph Hellwig <hch@lst.de> Link: https://lore.kernel.org/r/20220401102752.8599-5-jack@suse.cz Signed-off-by: Jens Axboe <axboe@kernel.dk> Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Diffstat (limited to 'block')
-rw-r--r--block/bfq-iosched.c3
1 files changed, 0 insertions, 3 deletions
diff --git a/block/bfq-iosched.c b/block/bfq-iosched.c
index e1183596de06..4ecfcb6e9a2d 100644
--- a/block/bfq-iosched.c
+++ b/block/bfq-iosched.c
@@ -6012,11 +6012,8 @@ static void bfq_insert_request(struct blk_mq_hw_ctx *hctx, struct request *rq,
return;
}
- spin_unlock_irq(&bfqd->lock);
-
trace_block_rq_insert(rq);
- spin_lock_irq(&bfqd->lock);
bfqq = bfq_init_rq(rq);
if (!bfqq || at_head) {
if (at_head)