diff --git a/block/blk-mq-tag.c b/block/blk-mq-tag.c index aacf10decdbd147c3d627472ac9e200c9ab7667f..9c92053e704dc105312a01e5f063a5643cc7fde7 100644 --- a/block/blk-mq-tag.c +++ b/block/blk-mq-tag.c @@ -416,9 +416,7 @@ void blk_mq_queue_tag_busy_iter(struct request_queue *q, busy_iter_fn *fn, /* * __blk_mq_update_nr_hw_queues() updates nr_hw_queues and queue_hw_ctx * while the queue is frozen. So we can use q_usage_counter to avoid - * racing with it. __blk_mq_update_nr_hw_queues() uses - * synchronize_rcu() to ensure this function left the critical section - * below. + * racing with it. */ if (!percpu_ref_tryget(&q->q_usage_counter)) return;