block: Protect queue flag changes with the queue lock
Since the queue flags may be changed concurrently from multiple contexts after a queue becomes visible in sysfs, make these changes safe by protecting these with the queue lock. Cc: Christoph Hellwig <hch@lst.de> Cc: Hannes Reinecke <hare@suse.de> Cc: Ming Lei <ming.lei@redhat.com> Reviewed-by: Martin K. Petersen <martin.petersen@oracle.com> Reviewed-by: Johannes Thumshirn <jthumshirn@suse.de> Signed-off-by: Bart Van Assche <bart.vanassche@wdc.com> Signed-off-by: Jens Axboe <axboe@kernel.dk>hifive-unleashed-5.1
parent
8814ce8a0f
commit
7dfdbc7367
|
@ -2997,7 +2997,7 @@ EXPORT_SYMBOL_GPL(blk_mq_update_nr_hw_queues);
|
||||||
static bool blk_poll_stats_enable(struct request_queue *q)
|
static bool blk_poll_stats_enable(struct request_queue *q)
|
||||||
{
|
{
|
||||||
if (test_bit(QUEUE_FLAG_POLL_STATS, &q->queue_flags) ||
|
if (test_bit(QUEUE_FLAG_POLL_STATS, &q->queue_flags) ||
|
||||||
test_and_set_bit(QUEUE_FLAG_POLL_STATS, &q->queue_flags))
|
blk_queue_flag_test_and_set(QUEUE_FLAG_POLL_STATS, q))
|
||||||
return true;
|
return true;
|
||||||
blk_stat_add_callback(q, q->poll_cb);
|
blk_stat_add_callback(q, q->poll_cb);
|
||||||
return false;
|
return false;
|
||||||
|
|
|
@ -152,7 +152,7 @@ void blk_stat_add_callback(struct request_queue *q,
|
||||||
|
|
||||||
spin_lock(&q->stats->lock);
|
spin_lock(&q->stats->lock);
|
||||||
list_add_tail_rcu(&cb->list, &q->stats->callbacks);
|
list_add_tail_rcu(&cb->list, &q->stats->callbacks);
|
||||||
queue_flag_set(QUEUE_FLAG_STATS, q);
|
blk_queue_flag_set(QUEUE_FLAG_STATS, q);
|
||||||
spin_unlock(&q->stats->lock);
|
spin_unlock(&q->stats->lock);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(blk_stat_add_callback);
|
EXPORT_SYMBOL_GPL(blk_stat_add_callback);
|
||||||
|
@ -163,7 +163,7 @@ void blk_stat_remove_callback(struct request_queue *q,
|
||||||
spin_lock(&q->stats->lock);
|
spin_lock(&q->stats->lock);
|
||||||
list_del_rcu(&cb->list);
|
list_del_rcu(&cb->list);
|
||||||
if (list_empty(&q->stats->callbacks) && !q->stats->enable_accounting)
|
if (list_empty(&q->stats->callbacks) && !q->stats->enable_accounting)
|
||||||
queue_flag_clear(QUEUE_FLAG_STATS, q);
|
blk_queue_flag_clear(QUEUE_FLAG_STATS, q);
|
||||||
spin_unlock(&q->stats->lock);
|
spin_unlock(&q->stats->lock);
|
||||||
|
|
||||||
del_timer_sync(&cb->timer);
|
del_timer_sync(&cb->timer);
|
||||||
|
@ -191,7 +191,7 @@ void blk_stat_enable_accounting(struct request_queue *q)
|
||||||
{
|
{
|
||||||
spin_lock(&q->stats->lock);
|
spin_lock(&q->stats->lock);
|
||||||
q->stats->enable_accounting = true;
|
q->stats->enable_accounting = true;
|
||||||
queue_flag_set(QUEUE_FLAG_STATS, q);
|
blk_queue_flag_set(QUEUE_FLAG_STATS, q);
|
||||||
spin_unlock(&q->stats->lock);
|
spin_unlock(&q->stats->lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue