[RFC v2 5/5] block: revert back to synchronous request_queue removal

From: Luis Chamberlain
Date: Thu Apr 09 2020 - 17:45:52 EST


Commit dc9edc44de6c ("block: Fix a blk_exit_rl() regression") merged on
v4.12 moved the work behind blk_release_queue() into a workqueue after a
splat floated around which indicated some work on blk_release_queue()
could sleep in blk_exit_rl(). This splat would be possible when a driver
called blk_put_queue() or blk_cleanup_queue() (which calls blk_put_queue()
as its final call) from an atomic context.

blk_put_queue() puts decrements the refcount for the request_queue
kobject, and upon reaching 0 blk_release_queue() is called. Although
blk_exit_rl() is now removed through commit db6d9952356 ("block: remove
request_list code"), we reserve the right to be able to sleep within
blk_release_queue() context. There should be little reason to
defer removal from atomic context these days, as you can always just
increase your block device's reference count even in atomic context and
leave the removal for the request_queue to the upper layers later.
However if you really need to defer removal of the request_queue, you can
set the queue flag QUEUE_FLAG_DEFER_REMOVAL now.

Cc: Bart Van Assche <bvanassche@xxxxxxx>
Cc: Omar Sandoval <osandov@xxxxxx>
Cc: Hannes Reinecke <hare@xxxxxxxx>
Cc: Nicolai Stange <nstange@xxxxxxx>
Cc: Greg Kroah-Hartman <gregkh@xxxxxxxxxxxxxxxxxxx>
Cc: Michal Hocko <mhocko@xxxxxxxxxx>
Cc: yu kuai <yukuai3@xxxxxxxxxx>
Signed-off-by: Luis Chamberlain <mcgrof@xxxxxxxxxx>
---
block/blk-sysfs.c | 40 ++++++++++++++++++++++++++++++++--------
include/linux/blkdev.h | 3 +++
2 files changed, 35 insertions(+), 8 deletions(-)

diff --git a/block/blk-sysfs.c b/block/blk-sysfs.c
index 20f20b0fa0b9..2ae8c39c88ef 100644
--- a/block/blk-sysfs.c
+++ b/block/blk-sysfs.c
@@ -860,10 +860,9 @@ static void blk_exit_queue(struct request_queue *q)
bdi_put(q->backing_dev_info);
}

-
/**
- * __blk_release_queue - release a request queue
- * @work: pointer to the release_work member of the request queue to be released
+ * blk_release_queue_sync- release a request queue
+ * @q: pointer to the request queue to be released
*
* Description:
* This function is called when a block device is being unregistered. The
@@ -872,11 +871,27 @@ static void blk_exit_queue(struct request_queue *q)
* the reference counter of the request queue. Once the reference counter
* of the request queue reaches zero, blk_release_queue is called to release
* all allocated resources of the request queue.
+ *
+ * There are two approaches to releasing the request queue, by default
+ * we reserve the right to sleep on release and so release is synchronous.
+ * If you know the path under which blk_cleanup_queue() or your last
+ * blk_put_queue() is called can be called in atomic context you want to
+ * ensure to defer the removal by setting the QUEUE_FLAG_DEFER_REMOVAL
+ * flag as follows upon initialization:
+ *
+ * blk_queue_flag_set(QUEUE_FLAG_DEFER_REMOVAL, q)
+ *
+ * Note that deferring removal may have implications for userspace. An
+ * example is if you are using an ioctl to allow removal of a block device,
+ * and the kernel returns immediately even though the device may only
+ * disappear after the full removal is completed.
+ *
+ * You should also be able to work around this by just increasing the
+ * refcount for the block device instead during your atomic operation,
+ * and so QUEUE_FLAG_DEFER_REMOVAL should almost never be required.
*/
-static void __blk_release_queue(struct work_struct *work)
+static void blk_release_queue_sync(struct request_queue *q)
{
- struct request_queue *q = container_of(work, typeof(*q), release_work);
-
if (test_bit(QUEUE_FLAG_POLL_STATS, &q->queue_flags))
blk_stat_remove_callback(q, q->poll_cb);
blk_stat_free_callback(q->poll_cb);
@@ -905,13 +920,22 @@ static void __blk_release_queue(struct work_struct *work)
call_rcu(&q->rcu_head, blk_free_queue_rcu);
}

+void __blk_release_queue(struct work_struct *work)
+{
+ struct request_queue *q = container_of(work, typeof(*q), release_work);
+
+ blk_release_queue_sync(q);
+}
+
static void blk_release_queue(struct kobject *kobj)
{
struct request_queue *q =
container_of(kobj, struct request_queue, kobj);

- INIT_WORK(&q->release_work, __blk_release_queue);
- schedule_work(&q->release_work);
+ if (blk_queue_defer_removal(q))
+ schedule_work(&q->release_work);
+ else
+ blk_release_queue_sync(q);
}

static const struct sysfs_ops queue_sysfs_ops = {
diff --git a/include/linux/blkdev.h b/include/linux/blkdev.h
index 8b1cab52cef9..46fee1ef92e3 100644
--- a/include/linux/blkdev.h
+++ b/include/linux/blkdev.h
@@ -614,6 +614,7 @@ struct request_queue {
#define QUEUE_FLAG_PCI_P2PDMA 25 /* device supports PCI p2p requests */
#define QUEUE_FLAG_ZONE_RESETALL 26 /* supports Zone Reset All */
#define QUEUE_FLAG_RQ_ALLOC_TIME 27 /* record rq->alloc_time_ns */
+#define QUEUE_FLAG_DEFER_REMOVAL 28 /* defer queue removal */

#define QUEUE_FLAG_MQ_DEFAULT ((1 << QUEUE_FLAG_IO_STAT) | \
(1 << QUEUE_FLAG_SAME_COMP))
@@ -648,6 +649,8 @@ bool blk_queue_flag_test_and_set(unsigned int flag, struct request_queue *q);
#else
#define blk_queue_rq_alloc_time(q) false
#endif
+#define blk_queue_defer_removal(q) \
+ test_bit(QUEUE_FLAG_DEFER_REMOVAL, &(q)->queue_flags)

#define blk_noretry_request(rq) \
((rq)->cmd_flags & (REQ_FAILFAST_DEV|REQ_FAILFAST_TRANSPORT| \
--
2.25.1