iowq             2941 fs/io_uring.c  static inline bool io_should_wake(struct io_wait_queue *iowq)
iowq             2943 fs/io_uring.c  	struct io_ring_ctx *ctx = iowq->ctx;
iowq             2950 fs/io_uring.c  	return io_cqring_events(ctx->rings) >= iowq->to_wait ||
iowq             2951 fs/io_uring.c  			atomic_read(&ctx->cq_timeouts) != iowq->nr_timeouts;
iowq             2957 fs/io_uring.c  	struct io_wait_queue *iowq = container_of(curr, struct io_wait_queue,
iowq             2960 fs/io_uring.c  	if (!io_should_wake(iowq))
iowq             2973 fs/io_uring.c  	struct io_wait_queue iowq = {
iowq             2977 fs/io_uring.c  			.entry		= LIST_HEAD_INIT(iowq.wq.entry),
iowq             3002 fs/io_uring.c  	iowq.nr_timeouts = atomic_read(&ctx->cq_timeouts);
iowq             3004 fs/io_uring.c  		prepare_to_wait_exclusive(&ctx->wait, &iowq.wq,
iowq             3006 fs/io_uring.c  		if (io_should_wake(&iowq))
iowq             3014 fs/io_uring.c  	finish_wait(&ctx->wait, &iowq.wq);