iowq 2941 fs/io_uring.c static inline bool io_should_wake(struct io_wait_queue *iowq) iowq 2943 fs/io_uring.c struct io_ring_ctx *ctx = iowq->ctx; iowq 2950 fs/io_uring.c return io_cqring_events(ctx->rings) >= iowq->to_wait || iowq 2951 fs/io_uring.c atomic_read(&ctx->cq_timeouts) != iowq->nr_timeouts; iowq 2957 fs/io_uring.c struct io_wait_queue *iowq = container_of(curr, struct io_wait_queue, iowq 2960 fs/io_uring.c if (!io_should_wake(iowq)) iowq 2973 fs/io_uring.c struct io_wait_queue iowq = { iowq 2977 fs/io_uring.c .entry = LIST_HEAD_INIT(iowq.wq.entry), iowq 3002 fs/io_uring.c iowq.nr_timeouts = atomic_read(&ctx->cq_timeouts); iowq 3004 fs/io_uring.c prepare_to_wait_exclusive(&ctx->wait, &iowq.wq, iowq 3006 fs/io_uring.c if (io_should_wake(&iowq)) iowq 3014 fs/io_uring.c finish_wait(&ctx->wait, &iowq.wq);