io_uring: IRQ rw completion batching

Employ inline completion logic for read/write completions done via
io_req_task_complete(). If ->uring_lock is contended, just do normal
request completion, but if not, make tctx_task_work() to grab the lock
and do batched inline completions in io_req_task_complete().

Signed-off-by: Pavel Begunkov <asml.silence@gmail.com>
Link: https://lore.kernel.org/r/94589c3ce69eaed86a21bb1ec696407a54fab1aa.1629286357.git.asml.silence@gmail.com
Signed-off-by: Jens Axboe <axboe@kernel.dk>
This commit is contained in:
Pavel Begunkov 2021-08-18 12:42:47 +01:00 committed by Jens Axboe
parent f237c30a56
commit 126180b95f
1 changed files with 16 additions and 1 deletions

View File

@ -2067,6 +2067,8 @@ static void tctx_task_work(struct callback_head *cb)
if (req->ctx != ctx) { if (req->ctx != ctx) {
ctx_flush_and_put(ctx, &locked); ctx_flush_and_put(ctx, &locked);
ctx = req->ctx; ctx = req->ctx;
/* if not contended, grab and improve batching */
locked = mutex_trylock(&ctx->uring_lock);
percpu_ref_get(&ctx->refs); percpu_ref_get(&ctx->refs);
} }
req->io_task_work.func(req, &locked); req->io_task_work.func(req, &locked);
@ -2582,7 +2584,20 @@ static bool __io_complete_rw_common(struct io_kiocb *req, long res)
static void io_req_task_complete(struct io_kiocb *req, bool *locked) static void io_req_task_complete(struct io_kiocb *req, bool *locked)
{ {
__io_req_complete(req, 0, req->result, io_put_rw_kbuf(req)); unsigned int cflags = io_put_rw_kbuf(req);
long res = req->result;
if (*locked) {
struct io_ring_ctx *ctx = req->ctx;
struct io_submit_state *state = &ctx->submit_state;
io_req_complete_state(req, res, cflags);
state->compl_reqs[state->compl_nr++] = req;
if (state->compl_nr == ARRAY_SIZE(state->compl_reqs))
io_submit_flush_completions(ctx);
} else {
io_req_complete_post(req, res, cflags);
}
} }
static void __io_complete_rw(struct io_kiocb *req, long res, long res2, static void __io_complete_rw(struct io_kiocb *req, long res, long res2,