io_uring: remove sequence from io_kiocb
req->sequence is used only for deferred (i.e. DRAIN) requests, but initialised for every request. Remove req->sequence from io_kiocb together with its initialisation in io_init_req(). Replace it with a new field in struct io_defer_entry, that will be calculated only when needed in io_req_defer(), which is a slow path. Signed-off-by: Pavel Begunkov <asml.silence@gmail.com> Signed-off-by: Jens Axboe <axboe@kernel.dk>
This commit is contained in:
parent
27dc8338e5
commit
9cf7c104de
|
@ -639,6 +639,7 @@ struct io_kiocb {
|
||||||
u8 iopoll_completed;
|
u8 iopoll_completed;
|
||||||
|
|
||||||
u16 buf_index;
|
u16 buf_index;
|
||||||
|
u32 result;
|
||||||
|
|
||||||
struct io_ring_ctx *ctx;
|
struct io_ring_ctx *ctx;
|
||||||
unsigned int flags;
|
unsigned int flags;
|
||||||
|
@ -646,8 +647,6 @@ struct io_kiocb {
|
||||||
struct task_struct *task;
|
struct task_struct *task;
|
||||||
unsigned long fsize;
|
unsigned long fsize;
|
||||||
u64 user_data;
|
u64 user_data;
|
||||||
u32 result;
|
|
||||||
u32 sequence;
|
|
||||||
|
|
||||||
struct list_head link_list;
|
struct list_head link_list;
|
||||||
|
|
||||||
|
@ -678,6 +677,7 @@ struct io_kiocb {
|
||||||
struct io_defer_entry {
|
struct io_defer_entry {
|
||||||
struct list_head list;
|
struct list_head list;
|
||||||
struct io_kiocb *req;
|
struct io_kiocb *req;
|
||||||
|
u32 seq;
|
||||||
};
|
};
|
||||||
|
|
||||||
#define IO_IOPOLL_BATCH 8
|
#define IO_IOPOLL_BATCH 8
|
||||||
|
@ -1090,13 +1090,13 @@ err:
|
||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline bool req_need_defer(struct io_kiocb *req)
|
static bool req_need_defer(struct io_kiocb *req, u32 seq)
|
||||||
{
|
{
|
||||||
if (unlikely(req->flags & REQ_F_IO_DRAIN)) {
|
if (unlikely(req->flags & REQ_F_IO_DRAIN)) {
|
||||||
struct io_ring_ctx *ctx = req->ctx;
|
struct io_ring_ctx *ctx = req->ctx;
|
||||||
|
|
||||||
return req->sequence != ctx->cached_cq_tail
|
return seq != ctx->cached_cq_tail
|
||||||
+ atomic_read(&ctx->cached_cq_overflow);
|
+ atomic_read(&ctx->cached_cq_overflow);
|
||||||
}
|
}
|
||||||
|
|
||||||
return false;
|
return false;
|
||||||
|
@ -1241,7 +1241,7 @@ static void __io_queue_deferred(struct io_ring_ctx *ctx)
|
||||||
struct io_defer_entry *de = list_first_entry(&ctx->defer_list,
|
struct io_defer_entry *de = list_first_entry(&ctx->defer_list,
|
||||||
struct io_defer_entry, list);
|
struct io_defer_entry, list);
|
||||||
|
|
||||||
if (req_need_defer(de->req))
|
if (req_need_defer(de->req, de->seq))
|
||||||
break;
|
break;
|
||||||
list_del_init(&de->list);
|
list_del_init(&de->list);
|
||||||
/* punt-init is done before queueing for defer */
|
/* punt-init is done before queueing for defer */
|
||||||
|
@ -5396,14 +5396,35 @@ static int io_req_defer_prep(struct io_kiocb *req,
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static u32 io_get_sequence(struct io_kiocb *req)
|
||||||
|
{
|
||||||
|
struct io_kiocb *pos;
|
||||||
|
struct io_ring_ctx *ctx = req->ctx;
|
||||||
|
u32 total_submitted, nr_reqs = 1;
|
||||||
|
|
||||||
|
if (req->flags & REQ_F_LINK_HEAD)
|
||||||
|
list_for_each_entry(pos, &req->link_list, link_list)
|
||||||
|
nr_reqs++;
|
||||||
|
|
||||||
|
total_submitted = ctx->cached_sq_head - ctx->cached_sq_dropped;
|
||||||
|
return total_submitted - nr_reqs;
|
||||||
|
}
|
||||||
|
|
||||||
static int io_req_defer(struct io_kiocb *req, const struct io_uring_sqe *sqe)
|
static int io_req_defer(struct io_kiocb *req, const struct io_uring_sqe *sqe)
|
||||||
{
|
{
|
||||||
struct io_ring_ctx *ctx = req->ctx;
|
struct io_ring_ctx *ctx = req->ctx;
|
||||||
struct io_defer_entry *de;
|
struct io_defer_entry *de;
|
||||||
int ret;
|
int ret;
|
||||||
|
u32 seq;
|
||||||
|
|
||||||
/* Still need defer if there is pending req in defer list. */
|
/* Still need defer if there is pending req in defer list. */
|
||||||
if (!req_need_defer(req) && list_empty_careful(&ctx->defer_list))
|
if (likely(list_empty_careful(&ctx->defer_list) &&
|
||||||
|
!(req->flags & REQ_F_IO_DRAIN)))
|
||||||
|
return 0;
|
||||||
|
|
||||||
|
seq = io_get_sequence(req);
|
||||||
|
/* Still a chance to pass the sequence check */
|
||||||
|
if (!req_need_defer(req, seq) && list_empty_careful(&ctx->defer_list))
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
if (!req->io) {
|
if (!req->io) {
|
||||||
|
@ -5419,7 +5440,7 @@ static int io_req_defer(struct io_kiocb *req, const struct io_uring_sqe *sqe)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
spin_lock_irq(&ctx->completion_lock);
|
spin_lock_irq(&ctx->completion_lock);
|
||||||
if (!req_need_defer(req) && list_empty(&ctx->defer_list)) {
|
if (!req_need_defer(req, seq) && list_empty(&ctx->defer_list)) {
|
||||||
spin_unlock_irq(&ctx->completion_lock);
|
spin_unlock_irq(&ctx->completion_lock);
|
||||||
kfree(de);
|
kfree(de);
|
||||||
return 0;
|
return 0;
|
||||||
|
@ -5427,6 +5448,7 @@ static int io_req_defer(struct io_kiocb *req, const struct io_uring_sqe *sqe)
|
||||||
|
|
||||||
trace_io_uring_defer(ctx, req, req->user_data);
|
trace_io_uring_defer(ctx, req, req->user_data);
|
||||||
de->req = req;
|
de->req = req;
|
||||||
|
de->seq = seq;
|
||||||
list_add_tail(&de->list, &ctx->defer_list);
|
list_add_tail(&de->list, &ctx->defer_list);
|
||||||
spin_unlock_irq(&ctx->completion_lock);
|
spin_unlock_irq(&ctx->completion_lock);
|
||||||
return -EIOCBQUEUED;
|
return -EIOCBQUEUED;
|
||||||
|
@ -6204,12 +6226,6 @@ static int io_init_req(struct io_ring_ctx *ctx, struct io_kiocb *req,
|
||||||
unsigned int sqe_flags;
|
unsigned int sqe_flags;
|
||||||
int id;
|
int id;
|
||||||
|
|
||||||
/*
|
|
||||||
* All io need record the previous position, if LINK vs DARIN,
|
|
||||||
* it can be used to mark the position of the first IO in the
|
|
||||||
* link list.
|
|
||||||
*/
|
|
||||||
req->sequence = ctx->cached_sq_head - ctx->cached_sq_dropped;
|
|
||||||
req->opcode = READ_ONCE(sqe->opcode);
|
req->opcode = READ_ONCE(sqe->opcode);
|
||||||
req->user_data = READ_ONCE(sqe->user_data);
|
req->user_data = READ_ONCE(sqe->user_data);
|
||||||
req->io = NULL;
|
req->io = NULL;
|
||||||
|
|
Loading…
Reference in New Issue