diff options
author | Pavel Begunkov <asml.silence@gmail.com> | 2024-10-18 18:14:00 +0200 |
---|---|---|
committer | Jens Axboe <axboe@kernel.dk> | 2024-10-29 20:43:27 +0100 |
commit | 2946f08ae9ed650b94e0ffebcdfdda8de76bd926 (patch) | |
tree | 464f90750fe641ba540963da88e050789c8689c0 /io_uring/io_uring.h | |
parent | io_uring: static_key for !IORING_SETUP_NO_SQARRAY (diff) | |
download | linux-2946f08ae9ed650b94e0ffebcdfdda8de76bd926.tar.xz linux-2946f08ae9ed650b94e0ffebcdfdda8de76bd926.zip |
io_uring: clean up cqe trace points
We have too many helpers posting CQEs, instead of tracing completion
events before filling in a CQE and thus having to pass all the data,
set the CQE first, pass it to the tracing helper and let it extract
everything it needs.
Signed-off-by: Pavel Begunkov <asml.silence@gmail.com>
Link: https://lore.kernel.org/r/b83c1ca9ee5aed2df0f3bb743bf5ed699cce4c86.1729267437.git.asml.silence@gmail.com
Signed-off-by: Jens Axboe <axboe@kernel.dk>
Diffstat (limited to '')
-rw-r--r-- | io_uring/io_uring.h | 7 |
1 files changed, 3 insertions, 4 deletions
diff --git a/io_uring/io_uring.h b/io_uring/io_uring.h index 70b6675941ff..9cd9a127e9ed 100644 --- a/io_uring/io_uring.h +++ b/io_uring/io_uring.h @@ -189,16 +189,15 @@ static __always_inline bool io_fill_cqe_req(struct io_ring_ctx *ctx, if (unlikely(!io_get_cqe(ctx, &cqe))) return false; - if (trace_io_uring_complete_enabled()) - trace_io_uring_complete(req->ctx, req, req->cqe.user_data, - req->cqe.res, req->cqe.flags, - req->big_cqe.extra1, req->big_cqe.extra2); memcpy(cqe, &req->cqe, sizeof(*cqe)); if (ctx->flags & IORING_SETUP_CQE32) { memcpy(cqe->big_cqe, &req->big_cqe, sizeof(*cqe)); memset(&req->big_cqe, 0, sizeof(req->big_cqe)); } + + if (trace_io_uring_complete_enabled()) + trace_io_uring_complete(req->ctx, req, cqe); return true; } |