mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-12-27 12:21:22 -05:00
selftests: ublk: pass 'ublk_thread *' to ->queue_io() and ->tgt_io_done()
'struct thread' is task local structure, and the related code will become more readable if we pass it via parameter. Meantime pass 'ublk_thread *' to ublk_io_alloc_sqes(), and this way is natural since we use per-thread io_uring for handling IO. More importantly it helps much for removing the current ubq_daemon or per-io-task limit. Signed-off-by: Ming Lei <ming.lei@redhat.com> Link: https://lore.kernel.org/r/20250713143415.2857561-13-ming.lei@redhat.com Signed-off-by: Jens Axboe <axboe@kernel.dk>
This commit is contained in:
@@ -38,7 +38,8 @@ static int ublk_fault_inject_tgt_init(const struct dev_ctx *ctx,
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int ublk_fault_inject_queue_io(struct ublk_queue *q, int tag)
|
||||
static int ublk_fault_inject_queue_io(struct ublk_thread *t,
|
||||
struct ublk_queue *q, int tag)
|
||||
{
|
||||
const struct ublksrv_io_desc *iod = ublk_get_iod(q, tag);
|
||||
struct io_uring_sqe *sqe;
|
||||
@@ -46,7 +47,7 @@ static int ublk_fault_inject_queue_io(struct ublk_queue *q, int tag)
|
||||
.tv_nsec = (long long)q->dev->private_data,
|
||||
};
|
||||
|
||||
ublk_io_alloc_sqes(ublk_get_io(q, tag), &sqe, 1);
|
||||
ublk_io_alloc_sqes(t, &sqe, 1);
|
||||
io_uring_prep_timeout(sqe, &ts, 1, 0);
|
||||
sqe->user_data = build_user_data(tag, ublksrv_get_op(iod), 0, q->q_id, 1);
|
||||
|
||||
@@ -55,7 +56,8 @@ static int ublk_fault_inject_queue_io(struct ublk_queue *q, int tag)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void ublk_fault_inject_tgt_io_done(struct ublk_queue *q,
|
||||
static void ublk_fault_inject_tgt_io_done(struct ublk_thread *t,
|
||||
struct ublk_queue *q,
|
||||
const struct io_uring_cqe *cqe)
|
||||
{
|
||||
unsigned tag = user_data_to_tag(cqe->user_data);
|
||||
|
||||
@@ -13,12 +13,13 @@ static enum io_uring_op ublk_to_uring_op(const struct ublksrv_io_desc *iod, int
|
||||
assert(0);
|
||||
}
|
||||
|
||||
static int loop_queue_flush_io(struct ublk_queue *q, const struct ublksrv_io_desc *iod, int tag)
|
||||
static int loop_queue_flush_io(struct ublk_thread *t, struct ublk_queue *q,
|
||||
const struct ublksrv_io_desc *iod, int tag)
|
||||
{
|
||||
unsigned ublk_op = ublksrv_get_op(iod);
|
||||
struct io_uring_sqe *sqe[1];
|
||||
|
||||
ublk_io_alloc_sqes(ublk_get_io(q, tag), sqe, 1);
|
||||
ublk_io_alloc_sqes(t, sqe, 1);
|
||||
io_uring_prep_fsync(sqe[0], 1 /*fds[1]*/, IORING_FSYNC_DATASYNC);
|
||||
io_uring_sqe_set_flags(sqe[0], IOSQE_FIXED_FILE);
|
||||
/* bit63 marks us as tgt io */
|
||||
@@ -26,7 +27,8 @@ static int loop_queue_flush_io(struct ublk_queue *q, const struct ublksrv_io_des
|
||||
return 1;
|
||||
}
|
||||
|
||||
static int loop_queue_tgt_rw_io(struct ublk_queue *q, const struct ublksrv_io_desc *iod, int tag)
|
||||
static int loop_queue_tgt_rw_io(struct ublk_thread *t, struct ublk_queue *q,
|
||||
const struct ublksrv_io_desc *iod, int tag)
|
||||
{
|
||||
unsigned ublk_op = ublksrv_get_op(iod);
|
||||
unsigned zc = ublk_queue_use_zc(q);
|
||||
@@ -36,7 +38,7 @@ static int loop_queue_tgt_rw_io(struct ublk_queue *q, const struct ublksrv_io_de
|
||||
void *addr = (zc | auto_zc) ? NULL : (void *)iod->addr;
|
||||
|
||||
if (!zc || auto_zc) {
|
||||
ublk_io_alloc_sqes(ublk_get_io(q, tag), sqe, 1);
|
||||
ublk_io_alloc_sqes(t, sqe, 1);
|
||||
if (!sqe[0])
|
||||
return -ENOMEM;
|
||||
|
||||
@@ -52,7 +54,7 @@ static int loop_queue_tgt_rw_io(struct ublk_queue *q, const struct ublksrv_io_de
|
||||
return 1;
|
||||
}
|
||||
|
||||
ublk_io_alloc_sqes(ublk_get_io(q, tag), sqe, 3);
|
||||
ublk_io_alloc_sqes(t, sqe, 3);
|
||||
|
||||
io_uring_prep_buf_register(sqe[0], 0, tag, q->q_id, ublk_get_io(q, tag)->buf_index);
|
||||
sqe[0]->flags |= IOSQE_CQE_SKIP_SUCCESS | IOSQE_IO_HARDLINK;
|
||||
@@ -72,7 +74,7 @@ static int loop_queue_tgt_rw_io(struct ublk_queue *q, const struct ublksrv_io_de
|
||||
return 2;
|
||||
}
|
||||
|
||||
static int loop_queue_tgt_io(struct ublk_queue *q, int tag)
|
||||
static int loop_queue_tgt_io(struct ublk_thread *t, struct ublk_queue *q, int tag)
|
||||
{
|
||||
const struct ublksrv_io_desc *iod = ublk_get_iod(q, tag);
|
||||
unsigned ublk_op = ublksrv_get_op(iod);
|
||||
@@ -80,7 +82,7 @@ static int loop_queue_tgt_io(struct ublk_queue *q, int tag)
|
||||
|
||||
switch (ublk_op) {
|
||||
case UBLK_IO_OP_FLUSH:
|
||||
ret = loop_queue_flush_io(q, iod, tag);
|
||||
ret = loop_queue_flush_io(t, q, iod, tag);
|
||||
break;
|
||||
case UBLK_IO_OP_WRITE_ZEROES:
|
||||
case UBLK_IO_OP_DISCARD:
|
||||
@@ -88,7 +90,7 @@ static int loop_queue_tgt_io(struct ublk_queue *q, int tag)
|
||||
break;
|
||||
case UBLK_IO_OP_READ:
|
||||
case UBLK_IO_OP_WRITE:
|
||||
ret = loop_queue_tgt_rw_io(q, iod, tag);
|
||||
ret = loop_queue_tgt_rw_io(t, q, iod, tag);
|
||||
break;
|
||||
default:
|
||||
ret = -EINVAL;
|
||||
@@ -100,15 +102,16 @@ static int loop_queue_tgt_io(struct ublk_queue *q, int tag)
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int ublk_loop_queue_io(struct ublk_queue *q, int tag)
|
||||
static int ublk_loop_queue_io(struct ublk_thread *t, struct ublk_queue *q,
|
||||
int tag)
|
||||
{
|
||||
int queued = loop_queue_tgt_io(q, tag);
|
||||
int queued = loop_queue_tgt_io(t, q, tag);
|
||||
|
||||
ublk_queued_tgt_io(q, tag, queued);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void ublk_loop_io_done(struct ublk_queue *q,
|
||||
static void ublk_loop_io_done(struct ublk_thread *t, struct ublk_queue *q,
|
||||
const struct io_uring_cqe *cqe)
|
||||
{
|
||||
unsigned tag = user_data_to_tag(cqe->user_data);
|
||||
|
||||
@@ -620,7 +620,7 @@ int ublk_queue_io_cmd(struct ublk_io *io)
|
||||
if (io_uring_sq_space_left(&t->ring) < 1)
|
||||
io_uring_submit(&t->ring);
|
||||
|
||||
ublk_io_alloc_sqes(io, sqe, 1);
|
||||
ublk_io_alloc_sqes(t, sqe, 1);
|
||||
if (!sqe[0]) {
|
||||
ublk_err("%s: run out of sqe. thread %u, tag %d\n",
|
||||
__func__, t->idx, io->tag);
|
||||
@@ -714,8 +714,9 @@ static int ublk_thread_is_done(struct ublk_thread *t)
|
||||
return (t->state & UBLKSRV_THREAD_STOPPING) && ublk_thread_is_idle(t);
|
||||
}
|
||||
|
||||
static inline void ublksrv_handle_tgt_cqe(struct ublk_queue *q,
|
||||
struct io_uring_cqe *cqe)
|
||||
static inline void ublksrv_handle_tgt_cqe(struct ublk_thread *t,
|
||||
struct ublk_queue *q,
|
||||
struct io_uring_cqe *cqe)
|
||||
{
|
||||
if (cqe->res < 0 && cqe->res != -EAGAIN)
|
||||
ublk_err("%s: failed tgt io: res %d qid %u tag %u, cmd_op %u\n",
|
||||
@@ -724,7 +725,7 @@ static inline void ublksrv_handle_tgt_cqe(struct ublk_queue *q,
|
||||
user_data_to_op(cqe->user_data));
|
||||
|
||||
if (q->tgt_ops->tgt_io_done)
|
||||
q->tgt_ops->tgt_io_done(q, cqe);
|
||||
q->tgt_ops->tgt_io_done(t, q, cqe);
|
||||
}
|
||||
|
||||
static void ublk_handle_cqe(struct ublk_thread *t,
|
||||
@@ -751,7 +752,7 @@ static void ublk_handle_cqe(struct ublk_thread *t,
|
||||
|
||||
/* Don't retrieve io in case of target io */
|
||||
if (is_target_io(cqe->user_data)) {
|
||||
ublksrv_handle_tgt_cqe(q, cqe);
|
||||
ublksrv_handle_tgt_cqe(t, q, cqe);
|
||||
return;
|
||||
}
|
||||
|
||||
@@ -766,7 +767,7 @@ static void ublk_handle_cqe(struct ublk_thread *t,
|
||||
if (cqe->res == UBLK_IO_RES_OK) {
|
||||
assert(tag < q->q_depth);
|
||||
if (q->tgt_ops->queue_io)
|
||||
q->tgt_ops->queue_io(q, tag);
|
||||
q->tgt_ops->queue_io(t, q, tag);
|
||||
} else if (cqe->res == UBLK_IO_RES_NEED_GET_DATA) {
|
||||
io->flags |= UBLKSRV_NEED_GET_DATA | UBLKSRV_IO_FREE;
|
||||
ublk_queue_io_cmd(io);
|
||||
|
||||
@@ -144,8 +144,9 @@ struct ublk_tgt_ops {
|
||||
int (*init_tgt)(const struct dev_ctx *ctx, struct ublk_dev *);
|
||||
void (*deinit_tgt)(struct ublk_dev *);
|
||||
|
||||
int (*queue_io)(struct ublk_queue *, int tag);
|
||||
void (*tgt_io_done)(struct ublk_queue *, const struct io_uring_cqe *);
|
||||
int (*queue_io)(struct ublk_thread *, struct ublk_queue *, int tag);
|
||||
void (*tgt_io_done)(struct ublk_thread *, struct ublk_queue *,
|
||||
const struct io_uring_cqe *);
|
||||
|
||||
/*
|
||||
* Target specific command line handling
|
||||
@@ -313,10 +314,10 @@ static inline struct ublk_queue *ublk_io_to_queue(const struct ublk_io *io)
|
||||
return container_of(io, struct ublk_queue, ios[io->tag]);
|
||||
}
|
||||
|
||||
static inline int ublk_io_alloc_sqes(struct ublk_io *io,
|
||||
static inline int ublk_io_alloc_sqes(struct ublk_thread *t,
|
||||
struct io_uring_sqe *sqes[], int nr_sqes)
|
||||
{
|
||||
struct io_uring *ring = &io->t->ring;
|
||||
struct io_uring *ring = &t->ring;
|
||||
unsigned left = io_uring_sq_space_left(ring);
|
||||
int i;
|
||||
|
||||
|
||||
@@ -55,12 +55,13 @@ static void __setup_nop_io(int tag, const struct ublksrv_io_desc *iod,
|
||||
sqe->user_data = build_user_data(tag, ublk_op, 0, q_id, 1);
|
||||
}
|
||||
|
||||
static int null_queue_zc_io(struct ublk_queue *q, int tag)
|
||||
static int null_queue_zc_io(struct ublk_thread *t, struct ublk_queue *q,
|
||||
int tag)
|
||||
{
|
||||
const struct ublksrv_io_desc *iod = ublk_get_iod(q, tag);
|
||||
struct io_uring_sqe *sqe[3];
|
||||
|
||||
ublk_io_alloc_sqes(ublk_get_io(q, tag), sqe, 3);
|
||||
ublk_io_alloc_sqes(t, sqe, 3);
|
||||
|
||||
io_uring_prep_buf_register(sqe[0], 0, tag, q->q_id, ublk_get_io(q, tag)->buf_index);
|
||||
sqe[0]->user_data = build_user_data(tag,
|
||||
@@ -77,18 +78,19 @@ static int null_queue_zc_io(struct ublk_queue *q, int tag)
|
||||
return 2;
|
||||
}
|
||||
|
||||
static int null_queue_auto_zc_io(struct ublk_queue *q, int tag)
|
||||
static int null_queue_auto_zc_io(struct ublk_thread *t, struct ublk_queue *q,
|
||||
int tag)
|
||||
{
|
||||
const struct ublksrv_io_desc *iod = ublk_get_iod(q, tag);
|
||||
struct io_uring_sqe *sqe[1];
|
||||
|
||||
ublk_io_alloc_sqes(ublk_get_io(q, tag), sqe, 1);
|
||||
ublk_io_alloc_sqes(t, sqe, 1);
|
||||
__setup_nop_io(tag, iod, sqe[0], q->q_id);
|
||||
return 1;
|
||||
}
|
||||
|
||||
static void ublk_null_io_done(struct ublk_queue *q,
|
||||
const struct io_uring_cqe *cqe)
|
||||
static void ublk_null_io_done(struct ublk_thread *t, struct ublk_queue *q,
|
||||
const struct io_uring_cqe *cqe)
|
||||
{
|
||||
unsigned tag = user_data_to_tag(cqe->user_data);
|
||||
unsigned op = user_data_to_op(cqe->user_data);
|
||||
@@ -110,7 +112,8 @@ static void ublk_null_io_done(struct ublk_queue *q,
|
||||
ublk_complete_io(q, tag, io->result);
|
||||
}
|
||||
|
||||
static int ublk_null_queue_io(struct ublk_queue *q, int tag)
|
||||
static int ublk_null_queue_io(struct ublk_thread *t, struct ublk_queue *q,
|
||||
int tag)
|
||||
{
|
||||
const struct ublksrv_io_desc *iod = ublk_get_iod(q, tag);
|
||||
unsigned auto_zc = ublk_queue_use_auto_zc(q);
|
||||
@@ -118,9 +121,9 @@ static int ublk_null_queue_io(struct ublk_queue *q, int tag)
|
||||
int queued;
|
||||
|
||||
if (auto_zc && !ublk_io_auto_zc_fallback(iod))
|
||||
queued = null_queue_auto_zc_io(q, tag);
|
||||
queued = null_queue_auto_zc_io(t, q, tag);
|
||||
else if (zc)
|
||||
queued = null_queue_zc_io(q, tag);
|
||||
queued = null_queue_zc_io(t, q, tag);
|
||||
else {
|
||||
ublk_complete_io(q, tag, iod->nr_sectors << 9);
|
||||
return 0;
|
||||
|
||||
@@ -123,7 +123,8 @@ static inline enum io_uring_op stripe_to_uring_op(
|
||||
assert(0);
|
||||
}
|
||||
|
||||
static int stripe_queue_tgt_rw_io(struct ublk_queue *q, const struct ublksrv_io_desc *iod, int tag)
|
||||
static int stripe_queue_tgt_rw_io(struct ublk_thread *t, struct ublk_queue *q,
|
||||
const struct ublksrv_io_desc *iod, int tag)
|
||||
{
|
||||
const struct stripe_conf *conf = get_chunk_shift(q);
|
||||
unsigned auto_zc = (ublk_queue_use_auto_zc(q) != 0);
|
||||
@@ -138,7 +139,7 @@ static int stripe_queue_tgt_rw_io(struct ublk_queue *q, const struct ublksrv_io_
|
||||
io->private_data = s;
|
||||
calculate_stripe_array(conf, iod, s, base);
|
||||
|
||||
ublk_io_alloc_sqes(ublk_get_io(q, tag), sqe, s->nr + extra);
|
||||
ublk_io_alloc_sqes(t, sqe, s->nr + extra);
|
||||
|
||||
if (zc) {
|
||||
io_uring_prep_buf_register(sqe[0], 0, tag, q->q_id, io->buf_index);
|
||||
@@ -176,13 +177,14 @@ static int stripe_queue_tgt_rw_io(struct ublk_queue *q, const struct ublksrv_io_
|
||||
return s->nr + zc;
|
||||
}
|
||||
|
||||
static int handle_flush(struct ublk_queue *q, const struct ublksrv_io_desc *iod, int tag)
|
||||
static int handle_flush(struct ublk_thread *t, struct ublk_queue *q,
|
||||
const struct ublksrv_io_desc *iod, int tag)
|
||||
{
|
||||
const struct stripe_conf *conf = get_chunk_shift(q);
|
||||
struct io_uring_sqe *sqe[NR_STRIPE];
|
||||
int i;
|
||||
|
||||
ublk_io_alloc_sqes(ublk_get_io(q, tag), sqe, conf->nr_files);
|
||||
ublk_io_alloc_sqes(t, sqe, conf->nr_files);
|
||||
for (i = 0; i < conf->nr_files; i++) {
|
||||
io_uring_prep_fsync(sqe[i], i + 1, IORING_FSYNC_DATASYNC);
|
||||
io_uring_sqe_set_flags(sqe[i], IOSQE_FIXED_FILE);
|
||||
@@ -191,7 +193,8 @@ static int handle_flush(struct ublk_queue *q, const struct ublksrv_io_desc *iod,
|
||||
return conf->nr_files;
|
||||
}
|
||||
|
||||
static int stripe_queue_tgt_io(struct ublk_queue *q, int tag)
|
||||
static int stripe_queue_tgt_io(struct ublk_thread *t, struct ublk_queue *q,
|
||||
int tag)
|
||||
{
|
||||
const struct ublksrv_io_desc *iod = ublk_get_iod(q, tag);
|
||||
unsigned ublk_op = ublksrv_get_op(iod);
|
||||
@@ -199,7 +202,7 @@ static int stripe_queue_tgt_io(struct ublk_queue *q, int tag)
|
||||
|
||||
switch (ublk_op) {
|
||||
case UBLK_IO_OP_FLUSH:
|
||||
ret = handle_flush(q, iod, tag);
|
||||
ret = handle_flush(t, q, iod, tag);
|
||||
break;
|
||||
case UBLK_IO_OP_WRITE_ZEROES:
|
||||
case UBLK_IO_OP_DISCARD:
|
||||
@@ -207,7 +210,7 @@ static int stripe_queue_tgt_io(struct ublk_queue *q, int tag)
|
||||
break;
|
||||
case UBLK_IO_OP_READ:
|
||||
case UBLK_IO_OP_WRITE:
|
||||
ret = stripe_queue_tgt_rw_io(q, iod, tag);
|
||||
ret = stripe_queue_tgt_rw_io(t, q, iod, tag);
|
||||
break;
|
||||
default:
|
||||
ret = -EINVAL;
|
||||
@@ -218,16 +221,17 @@ static int stripe_queue_tgt_io(struct ublk_queue *q, int tag)
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int ublk_stripe_queue_io(struct ublk_queue *q, int tag)
|
||||
static int ublk_stripe_queue_io(struct ublk_thread *t, struct ublk_queue *q,
|
||||
int tag)
|
||||
{
|
||||
int queued = stripe_queue_tgt_io(q, tag);
|
||||
int queued = stripe_queue_tgt_io(t, q, tag);
|
||||
|
||||
ublk_queued_tgt_io(q, tag, queued);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void ublk_stripe_io_done(struct ublk_queue *q,
|
||||
const struct io_uring_cqe *cqe)
|
||||
static void ublk_stripe_io_done(struct ublk_thread *t, struct ublk_queue *q,
|
||||
const struct io_uring_cqe *cqe)
|
||||
{
|
||||
unsigned tag = user_data_to_tag(cqe->user_data);
|
||||
const struct ublksrv_io_desc *iod = ublk_get_iod(q, tag);
|
||||
|
||||
Reference in New Issue
Block a user