mirror of
https://github.com/torvalds/linux.git
synced 2024-11-21 19:46:16 +00:00
io_uring: move cancelations to be io_uring_task based
Right now the task_struct pointer is used as the key to match a task, but in preparation for some io_kiocb changes, move it to using struct io_uring_task instead. No functional changes intended in this patch. Signed-off-by: Jens Axboe <axboe@kernel.dk>
This commit is contained in:
parent
6f94cbc29a
commit
f03baece08
@ -141,7 +141,7 @@ int io_futex_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd,
|
|||||||
return -ENOENT;
|
return -ENOENT;
|
||||||
}
|
}
|
||||||
|
|
||||||
bool io_futex_remove_all(struct io_ring_ctx *ctx, struct task_struct *task,
|
bool io_futex_remove_all(struct io_ring_ctx *ctx, struct io_uring_task *tctx,
|
||||||
bool cancel_all)
|
bool cancel_all)
|
||||||
{
|
{
|
||||||
struct hlist_node *tmp;
|
struct hlist_node *tmp;
|
||||||
@ -151,7 +151,7 @@ bool io_futex_remove_all(struct io_ring_ctx *ctx, struct task_struct *task,
|
|||||||
lockdep_assert_held(&ctx->uring_lock);
|
lockdep_assert_held(&ctx->uring_lock);
|
||||||
|
|
||||||
hlist_for_each_entry_safe(req, tmp, &ctx->futex_list, hash_node) {
|
hlist_for_each_entry_safe(req, tmp, &ctx->futex_list, hash_node) {
|
||||||
if (!io_match_task_safe(req, task, cancel_all))
|
if (!io_match_task_safe(req, tctx, cancel_all))
|
||||||
continue;
|
continue;
|
||||||
hlist_del_init(&req->hash_node);
|
hlist_del_init(&req->hash_node);
|
||||||
__io_futex_cancel(ctx, req);
|
__io_futex_cancel(ctx, req);
|
||||||
|
@ -11,7 +11,7 @@ int io_futex_wake(struct io_kiocb *req, unsigned int issue_flags);
|
|||||||
#if defined(CONFIG_FUTEX)
|
#if defined(CONFIG_FUTEX)
|
||||||
int io_futex_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd,
|
int io_futex_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd,
|
||||||
unsigned int issue_flags);
|
unsigned int issue_flags);
|
||||||
bool io_futex_remove_all(struct io_ring_ctx *ctx, struct task_struct *task,
|
bool io_futex_remove_all(struct io_ring_ctx *ctx, struct io_uring_task *tctx,
|
||||||
bool cancel_all);
|
bool cancel_all);
|
||||||
bool io_futex_cache_init(struct io_ring_ctx *ctx);
|
bool io_futex_cache_init(struct io_ring_ctx *ctx);
|
||||||
void io_futex_cache_free(struct io_ring_ctx *ctx);
|
void io_futex_cache_free(struct io_ring_ctx *ctx);
|
||||||
@ -23,7 +23,7 @@ static inline int io_futex_cancel(struct io_ring_ctx *ctx,
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
static inline bool io_futex_remove_all(struct io_ring_ctx *ctx,
|
static inline bool io_futex_remove_all(struct io_ring_ctx *ctx,
|
||||||
struct task_struct *task, bool cancel_all)
|
struct io_uring_task *tctx, bool cancel_all)
|
||||||
{
|
{
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
@ -142,7 +142,7 @@ struct io_defer_entry {
|
|||||||
#define IO_CQ_WAKE_FORCE (IO_CQ_WAKE_INIT >> 1)
|
#define IO_CQ_WAKE_FORCE (IO_CQ_WAKE_INIT >> 1)
|
||||||
|
|
||||||
static bool io_uring_try_cancel_requests(struct io_ring_ctx *ctx,
|
static bool io_uring_try_cancel_requests(struct io_ring_ctx *ctx,
|
||||||
struct task_struct *task,
|
struct io_uring_task *tctx,
|
||||||
bool cancel_all);
|
bool cancel_all);
|
||||||
|
|
||||||
static void io_queue_sqe(struct io_kiocb *req);
|
static void io_queue_sqe(struct io_kiocb *req);
|
||||||
@ -201,12 +201,12 @@ static bool io_match_linked(struct io_kiocb *head)
|
|||||||
* As io_match_task() but protected against racing with linked timeouts.
|
* As io_match_task() but protected against racing with linked timeouts.
|
||||||
* User must not hold timeout_lock.
|
* User must not hold timeout_lock.
|
||||||
*/
|
*/
|
||||||
bool io_match_task_safe(struct io_kiocb *head, struct task_struct *task,
|
bool io_match_task_safe(struct io_kiocb *head, struct io_uring_task *tctx,
|
||||||
bool cancel_all)
|
bool cancel_all)
|
||||||
{
|
{
|
||||||
bool matched;
|
bool matched;
|
||||||
|
|
||||||
if (task && head->task != task)
|
if (tctx && head->task->io_uring != tctx)
|
||||||
return false;
|
return false;
|
||||||
if (cancel_all)
|
if (cancel_all)
|
||||||
return true;
|
return true;
|
||||||
@ -2987,7 +2987,7 @@ static int io_uring_release(struct inode *inode, struct file *file)
|
|||||||
}
|
}
|
||||||
|
|
||||||
struct io_task_cancel {
|
struct io_task_cancel {
|
||||||
struct task_struct *task;
|
struct io_uring_task *tctx;
|
||||||
bool all;
|
bool all;
|
||||||
};
|
};
|
||||||
|
|
||||||
@ -2996,11 +2996,11 @@ static bool io_cancel_task_cb(struct io_wq_work *work, void *data)
|
|||||||
struct io_kiocb *req = container_of(work, struct io_kiocb, work);
|
struct io_kiocb *req = container_of(work, struct io_kiocb, work);
|
||||||
struct io_task_cancel *cancel = data;
|
struct io_task_cancel *cancel = data;
|
||||||
|
|
||||||
return io_match_task_safe(req, cancel->task, cancel->all);
|
return io_match_task_safe(req, cancel->tctx, cancel->all);
|
||||||
}
|
}
|
||||||
|
|
||||||
static __cold bool io_cancel_defer_files(struct io_ring_ctx *ctx,
|
static __cold bool io_cancel_defer_files(struct io_ring_ctx *ctx,
|
||||||
struct task_struct *task,
|
struct io_uring_task *tctx,
|
||||||
bool cancel_all)
|
bool cancel_all)
|
||||||
{
|
{
|
||||||
struct io_defer_entry *de;
|
struct io_defer_entry *de;
|
||||||
@ -3008,7 +3008,7 @@ static __cold bool io_cancel_defer_files(struct io_ring_ctx *ctx,
|
|||||||
|
|
||||||
spin_lock(&ctx->completion_lock);
|
spin_lock(&ctx->completion_lock);
|
||||||
list_for_each_entry_reverse(de, &ctx->defer_list, list) {
|
list_for_each_entry_reverse(de, &ctx->defer_list, list) {
|
||||||
if (io_match_task_safe(de->req, task, cancel_all)) {
|
if (io_match_task_safe(de->req, tctx, cancel_all)) {
|
||||||
list_cut_position(&list, &ctx->defer_list, &de->list);
|
list_cut_position(&list, &ctx->defer_list, &de->list);
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
@ -3051,11 +3051,10 @@ static __cold bool io_uring_try_cancel_iowq(struct io_ring_ctx *ctx)
|
|||||||
}
|
}
|
||||||
|
|
||||||
static __cold bool io_uring_try_cancel_requests(struct io_ring_ctx *ctx,
|
static __cold bool io_uring_try_cancel_requests(struct io_ring_ctx *ctx,
|
||||||
struct task_struct *task,
|
struct io_uring_task *tctx,
|
||||||
bool cancel_all)
|
bool cancel_all)
|
||||||
{
|
{
|
||||||
struct io_task_cancel cancel = { .task = task, .all = cancel_all, };
|
struct io_task_cancel cancel = { .tctx = tctx, .all = cancel_all, };
|
||||||
struct io_uring_task *tctx = task ? task->io_uring : NULL;
|
|
||||||
enum io_wq_cancel cret;
|
enum io_wq_cancel cret;
|
||||||
bool ret = false;
|
bool ret = false;
|
||||||
|
|
||||||
@ -3069,9 +3068,9 @@ static __cold bool io_uring_try_cancel_requests(struct io_ring_ctx *ctx,
|
|||||||
if (!ctx->rings)
|
if (!ctx->rings)
|
||||||
return false;
|
return false;
|
||||||
|
|
||||||
if (!task) {
|
if (!tctx) {
|
||||||
ret |= io_uring_try_cancel_iowq(ctx);
|
ret |= io_uring_try_cancel_iowq(ctx);
|
||||||
} else if (tctx && tctx->io_wq) {
|
} else if (tctx->io_wq) {
|
||||||
/*
|
/*
|
||||||
* Cancels requests of all rings, not only @ctx, but
|
* Cancels requests of all rings, not only @ctx, but
|
||||||
* it's fine as the task is in exit/exec.
|
* it's fine as the task is in exit/exec.
|
||||||
@ -3094,15 +3093,15 @@ static __cold bool io_uring_try_cancel_requests(struct io_ring_ctx *ctx,
|
|||||||
if ((ctx->flags & IORING_SETUP_DEFER_TASKRUN) &&
|
if ((ctx->flags & IORING_SETUP_DEFER_TASKRUN) &&
|
||||||
io_allowed_defer_tw_run(ctx))
|
io_allowed_defer_tw_run(ctx))
|
||||||
ret |= io_run_local_work(ctx, INT_MAX) > 0;
|
ret |= io_run_local_work(ctx, INT_MAX) > 0;
|
||||||
ret |= io_cancel_defer_files(ctx, task, cancel_all);
|
ret |= io_cancel_defer_files(ctx, tctx, cancel_all);
|
||||||
mutex_lock(&ctx->uring_lock);
|
mutex_lock(&ctx->uring_lock);
|
||||||
ret |= io_poll_remove_all(ctx, task, cancel_all);
|
ret |= io_poll_remove_all(ctx, tctx, cancel_all);
|
||||||
ret |= io_waitid_remove_all(ctx, task, cancel_all);
|
ret |= io_waitid_remove_all(ctx, tctx, cancel_all);
|
||||||
ret |= io_futex_remove_all(ctx, task, cancel_all);
|
ret |= io_futex_remove_all(ctx, tctx, cancel_all);
|
||||||
ret |= io_uring_try_cancel_uring_cmd(ctx, task, cancel_all);
|
ret |= io_uring_try_cancel_uring_cmd(ctx, tctx, cancel_all);
|
||||||
mutex_unlock(&ctx->uring_lock);
|
mutex_unlock(&ctx->uring_lock);
|
||||||
ret |= io_kill_timeouts(ctx, task, cancel_all);
|
ret |= io_kill_timeouts(ctx, tctx, cancel_all);
|
||||||
if (task)
|
if (tctx)
|
||||||
ret |= io_run_task_work() > 0;
|
ret |= io_run_task_work() > 0;
|
||||||
else
|
else
|
||||||
ret |= flush_delayed_work(&ctx->fallback_work);
|
ret |= flush_delayed_work(&ctx->fallback_work);
|
||||||
@ -3155,12 +3154,13 @@ __cold void io_uring_cancel_generic(bool cancel_all, struct io_sq_data *sqd)
|
|||||||
if (node->ctx->sq_data)
|
if (node->ctx->sq_data)
|
||||||
continue;
|
continue;
|
||||||
loop |= io_uring_try_cancel_requests(node->ctx,
|
loop |= io_uring_try_cancel_requests(node->ctx,
|
||||||
current, cancel_all);
|
current->io_uring,
|
||||||
|
cancel_all);
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
list_for_each_entry(ctx, &sqd->ctx_list, sqd_list)
|
list_for_each_entry(ctx, &sqd->ctx_list, sqd_list)
|
||||||
loop |= io_uring_try_cancel_requests(ctx,
|
loop |= io_uring_try_cancel_requests(ctx,
|
||||||
current,
|
current->io_uring,
|
||||||
cancel_all);
|
cancel_all);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -115,7 +115,7 @@ void io_queue_next(struct io_kiocb *req);
|
|||||||
void io_task_refs_refill(struct io_uring_task *tctx);
|
void io_task_refs_refill(struct io_uring_task *tctx);
|
||||||
bool __io_alloc_req_refill(struct io_ring_ctx *ctx);
|
bool __io_alloc_req_refill(struct io_ring_ctx *ctx);
|
||||||
|
|
||||||
bool io_match_task_safe(struct io_kiocb *head, struct task_struct *task,
|
bool io_match_task_safe(struct io_kiocb *head, struct io_uring_task *tctx,
|
||||||
bool cancel_all);
|
bool cancel_all);
|
||||||
|
|
||||||
void io_activate_pollwq(struct io_ring_ctx *ctx);
|
void io_activate_pollwq(struct io_ring_ctx *ctx);
|
||||||
|
@ -714,7 +714,7 @@ int io_arm_poll_handler(struct io_kiocb *req, unsigned issue_flags)
|
|||||||
/*
|
/*
|
||||||
* Returns true if we found and killed one or more poll requests
|
* Returns true if we found and killed one or more poll requests
|
||||||
*/
|
*/
|
||||||
__cold bool io_poll_remove_all(struct io_ring_ctx *ctx, struct task_struct *tsk,
|
__cold bool io_poll_remove_all(struct io_ring_ctx *ctx, struct io_uring_task *tctx,
|
||||||
bool cancel_all)
|
bool cancel_all)
|
||||||
{
|
{
|
||||||
unsigned nr_buckets = 1U << ctx->cancel_table.hash_bits;
|
unsigned nr_buckets = 1U << ctx->cancel_table.hash_bits;
|
||||||
@ -729,7 +729,7 @@ __cold bool io_poll_remove_all(struct io_ring_ctx *ctx, struct task_struct *tsk,
|
|||||||
struct io_hash_bucket *hb = &ctx->cancel_table.hbs[i];
|
struct io_hash_bucket *hb = &ctx->cancel_table.hbs[i];
|
||||||
|
|
||||||
hlist_for_each_entry_safe(req, tmp, &hb->list, hash_node) {
|
hlist_for_each_entry_safe(req, tmp, &hb->list, hash_node) {
|
||||||
if (io_match_task_safe(req, tsk, cancel_all)) {
|
if (io_match_task_safe(req, tctx, cancel_all)) {
|
||||||
hlist_del_init(&req->hash_node);
|
hlist_del_init(&req->hash_node);
|
||||||
io_poll_cancel_req(req);
|
io_poll_cancel_req(req);
|
||||||
found = true;
|
found = true;
|
||||||
|
@ -40,7 +40,7 @@ struct io_cancel_data;
|
|||||||
int io_poll_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd,
|
int io_poll_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd,
|
||||||
unsigned issue_flags);
|
unsigned issue_flags);
|
||||||
int io_arm_poll_handler(struct io_kiocb *req, unsigned issue_flags);
|
int io_arm_poll_handler(struct io_kiocb *req, unsigned issue_flags);
|
||||||
bool io_poll_remove_all(struct io_ring_ctx *ctx, struct task_struct *tsk,
|
bool io_poll_remove_all(struct io_ring_ctx *ctx, struct io_uring_task *tctx,
|
||||||
bool cancel_all);
|
bool cancel_all);
|
||||||
|
|
||||||
void io_poll_task_func(struct io_kiocb *req, struct io_tw_state *ts);
|
void io_poll_task_func(struct io_kiocb *req, struct io_tw_state *ts);
|
||||||
|
@ -637,13 +637,13 @@ void io_queue_linked_timeout(struct io_kiocb *req)
|
|||||||
io_put_req(req);
|
io_put_req(req);
|
||||||
}
|
}
|
||||||
|
|
||||||
static bool io_match_task(struct io_kiocb *head, struct task_struct *task,
|
static bool io_match_task(struct io_kiocb *head, struct io_uring_task *tctx,
|
||||||
bool cancel_all)
|
bool cancel_all)
|
||||||
__must_hold(&head->ctx->timeout_lock)
|
__must_hold(&head->ctx->timeout_lock)
|
||||||
{
|
{
|
||||||
struct io_kiocb *req;
|
struct io_kiocb *req;
|
||||||
|
|
||||||
if (task && head->task != task)
|
if (tctx && head->task->io_uring != tctx)
|
||||||
return false;
|
return false;
|
||||||
if (cancel_all)
|
if (cancel_all)
|
||||||
return true;
|
return true;
|
||||||
@ -656,7 +656,7 @@ static bool io_match_task(struct io_kiocb *head, struct task_struct *task,
|
|||||||
}
|
}
|
||||||
|
|
||||||
/* Returns true if we found and killed one or more timeouts */
|
/* Returns true if we found and killed one or more timeouts */
|
||||||
__cold bool io_kill_timeouts(struct io_ring_ctx *ctx, struct task_struct *tsk,
|
__cold bool io_kill_timeouts(struct io_ring_ctx *ctx, struct io_uring_task *tctx,
|
||||||
bool cancel_all)
|
bool cancel_all)
|
||||||
{
|
{
|
||||||
struct io_timeout *timeout, *tmp;
|
struct io_timeout *timeout, *tmp;
|
||||||
@ -671,7 +671,7 @@ __cold bool io_kill_timeouts(struct io_ring_ctx *ctx, struct task_struct *tsk,
|
|||||||
list_for_each_entry_safe(timeout, tmp, &ctx->timeout_list, list) {
|
list_for_each_entry_safe(timeout, tmp, &ctx->timeout_list, list) {
|
||||||
struct io_kiocb *req = cmd_to_io_kiocb(timeout);
|
struct io_kiocb *req = cmd_to_io_kiocb(timeout);
|
||||||
|
|
||||||
if (io_match_task(req, tsk, cancel_all) &&
|
if (io_match_task(req, tctx, cancel_all) &&
|
||||||
io_kill_timeout(req, -ECANCELED))
|
io_kill_timeout(req, -ECANCELED))
|
||||||
canceled++;
|
canceled++;
|
||||||
}
|
}
|
||||||
|
@ -24,7 +24,7 @@ static inline struct io_kiocb *io_disarm_linked_timeout(struct io_kiocb *req)
|
|||||||
__cold void io_flush_timeouts(struct io_ring_ctx *ctx);
|
__cold void io_flush_timeouts(struct io_ring_ctx *ctx);
|
||||||
struct io_cancel_data;
|
struct io_cancel_data;
|
||||||
int io_timeout_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd);
|
int io_timeout_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd);
|
||||||
__cold bool io_kill_timeouts(struct io_ring_ctx *ctx, struct task_struct *tsk,
|
__cold bool io_kill_timeouts(struct io_ring_ctx *ctx, struct io_uring_task *tctx,
|
||||||
bool cancel_all);
|
bool cancel_all);
|
||||||
void io_queue_linked_timeout(struct io_kiocb *req);
|
void io_queue_linked_timeout(struct io_kiocb *req);
|
||||||
void io_disarm_next(struct io_kiocb *req);
|
void io_disarm_next(struct io_kiocb *req);
|
||||||
|
@ -47,7 +47,7 @@ static void io_req_uring_cleanup(struct io_kiocb *req, unsigned int issue_flags)
|
|||||||
}
|
}
|
||||||
|
|
||||||
bool io_uring_try_cancel_uring_cmd(struct io_ring_ctx *ctx,
|
bool io_uring_try_cancel_uring_cmd(struct io_ring_ctx *ctx,
|
||||||
struct task_struct *task, bool cancel_all)
|
struct io_uring_task *tctx, bool cancel_all)
|
||||||
{
|
{
|
||||||
struct hlist_node *tmp;
|
struct hlist_node *tmp;
|
||||||
struct io_kiocb *req;
|
struct io_kiocb *req;
|
||||||
@ -61,7 +61,7 @@ bool io_uring_try_cancel_uring_cmd(struct io_ring_ctx *ctx,
|
|||||||
struct io_uring_cmd);
|
struct io_uring_cmd);
|
||||||
struct file *file = req->file;
|
struct file *file = req->file;
|
||||||
|
|
||||||
if (!cancel_all && req->task != task)
|
if (!cancel_all && req->task->io_uring != tctx)
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
if (cmd->flags & IORING_URING_CMD_CANCELABLE) {
|
if (cmd->flags & IORING_URING_CMD_CANCELABLE) {
|
||||||
|
@ -8,4 +8,4 @@ int io_uring_cmd(struct io_kiocb *req, unsigned int issue_flags);
|
|||||||
int io_uring_cmd_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
int io_uring_cmd_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
||||||
|
|
||||||
bool io_uring_try_cancel_uring_cmd(struct io_ring_ctx *ctx,
|
bool io_uring_try_cancel_uring_cmd(struct io_ring_ctx *ctx,
|
||||||
struct task_struct *task, bool cancel_all);
|
struct io_uring_task *tctx, bool cancel_all);
|
||||||
|
@ -184,7 +184,7 @@ int io_waitid_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd,
|
|||||||
return -ENOENT;
|
return -ENOENT;
|
||||||
}
|
}
|
||||||
|
|
||||||
bool io_waitid_remove_all(struct io_ring_ctx *ctx, struct task_struct *task,
|
bool io_waitid_remove_all(struct io_ring_ctx *ctx, struct io_uring_task *tctx,
|
||||||
bool cancel_all)
|
bool cancel_all)
|
||||||
{
|
{
|
||||||
struct hlist_node *tmp;
|
struct hlist_node *tmp;
|
||||||
@ -194,7 +194,7 @@ bool io_waitid_remove_all(struct io_ring_ctx *ctx, struct task_struct *task,
|
|||||||
lockdep_assert_held(&ctx->uring_lock);
|
lockdep_assert_held(&ctx->uring_lock);
|
||||||
|
|
||||||
hlist_for_each_entry_safe(req, tmp, &ctx->waitid_list, hash_node) {
|
hlist_for_each_entry_safe(req, tmp, &ctx->waitid_list, hash_node) {
|
||||||
if (!io_match_task_safe(req, task, cancel_all))
|
if (!io_match_task_safe(req, tctx, cancel_all))
|
||||||
continue;
|
continue;
|
||||||
hlist_del_init(&req->hash_node);
|
hlist_del_init(&req->hash_node);
|
||||||
__io_waitid_cancel(ctx, req);
|
__io_waitid_cancel(ctx, req);
|
||||||
|
@ -11,5 +11,5 @@ int io_waitid_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
|||||||
int io_waitid(struct io_kiocb *req, unsigned int issue_flags);
|
int io_waitid(struct io_kiocb *req, unsigned int issue_flags);
|
||||||
int io_waitid_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd,
|
int io_waitid_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd,
|
||||||
unsigned int issue_flags);
|
unsigned int issue_flags);
|
||||||
bool io_waitid_remove_all(struct io_ring_ctx *ctx, struct task_struct *task,
|
bool io_waitid_remove_all(struct io_ring_ctx *ctx, struct io_uring_task *tctx,
|
||||||
bool cancel_all);
|
bool cancel_all);
|
||||||
|
Loading…
Reference in New Issue
Block a user