io_uring: pass hash table into poll_find

In preparation for having multiple cancellation hash tables, pass a
table pointer into io_poll_find() and other poll cancel functions.

Signed-off-by: Pavel Begunkov <asml.silence@gmail.com>
Link: https://lore.kernel.org/r/a31c88502463dce09254240fa037352927d7ecc3.1655371007.git.asml.silence@gmail.com
Reviewed-by: Hao Xu <howeyxu@tencent.com>
Signed-off-by: Jens Axboe <axboe@kernel.dk>
This commit is contained in:
Pavel Begunkov 2022-06-16 10:22:09 +01:00 коммит произвёл Jens Axboe
Родитель 97bbdc06a4
Коммит a2cdd51932
1 изменённых файлов: 14 добавлений и 6 удалений

Просмотреть файл

@ -562,11 +562,12 @@ __cold bool io_poll_remove_all(struct io_ring_ctx *ctx, struct task_struct *tsk,
static struct io_kiocb *io_poll_find(struct io_ring_ctx *ctx, bool poll_only,
struct io_cancel_data *cd,
struct io_hash_bucket hash_table[],
struct io_hash_bucket **out_bucket)
{
struct io_kiocb *req;
u32 index = hash_long(cd->data, ctx->cancel_hash_bits);
struct io_hash_bucket *hb = &ctx->cancel_hash[index];
struct io_hash_bucket *hb = &hash_table[index];
*out_bucket = NULL;
@ -590,6 +591,7 @@ static struct io_kiocb *io_poll_find(struct io_ring_ctx *ctx, bool poll_only,
static struct io_kiocb *io_poll_file_find(struct io_ring_ctx *ctx,
struct io_cancel_data *cd,
struct io_hash_bucket hash_table[],
struct io_hash_bucket **out_bucket)
{
struct io_kiocb *req;
@ -598,7 +600,7 @@ static struct io_kiocb *io_poll_file_find(struct io_ring_ctx *ctx,
*out_bucket = NULL;
for (i = 0; i < (1U << ctx->cancel_hash_bits); i++) {
struct io_hash_bucket *hb = &ctx->cancel_hash[i];
struct io_hash_bucket *hb = &hash_table[i];
spin_lock(&hb->lock);
hlist_for_each_entry(req, &hb->list, hash_node) {
@ -625,15 +627,16 @@ static bool io_poll_disarm(struct io_kiocb *req)
return true;
}
int io_poll_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd)
static int __io_poll_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd,
struct io_hash_bucket hash_table[])
{
struct io_hash_bucket *bucket;
struct io_kiocb *req;
if (cd->flags & (IORING_ASYNC_CANCEL_FD|IORING_ASYNC_CANCEL_ANY))
req = io_poll_file_find(ctx, cd, &bucket);
req = io_poll_file_find(ctx, cd, ctx->cancel_hash, &bucket);
else
req = io_poll_find(ctx, false, cd, &bucket);
req = io_poll_find(ctx, false, cd, ctx->cancel_hash, &bucket);
if (req)
io_poll_cancel_req(req);
@ -642,6 +645,11 @@ int io_poll_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd)
return req ? 0 : -ENOENT;
}
int io_poll_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd)
{
return __io_poll_cancel(ctx, cd, ctx->cancel_hash);
}
static __poll_t io_poll_parse_events(const struct io_uring_sqe *sqe,
unsigned int flags)
{
@ -737,7 +745,7 @@ int io_poll_remove(struct io_kiocb *req, unsigned int issue_flags)
int ret2, ret = 0;
bool locked;
preq = io_poll_find(ctx, true, &cd, &bucket);
preq = io_poll_find(ctx, true, &cd, ctx->cancel_hash, &bucket);
if (preq)
ret2 = io_poll_disarm(preq);
if (bucket)