Lines Matching defs:tmp
1288 struct io_kiocb *req, *tmp;
1292 llist_for_each_entry_safe(req, tmp, node, io_task_work.fallback_node)
1542 struct io_kiocb *req, *tmp;
1545 list_for_each_entry_safe(req, tmp, &ctx->timeout_list, timeout.list) {
2549 struct io_kiocb *req, *tmp;
2559 list_for_each_entry_safe(req, tmp, &ctx->iopoll_list, inflight_entry) {
4301 u64 tmp;
4307 tmp = READ_ONCE(sqe->fd);
4308 if (!tmp || tmp > USHRT_MAX)
4312 p->nbufs = tmp;
4374 u64 tmp;
4379 tmp = READ_ONCE(sqe->fd);
4380 if (!tmp || tmp > USHRT_MAX)
4382 p->nbufs = tmp;
4397 tmp = READ_ONCE(sqe->off);
4398 if (tmp > USHRT_MAX)
4400 p->bid = tmp;
5861 struct hlist_node *tmp;
5871 hlist_for_each_entry_safe(req, tmp, list, hash_node) {
8331 struct io_rsrc_put *prsrc, *tmp;
8333 list_for_each_entry_safe(prsrc, tmp, &ref_node->rsrc_list, list) {
9158 unsigned long tmp, acct_len = iov->iov_len + (PAGE_SIZE - 1);
9174 if (check_add_overflow((unsigned long)iov->iov_base, acct_len, &tmp))
9587 struct io_kiocb *req, *tmp;
9592 list_for_each_entry_safe(req, tmp, &ctx->timeout_list, timeout.list) {
10670 __u32 tmp;
10673 if (check_add_overflow(up->offset, nr_args, &tmp))