Lines Matching defs:int
205 unsigned int nr_bvecs;
251 unsigned int nr;
290 int sq_cpu;
315 unsigned int free_reqs;
323 unsigned int compl_nr;
327 unsigned int ios_left;
336 unsigned int flags;
337 unsigned int compat: 1;
338 unsigned int drain_next: 1;
339 unsigned int eventfd_async: 1;
340 unsigned int restricted: 1;
341 unsigned int off_timeout_used: 1;
342 unsigned int drain_active: 1;
388 unsigned int locked_free_nr;
466 int cached_refs;
489 int retries;
504 int fd;
519 int __user *addr_len;
520 int flags;
529 int flags;
530 int mode;
569 int addr_len;
579 int msg_flags;
580 int bgid;
589 int dfd;
619 int epfd;
620 int op;
621 int fd;
630 int splice_fd_in;
631 unsigned int flags;
645 int dfd;
646 unsigned int mask;
647 unsigned int flags;
654 int how;
659 int old_dfd;
660 int new_dfd;
663 int flags;
668 int dfd;
669 int flags;
675 int dfd;
682 int new_dfd;
689 int old_dfd;
690 int new_dfd;
693 int flags;
873 unsigned int flags;
1092 static int __io_register_rsrc_update(struct io_ring_ctx *ctx, unsigned type,
1097 struct io_kiocb *req, int fd, bool fixed,
1098 unsigned int issue_flags);
1104 static int io_req_prep_async(struct io_kiocb *req);
1106 static int io_install_fixed_file(struct io_kiocb *req, struct file *file,
1107 unsigned int issue_flags, u32 slot_index);
1108 static int io_close_fixed(struct io_kiocb *req, unsigned int issue_flags);
1145 ((unsigned int) atomic_read(&(req->refs)) + 127u <= 127u)
1169 static inline void __io_req_set_refcount(struct io_kiocb *req, int nr)
1265 static inline void req_fail_link_node(struct io_kiocb *req, int res)
1307 int hash_bits;
1505 static void io_kill_timeout(struct io_kiocb *req, int status)
1602 static inline unsigned int __io_cqring_events(struct io_ring_ctx *ctx)
1736 static inline void io_put_task(struct task_struct *task, int nr)
1752 unsigned int refill = -tctx->cached_refs + IO_TCTX_REFS_CACHE_NR;
1759 static inline void io_get_task_refs(int nr)
1771 unsigned int refs = tctx->cached_refs;
1951 int nr;
1987 int ret, i;
2024 unsigned int flags = req->flags;
2302 static void io_req_task_queue_fail(struct io_kiocb *req, int ret)
2342 int task_refs;
2343 int ctx_refs;
2387 int i, nr = state->compl_nr;
2449 static inline unsigned int io_sqring_entries(struct io_ring_ctx *ctx)
2457 static unsigned int io_put_kbuf(struct io_kiocb *req, struct io_buffer *kbuf)
2459 unsigned int cflags;
2468 static inline unsigned int io_put_rw_kbuf(struct io_kiocb *req)
2501 static void io_iopoll_complete(struct io_ring_ctx *ctx, unsigned int *nr_events,
2546 static int io_do_iopoll(struct io_ring_ctx *ctx, unsigned int *nr_events,
2561 int ret;
2603 unsigned int nr_events = 0;
2624 static int io_iopoll_check(struct io_ring_ctx *ctx, long min)
2626 unsigned int nr_events = 0;
2627 int ret = 0;
2775 static inline int io_fixup_rw_res(struct io_kiocb *req, long res)
2791 unsigned int cflags = io_put_rw_kbuf(req);
2792 int res = req->result;
2867 unsigned int queue_num0, queue_num1;
2916 static bool __io_file_supports_nowait(struct file *file, int rw)
2949 static bool io_file_supports_nowait(struct io_kiocb *req, int rw)
2959 static int io_prep_rw(struct io_kiocb *req, const struct io_uring_sqe *sqe,
2960 int rw)
2966 int ret;
3071 unsigned int issue_flags)
3097 unsigned int cflags = io_put_rw_kbuf(req);
3113 static int __io_import_fixed(struct io_kiocb *req, int rw, struct iov_iter *iter,
3171 static int io_import_fixed(struct io_kiocb *req, int rw, struct iov_iter *iter)
3197 int bgid, struct io_buffer *kbuf,
3316 static int io_import_iovec(int rw, struct io_kiocb *req, struct iovec **iovec,
3367 static ssize_t loop_rw_iter(int rw, struct io_kiocb *req, struct iov_iter *iter)
3453 static inline int io_alloc_async_data(struct io_kiocb *req)
3460 static int io_setup_async_rw(struct io_kiocb *req, const struct iovec *iovec,
3482 static inline int io_rw_prep_async(struct io_kiocb *req, int rw)
3486 int ret;
3503 static int io_read_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe)
3520 static int io_async_buf_func(struct wait_queue_entry *wait, unsigned mode,
3521 int sync, void *arg)
3581 static inline int io_iter_do_read(struct io_kiocb *req, struct iov_iter *iter)
3597 static int io_read(struct io_kiocb *req, unsigned int issue_flags)
3729 static int io_write_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe)
3736 static int io_write(struct io_kiocb *req, unsigned int issue_flags)
3841 static int io_renameat_prep(struct io_kiocb *req,
3874 static int io_renameat(struct io_kiocb *req, unsigned int issue_flags)
3877 int ret;
3892 static int io_unlinkat_prep(struct io_kiocb *req,
3921 static int io_unlinkat(struct io_kiocb *req, unsigned int issue_flags)
3924 int ret;
3941 static int io_shutdown_prep(struct io_kiocb *req,
3958 static int io_shutdown(struct io_kiocb *req, unsigned int issue_flags)
3962 int ret;
3981 static int __io_splice_prep(struct io_kiocb *req,
3985 unsigned int valid_flags = SPLICE_F_FD_IN_FIXED | SPLICE_F_ALL;
3998 static int io_tee_prep(struct io_kiocb *req,
4006 static int io_tee(struct io_kiocb *req, unsigned int issue_flags)
4010 unsigned int flags = sp->flags & ~SPLICE_F_FD_IN_FIXED;
4036 static int io_splice_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe)
4045 static int io_splice(struct io_kiocb *req, unsigned int issue_flags)
4049 unsigned int flags = sp->flags & ~SPLICE_F_FD_IN_FIXED;
4082 static int io_nop(struct io_kiocb *req, unsigned int issue_flags)
4093 static int io_fsync_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe)
4112 static int io_fsync(struct io_kiocb *req, unsigned int issue_flags)
4115 int ret;
4130 static int io_fallocate_prep(struct io_kiocb *req,
4145 static int io_fallocate(struct io_kiocb *req, unsigned int issue_flags)
4147 int ret;
4162 static int __io_openat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe)
4165 int ret;
4196 static int io_openat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe)
4205 static int io_openat2_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe)
4209 int ret;
4224 static int io_openat2(struct io_kiocb *req, unsigned int issue_flags)
4230 int ret;
4292 static int io_openat(struct io_kiocb *req, unsigned int issue_flags)
4297 static int io_remove_buffers_prep(struct io_kiocb *req,
4317 static int __io_remove_buffers(struct io_ring_ctx *ctx, struct io_buffer *buf,
4318 int bgid, unsigned nbufs)
4344 static int io_remove_buffers(struct io_kiocb *req, unsigned int issue_flags)
4349 int ret = 0;
4369 static int io_provide_buffers_prep(struct io_kiocb *req,
4404 static int io_add_buffers(struct io_provide_buf *pbuf, struct io_buffer **head)
4408 int i, bid = pbuf->bid;
4432 static int io_provide_buffers(struct io_kiocb *req, unsigned int issue_flags)
4437 int ret = 0;
4461 static int io_epoll_ctl_prep(struct io_kiocb *req,
4488 static int io_epoll_ctl(struct io_kiocb *req, unsigned int issue_flags)
4492 int ret;
4508 static int io_madvise_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe)
4525 static int io_madvise(struct io_kiocb *req, unsigned int issue_flags)
4529 int ret;
4544 static int io_fadvise_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe)
4557 static int io_fadvise(struct io_kiocb *req, unsigned int issue_flags)
4560 int ret;
4580 static int io_statx_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe)
4598 static int io_statx(struct io_kiocb *req, unsigned int issue_flags)
4601 int ret;
4615 static int io_close_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe)
4633 static int io_close(struct io_kiocb *req, unsigned int issue_flags)
4639 int ret = -EBADF;
4684 static int io_sfr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe)
4700 static int io_sync_file_range(struct io_kiocb *req, unsigned int issue_flags)
4702 int ret;
4717 static bool io_net_retry(struct socket *sock, int flags)
4724 static int io_setup_async_msg(struct io_kiocb *req,
4749 static int io_sendmsg_copy_hdr(struct io_kiocb *req,
4753 int ret;
4764 static int io_sendmsg_prep_async(struct io_kiocb *req)
4766 int ret;
4774 static int io_sendmsg_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe)
4799 static int io_sendmsg(struct io_kiocb *req, unsigned int issue_flags)
4805 int min_ret = 0;
4806 int ret;
4854 static int io_send(struct io_kiocb *req, unsigned int issue_flags)
4861 int min_ret = 0;
4862 int ret;
4907 static int __io_recvmsg_copy_hdr(struct io_kiocb *req,
4913 int ret;
4940 static int __io_compat_recvmsg_copy_hdr(struct io_kiocb *req,
4947 int ret;
4981 static int io_recvmsg_copy_hdr(struct io_kiocb *req,
5009 static inline unsigned int io_put_recv_kbuf(struct io_kiocb *req)
5014 static int io_recvmsg_prep_async(struct io_kiocb *req)
5016 int ret;
5024 static int io_recvmsg_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe)
5050 static int io_recvmsg(struct io_kiocb *req, unsigned int issue_flags)
5057 int min_ret = 0;
5058 int ret, cflags = 0;
5122 static int io_recv(struct io_kiocb *req, unsigned int issue_flags)
5131 int min_ret = 0;
5132 int ret, cflags = 0;
5191 static int io_accept_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe)
5215 static int io_accept(struct io_kiocb *req, unsigned int issue_flags)
5219 unsigned int file_flags = force_nonblock ? O_NONBLOCK : 0;
5222 int ret, fd;
5254 static int io_connect_prep_async(struct io_kiocb *req)
5262 static int io_connect_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe)
5277 static int io_connect(struct io_kiocb *req, unsigned int issue_flags)
5281 int ret;
5319 static int io_##op(struct io_kiocb *req, unsigned int issue_flags) \
5326 static int io_##op##_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) \
5333 static int io_##op##_prep_async(struct io_kiocb *req) \
5349 int nr_entries;
5350 int error;
5365 int v;
5478 static int io_poll_check_events(struct io_kiocb *req)
5482 int v;
5553 int ret;
5576 int ret;
5594 static void __io_poll_execute(struct io_kiocb *req, int mask)
5606 static inline void io_poll_execute(struct io_kiocb *req, int res)
5619 static int io_poll_wake(struct wait_queue_entry *wait, unsigned mode, int sync,
5722 static int __io_arm_poll_handler(struct io_kiocb *req,
5803 static int io_arm_poll_handler(struct io_kiocb *req)
5810 int ret;
5864 int i;
5911 static int io_poll_cancel(struct io_ring_ctx *ctx, __u64 sqe_addr,
5924 unsigned int flags)
5937 static int io_poll_update_prep(struct io_kiocb *req,
5970 static int io_poll_add_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe)
5988 static int io_poll_add(struct io_kiocb *req, unsigned int issue_flags)
5992 int ret;
6005 static int io_poll_update(struct io_kiocb *req, unsigned int issue_flags)
6009 int ret2, ret = 0;
6096 static int io_timeout_cancel(struct io_ring_ctx *ctx, __u64 user_data)
6127 static int io_linked_timeout_update(struct io_ring_ctx *ctx, __u64 user_data,
6152 static int io_timeout_update(struct io_ring_ctx *ctx, __u64 user_data,
6171 static int io_timeout_remove_prep(struct io_kiocb *req,
6203 static inline enum hrtimer_mode io_translate_timeout_mode(unsigned int flags)
6212 static int io_timeout_remove(struct io_kiocb *req, unsigned int issue_flags)
6216 int ret;
6241 static int io_timeout_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe,
6294 static int io_timeout(struct io_kiocb *req, unsigned int issue_flags)
6357 static int io_async_cancel_one(struct io_uring_task *tctx, u64 user_data,
6362 int ret = 0;
6383 static int io_try_cancel_userdata(struct io_kiocb *req, u64 sqe_addr)
6386 int ret;
6406 static int io_async_cancel_prep(struct io_kiocb *req,
6421 static int io_async_cancel(struct io_kiocb *req, unsigned int issue_flags)
6426 int ret;
6450 static int io_rsrc_update_prep(struct io_kiocb *req,
6466 static int io_files_update(struct io_kiocb *req, unsigned int issue_flags)
6470 int ret;
6490 static int io_req_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe)
6568 static int io_req_prep_async(struct io_kiocb *req)
6609 int ret;
6745 static int io_issue_sqe(struct io_kiocb *req, unsigned int issue_flags)
6749 int ret;
6883 int ret = 0;
6933 int index)
6954 struct io_kiocb *req, int fd,
6955 unsigned int issue_flags)
6962 if (unlikely((unsigned int)fd >= ctx->nr_user_files))
6977 struct io_kiocb *req, int fd)
6990 struct io_kiocb *req, int fd, bool fixed,
6991 unsigned int issue_flags)
7002 int ret = -ENOENT;
7070 int ret;
7128 int ret = io_req_prep_async(req);
7144 unsigned int sqe_flags)
7163 static int io_init_req(struct io_ring_ctx *ctx, struct io_kiocb *req,
7168 unsigned int sqe_flags;
7169 int personality, ret = 0;
7226 static int io_submit_sqe(struct io_ring_ctx *ctx, struct io_kiocb *req,
7231 int ret;
7326 unsigned int max_ios)
7378 static int io_submit_sqes(struct io_ring_ctx *ctx, unsigned int nr)
7381 int submitted = 0;
7412 int ref_used = (submitted == -EAGAIN) ? 0 : submitted;
7413 int unused = nr - ref_used;
7448 static int __io_sq_thread(struct io_ring_ctx *ctx, bool cap_entries)
7450 unsigned int to_submit;
7451 int ret = 0;
7513 static int io_sq_thread(void *data)
7542 int ret = __io_sq_thread(ctx, cap_entries);
7609 int dist = ctx->cached_cq_tail - (int) iowq->cq_tail;
7619 static int io_wake_function(struct wait_queue_entry *curr, unsigned int mode,
7620 int wake_flags, void *key)
7634 static int io_run_task_work_sig(void)
7655 static inline int io_cqring_wait_schedule(struct io_ring_ctx *ctx,
7659 int io_wait, ret;
7688 static int io_cqring_wait(struct io_ring_ctx *ctx, int min_events,
7695 int ret;
7772 unsigned int this_size = min_t(size_t, size, PAGE_SIZE);
7864 static int io_rsrc_node_switch_start(struct io_ring_ctx *ctx)
7872 static int io_rsrc_ref_quiesce(struct io_rsrc_data *data, struct io_ring_ctx *ctx)
7874 int ret;
7919 static u64 *io_get_tag_slot(struct io_rsrc_data *data, unsigned int idx)
7921 unsigned int off = idx & IO_RSRC_TAG_TABLE_MASK;
7922 unsigned int table_idx = idx >> IO_RSRC_TAG_TABLE_SHIFT;
7936 static int io_rsrc_data_alloc(struct io_ring_ctx *ctx, rsrc_put_fn *do_put,
7941 int ret = -ENOMEM;
8000 int i;
8016 static int io_sqe_files_unregister(struct io_ring_ctx *ctx)
8019 int ret;
8167 static int __io_sqe_files_scm(struct io_ring_ctx *ctx, int nr, int offset)
8172 int i, nr_files;
8227 static int io_sqe_files_scm(struct io_ring_ctx *ctx)
8230 int ret = 0;
8258 static int io_sqe_files_scm(struct io_ring_ctx *ctx)
8271 int i;
8285 int left;
8375 static int io_sqe_files_register(struct io_ring_ctx *ctx, void __user *arg,
8380 int fd, ret;
8457 static int io_queue_rsrc_removal(struct io_rsrc_data *data, unsigned idx,
8474 static int io_install_fixed_file(struct io_kiocb *req, struct file *file,
8475 unsigned int issue_flags, u32 slot_index)
8481 int ret = -EBADF;
8524 static int io_close_fixed(struct io_kiocb *req, unsigned int issue_flags)
8526 unsigned int offset = req->close.file_slot - 1;
8530 int ret;
8562 static int __io_sqe_files_update(struct io_ring_ctx *ctx,
8571 int fd, i, err = 0;
8572 unsigned int done;
8640 unsigned int concurrency;
8667 static int io_uring_alloc_task_context(struct task_struct *task,
8671 int ret;
8715 static int io_sq_offload_create(struct io_ring_ctx *ctx,
8718 int ret;
8764 int cpu = p->sq_thread_cpu;
8807 static inline int __io_account_mem(struct user_struct *user,
8835 static int io_account_mem(struct io_ring_ctx *ctx, unsigned long nr_pages)
8837 int ret;
8902 unsigned int i;
8922 unsigned int i;
8933 static int io_sqe_buffers_unregister(struct io_ring_ctx *ctx)
8936 int ret;
8953 static int io_copy_iov(struct io_ring_ctx *ctx, struct iovec *dst,
8988 int nr_pages, struct page *hpage)
8990 int i, j;
9015 static int io_buffer_account_pin(struct io_ring_ctx *ctx, struct page **pages,
9016 int nr_pages, struct io_mapped_ubuf *imu,
9019 int i, ret;
9047 static int io_sqe_buffer_register(struct io_ring_ctx *ctx, struct iovec *iov,
9056 int ret, pret, nr_pages, i;
9150 static int io_buffers_map_alloc(struct io_ring_ctx *ctx, unsigned int nr_args)
9156 static int io_buffer_validate(struct iovec *iov)
9180 static int io_sqe_buffers_register(struct io_ring_ctx *ctx, void __user *arg,
9181 unsigned int nr_args, u64 __user *tags)
9185 int i, ret;
9232 static int __io_sqe_buffers_update(struct io_ring_ctx *ctx,
9234 unsigned int nr_args)
9241 int i, err;
9250 int offset = up->offset + done;
9292 static int io_eventfd_register(struct io_ring_ctx *ctx, void __user *arg)
9295 int fd;
9305 int ret = PTR_ERR(ctx->cq_ev_fd);
9314 static int io_eventfd_unregister(struct io_ring_ctx *ctx)
9456 static int io_unregister_personality(struct io_ring_ctx *ctx, unsigned id)
9506 int ret;
9588 int canceled = 0;
9640 static int io_uring_release(struct inode *inode, struct file *file)
9756 static int __io_uring_add_tctx_node(struct io_ring_ctx *ctx)
9760 int ret;
9769 unsigned int limits[2] = { ctx->iowq_limits[0],
9802 static inline int io_uring_add_tctx_node(struct io_ring_ctx *ctx)
9967 static int io_uring_mmap(struct file *file, struct vm_area_struct *vma)
9983 static int io_uring_mmap(struct file *file, struct vm_area_struct *vma)
9988 static unsigned int io_uring_nommu_mmap_capabilities(struct file *file)
10008 static int io_sqpoll_wait_sq(struct io_ring_ctx *ctx)
10026 static int io_get_ext_arg(unsigned flags, const void __user *argp, size_t *argsz,
10058 SYSCALL_DEFINE6(io_uring_enter, unsigned int, fd, u32, to_submit,
10063 int submitted = 0;
10154 static int io_uring_show_cred(struct seq_file *m, unsigned int id,
10161 int g;
10188 int sq_pid = -1, sq_cpu = -1;
10190 int i;
10226 unsigned int len = buf->ubuf_end - buf->ubuf;
10277 static int io_allocate_scq_urings(struct io_ring_ctx *ctx,
10319 static int io_uring_install_fd(struct io_ring_ctx *ctx, struct file *file)
10321 int ret, fd;
10346 int ret;
10367 static int io_uring_create(unsigned entries, struct io_uring_params *p,
10372 int ret;
10502 int i;
10526 static int io_probe(struct io_ring_ctx *ctx, void __user *arg, unsigned nr_args)
10530 int i, ret;
10565 static int io_register_personality(struct io_ring_ctx *ctx)
10569 int ret;
10582 static int io_register_restrictions(struct io_ring_ctx *ctx, void __user *arg,
10583 unsigned int nr_args)
10587 int i, ret;
10652 static int io_register_enable_rings(struct io_ring_ctx *ctx)
10666 static int __io_register_rsrc_update(struct io_ring_ctx *ctx, unsigned type,
10671 int err;
10688 static int io_register_files_update(struct io_ring_ctx *ctx, void __user *arg,
10703 static int io_register_rsrc_update(struct io_ring_ctx *ctx, void __user *arg,
10717 static int io_register_rsrc(struct io_ring_ctx *ctx, void __user *arg,
10718 unsigned int size, unsigned int type)
10743 static int io_register_iowq_aff(struct io_ring_ctx *ctx, void __user *arg,
10748 int ret;
10782 static int io_unregister_iowq_aff(struct io_ring_ctx *ctx)
10792 static int io_register_iowq_max_workers(struct io_ring_ctx *ctx,
10800 int i, ret;
10876 static bool io_register_op_must_quiesce(int op)
10900 static int io_ctx_quiesce(struct io_ring_ctx *ctx)
10927 static int __io_uring_register(struct io_ring_ctx *ctx, unsigned opcode,
10932 int ret;
11067 SYSCALL_DEFINE4(io_uring_register, unsigned int, fd, unsigned int, opcode,
11068 void __user *, arg, unsigned int, nr_args)
11099 static int __init io_uring_init(void)
11119 BUILD_BUG_SQE_ELEM(28, /* compat */ int, rw_flags);
11152 BUILD_BUG_ON(__REQ_F_LAST_BIT > 8 * sizeof(int));