/kernel/linux/linux-5.10/drivers/infiniband/hw/hns/ |
H A D | hns_roce_cq.c | 251 u32 cq_entries = attr->cqe; in hns_roce_create_cq() local 254 if (cq_entries < 1 || cq_entries > hr_dev->caps.max_cqes) { in hns_roce_create_cq() 256 cq_entries, hr_dev->caps.max_cqes); in hns_roce_create_cq() 266 cq_entries = max(cq_entries, hr_dev->caps.min_cqes); in hns_roce_create_cq() 267 cq_entries = roundup_pow_of_two(cq_entries); in hns_roce_create_cq() 268 hr_cq->ib_cq.cqe = cq_entries - 1; /* used as cqe index */ in hns_roce_create_cq() 269 hr_cq->cq_depth = cq_entries; in hns_roce_create_cq() [all...] |
/kernel/linux/linux-6.6/io_uring/ |
H A D | fdinfo.c | 58 unsigned int sq_mask = ctx->sq_entries - 1, cq_mask = ctx->cq_entries - 1; in io_uring_show_fdinfo() 65 unsigned int sq_entries, cq_entries; in io_uring_show_fdinfo() local 123 cq_entries = min(cq_tail - cq_head, ctx->cq_entries); in io_uring_show_fdinfo() 124 for (i = 0; i < cq_entries; i++) { in io_uring_show_fdinfo()
|
H A D | io_uring.c | 299 hash_bits = ilog2(p->cq_entries) - 5; in io_ring_ctx_alloc() 686 if (__io_cqring_events(ctx) == ctx->cq_entries) in __io_cqring_overflow_flush() 832 unsigned int off = ctx->cached_cq_tail & (ctx->cq_entries - 1); in io_cqe_cache_refill() 844 queued = min(__io_cqring_events(ctx), ctx->cq_entries); in io_cqe_cache_refill() 845 free = ctx->cq_entries - queued; in io_cqe_cache_refill() 847 len = min(free, ctx->cq_entries - off); in io_cqe_cache_refill() 2775 unsigned int cq_entries, size_t *sq_offset) in rings_size() 2780 off = struct_size(rings, cqes, cq_entries); in rings_size() 3682 ctx->cq_entries); in SYSCALL_DEFINE6() 3693 ctx->cq_entries); in SYSCALL_DEFINE6() 2774 rings_size(struct io_ring_ctx *ctx, unsigned int sq_entries, unsigned int cq_entries, size_t *sq_offset) rings_size() argument [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/hns/ |
H A D | hns_roce_cq.c | 314 static void set_cq_param(struct hns_roce_cq *hr_cq, u32 cq_entries, int vector, in set_cq_param() argument 319 cq_entries = max(cq_entries, hr_dev->caps.min_cqes); in set_cq_param() 320 cq_entries = roundup_pow_of_two(cq_entries); in set_cq_param() 321 hr_cq->ib_cq.cqe = cq_entries - 1; /* used as cqe index */ in set_cq_param() 322 hr_cq->cq_depth = cq_entries; in set_cq_param()
|
/kernel/linux/linux-5.10/include/trace/events/ |
H A D | io_uring.h | 18 * @cq_entries: actual CQ size 26 TP_PROTO(int fd, void *ctx, u32 sq_entries, u32 cq_entries, u32 flags), 28 TP_ARGS(fd, ctx, sq_entries, cq_entries, flags), 34 __field( u32, cq_entries ) 42 __entry->cq_entries = cq_entries; 48 __entry->cq_entries, __entry->flags)
|
/kernel/linux/linux-6.6/include/trace/events/ |
H A D | io_uring.h | 21 * @cq_entries: actual CQ size 29 TP_PROTO(int fd, void *ctx, u32 sq_entries, u32 cq_entries, u32 flags), 31 TP_ARGS(fd, ctx, sq_entries, cq_entries, flags), 37 __field( u32, cq_entries ) 45 __entry->cq_entries = cq_entries; 51 __entry->cq_entries, __entry->flags)
|
/kernel/linux/linux-5.10/tools/io_uring/ |
H A D | setup.c | 41 cq->ring_sz = p->cq_off.cqes + p->cq_entries * sizeof(struct io_uring_cqe); in io_uring_mmap()
|
H A D | io_uring-bench.c | 449 ptr = mmap(0, p.cq_off.cqes + p.cq_entries * sizeof(struct io_uring_cqe), in setup_ring()
|
/kernel/linux/patches/linux-5.10/prebuilts/usr/include/linux/ |
H A D | io_uring.h | 168 __u32 cq_entries; member
|
/kernel/linux/patches/linux-6.6/prebuilts/usr/include/linux/ |
H A D | io_uring.h | 168 __u32 cq_entries; member
|
/kernel/linux/linux-5.10/include/uapi/linux/ |
H A D | io_uring.h | 263 __u32 cq_entries; member
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | io_uring_types.h | 279 unsigned cq_entries; member
|
/kernel/linux/linux-6.6/include/uapi/linux/ |
H A D | io_uring.h | 475 __u32 cq_entries; member
|
/kernel/linux/linux-5.10/io_uring/ |
H A D | io_uring.c | 400 unsigned cq_entries; member 1317 hash_bits = ilog2(p->cq_entries); in io_ring_ctx_alloc() 1610 unsigned tail, mask = ctx->cq_entries - 1; in io_get_cqe() 1617 if (__io_cqring_events(ctx) == ctx->cq_entries) in io_get_cqe() 1681 if (!force && __io_cqring_events(ctx) == ctx->cq_entries) in __io_cqring_overflow_flush() 8870 static unsigned long rings_size(unsigned sq_entries, unsigned cq_entries, in rings_size() argument 8876 off = struct_size(rings, cqes, cq_entries); in rings_size() 10130 min_complete = min(min_complete, ctx->cq_entries); in SYSCALL_DEFINE6() 10285 ctx->cq_entries = p->cq_entries; in io_allocate_scq_urings() [all...] |
/kernel/linux/linux-6.6/tools/testing/selftests/net/ |
H A D | io_uring_zerocopy_tx.c | 198 cq->ring_sz = p->cq_off.cqes + p->cq_entries * sizeof(struct io_uring_cqe); in io_uring_mmap()
|
/kernel/linux/linux-6.6/tools/testing/selftests/x86/ |
H A D | lam.c | 390 cring->ring_sz = p.cq_off.cqes + p.cq_entries * sizeof(struct io_uring_cqe); in mmap_io_uring()
|