/kernel/linux/linux-5.10/tools/io_uring/ |
H A D | setup.c | 41 cq->ring_sz = p->cq_off.cqes + p->cq_entries * sizeof(struct io_uring_cqe); in io_uring_mmap() 54 cq->cqes = ptr + p->cq_off.cqes; in io_uring_mmap()
|
H A D | io_uring-bench.c | 51 struct io_uring_cqe *cqes; member 257 cqe = &ring->cqes[head & cq_ring_mask]; in reap_events() 449 ptr = mmap(0, p.cq_off.cqes + p.cq_entries * sizeof(struct io_uring_cqe), in setup_ring() 457 cring->cqes = ptr + p.cq_off.cqes; in setup_ring()
|
H A D | queue.c | 31 *cqe_ptr = &cq->cqes[head & mask]; in __io_uring_get_cqe()
|
H A D | liburing.h | 41 struct io_uring_cqe *cqes; member
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/cxgb4/ |
H A D | restrack.c | 334 struct t4_cqe *cqes) in fill_hwcqes() 339 if (fill_cqe(msg, cqes, idx, "hwcq_idx")) in fill_hwcqes() 342 if (fill_cqe(msg, cqes + 1, idx, "hwcq_idx")) in fill_hwcqes() 351 struct t4_cqe *cqes) in fill_swcqes() 359 if (fill_cqe(msg, cqes, idx, "swcq_idx")) in fill_swcqes() 364 if (fill_cqe(msg, cqes + 1, idx, "swcq_idx")) in fill_swcqes() 395 /* get 2 hw cqes: cidx-1, and cidx */ in c4iw_fill_res_cq_entry() 402 /* get first and last sw cqes */ in c4iw_fill_res_cq_entry() 333 fill_hwcqes(struct sk_buff *msg, struct t4_cq *cq, struct t4_cqe *cqes) fill_hwcqes() argument 350 fill_swcqes(struct sk_buff *msg, struct t4_cq *cq, struct t4_cqe *cqes) fill_swcqes() argument
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/cxgb4/ |
H A D | restrack.c | 334 struct t4_cqe *cqes) in fill_hwcqes() 339 if (fill_cqe(msg, cqes, idx, "hwcq_idx")) in fill_hwcqes() 342 if (fill_cqe(msg, cqes + 1, idx, "hwcq_idx")) in fill_hwcqes() 351 struct t4_cqe *cqes) in fill_swcqes() 359 if (fill_cqe(msg, cqes, idx, "swcq_idx")) in fill_swcqes() 364 if (fill_cqe(msg, cqes + 1, idx, "swcq_idx")) in fill_swcqes() 395 /* get 2 hw cqes: cidx-1, and cidx */ in c4iw_fill_res_cq_entry() 402 /* get first and last sw cqes */ in c4iw_fill_res_cq_entry() 333 fill_hwcqes(struct sk_buff *msg, struct t4_cq *cq, struct t4_cqe *cqes) fill_hwcqes() argument 350 fill_swcqes(struct sk_buff *msg, struct t4_cq *cq, struct t4_cqe *cqes) fill_swcqes() argument
|
/kernel/linux/linux-6.6/tools/testing/selftests/net/ |
H A D | io_uring_zerocopy_tx.c | 77 struct io_uring_cqe *cqes; member 102 struct io_uring_cqe *cqes; member 198 cq->ring_sz = p->cq_off.cqes + p->cq_entries * sizeof(struct io_uring_cqe); in io_uring_mmap() 211 cq->cqes = ptr + p->cq_off.cqes; in io_uring_mmap() 311 *cqe_ptr = &cq->cqes[head & mask]; in io_uring_wait_cqe()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/fungible/funeth/ |
H A D | funeth_rx.c | 353 q->next_cqe_info = cqe_to_info(q->cqes); in advance_cq() 644 q->cqes = fun_alloc_ring_mem(q->dma_dev, ncqe, FUNETH_CQE_SIZE, 0, in fun_rxq_create_sw() 647 if (!q->cqes) in fun_rxq_create_sw() 665 dma_free_coherent(q->dma_dev, ncqe * FUNETH_CQE_SIZE, q->cqes, in fun_rxq_create_sw() 686 q->cqes, q->cq_dma_addr); in fun_rxq_free_sw() 722 q->next_cqe_info = cqe_to_info(q->cqes); in fun_rxq_create_dev()
|
H A D | funeth_txrx.h | 167 void *cqes; /* base of CQ descriptor ring */ member
|
/kernel/linux/linux-6.6/drivers/net/ethernet/fungible/funcore/ |
H A D | fun_queue.c | 295 cqe = funq->cqes + (funq->cq_head << funq->cqe_size_log2); in __fun_process_cq() 366 funq->cqes = fun_alloc_ring_mem(funq->fdev->dev, funq->cq_depth, in fun_alloc_cqes() 370 return funq->cqes ? 0 : -ENOMEM; in fun_alloc_cqes() 389 funq->cqes, funq->cq_dma_addr, NULL); in fun_free_queue()
|
H A D | fun_queue.h | 36 void *cqes; member
|
/kernel/linux/linux-6.6/tools/testing/selftests/x86/ |
H A D | lam.c | 96 struct io_uring_cqe *cqes; member 390 cring->ring_sz = p.cq_off.cqes + p.cq_entries * sizeof(struct io_uring_cqe); in mmap_io_uring() 445 cring->queue.cqes = cq_ptr + p.cq_off.cqes; in mmap_io_uring() 482 cqe = &cring->queue.cqes[head & *s->cq_ring.ring_mask]; in handle_uring_cq()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_stats.c | 289 s->rx_xdp_tx_cqe += xdpsq_stats->cqes; in MLX5E_DECLARE_STATS_GRP_OP_UPDATE_STATS() 331 s->tx_xdp_cqes += xdpsq_red_stats->cqes; in MLX5E_DECLARE_STATS_GRP_OP_UPDATE_STATS() 357 s->tx_xsk_cqes += xsksq_stats->cqes; in MLX5E_DECLARE_STATS_GRP_OP_UPDATE_STATS() 393 s->tx_cqes += sq_stats->cqes; in MLX5E_DECLARE_STATS_GRP_OP_UPDATE_STATS() 1594 { MLX5E_DECLARE_TX_STAT(struct mlx5e_sq_stats, cqes) }, 1606 { MLX5E_DECLARE_RQ_XDPSQ_STAT(struct mlx5e_xdpsq_stats, cqes) }, 1616 { MLX5E_DECLARE_XDPSQ_STAT(struct mlx5e_xdpsq_stats, cqes) }, 1647 { MLX5E_DECLARE_XSKSQ_STAT(struct mlx5e_xdpsq_stats, cqes) },
|
H A D | en_stats.h | 376 u64 cqes ____cacheline_aligned_in_smp; 389 u64 cqes ____cacheline_aligned_in_smp;
|
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_stats.c | 282 s->tx_xdp_cqes += xdpsq_red_stats->cqes; in mlx5e_stats_grp_sw_update_stats_xdp_red() 294 s->rx_xdp_tx_cqe += xdpsq_stats->cqes; in mlx5e_stats_grp_sw_update_stats_xdpsq() 305 s->tx_xsk_cqes += xsksq_stats->cqes; in mlx5e_stats_grp_sw_update_stats_xsksq() 440 s->tx_cqes += sq_stats->cqes; in mlx5e_stats_grp_sw_update_stats_sq() 2067 { MLX5E_DECLARE_TX_STAT(struct mlx5e_sq_stats, cqes) }, 2079 { MLX5E_DECLARE_RQ_XDPSQ_STAT(struct mlx5e_xdpsq_stats, cqes) }, 2089 { MLX5E_DECLARE_XDPSQ_STAT(struct mlx5e_xdpsq_stats, cqes) }, 2119 { MLX5E_DECLARE_XSKSQ_STAT(struct mlx5e_xdpsq_stats, cqes) }, 2143 { MLX5E_DECLARE_PTP_TX_STAT(struct mlx5e_sq_stats, cqes) }, 2218 { MLX5E_DECLARE_QOS_TX_STAT(struct mlx5e_sq_stats, cqes) }, [all...] |
H A D | en_stats.h | 433 u64 cqes ____cacheline_aligned_in_smp; 446 u64 cqes ____cacheline_aligned_in_smp;
|
/kernel/linux/patches/linux-5.10/prebuilts/usr/include/linux/ |
H A D | io_uring.h | 157 __u32 cqes; member
|
/kernel/linux/patches/linux-6.6/prebuilts/usr/include/linux/ |
H A D | io_uring.h | 157 __u32 cqes; member
|
/kernel/linux/linux-5.10/drivers/net/ethernet/broadcom/ |
H A D | cnic.c | 1425 struct kcqe *cqes[], u32 num_cqes) in cnic_reply_bnx2x_kcqes() 1434 cqes, num_cqes); in cnic_reply_bnx2x_kcqes() 1552 struct kcqe *cqes[1]; in cnic_bnx2x_iscsi_init2() local 1582 cqes[0] = (struct kcqe *) &kcqe; in cnic_bnx2x_iscsi_init2() 1583 cnic_reply_bnx2x_kcqes(dev, CNIC_ULP_ISCSI, cqes, 1); in cnic_bnx2x_iscsi_init2() 1881 struct kcqe *cqes[1]; in cnic_bnx2x_iscsi_ofld1() local 1934 cqes[0] = (struct kcqe *) &kcqe; in cnic_bnx2x_iscsi_ofld1() 1935 cnic_reply_bnx2x_kcqes(dev, CNIC_ULP_ISCSI, cqes, 1); in cnic_bnx2x_iscsi_ofld1() 1999 struct kcqe *cqes[1]; in cnic_bnx2x_iscsi_destroy() local 2032 cqes[ in cnic_bnx2x_iscsi_destroy() 1424 cnic_reply_bnx2x_kcqes(struct cnic_dev *dev, int ulp_type, struct kcqe *cqes[], u32 num_cqes) cnic_reply_bnx2x_kcqes() argument 2228 struct kcqe *cqes[1]; cnic_bnx2x_offload_pg() local 2243 struct kcqe *cqes[1]; cnic_bnx2x_update_pg() local 2353 struct kcqe *cqes[1]; cnic_bnx2x_fcoe_ofld1() local 2500 struct kcqe *cqes[1]; cnic_bnx2x_fcoe_destroy() local 2584 struct kcqe *cqes[1]; cnic_bnx2x_kwqe_err() local [all...] |
H A D | cnic_if.h | 369 void (*indicate_kcqes)(void *ulp_ctx, struct kcqe *cqes[],
|
/kernel/linux/linux-6.6/drivers/net/ethernet/broadcom/ |
H A D | cnic.c | 1423 struct kcqe *cqes[], u32 num_cqes) in cnic_reply_bnx2x_kcqes() 1432 cqes, num_cqes); in cnic_reply_bnx2x_kcqes() 1550 struct kcqe *cqes[1]; in cnic_bnx2x_iscsi_init2() local 1580 cqes[0] = (struct kcqe *) &kcqe; in cnic_bnx2x_iscsi_init2() 1581 cnic_reply_bnx2x_kcqes(dev, CNIC_ULP_ISCSI, cqes, 1); in cnic_bnx2x_iscsi_init2() 1879 struct kcqe *cqes[1]; in cnic_bnx2x_iscsi_ofld1() local 1932 cqes[0] = (struct kcqe *) &kcqe; in cnic_bnx2x_iscsi_ofld1() 1933 cnic_reply_bnx2x_kcqes(dev, CNIC_ULP_ISCSI, cqes, 1); in cnic_bnx2x_iscsi_ofld1() 1997 struct kcqe *cqes[1]; in cnic_bnx2x_iscsi_destroy() local 2030 cqes[ in cnic_bnx2x_iscsi_destroy() 1422 cnic_reply_bnx2x_kcqes(struct cnic_dev *dev, int ulp_type, struct kcqe *cqes[], u32 num_cqes) cnic_reply_bnx2x_kcqes() argument 2226 struct kcqe *cqes[1]; cnic_bnx2x_offload_pg() local 2241 struct kcqe *cqes[1]; cnic_bnx2x_update_pg() local 2351 struct kcqe *cqes[1]; cnic_bnx2x_fcoe_ofld1() local 2498 struct kcqe *cqes[1]; cnic_bnx2x_fcoe_destroy() local 2582 struct kcqe *cqes[1]; cnic_bnx2x_kwqe_err() local [all...] |
/kernel/linux/linux-5.10/drivers/nvme/target/ |
H A D | passthru.c | 77 id->cqes = min_t(__u8, ((0x4 << 4) | 0x4), id->cqes); in nvmet_passthru_override_id_ctrl()
|
/kernel/linux/linux-5.10/include/uapi/linux/ |
H A D | io_uring.h | 237 __u32 cqes; member
|
/kernel/linux/linux-6.6/drivers/nvme/host/ |
H A D | apple.c | 135 struct nvme_completion *cqes; member 571 struct nvme_completion *hcqe = &q->cqes[q->cq_head]; in apple_nvme_cqe_pending() 589 struct nvme_completion *cqe = &q->cqes[idx]; in apple_nvme_handle_cqe() 976 memset(q->cqes, 0, depth * sizeof(struct nvme_completion)); in apple_nvme_init_queue() 1299 q->cqes = dmam_alloc_coherent(anv->dev, in apple_nvme_queue_alloc() 1302 if (!q->cqes) in apple_nvme_queue_alloc()
|
/kernel/linux/linux-6.6/drivers/nvme/target/ |
H A D | passthru.c | 134 id->cqes = min_t(__u8, ((0x4 << 4) | 0x4), id->cqes); in nvmet_passthru_override_id_ctrl()
|