/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx4/ |
H A D | en_cq.c | 34 #include <linux/mlx4/cq.h> 40 static void mlx4_en_cq_event(struct mlx4_cq *cq, enum mlx4_event event) in mlx4_en_cq_event() argument 52 struct mlx4_en_cq *cq; in mlx4_en_create_cq() local 55 cq = kzalloc_node(sizeof(*cq), GFP_KERNEL, node); in mlx4_en_create_cq() 56 if (!cq) { in mlx4_en_create_cq() 61 cq->size = entries; in mlx4_en_create_cq() 62 cq->buf_size = cq->size * mdev->dev->caps.cqe_size; in mlx4_en_create_cq() 64 cq in mlx4_en_create_cq() 89 mlx4_en_activate_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq, int cq_idx) mlx4_en_activate_cq() argument 179 struct mlx4_en_cq *cq = *pcq; mlx4_en_destroy_cq() local 192 mlx4_en_deactivate_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq) mlx4_en_deactivate_cq() argument 203 mlx4_en_set_cq_moder(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq) mlx4_en_set_cq_moder() argument 209 mlx4_en_arm_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq) mlx4_en_arm_cq() argument [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx4/ |
H A D | en_cq.c | 34 #include <linux/mlx4/cq.h> 40 static void mlx4_en_cq_event(struct mlx4_cq *cq, enum mlx4_event event) in mlx4_en_cq_event() argument 52 struct mlx4_en_cq *cq; in mlx4_en_create_cq() local 55 cq = kzalloc_node(sizeof(*cq), GFP_KERNEL, node); in mlx4_en_create_cq() 56 if (!cq) { in mlx4_en_create_cq() 61 cq->size = entries; in mlx4_en_create_cq() 62 cq->buf_size = cq->size * mdev->dev->caps.cqe_size; in mlx4_en_create_cq() 64 cq in mlx4_en_create_cq() 89 mlx4_en_activate_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq, int cq_idx) mlx4_en_activate_cq() argument 176 struct mlx4_en_cq *cq = *pcq; mlx4_en_destroy_cq() local 189 mlx4_en_deactivate_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq) mlx4_en_deactivate_cq() argument 200 mlx4_en_set_cq_moder(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq) mlx4_en_set_cq_moder() argument 206 mlx4_en_arm_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq) mlx4_en_arm_cq() argument [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/core/ |
H A D | cq.c | 42 struct ib_cq *cq = dim->priv; in ib_cq_rdma_dim_work() local 49 trace_cq_modify(cq, comps, usec); in ib_cq_rdma_dim_work() 50 cq->device->ops.modify_cq(cq, comps, usec); in ib_cq_rdma_dim_work() 53 static void rdma_dim_init(struct ib_cq *cq) in rdma_dim_init() argument 57 if (!cq->device->ops.modify_cq || !cq->device->use_cq_dim || in rdma_dim_init() 58 cq->poll_ctx == IB_POLL_DIRECT) in rdma_dim_init() 68 dim->priv = cq; in rdma_dim_init() 69 cq in rdma_dim_init() 74 rdma_dim_destroy(struct ib_cq *cq) rdma_dim_destroy() argument 83 __poll_cq(struct ib_cq *cq, int num_entries, struct ib_wc *wc) __poll_cq() argument 92 __ib_process_cq(struct ib_cq *cq, int budget, struct ib_wc *wcs, int batch) __ib_process_cq() argument 138 ib_process_cq_direct(struct ib_cq *cq, int budget) ib_process_cq_direct() argument 146 ib_cq_completion_direct(struct ib_cq *cq, void *private) ib_cq_completion_direct() argument 153 struct ib_cq *cq = container_of(iop, struct ib_cq, iop); ib_poll_handler() local 172 ib_cq_completion_softirq(struct ib_cq *cq, void *private) ib_cq_completion_softirq() argument 180 struct ib_cq *cq = container_of(work, struct ib_cq, work); ib_cq_poll_work() local 192 ib_cq_completion_workqueue(struct ib_cq *cq, void *private) ib_cq_completion_workqueue() argument 220 struct ib_cq *cq; __ib_alloc_cq() local 318 ib_free_cq(struct ib_cq *cq) ib_free_cq() argument 353 struct ib_cq *cq, *n; ib_cq_pool_cleanup() local 372 struct ib_cq *cq, *n; ib_alloc_cqs() local 434 struct ib_cq *cq, *found = NULL; ib_cq_pool_get() local 498 ib_cq_pool_put(struct ib_cq *cq, unsigned int nr_cqe) ib_cq_pool_put() argument [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/core/ |
H A D | cq.c | 43 struct ib_cq *cq = dim->priv; in ib_cq_rdma_dim_work() local 50 trace_cq_modify(cq, comps, usec); in ib_cq_rdma_dim_work() 51 cq->device->ops.modify_cq(cq, comps, usec); in ib_cq_rdma_dim_work() 54 static void rdma_dim_init(struct ib_cq *cq) in rdma_dim_init() argument 58 if (!cq->device->ops.modify_cq || !cq->device->use_cq_dim || in rdma_dim_init() 59 cq->poll_ctx == IB_POLL_DIRECT) in rdma_dim_init() 69 dim->priv = cq; in rdma_dim_init() 70 cq in rdma_dim_init() 75 rdma_dim_destroy(struct ib_cq *cq) rdma_dim_destroy() argument 84 __poll_cq(struct ib_cq *cq, int num_entries, struct ib_wc *wc) __poll_cq() argument 93 __ib_process_cq(struct ib_cq *cq, int budget, struct ib_wc *wcs, int batch) __ib_process_cq() argument 139 ib_process_cq_direct(struct ib_cq *cq, int budget) ib_process_cq_direct() argument 147 ib_cq_completion_direct(struct ib_cq *cq, void *private) ib_cq_completion_direct() argument 154 struct ib_cq *cq = container_of(iop, struct ib_cq, iop); ib_poll_handler() local 173 ib_cq_completion_softirq(struct ib_cq *cq, void *private) ib_cq_completion_softirq() argument 181 struct ib_cq *cq = container_of(work, struct ib_cq, work); ib_cq_poll_work() local 193 ib_cq_completion_workqueue(struct ib_cq *cq, void *private) ib_cq_completion_workqueue() argument 221 struct ib_cq *cq; __ib_alloc_cq() local 319 ib_free_cq(struct ib_cq *cq) ib_free_cq() argument 363 struct ib_cq *cq, *n; ib_cq_pool_destroy() local 381 struct ib_cq *cq, *n; ib_alloc_cqs() local 443 struct ib_cq *cq, *found = NULL; ib_cq_pool_get() local 507 ib_cq_pool_put(struct ib_cq *cq, unsigned int nr_cqe) ib_cq_pool_put() argument [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/sw/rxe/ |
H A D | rxe_cq.c | 11 int rxe_cq_chk_attr(struct rxe_dev *rxe, struct rxe_cq *cq, in rxe_cq_chk_attr() argument 27 if (cq) { in rxe_cq_chk_attr() 28 count = queue_count(cq->queue); in rxe_cq_chk_attr() 44 struct rxe_cq *cq = from_tasklet(cq, t, comp_task); in rxe_send_complete() local 47 spin_lock_irqsave(&cq->cq_lock, flags); in rxe_send_complete() 48 if (cq->is_dying) { in rxe_send_complete() 49 spin_unlock_irqrestore(&cq->cq_lock, flags); in rxe_send_complete() 52 spin_unlock_irqrestore(&cq->cq_lock, flags); in rxe_send_complete() 54 cq in rxe_send_complete() 57 rxe_cq_from_init(struct rxe_dev *rxe, struct rxe_cq *cq, int cqe, int comp_vector, struct ib_udata *udata, struct rxe_create_cq_resp __user *uresp) rxe_cq_from_init() argument 90 rxe_cq_resize_queue(struct rxe_cq *cq, int cqe, struct rxe_resize_cq_resp __user *uresp, struct ib_udata *udata) rxe_cq_resize_queue() argument 105 rxe_cq_post(struct rxe_cq *cq, struct rxe_cqe *cqe, int solicited) rxe_cq_post() argument 143 rxe_cq_disable(struct rxe_cq *cq) rxe_cq_disable() argument 154 struct rxe_cq *cq = container_of(arg, typeof(*cq), pelem); rxe_cq_cleanup() local [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/sw/rxe/ |
H A D | rxe_cq.c | 11 int rxe_cq_chk_attr(struct rxe_dev *rxe, struct rxe_cq *cq, in rxe_cq_chk_attr() argument 27 if (cq) { in rxe_cq_chk_attr() 28 count = queue_count(cq->queue, QUEUE_TYPE_TO_CLIENT); in rxe_cq_chk_attr() 30 rxe_dbg_cq(cq, "cqe(%d) < current # elements in queue (%d)", in rxe_cq_chk_attr() 42 int rxe_cq_from_init(struct rxe_dev *rxe, struct rxe_cq *cq, int cqe, in rxe_cq_from_init() argument 50 cq->queue = rxe_queue_init(rxe, &cqe, in rxe_cq_from_init() 52 if (!cq->queue) { in rxe_cq_from_init() 53 rxe_dbg_dev(rxe, "unable to create cq\n"); in rxe_cq_from_init() 58 cq->queue->buf, cq in rxe_cq_from_init() 72 rxe_cq_resize_queue(struct rxe_cq *cq, int cqe, struct rxe_resize_cq_resp __user *uresp, struct ib_udata *udata) rxe_cq_resize_queue() argument 88 rxe_cq_post(struct rxe_cq *cq, struct rxe_cqe *cqe, int solicited) rxe_cq_post() argument 129 struct rxe_cq *cq = container_of(elem, typeof(*cq), elem); rxe_cq_cleanup() local [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/hw/mthca/ |
H A D | mthca_cq.c | 169 static inline struct mthca_cqe *get_cqe(struct mthca_cq *cq, int entry) in get_cqe() argument 171 return get_cqe_from_buf(&cq->buf, entry); in get_cqe() 179 static inline struct mthca_cqe *next_cqe_sw(struct mthca_cq *cq) in next_cqe_sw() argument 181 return cqe_sw(get_cqe(cq, cq->cons_index & cq->ibcq.cqe)); in next_cqe_sw() 201 * incr is ignored in native Arbel (mem-free) mode, so cq->cons_index 204 static inline void update_cons_index(struct mthca_dev *dev, struct mthca_cq *cq, in update_cons_index() argument 208 *cq->set_ci_db = cpu_to_be32(cq in update_cons_index() 219 struct mthca_cq *cq; mthca_cq_completion() local 236 struct mthca_cq *cq; mthca_cq_event() local 273 mthca_cq_clean(struct mthca_dev *dev, struct mthca_cq *cq, u32 qpn, struct mthca_srq *srq) mthca_cq_clean() argument 325 mthca_cq_resize_copy_cqes(struct mthca_cq *cq) mthca_cq_resize_copy_cqes() argument 372 handle_error_cqe(struct mthca_dev *dev, struct mthca_cq *cq, struct mthca_qp *qp, int wqe_index, int is_send, struct mthca_err_cqe *cqe, struct ib_wc *entry, int *free_cqe) handle_error_cqe() argument 478 mthca_poll_one(struct mthca_dev *dev, struct mthca_cq *cq, struct mthca_qp **cur_qp, int *freed, struct ib_wc *entry) mthca_poll_one() argument 658 struct mthca_cq *cq = to_mcq(ibcq); mthca_poll_cq() local 723 mthca_tavor_arm_cq(struct ib_cq *cq, enum ib_cq_notify_flags flags) mthca_tavor_arm_cq() argument 738 struct mthca_cq *cq = to_mcq(ibcq); mthca_arbel_arm_cq() local 768 mthca_init_cq(struct mthca_dev *dev, int nent, struct mthca_ucontext *ctx, u32 pdn, struct mthca_cq *cq) mthca_init_cq() argument 889 get_cq_refcount(struct mthca_dev *dev, struct mthca_cq *cq) get_cq_refcount() argument 900 mthca_free_cq(struct mthca_dev *dev, struct mthca_cq *cq) mthca_free_cq() argument [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/mthca/ |
H A D | mthca_cq.c | 169 static inline struct mthca_cqe *get_cqe(struct mthca_cq *cq, int entry) in get_cqe() argument 171 return get_cqe_from_buf(&cq->buf, entry); in get_cqe() 179 static inline struct mthca_cqe *next_cqe_sw(struct mthca_cq *cq) in next_cqe_sw() argument 181 return cqe_sw(get_cqe(cq, cq->cons_index & cq->ibcq.cqe)); in next_cqe_sw() 201 * incr is ignored in native Arbel (mem-free) mode, so cq->cons_index 204 static inline void update_cons_index(struct mthca_dev *dev, struct mthca_cq *cq, in update_cons_index() argument 208 *cq->set_ci_db = cpu_to_be32(cq in update_cons_index() 219 struct mthca_cq *cq; mthca_cq_completion() local 236 struct mthca_cq *cq; mthca_cq_event() local 273 mthca_cq_clean(struct mthca_dev *dev, struct mthca_cq *cq, u32 qpn, struct mthca_srq *srq) mthca_cq_clean() argument 325 mthca_cq_resize_copy_cqes(struct mthca_cq *cq) mthca_cq_resize_copy_cqes() argument 372 handle_error_cqe(struct mthca_dev *dev, struct mthca_cq *cq, struct mthca_qp *qp, int wqe_index, int is_send, struct mthca_err_cqe *cqe, struct ib_wc *entry, int *free_cqe) handle_error_cqe() argument 478 mthca_poll_one(struct mthca_dev *dev, struct mthca_cq *cq, struct mthca_qp **cur_qp, int *freed, struct ib_wc *entry) mthca_poll_one() argument 658 struct mthca_cq *cq = to_mcq(ibcq); mthca_poll_cq() local 723 mthca_tavor_arm_cq(struct ib_cq *cq, enum ib_cq_notify_flags flags) mthca_tavor_arm_cq() argument 738 struct mthca_cq *cq = to_mcq(ibcq); mthca_arbel_arm_cq() local 768 mthca_init_cq(struct mthca_dev *dev, int nent, struct mthca_ucontext *ctx, u32 pdn, struct mthca_cq *cq) mthca_init_cq() argument 889 get_cq_refcount(struct mthca_dev *dev, struct mthca_cq *cq) get_cq_refcount() argument 900 mthca_free_cq(struct mthca_dev *dev, struct mthca_cq *cq) mthca_free_cq() argument [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/intel/ice/ |
H A D | ice_controlq.c | 36 struct ice_ctl_q_info *cq = &hw->adminq; in ice_adminq_init_regs() local 38 ICE_CQ_INIT_REGS(cq, PF_FW); in ice_adminq_init_regs() 49 struct ice_ctl_q_info *cq = &hw->mailboxq; in ice_mailbox_init_regs() local 51 ICE_CQ_INIT_REGS(cq, PF_MBX); in ice_mailbox_init_regs() 57 * @cq: pointer to the specific Control queue 61 bool ice_check_sq_alive(struct ice_hw *hw, struct ice_ctl_q_info *cq) in ice_check_sq_alive() argument 64 if (cq->sq.len && cq->sq.len_mask && cq->sq.len_ena_mask) in ice_check_sq_alive() 65 return (rd32(hw, cq in ice_check_sq_alive() 78 ice_alloc_ctrlq_sq_ring(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_alloc_ctrlq_sq_ring() argument 109 ice_alloc_ctrlq_rq_ring(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_alloc_ctrlq_rq_ring() argument 145 ice_alloc_rq_bufs(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_alloc_rq_bufs() argument 217 ice_alloc_sq_bufs(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_alloc_sq_bufs() argument 286 ice_cfg_sq_regs(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_cfg_sq_regs() argument 299 ice_cfg_rq_regs(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_cfg_rq_regs() argument 351 ice_init_sq(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_init_sq() argument 411 ice_init_rq(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_init_rq() argument 465 ice_shutdown_sq(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_shutdown_sq() argument 532 ice_shutdown_rq(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_shutdown_rq() argument 568 struct ice_ctl_q_info *cq = &hw->adminq; ice_init_check_adminq() local 604 struct ice_ctl_q_info *cq; ice_init_ctrlq() local 657 struct ice_ctl_q_info *cq; ice_shutdown_ctrlq() local 738 ice_init_ctrlq_locks(struct ice_ctl_q_info *cq) ice_init_ctrlq_locks() argument 774 ice_destroy_ctrlq_locks(struct ice_ctl_q_info *cq) ice_destroy_ctrlq_locks() argument 805 ice_clean_sq(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_clean_sq() argument 886 ice_sq_done(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_sq_done() argument 907 ice_sq_send_cmd(struct ice_hw *hw, struct ice_ctl_q_info *cq, struct ice_aq_desc *desc, void *buf, u16 buf_size, struct ice_sq_cd *cd) ice_sq_send_cmd() argument 1108 ice_clean_rq_elem(struct ice_hw *hw, struct ice_ctl_q_info *cq, struct ice_rq_event_info *e, u16 *pending) ice_clean_rq_elem() argument [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/intel/ice/ |
H A D | ice_controlq.c | 36 struct ice_ctl_q_info *cq = &hw->adminq; in ice_adminq_init_regs() local 38 ICE_CQ_INIT_REGS(cq, PF_FW); in ice_adminq_init_regs() 49 struct ice_ctl_q_info *cq = &hw->mailboxq; in ice_mailbox_init_regs() local 51 ICE_CQ_INIT_REGS(cq, PF_MBX); in ice_mailbox_init_regs() 62 struct ice_ctl_q_info *cq = &hw->sbq; in ice_sb_init_regs() local 64 ICE_CQ_INIT_REGS(cq, PF_SB); in ice_sb_init_regs() 70 * @cq: pointer to the specific Control queue 74 bool ice_check_sq_alive(struct ice_hw *hw, struct ice_ctl_q_info *cq) in ice_check_sq_alive() argument 77 if (cq->sq.len && cq in ice_check_sq_alive() 91 ice_alloc_ctrlq_sq_ring(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_alloc_ctrlq_sq_ring() argument 122 ice_alloc_ctrlq_rq_ring(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_alloc_ctrlq_rq_ring() argument 158 ice_alloc_rq_bufs(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_alloc_rq_bufs() argument 230 ice_alloc_sq_bufs(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_alloc_sq_bufs() argument 298 ice_cfg_sq_regs(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_cfg_sq_regs() argument 310 ice_cfg_rq_regs(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_cfg_rq_regs() argument 361 ice_init_sq(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_init_sq() argument 421 ice_init_rq(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_init_rq() argument 474 ice_shutdown_sq(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_shutdown_sq() argument 540 ice_shutdown_rq(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_shutdown_rq() argument 576 struct ice_ctl_q_info *cq = &hw->adminq; ice_init_check_adminq() local 612 struct ice_ctl_q_info *cq; ice_init_ctrlq() local 692 struct ice_ctl_q_info *cq; ice_shutdown_ctrlq() local 787 ice_init_ctrlq_locks(struct ice_ctl_q_info *cq) ice_init_ctrlq_locks() argument 825 ice_destroy_ctrlq_locks(struct ice_ctl_q_info *cq) ice_destroy_ctrlq_locks() argument 858 ice_clean_sq(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_clean_sq() argument 937 ice_sq_done(struct ice_hw *hw, struct ice_ctl_q_info *cq) ice_sq_done() argument 958 ice_sq_send_cmd(struct ice_hw *hw, struct ice_ctl_q_info *cq, struct ice_aq_desc *desc, void *buf, u16 buf_size, struct ice_sq_cd *cd) ice_sq_send_cmd() argument 1156 ice_clean_rq_elem(struct ice_hw *hw, struct ice_ctl_q_info *cq, struct ice_rq_event_info *e, u16 *pending) ice_clean_rq_elem() argument [all...] |
/kernel/linux/linux-5.10/drivers/scsi/fnic/ |
H A D | vnic_cq.c | 24 void vnic_cq_free(struct vnic_cq *cq) in vnic_cq_free() argument 26 vnic_dev_free_desc_ring(cq->vdev, &cq->ring); in vnic_cq_free() 28 cq->ctrl = NULL; in vnic_cq_free() 31 int vnic_cq_alloc(struct vnic_dev *vdev, struct vnic_cq *cq, unsigned int index, in vnic_cq_alloc() argument 36 cq->index = index; in vnic_cq_alloc() 37 cq->vdev = vdev; in vnic_cq_alloc() 39 cq->ctrl = vnic_dev_get_res(vdev, RES_TYPE_CQ, index); in vnic_cq_alloc() 40 if (!cq->ctrl) { in vnic_cq_alloc() 45 err = vnic_dev_alloc_desc_ring(vdev, &cq in vnic_cq_alloc() 52 vnic_cq_init(struct vnic_cq *cq, unsigned int flow_control_enable, unsigned int color_enable, unsigned int cq_head, unsigned int cq_tail, unsigned int cq_tail_color, unsigned int interrupt_enable, unsigned int cq_entry_enable, unsigned int cq_message_enable, unsigned int interrupt_offset, u64 cq_message_addr) vnic_cq_init() argument 75 vnic_cq_clean(struct vnic_cq *cq) vnic_cq_clean() argument [all...] |
/kernel/linux/linux-5.10/drivers/scsi/snic/ |
H A D | vnic_cq.c | 24 void svnic_cq_free(struct vnic_cq *cq) in svnic_cq_free() argument 26 svnic_dev_free_desc_ring(cq->vdev, &cq->ring); in svnic_cq_free() 28 cq->ctrl = NULL; in svnic_cq_free() 31 int svnic_cq_alloc(struct vnic_dev *vdev, struct vnic_cq *cq, in svnic_cq_alloc() argument 34 cq->index = index; in svnic_cq_alloc() 35 cq->vdev = vdev; in svnic_cq_alloc() 37 cq->ctrl = svnic_dev_get_res(vdev, RES_TYPE_CQ, index); in svnic_cq_alloc() 38 if (!cq->ctrl) { in svnic_cq_alloc() 44 return svnic_dev_alloc_desc_ring(vdev, &cq in svnic_cq_alloc() 47 svnic_cq_init(struct vnic_cq *cq, unsigned int flow_control_enable, unsigned int color_enable, unsigned int cq_head, unsigned int cq_tail, unsigned int cq_tail_color, unsigned int interrupt_enable, unsigned int cq_entry_enable, unsigned int cq_message_enable, unsigned int interrupt_offset, u64 cq_message_addr) svnic_cq_init() argument 70 svnic_cq_clean(struct vnic_cq *cq) svnic_cq_clean() argument [all...] |
/kernel/linux/linux-6.6/drivers/scsi/snic/ |
H A D | vnic_cq.c | 10 void svnic_cq_free(struct vnic_cq *cq) in svnic_cq_free() argument 12 svnic_dev_free_desc_ring(cq->vdev, &cq->ring); in svnic_cq_free() 14 cq->ctrl = NULL; in svnic_cq_free() 17 int svnic_cq_alloc(struct vnic_dev *vdev, struct vnic_cq *cq, in svnic_cq_alloc() argument 20 cq->index = index; in svnic_cq_alloc() 21 cq->vdev = vdev; in svnic_cq_alloc() 23 cq->ctrl = svnic_dev_get_res(vdev, RES_TYPE_CQ, index); in svnic_cq_alloc() 24 if (!cq->ctrl) { in svnic_cq_alloc() 30 return svnic_dev_alloc_desc_ring(vdev, &cq in svnic_cq_alloc() 33 svnic_cq_init(struct vnic_cq *cq, unsigned int flow_control_enable, unsigned int color_enable, unsigned int cq_head, unsigned int cq_tail, unsigned int cq_tail_color, unsigned int interrupt_enable, unsigned int cq_entry_enable, unsigned int cq_message_enable, unsigned int interrupt_offset, u64 cq_message_addr) svnic_cq_init() argument 56 svnic_cq_clean(struct vnic_cq *cq) svnic_cq_clean() argument [all...] |
/kernel/linux/linux-6.6/drivers/scsi/fnic/ |
H A D | vnic_cq.c | 12 void vnic_cq_free(struct vnic_cq *cq) in vnic_cq_free() argument 14 vnic_dev_free_desc_ring(cq->vdev, &cq->ring); in vnic_cq_free() 16 cq->ctrl = NULL; in vnic_cq_free() 19 int vnic_cq_alloc(struct vnic_dev *vdev, struct vnic_cq *cq, unsigned int index, in vnic_cq_alloc() argument 24 cq->index = index; in vnic_cq_alloc() 25 cq->vdev = vdev; in vnic_cq_alloc() 27 cq->ctrl = vnic_dev_get_res(vdev, RES_TYPE_CQ, index); in vnic_cq_alloc() 28 if (!cq->ctrl) { in vnic_cq_alloc() 33 err = vnic_dev_alloc_desc_ring(vdev, &cq in vnic_cq_alloc() 40 vnic_cq_init(struct vnic_cq *cq, unsigned int flow_control_enable, unsigned int color_enable, unsigned int cq_head, unsigned int cq_tail, unsigned int cq_tail_color, unsigned int interrupt_enable, unsigned int cq_entry_enable, unsigned int cq_message_enable, unsigned int interrupt_offset, u64 cq_message_addr) vnic_cq_init() argument 63 vnic_cq_clean(struct vnic_cq *cq) vnic_cq_clean() argument [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/sw/rdmavt/ |
H A D | cq.c | 50 #include "cq.h" 58 * @cq: completion queue 65 * false if cq is full. 67 bool rvt_cq_enter(struct rvt_cq *cq, struct ib_wc *entry, bool solicited) in rvt_cq_enter() argument 78 spin_lock_irqsave(&cq->lock, flags); in rvt_cq_enter() 80 if (cq->ip) { in rvt_cq_enter() 81 u_wc = cq->queue; in rvt_cq_enter() 86 k_wc = cq->kqueue; in rvt_cq_enter() 96 if (head >= (unsigned)cq->ibcq.cqe) { in rvt_cq_enter() 97 head = cq in rvt_cq_enter() 163 struct rvt_cq *cq = container_of(work, struct rvt_cq, comptask); send_complete() local 205 struct rvt_cq *cq = ibcq_to_rvtcq(ibcq); rvt_create_cq() local 320 struct rvt_cq *cq = ibcq_to_rvtcq(ibcq); rvt_destroy_cq() local 346 struct rvt_cq *cq = ibcq_to_rvtcq(ibcq); rvt_req_notify_cq() local 382 struct rvt_cq *cq = ibcq_to_rvtcq(ibcq); rvt_resize_cq() local 520 struct rvt_cq *cq = ibcq_to_rvtcq(ibcq); rvt_poll_cq() local [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/sw/rdmavt/ |
H A D | cq.c | 8 #include "cq.h" 16 * @cq: completion queue 23 * false if cq is full. 25 bool rvt_cq_enter(struct rvt_cq *cq, struct ib_wc *entry, bool solicited) in rvt_cq_enter() argument 36 spin_lock_irqsave(&cq->lock, flags); in rvt_cq_enter() 38 if (cq->ip) { in rvt_cq_enter() 39 u_wc = cq->queue; in rvt_cq_enter() 44 k_wc = cq->kqueue; in rvt_cq_enter() 54 if (head >= (unsigned)cq->ibcq.cqe) { in rvt_cq_enter() 55 head = cq in rvt_cq_enter() 121 struct rvt_cq *cq = container_of(work, struct rvt_cq, comptask); send_complete() local 163 struct rvt_cq *cq = ibcq_to_rvtcq(ibcq); rvt_create_cq() local 278 struct rvt_cq *cq = ibcq_to_rvtcq(ibcq); rvt_destroy_cq() local 304 struct rvt_cq *cq = ibcq_to_rvtcq(ibcq); rvt_req_notify_cq() local 340 struct rvt_cq *cq = ibcq_to_rvtcq(ibcq); rvt_resize_cq() local 478 struct rvt_cq *cq = ibcq_to_rvtcq(ibcq); rvt_poll_cq() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/cisco/enic/ |
H A D | vnic_cq.c | 29 void vnic_cq_free(struct vnic_cq *cq) in vnic_cq_free() argument 31 vnic_dev_free_desc_ring(cq->vdev, &cq->ring); in vnic_cq_free() 33 cq->ctrl = NULL; in vnic_cq_free() 36 int vnic_cq_alloc(struct vnic_dev *vdev, struct vnic_cq *cq, unsigned int index, in vnic_cq_alloc() argument 41 cq->index = index; in vnic_cq_alloc() 42 cq->vdev = vdev; in vnic_cq_alloc() 44 cq->ctrl = vnic_dev_get_res(vdev, RES_TYPE_CQ, index); in vnic_cq_alloc() 45 if (!cq->ctrl) { in vnic_cq_alloc() 50 err = vnic_dev_alloc_desc_ring(vdev, &cq in vnic_cq_alloc() 57 vnic_cq_init(struct vnic_cq *cq, unsigned int flow_control_enable, unsigned int color_enable, unsigned int cq_head, unsigned int cq_tail, unsigned int cq_tail_color, unsigned int interrupt_enable, unsigned int cq_entry_enable, unsigned int cq_message_enable, unsigned int interrupt_offset, u64 cq_message_addr) vnic_cq_init() argument 82 vnic_cq_clean(struct vnic_cq *cq) vnic_cq_clean() argument [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/cisco/enic/ |
H A D | vnic_cq.c | 16 void vnic_cq_free(struct vnic_cq *cq) in vnic_cq_free() argument 18 vnic_dev_free_desc_ring(cq->vdev, &cq->ring); in vnic_cq_free() 20 cq->ctrl = NULL; in vnic_cq_free() 23 int vnic_cq_alloc(struct vnic_dev *vdev, struct vnic_cq *cq, unsigned int index, in vnic_cq_alloc() argument 26 cq->index = index; in vnic_cq_alloc() 27 cq->vdev = vdev; in vnic_cq_alloc() 29 cq->ctrl = vnic_dev_get_res(vdev, RES_TYPE_CQ, index); in vnic_cq_alloc() 30 if (!cq->ctrl) { in vnic_cq_alloc() 35 return vnic_dev_alloc_desc_ring(vdev, &cq in vnic_cq_alloc() 38 vnic_cq_init(struct vnic_cq *cq, unsigned int flow_control_enable, unsigned int color_enable, unsigned int cq_head, unsigned int cq_tail, unsigned int cq_tail_color, unsigned int interrupt_enable, unsigned int cq_entry_enable, unsigned int cq_message_enable, unsigned int interrupt_offset, u64 cq_message_addr) vnic_cq_init() argument 63 vnic_cq_clean(struct vnic_cq *cq) vnic_cq_clean() argument [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx4/ |
H A D | cq.c | 34 #include <linux/mlx4/cq.h> 43 static void mlx4_ib_cq_comp(struct mlx4_cq *cq) in mlx4_ib_cq_comp() argument 45 struct ib_cq *ibcq = &to_mibcq(cq)->ibcq; in mlx4_ib_cq_comp() 49 static void mlx4_ib_cq_event(struct mlx4_cq *cq, enum mlx4_event type) in mlx4_ib_cq_event() argument 56 "on CQ %06x\n", type, cq->cqn); in mlx4_ib_cq_event() 60 ibcq = &to_mibcq(cq)->ibcq; in mlx4_ib_cq_event() 64 event.element.cq = ibcq; in mlx4_ib_cq_event() 74 static void *get_cqe(struct mlx4_ib_cq *cq, int n) in get_cqe() argument 76 return get_cqe_from_buf(&cq->buf, n); in get_cqe() 79 static void *get_sw_cqe(struct mlx4_ib_cq *cq, in argument 88 next_cqe_sw(struct mlx4_ib_cq *cq) next_cqe_sw() argument 93 mlx4_ib_modify_cq(struct ib_cq *cq, u16 cq_count, u16 cq_period) mlx4_ib_modify_cq() argument 181 struct mlx4_ib_cq *cq = to_mcq(ibcq); mlx4_ib_create_cq() local 290 mlx4_alloc_resize_buf(struct mlx4_ib_dev *dev, struct mlx4_ib_cq *cq, int entries) mlx4_alloc_resize_buf() argument 314 mlx4_alloc_resize_umem(struct mlx4_ib_dev *dev, struct mlx4_ib_cq *cq, int entries, struct ib_udata *udata) mlx4_alloc_resize_umem() argument 343 mlx4_ib_get_outstanding_cqes(struct mlx4_ib_cq *cq) mlx4_ib_get_outstanding_cqes() argument 354 mlx4_ib_cq_resize_copy_cqes(struct mlx4_ib_cq *cq) mlx4_ib_cq_resize_copy_cqes() argument 382 struct mlx4_ib_cq *cq = to_mcq(ibcq); mlx4_ib_resize_cq() local 477 mlx4_ib_destroy_cq(struct ib_cq *cq, struct ib_udata *udata) mlx4_ib_destroy_cq() argument 580 use_tunnel_data(struct mlx4_ib_qp *qp, struct mlx4_ib_cq *cq, struct ib_wc *wc, unsigned tail, struct mlx4_cqe *cqe, int is_eth) use_tunnel_data() argument 631 mlx4_ib_poll_sw_comp(struct mlx4_ib_cq *cq, int num_entries, struct ib_wc *wc, int *npolled) mlx4_ib_poll_sw_comp() argument 656 mlx4_ib_poll_one(struct mlx4_ib_cq *cq, struct mlx4_ib_qp **cur_qp, struct ib_wc *wc) mlx4_ib_poll_one() argument 879 struct mlx4_ib_cq *cq = to_mcq(ibcq); mlx4_ib_poll_cq() local 915 __mlx4_ib_cq_clean(struct mlx4_ib_cq *cq, u32 qpn, struct mlx4_ib_srq *srq) __mlx4_ib_cq_clean() argument 968 mlx4_ib_cq_clean(struct mlx4_ib_cq *cq, u32 qpn, struct mlx4_ib_srq *srq) mlx4_ib_cq_clean() argument [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/mlx4/ |
H A D | cq.c | 34 #include <linux/mlx4/cq.h> 43 static void mlx4_ib_cq_comp(struct mlx4_cq *cq) in mlx4_ib_cq_comp() argument 45 struct ib_cq *ibcq = &to_mibcq(cq)->ibcq; in mlx4_ib_cq_comp() 49 static void mlx4_ib_cq_event(struct mlx4_cq *cq, enum mlx4_event type) in mlx4_ib_cq_event() argument 56 "on CQ %06x\n", type, cq->cqn); in mlx4_ib_cq_event() 60 ibcq = &to_mibcq(cq)->ibcq; in mlx4_ib_cq_event() 64 event.element.cq = ibcq; in mlx4_ib_cq_event() 74 static void *get_cqe(struct mlx4_ib_cq *cq, int n) in get_cqe() argument 76 return get_cqe_from_buf(&cq->buf, n); in get_cqe() 79 static void *get_sw_cqe(struct mlx4_ib_cq *cq, in argument 88 next_cqe_sw(struct mlx4_ib_cq *cq) next_cqe_sw() argument 93 mlx4_ib_modify_cq(struct ib_cq *cq, u16 cq_count, u16 cq_period) mlx4_ib_modify_cq() argument 181 struct mlx4_ib_cq *cq = to_mcq(ibcq); mlx4_ib_create_cq() local 290 mlx4_alloc_resize_buf(struct mlx4_ib_dev *dev, struct mlx4_ib_cq *cq, int entries) mlx4_alloc_resize_buf() argument 314 mlx4_alloc_resize_umem(struct mlx4_ib_dev *dev, struct mlx4_ib_cq *cq, int entries, struct ib_udata *udata) mlx4_alloc_resize_umem() argument 343 mlx4_ib_get_outstanding_cqes(struct mlx4_ib_cq *cq) mlx4_ib_get_outstanding_cqes() argument 354 mlx4_ib_cq_resize_copy_cqes(struct mlx4_ib_cq *cq) mlx4_ib_cq_resize_copy_cqes() argument 382 struct mlx4_ib_cq *cq = to_mcq(ibcq); mlx4_ib_resize_cq() local 477 mlx4_ib_destroy_cq(struct ib_cq *cq, struct ib_udata *udata) mlx4_ib_destroy_cq() argument 580 use_tunnel_data(struct mlx4_ib_qp *qp, struct mlx4_ib_cq *cq, struct ib_wc *wc, unsigned tail, struct mlx4_cqe *cqe, int is_eth) use_tunnel_data() argument 631 mlx4_ib_poll_sw_comp(struct mlx4_ib_cq *cq, int num_entries, struct ib_wc *wc, int *npolled) mlx4_ib_poll_sw_comp() argument 656 mlx4_ib_poll_one(struct mlx4_ib_cq *cq, struct mlx4_ib_qp **cur_qp, struct ib_wc *wc) mlx4_ib_poll_one() argument 879 struct mlx4_ib_cq *cq = to_mcq(ibcq); mlx4_ib_poll_cq() local 915 __mlx4_ib_cq_clean(struct mlx4_ib_cq *cq, u32 qpn, struct mlx4_ib_srq *srq) __mlx4_ib_cq_clean() argument 968 mlx4_ib_cq_clean(struct mlx4_ib_cq *cq, u32 qpn, struct mlx4_ib_srq *srq) mlx4_ib_cq_clean() argument [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | cq.c | 38 #include <linux/mlx5/cq.h> 70 static void mlx5_add_cq_to_tasklet(struct mlx5_core_cq *cq, in mlx5_add_cq_to_tasklet() argument 74 struct mlx5_eq_tasklet *tasklet_ctx = cq->tasklet_ctx.priv; in mlx5_add_cq_to_tasklet() 82 if (list_empty_careful(&cq->tasklet_ctx.list)) { in mlx5_add_cq_to_tasklet() 83 mlx5_cq_hold(cq); in mlx5_add_cq_to_tasklet() 84 list_add_tail(&cq->tasklet_ctx.list, &tasklet_ctx->list); in mlx5_add_cq_to_tasklet() 89 int mlx5_core_create_cq(struct mlx5_core_dev *dev, struct mlx5_core_cq *cq, in mlx5_core_create_cq() argument 107 cq->cqn = MLX5_GET(create_cq_out, out, cqn); in mlx5_core_create_cq() 108 cq->cons_index = 0; in mlx5_core_create_cq() 109 cq in mlx5_core_create_cq() 152 mlx5_core_destroy_cq(struct mlx5_core_dev *dev, struct mlx5_core_cq *cq) mlx5_core_destroy_cq() argument 159 mlx5_eq_del_cq(mlx5_get_async_eq(dev), cq); mlx5_core_destroy_cq() local 177 mlx5_core_query_cq(struct mlx5_core_dev *dev, struct mlx5_core_cq *cq, u32 *out) mlx5_core_query_cq() argument 188 mlx5_core_modify_cq(struct mlx5_core_dev *dev, struct mlx5_core_cq *cq, u32 *in, int inlen) mlx5_core_modify_cq() argument 199 mlx5_core_modify_cq_moderation(struct mlx5_core_dev *dev, struct mlx5_core_cq *cq, u16 cq_period, u16 cq_max_count) mlx5_core_modify_cq_moderation() argument [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_cq.c | 67 struct pvrdma_cq *cq = to_vcq(ibcq); in pvrdma_req_notify_cq() local 68 u32 val = cq->cq_handle; in pvrdma_req_notify_cq() 75 spin_lock_irqsave(&cq->cq_lock, flags); in pvrdma_req_notify_cq() 82 has_data = pvrdma_idx_ring_has_data(&cq->ring_state->rx, in pvrdma_req_notify_cq() 83 cq->ibcq.cqe, &head); in pvrdma_req_notify_cq() 88 spin_unlock_irqrestore(&cq->cq_lock, flags); in pvrdma_req_notify_cq() 107 struct pvrdma_cq *cq = to_vcq(ibcq); in pvrdma_create_cq() local 129 cq->ibcq.cqe = entries; in pvrdma_create_cq() 130 cq->is_kernel = !udata; in pvrdma_create_cq() 132 if (!cq in pvrdma_create_cq() 222 pvrdma_free_cq(struct pvrdma_dev *dev, struct pvrdma_cq *cq) pvrdma_free_cq() argument 238 pvrdma_destroy_cq(struct ib_cq *cq, struct ib_udata *udata) pvrdma_destroy_cq() argument 267 get_cqe(struct pvrdma_cq *cq, int i) get_cqe() argument 275 _pvrdma_flush_cqe(struct pvrdma_qp *qp, struct pvrdma_cq *cq) _pvrdma_flush_cqe() argument 319 pvrdma_poll_one(struct pvrdma_cq *cq, struct pvrdma_qp **cur_qp, struct ib_wc *wc) pvrdma_poll_one() argument 385 struct pvrdma_cq *cq = to_vcq(ibcq); pvrdma_poll_cq() local [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_cq.c | 67 struct pvrdma_cq *cq = to_vcq(ibcq); in pvrdma_req_notify_cq() local 68 u32 val = cq->cq_handle; in pvrdma_req_notify_cq() 75 spin_lock_irqsave(&cq->cq_lock, flags); in pvrdma_req_notify_cq() 82 has_data = pvrdma_idx_ring_has_data(&cq->ring_state->rx, in pvrdma_req_notify_cq() 83 cq->ibcq.cqe, &head); in pvrdma_req_notify_cq() 88 spin_unlock_irqrestore(&cq->cq_lock, flags); in pvrdma_req_notify_cq() 107 struct pvrdma_cq *cq = to_vcq(ibcq); in pvrdma_create_cq() local 132 cq->ibcq.cqe = entries; in pvrdma_create_cq() 133 cq->is_kernel = !udata; in pvrdma_create_cq() 135 if (!cq in pvrdma_create_cq() 225 pvrdma_free_cq(struct pvrdma_dev *dev, struct pvrdma_cq *cq) pvrdma_free_cq() argument 241 pvrdma_destroy_cq(struct ib_cq *cq, struct ib_udata *udata) pvrdma_destroy_cq() argument 270 get_cqe(struct pvrdma_cq *cq, int i) get_cqe() argument 278 _pvrdma_flush_cqe(struct pvrdma_qp *qp, struct pvrdma_cq *cq) _pvrdma_flush_cqe() argument 322 pvrdma_poll_one(struct pvrdma_cq *cq, struct pvrdma_qp **cur_qp, struct ib_wc *wc) pvrdma_poll_one() argument 388 struct pvrdma_cq *cq = to_vcq(ibcq); pvrdma_poll_cq() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | cq.c | 37 #include <linux/mlx5/cq.h> 69 static void mlx5_add_cq_to_tasklet(struct mlx5_core_cq *cq, in mlx5_add_cq_to_tasklet() argument 73 struct mlx5_eq_tasklet *tasklet_ctx = cq->tasklet_ctx.priv; in mlx5_add_cq_to_tasklet() 81 if (list_empty_careful(&cq->tasklet_ctx.list)) { in mlx5_add_cq_to_tasklet() 82 mlx5_cq_hold(cq); in mlx5_add_cq_to_tasklet() 83 list_add_tail(&cq->tasklet_ctx.list, &tasklet_ctx->list); in mlx5_add_cq_to_tasklet() 89 int mlx5_create_cq(struct mlx5_core_dev *dev, struct mlx5_core_cq *cq, in mlx5_create_cq() argument 108 cq->cqn = MLX5_GET(create_cq_out, out, cqn); in mlx5_create_cq() 109 cq->cons_index = 0; in mlx5_create_cq() 110 cq in mlx5_create_cq() 154 mlx5_core_create_cq(struct mlx5_core_dev *dev, struct mlx5_core_cq *cq, u32 *in, int inlen, u32 *out, int outlen) mlx5_core_create_cq() argument 163 mlx5_core_destroy_cq(struct mlx5_core_dev *dev, struct mlx5_core_cq *cq) mlx5_core_destroy_cq() argument 170 mlx5_eq_del_cq(mlx5_get_async_eq(dev), cq); mlx5_core_destroy_cq() local 188 mlx5_core_query_cq(struct mlx5_core_dev *dev, struct mlx5_core_cq *cq, u32 *out) mlx5_core_query_cq() argument 199 mlx5_core_modify_cq(struct mlx5_core_dev *dev, struct mlx5_core_cq *cq, u32 *in, int inlen) mlx5_core_modify_cq() argument 210 mlx5_core_modify_cq_moderation(struct mlx5_core_dev *dev, struct mlx5_core_cq *cq, u16 cq_period, u16 cq_max_count) mlx5_core_modify_cq_moderation() argument [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/mlx5/ |
H A D | cq.c | 41 static void mlx5_ib_cq_comp(struct mlx5_core_cq *cq, struct mlx5_eqe *eqe) in mlx5_ib_cq_comp() argument 43 struct ib_cq *ibcq = &to_mibcq(cq)->ibcq; in mlx5_ib_cq_comp() 50 struct mlx5_ib_cq *cq = container_of(mcq, struct mlx5_ib_cq, mcq); in mlx5_ib_cq_event() local 51 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in mlx5_ib_cq_event() 52 struct ib_cq *ibcq = &cq->ibcq; in mlx5_ib_cq_event() 64 event.element.cq = ibcq; in mlx5_ib_cq_event() 69 static void *get_cqe(struct mlx5_ib_cq *cq, int n) in get_cqe() argument 71 return mlx5_frag_buf_get_wqe(&cq->buf.fbc, n); in get_cqe() 79 static void *get_sw_cqe(struct mlx5_ib_cq *cq, int n) in get_sw_cqe() argument 81 void *cqe = get_cqe(cq, in get_sw_cqe() 94 next_cqe_sw(struct mlx5_ib_cq *cq) next_cqe_sw() argument 427 mlx5_ib_poll_sw_comp(struct mlx5_ib_cq *cq, int num_entries, struct ib_wc *wc, int *npolled) mlx5_ib_poll_sw_comp() argument 447 mlx5_poll_one(struct mlx5_ib_cq *cq, struct mlx5_ib_qp **cur_qp, struct ib_wc *wc) mlx5_poll_one() argument 581 poll_soft_wc(struct mlx5_ib_cq *cq, int num_entries, struct ib_wc *wc, bool is_fatal_err) poll_soft_wc() argument 609 struct mlx5_ib_cq *cq = to_mcq(ibcq); mlx5_ib_poll_cq() local 647 struct mlx5_ib_cq *cq = to_mcq(ibcq); mlx5_ib_arm_cq() local 715 create_cq_user(struct mlx5_ib_dev *dev, struct ib_udata *udata, struct mlx5_ib_cq *cq, int entries, u32 **cqb, int *cqe_size, int *index, int *inlen) create_cq_user() argument 857 destroy_cq_user(struct mlx5_ib_cq *cq, struct ib_udata *udata) destroy_cq_user() argument 879 create_cq_kernel(struct mlx5_ib_dev *dev, struct mlx5_ib_cq *cq, int entries, int cqe_size, u32 **cqb, int *index, int *inlen) create_cq_kernel() argument 930 destroy_cq_kernel(struct mlx5_ib_dev *dev, struct mlx5_ib_cq *cq) destroy_cq_kernel() argument 938 struct mlx5_ib_cq *cq = container_of(work, struct mlx5_ib_cq, notify_soft_wc_handler() local 951 struct mlx5_ib_cq *cq = to_mcq(ibcq); mlx5_ib_create_cq() local 1049 mlx5_ib_destroy_cq(struct ib_cq *cq, struct ib_udata *udata) mlx5_ib_destroy_cq() argument 1071 __mlx5_ib_cq_clean(struct mlx5_ib_cq *cq, u32 rsn, struct mlx5_ib_srq *srq) __mlx5_ib_cq_clean() argument 1122 mlx5_ib_cq_clean(struct mlx5_ib_cq *cq, u32 qpn, struct mlx5_ib_srq *srq) mlx5_ib_cq_clean() argument 1132 mlx5_ib_modify_cq(struct ib_cq *cq, u16 cq_count, u16 cq_period) mlx5_ib_modify_cq() argument 1152 resize_user(struct mlx5_ib_dev *dev, struct mlx5_ib_cq *cq, int entries, struct ib_udata *udata, int *cqe_size) resize_user() argument 1185 resize_kernel(struct mlx5_ib_dev *dev, struct mlx5_ib_cq *cq, int entries, int cqe_size) resize_kernel() argument 1207 copy_resize_cqes(struct mlx5_ib_cq *cq) copy_resize_cqes() argument 1265 struct mlx5_ib_cq *cq = to_mcq(ibcq); mlx5_ib_resize_cq() local 1413 struct mlx5_ib_cq *cq; mlx5_ib_get_cqe_size() local 1426 struct mlx5_ib_cq *cq = to_mcq(ibcq); mlx5_ib_generate_wc() local [all...] |