/kernel/linux/linux-5.10/drivers/infiniband/hw/mthca/ |
H A D | mthca_cq.c | 181 return cqe_sw(get_cqe(cq, cq->cons_index & cq->ibcq.cqe)); in next_cqe_sw() 201 * incr is ignored in native Arbel (mem-free) mode, so cq->cons_index 208 *cq->set_ci_db = cpu_to_be32(cq->cons_index); in update_cons_index() 289 for (prod_index = cq->cons_index; in mthca_cq_clean() 292 if (prod_index == cq->cons_index + cq->ibcq.cqe) in mthca_cq_clean() 297 qpn, cq->cqn, cq->cons_index, prod_index); in mthca_cq_clean() 303 while ((int) --prod_index - (int) cq->cons_index >= 0) { in mthca_cq_clean() 316 set_cqe_hw(get_cqe(cq, (cq->cons_index + i) & cq->ibcq.cqe)); in mthca_cq_clean() 318 cq->cons_index += nfreed; in mthca_cq_clean() 337 cq->cons_index in mthca_cq_resize_copy_cqes() [all...] |
H A D | mthca_eq.c | 237 eqe = get_eqe(eq, eq->cons_index); in next_eqe_sw() 363 ++eq->cons_index; in mthca_eq_int() 379 set_eq_ci(dev, eq, eq->cons_index); in mthca_eq_int() 411 dev->eq_table.eq[i].cons_index); in mthca_tavor_interrupt() 424 tavor_set_eq_ci(dev, eq, eq->cons_index); in mthca_tavor_msi_x_interrupt() 444 dev->eq_table.eq[i].cons_index); in mthca_arbel_interrupt() 458 arbel_set_eq_ci(dev, eq, eq->cons_index); in mthca_arbel_msi_x_interrupt() 555 eq->cons_index = 0; in mthca_create_eq()
|
H A D | mthca_provider.h | 91 u32 cons_index; member 188 u32 cons_index; member
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/mthca/ |
H A D | mthca_cq.c | 181 return cqe_sw(get_cqe(cq, cq->cons_index & cq->ibcq.cqe)); in next_cqe_sw() 201 * incr is ignored in native Arbel (mem-free) mode, so cq->cons_index 208 *cq->set_ci_db = cpu_to_be32(cq->cons_index); in update_cons_index() 289 for (prod_index = cq->cons_index; in mthca_cq_clean() 292 if (prod_index == cq->cons_index + cq->ibcq.cqe) in mthca_cq_clean() 297 qpn, cq->cqn, cq->cons_index, prod_index); in mthca_cq_clean() 303 while ((int) --prod_index - (int) cq->cons_index >= 0) { in mthca_cq_clean() 316 set_cqe_hw(get_cqe(cq, (cq->cons_index + i) & cq->ibcq.cqe)); in mthca_cq_clean() 318 cq->cons_index += nfreed; in mthca_cq_clean() 337 cq->cons_index in mthca_cq_resize_copy_cqes() [all...] |
H A D | mthca_eq.c | 237 eqe = get_eqe(eq, eq->cons_index); in next_eqe_sw() 363 ++eq->cons_index; in mthca_eq_int() 379 set_eq_ci(dev, eq, eq->cons_index); in mthca_eq_int() 411 dev->eq_table.eq[i].cons_index); in mthca_tavor_interrupt() 424 tavor_set_eq_ci(dev, eq, eq->cons_index); in mthca_tavor_msi_x_interrupt() 444 dev->eq_table.eq[i].cons_index); in mthca_arbel_interrupt() 458 arbel_set_eq_ci(dev, eq, eq->cons_index); in mthca_arbel_msi_x_interrupt() 555 eq->cons_index = 0; in mthca_create_eq()
|
H A D | mthca_provider.h | 91 u32 cons_index; member 188 u32 cons_index; member
|
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/lib/ |
H A D | eq.h | 28 u32 cons_index; member 57 struct mlx5_eqe *eqe = get_eqe(eq, eq->cons_index & (eq->nent - 1)); in next_eqe_sw() 59 return ((eqe->owner & 1) ^ !!(eq->cons_index & eq->nent)) ? NULL : eqe; in next_eqe_sw() 65 u32 val = (eq->cons_index & 0xffffff) | (eq->eqn << 24); in eq_update_ci()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/lib/ |
H A D | eq.h | 30 u32 cons_index; member 63 struct mlx5_eqe *eqe = get_eqe(eq, eq->cons_index & eq->fbc.sz_m1); in next_eqe_sw() 65 return (eqe->owner ^ (eq->cons_index >> eq->fbc.log_sz)) & 1 ? NULL : eqe; in next_eqe_sw() 71 u32 val = (eq->cons_index & 0xffffff) | (eq->eqn << 24); in eq_update_ci()
|
/kernel/linux/linux-5.10/include/linux/mlx5/ |
H A D | cq.h | 51 u32 cons_index; member 143 *cq->set_ci_db = cpu_to_be32(cq->cons_index & 0xffffff); in mlx5_cq_set_ci() 153 u32 cons_index) in mlx5_cq_arm() 160 ci = cons_index & 0xffffff; in mlx5_cq_arm() 151 mlx5_cq_arm(struct mlx5_core_cq *cq, u32 cmd, void __iomem *uar_page, u32 cons_index) mlx5_cq_arm() argument
|
/kernel/linux/linux-6.6/include/linux/mlx5/ |
H A D | cq.h | 51 u32 cons_index; member 143 *cq->set_ci_db = cpu_to_be32(cq->cons_index & 0xffffff); in mlx5_cq_set_ci() 153 u32 cons_index) in mlx5_cq_arm() 160 ci = cons_index & 0xffffff; in mlx5_cq_arm() 151 mlx5_cq_arm(struct mlx5_core_cq *cq, u32 cmd, void __iomem *uar_page, u32 cons_index) mlx5_cq_arm() argument
|
/kernel/linux/linux-5.10/include/linux/mlx4/ |
H A D | cq.h | 156 ci = cq->cons_index & 0xffffff; in mlx4_cq_arm() 174 *cq->set_ci_db = cpu_to_be32(cq->cons_index & 0xffffff); in mlx4_cq_set_ci()
|
/kernel/linux/linux-6.6/include/linux/mlx4/ |
H A D | cq.h | 156 ci = cq->cons_index & 0xffffff; in mlx4_cq_arm() 174 *cq->set_ci_db = cpu_to_be32(cq->cons_index & 0xffffff); in mlx4_cq_set_ci()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | eq.c | 161 ++eq->cons_index; in mlx5_eq_comp_int() 184 count_eqe = eq->core.cons_index; in mlx5_eq_poll_irq_disabled() 186 count_eqe = eq->core.cons_index - count_eqe; in mlx5_eq_poll_irq_disabled() 246 ++eq->cons_index; in mlx5_eq_async_int() 299 eq->cons_index = 0; in create_map_eq() 761 u32 ci = eq->cons_index + cc; in mlx5_eq_get_eqe() 781 eq->cons_index += cc; in mlx5_eq_update_ci() 782 val = (eq->cons_index & 0xffffff) | (eq->eqn << 24); in mlx5_eq_update_ci()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx4/ |
H A D | cq.c | 90 return get_sw_cqe(cq, cq->mcq.cons_index); in next_cqe_sw() 347 i = cq->mcq.cons_index; in mlx4_ib_get_outstanding_cqes() 351 return i - cq->mcq.cons_index; in mlx4_ib_get_outstanding_cqes() 361 i = cq->mcq.cons_index; in mlx4_ib_cq_resize_copy_cqes() 376 ++cq->mcq.cons_index; in mlx4_ib_cq_resize_copy_cqes() 680 ++cq->mcq.cons_index; in mlx4_ib_poll_one() 930 for (prod_index = cq->mcq.cons_index; get_sw_cqe(cq, prod_index); ++prod_index) in __mlx4_ib_cq_clean() 931 if (prod_index == cq->mcq.cons_index + cq->ibcq.cqe) in __mlx4_ib_cq_clean() 938 while ((int) --prod_index - (int) cq->mcq.cons_index >= 0) { in __mlx4_ib_cq_clean() 958 cq->mcq.cons_index in __mlx4_ib_cq_clean() [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/mlx4/ |
H A D | cq.c | 90 return get_sw_cqe(cq, cq->mcq.cons_index); in next_cqe_sw() 347 i = cq->mcq.cons_index; in mlx4_ib_get_outstanding_cqes() 351 return i - cq->mcq.cons_index; in mlx4_ib_get_outstanding_cqes() 361 i = cq->mcq.cons_index; in mlx4_ib_cq_resize_copy_cqes() 376 ++cq->mcq.cons_index; in mlx4_ib_cq_resize_copy_cqes() 680 ++cq->mcq.cons_index; in mlx4_ib_poll_one() 930 for (prod_index = cq->mcq.cons_index; get_sw_cqe(cq, prod_index); ++prod_index) in __mlx4_ib_cq_clean() 931 if (prod_index == cq->mcq.cons_index + cq->ibcq.cqe) in __mlx4_ib_cq_clean() 938 while ((int) --prod_index - (int) cq->mcq.cons_index >= 0) { in __mlx4_ib_cq_clean() 958 cq->mcq.cons_index in __mlx4_ib_cq_clean() [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx4/ |
H A D | eq.c | 99 __raw_writel((__force u32) cpu_to_be32((eq->cons_index & 0xffffff) | in eq_set_ci() 123 struct mlx4_eqe *eqe = get_eqe(eq, eq->cons_index, eqe_factor, size); in next_eqe_sw() 124 return !!(eqe->owner & 0x80) ^ !!(eq->cons_index & eq->nent) ? NULL : eqe; in next_eqe_sw() 543 eq->eqn, eq->cons_index, ret); in mlx4_eq_int() 573 eq->eqn, eq->cons_index, ret); in mlx4_eq_int() 697 eq->eqn, eq->cons_index, ret); in mlx4_eq_int() 782 eq->cons_index, eqe->owner, eq->nent, in mlx4_eq_int() 785 !!(eq->cons_index & eq->nent) ? "HW" : "SW"); in mlx4_eq_int() 807 eq->cons_index, eqe->owner, eq->nent, in mlx4_eq_int() 809 !!(eq->cons_index in mlx4_eq_int() [all...] |
H A D | en_tx.c | 434 u32 cons_index = mcq->cons_index; in mlx4_en_process_tx_cq() local 451 index = cons_index & size_mask; in mlx4_en_process_tx_cq() 460 cons_index & size) && (done < budget)) { in mlx4_en_process_tx_cq() 503 ++cons_index; in mlx4_en_process_tx_cq() 504 index = cons_index & size_mask; in mlx4_en_process_tx_cq() 512 mcq->cons_index = cons_index; in mlx4_en_process_tx_cq()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx4/ |
H A D | eq.c | 99 __raw_writel((__force u32) cpu_to_be32((eq->cons_index & 0xffffff) | in eq_set_ci() 123 struct mlx4_eqe *eqe = get_eqe(eq, eq->cons_index, eqe_factor, size); in next_eqe_sw() 124 return !!(eqe->owner & 0x80) ^ !!(eq->cons_index & eq->nent) ? NULL : eqe; in next_eqe_sw() 543 eq->eqn, eq->cons_index, ret); in mlx4_eq_int() 573 eq->eqn, eq->cons_index, ret); in mlx4_eq_int() 698 eq->eqn, eq->cons_index, ret); in mlx4_eq_int() 783 eq->cons_index, eqe->owner, eq->nent, in mlx4_eq_int() 786 !!(eq->cons_index & eq->nent) ? "HW" : "SW"); in mlx4_eq_int() 808 eq->cons_index, eqe->owner, eq->nent, in mlx4_eq_int() 810 !!(eq->cons_index in mlx4_eq_int() [all...] |
H A D | en_tx.c | 439 u32 cons_index = mcq->cons_index; in mlx4_en_process_tx_cq() local 456 index = cons_index & size_mask; in mlx4_en_process_tx_cq() 465 cons_index & size) && (done < budget)) { in mlx4_en_process_tx_cq() 508 ++cons_index; in mlx4_en_process_tx_cq() 509 index = cons_index & size_mask; in mlx4_en_process_tx_cq() 517 mcq->cons_index = cons_index; in mlx4_en_process_tx_cq()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | eq.c | 143 ++eq->cons_index; in mlx5_eq_comp_int() 166 count_eqe = eq->core.cons_index; in mlx5_eq_poll_irq_disabled() 168 count_eqe = eq->core.cons_index - count_eqe; in mlx5_eq_poll_irq_disabled() 232 ++eq->cons_index; in mlx5_eq_async_int() 286 eq->cons_index = 0; in create_map_eq() 788 u32 ci = eq->cons_index + cc; in mlx5_eq_get_eqe() 809 eq->cons_index += cc; in mlx5_eq_update_ci() 810 val = (eq->cons_index & 0xffffff) | (eq->eqn << 24); in mlx5_eq_update_ci()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/hns/ |
H A D | hns_roce_hw_v1.c | 1989 return get_sw_cqe(hr_cq, hr_cq->cons_index); in next_cqe_sw() 1992 static void hns_roce_v1_cq_set_ci(struct hns_roce_cq *hr_cq, u32 cons_index) in hns_roce_v1_cq_set_ci() argument 1996 doorbell[0] = cpu_to_le32(cons_index & ((hr_cq->cq_depth << 1) - 1)); in hns_roce_v1_cq_set_ci() 2017 for (prod_index = hr_cq->cons_index; get_sw_cqe(hr_cq, prod_index); in __hns_roce_v1_cq_clean() 2019 if (prod_index == hr_cq->cons_index + hr_cq->ib_cq.cqe) in __hns_roce_v1_cq_clean() 2027 while ((int) --prod_index - (int) hr_cq->cons_index >= 0) { in __hns_roce_v1_cq_clean() 2046 hr_cq->cons_index += nfreed; in __hns_roce_v1_cq_clean() 2053 hns_roce_v1_cq_set_ci(hr_cq, hr_cq->cons_index); in __hns_roce_v1_cq_clean() 2155 cpu_to_le32(hr_cq->cons_index & ((hr_cq->cq_depth << 1) - 1)); in hns_roce_v1_req_notify_cq() 2190 ++hr_cq->cons_index; in hns_roce_v1_poll_one() [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | health.c | 128 err = devlink_fmsg_u32_pair_put(fmsg, "ci", eq->core.cons_index); in mlx5e_health_eq_diag_fmsg() 214 eq->core.eqn, eq->core.cons_index, eq->core.irqn); in mlx5e_health_channel_eq_recover()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | health.c | 127 err = devlink_fmsg_u32_pair_put(fmsg, "ci", eq->core.cons_index); in mlx5e_health_eq_diag_fmsg() 212 eq->core.eqn, eq->core.cons_index, eq->core.irqn); in mlx5e_health_channel_eq_recover()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx5/ |
H A D | cq.c | 96 return get_sw_cqe(cq, cq->mcq.cons_index); in next_cqe_sw() 463 ++cq->mcq.cons_index; in mlx5_poll_one() 653 uar_page, to_mcq(ibcq)->mcq.cons_index); in mlx5_ib_arm_cq() 1063 for (prod_index = cq->mcq.cons_index; get_sw_cqe(cq, prod_index); prod_index++) in __mlx5_ib_cq_clean() 1064 if (prod_index == cq->mcq.cons_index + cq->ibcq.cqe) in __mlx5_ib_cq_clean() 1070 while ((int) --prod_index - (int) cq->mcq.cons_index >= 0) { in __mlx5_ib_cq_clean() 1088 cq->mcq.cons_index += nfreed; in __mlx5_ib_cq_clean() 1206 i = cq->mcq.cons_index; in copy_resize_cqes() 1237 ++cq->mcq.cons_index; in copy_resize_cqes()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/mlx5/ |
H A D | cq.c | 96 return get_sw_cqe(cq, cq->mcq.cons_index); in next_cqe_sw() 469 ++cq->mcq.cons_index; in mlx5_poll_one() 663 uar_page, to_mcq(ibcq)->mcq.cons_index); in mlx5_ib_arm_cq() 1088 for (prod_index = cq->mcq.cons_index; get_sw_cqe(cq, prod_index); prod_index++) in __mlx5_ib_cq_clean() 1089 if (prod_index == cq->mcq.cons_index + cq->ibcq.cqe) in __mlx5_ib_cq_clean() 1095 while ((int) --prod_index - (int) cq->mcq.cons_index >= 0) { in __mlx5_ib_cq_clean() 1113 cq->mcq.cons_index += nfreed; in __mlx5_ib_cq_clean() 1227 i = cq->mcq.cons_index; in copy_resize_cqes() 1258 ++cq->mcq.cons_index; in copy_resize_cqes()
|