/kernel/linux/linux-6.6/drivers/infiniband/hw/erdma/ |
H A D | erdma_eq.c | 32 struct erdma_aeqe *aeqe; in erdma_aeq_event_handler() local 42 aeqe = get_next_valid_eqe(&dev->aeq); in erdma_aeq_event_handler() 43 if (!aeqe) in erdma_aeq_event_handler() 53 le32_to_cpu(aeqe->hdr)) == ERDMA_AE_TYPE_CQ_ERR) { in erdma_aeq_event_handler() 54 cqn = le32_to_cpu(aeqe->event_data0); in erdma_aeq_event_handler() 66 qpn = le32_to_cpu(aeqe->event_data0); in erdma_aeq_event_handler()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/hns/ |
H A D | hns_roce_hw_v1.c | 3684 struct hns_roce_aeqe *aeqe, int qpn) in hns_roce_v1_wq_catas_err_handle() 3689 switch (roce_get_field(aeqe->asyn, HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_M, in hns_roce_v1_wq_catas_err_handle() 3718 struct hns_roce_aeqe *aeqe, in hns_roce_v1_local_wq_access_err_handle() 3724 switch (roce_get_field(aeqe->asyn, HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_M, in hns_roce_v1_local_wq_access_err_handle() 3753 struct hns_roce_aeqe *aeqe, in hns_roce_v1_qp_err_handle() 3760 qpn = roce_get_field(aeqe->event.qp_event.qp, in hns_roce_v1_qp_err_handle() 3763 phy_port = roce_get_field(aeqe->event.qp_event.qp, in hns_roce_v1_qp_err_handle() 3775 hns_roce_v1_wq_catas_err_handle(hr_dev, aeqe, qpn); in hns_roce_v1_qp_err_handle() 3778 hns_roce_v1_local_wq_access_err_handle(hr_dev, aeqe, qpn); in hns_roce_v1_qp_err_handle() 3788 struct hns_roce_aeqe *aeqe, in hns_roce_v1_cq_err_handle() 3683 hns_roce_v1_wq_catas_err_handle(struct hns_roce_dev *hr_dev, struct hns_roce_aeqe *aeqe, int qpn) hns_roce_v1_wq_catas_err_handle() argument 3717 hns_roce_v1_local_wq_access_err_handle(struct hns_roce_dev *hr_dev, struct hns_roce_aeqe *aeqe, int qpn) hns_roce_v1_local_wq_access_err_handle() argument 3752 hns_roce_v1_qp_err_handle(struct hns_roce_dev *hr_dev, struct hns_roce_aeqe *aeqe, int event_type) hns_roce_v1_qp_err_handle() argument 3787 hns_roce_v1_cq_err_handle(struct hns_roce_dev *hr_dev, struct hns_roce_aeqe *aeqe, int event_type) hns_roce_v1_cq_err_handle() argument 3815 hns_roce_v1_db_overflow_handle(struct hns_roce_dev *hr_dev, struct hns_roce_aeqe *aeqe) hns_roce_v1_db_overflow_handle() argument 3856 struct hns_roce_aeqe *aeqe = get_aeqe_v1(eq, eq->cons_index); next_aeqe_sw_v1() local 3866 struct hns_roce_aeqe *aeqe; hns_roce_v1_aeq_int() local [all...] |
H A D | hns_roce_hw_v2.c | 5479 struct hns_roce_aeqe *aeqe; in next_aeqe_sw_v2() local 5481 aeqe = hns_roce_buf_offset(eq->mtr.kmem, in next_aeqe_sw_v2() 5485 return (roce_get_bit(aeqe->asyn, HNS_ROCE_V2_AEQ_AEQE_OWNER_S) ^ in next_aeqe_sw_v2() 5486 !!(eq->cons_index & eq->entries)) ? aeqe : NULL; in next_aeqe_sw_v2() 5493 struct hns_roce_aeqe *aeqe = next_aeqe_sw_v2(eq); in hns_roce_v2_aeq_int() local 5501 while (aeqe) { in hns_roce_v2_aeq_int() 5507 event_type = roce_get_field(aeqe->asyn, in hns_roce_v2_aeq_int() 5510 sub_type = roce_get_field(aeqe->asyn, in hns_roce_v2_aeq_int() 5513 qpn = roce_get_field(aeqe->event.qp_event.qp, in hns_roce_v2_aeq_int() 5516 cqn = roce_get_field(aeqe in hns_roce_v2_aeq_int() [all...] |
/kernel/linux/linux-5.10/drivers/crypto/hisilicon/ |
H A D | qm.c | 80 #define QM_AEQE_PHASE(aeqe) ((le32_to_cpu((aeqe)->dw0) >> 16) & 0x1) 706 struct qm_aeqe *aeqe = qm->aeqe + qm->status.aeq_head; in qm_aeq_irq() local 713 while (QM_AEQE_PHASE(aeqe) == qm->status.aeqc_phase) { in qm_aeq_irq() 714 type = le32_to_cpu(aeqe->dw0) >> QM_AEQE_TYPE_SHIFT; in qm_aeq_irq() 724 aeqe = qm->aeqe; in qm_aeq_irq() 727 aeqe++; in qm_aeq_irq() 1373 dev_err(dev, "Please input aeqe nu in qm_eq_aeq_dump() [all...] |
H A D | qm.h | 219 struct qm_aeqe *aeqe; member
|
/kernel/linux/linux-5.10/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_hw_eqs.c | 594 struct hinic_aeq_elem *aeqe; in aeq_elements_init() local 598 aeqe = GET_AEQ_ELEM(eq, i); in aeq_elements_init() 599 aeqe->desc = cpu_to_be32(init_val); in aeq_elements_init()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_hw_eqs.c | 594 struct hinic_aeq_elem *aeqe; in aeq_elements_init() local 598 aeqe = GET_AEQ_ELEM(eq, i); in aeq_elements_init() 599 aeqe->desc = cpu_to_be32(init_val); in aeq_elements_init()
|
/kernel/linux/linux-6.6/drivers/crypto/hisilicon/ |
H A D | qm.c | 70 #define QM_AEQE_PHASE(aeqe) ((le32_to_cpu((aeqe)->dw0) >> 16) & 0x1) 1058 struct qm_aeqe *aeqe = qm->aeqe + qm->status.aeq_head; in qm_aeq_thread() local 1064 while (QM_AEQE_PHASE(aeqe) == qm->status.aeqc_phase) { in qm_aeq_thread() 1065 type = le32_to_cpu(aeqe->dw0) >> QM_AEQE_TYPE_SHIFT; in qm_aeq_thread() 1066 qp_id = le32_to_cpu(aeqe->dw0) & QM_AEQE_CQN_MASK; in qm_aeq_thread() 1088 aeqe = qm->aeqe; in qm_aeq_thread() 1091 aeqe in qm_aeq_thread() [all...] |
H A D | debugfs.c | 368 dev_err(dev, "Please input eqe or aeqe num (0-%u)", xeq_depth - 1); in qm_eq_aeq_dump() 376 } else if (qm->aeqe && !strcmp(name, "AEQE")) { in qm_eq_aeq_dump() 377 xeqe = qm->aeqe + xeqe_id; in qm_eq_aeq_dump()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/bnxt_re/ |
H A D | main.c | 873 void *aeqe, void *obj) in bnxt_re_aeq_handler() 880 type = ((struct creq_base *)aeqe)->type; in bnxt_re_aeq_handler() 882 unaffi_async = aeqe; in bnxt_re_aeq_handler() 885 affi_async = aeqe; in bnxt_re_aeq_handler() 872 bnxt_re_aeq_handler(struct bnxt_qplib_rcfw *rcfw, void *aeqe, void *obj) bnxt_re_aeq_handler() argument
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/bnxt_re/ |
H A D | main.c | 1015 void *aeqe, void *obj) in bnxt_re_aeq_handler() 1022 type = ((struct creq_base *)aeqe)->type; in bnxt_re_aeq_handler() 1024 unaffi_async = aeqe; in bnxt_re_aeq_handler() 1027 affi_async = aeqe; in bnxt_re_aeq_handler() 1014 bnxt_re_aeq_handler(struct bnxt_qplib_rcfw *rcfw, void *aeqe, void *obj) bnxt_re_aeq_handler() argument
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/hns/ |
H A D | hns_roce_hw_v2.c | 5917 struct hns_roce_aeqe *aeqe; in next_aeqe_sw_v2() local 5919 aeqe = hns_roce_buf_offset(eq->mtr.kmem, in next_aeqe_sw_v2() 5923 return (hr_reg_read(aeqe, AEQE_OWNER) ^ in next_aeqe_sw_v2() 5924 !!(eq->cons_index & eq->entries)) ? aeqe : NULL; in next_aeqe_sw_v2() 5931 struct hns_roce_aeqe *aeqe = next_aeqe_sw_v2(eq); in hns_roce_v2_aeq_int() local 5937 while (aeqe) { in hns_roce_v2_aeq_int() 5943 event_type = hr_reg_read(aeqe, AEQE_EVENT_TYPE); in hns_roce_v2_aeq_int() 5944 sub_type = hr_reg_read(aeqe, AEQE_SUB_TYPE); in hns_roce_v2_aeq_int() 5945 queue_num = hr_reg_read(aeqe, AEQE_EVENT_QUEUE_NUM); in hns_roce_v2_aeq_int() 5970 le16_to_cpu(aeqe in hns_roce_v2_aeq_int() [all...] |
/kernel/linux/linux-6.6/include/linux/ |
H A D | hisi_acc_qm.h | 338 struct qm_aeqe *aeqe; member
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/i40iw/ |
H A D | i40iw_ctrl.c | 1857 * @info: aeqe info to be returned 1863 u64 *aeqe; in i40iw_sc_get_next_aeqe() local 1868 aeqe = (u64 *)I40IW_GET_CURRENT_AEQ_ELEMENT(aeq); in i40iw_sc_get_next_aeqe() 1869 get_64bit_val(aeqe, 0, &compl_ctx); in i40iw_sc_get_next_aeqe() 1870 get_64bit_val(aeqe, 8, &temp); in i40iw_sc_get_next_aeqe() 1876 i40iw_debug_buf(aeq->dev, I40IW_DEBUG_WQE, "AEQ_ENTRY", aeqe, 16); in i40iw_sc_get_next_aeqe()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/irdma/ |
H A D | ctrl.c | 4003 * @info: aeqe info to be returned 4009 __le64 *aeqe; in irdma_sc_get_next_aeqe() local 4013 aeqe = IRDMA_GET_CURRENT_AEQ_ELEM(aeq); in irdma_sc_get_next_aeqe() 4014 get_64bit_val(aeqe, 8, &temp); in irdma_sc_get_next_aeqe() 4023 get_64bit_val(aeqe, 0, &compl_ctx); in irdma_sc_get_next_aeqe() 4026 aeqe, 16, false); in irdma_sc_get_next_aeqe()
|