/kernel/linux/linux-5.10/drivers/infiniband/hw/i40iw/ |
H A D | i40iw_ctrl.c | 48 * @wqe: cqp wqe for header 49 * @header: header for the cqp wqe 69 * i40iw_get_cqp_reg_info - get head and tail for cqp using registers 70 * @cqp: struct for cqp hw 71 * @val: cqp tail register value 73 * @error: cqp processing err 75 static inline void i40iw_get_cqp_reg_info(struct i40iw_sc_cqp *cqp, in i40iw_get_cqp_reg_info() argument 80 if (cqp->dev->is_pf) { in i40iw_get_cqp_reg_info() 81 *val = i40iw_rd32(cqp in i40iw_get_cqp_reg_info() 97 i40iw_cqp_poll_registers( struct i40iw_sc_cqp *cqp, u32 tail, u32 count) i40iw_cqp_poll_registers() argument 455 i40iw_sc_cqp_init(struct i40iw_sc_cqp *cqp, struct i40iw_cqp_init_info *info) i40iw_sc_cqp_init() argument 503 i40iw_sc_cqp_create(struct i40iw_sc_cqp *cqp, u16 *maj_err, u16 *min_err) i40iw_sc_cqp_create() argument 582 i40iw_sc_cqp_post_sq(struct i40iw_sc_cqp *cqp) i40iw_sc_cqp_post_sq() argument 604 i40iw_sc_cqp_get_next_send_wqe_idx(struct i40iw_sc_cqp *cqp, u64 scratch, u32 *wqe_idx) i40iw_sc_cqp_get_next_send_wqe_idx() argument 639 i40iw_sc_cqp_get_next_send_wqe(struct i40iw_sc_cqp *cqp, u64 scratch) i40iw_sc_cqp_get_next_send_wqe() argument 650 i40iw_sc_cqp_destroy(struct i40iw_sc_cqp *cqp) i40iw_sc_cqp_destroy() argument 725 struct i40iw_sc_cqp *cqp; i40iw_sc_ccq_get_cqe_info() local 780 i40iw_sc_poll_for_cqp_op_done( struct i40iw_sc_cqp *cqp, u8 op_code, struct i40iw_ccq_cqe_info *compl_info) i40iw_sc_poll_for_cqp_op_done() argument 829 i40iw_sc_manage_push_page( struct i40iw_sc_cqp *cqp, struct i40iw_cqp_manage_push_page_info *info, u64 scratch, bool post_sq) i40iw_sc_manage_push_page() argument 870 i40iw_sc_manage_hmc_pm_func_table( struct i40iw_sc_cqp *cqp, u64 scratch, u8 vf_index, bool free_pm_fcn, bool post_sq) i40iw_sc_manage_hmc_pm_func_table() argument 908 i40iw_sc_set_hmc_resource_profile( struct i40iw_sc_cqp *cqp, u64 scratch, u8 hmc_profile_type, u8 vf_num, bool post_sq, bool poll_registers) i40iw_sc_set_hmc_resource_profile() argument 957 i40iw_sc_manage_hmc_pm_func_table_done(struct i40iw_sc_cqp *cqp) i40iw_sc_manage_hmc_pm_func_table_done() argument 966 i40iw_sc_commit_fpm_values_done(struct i40iw_sc_cqp *cqp) i40iw_sc_commit_fpm_values_done() argument 980 i40iw_sc_commit_fpm_values( struct i40iw_sc_cqp *cqp, u64 scratch, u8 hmc_fn_id, struct i40iw_dma_mem *commit_fpm_mem, bool post_sq, u8 wait_type) i40iw_sc_commit_fpm_values() argument 1029 i40iw_sc_query_rdma_features_done(struct i40iw_sc_cqp *cqp) i40iw_sc_query_rdma_features_done() argument 1042 i40iw_sc_query_rdma_features(struct i40iw_sc_cqp *cqp, struct i40iw_dma_mem *feat_mem, u64 scratch) i40iw_sc_query_rdma_features() argument 1117 i40iw_sc_query_fpm_values_done(struct i40iw_sc_cqp *cqp) i40iw_sc_query_fpm_values_done() argument 1131 i40iw_sc_query_fpm_values( struct i40iw_sc_cqp *cqp, u64 scratch, u8 hmc_fn_id, struct i40iw_dma_mem *query_fpm_mem, bool post_sq, u8 wait_type) i40iw_sc_query_fpm_values() argument 1183 i40iw_sc_add_arp_cache_entry( struct i40iw_sc_cqp *cqp, struct i40iw_add_arp_cache_entry_info *info, u64 scratch, bool post_sq) i40iw_sc_add_arp_cache_entry() argument 1229 i40iw_sc_del_arp_cache_entry( struct i40iw_sc_cqp *cqp, u64 scratch, u16 arp_index, bool post_sq) i40iw_sc_del_arp_cache_entry() argument 1262 i40iw_sc_query_arp_cache_entry( struct i40iw_sc_cqp *cqp, u64 scratch, u16 arp_index, bool post_sq) i40iw_sc_query_arp_cache_entry() argument 1297 i40iw_sc_manage_apbvt_entry( struct i40iw_sc_cqp *cqp, struct i40iw_apbvt_info *info, u64 scratch, bool post_sq) i40iw_sc_manage_apbvt_entry() argument 1342 i40iw_sc_manage_qhash_table_entry( struct i40iw_sc_cqp *cqp, struct i40iw_qhash_table_info *info, u64 scratch, bool post_sq) i40iw_sc_manage_qhash_table_entry() argument 1430 i40iw_sc_alloc_local_mac_ipaddr_entry( struct i40iw_sc_cqp *cqp, u64 scratch, bool post_sq) i40iw_sc_alloc_local_mac_ipaddr_entry() argument 1459 i40iw_sc_add_local_mac_ipaddr_entry( struct i40iw_sc_cqp *cqp, struct i40iw_local_mac_ipaddr_entry_info *info, u64 scratch, bool post_sq) i40iw_sc_add_local_mac_ipaddr_entry() argument 1502 i40iw_sc_del_local_mac_ipaddr_entry( struct i40iw_sc_cqp *cqp, u64 scratch, u8 entry_idx, u8 ignore_ref_count, bool post_sq) i40iw_sc_del_local_mac_ipaddr_entry() argument 1537 i40iw_sc_cqp_nop(struct i40iw_sc_cqp *cqp, u64 scratch, bool post_sq) i40iw_sc_cqp_nop() argument 1611 struct i40iw_sc_cqp *cqp; i40iw_sc_ceq_create() local 1647 struct i40iw_sc_cqp *cqp; i40iw_sc_cceq_create_done() local 1659 struct i40iw_sc_cqp *cqp; i40iw_sc_cceq_destroy_done() local 1691 struct i40iw_sc_cqp *cqp; i40iw_sc_ceq_destroy() local 1795 struct i40iw_sc_cqp *cqp; i40iw_sc_aeq_create() local 1832 struct i40iw_sc_cqp *cqp; i40iw_sc_aeq_destroy() local 1999 struct i40iw_sc_cqp *cqp; i40iw_sc_aeq_create_done() local 2011 struct i40iw_sc_cqp *cqp; i40iw_sc_aeq_destroy_done() local 2074 struct i40iw_sc_cqp *cqp; i40iw_sc_ccq_create_done() local 2093 struct i40iw_sc_cqp *cqp; i40iw_sc_ccq_create() local 2150 struct i40iw_sc_cqp *cqp; i40iw_sc_ccq_destroy() local 2251 struct i40iw_sc_cqp *cqp; i40iw_sc_cq_create() local 2309 struct i40iw_sc_cqp *cqp; i40iw_sc_cq_destroy() local 2355 struct i40iw_sc_cqp *cqp; i40iw_sc_cq_modify() local 2532 struct i40iw_sc_cqp *cqp; i40iw_sc_qp_create() local 2583 struct i40iw_sc_cqp *cqp; i40iw_sc_qp_modify() local 2651 struct i40iw_sc_cqp *cqp; i40iw_sc_qp_destroy() local 2693 struct i40iw_sc_cqp *cqp; i40iw_sc_qp_flush_wqes() local 2762 struct i40iw_sc_cqp *cqp; i40iw_sc_gen_ae() local 2804 struct i40iw_sc_cqp *cqp; i40iw_sc_qp_upload_context() local 3032 struct i40iw_sc_cqp *cqp; i40iw_sc_alloc_stag() local 3090 struct i40iw_sc_cqp *cqp; i40iw_sc_mr_reg_non_shared() local 3175 struct i40iw_sc_cqp *cqp; i40iw_sc_mr_reg_shared() local 3239 struct i40iw_sc_cqp *cqp; i40iw_sc_dealloc_stag() local 3280 struct i40iw_sc_cqp *cqp; i40iw_sc_query_stag() local 3319 struct i40iw_sc_cqp *cqp; i40iw_sc_mw_alloc() local 3727 cqp_sds_wqe_fill(struct i40iw_sc_cqp *cqp, struct i40iw_update_sds_info *info, u64 scratch) cqp_sds_wqe_fill() argument 3805 struct i40iw_sc_cqp *cqp = dev->cqp; i40iw_update_pe_sds() local 3824 struct i40iw_sc_cqp *cqp = dev->cqp; i40iw_update_sds_noccq() local 3846 i40iw_sc_suspend_qp(struct i40iw_sc_cqp *cqp, struct i40iw_sc_qp *qp, u64 scratch) i40iw_sc_suspend_qp() argument 3875 i40iw_sc_resume_qp(struct i40iw_sc_cqp *cqp, struct i40iw_sc_qp *qp, u64 scratch) i40iw_sc_resume_qp() argument 3910 i40iw_sc_static_hmc_pages_allocated( struct i40iw_sc_cqp *cqp, u64 scratch, u8 hmc_fn_id, bool post_sq, bool poll_registers) i40iw_sc_static_hmc_pages_allocated() argument 3959 i40iw_ring_full(struct i40iw_sc_cqp *cqp) i40iw_ring_full() argument [all...] |
H A D | i40iw_utils.c | 356 * i40iw_get_cqp_request - get cqp struct 357 * @cqp: device cqp ptr 358 * @wait: cqp to be used in wait mode 360 struct i40iw_cqp_request *i40iw_get_cqp_request(struct i40iw_cqp *cqp, bool wait) in i40iw_get_cqp_request() argument 365 spin_lock_irqsave(&cqp->req_lock, flags); in i40iw_get_cqp_request() 366 if (!list_empty(&cqp->cqp_avail_reqs)) { in i40iw_get_cqp_request() 367 cqp_request = list_entry(cqp->cqp_avail_reqs.next, in i40iw_get_cqp_request() 371 spin_unlock_irqrestore(&cqp->req_lock, flags); in i40iw_get_cqp_request() 395 * i40iw_free_cqp_request - free cqp reques 399 i40iw_free_cqp_request(struct i40iw_cqp *cqp, struct i40iw_cqp_request *cqp_request) i40iw_free_cqp_request() argument 423 i40iw_put_cqp_request(struct i40iw_cqp *cqp, struct i40iw_cqp_request *cqp_request) i40iw_put_cqp_request() argument 435 i40iw_free_pending_cqp_request(struct i40iw_cqp *cqp, struct i40iw_cqp_request *cqp_request) i40iw_free_pending_cqp_request() argument 458 struct i40iw_cqp *cqp = &iwdev->cqp; i40iw_cleanup_pending_cqp_op() local 824 struct i40iw_sc_cqp *cqp = dev->cqp; i40iw_qp_suspend_resume() local [all...] |
H A D | i40iw_vf.c | 46 * @cqp: cqp for cqp' sq wqe 51 enum i40iw_status_code i40iw_manage_vf_pble_bp(struct i40iw_sc_cqp *cqp, in i40iw_manage_vf_pble_bp() argument 59 wqe = i40iw_sc_cqp_get_next_send_wqe(cqp, scratch); in i40iw_manage_vf_pble_bp() 70 LS_64(cqp->polarity, I40IW_CQPSQ_WQEVALID); in i40iw_manage_vf_pble_bp() 76 i40iw_debug_buf(cqp->dev, I40IW_DEBUG_WQE, "MANAGE VF_PBLE_BP WQE", wqe, I40IW_CQP_WQE_SIZE * 8); in i40iw_manage_vf_pble_bp() 79 i40iw_sc_cqp_post_sq(cqp); in i40iw_manage_vf_pble_bp()
|
H A D | i40iw_p.h | 52 void i40iw_sc_cqp_post_sq(struct i40iw_sc_cqp *cqp); 54 u64 *i40iw_sc_cqp_get_next_send_wqe(struct i40iw_sc_cqp *cqp, u64 scratch); 96 enum i40iw_status_code i40iw_sc_suspend_qp(struct i40iw_sc_cqp *cqp, 99 enum i40iw_status_code i40iw_sc_resume_qp(struct i40iw_sc_cqp *cqp, 102 enum i40iw_status_code i40iw_sc_static_hmc_pages_allocated(struct i40iw_sc_cqp *cqp,
|
H A D | i40iw_main.c | 230 * @create_done: 1 if cqp create poll was success 232 * Issue destroy cqp request and 233 * free the resources associated with the cqp 238 struct i40iw_cqp *cqp = &iwdev->cqp; in i40iw_destroy_cqp() local 241 dev->cqp_ops->cqp_destroy(dev->cqp); in i40iw_destroy_cqp() 245 i40iw_free_dma_mem(dev->hw, &cqp->sq); in i40iw_destroy_cqp() 246 kfree(cqp->scratch_array); in i40iw_destroy_cqp() 247 iwdev->cqp.scratch_array = NULL; in i40iw_destroy_cqp() 249 kfree(cqp in i40iw_destroy_cqp() 560 struct i40iw_cqp *cqp = &iwdev->cqp; i40iw_create_cqp() local [all...] |
H A D | i40iw_hw.c | 117 * i40iw_cqp_ce_handler - handle cqp completions 120 * @cq: cq for cqp completions 148 i40iw_put_cqp_request(&iwdev->cqp, cqp_request); in i40iw_cqp_ce_handler() 152 i40iw_put_cqp_request(&iwdev->cqp, cqp_request); in i40iw_cqp_ce_handler() 437 * i40iw_cqp_manage_abvpt_cmd - send cqp command manage abpvt 452 cqp_request = i40iw_get_cqp_request(&iwdev->cqp, add_port); in i40iw_cqp_manage_abvpt_cmd() 465 cqp_info->in.u.manage_apbvt_entry.cqp = &iwdev->cqp.sc_cqp; in i40iw_cqp_manage_abvpt_cmd() 538 cqp_request = i40iw_get_cqp_request(&iwdev->cqp, false); in i40iw_manage_arp_cache() 551 cqp_info->in.u.add_arp_cache_entry.cqp in i40iw_manage_arp_cache() [all...] |
H A D | i40iw_type.h | 478 spinlock_t cqp_lock; /* cqp list sync */ 492 struct i40iw_sc_cqp *cqp; member 566 struct i40iw_sc_cqp *cqp; member 1228 struct i40iw_sc_cqp *cqp; member 1234 struct i40iw_sc_cqp *cqp; member 1240 struct i40iw_sc_cqp *cqp; member 1247 struct i40iw_sc_cqp *cqp; member 1253 struct i40iw_sc_cqp *cqp; member 1259 struct i40iw_sc_cqp *cqp; member 1271 struct i40iw_sc_cqp *cqp; member 1314 struct i40iw_sc_cqp *cqp; global() member 1322 struct i40iw_sc_cqp *cqp; global() member 1330 struct i40iw_sc_cqp *cqp; global() member 1336 struct i40iw_sc_cqp *cqp; global() member 1348 struct i40iw_sc_cqp *cqp; global() member 1353 struct i40iw_sc_cqp *cqp; global() member [all...] |
H A D | i40iw_puda.c | 518 struct i40iw_sc_cqp *cqp; in i40iw_puda_qp_wqe() local 524 cqp = dev->cqp; in i40iw_puda_qp_wqe() 525 wqe = i40iw_sc_cqp_get_next_send_wqe(cqp, 0); in i40iw_puda_qp_wqe() 536 LS_64(cqp->polarity, I40IW_CQPSQ_WQEVALID); in i40iw_puda_qp_wqe() 540 i40iw_debug_buf(cqp->dev, I40IW_DEBUG_PUDA, "PUDA CQE", wqe, 32); in i40iw_puda_qp_wqe() 541 i40iw_sc_cqp_post_sq(cqp); in i40iw_puda_qp_wqe() 542 status = dev->cqp_ops->poll_for_cqp_op_done(dev->cqp, in i40iw_puda_qp_wqe() 631 struct i40iw_sc_cqp *cqp; in i40iw_puda_cq_wqe() local 636 cqp in i40iw_puda_cq_wqe() [all...] |
H A D | i40iw_vf.h | 55 enum i40iw_status_code i40iw_manage_vf_pble_bp(struct i40iw_sc_cqp *cqp,
|
H A D | i40iw.h | 151 spinlock_t req_lock; /*cqp request list */ 256 struct i40iw_cqp cqp; member 428 struct i40iw_cqp_request *i40iw_get_cqp_request(struct i40iw_cqp *cqp, bool wait); 429 void i40iw_free_cqp_request(struct i40iw_cqp *cqp, struct i40iw_cqp_request *cqp_request); 430 void i40iw_put_cqp_request(struct i40iw_cqp *cqp, struct i40iw_cqp_request *cqp_request);
|
H A D | i40iw_verbs.c | 196 cqp_request = i40iw_get_cqp_request(&iwdev->cqp, true); in i40iw_alloc_push_page() 208 cqp_info->in.u.manage_push_page.cqp = &iwdev->cqp.sc_cqp; in i40iw_alloc_push_page() 216 i40iw_put_cqp_request(&iwdev->cqp, cqp_request); in i40iw_alloc_push_page() 233 cqp_request = i40iw_get_cqp_request(&iwdev->cqp, false); in i40iw_dealloc_push_page() 244 cqp_info->in.u.manage_push_page.cqp = &iwdev->cqp.sc_cqp; in i40iw_dealloc_push_page() 510 struct i40iw_cqp *iwcqp = &iwdev->cqp; in i40iw_create_qp() 769 * i40iw_hw_modify_qp - setup cqp for modify qp 783 cqp_request = i40iw_get_cqp_request(&iwdev->cqp, wai in i40iw_hw_modify_qp() [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/irdma/ |
H A D | uda.h | 39 int irdma_sc_access_ah(struct irdma_sc_cqp *cqp, struct irdma_ah_info *info, 41 int irdma_access_mcast_grp(struct irdma_sc_cqp *cqp, 50 static inline int irdma_sc_create_ah(struct irdma_sc_cqp *cqp, in irdma_sc_create_ah() argument 53 return irdma_sc_access_ah(cqp, info, IRDMA_CQP_OP_CREATE_ADDR_HANDLE, in irdma_sc_create_ah() 57 static inline int irdma_sc_destroy_ah(struct irdma_sc_cqp *cqp, in irdma_sc_destroy_ah() argument 60 return irdma_sc_access_ah(cqp, info, IRDMA_CQP_OP_DESTROY_ADDR_HANDLE, in irdma_sc_destroy_ah() 64 static inline int irdma_sc_create_mcast_grp(struct irdma_sc_cqp *cqp, in irdma_sc_create_mcast_grp() argument 68 return irdma_access_mcast_grp(cqp, info, IRDMA_CQP_OP_CREATE_MCAST_GRP, in irdma_sc_create_mcast_grp() 72 static inline int irdma_sc_modify_mcast_grp(struct irdma_sc_cqp *cqp, in irdma_sc_modify_mcast_grp() argument 76 return irdma_access_mcast_grp(cqp, inf in irdma_sc_modify_mcast_grp() 80 irdma_sc_destroy_mcast_grp(struct irdma_sc_cqp *cqp, struct irdma_mcast_grp_info *info, u64 scratch) irdma_sc_destroy_mcast_grp() argument [all...] |
H A D | ctrl.c | 62 /* issue cqp suspend command */ in irdma_sc_suspend_resume_qps() 176 * irdma_sc_add_arp_cache_entry - cqp wqe add arp cache entry 177 * @cqp: struct for cqp hw 179 * @scratch: u64 saved to be used during cqp completion 180 * @post_sq: flag for cqp db to ring 182 static int irdma_sc_add_arp_cache_entry(struct irdma_sc_cqp *cqp, in irdma_sc_add_arp_cache_entry() argument 189 wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch); in irdma_sc_add_arp_cache_entry() 199 FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity); in irdma_sc_add_arp_cache_entry() 207 irdma_sc_cqp_post_sq(cqp); in irdma_sc_add_arp_cache_entry() 219 irdma_sc_del_arp_cache_entry(struct irdma_sc_cqp *cqp, u64 scratch, u16 arp_index, bool post_sq) irdma_sc_del_arp_cache_entry() argument 252 irdma_sc_manage_apbvt_entry(struct irdma_sc_cqp *cqp, struct irdma_apbvt_info *info, u64 scratch, bool post_sq) irdma_sc_manage_apbvt_entry() argument 300 irdma_sc_manage_qhash_table_entry(struct irdma_sc_cqp *cqp, struct irdma_qhash_table_info *info, u64 scratch, bool post_sq) irdma_sc_manage_qhash_table_entry() argument 451 struct irdma_sc_cqp *cqp; irdma_sc_qp_create() local 503 struct irdma_sc_cqp *cqp; irdma_sc_qp_modify() local 573 struct irdma_sc_cqp *cqp; irdma_sc_qp_destroy() local 761 irdma_sc_alloc_local_mac_entry(struct irdma_sc_cqp *cqp, u64 scratch, bool post_sq) irdma_sc_alloc_local_mac_entry() argument 795 irdma_sc_add_local_mac_entry(struct irdma_sc_cqp *cqp, struct irdma_local_mac_entry_info *info, u64 scratch, bool post_sq) irdma_sc_add_local_mac_entry() argument 833 irdma_sc_del_local_mac_entry(struct irdma_sc_cqp *cqp, u64 scratch, u16 entry_idx, u8 ignore_ref_count, bool post_sq) irdma_sc_del_local_mac_entry() argument 1060 struct irdma_sc_cqp *cqp; irdma_sc_alloc_stag() local 1125 struct irdma_sc_cqp *cqp; irdma_sc_mr_reg_non_shared() local 1216 struct irdma_sc_cqp *cqp; irdma_sc_dealloc_stag() local 1255 struct irdma_sc_cqp *cqp; irdma_sc_mw_alloc() local 2047 irdma_sc_gather_stats(struct irdma_sc_cqp *cqp, struct irdma_stats_gather_info *info, u64 scratch) irdma_sc_gather_stats() argument 2094 irdma_sc_manage_stats_inst(struct irdma_sc_cqp *cqp, struct irdma_stats_inst_info *info, bool alloc, u64 scratch) irdma_sc_manage_stats_inst() argument 2131 irdma_sc_set_up_map(struct irdma_sc_cqp *cqp, struct irdma_up_info *info, u64 scratch) irdma_sc_set_up_map() argument 2173 irdma_sc_manage_ws_node(struct irdma_sc_cqp *cqp, struct irdma_ws_node_info *info, enum irdma_ws_node_op node_op, u64 scratch) irdma_sc_manage_ws_node() argument 2221 struct irdma_sc_cqp *cqp; irdma_sc_qp_flush_wqes() local 2295 struct irdma_sc_cqp *cqp; irdma_sc_gen_ae() local 2334 struct irdma_sc_cqp *cqp; irdma_sc_qp_upload_context() local 2369 irdma_sc_manage_push_page(struct irdma_sc_cqp *cqp, struct irdma_cqp_manage_push_page_info *info, u64 scratch, bool post_sq) irdma_sc_manage_push_page() argument 2408 irdma_sc_suspend_qp(struct irdma_sc_cqp *cqp, struct irdma_sc_qp *qp, u64 scratch) irdma_sc_suspend_qp() argument 2438 irdma_sc_resume_qp(struct irdma_sc_cqp *cqp, struct irdma_sc_qp *qp, u64 scratch) irdma_sc_resume_qp() argument 2520 struct irdma_sc_cqp *cqp; irdma_sc_cq_create() local 2591 struct irdma_sc_cqp *cqp; irdma_sc_cq_destroy() local 2659 struct irdma_sc_cqp *cqp; irdma_sc_cq_modify() local 2732 irdma_get_cqp_reg_info(struct irdma_sc_cqp *cqp, u32 *val, u32 *tail, u32 *error) irdma_get_cqp_reg_info() argument 2746 irdma_cqp_poll_registers(struct irdma_sc_cqp *cqp, u32 tail, u32 count) irdma_cqp_poll_registers() argument 3088 irdma_sc_cqp_init(struct irdma_sc_cqp *cqp, struct irdma_cqp_init_info *info) irdma_sc_cqp_init() argument 3136 (u64 *)(uintptr_t)cqp->sq_pa, cqp, cqp->polarity); irdma_sc_cqp_init() local 3146 irdma_sc_cqp_create(struct irdma_sc_cqp *cqp, u16 *maj_err, u16 *min_err) irdma_sc_cqp_create() argument 3244 irdma_sc_cqp_post_sq(struct irdma_sc_cqp *cqp) irdma_sc_cqp_post_sq() argument 3260 irdma_sc_cqp_get_next_send_wqe_idx(struct irdma_sc_cqp *cqp, u64 scratch, u32 *wqe_idx) irdma_sc_cqp_get_next_send_wqe_idx() argument 3291 irdma_sc_cqp_destroy(struct irdma_sc_cqp *cqp) irdma_sc_cqp_destroy() argument 3350 struct irdma_sc_cqp *cqp; irdma_sc_ccq_get_cqe_info() local 3414 irdma_sc_poll_for_cqp_op_done(struct irdma_sc_cqp *cqp, u8 op_code, struct irdma_ccq_cqe_info *compl_info) irdma_sc_poll_for_cqp_op_done() argument 3456 irdma_sc_manage_hmc_pm_func_table(struct irdma_sc_cqp *cqp, struct irdma_hmc_fcn_info *info, u64 scratch, bool post_sq) irdma_sc_manage_hmc_pm_func_table() argument 3498 irdma_sc_commit_fpm_val_done(struct irdma_sc_cqp *cqp) irdma_sc_commit_fpm_val_done() argument 3513 irdma_sc_commit_fpm_val(struct irdma_sc_cqp *cqp, u64 scratch, u8 hmc_fn_id, struct irdma_dma_mem *commit_fpm_mem, bool post_sq, u8 wait_type) irdma_sc_commit_fpm_val() argument 3559 irdma_sc_query_fpm_val_done(struct irdma_sc_cqp *cqp) irdma_sc_query_fpm_val_done() argument 3574 irdma_sc_query_fpm_val(struct irdma_sc_cqp *cqp, u64 scratch, u8 hmc_fn_id, struct irdma_dma_mem *query_fpm_mem, bool post_sq, u8 wait_type) irdma_sc_query_fpm_val() argument 3668 struct irdma_sc_cqp *cqp; irdma_sc_ceq_create() local 3709 struct irdma_sc_cqp *cqp; irdma_sc_cceq_create_done() local 3722 struct irdma_sc_cqp *cqp; irdma_sc_cceq_destroy_done() local 3766 struct irdma_sc_cqp *cqp; irdma_sc_ceq_destroy() local 3933 struct irdma_sc_cqp *cqp; irdma_sc_aeq_create() local 3972 struct irdma_sc_cqp *cqp; irdma_sc_aeq_destroy() local 4219 struct irdma_sc_cqp *cqp; irdma_sc_ccq_create_done() local 4260 struct irdma_sc_cqp *cqp; irdma_sc_ccq_destroy() local 4413 cqp_sds_wqe_fill(struct irdma_sc_cqp *cqp, struct irdma_update_sds_info *info, u64 scratch) cqp_sds_wqe_fill() argument 4495 struct irdma_sc_cqp *cqp = dev->cqp; irdma_update_pe_sds() local 4514 struct irdma_sc_cqp *cqp = dev->cqp; irdma_update_sds_noccq() local 4536 irdma_sc_static_hmc_pages_allocated(struct irdma_sc_cqp *cqp, u64 scratch, u8 hmc_fn_id, bool post_sq, bool poll_registers) irdma_sc_static_hmc_pages_allocated() argument 4582 irdma_cqp_ring_full(struct irdma_sc_cqp *cqp) irdma_cqp_ring_full() argument 4621 irdma_sc_query_rdma_features_done(struct irdma_sc_cqp *cqp) irdma_sc_query_rdma_features_done() argument 4634 irdma_sc_query_rdma_features(struct irdma_sc_cqp *cqp, struct irdma_dma_mem *buf, u64 scratch) irdma_sc_query_rdma_features() argument [all...] |
H A D | hw.c | 587 * Issue destroy cqp request and 588 * free the resources associated with the cqp 593 struct irdma_cqp *cqp = &rf->cqp; in irdma_destroy_cqp() local 596 status = irdma_sc_cqp_destroy(dev->cqp); in irdma_destroy_cqp() 601 dma_free_coherent(dev->hw->device, cqp->sq.size, cqp->sq.va, in irdma_destroy_cqp() 602 cqp->sq.pa); in irdma_destroy_cqp() 603 cqp->sq.va = NULL; in irdma_destroy_cqp() 604 kfree(cqp in irdma_destroy_cqp() 933 struct irdma_cqp *cqp = &rf->cqp; irdma_create_cqp() local [all...] |
H A D | utils.c | 425 * irdma_alloc_and_get_cqp_request - get cqp struct 426 * @cqp: device cqp ptr 427 * @wait: cqp to be used in wait mode 429 struct irdma_cqp_request *irdma_alloc_and_get_cqp_request(struct irdma_cqp *cqp, in irdma_alloc_and_get_cqp_request() argument 435 spin_lock_irqsave(&cqp->req_lock, flags); in irdma_alloc_and_get_cqp_request() 436 if (!list_empty(&cqp->cqp_avail_reqs)) { in irdma_alloc_and_get_cqp_request() 437 cqp_request = list_first_entry(&cqp->cqp_avail_reqs, in irdma_alloc_and_get_cqp_request() 441 spin_unlock_irqrestore(&cqp->req_lock, flags); in irdma_alloc_and_get_cqp_request() 451 ibdev_dbg(to_ibdev(cqp in irdma_alloc_and_get_cqp_request() 476 irdma_free_cqp_request(struct irdma_cqp *cqp, struct irdma_cqp_request *cqp_request) irdma_free_cqp_request() argument 500 irdma_put_cqp_request(struct irdma_cqp *cqp, struct irdma_cqp_request *cqp_request) irdma_put_cqp_request() argument 513 irdma_free_pending_cqp_request(struct irdma_cqp *cqp, struct irdma_cqp_request *cqp_request) irdma_free_pending_cqp_request() argument 534 struct irdma_cqp *cqp = &rf->cqp; irdma_cleanup_pending_cqp_op() local 867 struct irdma_sc_cqp *cqp = dev->cqp; irdma_cqp_qp_suspend_resume() local 1897 struct irdma_sc_cqp *cqp = &iwcqp->sc_cqp; irdma_cqp_ws_node_cmd() local [all...] |
H A D | uda.c | 15 * @cqp: struct for cqp hw 18 * @scratch: u64 saved to be used during cqp completion 20 int irdma_sc_access_ah(struct irdma_sc_cqp *cqp, struct irdma_ah_info *info, in irdma_sc_access_ah() argument 26 wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch); in irdma_sc_access_ah() 69 FIELD_PREP(IRDMA_UDA_CQPSQ_MAV_WQEVALID, cqp->polarity) | in irdma_sc_access_ah() 78 irdma_sc_cqp_post_sq(cqp); in irdma_sc_access_ah() 110 * @cqp: Control QP 113 * @scratch: u64 saved to be used during cqp completion 115 int irdma_access_mcast_grp(struct irdma_sc_cqp *cqp, in irdma_access_mcast_grp() argument [all...] |
H A D | type.h | 604 struct irdma_sc_cqp *cqp; member 668 struct irdma_sc_cqp *cqp; member 1192 int irdma_sc_cqp_create(struct irdma_sc_cqp *cqp, u16 *maj_err, u16 *min_err); 1193 int irdma_sc_cqp_destroy(struct irdma_sc_cqp *cqp); 1194 int irdma_sc_cqp_init(struct irdma_sc_cqp *cqp, 1196 void irdma_sc_cqp_post_sq(struct irdma_sc_cqp *cqp); 1197 int irdma_sc_poll_for_cqp_op_done(struct irdma_sc_cqp *cqp, u8 opcode, 1224 int irdma_sc_static_hmc_pages_allocated(struct irdma_sc_cqp *cqp, u64 scratch, 1292 struct irdma_sc_cqp *cqp; member 1298 struct irdma_sc_cqp *cqp; member 1304 struct irdma_sc_cqp *cqp; global() member 1310 struct irdma_sc_cqp *cqp; global() member 1317 struct irdma_sc_cqp *cqp; global() member 1322 struct irdma_sc_cqp *cqp; global() member 1372 struct irdma_sc_cqp *cqp; global() member 1380 struct irdma_sc_cqp *cqp; global() member 1388 struct irdma_sc_cqp *cqp; global() member 1394 struct irdma_sc_cqp *cqp; global() member 1406 struct irdma_sc_cqp *cqp; global() member 1412 struct irdma_sc_cqp *cqp; global() member 1418 struct irdma_sc_cqp *cqp; global() member 1424 struct irdma_sc_cqp *cqp; global() member 1430 struct irdma_sc_cqp *cqp; global() member 1436 struct irdma_sc_cqp *cqp; global() member 1442 struct irdma_sc_cqp *cqp; global() member 1448 struct irdma_sc_cqp *cqp; global() member 1454 struct irdma_sc_cqp *cqp; global() member 1460 struct irdma_sc_cqp *cqp; global() member 1466 struct irdma_sc_cqp *cqp; global() member 1488 irdma_sc_cqp_get_next_send_wqe(struct irdma_sc_cqp *cqp, u64 scratch) irdma_sc_cqp_get_next_send_wqe() argument [all...] |
H A D | protos.h | 18 void irdma_sc_cqp_post_sq(struct irdma_sc_cqp *cqp); 19 __le64 *irdma_sc_cqp_get_next_send_wqe(struct irdma_sc_cqp *cqp, u64 scratch); 77 int irdma_sc_static_hmc_pages_allocated(struct irdma_sc_cqp *cqp, u64 scratch,
|
H A D | puda.c | 607 struct irdma_sc_cqp *cqp; in irdma_puda_qp_wqe() local 613 cqp = dev->cqp; in irdma_puda_qp_wqe() 614 wqe = irdma_sc_cqp_get_next_send_wqe(cqp, 0); in irdma_puda_qp_wqe() 626 FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity); in irdma_puda_qp_wqe() 633 irdma_sc_cqp_post_sq(cqp); in irdma_puda_qp_wqe() 634 status = irdma_sc_poll_for_cqp_op_done(dev->cqp, IRDMA_CQP_OP_CREATE_QP, in irdma_puda_qp_wqe() 726 struct irdma_sc_cqp *cqp; in irdma_puda_cq_wqe() local 731 cqp = dev->cqp; in irdma_puda_cq_wqe() [all...] |
H A D | main.h | 303 struct irdma_cqp cqp; member 482 struct irdma_cqp_request *irdma_alloc_and_get_cqp_request(struct irdma_cqp *cqp, 484 void irdma_free_cqp_request(struct irdma_cqp *cqp, 486 void irdma_put_cqp_request(struct irdma_cqp *cqp,
|
H A D | verbs.c | 237 cqp_request = irdma_alloc_and_get_cqp_request(&iwdev->rf->cqp, true); in irdma_alloc_push_page() 249 cqp_info->in.u.manage_push_page.cqp = &iwdev->rf->cqp.sc_cqp; in irdma_alloc_push_page() 259 irdma_put_cqp_request(&iwdev->rf->cqp, cqp_request); in irdma_alloc_push_page() 739 cqp_request = irdma_alloc_and_get_cqp_request(&rf->cqp, true); in irdma_cqp_create_qp_cmd() 755 irdma_put_cqp_request(&rf->cqp, cqp_request); in irdma_cqp_create_qp_cmd() 1977 cqp_request = irdma_alloc_and_get_cqp_request(&rf->cqp, true); in irdma_resize_cq() 1995 irdma_put_cqp_request(&rf->cqp, cqp_request); in irdma_resize_cq() 2197 cqp_request = irdma_alloc_and_get_cqp_request(&rf->cqp, true); in irdma_create_cq() 2210 irdma_put_cqp_request(&rf->cqp, cqp_reques in irdma_create_cq() [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | params.h | 22 struct mlx5e_cq_param cqp; member 29 struct mlx5e_cq_param cqp; member
|
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
H A D | setup.c | 63 err = mlx5e_open_cq(c, params->rx_cq_moderation, &cparam->rq.cqp, &c->xskrq.cq); in mlx5e_open_xsk() 71 err = mlx5e_open_cq(c, params->tx_cq_moderation, &cparam->xdp_sq.cqp, &c->xsksq.cq); in mlx5e_open_xsk()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
H A D | setup.c | 130 err = mlx5e_open_cq(c->priv, params->rx_cq_moderation, &cparam->rq.cqp, &ccp, in mlx5e_open_xsk() 139 err = mlx5e_open_cq(c->priv, params->tx_cq_moderation, &cparam->xdp_sq.cqp, &ccp, in mlx5e_open_xsk()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | params.h | 23 struct mlx5e_cq_param cqp; member 31 struct mlx5e_cq_param cqp; member
|