Home
last modified time | relevance | path

Searched refs:cqp (Results 1 - 25 of 36) sorted by relevance

12

/kernel/linux/linux-5.10/drivers/infiniband/hw/i40iw/
H A Di40iw_ctrl.c48 * @wqe: cqp wqe for header
49 * @header: header for the cqp wqe
69 * i40iw_get_cqp_reg_info - get head and tail for cqp using registers
70 * @cqp: struct for cqp hw
71 * @val: cqp tail register value
73 * @error: cqp processing err
75 static inline void i40iw_get_cqp_reg_info(struct i40iw_sc_cqp *cqp, in i40iw_get_cqp_reg_info() argument
80 if (cqp->dev->is_pf) { in i40iw_get_cqp_reg_info()
81 *val = i40iw_rd32(cqp in i40iw_get_cqp_reg_info()
97 i40iw_cqp_poll_registers( struct i40iw_sc_cqp *cqp, u32 tail, u32 count) i40iw_cqp_poll_registers() argument
455 i40iw_sc_cqp_init(struct i40iw_sc_cqp *cqp, struct i40iw_cqp_init_info *info) i40iw_sc_cqp_init() argument
503 i40iw_sc_cqp_create(struct i40iw_sc_cqp *cqp, u16 *maj_err, u16 *min_err) i40iw_sc_cqp_create() argument
582 i40iw_sc_cqp_post_sq(struct i40iw_sc_cqp *cqp) i40iw_sc_cqp_post_sq() argument
604 i40iw_sc_cqp_get_next_send_wqe_idx(struct i40iw_sc_cqp *cqp, u64 scratch, u32 *wqe_idx) i40iw_sc_cqp_get_next_send_wqe_idx() argument
639 i40iw_sc_cqp_get_next_send_wqe(struct i40iw_sc_cqp *cqp, u64 scratch) i40iw_sc_cqp_get_next_send_wqe() argument
650 i40iw_sc_cqp_destroy(struct i40iw_sc_cqp *cqp) i40iw_sc_cqp_destroy() argument
725 struct i40iw_sc_cqp *cqp; i40iw_sc_ccq_get_cqe_info() local
780 i40iw_sc_poll_for_cqp_op_done( struct i40iw_sc_cqp *cqp, u8 op_code, struct i40iw_ccq_cqe_info *compl_info) i40iw_sc_poll_for_cqp_op_done() argument
829 i40iw_sc_manage_push_page( struct i40iw_sc_cqp *cqp, struct i40iw_cqp_manage_push_page_info *info, u64 scratch, bool post_sq) i40iw_sc_manage_push_page() argument
870 i40iw_sc_manage_hmc_pm_func_table( struct i40iw_sc_cqp *cqp, u64 scratch, u8 vf_index, bool free_pm_fcn, bool post_sq) i40iw_sc_manage_hmc_pm_func_table() argument
908 i40iw_sc_set_hmc_resource_profile( struct i40iw_sc_cqp *cqp, u64 scratch, u8 hmc_profile_type, u8 vf_num, bool post_sq, bool poll_registers) i40iw_sc_set_hmc_resource_profile() argument
957 i40iw_sc_manage_hmc_pm_func_table_done(struct i40iw_sc_cqp *cqp) i40iw_sc_manage_hmc_pm_func_table_done() argument
966 i40iw_sc_commit_fpm_values_done(struct i40iw_sc_cqp *cqp) i40iw_sc_commit_fpm_values_done() argument
980 i40iw_sc_commit_fpm_values( struct i40iw_sc_cqp *cqp, u64 scratch, u8 hmc_fn_id, struct i40iw_dma_mem *commit_fpm_mem, bool post_sq, u8 wait_type) i40iw_sc_commit_fpm_values() argument
1029 i40iw_sc_query_rdma_features_done(struct i40iw_sc_cqp *cqp) i40iw_sc_query_rdma_features_done() argument
1042 i40iw_sc_query_rdma_features(struct i40iw_sc_cqp *cqp, struct i40iw_dma_mem *feat_mem, u64 scratch) i40iw_sc_query_rdma_features() argument
1117 i40iw_sc_query_fpm_values_done(struct i40iw_sc_cqp *cqp) i40iw_sc_query_fpm_values_done() argument
1131 i40iw_sc_query_fpm_values( struct i40iw_sc_cqp *cqp, u64 scratch, u8 hmc_fn_id, struct i40iw_dma_mem *query_fpm_mem, bool post_sq, u8 wait_type) i40iw_sc_query_fpm_values() argument
1183 i40iw_sc_add_arp_cache_entry( struct i40iw_sc_cqp *cqp, struct i40iw_add_arp_cache_entry_info *info, u64 scratch, bool post_sq) i40iw_sc_add_arp_cache_entry() argument
1229 i40iw_sc_del_arp_cache_entry( struct i40iw_sc_cqp *cqp, u64 scratch, u16 arp_index, bool post_sq) i40iw_sc_del_arp_cache_entry() argument
1262 i40iw_sc_query_arp_cache_entry( struct i40iw_sc_cqp *cqp, u64 scratch, u16 arp_index, bool post_sq) i40iw_sc_query_arp_cache_entry() argument
1297 i40iw_sc_manage_apbvt_entry( struct i40iw_sc_cqp *cqp, struct i40iw_apbvt_info *info, u64 scratch, bool post_sq) i40iw_sc_manage_apbvt_entry() argument
1342 i40iw_sc_manage_qhash_table_entry( struct i40iw_sc_cqp *cqp, struct i40iw_qhash_table_info *info, u64 scratch, bool post_sq) i40iw_sc_manage_qhash_table_entry() argument
1430 i40iw_sc_alloc_local_mac_ipaddr_entry( struct i40iw_sc_cqp *cqp, u64 scratch, bool post_sq) i40iw_sc_alloc_local_mac_ipaddr_entry() argument
1459 i40iw_sc_add_local_mac_ipaddr_entry( struct i40iw_sc_cqp *cqp, struct i40iw_local_mac_ipaddr_entry_info *info, u64 scratch, bool post_sq) i40iw_sc_add_local_mac_ipaddr_entry() argument
1502 i40iw_sc_del_local_mac_ipaddr_entry( struct i40iw_sc_cqp *cqp, u64 scratch, u8 entry_idx, u8 ignore_ref_count, bool post_sq) i40iw_sc_del_local_mac_ipaddr_entry() argument
1537 i40iw_sc_cqp_nop(struct i40iw_sc_cqp *cqp, u64 scratch, bool post_sq) i40iw_sc_cqp_nop() argument
1611 struct i40iw_sc_cqp *cqp; i40iw_sc_ceq_create() local
1647 struct i40iw_sc_cqp *cqp; i40iw_sc_cceq_create_done() local
1659 struct i40iw_sc_cqp *cqp; i40iw_sc_cceq_destroy_done() local
1691 struct i40iw_sc_cqp *cqp; i40iw_sc_ceq_destroy() local
1795 struct i40iw_sc_cqp *cqp; i40iw_sc_aeq_create() local
1832 struct i40iw_sc_cqp *cqp; i40iw_sc_aeq_destroy() local
1999 struct i40iw_sc_cqp *cqp; i40iw_sc_aeq_create_done() local
2011 struct i40iw_sc_cqp *cqp; i40iw_sc_aeq_destroy_done() local
2074 struct i40iw_sc_cqp *cqp; i40iw_sc_ccq_create_done() local
2093 struct i40iw_sc_cqp *cqp; i40iw_sc_ccq_create() local
2150 struct i40iw_sc_cqp *cqp; i40iw_sc_ccq_destroy() local
2251 struct i40iw_sc_cqp *cqp; i40iw_sc_cq_create() local
2309 struct i40iw_sc_cqp *cqp; i40iw_sc_cq_destroy() local
2355 struct i40iw_sc_cqp *cqp; i40iw_sc_cq_modify() local
2532 struct i40iw_sc_cqp *cqp; i40iw_sc_qp_create() local
2583 struct i40iw_sc_cqp *cqp; i40iw_sc_qp_modify() local
2651 struct i40iw_sc_cqp *cqp; i40iw_sc_qp_destroy() local
2693 struct i40iw_sc_cqp *cqp; i40iw_sc_qp_flush_wqes() local
2762 struct i40iw_sc_cqp *cqp; i40iw_sc_gen_ae() local
2804 struct i40iw_sc_cqp *cqp; i40iw_sc_qp_upload_context() local
3032 struct i40iw_sc_cqp *cqp; i40iw_sc_alloc_stag() local
3090 struct i40iw_sc_cqp *cqp; i40iw_sc_mr_reg_non_shared() local
3175 struct i40iw_sc_cqp *cqp; i40iw_sc_mr_reg_shared() local
3239 struct i40iw_sc_cqp *cqp; i40iw_sc_dealloc_stag() local
3280 struct i40iw_sc_cqp *cqp; i40iw_sc_query_stag() local
3319 struct i40iw_sc_cqp *cqp; i40iw_sc_mw_alloc() local
3727 cqp_sds_wqe_fill(struct i40iw_sc_cqp *cqp, struct i40iw_update_sds_info *info, u64 scratch) cqp_sds_wqe_fill() argument
3805 struct i40iw_sc_cqp *cqp = dev->cqp; i40iw_update_pe_sds() local
3824 struct i40iw_sc_cqp *cqp = dev->cqp; i40iw_update_sds_noccq() local
3846 i40iw_sc_suspend_qp(struct i40iw_sc_cqp *cqp, struct i40iw_sc_qp *qp, u64 scratch) i40iw_sc_suspend_qp() argument
3875 i40iw_sc_resume_qp(struct i40iw_sc_cqp *cqp, struct i40iw_sc_qp *qp, u64 scratch) i40iw_sc_resume_qp() argument
3910 i40iw_sc_static_hmc_pages_allocated( struct i40iw_sc_cqp *cqp, u64 scratch, u8 hmc_fn_id, bool post_sq, bool poll_registers) i40iw_sc_static_hmc_pages_allocated() argument
3959 i40iw_ring_full(struct i40iw_sc_cqp *cqp) i40iw_ring_full() argument
[all...]
H A Di40iw_utils.c356 * i40iw_get_cqp_request - get cqp struct
357 * @cqp: device cqp ptr
358 * @wait: cqp to be used in wait mode
360 struct i40iw_cqp_request *i40iw_get_cqp_request(struct i40iw_cqp *cqp, bool wait) in i40iw_get_cqp_request() argument
365 spin_lock_irqsave(&cqp->req_lock, flags); in i40iw_get_cqp_request()
366 if (!list_empty(&cqp->cqp_avail_reqs)) { in i40iw_get_cqp_request()
367 cqp_request = list_entry(cqp->cqp_avail_reqs.next, in i40iw_get_cqp_request()
371 spin_unlock_irqrestore(&cqp->req_lock, flags); in i40iw_get_cqp_request()
395 * i40iw_free_cqp_request - free cqp reques
399 i40iw_free_cqp_request(struct i40iw_cqp *cqp, struct i40iw_cqp_request *cqp_request) i40iw_free_cqp_request() argument
423 i40iw_put_cqp_request(struct i40iw_cqp *cqp, struct i40iw_cqp_request *cqp_request) i40iw_put_cqp_request() argument
435 i40iw_free_pending_cqp_request(struct i40iw_cqp *cqp, struct i40iw_cqp_request *cqp_request) i40iw_free_pending_cqp_request() argument
458 struct i40iw_cqp *cqp = &iwdev->cqp; i40iw_cleanup_pending_cqp_op() local
824 struct i40iw_sc_cqp *cqp = dev->cqp; i40iw_qp_suspend_resume() local
[all...]
H A Di40iw_vf.c46 * @cqp: cqp for cqp' sq wqe
51 enum i40iw_status_code i40iw_manage_vf_pble_bp(struct i40iw_sc_cqp *cqp, in i40iw_manage_vf_pble_bp() argument
59 wqe = i40iw_sc_cqp_get_next_send_wqe(cqp, scratch); in i40iw_manage_vf_pble_bp()
70 LS_64(cqp->polarity, I40IW_CQPSQ_WQEVALID); in i40iw_manage_vf_pble_bp()
76 i40iw_debug_buf(cqp->dev, I40IW_DEBUG_WQE, "MANAGE VF_PBLE_BP WQE", wqe, I40IW_CQP_WQE_SIZE * 8); in i40iw_manage_vf_pble_bp()
79 i40iw_sc_cqp_post_sq(cqp); in i40iw_manage_vf_pble_bp()
H A Di40iw_p.h52 void i40iw_sc_cqp_post_sq(struct i40iw_sc_cqp *cqp);
54 u64 *i40iw_sc_cqp_get_next_send_wqe(struct i40iw_sc_cqp *cqp, u64 scratch);
96 enum i40iw_status_code i40iw_sc_suspend_qp(struct i40iw_sc_cqp *cqp,
99 enum i40iw_status_code i40iw_sc_resume_qp(struct i40iw_sc_cqp *cqp,
102 enum i40iw_status_code i40iw_sc_static_hmc_pages_allocated(struct i40iw_sc_cqp *cqp,
H A Di40iw_main.c230 * @create_done: 1 if cqp create poll was success
232 * Issue destroy cqp request and
233 * free the resources associated with the cqp
238 struct i40iw_cqp *cqp = &iwdev->cqp; in i40iw_destroy_cqp() local
241 dev->cqp_ops->cqp_destroy(dev->cqp); in i40iw_destroy_cqp()
245 i40iw_free_dma_mem(dev->hw, &cqp->sq); in i40iw_destroy_cqp()
246 kfree(cqp->scratch_array); in i40iw_destroy_cqp()
247 iwdev->cqp.scratch_array = NULL; in i40iw_destroy_cqp()
249 kfree(cqp in i40iw_destroy_cqp()
560 struct i40iw_cqp *cqp = &iwdev->cqp; i40iw_create_cqp() local
[all...]
H A Di40iw_hw.c117 * i40iw_cqp_ce_handler - handle cqp completions
120 * @cq: cq for cqp completions
148 i40iw_put_cqp_request(&iwdev->cqp, cqp_request); in i40iw_cqp_ce_handler()
152 i40iw_put_cqp_request(&iwdev->cqp, cqp_request); in i40iw_cqp_ce_handler()
437 * i40iw_cqp_manage_abvpt_cmd - send cqp command manage abpvt
452 cqp_request = i40iw_get_cqp_request(&iwdev->cqp, add_port); in i40iw_cqp_manage_abvpt_cmd()
465 cqp_info->in.u.manage_apbvt_entry.cqp = &iwdev->cqp.sc_cqp; in i40iw_cqp_manage_abvpt_cmd()
538 cqp_request = i40iw_get_cqp_request(&iwdev->cqp, false); in i40iw_manage_arp_cache()
551 cqp_info->in.u.add_arp_cache_entry.cqp in i40iw_manage_arp_cache()
[all...]
H A Di40iw_type.h478 spinlock_t cqp_lock; /* cqp list sync */
492 struct i40iw_sc_cqp *cqp; member
566 struct i40iw_sc_cqp *cqp; member
1228 struct i40iw_sc_cqp *cqp; member
1234 struct i40iw_sc_cqp *cqp; member
1240 struct i40iw_sc_cqp *cqp; member
1247 struct i40iw_sc_cqp *cqp; member
1253 struct i40iw_sc_cqp *cqp; member
1259 struct i40iw_sc_cqp *cqp; member
1271 struct i40iw_sc_cqp *cqp; member
1314 struct i40iw_sc_cqp *cqp; global() member
1322 struct i40iw_sc_cqp *cqp; global() member
1330 struct i40iw_sc_cqp *cqp; global() member
1336 struct i40iw_sc_cqp *cqp; global() member
1348 struct i40iw_sc_cqp *cqp; global() member
1353 struct i40iw_sc_cqp *cqp; global() member
[all...]
H A Di40iw_puda.c518 struct i40iw_sc_cqp *cqp; in i40iw_puda_qp_wqe() local
524 cqp = dev->cqp; in i40iw_puda_qp_wqe()
525 wqe = i40iw_sc_cqp_get_next_send_wqe(cqp, 0); in i40iw_puda_qp_wqe()
536 LS_64(cqp->polarity, I40IW_CQPSQ_WQEVALID); in i40iw_puda_qp_wqe()
540 i40iw_debug_buf(cqp->dev, I40IW_DEBUG_PUDA, "PUDA CQE", wqe, 32); in i40iw_puda_qp_wqe()
541 i40iw_sc_cqp_post_sq(cqp); in i40iw_puda_qp_wqe()
542 status = dev->cqp_ops->poll_for_cqp_op_done(dev->cqp, in i40iw_puda_qp_wqe()
631 struct i40iw_sc_cqp *cqp; in i40iw_puda_cq_wqe() local
636 cqp in i40iw_puda_cq_wqe()
[all...]
H A Di40iw_vf.h55 enum i40iw_status_code i40iw_manage_vf_pble_bp(struct i40iw_sc_cqp *cqp,
H A Di40iw.h151 spinlock_t req_lock; /*cqp request list */
256 struct i40iw_cqp cqp; member
428 struct i40iw_cqp_request *i40iw_get_cqp_request(struct i40iw_cqp *cqp, bool wait);
429 void i40iw_free_cqp_request(struct i40iw_cqp *cqp, struct i40iw_cqp_request *cqp_request);
430 void i40iw_put_cqp_request(struct i40iw_cqp *cqp, struct i40iw_cqp_request *cqp_request);
H A Di40iw_verbs.c196 cqp_request = i40iw_get_cqp_request(&iwdev->cqp, true); in i40iw_alloc_push_page()
208 cqp_info->in.u.manage_push_page.cqp = &iwdev->cqp.sc_cqp; in i40iw_alloc_push_page()
216 i40iw_put_cqp_request(&iwdev->cqp, cqp_request); in i40iw_alloc_push_page()
233 cqp_request = i40iw_get_cqp_request(&iwdev->cqp, false); in i40iw_dealloc_push_page()
244 cqp_info->in.u.manage_push_page.cqp = &iwdev->cqp.sc_cqp; in i40iw_dealloc_push_page()
510 struct i40iw_cqp *iwcqp = &iwdev->cqp; in i40iw_create_qp()
769 * i40iw_hw_modify_qp - setup cqp for modify qp
783 cqp_request = i40iw_get_cqp_request(&iwdev->cqp, wai in i40iw_hw_modify_qp()
[all...]
/kernel/linux/linux-6.6/drivers/infiniband/hw/irdma/
H A Duda.h39 int irdma_sc_access_ah(struct irdma_sc_cqp *cqp, struct irdma_ah_info *info,
41 int irdma_access_mcast_grp(struct irdma_sc_cqp *cqp,
50 static inline int irdma_sc_create_ah(struct irdma_sc_cqp *cqp, in irdma_sc_create_ah() argument
53 return irdma_sc_access_ah(cqp, info, IRDMA_CQP_OP_CREATE_ADDR_HANDLE, in irdma_sc_create_ah()
57 static inline int irdma_sc_destroy_ah(struct irdma_sc_cqp *cqp, in irdma_sc_destroy_ah() argument
60 return irdma_sc_access_ah(cqp, info, IRDMA_CQP_OP_DESTROY_ADDR_HANDLE, in irdma_sc_destroy_ah()
64 static inline int irdma_sc_create_mcast_grp(struct irdma_sc_cqp *cqp, in irdma_sc_create_mcast_grp() argument
68 return irdma_access_mcast_grp(cqp, info, IRDMA_CQP_OP_CREATE_MCAST_GRP, in irdma_sc_create_mcast_grp()
72 static inline int irdma_sc_modify_mcast_grp(struct irdma_sc_cqp *cqp, in irdma_sc_modify_mcast_grp() argument
76 return irdma_access_mcast_grp(cqp, inf in irdma_sc_modify_mcast_grp()
80 irdma_sc_destroy_mcast_grp(struct irdma_sc_cqp *cqp, struct irdma_mcast_grp_info *info, u64 scratch) irdma_sc_destroy_mcast_grp() argument
[all...]
H A Dctrl.c62 /* issue cqp suspend command */ in irdma_sc_suspend_resume_qps()
176 * irdma_sc_add_arp_cache_entry - cqp wqe add arp cache entry
177 * @cqp: struct for cqp hw
179 * @scratch: u64 saved to be used during cqp completion
180 * @post_sq: flag for cqp db to ring
182 static int irdma_sc_add_arp_cache_entry(struct irdma_sc_cqp *cqp, in irdma_sc_add_arp_cache_entry() argument
189 wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch); in irdma_sc_add_arp_cache_entry()
199 FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity); in irdma_sc_add_arp_cache_entry()
207 irdma_sc_cqp_post_sq(cqp); in irdma_sc_add_arp_cache_entry()
219 irdma_sc_del_arp_cache_entry(struct irdma_sc_cqp *cqp, u64 scratch, u16 arp_index, bool post_sq) irdma_sc_del_arp_cache_entry() argument
252 irdma_sc_manage_apbvt_entry(struct irdma_sc_cqp *cqp, struct irdma_apbvt_info *info, u64 scratch, bool post_sq) irdma_sc_manage_apbvt_entry() argument
300 irdma_sc_manage_qhash_table_entry(struct irdma_sc_cqp *cqp, struct irdma_qhash_table_info *info, u64 scratch, bool post_sq) irdma_sc_manage_qhash_table_entry() argument
451 struct irdma_sc_cqp *cqp; irdma_sc_qp_create() local
503 struct irdma_sc_cqp *cqp; irdma_sc_qp_modify() local
573 struct irdma_sc_cqp *cqp; irdma_sc_qp_destroy() local
761 irdma_sc_alloc_local_mac_entry(struct irdma_sc_cqp *cqp, u64 scratch, bool post_sq) irdma_sc_alloc_local_mac_entry() argument
795 irdma_sc_add_local_mac_entry(struct irdma_sc_cqp *cqp, struct irdma_local_mac_entry_info *info, u64 scratch, bool post_sq) irdma_sc_add_local_mac_entry() argument
833 irdma_sc_del_local_mac_entry(struct irdma_sc_cqp *cqp, u64 scratch, u16 entry_idx, u8 ignore_ref_count, bool post_sq) irdma_sc_del_local_mac_entry() argument
1060 struct irdma_sc_cqp *cqp; irdma_sc_alloc_stag() local
1125 struct irdma_sc_cqp *cqp; irdma_sc_mr_reg_non_shared() local
1216 struct irdma_sc_cqp *cqp; irdma_sc_dealloc_stag() local
1255 struct irdma_sc_cqp *cqp; irdma_sc_mw_alloc() local
2047 irdma_sc_gather_stats(struct irdma_sc_cqp *cqp, struct irdma_stats_gather_info *info, u64 scratch) irdma_sc_gather_stats() argument
2094 irdma_sc_manage_stats_inst(struct irdma_sc_cqp *cqp, struct irdma_stats_inst_info *info, bool alloc, u64 scratch) irdma_sc_manage_stats_inst() argument
2131 irdma_sc_set_up_map(struct irdma_sc_cqp *cqp, struct irdma_up_info *info, u64 scratch) irdma_sc_set_up_map() argument
2173 irdma_sc_manage_ws_node(struct irdma_sc_cqp *cqp, struct irdma_ws_node_info *info, enum irdma_ws_node_op node_op, u64 scratch) irdma_sc_manage_ws_node() argument
2221 struct irdma_sc_cqp *cqp; irdma_sc_qp_flush_wqes() local
2295 struct irdma_sc_cqp *cqp; irdma_sc_gen_ae() local
2334 struct irdma_sc_cqp *cqp; irdma_sc_qp_upload_context() local
2369 irdma_sc_manage_push_page(struct irdma_sc_cqp *cqp, struct irdma_cqp_manage_push_page_info *info, u64 scratch, bool post_sq) irdma_sc_manage_push_page() argument
2408 irdma_sc_suspend_qp(struct irdma_sc_cqp *cqp, struct irdma_sc_qp *qp, u64 scratch) irdma_sc_suspend_qp() argument
2438 irdma_sc_resume_qp(struct irdma_sc_cqp *cqp, struct irdma_sc_qp *qp, u64 scratch) irdma_sc_resume_qp() argument
2520 struct irdma_sc_cqp *cqp; irdma_sc_cq_create() local
2591 struct irdma_sc_cqp *cqp; irdma_sc_cq_destroy() local
2659 struct irdma_sc_cqp *cqp; irdma_sc_cq_modify() local
2732 irdma_get_cqp_reg_info(struct irdma_sc_cqp *cqp, u32 *val, u32 *tail, u32 *error) irdma_get_cqp_reg_info() argument
2746 irdma_cqp_poll_registers(struct irdma_sc_cqp *cqp, u32 tail, u32 count) irdma_cqp_poll_registers() argument
3088 irdma_sc_cqp_init(struct irdma_sc_cqp *cqp, struct irdma_cqp_init_info *info) irdma_sc_cqp_init() argument
3136 (u64 *)(uintptr_t)cqp->sq_pa, cqp, cqp->polarity); irdma_sc_cqp_init() local
3146 irdma_sc_cqp_create(struct irdma_sc_cqp *cqp, u16 *maj_err, u16 *min_err) irdma_sc_cqp_create() argument
3244 irdma_sc_cqp_post_sq(struct irdma_sc_cqp *cqp) irdma_sc_cqp_post_sq() argument
3260 irdma_sc_cqp_get_next_send_wqe_idx(struct irdma_sc_cqp *cqp, u64 scratch, u32 *wqe_idx) irdma_sc_cqp_get_next_send_wqe_idx() argument
3291 irdma_sc_cqp_destroy(struct irdma_sc_cqp *cqp) irdma_sc_cqp_destroy() argument
3350 struct irdma_sc_cqp *cqp; irdma_sc_ccq_get_cqe_info() local
3414 irdma_sc_poll_for_cqp_op_done(struct irdma_sc_cqp *cqp, u8 op_code, struct irdma_ccq_cqe_info *compl_info) irdma_sc_poll_for_cqp_op_done() argument
3456 irdma_sc_manage_hmc_pm_func_table(struct irdma_sc_cqp *cqp, struct irdma_hmc_fcn_info *info, u64 scratch, bool post_sq) irdma_sc_manage_hmc_pm_func_table() argument
3498 irdma_sc_commit_fpm_val_done(struct irdma_sc_cqp *cqp) irdma_sc_commit_fpm_val_done() argument
3513 irdma_sc_commit_fpm_val(struct irdma_sc_cqp *cqp, u64 scratch, u8 hmc_fn_id, struct irdma_dma_mem *commit_fpm_mem, bool post_sq, u8 wait_type) irdma_sc_commit_fpm_val() argument
3559 irdma_sc_query_fpm_val_done(struct irdma_sc_cqp *cqp) irdma_sc_query_fpm_val_done() argument
3574 irdma_sc_query_fpm_val(struct irdma_sc_cqp *cqp, u64 scratch, u8 hmc_fn_id, struct irdma_dma_mem *query_fpm_mem, bool post_sq, u8 wait_type) irdma_sc_query_fpm_val() argument
3668 struct irdma_sc_cqp *cqp; irdma_sc_ceq_create() local
3709 struct irdma_sc_cqp *cqp; irdma_sc_cceq_create_done() local
3722 struct irdma_sc_cqp *cqp; irdma_sc_cceq_destroy_done() local
3766 struct irdma_sc_cqp *cqp; irdma_sc_ceq_destroy() local
3933 struct irdma_sc_cqp *cqp; irdma_sc_aeq_create() local
3972 struct irdma_sc_cqp *cqp; irdma_sc_aeq_destroy() local
4219 struct irdma_sc_cqp *cqp; irdma_sc_ccq_create_done() local
4260 struct irdma_sc_cqp *cqp; irdma_sc_ccq_destroy() local
4413 cqp_sds_wqe_fill(struct irdma_sc_cqp *cqp, struct irdma_update_sds_info *info, u64 scratch) cqp_sds_wqe_fill() argument
4495 struct irdma_sc_cqp *cqp = dev->cqp; irdma_update_pe_sds() local
4514 struct irdma_sc_cqp *cqp = dev->cqp; irdma_update_sds_noccq() local
4536 irdma_sc_static_hmc_pages_allocated(struct irdma_sc_cqp *cqp, u64 scratch, u8 hmc_fn_id, bool post_sq, bool poll_registers) irdma_sc_static_hmc_pages_allocated() argument
4582 irdma_cqp_ring_full(struct irdma_sc_cqp *cqp) irdma_cqp_ring_full() argument
4621 irdma_sc_query_rdma_features_done(struct irdma_sc_cqp *cqp) irdma_sc_query_rdma_features_done() argument
4634 irdma_sc_query_rdma_features(struct irdma_sc_cqp *cqp, struct irdma_dma_mem *buf, u64 scratch) irdma_sc_query_rdma_features() argument
[all...]
H A Dhw.c587 * Issue destroy cqp request and
588 * free the resources associated with the cqp
593 struct irdma_cqp *cqp = &rf->cqp; in irdma_destroy_cqp() local
596 status = irdma_sc_cqp_destroy(dev->cqp); in irdma_destroy_cqp()
601 dma_free_coherent(dev->hw->device, cqp->sq.size, cqp->sq.va, in irdma_destroy_cqp()
602 cqp->sq.pa); in irdma_destroy_cqp()
603 cqp->sq.va = NULL; in irdma_destroy_cqp()
604 kfree(cqp in irdma_destroy_cqp()
933 struct irdma_cqp *cqp = &rf->cqp; irdma_create_cqp() local
[all...]
H A Dutils.c425 * irdma_alloc_and_get_cqp_request - get cqp struct
426 * @cqp: device cqp ptr
427 * @wait: cqp to be used in wait mode
429 struct irdma_cqp_request *irdma_alloc_and_get_cqp_request(struct irdma_cqp *cqp, in irdma_alloc_and_get_cqp_request() argument
435 spin_lock_irqsave(&cqp->req_lock, flags); in irdma_alloc_and_get_cqp_request()
436 if (!list_empty(&cqp->cqp_avail_reqs)) { in irdma_alloc_and_get_cqp_request()
437 cqp_request = list_first_entry(&cqp->cqp_avail_reqs, in irdma_alloc_and_get_cqp_request()
441 spin_unlock_irqrestore(&cqp->req_lock, flags); in irdma_alloc_and_get_cqp_request()
451 ibdev_dbg(to_ibdev(cqp in irdma_alloc_and_get_cqp_request()
476 irdma_free_cqp_request(struct irdma_cqp *cqp, struct irdma_cqp_request *cqp_request) irdma_free_cqp_request() argument
500 irdma_put_cqp_request(struct irdma_cqp *cqp, struct irdma_cqp_request *cqp_request) irdma_put_cqp_request() argument
513 irdma_free_pending_cqp_request(struct irdma_cqp *cqp, struct irdma_cqp_request *cqp_request) irdma_free_pending_cqp_request() argument
534 struct irdma_cqp *cqp = &rf->cqp; irdma_cleanup_pending_cqp_op() local
867 struct irdma_sc_cqp *cqp = dev->cqp; irdma_cqp_qp_suspend_resume() local
1897 struct irdma_sc_cqp *cqp = &iwcqp->sc_cqp; irdma_cqp_ws_node_cmd() local
[all...]
H A Duda.c15 * @cqp: struct for cqp hw
18 * @scratch: u64 saved to be used during cqp completion
20 int irdma_sc_access_ah(struct irdma_sc_cqp *cqp, struct irdma_ah_info *info, in irdma_sc_access_ah() argument
26 wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch); in irdma_sc_access_ah()
69 FIELD_PREP(IRDMA_UDA_CQPSQ_MAV_WQEVALID, cqp->polarity) | in irdma_sc_access_ah()
78 irdma_sc_cqp_post_sq(cqp); in irdma_sc_access_ah()
110 * @cqp: Control QP
113 * @scratch: u64 saved to be used during cqp completion
115 int irdma_access_mcast_grp(struct irdma_sc_cqp *cqp, in irdma_access_mcast_grp() argument
[all...]
H A Dtype.h604 struct irdma_sc_cqp *cqp; member
668 struct irdma_sc_cqp *cqp; member
1192 int irdma_sc_cqp_create(struct irdma_sc_cqp *cqp, u16 *maj_err, u16 *min_err);
1193 int irdma_sc_cqp_destroy(struct irdma_sc_cqp *cqp);
1194 int irdma_sc_cqp_init(struct irdma_sc_cqp *cqp,
1196 void irdma_sc_cqp_post_sq(struct irdma_sc_cqp *cqp);
1197 int irdma_sc_poll_for_cqp_op_done(struct irdma_sc_cqp *cqp, u8 opcode,
1224 int irdma_sc_static_hmc_pages_allocated(struct irdma_sc_cqp *cqp, u64 scratch,
1292 struct irdma_sc_cqp *cqp; member
1298 struct irdma_sc_cqp *cqp; member
1304 struct irdma_sc_cqp *cqp; global() member
1310 struct irdma_sc_cqp *cqp; global() member
1317 struct irdma_sc_cqp *cqp; global() member
1322 struct irdma_sc_cqp *cqp; global() member
1372 struct irdma_sc_cqp *cqp; global() member
1380 struct irdma_sc_cqp *cqp; global() member
1388 struct irdma_sc_cqp *cqp; global() member
1394 struct irdma_sc_cqp *cqp; global() member
1406 struct irdma_sc_cqp *cqp; global() member
1412 struct irdma_sc_cqp *cqp; global() member
1418 struct irdma_sc_cqp *cqp; global() member
1424 struct irdma_sc_cqp *cqp; global() member
1430 struct irdma_sc_cqp *cqp; global() member
1436 struct irdma_sc_cqp *cqp; global() member
1442 struct irdma_sc_cqp *cqp; global() member
1448 struct irdma_sc_cqp *cqp; global() member
1454 struct irdma_sc_cqp *cqp; global() member
1460 struct irdma_sc_cqp *cqp; global() member
1466 struct irdma_sc_cqp *cqp; global() member
1488 irdma_sc_cqp_get_next_send_wqe(struct irdma_sc_cqp *cqp, u64 scratch) irdma_sc_cqp_get_next_send_wqe() argument
[all...]
H A Dprotos.h18 void irdma_sc_cqp_post_sq(struct irdma_sc_cqp *cqp);
19 __le64 *irdma_sc_cqp_get_next_send_wqe(struct irdma_sc_cqp *cqp, u64 scratch);
77 int irdma_sc_static_hmc_pages_allocated(struct irdma_sc_cqp *cqp, u64 scratch,
H A Dpuda.c607 struct irdma_sc_cqp *cqp; in irdma_puda_qp_wqe() local
613 cqp = dev->cqp; in irdma_puda_qp_wqe()
614 wqe = irdma_sc_cqp_get_next_send_wqe(cqp, 0); in irdma_puda_qp_wqe()
626 FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity); in irdma_puda_qp_wqe()
633 irdma_sc_cqp_post_sq(cqp); in irdma_puda_qp_wqe()
634 status = irdma_sc_poll_for_cqp_op_done(dev->cqp, IRDMA_CQP_OP_CREATE_QP, in irdma_puda_qp_wqe()
726 struct irdma_sc_cqp *cqp; in irdma_puda_cq_wqe() local
731 cqp = dev->cqp; in irdma_puda_cq_wqe()
[all...]
H A Dmain.h303 struct irdma_cqp cqp; member
482 struct irdma_cqp_request *irdma_alloc_and_get_cqp_request(struct irdma_cqp *cqp,
484 void irdma_free_cqp_request(struct irdma_cqp *cqp,
486 void irdma_put_cqp_request(struct irdma_cqp *cqp,
H A Dverbs.c237 cqp_request = irdma_alloc_and_get_cqp_request(&iwdev->rf->cqp, true); in irdma_alloc_push_page()
249 cqp_info->in.u.manage_push_page.cqp = &iwdev->rf->cqp.sc_cqp; in irdma_alloc_push_page()
259 irdma_put_cqp_request(&iwdev->rf->cqp, cqp_request); in irdma_alloc_push_page()
739 cqp_request = irdma_alloc_and_get_cqp_request(&rf->cqp, true); in irdma_cqp_create_qp_cmd()
755 irdma_put_cqp_request(&rf->cqp, cqp_request); in irdma_cqp_create_qp_cmd()
1977 cqp_request = irdma_alloc_and_get_cqp_request(&rf->cqp, true); in irdma_resize_cq()
1995 irdma_put_cqp_request(&rf->cqp, cqp_request); in irdma_resize_cq()
2197 cqp_request = irdma_alloc_and_get_cqp_request(&rf->cqp, true); in irdma_create_cq()
2210 irdma_put_cqp_request(&rf->cqp, cqp_reques in irdma_create_cq()
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/en/
H A Dparams.h22 struct mlx5e_cq_param cqp; member
29 struct mlx5e_cq_param cqp; member
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/
H A Dsetup.c63 err = mlx5e_open_cq(c, params->rx_cq_moderation, &cparam->rq.cqp, &c->xskrq.cq); in mlx5e_open_xsk()
71 err = mlx5e_open_cq(c, params->tx_cq_moderation, &cparam->xdp_sq.cqp, &c->xsksq.cq); in mlx5e_open_xsk()
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/
H A Dsetup.c130 err = mlx5e_open_cq(c->priv, params->rx_cq_moderation, &cparam->rq.cqp, &ccp, in mlx5e_open_xsk()
139 err = mlx5e_open_cq(c->priv, params->tx_cq_moderation, &cparam->xdp_sq.cqp, &ccp, in mlx5e_open_xsk()
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/en/
H A Dparams.h23 struct mlx5e_cq_param cqp; member
31 struct mlx5e_cq_param cqp; member

Completed in 34 milliseconds

12