Searched refs:new_req (Results 1 - 7 of 7) sorted by relevance
/kernel/linux/linux-6.6/drivers/net/ethernet/broadcom/bnxt/ |
H A D | bnxt_hwrm.c | 170 * an assignment of new_req to req. Subsequent calls to HWRM API functions, 171 * such as hwrm_req_send(), should thus use req and not new_req (in fact, 174 * @len: The length of new_req. 175 * @new_req: The pre-built request to copy or reference. 177 * Replaces the request data in req with that of new_req. This is useful in 184 * req and is independent of and does not apply to new_req. The caller must 185 * ensure that the lifetime of new_req is least as long as req. Any slices 192 int hwrm_req_replace(struct bnxt *bp, void *req, void *new_req, u32 len) in hwrm_req_replace() argument 214 memcpy(internal_req, new_req, len); in hwrm_req_replace() 216 internal_req->req_type = ((struct input *)new_req) in hwrm_req_replace() [all...] |
H A D | bnxt_hwrm.h | 137 int hwrm_req_replace(struct bnxt *bp, void *req, void *new_req, u32 len);
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/irdma/ |
H A D | hw.c | 2633 struct irdma_cqp_request *new_req; in irdma_hw_flush_wqes() local 2642 new_req = irdma_alloc_and_get_cqp_request(&rf->cqp, true); in irdma_hw_flush_wqes() 2643 if (!new_req) { in irdma_hw_flush_wqes() 2647 cqp_info = &new_req->info; in irdma_hw_flush_wqes() 2648 hw_info = &new_req->info.in.u.qp_flush_wqes.info; in irdma_hw_flush_wqes() 2653 cqp_info->in.u.qp_flush_wqes.scratch = (uintptr_t)new_req; in irdma_hw_flush_wqes() 2655 status = irdma_handle_cqp_op(rf, new_req); in irdma_hw_flush_wqes() 2656 if (new_req->compl_info.maj_err_code || in irdma_hw_flush_wqes() 2657 new_req->compl_info.min_err_code != IRDMA_CQP_COMPL_SQ_WQE_FLUSHED || in irdma_hw_flush_wqes() 2664 irdma_put_cqp_request(&rf->cqp, new_req); in irdma_hw_flush_wqes() [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/ |
H A D | drm_plane.c | 1023 struct drm_mode_cursor2 new_req; in drm_mode_cursor_ioctl() local 1025 memcpy(&new_req, req, sizeof(struct drm_mode_cursor)); in drm_mode_cursor_ioctl() 1026 new_req.hot_x = new_req.hot_y = 0; in drm_mode_cursor_ioctl() 1028 return drm_mode_cursor_common(dev, &new_req, file_priv); in drm_mode_cursor_ioctl()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/ |
H A D | drm_plane.c | 1203 struct drm_mode_cursor2 new_req; in drm_mode_cursor_ioctl() local 1205 memcpy(&new_req, req, sizeof(struct drm_mode_cursor)); in drm_mode_cursor_ioctl() 1206 new_req.hot_x = new_req.hot_y = 0; in drm_mode_cursor_ioctl() 1208 return drm_mode_cursor_common(dev, &new_req, file_priv); in drm_mode_cursor_ioctl()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/hisilicon/hns3/hns3pf/ |
H A D | hclge_main.c | 7210 static void hclge_prepare_mac_addr(struct hclge_mac_vlan_tbl_entry_cmd *new_req, in hclge_prepare_mac_addr() argument 7218 hnae3_set_bit(new_req->flags, HCLGE_MAC_VLAN_BIT0_EN_B, 1); in hclge_prepare_mac_addr() 7220 hnae3_set_bit(new_req->entry_type, HCLGE_MAC_VLAN_BIT1_EN_B, 1); in hclge_prepare_mac_addr() 7221 hnae3_set_bit(new_req->mc_mac_en, HCLGE_MAC_VLAN_BIT0_EN_B, 1); in hclge_prepare_mac_addr() 7224 new_req->mac_addr_hi32 = cpu_to_le32(high_val); in hclge_prepare_mac_addr() 7225 new_req->mac_addr_lo16 = cpu_to_le16(low_val & 0xffff); in hclge_prepare_mac_addr()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/hisilicon/hns3/hns3pf/ |
H A D | hclge_main.c | 8151 static void hclge_prepare_mac_addr(struct hclge_mac_vlan_tbl_entry_cmd *new_req, in hclge_prepare_mac_addr() argument 8159 hnae3_set_bit(new_req->flags, HCLGE_MAC_VLAN_BIT0_EN_B, 1); in hclge_prepare_mac_addr() 8161 hnae3_set_bit(new_req->entry_type, HCLGE_MAC_VLAN_BIT1_EN_B, 1); in hclge_prepare_mac_addr() 8162 hnae3_set_bit(new_req->mc_mac_en, HCLGE_MAC_VLAN_BIT0_EN_B, 1); in hclge_prepare_mac_addr() 8165 new_req->mac_addr_hi32 = cpu_to_le32(high_val); in hclge_prepare_mac_addr() 8166 new_req->mac_addr_lo16 = cpu_to_le16(low_val & 0xffff); in hclge_prepare_mac_addr()
|
Completed in 36 milliseconds