/kernel/linux/linux-5.10/drivers/net/ethernet/hisilicon/hns3/hns3pf/ |
H A D | hclge_cmd.c | 63 (ring_type == HCLGE_TYPE_CSQ) ? &hw->cmq.csq : &hw->cmq.crq; in hclge_alloc_cmd_queue() 131 hclge_cmd_config_regs(&hw->cmq.csq); in hclge_cmd_init_regs() 132 hclge_cmd_config_regs(&hw->cmq.crq); in hclge_cmd_init_regs() 138 struct hclge_cmq_ring *csq = &hw->cmq.csq; in hclge_cmd_csq_clean() 164 return head == hw->cmq.csq.next_to_use; in hclge_cmd_csq_done() 232 desc[handle] = hw->cmq.csq.desc[ntc]; in hclge_cmd_check_retval() 234 if (ntc >= hw->cmq.csq.desc_num) in hclge_cmd_check_retval() 242 hw->cmq.last_status = desc_ret; in hclge_cmd_check_retval() 259 struct hclge_cmq_ring *csq = &hw->cmq in hclge_cmd_send() [all...] |
H A D | hclge_mbx.c | 666 return tail == hw->cmq.crq.next_to_use; in hclge_cmd_crq_empty() 697 struct hclge_cmq_ring *crq = &hdev->hw.cmq.crq; in hclge_mbx_handler()
|
H A D | hclge_main.h | 285 struct hclge_cmq cmq; member
|
/kernel/linux/linux-5.10/drivers/net/ethernet/hisilicon/hns3/hns3vf/ |
H A D | hclgevf_cmd.c | 42 struct hclgevf_cmq_ring *csq = &hw->cmq.csq; in hclgevf_cmd_csq_clean() 69 return head == hw->cmq.csq.next_to_use; in hclgevf_cmd_csq_done() 120 hclgevf_cmd_config_regs(&hw->cmq.csq); in hclgevf_cmd_init_regs() 121 hclgevf_cmd_config_regs(&hw->cmq.crq); in hclgevf_cmd_init_regs() 151 (ring_type == HCLGEVF_TYPE_CSQ) ? &hw->cmq.csq : &hw->cmq.crq; in hclgevf_alloc_cmd_queue() 222 struct hclgevf_cmq_ring *csq = &hw->cmq.csq; in hclgevf_cmd_send() 232 spin_lock_bh(&hw->cmq.csq.lock); in hclgevf_cmd_send() 235 spin_unlock_bh(&hw->cmq.csq.lock); in hclgevf_cmd_send() 239 if (num > hclgevf_ring_space(&hw->cmq in hclgevf_cmd_send() [all...] |
H A D | hclgevf_mbx.c | 150 return tail == hw->cmq.crq.next_to_use; in hclgevf_cmd_crq_empty() 165 crq = &hdev->hw.cmq.crq; in hclgevf_mbx_handler()
|
H A D | hclgevf_main.h | 172 struct hclgevf_cmq cmq; member
|
/kernel/linux/linux-6.6/drivers/net/ethernet/hisilicon/hns3/hns3_common/ |
H A D | hclge_comm_cmd.c | 38 hclge_comm_cmd_config_regs(hw, &hw->cmq.csq); in hclge_comm_cmd_init_regs() 39 hclge_comm_cmd_config_regs(hw, &hw->cmq.crq); in hclge_comm_cmd_init_regs() 209 (ring_type == HCLGE_COMM_TYPE_CSQ) ? &hw->cmq.csq : in hclge_comm_alloc_cmd_queue() 210 &hw->cmq.crq; in hclge_comm_alloc_cmd_queue() 298 desc_to_use = &hw->cmq.csq.desc[hw->cmq.csq.next_to_use]; in hclge_comm_cmd_copy_desc() 300 (hw->cmq.csq.next_to_use)++; in hclge_comm_cmd_copy_desc() 301 if (hw->cmq.csq.next_to_use >= hw->cmq.csq.desc_num) in hclge_comm_cmd_copy_desc() 302 hw->cmq in hclge_comm_cmd_copy_desc() [all...] |
H A D | hclge_comm_rss.c | 111 dev_err(&hw->cmq.csq.pdev->dev, in hclge_comm_set_rss_tc_mode() 166 dev_err(&hw->cmq.csq.pdev->dev, in hclge_comm_set_rss_tuple() 173 dev_err(&hw->cmq.csq.pdev->dev, in hclge_comm_set_rss_tuple() 299 dev_err(&hw->cmq.csq.pdev->dev, in hclge_comm_set_rss_indir_table() 331 dev_err(&hw->cmq.csq.pdev->dev, in hclge_comm_set_rss_input_tuple() 402 dev_err(&hw->cmq.csq.pdev->dev, in hclge_comm_set_rss_algo_key()
|
H A D | hclge_comm_tqp_stats.c | 77 dev_err(&hw->cmq.csq.pdev->dev, in hclge_comm_tqps_update_stats() 91 dev_err(&hw->cmq.csq.pdev->dev, in hclge_comm_tqps_update_stats()
|
H A D | hclge_comm_cmd.h | 413 u8 ring_type; /* cmq ring type */ 434 struct hclge_comm_cmq cmq; member
|
/kernel/linux/linux-6.6/drivers/net/ethernet/hisilicon/hns3/hns3vf/ |
H A D | hclgevf_mbx.c | 162 return tail == hw->hw.cmq.crq.next_to_use; in hclgevf_cmd_crq_empty() 232 crq = &hdev->hw.hw.cmq.crq; in hclgevf_mbx_handler()
|
H A D | hclgevf_main.c | 52 struct hclge_comm_cmq *cmdq = &hdev->hw.hw.cmq; in hclgevf_arq_init()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/hns/ |
H A D | hns_roce_hw_v2.c | 954 * sending mailbox&cmq&doorbell to hardware. If now in .init_instance() in hns_roce_v2_cmd_hw_reseted() 985 /* When hardware reset is detected, we should stop sending mailbox&cmq& in hns_roce_v2_cmd_hw_resetting() 1016 * should stop sending mailbox&cmq&doorbell to hardware, and exit in hns_roce_v2_cmd_sw_resetting() 1110 &priv->cmq.csq : &priv->cmq.crq; in hns_roce_init_cmq_ring() 1123 &priv->cmq.csq : &priv->cmq.crq; in hns_roce_cmq_init_regs() 1151 priv->cmq.csq.desc_num = CMD_CSQ_DESC_NUM; in hns_roce_v2_cmq_init() 1152 priv->cmq.crq.desc_num = CMD_CRQ_DESC_NUM; in hns_roce_v2_cmq_init() 1155 spin_lock_init(&priv->cmq in hns_roce_v2_cmq_init() [all...] |
H A D | hns_roce_hw_v2.h | 1787 struct hns_roce_v2_cmq cmq; member
|
/kernel/linux/linux-6.6/drivers/net/ethernet/hisilicon/hns3/hns3pf/ |
H A D | hclge_mbx.c | 805 return tail == hw->hw.cmq.crq.next_to_use; in hclge_cmd_crq_empty() 1105 struct hclge_comm_cmq_ring *crq = &hdev->hw.hw.cmq.crq; in hclge_mbx_handler()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/hns/ |
H A D | hns_roce_hw_v2.c | 1003 * sending mailbox&cmq&doorbell to hardware. If now in .init_instance() in hns_roce_v2_cmd_hw_reseted() 1034 /* When hardware reset is detected, we should stop sending mailbox&cmq& in hns_roce_v2_cmd_hw_resetting() 1065 * should stop sending mailbox&cmq&doorbell to hardware, and exit in hns_roce_v2_cmd_sw_resetting() 1200 priv->cmq.tx_timeout = HNS_ROCE_CMQ_TX_TIMEOUT; in hns_roce_v2_cmq_init() 1202 ret = init_csq(hr_dev, &priv->cmq.csq); in hns_roce_v2_cmq_init() 1213 hns_roce_free_cmq_desc(hr_dev, &priv->cmq.csq); in hns_roce_v2_cmq_exit() 1234 return tail == priv->cmq.csq.head; in hns_roce_cmq_csq_done() 1275 struct hns_roce_v2_cmq_ring *csq = &priv->cmq.csq; in __hns_roce_cmq_send() 1299 } while (++timeout < priv->cmq.tx_timeout); in __hns_roce_cmq_send() 5518 ibdev_err(ibdev, "failed to query clr cmq, re in hns_roce_v2_qp_flow_control_init() [all...] |
H A D | hns_roce_hw_v2.h | 1311 struct hns_roce_v2_cmq cmq; member
|