/kernel/linux/linux-6.6/drivers/ufs/core/ |
H A D | ufs-mcq.c | 115 u32 hwq = blk_mq_unique_tag_to_hwq(utag); in ufshcd_mcq_req_to_hwq() local 117 return &hba->uhq[hwq]; in ufshcd_mcq_req_to_hwq() 201 struct ufs_hw_queue *hwq; in ufshcd_mcq_memory_alloc() local 206 hwq = &hba->uhq[i]; in ufshcd_mcq_memory_alloc() 209 hwq->max_entries; in ufshcd_mcq_memory_alloc() 210 hwq->sqe_base_addr = dmam_alloc_coherent(hba->dev, utrdl_size, in ufshcd_mcq_memory_alloc() 211 &hwq->sqe_dma_addr, in ufshcd_mcq_memory_alloc() 213 if (!hwq->sqe_dma_addr) { in ufshcd_mcq_memory_alloc() 218 cqe_size = sizeof(struct cq_entry) * hwq->max_entries; in ufshcd_mcq_memory_alloc() 219 hwq in ufshcd_mcq_memory_alloc() 261 ufshcd_mcq_get_tag(struct ufs_hba *hba, struct ufs_hw_queue *hwq, struct cq_entry *cqe) ufshcd_mcq_get_tag() argument 277 ufshcd_mcq_process_cqe(struct ufs_hba *hba, struct ufs_hw_queue *hwq) ufshcd_mcq_process_cqe() argument 290 ufshcd_mcq_compl_all_cqes_lock(struct ufs_hba *hba, struct ufs_hw_queue *hwq) ufshcd_mcq_compl_all_cqes_lock() argument 308 ufshcd_mcq_poll_cqe_lock(struct ufs_hba *hba, struct ufs_hw_queue *hwq) ufshcd_mcq_poll_cqe_lock() argument 332 struct ufs_hw_queue *hwq; ufshcd_mcq_make_queues_operational() local 412 struct ufs_hw_queue *hwq; ufshcd_mcq_init() local 452 ufshcd_mcq_sq_stop(struct ufs_hba *hba, struct ufs_hw_queue *hwq) ufshcd_mcq_sq_stop() argument 471 ufshcd_mcq_sq_start(struct ufs_hba *hba, struct ufs_hw_queue *hwq) ufshcd_mcq_sq_start() argument 502 struct ufs_hw_queue *hwq; ufshcd_mcq_sq_cleanup() local 576 ufshcd_mcq_sqe_search(struct ufs_hba *hba, struct ufs_hw_queue *hwq, int task_tag) ufshcd_mcq_sqe_search() argument 632 struct ufs_hw_queue *hwq; ufshcd_mcq_abort() local [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/hw/bnxt_re/ |
H A D | qplib_res.h | 72 #define HWQ_CMP(idx, hwq) ((idx) & ((hwq)->max_elements - 1)) 74 #define HWQ_FREE_SLOTS(hwq) (hwq->max_elements - \ 75 ((HWQ_CMP(hwq->prod, hwq)\ 76 - HWQ_CMP(hwq->cons, hwq))\ 77 & (hwq->max_elements - 1))) 174 struct bnxt_qplib_hwq *hwq; member 292 bnxt_qplib_base_pg_size(struct bnxt_qplib_hwq *hwq) bnxt_qplib_base_pg_size() argument 324 bnxt_qplib_get_qe(struct bnxt_qplib_hwq *hwq, u32 indx, u64 *pg) bnxt_qplib_get_qe() argument 336 bnxt_qplib_get_prod_qe(struct bnxt_qplib_hwq *hwq, u32 idx) bnxt_qplib_get_prod_qe() argument 378 bnxt_qplib_hwq_incr_prod(struct bnxt_qplib_hwq *hwq, u32 cnt) bnxt_qplib_hwq_incr_prod() argument 383 bnxt_qplib_hwq_incr_cons(struct bnxt_qplib_hwq *hwq, u32 cnt) bnxt_qplib_hwq_incr_cons() argument [all...] |
H A D | qplib_fp.c | 75 dev_dbg(&scq->hwq.pdev->dev, in __bnxt_qplib_add_flush_qp() 83 dev_dbg(&rcq->hwq.pdev->dev, in __bnxt_qplib_add_flush_qp() 142 qp->sq.hwq.prod = 0; in bnxt_qplib_clean_qp() 143 qp->sq.hwq.cons = 0; in bnxt_qplib_clean_qp() 145 qp->rq.hwq.prod = 0; in bnxt_qplib_clean_qp() 146 qp->rq.hwq.cons = 0; in bnxt_qplib_clean_qp() 236 struct bnxt_qplib_hwq *hwq = &nq->hwq; in clean_nq() local 243 spin_lock_bh(&hwq->lock); in clean_nq() 245 raw_cons = hwq in clean_nq() 301 struct bnxt_qplib_hwq *hwq = &nq->hwq; bnxt_qplib_service_nq() local 392 struct bnxt_qplib_hwq *hwq = &nq->hwq; bnxt_qplib_nq_irq() local 945 struct bnxt_qplib_hwq *hwq; bnxt_qplib_init_psn_ptr() local 1599 struct bnxt_qplib_hwq *hwq; bnxt_qplib_put_inline() local 1640 bnxt_qplib_put_sges(struct bnxt_qplib_hwq *hwq, struct bnxt_qplib_sge *ssge, u16 nsge, u16 *idx) bnxt_qplib_put_sges() argument 1686 struct bnxt_qplib_hwq *hwq; bnxt_qplib_pull_psn_buff() local 1715 struct bnxt_qplib_hwq *hwq; bnxt_qplib_post_send() local 1968 struct bnxt_qplib_hwq *hwq; bnxt_qplib_post_recv() local [all...] |
H A D | qplib_res.c | 157 struct bnxt_qplib_hwq *hwq) in bnxt_qplib_free_hwq() 161 if (!hwq->max_elements) in bnxt_qplib_free_hwq() 163 if (hwq->level >= PBL_LVL_MAX) in bnxt_qplib_free_hwq() 166 for (i = 0; i < hwq->level + 1; i++) { in bnxt_qplib_free_hwq() 167 if (i == hwq->level) in bnxt_qplib_free_hwq() 168 __free_pbl(res, &hwq->pbl[i], hwq->is_user); in bnxt_qplib_free_hwq() 170 __free_pbl(res, &hwq->pbl[i], false); in bnxt_qplib_free_hwq() 173 hwq->level = PBL_LVL_MAX; in bnxt_qplib_free_hwq() 174 hwq in bnxt_qplib_free_hwq() 156 bnxt_qplib_free_hwq(struct bnxt_qplib_res *res, struct bnxt_qplib_hwq *hwq) bnxt_qplib_free_hwq() argument 183 bnxt_qplib_alloc_init_hwq(struct bnxt_qplib_hwq *hwq, struct bnxt_qplib_hwq_attr *hwq_attr) bnxt_qplib_alloc_init_hwq() argument [all...] |
H A D | qplib_rcfw.c | 92 struct bnxt_qplib_hwq *hwq = &cmdq->hwq; in __send_message() local 126 spin_lock_irqsave(&hwq->lock, flags); in __send_message() 127 if (req->cmd_size >= HWQ_FREE_SLOTS(hwq)) { in __send_message() 129 spin_unlock_irqrestore(&hwq->lock, flags); in __send_message() 143 spin_unlock_irqrestore(&hwq->lock, flags); in __send_message() 168 sw_prod = HWQ_CMP(hwq->prod, hwq); in __send_message() 169 cmdqe = bnxt_qplib_get_qe(hwq, sw_prod, NULL); in __send_message() 180 hwq in __send_message() 302 struct bnxt_qplib_hwq *hwq = &rcfw->cmdq.hwq; bnxt_qplib_process_qp_event() local 381 struct bnxt_qplib_hwq *hwq = &creq->hwq; bnxt_qplib_service_creq() local 441 struct bnxt_qplib_hwq *hwq; bnxt_qplib_creq_irq() local [all...] |
H A D | qplib_fp.h | 99 struct bnxt_qplib_hwq hwq; member 247 struct bnxt_qplib_hwq hwq; member 356 struct bnxt_qplib_hwq *hwq; in bnxt_qplib_queue_full() local 359 hwq = &que->hwq; in bnxt_qplib_queue_full() 361 avail = hwq->cons - hwq->prod; in bnxt_qplib_queue_full() 362 if (hwq->cons <= hwq->prod) in bnxt_qplib_queue_full() 363 avail += hwq in bnxt_qplib_queue_full() 400 struct bnxt_qplib_hwq hwq; global() member 473 struct bnxt_qplib_hwq hwq; global() member [all...] |
H A D | qplib_sp.c | 584 if (mrw->hwq.max_elements) in bnxt_qplib_free_mrw() 585 bnxt_qplib_free_hwq(res, &mrw->hwq); in bnxt_qplib_free_mrw() 643 if (mrw->hwq.max_elements) { in bnxt_qplib_dereg_mrw() 646 bnxt_qplib_free_hwq(res, &mrw->hwq); in bnxt_qplib_dereg_mrw() 679 /* Free the hwq if it already exist, must be a rereg */ in bnxt_qplib_reg_mr() 680 if (mr->hwq.max_elements) in bnxt_qplib_reg_mr() 681 bnxt_qplib_free_hwq(res, &mr->hwq); in bnxt_qplib_reg_mr() 691 rc = bnxt_qplib_alloc_init_hwq(&mr->hwq, &hwq_attr); in bnxt_qplib_reg_mr() 702 if (mr->hwq.level == PBL_LVL_MAX) { in bnxt_qplib_reg_mr() 708 level = mr->hwq in bnxt_qplib_reg_mr() [all...] |
/kernel/linux/linux-5.10/drivers/scsi/cxlflash/ |
H A D | main.c | 159 struct hwq *hwq = get_hwq(afu, cmd->hwq_index); in cmd_complete() local 161 spin_lock_irqsave(&hwq->hsq_slock, lock_flags); in cmd_complete() 163 spin_unlock_irqrestore(&hwq->hsq_slock, lock_flags); in cmd_complete() 186 * @hwq: Hardware queue to flush. 191 static void flush_pending_cmds(struct hwq *hwq) in flush_pending_cmds() argument 193 struct cxlflash_cfg *cfg = hwq->afu->parent; in flush_pending_cmds() 198 list_for_each_entry_safe(cmd, tmp, &hwq->pending_cmds, list) { in flush_pending_cmds() 226 * @hwq 235 context_reset(struct hwq *hwq, __be64 __iomem *reset_reg) context_reset() argument 276 context_reset_ioarrin(struct hwq *hwq) context_reset_ioarrin() argument 287 context_reset_sq(struct hwq *hwq) context_reset_sq() argument 304 struct hwq *hwq = get_hwq(afu, cmd->hwq_index); send_cmd_ioarrin() local 348 struct hwq *hwq = get_hwq(afu, cmd->hwq_index); send_cmd_sq() local 426 u32 hwq = 0; cmd_to_target_hwq() local 464 struct hwq *hwq = get_hwq(afu, PRIMARY_HWQ); send_tmf() local 568 struct hwq *hwq = get_hwq(afu, hwq_index); cxlflash_queuecommand() local 693 struct hwq *hwq; stop_afu() local 732 struct hwq *hwq; term_intr() local 778 struct hwq *hwq; term_mc() local 1265 struct hwq *hwq = get_hwq(afu, PRIMARY_HWQ); afu_err_intr_init() local 1334 struct hwq *hwq = (struct hwq *)data; cxlflash_sync_err_irq() local 1368 process_hrrq(struct hwq *hwq, struct list_head *doneq, int budget) process_hrrq() argument 1446 struct hwq *hwq = container_of(irqpoll, struct hwq, irqpoll); cxlflash_irqpoll() local 1472 struct hwq *hwq = (struct hwq *)data; cxlflash_rrq_irq() local 1540 struct hwq *hwq = (struct hwq *)data; cxlflash_async_err_irq() local 1734 struct hwq *hwq; init_pcr() local 1769 struct hwq *hwq; init_global() local 1895 struct hwq *hwq; start_afu() local 1948 init_intr(struct cxlflash_cfg *cfg, struct hwq *hwq) init_intr() argument 2008 struct hwq *hwq = get_hwq(cfg->afu, index); init_mc() local 2120 struct hwq *hwq; init_afu() local 2304 struct hwq *hwq = get_hwq(afu, PRIMARY_HWQ); send_afu_cmd() local 2432 struct hwq *hwq = get_hwq(afu, cmd->hwq_index); cxlflash_eh_abort_handler() local 2909 struct hwq *hwq; irqpoll_weight_store() local [all...] |
H A D | common.h | 196 struct hwq { struct 209 u32 index; /* Index of this hwq */ 231 struct hwq hwqs[CXLFLASH_MAX_HWQS]; 233 int (*context_reset)(struct hwq *hwq); 255 static inline struct hwq *get_hwq(struct afu *afu, u32 index) in get_hwq()
|
/kernel/linux/linux-6.6/drivers/scsi/cxlflash/ |
H A D | main.c | 159 struct hwq *hwq = get_hwq(afu, cmd->hwq_index); in cmd_complete() local 161 spin_lock_irqsave(&hwq->hsq_slock, lock_flags); in cmd_complete() 163 spin_unlock_irqrestore(&hwq->hsq_slock, lock_flags); in cmd_complete() 186 * @hwq: Hardware queue to flush. 191 static void flush_pending_cmds(struct hwq *hwq) in flush_pending_cmds() argument 193 struct cxlflash_cfg *cfg = hwq->afu->parent; in flush_pending_cmds() 198 list_for_each_entry_safe(cmd, tmp, &hwq->pending_cmds, list) { in flush_pending_cmds() 226 * @hwq 235 context_reset(struct hwq *hwq, __be64 __iomem *reset_reg) context_reset() argument 276 context_reset_ioarrin(struct hwq *hwq) context_reset_ioarrin() argument 287 context_reset_sq(struct hwq *hwq) context_reset_sq() argument 304 struct hwq *hwq = get_hwq(afu, cmd->hwq_index); send_cmd_ioarrin() local 348 struct hwq *hwq = get_hwq(afu, cmd->hwq_index); send_cmd_sq() local 426 u32 hwq = 0; cmd_to_target_hwq() local 464 struct hwq *hwq = get_hwq(afu, PRIMARY_HWQ); send_tmf() local 568 struct hwq *hwq = get_hwq(afu, hwq_index); cxlflash_queuecommand() local 693 struct hwq *hwq; stop_afu() local 732 struct hwq *hwq; term_intr() local 778 struct hwq *hwq; term_mc() local 1265 struct hwq *hwq = get_hwq(afu, PRIMARY_HWQ); afu_err_intr_init() local 1334 struct hwq *hwq = (struct hwq *)data; cxlflash_sync_err_irq() local 1368 process_hrrq(struct hwq *hwq, struct list_head *doneq, int budget) process_hrrq() argument 1446 struct hwq *hwq = container_of(irqpoll, struct hwq, irqpoll); cxlflash_irqpoll() local 1472 struct hwq *hwq = (struct hwq *)data; cxlflash_rrq_irq() local 1540 struct hwq *hwq = (struct hwq *)data; cxlflash_async_err_irq() local 1711 struct hwq *hwq; init_pcr() local 1746 struct hwq *hwq; init_global() local 1872 struct hwq *hwq; start_afu() local 1925 init_intr(struct cxlflash_cfg *cfg, struct hwq *hwq) init_intr() argument 1985 struct hwq *hwq = get_hwq(cfg->afu, index); init_mc() local 2097 struct hwq *hwq; init_afu() local 2281 struct hwq *hwq = get_hwq(afu, PRIMARY_HWQ); send_afu_cmd() local 2409 struct hwq *hwq = get_hwq(afu, cmd->hwq_index); cxlflash_eh_abort_handler() local 2886 struct hwq *hwq; irqpoll_weight_store() local [all...] |
H A D | common.h | 196 struct hwq { struct 209 u32 index; /* Index of this hwq */ 231 struct hwq hwqs[CXLFLASH_MAX_HWQS]; 233 int (*context_reset)(struct hwq *hwq); 255 static inline struct hwq *get_hwq(struct afu *afu, u32 index) in get_hwq()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/bnxt_re/ |
H A D | qplib_res.h | 83 #define HWQ_CMP(idx, hwq) ((idx) & ((hwq)->max_elements - 1)) 85 #define HWQ_FREE_SLOTS(hwq) (hwq->max_elements - \ 86 ((HWQ_CMP(hwq->prod, hwq)\ 87 - HWQ_CMP(hwq->cons, hwq))\ 88 & (hwq->max_elements - 1))) 186 struct bnxt_qplib_hwq *hwq; member 311 bnxt_qplib_base_pg_size(struct bnxt_qplib_hwq *hwq) bnxt_qplib_base_pg_size() argument 343 bnxt_qplib_get_qe(struct bnxt_qplib_hwq *hwq, u32 indx, u64 *pg) bnxt_qplib_get_qe() argument 355 bnxt_qplib_get_prod_qe(struct bnxt_qplib_hwq *hwq, u32 idx) bnxt_qplib_get_prod_qe() argument 399 bnxt_qplib_hwq_incr_prod(struct bnxt_qplib_hwq *hwq, u32 cnt) bnxt_qplib_hwq_incr_prod() argument 404 bnxt_qplib_hwq_incr_cons(struct bnxt_qplib_hwq *hwq, u32 cnt) bnxt_qplib_hwq_incr_cons() argument [all...] |
H A D | qplib_fp.c | 76 dev_dbg(&scq->hwq.pdev->dev, in __bnxt_qplib_add_flush_qp() 84 dev_dbg(&rcq->hwq.pdev->dev, in __bnxt_qplib_add_flush_qp() 143 qp->sq.hwq.prod = 0; in bnxt_qplib_clean_qp() 144 qp->sq.hwq.cons = 0; in bnxt_qplib_clean_qp() 146 qp->rq.hwq.prod = 0; in bnxt_qplib_clean_qp() 147 qp->rq.hwq.cons = 0; in bnxt_qplib_clean_qp() 237 struct bnxt_qplib_hwq *hwq = &nq->hwq; in clean_nq() local 244 spin_lock_bh(&hwq->lock); in clean_nq() 246 raw_cons = hwq in clean_nq() 302 struct bnxt_qplib_hwq *hwq = &nq->hwq; bnxt_qplib_service_nq() local 405 struct bnxt_qplib_hwq *hwq = &nq->hwq; bnxt_qplib_nq_irq() local 962 struct bnxt_qplib_hwq *hwq; bnxt_qplib_init_psn_ptr() local 1629 struct bnxt_qplib_hwq *hwq; bnxt_qplib_put_inline() local 1670 bnxt_qplib_put_sges(struct bnxt_qplib_hwq *hwq, struct bnxt_qplib_sge *ssge, u16 nsge, u16 *idx) bnxt_qplib_put_sges() argument 1716 struct bnxt_qplib_hwq *hwq; bnxt_qplib_pull_psn_buff() local 1745 struct bnxt_qplib_hwq *hwq; bnxt_qplib_post_send() local 1998 struct bnxt_qplib_hwq *hwq; bnxt_qplib_post_recv() local [all...] |
H A D | qplib_res.c | 157 struct bnxt_qplib_hwq *hwq) in bnxt_qplib_free_hwq() 161 if (!hwq->max_elements) in bnxt_qplib_free_hwq() 163 if (hwq->level >= PBL_LVL_MAX) in bnxt_qplib_free_hwq() 166 for (i = 0; i < hwq->level + 1; i++) { in bnxt_qplib_free_hwq() 167 if (i == hwq->level) in bnxt_qplib_free_hwq() 168 __free_pbl(res, &hwq->pbl[i], hwq->is_user); in bnxt_qplib_free_hwq() 170 __free_pbl(res, &hwq->pbl[i], false); in bnxt_qplib_free_hwq() 173 hwq->level = PBL_LVL_MAX; in bnxt_qplib_free_hwq() 174 hwq in bnxt_qplib_free_hwq() 156 bnxt_qplib_free_hwq(struct bnxt_qplib_res *res, struct bnxt_qplib_hwq *hwq) bnxt_qplib_free_hwq() argument 183 bnxt_qplib_alloc_init_hwq(struct bnxt_qplib_hwq *hwq, struct bnxt_qplib_hwq_attr *hwq_attr) bnxt_qplib_alloc_init_hwq() argument [all...] |
H A D | qplib_rcfw.c | 224 * user must hold the completion queue hwq->lock. 239 struct bnxt_qplib_hwq *hwq = &cmdq->hwq; in __send_message_no_waiter() local 264 sw_prod = HWQ_CMP(hwq->prod, hwq); in __send_message_no_waiter() 265 cmdqe = bnxt_qplib_get_qe(hwq, sw_prod, NULL); in __send_message_no_waiter() 271 hwq->prod++; in __send_message_no_waiter() 275 cmdq_prod = hwq->prod; in __send_message_no_waiter() 290 struct bnxt_qplib_hwq *hwq; in __send_message() local 298 hwq in __send_message() 624 struct bnxt_qplib_hwq *hwq = &rcfw->cmdq.hwq; bnxt_qplib_process_qp_event() local 735 struct bnxt_qplib_hwq *hwq = &creq->hwq; bnxt_qplib_service_creq() local 796 struct bnxt_qplib_hwq *hwq; bnxt_qplib_creq_irq() local [all...] |
H A D | qplib_fp.h | 101 struct bnxt_qplib_hwq hwq; member 249 struct bnxt_qplib_hwq hwq; member 358 struct bnxt_qplib_hwq *hwq; in bnxt_qplib_queue_full() local 361 hwq = &que->hwq; in bnxt_qplib_queue_full() 363 avail = hwq->cons - hwq->prod; in bnxt_qplib_queue_full() 364 if (hwq->cons <= hwq->prod) in bnxt_qplib_queue_full() 365 avail += hwq in bnxt_qplib_queue_full() 402 struct bnxt_qplib_hwq hwq; global() member 476 struct bnxt_qplib_hwq hwq; global() member [all...] |
H A D | qplib_sp.c | 520 if (mrw->hwq.max_elements) in bnxt_qplib_free_mrw() 521 bnxt_qplib_free_hwq(res, &mrw->hwq); in bnxt_qplib_free_mrw() 585 if (mrw->hwq.max_elements) { in bnxt_qplib_dereg_mrw() 588 bnxt_qplib_free_hwq(res, &mrw->hwq); in bnxt_qplib_dereg_mrw() 612 /* Free the hwq if it already exist, must be a rereg */ in bnxt_qplib_reg_mr() 613 if (mr->hwq.max_elements) in bnxt_qplib_reg_mr() 614 bnxt_qplib_free_hwq(res, &mr->hwq); in bnxt_qplib_reg_mr() 624 rc = bnxt_qplib_alloc_init_hwq(&mr->hwq, &hwq_attr); in bnxt_qplib_reg_mr() 637 if (mr->hwq.level == PBL_LVL_MAX) { in bnxt_qplib_reg_mr() 643 level = mr->hwq in bnxt_qplib_reg_mr() [all...] |
/kernel/linux/linux-5.10/include/scsi/ |
H A D | scsi_tcq.h | 26 u16 hwq; in scsi_host_find_tag() local 31 hwq = blk_mq_unique_tag_to_hwq(tag); in scsi_host_find_tag() 32 if (hwq < shost->tag_set.nr_hw_queues) { in scsi_host_find_tag() 33 req = blk_mq_tag_to_rq(shost->tag_set.tags[hwq], in scsi_host_find_tag()
|
/kernel/linux/linux-6.6/include/scsi/ |
H A D | scsi_tcq.h | 26 u16 hwq; in scsi_host_find_tag() local 31 hwq = blk_mq_unique_tag_to_hwq(tag); in scsi_host_find_tag() 32 if (hwq < shost->tag_set.nr_hw_queues) { in scsi_host_find_tag() 33 req = blk_mq_tag_to_rq(shost->tag_set.tags[hwq], in scsi_host_find_tag()
|
/kernel/linux/linux-5.10/drivers/net/wireless/mediatek/mt76/mt7915/ |
H A D | dma.c | 11 struct mt76_queue *hwq; in mt7915_init_tx_queues() local 14 hwq = devm_kzalloc(dev->mt76.dev, sizeof(*hwq), GFP_KERNEL); in mt7915_init_tx_queues() 15 if (!hwq) in mt7915_init_tx_queues() 18 err = mt76_queue_alloc(dev, hwq, MT7915_TXQ_BAND0, n_desc, 0, in mt7915_init_tx_queues() 24 dev->mt76.q_tx[i] = hwq; in mt7915_init_tx_queues() 32 struct mt76_queue *hwq; in mt7915_init_mcu_queue() local 35 hwq = devm_kzalloc(dev->mt76.dev, sizeof(*hwq), GFP_KERNEL); in mt7915_init_mcu_queue() 36 if (!hwq) in mt7915_init_mcu_queue() [all...] |
/kernel/linux/linux-5.10/drivers/net/wireless/mediatek/mt76/mt7603/ |
H A D | dma.c | 10 struct mt76_queue *hwq; in mt7603_init_tx_queue() local 13 hwq = devm_kzalloc(dev->mt76.dev, sizeof(*hwq), GFP_KERNEL); in mt7603_init_tx_queue() 14 if (!hwq) in mt7603_init_tx_queue() 17 err = mt76_queue_alloc(dev, hwq, idx, n_desc, 0, MT_TX_RING_BASE); in mt7603_init_tx_queue() 21 dev->mt76.q_tx[qid] = hwq; in mt7603_init_tx_queue()
|
/kernel/linux/linux-5.10/drivers/net/wireless/mediatek/mt76/mt7615/ |
H A D | dma.c | 17 struct mt76_queue *hwq; in mt7615_init_tx_queue() local 20 hwq = devm_kzalloc(dev->mt76.dev, sizeof(*hwq), GFP_KERNEL); in mt7615_init_tx_queue() 21 if (!hwq) in mt7615_init_tx_queue() 24 err = mt76_queue_alloc(dev, hwq, idx, n_desc, 0, MT_TX_RING_BASE); in mt7615_init_tx_queue() 28 dev->mt76.q_tx[qid] = hwq; in mt7615_init_tx_queue()
|
/kernel/linux/linux-5.10/drivers/net/wireless/mediatek/mt76/ |
H A D | tx.c | 347 struct mt76_queue *hwq = dev->q_tx[MT_TXQ_PSD]; in mt76_release_buffered_frames() local 350 spin_lock_bh(&hwq->lock); in mt76_release_buffered_frames() 374 dev->queue_ops->kick(dev, hwq); in mt76_release_buffered_frames() 379 spin_unlock_bh(&hwq->lock); in mt76_release_buffered_frames() 548 struct mt76_queue *hwq; in mt76_stop_tx_queues() local 554 hwq = dev->q_tx[mt76_txq_get_qid(txq)]; in mt76_stop_tx_queues() 557 spin_lock_bh(&hwq->lock); in mt76_stop_tx_queues() 559 spin_unlock_bh(&hwq->lock); in mt76_stop_tx_queues()
|
H A D | mt76x02_mmio.c | 109 struct mt76_queue *hwq; in mt76x02_init_tx_queue() local 112 hwq = devm_kzalloc(dev->mt76.dev, sizeof(*hwq), GFP_KERNEL); in mt76x02_init_tx_queue() 113 if (!hwq) in mt76x02_init_tx_queue() 116 err = mt76_queue_alloc(dev, hwq, idx, n_desc, 0, MT_TX_RING_BASE); in mt76x02_init_tx_queue() 120 dev->mt76.q_tx[qid] = hwq; in mt76x02_init_tx_queue()
|
/kernel/linux/linux-5.10/drivers/net/wireless/ti/wlcore/ |
H A D | tx.c | 1201 int hwq = wlcore_tx_get_mac80211_queue(wlvif, queue); in wlcore_stop_queue_locked() local 1202 bool stopped = !!wl->queue_stop_reasons[hwq]; in wlcore_stop_queue_locked() 1205 WARN_ON_ONCE(test_and_set_bit(reason, &wl->queue_stop_reasons[hwq])); in wlcore_stop_queue_locked() 1210 ieee80211_stop_queue(wl->hw, hwq); in wlcore_stop_queue_locked() 1227 int hwq = wlcore_tx_get_mac80211_queue(wlvif, queue); in wlcore_wake_queue() local 1232 WARN_ON_ONCE(!test_and_clear_bit(reason, &wl->queue_stop_reasons[hwq])); in wlcore_wake_queue() 1234 if (wl->queue_stop_reasons[hwq]) in wlcore_wake_queue() 1237 ieee80211_wake_queue(wl->hw, hwq); in wlcore_wake_queue() 1304 int hwq = wlcore_tx_get_mac80211_queue(wlvif, queue); in wlcore_is_queue_stopped_by_reason_locked() local 1307 return test_bit(reason, &wl->queue_stop_reasons[hwq]); in wlcore_is_queue_stopped_by_reason_locked() 1313 int hwq = wlcore_tx_get_mac80211_queue(wlvif, queue); wlcore_is_queue_stopped_locked() local [all...] |