/kernel/linux/linux-5.10/drivers/net/vmxnet3/ |
H A D | vmxnet3_drv.c | 103 vmxnet3_tq_stopped(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) in vmxnet3_tq_stopped() argument 105 return tq->stopped; in vmxnet3_tq_stopped() 110 vmxnet3_tq_start(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) in vmxnet3_tq_start() argument 112 tq->stopped = false; in vmxnet3_tq_start() 113 netif_start_subqueue(adapter->netdev, tq - adapter->tx_queue); in vmxnet3_tq_start() 118 vmxnet3_tq_wake(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) in vmxnet3_tq_wake() argument 120 tq->stopped = false; in vmxnet3_tq_wake() 121 netif_wake_subqueue(adapter->netdev, (tq - adapter->tx_queue)); in vmxnet3_tq_wake() 126 vmxnet3_tq_stop(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) in vmxnet3_tq_stop() argument 128 tq in vmxnet3_tq_stop() 329 vmxnet3_unmap_pkt(u32 eop_idx, struct vmxnet3_tx_queue *tq, struct pci_dev *pdev, struct vmxnet3_adapter *adapter) vmxnet3_unmap_pkt() argument 364 vmxnet3_tq_tx_complete(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) vmxnet3_tq_tx_complete() argument 400 vmxnet3_tq_cleanup(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) vmxnet3_tq_cleanup() argument 433 vmxnet3_tq_destroy(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) vmxnet3_tq_destroy() argument 475 vmxnet3_tq_init(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) vmxnet3_tq_init() argument 505 vmxnet3_tq_create(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) vmxnet3_tq_create() argument 676 vmxnet3_map_pkt(struct sk_buff *skb, struct vmxnet3_tx_ctx *ctx, struct vmxnet3_tx_queue *tq, struct pci_dev *pdev, struct vmxnet3_adapter *adapter) vmxnet3_map_pkt() argument 840 vmxnet3_parse_hdr(struct sk_buff *skb, struct vmxnet3_tx_queue *tq, struct vmxnet3_tx_ctx *ctx, struct vmxnet3_adapter *adapter) vmxnet3_parse_hdr() argument 943 vmxnet3_copy_hdr(struct sk_buff *skb, struct vmxnet3_tx_queue *tq, struct vmxnet3_tx_ctx *ctx, struct vmxnet3_adapter *adapter) vmxnet3_copy_hdr() argument 1023 vmxnet3_tq_xmit(struct sk_buff *skb, struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter, struct net_device *netdev) vmxnet3_tq_xmit() argument 2003 struct vmxnet3_tx_queue *tq = vmxnet3_poll_rx_only() local 2028 struct vmxnet3_tx_queue *tq = data; vmxnet3_msix_tx() local 2540 struct vmxnet3_tx_queue *tq = &adapter->tx_queue[i]; vmxnet3_setup_driver_shared() local 2957 struct vmxnet3_tx_queue *tq = &adapter->tx_queue[i]; vmxnet3_create_queues() local [all...] |
H A D | vmxnet3_ethtool.c | 40 /* per tq stats maintained by the device */ 57 /* per tq stats maintained by the driver */ 478 struct vmxnet3_tx_queue *tq = &adapter->tx_queue[i]; in vmxnet3_get_regs() local 483 buf[j++] = VMXNET3_GET_ADDR_LO(tq->tx_ring.basePA); in vmxnet3_get_regs() 484 buf[j++] = VMXNET3_GET_ADDR_HI(tq->tx_ring.basePA); in vmxnet3_get_regs() 485 buf[j++] = tq->tx_ring.size; in vmxnet3_get_regs() 486 buf[j++] = tq->tx_ring.next2fill; in vmxnet3_get_regs() 487 buf[j++] = tq->tx_ring.next2comp; in vmxnet3_get_regs() 488 buf[j++] = tq->tx_ring.gen; in vmxnet3_get_regs() 490 buf[j++] = VMXNET3_GET_ADDR_LO(tq in vmxnet3_get_regs() [all...] |
/kernel/linux/linux-6.6/drivers/net/vmxnet3/ |
H A D | vmxnet3_xdp.c | 22 struct vmxnet3_tx_queue *tq; in vmxnet3_xdp_get_tq() local 29 tq = &adapter->tx_queue[cpu]; in vmxnet3_xdp_get_tq() 31 tq = &adapter->tx_queue[reciprocal_scale(cpu, tq_number)]; in vmxnet3_xdp_get_tq() 33 return tq; in vmxnet3_xdp_get_tq() 117 struct vmxnet3_tx_queue *tq, bool dma_map) in vmxnet3_xdp_xmit_frame() 127 dw2 = (tq->tx_ring.gen ^ 0x1) << VMXNET3_TXD_GEN_SHIFT; in vmxnet3_xdp_xmit_frame() 129 ctx.sop_txd = tq->tx_ring.base + tq->tx_ring.next2fill; in vmxnet3_xdp_xmit_frame() 133 tbi = tq->buf_info + tq in vmxnet3_xdp_xmit_frame() 115 vmxnet3_xdp_xmit_frame(struct vmxnet3_adapter *adapter, struct xdp_frame *xdpf, struct vmxnet3_tx_queue *tq, bool dma_map) vmxnet3_xdp_xmit_frame() argument 204 struct vmxnet3_tx_queue *tq; vmxnet3_xdp_xmit_back() local 227 struct vmxnet3_tx_queue *tq; vmxnet3_xdp_xmit() local [all...] |
H A D | vmxnet3_drv.c | 116 vmxnet3_tq_stopped(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) in vmxnet3_tq_stopped() argument 118 return tq->stopped; in vmxnet3_tq_stopped() 123 vmxnet3_tq_start(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) in vmxnet3_tq_start() argument 125 tq->stopped = false; in vmxnet3_tq_start() 126 netif_start_subqueue(adapter->netdev, tq - adapter->tx_queue); in vmxnet3_tq_start() 131 vmxnet3_tq_wake(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) in vmxnet3_tq_wake() argument 133 tq->stopped = false; in vmxnet3_tq_wake() 134 netif_wake_subqueue(adapter->netdev, (tq - adapter->tx_queue)); in vmxnet3_tq_wake() 139 vmxnet3_tq_stop(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) in vmxnet3_tq_stop() argument 141 tq in vmxnet3_tq_stop() 358 vmxnet3_unmap_pkt(u32 eop_idx, struct vmxnet3_tx_queue *tq, struct pci_dev *pdev, struct vmxnet3_adapter *adapter, struct xdp_frame_bulk *bq) vmxnet3_unmap_pkt() argument 401 vmxnet3_tq_tx_complete(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) vmxnet3_tq_tx_complete() argument 443 vmxnet3_tq_cleanup(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) vmxnet3_tq_cleanup() argument 486 vmxnet3_tq_destroy(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) vmxnet3_tq_destroy() argument 524 vmxnet3_tq_init(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) vmxnet3_tq_init() argument 554 vmxnet3_tq_create(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) vmxnet3_tq_create() argument 734 vmxnet3_map_pkt(struct sk_buff *skb, struct vmxnet3_tx_ctx *ctx, struct vmxnet3_tx_queue *tq, struct pci_dev *pdev, struct vmxnet3_adapter *adapter) vmxnet3_map_pkt() argument 898 vmxnet3_parse_hdr(struct sk_buff *skb, struct vmxnet3_tx_queue *tq, struct vmxnet3_tx_ctx *ctx, struct vmxnet3_adapter *adapter) vmxnet3_parse_hdr() argument 1001 vmxnet3_copy_hdr(struct sk_buff *skb, struct vmxnet3_tx_queue *tq, struct vmxnet3_tx_ctx *ctx, struct vmxnet3_adapter *adapter) vmxnet3_copy_hdr() argument 1081 vmxnet3_tq_xmit(struct sk_buff *skb, struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter, struct net_device *netdev) vmxnet3_tq_xmit() argument 2267 struct vmxnet3_tx_queue *tq = vmxnet3_poll_rx_only() local 2292 struct vmxnet3_tx_queue *tq = data; vmxnet3_msix_tx() local 2805 struct vmxnet3_tx_queue *tq = &adapter->tx_queue[i]; vmxnet3_setup_driver_shared() local 3293 struct vmxnet3_tx_queue *tq = &adapter->tx_queue[i]; vmxnet3_create_queues() local [all...] |
H A D | vmxnet3_ethtool.c | 41 /* per tq stats maintained by the device */ 58 /* per tq stats maintained by the driver */ 550 struct vmxnet3_tx_queue *tq = &adapter->tx_queue[i]; in vmxnet3_get_regs() local 555 buf[j++] = VMXNET3_GET_ADDR_LO(tq->tx_ring.basePA); in vmxnet3_get_regs() 556 buf[j++] = VMXNET3_GET_ADDR_HI(tq->tx_ring.basePA); in vmxnet3_get_regs() 557 buf[j++] = tq->tx_ring.size; in vmxnet3_get_regs() 558 buf[j++] = tq->tx_ring.next2fill; in vmxnet3_get_regs() 559 buf[j++] = tq->tx_ring.next2comp; in vmxnet3_get_regs() 560 buf[j++] = tq->tx_ring.gen; in vmxnet3_get_regs() 562 buf[j++] = VMXNET3_GET_ADDR_LO(tq in vmxnet3_get_regs() [all...] |
/kernel/linux/linux-5.10/drivers/net/wireless/ath/ath5k/ |
H A D | qcu.c | 286 struct ath5k_txq_info *tq = &ah->ah_txq[queue]; in ath5k_hw_set_tx_retry_limits() local 292 (tq->tqi_cw_min << AR5K_NODCU_RETRY_LMT_CW_MIN_S) in ath5k_hw_set_tx_retry_limits() 326 struct ath5k_txq_info *tq = &ah->ah_txq[queue]; in ath5k_hw_reset_tx_queue() local 333 (tq->tqi_type == AR5K_TX_QUEUE_INACTIVE)) in ath5k_hw_reset_tx_queue() 341 AR5K_REG_SM(tq->tqi_cw_min, AR5K_DCU_LCL_IFS_CW_MIN) | in ath5k_hw_reset_tx_queue() 342 AR5K_REG_SM(tq->tqi_cw_max, AR5K_DCU_LCL_IFS_CW_MAX) | in ath5k_hw_reset_tx_queue() 343 AR5K_REG_SM(tq->tqi_aifs, AR5K_DCU_LCL_IFS_AIFS), in ath5k_hw_reset_tx_queue() 366 if (tq->tqi_cbr_period) { in ath5k_hw_reset_tx_queue() 367 ath5k_hw_reg_write(ah, AR5K_REG_SM(tq->tqi_cbr_period, in ath5k_hw_reset_tx_queue() 369 AR5K_REG_SM(tq in ath5k_hw_reset_tx_queue() [all...] |
/kernel/linux/linux-6.6/drivers/net/wireless/ath/ath5k/ |
H A D | qcu.c | 286 struct ath5k_txq_info *tq = &ah->ah_txq[queue]; in ath5k_hw_set_tx_retry_limits() local 292 (tq->tqi_cw_min << AR5K_NODCU_RETRY_LMT_CW_MIN_S) in ath5k_hw_set_tx_retry_limits() 326 struct ath5k_txq_info *tq = &ah->ah_txq[queue]; in ath5k_hw_reset_tx_queue() local 333 (tq->tqi_type == AR5K_TX_QUEUE_INACTIVE)) in ath5k_hw_reset_tx_queue() 341 AR5K_REG_SM(tq->tqi_cw_min, AR5K_DCU_LCL_IFS_CW_MIN) | in ath5k_hw_reset_tx_queue() 342 AR5K_REG_SM(tq->tqi_cw_max, AR5K_DCU_LCL_IFS_CW_MAX) | in ath5k_hw_reset_tx_queue() 343 AR5K_REG_SM(tq->tqi_aifs, AR5K_DCU_LCL_IFS_AIFS), in ath5k_hw_reset_tx_queue() 366 if (tq->tqi_cbr_period) { in ath5k_hw_reset_tx_queue() 367 ath5k_hw_reg_write(ah, AR5K_REG_SM(tq->tqi_cbr_period, in ath5k_hw_reset_tx_queue() 369 AR5K_REG_SM(tq in ath5k_hw_reset_tx_queue() [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/chelsio/cxgb4vf/ |
H A D | sge.c | 229 * @tq: the TX queue 233 static inline unsigned int txq_avail(const struct sge_txq *tq) in txq_avail() argument 235 return tq->size - 1 - tq->in_use; in txq_avail() 307 const struct ulptx_sgl *sgl, const struct sge_txq *tq) in unmap_sgl() 326 if (likely((u8 *)(p + 1) <= (u8 *)tq->stat)) { in unmap_sgl() 333 } else if ((u8 *)p == (u8 *)tq->stat) { in unmap_sgl() 334 p = (const struct ulptx_sge_pair *)tq->desc; in unmap_sgl() 336 } else if ((u8 *)p + 8 == (u8 *)tq->stat) { in unmap_sgl() 337 const __be64 *addr = (const __be64 *)tq in unmap_sgl() 306 unmap_sgl(struct device *dev, const struct sk_buff *skb, const struct ulptx_sgl *sgl, const struct sge_txq *tq) unmap_sgl() argument 377 free_tx_desc(struct adapter *adapter, struct sge_txq *tq, unsigned int n, bool unmap) free_tx_desc() argument 411 reclaimable(const struct sge_txq *tq) reclaimable() argument 430 reclaim_completed_tx(struct adapter *adapter, struct sge_txq *tq, bool unmap) reclaim_completed_tx() argument 901 write_sgl(const struct sk_buff *skb, struct sge_txq *tq, struct ulptx_sgl *sgl, u64 *end, unsigned int start, const dma_addr_t *addr) write_sgl() argument 964 ring_tx_db(struct adapter *adapter, struct sge_txq *tq, int n) ring_tx_db() argument 1052 inline_tx_skb(const struct sk_buff *skb, const struct sge_txq *tq, void *pos) inline_tx_skb() argument 1142 txq_advance(struct sge_txq *tq, unsigned int n) txq_advance() argument 1420 struct sge_txq *tq = &txq->q; t4vf_eth_xmit() local 2514 free_txq(struct adapter *adapter, struct sge_txq *tq) free_txq() argument [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/chelsio/cxgb4vf/ |
H A D | sge.c | 229 * @tq: the TX queue 233 static inline unsigned int txq_avail(const struct sge_txq *tq) in txq_avail() argument 235 return tq->size - 1 - tq->in_use; in txq_avail() 307 const struct ulptx_sgl *sgl, const struct sge_txq *tq) in unmap_sgl() 326 if (likely((u8 *)(p + 1) <= (u8 *)tq->stat)) { in unmap_sgl() 333 } else if ((u8 *)p == (u8 *)tq->stat) { in unmap_sgl() 334 p = (const struct ulptx_sge_pair *)tq->desc; in unmap_sgl() 336 } else if ((u8 *)p + 8 == (u8 *)tq->stat) { in unmap_sgl() 337 const __be64 *addr = (const __be64 *)tq in unmap_sgl() 306 unmap_sgl(struct device *dev, const struct sk_buff *skb, const struct ulptx_sgl *sgl, const struct sge_txq *tq) unmap_sgl() argument 377 free_tx_desc(struct adapter *adapter, struct sge_txq *tq, unsigned int n, bool unmap) free_tx_desc() argument 411 reclaimable(const struct sge_txq *tq) reclaimable() argument 430 reclaim_completed_tx(struct adapter *adapter, struct sge_txq *tq, bool unmap) reclaim_completed_tx() argument 901 write_sgl(const struct sk_buff *skb, struct sge_txq *tq, struct ulptx_sgl *sgl, u64 *end, unsigned int start, const dma_addr_t *addr) write_sgl() argument 964 ring_tx_db(struct adapter *adapter, struct sge_txq *tq, int n) ring_tx_db() argument 1052 inline_tx_skb(const struct sk_buff *skb, const struct sge_txq *tq, void *pos) inline_tx_skb() argument 1142 txq_advance(struct sge_txq *tq, unsigned int n) txq_advance() argument 1417 struct sge_txq *tq = &txq->q; t4vf_eth_xmit() local 2511 free_txq(struct adapter *adapter, struct sge_txq *tq) free_txq() argument [all...] |
/kernel/linux/linux-6.6/drivers/net/can/dev/ |
H A D | bittiming.c | 85 brp64 = (u64)priv->clock.freq * (u64)bt->tq; in can_fixup_bittiming() 107 bt->tq = DIV_U64_ROUND_CLOSEST(mul_u32_u32(bt->brp, NSEC_PER_SEC), in can_fixup_bittiming() 141 * alternatively the CAN timing parameters (tq, prop_seg, etc.) are in can_get_bittiming() 144 if (!bt->tq && bt->bitrate && btc) in can_get_bittiming() 146 if (bt->tq && !bt->bitrate && btc) in can_get_bittiming() 148 if (!bt->tq && bt->bitrate && bitrate_const) in can_get_bittiming()
|
/kernel/linux/linux-5.10/drivers/media/v4l2-core/ |
H A D | v4l2-jpeg.c | 211 int c, h_v, tq; in jpeg_parse_frame_header() local 232 tq = jpeg_get_byte(stream); in jpeg_parse_frame_header() 233 if (tq < 0) in jpeg_parse_frame_header() 234 return tq; in jpeg_parse_frame_header() 241 component->quantization_table_selector = tq; in jpeg_parse_frame_header() 317 u8 pq, tq, *qk; in jpeg_parse_quantization_tables() local 335 tq = pq_tq & 0xf; in jpeg_parse_quantization_tables() 336 if (tq > 3) in jpeg_parse_quantization_tables() 346 tables[tq].start = qk; in jpeg_parse_quantization_tables() 347 tables[tq] in jpeg_parse_quantization_tables() [all...] |
/kernel/linux/linux-6.6/drivers/media/v4l2-core/ |
H A D | v4l2-jpeg.c | 212 int c, h_v, tq; in jpeg_parse_frame_header() local 233 tq = jpeg_get_byte(stream); in jpeg_parse_frame_header() 234 if (tq < 0) in jpeg_parse_frame_header() 235 return tq; in jpeg_parse_frame_header() 242 component->quantization_table_selector = tq; in jpeg_parse_frame_header() 318 u8 pq, tq, *qk; in jpeg_parse_quantization_tables() local 336 tq = pq_tq & 0xf; in jpeg_parse_quantization_tables() 337 if (tq > 3) in jpeg_parse_quantization_tables() 347 tables[tq].start = qk; in jpeg_parse_quantization_tables() 348 tables[tq] in jpeg_parse_quantization_tables() [all...] |
/kernel/linux/linux-5.10/drivers/net/ |
H A D | ifb.c | 51 struct sk_buff_head tq; member 69 skb = skb_peek(&txp->tq); in ifb_ri_tasklet() 73 skb_queue_splice_tail_init(&txp->rq, &txp->tq); in ifb_ri_tasklet() 77 while ((skb = __skb_dequeue(&txp->tq)) != NULL) { in ifb_ri_tasklet() 94 if (skb_queue_len(&txp->tq) != 0) in ifb_ri_tasklet() 172 __skb_queue_head_init(&txp->tq); in ifb_dev_init() 206 __skb_queue_purge(&txp->tq); in ifb_dev_free()
|
/kernel/linux/linux-5.10/drivers/input/serio/ |
H A D | hp_sdc.c | 190 curr = hp_sdc.tq[hp_sdc.rcurr]; in hp_sdc_take() 316 curr = hp_sdc.tq[hp_sdc.rcurr]; in hp_sdc_tasklet() 378 if (hp_sdc.tq[curridx] != NULL) in hp_sdc_put() 392 if (hp_sdc.tq[curridx] != NULL) in hp_sdc_put() 412 curr = hp_sdc.tq[curridx]; in hp_sdc_put() 416 hp_sdc.tq[curridx] = NULL; in hp_sdc_put() 430 hp_sdc.tq[curridx] = NULL; in hp_sdc_put() 573 hp_sdc.tq[curridx] = NULL; in hp_sdc_put() 609 if (hp_sdc.tq[i] == this) in __hp_sdc_enqueue_transaction() 617 if (hp_sdc.tq[ in __hp_sdc_enqueue_transaction() [all...] |
/kernel/linux/linux-6.6/drivers/input/serio/ |
H A D | hp_sdc.c | 190 curr = hp_sdc.tq[hp_sdc.rcurr]; in hp_sdc_take() 316 curr = hp_sdc.tq[hp_sdc.rcurr]; in hp_sdc_tasklet() 378 if (hp_sdc.tq[curridx] != NULL) in hp_sdc_put() 392 if (hp_sdc.tq[curridx] != NULL) in hp_sdc_put() 412 curr = hp_sdc.tq[curridx]; in hp_sdc_put() 416 hp_sdc.tq[curridx] = NULL; in hp_sdc_put() 430 hp_sdc.tq[curridx] = NULL; in hp_sdc_put() 573 hp_sdc.tq[curridx] = NULL; in hp_sdc_put() 609 if (hp_sdc.tq[i] == this) in __hp_sdc_enqueue_transaction() 617 if (hp_sdc.tq[ in __hp_sdc_enqueue_transaction() [all...] |
/kernel/linux/linux-5.10/net/tipc/ |
H A D | trace.h | 284 TP_PROTO(struct tipc_link *r, u16 f, u16 t, struct sk_buff_head *tq), 286 TP_ARGS(r, f, t, tq), 301 __entry->len = skb_queue_len(tq); 303 msg_seqno(buf_msg(skb_peek(tq))) : 0; 305 msg_seqno(buf_msg(skb_peek_tail(tq))) : 0; 314 TP_PROTO(struct tipc_link *r, u16 f, u16 t, struct sk_buff_head *tq), 315 TP_ARGS(r, f, t, tq), 320 TP_PROTO(struct tipc_link *r, u16 f, u16 t, struct sk_buff_head *tq), 321 TP_ARGS(r, f, t, tq),
|
/kernel/linux/linux-6.6/net/tipc/ |
H A D | trace.h | 284 TP_PROTO(struct tipc_link *r, u16 f, u16 t, struct sk_buff_head *tq), 286 TP_ARGS(r, f, t, tq), 301 __entry->len = skb_queue_len(tq); 303 msg_seqno(buf_msg(skb_peek(tq))) : 0; 305 msg_seqno(buf_msg(skb_peek_tail(tq))) : 0; 314 TP_PROTO(struct tipc_link *r, u16 f, u16 t, struct sk_buff_head *tq), 315 TP_ARGS(r, f, t, tq), 320 TP_PROTO(struct tipc_link *r, u16 f, u16 t, struct sk_buff_head *tq), 321 TP_ARGS(r, f, t, tq),
|
/kernel/linux/linux-5.10/drivers/input/keyboard/ |
H A D | sunkbd.c | 63 struct work_struct tq; member 103 schedule_work(&sunkbd->tq); in sunkbd_interrupt() 230 struct sunkbd *sunkbd = container_of(work, struct sunkbd, tq); in sunkbd_reinit() 253 cancel_work_sync(&sunkbd->tq); in sunkbd_enable() 277 INIT_WORK(&sunkbd->tq, sunkbd_reinit); in sunkbd_connect()
|
/kernel/linux/linux-6.6/drivers/input/keyboard/ |
H A D | sunkbd.c | 60 struct work_struct tq; member 100 schedule_work(&sunkbd->tq); in sunkbd_interrupt() 227 struct sunkbd *sunkbd = container_of(work, struct sunkbd, tq); in sunkbd_reinit() 250 cancel_work_sync(&sunkbd->tq); in sunkbd_enable() 274 INIT_WORK(&sunkbd->tq, sunkbd_reinit); in sunkbd_connect()
|
/kernel/linux/linux-6.6/drivers/net/ |
H A D | ifb.c | 53 struct sk_buff_head tq; member 96 skb = skb_peek(&txp->tq); in ifb_ri_tasklet() 100 skb_queue_splice_tail_init(&txp->rq, &txp->tq); in ifb_ri_tasklet() 104 while ((skb = __skb_dequeue(&txp->tq)) != NULL) { in ifb_ri_tasklet() 120 if (skb_queue_len(&txp->tq) != 0) in ifb_ri_tasklet() 198 __skb_queue_head_init(&txp->tq); in ifb_dev_init() 305 __skb_queue_purge(&txp->tq); in ifb_dev_free()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gem/selftests/ |
H A D | i915_gem_context.c | 656 struct i915_request *tq[5] = {}; in igt_ctx_exec() local 714 err = throttle(ce, tq, ARRAY_SIZE(tq)); in igt_ctx_exec() 749 throttle_release(tq, ARRAY_SIZE(tq)); in igt_ctx_exec() 766 struct i915_request *tq[5] = {}; in igt_shared_ctx_exec() local 852 err = throttle(ce, tq, ARRAY_SIZE(tq)); in igt_shared_ctx_exec() 888 throttle_release(tq, ARRAY_SIZE(tq)); in igt_shared_ctx_exec() 1355 struct i915_request *tq[5] = {}; igt_ctx_readonly() local [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/gem/selftests/ |
H A D | i915_gem_context.c | 677 struct i915_request *tq[5] = {}; in igt_ctx_exec() local 735 err = throttle(ce, tq, ARRAY_SIZE(tq)); in igt_ctx_exec() 770 throttle_release(tq, ARRAY_SIZE(tq)); in igt_ctx_exec() 787 struct i915_request *tq[5] = {}; in igt_shared_ctx_exec() local 869 err = throttle(ce, tq, ARRAY_SIZE(tq)); in igt_shared_ctx_exec() 905 throttle_release(tq, ARRAY_SIZE(tq)); in igt_shared_ctx_exec() 1371 struct i915_request *tq[5] = {}; igt_ctx_readonly() local [all...] |
/kernel/linux/linux-5.10/net/batman-adv/ |
H A D | bat_iv_ogm.c | 220 batadv_ogm_packet->tq = BATADV_TQ_MAX_VALUE; in batadv_iv_ogm_iface_enable() 296 static u8 batadv_hop_penalty(u8 tq, const struct batadv_priv *bat_priv) in batadv_hop_penalty() argument 301 new_tq = tq * (BATADV_TQ_MAX_VALUE - hop_penalty); in batadv_hop_penalty() 375 batadv_ogm_packet->tq, batadv_ogm_packet->ttl, in batadv_iv_ogm_send_to_if() 719 batadv_ogm_packet->tq = batadv_hop_penalty(batadv_ogm_packet->tq, in batadv_iv_ogm_forward() 723 "Forwarding packet: tq: %i, ttl: %i\n", in batadv_iv_ogm_forward() 724 batadv_ogm_packet->tq, batadv_ogm_packet->ttl); in batadv_iv_ogm_forward() 1002 batadv_ogm_packet->tq); in batadv_iv_ogm_orig_update() 1061 * batadv_iv_ogm_calc_tq() - calculate tq fo [all...] |
/kernel/linux/linux-6.6/net/batman-adv/ |
H A D | bat_iv_ogm.c | 218 batadv_ogm_packet->tq = BATADV_TQ_MAX_VALUE; in batadv_iv_ogm_iface_enable() 294 static u8 batadv_hop_penalty(u8 tq, const struct batadv_priv *bat_priv) in batadv_hop_penalty() argument 299 new_tq = tq * (BATADV_TQ_MAX_VALUE - hop_penalty); in batadv_hop_penalty() 373 batadv_ogm_packet->tq, batadv_ogm_packet->ttl, in batadv_iv_ogm_send_to_if() 716 batadv_ogm_packet->tq = batadv_hop_penalty(batadv_ogm_packet->tq, in batadv_iv_ogm_forward() 720 "Forwarding packet: tq: %i, ttl: %i\n", in batadv_iv_ogm_forward() 721 batadv_ogm_packet->tq, batadv_ogm_packet->ttl); in batadv_iv_ogm_forward() 998 batadv_ogm_packet->tq); in batadv_iv_ogm_orig_update() 1053 * batadv_iv_ogm_calc_tq() - calculate tq fo [all...] |
/kernel/linux/linux-5.10/include/uapi/linux/can/ |
H A D | netlink.h | 34 __u32 tq; /* Time quanta (TQ) in nanoseconds */ member
|