/kernel/linux/linux-5.10/drivers/media/v4l2-core/ |
H A D | v4l2-mem2mem.c | 126 struct v4l2_m2m_queue_ctx *q_ctx; in v4l2_m2m_get_vq() local 128 q_ctx = get_queue_ctx(m2m_ctx, type); in v4l2_m2m_get_vq() 129 if (!q_ctx) in v4l2_m2m_get_vq() 132 return &q_ctx->q; in v4l2_m2m_get_vq() 136 struct vb2_v4l2_buffer *v4l2_m2m_next_buf(struct v4l2_m2m_queue_ctx *q_ctx) in v4l2_m2m_next_buf() argument 141 spin_lock_irqsave(&q_ctx->rdy_spinlock, flags); in v4l2_m2m_next_buf() 143 if (list_empty(&q_ctx->rdy_queue)) { in v4l2_m2m_next_buf() 144 spin_unlock_irqrestore(&q_ctx->rdy_spinlock, flags); in v4l2_m2m_next_buf() 148 b = list_first_entry(&q_ctx->rdy_queue, struct v4l2_m2m_buffer, list); in v4l2_m2m_next_buf() 149 spin_unlock_irqrestore(&q_ctx in v4l2_m2m_next_buf() 154 v4l2_m2m_last_buf(struct v4l2_m2m_queue_ctx *q_ctx) v4l2_m2m_last_buf() argument 172 v4l2_m2m_buf_remove(struct v4l2_m2m_queue_ctx *q_ctx) v4l2_m2m_buf_remove() argument 191 v4l2_m2m_buf_remove_by_buf(struct v4l2_m2m_queue_ctx *q_ctx, struct vb2_v4l2_buffer *vbuf) v4l2_m2m_buf_remove_by_buf() argument 206 v4l2_m2m_buf_remove_by_idx(struct v4l2_m2m_queue_ctx *q_ctx, unsigned int idx) v4l2_m2m_buf_remove_by_idx() argument 872 struct v4l2_m2m_queue_ctx *q_ctx; v4l2_m2m_streamoff() local 1254 struct v4l2_m2m_queue_ctx *q_ctx; v4l2_m2m_buf_queue() local [all...] |
/kernel/linux/linux-6.6/drivers/media/v4l2-core/ |
H A D | v4l2-mem2mem.c | 126 struct v4l2_m2m_queue_ctx *q_ctx; in v4l2_m2m_get_vq() local 128 q_ctx = get_queue_ctx(m2m_ctx, type); in v4l2_m2m_get_vq() 129 if (!q_ctx) in v4l2_m2m_get_vq() 132 return &q_ctx->q; in v4l2_m2m_get_vq() 136 struct vb2_v4l2_buffer *v4l2_m2m_next_buf(struct v4l2_m2m_queue_ctx *q_ctx) in v4l2_m2m_next_buf() argument 141 spin_lock_irqsave(&q_ctx->rdy_spinlock, flags); in v4l2_m2m_next_buf() 143 if (list_empty(&q_ctx->rdy_queue)) { in v4l2_m2m_next_buf() 144 spin_unlock_irqrestore(&q_ctx->rdy_spinlock, flags); in v4l2_m2m_next_buf() 148 b = list_first_entry(&q_ctx->rdy_queue, struct v4l2_m2m_buffer, list); in v4l2_m2m_next_buf() 149 spin_unlock_irqrestore(&q_ctx in v4l2_m2m_next_buf() 154 v4l2_m2m_last_buf(struct v4l2_m2m_queue_ctx *q_ctx) v4l2_m2m_last_buf() argument 172 v4l2_m2m_buf_remove(struct v4l2_m2m_queue_ctx *q_ctx) v4l2_m2m_buf_remove() argument 191 v4l2_m2m_buf_remove_by_buf(struct v4l2_m2m_queue_ctx *q_ctx, struct vb2_v4l2_buffer *vbuf) v4l2_m2m_buf_remove_by_buf() argument 206 v4l2_m2m_buf_remove_by_idx(struct v4l2_m2m_queue_ctx *q_ctx, unsigned int idx) v4l2_m2m_buf_remove_by_idx() argument 873 struct v4l2_m2m_queue_ctx *q_ctx; v4l2_m2m_streamoff() local 1285 struct v4l2_m2m_queue_ctx *q_ctx; v4l2_m2m_buf_queue() local [all...] |
/kernel/linux/linux-5.10/include/media/ |
H A D | v4l2-mem2mem.h | 623 * @q_ctx: pointer to struct @v4l2_m2m_queue_ctx 625 struct vb2_v4l2_buffer *v4l2_m2m_next_buf(struct v4l2_m2m_queue_ctx *q_ctx); 654 * @q_ctx: pointer to struct @v4l2_m2m_queue_ctx 656 struct vb2_v4l2_buffer *v4l2_m2m_last_buf(struct v4l2_m2m_queue_ctx *q_ctx); 749 * @q_ctx: pointer to struct @v4l2_m2m_queue_ctx 751 struct vb2_v4l2_buffer *v4l2_m2m_buf_remove(struct v4l2_m2m_queue_ctx *q_ctx); 781 * @q_ctx: pointer to struct @v4l2_m2m_queue_ctx 784 void v4l2_m2m_buf_remove_by_buf(struct v4l2_m2m_queue_ctx *q_ctx, 814 v4l2_m2m_buf_remove_by_idx(struct v4l2_m2m_queue_ctx *q_ctx, unsigned int idx);
|
/kernel/linux/linux-6.6/include/media/ |
H A D | v4l2-mem2mem.h | 628 * @q_ctx: pointer to struct @v4l2_m2m_queue_ctx 630 struct vb2_v4l2_buffer *v4l2_m2m_next_buf(struct v4l2_m2m_queue_ctx *q_ctx); 659 * @q_ctx: pointer to struct @v4l2_m2m_queue_ctx 661 struct vb2_v4l2_buffer *v4l2_m2m_last_buf(struct v4l2_m2m_queue_ctx *q_ctx); 754 * @q_ctx: pointer to struct @v4l2_m2m_queue_ctx 756 struct vb2_v4l2_buffer *v4l2_m2m_buf_remove(struct v4l2_m2m_queue_ctx *q_ctx); 786 * @q_ctx: pointer to struct @v4l2_m2m_queue_ctx 789 void v4l2_m2m_buf_remove_by_buf(struct v4l2_m2m_queue_ctx *q_ctx, 819 v4l2_m2m_buf_remove_by_idx(struct v4l2_m2m_queue_ctx *q_ctx, unsigned int idx);
|
/kernel/linux/linux-6.6/drivers/net/ethernet/intel/ice/ |
H A D | ice_sched.c | 571 struct ice_q_ctx *q_ctx; in ice_alloc_lan_q_ctx() local 579 q_ctx = devm_kcalloc(ice_hw_to_dev(hw), new_numqs, in ice_alloc_lan_q_ctx() 580 sizeof(*q_ctx), GFP_KERNEL); in ice_alloc_lan_q_ctx() 581 if (!q_ctx) in ice_alloc_lan_q_ctx() 585 q_ctx[idx].q_handle = ICE_INVAL_Q_HANDLE; in ice_alloc_lan_q_ctx() 586 q_ctx[idx].q_teid = ICE_INVAL_TEID; in ice_alloc_lan_q_ctx() 589 vsi_ctx->lan_q_ctx[tc] = q_ctx; in ice_alloc_lan_q_ctx() 597 q_ctx = devm_kcalloc(ice_hw_to_dev(hw), new_numqs, in ice_alloc_lan_q_ctx() 598 sizeof(*q_ctx), GFP_KERNEL); in ice_alloc_lan_q_ctx() 599 if (!q_ctx) in ice_alloc_lan_q_ctx() 628 struct ice_q_ctx *q_ctx; ice_alloc_rdma_q_ctx() local 3796 ice_sched_save_q_bw(struct ice_q_ctx *q_ctx, enum ice_rl_type rl_type, u32 bw) ice_sched_save_q_bw() argument 3830 struct ice_q_ctx *q_ctx; ice_sched_set_q_bw_lmt() local 4345 ice_sched_replay_q_bw(struct ice_port_info *pi, struct ice_q_ctx *q_ctx) ice_sched_replay_q_bw() argument [all...] |
H A D | ice_common.c | 4592 struct ice_q_ctx *q_ctx; in ice_get_lan_q_ctx() local 4601 q_ctx = vsi->lan_q_ctx[tc]; in ice_get_lan_q_ctx() 4602 return &q_ctx[q_handle]; in ice_get_lan_q_ctx() 4625 struct ice_q_ctx *q_ctx; in ice_ena_vsi_txq() local 4642 q_ctx = ice_get_lan_q_ctx(hw, vsi_handle, tc, q_handle); in ice_ena_vsi_txq() 4643 if (!q_ctx) { in ice_ena_vsi_txq() 4695 q_ctx->q_handle = q_handle; in ice_ena_vsi_txq() 4696 q_ctx->q_teid = le32_to_cpu(node.node_teid); in ice_ena_vsi_txq() 4701 status = ice_sched_replay_q_bw(pi, q_ctx); in ice_ena_vsi_txq() 4730 struct ice_q_ctx *q_ctx; in ice_dis_vsi_txq() local [all...] |
H A D | ice_lag.c | 318 struct ice_q_ctx *q_ctx; in ice_lag_qbuf_recfg() local 325 q_ctx = ice_get_lan_q_ctx(hw, vsi_num, tc, i); in ice_lag_qbuf_recfg() 326 if (!q_ctx) { in ice_lag_qbuf_recfg() 331 if (q_ctx->q_teid == ICE_INVAL_TEID) { in ice_lag_qbuf_recfg() 336 if (q_ctx->q_handle == ICE_INVAL_Q_HANDLE) { in ice_lag_qbuf_recfg() 342 qid = pf->vsi[vsi_num]->txq_map[q_ctx->q_handle]; in ice_lag_qbuf_recfg() 345 qbuf->queue_info[count].q_teid = cpu_to_le32(q_ctx->q_teid); in ice_lag_qbuf_recfg()
|
H A D | ice_sched.h | 169 int ice_sched_replay_q_bw(struct ice_port_info *pi, struct ice_q_ctx *q_ctx);
|
/kernel/linux/linux-5.10/drivers/net/ethernet/intel/ice/ |
H A D | ice_sched.c | 545 struct ice_q_ctx *q_ctx; in ice_alloc_lan_q_ctx() local 554 sizeof(*q_ctx), in ice_alloc_lan_q_ctx() 565 q_ctx = devm_kcalloc(ice_hw_to_dev(hw), new_numqs, in ice_alloc_lan_q_ctx() 566 sizeof(*q_ctx), GFP_KERNEL); in ice_alloc_lan_q_ctx() 567 if (!q_ctx) in ice_alloc_lan_q_ctx() 569 memcpy(q_ctx, vsi_ctx->lan_q_ctx[tc], in ice_alloc_lan_q_ctx() 570 prev_num * sizeof(*q_ctx)); in ice_alloc_lan_q_ctx() 572 vsi_ctx->lan_q_ctx[tc] = q_ctx; in ice_alloc_lan_q_ctx() 2698 * @q_ctx: queue context structure 2705 ice_sched_save_q_bw(struct ice_q_ctx *q_ctx, enu argument 2740 struct ice_q_ctx *q_ctx; ice_sched_set_q_bw_lmt() local 2959 ice_sched_replay_q_bw(struct ice_port_info *pi, struct ice_q_ctx *q_ctx) ice_sched_replay_q_bw() argument [all...] |
H A D | ice_common.c | 3836 struct ice_q_ctx *q_ctx; in ice_get_lan_q_ctx() local 3845 q_ctx = vsi->lan_q_ctx[tc]; in ice_get_lan_q_ctx() 3846 return &q_ctx[q_handle]; in ice_get_lan_q_ctx() 3869 struct ice_q_ctx *q_ctx; in ice_ena_vsi_txq() local 3886 q_ctx = ice_get_lan_q_ctx(hw, vsi_handle, tc, q_handle); in ice_ena_vsi_txq() 3887 if (!q_ctx) { in ice_ena_vsi_txq() 3939 q_ctx->q_handle = q_handle; in ice_ena_vsi_txq() 3940 q_ctx->q_teid = le32_to_cpu(node.node_teid); in ice_ena_vsi_txq() 3945 status = ice_sched_replay_q_bw(pi, q_ctx); in ice_ena_vsi_txq() 3975 struct ice_q_ctx *q_ctx; in ice_dis_vsi_txq() local [all...] |
H A D | ice_sched.h | 89 ice_sched_replay_q_bw(struct ice_port_info *pi, struct ice_q_ctx *q_ctx);
|