Home
last modified time | relevance | path

Searched refs:q_ctx (Results 1 - 11 of 11) sorted by relevance

/kernel/linux/linux-5.10/drivers/media/v4l2-core/
H A Dv4l2-mem2mem.c126 struct v4l2_m2m_queue_ctx *q_ctx; in v4l2_m2m_get_vq() local
128 q_ctx = get_queue_ctx(m2m_ctx, type); in v4l2_m2m_get_vq()
129 if (!q_ctx) in v4l2_m2m_get_vq()
132 return &q_ctx->q; in v4l2_m2m_get_vq()
136 struct vb2_v4l2_buffer *v4l2_m2m_next_buf(struct v4l2_m2m_queue_ctx *q_ctx) in v4l2_m2m_next_buf() argument
141 spin_lock_irqsave(&q_ctx->rdy_spinlock, flags); in v4l2_m2m_next_buf()
143 if (list_empty(&q_ctx->rdy_queue)) { in v4l2_m2m_next_buf()
144 spin_unlock_irqrestore(&q_ctx->rdy_spinlock, flags); in v4l2_m2m_next_buf()
148 b = list_first_entry(&q_ctx->rdy_queue, struct v4l2_m2m_buffer, list); in v4l2_m2m_next_buf()
149 spin_unlock_irqrestore(&q_ctx in v4l2_m2m_next_buf()
154 v4l2_m2m_last_buf(struct v4l2_m2m_queue_ctx *q_ctx) v4l2_m2m_last_buf() argument
172 v4l2_m2m_buf_remove(struct v4l2_m2m_queue_ctx *q_ctx) v4l2_m2m_buf_remove() argument
191 v4l2_m2m_buf_remove_by_buf(struct v4l2_m2m_queue_ctx *q_ctx, struct vb2_v4l2_buffer *vbuf) v4l2_m2m_buf_remove_by_buf() argument
206 v4l2_m2m_buf_remove_by_idx(struct v4l2_m2m_queue_ctx *q_ctx, unsigned int idx) v4l2_m2m_buf_remove_by_idx() argument
872 struct v4l2_m2m_queue_ctx *q_ctx; v4l2_m2m_streamoff() local
1254 struct v4l2_m2m_queue_ctx *q_ctx; v4l2_m2m_buf_queue() local
[all...]
/kernel/linux/linux-6.6/drivers/media/v4l2-core/
H A Dv4l2-mem2mem.c126 struct v4l2_m2m_queue_ctx *q_ctx; in v4l2_m2m_get_vq() local
128 q_ctx = get_queue_ctx(m2m_ctx, type); in v4l2_m2m_get_vq()
129 if (!q_ctx) in v4l2_m2m_get_vq()
132 return &q_ctx->q; in v4l2_m2m_get_vq()
136 struct vb2_v4l2_buffer *v4l2_m2m_next_buf(struct v4l2_m2m_queue_ctx *q_ctx) in v4l2_m2m_next_buf() argument
141 spin_lock_irqsave(&q_ctx->rdy_spinlock, flags); in v4l2_m2m_next_buf()
143 if (list_empty(&q_ctx->rdy_queue)) { in v4l2_m2m_next_buf()
144 spin_unlock_irqrestore(&q_ctx->rdy_spinlock, flags); in v4l2_m2m_next_buf()
148 b = list_first_entry(&q_ctx->rdy_queue, struct v4l2_m2m_buffer, list); in v4l2_m2m_next_buf()
149 spin_unlock_irqrestore(&q_ctx in v4l2_m2m_next_buf()
154 v4l2_m2m_last_buf(struct v4l2_m2m_queue_ctx *q_ctx) v4l2_m2m_last_buf() argument
172 v4l2_m2m_buf_remove(struct v4l2_m2m_queue_ctx *q_ctx) v4l2_m2m_buf_remove() argument
191 v4l2_m2m_buf_remove_by_buf(struct v4l2_m2m_queue_ctx *q_ctx, struct vb2_v4l2_buffer *vbuf) v4l2_m2m_buf_remove_by_buf() argument
206 v4l2_m2m_buf_remove_by_idx(struct v4l2_m2m_queue_ctx *q_ctx, unsigned int idx) v4l2_m2m_buf_remove_by_idx() argument
873 struct v4l2_m2m_queue_ctx *q_ctx; v4l2_m2m_streamoff() local
1285 struct v4l2_m2m_queue_ctx *q_ctx; v4l2_m2m_buf_queue() local
[all...]
/kernel/linux/linux-5.10/include/media/
H A Dv4l2-mem2mem.h623 * @q_ctx: pointer to struct @v4l2_m2m_queue_ctx
625 struct vb2_v4l2_buffer *v4l2_m2m_next_buf(struct v4l2_m2m_queue_ctx *q_ctx);
654 * @q_ctx: pointer to struct @v4l2_m2m_queue_ctx
656 struct vb2_v4l2_buffer *v4l2_m2m_last_buf(struct v4l2_m2m_queue_ctx *q_ctx);
749 * @q_ctx: pointer to struct @v4l2_m2m_queue_ctx
751 struct vb2_v4l2_buffer *v4l2_m2m_buf_remove(struct v4l2_m2m_queue_ctx *q_ctx);
781 * @q_ctx: pointer to struct @v4l2_m2m_queue_ctx
784 void v4l2_m2m_buf_remove_by_buf(struct v4l2_m2m_queue_ctx *q_ctx,
814 v4l2_m2m_buf_remove_by_idx(struct v4l2_m2m_queue_ctx *q_ctx, unsigned int idx);
/kernel/linux/linux-6.6/include/media/
H A Dv4l2-mem2mem.h628 * @q_ctx: pointer to struct @v4l2_m2m_queue_ctx
630 struct vb2_v4l2_buffer *v4l2_m2m_next_buf(struct v4l2_m2m_queue_ctx *q_ctx);
659 * @q_ctx: pointer to struct @v4l2_m2m_queue_ctx
661 struct vb2_v4l2_buffer *v4l2_m2m_last_buf(struct v4l2_m2m_queue_ctx *q_ctx);
754 * @q_ctx: pointer to struct @v4l2_m2m_queue_ctx
756 struct vb2_v4l2_buffer *v4l2_m2m_buf_remove(struct v4l2_m2m_queue_ctx *q_ctx);
786 * @q_ctx: pointer to struct @v4l2_m2m_queue_ctx
789 void v4l2_m2m_buf_remove_by_buf(struct v4l2_m2m_queue_ctx *q_ctx,
819 v4l2_m2m_buf_remove_by_idx(struct v4l2_m2m_queue_ctx *q_ctx, unsigned int idx);
/kernel/linux/linux-6.6/drivers/net/ethernet/intel/ice/
H A Dice_sched.c571 struct ice_q_ctx *q_ctx; in ice_alloc_lan_q_ctx() local
579 q_ctx = devm_kcalloc(ice_hw_to_dev(hw), new_numqs, in ice_alloc_lan_q_ctx()
580 sizeof(*q_ctx), GFP_KERNEL); in ice_alloc_lan_q_ctx()
581 if (!q_ctx) in ice_alloc_lan_q_ctx()
585 q_ctx[idx].q_handle = ICE_INVAL_Q_HANDLE; in ice_alloc_lan_q_ctx()
586 q_ctx[idx].q_teid = ICE_INVAL_TEID; in ice_alloc_lan_q_ctx()
589 vsi_ctx->lan_q_ctx[tc] = q_ctx; in ice_alloc_lan_q_ctx()
597 q_ctx = devm_kcalloc(ice_hw_to_dev(hw), new_numqs, in ice_alloc_lan_q_ctx()
598 sizeof(*q_ctx), GFP_KERNEL); in ice_alloc_lan_q_ctx()
599 if (!q_ctx) in ice_alloc_lan_q_ctx()
628 struct ice_q_ctx *q_ctx; ice_alloc_rdma_q_ctx() local
3796 ice_sched_save_q_bw(struct ice_q_ctx *q_ctx, enum ice_rl_type rl_type, u32 bw) ice_sched_save_q_bw() argument
3830 struct ice_q_ctx *q_ctx; ice_sched_set_q_bw_lmt() local
4345 ice_sched_replay_q_bw(struct ice_port_info *pi, struct ice_q_ctx *q_ctx) ice_sched_replay_q_bw() argument
[all...]
H A Dice_common.c4592 struct ice_q_ctx *q_ctx; in ice_get_lan_q_ctx() local
4601 q_ctx = vsi->lan_q_ctx[tc]; in ice_get_lan_q_ctx()
4602 return &q_ctx[q_handle]; in ice_get_lan_q_ctx()
4625 struct ice_q_ctx *q_ctx; in ice_ena_vsi_txq() local
4642 q_ctx = ice_get_lan_q_ctx(hw, vsi_handle, tc, q_handle); in ice_ena_vsi_txq()
4643 if (!q_ctx) { in ice_ena_vsi_txq()
4695 q_ctx->q_handle = q_handle; in ice_ena_vsi_txq()
4696 q_ctx->q_teid = le32_to_cpu(node.node_teid); in ice_ena_vsi_txq()
4701 status = ice_sched_replay_q_bw(pi, q_ctx); in ice_ena_vsi_txq()
4730 struct ice_q_ctx *q_ctx; in ice_dis_vsi_txq() local
[all...]
H A Dice_lag.c318 struct ice_q_ctx *q_ctx; in ice_lag_qbuf_recfg() local
325 q_ctx = ice_get_lan_q_ctx(hw, vsi_num, tc, i); in ice_lag_qbuf_recfg()
326 if (!q_ctx) { in ice_lag_qbuf_recfg()
331 if (q_ctx->q_teid == ICE_INVAL_TEID) { in ice_lag_qbuf_recfg()
336 if (q_ctx->q_handle == ICE_INVAL_Q_HANDLE) { in ice_lag_qbuf_recfg()
342 qid = pf->vsi[vsi_num]->txq_map[q_ctx->q_handle]; in ice_lag_qbuf_recfg()
345 qbuf->queue_info[count].q_teid = cpu_to_le32(q_ctx->q_teid); in ice_lag_qbuf_recfg()
H A Dice_sched.h169 int ice_sched_replay_q_bw(struct ice_port_info *pi, struct ice_q_ctx *q_ctx);
/kernel/linux/linux-5.10/drivers/net/ethernet/intel/ice/
H A Dice_sched.c545 struct ice_q_ctx *q_ctx; in ice_alloc_lan_q_ctx() local
554 sizeof(*q_ctx), in ice_alloc_lan_q_ctx()
565 q_ctx = devm_kcalloc(ice_hw_to_dev(hw), new_numqs, in ice_alloc_lan_q_ctx()
566 sizeof(*q_ctx), GFP_KERNEL); in ice_alloc_lan_q_ctx()
567 if (!q_ctx) in ice_alloc_lan_q_ctx()
569 memcpy(q_ctx, vsi_ctx->lan_q_ctx[tc], in ice_alloc_lan_q_ctx()
570 prev_num * sizeof(*q_ctx)); in ice_alloc_lan_q_ctx()
572 vsi_ctx->lan_q_ctx[tc] = q_ctx; in ice_alloc_lan_q_ctx()
2698 * @q_ctx: queue context structure
2705 ice_sched_save_q_bw(struct ice_q_ctx *q_ctx, enu argument
2740 struct ice_q_ctx *q_ctx; ice_sched_set_q_bw_lmt() local
2959 ice_sched_replay_q_bw(struct ice_port_info *pi, struct ice_q_ctx *q_ctx) ice_sched_replay_q_bw() argument
[all...]
H A Dice_common.c3836 struct ice_q_ctx *q_ctx; in ice_get_lan_q_ctx() local
3845 q_ctx = vsi->lan_q_ctx[tc]; in ice_get_lan_q_ctx()
3846 return &q_ctx[q_handle]; in ice_get_lan_q_ctx()
3869 struct ice_q_ctx *q_ctx; in ice_ena_vsi_txq() local
3886 q_ctx = ice_get_lan_q_ctx(hw, vsi_handle, tc, q_handle); in ice_ena_vsi_txq()
3887 if (!q_ctx) { in ice_ena_vsi_txq()
3939 q_ctx->q_handle = q_handle; in ice_ena_vsi_txq()
3940 q_ctx->q_teid = le32_to_cpu(node.node_teid); in ice_ena_vsi_txq()
3945 status = ice_sched_replay_q_bw(pi, q_ctx); in ice_ena_vsi_txq()
3975 struct ice_q_ctx *q_ctx; in ice_dis_vsi_txq() local
[all...]
H A Dice_sched.h89 ice_sched_replay_q_bw(struct ice_port_info *pi, struct ice_q_ctx *q_ctx);

Completed in 26 milliseconds