Home
last modified time | relevance | path

Searched refs:fq (Results 1 - 25 of 107) sorted by relevance

12345

/kernel/linux/linux-5.10/include/net/
H A Dfq_impl.h10 #include <net/fq.h>
14 static void fq_adjust_removal(struct fq *fq, in fq_adjust_removal() argument
23 fq->backlog--; in fq_adjust_removal()
24 fq->memory_usage -= skb->truesize; in fq_adjust_removal()
27 static void fq_rejigger_backlog(struct fq *fq, struct fq_flow *flow) in fq_rejigger_backlog() argument
36 list_for_each_entry_continue(i, &fq->backlogs, backlogchain) in fq_rejigger_backlog()
45 static struct sk_buff *fq_flow_dequeue(struct fq *fq, in fq_flow_dequeue() argument
62 fq_tin_dequeue(struct fq *fq, struct fq_tin *tin, fq_tin_dequeue_t dequeue_func) fq_tin_dequeue() argument
109 fq_flow_idx(struct fq *fq, struct sk_buff *skb) fq_flow_idx() argument
116 fq_flow_classify(struct fq *fq, struct fq_tin *tin, u32 idx, struct sk_buff *skb, fq_flow_get_default_t get_default_func) fq_flow_classify() argument
138 fq_recalc_backlog(struct fq *fq, struct fq_tin *tin, struct fq_flow *flow) fq_recalc_backlog() argument
156 fq_tin_enqueue(struct fq *fq, struct fq_tin *tin, u32 idx, struct sk_buff *skb, fq_skb_free_t free_func, fq_flow_get_default_t get_default_func) fq_tin_enqueue() argument
208 fq_flow_filter(struct fq *fq, struct fq_flow *flow, fq_skb_filter_t filter_func, void *filter_data, fq_skb_free_t free_func) fq_flow_filter() argument
231 fq_tin_filter(struct fq *fq, struct fq_tin *tin, fq_skb_filter_t filter_func, void *filter_data, fq_skb_free_t free_func) fq_tin_filter() argument
247 fq_flow_reset(struct fq *fq, struct fq_flow *flow, fq_skb_free_t free_func) fq_flow_reset() argument
267 fq_tin_reset(struct fq *fq, struct fq_tin *tin, fq_skb_free_t free_func) fq_tin_reset() argument
303 fq_init(struct fq *fq, int flows_cnt) fq_init() argument
325 fq_reset(struct fq *fq, fq_skb_free_t free_func) fq_reset() argument
[all...]
H A Dipv6_frag.h33 struct frag_queue *fq = container_of(q, struct frag_queue, q); in ip6frag_init() local
37 fq->ecn = 0; in ip6frag_init()
48 const struct inet_frag_queue *fq = data; in ip6frag_obj_hashfn() local
50 return jhash2((const u32 *)&fq->key.v6, in ip6frag_obj_hashfn()
58 const struct inet_frag_queue *fq = ptr; in ip6frag_obj_cmpfn() local
60 return !!memcmp(&fq->key, key, sizeof(*key)); in ip6frag_obj_cmpfn()
64 ip6frag_expire_frag_queue(struct net *net, struct frag_queue *fq) in ip6frag_expire_frag_queue() argument
71 if (READ_ONCE(fq->q.fqdir->dead)) in ip6frag_expire_frag_queue()
73 spin_lock(&fq->q.lock); in ip6frag_expire_frag_queue()
75 if (fq in ip6frag_expire_frag_queue()
[all...]
/kernel/linux/linux-6.6/include/net/
H A Dfq_impl.h10 #include <net/fq.h>
16 __fq_adjust_removal(struct fq *fq, struct fq_flow *flow, unsigned int packets, in __fq_adjust_removal() argument
25 fq->backlog -= packets; in __fq_adjust_removal()
26 fq->memory_usage -= truesize; in __fq_adjust_removal()
36 idx = flow - fq->flows; in __fq_adjust_removal()
37 __clear_bit(idx, fq->flows_bitmap); in __fq_adjust_removal()
40 static void fq_adjust_removal(struct fq *fq, in fq_adjust_removal() argument
44 __fq_adjust_removal(fq, flo in fq_adjust_removal()
47 fq_flow_dequeue(struct fq *fq, struct fq_flow *flow) fq_flow_dequeue() argument
63 fq_flow_drop(struct fq *fq, struct fq_flow *flow, fq_skb_free_t free_func) fq_flow_drop() argument
90 fq_tin_dequeue(struct fq *fq, struct fq_tin *tin, fq_tin_dequeue_t dequeue_func) fq_tin_dequeue() argument
137 fq_flow_idx(struct fq *fq, struct sk_buff *skb) fq_flow_idx() argument
144 fq_flow_classify(struct fq *fq, struct fq_tin *tin, u32 idx, struct sk_buff *skb) fq_flow_classify() argument
165 fq_find_fattest_flow(struct fq *fq) fq_find_fattest_flow() argument
197 fq_tin_enqueue(struct fq *fq, struct fq_tin *tin, u32 idx, struct sk_buff *skb, fq_skb_free_t free_func) fq_tin_enqueue() argument
252 fq_flow_filter(struct fq *fq, struct fq_flow *flow, fq_skb_filter_t filter_func, void *filter_data, fq_skb_free_t free_func) fq_flow_filter() argument
273 fq_tin_filter(struct fq *fq, struct fq_tin *tin, fq_skb_filter_t filter_func, void *filter_data, fq_skb_free_t free_func) fq_tin_filter() argument
289 fq_flow_reset(struct fq *fq, struct fq_flow *flow, fq_skb_free_t free_func) fq_flow_reset() argument
311 fq_tin_reset(struct fq *fq, struct fq_tin *tin, fq_skb_free_t free_func) fq_tin_reset() argument
349 fq_init(struct fq *fq, int flows_cnt) fq_init() argument
378 fq_reset(struct fq *fq, fq_skb_free_t free_func) fq_reset() argument
[all...]
H A Dipv6_frag.h34 struct frag_queue *fq = container_of(q, struct frag_queue, q); in ip6frag_init() local
38 fq->ecn = 0; in ip6frag_init()
49 const struct inet_frag_queue *fq = data; in ip6frag_obj_hashfn() local
51 return jhash2((const u32 *)&fq->key.v6, in ip6frag_obj_hashfn()
59 const struct inet_frag_queue *fq = ptr; in ip6frag_obj_cmpfn() local
61 return !!memcmp(&fq->key, key, sizeof(*key)); in ip6frag_obj_cmpfn()
65 ip6frag_expire_frag_queue(struct net *net, struct frag_queue *fq) in ip6frag_expire_frag_queue() argument
72 if (READ_ONCE(fq->q.fqdir->dead)) in ip6frag_expire_frag_queue()
74 spin_lock(&fq->q.lock); in ip6frag_expire_frag_queue()
76 if (fq in ip6frag_expire_frag_queue()
[all...]
/kernel/linux/linux-6.6/block/
H A Dblk-flush.c30 * fq->flush_queue[fq->flush_pending_idx]. Once certain criteria are met, a
95 struct blk_flush_queue *fq, blk_opf_t flags);
100 return blk_mq_map_queue(q, REQ_OP_FLUSH, ctx)->fq; in blk_get_flush_queue()
153 * @fq: flush queue
161 * spin_lock_irq(fq->mq_flush_lock)
164 struct blk_flush_queue *fq, in blk_flush_complete_seq()
168 struct list_head *pending = &fq->flush_queue[fq->flush_pending_idx]; in blk_flush_complete_seq()
185 fq in blk_flush_complete_seq()
163 blk_flush_complete_seq(struct request *rq, struct blk_flush_queue *fq, unsigned int seq, blk_status_t error) blk_flush_complete_seq() argument
223 struct blk_flush_queue *fq = blk_get_flush_queue(q, flush_rq->mq_ctx); flush_end_io() local
289 blk_kick_flush(struct request_queue *q, struct blk_flush_queue *fq, blk_opf_t flags) blk_kick_flush() argument
365 struct blk_flush_queue *fq = blk_get_flush_queue(q, ctx); mq_flush_data_end_io() local
408 struct blk_flush_queue *fq = blk_get_flush_queue(q, rq->mq_ctx); blk_insert_flush() local
488 struct blk_flush_queue *fq; blk_alloc_flush_queue() local
513 blk_free_flush_queue(struct blk_flush_queue *fq) blk_free_flush_queue() argument
[all...]
/kernel/linux/linux-5.10/block/
H A Dblk-flush.c30 * fq->flush_queue[fq->flush_pending_idx]. Once certain criteria are met, a
96 struct blk_flush_queue *fq, unsigned int flags);
153 * @fq: flush queue
161 * spin_lock_irq(fq->mq_flush_lock)
164 struct blk_flush_queue *fq, in blk_flush_complete_seq()
168 struct list_head *pending = &fq->flush_queue[fq->flush_pending_idx]; in blk_flush_complete_seq()
185 fq->flush_pending_since = jiffies; in blk_flush_complete_seq()
190 list_move_tail(&rq->flush.list, &fq in blk_flush_complete_seq()
163 blk_flush_complete_seq(struct request *rq, struct blk_flush_queue *fq, unsigned int seq, blk_status_t error) blk_flush_complete_seq() argument
220 struct blk_flush_queue *fq = blk_get_flush_queue(q, flush_rq->mq_ctx); flush_end_io() local
285 blk_kick_flush(struct request_queue *q, struct blk_flush_queue *fq, unsigned int flags) blk_kick_flush() argument
357 struct blk_flush_queue *fq = blk_get_flush_queue(q, ctx); mq_flush_data_end_io() local
389 struct blk_flush_queue *fq = blk_get_flush_queue(q, rq->mq_ctx); blk_insert_flush() local
472 struct blk_flush_queue *fq; blk_alloc_flush_queue() local
498 blk_free_flush_queue(struct blk_flush_queue *fq) blk_free_flush_queue() argument
[all...]
/kernel/linux/linux-6.6/net/ipv6/netfilter/
H A Dnf_conntrack_reasm.c126 static int nf_ct_frag6_reasm(struct frag_queue *fq, struct sk_buff *skb,
137 struct frag_queue *fq; in nf_ct_frag6_expire() local
139 fq = container_of(frag, struct frag_queue, q); in nf_ct_frag6_expire()
141 ip6frag_expire_frag_queue(fq->q.fqdir->net, fq); in nf_ct_frag6_expire()
166 static int nf_ct_frag6_queue(struct frag_queue *fq, struct sk_buff *skb, in nf_ct_frag6_queue() argument
175 if (fq->q.flags & INET_FRAG_COMPLETE) { in nf_ct_frag6_queue()
205 if (end < fq->q.len || in nf_ct_frag6_queue()
206 ((fq->q.flags & INET_FRAG_LAST_IN) && end != fq in nf_ct_frag6_queue()
313 nf_ct_frag6_reasm(struct frag_queue *fq, struct sk_buff *skb, struct sk_buff *prev_tail, struct net_device *dev) nf_ct_frag6_reasm() argument
444 struct frag_queue *fq; nf_ct_frag6_gather() local
[all...]
/kernel/linux/linux-5.10/net/ipv6/netfilter/
H A Dnf_conntrack_reasm.c130 static int nf_ct_frag6_reasm(struct frag_queue *fq, struct sk_buff *skb,
141 struct frag_queue *fq; in nf_ct_frag6_expire() local
143 fq = container_of(frag, struct frag_queue, q); in nf_ct_frag6_expire()
145 ip6frag_expire_frag_queue(fq->q.fqdir->net, fq); in nf_ct_frag6_expire()
169 static int nf_ct_frag6_queue(struct frag_queue *fq, struct sk_buff *skb, in nf_ct_frag6_queue() argument
178 if (fq->q.flags & INET_FRAG_COMPLETE) { in nf_ct_frag6_queue()
208 if (end < fq->q.len || in nf_ct_frag6_queue()
209 ((fq->q.flags & INET_FRAG_LAST_IN) && end != fq in nf_ct_frag6_queue()
315 nf_ct_frag6_reasm(struct frag_queue *fq, struct sk_buff *skb, struct sk_buff *prev_tail, struct net_device *dev) nf_ct_frag6_reasm() argument
446 struct frag_queue *fq; nf_ct_frag6_gather() local
[all...]
/kernel/linux/linux-6.6/net/ieee802154/6lowpan/
H A Dreassembly.c33 static int lowpan_frag_reasm(struct lowpan_frag_queue *fq, struct sk_buff *skb,
47 struct frag_queue *fq; in lowpan_frag_expire() local
49 fq = container_of(frag, struct frag_queue, q); in lowpan_frag_expire()
51 spin_lock(&fq->q.lock); in lowpan_frag_expire()
53 if (fq->q.flags & INET_FRAG_COMPLETE) in lowpan_frag_expire()
56 inet_frag_kill(&fq->q); in lowpan_frag_expire()
58 spin_unlock(&fq->q.lock); in lowpan_frag_expire()
59 inet_frag_put(&fq->q); in lowpan_frag_expire()
84 static int lowpan_frag_queue(struct lowpan_frag_queue *fq, in lowpan_frag_queue() argument
97 if (fq in lowpan_frag_queue()
164 lowpan_frag_reasm(struct lowpan_frag_queue *fq, struct sk_buff *skb, struct sk_buff *prev_tail, struct net_device *ldev) lowpan_frag_reasm() argument
279 struct lowpan_frag_queue *fq; lowpan_frag_rcv() local
498 const struct inet_frag_queue *fq = data; lowpan_obj_hashfn() local
507 const struct inet_frag_queue *fq = ptr; lowpan_obj_cmpfn() local
[all...]
/kernel/linux/linux-5.10/net/ieee802154/6lowpan/
H A Dreassembly.c33 static int lowpan_frag_reasm(struct lowpan_frag_queue *fq, struct sk_buff *skb,
47 struct frag_queue *fq; in lowpan_frag_expire() local
49 fq = container_of(frag, struct frag_queue, q); in lowpan_frag_expire()
51 spin_lock(&fq->q.lock); in lowpan_frag_expire()
53 if (fq->q.flags & INET_FRAG_COMPLETE) in lowpan_frag_expire()
56 inet_frag_kill(&fq->q); in lowpan_frag_expire()
58 spin_unlock(&fq->q.lock); in lowpan_frag_expire()
59 inet_frag_put(&fq->q); in lowpan_frag_expire()
84 static int lowpan_frag_queue(struct lowpan_frag_queue *fq, in lowpan_frag_queue() argument
97 if (fq in lowpan_frag_queue()
163 lowpan_frag_reasm(struct lowpan_frag_queue *fq, struct sk_buff *skb, struct sk_buff *prev_tail, struct net_device *ldev) lowpan_frag_reasm() argument
278 struct lowpan_frag_queue *fq; lowpan_frag_rcv() local
493 const struct inet_frag_queue *fq = data; lowpan_obj_hashfn() local
502 const struct inet_frag_queue *fq = ptr; lowpan_obj_cmpfn() local
[all...]
/kernel/linux/linux-6.6/net/ipv6/
H A Dreassembly.c70 static int ip6_frag_reasm(struct frag_queue *fq, struct sk_buff *skb,
76 struct frag_queue *fq; in ip6_frag_expire() local
78 fq = container_of(frag, struct frag_queue, q); in ip6_frag_expire()
80 ip6frag_expire_frag_queue(fq->q.fqdir->net, fq); in ip6_frag_expire()
106 static int ip6_frag_queue(struct frag_queue *fq, struct sk_buff *skb, in ip6_frag_queue() argument
119 if (fq->q.flags & INET_FRAG_COMPLETE) { in ip6_frag_queue()
151 if (end < fq->q.len || in ip6_frag_queue()
152 ((fq->q.flags & INET_FRAG_LAST_IN) && end != fq in ip6_frag_queue()
256 ip6_frag_reasm(struct frag_queue *fq, struct sk_buff *skb, struct sk_buff *prev_tail, struct net_device *dev) ip6_frag_reasm() argument
330 struct frag_queue *fq; ipv6_frag_rcv() local
[all...]
/kernel/linux/linux-5.10/net/ipv6/
H A Dreassembly.c70 static int ip6_frag_reasm(struct frag_queue *fq, struct sk_buff *skb,
76 struct frag_queue *fq; in ip6_frag_expire() local
78 fq = container_of(frag, struct frag_queue, q); in ip6_frag_expire()
80 ip6frag_expire_frag_queue(fq->q.fqdir->net, fq); in ip6_frag_expire()
106 static int ip6_frag_queue(struct frag_queue *fq, struct sk_buff *skb, in ip6_frag_queue() argument
117 if (fq->q.flags & INET_FRAG_COMPLETE) in ip6_frag_queue()
147 if (end < fq->q.len || in ip6_frag_queue()
148 ((fq->q.flags & INET_FRAG_LAST_IN) && end != fq in ip6_frag_queue()
250 ip6_frag_reasm(struct frag_queue *fq, struct sk_buff *skb, struct sk_buff *prev_tail, struct net_device *dev) ip6_frag_reasm() argument
324 struct frag_queue *fq; ipv6_frag_rcv() local
[all...]
/kernel/linux/linux-5.10/drivers/soc/fsl/qbman/
H A Dqman.c268 struct qm_mcc_fq fq; member
957 static inline void fq_set(struct qman_fq *fq, u32 mask) in fq_set() argument
959 fq->flags |= mask; in fq_set()
962 static inline void fq_clear(struct qman_fq *fq, u32 mask) in fq_clear() argument
964 fq->flags &= ~mask; in fq_clear()
967 static inline int fq_isset(struct qman_fq *fq, u32 mask) in fq_isset() argument
969 return fq->flags & mask; in fq_isset()
972 static inline int fq_isclear(struct qman_fq *fq, u32 mask) in fq_isclear() argument
974 return !(fq->flags & mask); in fq_isclear()
1114 pr_debug("Allocated fq looku in qman_alloc_fq_table()
1121 struct qman_fq *fq; idx_to_fq() local
1151 fq_to_tag(struct qman_fq *fq) fq_to_tag() argument
1425 fq_state_change(struct qman_portal *p, struct qman_fq *fq, const union qm_mr_entry *msg, u8 verb) fq_state_change() argument
1488 struct qman_fq *fq; qm_mr_process_task() local
1571 clear_vdqcr(struct qman_portal *p, struct qman_fq *fq) clear_vdqcr() argument
1608 struct qman_fq *fq; __poll_portal_fast() local
1795 qman_create_fq(u32 fqid, u32 flags, struct qman_fq *fq) qman_create_fq() argument
1825 qman_destroy_fq(struct qman_fq *fq) qman_destroy_fq() argument
1847 qman_fq_fqid(struct qman_fq *fq) qman_fq_fqid() argument
1853 qman_init_fq(struct qman_fq *fq, u32 flags, struct qm_mcc_initfq *opts) qman_init_fq() argument
1966 qman_schedule_fq(struct qman_fq *fq) qman_schedule_fq() argument
2007 qman_retire_fq(struct qman_fq *fq, u32 *flags) qman_retire_fq() argument
2089 qman_oos_fq(struct qman_fq *fq) qman_oos_fq() argument
2127 qman_query_fq(struct qman_fq *fq, struct qm_fqd *fqd) qman_query_fq() argument
2152 qman_query_fq_np(struct qman_fq *fq, struct qm_mcr_queryfq_np *np) qman_query_fq_np() argument
2223 set_p_vdqcr(struct qman_portal *p, struct qman_fq *fq, u32 vdqcr) set_p_vdqcr() argument
2243 set_vdqcr(struct qman_portal **p, struct qman_fq *fq, u32 vdqcr) set_vdqcr() argument
2253 wait_vdqcr_start(struct qman_portal **p, struct qman_fq *fq, u32 vdqcr, u32 flags) wait_vdqcr_start() argument
2266 qman_volatile_dequeue(struct qman_fq *fq, u32 flags, u32 vdqcr) qman_volatile_dequeue() argument
2312 qman_enqueue(struct qman_fq *fq, const struct qm_fd *fd) qman_enqueue() argument
2896 struct qman_fq fq = { qpool_cleanup() local
2954 struct qman_fq fq = { cgr_cleanup() local
[all...]
H A Dqman_test_api.c105 static int do_enqueues(struct qman_fq *fq) in do_enqueues() argument
111 if (qman_enqueue(fq, &fd)) { in do_enqueues()
125 struct qman_fq *fq = &fq_base; in qman_test_api() local
132 err = qman_create_fq(0, FQ_FLAGS, fq); in qman_test_api()
137 err = qman_init_fq(fq, QMAN_INITFQ_FLAG_LOCAL, NULL); in qman_test_api()
143 err = do_enqueues(fq); in qman_test_api()
148 err = qman_volatile_dequeue(fq, VDQCR_FLAGS, frmcnt); in qman_test_api()
153 err = do_enqueues(fq); in qman_test_api()
158 err = qman_volatile_dequeue(fq, VDQCR_FLAGS, frmcnt); in qman_test_api()
166 err = qman_volatile_dequeue(fq, VDQCR_FLAG in qman_test_api()
209 cb_dqrr(struct qman_portal *p, struct qman_fq *fq, const struct qm_dqrr_entry *dq) cb_dqrr() argument
225 cb_ern(struct qman_portal *p, struct qman_fq *fq, const union qm_mr_entry *msg) cb_ern() argument
232 cb_fqs(struct qman_portal *p, struct qman_fq *fq, const union qm_mr_entry *msg) cb_fqs() argument
[all...]
/kernel/linux/linux-6.6/drivers/soc/fsl/qbman/
H A Dqman.c268 struct qm_mcc_fq fq; member
957 static inline void fq_set(struct qman_fq *fq, u32 mask) in fq_set() argument
959 fq->flags |= mask; in fq_set()
962 static inline void fq_clear(struct qman_fq *fq, u32 mask) in fq_clear() argument
964 fq->flags &= ~mask; in fq_clear()
967 static inline int fq_isset(struct qman_fq *fq, u32 mask) in fq_isset() argument
969 return fq->flags & mask; in fq_isset()
972 static inline int fq_isclear(struct qman_fq *fq, u32 mask) in fq_isclear() argument
974 return !(fq->flags & mask); in fq_isclear()
1114 pr_debug("Allocated fq looku in qman_alloc_fq_table()
1121 struct qman_fq *fq; idx_to_fq() local
1151 fq_to_tag(struct qman_fq *fq) fq_to_tag() argument
1425 fq_state_change(struct qman_portal *p, struct qman_fq *fq, const union qm_mr_entry *msg, u8 verb) fq_state_change() argument
1488 struct qman_fq *fq; qm_mr_process_task() local
1571 clear_vdqcr(struct qman_portal *p, struct qman_fq *fq) clear_vdqcr() argument
1608 struct qman_fq *fq; __poll_portal_fast() local
1795 qman_create_fq(u32 fqid, u32 flags, struct qman_fq *fq) qman_create_fq() argument
1825 qman_destroy_fq(struct qman_fq *fq) qman_destroy_fq() argument
1847 qman_fq_fqid(struct qman_fq *fq) qman_fq_fqid() argument
1853 qman_init_fq(struct qman_fq *fq, u32 flags, struct qm_mcc_initfq *opts) qman_init_fq() argument
1966 qman_schedule_fq(struct qman_fq *fq) qman_schedule_fq() argument
2007 qman_retire_fq(struct qman_fq *fq, u32 *flags) qman_retire_fq() argument
2089 qman_oos_fq(struct qman_fq *fq) qman_oos_fq() argument
2127 qman_query_fq(struct qman_fq *fq, struct qm_fqd *fqd) qman_query_fq() argument
2152 qman_query_fq_np(struct qman_fq *fq, struct qm_mcr_queryfq_np *np) qman_query_fq_np() argument
2223 set_p_vdqcr(struct qman_portal *p, struct qman_fq *fq, u32 vdqcr) set_p_vdqcr() argument
2243 set_vdqcr(struct qman_portal **p, struct qman_fq *fq, u32 vdqcr) set_vdqcr() argument
2253 wait_vdqcr_start(struct qman_portal **p, struct qman_fq *fq, u32 vdqcr, u32 flags) wait_vdqcr_start() argument
2266 qman_volatile_dequeue(struct qman_fq *fq, u32 flags, u32 vdqcr) qman_volatile_dequeue() argument
2312 qman_enqueue(struct qman_fq *fq, const struct qm_fd *fd) qman_enqueue() argument
2949 struct qman_fq fq = { qpool_cleanup() local
3007 struct qman_fq fq = { cgr_cleanup() local
[all...]
H A Dqman_test_api.c106 static int do_enqueues(struct qman_fq *fq) in do_enqueues() argument
112 if (qman_enqueue(fq, &fd)) { in do_enqueues()
126 struct qman_fq *fq = &fq_base; in qman_test_api() local
133 err = qman_create_fq(0, FQ_FLAGS, fq); in qman_test_api()
138 err = qman_init_fq(fq, QMAN_INITFQ_FLAG_LOCAL, NULL); in qman_test_api()
144 err = do_enqueues(fq); in qman_test_api()
149 err = qman_volatile_dequeue(fq, VDQCR_FLAGS, frmcnt); in qman_test_api()
154 err = do_enqueues(fq); in qman_test_api()
159 err = qman_volatile_dequeue(fq, VDQCR_FLAGS, frmcnt); in qman_test_api()
167 err = qman_volatile_dequeue(fq, VDQCR_FLAG in qman_test_api()
210 cb_dqrr(struct qman_portal *p, struct qman_fq *fq, const struct qm_dqrr_entry *dq, bool sched_napi) cb_dqrr() argument
227 cb_ern(struct qman_portal *p, struct qman_fq *fq, const union qm_mr_entry *msg) cb_ern() argument
234 cb_fqs(struct qman_portal *p, struct qman_fq *fq, const union qm_mr_entry *msg) cb_fqs() argument
[all...]
/kernel/linux/linux-5.10/drivers/iommu/
H A Diova.c49 iovad->fq = NULL; in init_iova_domain()
59 return !!iovad->fq; in has_iova_flush_queue()
71 free_percpu(iovad->fq); in free_iova_flush_queue()
73 iovad->fq = NULL; in free_iova_flush_queue()
95 struct iova_fq *fq; in init_iova_flush_queue() local
97 fq = per_cpu_ptr(queue, cpu); in init_iova_flush_queue()
98 fq->head = 0; in init_iova_flush_queue()
99 fq->tail = 0; in init_iova_flush_queue()
101 spin_lock_init(&fq->lock); in init_iova_flush_queue()
106 iovad->fq in init_iova_flush_queue()
462 fq_full(struct iova_fq *fq) fq_full() argument
468 fq_ring_add(struct iova_fq *fq) fq_ring_add() argument
479 fq_ring_free(struct iova_domain *iovad, struct iova_fq *fq) fq_ring_free() argument
522 struct iova_fq *fq = per_cpu_ptr(iovad->fq, cpu); fq_destroy_all_entries() local
540 struct iova_fq *fq; fq_flush_timeout() local
553 struct iova_fq *fq = raw_cpu_ptr(iovad->fq); queue_iova() local
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/freescale/dpaa2/
H A Ddpaa2-eth-debugfs.c47 static char *fq_type_to_str(struct dpaa2_eth_fq *fq) in fq_type_to_str() argument
49 switch (fq->type) { in fq_type_to_str()
62 struct dpaa2_eth_fq *fq; in dpaa2_dbg_fqs_show() local
71 fq = &priv->fq[i]; in dpaa2_dbg_fqs_show()
72 err = dpaa2_io_query_fq_count(NULL, fq->fqid, &fcnt, &bcnt); in dpaa2_dbg_fqs_show()
77 if (!fq->stats.frames && !fcnt) in dpaa2_dbg_fqs_show()
81 fq->fqid, in dpaa2_dbg_fqs_show()
82 fq->target_cpu, in dpaa2_dbg_fqs_show()
83 fq in dpaa2_dbg_fqs_show()
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/freescale/dpaa/
H A Ddpaa_eth_trace.h58 struct qman_fq *fq,
62 TP_ARGS(netdev, fq, fd),
80 __entry->fqid = fq->fqid;
104 struct qman_fq *fq,
107 TP_ARGS(netdev, fq, fd)
114 struct qman_fq *fq,
117 TP_ARGS(netdev, fq, fd)
124 struct qman_fq *fq,
127 TP_ARGS(netdev, fq, fd)
H A Ddpaa_eth_sysfs.c59 struct dpaa_fq *fq; in dpaa_eth_show_fqids() local
66 list_for_each_entry_safe(fq, tmp, &priv->dpaa_fq_list, list) { in dpaa_eth_show_fqids()
67 switch (fq->fq_type) { in dpaa_eth_show_fqids()
93 if (prev && (abs(fq->fqid - prev->fqid) != 1 || in dpaa_eth_show_fqids()
104 if (prev && abs(fq->fqid - prev->fqid) == 1 && in dpaa_eth_show_fqids()
106 last_fqid = fq->fqid; in dpaa_eth_show_fqids()
108 first_fqid = fq->fqid; in dpaa_eth_show_fqids()
109 last_fqid = fq->fqid; in dpaa_eth_show_fqids()
112 prev = fq; in dpaa_eth_show_fqids()
/kernel/linux/linux-6.6/drivers/net/ethernet/freescale/dpaa/
H A Ddpaa_eth_trace.h32 struct qman_fq *fq,
36 TP_ARGS(netdev, fq, fd),
54 __entry->fqid = fq->fqid;
78 struct qman_fq *fq,
81 TP_ARGS(netdev, fq, fd)
88 struct qman_fq *fq,
91 TP_ARGS(netdev, fq, fd)
98 struct qman_fq *fq,
101 TP_ARGS(netdev, fq, fd)
H A Ddpaa_eth_sysfs.c33 struct dpaa_fq *fq; in dpaa_eth_show_fqids() local
40 list_for_each_entry_safe(fq, tmp, &priv->dpaa_fq_list, list) { in dpaa_eth_show_fqids()
41 switch (fq->fq_type) { in dpaa_eth_show_fqids()
67 if (prev && (abs(fq->fqid - prev->fqid) != 1 || in dpaa_eth_show_fqids()
78 if (prev && abs(fq->fqid - prev->fqid) == 1 && in dpaa_eth_show_fqids()
80 last_fqid = fq->fqid; in dpaa_eth_show_fqids()
82 first_fqid = fq->fqid; in dpaa_eth_show_fqids()
83 last_fqid = fq->fqid; in dpaa_eth_show_fqids()
86 prev = fq; in dpaa_eth_show_fqids()
/kernel/linux/linux-6.6/drivers/net/ethernet/freescale/dpaa2/
H A Ddpaa2-eth-debugfs.c47 static char *fq_type_to_str(struct dpaa2_eth_fq *fq) in fq_type_to_str() argument
49 switch (fq->type) { in fq_type_to_str()
62 struct dpaa2_eth_fq *fq; in dpaa2_dbg_fqs_show() local
71 fq = &priv->fq[i]; in dpaa2_dbg_fqs_show()
72 err = dpaa2_io_query_fq_count(NULL, fq->fqid, &fcnt, &bcnt); in dpaa2_dbg_fqs_show()
77 if (!fq->stats.frames && !fcnt) in dpaa2_dbg_fqs_show()
81 fq->fqid, in dpaa2_dbg_fqs_show()
82 fq->target_cpu, in dpaa2_dbg_fqs_show()
83 fq in dpaa2_dbg_fqs_show()
[all...]
/kernel/linux/linux-5.10/drivers/crypto/caam/
H A Dqi.c135 static void caam_fq_ern_cb(struct qman_portal *qm, struct qman_fq *fq, in caam_fq_ern_cb() argument
218 static int empty_retired_fq(struct device *qidev, struct qman_fq *fq) in empty_retired_fq() argument
222 ret = qman_volatile_dequeue(fq, QMAN_VOLATILE_FLAG_WAIT_INT | in empty_retired_fq()
227 dev_err(qidev, "Volatile dequeue fail for FQ: %u\n", fq->fqid); in empty_retired_fq()
236 } while (fq->flags & QMAN_FQ_STATE_NE); in empty_retired_fq()
241 static int kill_fq(struct device *qidev, struct qman_fq *fq) in kill_fq() argument
246 ret = qman_retire_fq(fq, &flags); in kill_fq()
260 } while (fq->state != qman_fq_state_retired); in kill_fq()
262 WARN_ON(fq->flags & QMAN_FQ_STATE_BLOCKOOS); in kill_fq()
263 WARN_ON(fq in kill_fq()
286 empty_caam_fq(struct qman_fq *fq, struct caam_drv_ctx *drv_ctx) empty_caam_fq() argument
617 struct qman_fq *fq; alloc_rsp_fq_cpu() local
[all...]
/kernel/linux/linux-6.6/drivers/crypto/caam/
H A Dqi.c142 static void caam_fq_ern_cb(struct qman_portal *qm, struct qman_fq *fq, in caam_fq_ern_cb() argument
225 static int empty_retired_fq(struct device *qidev, struct qman_fq *fq) in empty_retired_fq() argument
229 ret = qman_volatile_dequeue(fq, QMAN_VOLATILE_FLAG_WAIT_INT | in empty_retired_fq()
234 dev_err(qidev, "Volatile dequeue fail for FQ: %u\n", fq->fqid); in empty_retired_fq()
243 } while (fq->flags & QMAN_FQ_STATE_NE); in empty_retired_fq()
248 static int kill_fq(struct device *qidev, struct qman_fq *fq) in kill_fq() argument
253 ret = qman_retire_fq(fq, &flags); in kill_fq()
267 } while (fq->state != qman_fq_state_retired); in kill_fq()
269 WARN_ON(fq->flags & QMAN_FQ_STATE_BLOCKOOS); in kill_fq()
270 WARN_ON(fq in kill_fq()
293 empty_caam_fq(struct qman_fq *fq, struct caam_drv_ctx *drv_ctx) empty_caam_fq() argument
621 struct qman_fq *fq; alloc_rsp_fq_cpu() local
[all...]

Completed in 17 milliseconds

12345