/kernel/linux/linux-5.10/include/net/ |
H A D | fq_impl.h | 10 #include <net/fq.h> 14 static void fq_adjust_removal(struct fq *fq, in fq_adjust_removal() argument 23 fq->backlog--; in fq_adjust_removal() 24 fq->memory_usage -= skb->truesize; in fq_adjust_removal() 27 static void fq_rejigger_backlog(struct fq *fq, struct fq_flow *flow) in fq_rejigger_backlog() argument 36 list_for_each_entry_continue(i, &fq->backlogs, backlogchain) in fq_rejigger_backlog() 45 static struct sk_buff *fq_flow_dequeue(struct fq *fq, in fq_flow_dequeue() argument 62 fq_tin_dequeue(struct fq *fq, struct fq_tin *tin, fq_tin_dequeue_t dequeue_func) fq_tin_dequeue() argument 109 fq_flow_idx(struct fq *fq, struct sk_buff *skb) fq_flow_idx() argument 116 fq_flow_classify(struct fq *fq, struct fq_tin *tin, u32 idx, struct sk_buff *skb, fq_flow_get_default_t get_default_func) fq_flow_classify() argument 138 fq_recalc_backlog(struct fq *fq, struct fq_tin *tin, struct fq_flow *flow) fq_recalc_backlog() argument 156 fq_tin_enqueue(struct fq *fq, struct fq_tin *tin, u32 idx, struct sk_buff *skb, fq_skb_free_t free_func, fq_flow_get_default_t get_default_func) fq_tin_enqueue() argument 208 fq_flow_filter(struct fq *fq, struct fq_flow *flow, fq_skb_filter_t filter_func, void *filter_data, fq_skb_free_t free_func) fq_flow_filter() argument 231 fq_tin_filter(struct fq *fq, struct fq_tin *tin, fq_skb_filter_t filter_func, void *filter_data, fq_skb_free_t free_func) fq_tin_filter() argument 247 fq_flow_reset(struct fq *fq, struct fq_flow *flow, fq_skb_free_t free_func) fq_flow_reset() argument 267 fq_tin_reset(struct fq *fq, struct fq_tin *tin, fq_skb_free_t free_func) fq_tin_reset() argument 303 fq_init(struct fq *fq, int flows_cnt) fq_init() argument 325 fq_reset(struct fq *fq, fq_skb_free_t free_func) fq_reset() argument [all...] |
H A D | ipv6_frag.h | 33 struct frag_queue *fq = container_of(q, struct frag_queue, q); in ip6frag_init() local 37 fq->ecn = 0; in ip6frag_init() 48 const struct inet_frag_queue *fq = data; in ip6frag_obj_hashfn() local 50 return jhash2((const u32 *)&fq->key.v6, in ip6frag_obj_hashfn() 58 const struct inet_frag_queue *fq = ptr; in ip6frag_obj_cmpfn() local 60 return !!memcmp(&fq->key, key, sizeof(*key)); in ip6frag_obj_cmpfn() 64 ip6frag_expire_frag_queue(struct net *net, struct frag_queue *fq) in ip6frag_expire_frag_queue() argument 71 if (READ_ONCE(fq->q.fqdir->dead)) in ip6frag_expire_frag_queue() 73 spin_lock(&fq->q.lock); in ip6frag_expire_frag_queue() 75 if (fq in ip6frag_expire_frag_queue() [all...] |
/kernel/linux/linux-6.6/include/net/ |
H A D | fq_impl.h | 10 #include <net/fq.h> 16 __fq_adjust_removal(struct fq *fq, struct fq_flow *flow, unsigned int packets, in __fq_adjust_removal() argument 25 fq->backlog -= packets; in __fq_adjust_removal() 26 fq->memory_usage -= truesize; in __fq_adjust_removal() 36 idx = flow - fq->flows; in __fq_adjust_removal() 37 __clear_bit(idx, fq->flows_bitmap); in __fq_adjust_removal() 40 static void fq_adjust_removal(struct fq *fq, in fq_adjust_removal() argument 44 __fq_adjust_removal(fq, flo in fq_adjust_removal() 47 fq_flow_dequeue(struct fq *fq, struct fq_flow *flow) fq_flow_dequeue() argument 63 fq_flow_drop(struct fq *fq, struct fq_flow *flow, fq_skb_free_t free_func) fq_flow_drop() argument 90 fq_tin_dequeue(struct fq *fq, struct fq_tin *tin, fq_tin_dequeue_t dequeue_func) fq_tin_dequeue() argument 137 fq_flow_idx(struct fq *fq, struct sk_buff *skb) fq_flow_idx() argument 144 fq_flow_classify(struct fq *fq, struct fq_tin *tin, u32 idx, struct sk_buff *skb) fq_flow_classify() argument 165 fq_find_fattest_flow(struct fq *fq) fq_find_fattest_flow() argument 197 fq_tin_enqueue(struct fq *fq, struct fq_tin *tin, u32 idx, struct sk_buff *skb, fq_skb_free_t free_func) fq_tin_enqueue() argument 252 fq_flow_filter(struct fq *fq, struct fq_flow *flow, fq_skb_filter_t filter_func, void *filter_data, fq_skb_free_t free_func) fq_flow_filter() argument 273 fq_tin_filter(struct fq *fq, struct fq_tin *tin, fq_skb_filter_t filter_func, void *filter_data, fq_skb_free_t free_func) fq_tin_filter() argument 289 fq_flow_reset(struct fq *fq, struct fq_flow *flow, fq_skb_free_t free_func) fq_flow_reset() argument 311 fq_tin_reset(struct fq *fq, struct fq_tin *tin, fq_skb_free_t free_func) fq_tin_reset() argument 349 fq_init(struct fq *fq, int flows_cnt) fq_init() argument 378 fq_reset(struct fq *fq, fq_skb_free_t free_func) fq_reset() argument [all...] |
H A D | ipv6_frag.h | 34 struct frag_queue *fq = container_of(q, struct frag_queue, q); in ip6frag_init() local 38 fq->ecn = 0; in ip6frag_init() 49 const struct inet_frag_queue *fq = data; in ip6frag_obj_hashfn() local 51 return jhash2((const u32 *)&fq->key.v6, in ip6frag_obj_hashfn() 59 const struct inet_frag_queue *fq = ptr; in ip6frag_obj_cmpfn() local 61 return !!memcmp(&fq->key, key, sizeof(*key)); in ip6frag_obj_cmpfn() 65 ip6frag_expire_frag_queue(struct net *net, struct frag_queue *fq) in ip6frag_expire_frag_queue() argument 72 if (READ_ONCE(fq->q.fqdir->dead)) in ip6frag_expire_frag_queue() 74 spin_lock(&fq->q.lock); in ip6frag_expire_frag_queue() 76 if (fq in ip6frag_expire_frag_queue() [all...] |
/kernel/linux/linux-6.6/block/ |
H A D | blk-flush.c | 30 * fq->flush_queue[fq->flush_pending_idx]. Once certain criteria are met, a 95 struct blk_flush_queue *fq, blk_opf_t flags); 100 return blk_mq_map_queue(q, REQ_OP_FLUSH, ctx)->fq; in blk_get_flush_queue() 153 * @fq: flush queue 161 * spin_lock_irq(fq->mq_flush_lock) 164 struct blk_flush_queue *fq, in blk_flush_complete_seq() 168 struct list_head *pending = &fq->flush_queue[fq->flush_pending_idx]; in blk_flush_complete_seq() 185 fq in blk_flush_complete_seq() 163 blk_flush_complete_seq(struct request *rq, struct blk_flush_queue *fq, unsigned int seq, blk_status_t error) blk_flush_complete_seq() argument 223 struct blk_flush_queue *fq = blk_get_flush_queue(q, flush_rq->mq_ctx); flush_end_io() local 289 blk_kick_flush(struct request_queue *q, struct blk_flush_queue *fq, blk_opf_t flags) blk_kick_flush() argument 365 struct blk_flush_queue *fq = blk_get_flush_queue(q, ctx); mq_flush_data_end_io() local 408 struct blk_flush_queue *fq = blk_get_flush_queue(q, rq->mq_ctx); blk_insert_flush() local 488 struct blk_flush_queue *fq; blk_alloc_flush_queue() local 513 blk_free_flush_queue(struct blk_flush_queue *fq) blk_free_flush_queue() argument [all...] |
/kernel/linux/linux-5.10/block/ |
H A D | blk-flush.c | 30 * fq->flush_queue[fq->flush_pending_idx]. Once certain criteria are met, a 96 struct blk_flush_queue *fq, unsigned int flags); 153 * @fq: flush queue 161 * spin_lock_irq(fq->mq_flush_lock) 164 struct blk_flush_queue *fq, in blk_flush_complete_seq() 168 struct list_head *pending = &fq->flush_queue[fq->flush_pending_idx]; in blk_flush_complete_seq() 185 fq->flush_pending_since = jiffies; in blk_flush_complete_seq() 190 list_move_tail(&rq->flush.list, &fq in blk_flush_complete_seq() 163 blk_flush_complete_seq(struct request *rq, struct blk_flush_queue *fq, unsigned int seq, blk_status_t error) blk_flush_complete_seq() argument 220 struct blk_flush_queue *fq = blk_get_flush_queue(q, flush_rq->mq_ctx); flush_end_io() local 285 blk_kick_flush(struct request_queue *q, struct blk_flush_queue *fq, unsigned int flags) blk_kick_flush() argument 357 struct blk_flush_queue *fq = blk_get_flush_queue(q, ctx); mq_flush_data_end_io() local 389 struct blk_flush_queue *fq = blk_get_flush_queue(q, rq->mq_ctx); blk_insert_flush() local 472 struct blk_flush_queue *fq; blk_alloc_flush_queue() local 498 blk_free_flush_queue(struct blk_flush_queue *fq) blk_free_flush_queue() argument [all...] |
/kernel/linux/linux-6.6/net/ipv6/netfilter/ |
H A D | nf_conntrack_reasm.c | 126 static int nf_ct_frag6_reasm(struct frag_queue *fq, struct sk_buff *skb, 137 struct frag_queue *fq; in nf_ct_frag6_expire() local 139 fq = container_of(frag, struct frag_queue, q); in nf_ct_frag6_expire() 141 ip6frag_expire_frag_queue(fq->q.fqdir->net, fq); in nf_ct_frag6_expire() 166 static int nf_ct_frag6_queue(struct frag_queue *fq, struct sk_buff *skb, in nf_ct_frag6_queue() argument 175 if (fq->q.flags & INET_FRAG_COMPLETE) { in nf_ct_frag6_queue() 205 if (end < fq->q.len || in nf_ct_frag6_queue() 206 ((fq->q.flags & INET_FRAG_LAST_IN) && end != fq in nf_ct_frag6_queue() 313 nf_ct_frag6_reasm(struct frag_queue *fq, struct sk_buff *skb, struct sk_buff *prev_tail, struct net_device *dev) nf_ct_frag6_reasm() argument 444 struct frag_queue *fq; nf_ct_frag6_gather() local [all...] |
/kernel/linux/linux-5.10/net/ipv6/netfilter/ |
H A D | nf_conntrack_reasm.c | 130 static int nf_ct_frag6_reasm(struct frag_queue *fq, struct sk_buff *skb, 141 struct frag_queue *fq; in nf_ct_frag6_expire() local 143 fq = container_of(frag, struct frag_queue, q); in nf_ct_frag6_expire() 145 ip6frag_expire_frag_queue(fq->q.fqdir->net, fq); in nf_ct_frag6_expire() 169 static int nf_ct_frag6_queue(struct frag_queue *fq, struct sk_buff *skb, in nf_ct_frag6_queue() argument 178 if (fq->q.flags & INET_FRAG_COMPLETE) { in nf_ct_frag6_queue() 208 if (end < fq->q.len || in nf_ct_frag6_queue() 209 ((fq->q.flags & INET_FRAG_LAST_IN) && end != fq in nf_ct_frag6_queue() 315 nf_ct_frag6_reasm(struct frag_queue *fq, struct sk_buff *skb, struct sk_buff *prev_tail, struct net_device *dev) nf_ct_frag6_reasm() argument 446 struct frag_queue *fq; nf_ct_frag6_gather() local [all...] |
/kernel/linux/linux-6.6/net/ieee802154/6lowpan/ |
H A D | reassembly.c | 33 static int lowpan_frag_reasm(struct lowpan_frag_queue *fq, struct sk_buff *skb, 47 struct frag_queue *fq; in lowpan_frag_expire() local 49 fq = container_of(frag, struct frag_queue, q); in lowpan_frag_expire() 51 spin_lock(&fq->q.lock); in lowpan_frag_expire() 53 if (fq->q.flags & INET_FRAG_COMPLETE) in lowpan_frag_expire() 56 inet_frag_kill(&fq->q); in lowpan_frag_expire() 58 spin_unlock(&fq->q.lock); in lowpan_frag_expire() 59 inet_frag_put(&fq->q); in lowpan_frag_expire() 84 static int lowpan_frag_queue(struct lowpan_frag_queue *fq, in lowpan_frag_queue() argument 97 if (fq in lowpan_frag_queue() 164 lowpan_frag_reasm(struct lowpan_frag_queue *fq, struct sk_buff *skb, struct sk_buff *prev_tail, struct net_device *ldev) lowpan_frag_reasm() argument 279 struct lowpan_frag_queue *fq; lowpan_frag_rcv() local 498 const struct inet_frag_queue *fq = data; lowpan_obj_hashfn() local 507 const struct inet_frag_queue *fq = ptr; lowpan_obj_cmpfn() local [all...] |
/kernel/linux/linux-5.10/net/ieee802154/6lowpan/ |
H A D | reassembly.c | 33 static int lowpan_frag_reasm(struct lowpan_frag_queue *fq, struct sk_buff *skb, 47 struct frag_queue *fq; in lowpan_frag_expire() local 49 fq = container_of(frag, struct frag_queue, q); in lowpan_frag_expire() 51 spin_lock(&fq->q.lock); in lowpan_frag_expire() 53 if (fq->q.flags & INET_FRAG_COMPLETE) in lowpan_frag_expire() 56 inet_frag_kill(&fq->q); in lowpan_frag_expire() 58 spin_unlock(&fq->q.lock); in lowpan_frag_expire() 59 inet_frag_put(&fq->q); in lowpan_frag_expire() 84 static int lowpan_frag_queue(struct lowpan_frag_queue *fq, in lowpan_frag_queue() argument 97 if (fq in lowpan_frag_queue() 163 lowpan_frag_reasm(struct lowpan_frag_queue *fq, struct sk_buff *skb, struct sk_buff *prev_tail, struct net_device *ldev) lowpan_frag_reasm() argument 278 struct lowpan_frag_queue *fq; lowpan_frag_rcv() local 493 const struct inet_frag_queue *fq = data; lowpan_obj_hashfn() local 502 const struct inet_frag_queue *fq = ptr; lowpan_obj_cmpfn() local [all...] |
/kernel/linux/linux-6.6/net/ipv6/ |
H A D | reassembly.c | 70 static int ip6_frag_reasm(struct frag_queue *fq, struct sk_buff *skb, 76 struct frag_queue *fq; in ip6_frag_expire() local 78 fq = container_of(frag, struct frag_queue, q); in ip6_frag_expire() 80 ip6frag_expire_frag_queue(fq->q.fqdir->net, fq); in ip6_frag_expire() 106 static int ip6_frag_queue(struct frag_queue *fq, struct sk_buff *skb, in ip6_frag_queue() argument 119 if (fq->q.flags & INET_FRAG_COMPLETE) { in ip6_frag_queue() 151 if (end < fq->q.len || in ip6_frag_queue() 152 ((fq->q.flags & INET_FRAG_LAST_IN) && end != fq in ip6_frag_queue() 256 ip6_frag_reasm(struct frag_queue *fq, struct sk_buff *skb, struct sk_buff *prev_tail, struct net_device *dev) ip6_frag_reasm() argument 330 struct frag_queue *fq; ipv6_frag_rcv() local [all...] |
/kernel/linux/linux-5.10/net/ipv6/ |
H A D | reassembly.c | 70 static int ip6_frag_reasm(struct frag_queue *fq, struct sk_buff *skb, 76 struct frag_queue *fq; in ip6_frag_expire() local 78 fq = container_of(frag, struct frag_queue, q); in ip6_frag_expire() 80 ip6frag_expire_frag_queue(fq->q.fqdir->net, fq); in ip6_frag_expire() 106 static int ip6_frag_queue(struct frag_queue *fq, struct sk_buff *skb, in ip6_frag_queue() argument 117 if (fq->q.flags & INET_FRAG_COMPLETE) in ip6_frag_queue() 147 if (end < fq->q.len || in ip6_frag_queue() 148 ((fq->q.flags & INET_FRAG_LAST_IN) && end != fq in ip6_frag_queue() 250 ip6_frag_reasm(struct frag_queue *fq, struct sk_buff *skb, struct sk_buff *prev_tail, struct net_device *dev) ip6_frag_reasm() argument 324 struct frag_queue *fq; ipv6_frag_rcv() local [all...] |
/kernel/linux/linux-5.10/drivers/soc/fsl/qbman/ |
H A D | qman.c | 268 struct qm_mcc_fq fq; member 957 static inline void fq_set(struct qman_fq *fq, u32 mask) in fq_set() argument 959 fq->flags |= mask; in fq_set() 962 static inline void fq_clear(struct qman_fq *fq, u32 mask) in fq_clear() argument 964 fq->flags &= ~mask; in fq_clear() 967 static inline int fq_isset(struct qman_fq *fq, u32 mask) in fq_isset() argument 969 return fq->flags & mask; in fq_isset() 972 static inline int fq_isclear(struct qman_fq *fq, u32 mask) in fq_isclear() argument 974 return !(fq->flags & mask); in fq_isclear() 1114 pr_debug("Allocated fq looku in qman_alloc_fq_table() 1121 struct qman_fq *fq; idx_to_fq() local 1151 fq_to_tag(struct qman_fq *fq) fq_to_tag() argument 1425 fq_state_change(struct qman_portal *p, struct qman_fq *fq, const union qm_mr_entry *msg, u8 verb) fq_state_change() argument 1488 struct qman_fq *fq; qm_mr_process_task() local 1571 clear_vdqcr(struct qman_portal *p, struct qman_fq *fq) clear_vdqcr() argument 1608 struct qman_fq *fq; __poll_portal_fast() local 1795 qman_create_fq(u32 fqid, u32 flags, struct qman_fq *fq) qman_create_fq() argument 1825 qman_destroy_fq(struct qman_fq *fq) qman_destroy_fq() argument 1847 qman_fq_fqid(struct qman_fq *fq) qman_fq_fqid() argument 1853 qman_init_fq(struct qman_fq *fq, u32 flags, struct qm_mcc_initfq *opts) qman_init_fq() argument 1966 qman_schedule_fq(struct qman_fq *fq) qman_schedule_fq() argument 2007 qman_retire_fq(struct qman_fq *fq, u32 *flags) qman_retire_fq() argument 2089 qman_oos_fq(struct qman_fq *fq) qman_oos_fq() argument 2127 qman_query_fq(struct qman_fq *fq, struct qm_fqd *fqd) qman_query_fq() argument 2152 qman_query_fq_np(struct qman_fq *fq, struct qm_mcr_queryfq_np *np) qman_query_fq_np() argument 2223 set_p_vdqcr(struct qman_portal *p, struct qman_fq *fq, u32 vdqcr) set_p_vdqcr() argument 2243 set_vdqcr(struct qman_portal **p, struct qman_fq *fq, u32 vdqcr) set_vdqcr() argument 2253 wait_vdqcr_start(struct qman_portal **p, struct qman_fq *fq, u32 vdqcr, u32 flags) wait_vdqcr_start() argument 2266 qman_volatile_dequeue(struct qman_fq *fq, u32 flags, u32 vdqcr) qman_volatile_dequeue() argument 2312 qman_enqueue(struct qman_fq *fq, const struct qm_fd *fd) qman_enqueue() argument 2896 struct qman_fq fq = { qpool_cleanup() local 2954 struct qman_fq fq = { cgr_cleanup() local [all...] |
H A D | qman_test_api.c | 105 static int do_enqueues(struct qman_fq *fq) in do_enqueues() argument 111 if (qman_enqueue(fq, &fd)) { in do_enqueues() 125 struct qman_fq *fq = &fq_base; in qman_test_api() local 132 err = qman_create_fq(0, FQ_FLAGS, fq); in qman_test_api() 137 err = qman_init_fq(fq, QMAN_INITFQ_FLAG_LOCAL, NULL); in qman_test_api() 143 err = do_enqueues(fq); in qman_test_api() 148 err = qman_volatile_dequeue(fq, VDQCR_FLAGS, frmcnt); in qman_test_api() 153 err = do_enqueues(fq); in qman_test_api() 158 err = qman_volatile_dequeue(fq, VDQCR_FLAGS, frmcnt); in qman_test_api() 166 err = qman_volatile_dequeue(fq, VDQCR_FLAG in qman_test_api() 209 cb_dqrr(struct qman_portal *p, struct qman_fq *fq, const struct qm_dqrr_entry *dq) cb_dqrr() argument 225 cb_ern(struct qman_portal *p, struct qman_fq *fq, const union qm_mr_entry *msg) cb_ern() argument 232 cb_fqs(struct qman_portal *p, struct qman_fq *fq, const union qm_mr_entry *msg) cb_fqs() argument [all...] |
/kernel/linux/linux-6.6/drivers/soc/fsl/qbman/ |
H A D | qman.c | 268 struct qm_mcc_fq fq; member 957 static inline void fq_set(struct qman_fq *fq, u32 mask) in fq_set() argument 959 fq->flags |= mask; in fq_set() 962 static inline void fq_clear(struct qman_fq *fq, u32 mask) in fq_clear() argument 964 fq->flags &= ~mask; in fq_clear() 967 static inline int fq_isset(struct qman_fq *fq, u32 mask) in fq_isset() argument 969 return fq->flags & mask; in fq_isset() 972 static inline int fq_isclear(struct qman_fq *fq, u32 mask) in fq_isclear() argument 974 return !(fq->flags & mask); in fq_isclear() 1114 pr_debug("Allocated fq looku in qman_alloc_fq_table() 1121 struct qman_fq *fq; idx_to_fq() local 1151 fq_to_tag(struct qman_fq *fq) fq_to_tag() argument 1425 fq_state_change(struct qman_portal *p, struct qman_fq *fq, const union qm_mr_entry *msg, u8 verb) fq_state_change() argument 1488 struct qman_fq *fq; qm_mr_process_task() local 1571 clear_vdqcr(struct qman_portal *p, struct qman_fq *fq) clear_vdqcr() argument 1608 struct qman_fq *fq; __poll_portal_fast() local 1795 qman_create_fq(u32 fqid, u32 flags, struct qman_fq *fq) qman_create_fq() argument 1825 qman_destroy_fq(struct qman_fq *fq) qman_destroy_fq() argument 1847 qman_fq_fqid(struct qman_fq *fq) qman_fq_fqid() argument 1853 qman_init_fq(struct qman_fq *fq, u32 flags, struct qm_mcc_initfq *opts) qman_init_fq() argument 1966 qman_schedule_fq(struct qman_fq *fq) qman_schedule_fq() argument 2007 qman_retire_fq(struct qman_fq *fq, u32 *flags) qman_retire_fq() argument 2089 qman_oos_fq(struct qman_fq *fq) qman_oos_fq() argument 2127 qman_query_fq(struct qman_fq *fq, struct qm_fqd *fqd) qman_query_fq() argument 2152 qman_query_fq_np(struct qman_fq *fq, struct qm_mcr_queryfq_np *np) qman_query_fq_np() argument 2223 set_p_vdqcr(struct qman_portal *p, struct qman_fq *fq, u32 vdqcr) set_p_vdqcr() argument 2243 set_vdqcr(struct qman_portal **p, struct qman_fq *fq, u32 vdqcr) set_vdqcr() argument 2253 wait_vdqcr_start(struct qman_portal **p, struct qman_fq *fq, u32 vdqcr, u32 flags) wait_vdqcr_start() argument 2266 qman_volatile_dequeue(struct qman_fq *fq, u32 flags, u32 vdqcr) qman_volatile_dequeue() argument 2312 qman_enqueue(struct qman_fq *fq, const struct qm_fd *fd) qman_enqueue() argument 2949 struct qman_fq fq = { qpool_cleanup() local 3007 struct qman_fq fq = { cgr_cleanup() local [all...] |
H A D | qman_test_api.c | 106 static int do_enqueues(struct qman_fq *fq) in do_enqueues() argument 112 if (qman_enqueue(fq, &fd)) { in do_enqueues() 126 struct qman_fq *fq = &fq_base; in qman_test_api() local 133 err = qman_create_fq(0, FQ_FLAGS, fq); in qman_test_api() 138 err = qman_init_fq(fq, QMAN_INITFQ_FLAG_LOCAL, NULL); in qman_test_api() 144 err = do_enqueues(fq); in qman_test_api() 149 err = qman_volatile_dequeue(fq, VDQCR_FLAGS, frmcnt); in qman_test_api() 154 err = do_enqueues(fq); in qman_test_api() 159 err = qman_volatile_dequeue(fq, VDQCR_FLAGS, frmcnt); in qman_test_api() 167 err = qman_volatile_dequeue(fq, VDQCR_FLAG in qman_test_api() 210 cb_dqrr(struct qman_portal *p, struct qman_fq *fq, const struct qm_dqrr_entry *dq, bool sched_napi) cb_dqrr() argument 227 cb_ern(struct qman_portal *p, struct qman_fq *fq, const union qm_mr_entry *msg) cb_ern() argument 234 cb_fqs(struct qman_portal *p, struct qman_fq *fq, const union qm_mr_entry *msg) cb_fqs() argument [all...] |
/kernel/linux/linux-5.10/drivers/iommu/ |
H A D | iova.c | 49 iovad->fq = NULL; in init_iova_domain() 59 return !!iovad->fq; in has_iova_flush_queue() 71 free_percpu(iovad->fq); in free_iova_flush_queue() 73 iovad->fq = NULL; in free_iova_flush_queue() 95 struct iova_fq *fq; in init_iova_flush_queue() local 97 fq = per_cpu_ptr(queue, cpu); in init_iova_flush_queue() 98 fq->head = 0; in init_iova_flush_queue() 99 fq->tail = 0; in init_iova_flush_queue() 101 spin_lock_init(&fq->lock); in init_iova_flush_queue() 106 iovad->fq in init_iova_flush_queue() 462 fq_full(struct iova_fq *fq) fq_full() argument 468 fq_ring_add(struct iova_fq *fq) fq_ring_add() argument 479 fq_ring_free(struct iova_domain *iovad, struct iova_fq *fq) fq_ring_free() argument 522 struct iova_fq *fq = per_cpu_ptr(iovad->fq, cpu); fq_destroy_all_entries() local 540 struct iova_fq *fq; fq_flush_timeout() local 553 struct iova_fq *fq = raw_cpu_ptr(iovad->fq); queue_iova() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/freescale/dpaa2/ |
H A D | dpaa2-eth-debugfs.c | 47 static char *fq_type_to_str(struct dpaa2_eth_fq *fq) in fq_type_to_str() argument 49 switch (fq->type) { in fq_type_to_str() 62 struct dpaa2_eth_fq *fq; in dpaa2_dbg_fqs_show() local 71 fq = &priv->fq[i]; in dpaa2_dbg_fqs_show() 72 err = dpaa2_io_query_fq_count(NULL, fq->fqid, &fcnt, &bcnt); in dpaa2_dbg_fqs_show() 77 if (!fq->stats.frames && !fcnt) in dpaa2_dbg_fqs_show() 81 fq->fqid, in dpaa2_dbg_fqs_show() 82 fq->target_cpu, in dpaa2_dbg_fqs_show() 83 fq in dpaa2_dbg_fqs_show() [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/freescale/dpaa/ |
H A D | dpaa_eth_trace.h | 58 struct qman_fq *fq, 62 TP_ARGS(netdev, fq, fd), 80 __entry->fqid = fq->fqid; 104 struct qman_fq *fq, 107 TP_ARGS(netdev, fq, fd) 114 struct qman_fq *fq, 117 TP_ARGS(netdev, fq, fd) 124 struct qman_fq *fq, 127 TP_ARGS(netdev, fq, fd)
|
H A D | dpaa_eth_sysfs.c | 59 struct dpaa_fq *fq; in dpaa_eth_show_fqids() local 66 list_for_each_entry_safe(fq, tmp, &priv->dpaa_fq_list, list) { in dpaa_eth_show_fqids() 67 switch (fq->fq_type) { in dpaa_eth_show_fqids() 93 if (prev && (abs(fq->fqid - prev->fqid) != 1 || in dpaa_eth_show_fqids() 104 if (prev && abs(fq->fqid - prev->fqid) == 1 && in dpaa_eth_show_fqids() 106 last_fqid = fq->fqid; in dpaa_eth_show_fqids() 108 first_fqid = fq->fqid; in dpaa_eth_show_fqids() 109 last_fqid = fq->fqid; in dpaa_eth_show_fqids() 112 prev = fq; in dpaa_eth_show_fqids()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/freescale/dpaa/ |
H A D | dpaa_eth_trace.h | 32 struct qman_fq *fq, 36 TP_ARGS(netdev, fq, fd), 54 __entry->fqid = fq->fqid; 78 struct qman_fq *fq, 81 TP_ARGS(netdev, fq, fd) 88 struct qman_fq *fq, 91 TP_ARGS(netdev, fq, fd) 98 struct qman_fq *fq, 101 TP_ARGS(netdev, fq, fd)
|
H A D | dpaa_eth_sysfs.c | 33 struct dpaa_fq *fq; in dpaa_eth_show_fqids() local 40 list_for_each_entry_safe(fq, tmp, &priv->dpaa_fq_list, list) { in dpaa_eth_show_fqids() 41 switch (fq->fq_type) { in dpaa_eth_show_fqids() 67 if (prev && (abs(fq->fqid - prev->fqid) != 1 || in dpaa_eth_show_fqids() 78 if (prev && abs(fq->fqid - prev->fqid) == 1 && in dpaa_eth_show_fqids() 80 last_fqid = fq->fqid; in dpaa_eth_show_fqids() 82 first_fqid = fq->fqid; in dpaa_eth_show_fqids() 83 last_fqid = fq->fqid; in dpaa_eth_show_fqids() 86 prev = fq; in dpaa_eth_show_fqids()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/freescale/dpaa2/ |
H A D | dpaa2-eth-debugfs.c | 47 static char *fq_type_to_str(struct dpaa2_eth_fq *fq) in fq_type_to_str() argument 49 switch (fq->type) { in fq_type_to_str() 62 struct dpaa2_eth_fq *fq; in dpaa2_dbg_fqs_show() local 71 fq = &priv->fq[i]; in dpaa2_dbg_fqs_show() 72 err = dpaa2_io_query_fq_count(NULL, fq->fqid, &fcnt, &bcnt); in dpaa2_dbg_fqs_show() 77 if (!fq->stats.frames && !fcnt) in dpaa2_dbg_fqs_show() 81 fq->fqid, in dpaa2_dbg_fqs_show() 82 fq->target_cpu, in dpaa2_dbg_fqs_show() 83 fq in dpaa2_dbg_fqs_show() [all...] |
/kernel/linux/linux-5.10/drivers/crypto/caam/ |
H A D | qi.c | 135 static void caam_fq_ern_cb(struct qman_portal *qm, struct qman_fq *fq, in caam_fq_ern_cb() argument 218 static int empty_retired_fq(struct device *qidev, struct qman_fq *fq) in empty_retired_fq() argument 222 ret = qman_volatile_dequeue(fq, QMAN_VOLATILE_FLAG_WAIT_INT | in empty_retired_fq() 227 dev_err(qidev, "Volatile dequeue fail for FQ: %u\n", fq->fqid); in empty_retired_fq() 236 } while (fq->flags & QMAN_FQ_STATE_NE); in empty_retired_fq() 241 static int kill_fq(struct device *qidev, struct qman_fq *fq) in kill_fq() argument 246 ret = qman_retire_fq(fq, &flags); in kill_fq() 260 } while (fq->state != qman_fq_state_retired); in kill_fq() 262 WARN_ON(fq->flags & QMAN_FQ_STATE_BLOCKOOS); in kill_fq() 263 WARN_ON(fq in kill_fq() 286 empty_caam_fq(struct qman_fq *fq, struct caam_drv_ctx *drv_ctx) empty_caam_fq() argument 617 struct qman_fq *fq; alloc_rsp_fq_cpu() local [all...] |
/kernel/linux/linux-6.6/drivers/crypto/caam/ |
H A D | qi.c | 142 static void caam_fq_ern_cb(struct qman_portal *qm, struct qman_fq *fq, in caam_fq_ern_cb() argument 225 static int empty_retired_fq(struct device *qidev, struct qman_fq *fq) in empty_retired_fq() argument 229 ret = qman_volatile_dequeue(fq, QMAN_VOLATILE_FLAG_WAIT_INT | in empty_retired_fq() 234 dev_err(qidev, "Volatile dequeue fail for FQ: %u\n", fq->fqid); in empty_retired_fq() 243 } while (fq->flags & QMAN_FQ_STATE_NE); in empty_retired_fq() 248 static int kill_fq(struct device *qidev, struct qman_fq *fq) in kill_fq() argument 253 ret = qman_retire_fq(fq, &flags); in kill_fq() 267 } while (fq->state != qman_fq_state_retired); in kill_fq() 269 WARN_ON(fq->flags & QMAN_FQ_STATE_BLOCKOOS); in kill_fq() 270 WARN_ON(fq in kill_fq() 293 empty_caam_fq(struct qman_fq *fq, struct caam_drv_ctx *drv_ctx) empty_caam_fq() argument 621 struct qman_fq *fq; alloc_rsp_fq_cpu() local [all...] |