Lines Matching refs:fq_flow
68 struct fq_flow {
85 struct fq_flow *next; /* next pointer in RR lists */
92 struct fq_flow *first;
93 struct fq_flow *last;
106 struct fq_flow internal; /* for non classified or high prio packets */
144 static void fq_flow_set_detached(struct fq_flow *f)
149 static bool fq_flow_is_detached(const struct fq_flow *f)
155 static struct fq_flow throttled;
157 static bool fq_flow_is_throttled(const struct fq_flow *f)
162 static void fq_flow_add_tail(struct fq_flow_head *head, struct fq_flow *flow)
172 static void fq_flow_unset_throttled(struct fq_sched_data *q, struct fq_flow *f)
179 static void fq_flow_set_throttled(struct fq_sched_data *q, struct fq_flow *f)
184 struct fq_flow *aux;
187 aux = rb_entry(parent, struct fq_flow, rate_node);
211 static bool fq_gc_candidate(const struct fq_flow *f)
223 struct fq_flow *f;
231 f = rb_entry(parent, struct fq_flow, fq_node);
261 static struct fq_flow *fq_classify(struct sk_buff *skb, struct fq_sched_data *q)
266 struct fq_flow *f;
313 f = rb_entry(parent, struct fq_flow, fq_node);
364 static struct sk_buff *fq_peek(struct fq_flow *flow)
380 static void fq_erase_head(struct Qdisc *sch, struct fq_flow *flow,
394 static void fq_dequeue_skb(struct Qdisc *sch, struct fq_flow *flow,
404 static void flow_queue_add(struct fq_flow *flow, struct sk_buff *skb)
446 struct fq_flow *f;
516 struct fq_flow *f = rb_entry(p, struct fq_flow, rate_node);
531 struct fq_flow *f;
646 static void fq_flow_purge(struct fq_flow *flow)
667 struct fq_flow *f;
681 f = rb_entry(p, struct fq_flow, fq_node);
703 struct fq_flow *of, *nf;
711 of = rb_entry(op, struct fq_flow, fq_node);
724 nf = rb_entry(parent, struct fq_flow, fq_node);
1058 sizeof(struct fq_flow),