Lines Matching defs:queue
59 * from the queue itself.
246 * The stochastic-multi-queue is a set of lru lists stacked into levels.
253 struct queue {
270 static void q_init(struct queue *q, struct entry_space *es, unsigned int nr_levels)
288 static unsigned int q_size(struct queue *q)
296 static void q_push(struct queue *q, struct entry *e)
306 static void q_push_front(struct queue *q, struct entry *e)
316 static void q_push_before(struct queue *q, struct entry *old, struct entry *e)
326 static void q_del(struct queue *q, struct entry *e)
336 static struct entry *q_peek(struct queue *q, unsigned int max_level, bool can_cross_sentinel)
358 static struct entry *q_pop(struct queue *q)
373 static struct entry *__redist_pop_from(struct queue *q, unsigned int level)
387 static void q_set_targets_subrange_(struct queue *q, unsigned int nr_elts,
407 static void q_set_targets(struct queue *q)
429 static void q_redistribute(struct queue *q)
472 static void q_requeue(struct queue *q, struct entry *e, unsigned int extra_levels,
554 * There are times when we don't have any confidence in the hotspot queue.
557 * seeing how often a lookup is in the top levels of the hotspot queue.
821 * consisting of a clean and dirty queue, containing the currently
822 * active mappings. The hotspot queue uses a larger block size to
826 struct queue hotspot;
827 struct queue clean;
828 struct queue dirty;
889 struct queue *q = &mq->dirty;
902 struct queue *q = &mq->clean;
1060 * If the hotspot queue is performing badly then we have little
1082 * If the hotspot queue is performing badly, then we try and move entries
1145 * size of the clean queue.
1397 * The hotspot queue only gets updated with misses.