Lines Matching refs:data
327 * we don't want to throttle again until we get more data, so we
554 static void kyber_limit_depth(blk_opf_t opf, struct blk_mq_alloc_data *data)
561 struct kyber_queue_data *kqd = data->q->elevator->elevator_data;
563 data->shallow_depth = kqd->async_depth;
667 static bool flush_busy_kcq(struct sbitmap *sb, unsigned int bitnr, void *data)
669 struct flush_kcq_data *flush_data = data;
685 struct flush_kcq_data data = {
692 flush_busy_kcq, &data);
901 static int kyber_##name##_tokens_show(void *data, struct seq_file *m) \
903 struct request_queue *q = data; \
945 static int kyber_##name##_waiting_show(void *data, struct seq_file *m) \
947 struct blk_mq_hw_ctx *hctx = data; \
960 static int kyber_async_depth_show(void *data, struct seq_file *m)
962 struct request_queue *q = data;
969 static int kyber_cur_domain_show(void *data, struct seq_file *m)
971 struct blk_mq_hw_ctx *hctx = data;
978 static int kyber_batching_show(void *data, struct seq_file *m)
980 struct blk_mq_hw_ctx *hctx = data;