Lines Matching refs:next_rq
578 if (sector > blk_rq_pos(bfqq->next_rq))
580 else if (sector < blk_rq_pos(bfqq->next_rq))
639 if (!bfqq->next_rq)
644 blk_rq_pos(bfqq->next_rq), &parent, &p);
937 * bfq_updated_next_req - update the queue after a new next_rq selection.
951 struct request *next_rq = bfqq->next_rq;
954 if (!next_rq)
966 bfq_serv_to_charge(next_rq, bfqq)),
1508 bfq_serv_to_charge(bfqq->next_rq, bfqq));
1876 struct request *next_rq, *prev;
2063 prev = bfqq->next_rq;
2064 next_rq = bfq_choose_req(bfqd, bfqq->next_rq, rq, bfqd->last_position);
2065 bfqq->next_rq = next_rq;
2068 * Adjust priority tree position, if next_rq changes.
2071 if (unlikely(!bfqd->nonrot_with_queueing && prev != bfqq->next_rq))
2088 if (prev != bfqq->next_rq)
2167 if (bfqq->next_rq == rq) {
2168 bfqq->next_rq = bfq_find_next_rq(bfqd, bfqq, rq);
2183 bfqq->next_rq = NULL;
2288 struct request *prev, *next_rq;
2300 prev = bfqq->next_rq;
2301 next_rq = bfq_choose_req(bfqd, bfqq->next_rq, req,
2303 bfqq->next_rq = next_rq;
2305 * If next_rq changes, update both the queue's budget to
2309 if (prev != bfqq->next_rq) {
2361 if (bfqq->next_rq == next)
2362 bfqq->next_rq = rq;
2458 if (bfq_rq_close_to_sector(__bfqq->next_rq, true, sector))
2461 if (blk_rq_pos(__bfqq->next_rq) < sector)
2469 if (bfq_rq_close_to_sector(__bfqq->next_rq, true, sector))
3575 struct request *next_rq;
3722 next_rq = bfqq->next_rq;
3723 if (next_rq)
3725 bfq_serv_to_charge(next_rq, bfqq));
3728 next_rq ? blk_rq_sectors(next_rq) : 0,
4337 bfq_serv_to_charge(bfqq->next_rq, bfqq) <=
4357 blk_rq_sectors(bfqq->next_rq) >=
4379 struct request *next_rq;
4406 next_rq = bfqq->next_rq;
4411 if (next_rq) {
4412 if (bfq_serv_to_charge(next_rq, bfqq) >
4462 bfqq->bic->bfqq[0]->next_rq ?
4538 icq_to_bic(async_bfqq->next_rq->elv.icq) == bfqq->bic &&
4539 bfq_serv_to_charge(async_bfqq->next_rq, async_bfqq) <=
4544 bfqq->waker_bfqq->next_rq &&
4545 bfq_serv_to_charge(bfqq->waker_bfqq->next_rq,
4637 struct request *rq = bfqq->next_rq;