Lines Matching refs:pr

76 				       const struct kvec_test_range *pr)
81 for (i = 0; i < kvmax; i++, pr++) {
82 if (pr->from < 0)
84 KUNIT_ASSERT_GE(test, pr->to, pr->from);
85 KUNIT_ASSERT_LE(test, pr->to, bufsize);
86 kvec[i].iov_base = buffer + pr->from;
87 kvec[i].iov_len = pr->to - pr->from;
88 size += pr->to - pr->from;
100 const struct kvec_test_range *pr;
131 for (pr = kvec_test_ranges; pr->from >= 0; pr++)
132 for (i = pr->from; i < pr->to; i++)
150 const struct kvec_test_range *pr;
181 for (pr = kvec_test_ranges; pr->from >= 0; pr++) {
182 for (j = pr->from; j < pr->to; j++) {
221 const struct bvec_test_range *pr)
227 for (i = 0; i < bvmax; i++, pr++) {
228 if (pr->from < 0)
230 KUNIT_ASSERT_LT(test, pr->page, npages);
231 KUNIT_ASSERT_LT(test, pr->page * PAGE_SIZE, bufsize);
232 KUNIT_ASSERT_GE(test, pr->from, 0);
233 KUNIT_ASSERT_GE(test, pr->to, pr->from);
234 KUNIT_ASSERT_LE(test, pr->to, PAGE_SIZE);
236 page = pages[pr->page];
237 if (pr->from == 0 && pr->from != pr->to && page == can_merge) {
239 bvec[i].bv_len += pr->to;
241 bvec_set_page(&bvec[i], page, pr->to - pr->from, pr->from);
244 size += pr->to - pr->from;
245 if ((pr->to & ~PAGE_MASK) == 0)
246 can_merge = page + pr->to / PAGE_SIZE;
259 const struct bvec_test_range *pr;
291 for (pr = bvec_test_ranges; pr->from >= 0; pr++, b++) {
292 u8 *p = scratch + pr->page * PAGE_SIZE;
294 for (i = pr->from; i < pr->to; i++)
313 const struct bvec_test_range *pr;
344 for (pr = bvec_test_ranges; pr->from >= 0; pr++) {
345 size_t patt = pr->page * PAGE_SIZE;
347 for (j = pr->from; j < pr->to; j++) {
406 const struct kvec_test_range *pr;
429 for (pr = kvec_test_ranges; pr->from >= 0; pr++) {
430 size = pr->to - pr->from;
431 KUNIT_ASSERT_LE(test, pr->to, bufsize);
433 iov_iter_xarray(&iter, READ, xarray, pr->from, size);
445 for (pr = kvec_test_ranges; pr->from >= 0; pr++)
446 for (i = pr->from; i < pr->to; i++)
464 const struct kvec_test_range *pr;
487 for (pr = kvec_test_ranges; pr->from >= 0; pr++) {
488 size = pr->to - pr->from;
489 KUNIT_ASSERT_LE(test, pr->to, bufsize);
491 iov_iter_xarray(&iter, WRITE, xarray, pr->from, size);
503 for (pr = kvec_test_ranges; pr->from >= 0; pr++) {
504 for (j = pr->from; j < pr->to; j++) {
527 const struct kvec_test_range *pr;
545 pr = kvec_test_ranges;
546 from = pr->from;
573 while (from == pr->to) {
574 pr++;
575 from = pr->from;
607 const struct bvec_test_range *pr;
623 pr = bvec_test_ranges;
624 from = pr->from;
651 while (from == pr->to) {
652 pr++;
653 from = pr->from;
657 ix = pr->page + from / PAGE_SIZE;
685 const struct kvec_test_range *pr;
701 for (pr = kvec_test_ranges; pr->from >= 0; pr++) {
702 from = pr->from;
703 size = pr->to - from;
704 KUNIT_ASSERT_LE(test, pr->to, bufsize);
752 KUNIT_EXPECT_EQ(test, iter.iov_offset, pr->to - pr->from);