/kernel/linux/linux-6.6/include/linux/ |
H A D | bvec.h | 22 * @bv_len: Number of bytes in the address range. 25 * The following holds for a bvec if n * PAGE_SIZE < bv_offset + bv_len: 33 unsigned int bv_len; member 48 bv->bv_len = len; in bvec_set_page() 106 __bvec_iter_bvec((bvec), (iter))->bv_len - (iter).bi_bvec_done) 117 .bv_len = mp_bvec_iter_len((bvec), (iter)), \ 136 .bv_len = bvec_iter_len((bvec), (iter)), \ 154 while (bytes && bytes >= bv[idx].bv_len) { in bvec_iter_advance() 155 bytes -= bv[idx].bv_len; in bvec_iter_advance() 166 * across multiple bvec entries, i.e. bytes <= bv[i->bi_idx].bv_len [all...] |
/kernel/linux/linux-5.10/include/linux/ |
H A D | bvec.h | 22 * @bv_len: Number of bytes in the address range. 25 * The following holds for a bvec if n * PAGE_SIZE < bv_offset + bv_len: 33 unsigned int bv_len; member 66 __bvec_iter_bvec((bvec), (iter))->bv_len - (iter).bi_bvec_done) 77 .bv_len = mp_bvec_iter_len((bvec), (iter)), \ 96 .bv_len = bvec_iter_len((bvec), (iter)), \ 114 while (bytes && bytes >= bv[idx].bv_len) { in bvec_iter_advance() 115 bytes -= bv[idx].bv_len; in bvec_iter_advance() 134 (bvl).bv_len ? (void)bvec_iter_advance((bio_vec), &(iter), \ 135 (bvl).bv_len) [all...] |
H A D | bio.h | 87 return bio_iovec(bio).bv_len; in bio_cur_bytes() 152 bio_advance_iter((bio), &(iter), (bvl).bv_len)) 161 bio_advance_iter((bio), &(iter), (bvl).bv_len)) 175 #define bio_iter_last(bvec, iter) ((iter).bi_size == (bvec).bv_len) 261 if (bv->bv_len == bio->bi_iter.bi_size) in bio_get_last_bvec() 278 bv->bv_len = iter.bi_bvec_done; in bio_get_last_bvec()
|
/kernel/linux/linux-5.10/lib/ |
H A D | iov_iter.c | 75 if (!__v.bv_len) \ 628 (from += v.bv_len) - v.bv_len, v.bv_len), in _copy_to_iter() 730 (from += v.bv_len) - v.bv_len, v.bv_len); in _copy_mc_to_iter() 764 memcpy_from_page((to += v.bv_len) - v.bv_len, v.bv_page, in _copy_from_iter() 765 v.bv_offset, v.bv_len), in _copy_from_iter() [all...] |
/kernel/linux/linux-6.6/drivers/block/ |
H A D | n64cart.c | 69 (bv->bv_len & (MIN_ALIGNMENT - 1))); in n64cart_do_bvec() 79 n64cart_write_reg(PI_WRITE_REG, bv->bv_len - 1); in n64cart_do_bvec() 83 dma_unmap_page(dev, dma_addr, bv->bv_len, DMA_FROM_DEVICE); in n64cart_do_bvec() 99 pos += bvec.bv_len; in n64cart_submit_bio()
|
/kernel/linux/linux-5.10/block/ |
H A D | bounce.c | 81 memcpy(vto + to->bv_offset, vfrom, to->bv_len); in bounce_copy_vec() 88 memcpy(page_address((to)->bv_page) + (to)->bv_offset, vfrom, (to)->bv_len) 148 * fromvec->bv_offset and fromvec->bv_len might have in copy_to_high_bio_irq() 150 * copy, bounce_copy_vec already uses tovec->bv_len in copy_to_high_bio_irq() 158 bio_advance_iter(from, &from_iter, tovec.bv_len); in copy_to_high_bio_irq() 178 bio_advance_iter(bio_orig, &orig_iter, orig_vec.bv_len); in bounce_end_io() 301 sectors += from.bv_len >> 9; in __blk_queue_bounce() 338 memcpy(vto, vfrom, to->bv_len); in __blk_queue_bounce()
|
H A D | blk-map.c | 56 bvec->bv_len, in bio_copy_from_iter() 62 if (ret < bvec->bv_len) in bio_copy_from_iter() 87 bvec->bv_len, in bio_copy_to_iter() 93 if (ret < bvec->bv_len) in bio_copy_to_iter() 359 len += bio->bi_io_vec[i].bv_len; in bio_invalidate_vmalloc_pages() 445 memcpy(p, page_address(bvec->bv_page), bvec->bv_len); in bio_copy_kern_endio_read() 446 p += bvec->bv_len; in bio_copy_kern_endio_read()
|
H A D | blk-merge.c | 204 unsigned len = min(bv->bv_len, max_len); in bvec_split_segs() 224 return len > 0 || bv->bv_len > max_len; in bvec_split_segs() 266 sectors + (bv.bv_len >> 9) <= max_sectors && in blk_bio_segment_split() 267 bv.bv_offset + bv.bv_len <= PAGE_SIZE) { in blk_bio_segment_split() 269 sectors += bv.bv_len >> 9; in blk_bio_segment_split() 328 ((*bio)->bi_io_vec[0].bv_len + in __blk_queue_split() 422 unsigned nbytes = bvec->bv_len; in blk_bvec_map_sg() 457 sg_set_page(*sg, bv.bv_page, bv.bv_len, bv.bv_offset); in __blk_bvec_map_sg() 467 int nbytes = bvec->bv_len; in __blk_segment_map_sg_merge() 503 if (bvec.bv_offset + bvec.bv_len < in __blk_bios_map_sg() [all...] |
H A D | blk-integrity.c | 40 if (seg_size + iv.bv_len > queue_max_segment_size(q)) in blk_rq_count_integrity_sg() 43 seg_size += iv.bv_len; in blk_rq_count_integrity_sg() 47 seg_size = iv.bv_len; in blk_rq_count_integrity_sg() 82 if (sg->length + iv.bv_len > queue_max_segment_size(q)) in blk_rq_map_integrity_sg() 85 sg->length += iv.bv_len; in blk_rq_map_integrity_sg() 95 sg_set_page(sg, iv.bv_page, iv.bv_len, iv.bv_offset); in blk_rq_map_integrity_sg()
|
H A D | bio.c | 540 memset(data, 0, bv.bv_len); in zero_fill_bio_iter() 571 if (done + bv.bv_len > new_size) { in bio_truncate() 579 bv.bv_len - offset); in bio_truncate() 582 done += bv.bv_len; in bio_truncate() 742 size_t bv_end = bv->bv_offset + bv->bv_len; in page_is_mergeable() 773 if (len > queue_max_segment_size(q) - bv->bv_len) in bio_try_merge_hw_seg() 824 bvec->bv_len = len; in bio_add_hw_page() 885 bv->bv_len += len; in __bio_try_merge_page() 914 bv->bv_len = len; in __bio_add_page() 970 if (WARN_ON_ONCE(iter->iov_offset > bv->bv_len)) in __bio_iov_bvec_add_pages() [all...] |
/kernel/linux/linux-6.6/block/ |
H A D | blk-integrity.c | 40 if (seg_size + iv.bv_len > queue_max_segment_size(q)) in blk_rq_count_integrity_sg() 43 seg_size += iv.bv_len; in blk_rq_count_integrity_sg() 47 seg_size = iv.bv_len; in blk_rq_count_integrity_sg() 82 if (sg->length + iv.bv_len > queue_max_segment_size(q)) in blk_rq_map_integrity_sg() 85 sg->length += iv.bv_len; in blk_rq_map_integrity_sg() 95 sg_set_page(sg, iv.bv_page, iv.bv_len, iv.bv_offset); in blk_rq_map_integrity_sg()
|
H A D | blk-map.c | 58 bvec->bv_len, in bio_copy_from_iter() 64 if (ret < bvec->bv_len) in bio_copy_from_iter() 89 bvec->bv_len, in bio_copy_to_iter() 95 if (ret < bvec->bv_len) in bio_copy_to_iter() 366 len += bio->bi_io_vec[i].bv_len; in bio_invalidate_vmalloc_pages() 457 p += bvec->bv_len; in bio_copy_kern_endio_read() 600 if (nsegs >= nr_segs || bytes > UINT_MAX - bv->bv_len) in blk_rq_map_user_bvec() 602 if (bytes + bv->bv_len > nr_iter) in blk_rq_map_user_bvec() 604 if (bv->bv_offset + bv->bv_len > PAGE_SIZE) in blk_rq_map_user_bvec() 608 bytes += bv->bv_len; in blk_rq_map_user_bvec() [all...] |
H A D | blk-merge.c | 32 if (bv->bv_len == bio->bi_iter.bi_size) in bio_get_last_bvec() 49 bv->bv_len = iter.bi_bvec_done; in bio_get_last_bvec() 236 unsigned len = min(bv->bv_len, max_len); in bvec_split_segs() 256 return len > 0 || bv->bv_len > max_len; in bvec_split_segs() 295 bytes + bv.bv_len <= max_bytes && in bio_split_rw() 296 bv.bv_offset + bv.bv_len <= PAGE_SIZE) { in bio_split_rw() 298 bytes += bv.bv_len; in bio_split_rw() 466 unsigned nbytes = bvec->bv_len; in blk_bvec_map_sg() 501 sg_set_page(*sg, bv.bv_page, bv.bv_len, bv.bv_offset); in __blk_bvec_map_sg() 511 int nbytes = bvec->bv_len; in __blk_segment_map_sg_merge() [all...] |
H A D | bounce.c | 91 * fromvec->bv_offset and fromvec->bv_len might have in copy_to_high_bio_irq() 93 * copy, bounce_copy_vec already uses tovec->bv_len in copy_to_high_bio_irq() 98 bio_advance_iter(from, &from_iter, tovec.bv_len); in copy_to_high_bio_irq() 118 bio_advance_iter(bio_orig, &orig_iter, orig_vec.bv_len); in bounce_end_io() 214 bytes += from.bv_len; in __blk_queue_bounce()
|
/kernel/linux/linux-6.6/drivers/md/ |
H A D | dm-io-rewind.c | 27 while (idx >= 0 && bytes && bytes > bv[idx].bv_len) { in dm_bvec_iter_rewind() 28 bytes -= bv[idx].bv_len; in dm_bvec_iter_rewind() 41 iter->bi_bvec_done = bv[idx].bv_len - bytes; in dm_bvec_iter_rewind()
|
H A D | dm-ebs-target.c | 71 unsigned int bv_len = bv->bv_len; in __ebs_rw_bvec() local 76 if (unlikely(!bv->bv_page || !bv_len)) in __ebs_rw_bvec() 82 while (bv_len) { in __ebs_rw_bvec() 83 cur_len = min(dm_bufio_get_block_size(ec->bufio) - buf_off, bv_len); in __ebs_rw_bvec() 86 if (op == REQ_OP_READ || buf_off || bv_len < dm_bufio_get_block_size(ec->bufio)) in __ebs_rw_bvec() 113 bv_len -= cur_len; in __ebs_rw_bvec()
|
H A D | dm-log-writes.c | 367 block->vecs[i].bv_len, 0); in log_one_block() 368 if (ret != block->vecs[i].bv_len) { in log_one_block() 380 block->vecs[i].bv_len, 0); in log_one_block() 381 if (ret != block->vecs[i].bv_len) { in log_one_block() 387 sector += block->vecs[i].bv_len >> SECTOR_SHIFT; in log_one_block() 750 block->vecs[i].bv_len = bv.bv_len; in log_writes_map()
|
/kernel/linux/linux-5.10/drivers/md/ |
H A D | dm-ebs-target.c | 69 unsigned int bv_len = bv->bv_len; in __ebs_rw_bvec() local 74 if (unlikely(!bv->bv_page || !bv_len)) in __ebs_rw_bvec() 80 while (bv_len) { in __ebs_rw_bvec() 81 cur_len = min(dm_bufio_get_block_size(ec->bufio) - buf_off, bv_len); in __ebs_rw_bvec() 84 if (rw == READ || buf_off || bv_len < dm_bufio_get_block_size(ec->bufio)) in __ebs_rw_bvec() 111 bv_len -= cur_len; in __ebs_rw_bvec()
|
H A D | dm-log-writes.c | 385 block->vecs[i].bv_len, 0); in log_one_block() 386 if (ret != block->vecs[i].bv_len) { in log_one_block() 402 block->vecs[i].bv_len, 0); in log_one_block() 403 if (ret != block->vecs[i].bv_len) { in log_one_block() 409 sector += block->vecs[i].bv_len >> SECTOR_SHIFT; in log_one_block() 770 memcpy(dst, src + bv.bv_offset, bv.bv_len); in log_writes_map() 774 block->vecs[i].bv_len = bv.bv_len; in log_writes_map()
|
/kernel/linux/linux-5.10/drivers/md/bcache/ |
H A D | util.c | 244 start: bv->bv_len = min_t(size_t, PAGE_SIZE - bv->bv_offset, in bch_bio_map() 251 base += bv->bv_len; in bch_bio_map() 254 size -= bv->bv_len; in bch_bio_map()
|
/kernel/linux/linux-5.10/fs/squashfs/ |
H A D | block.c | 47 int bytes_to_copy = min_t(int, bvec->bv_len - offset, in copy_bio_to_actor() 66 if (offset >= bvec->bv_len) { in copy_bio_to_actor() 182 if (offset < bvec->bv_len - 1) { in squashfs_read_data()
|
/kernel/linux/linux-6.6/drivers/md/bcache/ |
H A D | util.c | 244 start: bv->bv_len = min_t(size_t, PAGE_SIZE - bv->bv_offset, in bch_bio_map() 251 base += bv->bv_len; in bch_bio_map() 254 size -= bv->bv_len; in bch_bio_map()
|
/kernel/linux/linux-5.10/drivers/s390/block/ |
H A D | dasd_fba.c | 471 if (bv.bv_len & (blksize - 1)) in dasd_fba_build_cp_regular() 474 count += bv.bv_len >> (block->s2b_shift + 9); in dasd_fba_build_cp_regular() 475 if (idal_is_needed (page_address(bv.bv_page), bv.bv_len)) in dasd_fba_build_cp_regular() 476 cidaw += bv.bv_len / blksize; in dasd_fba_build_cp_regular() 518 memcpy(copy + bv.bv_offset, dst, bv.bv_len); in dasd_fba_build_cp_regular() 522 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_fba_build_cp_regular() 596 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_fba_free_cp() 607 memcpy(dst, cda, bv.bv_len); in dasd_fba_free_cp()
|
/kernel/linux/linux-6.6/drivers/s390/block/ |
H A D | dasd_fba.c | 462 if (bv.bv_len & (blksize - 1)) in dasd_fba_build_cp_regular() 465 count += bv.bv_len >> (block->s2b_shift + 9); in dasd_fba_build_cp_regular() 466 if (idal_is_needed (page_address(bv.bv_page), bv.bv_len)) in dasd_fba_build_cp_regular() 467 cidaw += bv.bv_len / blksize; in dasd_fba_build_cp_regular() 509 memcpy(copy + bv.bv_offset, dst, bv.bv_len); in dasd_fba_build_cp_regular() 513 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_fba_build_cp_regular() 587 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_fba_free_cp() 598 memcpy(dst, cda, bv.bv_len); in dasd_fba_free_cp()
|
/kernel/linux/linux-5.10/drivers/block/zram/ |
H A D | zram_drv.c | 97 return bvec->bv_len != PAGE_SIZE; in is_partial_io() 132 *index += (*offset + bvec->bv_len) / PAGE_SIZE; in update_position() 133 *offset = (*offset + bvec->bv_len) % PAGE_SIZE; in update_position() 552 if (!bio_add_page(bio, bvec->bv_page, bvec->bv_len, bvec->bv_offset)) { in read_from_bdev_async() 613 bvec.bv_len = PAGE_SIZE; in writeback_store() 668 bio_add_page(&bio, bvec.bv_page, bvec.bv_len, in writeback_store() 1255 bvec.bv_len = PAGE_SIZE; in __zram_bvec_read() 1337 memcpy(dst + bvec->bv_offset, src + offset, bvec->bv_len); in zram_bvec_read() 1492 memcpy(dst + offset, src + bvec->bv_offset, bvec->bv_len); in zram_bvec_write() 1497 vec.bv_len in zram_bvec_write() [all...] |