Home
last modified time | relevance | path

Searched refs:bv (Results 1 - 25 of 185) sorted by relevance

12345678

/kernel/linux/linux-6.6/include/linux/
H A Dbvec.h39 * @bv: bvec to initialize
44 static inline void bvec_set_page(struct bio_vec *bv, struct page *page, in bvec_set_page() argument
47 bv->bv_page = page; in bvec_set_page()
48 bv->bv_len = len; in bvec_set_page()
49 bv->bv_offset = offset; in bvec_set_page()
54 * @bv: bvec to initialize
59 static inline void bvec_set_folio(struct bio_vec *bv, struct folio *folio, in bvec_set_folio() argument
62 bvec_set_page(bv, &folio->page, len, offset); in bvec_set_folio()
67 * @bv: bvec to initialize
71 static inline void bvec_set_virt(struct bio_vec *bv, voi argument
89 struct bio_vec bv; global() member
140 bvec_iter_advance(const struct bio_vec *bv, struct bvec_iter *iter, unsigned bytes) bvec_iter_advance() argument
168 bvec_iter_advance_single(const struct bio_vec *bv, struct bvec_iter *iter, unsigned int bytes) bvec_iter_advance_single() argument
207 struct bio_vec *bv = &iter_all->bv; bvec_advance() local
[all...]
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gt/
H A Dgen7_renderclear.c43 static inline int num_primitives(const struct batch_vals *bv) in num_primitives() argument
51 return bv->max_threads; in num_primitives()
55 batch_get_defaults(struct drm_i915_private *i915, struct batch_vals *bv) in batch_get_defaults() argument
61 bv->max_threads = 70; in batch_get_defaults()
64 bv->max_threads = 140; in batch_get_defaults()
67 bv->max_threads = 280; in batch_get_defaults()
70 bv->surface_height = 16 * 16; in batch_get_defaults()
71 bv->surface_width = 32 * 2 * 16; in batch_get_defaults()
76 bv->max_threads = 36; in batch_get_defaults()
79 bv in batch_get_defaults()
141 gen7_fill_surface_state(struct batch_chunk *state, const u32 dst_offset, const struct batch_vals *bv) gen7_fill_surface_state() argument
174 gen7_fill_binding_table(struct batch_chunk *state, const struct batch_vals *bv) gen7_fill_binding_table() argument
206 gen7_fill_interface_descriptor(struct batch_chunk *state, const struct batch_vals *bv, const struct cb_kernel *kernel, unsigned int count) gen7_fill_interface_descriptor() argument
260 gen7_emit_vfe_state(struct batch_chunk *batch, const struct batch_vals *bv, u32 urb_size, u32 curbe_size, u32 mode) gen7_emit_vfe_state() argument
375 emit_batch(struct i915_vma * const vma, u32 *start, const struct batch_vals *bv) emit_batch() argument
433 struct batch_vals bv; gen7_setup_clear_gpr_bb() local
[all...]
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/gt/
H A Dgen7_renderclear.c44 static int num_primitives(const struct batch_vals *bv) in num_primitives() argument
52 return bv->max_threads; in num_primitives()
56 batch_get_defaults(struct drm_i915_private *i915, struct batch_vals *bv) in batch_get_defaults() argument
62 bv->max_threads = 70; in batch_get_defaults()
65 bv->max_threads = 140; in batch_get_defaults()
68 bv->max_threads = 280; in batch_get_defaults()
71 bv->surface_height = 16 * 16; in batch_get_defaults()
72 bv->surface_width = 32 * 2 * 16; in batch_get_defaults()
77 bv->max_threads = 36; in batch_get_defaults()
80 bv in batch_get_defaults()
142 gen7_fill_surface_state(struct batch_chunk *state, const u32 dst_offset, const struct batch_vals *bv) gen7_fill_surface_state() argument
175 gen7_fill_binding_table(struct batch_chunk *state, const struct batch_vals *bv) gen7_fill_binding_table() argument
207 gen7_fill_interface_descriptor(struct batch_chunk *state, const struct batch_vals *bv, const struct cb_kernel *kernel, unsigned int count) gen7_fill_interface_descriptor() argument
261 gen7_emit_vfe_state(struct batch_chunk *batch, const struct batch_vals *bv, u32 urb_size, u32 curbe_size, u32 mode) gen7_emit_vfe_state() argument
376 emit_batch(struct i915_vma * const vma, u32 *start, const struct batch_vals *bv) emit_batch() argument
434 struct batch_vals bv; gen7_setup_clear_gpr_bb() local
[all...]
/kernel/linux/linux-5.10/include/linux/
H A Dbvec.h49 struct bio_vec bv; member
100 static inline bool bvec_iter_advance(const struct bio_vec *bv, in bvec_iter_advance() argument
114 while (bytes && bytes >= bv[idx].bv_len) { in bvec_iter_advance()
115 bytes -= bv[idx].bv_len; in bvec_iter_advance()
151 return &iter_all->bv; in bvec_init_iter_all()
157 struct bio_vec *bv = &iter_all->bv; in bvec_advance() local
160 bv->bv_page++; in bvec_advance()
161 bv->bv_offset = 0; in bvec_advance()
163 bv in bvec_advance()
[all...]
/kernel/linux/linux-5.10/drivers/md/bcache/
H A Dutil.c234 struct bio_vec *bv = bio->bi_io_vec; in bch_bio_map() local
239 bv->bv_offset = base ? offset_in_page(base) : 0; in bch_bio_map()
242 for (; size; bio->bi_vcnt++, bv++) { in bch_bio_map()
243 bv->bv_offset = 0; in bch_bio_map()
244 start: bv->bv_len = min_t(size_t, PAGE_SIZE - bv->bv_offset, in bch_bio_map()
247 bv->bv_page = is_vmalloc_addr(base) in bch_bio_map()
251 base += bv->bv_len; in bch_bio_map()
254 size -= bv->bv_len; in bch_bio_map()
271 struct bio_vec *bv; in bch_bio_alloc_pages() local
[all...]
/kernel/linux/linux-6.6/drivers/md/bcache/
H A Dutil.c234 struct bio_vec *bv = bio->bi_io_vec; in bch_bio_map() local
239 bv->bv_offset = base ? offset_in_page(base) : 0; in bch_bio_map()
242 for (; size; bio->bi_vcnt++, bv++) { in bch_bio_map()
243 bv->bv_offset = 0; in bch_bio_map()
244 start: bv->bv_len = min_t(size_t, PAGE_SIZE - bv->bv_offset, in bch_bio_map()
247 bv->bv_page = is_vmalloc_addr(base) in bch_bio_map()
251 base += bv->bv_len; in bch_bio_map()
254 size -= bv->bv_len; in bch_bio_map()
271 struct bio_vec *bv; in bch_bio_alloc_pages() local
[all...]
/kernel/linux/linux-6.6/fs/netfs/
H A Diterator.c40 struct bio_vec *bv = NULL; in netfs_extract_user_iter() local
54 bv_size = array_size(max_pages, sizeof(*bv)); in netfs_extract_user_iter()
55 bv = kvmalloc(bv_size, GFP_KERNEL); in netfs_extract_user_iter()
56 if (!bv) in netfs_extract_user_iter()
64 pages = (void *)bv + bv_size - pg_size; in netfs_extract_user_iter()
92 bvec_set_page(bv + npages + i, *pages++, len - offset, offset); in netfs_extract_user_iter()
100 iov_iter_bvec(new, orig->data_source, bv, npages, orig_len - count); in netfs_extract_user_iter()
/kernel/linux/linux-5.10/block/
H A Dbio.c154 void bvec_free(mempool_t *pool, struct bio_vec *bv, unsigned int idx) in bvec_free() argument
163 mempool_free(bv, pool); in bvec_free()
167 kmem_cache_free(bvs->slab, bv); in bvec_free()
535 struct bio_vec bv; in zero_fill_bio_iter() local
538 __bio_for_each_segment(bv, bio, iter, start) { in zero_fill_bio_iter()
539 char *data = bvec_kmap_irq(&bv, &flags); in zero_fill_bio_iter()
540 memset(data, 0, bv.bv_len); in zero_fill_bio_iter()
541 flush_dcache_page(bv.bv_page); in zero_fill_bio_iter()
559 struct bio_vec bv; in bio_truncate() local
570 bio_for_each_segment(bv, bi in bio_truncate()
738 page_is_mergeable(const struct bio_vec *bv, struct page *page, unsigned int len, unsigned int off, bool *same_page) page_is_mergeable() argument
766 struct bio_vec *bv = &bio->bi_io_vec[bio->bi_vcnt - 1]; bio_try_merge_hw_seg() local
878 struct bio_vec *bv = &bio->bi_io_vec[bio->bi_vcnt - 1]; __bio_try_merge_page() local
907 struct bio_vec *bv = &bio->bi_io_vec[bio->bi_vcnt]; __bio_add_page() local
966 const struct bio_vec *bv = iter->bvec; __bio_iov_bvec_add_pages() local
1006 struct bio_vec *bv = bio->bi_io_vec + bio->bi_vcnt; __bio_iov_iter_get_pages() local
1053 struct bio_vec *bv = bio->bi_io_vec + bio->bi_vcnt; __bio_iov_append_get_pages() local
[all...]
H A Dblk-merge.c180 * @q: [in] request queue associated with the bio associated with @bv
181 * @bv: [in] bvec to examine
183 * by the number of segments from @bv that may be appended to that
186 * by the number of sectors from @bv that may be appended to that
194 * %true is returned if and only if appending the entire @bv to a bio with
199 const struct bio_vec *bv, unsigned *nsegs, in bvec_split_segs()
204 unsigned len = min(bv->bv_len, max_len); in bvec_split_segs()
209 seg_size = get_max_segment_size(q, bv->bv_page, in bvec_split_segs()
210 bv->bv_offset + total_len); in bvec_split_segs()
217 if ((bv in bvec_split_segs()
198 bvec_split_segs(const struct request_queue *q, const struct bio_vec *bv, unsigned *nsegs, unsigned *sectors, unsigned max_segs, unsigned max_sectors) bvec_split_segs() argument
251 struct bio_vec bv, bvprv, *bvprvp = NULL; blk_bio_segment_split() local
373 struct bio_vec bv; blk_recalc_rq_segments() local
453 __blk_bvec_map_sg(struct bio_vec bv, struct scatterlist *sglist, struct scatterlist **sg) __blk_bvec_map_sg() argument
[all...]
H A Dblk-crypto-fallback.c164 struct bio_vec bv; in blk_crypto_clone_bio() local
177 bio_for_each_segment(bv, bio_src, iter) in blk_crypto_clone_bio()
178 bio->bi_io_vec[bio->bi_vcnt++] = bv; in blk_crypto_clone_bio()
214 struct bio_vec bv; in blk_crypto_split_bio_if_needed() local
217 bio_for_each_segment(bv, bio, iter) { in blk_crypto_split_bio_if_needed()
218 num_sectors += bv.bv_len >> SECTOR_SHIFT; in blk_crypto_split_bio_if_needed()
385 struct bio_vec bv; in blk_crypto_fallback_decrypt_bio() local
413 __bio_for_each_segment(bv, bio, iter, f_ctx->crypt_iter) { in blk_crypto_fallback_decrypt_bio()
414 struct page *page = bv.bv_page; in blk_crypto_fallback_decrypt_bio()
416 sg_set_page(&sg, page, data_unit_size, bv in blk_crypto_fallback_decrypt_bio()
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/netronome/nfp/bpf/
H A Dmain.c66 struct nfp_bpf_vnic *bv; in nfp_bpf_vnic_alloc() local
79 bv = kzalloc(sizeof(*bv), GFP_KERNEL); in nfp_bpf_vnic_alloc()
80 if (!bv) in nfp_bpf_vnic_alloc()
82 nn->app_priv = bv; in nfp_bpf_vnic_alloc()
88 bv->start_off = nn_readw(nn, NFP_NET_CFG_BPF_START); in nfp_bpf_vnic_alloc()
89 bv->tgt_done = nn_readw(nn, NFP_NET_CFG_BPF_DONE); in nfp_bpf_vnic_alloc()
99 struct nfp_bpf_vnic *bv = nn->app_priv; in nfp_bpf_vnic_free() local
101 WARN_ON(bv->tc_prog); in nfp_bpf_vnic_free()
102 kfree(bv); in nfp_bpf_vnic_free()
111 struct nfp_bpf_vnic *bv; nfp_bpf_setup_tc_block_cb() local
185 struct nfp_bpf_vnic *bv; nfp_bpf_check_mtu() local
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/netronome/nfp/bpf/
H A Dmain.c66 struct nfp_bpf_vnic *bv; in nfp_bpf_vnic_alloc() local
79 bv = kzalloc(sizeof(*bv), GFP_KERNEL); in nfp_bpf_vnic_alloc()
80 if (!bv) in nfp_bpf_vnic_alloc()
82 nn->app_priv = bv; in nfp_bpf_vnic_alloc()
88 bv->start_off = nn_readw(nn, NFP_NET_CFG_BPF_START); in nfp_bpf_vnic_alloc()
89 bv->tgt_done = nn_readw(nn, NFP_NET_CFG_BPF_DONE); in nfp_bpf_vnic_alloc()
99 struct nfp_bpf_vnic *bv = nn->app_priv; in nfp_bpf_vnic_free() local
101 WARN_ON(bv->tc_prog); in nfp_bpf_vnic_free()
102 kfree(bv); in nfp_bpf_vnic_free()
111 struct nfp_bpf_vnic *bv; nfp_bpf_setup_tc_block_cb() local
185 struct nfp_bpf_vnic *bv; nfp_bpf_check_mtu() local
[all...]
/kernel/linux/linux-6.6/block/
H A Dblk-merge.c21 static inline void bio_get_first_bvec(struct bio *bio, struct bio_vec *bv) in bio_get_first_bvec() argument
23 *bv = mp_bvec_iter_bvec(bio->bi_io_vec, bio->bi_iter); in bio_get_first_bvec()
26 static inline void bio_get_last_bvec(struct bio *bio, struct bio_vec *bv) in bio_get_last_bvec() argument
31 bio_get_first_bvec(bio, bv); in bio_get_last_bvec()
32 if (bv->bv_len == bio->bi_iter.bi_size) in bio_get_last_bvec()
42 *bv = bio->bi_io_vec[idx]; in bio_get_last_bvec()
49 bv->bv_len = iter.bi_bvec_done; in bio_get_last_bvec()
214 * @bv: [in] bvec to examine
216 * by the number of segments from @bv that may be appended to that
219 * by the number of bytes from @bv tha
231 bvec_split_segs(const struct queue_limits *lim, const struct bio_vec *bv, unsigned *nsegs, unsigned *bytes, unsigned max_segs, unsigned max_bytes) bvec_split_segs() argument
282 struct bio_vec bv, bvprv, *bvprvp = NULL; bio_split_rw() local
417 struct bio_vec bv; blk_recalc_rq_segments() local
497 __blk_bvec_map_sg(struct bio_vec bv, struct scatterlist *sglist, struct scatterlist **sg) __blk_bvec_map_sg() argument
[all...]
H A Dbio.c163 void bvec_free(mempool_t *pool, struct bio_vec *bv, unsigned short nr_vecs) in bvec_free() argument
168 mempool_free(bv, pool); in bvec_free()
170 kmem_cache_free(biovec_slab(nr_vecs)->slab, bv); in bvec_free()
611 struct bio_vec bv; in zero_fill_bio_iter() local
614 __bio_for_each_segment(bv, bio, iter, start) in zero_fill_bio_iter()
615 memzero_bvec(&bv); in zero_fill_bio_iter()
631 struct bio_vec bv; in bio_truncate() local
642 bio_for_each_segment(bv, bio, iter) { in bio_truncate()
643 if (done + bv.bv_len > new_size) { in bio_truncate()
650 zero_user(bv in bio_truncate()
906 bvec_try_merge_page(struct bio_vec *bv, struct page *page, unsigned int len, unsigned int off, bool *same_page) bvec_try_merge_page() argument
937 bvec_try_merge_hw_page(struct request_queue *q, struct bio_vec *bv, struct page *page, unsigned len, unsigned offset, bool *same_page) bvec_try_merge_hw_page() argument
976 struct bio_vec *bv = &bio->bi_io_vec[bio->bi_vcnt - 1]; bio_add_hw_page() local
1240 struct bio_vec *bv = bio->bi_io_vec + bio->bi_vcnt; __bio_iov_iter_get_pages() local
[all...]
H A Dblk-crypto-fallback.c164 struct bio_vec bv; in blk_crypto_fallback_clone_bio() local
178 bio_for_each_segment(bv, bio_src, iter) in blk_crypto_fallback_clone_bio()
179 bio->bi_io_vec[bio->bi_vcnt++] = bv; in blk_crypto_fallback_clone_bio()
215 struct bio_vec bv; in blk_crypto_fallback_split_bio_if_needed() local
218 bio_for_each_segment(bv, bio, iter) { in blk_crypto_fallback_split_bio_if_needed()
219 num_sectors += bv.bv_len >> SECTOR_SHIFT; in blk_crypto_fallback_split_bio_if_needed()
388 struct bio_vec bv; in blk_crypto_fallback_decrypt_bio() local
417 __bio_for_each_segment(bv, bio, iter, f_ctx->crypt_iter) { in blk_crypto_fallback_decrypt_bio()
418 struct page *page = bv.bv_page; in blk_crypto_fallback_decrypt_bio()
420 sg_set_page(&sg, page, data_unit_size, bv in blk_crypto_fallback_decrypt_bio()
[all...]
/kernel/linux/linux-5.10/arch/parisc/kernel/
H A Dentry.S169 bv,n 0(%r3)
791 bv %r0(%r2)
800 bv %r0(%r2)
960 bv %r0(%r20)
989 bv %r0(%r1)
1774 bv %r0(%r2)
1971 bv %r0(%r19) /* jumps to schedule() */
2002 bv %r0(%rp)
2070 bv,n (%r1)
2166 bv,
[all...]
H A Dreal2.S92 bv 0(%r31)
114 bv 0(%rp)
134 bv 0(%r2)
149 bv 0(%r2)
190 bv 0(%r2)
227 bv 0(%r2)
273 bv 0(%r31)
286 bv 0(%rp)
301 bv %r0(%r2)
/kernel/linux/linux-6.6/arch/parisc/kernel/
H A Dentry.S161 bv,n 0(%r3)
772 bv %r0(%r2)
781 bv %r0(%r2)
940 bv %r0(%r20)
969 bv %r0(%r1)
1701 bv %r0(%r2)
1897 bv %r0(%r19) /* jumps to schedule() */
1928 bv %r0(%rp)
1996 bv,n (%r1)
2092 bv,
[all...]
H A Dreal2.S79 bv 0(%r31)
101 bv 0(%rp)
121 bv 0(%r2)
136 bv 0(%r2)
177 bv 0(%r2)
214 bv 0(%r2)
260 bv 0(%r31)
273 bv 0(%rp)
288 bv %r0(%r2)
/kernel/linux/linux-6.6/drivers/block/
H A Dn64cart.c62 static bool n64cart_do_bvec(struct device *dev, struct bio_vec *bv, u32 pos) in n64cart_do_bvec() argument
68 WARN_ON_ONCE((bv->bv_offset & (MIN_ALIGNMENT - 1)) || in n64cart_do_bvec()
69 (bv->bv_len & (MIN_ALIGNMENT - 1))); in n64cart_do_bvec()
71 dma_addr = dma_map_bvec(dev, bv, DMA_FROM_DEVICE, 0); in n64cart_do_bvec()
79 n64cart_write_reg(PI_WRITE_REG, bv->bv_len - 1); in n64cart_do_bvec()
83 dma_unmap_page(dev, dma_addr, bv->bv_len, DMA_FROM_DEVICE); in n64cart_do_bvec()
/kernel/linux/linux-5.10/drivers/s390/block/
H A Ddasd_fba.c449 struct bio_vec bv; in dasd_fba_build_cp_regular() local
470 rq_for_each_segment(bv, req, iter) { in dasd_fba_build_cp_regular()
471 if (bv.bv_len & (blksize - 1)) in dasd_fba_build_cp_regular()
474 count += bv.bv_len >> (block->s2b_shift + 9); in dasd_fba_build_cp_regular()
475 if (idal_is_needed (page_address(bv.bv_page), bv.bv_len)) in dasd_fba_build_cp_regular()
476 cidaw += bv.bv_len / blksize; in dasd_fba_build_cp_regular()
512 rq_for_each_segment(bv, req, iter) { in dasd_fba_build_cp_regular()
513 dst = page_address(bv.bv_page) + bv in dasd_fba_build_cp_regular()
581 struct bio_vec bv; dasd_fba_free_cp() local
[all...]
/kernel/linux/linux-5.10/fs/cifs/
H A Dmisc.c791 * Must use kzalloc to initialize ctx->bv to NULL and ctx->direct_io in cifs_aio_ctx_alloc()
815 * ctx->bv is only set if setup_aio_ctx_iter() was call successfuly in cifs_aio_ctx_release()
819 if (ctx->bv) { in cifs_aio_ctx_release()
824 set_page_dirty(ctx->bv[i].bv_page); in cifs_aio_ctx_release()
825 put_page(ctx->bv[i].bv_page); in cifs_aio_ctx_release()
827 kvfree(ctx->bv); in cifs_aio_ctx_release()
848 struct bio_vec *bv = NULL; in setup_aio_ctx_iter() local
857 if (array_size(max_pages, sizeof(*bv)) <= CIFS_AIO_KMALLOC_LIMIT) in setup_aio_ctx_iter()
858 bv = kmalloc_array(max_pages, sizeof(*bv), GFP_KERNE in setup_aio_ctx_iter()
[all...]
/kernel/linux/linux-6.6/drivers/s390/block/
H A Ddasd_fba.c440 struct bio_vec bv; in dasd_fba_build_cp_regular() local
461 rq_for_each_segment(bv, req, iter) { in dasd_fba_build_cp_regular()
462 if (bv.bv_len & (blksize - 1)) in dasd_fba_build_cp_regular()
465 count += bv.bv_len >> (block->s2b_shift + 9); in dasd_fba_build_cp_regular()
466 if (idal_is_needed (page_address(bv.bv_page), bv.bv_len)) in dasd_fba_build_cp_regular()
467 cidaw += bv.bv_len / blksize; in dasd_fba_build_cp_regular()
503 rq_for_each_segment(bv, req, iter) { in dasd_fba_build_cp_regular()
504 dst = bvec_virt(&bv); in dasd_fba_build_cp_regular()
509 memcpy(copy + bv in dasd_fba_build_cp_regular()
572 struct bio_vec bv; dasd_fba_free_cp() local
[all...]
/kernel/linux/linux-5.10/fs/crypto/
H A Dbio.c31 struct bio_vec *bv; in fscrypt_decrypt_bio() local
34 bio_for_each_segment_all(bv, bio, iter_all) { in fscrypt_decrypt_bio()
35 struct page *page = bv->bv_page; in fscrypt_decrypt_bio()
36 int ret = fscrypt_decrypt_pagecache_blocks(page, bv->bv_len, in fscrypt_decrypt_bio()
37 bv->bv_offset); in fscrypt_decrypt_bio()
/kernel/linux/linux-5.10/drivers/md/
H A Ddm-ebs-target.c64 static int __ebs_rw_bvec(struct ebs_c *ec, int rw, struct bio_vec *bv, struct bvec_iter *iter) in __ebs_rw_bvec() argument
69 unsigned int bv_len = bv->bv_len; in __ebs_rw_bvec()
74 if (unlikely(!bv->bv_page || !bv_len)) in __ebs_rw_bvec()
77 pa = page_address(bv->bv_page) + bv->bv_offset; in __ebs_rw_bvec()
100 flush_dcache_page(bv->bv_page); in __ebs_rw_bvec()
102 flush_dcache_page(bv->bv_page); in __ebs_rw_bvec()
123 struct bio_vec bv; in __ebs_rw_bio() local
126 bio_for_each_bvec(bv, bio, iter) { in __ebs_rw_bio()
127 rr = __ebs_rw_bvec(ec, rw, &bv, in __ebs_rw_bio()
[all...]

Completed in 19 milliseconds

12345678