/kernel/linux/linux-5.10/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_page_dirty.c | 77 * @vbo: The buffer object to scan 83 static void vmw_bo_dirty_scan_pagetable(struct vmw_buffer_object *vbo) in vmw_bo_dirty_scan_pagetable() argument 85 struct vmw_bo_dirty *dirty = vbo->dirty; in vmw_bo_dirty_scan_pagetable() 86 pgoff_t offset = drm_vma_node_start(&vbo->base.base.vma_node); in vmw_bo_dirty_scan_pagetable() 87 struct address_space *mapping = vbo->base.bdev->dev_mapping; in vmw_bo_dirty_scan_pagetable() 114 * @vbo: The buffer object to scan 121 static void vmw_bo_dirty_scan_mkwrite(struct vmw_buffer_object *vbo) in vmw_bo_dirty_scan_mkwrite() argument 123 struct vmw_bo_dirty *dirty = vbo->dirty; in vmw_bo_dirty_scan_mkwrite() 124 unsigned long offset = drm_vma_node_start(&vbo->base.base.vma_node); in vmw_bo_dirty_scan_mkwrite() 125 struct address_space *mapping = vbo in vmw_bo_dirty_scan_mkwrite() 165 vmw_bo_dirty_scan(struct vmw_buffer_object *vbo) vmw_bo_dirty_scan() argument 186 vmw_bo_dirty_pre_unmap(struct vmw_buffer_object *vbo, pgoff_t start, pgoff_t end) vmw_bo_dirty_pre_unmap() argument 211 vmw_bo_dirty_unmap(struct vmw_buffer_object *vbo, pgoff_t start, pgoff_t end) vmw_bo_dirty_unmap() argument 232 vmw_bo_dirty_add(struct vmw_buffer_object *vbo) vmw_bo_dirty_add() argument 302 vmw_bo_dirty_release(struct vmw_buffer_object *vbo) vmw_bo_dirty_release() argument 327 struct vmw_buffer_object *vbo = res->backup; vmw_bo_dirty_transfer_to_res() local 374 struct vmw_buffer_object *vbo = res->backup; vmw_bo_dirty_clear_res() local 401 struct vmw_buffer_object *vbo = vmw_bo_vm_mkwrite() local 440 struct vmw_buffer_object *vbo = vmw_bo_vm_fault() local 497 struct vmw_buffer_object *vbo = vmw_bo_vm_huge_fault() local [all...] |
H A D | vmwgfx_bo.c | 39 * @vbo: The struct vmw_buffer_object 43 struct vmw_buffer_object vbo; member 75 return container_of(vmw_bo, struct vmw_user_buffer_object, vbo); in vmw_user_buffer_object() 331 * @vbo: The buffer object. Must be reserved. 335 void vmw_bo_pin_reserved(struct vmw_buffer_object *vbo, bool pin) in vmw_bo_pin_reserved() argument 340 struct ttm_buffer_object *bo = &vbo->base; in vmw_bo_pin_reserved() 347 if (vbo->pin_count++ > 0) in vmw_bo_pin_reserved() 350 WARN_ON(vbo->pin_count <= 0); in vmw_bo_pin_reserved() 351 if (--vbo->pin_count > 0) in vmw_bo_pin_reserved() 377 * @vbo 389 vmw_bo_map_and_cache(struct vmw_buffer_object *vbo) vmw_bo_map_and_cache() argument 416 vmw_bo_unmap(struct vmw_buffer_object *vbo) vmw_bo_unmap() argument 482 struct vmw_buffer_object *vbo = &vmw_user_bo->vbo; vmw_user_bo_destroy() local 766 struct vmw_buffer_object *vbo; vmw_user_bo_synccpu_ioctl() local 837 struct vmw_buffer_object *vbo; vmw_bo_alloc_ioctl() local 985 vmw_user_bo_reference(struct ttm_object_file *tfile, struct vmw_buffer_object *vbo, uint32_t *handle) vmw_user_bo_reference() argument 1048 struct vmw_buffer_object *vbo; vmw_dumb_create() local 1149 struct vmw_buffer_object *vbo; vmw_bo_move_notify() local [all...] |
H A D | vmwgfx_validation.c | 171 * @vbo: The buffer object to search for. 178 struct vmw_buffer_object *vbo) in vmw_validation_find_bo_dup() 188 if (!drm_ht_find_item(ctx->ht, (unsigned long) vbo, &hash)) in vmw_validation_find_bo_dup() 194 if (entry->base.bo == &vbo->base) { in vmw_validation_find_bo_dup() 208 * @vbo: The buffer object to search for. 252 * @vbo: The buffer object. 259 struct vmw_buffer_object *vbo, in vmw_validation_add_bo() 265 bo_node = vmw_validation_find_bo_dup(ctx, vbo); in vmw_validation_add_bo() 281 bo_node->hash.key = (unsigned long) vbo; in vmw_validation_add_bo() 290 val_buf->bo = ttm_bo_get_unless_zero(&vbo in vmw_validation_add_bo() 177 vmw_validation_find_bo_dup(struct vmw_validation_context *ctx, struct vmw_buffer_object *vbo) vmw_validation_find_bo_dup() argument 258 vmw_validation_add_bo(struct vmw_validation_context *ctx, struct vmw_buffer_object *vbo, bool as_mob, bool cpu_blit) vmw_validation_add_bo() argument 414 vmw_validation_res_switch_backup(struct vmw_validation_context *ctx, void *val_private, struct vmw_buffer_object *vbo, unsigned long backup_offset) vmw_validation_res_switch_backup() argument 457 struct vmw_buffer_object *vbo = res->backup; vmw_validation_res_reserve() local 532 struct vmw_buffer_object *vbo = vmw_validation_bo_validate_single() local 584 struct vmw_buffer_object *vbo = vmw_validation_bo_validate() local 657 struct vmw_buffer_object *vbo = res->backup; vmw_validation_res_validate() local 895 struct vmw_buffer_object *vbo = vmw_validation_bo_backoff() local [all...] |
H A D | vmwgfx_resource.c | 772 * @vbo: Pointer to the current backing MOB. 780 void vmw_resource_unbind_list(struct vmw_buffer_object *vbo) in vmw_resource_unbind_list() argument 783 .bo = &vbo->base, in vmw_resource_unbind_list() 787 dma_resv_assert_held(vbo->base.base.resv); in vmw_resource_unbind_list() 788 while (!RB_EMPTY_ROOT(&vbo->res_tree)) { in vmw_resource_unbind_list() 789 struct rb_node *node = vbo->res_tree.rb_node; in vmw_resource_unbind_list() 801 (void) ttm_bo_wait(&vbo->base, false, false); in vmw_resource_unbind_list() 999 struct vmw_buffer_object *vbo = NULL; in vmw_resource_pin() local 1002 vbo = res->backup; in vmw_resource_pin() 1004 ttm_bo_reserve(&vbo in vmw_resource_pin() 1057 struct vmw_buffer_object *vbo = res->backup; vmw_resource_unpin() local 1103 vmw_resources_clean(struct vmw_buffer_object *vbo, pgoff_t start, pgoff_t end, pgoff_t *num_prefault) vmw_resources_clean() argument [all...] |
H A D | vmwgfx_drv.h | 797 extern void vmw_resource_unbind_list(struct vmw_buffer_object *vbo); 802 int vmw_resources_clean(struct vmw_buffer_object *vbo, pgoff_t start, 876 extern void *vmw_bo_map_and_cache(struct vmw_buffer_object *vbo); 877 extern void vmw_bo_unmap(struct vmw_buffer_object *vbo); 896 * @vbo: The struct vmw_buffer_object 898 static inline void vmw_bo_prio_adjust(struct vmw_buffer_object *vbo) in vmw_bo_prio_adjust() argument 900 int i = ARRAY_SIZE(vbo->res_prios); in vmw_bo_prio_adjust() 903 if (vbo->res_prios[i]) { in vmw_bo_prio_adjust() 904 vbo->base.priority = i; in vmw_bo_prio_adjust() 909 vbo in vmw_bo_prio_adjust() 921 vmw_bo_prio_add(struct vmw_buffer_object *vbo, int prio) vmw_bo_prio_add() argument 936 vmw_bo_prio_del(struct vmw_buffer_object *vbo, int prio) vmw_bo_prio_del() argument [all...] |
H A D | vmwgfx_drv.c | 360 struct vmw_buffer_object *vbo; in vmw_dummy_query_bo_create() local 366 * Create the vbo as pinned, so that a tryreserve will in vmw_dummy_query_bo_create() 370 vbo = kzalloc(sizeof(*vbo), GFP_KERNEL); in vmw_dummy_query_bo_create() 371 if (!vbo) in vmw_dummy_query_bo_create() 374 ret = vmw_bo_init(dev_priv, vbo, PAGE_SIZE, in vmw_dummy_query_bo_create() 380 ret = ttm_bo_reserve(&vbo->base, false, true, NULL); in vmw_dummy_query_bo_create() 382 vmw_bo_pin_reserved(vbo, true); in vmw_dummy_query_bo_create() 384 ret = ttm_bo_kmap(&vbo->base, 0, 1, &map); in vmw_dummy_query_bo_create() 392 vmw_bo_pin_reserved(vbo, fals in vmw_dummy_query_bo_create() [all...] |
H A D | vmwgfx_validation.h | 221 struct vmw_buffer_object *vbo, 241 struct vmw_buffer_object *vbo,
|
/kernel/linux/linux-6.6/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_page_dirty.c | 76 * @vbo: The buffer object to scan 82 static void vmw_bo_dirty_scan_pagetable(struct vmw_bo *vbo) in vmw_bo_dirty_scan_pagetable() argument 84 struct vmw_bo_dirty *dirty = vbo->dirty; in vmw_bo_dirty_scan_pagetable() 85 pgoff_t offset = drm_vma_node_start(&vbo->tbo.base.vma_node); in vmw_bo_dirty_scan_pagetable() 86 struct address_space *mapping = vbo->tbo.bdev->dev_mapping; in vmw_bo_dirty_scan_pagetable() 113 * @vbo: The buffer object to scan 120 static void vmw_bo_dirty_scan_mkwrite(struct vmw_bo *vbo) in vmw_bo_dirty_scan_mkwrite() argument 122 struct vmw_bo_dirty *dirty = vbo->dirty; in vmw_bo_dirty_scan_mkwrite() 123 unsigned long offset = drm_vma_node_start(&vbo->tbo.base.vma_node); in vmw_bo_dirty_scan_mkwrite() 124 struct address_space *mapping = vbo in vmw_bo_dirty_scan_mkwrite() 163 vmw_bo_dirty_scan(struct vmw_bo *vbo) vmw_bo_dirty_scan() argument 184 vmw_bo_dirty_pre_unmap(struct vmw_bo *vbo, pgoff_t start, pgoff_t end) vmw_bo_dirty_pre_unmap() argument 209 vmw_bo_dirty_unmap(struct vmw_bo *vbo, pgoff_t start, pgoff_t end) vmw_bo_dirty_unmap() argument 230 vmw_bo_dirty_add(struct vmw_bo *vbo) vmw_bo_dirty_add() argument 287 vmw_bo_dirty_release(struct vmw_bo *vbo) vmw_bo_dirty_release() argument 309 struct vmw_bo *vbo = res->guest_memory_bo; vmw_bo_dirty_transfer_to_res() local 356 struct vmw_bo *vbo = res->guest_memory_bo; vmw_bo_dirty_clear_res() local 383 struct vmw_bo *vbo = to_vmw_bo(&bo->base); vmw_bo_vm_mkwrite() local 421 struct vmw_bo *vbo = to_vmw_bo(&bo->base); vmw_bo_vm_fault() local [all...] |
H A D | vmwgfx_bo.h | 124 void *vmw_bo_map_and_cache(struct vmw_bo *vbo); 125 void vmw_bo_unmap(struct vmw_bo *vbo); 137 * @vbo: The struct vmw_bo 139 static inline void vmw_bo_prio_adjust(struct vmw_bo *vbo) in vmw_bo_prio_adjust() argument 141 int i = ARRAY_SIZE(vbo->res_prios); in vmw_bo_prio_adjust() 144 if (vbo->res_prios[i]) { in vmw_bo_prio_adjust() 145 vbo->tbo.priority = i; in vmw_bo_prio_adjust() 150 vbo->tbo.priority = 3; in vmw_bo_prio_adjust() 156 * @vbo: The struct vmw_bo 162 static inline void vmw_bo_prio_add(struct vmw_bo *vbo, in argument 177 vmw_bo_prio_del(struct vmw_bo *vbo, int prio) vmw_bo_prio_del() argument 198 vmw_user_bo_ref(struct vmw_bo *vbo) vmw_user_bo_ref() argument [all...] |
H A D | vmwgfx_bo.c | 35 static void vmw_bo_release(struct vmw_bo *vbo) in vmw_bo_release() argument 37 WARN_ON(vbo->tbo.base.funcs && in vmw_bo_release() 38 kref_read(&vbo->tbo.base.refcount) != 0); in vmw_bo_release() 39 vmw_bo_unmap(vbo); in vmw_bo_release() 40 drm_gem_object_release(&vbo->tbo.base); in vmw_bo_release() 50 struct vmw_bo *vbo = to_vmw_bo(&bo->base); in vmw_bo_free() local 52 WARN_ON(vbo->dirty); in vmw_bo_free() 53 WARN_ON(!RB_EMPTY_ROOT(&vbo->res_tree)); in vmw_bo_free() 54 vmw_bo_release(vbo); in vmw_bo_free() 55 kfree(vbo); in vmw_bo_free() 278 vmw_bo_pin_reserved(struct vmw_bo *vbo, bool pin) vmw_bo_pin_reserved() argument 326 vmw_bo_map_and_cache(struct vmw_bo *vbo) vmw_bo_map_and_cache() argument 353 vmw_bo_unmap(struct vmw_bo *vbo) vmw_bo_unmap() argument 526 struct vmw_bo *vbo; vmw_user_bo_synccpu_ioctl() local 673 struct vmw_bo *vbo; vmw_dumb_create() local 726 struct vmw_bo *vbo = to_vmw_bo(&bo->base); vmw_bo_move_notify() local [all...] |
H A D | vmwgfx_validation.c | 168 * @vbo: The buffer object to search for. 175 struct vmw_bo *vbo) in vmw_validation_find_bo_dup() 184 unsigned long key = (unsigned long) vbo; in vmw_validation_find_bo_dup() 196 if (entry->base.bo == &vbo->tbo) { in vmw_validation_find_bo_dup() 259 * @vbo: The buffer object. 264 struct vmw_bo *vbo) in vmw_validation_add_bo() 268 bo_node = vmw_validation_find_bo_dup(ctx, vbo); in vmw_validation_add_bo() 277 bo_node->hash.key = (unsigned long) vbo; in vmw_validation_add_bo() 282 val_buf->bo = ttm_bo_get_unless_zero(&vbo->tbo); in vmw_validation_add_bo() 394 * @vbo 174 vmw_validation_find_bo_dup(struct vmw_validation_context *ctx, struct vmw_bo *vbo) vmw_validation_find_bo_dup() argument 263 vmw_validation_add_bo(struct vmw_validation_context *ctx, struct vmw_bo *vbo) vmw_validation_add_bo() argument 398 vmw_validation_res_switch_backup(struct vmw_validation_context *ctx, void *val_private, struct vmw_bo *vbo, unsigned long guest_memory_offset) vmw_validation_res_switch_backup() argument 441 struct vmw_bo *vbo = res->guest_memory_bo; vmw_validation_res_reserve() local 515 struct vmw_bo *vbo = to_vmw_bo(&bo->base); vmw_validation_bo_validate_single() local 556 struct vmw_bo *vbo = to_vmw_bo(&entry->base.bo->base); vmw_validation_bo_validate() local 618 struct vmw_bo *vbo = res->guest_memory_bo; vmw_validation_res_validate() local 856 struct vmw_bo *vbo = to_vmw_bo(&entry->base.bo->base); vmw_validation_bo_backoff() local [all...] |
H A D | vmwgfx_resource.c | 743 * @vbo: Pointer to the current backing MOB. 751 void vmw_resource_unbind_list(struct vmw_bo *vbo) in vmw_resource_unbind_list() argument 754 .bo = &vbo->tbo, in vmw_resource_unbind_list() 758 dma_resv_assert_held(vbo->tbo.base.resv); in vmw_resource_unbind_list() 759 while (!RB_EMPTY_ROOT(&vbo->res_tree)) { in vmw_resource_unbind_list() 760 struct rb_node *node = vbo->res_tree.rb_node; in vmw_resource_unbind_list() 772 (void) ttm_bo_wait(&vbo->tbo, false, false); in vmw_resource_unbind_list() 969 struct vmw_bo *vbo = NULL; in vmw_resource_pin() local 972 vbo = res->guest_memory_bo; in vmw_resource_pin() 974 ret = ttm_bo_reserve(&vbo in vmw_resource_pin() 1030 struct vmw_bo *vbo = res->guest_memory_bo; vmw_resource_unpin() local 1075 vmw_resources_clean(struct vmw_bo *vbo, pgoff_t start, pgoff_t end, pgoff_t *num_prefault) vmw_resources_clean() argument [all...] |
H A D | vmwgfx_gem.c | 54 struct vmw_bo *vbo = to_vmw_bo(obj); in vmw_gem_pin_private() local 61 vmw_bo_pin_reserved(vbo, do_pin); in vmw_gem_pin_private() 161 struct vmw_bo *vbo; in vmw_gem_object_create_ioctl() local 166 req->size, &handle, &vbo); in vmw_gem_object_create_ioctl() 171 rep->map_handle = drm_vma_node_offset_addr(&vbo->tbo.base.vma_node); in vmw_gem_object_create_ioctl() 175 drm_gem_object_put(&vbo->tbo.base); in vmw_gem_object_create_ioctl()
|
H A D | vmwgfx_ttm_buffer.c | 584 struct vmw_bo *vbo; in vmw_bo_create_and_populate() local 594 ret = vmw_bo_create(dev_priv, &bo_params, &vbo); in vmw_bo_create_and_populate() 598 ret = ttm_bo_reserve(&vbo->tbo, false, true, NULL); in vmw_bo_create_and_populate() 600 ret = vmw_ttm_populate(vbo->tbo.bdev, vbo->tbo.ttm, &ctx); in vmw_bo_create_and_populate() 603 container_of(vbo->tbo.ttm, struct vmw_ttm_tt, dma_ttm); in vmw_bo_create_and_populate() 607 ttm_bo_unreserve(&vbo->tbo); in vmw_bo_create_and_populate() 610 *bo_p = vbo; in vmw_bo_create_and_populate()
|
H A D | vmwgfx_validation.h | 162 struct vmw_bo *vbo); 178 struct vmw_bo *vbo,
|
/kernel/linux/linux-6.6/fs/ntfs3/ |
H A D | attrib.c | 1228 u64 vbo; in attr_data_read_resident() local 1239 vbo = page->index << PAGE_SHIFT; in attr_data_read_resident() 1241 if (vbo < data_size) { in attr_data_read_resident() 1244 u32 use = data_size - vbo; in attr_data_read_resident() 1249 memcpy(kaddr, data + vbo, use); in attr_data_read_resident() 1264 u64 vbo; in attr_data_write_resident() local 1278 vbo = page->index << PAGE_SHIFT; in attr_data_write_resident() 1280 if (vbo < data_size) { in attr_data_write_resident() 1283 u32 use = data_size - vbo; in attr_data_write_resident() 1287 memcpy(data + vbo, kadd in attr_data_write_resident() 1377 u64 vbo[2], off[2], wof_size; attr_wof_frame_info() local 1847 attr_collapse_range(struct ntfs_inode *ni, u64 vbo, u64 bytes) attr_collapse_range() argument 2105 attr_punch_hole(struct ntfs_inode *ni, u64 vbo, u64 bytes, u32 *frame_size) attr_punch_hole() argument 2323 attr_insert_range(struct ntfs_inode *ni, u64 vbo, u64 bytes) attr_insert_range() argument [all...] |
H A D | dir.c | 345 const struct INDEX_HDR *hdr, u64 vbo, u64 pos, in ntfs_read_hdr() 367 if (vbo + off < pos) in ntfs_read_hdr() 373 ctx->pos = vbo + off; in ntfs_read_hdr() 392 u64 vbo; in ntfs_readdir() local 460 vbo = (u64)bit << index_bits; in ntfs_readdir() 461 if (vbo >= i_size) { in ntfs_readdir() 475 vbo = (u64)bit << index_bits; in ntfs_readdir() 476 if (vbo >= i_size) { in ntfs_readdir() 488 vbo + sbi->record_size, pos, name, ctx); in ntfs_readdir() 344 ntfs_read_hdr(struct ntfs_sb_info *sbi, struct ntfs_inode *ni, const struct INDEX_HDR *hdr, u64 vbo, u64 pos, u8 *name, struct dir_context *ctx) ntfs_read_hdr() argument
|
H A D | fsntfs.c | 774 u64 vbo; in ntfs_clear_mft_tail() local 786 vbo = (u64)from * rs; in ntfs_clear_mft_tail() 787 for (; from < to; from++, vbo += rs) { in ntfs_clear_mft_tail() 790 err = ntfs_get_bh(sbi, run, vbo, rs, &nb); in ntfs_clear_mft_tail() 1128 u64 vbo, const void *buf, size_t bytes, int sync) in ntfs_sb_write_run() 1132 u32 off = vbo & sbi->cluster_mask; in ntfs_sb_write_run() 1133 CLST lcn, clen, vcn = vbo >> cluster_bits, vcn_next; in ntfs_sb_write_run() 1176 const struct runs_tree *run, u64 vbo) in ntfs_bread_run() 1183 if (!run_lookup_entry(run, vbo >> cluster_bits, &lcn, NULL, NULL)) in ntfs_bread_run() 1186 lbo = ((u64)lcn << cluster_bits) + (vbo in ntfs_bread_run() 1127 ntfs_sb_write_run(struct ntfs_sb_info *sbi, const struct runs_tree *run, u64 vbo, const void *buf, size_t bytes, int sync) ntfs_sb_write_run() argument 1175 ntfs_bread_run(struct ntfs_sb_info *sbi, const struct runs_tree *run, u64 vbo) ntfs_bread_run() argument 1191 ntfs_read_run_nb(struct ntfs_sb_info *sbi, const struct runs_tree *run, u64 vbo, void *buf, u32 bytes, struct ntfs_buffers *nb) ntfs_read_run_nb() argument 1309 ntfs_read_bh(struct ntfs_sb_info *sbi, const struct runs_tree *run, u64 vbo, struct NTFS_RECORD_HEADER *rhdr, u32 bytes, struct ntfs_buffers *nb) ntfs_read_bh() argument 1320 ntfs_get_bh(struct ntfs_sb_info *sbi, const struct runs_tree *run, u64 vbo, u32 bytes, struct ntfs_buffers *nb) ntfs_get_bh() argument 1497 ntfs_bio_pages(struct ntfs_sb_info *sbi, const struct runs_tree *run, struct page **pages, u32 nr_pages, u64 vbo, u32 bytes, enum req_op op) ntfs_bio_pages() argument 1666 ntfs_vbo_to_lbo(struct ntfs_sb_info *sbi, const struct runs_tree *run, u64 vbo, u64 *lbo, u64 *bytes) ntfs_vbo_to_lbo() argument [all...] |
H A D | file.c | 176 * It zeroes a range [vbo, vbo_to). 178 static int ntfs_zero_range(struct inode *inode, u64 vbo, u64 vbo_to) in ntfs_zero_range() argument 183 pgoff_t idx = vbo >> PAGE_SHIFT; in ntfs_zero_range() 184 u32 from = vbo & (PAGE_SIZE - 1); in ntfs_zero_range() 428 * space requests 'len' bytes at 'vbo'. If FALLOC_FL_KEEP_SIZE is set 432 static long ntfs_fallocate(struct file *file, int mode, loff_t vbo, loff_t len) in ntfs_fallocate() argument 439 loff_t end = vbo + len; in ntfs_fallocate() 440 loff_t vbo_down = round_down(vbo, max_t(unsigned long, in ntfs_fallocate() 502 err = attr_punch_hole(ni, vbo, len, &frame_size); in ntfs_fallocate() 513 vbo_a = (vbo in ntfs_fallocate() [all...] |
H A D | fslog.c | 892 static inline void *alloc_rsttbl_from_idx(struct RESTART_TABLE **tbl, u32 vbo) in alloc_rsttbl_from_idx() argument 901 if (vbo >= bytes) { in alloc_rsttbl_from_idx() 906 u32 bytes2idx = vbo - bytes; in alloc_rsttbl_from_idx() 918 e = Add2Ptr(rt, vbo); in alloc_rsttbl_from_idx() 930 if (off == vbo) { in alloc_rsttbl_from_idx() 952 if (off == vbo) { in alloc_rsttbl_from_idx() 982 u32 vbo; member 1062 u32 vbo = (lsn << log->seq_num_bits) >> (log->seq_num_bits - 3); in lsn_to_vbo() local 1064 return vbo; in lsn_to_vbo() 1123 static int read_log_page(struct ntfs_log *log, u32 vbo, in read_log_page() argument 1187 u32 skip, vbo; log_read_rst() local 1449 u32 vbo = lsn_to_vbo(log, this_lsn); next_log_lsn() local 1544 check_subseq_log_page(struct ntfs_log *log, const struct RECORD_PAGE_HDR *rp, u32 vbo, u64 seq) check_subseq_log_page() argument 2236 u32 vbo = lsn_to_vbo(log, lsn) & ~log->page_mask; read_log_rec_buf() local 3042 u64 vbo = cbo + tvo; do_action() local 3741 u32 vbo, tail, off, dlen; log_replay() local [all...] |
H A D | frecord.c | 948 u64 vbo; in ni_ins_attr_ext() local 979 vbo = is_mft_data ? ((u64)svcn << sbi->cluster_bits) : 0; in ni_ins_attr_ext() 995 vbo <= ((u64)mi->rno << sbi->record_bits))) { in ni_ins_attr_ext() 1042 if (is_mft_data && vbo <= ((u64)rno << sbi->record_bits)) { in ni_ins_attr_ext() 1906 __u64 vbo, __u64 len) in ni_fiemap() 1914 CLST vcn = vbo >> cluster_bits; in ni_fiemap() 1956 end = vbo + len; in ni_fiemap() 1963 while (vbo < end) { in ni_fiemap() 2004 vbo = (u64)vcn << cluster_bits; in ni_fiemap() 2024 vbo in ni_fiemap() 1905 ni_fiemap(struct ntfs_inode *ni, struct fiemap_extent_info *fieinfo, __u64 vbo, __u64 len) ni_fiemap() argument 2094 u64 frame_vbo, vbo = (u64)index << PAGE_SHIFT; ni_readpage_cmpr() local 2184 u64 vbo; ni_decompress_file() local [all...] |
H A D | index.c | 208 size_t data_size, valid_size, vbo, off = bit >> 3; in bmp_buf_get() local 263 vbo = off & ~(size_t)sbi->block_mask; in bmp_buf_get() 265 bbuf->new_valid = vbo + blocksize; in bmp_buf_get() 271 if (vbo >= valid_size) { in bmp_buf_get() 273 } else if (vbo + blocksize > valid_size) { in bmp_buf_get() 375 size_t vbo = from >> 3; in scan_nres_bitmap() local 376 sector_t blk = (vbo & sbi->cluster_mask) >> sb->s_blocksize_bits; in scan_nres_bitmap() 377 sector_t vblock = vbo >> sb->s_blocksize_bits; in scan_nres_bitmap() 390 vcn = vbo >> sbi->cluster_bits; in scan_nres_bitmap() 422 vbo in scan_nres_bitmap() 940 u64 vbo = (u64)vbn << indx->vbn2vbo_bits; indx_new() local 1044 u64 vbo = (u64)vbn << indx->vbn2vbo_bits; indx_read() local [all...] |
H A D | ntfs_fs.h | 452 int attr_collapse_range(struct ntfs_inode *ni, u64 vbo, u64 bytes); 453 int attr_insert_range(struct ntfs_inode *ni, u64 vbo, u64 bytes); 454 int attr_punch_hole(struct ntfs_inode *ni, u64 vbo, u64 bytes, u32 *frame_size); 561 __u64 vbo, __u64 len); 617 u64 vbo, const void *buf, size_t bytes, int sync); 619 const struct runs_tree *run, u64 vbo); 621 u64 vbo, void *buf, u32 bytes, struct ntfs_buffers *nb); 622 int ntfs_read_bh(struct ntfs_sb_info *sbi, const struct runs_tree *run, u64 vbo, 625 int ntfs_get_bh(struct ntfs_sb_info *sbi, const struct runs_tree *run, u64 vbo, 630 struct page **pages, u32 nr_pages, u64 vbo, u3 [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/imx/ |
H A D | ipuv3-plane.c | 354 unsigned long eba, ubo, vbo, old_ubo, old_vbo, alpha_eba; in ipu_plane_atomic_check() local 436 vbo = drm_plane_state_to_vbo(state); in ipu_plane_atomic_check() 438 if (vbo & 0x7 || vbo > 0xfffff8) in ipu_plane_atomic_check() 443 if (vbo != old_vbo) in ipu_plane_atomic_check() 553 unsigned long eba, ubo, vbo; in ipu_plane_atomic_update() local 652 vbo = drm_plane_state_to_vbo(state); in ipu_plane_atomic_update() 656 swap(ubo, vbo); in ipu_plane_atomic_update() 659 fb->pitches[1], ubo, vbo); in ipu_plane_atomic_update() 662 "phy = %lu %lu %lu, x = %d, y = %d", eba, ubo, vbo, in ipu_plane_atomic_update() [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/imx/ipuv3/ |
H A D | ipuv3-plane.c | 377 unsigned long eba, ubo, vbo, old_ubo, old_vbo, alpha_eba; in ipu_plane_atomic_check() local 466 vbo = drm_plane_state_to_vbo(new_state); in ipu_plane_atomic_check() 468 if (vbo & 0x7 || vbo > 0xfffff8) in ipu_plane_atomic_check() 473 if (vbo != old_vbo) in ipu_plane_atomic_check() 586 unsigned long eba, ubo, vbo; in ipu_plane_atomic_update() local 710 vbo = drm_plane_state_to_vbo(new_state); in ipu_plane_atomic_update() 714 swap(ubo, vbo); in ipu_plane_atomic_update() 717 fb->pitches[1], ubo, vbo); in ipu_plane_atomic_update() 720 "phy = %lu %lu %lu, x = %d, y = %d", eba, ubo, vbo, in ipu_plane_atomic_update() [all...] |