/kernel/linux/linux-6.6/drivers/gpu/drm/ttm/tests/ |
H A D | ttm_pool_test.c | 52 enum ttm_caching caching, in ttm_tt_kunit_init() 67 err = ttm_tt_init(tt, priv->mock_bo, page_flags, caching, 0); in ttm_tt_kunit_init() 75 enum ttm_caching caching) in ttm_pool_pre_populated() 84 tt = ttm_tt_kunit_init(test, order, caching, size); in ttm_pool_pre_populated() 143 enum ttm_caching caching = ttm_uncached; in ttm_pool_alloc_basic() local 148 tt = ttm_tt_kunit_init(test, 0, caching, size); in ttm_pool_alloc_basic() 205 enum ttm_caching caching = ttm_uncached; in ttm_pool_alloc_basic_dma_addr() local 216 err = ttm_sg_tt_init(tt, bo, 0, caching); in ttm_pool_alloc_basic_dma_addr() 244 enum ttm_caching caching = ttm_uncached; in ttm_pool_alloc_order_caching_match() local 249 pool = ttm_pool_pre_populated(test, size, caching); in ttm_pool_alloc_order_caching_match() 50 ttm_tt_kunit_init(struct kunit *test, uint32_t page_flags, enum ttm_caching caching, size_t size) ttm_tt_kunit_init() argument 73 ttm_pool_pre_populated(struct kunit *test, size_t size, enum ttm_caching caching) ttm_pool_pre_populated() argument 306 enum ttm_caching caching = ttm_uncached; ttm_pool_alloc_order_mismatch() local 342 enum ttm_caching caching = ttm_uncached; ttm_pool_free_dma_alloc() local 373 enum ttm_caching caching = ttm_uncached; ttm_pool_free_no_dma_alloc() local 401 enum ttm_caching caching = ttm_uncached; ttm_pool_fini_basic() local [all...] |
H A D | ttm_device_test.c | 179 pt = pool->caching[i].orders[j]; in ttm_device_init_pools() 181 KUNIT_EXPECT_EQ(test, pt.caching, i); in ttm_device_init_pools()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/ttm/ |
H A D | ttm_pool.c | 26 /* Pooling of allocated pages is necessary because changing the caching 131 /* Reset the caching and pages of size 1 << order */ 132 static void ttm_pool_free_page(struct ttm_pool *pool, enum ttm_caching caching, in ttm_pool_free_page() argument 143 if (caching != ttm_cached && !PageHighMem(p)) in ttm_pool_free_page() 162 /* Apply a new caching to an array of pages */ 164 enum ttm_caching caching) in ttm_pool_apply_caching() 172 switch (caching) { in ttm_pool_apply_caching() 259 enum ttm_caching caching, unsigned int order) in ttm_pool_type_init() 262 pt->caching = caching; in ttm_pool_type_init() 163 ttm_pool_apply_caching(struct page **first, struct page **last, enum ttm_caching caching) ttm_pool_apply_caching() argument 258 ttm_pool_type_init(struct ttm_pool_type *pt, struct ttm_pool *pool, enum ttm_caching caching, unsigned int order) ttm_pool_type_init() argument 286 ttm_pool_select_type(struct ttm_pool *pool, enum ttm_caching caching, unsigned int order) ttm_pool_select_type() argument 383 ttm_pool_free_range(struct ttm_pool *pool, struct ttm_tt *tt, enum ttm_caching caching, pgoff_t start_page, pgoff_t end_page) ttm_pool_free_range() argument 424 struct page **caching = tt->pages; ttm_pool_alloc() local [all...] |
H A D | ttm_module.c | 57 * @caching: The ttm caching mode 62 pgprot_t ttm_prot_from_caching(enum ttm_caching caching, pgprot_t tmp) in ttm_prot_from_caching() argument 65 if (caching == ttm_cached) in ttm_prot_from_caching() 69 if (caching == ttm_write_combined) in ttm_prot_from_caching() 78 if (caching == ttm_write_combined) in ttm_prot_from_caching()
|
H A D | ttm_tt.c | 136 enum ttm_caching caching, in ttm_tt_init_fields() 144 ttm->caching = caching; in ttm_tt_init_fields() 148 uint32_t page_flags, enum ttm_caching caching, in ttm_tt_init() 151 ttm_tt_init_fields(ttm, bo, page_flags, caching, extra_pages); in ttm_tt_init() 179 uint32_t page_flags, enum ttm_caching caching) in ttm_sg_tt_init() 183 ttm_tt_init_fields(ttm, bo, page_flags, caching, 0); in ttm_sg_tt_init() 445 iter_tt->prot = ttm_prot_from_caching(tt->caching, PAGE_KERNEL); in ttm_kmap_iter_tt_init() 133 ttm_tt_init_fields(struct ttm_tt *ttm, struct ttm_buffer_object *bo, uint32_t page_flags, enum ttm_caching caching, unsigned long extra_pages) ttm_tt_init_fields() argument 147 ttm_tt_init(struct ttm_tt *ttm, struct ttm_buffer_object *bo, uint32_t page_flags, enum ttm_caching caching, unsigned long extra_pages) ttm_tt_init() argument 178 ttm_sg_tt_init(struct ttm_tt *ttm, struct ttm_buffer_object *bo, uint32_t page_flags, enum ttm_caching caching) ttm_sg_tt_init() argument
|
H A D | ttm_bo_util.c | 288 * setting up a PTE with the caching model indicated by @c_state. 294 enum ttm_caching caching; in ttm_io_prot() local 297 caching = man->use_tt ? bo->ttm->caching : res->bus.caching; in ttm_io_prot() 299 return ttm_prot_from_caching(caching, tmp); in ttm_io_prot() 317 if (mem->bus.caching == ttm_write_combined) in ttm_bo_ioremap() 320 else if (mem->bus.caching == ttm_cached) in ttm_bo_ioremap() 349 if (num_pages == 1 && ttm->caching == ttm_cached) { in ttm_bo_kmap_ttm() 476 else if (mem->bus.caching in ttm_bo_vmap() [all...] |
H A D | ttm_agp_backend.c | 55 int ret, cached = ttm->caching == ttm_cached; in ttm_agp_bind()
|
H A D | ttm_resource.c | 190 res->bus.caching = ttm_cached; in ttm_resource_init() 673 if (mem->bus.caching == ttm_write_combined) in ttm_kmap_iter_linear_io_init() 677 else if (mem->bus.caching == ttm_cached) in ttm_kmap_iter_linear_io_init()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/ |
H A D | nouveau_sgdma.c | 73 enum ttm_caching caching; in nouveau_sgdma_create_ttm() local 76 caching = ttm_uncached; in nouveau_sgdma_create_ttm() 78 caching = ttm_write_combined; in nouveau_sgdma_create_ttm() 80 caching = ttm_cached; in nouveau_sgdma_create_ttm() 86 if (ttm_sg_tt_init(&nvbe->ttm, bo, page_flags, caching)) { in nouveau_sgdma_create_ttm()
|
/kernel/linux/linux-6.6/include/drm/ttm/ |
H A D | ttm_tt.h | 42 * struct ttm_tt - This is a structure holding the pages, caching- and aperture 102 * @caching: The current caching state of the pages, see enum 105 enum ttm_caching caching; member 142 * @caching: the desired caching state of the pages 151 uint32_t page_flags, enum ttm_caching caching, 154 uint32_t page_flags, enum ttm_caching caching);
|
H A D | ttm_pool.h | 44 * @caching: the caching type our pages have 52 enum ttm_caching caching; member 61 * struct ttm_pool - Pool for all caching and orders 67 * @caching: pools for each caching/order 78 } caching[TTM_NUM_CACHING_TYPES]; member
|
H A D | ttm_caching.h | 31 * enum ttm_caching - CPU caching and BUS snooping behavior. 53 pgprot_t ttm_prot_from_caching(enum ttm_caching caching, pgprot_t tmp);
|
/kernel/linux/linux-5.10/drivers/acpi/ |
H A D | acpi_memhotplug.c | 49 unsigned short caching; /* memory cache attribute */ member 74 if ((info->caching == address64.info.mem.caching) && in acpi_memory_get_resource() 87 new->caching = address64.info.mem.caching; in acpi_memory_get_resource()
|
/kernel/linux/linux-6.6/fs/afs/ |
H A D | write.c | 23 loff_t i_size, bool caching); 35 static void afs_folio_start_fscache(bool caching, struct folio *folio) in afs_folio_start_fscache() argument 37 if (caching) in afs_folio_start_fscache() 41 static void afs_folio_start_fscache(bool caching, struct folio *folio) in afs_folio_start_fscache() argument 83 /* Prefetch area to be written into the cache if we're caching this in afs_write_begin() 467 bool caching, in afs_extend_writeback() 551 * writable and mark them for caching. in afs_extend_writeback() 564 afs_folio_start_fscache(caching, folio); in afs_extend_writeback() 592 bool caching = fscache_cookie_enabled(afs_vnode_cache(vnode)); in afs_write_back_from_locked_folio() local 600 afs_folio_start_fscache(caching, foli in afs_write_back_from_locked_folio() 461 afs_extend_writeback(struct address_space *mapping, struct afs_vnode *vnode, long *_count, loff_t start, loff_t max_len, bool new_content, bool caching, unsigned int *_len) afs_extend_writeback() argument 1047 afs_write_to_cache(struct afs_vnode *vnode, loff_t start, size_t len, loff_t i_size, bool caching) afs_write_to_cache() argument 1053 afs_write_to_cache_done, vnode, caching); afs_write_to_cache() local [all...] |
/kernel/linux/linux-5.10/drivers/xen/ |
H A D | xen-acpi-memhotplug.c | 25 unsigned short caching; /* memory cache attribute */ member 108 if ((info->caching == address64.info.mem.caching) && in acpi_memory_get_resource() 121 new->caching = address64.info.mem.caching; in acpi_memory_get_resource()
|
/kernel/linux/linux-6.6/drivers/acpi/ |
H A D | acpi_memhotplug.c | 49 unsigned short caching; /* memory cache attribute */ member 75 if ((info->caching == address64.info.mem.caching) && in acpi_memory_get_resource() 88 new->caching = address64.info.mem.caching; in acpi_memory_get_resource()
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | fscache.h | 2 /* General filesystem caching interface 9 * Documentation/filesystems/caching/netfs-api.rst 151 * slow-path functions for when there is actually caching available, and the 181 * fscache_acquire_volume - Register a volume as desiring caching services 187 * Register a volume as desiring caching services if they're available. The 194 * cookie pointer and can be returned if caching is refused. 209 * fscache_relinquish_volume - Cease caching a volume 214 * Indicate that a filesystem no longer desires caching services for a volume. 239 * See Documentation/filesystems/caching/netfs-api.rst for a complete 301 * See Documentation/filesystems/caching/netf 586 fscache_clear_page_bits(struct address_space *mapping, loff_t start, size_t len, bool caching) fscache_clear_page_bits() argument 617 fscache_write_to_cache(struct fscache_cookie *cookie, struct address_space *mapping, loff_t start, size_t len, loff_t i_size, netfs_io_terminated_t term_func, void *term_func_priv, bool caching) fscache_write_to_cache() argument [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_domain.c | 235 args->caching = I915_CACHING_CACHED; in i915_gem_get_caching_ioctl() 239 args->caching = I915_CACHING_DISPLAY; in i915_gem_get_caching_ioctl() 243 args->caching = I915_CACHING_NONE; in i915_gem_get_caching_ioctl() 260 switch (args->caching) { in i915_gem_set_caching_ioctl() 288 * The caching mode of proxy object is handled by its generator, and in i915_gem_set_caching_ioctl()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_domain.c | 329 args->caching = I915_CACHING_CACHED; in i915_gem_get_caching_ioctl() 331 args->caching = I915_CACHING_DISPLAY; in i915_gem_get_caching_ioctl() 333 args->caching = I915_CACHING_NONE; in i915_gem_get_caching_ioctl() 354 switch (args->caching) { in i915_gem_set_caching_ioctl() 391 * The caching mode of proxy object is handled by its generator, and in i915_gem_set_caching_ioctl() 400 args->caching != I915_CACHING_CACHED) in i915_gem_set_caching_ioctl()
|
H A D | i915_gem_ttm.c | 279 enum ttm_caching caching; in i915_ttm_tt_create() local 294 caching = i915_ttm_select_tt_caching(obj); in i915_ttm_tt_create() 295 if (i915_gem_object_is_shrinkable(obj) && caching == ttm_cached) { in i915_ttm_tt_create() 306 ret = ttm_tt_init(&i915_tt->ttm, bo, page_flags, caching, ccs_pages); in i915_ttm_tt_create() 557 * @obj: The GEM object used for sg-table caching 684 mem->bus.caching = ttm_write_combined; in i915_ttm_io_mem_reserve()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/radeon/ |
H A D | radeon_ttm.c | 280 mem->bus.caching = ttm_write_combined; in radeon_ttm_io_mem_reserve() 291 mem->bus.caching = ttm_write_combined; in radeon_ttm_io_mem_reserve() 451 if (ttm->caching == ttm_cached) in radeon_ttm_backend_bind() 492 enum ttm_caching caching; in radeon_ttm_tt_create() local 509 caching = ttm_uncached; in radeon_ttm_tt_create() 511 caching = ttm_write_combined; in radeon_ttm_tt_create() 513 caching = ttm_cached; in radeon_ttm_tt_create() 515 if (ttm_sg_tt_init(>t->ttm, bo, page_flags, caching)) { in radeon_ttm_tt_create()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/qxl/ |
H A D | qxl_ttm.c | 85 mem->bus.caching = ttm_write_combined; in qxl_ttm_io_mem_reserve() 91 mem->bus.caching = ttm_write_combined; in qxl_ttm_io_mem_reserve()
|
/kernel/linux/linux-6.6/fs/ceph/ |
H A D | addr.c | 530 static void ceph_fscache_write_to_cache(struct inode *inode, u64 off, u64 len, bool caching) in ceph_fscache_write_to_cache() argument 536 ceph_fscache_write_terminated, inode, caching); in ceph_fscache_write_to_cache() 543 static inline void ceph_fscache_write_to_cache(struct inode *inode, u64 off, u64 len, bool caching) in ceph_fscache_write_to_cache() argument 670 bool caching = ceph_is_cache_enabled(inode); in writepage_nounlock() local 729 if (caching) in writepage_nounlock() 731 ceph_fscache_write_to_cache(inode, page_off, len, caching); in writepage_nounlock() 940 bool caching = ceph_is_cache_enabled(inode); in ceph_writepages_start() local 1252 ceph_fscache_write_to_cache(inode, offset, len, caching); in ceph_writepages_start() 1271 if (caching) in ceph_writepages_start() 1275 ceph_fscache_write_to_cache(inode, offset, len, caching); in ceph_writepages_start() [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/ttm/ |
H A D | ttm_bo.c | 878 uint32_t caching = proposed_placement & TTM_PL_MASK_CACHING; in ttm_bo_select_caching() local 882 * Keep current caching if possible. in ttm_bo_select_caching() 885 if ((cur_placement & caching) != 0) in ttm_bo_select_caching() 886 result |= (cur_placement & caching); in ttm_bo_select_caching() 887 else if ((TTM_PL_FLAG_CACHED & caching) != 0) in ttm_bo_select_caching() 889 else if ((TTM_PL_FLAG_WC & caching) != 0) in ttm_bo_select_caching() 891 else if ((TTM_PL_FLAG_UNCACHED & caching) != 0) in ttm_bo_select_caching()
|
/kernel/linux/linux-5.10/drivers/acpi/acpica/ |
H A D | rsaddr.c | 218 {ACPI_RSC_2BITFLAG, ACPI_RS_OFFSET(data.address.info.mem.caching),
|