Home
last modified time | relevance | path

Searched refs:kbase_va_region (Results 1 - 25 of 77) sorted by relevance

1234

/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/
H A Dmali_kbase_mem.h69 struct kbase_va_region *region;
135 struct kbase_va_region *reg;
272 struct kbase_va_region { struct
455 #define KBASE_RESERVED_REG_JIT_ALLOC ((struct kbase_va_region *)-1)
457 static inline bool kbase_is_region_free(struct kbase_va_region *reg) in kbase_is_region_free()
462 static inline bool kbase_is_region_invalid(struct kbase_va_region *reg) in kbase_is_region_invalid()
467 static inline bool kbase_is_region_invalid_or_free(struct kbase_va_region *reg) in kbase_is_region_invalid_or_free()
475 int kbase_remove_va_region(struct kbase_va_region *reg);
476 static inline void kbase_region_refcnt_free(struct kbase_va_region *reg) in kbase_region_refcnt_free()
488 static inline struct kbase_va_region *kbase_va_region_alloc_ge
[all...]
H A Dmali_kbase_mem.c135 static void kbase_region_tracker_insert(struct kbase_va_region *new_reg) in kbase_region_tracker_insert()
147 struct kbase_va_region *old_reg; in kbase_region_tracker_insert()
150 old_reg = rb_entry(parent, struct kbase_va_region, rblink); in kbase_region_tracker_insert()
168 static struct kbase_va_region *find_region_enclosing_range_rbtree(struct rb_root *rbtree, u64 start_pfn, in find_region_enclosing_range_rbtree()
172 struct kbase_va_region *reg; in find_region_enclosing_range_rbtree()
180 reg = rb_entry(rbnode, struct kbase_va_region, rblink); in find_region_enclosing_range_rbtree()
197 struct kbase_va_region *kbase_find_region_enclosing_address(struct rb_root *rbtree, u64 gpu_addr) in kbase_find_region_enclosing_address()
201 struct kbase_va_region *reg; in kbase_find_region_enclosing_address()
208 reg = rb_entry(rbnode, struct kbase_va_region, rblink); in kbase_find_region_enclosing_address()
226 struct kbase_va_region *kbase_region_tracker_find_region_enclosing_addres
[all...]
H A Dmali_kbase_mem_linux.h53 struct kbase_va_region *kbase_mem_alloc(struct kbase_context *kctx,
142 struct kbase_va_region *const reg, u64 const new_pages);
186 struct kbase_va_region *reg, u64 new_pages,
347 struct kbase_va_region *reg, u64 new_pages,
H A Dmali_kbase_mem_linux.c98 struct kbase_va_region *reg, u64 offset_bytes,
107 struct kbase_va_region *const reg,
115 static struct kbase_va_region *
120 struct kbase_va_region *reg; in kbase_find_event_mem_region()
180 struct kbase_va_region *reg, in kbase_phy_alloc_mapping_init()
248 struct kbase_va_region *reg; in kbase_phy_alloc_mapping_get()
307 struct kbase_va_region *kbase_mem_alloc(struct kbase_context *kctx, in kbase_mem_alloc()
312 struct kbase_va_region *reg; in kbase_mem_alloc()
555 struct kbase_va_region *reg; in kbase_mem_query()
927 struct kbase_va_region *re in kbase_mem_flags_change()
[all...]
H A Dmali_linux_trace.h155 mali_mmu_page_fault_grow, TP_PROTO(struct kbase_va_region *reg, struct kbase_fault *fault, size_t new_pages),
177 DECLARE_EVENT_CLASS(mali_jit_softjob_template, TP_PROTO(struct kbase_va_region *reg, u8 jit_id), TP_ARGS(reg, jit_id),
192 DEFINE_EVENT(mali_jit_softjob_template, mali_jit_alloc, TP_PROTO(struct kbase_va_region *reg, u8 jit_id),
204 DEFINE_EVENT_PRINT(mali_jit_softjob_template, mali_jit_free, TP_PROTO(struct kbase_va_region *reg, u8 jit_id),
218 TP_PROTO(struct kbase_jd_atom *katom, struct kbase_va_region *reg, unsigned int id_idx, u64 read_val,
249 TP_PROTO(struct kbase_va_region *reg, u64 new_used_pages, u64 new_pressure, unsigned int flags),
328 TP_PROTO(struct kbase_va_region *reg, size_t freed_pages, size_t old_pages, size_t available_pages,
H A Dmali_kbase_gwt.c32 struct kbase_va_region *reg; in kbase_gpu_gwt_setup_page_permission()
35 reg = rb_entry(rbnode, struct kbase_va_region, rblink); in kbase_gpu_gwt_setup_page_permission()
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/
H A Dmali_kbase_mem.h70 struct kbase_va_region *region;
142 struct kbase_va_region *reg;
282 * struct kbase_va_region - A GPU memory region, and attributes for CPU mappings
324 struct kbase_va_region { struct
531 #define KBASE_RESERVED_REG_JIT_ALLOC ((struct kbase_va_region *)-1)
533 static inline bool kbase_is_region_free(struct kbase_va_region *reg) in kbase_is_region_free()
538 static inline bool kbase_is_region_invalid(struct kbase_va_region *reg) in kbase_is_region_invalid()
543 static inline bool kbase_is_region_invalid_or_free(struct kbase_va_region *reg) in kbase_is_region_invalid_or_free()
552 struct kbase_va_region *reg);
554 struct kbase_va_region *re in kbase_region_refcnt_free()
[all...]
H A Dmali_kbase_mem.c141 static void kbase_region_tracker_insert(struct kbase_va_region *new_reg) in kbase_region_tracker_insert()
153 struct kbase_va_region *old_reg; in kbase_region_tracker_insert()
156 old_reg = rb_entry(parent, struct kbase_va_region, rblink); in kbase_region_tracker_insert()
173 static struct kbase_va_region *find_region_enclosing_range_rbtree( in find_region_enclosing_range_rbtree()
177 struct kbase_va_region *reg; in find_region_enclosing_range_rbtree()
185 reg = rb_entry(rbnode, struct kbase_va_region, rblink); in find_region_enclosing_range_rbtree()
202 struct kbase_va_region *kbase_find_region_enclosing_address( in kbase_find_region_enclosing_address()
207 struct kbase_va_region *reg; in kbase_find_region_enclosing_address()
214 reg = rb_entry(rbnode, struct kbase_va_region, rblink); in kbase_find_region_enclosing_address()
232 struct kbase_va_region *kbase_region_tracker_find_region_enclosing_addres
[all...]
H A Dmali_kbase_mem_linux.h52 struct kbase_va_region *
139 struct kbase_va_region *reg, u64 new_pages);
184 struct kbase_va_region *reg, u64 new_pages,
345 struct kbase_va_region *reg,
H A Dmali_kbase_mem_linux.c96 struct kbase_va_region *reg, u64 offset_bytes, size_t size,
104 struct kbase_va_region *reg,
111 static struct kbase_va_region *kbase_find_event_mem_region( in kbase_find_event_mem_region()
116 struct kbase_va_region *reg; in kbase_find_event_mem_region()
173 struct kbase_va_region *reg, size_t vsize, size_t size) in kbase_phy_alloc_mapping_init()
237 struct kbase_va_region *reg; in kbase_phy_alloc_mapping_get()
294 struct kbase_va_region *
300 struct kbase_va_region *reg; in kbase_mem_alloc()
523 struct kbase_va_region *reg; in kbase_mem_query()
883 struct kbase_va_region *re in kbase_mem_flags_change()
[all...]
H A Dmali_linux_trace.h224 TP_PROTO(struct kbase_va_region *reg, struct kbase_fault *fault,
265 TP_PROTO(struct kbase_va_region *reg, u8 jit_id),
290 TP_PROTO(struct kbase_va_region *reg, u8 jit_id),
303 TP_PROTO(struct kbase_va_region *reg, u8 jit_id),
316 TP_PROTO(struct kbase_jd_atom *katom, struct kbase_va_region *reg,
361 TP_PROTO(struct kbase_va_region *reg, u64 new_used_pages,
496 TP_PROTO(struct kbase_va_region *reg, size_t freed_pages,
/device/soc/rockchip/common/vendor/drivers/gpu/arm/midgard/
H A Dmali_kbase_mem.h69 struct kbase_va_region *region;
123 struct kbase_va_region *reg;
217 struct kbase_va_region { struct
312 static inline phys_addr_t *kbase_get_cpu_phy_pages(struct kbase_va_region *reg) in kbase_get_cpu_phy_pages()
322 static inline phys_addr_t *kbase_get_gpu_phy_pages(struct kbase_va_region *reg) in kbase_get_gpu_phy_pages()
332 static inline size_t kbase_reg_current_backed_size(struct kbase_va_region *reg) in kbase_reg_current_backed_size()
399 static inline int kbase_reg_prepare_native(struct kbase_va_region *reg, struct kbase_context *kctx) in kbase_reg_prepare_native()
618 struct kbase_va_region *kbase_region_tracker_find_region_enclosing_address(struct kbase_context *kctx, u64 gpu_addr);
625 struct kbase_va_region *kbase_region_tracker_find_region_base_address(struct kbase_context *kctx, u64 gpu_addr);
627 struct kbase_va_region *kbase_alloc_free_regio
[all...]
H A Dmali_kbase_mem.c42 static struct rb_root *kbase_reg_flags_to_rbtree(struct kbase_context *kctx, struct kbase_va_region *reg) in kbase_reg_flags_to_rbtree()
94 static void kbase_region_tracker_insert(struct kbase_context *kctx, struct kbase_va_region *new_reg) in kbase_region_tracker_insert()
106 struct kbase_va_region *old_reg; in kbase_region_tracker_insert()
109 old_reg = rb_entry(parent, struct kbase_va_region, rblink); in kbase_region_tracker_insert()
128 static struct kbase_va_region *kbase_region_tracker_find_region_enclosing_range_free(struct kbase_context *kctx, in kbase_region_tracker_find_region_enclosing_range_free()
132 struct kbase_va_region *reg = NULL; in kbase_region_tracker_find_region_enclosing_range_free()
144 reg = rb_entry(rbnode, struct kbase_va_region, rblink); in kbase_region_tracker_find_region_enclosing_range_free()
162 struct kbase_va_region *kbase_region_tracker_find_region_enclosing_address(struct kbase_context *kctx, u64 gpu_addr) in kbase_region_tracker_find_region_enclosing_address()
165 struct kbase_va_region *reg; in kbase_region_tracker_find_region_enclosing_address()
180 reg = rb_entry(rbnode, struct kbase_va_region, rblin in kbase_region_tracker_find_region_enclosing_address()
[all...]
H A Dmali_kbase_mem_linux.c58 static void kbase_mem_shrink_cpu_mapping(struct kbase_context *kctx, struct kbase_va_region *reg, u64 new_pages,
74 static int kbase_mem_shrink_gpu_mapping(struct kbase_context *kctx, struct kbase_va_region *reg, u64 new_pages,
77 struct kbase_va_region *kbase_mem_alloc(struct kbase_context *kctx, u64 va_pages, u64 commit_pages, u64 extent, in kbase_mem_alloc()
83 struct kbase_va_region *reg; in kbase_mem_alloc()
253 struct kbase_va_region *reg; in kbase_mem_query()
569 struct kbase_va_region *reg; in kbase_mem_flags_change()
692 static struct kbase_va_region *kbase_mem_from_ump(struct kbase_context *kctx, ump_secure_id id, u64 *va_pages, in kbase_mem_from_ump()
695 struct kbase_va_region *reg; in kbase_mem_from_ump()
816 static struct kbase_va_region *kbase_mem_from_umm(struct kbase_context *kctx, int fd, u64 *va_pages, u64 *flags, in kbase_mem_from_umm()
819 struct kbase_va_region *re in kbase_mem_from_umm()
[all...]
H A Dmali_kbase_mem_linux.h36 struct kbase_va_region *kbase_mem_alloc(struct kbase_context *kctx, u64 va_pages, u64 commit_pages, u64 extent,
88 int kbase_mem_grow_gpu_mapping(struct kbase_context *kctx, struct kbase_va_region *reg, u64 new_pages, u64 old_pages);
/device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/
H A Dmali_kbase_mem.h71 struct kbase_va_region *region;
125 struct kbase_va_region *reg;
217 struct kbase_va_region { struct
311 static inline phys_addr_t *kbase_get_cpu_phy_pages(struct kbase_va_region *reg) in kbase_get_cpu_phy_pages()
321 static inline phys_addr_t *kbase_get_gpu_phy_pages(struct kbase_va_region *reg) in kbase_get_gpu_phy_pages()
331 static inline size_t kbase_reg_current_backed_size(struct kbase_va_region *reg) in kbase_reg_current_backed_size()
395 static inline int kbase_reg_prepare_native(struct kbase_va_region *reg, in kbase_reg_prepare_native()
622 struct kbase_va_region *kbase_region_tracker_find_region_enclosing_address(struct kbase_context *kctx, u64 gpu_addr);
629 struct kbase_va_region *kbase_region_tracker_find_region_base_address(struct kbase_context *kctx, u64 gpu_addr);
631 struct kbase_va_region *kbase_alloc_free_regio
[all...]
H A Dmali_kbase_mem.c47 struct kbase_va_region *reg) in kbase_reg_flags_to_rbtree()
99 struct kbase_va_region *new_reg) in kbase_region_tracker_insert()
111 struct kbase_va_region *old_reg; in kbase_region_tracker_insert()
114 old_reg = rb_entry(parent, struct kbase_va_region, rblink); in kbase_region_tracker_insert()
132 static struct kbase_va_region *kbase_region_tracker_find_region_enclosing_range_free( in kbase_region_tracker_find_region_enclosing_range_free()
136 struct kbase_va_region *reg = NULL; in kbase_region_tracker_find_region_enclosing_range_free()
148 reg = rb_entry(rbnode, struct kbase_va_region, rblink); in kbase_region_tracker_find_region_enclosing_range_free()
166 struct kbase_va_region *kbase_region_tracker_find_region_enclosing_address(struct kbase_context *kctx, u64 gpu_addr) in kbase_region_tracker_find_region_enclosing_address()
169 struct kbase_va_region *reg; in kbase_region_tracker_find_region_enclosing_address()
184 reg = rb_entry(rbnode, struct kbase_va_region, rblin in kbase_region_tracker_find_region_enclosing_address()
[all...]
H A Dmali_kbase_mem_linux.c64 struct kbase_va_region *reg,
81 struct kbase_va_region *reg,
84 struct kbase_va_region *kbase_mem_alloc(struct kbase_context *kctx, in kbase_mem_alloc()
91 struct kbase_va_region *reg; in kbase_mem_alloc()
261 struct kbase_va_region *reg; in kbase_mem_query()
582 struct kbase_va_region *reg; in kbase_mem_flags_change()
694 static struct kbase_va_region *kbase_mem_from_ump(struct kbase_context *kctx, ump_secure_id id, u64 *va_pages, u64 *flags) in kbase_mem_from_ump()
696 struct kbase_va_region *reg; in kbase_mem_from_ump()
812 static struct kbase_va_region *kbase_mem_from_umm(struct kbase_context *kctx, in kbase_mem_from_umm()
815 struct kbase_va_region *re in kbase_mem_from_umm()
[all...]
H A Dmali_kbase_mem_linux.h35 struct kbase_va_region *kbase_mem_alloc(struct kbase_context *kctx,
89 struct kbase_va_region *reg,
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/csf/
H A Dmali_kbase_csf_defs.h277 struct kbase_va_region *reg;
286 struct kbase_va_region *queue_reg;
309 struct kbase_va_region *reg;
322 struct kbase_va_region *reg;
465 struct kbase_va_region *region;
H A Dmali_kbase_csf_tiler_heap_def.h74 struct kbase_va_region *region;
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/csf/
H A Dmali_kbase_csf_defs.h339 struct kbase_va_region *reg;
348 struct kbase_va_region *queue_reg;
380 struct kbase_va_region *reg;
393 struct kbase_va_region *reg;
562 struct kbase_va_region *region;
H A Dmali_kbase_csf_tiler_heap_def.h77 struct kbase_va_region *region;
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/mmu/
H A Dmali_kbase_mmu_internal.h56 struct kbase_va_region *reg);
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/mmu/
H A Dmali_kbase_mmu_internal.h52 int kbase_mmu_switch_to_ir(struct kbase_context *kctx, struct kbase_va_region *reg);

Completed in 33 milliseconds

1234