/device/soc/rockchip/common/vendor/drivers/gpu/arm/midgard/ |
H A D | mali_malisw.h | 34 #define U64_MAX ((u64)~0ULL) macro 35 #define S64_MAX ((s64)(U64_MAX >> 1))
|
H A D | mali_kbase_mmu.c | 535 KBASE_DEBUG_ASSERT(vpfn <= (U64_MAX / PAGE_SIZE));
in mmu_insert_pages_failure_recovery() 592 KBASE_DEBUG_ASSERT(vpfn <= (U64_MAX / PAGE_SIZE));
in kbase_mmu_insert_single_page() 698 KBASE_DEBUG_ASSERT(vpfn <= (U64_MAX / PAGE_SIZE));
in kbase_mmu_insert_pages_no_flush() 1082 KBASE_DEBUG_ASSERT(vpfn <= (U64_MAX / PAGE_SIZE));
in kbase_mmu_update_pages()
|
H A D | mali_kbase_mem_linux.c | 101 if (va_pages > (U64_MAX / PAGE_SIZE)) {
in kbase_mem_alloc() 725 if (*va_pages > (U64_MAX / PAGE_SIZE)) {
in kbase_mem_from_ump() 839 if (*va_pages > (U64_MAX / PAGE_SIZE)) {
in kbase_mem_from_umm() 1138 if ((nents * stride) > (U64_MAX / PAGE_SIZE)) {
|
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/ |
H A D | mali_malisw.h | 41 #define U64_MAX ((u64)~0ULL) macro 42 #define S64_MAX ((s64)(U64_MAX >> 1))
|
H A D | mali_kbase_hwcnt_types.h | 538 u64 block_enable_map_mask = U64_MAX; in kbase_hwcnt_enable_map_block_enable_all() 625 u64 block_enable_map_mask = U64_MAX; in kbase_hwcnt_enable_map_block_enabled()
|
H A D | mali_kbase_vinstr.c | 261 u64 earliest_next_ns = U64_MAX; in kbasep_vinstr_reschedule_worker() 290 if ((earliest_next_ns != U64_MAX) && (vctx->suspend_count == 0) && !WARN_ON(earliest_next_ns < cur_ts_ns)) { in kbasep_vinstr_reschedule_worker()
|
H A D | mali_kbase_mem.c | 746 kctx->exec_va_start = U64_MAX; in kbase_region_tracker_init() 1086 KBASE_DEBUG_ASSERT(start_pfn + nr_pages <= (U64_MAX / PAGE_SIZE)); in kbase_alloc_free_region() 4049 if (!WARN_ON(diff > U64_MAX - kctx->jit_current_phys_pressure)) { in kbase_jit_report_update_pressure() 4063 has_exec_va_zone = (kctx->exec_va_start != U64_MAX); in kbase_has_exec_va_zone()
|
H A D | mali_kbase_mem_linux.c | 1421 if (*va_pages > (U64_MAX / PAGE_SIZE)) { in kbase_mem_from_umm() 1816 if ((nents * stride) > (U64_MAX / PAGE_SIZE)) {
|
/device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/ |
H A D | mali_malisw.h | 36 #define U64_MAX ((u64)~0ULL) macro 37 #define S64_MAX ((s64)(U64_MAX>>1))
|
H A D | mali_kbase_mmu.c | 573 KBASE_DEBUG_ASSERT(vpfn <= (U64_MAX / PAGE_SIZE)); in mmu_insert_pages_failure_recovery() 630 KBASE_DEBUG_ASSERT(vpfn <= (U64_MAX / PAGE_SIZE)); in kbase_mmu_insert_single_page() 742 KBASE_DEBUG_ASSERT(vpfn <= (U64_MAX / PAGE_SIZE)); in kbase_mmu_insert_pages_no_flush() 1135 KBASE_DEBUG_ASSERT(vpfn <= (U64_MAX / PAGE_SIZE)); in kbase_mmu_update_pages()
|
H A D | mali_kbase_mem_linux.c | 109 if (va_pages > (U64_MAX / PAGE_SIZE)) in kbase_mem_alloc() 724 if (*va_pages > (U64_MAX / PAGE_SIZE)) in kbase_mem_from_ump() 832 if (*va_pages > (U64_MAX / PAGE_SIZE)) in kbase_mem_from_umm() 1137 if ((nents * stride) > (U64_MAX / PAGE_SIZE))
|
/device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/tests/mali_kutf_irq_test/ |
H A D | mali_kutf_irq_test_main.c | 171 u64 min_time = U64_MAX, max_time = 0, average_time = 0; in mali_kutf_irq_latency()
|
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/tests/mali_kutf_irq_test/ |
H A D | mali_kutf_irq_test_main.c | 182 u64 min_time = U64_MAX, max_time = 0, average_time = 0; in mali_kutf_irq_latency()
|
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/ |
H A D | mali_kbase_hwcnt_types.h | 662 u64 block_enable_map_mask = U64_MAX; in kbase_hwcnt_enable_map_block_enable_all() 759 u64 block_enable_map_mask = U64_MAX; in kbase_hwcnt_enable_map_block_enabled()
|
H A D | mali_kbase_kinstr_prfcnt.c | 308 u64 shortest_period_ns = U64_MAX; in kbasep_kinstr_prfcnt_reschedule_worker() 354 if ((shortest_period_ns != U64_MAX) && in kbasep_kinstr_prfcnt_reschedule_worker() 1959 * Temporarily, it is set to U64_MAX, waiting for the in kbasep_kinstr_prfcnt_get_block_info_list() 1963 U64_MAX; in kbasep_kinstr_prfcnt_get_block_info_list() 1965 U64_MAX; in kbasep_kinstr_prfcnt_get_block_info_list()
|
H A D | mali_kbase_vinstr.c | 283 u64 earliest_next_ns = U64_MAX; in kbasep_vinstr_reschedule_worker() 313 if ((earliest_next_ns != U64_MAX) && in kbasep_vinstr_reschedule_worker()
|
H A D | mali_kbase_mem.c | 831 * initially U64_MAX in kbase_region_tracker_init() 833 kbase_ctx_reg_zone_init(kctx, KBASE_REG_ZONE_EXEC_VA, U64_MAX, 0u); in kbase_region_tracker_init() 862 return (exec_va_zone->base_pfn != U64_MAX); in kbase_has_exec_va_zone_locked() 1356 KBASE_DEBUG_ASSERT(start_pfn + nr_pages <= (U64_MAX / PAGE_SIZE)); in kbase_alloc_free_region() 4521 if (!WARN_ON(diff > U64_MAX - kctx->jit_current_phys_pressure)) in kbase_jit_report_update_pressure()
|
H A D | mali_kbase_mem_linux.c | 1376 if (*va_pages > (U64_MAX / PAGE_SIZE)) in kbase_mem_from_umm() 1764 if ((nents * stride) > (U64_MAX / PAGE_SIZE))
|
/device/soc/rockchip/common/sdk_linux/include/linux/ |
H A D | memblock.h | 270 for (i = 0, __next_mem_pfn_range_in_zone(&i, zone, p_start, p_end); i != U64_MAX; \ 286 for (; i != U64_MAX; __next_mem_pfn_range_in_zone(&i, zone, p_start, p_end))
|
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/csf/ |
H A D | mali_kbase_csf_csg_debugfs.c | 197 wait_sync_live_value = U64_MAX; in kbasep_csf_scheduler_dump_active_queue() 254 wait_sync_live_value = U64_MAX; in kbasep_csf_scheduler_dump_active_queue()
|
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/mmu/ |
H A D | mali_kbase_mmu.c | 1283 KBASE_DEBUG_ASSERT(vpfn <= (U64_MAX / PAGE_SIZE)); in mmu_insert_pages_failure_recovery() 1389 KBASE_DEBUG_ASSERT(vpfn <= (U64_MAX / PAGE_SIZE)); in kbase_mmu_insert_single_page() 1557 KBASE_DEBUG_ASSERT(start_vpfn <= (U64_MAX / PAGE_SIZE)); in kbase_mmu_insert_pages_no_flush() 2186 KBASE_DEBUG_ASSERT(vpfn <= (U64_MAX / PAGE_SIZE)); in kbase_mmu_update_pages_no_flush()
|
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/mmu/ |
H A D | mali_kbase_mmu.c | 1038 KBASE_DEBUG_ASSERT(vpfn <= (U64_MAX / PAGE_SIZE)); in mmu_insert_pages_failure_recovery() 1121 KBASE_DEBUG_ASSERT(vpfn <= (U64_MAX / PAGE_SIZE)); in kbase_mmu_insert_single_page() 1260 KBASE_DEBUG_ASSERT(start_vpfn <= (U64_MAX / PAGE_SIZE)); in kbase_mmu_insert_pages_no_flush() 1719 KBASE_DEBUG_ASSERT(vpfn <= (U64_MAX / PAGE_SIZE)); in kbase_mmu_update_pages_no_flush()
|
/device/soc/rockchip/common/sdk_linux/drivers/gpu/drm/ |
H A D | drm_mode_config.c | 293 prop = drm_property_create_range(dev, DRM_MODE_PROP_ATOMIC, "OUT_FENCE_PTR", 0, U64_MAX); in drm_mode_create_standard_properties()
|
/device/soc/rockchip/common/sdk_linux/kernel/sched/ |
H A D | rt.c | 3044 } else if ((u64)rt_runtime_us > U64_MAX / NSEC_PER_USEC) { in sched_group_set_rt_runtime() 3068 if (rt_period_us > U64_MAX / NSEC_PER_USEC) { in sched_group_set_rt_period()
|
/device/soc/rockchip/common/sdk_linux/kernel/bpf/ |
H A D | verifier.c | 616 if (reg->umax_value != U64_MAX) {
in print_verifier_state() 1094 reg->umax_value = U64_MAX;
in verifier_mark_reg_unbounded() 1107 reg->umax_value = U64_MAX;
in verifier_mark_reg64_unbounded() 2253 reg->umin_value == 0 && reg->umax_value == U64_MAX && reg->s32_min_value == S32_MIN &&
in __is_scalar_unbounded() 6061 dst_reg->umax_value = U64_MAX;
in adjust_ptr_min_max_vals() 6115 dst_reg->umax_value = U64_MAX;
in adjust_ptr_min_max_vals() 6203 dst_reg->umax_value = U64_MAX;
in scalar_min_max_add() 6255 dst_reg->umax_value = U64_MAX;
in scalar_min_max_sub() 6575 dst_reg->umax_value = U64_MAX;
in __scalar64_min_max_lsh() 6698 dst_reg->umax_value = U64_MAX;
in scalar_min_max_arsh() [all...] |