/device/soc/rockchip/common/vendor/drivers/gpu/arm/mali400/mali/linux/ |
H A D | mali_osk_locks.c | 180 struct _mali_osk_lock_debug_s *curr; in remove_lock_from_log() local 188 curr = lock_lookup_list; in remove_lock_from_log() 190 if (curr == NULL) { in remove_lock_from_log() 195 MALI_DEBUG_ASSERT_POINTER(curr); in remove_lock_from_log() 197 while (lock != curr) { in remove_lock_from_log() 198 prev = curr; in remove_lock_from_log() 200 MALI_DEBUG_ASSERT_POINTER(curr); in remove_lock_from_log() 201 curr = curr->next; in remove_lock_from_log() 206 lock_lookup_list = curr in remove_lock_from_log() [all...] |
H A D | mali_internal_sync.c | 141 static int mali_internal_sync_fence_wake_up_wq(wait_queue_entry_t *curr, in mali_internal_sync_fence_wake_up_wq() argument 145 static int mali_internal_sync_fence_wake_up_wq(wait_queue_t *curr, in mali_internal_sync_fence_wake_up_wq() 155 wait = container_of(curr, struct mali_internal_sync_fence_waiter, work); in mali_internal_sync_fence_wake_up_wq()
|
/device/soc/rockchip/common/kernel/drivers/gpu/arm/mali400/mali/linux/ |
H A D | mali_osk_locks.c | 182 struct _mali_osk_lock_debug_s *curr; in remove_lock_from_log() local 190 curr = lock_lookup_list; in remove_lock_from_log() 192 if (NULL == curr) { in remove_lock_from_log() 197 MALI_DEBUG_ASSERT_POINTER(curr); in remove_lock_from_log() 200 while (lock != curr) { in remove_lock_from_log() 201 prev = curr; in remove_lock_from_log() 203 MALI_DEBUG_ASSERT_POINTER(curr); in remove_lock_from_log() 204 curr = curr->next; in remove_lock_from_log() 209 lock_lookup_list = curr in remove_lock_from_log() [all...] |
H A D | mali_internal_sync.c | 124 static int mali_internal_sync_fence_wake_up_wq(wait_queue_entry_t *curr, unsigned mode, in mali_internal_sync_fence_wake_up_wq() argument 127 static int mali_internal_sync_fence_wake_up_wq(wait_queue_t *curr, unsigned mode, in mali_internal_sync_fence_wake_up_wq() 136 wait = container_of(curr, struct mali_internal_sync_fence_waiter, work); in mali_internal_sync_fence_wake_up_wq()
|
/device/soc/rockchip/common/sdk_linux/kernel/sched/ |
H A D | rt.c | 101 rt_rq->highest_prio.curr = MAX_RT_PRIO; in init_rt_rq() 176 rt_rq->highest_prio.curr = MAX_RT_PRIO; in init_tg_rt_entry() 289 return rq->rt.highest_prio.curr > prev->prio && !cpu_isolated(cpu_of(rq)); in need_pull_rt_task() 558 struct task_struct *curr = rq_of_rt_rq(rt_rq)->curr; in sched_rt_rq_enqueue() local 573 if (rt_rq->highest_prio.curr < curr->prio) { in sched_rt_rq_enqueue() 954 if (rt_rq->rt_nr_running && rq->curr == rq->idle) { in do_sched_rt_period_timer() 991 return rt_rq->highest_prio.curr; in rt_se_prio() 1061 struct task_struct *curr in update_curr_rt() local 1535 struct task_struct *curr; select_task_rq_rt() local [all...] |
H A D | wait.c | 69 wait_queue_entry_t *curr, *next; in __wake_up_common() local 75 curr = list_next_entry(bookmark, entry); in __wake_up_common() 80 curr = list_first_entry(&wq_head->head, wait_queue_entry_t, entry); in __wake_up_common() 83 if (&curr->entry == &wq_head->head) { in __wake_up_common() 87 list_for_each_entry_safe_from(curr, next, &wq_head->head, entry) in __wake_up_common() 89 unsigned flags = curr->flags; in __wake_up_common() 96 ret = curr->func(curr, mode, wake_flags, key); in __wake_up_common()
|
H A D | fair.c | 577 struct sched_entity *curr = cfs_rq->curr; in update_min_vruntime() local 582 if (curr) { in update_min_vruntime() 583 if (curr->on_rq) { in update_min_vruntime() 584 vruntime = curr->vruntime; in update_min_vruntime() 586 curr = NULL; in update_min_vruntime() 594 if (!curr) { in update_min_vruntime() 896 struct sched_entity *curr = cfs_rq->curr; in update_curr() local 900 if (unlikely(!curr)) { in update_curr() 3018 task_tick_numa(struct rq *rq, struct task_struct *curr) task_tick_numa() argument 3089 task_tick_numa(struct rq *rq, struct task_struct *curr) task_tick_numa() argument 4419 bool curr = cfs_rq->curr == se; enqueue_entity() local 4590 check_preempt_tick(struct cfs_rq *cfs_rq, struct sched_entity *curr) check_preempt_tick() argument 4668 pick_next_entity(struct cfs_rq *cfs_rq, struct sched_entity *curr) pick_next_entity() argument 4748 entity_tick(struct cfs_rq *cfs_rq, struct sched_entity *curr, int queued) entity_tick() argument 5718 struct task_struct *curr = rq->curr; hrtick_update() local 7263 wakeup_latency_gran(struct sched_entity *curr, struct sched_entity *se) wakeup_latency_gran() argument 7314 wakeup_preempt_entity(struct sched_entity *curr, struct sched_entity *se) wakeup_preempt_entity() argument 7371 struct task_struct *curr = rq->curr; check_preempt_wakeup() local 7485 struct sched_entity *curr = cfs_rq->curr; pick_next_task_fair() local 7638 struct task_struct *curr = rq->curr; yield_task_fair() local 11544 task_tick_fair(struct rq *rq, struct task_struct *curr, int queued) task_tick_fair() argument 11570 struct sched_entity *se = &p->se, *curr; task_fork_fair() local [all...] |
H A D | core.c | 372 rq->curr->sched_class->task_tick(rq, rq->curr, 1); in hrtick() 633 struct task_struct *curr = rq->curr; in resched_curr() local 638 if (test_tsk_need_resched(curr)) { in resched_curr() 644 set_tsk_need_resched(curr); in resched_curr() 649 if (set_nr_and_not_polling(curr)) { in resched_curr() 1842 if (p->sched_class == rq->curr->sched_class) { in check_preempt_curr() 1843 rq->curr->sched_class->check_preempt_curr(rq, p, flags); in check_preempt_curr() 1844 } else if (p->sched_class > rq->curr in check_preempt_curr() 3802 __fire_sched_in_preempt_notifiers(struct task_struct *curr) __fire_sched_in_preempt_notifiers() argument 3811 fire_sched_in_preempt_notifiers(struct task_struct *curr) fire_sched_in_preempt_notifiers() argument 3818 __fire_sched_out_preempt_notifiers(struct task_struct *curr, struct task_struct *next) __fire_sched_out_preempt_notifiers() argument 3828 fire_sched_out_preempt_notifiers(struct task_struct *curr, struct task_struct *next) fire_sched_out_preempt_notifiers() argument 3838 fire_sched_in_preempt_notifiers(struct task_struct *curr) fire_sched_in_preempt_notifiers() argument 3842 fire_sched_out_preempt_notifiers(struct task_struct *curr, struct task_struct *next) fire_sched_out_preempt_notifiers() argument 4323 struct sched_entity *curr = (&p->se)->cfs_rq->curr; prefetch_curr_exec_start() local 4384 struct task_struct *curr = rq->curr; scheduler_tick() local 4466 struct task_struct *curr; sched_tick_remote() local 5250 default_wake_function(wait_queue_entry_t *curr, unsigned mode, int wake_flags, void *key) default_wake_function() argument 6784 struct task_struct *curr = current; yield_to() local [all...] |
H A D | pelt.c | 249 * se has been already dequeued but cfs_rq->curr still points to it. in ___update_load_sum() 350 if (___update_load_sum(now, &se->avg, !!se->on_rq, se_runnable(se), cfs_rq->curr == se)) { in __update_load_avg_se() 363 cfs_rq->curr != NULL)) { in __update_load_avg_cfs_rq()
|
H A D | sched.h | 605 * 'curr' points to currently running entity on this cfs_rq. 608 struct sched_entity *curr; member 703 int curr; /* highest queued rt task prio */ member 752 u64 curr; member 1033 struct task_struct __rcu *curr; member 1216 #define cpu_curr(cpu) (cpu_rq(cpu)->curr) 1792 return rq->curr == p; in task_current() 1944 WARN_ON_ONCE(rq->curr != prev); in put_prev_task() 1950 WARN_ON_ONCE(rq->curr != next); in set_next_task() 2735 * In the same way it provides those guarantees around store to rq->curr [all...] |
/device/soc/rockchip/common/sdk_linux/kernel/power/ |
H A D | process.c | 211 struct task_struct *curr = current; in thaw_processes() local 233 WARN_ON((p != curr) && (p->flags & PF_SUSPEND_TASK)); in thaw_processes() 238 WARN_ON(!(curr->flags & PF_SUSPEND_TASK)); in thaw_processes() 239 curr->flags &= ~PF_SUSPEND_TASK; in thaw_processes()
|
H A D | snapshot.c | 719 struct mem_zone_bm_rtree *curr, *zone; in memory_bm_find_bit() local 732 list_for_each_entry(curr, &bm->zones, list) in memory_bm_find_bit() 734 if (pfn >= curr->start_pfn && pfn < curr->end_pfn) { in memory_bm_find_bit() 735 zone = curr; in memory_bm_find_bit()
|
/device/soc/rockchip/common/sdk_linux/ipc/ |
H A D | sem.c | 295 struct sem *curr;
in unmerge_queues() local 296 curr = &sma->sems[q->sops[0].sem_num];
in unmerge_queues() 298 list_add_tail(&q->list, &curr->pending_alter);
in unmerge_queues() 654 struct sem *curr;
in perform_atomic_semop_slow() local 664 curr = &sma->sems[idx];
in perform_atomic_semop_slow() 666 result = curr->semval;
in perform_atomic_semop_slow() 689 curr->semval = result;
in perform_atomic_semop_slow() 732 struct sem *curr;
in perform_atomic_semop() local 753 curr = &sma->sems[idx];
in perform_atomic_semop() 755 result = curr in perform_atomic_semop() 1380 struct sem *curr; semctl_setval() local 1439 struct sem *curr; semctl_main() local 2192 struct sem *curr; do_semtimedop() local [all...] |
/device/soc/rockchip/rk3588/kernel/include/trace/hooks/ |
H A D | sched.h | 178 TP_PROTO(struct task_struct *curr, int cpu, s64 delta), 179 TP_ARGS(curr, cpu, delta), 1); 252 TP_PROTO(struct cfs_rq *cfs_rq, struct sched_entity *curr, 254 TP_ARGS(cfs_rq, curr, se), 1); 283 unsigned long delta_exec, struct cfs_rq *cfs_rq, struct sched_entity *curr, 285 TP_ARGS(p, ideal_runtime, skip_preempt, delta_exec, cfs_rq, curr, granularity), 1);
|
/device/soc/rockchip/rk3588/kernel/drivers/video/rockchip/rga3/include/ |
H A D | rga2_mmu_info.h | 28 int32_t curr; member
|
/device/soc/rockchip/common/hardware/mpp/include/ |
H A D | mpp_rc_defs.h | 129 EncFrmStatus curr; member
|
/device/soc/rockchip/rk3399/hardware/mpp/include/ |
H A D | mpp_rc_defs.h | 129 EncFrmStatus curr; member
|
/device/soc/rockchip/rk3568/hardware/mpp/include/ |
H A D | mpp_rc_defs.h | 129 EncFrmStatus curr; member
|
/device/soc/rockchip/rk3588/hardware/mpp/include/ |
H A D | mpp_rc_defs.h | 148 EncFrmStatus curr; member
|
/device/soc/rockchip/common/sdk_linux/include/linux/usb/ |
H A D | pd_vdo.h | 471 #define VDO_VPD(hw, fw, ver, vbm, curr, vbi, gi, ct) \ 472 (((hw)&0xf) << 28 | ((fw)&0xf) << 24 | ((ver)&0x7) << 21 | ((vbm)&0x3) << 15 | (curr) << 14 | ((vbi)&0x3f) << 7 | \
|
/device/soc/rockchip/common/sdk_linux/drivers/power/supply/ |
H A D | power_supply_core.c | 421 int curr;
in power_supply_set_input_current_limit_from_supplier() local 432 curr = class_for_each_device(power_supply_class, NULL, psy, power_supply_get_supplier_max_current_ext);
in power_supply_set_input_current_limit_from_supplier() 433 if (curr <= 0) {
in power_supply_set_input_current_limit_from_supplier() 434 return (curr == 0) ? -ENODEV : curr;
in power_supply_set_input_current_limit_from_supplier() 437 val.intval = curr;
in power_supply_set_input_current_limit_from_supplier()
|
/device/soc/rockchip/common/sdk_linux/include/drm/bridge/ |
H A D | dw_hdmi.h | 115 u16 curr[DW_HDMI_RES_MAX]; member
|
/device/soc/rockchip/common/kernel/drivers/net/wireless/rockchip_wlan/rkwifi/bcmdhd_wifi6/include/ |
H A D | bcmutils.h | 548 #define DELTA(curr, prev) ((curr) > (prev) ? ((curr) - (prev)) : \ 549 (0xffffffff - (prev) + (curr) + 1))
|
/device/soc/rockchip/common/kernel/drivers/net/wireless/rockchip_wlan/rkwifi/bcmdhd_wifi6/ |
H A D | dhd_pcie.c | 4604 uint32 curr[BCM_BUZZZ_COUNTERS_MAX], prev[BCM_BUZZZ_COUNTERS_MAX]; in dhd_buzzz_dump_cntrs() local 4610 curr[ctr] = *log++; in dhd_buzzz_dump_cntrs() 4611 core[ctr] = curr[ctr]; /* saved for next log */ in dhd_buzzz_dump_cntrs() 4613 if (curr[ctr] < prev[ctr]) in dhd_buzzz_dump_cntrs() 4614 delta[ctr] = curr[ctr] + (~0U - prev[ctr]); in dhd_buzzz_dump_cntrs() 4616 delta[ctr] = (curr[ctr] - prev[ctr]); in dhd_buzzz_dump_cntrs() 4645 uint32 curr, prev, delta; in dhd_bcm_buzzz_dump_cntrs6() local 4646 prev = core[0]; curr = *log++; core[0] = curr; in dhd_bcm_buzzz_dump_cntrs6() 4647 if (curr < pre in dhd_bcm_buzzz_dump_cntrs6() 4659 cm3_cnts_t curr, prev, delta; dhd_bcm_buzzz_dump_cntrs6() local 4672 uint8 curr, prev, delta, max8 = ~0; dhd_bcm_buzzz_dump_cntrs6() local [all...] |
/device/soc/rockchip/common/vendor/drivers/video/rockchip/rga2/ |
H A D | rga2.h | 568 int32_t curr;
member
|