/device/soc/rockchip/common/sdk_linux/kernel/sched/ |
H A D | rt.c | 149 void free_rt_sched_group(struct task_group *tg) in free_rt_sched_group() argument 153 if (tg->rt_se) { in free_rt_sched_group() 154 destroy_rt_bandwidth(&tg->rt_bandwidth); in free_rt_sched_group() 159 if (tg->rt_rq) { in free_rt_sched_group() 160 kfree(tg->rt_rq[i]); in free_rt_sched_group() 162 if (tg->rt_se) { in free_rt_sched_group() 163 kfree(tg->rt_se[i]); in free_rt_sched_group() 167 kfree(tg->rt_rq); in free_rt_sched_group() 168 kfree(tg->rt_se); in free_rt_sched_group() 171 void init_tg_rt_entry(struct task_group *tg, struc argument 199 alloc_rt_sched_group(struct task_group *tg, struct task_group *parent) alloc_rt_sched_group() argument 269 free_rt_sched_group(struct task_group *tg) free_rt_sched_group() argument 273 alloc_rt_sched_group(struct task_group *tg, struct task_group *parent) alloc_rt_sched_group() argument 527 next_task_group(struct task_group *tg) next_task_group() argument 2882 tg_has_rt_tasks(struct task_group *tg) tg_has_rt_tasks() argument 2905 struct task_group *tg; global() member 2910 tg_rt_schedulable(struct task_group *tg, void *data) tg_rt_schedulable() argument 2970 __rt_schedulable(struct task_group *tg, u64 period, u64 runtime) __rt_schedulable() argument 2987 tg_set_rt_bandwidth(struct task_group *tg, u64 rt_period, u64 rt_runtime) tg_set_rt_bandwidth() argument 3036 sched_group_set_rt_runtime(struct task_group *tg, long rt_runtime_us) sched_group_set_rt_runtime() argument 3051 sched_group_rt_runtime(struct task_group *tg) sched_group_rt_runtime() argument 3064 sched_group_set_rt_period(struct task_group *tg, u64 rt_period_us) sched_group_set_rt_period() argument 3078 sched_group_rt_period(struct task_group *tg) sched_group_rt_period() argument 3098 sched_rt_can_attach(struct task_group *tg, struct task_struct *tsk) sched_rt_can_attach() argument [all...] |
H A D | core.c | 902 int tg_nop(struct task_group *tg, void *data) in tg_nop() argument 1470 struct task_group *tg = &root_task_group; in uclamp_update_root_tg() local 1472 uclamp_se_set(&tg->uclamp_req[UCLAMP_MIN], sysctl_sched_uclamp_util_min, in uclamp_update_root_tg() 1474 uclamp_se_set(&tg->uclamp_req[UCLAMP_MAX], sysctl_sched_uclamp_util_max, in uclamp_update_root_tg() 3666 struct task_group *tg; in sched_post_fork() local 3671 tg = container_of(kargs->cset->subsys[cpu_cgrp_id], struct task_group, css); in sched_post_fork() 3672 p->sched_task_group = autogroup_task_group(p, tg); in sched_post_fork() 8427 static inline void alloc_uclamp_sched_group(struct task_group *tg, in alloc_uclamp_sched_group() argument 8434 uclamp_se_set(&tg->uclamp_req[clamp_id], uclamp_none(clamp_id), false); in alloc_uclamp_sched_group() 8435 tg in alloc_uclamp_sched_group() 8440 sched_free_group(struct task_group *tg) sched_free_group() argument 8451 struct task_group *tg; sched_create_group() local 8475 sched_online_group(struct task_group *tg, struct task_group *parent) sched_online_group() argument 8500 sched_destroy_group(struct task_group *tg) sched_destroy_group() argument 8506 sched_offline_group(struct task_group *tg) sched_offline_group() argument 8521 struct task_group *tg; sched_change_group() local 8594 struct task_group *tg; cpu_cgroup_css_alloc() local 8617 struct task_group *tg = css_tg(css); cpu_cgroup_css_online() local 8638 struct task_group *tg = css_tg(css); cpu_cgroup_css_released() local 8645 struct task_group *tg = css_tg(css); cpu_cgroup_css_free() local 8711 struct task_group *tg; schedgp_attach() local 8832 struct task_group *tg; cpu_uclamp_write() local 8879 struct task_group *tg; cpu_uclamp_print() local 8915 struct task_group *tg = css_tg(css); sched_colocate_read() local 8923 struct task_group *tg = css_tg(css); sched_colocate_write() local 8950 struct task_group *tg = css_tg(css); cpu_shares_read_u64() local 8965 tg_set_cfs_bandwidth(struct task_group *tg, u64 period, u64 quota) tg_set_cfs_bandwidth() argument 9057 tg_set_cfs_quota(struct task_group *tg, long cfs_quota_us) tg_set_cfs_quota() argument 9073 tg_get_cfs_quota(struct task_group *tg) tg_get_cfs_quota() argument 9087 tg_set_cfs_period(struct task_group *tg, long cfs_period_us) tg_set_cfs_period() argument 9101 tg_get_cfs_period(struct task_group *tg) tg_get_cfs_period() argument 9136 struct task_group *tg; global() member 9144 normalize_cfs_quota(struct task_group *tg, struct cfs_schedulable_data *d) normalize_cfs_quota() argument 9165 tg_cfs_schedulable_down(struct task_group *tg, void *data) tg_cfs_schedulable_down() argument 9199 __cfs_schedulable(struct task_group *tg, u64 period, u64 quota) __cfs_schedulable() argument 9222 struct task_group *tg = css_tg(seq_css(sf)); cpu_cfs_stat_show() local 9336 struct task_group *tg = css_tg(css); cpu_extra_stat_show() local 9357 struct task_group *tg = css_tg(css); cpu_weight_read_u64() local 9457 struct task_group *tg = css_tg(seq_css(sf)); cpu_max_show() local 9466 struct task_group *tg = css_tg(of_css(of)); cpu_max_write() local [all...] |
H A D | sched.h | 539 extern int tg_nop(struct task_group *tg, void *data); 541 extern void free_fair_sched_group(struct task_group *tg); 542 extern int alloc_fair_sched_group(struct task_group *tg, struct task_group *parent); 543 extern void online_fair_sched_group(struct task_group *tg); 544 extern void unregister_fair_sched_group(struct task_group *tg); 545 extern void init_tg_cfs_entry(struct task_group *tg, struct cfs_rq *cfs_rq, struct sched_entity *se, int cpu, 553 extern void free_rt_sched_group(struct task_group *tg); 554 extern int alloc_rt_sched_group(struct task_group *tg, struct task_group *parent); 555 extern void init_tg_rt_entry(struct task_group *tg, struct rt_rq *rt_rq, struct sched_rt_entity *rt_se, int cpu, 557 extern int sched_group_set_rt_runtime(struct task_group *tg, lon 663 struct task_group *tg; /* group that "owns" this runqueue */ global() member 728 struct task_group *tg; global() member 1664 struct task_group *tg = task_group(p); set_task_rq() local [all...] |
H A D | fair.c | 159 * Amount of runtime to allocate from global (tg) to local (per-cfs_rq) pool 332 if (cfs_rq && task_group_is_autogroup(cfs_rq->tg)) { in cfs_rq_tg_path() 333 autogroup_path(cfs_rq->tg, path, len); in cfs_rq_tg_path() 334 } else if (cfs_rq && cfs_rq->tg->css.cgroup) { in cfs_rq_tg_path() 335 cgroup_path(cfs_rq->tg->css.cgroup, path, len); in cfs_rq_tg_path() 361 if (cfs_rq->tg->parent && cfs_rq->tg->parent->cfs_rq[cpu]->on_list) { in list_add_leaf_cfs_rq() 368 list_add_tail_rcu(&cfs_rq->leaf_cfs_rq_list, &(cfs_rq->tg->parent->cfs_rq[cpu]->leaf_cfs_rq_list)); in list_add_leaf_cfs_rq() 378 if (!cfs_rq->tg->parent) { in list_add_leaf_cfs_rq() 3252 * tg 3321 struct task_group *tg = cfs_rq->tg; calc_group_shares() local 4848 tg_cfs_bandwidth(struct task_group *tg) tg_cfs_bandwidth() argument 4939 throttled_lb_pair(struct task_group *tg, int src_cpu, int dest_cpu) throttled_lb_pair() argument 4949 tg_unthrottle_up(struct task_group *tg, void *data) tg_unthrottle_up() argument 4967 tg_throttle_down(struct task_group *tg, void *data) tg_throttle_down() argument 5409 sync_throttle(struct task_group *tg, int cpu) sync_throttle() argument 5568 struct task_group *tg; update_runtime_enabled() local 5588 struct task_group *tg; unthrottle_offline_cfs_rqs() local 5636 sync_throttle(struct task_group *tg, int cpu) sync_throttle() argument 5653 throttled_lb_pair(struct task_group *tg, int src_cpu, int dest_cpu) throttled_lb_pair() argument 5668 tg_cfs_bandwidth(struct task_group *tg) tg_cfs_bandwidth() argument 11844 free_fair_sched_group(struct task_group *tg) free_fair_sched_group() argument 11864 alloc_fair_sched_group(struct task_group *tg, struct task_group *parent) alloc_fair_sched_group() argument 11908 online_fair_sched_group(struct task_group *tg) online_fair_sched_group() argument 11927 unregister_fair_sched_group(struct task_group *tg) unregister_fair_sched_group() argument 11955 init_tg_cfs_entry(struct task_group *tg, struct cfs_rq *cfs_rq, struct sched_entity *se, int cpu, struct sched_entity *parent) init_tg_cfs_entry() argument 11988 sched_group_set_shares(struct task_group *tg, unsigned long shares) sched_group_set_shares() argument 12028 free_fair_sched_group(struct task_group *tg) free_fair_sched_group() argument 12032 alloc_fair_sched_group(struct task_group *tg, struct task_group *parent) alloc_fair_sched_group() argument 12037 online_fair_sched_group(struct task_group *tg) online_fair_sched_group() argument 12041 unregister_fair_sched_group(struct task_group *tg) unregister_fair_sched_group() argument [all...] |
/device/soc/rockchip/common/sdk_linux/include/drm/ |
H A D | drm_connector.h | 1691 void drm_mode_put_tile_group(struct drm_device *dev, struct drm_tile_group *tg);
|
/device/soc/rockchip/common/sdk_linux/drivers/gpu/drm/ |
H A D | drm_edid.c | 6029 struct drm_tile_group *tg; in drm_parse_tiled_block() local 6056 tg = drm_mode_get_tile_group(connector->dev, tile->topology_id); in drm_parse_tiled_block() 6057 if (!tg) { in drm_parse_tiled_block() 6058 tg = drm_mode_create_tile_group(connector->dev, tile->topology_id); in drm_parse_tiled_block() 6060 if (!tg) { in drm_parse_tiled_block() 6064 if (connector->tile_group != tg) { in drm_parse_tiled_block() 6070 connector->tile_group = tg; in drm_parse_tiled_block() 6073 drm_mode_put_tile_group(connector->dev, tg); in drm_parse_tiled_block()
|