Home
last modified time | relevance | path

Searched refs:tg (Results 1 - 6 of 6) sorted by relevance

/device/soc/rockchip/common/sdk_linux/kernel/sched/
H A Drt.c149 void free_rt_sched_group(struct task_group *tg) in free_rt_sched_group() argument
153 if (tg->rt_se) { in free_rt_sched_group()
154 destroy_rt_bandwidth(&tg->rt_bandwidth); in free_rt_sched_group()
159 if (tg->rt_rq) { in free_rt_sched_group()
160 kfree(tg->rt_rq[i]); in free_rt_sched_group()
162 if (tg->rt_se) { in free_rt_sched_group()
163 kfree(tg->rt_se[i]); in free_rt_sched_group()
167 kfree(tg->rt_rq); in free_rt_sched_group()
168 kfree(tg->rt_se); in free_rt_sched_group()
171 void init_tg_rt_entry(struct task_group *tg, struc argument
199 alloc_rt_sched_group(struct task_group *tg, struct task_group *parent) alloc_rt_sched_group() argument
269 free_rt_sched_group(struct task_group *tg) free_rt_sched_group() argument
273 alloc_rt_sched_group(struct task_group *tg, struct task_group *parent) alloc_rt_sched_group() argument
527 next_task_group(struct task_group *tg) next_task_group() argument
2882 tg_has_rt_tasks(struct task_group *tg) tg_has_rt_tasks() argument
2905 struct task_group *tg; global() member
2910 tg_rt_schedulable(struct task_group *tg, void *data) tg_rt_schedulable() argument
2970 __rt_schedulable(struct task_group *tg, u64 period, u64 runtime) __rt_schedulable() argument
2987 tg_set_rt_bandwidth(struct task_group *tg, u64 rt_period, u64 rt_runtime) tg_set_rt_bandwidth() argument
3036 sched_group_set_rt_runtime(struct task_group *tg, long rt_runtime_us) sched_group_set_rt_runtime() argument
3051 sched_group_rt_runtime(struct task_group *tg) sched_group_rt_runtime() argument
3064 sched_group_set_rt_period(struct task_group *tg, u64 rt_period_us) sched_group_set_rt_period() argument
3078 sched_group_rt_period(struct task_group *tg) sched_group_rt_period() argument
3098 sched_rt_can_attach(struct task_group *tg, struct task_struct *tsk) sched_rt_can_attach() argument
[all...]
H A Dcore.c902 int tg_nop(struct task_group *tg, void *data) in tg_nop() argument
1470 struct task_group *tg = &root_task_group; in uclamp_update_root_tg() local
1472 uclamp_se_set(&tg->uclamp_req[UCLAMP_MIN], sysctl_sched_uclamp_util_min, in uclamp_update_root_tg()
1474 uclamp_se_set(&tg->uclamp_req[UCLAMP_MAX], sysctl_sched_uclamp_util_max, in uclamp_update_root_tg()
3666 struct task_group *tg; in sched_post_fork() local
3671 tg = container_of(kargs->cset->subsys[cpu_cgrp_id], struct task_group, css); in sched_post_fork()
3672 p->sched_task_group = autogroup_task_group(p, tg); in sched_post_fork()
8427 static inline void alloc_uclamp_sched_group(struct task_group *tg, in alloc_uclamp_sched_group() argument
8434 uclamp_se_set(&tg->uclamp_req[clamp_id], uclamp_none(clamp_id), false); in alloc_uclamp_sched_group()
8435 tg in alloc_uclamp_sched_group()
8440 sched_free_group(struct task_group *tg) sched_free_group() argument
8451 struct task_group *tg; sched_create_group() local
8475 sched_online_group(struct task_group *tg, struct task_group *parent) sched_online_group() argument
8500 sched_destroy_group(struct task_group *tg) sched_destroy_group() argument
8506 sched_offline_group(struct task_group *tg) sched_offline_group() argument
8521 struct task_group *tg; sched_change_group() local
8594 struct task_group *tg; cpu_cgroup_css_alloc() local
8617 struct task_group *tg = css_tg(css); cpu_cgroup_css_online() local
8638 struct task_group *tg = css_tg(css); cpu_cgroup_css_released() local
8645 struct task_group *tg = css_tg(css); cpu_cgroup_css_free() local
8711 struct task_group *tg; schedgp_attach() local
8832 struct task_group *tg; cpu_uclamp_write() local
8879 struct task_group *tg; cpu_uclamp_print() local
8915 struct task_group *tg = css_tg(css); sched_colocate_read() local
8923 struct task_group *tg = css_tg(css); sched_colocate_write() local
8950 struct task_group *tg = css_tg(css); cpu_shares_read_u64() local
8965 tg_set_cfs_bandwidth(struct task_group *tg, u64 period, u64 quota) tg_set_cfs_bandwidth() argument
9057 tg_set_cfs_quota(struct task_group *tg, long cfs_quota_us) tg_set_cfs_quota() argument
9073 tg_get_cfs_quota(struct task_group *tg) tg_get_cfs_quota() argument
9087 tg_set_cfs_period(struct task_group *tg, long cfs_period_us) tg_set_cfs_period() argument
9101 tg_get_cfs_period(struct task_group *tg) tg_get_cfs_period() argument
9136 struct task_group *tg; global() member
9144 normalize_cfs_quota(struct task_group *tg, struct cfs_schedulable_data *d) normalize_cfs_quota() argument
9165 tg_cfs_schedulable_down(struct task_group *tg, void *data) tg_cfs_schedulable_down() argument
9199 __cfs_schedulable(struct task_group *tg, u64 period, u64 quota) __cfs_schedulable() argument
9222 struct task_group *tg = css_tg(seq_css(sf)); cpu_cfs_stat_show() local
9336 struct task_group *tg = css_tg(css); cpu_extra_stat_show() local
9357 struct task_group *tg = css_tg(css); cpu_weight_read_u64() local
9457 struct task_group *tg = css_tg(seq_css(sf)); cpu_max_show() local
9466 struct task_group *tg = css_tg(of_css(of)); cpu_max_write() local
[all...]
H A Dsched.h539 extern int tg_nop(struct task_group *tg, void *data);
541 extern void free_fair_sched_group(struct task_group *tg);
542 extern int alloc_fair_sched_group(struct task_group *tg, struct task_group *parent);
543 extern void online_fair_sched_group(struct task_group *tg);
544 extern void unregister_fair_sched_group(struct task_group *tg);
545 extern void init_tg_cfs_entry(struct task_group *tg, struct cfs_rq *cfs_rq, struct sched_entity *se, int cpu,
553 extern void free_rt_sched_group(struct task_group *tg);
554 extern int alloc_rt_sched_group(struct task_group *tg, struct task_group *parent);
555 extern void init_tg_rt_entry(struct task_group *tg, struct rt_rq *rt_rq, struct sched_rt_entity *rt_se, int cpu,
557 extern int sched_group_set_rt_runtime(struct task_group *tg, lon
663 struct task_group *tg; /* group that "owns" this runqueue */ global() member
728 struct task_group *tg; global() member
1664 struct task_group *tg = task_group(p); set_task_rq() local
[all...]
H A Dfair.c159 * Amount of runtime to allocate from global (tg) to local (per-cfs_rq) pool
332 if (cfs_rq && task_group_is_autogroup(cfs_rq->tg)) { in cfs_rq_tg_path()
333 autogroup_path(cfs_rq->tg, path, len); in cfs_rq_tg_path()
334 } else if (cfs_rq && cfs_rq->tg->css.cgroup) { in cfs_rq_tg_path()
335 cgroup_path(cfs_rq->tg->css.cgroup, path, len); in cfs_rq_tg_path()
361 if (cfs_rq->tg->parent && cfs_rq->tg->parent->cfs_rq[cpu]->on_list) { in list_add_leaf_cfs_rq()
368 list_add_tail_rcu(&cfs_rq->leaf_cfs_rq_list, &(cfs_rq->tg->parent->cfs_rq[cpu]->leaf_cfs_rq_list)); in list_add_leaf_cfs_rq()
378 if (!cfs_rq->tg->parent) { in list_add_leaf_cfs_rq()
3252 * tg
3321 struct task_group *tg = cfs_rq->tg; calc_group_shares() local
4848 tg_cfs_bandwidth(struct task_group *tg) tg_cfs_bandwidth() argument
4939 throttled_lb_pair(struct task_group *tg, int src_cpu, int dest_cpu) throttled_lb_pair() argument
4949 tg_unthrottle_up(struct task_group *tg, void *data) tg_unthrottle_up() argument
4967 tg_throttle_down(struct task_group *tg, void *data) tg_throttle_down() argument
5409 sync_throttle(struct task_group *tg, int cpu) sync_throttle() argument
5568 struct task_group *tg; update_runtime_enabled() local
5588 struct task_group *tg; unthrottle_offline_cfs_rqs() local
5636 sync_throttle(struct task_group *tg, int cpu) sync_throttle() argument
5653 throttled_lb_pair(struct task_group *tg, int src_cpu, int dest_cpu) throttled_lb_pair() argument
5668 tg_cfs_bandwidth(struct task_group *tg) tg_cfs_bandwidth() argument
11844 free_fair_sched_group(struct task_group *tg) free_fair_sched_group() argument
11864 alloc_fair_sched_group(struct task_group *tg, struct task_group *parent) alloc_fair_sched_group() argument
11908 online_fair_sched_group(struct task_group *tg) online_fair_sched_group() argument
11927 unregister_fair_sched_group(struct task_group *tg) unregister_fair_sched_group() argument
11955 init_tg_cfs_entry(struct task_group *tg, struct cfs_rq *cfs_rq, struct sched_entity *se, int cpu, struct sched_entity *parent) init_tg_cfs_entry() argument
11988 sched_group_set_shares(struct task_group *tg, unsigned long shares) sched_group_set_shares() argument
12028 free_fair_sched_group(struct task_group *tg) free_fair_sched_group() argument
12032 alloc_fair_sched_group(struct task_group *tg, struct task_group *parent) alloc_fair_sched_group() argument
12037 online_fair_sched_group(struct task_group *tg) online_fair_sched_group() argument
12041 unregister_fair_sched_group(struct task_group *tg) unregister_fair_sched_group() argument
[all...]
/device/soc/rockchip/common/sdk_linux/include/drm/
H A Ddrm_connector.h1691 void drm_mode_put_tile_group(struct drm_device *dev, struct drm_tile_group *tg);
/device/soc/rockchip/common/sdk_linux/drivers/gpu/drm/
H A Ddrm_edid.c6029 struct drm_tile_group *tg; in drm_parse_tiled_block() local
6056 tg = drm_mode_get_tile_group(connector->dev, tile->topology_id); in drm_parse_tiled_block()
6057 if (!tg) { in drm_parse_tiled_block()
6058 tg = drm_mode_create_tile_group(connector->dev, tile->topology_id); in drm_parse_tiled_block()
6060 if (!tg) { in drm_parse_tiled_block()
6064 if (connector->tile_group != tg) { in drm_parse_tiled_block()
6070 connector->tile_group = tg; in drm_parse_tiled_block()
6073 drm_mode_put_tile_group(connector->dev, tg); in drm_parse_tiled_block()

Completed in 50 milliseconds