/kernel/linux/linux-6.6/kernel/sched/ |
H A D | deadline.c | 228 __dl_overflow(struct dl_bw *dl_b, unsigned long cap, u64 old_bw, u64 new_bw) in __dl_overflow() argument 231 cap_scale(dl_b->bw, cap) < dl_b->total_bw - old_bw + new_bw; in __dl_overflow() 312 static void dl_change_utilization(struct task_struct *p, u64 new_bw) in dl_change_utilization() argument 336 __add_rq_bw(new_bw, &rq->dl); in dl_change_utilization() 2770 u64 new_bw = to_ratio(period, runtime); in sched_dl_global_validate() local 2791 if (new_bw * cpus < dl_b->total_bw) in sched_dl_global_validate() 2820 u64 new_bw = -1; in sched_dl_do_global() local 2827 new_bw = to_ratio(global_rt_period(), global_rt_runtime()); in sched_dl_do_global() 2840 dl_b->bw = new_bw; in sched_dl_do_global() 2861 u64 new_bw in sched_dl_overflow() local [all...] |
/kernel/linux/linux-5.10/kernel/sched/ |
H A D | deadline.c | 219 static void dl_change_utilization(struct task_struct *p, u64 new_bw) in dl_change_utilization() argument 243 __add_rq_bw(new_bw, &rq->dl); in dl_change_utilization() 2577 u64 new_bw = to_ratio(period, runtime); in sched_dl_global_validate() local 2597 if (new_bw * cpus < dl_b->total_bw) in sched_dl_global_validate() 2625 u64 new_bw = -1; in sched_dl_do_global() local 2634 new_bw = to_ratio(global_rt_period(), global_rt_runtime()); in sched_dl_do_global() 2644 dl_b->bw = new_bw; in sched_dl_do_global() 2665 u64 new_bw = dl_policy(policy) ? to_ratio(period, runtime) : 0; in sched_dl_overflow() local 2674 if (new_bw == p->dl.dl_bw && task_has_dl_policy(p)) in sched_dl_overflow() 2687 !__dl_overflow(dl_b, cap, 0, new_bw)) { in sched_dl_overflow() [all...] |
H A D | sched.h | 403 u64 old_bw, u64 new_bw) in __dl_overflow() 406 cap_scale(dl_b->bw, cap) < dl_b->total_bw - old_bw + new_bw; in __dl_overflow() 402 __dl_overflow(struct dl_bw *dl_b, unsigned long cap, u64 old_bw, u64 new_bw) __dl_overflow() argument
|
/kernel/linux/linux-5.10/net/mac80211/ |
H A D | vht.c | 558 enum ieee80211_sta_rx_bandwidth new_bw; in __ieee80211_vht_handle_opmode() local 599 new_bw = ieee80211_sta_cur_vht_bw(sta); in __ieee80211_vht_handle_opmode() 600 if (new_bw != sta->sta.bandwidth) { in __ieee80211_vht_handle_opmode() 601 sta->sta.bandwidth = new_bw; in __ieee80211_vht_handle_opmode()
|
H A D | rx.c | 3297 enum ieee80211_sta_rx_bandwidth max_bw, new_bw; in ieee80211_rx_h_action() local 3312 new_bw = ieee80211_sta_cur_vht_bw(rx->sta); in ieee80211_rx_h_action() 3314 if (rx->sta->sta.bandwidth == new_bw) in ieee80211_rx_h_action() 3317 rx->sta->sta.bandwidth = new_bw; in ieee80211_rx_h_action()
|
/kernel/linux/linux-6.6/net/mac80211/ |
H A D | vht.c | 637 enum ieee80211_sta_rx_bandwidth new_bw; in __ieee80211_vht_handle_opmode() local 693 new_bw = ieee80211_sta_cur_vht_bw(link_sta); in __ieee80211_vht_handle_opmode() 694 if (new_bw != link_sta->pub->bandwidth) { in __ieee80211_vht_handle_opmode() 695 link_sta->pub->bandwidth = new_bw; in __ieee80211_vht_handle_opmode()
|
H A D | rx.c | 3540 enum ieee80211_sta_rx_bandwidth max_bw, new_bw; in ieee80211_rx_h_action() local 3555 new_bw = ieee80211_sta_cur_vht_bw(rx->link_sta); in ieee80211_rx_h_action() 3557 if (rx->link_sta->pub->bandwidth == new_bw) in ieee80211_rx_h_action() 3560 rx->link_sta->pub->bandwidth = new_bw; in ieee80211_rx_h_action()
|
/kernel/linux/linux-5.10/drivers/edac/ |
H A D | edac_mc_sysfs.c | 706 int new_bw = 0; in mci_sdram_scrub_rate_store() local 711 new_bw = mci->set_sdram_scrub_rate(mci, bandwidth); in mci_sdram_scrub_rate_store() 712 if (new_bw < 0) { in mci_sdram_scrub_rate_store()
|
H A D | i7core_edac.c | 1957 static int set_sdram_scrub_rate(struct mem_ctl_info *mci, u32 new_bw) in set_sdram_scrub_rate() argument 1971 if (new_bw == 0) { in set_sdram_scrub_rate() 1992 do_div(scrub_interval, new_bw); in set_sdram_scrub_rate() 2011 return new_bw; in set_sdram_scrub_rate()
|
H A D | e752x_edac.c | 989 static int set_sdram_scrub_rate(struct mem_ctl_info *mci, u32 new_bw) in set_sdram_scrub_rate() argument 1006 if (scrubrates[i].bandwidth >= new_bw) in set_sdram_scrub_rate()
|
H A D | amd64_edac.c | 189 static int __set_scrub_rate(struct amd64_pvt *pvt, u32 new_bw, u32 min_rate) in __set_scrub_rate() argument 195 * map the configured rate (new_bw) to a value specific to the AMD64 in __set_scrub_rate() 211 if (scrubrates[i].bandwidth <= new_bw) in __set_scrub_rate()
|
/kernel/linux/linux-6.6/drivers/edac/ |
H A D | edac_mc_sysfs.c | 730 int new_bw = 0; in mci_sdram_scrub_rate_store() local 735 new_bw = mci->set_sdram_scrub_rate(mci, bandwidth); in mci_sdram_scrub_rate_store() 736 if (new_bw < 0) { in mci_sdram_scrub_rate_store()
|
H A D | i7core_edac.c | 1957 static int set_sdram_scrub_rate(struct mem_ctl_info *mci, u32 new_bw) in set_sdram_scrub_rate() argument 1971 if (new_bw == 0) { in set_sdram_scrub_rate() 1992 do_div(scrub_interval, new_bw); in set_sdram_scrub_rate() 2011 return new_bw; in set_sdram_scrub_rate()
|
H A D | e752x_edac.c | 988 static int set_sdram_scrub_rate(struct mem_ctl_info *mci, u32 new_bw) in set_sdram_scrub_rate() argument 1005 if (scrubrates[i].bandwidth >= new_bw) in set_sdram_scrub_rate()
|
H A D | amd64_edac.c | 187 static int __set_scrub_rate(struct amd64_pvt *pvt, u32 new_bw, u32 min_rate) in __set_scrub_rate() argument 193 * map the configured rate (new_bw) to a value specific to the AMD64 in __set_scrub_rate() 209 if (scrubrates[i].bandwidth <= new_bw) in __set_scrub_rate()
|
/kernel/linux/linux-5.10/drivers/thunderbolt/ |
H A D | tunnel.c | 481 u32 new_rate, new_lanes, new_bw; in tb_dp_xchg_caps() local 491 new_bw = tb_dp_bandwidth(new_rate, new_lanes); in tb_dp_xchg_caps() 493 new_rate, new_lanes, new_bw); in tb_dp_xchg_caps()
|
/kernel/linux/linux-6.6/drivers/thunderbolt/ |
H A D | tunnel.c | 636 u32 new_rate, new_lanes, new_bw; in tb_dp_xchg_caps() local 646 new_bw = tb_dp_bandwidth(new_rate, new_lanes); in tb_dp_xchg_caps() 648 new_rate, new_lanes, new_bw); in tb_dp_xchg_caps()
|