/kernel/linux/linux-5.10/net/netfilter/ |
H A D | xt_limit.c | 105 || user2credits(r->avg * r->burst) < user2credits(r->avg)) { in limit_mt_check() 107 r->avg, r->burst); in limit_mt_check() 117 /* User avg in seconds * XT_LIMIT_SCALE: convert to jiffies * in limit_mt_check() 120 priv->credit = user2credits(r->avg * r->burst); /* Credits full. */ in limit_mt_check() 123 r->cost = user2credits(r->avg); in limit_mt_check() 139 u_int32_t avg; member 155 .avg = cm->avg, in limit_mt_compat_from_user() 169 .avg in limit_mt_compat_to_user() [all...] |
/kernel/linux/linux-6.6/net/netfilter/ |
H A D | xt_limit.c | 112 || user2credits(r->avg * r->burst) < user2credits(r->avg)) { in limit_mt_check() 114 r->avg, r->burst); in limit_mt_check() 124 /* User avg in seconds * XT_LIMIT_SCALE: convert to jiffies * in limit_mt_check() 127 priv->credit = user2credits(r->avg * r->burst); /* Credits full. */ in limit_mt_check() 130 r->cost = user2credits(r->avg); in limit_mt_check() 145 u_int32_t avg; member 161 .avg = cm->avg, in limit_mt_compat_from_user() 175 .avg in limit_mt_compat_to_user() [all...] |
/kernel/linux/linux-5.10/tools/perf/util/ |
H A D | stat-shadow.c | 450 struct evsel *evsel, double avg, in print_stalled_cycles_frontend() 461 ratio = avg / total * 100.0; in print_stalled_cycles_frontend() 474 struct evsel *evsel, double avg, in print_stalled_cycles_backend() 485 ratio = avg / total * 100.0; in print_stalled_cycles_backend() 495 double avg, in print_branch_misses() 506 ratio = avg / total * 100.0; in print_branch_misses() 516 double avg, in print_l1_dcache_misses() 528 ratio = avg / total * 100.0; in print_l1_dcache_misses() 538 double avg, in print_l1_icache_misses() 550 ratio = avg / tota in print_l1_icache_misses() 448 print_stalled_cycles_frontend(struct perf_stat_config *config, int cpu, struct evsel *evsel, double avg, struct perf_stat_output_ctx *out, struct runtime_stat *st) print_stalled_cycles_frontend() argument 472 print_stalled_cycles_backend(struct perf_stat_config *config, int cpu, struct evsel *evsel, double avg, struct perf_stat_output_ctx *out, struct runtime_stat *st) print_stalled_cycles_backend() argument 492 print_branch_misses(struct perf_stat_config *config, int cpu, struct evsel *evsel, double avg, struct perf_stat_output_ctx *out, struct runtime_stat *st) print_branch_misses() argument 513 print_l1_dcache_misses(struct perf_stat_config *config, int cpu, struct evsel *evsel, double avg, struct perf_stat_output_ctx *out, struct runtime_stat *st) print_l1_dcache_misses() argument 535 print_l1_icache_misses(struct perf_stat_config *config, int cpu, struct evsel *evsel, double avg, struct perf_stat_output_ctx *out, struct runtime_stat *st) print_l1_icache_misses() argument 556 print_dtlb_cache_misses(struct perf_stat_config *config, int cpu, struct evsel *evsel, double avg, struct perf_stat_output_ctx *out, struct runtime_stat *st) print_dtlb_cache_misses() argument 576 print_itlb_cache_misses(struct perf_stat_config *config, int cpu, struct evsel *evsel, double avg, struct perf_stat_output_ctx *out, struct runtime_stat *st) print_itlb_cache_misses() argument 596 print_ll_cache_misses(struct perf_stat_config *config, int cpu, struct evsel *evsel, double avg, struct perf_stat_output_ctx *out, struct runtime_stat *st) print_ll_cache_misses() argument 922 perf_stat__print_shadow_stats(struct perf_stat_config *config, struct evsel *evsel, double avg, int cpu, struct perf_stat_output_ctx *out, struct rblist *metric_events, struct runtime_stat *st) perf_stat__print_shadow_stats() argument [all...] |
/kernel/linux/linux-5.10/net/bridge/netfilter/ |
H A D | ebt_limit.c | 75 user2credits(info->avg * info->burst) < user2credits(info->avg)) { in ebt_limit_mt_check() 77 info->avg, info->burst); in ebt_limit_mt_check() 81 /* User avg in seconds * EBT_LIMIT_SCALE: convert to jiffies * 128. */ in ebt_limit_mt_check() 83 info->credit = user2credits(info->avg * info->burst); in ebt_limit_mt_check() 84 info->credit_cap = user2credits(info->avg * info->burst); in ebt_limit_mt_check() 85 info->cost = user2credits(info->avg); in ebt_limit_mt_check() 93 * only avg/burst have meaningful values in userspace. 96 compat_uint_t avg, burst; member
|
/kernel/linux/linux-6.6/net/bridge/netfilter/ |
H A D | ebt_limit.c | 75 user2credits(info->avg * info->burst) < user2credits(info->avg)) { in ebt_limit_mt_check() 77 info->avg, info->burst); in ebt_limit_mt_check() 81 /* User avg in seconds * EBT_LIMIT_SCALE: convert to jiffies * 128. */ in ebt_limit_mt_check() 83 info->credit = user2credits(info->avg * info->burst); in ebt_limit_mt_check() 84 info->credit_cap = user2credits(info->avg * info->burst); in ebt_limit_mt_check() 85 info->cost = user2credits(info->avg); in ebt_limit_mt_check() 93 * only avg/burst have meaningful values in userspace. 96 compat_uint_t avg, burst; member
|
/kernel/linux/linux-5.10/drivers/mfd/ |
H A D | pcf50633-adc.c | 27 int avg; member 54 static void adc_setup(struct pcf50633 *pcf, int channel, int avg) in adc_setup() argument 63 pcf50633_reg_write(pcf, PCF50633_REG_ADCC1, channel | avg | in adc_setup() 77 adc_setup(pcf, adc->queue[head]->mux, adc->queue[head]->avg); in trigger_next_adc_job_if_any() 116 int pcf50633_adc_sync_read(struct pcf50633 *pcf, int mux, int avg) in pcf50633_adc_sync_read() argument 123 ret = pcf50633_adc_async_read(pcf, mux, avg, in pcf50633_adc_sync_read() 134 int pcf50633_adc_async_read(struct pcf50633 *pcf, int mux, int avg, in pcf50633_adc_async_read() argument 147 req->avg = avg; in pcf50633_adc_async_read()
|
/kernel/linux/linux-6.6/drivers/mfd/ |
H A D | pcf50633-adc.c | 27 int avg; member 54 static void adc_setup(struct pcf50633 *pcf, int channel, int avg) in adc_setup() argument 63 pcf50633_reg_write(pcf, PCF50633_REG_ADCC1, channel | avg | in adc_setup() 77 adc_setup(pcf, adc->queue[head]->mux, adc->queue[head]->avg); in trigger_next_adc_job_if_any() 116 int pcf50633_adc_sync_read(struct pcf50633 *pcf, int mux, int avg) in pcf50633_adc_sync_read() argument 123 ret = pcf50633_adc_async_read(pcf, mux, avg, in pcf50633_adc_sync_read() 134 int pcf50633_adc_async_read(struct pcf50633 *pcf, int mux, int avg, in pcf50633_adc_async_read() argument 147 req->avg = avg; in pcf50633_adc_async_read()
|
/kernel/linux/linux-5.10/tools/testing/selftests/kvm/ |
H A D | dirty_log_perf_test.c | 43 struct timespec avg; in vcpu_worker() local 83 avg = timespec_div(total, vcpu_last_completed_iteration[vcpu_id]); in vcpu_worker() 86 total.tv_sec, total.tv_nsec, avg.tv_sec, avg.tv_nsec); in vcpu_worker() 108 struct timespec avg; in run_test() local 218 avg = timespec_div(get_dirty_log_total, iterations); in run_test() 221 get_dirty_log_total.tv_nsec, avg.tv_sec, avg.tv_nsec); in run_test() 224 avg = timespec_div(clear_dirty_log_total, iterations); in run_test() 227 clear_dirty_log_total.tv_nsec, avg in run_test() [all...] |
/kernel/linux/linux-6.6/tools/testing/selftests/kvm/ |
H A D | dirty_log_perf_test.c | 79 struct timespec avg; in vcpu_worker() local 122 avg = timespec_div(total, vcpu_last_completed_iteration[vcpu_idx]); in vcpu_worker() 125 total.tv_sec, total.tv_nsec, avg.tv_sec, avg.tv_nsec); in vcpu_worker() 151 struct timespec avg; in run_test() local 280 avg = timespec_div(get_dirty_log_total, p->iterations); in run_test() 283 get_dirty_log_total.tv_nsec, avg.tv_sec, avg.tv_nsec); in run_test() 286 avg = timespec_div(clear_dirty_log_total, p->iterations); in run_test() 289 clear_dirty_log_total.tv_nsec, avg in run_test() [all...] |
/kernel/linux/linux-5.10/tools/perf/scripts/python/Perf-Trace-Util/lib/Perf/Trace/ |
H A D | Util.py | 20 def avg(total, n): function 40 min, max, avg, count = dict[key] 45 avg = (avg + value) / 2 46 dict[key] = (min, max, avg, count + 1)
|
/kernel/linux/linux-6.6/tools/perf/scripts/python/Perf-Trace-Util/lib/Perf/Trace/ |
H A D | Util.py | 20 def avg(total, n): function 40 min, max, avg, count = dict[key] 45 avg = (avg + value) / 2 46 dict[key] = (min, max, avg, count + 1)
|
/kernel/linux/linux-5.10/kernel/sched/ |
H A D | pelt.h | 42 static inline u32 get_pelt_divider(struct sched_avg *avg) in get_pelt_divider() argument 44 return PELT_MIN_DIVIDER + avg->period_contrib; in get_pelt_divider() 47 static inline void cfs_se_util_change(struct sched_avg *avg) in cfs_se_util_change() argument 55 enqueued = avg->util_est.enqueued; in cfs_se_util_change() 61 WRITE_ONCE(avg->util_est.enqueued, enqueued); in cfs_se_util_change() 110 * (LOAD_AVG_MAX - 1024 + rq->cfs.avg.period_contrib) << SCHED_CAPACITY_SHIFT; 118 u32 util_sum = rq->cfs.avg.util_sum; in update_idle_rq_clock_pelt()
|
H A D | pelt.c | 292 * runnable_sum = \Sum se->avg.runnable_sum 293 * runnable_avg = \Sum se->avg.runnable_avg 295 * load_sum = \Sum se_weight(se) * se->avg.load_sum 296 * load_avg = \Sum se->avg.load_avg 301 if (___update_load_sum(now, &se->avg, 0, 0, 0)) { in __update_load_avg_blocked_se() 302 ___update_load_avg(&se->avg, se_weight(se)); in __update_load_avg_blocked_se() 312 if (___update_load_sum(now, &se->avg, !!se->on_rq, se_runnable(se), in __update_load_avg_se() 315 ___update_load_avg(&se->avg, se_weight(se)); in __update_load_avg_se() 316 cfs_se_util_change(&se->avg); in __update_load_avg_se() 326 if (___update_load_sum(now, &cfs_rq->avg, in __update_load_avg_cfs_rq() [all...] |
/kernel/linux/linux-5.10/drivers/iio/humidity/ |
H A D | hts221_core.c | 173 const struct hts221_avg *avg = &hts221_avg_list[type]; in hts221_update_avg() local 177 if (avg->avg_avl[i] == val) in hts221_update_avg() 183 data = ((i << __ffs(avg->mask)) & avg->mask); in hts221_update_avg() 184 err = regmap_update_bits(hw->regmap, avg->addr, in hts221_update_avg() 185 avg->mask, data); in hts221_update_avg() 214 const struct hts221_avg *avg = &hts221_avg_list[HTS221_SENSOR_H]; in hts221_sysfs_rh_oversampling_avail() local 218 for (i = 0; i < ARRAY_SIZE(avg->avg_avl); i++) in hts221_sysfs_rh_oversampling_avail() 220 avg->avg_avl[i]); in hts221_sysfs_rh_oversampling_avail() 231 const struct hts221_avg *avg in hts221_sysfs_temp_oversampling_avail() local 447 const struct hts221_avg *avg; hts221_read_raw() local [all...] |
/kernel/linux/linux-6.6/drivers/iio/humidity/ |
H A D | hts221_core.c | 174 const struct hts221_avg *avg = &hts221_avg_list[type]; in hts221_update_avg() local 178 if (avg->avg_avl[i] == val) in hts221_update_avg() 184 data = ((i << __ffs(avg->mask)) & avg->mask); in hts221_update_avg() 185 err = regmap_update_bits(hw->regmap, avg->addr, in hts221_update_avg() 186 avg->mask, data); in hts221_update_avg() 215 const struct hts221_avg *avg = &hts221_avg_list[HTS221_SENSOR_H]; in hts221_sysfs_rh_oversampling_avail() local 219 for (i = 0; i < ARRAY_SIZE(avg->avg_avl); i++) in hts221_sysfs_rh_oversampling_avail() 221 avg->avg_avl[i]); in hts221_sysfs_rh_oversampling_avail() 232 const struct hts221_avg *avg in hts221_sysfs_temp_oversampling_avail() local 448 const struct hts221_avg *avg; hts221_read_raw() local [all...] |
/kernel/linux/linux-6.6/kernel/sched/ |
H A D | pelt.h | 42 static inline u32 get_pelt_divider(struct sched_avg *avg) in get_pelt_divider() argument 44 return PELT_MIN_DIVIDER + avg->period_contrib; in get_pelt_divider() 47 static inline void cfs_se_util_change(struct sched_avg *avg) in cfs_se_util_change() argument 55 enqueued = avg->util_est.enqueued; in cfs_se_util_change() 61 WRITE_ONCE(avg->util_est.enqueued, enqueued); in cfs_se_util_change() 128 * (LOAD_AVG_MAX - 1024 + rq->cfs.avg.period_contrib) << SCHED_CAPACITY_SHIFT; 136 u32 util_sum = rq->cfs.avg.util_sum; in update_idle_rq_clock_pelt()
|
H A D | pelt.c | 288 * runnable_sum = \Sum se->avg.runnable_sum 289 * runnable_avg = \Sum se->avg.runnable_avg 291 * load_sum = \Sum se_weight(se) * se->avg.load_sum 292 * load_avg = \Sum se->avg.load_avg 297 if (___update_load_sum(now, &se->avg, 0, 0, 0)) { in __update_load_avg_blocked_se() 298 ___update_load_avg(&se->avg, se_weight(se)); in __update_load_avg_blocked_se() 308 if (___update_load_sum(now, &se->avg, !!se->on_rq, se_runnable(se), in __update_load_avg_se() 311 ___update_load_avg(&se->avg, se_weight(se)); in __update_load_avg_se() 312 cfs_se_util_change(&se->avg); in __update_load_avg_se() 322 if (___update_load_sum(now, &cfs_rq->avg, in __update_load_avg_cfs_rq() [all...] |
H A D | fair.c | 741 s64 avg = cfs_rq->avg_vruntime; in avg_vruntime() local 747 avg += entity_key(cfs_rq, curr) * weight; in avg_vruntime() 753 if (avg < 0) in avg_vruntime() 754 avg -= (load - 1); in avg_vruntime() 755 avg = div_s64(avg, load); in avg_vruntime() 758 return cfs_rq->min_vruntime + avg; in avg_vruntime() 808 s64 avg = cfs_rq->avg_vruntime; in entity_eligible() local 814 avg += entity_key(cfs_rq, curr) * weight; in entity_eligible() 818 return avg > in entity_eligible() [all...] |
/kernel/linux/linux-5.10/kernel/trace/ |
H A D | trace_benchmark.c | 45 unsigned int avg; in trace_do_benchmark() local 83 * We should have enough data for the avg and stddev anyway. in trace_do_benchmark() 87 "last=%llu first=%llu max=%llu min=%llu ** avg=%u std=%d std^2=%lld", in trace_do_benchmark() 109 avg = delta; in trace_do_benchmark() 124 seed = avg; in trace_do_benchmark() 139 "last=%llu first=%llu max=%llu min=%llu avg=%u std=%d std^2=%lld", in trace_do_benchmark() 140 bm_last, bm_first, bm_max, bm_min, avg, std, stddev); in trace_do_benchmark() 143 bm_avg = avg; in trace_do_benchmark()
|
/kernel/linux/linux-6.6/kernel/trace/ |
H A D | trace_benchmark.c | 45 unsigned int avg; in trace_do_benchmark() local 83 * We should have enough data for the avg and stddev anyway. in trace_do_benchmark() 87 "last=%llu first=%llu max=%llu min=%llu ** avg=%u std=%d std^2=%lld", in trace_do_benchmark() 109 avg = delta; in trace_do_benchmark() 124 seed = avg; in trace_do_benchmark() 139 "last=%llu first=%llu max=%llu min=%llu avg=%u std=%d std^2=%lld", in trace_do_benchmark() 140 bm_last, bm_first, bm_max, bm_min, avg, std, stddev); in trace_do_benchmark() 143 bm_avg = avg; in trace_do_benchmark()
|
/kernel/linux/linux-5.10/include/uapi/linux/netfilter/ |
H A D | xt_hashlimit.h | 34 __u32 avg; /* Average secs between packets * scale */ member 58 __u32 avg; /* Average secs between packets * scale */ member 71 __u64 avg; /* Average secs between packets * scale */ member 85 __u64 avg; /* Average secs between packets * scale */ member
|
/kernel/linux/linux-6.6/include/uapi/linux/netfilter/ |
H A D | xt_hashlimit.h | 34 __u32 avg; /* Average secs between packets * scale */ member 58 __u32 avg; /* Average secs between packets * scale */ member 71 __u64 avg; /* Average secs between packets * scale */ member 85 __u64 avg; /* Average secs between packets * scale */ member
|
/kernel/linux/patches/linux-5.10/prebuilts/usr/include/linux/netfilter/ |
H A D | xt_hashlimit.h | 39 __u32 avg; member 57 __u32 avg; member 66 __u64 avg; member 76 __u64 avg; member
|
/kernel/linux/patches/linux-6.6/prebuilts/usr/include/linux/netfilter/ |
H A D | xt_hashlimit.h | 39 __u32 avg; member 57 __u32 avg; member 66 __u64 avg; member 76 __u64 avg; member
|
/kernel/linux/patches/linux-4.19/prebuilts/usr/include/linux/netfilter/ |
H A D | xt_hashlimit.h | 26 __u32 avg; member 46 __u32 avg; member 56 __u64 avg; member 67 __u64 avg; member
|