Lines Matching defs:mult

93  * returns nanoseconds already so no conversion is required, hence mult=1
101 .mult = 1, \
272 /* Cap delta value to the max_cycles values to avoid mult overflows */
325 /* Do the ns -> cycle conversion first, using original mult */
329 tmp += clock->mult/2;
330 do_div(tmp, clock->mult);
338 tk->xtime_interval = interval * clock->mult;
340 tk->raw_interval = interval * clock->mult;
362 * The timekeeper keeps its own mult values for the currently
366 tk->tkr_mono.mult = clock->mult;
367 tk->tkr_raw.mult = clock->mult;
378 nsec = delta * tkr->mult + tkr->xtime_nsec;
803 tk->tkr_mono.xtime_nsec += delta * tk->tkr_mono.mult;
804 tk->tkr_raw.xtime_nsec += delta * tk->tkr_raw.mult;
866 nsecs = tk->tkr_mono.mult >> tk->tkr_mono.shift;
1086 /* Scale base by mult/div checking for overflow */
1087 static int scale64_check_overflow(u64 mult, u64 div, u64 *base)
1093 if (((int)sizeof(u64)*8 - fls64(mult) < fls64(tmp)) ||
1094 ((int)sizeof(u64)*8 - fls64(mult) < fls64(rem)))
1096 tmp *= mult;
1098 rem = div64_u64(rem * mult, div);
1154 * mult(real)/mult(raw) yielding the realtime correction
1160 (corr_raw, tk->tkr_mono.mult, tk->tkr_raw.mult);
1951 * xtime_interval = cycle_interval * mult
1952 * So if mult is being incremented by one:
1953 * xtime_interval = cycle_interval * (mult + 1)
1955 * xtime_interval = (cycle_interval * mult) + cycle_interval
1987 if ((mult_adj > 0) && (tk->tkr_mono.mult + mult_adj < mult_adj)) {
1988 /* NTP adjustment caused clocksource mult overflow */
1993 tk->tkr_mono.mult += mult_adj;
2004 u32 mult;
2011 mult = tk->tkr_mono.mult - tk->ntp_err_mult;
2014 mult = div64_u64((tk->ntp_tick >> tk->ntp_error_shift) -
2025 mult += tk->ntp_err_mult;
2027 timekeeping_apply_adjustment(tk, offset, mult - tk->tkr_mono.mult);
2030 (abs(tk->tkr_mono.mult - tk->tkr_mono.clock->mult)
2034 tk->tkr_mono.clock->name, (long)tk->tkr_mono.mult,
2035 (long)tk->tkr_mono.clock->mult + tk->tkr_mono.clock->maxadj);