Lines Matching refs:cycle
177 * - ic_size is the full size of the log buffer, minus the cycle headers.
379 * The reservation head lsn is not made up of a cycle number and block number.
380 * Instead, it uses a cycle number and byte number. Logs don't expect to
496 xlog_crack_atomic_lsn(atomic64_t *lsn, uint *cycle, uint *block)
500 *cycle = CYCLE_LSN(val);
508 xlog_assign_atomic_lsn(atomic64_t *lsn, uint cycle, uint block)
510 atomic64_set(lsn, xlog_assign_lsn(cycle, block));
519 xlog_crack_grant_head_val(int64_t val, int *cycle, int *space)
521 *cycle = val >> 32;
526 xlog_crack_grant_head(atomic64_t *head, int *cycle, int *space)
528 xlog_crack_grant_head_val(atomic64_read(head), cycle, space);
532 xlog_assign_grant_head_val(int cycle, int space)
534 return ((int64_t)cycle << 32) | space;
538 xlog_assign_grant_head(atomic64_t *head, int cycle, int space)
540 atomic64_set(head, xlog_assign_grant_head_val(cycle, space));
601 * contention from metadata I/O. The current cycle and block are updated
606 * The current block is always rewound before the cycle is bumped in
609 * transiently behind state if we happen to race with a cycle wrap.
619 * above raced with a wrap to the next log cycle. Grab the lock