Lines Matching refs:cycle
168 * - ic_size is the full size of the log buffer, minus the cycle headers.
394 * The reservation head lsn is not made up of a cycle number and block number.
395 * Instead, it uses a cycle number and byte number. Logs don't expect to
533 xlog_crack_atomic_lsn(atomic64_t *lsn, uint *cycle, uint *block)
537 *cycle = CYCLE_LSN(val);
545 xlog_assign_atomic_lsn(atomic64_t *lsn, uint cycle, uint block)
547 atomic64_set(lsn, xlog_assign_lsn(cycle, block));
556 xlog_crack_grant_head_val(int64_t val, int *cycle, int *space)
558 *cycle = val >> 32;
563 xlog_crack_grant_head(atomic64_t *head, int *cycle, int *space)
565 xlog_crack_grant_head_val(atomic64_read(head), cycle, space);
569 xlog_assign_grant_head_val(int cycle, int space)
571 return ((int64_t)cycle << 32) | space;
575 xlog_assign_grant_head(atomic64_t *head, int cycle, int space)
577 atomic64_set(head, xlog_assign_grant_head_val(cycle, space));
644 * contention from metadata I/O. The current cycle and block are updated
649 * The current block is always rewound before the cycle is bumped in
652 * transiently behind state if we happen to race with a cycle wrap.
662 * above raced with a wrap to the next log cycle. Grab the lock