Lines Matching refs:cycle
168 * - ic_size is the full size of the log buffer, minus the cycle headers.
394 * The reservation head lsn is not made up of a cycle number and block number.
395 * Instead, it uses a cycle number and byte number. Logs don't expect to
534 xlog_crack_atomic_lsn(atomic64_t *lsn, uint *cycle, uint *block)
538 *cycle = CYCLE_LSN(val);
546 xlog_assign_atomic_lsn(atomic64_t *lsn, uint cycle, uint block)
548 atomic64_set(lsn, xlog_assign_lsn(cycle, block));
557 xlog_crack_grant_head_val(int64_t val, int *cycle, int *space)
559 *cycle = val >> 32;
564 xlog_crack_grant_head(atomic64_t *head, int *cycle, int *space)
566 xlog_crack_grant_head_val(atomic64_read(head), cycle, space);
570 xlog_assign_grant_head_val(int cycle, int space)
572 return ((int64_t)cycle << 32) | space;
576 xlog_assign_grant_head(atomic64_t *head, int cycle, int space)
578 atomic64_set(head, xlog_assign_grant_head_val(cycle, space));
645 * contention from metadata I/O. The current cycle and block are updated
650 * The current block is always rewound before the cycle is bumped in
653 * transiently behind state if we happen to race with a cycle wrap.
663 * above raced with a wrap to the next log cycle. Grab the lock