tail_cycle 1194 fs/xfs/xfs_log.c int tail_cycle; tail_cycle 1199 fs/xfs/xfs_log.c xlog_crack_atomic_lsn(&log->l_tail_lsn, &tail_cycle, &tail_bytes); tail_cycle 1201 fs/xfs/xfs_log.c if (tail_cycle == head_cycle && head_bytes >= tail_bytes) tail_cycle 1203 fs/xfs/xfs_log.c else if (tail_cycle + 1 < head_cycle) tail_cycle 1205 fs/xfs/xfs_log.c else if (tail_cycle < head_cycle) { tail_cycle 1206 fs/xfs/xfs_log.c ASSERT(tail_cycle == (head_cycle - 1)); tail_cycle 1217 fs/xfs/xfs_log.c tail_cycle, tail_bytes); tail_cycle 3738 fs/xfs/xfs_log.c int tail_cycle, tail_blocks; tail_cycle 3742 fs/xfs/xfs_log.c xlog_crack_atomic_lsn(&log->l_tail_lsn, &tail_cycle, &tail_blocks); tail_cycle 3743 fs/xfs/xfs_log.c if (tail_cycle != cycle) { tail_cycle 3744 fs/xfs/xfs_log.c if (cycle - 1 != tail_cycle && tail_cycle 1557 fs/xfs/xfs_log_recover.c int tail_cycle, tail_cycle 1568 fs/xfs/xfs_log_recover.c recp->h_tail_lsn = cpu_to_be64(xlog_assign_lsn(tail_cycle, tail_block)); tail_cycle 1579 fs/xfs/xfs_log_recover.c int tail_cycle, tail_cycle 1641 fs/xfs/xfs_log_recover.c tail_cycle, tail_block); tail_cycle 1677 fs/xfs/xfs_log_recover.c int tail_cycle, head_cycle; tail_cycle 1683 fs/xfs/xfs_log_recover.c tail_cycle = CYCLE_LSN(tail_lsn); tail_cycle 1694 fs/xfs/xfs_log_recover.c if (head_cycle == tail_cycle) { tail_cycle 1714 fs/xfs/xfs_log_recover.c if (unlikely(head_block >= tail_block || head_cycle != (tail_cycle + 1))){ tail_cycle 1750 fs/xfs/xfs_log_recover.c head_block, max_distance, tail_cycle, tail_cycle 1764 fs/xfs/xfs_log_recover.c head_block, distance, tail_cycle, tail_cycle 1780 fs/xfs/xfs_log_recover.c tail_cycle, tail_block);