tail_cycle       1194 fs/xfs/xfs_log.c 	int		tail_cycle;
tail_cycle       1199 fs/xfs/xfs_log.c 	xlog_crack_atomic_lsn(&log->l_tail_lsn, &tail_cycle, &tail_bytes);
tail_cycle       1201 fs/xfs/xfs_log.c 	if (tail_cycle == head_cycle && head_bytes >= tail_bytes)
tail_cycle       1203 fs/xfs/xfs_log.c 	else if (tail_cycle + 1 < head_cycle)
tail_cycle       1205 fs/xfs/xfs_log.c 	else if (tail_cycle < head_cycle) {
tail_cycle       1206 fs/xfs/xfs_log.c 		ASSERT(tail_cycle == (head_cycle - 1));
tail_cycle       1217 fs/xfs/xfs_log.c 			  tail_cycle, tail_bytes);
tail_cycle       3738 fs/xfs/xfs_log.c 	int		tail_cycle, tail_blocks;
tail_cycle       3742 fs/xfs/xfs_log.c 	xlog_crack_atomic_lsn(&log->l_tail_lsn, &tail_cycle, &tail_blocks);
tail_cycle       3743 fs/xfs/xfs_log.c 	if (tail_cycle != cycle) {
tail_cycle       3744 fs/xfs/xfs_log.c 		if (cycle - 1 != tail_cycle &&
tail_cycle       1557 fs/xfs/xfs_log_recover.c 	int			tail_cycle,
tail_cycle       1568 fs/xfs/xfs_log_recover.c 	recp->h_tail_lsn = cpu_to_be64(xlog_assign_lsn(tail_cycle, tail_block));
tail_cycle       1579 fs/xfs/xfs_log_recover.c 	int		tail_cycle,
tail_cycle       1641 fs/xfs/xfs_log_recover.c 					tail_cycle, tail_block);
tail_cycle       1677 fs/xfs/xfs_log_recover.c 	int		tail_cycle, head_cycle;
tail_cycle       1683 fs/xfs/xfs_log_recover.c 	tail_cycle = CYCLE_LSN(tail_lsn);
tail_cycle       1694 fs/xfs/xfs_log_recover.c 	if (head_cycle == tail_cycle) {
tail_cycle       1714 fs/xfs/xfs_log_recover.c 		if (unlikely(head_block >= tail_block || head_cycle != (tail_cycle + 1))){
tail_cycle       1750 fs/xfs/xfs_log_recover.c 				head_block, max_distance, tail_cycle,
tail_cycle       1764 fs/xfs/xfs_log_recover.c 				head_block, distance, tail_cycle,
tail_cycle       1780 fs/xfs/xfs_log_recover.c 				tail_cycle, tail_block);