Lines Matching refs:iclog
45 struct xlog_in_core **iclog,
61 struct xlog_in_core *iclog);
67 STATIC void xlog_state_done_syncing(xlog_in_core_t *iclog, int);
72 struct xlog_in_core *iclog);
77 struct xlog_in_core **iclog,
84 struct xlog_in_core *iclog);
88 struct xlog_in_core *iclog,
93 struct xlog_in_core *iclog);
119 struct xlog_in_core *iclog,
125 struct xlog_in_core *iclog,
515 struct xlog_in_core **iclog, in xfs_log_done() argument
527 (xlog_commit_record(log, ticket, iclog, &lsn)))) { in xfs_log_done()
565 struct xlog_in_core *iclog, in xfs_log_notify() argument
570 spin_lock(&iclog->ic_callback_lock); in xfs_log_notify()
571 abortflg = (iclog->ic_state & XLOG_STATE_IOERROR); in xfs_log_notify()
573 ASSERT_ALWAYS((iclog->ic_state == XLOG_STATE_ACTIVE) || in xfs_log_notify()
574 (iclog->ic_state == XLOG_STATE_WANT_SYNC)); in xfs_log_notify()
576 *(iclog->ic_callback_tail) = cb; in xfs_log_notify()
577 iclog->ic_callback_tail = &(cb->cb_next); in xfs_log_notify()
579 spin_unlock(&iclog->ic_callback_lock); in xfs_log_notify()
586 struct xlog_in_core *iclog) in xfs_log_release_iclog() argument
588 if (xlog_state_release_iclog(mp->m_log, iclog)) { in xfs_log_release_iclog()
797 xlog_in_core_t *iclog; in xfs_log_unmount_write() local
816 first_iclog = iclog = log->l_iclog; in xfs_log_unmount_write()
818 if (!(iclog->ic_state & XLOG_STATE_IOERROR)) { in xfs_log_unmount_write()
819 ASSERT(iclog->ic_state & XLOG_STATE_ACTIVE); in xfs_log_unmount_write()
820 ASSERT(iclog->ic_offset == 0); in xfs_log_unmount_write()
822 iclog = iclog->ic_next; in xfs_log_unmount_write()
823 } while (iclog != first_iclog); in xfs_log_unmount_write()
864 iclog = log->l_iclog; in xfs_log_unmount_write()
865 atomic_inc(&iclog->ic_refcnt); in xfs_log_unmount_write()
866 xlog_state_want_sync(log, iclog); in xfs_log_unmount_write()
868 error = xlog_state_release_iclog(log, iclog); in xfs_log_unmount_write()
871 if (!(iclog->ic_state == XLOG_STATE_ACTIVE || in xfs_log_unmount_write()
872 iclog->ic_state == XLOG_STATE_DIRTY)) { in xfs_log_unmount_write()
874 xlog_wait(&iclog->ic_force_wait, in xfs_log_unmount_write()
902 iclog = log->l_iclog; in xfs_log_unmount_write()
903 atomic_inc(&iclog->ic_refcnt); in xfs_log_unmount_write()
905 xlog_state_want_sync(log, iclog); in xfs_log_unmount_write()
907 error = xlog_state_release_iclog(log, iclog); in xfs_log_unmount_write()
911 if ( ! ( iclog->ic_state == XLOG_STATE_ACTIVE in xfs_log_unmount_write()
912 || iclog->ic_state == XLOG_STATE_DIRTY in xfs_log_unmount_write()
913 || iclog->ic_state == XLOG_STATE_IOERROR) ) { in xfs_log_unmount_write()
915 xlog_wait(&iclog->ic_force_wait, in xfs_log_unmount_write()
1186 struct xlog_in_core *iclog = bp->b_fspriv; in xlog_iodone() local
1187 struct xlog *l = iclog->ic_log; in xlog_iodone()
1204 } else if (iclog->ic_state & XLOG_STATE_IOERROR) { in xlog_iodone()
1210 xlog_state_done_syncing(iclog, aborted); in xlog_iodone()
1349 xlog_in_core_t *iclog, *prev_iclog=NULL; in xlog_alloc_log() local
1449 iclog = *iclogp; in xlog_alloc_log()
1450 iclog->ic_prev = prev_iclog; in xlog_alloc_log()
1451 prev_iclog = iclog; in xlog_alloc_log()
1464 iclog->ic_bp = bp; in xlog_alloc_log()
1465 iclog->ic_data = bp->b_addr; in xlog_alloc_log()
1467 log->l_iclog_bak[i] = &iclog->ic_header; in xlog_alloc_log()
1469 head = &iclog->ic_header; in xlog_alloc_log()
1479 iclog->ic_size = BBTOB(bp->b_length) - log->l_iclog_hsize; in xlog_alloc_log()
1480 iclog->ic_state = XLOG_STATE_ACTIVE; in xlog_alloc_log()
1481 iclog->ic_log = log; in xlog_alloc_log()
1482 atomic_set(&iclog->ic_refcnt, 0); in xlog_alloc_log()
1483 spin_lock_init(&iclog->ic_callback_lock); in xlog_alloc_log()
1484 iclog->ic_callback_tail = &(iclog->ic_callback); in xlog_alloc_log()
1485 iclog->ic_datap = (char *)iclog->ic_data + log->l_iclog_hsize; in xlog_alloc_log()
1487 init_waitqueue_head(&iclog->ic_force_wait); in xlog_alloc_log()
1488 init_waitqueue_head(&iclog->ic_write_wait); in xlog_alloc_log()
1490 iclogp = &iclog->ic_next; in xlog_alloc_log()
1501 for (iclog = log->l_iclog; iclog; iclog = prev_iclog) { in xlog_alloc_log()
1502 prev_iclog = iclog->ic_next; in xlog_alloc_log()
1503 if (iclog->ic_bp) in xlog_alloc_log()
1504 xfs_buf_free(iclog->ic_bp); in xlog_alloc_log()
1505 kmem_free(iclog); in xlog_alloc_log()
1524 struct xlog_in_core **iclog, in xlog_commit_record() argument
1539 ASSERT_ALWAYS(iclog); in xlog_commit_record()
1540 error = xlog_write(log, &vec, ticket, commitlsnp, iclog, in xlog_commit_record()
1616 struct xlog_in_core *iclog, in xlog_pack_data() argument
1620 int size = iclog->ic_offset + roundoff; in xlog_pack_data()
1624 cycle_lsn = CYCLE_LSN_DISK(iclog->ic_header.h_lsn); in xlog_pack_data()
1626 dp = iclog->ic_datap; in xlog_pack_data()
1630 iclog->ic_header.h_cycle_data[i] = *(__be32 *)dp; in xlog_pack_data()
1636 xlog_in_core_2_t *xhdr = iclog->ic_data; in xlog_pack_data()
1711 struct xlog_in_core *iclog = bp->b_fspriv; in xlog_bdstrat() local
1714 if (iclog->ic_state & XLOG_STATE_IOERROR) { in xlog_bdstrat()
1759 struct xlog_in_core *iclog) in xlog_sync() argument
1772 ASSERT(atomic_read(&iclog->ic_refcnt) == 0); in xlog_sync()
1775 count_init = log->l_iclog_hsize + iclog->ic_offset; in xlog_sync()
1797 xlog_pack_data(log, iclog, roundoff); in xlog_sync()
1800 size = iclog->ic_offset; in xlog_sync()
1803 iclog->ic_header.h_len = cpu_to_be32(size); in xlog_sync()
1805 bp = iclog->ic_bp; in xlog_sync()
1806 XFS_BUF_SET_ADDR(bp, BLOCK_LSN(be64_to_cpu(iclog->ic_header.h_lsn))); in xlog_sync()
1816 iclog->ic_bwritecnt = 2; in xlog_sync()
1825 dptr = (char *)&iclog->ic_header + count; in xlog_sync()
1835 iclog->ic_bwritecnt = 1; in xlog_sync()
1839 iclog->ic_header.h_crc = xlog_cksum(log, &iclog->ic_header, in xlog_sync()
1840 iclog->ic_datap, size); in xlog_sync()
1843 bp->b_fspriv = iclog; in xlog_sync()
1869 xlog_verify_iclog(log, iclog, count, true); in xlog_sync()
1885 bp = iclog->ic_log->l_xbuf; in xlog_sync()
1888 (char *)&iclog->ic_header + count, split); in xlog_sync()
1889 bp->b_fspriv = iclog; in xlog_sync()
1918 xlog_in_core_t *iclog, *next_iclog; in xlog_dealloc_log() local
1927 iclog = log->l_iclog; in xlog_dealloc_log()
1929 xfs_buf_lock(iclog->ic_bp); in xlog_dealloc_log()
1930 xfs_buf_unlock(iclog->ic_bp); in xlog_dealloc_log()
1931 iclog = iclog->ic_next; in xlog_dealloc_log()
1944 iclog = log->l_iclog; in xlog_dealloc_log()
1946 xfs_buf_free(iclog->ic_bp); in xlog_dealloc_log()
1947 next_iclog = iclog->ic_next; in xlog_dealloc_log()
1948 kmem_free(iclog); in xlog_dealloc_log()
1949 iclog = next_iclog; in xlog_dealloc_log()
1964 struct xlog_in_core *iclog, in xlog_state_finish_copy() argument
1970 be32_add_cpu(&iclog->ic_header.h_num_logops, record_cnt); in xlog_state_finish_copy()
1971 iclog->ic_offset += copy_bytes; in xlog_state_finish_copy()
2236 struct xlog_in_core *iclog, in xlog_write_copy_finish() argument
2250 xlog_state_finish_copy(log, iclog, *record_cnt, *data_cnt); in xlog_write_copy_finish()
2253 return xlog_state_release_iclog(log, iclog); in xlog_write_copy_finish()
2259 if (iclog->ic_size - log_offset <= sizeof(xlog_op_header_t)) { in xlog_write_copy_finish()
2261 xlog_state_finish_copy(log, iclog, *record_cnt, *data_cnt); in xlog_write_copy_finish()
2266 xlog_state_want_sync(log, iclog); in xlog_write_copy_finish()
2270 return xlog_state_release_iclog(log, iclog); in xlog_write_copy_finish()
2272 *commit_iclog = iclog; in xlog_write_copy_finish()
2327 struct xlog_in_core *iclog = NULL; in xlog_write() local
2368 error = xlog_state_get_iclog_space(log, len, &iclog, ticket, in xlog_write()
2373 ASSERT(log_offset <= iclog->ic_size - 1); in xlog_write()
2374 ptr = iclog->ic_datap + log_offset; in xlog_write()
2378 *start_lsn = be64_to_cpu(iclog->ic_header.h_lsn); in xlog_write()
2418 iclog->ic_size-log_offset, in xlog_write()
2443 error = xlog_write_copy_finish(log, iclog, flags, in xlog_write()
2484 xlog_state_finish_copy(log, iclog, record_cnt, data_cnt); in xlog_write()
2486 return xlog_state_release_iclog(log, iclog); in xlog_write()
2489 *commit_iclog = iclog; in xlog_write()
2513 xlog_in_core_t *iclog; in xlog_state_clean_log() local
2516 iclog = log->l_iclog; in xlog_state_clean_log()
2518 if (iclog->ic_state == XLOG_STATE_DIRTY) { in xlog_state_clean_log()
2519 iclog->ic_state = XLOG_STATE_ACTIVE; in xlog_state_clean_log()
2520 iclog->ic_offset = 0; in xlog_state_clean_log()
2521 ASSERT(iclog->ic_callback == NULL); in xlog_state_clean_log()
2531 (be32_to_cpu(iclog->ic_header.h_num_logops) == in xlog_state_clean_log()
2542 iclog->ic_header.h_num_logops = 0; in xlog_state_clean_log()
2543 memset(iclog->ic_header.h_cycle_data, 0, in xlog_state_clean_log()
2544 sizeof(iclog->ic_header.h_cycle_data)); in xlog_state_clean_log()
2545 iclog->ic_header.h_lsn = 0; in xlog_state_clean_log()
2546 } else if (iclog->ic_state == XLOG_STATE_ACTIVE) in xlog_state_clean_log()
2550 iclog = iclog->ic_next; in xlog_state_clean_log()
2551 } while (iclog != log->l_iclog); in xlog_state_clean_log()
2618 xlog_in_core_t *iclog; in xlog_state_do_callback() local
2632 first_iclog = iclog = log->l_iclog; in xlog_state_do_callback()
2647 iclog = log->l_iclog; in xlog_state_do_callback()
2654 if (iclog->ic_state & in xlog_state_do_callback()
2656 iclog = iclog->ic_next; in xlog_state_do_callback()
2667 if (!(iclog->ic_state & XLOG_STATE_IOERROR)) { in xlog_state_do_callback()
2678 if (!(iclog->ic_state & in xlog_state_do_callback()
2704 be64_to_cpu(iclog->ic_header.h_lsn)) < 0) { in xlog_state_do_callback()
2705 iclog = iclog->ic_next; in xlog_state_do_callback()
2710 iclog->ic_state = XLOG_STATE_CALLBACK; in xlog_state_do_callback()
2731 be64_to_cpu(iclog->ic_header.h_lsn)) <= 0); in xlog_state_do_callback()
2732 if (iclog->ic_callback) in xlog_state_do_callback()
2734 be64_to_cpu(iclog->ic_header.h_lsn)); in xlog_state_do_callback()
2748 spin_lock(&iclog->ic_callback_lock); in xlog_state_do_callback()
2749 cb = iclog->ic_callback; in xlog_state_do_callback()
2751 iclog->ic_callback_tail = &(iclog->ic_callback); in xlog_state_do_callback()
2752 iclog->ic_callback = NULL; in xlog_state_do_callback()
2753 spin_unlock(&iclog->ic_callback_lock); in xlog_state_do_callback()
2760 spin_lock(&iclog->ic_callback_lock); in xlog_state_do_callback()
2761 cb = iclog->ic_callback; in xlog_state_do_callback()
2768 ASSERT(iclog->ic_callback == NULL); in xlog_state_do_callback()
2769 spin_unlock(&iclog->ic_callback_lock); in xlog_state_do_callback()
2770 if (!(iclog->ic_state & XLOG_STATE_IOERROR)) in xlog_state_do_callback()
2771 iclog->ic_state = XLOG_STATE_DIRTY; in xlog_state_do_callback()
2780 wake_up_all(&iclog->ic_force_wait); in xlog_state_do_callback()
2782 iclog = iclog->ic_next; in xlog_state_do_callback()
2783 } while (first_iclog != iclog); in xlog_state_do_callback()
2800 first_iclog = iclog = log->l_iclog; in xlog_state_do_callback()
2802 ASSERT(iclog->ic_state != XLOG_STATE_DO_CALLBACK); in xlog_state_do_callback()
2812 if (iclog->ic_state == XLOG_STATE_WANT_SYNC || in xlog_state_do_callback()
2813 iclog->ic_state == XLOG_STATE_SYNCING || in xlog_state_do_callback()
2814 iclog->ic_state == XLOG_STATE_DONE_SYNC || in xlog_state_do_callback()
2815 iclog->ic_state == XLOG_STATE_IOERROR ) in xlog_state_do_callback()
2817 iclog = iclog->ic_next; in xlog_state_do_callback()
2818 } while (first_iclog != iclog); in xlog_state_do_callback()
2846 xlog_in_core_t *iclog, in xlog_state_done_syncing() argument
2849 struct xlog *log = iclog->ic_log; in xlog_state_done_syncing()
2853 ASSERT(iclog->ic_state == XLOG_STATE_SYNCING || in xlog_state_done_syncing()
2854 iclog->ic_state == XLOG_STATE_IOERROR); in xlog_state_done_syncing()
2855 ASSERT(atomic_read(&iclog->ic_refcnt) == 0); in xlog_state_done_syncing()
2856 ASSERT(iclog->ic_bwritecnt == 1 || iclog->ic_bwritecnt == 2); in xlog_state_done_syncing()
2865 if (iclog->ic_state != XLOG_STATE_IOERROR) { in xlog_state_done_syncing()
2866 if (--iclog->ic_bwritecnt == 1) { in xlog_state_done_syncing()
2870 iclog->ic_state = XLOG_STATE_DONE_SYNC; in xlog_state_done_syncing()
2878 wake_up_all(&iclog->ic_write_wait); in xlog_state_done_syncing()
2880 xlog_state_do_callback(log, aborted, iclog); /* also cleans log */ in xlog_state_done_syncing()
2913 xlog_in_core_t *iclog; in xlog_state_get_iclog_space() local
2923 iclog = log->l_iclog; in xlog_state_get_iclog_space()
2924 if (iclog->ic_state != XLOG_STATE_ACTIVE) { in xlog_state_get_iclog_space()
2932 head = &iclog->ic_header; in xlog_state_get_iclog_space()
2934 atomic_inc(&iclog->ic_refcnt); /* prevents sync */ in xlog_state_get_iclog_space()
2935 log_offset = iclog->ic_offset; in xlog_state_get_iclog_space()
2962 if (iclog->ic_size - iclog->ic_offset < 2*sizeof(xlog_op_header_t)) { in xlog_state_get_iclog_space()
2963 xlog_state_switch_iclogs(log, iclog, iclog->ic_size); in xlog_state_get_iclog_space()
2972 if (!atomic_add_unless(&iclog->ic_refcnt, -1, 1)) { in xlog_state_get_iclog_space()
2975 error = xlog_state_release_iclog(log, iclog); in xlog_state_get_iclog_space()
2990 if (len <= iclog->ic_size - iclog->ic_offset) { in xlog_state_get_iclog_space()
2992 iclog->ic_offset += len; in xlog_state_get_iclog_space()
2995 xlog_state_switch_iclogs(log, iclog, iclog->ic_size); in xlog_state_get_iclog_space()
2997 *iclogp = iclog; in xlog_state_get_iclog_space()
2999 ASSERT(iclog->ic_offset <= iclog->ic_size); in xlog_state_get_iclog_space()
3103 struct xlog_in_core *iclog) in xlog_state_release_iclog() argument
3107 if (iclog->ic_state & XLOG_STATE_IOERROR) in xlog_state_release_iclog()
3110 ASSERT(atomic_read(&iclog->ic_refcnt) > 0); in xlog_state_release_iclog()
3111 if (!atomic_dec_and_lock(&iclog->ic_refcnt, &log->l_icloglock)) in xlog_state_release_iclog()
3114 if (iclog->ic_state & XLOG_STATE_IOERROR) { in xlog_state_release_iclog()
3118 ASSERT(iclog->ic_state == XLOG_STATE_ACTIVE || in xlog_state_release_iclog()
3119 iclog->ic_state == XLOG_STATE_WANT_SYNC); in xlog_state_release_iclog()
3121 if (iclog->ic_state == XLOG_STATE_WANT_SYNC) { in xlog_state_release_iclog()
3125 iclog->ic_state = XLOG_STATE_SYNCING; in xlog_state_release_iclog()
3126 iclog->ic_header.h_tail_lsn = cpu_to_be64(tail_lsn); in xlog_state_release_iclog()
3127 xlog_verify_tail_lsn(log, iclog, tail_lsn); in xlog_state_release_iclog()
3140 return xlog_sync(log, iclog); in xlog_state_release_iclog()
3155 struct xlog_in_core *iclog, in xlog_state_switch_iclogs() argument
3158 ASSERT(iclog->ic_state == XLOG_STATE_ACTIVE); in xlog_state_switch_iclogs()
3160 eventual_size = iclog->ic_offset; in xlog_state_switch_iclogs()
3161 iclog->ic_state = XLOG_STATE_WANT_SYNC; in xlog_state_switch_iclogs()
3162 iclog->ic_header.h_prev_block = cpu_to_be32(log->l_prev_block); in xlog_state_switch_iclogs()
3191 ASSERT(iclog == log->l_iclog); in xlog_state_switch_iclogs()
3192 log->l_iclog = iclog->ic_next; in xlog_state_switch_iclogs()
3229 struct xlog_in_core *iclog; in _xfs_log_force() local
3238 iclog = log->l_iclog; in _xfs_log_force()
3239 if (iclog->ic_state & XLOG_STATE_IOERROR) { in _xfs_log_force()
3247 if (iclog->ic_state == XLOG_STATE_ACTIVE || in _xfs_log_force()
3248 iclog->ic_state == XLOG_STATE_DIRTY) { in _xfs_log_force()
3256 if (iclog->ic_state == XLOG_STATE_DIRTY || in _xfs_log_force()
3257 (atomic_read(&iclog->ic_refcnt) == 0 in _xfs_log_force()
3258 && iclog->ic_offset == 0)) { in _xfs_log_force()
3259 iclog = iclog->ic_prev; in _xfs_log_force()
3260 if (iclog->ic_state == XLOG_STATE_ACTIVE || in _xfs_log_force()
3261 iclog->ic_state == XLOG_STATE_DIRTY) in _xfs_log_force()
3266 if (atomic_read(&iclog->ic_refcnt) == 0) { in _xfs_log_force()
3273 atomic_inc(&iclog->ic_refcnt); in _xfs_log_force()
3274 lsn = be64_to_cpu(iclog->ic_header.h_lsn); in _xfs_log_force()
3275 xlog_state_switch_iclogs(log, iclog, 0); in _xfs_log_force()
3278 if (xlog_state_release_iclog(log, iclog)) in _xfs_log_force()
3284 if (be64_to_cpu(iclog->ic_header.h_lsn) == lsn && in _xfs_log_force()
3285 iclog->ic_state != XLOG_STATE_DIRTY) in _xfs_log_force()
3295 xlog_state_switch_iclogs(log, iclog, 0); in _xfs_log_force()
3313 if (iclog->ic_state & XLOG_STATE_IOERROR) { in _xfs_log_force()
3318 xlog_wait(&iclog->ic_force_wait, &log->l_icloglock); in _xfs_log_force()
3324 if (iclog->ic_state & XLOG_STATE_IOERROR) in _xfs_log_force()
3377 struct xlog_in_core *iclog; in _xfs_log_force_lsn() local
3390 iclog = log->l_iclog; in _xfs_log_force_lsn()
3391 if (iclog->ic_state & XLOG_STATE_IOERROR) { in _xfs_log_force_lsn()
3397 if (be64_to_cpu(iclog->ic_header.h_lsn) != lsn) { in _xfs_log_force_lsn()
3398 iclog = iclog->ic_next; in _xfs_log_force_lsn()
3402 if (iclog->ic_state == XLOG_STATE_DIRTY) { in _xfs_log_force_lsn()
3407 if (iclog->ic_state == XLOG_STATE_ACTIVE) { in _xfs_log_force_lsn()
3427 (iclog->ic_prev->ic_state & in _xfs_log_force_lsn()
3429 ASSERT(!(iclog->ic_state & XLOG_STATE_IOERROR)); in _xfs_log_force_lsn()
3433 xlog_wait(&iclog->ic_prev->ic_write_wait, in _xfs_log_force_lsn()
3440 atomic_inc(&iclog->ic_refcnt); in _xfs_log_force_lsn()
3441 xlog_state_switch_iclogs(log, iclog, 0); in _xfs_log_force_lsn()
3443 if (xlog_state_release_iclog(log, iclog)) in _xfs_log_force_lsn()
3451 !(iclog->ic_state & in _xfs_log_force_lsn()
3457 if (iclog->ic_state & XLOG_STATE_IOERROR) { in _xfs_log_force_lsn()
3462 xlog_wait(&iclog->ic_force_wait, &log->l_icloglock); in _xfs_log_force_lsn()
3468 if (iclog->ic_state & XLOG_STATE_IOERROR) in _xfs_log_force_lsn()
3478 } while (iclog != log->l_iclog); in _xfs_log_force_lsn()
3510 struct xlog_in_core *iclog) in xlog_state_want_sync() argument
3514 if (iclog->ic_state == XLOG_STATE_ACTIVE) { in xlog_state_want_sync()
3515 xlog_state_switch_iclogs(log, iclog, 0); in xlog_state_want_sync()
3517 ASSERT(iclog->ic_state & in xlog_state_want_sync()
3760 struct xlog_in_core *iclog, in xlog_verify_tail_lsn() argument
3768 if (blocks < BTOBB(iclog->ic_offset)+BTOBB(log->l_iclog_hsize)) in xlog_verify_tail_lsn()
3777 if (blocks < BTOBB(iclog->ic_offset) + 1) in xlog_verify_tail_lsn()
3800 struct xlog_in_core *iclog, in xlog_verify_iclog() argument
3824 if (iclog->ic_header.h_magicno != cpu_to_be32(XLOG_HEADER_MAGIC_NUM)) in xlog_verify_iclog()
3827 base_ptr = ptr = &iclog->ic_header; in xlog_verify_iclog()
3828 p = &iclog->ic_header; in xlog_verify_iclog()
3836 len = be32_to_cpu(iclog->ic_header.h_num_logops); in xlog_verify_iclog()
3837 base_ptr = ptr = iclog->ic_datap; in xlog_verify_iclog()
3839 xhdr = iclog->ic_data; in xlog_verify_iclog()
3849 idx = BTOBBT((char *)&ophead->oh_clientid - iclog->ic_datap); in xlog_verify_iclog()
3857 iclog->ic_header.h_cycle_data[idx]); in xlog_verify_iclog()
3873 (uintptr_t)iclog->ic_datap); in xlog_verify_iclog()
3879 op_len = be32_to_cpu(iclog->ic_header.h_cycle_data[idx]); in xlog_verify_iclog()
3894 xlog_in_core_t *iclog, *ic; in xlog_state_ioerror() local
3896 iclog = log->l_iclog; in xlog_state_ioerror()
3897 if (! (iclog->ic_state & XLOG_STATE_IOERROR)) { in xlog_state_ioerror()
3902 ic = iclog; in xlog_state_ioerror()
3906 } while (ic != iclog); in xlog_state_ioerror()
4010 xlog_in_core_t *iclog; in xfs_log_force_umount() local
4013 iclog = log->l_iclog; in xfs_log_force_umount()
4015 ASSERT(iclog->ic_callback == 0); in xfs_log_force_umount()
4016 iclog = iclog->ic_next; in xfs_log_force_umount()
4017 } while (iclog != log->l_iclog); in xfs_log_force_umount()
4029 xlog_in_core_t *iclog; in xlog_iclogs_empty() local
4031 iclog = log->l_iclog; in xlog_iclogs_empty()
4036 if (iclog->ic_header.h_num_logops) in xlog_iclogs_empty()
4038 iclog = iclog->ic_next; in xlog_iclogs_empty()
4039 } while (iclog != log->l_iclog); in xlog_iclogs_empty()