refchk 97 fs/xfs/scrub/refcount.c struct xchk_refcnt_check *refchk = priv; refchk 103 fs/xfs/scrub/refcount.c if (xchk_should_terminate(refchk->sc, &error)) refchk 107 fs/xfs/scrub/refcount.c rc_last = refchk->bno + refchk->len - 1; refchk 110 fs/xfs/scrub/refcount.c if (refchk->refcount == 1 && rec->rm_owner != XFS_RMAP_OWN_COW) { refchk 111 fs/xfs/scrub/refcount.c xchk_btree_xref_set_corrupt(refchk->sc, cur, 0); refchk 115 fs/xfs/scrub/refcount.c if (rec->rm_startblock <= refchk->bno && rm_last >= rc_last) { refchk 120 fs/xfs/scrub/refcount.c refchk->seen++; refchk 133 fs/xfs/scrub/refcount.c list_add_tail(&frag->list, &refchk->fragments); refchk 148 fs/xfs/scrub/refcount.c struct xchk_refcnt_check *refchk) refchk 159 fs/xfs/scrub/refcount.c target_nr = refchk->refcount - refchk->seen; refchk 177 fs/xfs/scrub/refcount.c list_for_each_entry(frag, &refchk->fragments, list) { refchk 187 fs/xfs/scrub/refcount.c list_for_each_entry_safe(frag, n, &refchk->fragments, list) { refchk 188 fs/xfs/scrub/refcount.c if (frag->rm.rm_startblock > refchk->bno) refchk 206 fs/xfs/scrub/refcount.c while (!list_empty(&refchk->fragments)) { refchk 223 fs/xfs/scrub/refcount.c list_for_each_entry_safe(frag, n, &refchk->fragments, list) { refchk 251 fs/xfs/scrub/refcount.c if (rbno < refchk->bno + refchk->len) refchk 255 fs/xfs/scrub/refcount.c refchk->seen = refchk->refcount; refchk 262 fs/xfs/scrub/refcount.c list_for_each_entry_safe(frag, n, &refchk->fragments, list) { refchk 276 fs/xfs/scrub/refcount.c struct xchk_refcnt_check refchk = { refchk 298 fs/xfs/scrub/refcount.c INIT_LIST_HEAD(&refchk.fragments); refchk 300 fs/xfs/scrub/refcount.c &xchk_refcountbt_rmap_check, &refchk); refchk 304 fs/xfs/scrub/refcount.c xchk_refcountbt_process_rmap_fragments(&refchk); refchk 305 fs/xfs/scrub/refcount.c if (refcount != refchk.seen) refchk 309 fs/xfs/scrub/refcount.c list_for_each_entry_safe(frag, n, &refchk.fragments, list) {