jhead 197 fs/ubifs/debug.c const char *dbg_jhead(int jhead) jhead 199 fs/ubifs/debug.c switch (jhead) { jhead 405 fs/ubifs/debug.c pr_err("\tjhead %u\n", le32_to_cpu(ref->jhead)); jhead 618 fs/ubifs/debug.c dbg_jhead(c->jheads[i].wbuf.jhead), jhead 726 fs/ubifs/debug.c dbg_jhead(bud->jhead)); jhead 239 fs/ubifs/debug.h const char *dbg_jhead(int jhead); jhead 685 fs/ubifs/gc.c mutex_lock_nested(&wbuf->io_mutex, wbuf->jhead); jhead 946 fs/ubifs/gc.c mutex_lock_nested(&wbuf->io_mutex, wbuf->jhead); jhead 478 fs/ubifs/io.c dbg_io("jhead %s", dbg_jhead(wbuf->jhead)); jhead 504 fs/ubifs/io.c dbg_jhead(wbuf->jhead), jhead 547 fs/ubifs/io.c wbuf->lnum, wbuf->offs, wbuf->used, dbg_jhead(wbuf->jhead)); jhead 615 fs/ubifs/io.c dbg_io("LEB %d:%d, jhead %s", lnum, offs, dbg_jhead(wbuf->jhead)); jhead 673 fs/ubifs/io.c mutex_lock_nested(&wbuf->io_mutex, wbuf->jhead); jhead 695 fs/ubifs/io.c mutex_lock_nested(&wbuf->io_mutex, wbuf->jhead); jhead 725 fs/ubifs/io.c dbg_jhead(wbuf->jhead), wbuf->lnum, wbuf->offs + wbuf->used); jhead 758 fs/ubifs/io.c dbg_jhead(wbuf->jhead), wbuf->lnum, wbuf->offs); jhead 793 fs/ubifs/io.c dbg_jhead(wbuf->jhead), wbuf->lnum, wbuf->offs); jhead 972 fs/ubifs/io.c dbg_ntype(type), len, dbg_jhead(wbuf->jhead)); jhead 1202 fs/ubifs/io.c mutex_lock_nested(&wbuf->io_mutex, wbuf->jhead); jhead 98 fs/ubifs/journal.c static int reserve_space(struct ubifs_info *c, int jhead, int len) jhead 101 fs/ubifs/journal.c struct ubifs_wbuf *wbuf = &c->jheads[jhead].wbuf; jhead 109 fs/ubifs/journal.c squeeze = (jhead == BASEHD); jhead 111 fs/ubifs/journal.c mutex_lock_nested(&wbuf->io_mutex, wbuf->jhead); jhead 139 fs/ubifs/journal.c dbg_jnl("no free space in jhead %s, run GC", dbg_jhead(jhead)); jhead 155 fs/ubifs/journal.c dbg_jhead(jhead)); jhead 165 fs/ubifs/journal.c mutex_lock_nested(&wbuf->io_mutex, wbuf->jhead); jhead 166 fs/ubifs/journal.c dbg_jnl("got LEB %d for jhead %s", lnum, dbg_jhead(jhead)); jhead 197 fs/ubifs/journal.c err = ubifs_add_bud_to_log(c, jhead, lnum, offs); jhead 268 fs/ubifs/journal.c static int write_head(struct ubifs_info *c, int jhead, void *buf, int len, jhead 272 fs/ubifs/journal.c struct ubifs_wbuf *wbuf = &c->jheads[jhead].wbuf; jhead 274 fs/ubifs/journal.c ubifs_assert(c, jhead != GCHD); jhead 276 fs/ubifs/journal.c *lnum = c->jheads[jhead].wbuf.lnum; jhead 277 fs/ubifs/journal.c *offs = c->jheads[jhead].wbuf.offs + c->jheads[jhead].wbuf.used; jhead 279 fs/ubifs/journal.c dbg_jhead(jhead), *lnum, *offs, len); jhead 282 fs/ubifs/journal.c err = ubifs_hash_nodes(c, buf, len, c->jheads[jhead].log_hash); jhead 311 fs/ubifs/journal.c static int make_reservation(struct ubifs_info *c, int jhead, int len) jhead 317 fs/ubifs/journal.c err = reserve_space(c, jhead, len); jhead 373 fs/ubifs/journal.c len, jhead, err); jhead 395 fs/ubifs/journal.c static inline void release_head(struct ubifs_info *c, int jhead) jhead 397 fs/ubifs/journal.c mutex_unlock(&c->jheads[jhead].wbuf.io_mutex); jhead 63 fs/ubifs/log.c int jhead; jhead 77 fs/ubifs/log.c jhead = bud->jhead; jhead 79 fs/ubifs/log.c return &c->jheads[jhead].wbuf; jhead 116 fs/ubifs/log.c struct ubifs_jhead *jhead; jhead 133 fs/ubifs/log.c jhead = &c->jheads[bud->jhead]; jhead 134 fs/ubifs/log.c list_add_tail(&bud->list, &jhead->buds_list); jhead 147 fs/ubifs/log.c bud->start, dbg_jhead(bud->jhead), c->bud_bytes); jhead 164 fs/ubifs/log.c int ubifs_add_bud_to_log(struct ubifs_info *c, int jhead, int lnum, int offs) jhead 226 fs/ubifs/log.c bud->jhead = jhead; jhead 232 fs/ubifs/log.c ref->jhead = cpu_to_le32(jhead); jhead 271 fs/ubifs/log.c err = ubifs_shash_copy_state(c, c->log_hash, c->jheads[jhead].log_hash); jhead 312 fs/ubifs/log.c wbuf = &c->jheads[bud->jhead].wbuf; jhead 321 fs/ubifs/log.c bud->lnum, bud->start, dbg_jhead(bud->jhead), jhead 327 fs/ubifs/log.c bud->lnum, bud->start, dbg_jhead(bud->jhead), jhead 406 fs/ubifs/log.c ref->jhead = cpu_to_le32(i); jhead 131 fs/ubifs/misc.h mutex_lock_nested(&wbuf->io_mutex, wbuf->jhead); jhead 623 fs/ubifs/recovery.c int offs, void *sbuf, int jhead) jhead 626 fs/ubifs/recovery.c int grouped = jhead == -1 ? 0 : c->jheads[jhead].grouped; jhead 630 fs/ubifs/recovery.c dbg_rcvry("%d:%d, jhead %d, grouped %d", lnum, offs, jhead, grouped); jhead 710 fs/ubifs/recovery.c if (jhead == GCHD) { jhead 1192 fs/ubifs/recovery.c mutex_lock_nested(&wbuf->io_mutex, wbuf->jhead); jhead 147 fs/ubifs/replay.c err = ubifs_wbuf_seek_nolock(&c->jheads[b->bud->jhead].wbuf, jhead 515 fs/ubifs/replay.c struct ubifs_jhead *jh = &c->jheads[bud->jhead]; jhead 680 fs/ubifs/replay.c lnum, b->bud->jhead, offs, is_last); jhead 689 fs/ubifs/replay.c sleb = ubifs_recover_leb(c, lnum, offs, c->sbuf, b->bud->jhead); jhead 702 fs/ubifs/replay.c c->jheads[b->bud->jhead].log_hash); jhead 894 fs/ubifs/replay.c static int add_replay_bud(struct ubifs_info *c, int lnum, int offs, int jhead, jhead 901 fs/ubifs/replay.c dbg_mnt("add replay bud LEB %d:%d, head %d", lnum, offs, jhead); jhead 915 fs/ubifs/replay.c bud->jhead = jhead; jhead 954 fs/ubifs/replay.c unsigned int jhead = le32_to_cpu(ref->jhead); jhead 961 fs/ubifs/replay.c if (jhead >= c->jhead_cnt || lnum >= c->leb_cnt || jhead 969 fs/ubifs/replay.c if (bud->jhead == jhead && bud->start <= offs) jhead 1103 fs/ubifs/replay.c le32_to_cpu(ref->jhead), jhead 820 fs/ubifs/super.c c->jheads[i].wbuf.jhead = i; jhead 770 fs/ubifs/ubifs-media.h __le32 jhead; jhead 689 fs/ubifs/ubifs.h int jhead; jhead 712 fs/ubifs/ubifs.h int jhead; jhead 1754 fs/ubifs/ubifs.h int ubifs_add_bud_to_log(struct ubifs_info *c, int jhead, int lnum, int offs); jhead 2040 fs/ubifs/ubifs.h int offs, void *sbuf, int jhead);