m_la 54 fs/erofs/data.c u64 offset = map->m_la; m_la 74 fs/erofs/data.c map->m_pa = blknr_to_addr(vi->raw_blkaddr) + map->m_la; m_la 81 fs/erofs/data.c vi->xattr_isize + erofs_blkoff(map->m_la); m_la 98 fs/erofs/data.c vi->nid, inode->i_size, map->m_la); m_la 157 fs/erofs/data.c .m_la = blknr_to_addr(current_block), m_la 330 fs/erofs/data.c .m_la = iblock << 9, m_la 325 fs/erofs/internal.h erofs_off_t m_pa, m_la; m_la 362 fs/erofs/zdata.c if (cl->pageofs != (map->m_la & ~PAGE_MASK)) { m_la 433 fs/erofs/zdata.c cl->pageofs = map->m_la & ~PAGE_MASK; m_la 594 fs/erofs/zdata.c if (offset + cur >= map->m_la && m_la 595 fs/erofs/zdata.c offset + cur < map->m_la + map->m_llen) { m_la 608 fs/erofs/zdata.c map->m_la = offset + cur; m_la 623 fs/erofs/zdata.c if (should_alloc_managed_pages(fe, sbi->cache_strategy, map->m_la)) m_la 641 fs/erofs/zdata.c cur = end - min_t(unsigned int, offset + end - map->m_la, end); m_la 672 fs/erofs/zdata.c index = page->index - (map->m_la >> PAGE_SHIFT); m_la 682 fs/erofs/zdata.c map->m_llen = offset + cur - map->m_la; m_la 364 fs/erofs/zmap.c map->m_la = (lcn << lclusterbits) | m->clusterofs; m_la 392 fs/erofs/zmap.c if (map->m_la >= inode->i_size) { m_la 393 fs/erofs/zmap.c map->m_llen = map->m_la + 1 - inode->i_size; m_la 394 fs/erofs/zmap.c map->m_la = inode->i_size; m_la 404 fs/erofs/zmap.c ofs = map->m_la; m_la 422 fs/erofs/zmap.c map->m_la = (m.lcn << lclusterbits) | m.clusterofs; m_la 451 fs/erofs/zmap.c map->m_llen = end - map->m_la; m_la 462 fs/erofs/zmap.c __func__, map->m_la, map->m_pa, m_la 161 include/trace/events/erofs.h __entry->la = map->m_la; m_la 208 include/trace/events/erofs.h __entry->la = map->m_la;