rgn 104 arch/powerpc/platforms/powernv/opal-fadump.c fadump_conf->boot_mem_dest_addr = fdm->rgn[0].dest; rgn 130 arch/powerpc/platforms/powernv/opal-fadump.c base = fdm->rgn[i].src; rgn 131 arch/powerpc/platforms/powernv/opal-fadump.c size = fdm->rgn[i].size; rgn 146 arch/powerpc/platforms/powernv/opal-fadump.c fadump_conf->reserve_dump_area_start = fdm->rgn[0].dest; rgn 164 arch/powerpc/platforms/powernv/opal-fadump.c i, fdm->rgn[i].src, fdm->rgn[i].size); rgn 196 arch/powerpc/platforms/powernv/opal-fadump.c opal_fdm->rgn[i].src = fadump_conf->boot_mem_addr[i]; rgn 197 arch/powerpc/platforms/powernv/opal-fadump.c opal_fdm->rgn[i].dest = addr; rgn 198 arch/powerpc/platforms/powernv/opal-fadump.c opal_fdm->rgn[i].size = fadump_conf->boot_mem_sz[i]; rgn 208 arch/powerpc/platforms/powernv/opal-fadump.c opal_fdm->fadumphdr_addr = (opal_fdm->rgn[0].dest + rgn 276 arch/powerpc/platforms/powernv/opal-fadump.c opal_fdm->rgn[i].src, rgn 277 arch/powerpc/platforms/powernv/opal-fadump.c opal_fdm->rgn[i].dest, rgn 278 arch/powerpc/platforms/powernv/opal-fadump.c opal_fdm->rgn[i].size); rgn 573 arch/powerpc/platforms/powernv/opal-fadump.c dumped_bytes = fdm_ptr->rgn[i].size; rgn 576 arch/powerpc/platforms/powernv/opal-fadump.c fdm_ptr->rgn[i].src, fdm_ptr->rgn[i].dest); rgn 578 arch/powerpc/platforms/powernv/opal-fadump.c fdm_ptr->rgn[i].size, dumped_bytes); rgn 42 arch/powerpc/platforms/powernv/opal-fadump.h struct opal_mpipl_region rgn[FADUMP_MAX_MEM_REGS]; rgn 65 drivers/mtd/nand/onenand/onenand_bbt.c int rgn; rgn 113 drivers/mtd/nand/onenand/onenand_bbt.c rgn = flexonenand_region(mtd, from); rgn 114 drivers/mtd/nand/onenand/onenand_bbt.c from += mtd->eraseregions[rgn].erasesize; rgn 529 include/linux/memblock.h #define for_each_memblock_type(i, memblock_type, rgn) \ rgn 530 include/linux/memblock.h for (i = 0, rgn = &memblock_type->regions[0]; \ rgn 532 include/linux/memblock.h i++, rgn = &memblock_type->regions[i]) rgn 554 mm/memblock.c struct memblock_region *rgn = &type->regions[idx]; rgn 557 mm/memblock.c memmove(rgn + 1, rgn, (type->cnt - idx) * sizeof(*rgn)); rgn 558 mm/memblock.c rgn->base = base; rgn 559 mm/memblock.c rgn->size = size; rgn 560 mm/memblock.c rgn->flags = flags; rgn 561 mm/memblock.c memblock_set_region_node(rgn, nid); rgn 590 mm/memblock.c struct memblock_region *rgn; rgn 614 mm/memblock.c for_each_memblock_type(idx, type, rgn) { rgn 615 mm/memblock.c phys_addr_t rbase = rgn->base; rgn 616 mm/memblock.c phys_addr_t rend = rbase + rgn->size; rgn 628 mm/memblock.c WARN_ON(nid != memblock_get_region_node(rgn)); rgn 630 mm/memblock.c WARN_ON(flags != rgn->flags); rgn 729 mm/memblock.c struct memblock_region *rgn; rgn 741 mm/memblock.c for_each_memblock_type(idx, type, rgn) { rgn 742 mm/memblock.c phys_addr_t rbase = rgn->base; rgn 743 mm/memblock.c phys_addr_t rend = rbase + rgn->size; rgn 755 mm/memblock.c rgn->base = base; rgn 756 mm/memblock.c rgn->size -= base - rbase; rgn 759 mm/memblock.c memblock_get_region_node(rgn), rgn 760 mm/memblock.c rgn->flags); rgn 766 mm/memblock.c rgn->base = end; rgn 767 mm/memblock.c rgn->size -= end - rbase; rgn 770 mm/memblock.c memblock_get_region_node(rgn), rgn 771 mm/memblock.c rgn->flags); rgn 1844 mm/memblock.c struct memblock_region *rgn; rgn 1848 mm/memblock.c for_each_memblock_type(idx, type, rgn) { rgn 1851 mm/memblock.c base = rgn->base; rgn 1852 mm/memblock.c size = rgn->size; rgn 1854 mm/memblock.c flags = rgn->flags; rgn 1856 mm/memblock.c if (memblock_get_region_node(rgn) != MAX_NUMNODES) rgn 1858 mm/memblock.c memblock_get_region_node(rgn));