res_end 438 arch/arm/mm/init.c unsigned long res_start, res_end; res_end 441 arch/arm/mm/init.c res_end = memblock_region_reserved_end_pfn(res); res_end 443 arch/arm/mm/init.c if (res_end < start) res_end 449 arch/arm/mm/init.c if (res_end > end) res_end 450 arch/arm/mm/init.c res_end = end; res_end 453 arch/arm/mm/init.c start = res_end; res_end 479 arch/ia64/hp/common/sba_iommu.c unsigned long *res_end = (unsigned long *) &(ioc->res_map[ioc->res_size]); res_end 486 arch/ia64/hp/common/sba_iommu.c ASSERT(res_ptr < res_end); res_end 514 arch/ia64/hp/common/sba_iommu.c for(; res_ptr < res_end ; res_ptr++) { res_end 542 arch/ia64/hp/common/sba_iommu.c for(; res_ptr < res_end ; res_ptr++) res_end 571 arch/ia64/hp/common/sba_iommu.c end = res_end - qwords; res_end 226 arch/parisc/kernel/pci-dma.c for(; res_ptr < res_end; ++res_ptr) \ res_end 238 arch/parisc/kernel/pci-dma.c u##size *res_end = (u##size *)&pcxl_res_map[pcxl_res_size]; \ res_end 156 arch/powerpc/include/asm/iommu.h int nid, unsigned long res_start, unsigned long res_end); res_end 637 arch/powerpc/kernel/iommu.c unsigned long res_start, unsigned long res_end) res_end 641 arch/powerpc/kernel/iommu.c WARN_ON_ONCE(res_end < res_start); res_end 651 arch/powerpc/kernel/iommu.c tbl->it_reserved_end = res_end; res_end 654 arch/powerpc/kernel/iommu.c if (res_start && res_end && res_end 656 arch/powerpc/kernel/iommu.c res_end < tbl->it_offset)) res_end 683 arch/powerpc/kernel/iommu.c unsigned long res_start, unsigned long res_end) res_end 702 arch/powerpc/kernel/iommu.c iommu_table_reserve_pages(tbl, res_start, res_end); res_end 2396 arch/powerpc/platforms/powernv/pci-ioda.c unsigned long res_start, res_end; res_end 2444 arch/powerpc/platforms/powernv/pci-ioda.c res_end = 0; res_end 2447 arch/powerpc/platforms/powernv/pci-ioda.c res_end = min(window_size, SZ_4G) >> tbl->it_page_shift; res_end 2449 arch/powerpc/platforms/powernv/pci-ioda.c iommu_init_table(tbl, pe->phb->hose->node, res_start, res_end); res_end 113 arch/xtensa/mm/init.c unsigned long res_start, res_end; res_end 116 arch/xtensa/mm/init.c res_end = memblock_region_reserved_end_pfn(res); res_end 118 arch/xtensa/mm/init.c if (res_end < start) res_end 124 arch/xtensa/mm/init.c if (res_end > end) res_end 125 arch/xtensa/mm/init.c res_end = end; res_end 128 arch/xtensa/mm/init.c start = res_end; res_end 1188 drivers/nvdimm/region_devs.c resource_size_t res_end, region_end, region_start; res_end 1197 drivers/nvdimm/region_devs.c res_end = ctx->start + ctx->size; res_end 1202 drivers/nvdimm/region_devs.c if (res_end > region_start && res_end <= region_end) res_end 295 drivers/parisc/ccio-dma.c for(; res_ptr < res_end; ++res_ptr) { \ res_end 310 drivers/parisc/ccio-dma.c u##size *res_end = (u##size *)&(ioc)->res_map[ioa->res_size]; \ res_end 339 drivers/parisc/sba_iommu.c unsigned long *res_end = (unsigned long *) &(ioc->res_map[ioc->res_size]); res_end 357 drivers/parisc/sba_iommu.c for(; res_ptr < res_end; ++res_ptr) { res_end 389 drivers/parisc/sba_iommu.c while(res_ptr < res_end) res_end 415 drivers/parisc/sba_iommu.c if (res_end <= res_ptr) {