Lines Matching refs:size

80 static inline phys_addr_t memblock_cap_size(phys_addr_t base, phys_addr_t *size)  in memblock_cap_size()  argument
82 return *size = min(*size, (phys_addr_t)ULLONG_MAX - base); in memblock_cap_size()
95 phys_addr_t base, phys_addr_t size) in memblock_overlaps_region() argument
101 phys_addr_t rgnsize = type->regions[i].size; in memblock_overlaps_region()
102 if (memblock_addrs_overlap(base, size, rgnbase, rgnsize)) in memblock_overlaps_region()
125 phys_addr_t size, phys_addr_t align, int nid, in __memblock_find_range_bottom_up() argument
136 if (cand < this_end && this_end - cand >= size) in __memblock_find_range_bottom_up()
159 phys_addr_t size, phys_addr_t align, int nid, in __memblock_find_range_top_down() argument
170 if (this_end < size) in __memblock_find_range_top_down()
173 cand = round_down(this_end - size, align); in __memblock_find_range_top_down()
203 phys_addr_t __init_memblock memblock_find_in_range_node(phys_addr_t size, in memblock_find_in_range_node() argument
230 size, align, nid, flags); in memblock_find_in_range_node()
248 return __memblock_find_range_top_down(start, end, size, align, nid, in memblock_find_in_range_node()
265 phys_addr_t end, phys_addr_t size, in memblock_find_in_range() argument
272 ret = memblock_find_in_range_node(size, align, start, end, in memblock_find_in_range()
277 &size); in memblock_find_in_range()
287 type->total_size -= type->regions[r].size; in memblock_remove_region()
297 type->regions[0].size = 0; in memblock_remove_region()
463 if (this->base + this->size != next->base || in memblock_merge_regions()
467 BUG_ON(this->base + this->size > next->base); in memblock_merge_regions()
472 this->size += next->size; in memblock_merge_regions()
493 phys_addr_t size, in memblock_insert_region() argument
501 rgn->size = size; in memblock_insert_region()
505 type->total_size += size; in memblock_insert_region()
525 phys_addr_t base, phys_addr_t size, in memblock_add_range() argument
530 phys_addr_t end = base + memblock_cap_size(base, &size); in memblock_add_range()
533 if (!size) in memblock_add_range()
537 if (type->regions[0].size == 0) { in memblock_add_range()
540 type->regions[0].size = size; in memblock_add_range()
543 type->total_size = size; in memblock_add_range()
558 phys_addr_t rend = rbase + rgn->size; in memblock_add_range()
597 if (memblock_double_array(type, obase, size) < 0) in memblock_add_range()
607 int __init_memblock memblock_add_node(phys_addr_t base, phys_addr_t size, in memblock_add_node() argument
610 return memblock_add_range(&memblock.memory, base, size, nid, 0); in memblock_add_node()
614 phys_addr_t size, in memblock_add_region() argument
622 (unsigned long long)base + size - 1, in memblock_add_region()
625 return memblock_add_range(type, base, size, nid, flags); in memblock_add_region()
628 int __init_memblock memblock_add(phys_addr_t base, phys_addr_t size) in memblock_add() argument
630 return memblock_add_region(base, size, MAX_NUMNODES, 0); in memblock_add()
650 phys_addr_t base, phys_addr_t size, in memblock_isolate_range() argument
653 phys_addr_t end = base + memblock_cap_size(base, &size); in memblock_isolate_range()
658 if (!size) in memblock_isolate_range()
663 if (memblock_double_array(type, base, size) < 0) in memblock_isolate_range()
669 phys_addr_t rend = rbase + rgn->size; in memblock_isolate_range()
682 rgn->size -= base - rbase; in memblock_isolate_range()
693 rgn->size -= end - rbase; in memblock_isolate_range()
710 phys_addr_t base, phys_addr_t size) in memblock_remove_range() argument
715 ret = memblock_isolate_range(type, base, size, &start_rgn, &end_rgn); in memblock_remove_range()
724 int __init_memblock memblock_remove(phys_addr_t base, phys_addr_t size) in memblock_remove() argument
726 return memblock_remove_range(&memblock.memory, base, size); in memblock_remove()
730 int __init_memblock memblock_free(phys_addr_t base, phys_addr_t size) in memblock_free() argument
734 (unsigned long long)base + size - 1, in memblock_free()
737 kmemleak_free_part(__va(base), size); in memblock_free()
738 return memblock_remove_range(&memblock.reserved, base, size); in memblock_free()
742 phys_addr_t size, in memblock_reserve_region() argument
750 (unsigned long long)base + size - 1, in memblock_reserve_region()
753 return memblock_add_range(type, base, size, nid, flags); in memblock_reserve_region()
756 int __init_memblock memblock_reserve(phys_addr_t base, phys_addr_t size) in memblock_reserve() argument
758 return memblock_reserve_region(base, size, MAX_NUMNODES, 0); in memblock_reserve()
768 phys_addr_t size, int set, int flag) in memblock_setclr_flag() argument
773 ret = memblock_isolate_range(type, base, size, &start_rgn, &end_rgn); in memblock_setclr_flag()
794 int __init_memblock memblock_mark_hotplug(phys_addr_t base, phys_addr_t size) in memblock_mark_hotplug() argument
796 return memblock_setclr_flag(base, size, 1, MEMBLOCK_HOTPLUG); in memblock_mark_hotplug()
806 int __init_memblock memblock_clear_hotplug(phys_addr_t base, phys_addr_t size) in memblock_clear_hotplug() argument
808 return memblock_setclr_flag(base, size, 0, MEMBLOCK_HOTPLUG); in memblock_clear_hotplug()
818 int __init_memblock memblock_mark_mirror(phys_addr_t base, phys_addr_t size) in memblock_mark_mirror() argument
822 return memblock_setclr_flag(base, size, 1, MEMBLOCK_MIRROR); in memblock_mark_mirror()
843 phys_addr_t size = r->size; in __next_reserved_mem_region() local
848 *out_end = base + size - 1; in __next_reserved_mem_region()
901 phys_addr_t m_end = m->base + m->size; in __next_mem_range()
935 r_start = idx_b ? r[-1].base + r[-1].size : 0; in __next_mem_range()
1010 phys_addr_t m_end = m->base + m->size; in __next_mem_range_rev()
1044 r_start = idx_b ? r[-1].base + r[-1].size : 0; in __next_mem_range_rev()
1089 if (PFN_UP(r->base) >= PFN_DOWN(r->base + r->size)) in __next_mem_pfn_range()
1102 *out_end_pfn = PFN_DOWN(r->base + r->size); in __next_mem_pfn_range()
1120 int __init_memblock memblock_set_node(phys_addr_t base, phys_addr_t size, in memblock_set_node() argument
1126 ret = memblock_isolate_range(type, base, size, &start_rgn, &end_rgn); in memblock_set_node()
1138 static phys_addr_t __init memblock_alloc_range_nid(phys_addr_t size, in memblock_alloc_range_nid() argument
1147 found = memblock_find_in_range_node(size, align, start, end, nid, in memblock_alloc_range_nid()
1149 if (found && !memblock_reserve(found, size)) { in memblock_alloc_range_nid()
1154 kmemleak_alloc(__va(found), size, 0, 0); in memblock_alloc_range_nid()
1160 phys_addr_t __init memblock_alloc_range(phys_addr_t size, phys_addr_t align, in memblock_alloc_range() argument
1164 return memblock_alloc_range_nid(size, align, start, end, NUMA_NO_NODE, in memblock_alloc_range()
1168 static phys_addr_t __init memblock_alloc_base_nid(phys_addr_t size, in memblock_alloc_base_nid() argument
1172 return memblock_alloc_range_nid(size, align, 0, max_addr, nid, flags); in memblock_alloc_base_nid()
1175 phys_addr_t __init memblock_alloc_nid(phys_addr_t size, phys_addr_t align, int nid) in memblock_alloc_nid() argument
1181 ret = memblock_alloc_base_nid(size, align, MEMBLOCK_ALLOC_ACCESSIBLE, in memblock_alloc_nid()
1191 phys_addr_t __init __memblock_alloc_base(phys_addr_t size, phys_addr_t align, phys_addr_t max_addr) in __memblock_alloc_base() argument
1193 return memblock_alloc_base_nid(size, align, max_addr, NUMA_NO_NODE, in __memblock_alloc_base()
1197 phys_addr_t __init memblock_alloc_base(phys_addr_t size, phys_addr_t align, phys_addr_t max_addr) in memblock_alloc_base() argument
1201 alloc = __memblock_alloc_base(size, align, max_addr); in memblock_alloc_base()
1205 (unsigned long long) size, (unsigned long long) max_addr); in memblock_alloc_base()
1210 phys_addr_t __init memblock_alloc(phys_addr_t size, phys_addr_t align) in memblock_alloc() argument
1212 return memblock_alloc_base(size, align, MEMBLOCK_ALLOC_ACCESSIBLE); in memblock_alloc()
1215 phys_addr_t __init memblock_alloc_try_nid(phys_addr_t size, phys_addr_t align, int nid) in memblock_alloc_try_nid() argument
1217 phys_addr_t res = memblock_alloc_nid(size, align, nid); in memblock_alloc_try_nid()
1221 return memblock_alloc_base(size, align, MEMBLOCK_ALLOC_ACCESSIBLE); in memblock_alloc_try_nid()
1252 phys_addr_t size, phys_addr_t align, in memblock_virt_alloc_internal() argument
1269 return kzalloc_node(size, GFP_NOWAIT, nid); in memblock_virt_alloc_internal()
1278 alloc = memblock_find_in_range_node(size, align, min_addr, max_addr, in memblock_virt_alloc_internal()
1284 alloc = memblock_find_in_range_node(size, align, min_addr, in memblock_virt_alloc_internal()
1299 &size); in memblock_virt_alloc_internal()
1305 memblock_reserve(alloc, size); in memblock_virt_alloc_internal()
1307 memset(ptr, 0, size); in memblock_virt_alloc_internal()
1315 kmemleak_alloc(ptr, size, 0, 0); in memblock_virt_alloc_internal()
1338 phys_addr_t size, phys_addr_t align, in memblock_virt_alloc_try_nid_nopanic() argument
1343 __func__, (u64)size, (u64)align, nid, (u64)min_addr, in memblock_virt_alloc_try_nid_nopanic()
1345 return memblock_virt_alloc_internal(size, align, min_addr, in memblock_virt_alloc_try_nid_nopanic()
1368 phys_addr_t size, phys_addr_t align, in memblock_virt_alloc_try_nid() argument
1375 __func__, (u64)size, (u64)align, nid, (u64)min_addr, in memblock_virt_alloc_try_nid()
1377 ptr = memblock_virt_alloc_internal(size, align, in memblock_virt_alloc_try_nid()
1383 __func__, (u64)size, (u64)align, nid, (u64)min_addr, in memblock_virt_alloc_try_nid()
1396 void __init __memblock_free_early(phys_addr_t base, phys_addr_t size) in __memblock_free_early() argument
1399 __func__, (u64)base, (u64)base + size - 1, in __memblock_free_early()
1401 kmemleak_free_part(__va(base), size); in __memblock_free_early()
1402 memblock_remove_range(&memblock.reserved, base, size); in __memblock_free_early()
1414 void __init __memblock_free_late(phys_addr_t base, phys_addr_t size) in __memblock_free_late() argument
1419 __func__, (u64)base, (u64)base + size - 1, in __memblock_free_late()
1421 kmemleak_free_part(__va(base), size); in __memblock_free_late()
1423 end = PFN_DOWN(base + size); in __memblock_free_late()
1467 return (memblock.memory.regions[idx].base + memblock.memory.regions[idx].size); in memblock_end_of_DRAM()
1480 if (limit <= r->size) { in memblock_enforce_memory_limit()
1484 limit -= r->size; in memblock_enforce_memory_limit()
1504 type->regions[mid].size)) in memblock_search()
1533 *end_pfn = PFN_DOWN(type->regions[mid].base + type->regions[mid].size); in memblock_search_pfn_nid()
1549 int __init_memblock memblock_is_region_memory(phys_addr_t base, phys_addr_t size) in memblock_is_region_memory() argument
1552 phys_addr_t end = base + memblock_cap_size(base, &size); in memblock_is_region_memory()
1558 memblock.memory.regions[idx].size) >= end; in memblock_is_region_memory()
1571 bool __init_memblock memblock_is_region_reserved(phys_addr_t base, phys_addr_t size) in memblock_is_region_reserved() argument
1573 memblock_cap_size(base, &size); in memblock_is_region_reserved()
1574 return memblock_overlaps_region(&memblock.reserved, base, size); in memblock_is_region_reserved()
1584 orig_end = r->base + r->size; in memblock_trim_memory()
1593 r->size = end - start; in memblock_trim_memory()
1614 unsigned long long base, size; in memblock_dump() local
1625 size = rgn->size; in memblock_dump()
1633 name, i, base, base + size - 1, size, nid_buf, flags); in memblock_dump()
1675 (unsigned long)(reg->base + reg->size - 1)); in memblock_debug_show()
1679 (unsigned long long)(reg->base + reg->size - 1)); in memblock_debug_show()