/linux-4.1.27/include/linux/ |
D | memory_hotplug.h | 41 void pgdat_resize_lock(struct pglist_data *pgdat, unsigned long *flags) in pgdat_resize_lock() argument 43 spin_lock_irqsave(&pgdat->node_size_lock, *flags); in pgdat_resize_lock() 46 void pgdat_resize_unlock(struct pglist_data *pgdat, unsigned long *flags) in pgdat_resize_unlock() argument 48 spin_unlock_irqrestore(&pgdat->node_size_lock, *flags); in pgdat_resize_unlock() 51 void pgdat_resize_init(struct pglist_data *pgdat) in pgdat_resize_init() argument 53 spin_lock_init(&pgdat->node_size_lock); in pgdat_resize_init() 133 extern void arch_free_nodedata(pg_data_t *pgdat); 134 extern void arch_refresh_nodedata(int nid, pg_data_t *pgdat); 139 #define arch_free_nodedata(pgdat) generic_free_nodedata(pgdat) argument 156 #define generic_free_nodedata(pgdat) kfree(pgdat) argument [all …]
|
D | bootmem.h | 42 extern unsigned long init_bootmem_node(pg_data_t *pgdat, 49 extern void reset_node_managed_pages(pg_data_t *pgdat); 52 extern void free_bootmem_node(pg_data_t *pgdat, 72 extern int reserve_bootmem_node(pg_data_t *pgdat, 83 extern void *__alloc_bootmem_node(pg_data_t *pgdat, 87 void *__alloc_bootmem_node_high(pg_data_t *pgdat, 91 extern void *__alloc_bootmem_node_nopanic(pg_data_t *pgdat, 95 void *___alloc_bootmem_node_nopanic(pg_data_t *pgdat, 106 extern void *__alloc_bootmem_low_node(pg_data_t *pgdat, 128 #define alloc_bootmem_node(pgdat, x) \ argument [all …]
|
D | mmzone.h | 770 #define pgdat_page_nr(pgdat, pagenr) ((pgdat)->node_mem_map + (pagenr)) argument 772 #define pgdat_page_nr(pgdat, pagenr) pfn_to_page((pgdat)->node_start_pfn + (pagenr)) argument 779 static inline unsigned long pgdat_end_pfn(pg_data_t *pgdat) in pgdat_end_pfn() argument 781 return pgdat->node_start_pfn + pgdat->node_spanned_pages; in pgdat_end_pfn() 784 static inline bool pgdat_is_empty(pg_data_t *pgdat) in pgdat_is_empty() argument 786 return !pgdat->node_start_pfn && !pgdat->node_spanned_pages; in pgdat_is_empty() 792 void build_all_zonelists(pg_data_t *pgdat, struct zone *zone); 916 extern struct pglist_data *next_online_pgdat(struct pglist_data *pgdat); 923 #define for_each_online_pgdat(pgdat) \ argument 924 for (pgdat = first_online_pgdat(); \ [all …]
|
D | compaction.h | 43 extern void compact_pgdat(pg_data_t *pgdat, int order); 44 extern void reset_isolation_suitable(pg_data_t *pgdat); 63 static inline void compact_pgdat(pg_data_t *pgdat, int order) in compact_pgdat() argument 67 static inline void reset_isolation_suitable(pg_data_t *pgdat) in reset_isolation_suitable() argument
|
D | page_ext.h | 48 extern void pgdat_page_ext_init(struct pglist_data *pgdat); 67 static inline void pgdat_page_ext_init(struct pglist_data *pgdat) in pgdat_page_ext_init() argument
|
D | vmstat.h | 221 void set_pgdat_percpu_threshold(pg_data_t *pgdat, 270 #define set_pgdat_percpu_threshold(pgdat, callback) { } argument
|
/linux-4.1.27/mm/ |
D | nobootmem.c | 148 void reset_node_managed_pages(pg_data_t *pgdat) in reset_node_managed_pages() argument 152 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in reset_node_managed_pages() 158 struct pglist_data *pgdat; in reset_all_zones_managed_pages() local 163 for_each_online_pgdat(pgdat) in reset_all_zones_managed_pages() 164 reset_node_managed_pages(pgdat); in reset_all_zones_managed_pages() 201 void __init free_bootmem_node(pg_data_t *pgdat, unsigned long physaddr, in free_bootmem_node() argument 303 void * __init ___alloc_bootmem_node_nopanic(pg_data_t *pgdat, in ___alloc_bootmem_node_nopanic() argument 312 ptr = __alloc_memory_core_early(pgdat->node_id, size, align, in ___alloc_bootmem_node_nopanic() 330 void * __init __alloc_bootmem_node_nopanic(pg_data_t *pgdat, unsigned long size, in __alloc_bootmem_node_nopanic() argument 334 return kzalloc_node(size, GFP_NOWAIT, pgdat->node_id); in __alloc_bootmem_node_nopanic() [all …]
|
D | memory_hotplug.c | 245 void register_page_bootmem_info_node(struct pglist_data *pgdat) in register_page_bootmem_info_node() argument 248 int node = pgdat->node_id; in register_page_bootmem_info_node() 253 page = virt_to_page(pgdat); in register_page_bootmem_info_node() 258 zone = &pgdat->node_zones[0]; in register_page_bootmem_info_node() 259 for (; zone < pgdat->node_zones + MAX_NR_ZONES - 1; zone++) { in register_page_bootmem_info_node() 271 pfn = pgdat->node_start_pfn; in register_page_bootmem_info_node() 272 end_pfn = pgdat_end_pfn(pgdat); in register_page_bootmem_info_node() 431 static void __meminit grow_pgdat_span(struct pglist_data *pgdat, unsigned long start_pfn, in grow_pgdat_span() argument 434 unsigned long old_pgdat_end_pfn = pgdat_end_pfn(pgdat); in grow_pgdat_span() 436 if (!pgdat->node_spanned_pages || start_pfn < pgdat->node_start_pfn) in grow_pgdat_span() [all …]
|
D | mmzone.c | 17 struct pglist_data *next_online_pgdat(struct pglist_data *pgdat) in next_online_pgdat() argument 19 int nid = next_online_node(pgdat->node_id); in next_online_pgdat() 31 pg_data_t *pgdat = zone->zone_pgdat; in next_zone() local 33 if (zone < pgdat->node_zones + MAX_NR_ZONES - 1) in next_zone() 36 pgdat = next_online_pgdat(pgdat); in next_zone() 37 if (pgdat) in next_zone() 38 zone = pgdat->node_zones; in next_zone()
|
D | bootmem.c | 128 unsigned long __init init_bootmem_node(pg_data_t *pgdat, unsigned long freepfn, in init_bootmem_node() argument 131 return init_bootmem_core(pgdat->bdata, freepfn, startpfn, endpfn); in init_bootmem_node() 247 void reset_node_managed_pages(pg_data_t *pgdat) in reset_node_managed_pages() argument 251 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in reset_node_managed_pages() 257 struct pglist_data *pgdat; in reset_all_zones_managed_pages() local 262 for_each_online_pgdat(pgdat) in reset_all_zones_managed_pages() 263 reset_node_managed_pages(pgdat); in reset_all_zones_managed_pages() 393 void __init free_bootmem_node(pg_data_t *pgdat, unsigned long physaddr, in free_bootmem_node() argument 403 mark_bootmem_node(pgdat->bdata, start, end, 0, 0); in free_bootmem_node() 438 int __init reserve_bootmem_node(pg_data_t *pgdat, unsigned long physaddr, in reserve_bootmem_node() argument [all …]
|
D | vmstat.c | 196 void set_pgdat_percpu_threshold(pg_data_t *pgdat, in set_pgdat_percpu_threshold() argument 204 for (i = 0; i < pgdat->nr_zones; i++) { in set_pgdat_percpu_threshold() 205 zone = &pgdat->node_zones[i]; in set_pgdat_percpu_threshold() 857 pg_data_t *pgdat; in frag_start() local 860 for (pgdat = first_online_pgdat(); in frag_start() 861 pgdat && node; in frag_start() 862 pgdat = next_online_pgdat(pgdat)) in frag_start() 865 return pgdat; in frag_start() 870 pg_data_t *pgdat = (pg_data_t *)arg; in frag_next() local 873 return next_online_pgdat(pgdat); in frag_next() [all …]
|
D | vmscan.c | 2635 static bool pfmemalloc_watermark_ok(pg_data_t *pgdat) in pfmemalloc_watermark_ok() argument 2644 zone = &pgdat->node_zones[i]; in pfmemalloc_watermark_ok() 2659 if (!wmark_ok && waitqueue_active(&pgdat->kswapd_wait)) { in pfmemalloc_watermark_ok() 2660 pgdat->classzone_idx = min(pgdat->classzone_idx, in pfmemalloc_watermark_ok() 2662 wake_up_interruptible(&pgdat->kswapd_wait); in pfmemalloc_watermark_ok() 2682 pg_data_t *pgdat = NULL; in throttle_direct_reclaim() local 2721 pgdat = zone->zone_pgdat; in throttle_direct_reclaim() 2722 if (pfmemalloc_watermark_ok(pgdat)) in throttle_direct_reclaim() 2728 if (!pgdat) in throttle_direct_reclaim() 2743 wait_event_interruptible_timeout(pgdat->pfmemalloc_wait, in throttle_direct_reclaim() [all …]
|
D | page_owner.c | 207 static void init_pages_in_zone(pg_data_t *pgdat, struct zone *zone) in init_pages_in_zone() argument 265 pgdat->node_id, zone->name, count); in init_pages_in_zone() 268 static void init_zones_in_node(pg_data_t *pgdat) in init_zones_in_node() argument 271 struct zone *node_zones = pgdat->node_zones; in init_zones_in_node() 279 init_pages_in_zone(pgdat, zone); in init_zones_in_node() 286 pg_data_t *pgdat; in init_early_allocated_pages() local 289 for_each_online_pgdat(pgdat) in init_early_allocated_pages() 290 init_zones_in_node(pgdat); in init_early_allocated_pages()
|
D | page_alloc.c | 3201 pg_data_t *pgdat = NODE_DATA(nid); in si_meminfo_node() local 3204 managed_pages += pgdat->node_zones[zone_type].managed_pages; in si_meminfo_node() 3209 val->totalhigh = pgdat->node_zones[ZONE_HIGHMEM].managed_pages; in si_meminfo_node() 3210 val->freehigh = zone_page_state(&pgdat->node_zones[ZONE_HIGHMEM], in si_meminfo_node() 3451 static int build_zonelists_node(pg_data_t *pgdat, struct zonelist *zonelist, in build_zonelists_node() argument 3459 zone = pgdat->node_zones + zone_type; in build_zonelists_node() 3652 static void build_zonelists_in_node_order(pg_data_t *pgdat, int node) in build_zonelists_in_node_order() argument 3657 zonelist = &pgdat->node_zonelists[0]; in build_zonelists_in_node_order() 3668 static void build_thisnode_zonelists(pg_data_t *pgdat) in build_thisnode_zonelists() argument 3673 zonelist = &pgdat->node_zonelists[1]; in build_thisnode_zonelists() [all …]
|
D | sparse.c | 264 sparse_early_usemaps_alloc_pgdat_section(struct pglist_data *pgdat, in sparse_early_usemaps_alloc_pgdat_section() argument 280 goal = __pa(pgdat) & (PAGE_SECTION_MASK << PAGE_SHIFT); in sparse_early_usemaps_alloc_pgdat_section() 299 struct pglist_data *pgdat = NODE_DATA(nid); in check_usemap_section_nr() local 303 pgdat_snr = pfn_to_section_nr(__pa(pgdat) >> PAGE_SHIFT); in check_usemap_section_nr() 334 sparse_early_usemaps_alloc_pgdat_section(struct pglist_data *pgdat, in sparse_early_usemaps_alloc_pgdat_section() argument 337 return memblock_virt_alloc_node_nopanic(size, pgdat->node_id); in sparse_early_usemaps_alloc_pgdat_section() 696 struct pglist_data *pgdat = zone->zone_pgdat; in sparse_add_one_section() local 707 ret = sparse_index_init(section_nr, pgdat->node_id); in sparse_add_one_section() 710 memmap = kmalloc_section_memmap(section_nr, pgdat->node_id); in sparse_add_one_section() 719 pgdat_resize_lock(pgdat, &flags); in sparse_add_one_section() [all …]
|
D | migrate.c | 1531 static bool migrate_balanced_pgdat(struct pglist_data *pgdat, in migrate_balanced_pgdat() argument 1535 for (z = pgdat->nr_zones - 1; z >= 0; z--) { in migrate_balanced_pgdat() 1536 struct zone *zone = pgdat->node_zones + z; in migrate_balanced_pgdat() 1580 static bool numamigrate_update_ratelimit(pg_data_t *pgdat, in numamigrate_update_ratelimit() argument 1588 if (time_after(jiffies, pgdat->numabalancing_migrate_next_window)) { in numamigrate_update_ratelimit() 1589 spin_lock(&pgdat->numabalancing_migrate_lock); in numamigrate_update_ratelimit() 1590 pgdat->numabalancing_migrate_nr_pages = 0; in numamigrate_update_ratelimit() 1591 pgdat->numabalancing_migrate_next_window = jiffies + in numamigrate_update_ratelimit() 1593 spin_unlock(&pgdat->numabalancing_migrate_lock); in numamigrate_update_ratelimit() 1595 if (pgdat->numabalancing_migrate_nr_pages > ratelimit_pages) { in numamigrate_update_ratelimit() [all …]
|
D | mm_init.c | 32 pg_data_t *pgdat = NODE_DATA(nid); in mminit_verify_zonelist() local 44 zonelist = &pgdat->node_zonelists[listid]; in mminit_verify_zonelist() 45 zone = &pgdat->node_zones[zoneid]; in mminit_verify_zonelist()
|
D | highmem.c | 115 pg_data_t *pgdat; in nr_free_highpages() local 118 for_each_online_pgdat(pgdat) { in nr_free_highpages() 119 pages += zone_page_state(&pgdat->node_zones[ZONE_HIGHMEM], in nr_free_highpages() 123 &pgdat->node_zones[ZONE_MOVABLE], in nr_free_highpages()
|
D | page_ext.c | 93 void __meminit pgdat_page_ext_init(struct pglist_data *pgdat) in pgdat_page_ext_init() argument 95 pgdat->node_page_ext = NULL; in pgdat_page_ext_init() 399 void __meminit pgdat_page_ext_init(struct pglist_data *pgdat) in pgdat_page_ext_init() argument
|
D | compaction.c | 243 void reset_isolation_suitable(pg_data_t *pgdat) in reset_isolation_suitable() argument 248 struct zone *zone = &pgdat->node_zones[zoneid]; in reset_isolation_suitable() 1584 static void __compact_pgdat(pg_data_t *pgdat, struct compact_control *cc) in __compact_pgdat() argument 1591 zone = &pgdat->node_zones[zoneid]; in __compact_pgdat() 1623 void compact_pgdat(pg_data_t *pgdat, int order) in compact_pgdat() argument 1633 __compact_pgdat(pgdat, &cc); in compact_pgdat()
|
/linux-4.1.27/arch/avr32/mm/ |
D | init.c | 76 pg_data_t *pgdat = NODE_DATA(nid); in paging_init() local 80 start_pfn = pgdat->bdata->node_min_pfn; in paging_init() 81 low = pgdat->bdata->node_low_pfn; in paging_init() 92 pgdat->node_id, pgdat->node_mem_map); in paging_init() 103 pg_data_t *pgdat; in mem_init() local 106 for_each_online_pgdat(pgdat) in mem_init() 108 __va(pgdat_end_pfn(pgdat) << PAGE_SHIFT)); in mem_init()
|
/linux-4.1.27/arch/m68k/include/asm/ |
D | page_mm.h | 160 struct pglist_data *pgdat; \ 161 pgdat = __virt_to_node((unsigned long)pfn_to_virt(__pfn)); \ 162 pgdat->node_mem_map + (__pfn - pgdat->node_start_pfn); \ 166 struct pglist_data *pgdat; \ 167 pgdat = &pg_data_map[page_to_nid(__p)]; \ 168 ((__p) - pgdat->node_mem_map) + pgdat->node_start_pfn; \
|
/linux-4.1.27/lib/ |
D | show_mem.c | 14 pg_data_t *pgdat; in show_mem() local 20 for_each_online_pgdat(pgdat) { in show_mem() 24 pgdat_resize_lock(pgdat, &flags); in show_mem() 26 struct zone *zone = &pgdat->node_zones[zoneid]; in show_mem() 36 pgdat_resize_unlock(pgdat, &flags); in show_mem()
|
/linux-4.1.27/arch/ia64/mm/ |
D | init.c | 387 pg_data_t *pgdat = NODE_DATA(node); in vmemmap_find_next_valid_pfn() local 389 end_address = (unsigned long) &vmem_map[pgdat->node_start_pfn + i]; in vmemmap_find_next_valid_pfn() 391 stop_address = (unsigned long) &vmem_map[pgdat_end_pfn(pgdat)]; in vmemmap_find_next_valid_pfn() 434 return hole_next_pfn - pgdat->node_start_pfn; in vmemmap_find_next_valid_pfn() 657 pg_data_t *pgdat; in arch_add_memory() local 663 pgdat = NODE_DATA(nid); in arch_add_memory() 665 zone = pgdat->node_zones + in arch_add_memory() 705 pg_data_t *pgdat; in show_mem() local 710 for_each_online_pgdat(pgdat) { in show_mem() 714 int nid = pgdat->node_id; in show_mem() [all …]
|
D | discontig.c | 743 void arch_free_nodedata(pg_data_t *pgdat) in arch_free_nodedata() argument 745 kfree(pgdat); in arch_free_nodedata()
|
/linux-4.1.27/arch/sh/mm/ |
D | init.c | 382 pg_data_t *pgdat = NODE_DATA(nid); in paging_init() local 385 start_pfn = pgdat->bdata->node_min_pfn; in paging_init() 386 low = pgdat->bdata->node_low_pfn; in paging_init() 410 pg_data_t *pgdat; in mem_init() local 415 for_each_online_pgdat(pgdat) in mem_init() 417 __va(pgdat_end_pfn(pgdat) << PAGE_SHIFT)); in mem_init() 490 pg_data_t *pgdat; in arch_add_memory() local 495 pgdat = NODE_DATA(nid); in arch_add_memory() 498 ret = __add_pages(nid, pgdat->node_zones + in arch_add_memory()
|
/linux-4.1.27/arch/ia64/include/asm/ |
D | nodedata.h | 57 #define LOCAL_DATA_ADDR(pgdat) \ argument 58 ((struct ia64_node_data *)((u64)(pgdat) + \
|
/linux-4.1.27/arch/c6x/mm/ |
D | init.c | 38 struct pglist_data *pgdat = NODE_DATA(0); in paging_init() local 53 pgdat->node_zones[ZONE_NORMAL].zone_start_pfn = in paging_init()
|
/linux-4.1.27/arch/metag/mm/ |
D | init.c | 358 pg_data_t *pgdat = NODE_DATA(nid); in paging_init() local 361 start_pfn = pgdat->bdata->node_min_pfn; in paging_init() 362 low = pgdat->bdata->node_low_pfn; in paging_init()
|
/linux-4.1.27/arch/unicore32/mm/ |
D | init.c | 137 pg_data_t *pgdat; in uc32_bootmem_init() local 152 pgdat = NODE_DATA(0); in uc32_bootmem_init() 153 init_bootmem_node(pgdat, __phys_to_pfn(bitmap), start_pfn, end_pfn); in uc32_bootmem_init()
|
/linux-4.1.27/arch/parisc/mm/ |
D | init.c | 649 pg_data_t *pgdat; in show_mem() local 654 for_each_online_pgdat(pgdat) { in show_mem() 658 pgdat_resize_lock(pgdat, &flags); in show_mem() 660 struct zone *zone = &pgdat->node_zones[zoneid]; in show_mem() 667 pgdat_resize_unlock(pgdat, &flags); in show_mem()
|
/linux-4.1.27/arch/x86/mm/ |
D | init_64.c | 692 struct pglist_data *pgdat = NODE_DATA(nid); in arch_add_memory() local 693 struct zone *zone = pgdat->node_zones + in arch_add_memory()
|