Lines Matching refs:zone

94 int calculate_pressure_threshold(struct zone *zone)  in calculate_pressure_threshold()  argument
107 watermark_distance = low_wmark_pages(zone) - min_wmark_pages(zone); in calculate_pressure_threshold()
118 int calculate_normal_threshold(struct zone *zone) in calculate_normal_threshold() argument
153 mem = zone->managed_pages >> (27 - PAGE_SHIFT); in calculate_normal_threshold()
170 struct zone *zone; in refresh_zone_stat_thresholds() local
174 for_each_populated_zone(zone) { in refresh_zone_stat_thresholds()
177 threshold = calculate_normal_threshold(zone); in refresh_zone_stat_thresholds()
180 per_cpu_ptr(zone->pageset, cpu)->stat_threshold in refresh_zone_stat_thresholds()
188 tolerate_drift = low_wmark_pages(zone) - min_wmark_pages(zone); in refresh_zone_stat_thresholds()
191 zone->percpu_drift_mark = high_wmark_pages(zone) + in refresh_zone_stat_thresholds()
197 int (*calculate_pressure)(struct zone *)) in set_pgdat_percpu_threshold() argument
199 struct zone *zone; in set_pgdat_percpu_threshold() local
205 zone = &pgdat->node_zones[i]; in set_pgdat_percpu_threshold()
206 if (!zone->percpu_drift_mark) in set_pgdat_percpu_threshold()
209 threshold = (*calculate_pressure)(zone); in set_pgdat_percpu_threshold()
211 per_cpu_ptr(zone->pageset, cpu)->stat_threshold in set_pgdat_percpu_threshold()
221 void __mod_zone_page_state(struct zone *zone, enum zone_stat_item item, in __mod_zone_page_state() argument
224 struct per_cpu_pageset __percpu *pcp = zone->pageset; in __mod_zone_page_state()
234 zone_page_state_add(x, zone, item); in __mod_zone_page_state()
264 void __inc_zone_state(struct zone *zone, enum zone_stat_item item) in __inc_zone_state() argument
266 struct per_cpu_pageset __percpu *pcp = zone->pageset; in __inc_zone_state()
275 zone_page_state_add(v + overstep, zone, item); in __inc_zone_state()
286 void __dec_zone_state(struct zone *zone, enum zone_stat_item item) in __dec_zone_state() argument
288 struct per_cpu_pageset __percpu *pcp = zone->pageset; in __dec_zone_state()
297 zone_page_state_add(v - overstep, zone, item); in __dec_zone_state()
321 static inline void mod_state(struct zone *zone, enum zone_stat_item item, in mod_state() argument
324 struct per_cpu_pageset __percpu *pcp = zone->pageset; in mod_state()
356 zone_page_state_add(z, zone, item); in mod_state()
359 void mod_zone_page_state(struct zone *zone, enum zone_stat_item item, in mod_zone_page_state() argument
362 mod_state(zone, item, delta, 0); in mod_zone_page_state()
366 void inc_zone_state(struct zone *zone, enum zone_stat_item item) in inc_zone_state() argument
368 mod_state(zone, item, 1, 1); in inc_zone_state()
386 void mod_zone_page_state(struct zone *zone, enum zone_stat_item item, in mod_zone_page_state() argument
392 __mod_zone_page_state(zone, item, delta); in mod_zone_page_state()
397 void inc_zone_state(struct zone *zone, enum zone_stat_item item) in inc_zone_state() argument
402 __inc_zone_state(zone, item); in inc_zone_state()
409 struct zone *zone; in inc_zone_page_state() local
411 zone = page_zone(page); in inc_zone_page_state()
413 __inc_zone_state(zone, item); in inc_zone_page_state()
465 struct zone *zone; in refresh_cpu_vm_stats() local
470 for_each_populated_zone(zone) { in refresh_cpu_vm_stats()
471 struct per_cpu_pageset __percpu *p = zone->pageset; in refresh_cpu_vm_stats()
479 atomic_long_add(v, &zone->vm_stat[i]); in refresh_cpu_vm_stats()
503 if (zone_to_nid(zone) == numa_node_id()) { in refresh_cpu_vm_stats()
512 drain_zone_pages(zone, this_cpu_ptr(&p->pcp)); in refresh_cpu_vm_stats()
528 struct zone *zone; in cpu_vm_stats_fold() local
532 for_each_populated_zone(zone) { in cpu_vm_stats_fold()
535 p = per_cpu_ptr(zone->pageset, cpu); in cpu_vm_stats_fold()
543 atomic_long_add(v, &zone->vm_stat[i]); in cpu_vm_stats_fold()
555 void drain_zonestat(struct zone *zone, struct per_cpu_pageset *pset) in drain_zonestat() argument
563 atomic_long_add(v, &zone->vm_stat[i]); in drain_zonestat()
580 void zone_statistics(struct zone *preferred_zone, struct zone *z, gfp_t flags) in zone_statistics()
600 struct zone *zones = NODE_DATA(node)->node_zones; in node_page_state()
634 static void fill_contig_page_info(struct zone *zone, in fill_contig_page_info() argument
648 blocks = zone->free_area[order].nr_free; in fill_contig_page_info()
689 int fragmentation_index(struct zone *zone, unsigned int order) in fragmentation_index() argument
693 fill_contig_page_info(zone, order, &info); in fragmentation_index()
904 void (*print)(struct seq_file *m, pg_data_t *, struct zone *)) in walk_zones_in_node() argument
906 struct zone *zone; in walk_zones_in_node() local
907 struct zone *node_zones = pgdat->node_zones; in walk_zones_in_node()
910 for (zone = node_zones; zone - node_zones < MAX_NR_ZONES; ++zone) { in walk_zones_in_node()
911 if (!populated_zone(zone)) in walk_zones_in_node()
914 spin_lock_irqsave(&zone->lock, flags); in walk_zones_in_node()
915 print(m, pgdat, zone); in walk_zones_in_node()
916 spin_unlock_irqrestore(&zone->lock, flags); in walk_zones_in_node()
936 struct zone *zone) in frag_show_print() argument
940 seq_printf(m, "Node %d, zone %8s ", pgdat->node_id, zone->name); in frag_show_print()
942 seq_printf(m, "%6lu ", zone->free_area[order].nr_free); in frag_show_print()
957 pg_data_t *pgdat, struct zone *zone) in pagetypeinfo_showfree_print() argument
964 zone->name, in pagetypeinfo_showfree_print()
971 area = &(zone->free_area[order]); in pagetypeinfo_showfree_print()
999 pg_data_t *pgdat, struct zone *zone) in pagetypeinfo_showblockcount_print() argument
1003 unsigned long start_pfn = zone->zone_start_pfn; in pagetypeinfo_showblockcount_print()
1004 unsigned long end_pfn = zone_end_pfn(zone); in pagetypeinfo_showblockcount_print()
1016 if (!memmap_valid_within(pfn, page, zone)) in pagetypeinfo_showblockcount_print()
1026 seq_printf(m, "Node %d, zone %8s ", pgdat->node_id, zone->name); in pagetypeinfo_showblockcount_print()
1050 struct zone *zone) in pagetypeinfo_showmixedcount_print() argument
1054 unsigned long pfn = zone->zone_start_pfn, block_end_pfn; in pagetypeinfo_showmixedcount_print()
1055 unsigned long end_pfn = pfn + zone->spanned_pages; in pagetypeinfo_showmixedcount_print()
1061 pfn = zone->zone_start_pfn; in pagetypeinfo_showmixedcount_print()
1113 seq_printf(m, "Node %d, zone %8s ", pgdat->node_id, zone->name); in pagetypeinfo_showmixedcount_print()
1206 struct zone *zone) in zoneinfo_show_print() argument
1209 seq_printf(m, "Node %d, zone %8s", pgdat->node_id, zone->name); in zoneinfo_show_print()
1219 zone_page_state(zone, NR_FREE_PAGES), in zoneinfo_show_print()
1220 min_wmark_pages(zone), in zoneinfo_show_print()
1221 low_wmark_pages(zone), in zoneinfo_show_print()
1222 high_wmark_pages(zone), in zoneinfo_show_print()
1223 zone_page_state(zone, NR_PAGES_SCANNED), in zoneinfo_show_print()
1224 zone->spanned_pages, in zoneinfo_show_print()
1225 zone->present_pages, in zoneinfo_show_print()
1226 zone->managed_pages); in zoneinfo_show_print()
1230 zone_page_state(zone, i)); in zoneinfo_show_print()
1234 zone->lowmem_reserve[0]); in zoneinfo_show_print()
1235 for (i = 1; i < ARRAY_SIZE(zone->lowmem_reserve); i++) in zoneinfo_show_print()
1236 seq_printf(m, ", %ld", zone->lowmem_reserve[i]); in zoneinfo_show_print()
1243 pageset = per_cpu_ptr(zone->pageset, i); in zoneinfo_show_print()
1262 !zone_reclaimable(zone), in zoneinfo_show_print()
1263 zone->zone_start_pfn, in zoneinfo_show_print()
1264 zone->inactive_ratio); in zoneinfo_show_print()
1426 struct zone *zone; in need_update() local
1428 for_each_populated_zone(zone) { in need_update()
1429 struct per_cpu_pageset *p = per_cpu_ptr(zone->pageset, cpu); in need_update()
1590 pg_data_t *pgdat, struct zone *zone) in unusable_show_print() argument
1598 zone->name); in unusable_show_print()
1600 fill_contig_page_info(zone, order, &info); in unusable_show_print()
1650 pg_data_t *pgdat, struct zone *zone) in extfrag_show_print() argument
1660 zone->name); in extfrag_show_print()
1662 fill_contig_page_info(zone, order, &info); in extfrag_show_print()