zone_managed_pages 868 drivers/gpu/drm/amd/amdkfd/kfd_crat.c mem_in_bytes += zone_managed_pages(&pgdat->node_zones[zone_type]); zone_managed_pages 874 include/linux/mmzone.h return zone_managed_pages(zone); zone_managed_pages 28 lib/show_mem.c reserved += zone->present_pages - zone_managed_pages(zone); zone_managed_pages 2360 mm/page_alloc.c if ((pageblock_nr_pages * 4) > zone_managed_pages(zone)) zone_managed_pages 2521 mm/page_alloc.c max_managed = (zone_managed_pages(zone) / 100) + pageblock_nr_pages; zone_managed_pages 5055 mm/page_alloc.c unsigned long size = zone_managed_pages(zone); zone_managed_pages 5167 mm/page_alloc.c managed_pages += zone_managed_pages(&pgdat->node_zones[zone_type]); zone_managed_pages 5176 mm/page_alloc.c managed_highpages += zone_managed_pages(zone); zone_managed_pages 5383 mm/page_alloc.c K(zone_managed_pages(zone)), zone_managed_pages 6043 mm/page_alloc.c batch = zone_managed_pages(zone) / 1024; zone_managed_pages 6153 mm/page_alloc.c (zone_managed_pages(zone) / zone_managed_pages 7681 mm/page_alloc.c unsigned long managed_pages = zone_managed_pages(zone); zone_managed_pages 7717 mm/page_alloc.c unsigned long managed_pages = zone_managed_pages(zone); zone_managed_pages 7735 mm/page_alloc.c managed_pages += zone_managed_pages(lower_zone); zone_managed_pages 7754 mm/page_alloc.c lowmem_pages += zone_managed_pages(zone); zone_managed_pages 7761 mm/page_alloc.c tmp = (u64)pages_min * zone_managed_pages(zone); zone_managed_pages 7775 mm/page_alloc.c min_pages = zone_managed_pages(zone) / 1024; zone_managed_pages 7792 mm/page_alloc.c mult_frac(zone_managed_pages(zone), zone_managed_pages 7935 mm/page_alloc.c zone->zone_pgdat->min_unmapped_pages += (zone_managed_pages(zone) * zone_managed_pages 7963 mm/page_alloc.c zone->zone_pgdat->min_slab_pages += (zone_managed_pages(zone) * zone_managed_pages 231 mm/vmstat.c mem = zone_managed_pages(zone) >> (27 - PAGE_SHIFT); zone_managed_pages 1587 mm/vmstat.c zone_managed_pages(zone));