safe_needed 156 kernel/power/snapshot.c static void *get_image_page(gfp_t gfp_mask, int safe_needed) safe_needed 161 kernel/power/snapshot.c if (safe_needed) safe_needed 263 kernel/power/snapshot.c int safe_needed; /* if set, only "safe" pages are allocated */ safe_needed 267 kernel/power/snapshot.c int safe_needed) safe_needed 272 kernel/power/snapshot.c ca->safe_needed = safe_needed; safe_needed 282 kernel/power/snapshot.c lp = ca->safe_needed ? __get_safe_page(ca->gfp_mask) : safe_needed 402 kernel/power/snapshot.c static struct rtree_node *alloc_rtree_node(gfp_t gfp_mask, int safe_needed, safe_needed 412 kernel/power/snapshot.c node->data = get_image_page(gfp_mask, safe_needed); safe_needed 429 kernel/power/snapshot.c int safe_needed, struct chain_allocator *ca) safe_needed 446 kernel/power/snapshot.c node = alloc_rtree_node(gfp_mask, safe_needed, ca, safe_needed 457 kernel/power/snapshot.c block = alloc_rtree_node(gfp_mask, safe_needed, ca, &zone->leaves); safe_needed 469 kernel/power/snapshot.c node = alloc_rtree_node(gfp_mask, safe_needed, ca, safe_needed 499 kernel/power/snapshot.c int safe_needed, safe_needed 520 kernel/power/snapshot.c if (add_rtree_block(zone, gfp_mask, safe_needed, ca)) { safe_needed 644 kernel/power/snapshot.c int safe_needed) safe_needed 651 kernel/power/snapshot.c chain_init(&ca, gfp_mask, safe_needed); safe_needed 661 kernel/power/snapshot.c zone = create_zone_bm_rtree(gfp_mask, safe_needed, &ca, safe_needed 1898 kernel/power/snapshot.c static inline int get_highmem_buffer(int safe_needed) safe_needed 1900 kernel/power/snapshot.c buffer = get_image_page(GFP_ATOMIC, safe_needed); safe_needed 1928 kernel/power/snapshot.c static inline int get_highmem_buffer(int safe_needed) { return 0; }