safe_needed       156 kernel/power/snapshot.c static void *get_image_page(gfp_t gfp_mask, int safe_needed)
safe_needed       161 kernel/power/snapshot.c 	if (safe_needed)
safe_needed       263 kernel/power/snapshot.c 	int safe_needed;	/* if set, only "safe" pages are allocated */
safe_needed       267 kernel/power/snapshot.c 		       int safe_needed)
safe_needed       272 kernel/power/snapshot.c 	ca->safe_needed = safe_needed;
safe_needed       282 kernel/power/snapshot.c 		lp = ca->safe_needed ? __get_safe_page(ca->gfp_mask) :
safe_needed       402 kernel/power/snapshot.c static struct rtree_node *alloc_rtree_node(gfp_t gfp_mask, int safe_needed,
safe_needed       412 kernel/power/snapshot.c 	node->data = get_image_page(gfp_mask, safe_needed);
safe_needed       429 kernel/power/snapshot.c 			   int safe_needed, struct chain_allocator *ca)
safe_needed       446 kernel/power/snapshot.c 		node = alloc_rtree_node(gfp_mask, safe_needed, ca,
safe_needed       457 kernel/power/snapshot.c 	block = alloc_rtree_node(gfp_mask, safe_needed, ca, &zone->leaves);
safe_needed       469 kernel/power/snapshot.c 			node = alloc_rtree_node(gfp_mask, safe_needed, ca,
safe_needed       499 kernel/power/snapshot.c 						      int safe_needed,
safe_needed       520 kernel/power/snapshot.c 		if (add_rtree_block(zone, gfp_mask, safe_needed, ca)) {
safe_needed       644 kernel/power/snapshot.c 			    int safe_needed)
safe_needed       651 kernel/power/snapshot.c 	chain_init(&ca, gfp_mask, safe_needed);
safe_needed       661 kernel/power/snapshot.c 		zone = create_zone_bm_rtree(gfp_mask, safe_needed, &ca,
safe_needed      1898 kernel/power/snapshot.c static inline int get_highmem_buffer(int safe_needed)
safe_needed      1900 kernel/power/snapshot.c 	buffer = get_image_page(GFP_ATOMIC, safe_needed);
safe_needed      1928 kernel/power/snapshot.c static inline int get_highmem_buffer(int safe_needed) { return 0; }