gfpflags 2480 drivers/scsi/scsi_lib.c gfp_t gfpflags) gfpflags 2482 drivers/scsi/scsi_lib.c struct scsi_event *evt = kzalloc(sizeof(struct scsi_event), gfpflags); gfpflags 2517 drivers/scsi/scsi_lib.c enum scsi_device_event evt_type, gfp_t gfpflags) gfpflags 2519 drivers/scsi/scsi_lib.c struct scsi_event *evt = sdev_evt_alloc(evt_type, gfpflags); gfpflags 67 include/linux/fault-inject.h int should_failslab(struct kmem_cache *s, gfp_t gfpflags); gfpflags 69 include/linux/fault-inject.h extern bool __should_failslab(struct kmem_cache *s, gfp_t gfpflags); gfpflags 71 include/linux/fault-inject.h static inline bool __should_failslab(struct kmem_cache *s, gfp_t gfpflags) gfpflags 436 include/linux/slab.h gfp_t gfpflags, gfpflags 441 include/linux/slab.h gfp_t gfpflags, gfpflags 444 include/linux/slab.h return kmem_cache_alloc_trace(s, gfpflags, size); gfpflags 460 include/linux/slab.h gfp_t gfpflags, gfpflags 463 include/linux/slab.h void *ret = kmem_cache_alloc_node(s, gfpflags, node); gfpflags 465 include/linux/slab.h ret = kasan_kmalloc(s, ret, size, gfpflags); gfpflags 412 include/scsi/scsi_device.h gfp_t gfpflags); gfpflags 415 include/scsi/scsi_device.h enum scsi_device_event evt_type, gfp_t gfpflags); gfpflags 1204 kernel/trace/trace_functions_graph.c gfp_t gfpflags; gfpflags 1210 kernel/trace/trace_functions_graph.c gfpflags = (in_atomic() || irqs_disabled()) ? GFP_ATOMIC : GFP_KERNEL; gfpflags 1212 kernel/trace/trace_functions_graph.c data = kzalloc(sizeof(*data), gfpflags); gfpflags 1216 kernel/trace/trace_functions_graph.c data->cpu_data = alloc_percpu_gfp(struct fgraph_cpu_data, gfpflags); gfpflags 17 mm/failslab.c bool __should_failslab(struct kmem_cache *s, gfp_t gfpflags) gfpflags 23 mm/failslab.c if (gfpflags & __GFP_NOFAIL) gfpflags 27 mm/failslab.c (gfpflags & __GFP_DIRECT_RECLAIM)) gfpflags 1317 mm/slab.c slab_out_of_memory(struct kmem_cache *cachep, gfp_t gfpflags, int nodeid) gfpflags 1326 mm/slab.c if ((gfpflags & __GFP_NOWARN) || !__ratelimit(&slab_oom_rs)) gfpflags 1330 mm/slab.c nodeid, gfpflags, &gfpflags); gfpflags 1802 mm/slab_common.c int should_failslab(struct kmem_cache *s, gfp_t gfpflags) gfpflags 1804 mm/slab_common.c if (__should_failslab(s, gfpflags)) gfpflags 1830 mm/slub.c static inline bool pfmemalloc_match(struct page *page, gfp_t gfpflags); gfpflags 2405 mm/slub.c slab_out_of_memory(struct kmem_cache *s, gfp_t gfpflags, int nid) gfpflags 2413 mm/slub.c if ((gfpflags & __GFP_NOWARN) || !__ratelimit(&slub_oom_rs)) gfpflags 2417 mm/slub.c nid, gfpflags, &gfpflags); gfpflags 2476 mm/slub.c static inline bool pfmemalloc_match(struct page *page, gfp_t gfpflags) gfpflags 2479 mm/slub.c return gfp_pfmemalloc_allowed(gfpflags); gfpflags 2537 mm/slub.c static void *___slab_alloc(struct kmem_cache *s, gfp_t gfpflags, int node, gfpflags 2576 mm/slub.c if (unlikely(!pfmemalloc_match(page, gfpflags))) { gfpflags 2616 mm/slub.c freelist = new_slab_objects(s, gfpflags, node, &c); gfpflags 2619 mm/slub.c slab_out_of_memory(s, gfpflags, node); gfpflags 2624 mm/slub.c if (likely(!kmem_cache_debug(s) && pfmemalloc_match(page, gfpflags))) gfpflags 2640 mm/slub.c static void *__slab_alloc(struct kmem_cache *s, gfp_t gfpflags, int node, gfpflags 2656 mm/slub.c p = ___slab_alloc(s, gfpflags, node, addr, c); gfpflags 2683 mm/slub.c gfp_t gfpflags, int node, unsigned long addr) gfpflags 2690 mm/slub.c s = slab_pre_alloc_hook(s, gfpflags); gfpflags 2730 mm/slub.c object = __slab_alloc(s, gfpflags, node, addr, c); gfpflags 2763 mm/slub.c if (unlikely(slab_want_init_on_alloc(gfpflags, s)) && object) gfpflags 2766 mm/slub.c slab_post_alloc_hook(s, gfpflags, 1, &object); gfpflags 2772 mm/slub.c gfp_t gfpflags, unsigned long addr) gfpflags 2774 mm/slub.c return slab_alloc_node(s, gfpflags, NUMA_NO_NODE, addr); gfpflags 2777 mm/slub.c void *kmem_cache_alloc(struct kmem_cache *s, gfp_t gfpflags) gfpflags 2779 mm/slub.c void *ret = slab_alloc(s, gfpflags, _RET_IP_); gfpflags 2782 mm/slub.c s->size, gfpflags); gfpflags 2789 mm/slub.c void *kmem_cache_alloc_trace(struct kmem_cache *s, gfp_t gfpflags, size_t size) gfpflags 2791 mm/slub.c void *ret = slab_alloc(s, gfpflags, _RET_IP_); gfpflags 2792 mm/slub.c trace_kmalloc(_RET_IP_, ret, size, s->size, gfpflags); gfpflags 2793 mm/slub.c ret = kasan_kmalloc(s, ret, size, gfpflags); gfpflags 2800 mm/slub.c void *kmem_cache_alloc_node(struct kmem_cache *s, gfp_t gfpflags, int node) gfpflags 2802 mm/slub.c void *ret = slab_alloc_node(s, gfpflags, node, _RET_IP_); gfpflags 2805 mm/slub.c s->object_size, s->size, gfpflags, node); gfpflags 2813 mm/slub.c gfp_t gfpflags, gfpflags 2816 mm/slub.c void *ret = slab_alloc_node(s, gfpflags, node, _RET_IP_); gfpflags 2819 mm/slub.c size, s->size, gfpflags, node); gfpflags 2821 mm/slub.c ret = kasan_kmalloc(s, ret, size, gfpflags); gfpflags 4340 mm/slub.c void *__kmalloc_track_caller(size_t size, gfp_t gfpflags, unsigned long caller) gfpflags 4346 mm/slub.c return kmalloc_large(size, gfpflags); gfpflags 4348 mm/slub.c s = kmalloc_slab(size, gfpflags); gfpflags 4353 mm/slub.c ret = slab_alloc(s, gfpflags, caller); gfpflags 4356 mm/slub.c trace_kmalloc(caller, ret, size, s->size, gfpflags); gfpflags 4362 mm/slub.c void *__kmalloc_node_track_caller(size_t size, gfp_t gfpflags, gfpflags 4369 mm/slub.c ret = kmalloc_large_node(size, gfpflags, node); gfpflags 4373 mm/slub.c gfpflags, node); gfpflags 4378 mm/slub.c s = kmalloc_slab(size, gfpflags); gfpflags 4383 mm/slub.c ret = slab_alloc_node(s, gfpflags, node, caller); gfpflags 4386 mm/slub.c trace_kmalloc_node(caller, ret, size, s->size, gfpflags, node);