Home
last modified time | relevance | path

Searched refs:gfp (Results 1 – 200 of 389) sorted by relevance

12

/linux-4.4.14/lib/
Dkasprintf.c14 char *kvasprintf(gfp_t gfp, const char *fmt, va_list ap) in kvasprintf() argument
24 p = kmalloc_track_caller(len+1, gfp); in kvasprintf()
40 const char *kvasprintf_const(gfp_t gfp, const char *fmt, va_list ap) in kvasprintf_const() argument
43 return kstrdup_const(fmt, gfp); in kvasprintf_const()
45 return kstrdup_const(va_arg(ap, const char*), gfp); in kvasprintf_const()
46 return kvasprintf(gfp, fmt, ap); in kvasprintf_const()
50 char *kasprintf(gfp_t gfp, const char *fmt, ...) in kasprintf() argument
56 p = kvasprintf(gfp, fmt, ap); in kasprintf()
Dargv_split.c59 char **argv_split(gfp_t gfp, const char *str, int *argcp) in argv_split() argument
66 argv_str = kstrndup(str, KMALLOC_MAX_SIZE - 1, gfp); in argv_split()
71 argv = kmalloc(sizeof(*argv) * (argc + 2), gfp); in argv_split()
Dflex_proportions.c37 int fprop_global_init(struct fprop_global *p, gfp_t gfp) in fprop_global_init() argument
43 err = percpu_counter_init(&p->events, 1, gfp); in fprop_global_init()
171 int fprop_local_init_percpu(struct fprop_local_percpu *pl, gfp_t gfp) in fprop_local_init_percpu() argument
175 err = percpu_counter_init(&pl->events, 0, gfp); in fprop_local_init_percpu()
Dproportions.c76 int prop_descriptor_init(struct prop_descriptor *pd, int shift, gfp_t gfp) in prop_descriptor_init() argument
86 err = percpu_counter_init(&pd->pg[0].events, 0, gfp); in prop_descriptor_init()
90 err = percpu_counter_init(&pd->pg[1].events, 0, gfp); in prop_descriptor_init()
191 int prop_local_init_percpu(struct prop_local_percpu *pl, gfp_t gfp) in prop_local_init_percpu() argument
196 return percpu_counter_init(&pl->events, 0, gfp); in prop_local_init_percpu()
Dbtree.c93 static unsigned long *btree_node_alloc(struct btree_head *head, gfp_t gfp) in btree_node_alloc() argument
97 node = mempool_alloc(head->mempool, gfp); in btree_node_alloc()
414 gfp_t gfp) in btree_grow() argument
419 node = btree_node_alloc(head, gfp); in btree_grow()
450 gfp_t gfp) in btree_insert_level() argument
457 err = btree_grow(head, geo, gfp); in btree_insert_level()
473 new = btree_node_alloc(head, gfp); in btree_insert_level()
478 new, level + 1, gfp); in btree_insert_level()
511 unsigned long *key, void *val, gfp_t gfp) in btree_insert() argument
514 return btree_insert_level(head, geo, key, val, 1, gfp); in btree_insert()
[all …]
Drhashtable.c64 gfp_t gfp) in alloc_bucket_locks() argument
82 gfp == GFP_KERNEL) in alloc_bucket_locks()
87 gfp); in alloc_bucket_locks()
113 gfp_t gfp) in bucket_table_alloc() argument
121 gfp != GFP_KERNEL) in bucket_table_alloc()
122 tbl = kzalloc(size, gfp | __GFP_NOWARN | __GFP_NORETRY); in bucket_table_alloc()
123 if (tbl == NULL && gfp == GFP_KERNEL) in bucket_table_alloc()
130 if (alloc_bucket_locks(ht, tbl, gfp) < 0) { in bucket_table_alloc()
Dpercpu-refcount.c59 unsigned int flags, gfp_t gfp) in percpu_ref_init() argument
66 __alloc_percpu_gfp(sizeof(unsigned long), align, gfp); in percpu_ref_init()
Dpercpu_counter.c115 int __percpu_counter_init(struct percpu_counter *fbc, s64 amount, gfp_t gfp, in __percpu_counter_init() argument
123 fbc->counters = alloc_percpu_gfp(s32, gfp); in __percpu_counter_init()
/linux-4.4.14/net/wireless/
Dnl80211.h37 const u8 *buf, size_t len, gfp_t gfp);
40 const u8 *buf, size_t len, gfp_t gfp,
44 const u8 *buf, size_t len, gfp_t gfp);
47 const u8 *buf, size_t len, gfp_t gfp);
50 const u8 *addr, gfp_t gfp);
53 const u8 *addr, gfp_t gfp);
58 u16 status, gfp_t gfp);
62 const u8 *resp_ie, size_t resp_ie_len, gfp_t gfp);
71 int key_id, const u8 *tsc, gfp_t gfp);
80 gfp_t gfp);
[all …]
Dnl80211.c7585 gfp_t gfp) in __cfg80211_alloc_vendor_skb() argument
7591 skb = nlmsg_new(approxlen + 100, gfp); in __cfg80211_alloc_vendor_skb()
7641 int approxlen, gfp_t gfp) in __cfg80211_alloc_event_skb() argument
7664 cmd, attr, info, gfp); in __cfg80211_alloc_event_skb()
7668 void __cfg80211_send_event_skb(struct sk_buff *skb, gfp_t gfp) in __cfg80211_send_event_skb() argument
7685 mcgrp, gfp); in __cfg80211_send_event_skb()
11713 enum nl80211_commands cmd, gfp_t gfp, in nl80211_send_mlme_event() argument
11719 msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp); in nl80211_send_mlme_event()
11750 NL80211_MCGRP_MLME, gfp); in nl80211_send_mlme_event()
11760 size_t len, gfp_t gfp) in nl80211_send_rx_auth() argument
[all …]
Dmlme.c173 const u8 *tsc, gfp_t gfp) in cfg80211_michael_mic_failure() argument
179 char *buf = kmalloc(128, gfp); in cfg80211_michael_mic_failure()
194 nl80211_michael_mic_failure(rdev, dev, addr, key_type, key_id, tsc, gfp); in cfg80211_michael_mic_failure()
775 gfp_t gfp) in cfg80211_radar_event() argument
792 nl80211_radar_notify(rdev, chandef, NL80211_RADAR_DETECTED, NULL, gfp); in cfg80211_radar_event()
798 enum nl80211_radar_event event, gfp_t gfp) in cfg80211_cac_event() argument
828 nl80211_radar_notify(rdev, chandef, event, netdev, gfp); in cfg80211_cac_event()
Dsme.c747 u16 status, gfp_t gfp) in cfg80211_connect_result() argument
754 ev = kzalloc(sizeof(*ev) + req_ie_len + resp_ie_len, gfp); in cfg80211_connect_result()
842 const u8 *resp_ie, size_t resp_ie_len, gfp_t gfp) in cfg80211_roamed() argument
854 resp_ie_len, gfp); in cfg80211_roamed()
862 size_t resp_ie_len, gfp_t gfp) in cfg80211_roamed_bss() argument
872 ev = kzalloc(sizeof(*ev) + req_ie_len + resp_ie_len, gfp); in cfg80211_roamed_bss()
942 bool locally_generated, gfp_t gfp) in cfg80211_disconnected() argument
949 ev = kzalloc(sizeof(*ev) + ie_len, gfp); in cfg80211_disconnected()
Dreg.h83 gfp_t gfp);
/linux-4.4.14/include/linux/
Dkmemleak.h30 gfp_t gfp) __ref;
32 gfp_t gfp) __ref;
39 extern void kmemleak_scan_area(const void *ptr, size_t size, gfp_t gfp) __ref;
44 gfp_t gfp) in kmemleak_alloc_recursive() argument
47 kmemleak_alloc(ptr, size, min_count, gfp); in kmemleak_alloc_recursive()
67 gfp_t gfp) in kmemleak_alloc() argument
72 gfp_t gfp) in kmemleak_alloc_recursive() argument
76 gfp_t gfp) in kmemleak_alloc_percpu() argument
100 static inline void kmemleak_scan_area(const void *ptr, size_t size, gfp_t gfp) in kmemleak_scan_area() argument
Dvmpressure.h29 extern void vmpressure(gfp_t gfp, struct mem_cgroup *memcg,
31 extern void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, int prio);
43 static inline void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, in vmpressure() argument
45 static inline void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, in vmpressure_prio() argument
Ddevcoredump.h10 gfp_t gfp);
13 const void *data, size_t datalen, gfp_t gfp,
19 size_t datalen, gfp_t gfp) in dev_coredumpv() argument
26 const void *data, size_t datalen, gfp_t gfp, in dev_coredumpm() argument
Dfscache.h546 gfp_t gfp) in fscache_read_or_alloc_page() argument
550 context, gfp); in fscache_read_or_alloc_page()
597 gfp_t gfp) in fscache_read_or_alloc_pages() argument
602 context, gfp); in fscache_read_or_alloc_pages()
628 gfp_t gfp) in fscache_alloc_page() argument
631 return __fscache_alloc_page(cookie, page, gfp); in fscache_alloc_page()
677 gfp_t gfp) in fscache_write_page() argument
680 return __fscache_write_page(cookie, page, gfp); in fscache_write_page()
763 gfp_t gfp) in fscache_maybe_release_page() argument
766 return __fscache_maybe_release_page(cookie, page, gfp); in fscache_maybe_release_page()
Dzpool.h42 gfp_t gfp, const struct zpool_ops *ops);
48 int zpool_malloc(struct zpool *pool, size_t size, gfp_t gfp,
87 gfp_t gfp,
92 int (*malloc)(void *pool, size_t size, gfp_t gfp,
Dstring.h121 extern char *kstrdup(const char *s, gfp_t gfp);
122 extern const char *kstrdup_const(const char *s, gfp_t gfp);
123 extern char *kstrndup(const char *s, size_t len, gfp_t gfp);
124 extern void *kmemdup(const void *src, size_t len, gfp_t gfp);
126 extern char **argv_split(gfp_t gfp, const char *str, int *argcp);
Dbtree-type.h31 gfp_t gfp) in BTREE_FN()
33 return btree_merge(&target->h, &victim->h, BTREE_TYPE_GEO, gfp); in BTREE_FN()
44 void *val, gfp_t gfp) in BTREE_FN()
47 return btree_insert(&head->h, BTREE_TYPE_GEO, &_key, val, gfp); in BTREE_FN()
87 void *val, gfp_t gfp) in BTREE_FN()
90 val, gfp); in BTREE_FN()
Dmemcontrol.h754 int __memcg_kmem_charge_memcg(struct page *page, gfp_t gfp, int order,
756 int __memcg_kmem_charge(struct page *page, gfp_t gfp, int order);
772 static inline bool __memcg_kmem_bypass(gfp_t gfp) in __memcg_kmem_bypass() argument
776 if (gfp & __GFP_NOACCOUNT) in __memcg_kmem_bypass()
792 gfp_t gfp, int order) in memcg_kmem_charge() argument
794 if (__memcg_kmem_bypass(gfp)) in memcg_kmem_charge()
796 return __memcg_kmem_charge(page, gfp, order); in memcg_kmem_charge()
818 memcg_kmem_get_cache(struct kmem_cache *cachep, gfp_t gfp) in memcg_kmem_get_cache() argument
820 if (__memcg_kmem_bypass(gfp)) in memcg_kmem_get_cache()
844 static inline int memcg_kmem_charge(struct page *page, gfp_t gfp, int order) in memcg_kmem_charge() argument
[all …]
Dbtree-128.h41 void *val, gfp_t gfp) in btree_insert128() argument
45 (unsigned long *)&key, val, gfp); in btree_insert128()
78 gfp_t gfp) in btree_merge128() argument
80 return btree_merge(&target->h, &victim->h, &btree_geo128, gfp); in btree_merge128()
Dpercpu_counter.h30 int __percpu_counter_init(struct percpu_counter *fbc, s64 amount, gfp_t gfp,
33 #define percpu_counter_init(fbc, value, gfp) \ argument
37 __percpu_counter_init(fbc, value, gfp, &__key); \
99 gfp_t gfp) in percpu_counter_init() argument
Dpercpu.h126 extern void __percpu *__alloc_percpu_gfp(size_t size, size_t align, gfp_t gfp);
131 #define alloc_percpu_gfp(type, gfp) \ argument
133 __alignof__(type), gfp)
Dzbud.h12 struct zbud_pool *zbud_create_pool(gfp_t gfp, const struct zbud_ops *ops);
14 int zbud_alloc(struct zbud_pool *pool, size_t size, gfp_t gfp,
Dvirtio.h40 gfp_t gfp);
45 gfp_t gfp);
52 gfp_t gfp);
Dbacking-dev.h245 wb_congested_get_create(struct backing_dev_info *bdi, int blkcg_id, gfp_t gfp);
249 gfp_t gfp);
314 wb_get_create_current(struct backing_dev_info *bdi, gfp_t gfp) in wb_get_create_current() argument
328 wb = wb_get_create(bdi, memcg_css, gfp); in wb_get_create_current()
422 wb_congested_get_create(struct backing_dev_info *bdi, int blkcg_id, gfp_t gfp) in wb_congested_get_create() argument
440 wb_get_create_current(struct backing_dev_info *bdi, gfp_t gfp) in wb_get_create_current() argument
Ddma-mapping.h20 dma_addr_t *dma_handle, gfp_t gfp,
236 dma_addr_t *dma_handle, gfp_t gfp);
240 dma_addr_t *dma_handle, gfp_t gfp);
252 size_t size, gfp_t gfp) in dmam_declare_coherent_memory() argument
279 dma_addr_t *dma_addr, gfp_t gfp) in dma_alloc_writecombine() argument
283 return dma_alloc_attrs(dev, size, dma_addr, gfp, &attrs); in dma_alloc_writecombine()
Ddevice.h607 extern void *__devres_alloc_node(dr_release_t release, size_t size, gfp_t gfp,
609 #define devres_alloc(release, size, gfp) \ argument
610 __devres_alloc_node(release, size, gfp, NUMA_NO_NODE, #release)
611 #define devres_alloc_node(release, size, gfp, nid) \ argument
612 __devres_alloc_node(release, size, gfp, nid, #release)
614 extern void *devres_alloc_node(dr_release_t release, size_t size, gfp_t gfp,
616 static inline void *devres_alloc(dr_release_t release, size_t size, gfp_t gfp) in devres_alloc() argument
618 return devres_alloc_node(release, size, gfp, NUMA_NO_NODE); in devres_alloc()
641 gfp_t gfp);
647 extern void *devm_kmalloc(struct device *dev, size_t size, gfp_t gfp);
[all …]
Dflex_array.h76 #define flex_array_put_ptr(fa, nr, src, gfp) \ argument
77 flex_array_put(fa, nr, (void *)&(src), gfp)
Dflex_proportions.h36 int fprop_global_init(struct fprop_global *p, gfp_t gfp);
83 int fprop_local_init_percpu(struct fprop_local_percpu *pl, gfp_t gfp);
Dfirmware.h46 const char *name, struct device *device, gfp_t gfp, void *context,
61 const char *name, struct device *device, gfp_t gfp, void *context, in request_firmware_nowait() argument
Dproportions.h44 int prop_descriptor_init(struct prop_descriptor *pd, int shift, gfp_t gfp);
65 int prop_local_init_percpu(struct prop_local_percpu *pl, gfp_t gfp);
Dbtree.h115 unsigned long *key, void *val, gfp_t gfp);
158 struct btree_geo *geo, gfp_t gfp);
Dblk-cgroup.h138 typedef struct blkcg_policy_data *(blkcg_pol_alloc_cpd_fn)(gfp_t gfp);
142 typedef struct blkg_policy_data *(blkcg_pol_alloc_pd_fn)(gfp_t gfp, int node);
503 static inline int blkg_stat_init(struct blkg_stat *stat, gfp_t gfp) in blkg_stat_init() argument
507 ret = percpu_counter_init(&stat->cpu_cnt, 0, gfp); in blkg_stat_init()
566 static inline int blkg_rwstat_init(struct blkg_rwstat *rwstat, gfp_t gfp) in blkg_rwstat_init() argument
571 ret = percpu_counter_init(&rwstat->cpu_cnt[i], 0, gfp); in blkg_rwstat_init()
Dbuffer_head.h179 unsigned size, gfp_t gfp);
184 sector_t block, unsigned size, gfp_t gfp);
320 sb_getblk_gfp(struct super_block *sb, sector_t block, gfp_t gfp) in sb_getblk_gfp() argument
322 return __getblk_gfp(sb->s_bdev, block, sb->s_blocksize, gfp); in sb_getblk_gfp()
Dvmstat.h166 extern void zone_statistics(struct zone *, struct zone *, gfp_t gfp);
171 #define zone_statistics(_zl, _z, gfp) do { } while (0) argument
Ddma-iommu.h42 gfp_t gfp, int prot, dma_addr_t *handle,
Dbio.h409 gfp_t gfp, struct bio_set *bs);
422 gfp_t gfp, struct bio_set *bs) in bio_next_split() argument
427 return bio_split(bio, sectors, gfp, bs); in bio_next_split()
514 extern int bio_alloc_pages(struct bio *bio, gfp_t gfp);
Dpagemap.h222 extern struct page *__page_cache_alloc(gfp_t gfp);
224 static inline struct page *__page_cache_alloc(gfp_t gfp) in __page_cache_alloc() argument
226 return alloc_pages(gfp, 0); in __page_cache_alloc()
Dfscache-cache.h147 gfp_t gfp);
152 gfp_t gfp);
Dpercpu-refcount.h98 gfp_t gfp);
/linux-4.4.14/include/net/sctp/
Dulpevent.h86 gfp_t gfp);
94 gfp_t gfp);
100 gfp_t gfp);
106 gfp_t gfp);
111 gfp_t gfp);
115 __u32 indication, gfp_t gfp);
118 const struct sctp_association *asoc, gfp_t gfp);
122 gfp_t gfp);
126 __u32 indication, gfp_t gfp);
129 const struct sctp_association *asoc, gfp_t gfp);
Dauth.h83 struct sctp_shared_key *sctp_auth_shkey_create(__u16 key_id, gfp_t gfp);
85 int sctp_auth_asoc_init_active_key(struct sctp_association *asoc, gfp_t gfp);
91 gfp_t gfp);
92 int sctp_auth_init_hmacs(struct sctp_endpoint *ep, gfp_t gfp);
104 struct sctp_auth_chunk *auth, gfp_t gfp);
Dsm.h182 gfp_t gfp);
189 gfp_t gfp, int vparam_len);
192 const gfp_t gfp,
273 gfp_t gfp);
285 gfp_t gfp, int *err,
Dstructs.h404 gfp_t gfp);
1096 sctp_scope_t scope, gfp_t gfp,
1100 gfp_t gfp);
1102 __u8 addr_state, gfp_t gfp);
1116 gfp_t gfp);
1118 __u16 port, gfp_t gfp);
1284 sctp_init_chunk_t *init, gfp_t gfp);
1868 sctp_scope_t scope, gfp_t gfp);
1882 const gfp_t gfp,
1913 gfp_t gfp);
Dtsnmap.h100 __u32 initial_tsn, gfp_t gfp);
/linux-4.4.14/arch/x86/include/asm/
Ddma-mapping.h43 bool arch_dma_alloc_attrs(struct device **dev, gfp_t *gfp);
92 gfp_t gfp) in dma_alloc_coherent_mask() argument
98 dma_mask = (gfp & GFP_DMA) ? DMA_BIT_MASK(24) : DMA_BIT_MASK(32); in dma_alloc_coherent_mask()
103 static inline gfp_t dma_alloc_coherent_gfp_flags(struct device *dev, gfp_t gfp) in dma_alloc_coherent_gfp_flags() argument
105 unsigned long dma_mask = dma_alloc_coherent_mask(dev, gfp); in dma_alloc_coherent_gfp_flags()
108 gfp |= GFP_DMA; in dma_alloc_coherent_gfp_flags()
110 if (dma_mask <= DMA_BIT_MASK(32) && !(gfp & GFP_DMA)) in dma_alloc_coherent_gfp_flags()
111 gfp |= GFP_DMA32; in dma_alloc_coherent_gfp_flags()
113 return gfp; in dma_alloc_coherent_gfp_flags()
/linux-4.4.14/mm/
Dslob.c190 static void *slob_new_pages(gfp_t gfp, int order, int node) in slob_new_pages() argument
196 page = __alloc_pages_node(node, gfp, order); in slob_new_pages()
199 page = alloc_pages(gfp, order); in slob_new_pages()
268 static void *slob_alloc(size_t size, gfp_t gfp, int align, int node) in slob_alloc() argument
316 b = slob_new_pages(gfp & ~__GFP_ZERO, 0, node); in slob_alloc()
332 if (unlikely((gfp & __GFP_ZERO) && b)) in slob_alloc()
427 __do_kmalloc_node(size_t size, gfp_t gfp, int node, unsigned long caller) in __do_kmalloc_node() argument
433 gfp &= gfp_allowed_mask; in __do_kmalloc_node()
435 lockdep_trace_alloc(gfp); in __do_kmalloc_node()
441 m = slob_alloc(size + align, gfp, align, node); in __do_kmalloc_node()
[all …]
Dzbud.c140 static void *zbud_zpool_create(const char *name, gfp_t gfp, in zbud_zpool_create() argument
146 pool = zbud_create_pool(gfp, zpool_ops ? &zbud_zpool_ops : NULL); in zbud_zpool_create()
159 static int zbud_zpool_malloc(void *pool, size_t size, gfp_t gfp, in zbud_zpool_malloc() argument
162 return zbud_alloc(pool, size, gfp, handle); in zbud_zpool_malloc()
305 struct zbud_pool *zbud_create_pool(gfp_t gfp, const struct zbud_ops *ops) in zbud_create_pool() argument
310 pool = kzalloc(sizeof(struct zbud_pool), gfp); in zbud_create_pool()
353 int zbud_alloc(struct zbud_pool *pool, size_t size, gfp_t gfp, in zbud_alloc() argument
361 if (!size || (gfp & __GFP_HIGHMEM)) in zbud_alloc()
385 page = alloc_page(gfp); in zbud_alloc()
Dutil.c44 char *kstrdup(const char *s, gfp_t gfp) in kstrdup() argument
53 buf = kmalloc_track_caller(len, gfp); in kstrdup()
69 const char *kstrdup_const(const char *s, gfp_t gfp) in kstrdup_const() argument
74 return kstrdup(s, gfp); in kstrdup_const()
84 char *kstrndup(const char *s, size_t max, gfp_t gfp) in kstrndup() argument
93 buf = kmalloc_track_caller(len+1, gfp); in kstrndup()
109 void *kmemdup(const void *src, size_t len, gfp_t gfp) in kmemdup() argument
113 p = kmalloc_track_caller(len, gfp); in kmemdup()
Dzpool.c153 struct zpool *zpool_create_pool(const char *type, const char *name, gfp_t gfp, in zpool_create_pool() argument
173 zpool = kmalloc(sizeof(*zpool), gfp); in zpool_create_pool()
181 zpool->pool = driver->create(name, gfp, ops, zpool); in zpool_create_pool()
254 int zpool_malloc(struct zpool *zpool, size_t size, gfp_t gfp, in zpool_malloc() argument
257 return zpool->driver->malloc(zpool->pool, size, gfp, handle); in zpool_malloc()
Dbacking-dev.c290 int blkcg_id, gfp_t gfp) in wb_init() argument
314 wb->congested = wb_congested_get_create(bdi, blkcg_id, gfp); in wb_init()
318 err = fprop_local_init_percpu(&wb->completions, gfp); in wb_init()
323 err = percpu_counter_init(&wb->stat[i], 0, gfp); in wb_init()
399 wb_congested_get_create(struct backing_dev_info *bdi, int blkcg_id, gfp_t gfp) in wb_congested_get_create() argument
434 new_congested = kzalloc(sizeof(*new_congested), gfp); in wb_congested_get_create()
519 struct cgroup_subsys_state *memcg_css, gfp_t gfp) in cgwb_create() argument
547 wb = kmalloc(sizeof(*wb), gfp); in cgwb_create()
551 ret = wb_init(wb, bdi, blkcg_css->id, gfp); in cgwb_create()
555 ret = percpu_ref_init(&wb->refcnt, cgwb_release, 0, gfp); in cgwb_create()
[all …]
Dvmpressure.c211 void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, in vmpressure() argument
227 if (!(gfp & (__GFP_HIGHMEM | __GFP_MOVABLE | __GFP_IO | __GFP_FS))) in vmpressure()
263 void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, int prio) in vmpressure_prio() argument
279 vmpressure(gfp, memcg, vmpressure_win, 0); in vmpressure_prio()
Dshmem.c121 static bool shmem_should_replace_page(struct page *page, gfp_t gfp);
122 static int shmem_replace_page(struct page **pagep, gfp_t gfp,
125 struct page **pagep, enum sgp_type sgp, gfp_t gfp, int *fault_type);
640 gfp_t gfp; in shmem_unuse_inode() local
657 gfp = mapping_gfp_mask(mapping); in shmem_unuse_inode()
658 if (shmem_should_replace_page(*pagep, gfp)) { in shmem_unuse_inode()
660 error = shmem_replace_page(pagep, gfp, info, index); in shmem_unuse_inode()
896 static struct page *shmem_swapin(swp_entry_t swap, gfp_t gfp, in shmem_swapin() argument
909 page = swapin_readahead(swap, gfp, &pvma, 0); in shmem_swapin()
917 static struct page *shmem_alloc_page(gfp_t gfp, in shmem_alloc_page() argument
[all …]
Dmempolicy.c1635 static nodemask_t *policy_nodemask(gfp_t gfp, struct mempolicy *policy) in policy_nodemask() argument
1639 apply_policy_zone(policy, gfp_zone(gfp)) && in policy_nodemask()
1647 static struct zonelist *policy_zonelist(gfp_t gfp, struct mempolicy *policy, in policy_zonelist() argument
1662 if (unlikely(gfp & __GFP_THISNODE) && in policy_zonelist()
1669 return node_zonelist(nd, gfp); in policy_zonelist()
1922 static struct page *alloc_page_interleave(gfp_t gfp, unsigned order, in alloc_page_interleave() argument
1928 zl = node_zonelist(nid, gfp); in alloc_page_interleave()
1929 page = __alloc_pages(gfp, order, zl); in alloc_page_interleave()
1959 alloc_pages_vma(gfp_t gfp, int order, struct vm_area_struct *vma, in alloc_pages_vma() argument
1977 page = alloc_page_interleave(gfp, order, nid); in alloc_pages_vma()
[all …]
Dkmemleak.c125 #define gfp_kmemleak_mask(gfp) (((gfp) & (GFP_KERNEL | GFP_ATOMIC | \ argument
539 int min_count, gfp_t gfp) in create_object() argument
545 object = kmem_cache_alloc(object_cache, gfp_kmemleak_mask(gfp)); in create_object()
753 static void add_scan_area(unsigned long ptr, size_t size, gfp_t gfp) in add_scan_area() argument
766 area = kmem_cache_alloc(scan_area_cache, gfp_kmemleak_mask(gfp)); in add_scan_area()
910 gfp_t gfp) in kmemleak_alloc() argument
915 create_object((unsigned long)ptr, size, min_count, gfp); in kmemleak_alloc()
931 gfp_t gfp) in kmemleak_alloc_percpu() argument
944 size, 0, gfp); in kmemleak_alloc_percpu()
1093 void __ref kmemleak_scan_area(const void *ptr, size_t size, gfp_t gfp) in kmemleak_scan_area() argument
[all …]
Dslab.h240 gfp_t gfp, int order, in memcg_charge_slab() argument
247 return __memcg_kmem_charge_memcg(page, gfp, order, in memcg_charge_slab()
285 static inline int memcg_charge_slab(struct page *page, gfp_t gfp, int order, in memcg_charge_slab() argument
Dslab.c250 static int enable_cpucache(struct kmem_cache *cachep, gfp_t gfp);
680 int batchcount, gfp_t gfp) in alloc_arraycache() argument
685 ac = kmalloc_node(memsize, gfp, node); in alloc_arraycache()
835 int limit, gfp_t gfp) in alloc_alien_cache() argument
872 int batch, gfp_t gfp) in __alloc_alien_cache() argument
877 alc = kmalloc_node(memsize, gfp, node); in __alloc_alien_cache()
883 static struct alien_cache **alloc_alien_cache(int node, int limit, gfp_t gfp) in alloc_alien_cache() argument
891 alc_ptr = kzalloc_node(memsize, gfp, node); in alloc_alien_cache()
898 alc_ptr[i] = __alloc_alien_cache(node, limit, 0xbaadf00d, gfp); in alloc_alien_cache()
2013 static int __init_refok setup_cpu_cache(struct kmem_cache *cachep, gfp_t gfp) in setup_cpu_cache() argument
[all …]
Dpercpu-vm.c85 const gfp_t gfp = GFP_KERNEL | __GFP_HIGHMEM | __GFP_COLD; in pcpu_alloc_pages() local
93 *pagep = alloc_pages_node(cpu_to_node(cpu), gfp, 0); in pcpu_alloc_pages()
Dfilemap.c708 struct page *__page_cache_alloc(gfp_t gfp) in __page_cache_alloc() argument
718 page = __alloc_pages_node(n, gfp, 0); in __page_cache_alloc()
723 return alloc_pages(gfp, 0); in __page_cache_alloc()
2209 gfp_t gfp) in __read_cache_page() argument
2216 page = __page_cache_alloc(gfp | __GFP_COLD); in __read_cache_page()
2219 err = add_to_page_cache_lru(page, mapping, index, gfp); in __read_cache_page()
2242 gfp_t gfp) in do_read_cache_page() argument
2249 page = __read_cache_page(mapping, index, filler, data, gfp); in do_read_cache_page()
2313 gfp_t gfp) in read_cache_page_gfp() argument
2317 return do_read_cache_page(mapping, index, filler, NULL, gfp); in read_cache_page_gfp()
Dhuge_memory.c717 struct page *page, gfp_t gfp, in __do_huge_pmd_anonymous_page() argument
727 if (mem_cgroup_try_charge(page, mm, gfp, &memcg)) { in __do_huge_pmd_anonymous_page()
812 gfp_t gfp; in do_huge_pmd_anonymous_page() local
862 gfp = alloc_hugepage_gfpmask(transparent_hugepage_defrag(vma), 0); in do_huge_pmd_anonymous_page()
863 page = alloc_hugepage_vma(gfp, vma, haddr, HPAGE_PMD_ORDER); in do_huge_pmd_anonymous_page()
868 return __do_huge_pmd_anonymous_page(mm, vma, address, pmd, page, gfp, in do_huge_pmd_anonymous_page()
2413 khugepaged_alloc_page(struct page **hpage, gfp_t gfp, struct mm_struct *mm, in khugepaged_alloc_page() argument
2426 *hpage = __alloc_pages_node(node, gfp, HPAGE_PMD_ORDER); in khugepaged_alloc_page()
2480 khugepaged_alloc_page(struct page **hpage, gfp_t gfp, struct mm_struct *mm, in khugepaged_alloc_page() argument
2519 gfp_t gfp; in collapse_huge_page() local
[all …]
/linux-4.4.14/net/sctp/
Dauth.c73 static struct sctp_auth_bytes *sctp_auth_create_key(__u32 key_len, gfp_t gfp) in sctp_auth_create_key() argument
82 key = kmalloc(sizeof(struct sctp_auth_bytes) + key_len, gfp); in sctp_auth_create_key()
94 struct sctp_shared_key *sctp_auth_shkey_create(__u16 key_id, gfp_t gfp) in sctp_auth_shkey_create() argument
99 new = kzalloc(sizeof(struct sctp_shared_key), gfp); in sctp_auth_shkey_create()
191 gfp_t gfp) in sctp_auth_make_key_vector() argument
205 new = sctp_auth_create_key(len, gfp); in sctp_auth_make_key_vector()
226 gfp_t gfp) in sctp_auth_make_local_vector() argument
232 gfp); in sctp_auth_make_local_vector()
238 gfp_t gfp) in sctp_auth_make_peer_vector() argument
243 gfp); in sctp_auth_make_peer_vector()
[all …]
Dulpevent.c64 gfp_t gfp) in sctp_ulpevent_new() argument
69 skb = alloc_skb(size, gfp); in sctp_ulpevent_new()
129 __u16 inbound, struct sctp_chunk *chunk, gfp_t gfp) in sctp_ulpevent_make_assoc_change() argument
143 sizeof(struct sctp_assoc_change), 0, gfp); in sctp_ulpevent_make_assoc_change()
162 MSG_NOTIFICATION, gfp); in sctp_ulpevent_make_assoc_change()
258 int flags, int state, int error, gfp_t gfp) in sctp_ulpevent_make_peer_addr_change() argument
265 MSG_NOTIFICATION, gfp); in sctp_ulpevent_make_peer_addr_change()
372 gfp_t gfp) in sctp_ulpevent_make_remote_error() argument
391 skb = skb_copy_expand(chunk->skb, sizeof(*sre), 0, gfp); in sctp_ulpevent_make_remote_error()
428 __u16 flags, __u32 error, gfp_t gfp) in sctp_ulpevent_make_send_failed() argument
[all …]
Dbind_addr.c49 union sctp_addr *, sctp_scope_t scope, gfp_t gfp,
60 sctp_scope_t scope, gfp_t gfp, in sctp_bind_addr_copy() argument
72 gfp, flags); in sctp_bind_addr_copy()
84 SCTP_SCOPE_LINK, gfp, in sctp_bind_addr_copy()
105 gfp_t gfp) in sctp_bind_addr_dup() argument
114 error = sctp_add_bind_addr(dest, &addr->a, 1, gfp); in sctp_bind_addr_dup()
153 __u8 addr_state, gfp_t gfp) in sctp_add_bind_addr() argument
158 addr = kzalloc(sizeof(*addr), gfp); in sctp_add_bind_addr()
221 gfp_t gfp) in sctp_bind_addrs_to_raw() argument
248 retval.v = kmalloc(len, gfp); in sctp_bind_addrs_to_raw()
[all …]
Dendpointola.c59 gfp_t gfp) in sctp_endpoint_init() argument
67 ep->digest = kzalloc(SCTP_SIGNATURE_SIZE, gfp); in sctp_endpoint_init()
78 sizeof(__u16) * SCTP_AUTH_NUM_HMACS, gfp); in sctp_endpoint_init()
83 SCTP_NUM_CHUNK_TYPES, gfp); in sctp_endpoint_init()
151 null_key = sctp_auth_shkey_create(0, gfp); in sctp_endpoint_init()
158 err = sctp_auth_init_hmacs(ep, gfp); in sctp_endpoint_init()
184 struct sctp_endpoint *sctp_endpoint_new(struct sock *sk, gfp_t gfp) in sctp_endpoint_new() argument
189 ep = kzalloc(sizeof(*ep), gfp); in sctp_endpoint_new()
193 if (!sctp_endpoint_init(ep, sk, gfp)) in sctp_endpoint_new()
Dssnmap.c53 gfp_t gfp) in sctp_ssnmap_new() argument
60 retval = kmalloc(size, gfp); in sctp_ssnmap_new()
63 __get_free_pages(gfp, get_order(size)); in sctp_ssnmap_new()
Dassociola.c70 gfp_t gfp) in sctp_association_init() argument
266 err = sctp_auth_asoc_copy_shkeys(ep, asoc, gfp); in sctp_association_init()
298 gfp_t gfp) in sctp_association_new() argument
302 asoc = kzalloc(sizeof(*asoc), gfp); in sctp_association_new()
306 if (!sctp_association_init(asoc, ep, sk, scope, gfp)) in sctp_association_new()
576 const gfp_t gfp, in sctp_assoc_add_peer() argument
609 peer = sctp_transport_new(net, addr, gfp); in sctp_assoc_add_peer()
1544 sctp_scope_t scope, gfp_t gfp) in sctp_assoc_set_bind_addr_from_ep() argument
1560 scope, gfp, flags); in sctp_assoc_set_bind_addr_from_ep()
1566 gfp_t gfp) in sctp_assoc_set_bind_addr_from_cookie() argument
[all …]
Dulpqueue.c96 gfp_t gfp) in sctp_ulpq_tail_data() argument
103 event = sctp_ulpevent_make_rcvmsg(chunk->asoc, chunk, gfp); in sctp_ulpq_tail_data()
1035 gfp_t gfp) in sctp_ulpq_partial_delivery() argument
1081 gfp_t gfp) in sctp_ulpq_renege() argument
1105 retval = sctp_ulpq_tail_data(ulpq, chunk, gfp); in sctp_ulpq_renege()
1111 sctp_ulpq_partial_delivery(ulpq, gfp); in sctp_ulpq_renege()
1124 void sctp_ulpq_abort_pd(struct sctp_ulpq *ulpq, gfp_t gfp) in sctp_ulpq_abort_pd() argument
1137 gfp); in sctp_ulpq_abort_pd()
Dsm_sideeffect.c62 gfp_t gfp);
70 gfp_t gfp);
615 gfp_t gfp) in sctp_cmd_process_init() argument
624 if (!sctp_process_init(asoc, chunk, sctp_source(chunk), peer_init, gfp)) in sctp_cmd_process_init()
1102 gfp_t gfp) in sctp_do_sm() argument
1127 &commands, gfp); in sctp_do_sm()
1143 gfp_t gfp) in sctp_side_effects() argument
1156 commands, gfp))) in sctp_side_effects()
1228 gfp_t gfp) in sctp_cmd_interpreter() argument
1350 cmd->obj.init, gfp); in sctp_cmd_interpreter()
Dsm_make_chunk.c78 gfp_t gfp);
219 gfp_t gfp, int vparam_len) in sctp_make_init() argument
246 addrs = sctp_bind_addrs_to_raw(bp, &addrs_len, gfp); in sctp_make_init()
384 gfp_t gfp, int unkparam_len) in sctp_make_init_ack() argument
405 addrs = sctp_bind_addrs_to_raw(&asoc->base.bind_addr, &addrs_len, gfp); in sctp_make_init_ack()
1571 gfp_t gfp) in sctp_make_temp_asoc() argument
1580 asoc = sctp_association_new(ep, ep->base.sk, scope, gfp); in sctp_make_temp_asoc()
1692 struct sctp_chunk *chunk, gfp_t gfp, in sctp_unpack_cookie() argument
1812 retval = sctp_association_new(ep, ep->base.sk, scope, gfp); in sctp_unpack_cookie()
2292 sctp_init_chunk_t *peer_init, gfp_t gfp) in sctp_process_init() argument
[all …]
Dtransport.c58 gfp_t gfp) in sctp_transport_init() argument
108 gfp_t gfp) in sctp_transport_new() argument
112 transport = kzalloc(sizeof(*transport), gfp); in sctp_transport_new()
116 if (!sctp_transport_init(net, transport, addr, gfp)) in sctp_transport_new()
Dtsnmap.c51 __u32 initial_tsn, gfp_t gfp) in sctp_tsnmap_init() argument
54 map->tsn_map = kzalloc(len>>3, gfp); in sctp_tsnmap_init()
Dchunk.c62 static struct sctp_datamsg *sctp_datamsg_new(gfp_t gfp) in sctp_datamsg_new() argument
65 msg = kmalloc(sizeof(struct sctp_datamsg), gfp); in sctp_datamsg_new()
/linux-4.4.14/arch/mn10300/mm/
Ddma-alloc.c24 dma_addr_t *dma_handle, int gfp) in dma_alloc_coherent() argument
30 dev ? dev_name(dev) : "?", size, gfp); in dma_alloc_coherent()
41 gfp &= ~(__GFP_DMA | __GFP_HIGHMEM); in dma_alloc_coherent()
44 gfp |= GFP_DMA; in dma_alloc_coherent()
46 addr = __get_free_pages(gfp, get_order(size)); in dma_alloc_coherent()
/linux-4.4.14/scripts/coccinelle/locks/
Dcall_kern.cocci19 @gfp exists@
39 identifier gfp.fn;
74 position gfp.p;
81 position gfp.p;
87 p << gfp.p;
88 fn << gfp.fn;
98 p << gfp.p;
99 fn << gfp.fn;
/linux-4.4.14/tools/testing/fault-injection/
Dfailcmd.sh43 --reject-start=value, --reject-end=value, --ignore-gfp-wait=value
51 --ignore-gfp-highmem=value, --min-order=value
91 LONGOPTS=$LONGOPTS,ignore-gfp-wait:,cache-filter:
93 LONGOPTS=$LONGOPTS,ignore-gfp-wait:,ignore-gfp-highmem:,min-order:
181 --ignore-gfp-wait)
182 echo $2 > $FAULTATTR/ignore-gfp-wait
189 --ignore-gfp-highmem)
190 echo $2 > $FAULTATTR/ignore-gfp-highmem
/linux-4.4.14/arch/avr32/mm/
Ddma-coherent.c41 dma_addr_t *handle, gfp_t gfp) in __dma_alloc() argument
51 gfp &= ~(__GFP_COMP); in __dma_alloc()
56 page = alloc_pages(gfp, order); in __dma_alloc()
97 dma_addr_t *handle, gfp_t gfp) in dma_alloc_coherent() argument
102 page = __dma_alloc(dev, size, handle, gfp); in dma_alloc_coherent()
125 dma_addr_t *handle, gfp_t gfp) in dma_alloc_writecombine() argument
130 page = __dma_alloc(dev, size, handle, gfp); in dma_alloc_writecombine()
/linux-4.4.14/arch/mips/netlogic/common/
Dnlm-dma.c48 dma_addr_t *dma_handle, gfp_t gfp, struct dma_attrs *attrs) in nlm_dma_alloc_coherent() argument
51 gfp &= ~(__GFP_DMA | __GFP_DMA32 | __GFP_HIGHMEM); in nlm_dma_alloc_coherent()
55 gfp |= __GFP_DMA32; in nlm_dma_alloc_coherent()
59 gfp |= __GFP_NORETRY; in nlm_dma_alloc_coherent()
61 return swiotlb_alloc_coherent(dev, size, dma_handle, gfp); in nlm_dma_alloc_coherent()
/linux-4.4.14/arch/mips/loongson64/common/
Ddma-swiotlb.c13 dma_addr_t *dma_handle, gfp_t gfp, struct dma_attrs *attrs) in loongson_dma_alloc_coherent() argument
18 gfp &= ~(__GFP_DMA | __GFP_DMA32 | __GFP_HIGHMEM); in loongson_dma_alloc_coherent()
22 gfp |= __GFP_DMA; in loongson_dma_alloc_coherent()
27 gfp |= __GFP_DMA; in loongson_dma_alloc_coherent()
32 gfp |= __GFP_DMA32; in loongson_dma_alloc_coherent()
36 gfp |= __GFP_NORETRY; in loongson_dma_alloc_coherent()
38 ret = swiotlb_alloc_coherent(dev, size, dma_handle, gfp); in loongson_dma_alloc_coherent()
/linux-4.4.14/arch/arc/mm/
Ddma.c29 dma_addr_t *dma_handle, gfp_t gfp) in dma_alloc_noncoherent() argument
34 paddr = alloc_pages_exact(size, gfp); in dma_alloc_noncoherent()
53 dma_addr_t *dma_handle, gfp_t gfp) in dma_alloc_coherent() argument
69 return dma_alloc_noncoherent(dev, size, dma_handle, gfp); in dma_alloc_coherent()
72 paddr = alloc_pages_exact(size, gfp); in dma_alloc_coherent()
/linux-4.4.14/arch/mips/mm/
Ddma-default.c78 static gfp_t massage_gfp_flags(const struct device *dev, gfp_t gfp) in massage_gfp_flags() argument
83 gfp &= ~(__GFP_DMA | __GFP_DMA32 | __GFP_HIGHMEM); in massage_gfp_flags()
110 gfp |= __GFP_NORETRY; in massage_gfp_flags()
112 return gfp | dma_flag; in massage_gfp_flags()
116 dma_addr_t * dma_handle, gfp_t gfp) in mips_dma_alloc_noncoherent() argument
120 gfp = massage_gfp_flags(dev, gfp); in mips_dma_alloc_noncoherent()
122 ret = (void *) __get_free_pages(gfp, get_order(size)); in mips_dma_alloc_noncoherent()
133 dma_addr_t * dma_handle, gfp_t gfp, struct dma_attrs *attrs) in mips_dma_alloc_coherent() argument
144 return mips_dma_alloc_noncoherent(dev, size, dma_handle, gfp); in mips_dma_alloc_coherent()
146 gfp = massage_gfp_flags(dev, gfp); in mips_dma_alloc_coherent()
[all …]
/linux-4.4.14/arch/cris/arch-v32/drivers/pci/
Ddma.c20 dma_addr_t *dma_handle, gfp_t gfp) in dma_alloc_coherent() argument
25 gfp &= ~(__GFP_DMA | __GFP_HIGHMEM); in dma_alloc_coherent()
31 gfp |= GFP_DMA; in dma_alloc_coherent()
33 ret = (void *)__get_free_pages(gfp, order); in dma_alloc_coherent()
/linux-4.4.14/include/net/netfilter/
Dnf_conntrack_extend.h94 size_t var_alloc_len, gfp_t gfp);
96 #define nf_ct_ext_add(ct, id, gfp) \ argument
97 ((id##_TYPE *)__nf_ct_ext_add_length((ct), (id), 0, (gfp)))
98 #define nf_ct_ext_add_length(ct, id, len, gfp) \ argument
99 ((id##_TYPE *)__nf_ct_ext_add_length((ct), (id), (len), (gfp)))
Dnf_conntrack_timestamp.h26 struct nf_conn_tstamp *nf_ct_tstamp_ext_add(struct nf_conn *ct, gfp_t gfp) in nf_ct_tstamp_ext_add() argument
34 return nf_ct_ext_add(ct, NF_CT_EXT_TSTAMP, gfp); in nf_ct_tstamp_ext_add()
Dnf_conntrack_acct.h33 struct nf_conn_acct *nf_ct_acct_ext_add(struct nf_conn *ct, gfp_t gfp) in nf_ct_acct_ext_add() argument
41 acct = nf_ct_ext_add(ct, NF_CT_EXT_ACCT, gfp); in nf_ct_acct_ext_add()
Dnf_conntrack_timeout.h51 gfp_t gfp) in nf_ct_timeout_ext_add() argument
56 timeout_ext = nf_ct_ext_add(ct, NF_CT_EXT_TIMEOUT, gfp); in nf_ct_timeout_ext_add()
Dnf_conntrack_ecache.h34 nf_ct_ecache_ext_add(struct nf_conn *ct, u16 ctmask, u16 expmask, gfp_t gfp) in nf_ct_ecache_ext_add() argument
47 e = nf_ct_ext_add(ct, NF_CT_EXT_ECACHE, gfp); in nf_ct_ecache_ext_add()
Dnf_conntrack_helper.h67 gfp_t gfp);
Dnf_tables.h517 u64 timeout, gfp_t gfp);
549 gfp_t gfp);
560 gfp_t gfp) in nft_set_gc_batch_check() argument
567 return nft_set_gc_batch_alloc(set, gfp); in nft_set_gc_batch_check()
/linux-4.4.14/include/net/
Dcfg80211.h4029 gfp_t gfp);
4036 s32 signal, gfp_t gfp) in cfg80211_inform_bss_width_frame() argument
4044 return cfg80211_inform_bss_frame_data(wiphy, &data, mgmt, len, gfp); in cfg80211_inform_bss_width_frame()
4051 s32 signal, gfp_t gfp) in cfg80211_inform_bss_frame() argument
4059 return cfg80211_inform_bss_frame_data(wiphy, &data, mgmt, len, gfp); in cfg80211_inform_bss_frame()
4101 gfp_t gfp);
4110 s32 signal, gfp_t gfp) in cfg80211_inform_bss_width() argument
4120 gfp); in cfg80211_inform_bss_width()
4129 s32 signal, gfp_t gfp) in cfg80211_inform_bss() argument
4139 gfp); in cfg80211_inform_bss()
[all …]
Ddn_nsp.h22 void dn_send_conn_conf(struct sock *sk, gfp_t gfp);
24 unsigned short reason, gfp_t gfp);
33 void dn_nsp_queue_xmit(struct sock *sk, struct sk_buff *skb, gfp_t gfp,
/linux-4.4.14/tools/virtio/linux/
Dkernel.h45 static inline void *kmalloc(size_t s, gfp_t gfp) in kmalloc() argument
51 static inline void *kzalloc(size_t s, gfp_t gfp) in kzalloc() argument
53 void *p = kmalloc(s, gfp); in kzalloc()
66 static inline void *krealloc(void *p, size_t s, gfp_t gfp) in krealloc() argument
72 static inline unsigned long __get_free_page(gfp_t gfp) in __get_free_page() argument
Dvirtio.h29 gfp_t gfp);
34 gfp_t gfp);
39 gfp_t gfp);
/linux-4.4.14/arch/h8300/kernel/
Ddma.c14 dma_addr_t *dma_handle, gfp_t gfp, in dma_alloc() argument
20 gfp &= ~(__GFP_DMA | __GFP_HIGHMEM); in dma_alloc()
23 gfp |= GFP_DMA; in dma_alloc()
24 ret = (void *)__get_free_pages(gfp, get_order(size)); in dma_alloc()
/linux-4.4.14/drivers/base/
Ddevres.c85 size_t size, gfp_t gfp, int nid) in alloc_dr() argument
90 dr = kmalloc_node_track_caller(tot_size, gfp, nid); in alloc_dr()
109 void * __devres_alloc_node(dr_release_t release, size_t size, gfp_t gfp, int nid, in __devres_alloc_node() argument
114 dr = alloc_dr(release, size, gfp | __GFP_ZERO, nid); in __devres_alloc_node()
136 void * devres_alloc_node(dr_release_t release, size_t size, gfp_t gfp, int nid) in devres_alloc_node() argument
140 dr = alloc_dr(release, size, gfp | __GFP_ZERO, nid); in devres_alloc_node()
533 void * devres_open_group(struct device *dev, void *id, gfp_t gfp) in devres_open_group() argument
538 grp = kmalloc(sizeof(*grp), gfp); in devres_open_group()
775 void * devm_kmalloc(struct device *dev, size_t size, gfp_t gfp) in devm_kmalloc() argument
780 dr = alloc_dr(devm_kmalloc_release, size, gfp, dev_to_node(dev)); in devm_kmalloc()
[all …]
Ddma-mapping.c66 dma_addr_t *dma_handle, gfp_t gfp) in dmam_alloc_coherent() argument
71 dr = devres_alloc(dmam_coherent_release, sizeof(*dr), gfp); in dmam_alloc_coherent()
75 vaddr = dma_alloc_coherent(dev, size, dma_handle, gfp); in dmam_alloc_coherent()
125 dma_addr_t *dma_handle, gfp_t gfp) in dmam_alloc_noncoherent() argument
130 dr = devres_alloc(dmam_noncoherent_release, sizeof(*dr), gfp); in dmam_alloc_noncoherent()
134 vaddr = dma_alloc_noncoherent(dev, size, dma_handle, gfp); in dmam_alloc_noncoherent()
Ddevcoredump.c202 gfp_t gfp) in dev_coredumpv() argument
204 dev_coredumpm(dev, NULL, data, datalen, gfp, devcd_readv, vfree); in dev_coredumpv()
231 const void *data, size_t datalen, gfp_t gfp, in dev_coredumpm() argument
253 devcd = kzalloc(sizeof(*devcd), gfp); in dev_coredumpm()
/linux-4.4.14/net/decnet/
Ddn_nsp_out.c212 gfp_t gfp) in dn_nsp_clone_and_send() argument
218 if ((skb2 = skb_clone(skb, gfp)) != NULL) { in dn_nsp_clone_and_send()
352 gfp_t gfp, int oth) in dn_nsp_queue_xmit() argument
376 dn_nsp_clone_and_send(skb, gfp); in dn_nsp_queue_xmit()
515 void dn_send_conn_conf(struct sock *sk, gfp_t gfp) in dn_send_conn_conf() argument
522 if ((skb = dn_alloc_skb(sk, 50 + len, gfp)) == NULL) in dn_send_conn_conf()
547 unsigned short reason, gfp_t gfp, in dn_nsp_do_disc() argument
561 if ((skb = dn_alloc_skb(sk, size, gfp)) == NULL) in dn_nsp_do_disc()
590 unsigned short reason, gfp_t gfp) in dn_nsp_send_disc() argument
601 dn_nsp_do_disc(sk, msgflg, reason, gfp, __sk_dst_get(sk), ddl, in dn_nsp_send_disc()
[all …]
/linux-4.4.14/net/rds/
Dtcp_recv.c151 gfp_t gfp; member
174 arg->gfp); in rds_tcp_data_recv()
210 clone = skb_clone(skb, arg->gfp); in rds_tcp_data_recv()
244 arg->gfp); in rds_tcp_data_recv()
261 static int rds_tcp_read_sock(struct rds_connection *conn, gfp_t gfp) in rds_tcp_read_sock() argument
270 arg.gfp = gfp; in rds_tcp_read_sock()
276 rdsdebug("tcp_read_sock for tc %p gfp 0x%x returned %d\n", tc, gfp, in rds_tcp_read_sock()
Dpage.c98 gfp_t gfp) in rds_page_remainder_alloc() argument
105 gfp |= __GFP_HIGHMEM; in rds_page_remainder_alloc()
109 page = alloc_page(gfp); in rds_page_remainder_alloc()
151 page = alloc_page(gfp); in rds_page_remainder_alloc()
Dconnection.c123 struct rds_transport *trans, gfp_t gfp, in __rds_conn_create() argument
147 conn = kmem_cache_zalloc(rds_conn_slab, gfp); in __rds_conn_create()
191 ret = trans->conn_alloc(conn, gfp); in __rds_conn_create()
256 struct rds_transport *trans, gfp_t gfp) in rds_conn_create() argument
258 return __rds_conn_create(net, laddr, faddr, trans, gfp, 0); in rds_conn_create()
264 struct rds_transport *trans, gfp_t gfp) in rds_conn_create_outgoing() argument
266 return __rds_conn_create(net, laddr, faddr, trans, gfp, 1); in rds_conn_create_outgoing()
Dloop.c119 static int rds_loop_conn_alloc(struct rds_connection *conn, gfp_t gfp) in rds_loop_conn_alloc() argument
124 lc = kzalloc(sizeof(struct rds_loop_connection), gfp); in rds_loop_conn_alloc()
Drds.h438 int (*conn_alloc)(struct rds_connection *conn, gfp_t gfp);
633 struct rds_transport *trans, gfp_t gfp);
636 struct rds_transport *trans, gfp_t gfp);
676 struct rds_message *rds_message_alloc(unsigned int nents, gfp_t gfp);
708 gfp_t gfp);
723 struct rds_incoming *inc, gfp_t gfp);
Dib_recv.c300 struct rds_ib_recv_work *recv, gfp_t gfp) in rds_ib_recv_refill_one() argument
308 if (gfp & __GFP_DIRECT_RECLAIM) { in rds_ib_recv_refill_one()
375 void rds_ib_recv_refill(struct rds_connection *conn, int prefill, gfp_t gfp) in rds_ib_recv_refill() argument
382 bool can_wait = !!(gfp & __GFP_DIRECT_RECLAIM); in rds_ib_recv_refill()
401 ret = rds_ib_recv_refill_one(conn, recv, gfp); in rds_ib_recv_refill()
Dmessage.c193 struct rds_message *rds_message_alloc(unsigned int extra_len, gfp_t gfp) in rds_message_alloc() argument
200 rm = kzalloc(sizeof(struct rds_message) + extra_len, gfp); in rds_message_alloc()
/linux-4.4.14/block/
Dblk-mq.h99 gfp_t gfp; member
108 struct request_queue *q, gfp_t gfp, bool reserved, in blk_mq_set_alloc_data() argument
113 data->gfp = gfp; in blk_mq_set_alloc_data()
/linux-4.4.14/scripts/
Dgfp-translate66 grep -q ___GFP $SOURCE/include/linux/gfp.h
68 grep "^#define ___GFP" $SOURCE/include/linux/gfp.h | sed -e 's/u$//' | grep -v GFP_BITS > $TMPFILE
70 …grep "^#define __GFP" $SOURCE/include/linux/gfp.h | sed -e 's/(__force gfp_t)//' | sed -e 's/u)/)/…
/linux-4.4.14/fs/9p/
Dcache.h45 extern int __v9fs_fscache_release_page(struct page *page, gfp_t gfp);
58 gfp_t gfp) in v9fs_fscache_release_page() argument
60 return __v9fs_fscache_release_page(page, gfp); in v9fs_fscache_release_page()
118 gfp_t gfp) { in v9fs_fscache_release_page() argument
Dvfs_addr.c135 static int v9fs_release_page(struct page *page, gfp_t gfp) in v9fs_release_page() argument
139 return v9fs_fscache_release_page(page, gfp); in v9fs_release_page()
/linux-4.4.14/arch/ia64/kernel/
Dpci-swiotlb.c18 dma_addr_t *dma_handle, gfp_t gfp, in ia64_swiotlb_alloc_coherent() argument
22 gfp |= GFP_DMA; in ia64_swiotlb_alloc_coherent()
23 return swiotlb_alloc_coherent(dev, size, dma_handle, gfp); in ia64_swiotlb_alloc_coherent()
/linux-4.4.14/arch/arm/mm/
Ddma-mapping.c149 dma_addr_t *handle, gfp_t gfp, struct dma_attrs *attrs);
258 static struct page *__dma_alloc_buffer(struct device *dev, size_t size, gfp_t gfp) in __dma_alloc_buffer() argument
263 page = alloc_pages(gfp, order); in __dma_alloc_buffer()
298 static void *__alloc_remap_buffer(struct device *dev, size_t size, gfp_t gfp,
303 __dma_alloc_remap(struct page *page, size_t size, gfp_t gfp, pgprot_t prot, in __dma_alloc_remap() argument
354 gfp_t gfp = GFP_KERNEL | GFP_DMA; in atomic_pool_init() local
366 ptr = __alloc_remap_buffer(NULL, atomic_pool_size, gfp, prot, in atomic_pool_init()
472 static void *__alloc_remap_buffer(struct device *dev, size_t size, gfp_t gfp, in __alloc_remap_buffer() argument
478 page = __dma_alloc_buffer(dev, size, gfp); in __alloc_remap_buffer()
484 ptr = __dma_alloc_remap(page, size, gfp, prot, caller); in __alloc_remap_buffer()
[all …]
/linux-4.4.14/arch/nios2/mm/
Ddma-mapping.c25 dma_addr_t *dma_handle, gfp_t gfp) in dma_alloc_coherent() argument
30 gfp &= ~(__GFP_DMA | __GFP_HIGHMEM); in dma_alloc_coherent()
33 gfp |= __GFP_ZERO; in dma_alloc_coherent()
36 gfp |= GFP_DMA; in dma_alloc_coherent()
38 ret = (void *) __get_free_pages(gfp, get_order(size)); in dma_alloc_coherent()
/linux-4.4.14/arch/mips/cavium-octeon/
Ddma-octeon.c160 dma_addr_t *dma_handle, gfp_t gfp, struct dma_attrs *attrs) in octeon_dma_alloc_coherent() argument
165 gfp &= ~(__GFP_DMA | __GFP_DMA32 | __GFP_HIGHMEM); in octeon_dma_alloc_coherent()
169 gfp |= __GFP_DMA; in octeon_dma_alloc_coherent()
171 gfp |= __GFP_DMA; in octeon_dma_alloc_coherent()
176 gfp |= __GFP_DMA32; in octeon_dma_alloc_coherent()
182 gfp |= __GFP_NORETRY; in octeon_dma_alloc_coherent()
184 ret = swiotlb_alloc_coherent(dev, size, dma_handle, gfp); in octeon_dma_alloc_coherent()
/linux-4.4.14/net/ceph/
Dbuffer.c11 struct ceph_buffer *ceph_buffer_new(size_t len, gfp_t gfp) in ceph_buffer_new() argument
15 b = kmalloc(sizeof(*b), gfp); in ceph_buffer_new()
19 b->vec.iov_base = ceph_kvmalloc(len, gfp); in ceph_buffer_new()
/linux-4.4.14/net/netfilter/
Dnf_conntrack_extend.c48 size_t var_alloc_len, gfp_t gfp) in nf_ct_ext_create() argument
62 *ext = kzalloc(alloc_size, gfp); in nf_ct_ext_create()
73 size_t var_alloc_len, gfp_t gfp) in __nf_ct_ext_add_length() argument
84 return nf_ct_ext_create(&ct->ext, id, var_alloc_len, gfp); in __nf_ct_ext_add_length()
97 new = __krealloc(old, newlen, gfp); in __nf_ct_ext_add_length()
/linux-4.4.14/drivers/usb/storage/
Duas.c77 struct uas_dev_info *devinfo, gfp_t gfp);
399 static struct urb *uas_alloc_data_urb(struct uas_dev_info *devinfo, gfp_t gfp, in uas_alloc_data_urb() argument
405 struct urb *urb = usb_alloc_urb(0, gfp); in uas_alloc_data_urb()
423 static struct urb *uas_alloc_sense_urb(struct uas_dev_info *devinfo, gfp_t gfp, in uas_alloc_sense_urb() argument
428 struct urb *urb = usb_alloc_urb(0, gfp); in uas_alloc_sense_urb()
434 iu = kzalloc(sizeof(*iu), gfp); in uas_alloc_sense_urb()
450 static struct urb *uas_alloc_cmd_urb(struct uas_dev_info *devinfo, gfp_t gfp, in uas_alloc_cmd_urb() argument
456 struct urb *urb = usb_alloc_urb(0, gfp); in uas_alloc_cmd_urb()
467 iu = kzalloc(sizeof(*iu) + len, gfp); in uas_alloc_cmd_urb()
494 static struct urb *uas_submit_sense_urb(struct scsi_cmnd *cmnd, gfp_t gfp) in uas_submit_sense_urb() argument
[all …]
/linux-4.4.14/arch/powerpc/mm/
Ddma-noncoherent.c97 ppc_vm_region_alloc(struct ppc_vm_region *head, size_t size, gfp_t gfp) in ppc_vm_region_alloc() argument
103 new = kmalloc(sizeof(struct ppc_vm_region), gfp); in ppc_vm_region_alloc()
155 __dma_alloc_coherent(struct device *dev, size_t size, dma_addr_t *handle, gfp_t gfp) in __dma_alloc_coherent() argument
196 gfp |= GFP_DMA; in __dma_alloc_coherent()
198 page = alloc_pages(gfp, order); in __dma_alloc_coherent()
216 gfp & ~(__GFP_DMA | __GFP_HIGHMEM)); in __dma_alloc_coherent()
/linux-4.4.14/arch/tile/kernel/
Dpci-dma.c36 dma_addr_t *dma_handle, gfp_t gfp, in tile_dma_alloc_coherent() argument
46 gfp |= __GFP_ZERO; in tile_dma_alloc_coherent()
57 gfp |= GFP_DMA; in tile_dma_alloc_coherent()
61 pg = homecache_alloc_pages_node(node, gfp, order, PAGE_HOME_DMA); in tile_dma_alloc_coherent()
345 dma_addr_t *dma_handle, gfp_t gfp, in tile_pci_dma_alloc_coherent() argument
353 gfp |= __GFP_ZERO; in tile_pci_dma_alloc_coherent()
355 pg = homecache_alloc_pages_node(node, gfp, order, PAGE_HOME_DMA); in tile_pci_dma_alloc_coherent()
533 dma_addr_t *dma_handle, gfp_t gfp, in tile_swiotlb_alloc_coherent() argument
536 gfp |= GFP_DMA; in tile_swiotlb_alloc_coherent()
537 return swiotlb_alloc_coherent(dev, size, dma_handle, gfp); in tile_swiotlb_alloc_coherent()
/linux-4.4.14/fs/fscache/
Dpage.c65 gfp_t gfp) in __fscache_maybe_release_page() argument
70 _enter("%p,%p,%x", cookie, page, gfp); in __fscache_maybe_release_page()
125 if (!(gfp & __GFP_DIRECT_RECLAIM) || !(gfp & __GFP_FS)) { in __fscache_maybe_release_page()
135 gfp &= ~__GFP_DIRECT_RECLAIM; in __fscache_maybe_release_page()
414 gfp_t gfp) in __fscache_read_or_alloc_page() argument
479 ret = object->cache->ops->allocate_page(op, page, gfp); in __fscache_read_or_alloc_page()
485 ret = object->cache->ops->read_or_alloc_page(op, page, gfp); in __fscache_read_or_alloc_page()
544 gfp_t gfp) in __fscache_read_or_alloc_pages() argument
606 op, pages, nr_pages, gfp); in __fscache_read_or_alloc_pages()
611 op, pages, nr_pages, gfp); in __fscache_read_or_alloc_pages()
[all …]
/linux-4.4.14/arch/m68k/kernel/
Ddma.c76 dma_addr_t *dma_handle, gfp_t gfp) in dma_alloc_coherent() argument
80 gfp &= ~(__GFP_DMA | __GFP_HIGHMEM); in dma_alloc_coherent()
83 gfp |= GFP_DMA; in dma_alloc_coherent()
84 ret = (void *)__get_free_pages(gfp, get_order(size)); in dma_alloc_coherent()
/linux-4.4.14/arch/metag/kernel/
Ddma.c115 gfp_t gfp) in metag_vm_region_alloc() argument
121 new = kmalloc(sizeof(struct metag_vm_region), gfp); in metag_vm_region_alloc()
175 dma_addr_t *handle, gfp_t gfp) in dma_alloc_coherent() argument
203 gfp |= GFP_DMA; in dma_alloc_coherent()
205 page = alloc_pages(gfp, order); in dma_alloc_coherent()
223 gfp & ~(__GFP_DMA | __GFP_HIGHMEM)); in dma_alloc_coherent()
/linux-4.4.14/drivers/md/bcache/
Dutil.h40 #define init_heap(heap, _size, gfp) \ argument
48 (heap)->data = kmalloc(_bytes, (gfp)); \
49 if ((!(heap)->data) && ((gfp) & GFP_KERNEL)) \
129 #define __init_fifo(fifo, gfp) \ argument
142 (fifo)->data = kmalloc(_bytes, (gfp)); \
143 if ((!(fifo)->data) && ((gfp) & GFP_KERNEL)) \
148 #define init_fifo_exact(fifo, _size, gfp) \ argument
151 __init_fifo(fifo, gfp); \
154 #define init_fifo(fifo, _size, gfp) \ argument
159 __init_fifo(fifo, gfp); \
/linux-4.4.14/drivers/virtio/
Dvirtio_ring.c108 unsigned int total_sg, gfp_t gfp) in alloc_indirect() argument
118 gfp &= ~__GFP_HIGHMEM; in alloc_indirect()
120 desc = kmalloc(total_sg * sizeof(struct vring_desc), gfp); in alloc_indirect()
135 gfp_t gfp) in virtqueue_add() argument
174 desc = alloc_indirect(_vq, total_sg, gfp); in virtqueue_add()
284 gfp_t gfp) in virtqueue_add_sgs() argument
294 return virtqueue_add(_vq, sgs, total_sg, out_sgs, in_sgs, data, gfp); in virtqueue_add_sgs()
314 gfp_t gfp) in virtqueue_add_outbuf() argument
316 return virtqueue_add(vq, &sg, num, 1, 0, data, gfp); in virtqueue_add_outbuf()
336 gfp_t gfp) in virtqueue_add_inbuf() argument
[all …]
/linux-4.4.14/net/rxrpc/
Dar-connection.c35 static struct rxrpc_conn_bundle *rxrpc_alloc_bundle(gfp_t gfp) in rxrpc_alloc_bundle() argument
41 bundle = kzalloc(sizeof(struct rxrpc_conn_bundle), gfp); in rxrpc_alloc_bundle()
73 gfp_t gfp) in rxrpc_get_bundle() argument
106 candidate = rxrpc_alloc_bundle(gfp); in rxrpc_get_bundle()
198 static struct rxrpc_connection *rxrpc_alloc_connection(gfp_t gfp) in rxrpc_alloc_connection() argument
204 conn = kzalloc(sizeof(struct rxrpc_connection), gfp); in rxrpc_alloc_connection()
339 gfp_t gfp) in rxrpc_connect_exclusive() argument
350 conn = rxrpc_alloc_connection(gfp); in rxrpc_connect_exclusive()
433 gfp_t gfp) in rxrpc_connect_call() argument
444 call, gfp); in rxrpc_connect_call()
[all …]
Dar-transport.c36 gfp_t gfp) in rxrpc_alloc_transport() argument
42 trans = kzalloc(sizeof(struct rxrpc_transport), gfp); in rxrpc_alloc_transport()
80 gfp_t gfp) in rxrpc_get_transport() argument
102 candidate = rxrpc_alloc_transport(local, peer, gfp); in rxrpc_get_transport()
Dar-peer.c62 gfp_t gfp) in rxrpc_alloc_peer() argument
68 peer = kzalloc(sizeof(struct rxrpc_peer), gfp); in rxrpc_alloc_peer()
106 struct rxrpc_peer *rxrpc_get_peer(struct sockaddr_rxrpc *srx, gfp_t gfp) in rxrpc_get_peer() argument
139 candidate = rxrpc_alloc_peer(srx, gfp); in rxrpc_get_peer()
Dar-call.c202 static struct rxrpc_call *rxrpc_alloc_call(gfp_t gfp) in rxrpc_alloc_call() argument
206 call = kmem_cache_zalloc(rxrpc_call_jar, gfp); in rxrpc_alloc_call()
212 gfp); in rxrpc_alloc_call()
255 gfp_t gfp) in rxrpc_alloc_client_call() argument
266 call = rxrpc_alloc_call(gfp); in rxrpc_alloc_client_call()
274 ret = rxrpc_connect_call(rx, trans, bundle, call, gfp); in rxrpc_alloc_client_call()
320 gfp_t gfp) in rxrpc_get_client_call() argument
352 candidate = rxrpc_alloc_client_call(rx, trans, bundle, gfp); in rxrpc_get_client_call()
418 gfp_t gfp) in rxrpc_incoming_call() argument
424 _enter(",%d,,%x", conn->debug_id, gfp); in rxrpc_incoming_call()
[all …]
Daf_rxrpc.c226 gfp_t gfp) in rxrpc_name_to_transport() argument
244 peer = rxrpc_get_peer(srx, gfp); in rxrpc_name_to_transport()
249 trans = rxrpc_get_transport(rx->local, peer, gfp); in rxrpc_name_to_transport()
273 gfp_t gfp) in rxrpc_kernel_begin_call() argument
287 sizeof(*srx), 0, gfp); in rxrpc_kernel_begin_call()
311 bundle = rxrpc_get_bundle(rx, trans, key, service_id, gfp); in rxrpc_kernel_begin_call()
318 gfp); in rxrpc_kernel_begin_call()
/linux-4.4.14/arch/x86/kernel/
Dpci-dma.c132 bool arch_dma_alloc_attrs(struct device **dev, gfp_t *gfp) in arch_dma_alloc_attrs() argument
137 *gfp &= ~(__GFP_DMA | __GFP_HIGHMEM | __GFP_DMA32); in arch_dma_alloc_attrs()
138 *gfp = dma_alloc_coherent_gfp_flags(*dev, *gfp); in arch_dma_alloc_attrs()
/linux-4.4.14/drivers/usb/wusbcore/
Dwa-rpipe.c194 gfp_t gfp) in rpipe_get_idle() argument
201 rpipe = kzalloc(sizeof(*rpipe), gfp); in rpipe_get_idle()
314 struct usb_host_endpoint *ep, struct urb *urb, gfp_t gfp) in rpipe_aim() argument
398 const struct urb *urb, gfp_t gfp) in rpipe_check_aim() argument
438 struct urb *urb, gfp_t gfp) in rpipe_get_by_ep() argument
449 result = rpipe_check_aim(rpipe, wa, ep, urb, gfp); in rpipe_get_by_ep()
461 result = rpipe_get_idle(&rpipe, wa, 1 << eptype, gfp); in rpipe_get_by_ep()
464 result = rpipe_aim(rpipe, wa, ep, urb, gfp); in rpipe_get_by_ep()
/linux-4.4.14/fs/ceph/
Dcache.h69 static inline int ceph_release_fscache_page(struct page *page, gfp_t gfp) in ceph_release_fscache_page() argument
73 return fscache_maybe_release_page(ci->fscache, page, gfp); in ceph_release_fscache_page()
161 static inline int ceph_release_fscache_page(struct page *page, gfp_t gfp) in ceph_release_fscache_page() argument
/linux-4.4.14/arch/sh/mm/
Dconsistent.c36 dma_addr_t *dma_handle, gfp_t gfp, in dma_generic_alloc_coherent() argument
42 gfp |= __GFP_ZERO; in dma_generic_alloc_coherent()
44 ret = (void *)__get_free_pages(gfp, order); in dma_generic_alloc_coherent()
/linux-4.4.14/security/selinux/
Dxfrm.c79 gfp_t gfp) in selinux_xfrm_alloc_user() argument
95 ctx = kmalloc(sizeof(*ctx) + str_len + 1, gfp); in selinux_xfrm_alloc_user()
104 rc = security_context_to_sid(ctx->ctx_str, str_len, &ctx->ctx_sid, gfp); in selinux_xfrm_alloc_user()
284 gfp_t gfp) in selinux_xfrm_policy_alloc() argument
286 return selinux_xfrm_alloc_user(ctxp, uctx, gfp); in selinux_xfrm_policy_alloc()
/linux-4.4.14/arch/alpha/kernel/
Dpci-noop.c111 dma_addr_t *dma_handle, gfp_t gfp, in alpha_noop_alloc_coherent() argument
117 gfp &= ~GFP_DMA; in alpha_noop_alloc_coherent()
118 ret = (void *)__get_free_pages(gfp, get_order(size)); in alpha_noop_alloc_coherent()
Dpci_iommu.c435 dma_addr_t *dma_addrp, gfp_t gfp, in alpha_pci_alloc_coherent() argument
442 gfp &= ~GFP_DMA; in alpha_pci_alloc_coherent()
445 cpu_addr = (void *)__get_free_pages(gfp, order); in alpha_pci_alloc_coherent()
459 if (alpha_mv.mv_pci_tbi || (gfp & GFP_DMA)) in alpha_pci_alloc_coherent()
463 gfp |= GFP_DMA; in alpha_pci_alloc_coherent()
/linux-4.4.14/drivers/misc/mic/scif/
Dscif_map.h26 gfp_t gfp) in scif_alloc_coherent() argument
31 va = kmalloc(size, gfp); in scif_alloc_coherent()
36 size, dma_handle, gfp); in scif_alloc_coherent()
/linux-4.4.14/kernel/irq/
Dirqdesc.c39 static int alloc_masks(struct irq_desc *desc, gfp_t gfp, int node) in alloc_masks() argument
42 gfp, node)) in alloc_masks()
46 if (!zalloc_cpumask_var_node(&desc->pending_mask, gfp, node)) { in alloc_masks()
67 alloc_masks(struct irq_desc *desc, gfp_t gfp, int node) { return 0; } in alloc_masks() argument
147 gfp_t gfp = GFP_KERNEL; in alloc_desc() local
149 desc = kzalloc_node(sizeof(*desc), gfp, node); in alloc_desc()
157 if (alloc_masks(desc, gfp, node)) in alloc_desc()
/linux-4.4.14/include/crypto/
Dskcipher.h110 struct crypto_ablkcipher *tfm, gfp_t gfp) in skcipher_givcrypt_alloc() argument
115 crypto_ablkcipher_reqsize(tfm), gfp); in skcipher_givcrypt_alloc()
421 struct crypto_skcipher *tfm, gfp_t gfp) in skcipher_request_alloc() argument
426 crypto_skcipher_reqsize(tfm), gfp); in skcipher_request_alloc()
Daead.h411 gfp_t gfp) in aead_request_alloc() argument
415 req = kmalloc(sizeof(*req) + crypto_aead_reqsize(tfm), gfp); in aead_request_alloc()
Dakcipher.h192 struct crypto_akcipher *tfm, gfp_t gfp) in akcipher_request_alloc() argument
196 req = kmalloc(sizeof(*req) + crypto_akcipher_reqsize(tfm), gfp); in akcipher_request_alloc()
Dhash.h531 struct crypto_ahash *tfm, gfp_t gfp) in ahash_request_alloc() argument
536 crypto_ahash_reqsize(tfm), gfp); in ahash_request_alloc()
/linux-4.4.14/drivers/iommu/
Ddma-iommu.c193 static struct page **__iommu_dma_alloc_pages(unsigned int count, gfp_t gfp) in __iommu_dma_alloc_pages() argument
207 gfp |= __GFP_NOWARN | __GFP_HIGHMEM; in __iommu_dma_alloc_pages()
220 page = alloc_pages(gfp | __GFP_NORETRY, order); in __iommu_dma_alloc_pages()
233 page = alloc_page(gfp); in __iommu_dma_alloc_pages()
282 gfp_t gfp, int prot, dma_addr_t *handle, in iommu_dma_alloc() argument
295 pages = __iommu_dma_alloc_pages(count, gfp); in iommu_dma_alloc()
/linux-4.4.14/arch/frv/mb93090-mb00/
Dpci-dma.c21 void *dma_alloc_coherent(struct device *hwdev, size_t size, dma_addr_t *dma_handle, gfp_t gfp) in dma_alloc_coherent() argument
25 ret = consistent_alloc(gfp, size, dma_handle); in dma_alloc_coherent()
/linux-4.4.14/fs/
Dmpage.c143 gfp_t gfp) in do_mpage_readpage() argument
281 min_t(int, nr_pages, BIO_MAX_PAGES), gfp); in do_mpage_readpage()
364 gfp_t gfp = mapping_gfp_constraint(mapping, GFP_KERNEL); in mpage_readpages() local
375 gfp)) { in mpage_readpages()
380 get_block, gfp); in mpage_readpages()
400 gfp_t gfp = mapping_gfp_constraint(page->mapping, GFP_KERNEL); in mpage_readpage() local
405 &map_bh, &first_logical_block, get_block, gfp); in mpage_readpage()
Dposix_acl.c483 __posix_acl_create(struct posix_acl **acl, gfp_t gfp, umode_t *mode_p) in __posix_acl_create() argument
485 struct posix_acl *clone = posix_acl_clone(*acl, gfp); in __posix_acl_create()
501 __posix_acl_chmod(struct posix_acl **acl, gfp_t gfp, umode_t mode) in __posix_acl_chmod() argument
503 struct posix_acl *clone = posix_acl_clone(*acl, gfp); in __posix_acl_chmod()
/linux-4.4.14/drivers/net/ethernet/mellanox/mlx4/
Dalloc.c587 struct mlx4_buf *buf, gfp_t gfp) in mlx4_buf_alloc() argument
596 size, &t, gfp); in mlx4_buf_alloc()
616 gfp); in mlx4_buf_alloc()
624 &t, gfp); in mlx4_buf_alloc()
635 pages = kmalloc(sizeof *pages * buf->nbufs, gfp); in mlx4_buf_alloc()
680 gfp_t gfp) in mlx4_alloc_db_pgdir() argument
684 pgdir = kzalloc(sizeof *pgdir, gfp); in mlx4_alloc_db_pgdir()
692 &pgdir->db_dma, gfp); in mlx4_alloc_db_pgdir()
732 int mlx4_db_alloc(struct mlx4_dev *dev, struct mlx4_db *db, int order, gfp_t gfp) in mlx4_db_alloc() argument
744 pgdir = mlx4_alloc_db_pgdir(&dev->persist->pdev->dev, gfp); in mlx4_db_alloc()
Dqp.c298 int __mlx4_qp_alloc_icm(struct mlx4_dev *dev, int qpn, gfp_t gfp) in __mlx4_qp_alloc_icm() argument
304 err = mlx4_table_get(dev, &qp_table->qp_table, qpn, gfp); in __mlx4_qp_alloc_icm()
308 err = mlx4_table_get(dev, &qp_table->auxc_table, qpn, gfp); in __mlx4_qp_alloc_icm()
312 err = mlx4_table_get(dev, &qp_table->altc_table, qpn, gfp); in __mlx4_qp_alloc_icm()
316 err = mlx4_table_get(dev, &qp_table->rdmarc_table, qpn, gfp); in __mlx4_qp_alloc_icm()
320 err = mlx4_table_get(dev, &qp_table->cmpt_table, qpn, gfp); in __mlx4_qp_alloc_icm()
342 static int mlx4_qp_alloc_icm(struct mlx4_dev *dev, int qpn, gfp_t gfp) in mlx4_qp_alloc_icm() argument
352 return __mlx4_qp_alloc_icm(dev, qpn, gfp); in mlx4_qp_alloc_icm()
381 int mlx4_qp_alloc(struct mlx4_dev *dev, int qpn, struct mlx4_qp *qp, gfp_t gfp) in mlx4_qp_alloc() argument
392 err = mlx4_qp_alloc_icm(dev, qpn, gfp); in mlx4_qp_alloc()
Dmr.c485 int __mlx4_mpt_alloc_icm(struct mlx4_dev *dev, u32 index, gfp_t gfp) in __mlx4_mpt_alloc_icm() argument
489 return mlx4_table_get(dev, &mr_table->dmpt_table, index, gfp); in __mlx4_mpt_alloc_icm()
492 static int mlx4_mpt_alloc_icm(struct mlx4_dev *dev, u32 index, gfp_t gfp) in mlx4_mpt_alloc_icm() argument
503 return __mlx4_mpt_alloc_icm(dev, index, gfp); in mlx4_mpt_alloc_icm()
793 struct mlx4_buf *buf, gfp_t gfp) in mlx4_buf_write_mtt() argument
800 gfp); in mlx4_buf_write_mtt()
Den_rx.c61 gfp_t gfp = _gfp; in mlx4_alloc_pages() local
64 gfp |= __GFP_COMP | __GFP_NOWARN; in mlx4_alloc_pages()
65 page = alloc_pages(gfp, order); in mlx4_alloc_pages()
94 gfp_t gfp) in mlx4_en_alloc_frags() argument
111 if (mlx4_alloc_pages(priv, &page_alloc[i], frag_info, gfp)) in mlx4_en_alloc_frags()
238 gfp_t gfp) in mlx4_en_prepare_rx_desc() argument
244 return mlx4_en_alloc_frags(priv, rx_desc, frags, ring->page_alloc, gfp); in mlx4_en_prepare_rx_desc()
Dicm.h75 gfp_t gfp);
/linux-4.4.14/security/apparmor/include/
Dcontext.h44 static inline struct aa_file_cxt *aa_alloc_file_context(gfp_t gfp) in aa_alloc_file_context() argument
46 return kzalloc(sizeof(struct aa_file_cxt), gfp); in aa_alloc_file_context()
Daudit.h136 int aa_audit(int type, struct aa_profile *profile, gfp_t gfp,
/linux-4.4.14/fs/ramfs/
Dfile-nommu.c73 gfp_t gfp = mapping_gfp_mask(inode->i_mapping); in ramfs_nommu_expand_for_mapping() local
88 pages = alloc_pages(gfp, order); in ramfs_nommu_expand_for_mapping()
112 gfp); in ramfs_nommu_expand_for_mapping()
/linux-4.4.14/arch/frv/mm/
Ddma-alloc.c81 void *consistent_alloc(gfp_t gfp, size_t size, dma_addr_t *dma_handle) in consistent_alloc() argument
95 page = __get_free_pages(gfp, order); in consistent_alloc()
/linux-4.4.14/fs/cifs/
Dfscache.h51 extern int cifs_fscache_release_page(struct page *page, gfp_t gfp);
118 static inline int cifs_fscache_release_page(struct page *page, gfp_t gfp) in cifs_fscache_release_page() argument
Dfscache.c128 int cifs_fscache_release_page(struct page *page, gfp_t gfp) in cifs_fscache_release_page() argument
136 if (!fscache_maybe_release_page(cifsi->fscache, page, gfp)) in cifs_fscache_release_page()
/linux-4.4.14/drivers/staging/unisys/visorbus/
Dvisorchannel.c64 gfp_t gfp, unsigned long off, in visorchannel_create_guts() argument
74 channel = kzalloc(sizeof(*channel), gfp); in visorchannel_create_guts()
149 gfp_t gfp, uuid_le guid) in visorchannel_create() argument
151 return visorchannel_create_guts(physaddr, channel_bytes, gfp, 0, guid, in visorchannel_create()
158 gfp_t gfp, uuid_le guid) in visorchannel_create_with_lock() argument
160 return visorchannel_create_guts(physaddr, channel_bytes, gfp, 0, guid, in visorchannel_create_with_lock()
/linux-4.4.14/arch/powerpc/include/asm/
Ddma-mapping.h47 dma_addr_t *handle, gfp_t gfp);
59 #define __dma_alloc_coherent(dev, gfp, size, handle) NULL argument
/linux-4.4.14/net/mac80211/
Drate.h80 struct sta_info *sta, gfp_t gfp) in rate_control_alloc_sta() argument
83 return ref->ops->alloc_sta(ref->priv, &sta->sta, gfp); in rate_control_alloc_sta()
Dpm.c201 gfp_t gfp) in ieee80211_report_wowlan_wakeup() argument
205 cfg80211_report_wowlan_wakeup(&sdata->wdev, wakeup, gfp); in ieee80211_report_wowlan_wakeup()
Drc80211_minstrel.c569 minstrel_alloc_sta(void *priv, struct ieee80211_sta *sta, gfp_t gfp) in minstrel_alloc_sta() argument
578 mi = kzalloc(sizeof(struct minstrel_sta_info), gfp); in minstrel_alloc_sta()
588 mi->r = kzalloc(sizeof(struct minstrel_rate) * max_rates, gfp); in minstrel_alloc_sta()
592 mi->sample_table = kmalloc(SAMPLE_COLUMNS * max_rates, gfp); in minstrel_alloc_sta()
/linux-4.4.14/Documentation/fault-injection/
Dfault-injection.txt102 - /sys/kernel/debug/fail_page_alloc/ignore-gfp-highmem:
108 - /sys/kernel/debug/failslab/ignore-gfp-wait:
109 - /sys/kernel/debug/fail_page_alloc/ignore-gfp-wait:
191 echo 1 > /sys/kernel/debug/$FAILTYPE/ignore-gfp-wait
245 echo 1 > /sys/kernel/debug/$FAILTYPE/ignore-gfp-wait
246 echo 1 > /sys/kernel/debug/$FAILTYPE/ignore-gfp-highmem
/linux-4.4.14/drivers/infiniband/hw/qib/
Dqib_qp.c104 gfp_t gfp) in get_map_page() argument
106 unsigned long page = get_zeroed_page(gfp); in get_map_page()
125 enum ib_qp_type type, u8 port, gfp_t gfp) in alloc_qpn() argument
155 get_map_page(qpt, map, gfp); in alloc_qpn()
987 gfp_t gfp; in qib_create_qp() local
1000 gfp = init_attr->create_flags & IB_QP_CREATE_USE_GFP_NOIO ? in qib_create_qp()
1034 gfp, PAGE_KERNEL); in qib_create_qp()
1050 qp = kzalloc(sz + sg_list_sz, gfp); in qib_create_qp()
1056 qp->s_hdr = kzalloc(sizeof(*qp->s_hdr), gfp); in qib_create_qp()
1071 if (gfp != GFP_NOIO) in qib_create_qp()
[all …]
/linux-4.4.14/arch/microblaze/mm/
Dconsistent.c62 void *consistent_alloc(gfp_t gfp, size_t size, dma_addr_t *dma_handle) in consistent_alloc() argument
82 vaddr = __get_free_pages(gfp, order); in consistent_alloc()
/linux-4.4.14/drivers/crypto/ccp/
Dccp-crypto-main.c272 gfp_t gfp; in ccp_crypto_enqueue_request() local
274 gfp = req->flags & CRYPTO_TFM_REQ_MAY_SLEEP ? GFP_KERNEL : GFP_ATOMIC; in ccp_crypto_enqueue_request()
276 crypto_cmd = kzalloc(sizeof(*crypto_cmd), gfp); in ccp_crypto_enqueue_request()
Dccp-crypto-aes-cmac.c68 gfp_t gfp; in ccp_do_cmac_update() local
111 gfp = req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP ? in ccp_do_cmac_update()
113 ret = sg_alloc_table(&rctx->data_sg, sg_count, gfp); in ccp_do_cmac_update()
/linux-4.4.14/include/linux/ceph/
Dbuffer.h21 extern struct ceph_buffer *ceph_buffer_new(size_t len, gfp_t gfp);
Ddecode.h106 size_t *lenp, gfp_t gfp) in ceph_extract_encoded_string() argument
116 buf = kmalloc(len + 1, gfp); in ceph_extract_encoded_string()
/linux-4.4.14/drivers/net/
Dvirtio_net.c536 gfp_t gfp) in add_recvbuf_small() argument
542 skb = __netdev_alloc_skb_ip_align(vi->dev, GOOD_PACKET_LEN, gfp); in add_recvbuf_small()
553 err = virtqueue_add_inbuf(rq->vq, rq->sg, 2, skb, gfp); in add_recvbuf_small()
561 gfp_t gfp) in add_recvbuf_big() argument
571 first = get_a_page(rq, gfp); in add_recvbuf_big()
584 first = get_a_page(rq, gfp); in add_recvbuf_big()
602 first, gfp); in add_recvbuf_big()
619 static int add_recvbuf_mergeable(struct receive_queue *rq, gfp_t gfp) in add_recvbuf_mergeable() argument
628 if (unlikely(!skb_page_frag_refill(len, alloc_frag, gfp))) in add_recvbuf_mergeable()
647 err = virtqueue_add_inbuf(rq->vq, rq->sg, 1, (void *)ctx, gfp); in add_recvbuf_mergeable()
[all …]
/linux-4.4.14/drivers/staging/unisys/include/
Dvisorbus.h189 gfp_t gfp, uuid_le guid);
192 gfp_t gfp, uuid_le guid);
/linux-4.4.14/arch/openrisc/kernel/
Ddma.c85 dma_addr_t *dma_handle, gfp_t gfp, in or1k_dma_alloc() argument
95 page = alloc_pages_exact(size, gfp); in or1k_dma_alloc()
/linux-4.4.14/arch/frv/include/asm/
Dpci.h28 extern void *consistent_alloc(gfp_t gfp, size_t size, dma_addr_t *dma_handle);
/linux-4.4.14/arch/x86/kernel/cpu/
Dperf_event_intel_pt.c344 static struct topa *topa_alloc(int cpu, gfp_t gfp) in topa_alloc() argument
350 p = alloc_pages_node(node, gfp | __GFP_ZERO, 0); in topa_alloc()
434 static int topa_insert_pages(struct pt_buffer *buf, gfp_t gfp) in topa_insert_pages() argument
445 topa = topa_alloc(buf->cpu, gfp); in topa_insert_pages()
821 gfp_t gfp) in pt_buffer_init_topa() argument
826 topa = topa_alloc(buf->cpu, gfp); in pt_buffer_init_topa()
833 err = topa_insert_pages(buf, gfp); in pt_buffer_init_topa()
/linux-4.4.14/drivers/vhost/
Dvringh.c181 static int resize_iovec(struct vringh_kiov *iov, gfp_t gfp) in resize_iovec() argument
191 new = krealloc(iov->iov, new_num * sizeof(struct iovec), gfp); in resize_iovec()
193 new = kmalloc(new_num * sizeof(struct iovec), gfp); in resize_iovec()
262 gfp_t gfp, in __vringh_iov() argument
358 err = resize_iovec(iov, gfp); in __vringh_iov()
912 gfp_t gfp) in vringh_getdesc_kern() argument
926 gfp, copydesc_kern); in vringh_getdesc_kern()
/linux-4.4.14/arch/microblaze/include/asm/
Ddma-mapping.h34 #define __dma_alloc_coherent(dev, gfp, size, handle) NULL argument
/linux-4.4.14/arch/avr32/include/asm/
Ddma-mapping.h55 dma_addr_t *handle, gfp_t gfp);
85 dma_addr_t *handle, gfp_t gfp);
/linux-4.4.14/arch/sparc/kernel/
Dioport.c262 dma_addr_t *dma_addrp, gfp_t gfp, in sbus_alloc_coherent() argument
281 va = __get_free_pages(gfp, order); in sbus_alloc_coherent()
431 dma_addr_t *pba, gfp_t gfp, in pci32_alloc_coherent() argument
447 va = (void *) __get_free_pages(gfp, order); in pci32_alloc_coherent()
/linux-4.4.14/include/net/iucv/
Diucv.h223 static inline struct iucv_path *iucv_path_alloc(u16 msglim, u8 flags, gfp_t gfp) in iucv_path_alloc() argument
227 path = kzalloc(sizeof(struct iucv_path), gfp); in iucv_path_alloc()
/linux-4.4.14/arch/arc/include/asm/
Ddma-mapping.h18 dma_addr_t *dma_handle, gfp_t gfp);
24 dma_addr_t *dma_handle, gfp_t gfp);
/linux-4.4.14/include/scsi/
Dscsi_dh.h92 gfp_t gfp) in scsi_dh_attached_handler_name() argument
/linux-4.4.14/security/selinux/include/
Dsecurity.h167 u32 *out_sid, gfp_t gfp);
169 int security_context_str_to_sid(const char *scontext, u32 *out_sid, gfp_t gfp);
Dxfrm.h14 gfp_t gfp);
/linux-4.4.14/drivers/staging/lustre/lustre/libcfs/
Dtracefile.c68 static struct cfs_trace_page *cfs_tage_alloc(gfp_t gfp) in cfs_tage_alloc() argument
81 gfp |= __GFP_NOWARN; in cfs_tage_alloc()
82 page = alloc_page(gfp); in cfs_tage_alloc()
86 tage = kmalloc(sizeof(*tage), gfp); in cfs_tage_alloc()
116 int cfs_trace_refill_stock(struct cfs_trace_cpu_data *tcd, gfp_t gfp, in cfs_trace_refill_stock() argument
129 tage = cfs_tage_alloc(gfp); in cfs_trace_refill_stock()
/linux-4.4.14/arch/arm64/mm/
Ddma-mapping.c542 dma_addr_t *handle, gfp_t gfp, in __iommu_alloc_attrs() argument
559 gfp |= __GFP_ZERO; in __iommu_alloc_attrs()
561 if (gfpflags_allow_blocking(gfp)) { in __iommu_alloc_attrs()
565 pages = iommu_dma_alloc(dev, iosize, gfp, ioprot, handle, in __iommu_alloc_attrs()
582 page = alloc_pages(gfp, get_order(size)); in __iommu_alloc_attrs()
585 addr = __alloc_from_pool(size, &page, gfp); in __iommu_alloc_attrs()
/linux-4.4.14/drivers/net/wireless/mediatek/mt7601u/
Dusb.h68 struct mt7601u_dma_buf *buf, gfp_t gfp,
/linux-4.4.14/drivers/scsi/cxgbi/
Dlibcxgbi.h420 static inline struct sk_buff *alloc_wr(int wrlen, int dlen, gfp_t gfp) in alloc_wr() argument
422 struct sk_buff *skb = alloc_skb(wrlen + dlen, gfp); in alloc_wr()
676 gfp_t gfp) in cxgbi_alloc_big_mem() argument
678 void *p = kzalloc(size, gfp | __GFP_NOWARN); in cxgbi_alloc_big_mem()
/linux-4.4.14/drivers/md/
Ddm-bio-prison.h60 gfp_t gfp);
/linux-4.4.14/drivers/infiniband/hw/cxgb3/
Diwch_cm.c109 static struct sk_buff *get_skb(struct sk_buff *skb, int len, gfp_t gfp);
266 static void *alloc_ep(int size, gfp_t gfp) in alloc_ep() argument
270 epc = kzalloc(size, gfp); in alloc_ep()
324 static struct sk_buff *get_skb(struct sk_buff *skb, int len, gfp_t gfp) in get_skb() argument
330 skb = alloc_skb(len, gfp); in get_skb()
387 static int send_halfclose(struct iwch_ep *ep, gfp_t gfp) in send_halfclose() argument
393 skb = get_skb(NULL, sizeof(*req), gfp); in send_halfclose()
407 static int send_abort(struct iwch_ep *ep, struct sk_buff *skb, gfp_t gfp) in send_abort() argument
412 skb = get_skb(skb, sizeof(*req), gfp); in send_abort()
661 static void abort_connection(struct iwch_ep *ep, struct sk_buff *skb, gfp_t gfp) in abort_connection() argument
[all …]
/linux-4.4.14/drivers/crypto/qce/
Dablkcipher.c70 gfp_t gfp; in qce_ablkcipher_async_req_handle() local
89 gfp = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ? in qce_ablkcipher_async_req_handle()
92 ret = sg_alloc_table(&rctx->dst_tbl, rctx->dst_nents, gfp); in qce_ablkcipher_async_req_handle()
/linux-4.4.14/drivers/net/wireless/realtek/rtlwifi/
Drc.c258 struct ieee80211_sta *sta, gfp_t gfp) in rtl_rate_alloc_sta() argument
263 rate_priv = kzalloc(sizeof(struct rtl_rate_priv), gfp); in rtl_rate_alloc_sta()
/linux-4.4.14/drivers/net/wireless/ath/carl9170/
Dusb.c314 static int carl9170_usb_submit_rx_urb(struct ar9170 *ar, gfp_t gfp) in carl9170_usb_submit_rx_urb() argument
325 err = usb_submit_urb(urb, gfp); in carl9170_usb_submit_rx_urb()
457 static struct urb *carl9170_usb_alloc_rx_urb(struct ar9170 *ar, gfp_t gfp) in carl9170_usb_alloc_rx_urb() argument
462 buf = kmalloc(ar->fw.rx_size, gfp); in carl9170_usb_alloc_rx_urb()
466 urb = usb_alloc_urb(0, gfp); in carl9170_usb_alloc_rx_urb()
/linux-4.4.14/drivers/infiniband/core/
Diwpm_util.h108 u8 nl_client, gfp_t gfp);
/linux-4.4.14/drivers/power/
Dpower_supply_sysfs.c256 static char *kstruprdup(const char *str, gfp_t gfp) in kstruprdup() argument
260 ustr = ret = kmalloc(strlen(str) + 1, gfp); in kstruprdup()
/linux-4.4.14/drivers/scsi/
Dscsi_dh.c427 const char *scsi_dh_attached_handler_name(struct request_queue *q, gfp_t gfp) in scsi_dh_attached_handler_name() argument
437 handler_name = kstrdup(sdev->handler->name, gfp); in scsi_dh_attached_handler_name()
/linux-4.4.14/arch/c6x/mm/
Ddma-coherent.c77 dma_addr_t *handle, gfp_t gfp) in dma_alloc_coherent() argument
/linux-4.4.14/drivers/mfd/
Ddln2.c306 int *obuf_len, gfp_t gfp) in dln2_prep_buf() argument
313 buf = kmalloc(len, gfp); in dln2_prep_buf()
633 static int dln2_start_rx_urbs(struct dln2_dev *dln2, gfp_t gfp) in dln2_start_rx_urbs() argument
640 ret = usb_submit_urb(dln2->rx_urb[i], gfp); in dln2_start_rx_urbs()
/linux-4.4.14/security/apparmor/
Dlsm.c58 static int apparmor_cred_alloc_blank(struct cred *cred, gfp_t gfp) in apparmor_cred_alloc_blank() argument
61 struct aa_task_cxt *cxt = aa_alloc_task_context(gfp); in apparmor_cred_alloc_blank()
73 gfp_t gfp) in apparmor_cred_prepare() argument
76 struct aa_task_cxt *cxt = aa_alloc_task_context(gfp); in apparmor_cred_prepare()
Daudit.c172 int aa_audit(int type, struct aa_profile *profile, gfp_t gfp, in aa_audit() argument
/linux-4.4.14/security/tomoyo/
Dtomoyo.c18 static int tomoyo_cred_alloc_blank(struct cred *new, gfp_t gfp) in tomoyo_cred_alloc_blank() argument
34 gfp_t gfp) in tomoyo_cred_prepare() argument

12