Home
last modified time | relevance | path

Searched refs:gfp (Results 1 – 200 of 358) sorted by relevance

12

/linux-4.1.27/net/wireless/
Dnl80211.h37 const u8 *buf, size_t len, gfp_t gfp);
40 const u8 *buf, size_t len, gfp_t gfp,
44 const u8 *buf, size_t len, gfp_t gfp);
47 const u8 *buf, size_t len, gfp_t gfp);
50 const u8 *addr, gfp_t gfp);
53 const u8 *addr, gfp_t gfp);
58 u16 status, gfp_t gfp);
62 const u8 *resp_ie, size_t resp_ie_len, gfp_t gfp);
71 int key_id, const u8 *tsc, gfp_t gfp);
80 gfp_t gfp);
[all …]
Dnl80211.c7417 gfp_t gfp) in __cfg80211_alloc_vendor_skb() argument
7423 skb = nlmsg_new(approxlen + 100, gfp); in __cfg80211_alloc_vendor_skb()
7473 int approxlen, gfp_t gfp) in __cfg80211_alloc_event_skb() argument
7496 cmd, attr, info, gfp); in __cfg80211_alloc_event_skb()
7500 void __cfg80211_send_event_skb(struct sk_buff *skb, gfp_t gfp) in __cfg80211_send_event_skb() argument
7517 mcgrp, gfp); in __cfg80211_send_event_skb()
11329 enum nl80211_commands cmd, gfp_t gfp, in nl80211_send_mlme_event() argument
11335 msg = nlmsg_new(NLMSG_DEFAULT_SIZE, gfp); in nl80211_send_mlme_event()
11366 NL80211_MCGRP_MLME, gfp); in nl80211_send_mlme_event()
11376 size_t len, gfp_t gfp) in nl80211_send_rx_auth() argument
[all …]
Dmlme.c172 const u8 *tsc, gfp_t gfp) in cfg80211_michael_mic_failure() argument
178 char *buf = kmalloc(128, gfp); in cfg80211_michael_mic_failure()
193 nl80211_michael_mic_failure(rdev, dev, addr, key_type, key_id, tsc, gfp); in cfg80211_michael_mic_failure()
730 gfp_t gfp) in cfg80211_radar_event() argument
747 nl80211_radar_notify(rdev, chandef, NL80211_RADAR_DETECTED, NULL, gfp); in cfg80211_radar_event()
753 enum nl80211_radar_event event, gfp_t gfp) in cfg80211_cac_event() argument
783 nl80211_radar_notify(rdev, chandef, event, netdev, gfp); in cfg80211_cac_event()
Dsme.c747 u16 status, gfp_t gfp) in cfg80211_connect_result() argument
754 ev = kzalloc(sizeof(*ev) + req_ie_len + resp_ie_len, gfp); in cfg80211_connect_result()
842 const u8 *resp_ie, size_t resp_ie_len, gfp_t gfp) in cfg80211_roamed() argument
854 resp_ie_len, gfp); in cfg80211_roamed()
862 size_t resp_ie_len, gfp_t gfp) in cfg80211_roamed_bss() argument
872 ev = kzalloc(sizeof(*ev) + req_ie_len + resp_ie_len, gfp); in cfg80211_roamed_bss()
941 const u8 *ie, size_t ie_len, gfp_t gfp) in cfg80211_disconnected() argument
948 ev = kzalloc(sizeof(*ev) + ie_len, gfp); in cfg80211_disconnected()
Dreg.h83 gfp_t gfp);
Dscan.c948 s32 signal, gfp_t gfp) in cfg80211_inform_bss_width() argument
981 ies = kzalloc(sizeof(*ies) + ielen, gfp); in cfg80211_inform_bss_width()
1012 regulatory_hint_found_beacon(wiphy, channel, gfp); in cfg80211_inform_bss_width()
1015 regulatory_hint_found_beacon(wiphy, channel, gfp); in cfg80211_inform_bss_width()
1030 s32 signal, gfp_t gfp) in cfg80211_inform_bss_width_frame() argument
1064 ies = kzalloc(sizeof(*ies) + ielen, gfp); in cfg80211_inform_bss_width_frame()
1095 regulatory_hint_found_beacon(wiphy, channel, gfp); in cfg80211_inform_bss_width_frame()
1098 regulatory_hint_found_beacon(wiphy, channel, gfp); in cfg80211_inform_bss_width_frame()
/linux-4.1.27/include/linux/
Dkmemleak.h30 gfp_t gfp) __ref;
32 gfp_t gfp) __ref;
39 extern void kmemleak_scan_area(const void *ptr, size_t size, gfp_t gfp) __ref;
44 gfp_t gfp) in kmemleak_alloc_recursive() argument
47 kmemleak_alloc(ptr, size, min_count, gfp); in kmemleak_alloc_recursive()
67 gfp_t gfp) in kmemleak_alloc() argument
72 gfp_t gfp) in kmemleak_alloc_recursive() argument
76 gfp_t gfp) in kmemleak_alloc_percpu() argument
100 static inline void kmemleak_scan_area(const void *ptr, size_t size, gfp_t gfp) in kmemleak_scan_area() argument
Ddevcoredump.h10 gfp_t gfp);
13 const void *data, size_t datalen, gfp_t gfp,
19 size_t datalen, gfp_t gfp) in dev_coredumpv() argument
26 const void *data, size_t datalen, gfp_t gfp, in dev_coredumpm() argument
Dvmpressure.h29 extern void vmpressure(gfp_t gfp, struct mem_cgroup *memcg,
31 extern void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, int prio);
43 static inline void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, in vmpressure() argument
45 static inline void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, in vmpressure_prio() argument
Dfscache.h546 gfp_t gfp) in fscache_read_or_alloc_page() argument
550 context, gfp); in fscache_read_or_alloc_page()
597 gfp_t gfp) in fscache_read_or_alloc_pages() argument
602 context, gfp); in fscache_read_or_alloc_pages()
628 gfp_t gfp) in fscache_alloc_page() argument
631 return __fscache_alloc_page(cookie, page, gfp); in fscache_alloc_page()
677 gfp_t gfp) in fscache_write_page() argument
680 return __fscache_write_page(cookie, page, gfp); in fscache_write_page()
763 gfp_t gfp) in fscache_maybe_release_page() argument
766 return __fscache_maybe_release_page(cookie, page, gfp); in fscache_maybe_release_page()
Dmemcontrol.h431 bool __memcg_kmem_newpage_charge(gfp_t gfp, struct mem_cgroup **memcg,
444 int memcg_charge_kmem(struct mem_cgroup *memcg, gfp_t gfp,
461 memcg_kmem_newpage_charge(gfp_t gfp, struct mem_cgroup **memcg, int order) in memcg_kmem_newpage_charge() argument
466 if (gfp & __GFP_NOACCOUNT) in memcg_kmem_newpage_charge()
474 if (gfp & __GFP_NOFAIL) in memcg_kmem_newpage_charge()
483 return __memcg_kmem_newpage_charge(gfp, memcg, order); in memcg_kmem_newpage_charge()
523 memcg_kmem_get_cache(struct kmem_cache *cachep, gfp_t gfp) in memcg_kmem_get_cache() argument
527 if (gfp & __GFP_NOACCOUNT) in memcg_kmem_get_cache()
529 if (gfp & __GFP_NOFAIL) in memcg_kmem_get_cache()
566 memcg_kmem_newpage_charge(gfp_t gfp, struct mem_cgroup **memcg, int order) in memcg_kmem_newpage_charge() argument
[all …]
Dzpool.h40 gfp_t gfp, struct zpool_ops *ops);
46 int zpool_malloc(struct zpool *pool, size_t size, gfp_t gfp,
84 void *(*create)(char *name, gfp_t gfp, struct zpool_ops *ops);
87 int (*malloc)(void *pool, size_t size, gfp_t gfp,
Dstring.h117 extern char *kstrdup(const char *s, gfp_t gfp);
118 extern const char *kstrdup_const(const char *s, gfp_t gfp);
119 extern char *kstrndup(const char *s, size_t len, gfp_t gfp);
120 extern void *kmemdup(const void *src, size_t len, gfp_t gfp);
122 extern char **argv_split(gfp_t gfp, const char *str, int *argcp);
Dbtree-type.h31 gfp_t gfp) in BTREE_FN()
33 return btree_merge(&target->h, &victim->h, BTREE_TYPE_GEO, gfp); in BTREE_FN()
44 void *val, gfp_t gfp) in BTREE_FN()
47 return btree_insert(&head->h, BTREE_TYPE_GEO, &_key, val, gfp); in BTREE_FN()
87 void *val, gfp_t gfp) in BTREE_FN()
90 val, gfp); in BTREE_FN()
Dpercpu_counter.h30 int __percpu_counter_init(struct percpu_counter *fbc, s64 amount, gfp_t gfp,
33 #define percpu_counter_init(fbc, value, gfp) \ argument
37 __percpu_counter_init(fbc, value, gfp, &__key); \
99 gfp_t gfp) in percpu_counter_init() argument
Dbtree-128.h41 void *val, gfp_t gfp) in btree_insert128() argument
45 (unsigned long *)&key, val, gfp); in btree_insert128()
78 gfp_t gfp) in btree_merge128() argument
80 return btree_merge(&target->h, &victim->h, &btree_geo128, gfp); in btree_merge128()
Dpercpu.h126 extern void __percpu *__alloc_percpu_gfp(size_t size, size_t align, gfp_t gfp);
131 #define alloc_percpu_gfp(type, gfp) \ argument
133 __alignof__(type), gfp)
Dzbud.h12 struct zbud_pool *zbud_create_pool(gfp_t gfp, struct zbud_ops *ops);
14 int zbud_alloc(struct zbud_pool *pool, size_t size, gfp_t gfp,
Dvirtio.h40 gfp_t gfp);
45 gfp_t gfp);
52 gfp_t gfp);
Ddma-mapping.h19 dma_addr_t *dma_handle, gfp_t gfp,
232 dma_addr_t *dma_handle, gfp_t gfp);
236 dma_addr_t *dma_handle, gfp_t gfp);
248 size_t size, gfp_t gfp) in dmam_declare_coherent_memory() argument
275 dma_addr_t *dma_addr, gfp_t gfp) in dma_alloc_writecombine() argument
279 return dma_alloc_attrs(dev, size, dma_addr, gfp, &attrs); in dma_alloc_writecombine()
Dflex_array.h76 #define flex_array_put_ptr(fa, nr, src, gfp) \ argument
77 flex_array_put(fa, nr, (void *)&(src), gfp)
Dflex_proportions.h36 int fprop_global_init(struct fprop_global *p, gfp_t gfp);
83 int fprop_local_init_percpu(struct fprop_local_percpu *pl, gfp_t gfp);
Dfirmware.h46 const char *name, struct device *device, gfp_t gfp, void *context,
61 const char *name, struct device *device, gfp_t gfp, void *context, in request_firmware_nowait() argument
Dproportions.h44 int prop_descriptor_init(struct prop_descriptor *pd, int shift, gfp_t gfp);
65 int prop_local_init_percpu(struct prop_local_percpu *pl, gfp_t gfp);
Ddevice.h577 extern void *__devres_alloc(dr_release_t release, size_t size, gfp_t gfp,
579 #define devres_alloc(release, size, gfp) \ argument
580 __devres_alloc(release, size, gfp, #release)
582 extern void *devres_alloc(dr_release_t release, size_t size, gfp_t gfp);
603 gfp_t gfp);
609 extern void *devm_kmalloc(struct device *dev, size_t size, gfp_t gfp);
610 extern char *devm_kvasprintf(struct device *dev, gfp_t gfp, const char *fmt,
613 char *devm_kasprintf(struct device *dev, gfp_t gfp, const char *fmt, ...);
614 static inline void *devm_kzalloc(struct device *dev, size_t size, gfp_t gfp) in devm_kzalloc() argument
616 return devm_kmalloc(dev, size, gfp | __GFP_ZERO); in devm_kzalloc()
[all …]
Dbtree.h115 unsigned long *key, void *val, gfp_t gfp);
158 struct btree_geo *geo, gfp_t gfp);
Dbuffer_head.h179 unsigned size, gfp_t gfp);
184 sector_t block, unsigned size, gfp_t gfp);
322 sb_getblk_gfp(struct super_block *sb, sector_t block, gfp_t gfp) in sb_getblk_gfp() argument
324 return __getblk_gfp(sb->s_bdev, block, sb->s_blocksize, gfp); in sb_getblk_gfp()
Dvmstat.h188 extern void zone_statistics(struct zone *, struct zone *, gfp_t gfp);
193 #define zone_statistics(_zl, _z, gfp) do { } while (0) argument
Dpagemap.h215 extern struct page *__page_cache_alloc(gfp_t gfp);
217 static inline struct page *__page_cache_alloc(gfp_t gfp) in __page_cache_alloc() argument
219 return alloc_pages(gfp, 0); in __page_cache_alloc()
Dbio.h359 gfp_t gfp, struct bio_set *bs);
372 gfp_t gfp, struct bio_set *bs) in bio_next_split() argument
377 return bio_split(bio, sectors, gfp, bs); in bio_next_split()
459 extern int bio_alloc_pages(struct bio *bio, gfp_t gfp);
Dfscache-cache.h159 gfp_t gfp);
164 gfp_t gfp);
Dpercpu-refcount.h98 gfp_t gfp);
Dsecurity.h1597 int (*cred_alloc_blank) (struct cred *cred, gfp_t gfp);
1600 gfp_t gfp);
1723 struct xfrm_user_sec_ctx *sec_ctx, gfp_t gfp);
1877 int security_cred_alloc_blank(struct cred *cred, gfp_t gfp);
1879 int security_prepare_creds(struct cred *new, const struct cred *old, gfp_t gfp);
2402 static inline int security_cred_alloc_blank(struct cred *cred, gfp_t gfp) in security_cred_alloc_blank() argument
2412 gfp_t gfp) in security_prepare_creds() argument
2932 struct xfrm_user_sec_ctx *sec_ctx, gfp_t gfp);
2952 gfp_t gfp) in security_xfrm_policy_alloc() argument
/linux-4.1.27/include/net/sctp/
Dulpevent.h86 gfp_t gfp);
94 gfp_t gfp);
100 gfp_t gfp);
106 gfp_t gfp);
111 gfp_t gfp);
115 __u32 indication, gfp_t gfp);
118 const struct sctp_association *asoc, gfp_t gfp);
122 gfp_t gfp);
126 __u32 indication, gfp_t gfp);
129 const struct sctp_association *asoc, gfp_t gfp);
Dauth.h83 struct sctp_shared_key *sctp_auth_shkey_create(__u16 key_id, gfp_t gfp);
85 int sctp_auth_asoc_init_active_key(struct sctp_association *asoc, gfp_t gfp);
91 gfp_t gfp);
92 int sctp_auth_init_hmacs(struct sctp_endpoint *ep, gfp_t gfp);
104 struct sctp_auth_chunk *auth, gfp_t gfp);
Dsm.h182 gfp_t gfp);
189 gfp_t gfp, int vparam_len);
192 const gfp_t gfp,
273 gfp_t gfp);
285 gfp_t gfp, int *err,
Dstructs.h404 gfp_t gfp);
1096 sctp_scope_t scope, gfp_t gfp,
1100 gfp_t gfp);
1102 __u8 addr_state, gfp_t gfp);
1116 gfp_t gfp);
1118 __u16 port, gfp_t gfp);
1284 sctp_init_chunk_t *init, gfp_t gfp);
1867 sctp_scope_t scope, gfp_t gfp);
1881 const gfp_t gfp,
1912 gfp_t gfp);
Dtsnmap.h100 __u32 initial_tsn, gfp_t gfp);
/linux-4.1.27/mm/
Dslob.c190 static void *slob_new_pages(gfp_t gfp, int order, int node) in slob_new_pages() argument
196 page = alloc_pages_exact_node(node, gfp, order); in slob_new_pages()
199 page = alloc_pages(gfp, order); in slob_new_pages()
268 static void *slob_alloc(size_t size, gfp_t gfp, int align, int node) in slob_alloc() argument
316 b = slob_new_pages(gfp & ~__GFP_ZERO, 0, node); in slob_alloc()
332 if (unlikely((gfp & __GFP_ZERO) && b)) in slob_alloc()
427 __do_kmalloc_node(size_t size, gfp_t gfp, int node, unsigned long caller) in __do_kmalloc_node() argument
433 gfp &= gfp_allowed_mask; in __do_kmalloc_node()
435 lockdep_trace_alloc(gfp); in __do_kmalloc_node()
441 m = slob_alloc(size + align, gfp, align, node); in __do_kmalloc_node()
[all …]
Dzbud.c133 static void *zbud_zpool_create(char *name, gfp_t gfp, in zbud_zpool_create() argument
136 return zbud_create_pool(gfp, zpool_ops ? &zbud_zpool_ops : NULL); in zbud_zpool_create()
144 static int zbud_zpool_malloc(void *pool, size_t size, gfp_t gfp, in zbud_zpool_malloc() argument
147 return zbud_alloc(pool, size, gfp, handle); in zbud_zpool_malloc()
290 struct zbud_pool *zbud_create_pool(gfp_t gfp, struct zbud_ops *ops) in zbud_create_pool() argument
295 pool = kmalloc(sizeof(struct zbud_pool), gfp); in zbud_create_pool()
338 int zbud_alloc(struct zbud_pool *pool, size_t size, gfp_t gfp, in zbud_alloc() argument
346 if (!size || (gfp & __GFP_HIGHMEM)) in zbud_alloc()
370 page = alloc_page(gfp); in zbud_alloc()
Dutil.c44 char *kstrdup(const char *s, gfp_t gfp) in kstrdup() argument
53 buf = kmalloc_track_caller(len, gfp); in kstrdup()
69 const char *kstrdup_const(const char *s, gfp_t gfp) in kstrdup_const() argument
74 return kstrdup(s, gfp); in kstrdup_const()
84 char *kstrndup(const char *s, size_t max, gfp_t gfp) in kstrndup() argument
93 buf = kmalloc_track_caller(len+1, gfp); in kstrndup()
109 void *kmemdup(const void *src, size_t len, gfp_t gfp) in kmemdup() argument
113 p = kmalloc_track_caller(len, gfp); in kmemdup()
Dzpool.c144 struct zpool *zpool_create_pool(char *type, char *name, gfp_t gfp, in zpool_create_pool() argument
164 zpool = kmalloc(sizeof(*zpool), gfp); in zpool_create_pool()
173 zpool->pool = driver->create(name, gfp, ops); in zpool_create_pool()
246 int zpool_malloc(struct zpool *zpool, size_t size, gfp_t gfp, in zpool_malloc() argument
249 return zpool->driver->malloc(zpool->pool, size, gfp, handle); in zpool_malloc()
Dvmpressure.c211 void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, in vmpressure() argument
227 if (!(gfp & (__GFP_HIGHMEM | __GFP_MOVABLE | __GFP_IO | __GFP_FS))) in vmpressure()
263 void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, int prio) in vmpressure_prio() argument
279 vmpressure(gfp, memcg, vmpressure_win, 0); in vmpressure_prio()
Dshmem.c119 static bool shmem_should_replace_page(struct page *page, gfp_t gfp);
120 static int shmem_replace_page(struct page **pagep, gfp_t gfp,
123 struct page **pagep, enum sgp_type sgp, gfp_t gfp, int *fault_type);
618 gfp_t gfp; in shmem_unuse_inode() local
635 gfp = mapping_gfp_mask(mapping); in shmem_unuse_inode()
636 if (shmem_should_replace_page(*pagep, gfp)) { in shmem_unuse_inode()
638 error = shmem_replace_page(pagep, gfp, info, index); in shmem_unuse_inode()
874 static struct page *shmem_swapin(swp_entry_t swap, gfp_t gfp, in shmem_swapin() argument
887 page = swapin_readahead(swap, gfp, &pvma, 0); in shmem_swapin()
895 static struct page *shmem_alloc_page(gfp_t gfp, in shmem_alloc_page() argument
[all …]
Dmempolicy.c1638 static nodemask_t *policy_nodemask(gfp_t gfp, struct mempolicy *policy) in policy_nodemask() argument
1642 apply_policy_zone(policy, gfp_zone(gfp)) && in policy_nodemask()
1650 static struct zonelist *policy_zonelist(gfp_t gfp, struct mempolicy *policy, in policy_zonelist() argument
1665 if (unlikely(gfp & __GFP_THISNODE) && in policy_zonelist()
1672 return node_zonelist(nd, gfp); in policy_zonelist()
1925 static struct page *alloc_page_interleave(gfp_t gfp, unsigned order, in alloc_page_interleave() argument
1931 zl = node_zonelist(nid, gfp); in alloc_page_interleave()
1932 page = __alloc_pages(gfp, order, zl); in alloc_page_interleave()
1962 alloc_pages_vma(gfp_t gfp, int order, struct vm_area_struct *vma, in alloc_pages_vma() argument
1980 page = alloc_page_interleave(gfp, order, nid); in alloc_pages_vma()
[all …]
Dkmemleak.c118 #define gfp_kmemleak_mask(gfp) (((gfp) & (GFP_KERNEL | GFP_ATOMIC | \ argument
521 int min_count, gfp_t gfp) in create_object() argument
527 object = kmem_cache_alloc(object_cache, gfp_kmemleak_mask(gfp)); in create_object()
741 static void add_scan_area(unsigned long ptr, size_t size, gfp_t gfp) in add_scan_area() argument
754 area = kmem_cache_alloc(scan_area_cache, gfp_kmemleak_mask(gfp)); in add_scan_area()
897 gfp_t gfp) in kmemleak_alloc() argument
902 create_object((unsigned long)ptr, size, min_count, gfp); in kmemleak_alloc()
918 gfp_t gfp) in kmemleak_alloc_percpu() argument
931 size, 0, gfp); in kmemleak_alloc_percpu()
1080 void __ref kmemleak_scan_area(const void *ptr, size_t size, gfp_t gfp) in kmemleak_scan_area() argument
[all …]
Dslab.h234 gfp_t gfp, int order) in memcg_charge_slab() argument
240 return memcg_charge_kmem(s->memcg_params.memcg, gfp, 1 << order); in memcg_charge_slab()
288 static inline int memcg_charge_slab(struct kmem_cache *s, gfp_t gfp, int order) in memcg_charge_slab() argument
Dslab.c250 static int enable_cpucache(struct kmem_cache *cachep, gfp_t gfp);
679 int batchcount, gfp_t gfp) in alloc_arraycache() argument
684 ac = kmalloc_node(memsize, gfp, node); in alloc_arraycache()
834 int limit, gfp_t gfp) in alloc_alien_cache() argument
871 int batch, gfp_t gfp) in __alloc_alien_cache() argument
876 alc = kmalloc_node(memsize, gfp, node); in __alloc_alien_cache()
882 static struct alien_cache **alloc_alien_cache(int node, int limit, gfp_t gfp) in alloc_alien_cache() argument
890 alc_ptr = kzalloc_node(memsize, gfp, node); in alloc_alien_cache()
897 alc_ptr[i] = __alloc_alien_cache(node, limit, 0xbaadf00d, gfp); in alloc_alien_cache()
2022 static int __init_refok setup_cpu_cache(struct kmem_cache *cachep, gfp_t gfp) in setup_cpu_cache() argument
[all …]
Dzswap.c219 static struct zswap_entry *zswap_entry_cache_alloc(gfp_t gfp) in zswap_entry_cache_alloc() argument
222 entry = kmem_cache_alloc(zswap_entry_cache, gfp); in zswap_entry_cache_alloc()
902 gfp_t gfp = __GFP_NORETRY | __GFP_NOWARN; in init_zswap() local
909 zswap_pool = zpool_create_pool(zswap_zpool_type, "zswap", gfp, in init_zswap()
914 zswap_pool = zpool_create_pool(zswap_zpool_type, "zswap", gfp, in init_zswap()
Dpercpu-vm.c85 const gfp_t gfp = GFP_KERNEL | __GFP_HIGHMEM | __GFP_COLD; in pcpu_alloc_pages() local
93 *pagep = alloc_pages_node(cpu_to_node(cpu), gfp, 0); in pcpu_alloc_pages()
Dfilemap.c641 struct page *__page_cache_alloc(gfp_t gfp) in __page_cache_alloc() argument
651 page = alloc_pages_exact_node(n, gfp, 0); in __page_cache_alloc()
656 return alloc_pages(gfp, 0); in __page_cache_alloc()
2143 gfp_t gfp) in __read_cache_page() argument
2150 page = __page_cache_alloc(gfp | __GFP_COLD); in __read_cache_page()
2153 err = add_to_page_cache_lru(page, mapping, index, gfp); in __read_cache_page()
2176 gfp_t gfp) in do_read_cache_page() argument
2183 page = __read_cache_page(mapping, index, filler, data, gfp); in do_read_cache_page()
2247 gfp_t gfp) in read_cache_page_gfp() argument
2251 return do_read_cache_page(mapping, index, filler, NULL, gfp); in read_cache_page_gfp()
Dhuge_memory.c720 struct page *page, gfp_t gfp) in __do_huge_pmd_anonymous_page() argument
728 if (mem_cgroup_try_charge(page, mm, gfp, &memcg)) in __do_huge_pmd_anonymous_page()
793 gfp_t gfp; in do_huge_pmd_anonymous_page() local
828 gfp = alloc_hugepage_gfpmask(transparent_hugepage_defrag(vma), 0); in do_huge_pmd_anonymous_page()
829 page = alloc_hugepage_vma(gfp, vma, haddr, HPAGE_PMD_ORDER); in do_huge_pmd_anonymous_page()
834 if (unlikely(__do_huge_pmd_anonymous_page(mm, vma, haddr, pmd, page, gfp))) { in do_huge_pmd_anonymous_page()
2334 khugepaged_alloc_page(struct page **hpage, gfp_t gfp, struct mm_struct *mm, in khugepaged_alloc_page() argument
2348 *hpage = alloc_pages_exact_node(node, gfp, HPAGE_PMD_ORDER); in khugepaged_alloc_page()
2402 khugepaged_alloc_page(struct page **hpage, gfp_t gfp, struct mm_struct *mm, in khugepaged_alloc_page() argument
2442 gfp_t gfp; in collapse_huge_page() local
[all …]
/linux-4.1.27/arch/x86/include/asm/
Ddma-mapping.h104 gfp_t gfp) in dma_alloc_coherent_mask() argument
110 dma_mask = (gfp & GFP_DMA) ? DMA_BIT_MASK(24) : DMA_BIT_MASK(32); in dma_alloc_coherent_mask()
115 static inline gfp_t dma_alloc_coherent_gfp_flags(struct device *dev, gfp_t gfp) in dma_alloc_coherent_gfp_flags() argument
117 unsigned long dma_mask = dma_alloc_coherent_mask(dev, gfp); in dma_alloc_coherent_gfp_flags()
120 gfp |= GFP_DMA; in dma_alloc_coherent_gfp_flags()
122 if (dma_mask <= DMA_BIT_MASK(32) && !(gfp & GFP_DMA)) in dma_alloc_coherent_gfp_flags()
123 gfp |= GFP_DMA32; in dma_alloc_coherent_gfp_flags()
125 return gfp; in dma_alloc_coherent_gfp_flags()
132 gfp_t gfp, struct dma_attrs *attrs) in dma_alloc_attrs() argument
137 gfp &= ~(__GFP_DMA | __GFP_HIGHMEM | __GFP_DMA32); in dma_alloc_attrs()
[all …]
/linux-4.1.27/net/sctp/
Dauth.c73 static struct sctp_auth_bytes *sctp_auth_create_key(__u32 key_len, gfp_t gfp) in sctp_auth_create_key() argument
82 key = kmalloc(sizeof(struct sctp_auth_bytes) + key_len, gfp); in sctp_auth_create_key()
94 struct sctp_shared_key *sctp_auth_shkey_create(__u16 key_id, gfp_t gfp) in sctp_auth_shkey_create() argument
99 new = kzalloc(sizeof(struct sctp_shared_key), gfp); in sctp_auth_shkey_create()
191 gfp_t gfp) in sctp_auth_make_key_vector() argument
205 new = sctp_auth_create_key(len, gfp); in sctp_auth_make_key_vector()
226 gfp_t gfp) in sctp_auth_make_local_vector() argument
232 gfp); in sctp_auth_make_local_vector()
238 gfp_t gfp) in sctp_auth_make_peer_vector() argument
243 gfp); in sctp_auth_make_peer_vector()
[all …]
Dulpevent.c64 gfp_t gfp) in sctp_ulpevent_new() argument
69 skb = alloc_skb(size, gfp); in sctp_ulpevent_new()
129 __u16 inbound, struct sctp_chunk *chunk, gfp_t gfp) in sctp_ulpevent_make_assoc_change() argument
143 sizeof(struct sctp_assoc_change), 0, gfp); in sctp_ulpevent_make_assoc_change()
162 MSG_NOTIFICATION, gfp); in sctp_ulpevent_make_assoc_change()
258 int flags, int state, int error, gfp_t gfp) in sctp_ulpevent_make_peer_addr_change() argument
265 MSG_NOTIFICATION, gfp); in sctp_ulpevent_make_peer_addr_change()
372 gfp_t gfp) in sctp_ulpevent_make_remote_error() argument
391 skb = skb_copy_expand(chunk->skb, sizeof(*sre), 0, gfp); in sctp_ulpevent_make_remote_error()
428 __u16 flags, __u32 error, gfp_t gfp) in sctp_ulpevent_make_send_failed() argument
[all …]
Dbind_addr.c49 union sctp_addr *, sctp_scope_t scope, gfp_t gfp,
60 sctp_scope_t scope, gfp_t gfp, in sctp_bind_addr_copy() argument
72 gfp, flags); in sctp_bind_addr_copy()
84 SCTP_SCOPE_LINK, gfp, in sctp_bind_addr_copy()
105 gfp_t gfp) in sctp_bind_addr_dup() argument
114 error = sctp_add_bind_addr(dest, &addr->a, 1, gfp); in sctp_bind_addr_dup()
153 __u8 addr_state, gfp_t gfp) in sctp_add_bind_addr() argument
158 addr = kzalloc(sizeof(*addr), gfp); in sctp_add_bind_addr()
221 gfp_t gfp) in sctp_bind_addrs_to_raw() argument
248 retval.v = kmalloc(len, gfp); in sctp_bind_addrs_to_raw()
[all …]
Dendpointola.c59 gfp_t gfp) in sctp_endpoint_init() argument
67 ep->digest = kzalloc(SCTP_SIGNATURE_SIZE, gfp); in sctp_endpoint_init()
78 sizeof(__u16) * SCTP_AUTH_NUM_HMACS, gfp); in sctp_endpoint_init()
83 SCTP_NUM_CHUNK_TYPES, gfp); in sctp_endpoint_init()
151 null_key = sctp_auth_shkey_create(0, gfp); in sctp_endpoint_init()
158 err = sctp_auth_init_hmacs(ep, gfp); in sctp_endpoint_init()
184 struct sctp_endpoint *sctp_endpoint_new(struct sock *sk, gfp_t gfp) in sctp_endpoint_new() argument
189 ep = kzalloc(sizeof(*ep), gfp); in sctp_endpoint_new()
193 if (!sctp_endpoint_init(ep, sk, gfp)) in sctp_endpoint_new()
Dssnmap.c53 gfp_t gfp) in sctp_ssnmap_new() argument
60 retval = kmalloc(size, gfp); in sctp_ssnmap_new()
63 __get_free_pages(gfp, get_order(size)); in sctp_ssnmap_new()
Dassociola.c70 gfp_t gfp) in sctp_association_init() argument
266 err = sctp_auth_asoc_copy_shkeys(ep, asoc, gfp); in sctp_association_init()
298 gfp_t gfp) in sctp_association_new() argument
302 asoc = kzalloc(sizeof(*asoc), gfp); in sctp_association_new()
306 if (!sctp_association_init(asoc, ep, sk, scope, gfp)) in sctp_association_new()
576 const gfp_t gfp, in sctp_assoc_add_peer() argument
609 peer = sctp_transport_new(net, addr, gfp); in sctp_assoc_add_peer()
1542 sctp_scope_t scope, gfp_t gfp) in sctp_assoc_set_bind_addr_from_ep() argument
1558 scope, gfp, flags); in sctp_assoc_set_bind_addr_from_ep()
1564 gfp_t gfp) in sctp_assoc_set_bind_addr_from_cookie() argument
[all …]
Dulpqueue.c96 gfp_t gfp) in sctp_ulpq_tail_data() argument
103 event = sctp_ulpevent_make_rcvmsg(chunk->asoc, chunk, gfp); in sctp_ulpq_tail_data()
1035 gfp_t gfp) in sctp_ulpq_partial_delivery() argument
1081 gfp_t gfp) in sctp_ulpq_renege() argument
1105 retval = sctp_ulpq_tail_data(ulpq, chunk, gfp); in sctp_ulpq_renege()
1111 sctp_ulpq_partial_delivery(ulpq, gfp); in sctp_ulpq_renege()
1124 void sctp_ulpq_abort_pd(struct sctp_ulpq *ulpq, gfp_t gfp) in sctp_ulpq_abort_pd() argument
1137 gfp); in sctp_ulpq_abort_pd()
Dsm_sideeffect.c62 gfp_t gfp);
70 gfp_t gfp);
615 gfp_t gfp) in sctp_cmd_process_init() argument
624 if (!sctp_process_init(asoc, chunk, sctp_source(chunk), peer_init, gfp)) in sctp_cmd_process_init()
1102 gfp_t gfp) in sctp_do_sm() argument
1127 &commands, gfp); in sctp_do_sm()
1143 gfp_t gfp) in sctp_side_effects() argument
1156 commands, gfp))) in sctp_side_effects()
1228 gfp_t gfp) in sctp_cmd_interpreter() argument
1350 cmd->obj.init, gfp); in sctp_cmd_interpreter()
Dsm_make_chunk.c78 gfp_t gfp);
219 gfp_t gfp, int vparam_len) in sctp_make_init() argument
246 addrs = sctp_bind_addrs_to_raw(bp, &addrs_len, gfp); in sctp_make_init()
384 gfp_t gfp, int unkparam_len) in sctp_make_init_ack() argument
405 addrs = sctp_bind_addrs_to_raw(&asoc->base.bind_addr, &addrs_len, gfp); in sctp_make_init_ack()
1571 gfp_t gfp) in sctp_make_temp_asoc() argument
1580 asoc = sctp_association_new(ep, ep->base.sk, scope, gfp); in sctp_make_temp_asoc()
1692 struct sctp_chunk *chunk, gfp_t gfp, in sctp_unpack_cookie() argument
1812 retval = sctp_association_new(ep, ep->base.sk, scope, gfp); in sctp_unpack_cookie()
2292 sctp_init_chunk_t *peer_init, gfp_t gfp) in sctp_process_init() argument
[all …]
Dtransport.c58 gfp_t gfp) in sctp_transport_init() argument
108 gfp_t gfp) in sctp_transport_new() argument
112 transport = kzalloc(sizeof(*transport), gfp); in sctp_transport_new()
116 if (!sctp_transport_init(net, transport, addr, gfp)) in sctp_transport_new()
Dchunk.c62 static struct sctp_datamsg *sctp_datamsg_new(gfp_t gfp) in sctp_datamsg_new() argument
65 msg = kmalloc(sizeof(struct sctp_datamsg), gfp); in sctp_datamsg_new()
Dtsnmap.c51 __u32 initial_tsn, gfp_t gfp) in sctp_tsnmap_init() argument
54 map->tsn_map = kzalloc(len>>3, gfp); in sctp_tsnmap_init()
/linux-4.1.27/arch/mn10300/mm/
Ddma-alloc.c24 dma_addr_t *dma_handle, int gfp) in dma_alloc_coherent() argument
30 dev ? dev_name(dev) : "?", size, gfp); in dma_alloc_coherent()
41 gfp &= ~(__GFP_DMA | __GFP_HIGHMEM); in dma_alloc_coherent()
44 gfp |= GFP_DMA; in dma_alloc_coherent()
46 addr = __get_free_pages(gfp, get_order(size)); in dma_alloc_coherent()
/linux-4.1.27/scripts/coccinelle/locks/
Dcall_kern.cocci19 @gfp exists@
39 identifier gfp.fn;
74 position gfp.p;
81 position gfp.p;
87 p << gfp.p;
88 fn << gfp.fn;
98 p << gfp.p;
99 fn << gfp.fn;
/linux-4.1.27/tools/testing/fault-injection/
Dfailcmd.sh43 --reject-start=value, --reject-end=value, --ignore-gfp-wait=value
51 --ignore-gfp-highmem=value, --min-order=value
91 LONGOPTS=$LONGOPTS,ignore-gfp-wait:,cache-filter:
93 LONGOPTS=$LONGOPTS,ignore-gfp-wait:,ignore-gfp-highmem:,min-order:
181 --ignore-gfp-wait)
182 echo $2 > $FAULTATTR/ignore-gfp-wait
189 --ignore-gfp-highmem)
190 echo $2 > $FAULTATTR/ignore-gfp-highmem
/linux-4.1.27/arch/avr32/mm/
Ddma-coherent.c41 dma_addr_t *handle, gfp_t gfp) in __dma_alloc() argument
51 gfp &= ~(__GFP_COMP); in __dma_alloc()
56 page = alloc_pages(gfp, order); in __dma_alloc()
97 dma_addr_t *handle, gfp_t gfp) in dma_alloc_coherent() argument
102 page = __dma_alloc(dev, size, handle, gfp); in dma_alloc_coherent()
125 dma_addr_t *handle, gfp_t gfp) in dma_alloc_writecombine() argument
130 page = __dma_alloc(dev, size, handle, gfp); in dma_alloc_writecombine()
/linux-4.1.27/arch/mips/loongson/common/
Ddma-swiotlb.c13 dma_addr_t *dma_handle, gfp_t gfp, struct dma_attrs *attrs) in loongson_dma_alloc_coherent() argument
21 gfp &= ~(__GFP_DMA | __GFP_DMA32 | __GFP_HIGHMEM); in loongson_dma_alloc_coherent()
25 gfp |= __GFP_DMA; in loongson_dma_alloc_coherent()
30 gfp |= __GFP_DMA; in loongson_dma_alloc_coherent()
35 gfp |= __GFP_DMA32; in loongson_dma_alloc_coherent()
39 gfp |= __GFP_NORETRY; in loongson_dma_alloc_coherent()
41 ret = swiotlb_alloc_coherent(dev, size, dma_handle, gfp); in loongson_dma_alloc_coherent()
/linux-4.1.27/arch/mips/netlogic/common/
Dnlm-dma.c48 dma_addr_t *dma_handle, gfp_t gfp, struct dma_attrs *attrs) in nlm_dma_alloc_coherent() argument
56 gfp &= ~(__GFP_DMA | __GFP_DMA32 | __GFP_HIGHMEM); in nlm_dma_alloc_coherent()
60 gfp |= __GFP_DMA32; in nlm_dma_alloc_coherent()
64 gfp |= __GFP_NORETRY; in nlm_dma_alloc_coherent()
66 return swiotlb_alloc_coherent(dev, size, dma_handle, gfp); in nlm_dma_alloc_coherent()
/linux-4.1.27/arch/mips/mm/
Ddma-default.c78 static gfp_t massage_gfp_flags(const struct device *dev, gfp_t gfp) in massage_gfp_flags() argument
83 gfp &= ~(__GFP_DMA | __GFP_DMA32 | __GFP_HIGHMEM); in massage_gfp_flags()
110 gfp |= __GFP_NORETRY; in massage_gfp_flags()
112 return gfp | dma_flag; in massage_gfp_flags()
116 dma_addr_t * dma_handle, gfp_t gfp) in dma_alloc_noncoherent() argument
120 gfp = massage_gfp_flags(dev, gfp); in dma_alloc_noncoherent()
122 ret = (void *) __get_free_pages(gfp, get_order(size)); in dma_alloc_noncoherent()
134 dma_addr_t * dma_handle, gfp_t gfp, struct dma_attrs *attrs) in mips_dma_alloc_coherent() argument
143 gfp = massage_gfp_flags(dev, gfp); in mips_dma_alloc_coherent()
145 if (IS_ENABLED(CONFIG_DMA_CMA) && !(gfp & GFP_ATOMIC)) in mips_dma_alloc_coherent()
[all …]
/linux-4.1.27/lib/
Dkasprintf.c14 char *kvasprintf(gfp_t gfp, const char *fmt, va_list ap) in kvasprintf() argument
24 p = kmalloc_track_caller(len+1, gfp); in kvasprintf()
34 char *kasprintf(gfp_t gfp, const char *fmt, ...) in kasprintf() argument
40 p = kvasprintf(gfp, fmt, ap); in kasprintf()
Dargv_split.c59 char **argv_split(gfp_t gfp, const char *str, int *argcp) in argv_split() argument
66 argv_str = kstrndup(str, KMALLOC_MAX_SIZE - 1, gfp); in argv_split()
71 argv = kmalloc(sizeof(*argv) * (argc + 2), gfp); in argv_split()
Dflex_proportions.c37 int fprop_global_init(struct fprop_global *p, gfp_t gfp) in fprop_global_init() argument
43 err = percpu_counter_init(&p->events, 1, gfp); in fprop_global_init()
171 int fprop_local_init_percpu(struct fprop_local_percpu *pl, gfp_t gfp) in fprop_local_init_percpu() argument
175 err = percpu_counter_init(&pl->events, 0, gfp); in fprop_local_init_percpu()
Dproportions.c76 int prop_descriptor_init(struct prop_descriptor *pd, int shift, gfp_t gfp) in prop_descriptor_init() argument
86 err = percpu_counter_init(&pd->pg[0].events, 0, gfp); in prop_descriptor_init()
90 err = percpu_counter_init(&pd->pg[1].events, 0, gfp); in prop_descriptor_init()
191 int prop_local_init_percpu(struct prop_local_percpu *pl, gfp_t gfp) in prop_local_init_percpu() argument
196 return percpu_counter_init(&pl->events, 0, gfp); in prop_local_init_percpu()
Dbtree.c93 static unsigned long *btree_node_alloc(struct btree_head *head, gfp_t gfp) in btree_node_alloc() argument
97 node = mempool_alloc(head->mempool, gfp); in btree_node_alloc()
414 gfp_t gfp) in btree_grow() argument
419 node = btree_node_alloc(head, gfp); in btree_grow()
450 gfp_t gfp) in btree_insert_level() argument
457 err = btree_grow(head, geo, gfp); in btree_insert_level()
473 new = btree_node_alloc(head, gfp); in btree_insert_level()
478 new, level + 1, gfp); in btree_insert_level()
511 unsigned long *key, void *val, gfp_t gfp) in btree_insert() argument
514 return btree_insert_level(head, geo, key, val, 1, gfp); in btree_insert()
[all …]
Drhashtable.c64 gfp_t gfp) in alloc_bucket_locks() argument
82 gfp == GFP_KERNEL) in alloc_bucket_locks()
87 gfp); in alloc_bucket_locks()
113 gfp_t gfp) in bucket_table_alloc() argument
121 gfp != GFP_KERNEL) in bucket_table_alloc()
122 tbl = kzalloc(size, gfp | __GFP_NOWARN | __GFP_NORETRY); in bucket_table_alloc()
123 if (tbl == NULL && gfp == GFP_KERNEL) in bucket_table_alloc()
130 if (alloc_bucket_locks(ht, tbl, gfp) < 0) { in bucket_table_alloc()
Dpercpu-refcount.c59 unsigned int flags, gfp_t gfp) in percpu_ref_init() argument
66 __alloc_percpu_gfp(sizeof(unsigned long), align, gfp); in percpu_ref_init()
Dpercpu_counter.c115 int __percpu_counter_init(struct percpu_counter *fbc, s64 amount, gfp_t gfp, in __percpu_counter_init() argument
123 fbc->counters = alloc_percpu_gfp(s32, gfp); in __percpu_counter_init()
/linux-4.1.27/arch/cris/arch-v32/drivers/pci/
Ddma.c20 dma_addr_t *dma_handle, gfp_t gfp) in dma_alloc_coherent() argument
25 gfp &= ~(__GFP_DMA | __GFP_HIGHMEM); in dma_alloc_coherent()
31 gfp |= GFP_DMA; in dma_alloc_coherent()
33 ret = (void *)__get_free_pages(gfp, order); in dma_alloc_coherent()
/linux-4.1.27/include/net/netfilter/
Dnf_conntrack_extend.h94 size_t var_alloc_len, gfp_t gfp);
96 #define nf_ct_ext_add(ct, id, gfp) \ argument
97 ((id##_TYPE *)__nf_ct_ext_add_length((ct), (id), 0, (gfp)))
98 #define nf_ct_ext_add_length(ct, id, len, gfp) \ argument
99 ((id##_TYPE *)__nf_ct_ext_add_length((ct), (id), (len), (gfp)))
Dnf_conntrack_timestamp.h26 struct nf_conn_tstamp *nf_ct_tstamp_ext_add(struct nf_conn *ct, gfp_t gfp) in nf_ct_tstamp_ext_add() argument
34 return nf_ct_ext_add(ct, NF_CT_EXT_TSTAMP, gfp); in nf_ct_tstamp_ext_add()
Dnf_conntrack_acct.h33 struct nf_conn_acct *nf_ct_acct_ext_add(struct nf_conn *ct, gfp_t gfp) in nf_ct_acct_ext_add() argument
41 acct = nf_ct_ext_add(ct, NF_CT_EXT_ACCT, gfp); in nf_ct_acct_ext_add()
Dnf_conntrack_timeout.h41 gfp_t gfp) in nf_ct_timeout_ext_add() argument
46 timeout_ext = nf_ct_ext_add(ct, NF_CT_EXT_TIMEOUT, gfp); in nf_ct_timeout_ext_add()
Dnf_conntrack_ecache.h34 nf_ct_ecache_ext_add(struct nf_conn *ct, u16 ctmask, u16 expmask, gfp_t gfp) in nf_ct_ecache_ext_add() argument
47 e = nf_ct_ext_add(ct, NF_CT_EXT_ECACHE, gfp); in nf_ct_ecache_ext_add()
Dnf_tables.h516 u64 timeout, gfp_t gfp);
548 gfp_t gfp);
559 gfp_t gfp) in nft_set_gc_batch_check() argument
566 return nft_set_gc_batch_alloc(set, gfp); in nft_set_gc_batch_check()
Dnf_conntrack_helper.h67 gfp_t gfp);
/linux-4.1.27/arch/arc/mm/
Ddma.c30 dma_addr_t *dma_handle, gfp_t gfp) in dma_alloc_noncoherent() argument
35 paddr = alloc_pages_exact(size, gfp); in dma_alloc_noncoherent()
55 dma_addr_t *dma_handle, gfp_t gfp) in dma_alloc_coherent() argument
60 paddr = alloc_pages_exact(size, gfp); in dma_alloc_coherent()
/linux-4.1.27/drivers/base/
Ddevres.c85 size_t size, gfp_t gfp) in alloc_dr() argument
90 dr = kmalloc_track_caller(tot_size, gfp); in alloc_dr()
109 void * __devres_alloc(dr_release_t release, size_t size, gfp_t gfp, in __devres_alloc() argument
114 dr = alloc_dr(release, size, gfp | __GFP_ZERO); in __devres_alloc()
135 void * devres_alloc(dr_release_t release, size_t size, gfp_t gfp) in devres_alloc() argument
139 dr = alloc_dr(release, size, gfp | __GFP_ZERO); in devres_alloc()
532 void * devres_open_group(struct device *dev, void *id, gfp_t gfp) in devres_open_group() argument
537 grp = kmalloc(sizeof(*grp), gfp); in devres_open_group()
774 void * devm_kmalloc(struct device *dev, size_t size, gfp_t gfp) in devm_kmalloc() argument
779 dr = alloc_dr(devm_kmalloc_release, size, gfp); in devm_kmalloc()
[all …]
Ddma-mapping.c66 dma_addr_t *dma_handle, gfp_t gfp) in dmam_alloc_coherent() argument
71 dr = devres_alloc(dmam_coherent_release, sizeof(*dr), gfp); in dmam_alloc_coherent()
75 vaddr = dma_alloc_coherent(dev, size, dma_handle, gfp); in dmam_alloc_coherent()
125 dma_addr_t *dma_handle, gfp_t gfp) in dmam_alloc_noncoherent() argument
130 dr = devres_alloc(dmam_noncoherent_release, sizeof(*dr), gfp); in dmam_alloc_noncoherent()
134 vaddr = dma_alloc_noncoherent(dev, size, dma_handle, gfp); in dmam_alloc_noncoherent()
Ddevcoredump.c202 gfp_t gfp) in dev_coredumpv() argument
204 dev_coredumpm(dev, NULL, data, datalen, gfp, devcd_readv, vfree); in dev_coredumpv()
231 const void *data, size_t datalen, gfp_t gfp, in dev_coredumpm() argument
253 devcd = kzalloc(sizeof(*devcd), gfp); in dev_coredumpm()
/linux-4.1.27/include/net/
Dcfg80211.h3971 s32 signal, gfp_t gfp);
3977 s32 signal, gfp_t gfp) in cfg80211_inform_bss_frame() argument
3981 mgmt, len, signal, gfp); in cfg80211_inform_bss_frame()
4026 s32 signal, gfp_t gfp);
4034 s32 signal, gfp_t gfp) in cfg80211_inform_bss() argument
4040 gfp); in cfg80211_inform_bss()
4202 const u8 *tsc, gfp_t gfp);
4220 struct ieee80211_channel *channel, gfp_t gfp);
4236 const u8 *macaddr, const u8 *ie, u8 ie_len, gfp_t gfp);
4301 int approxlen, gfp_t gfp);
[all …]
Ddn_nsp.h22 void dn_send_conn_conf(struct sock *sk, gfp_t gfp);
24 unsigned short reason, gfp_t gfp);
33 void dn_nsp_queue_xmit(struct sock *sk, struct sk_buff *skb, gfp_t gfp,
/linux-4.1.27/net/decnet/
Ddn_nsp_out.c212 gfp_t gfp) in dn_nsp_clone_and_send() argument
218 if ((skb2 = skb_clone(skb, gfp)) != NULL) { in dn_nsp_clone_and_send()
352 gfp_t gfp, int oth) in dn_nsp_queue_xmit() argument
376 dn_nsp_clone_and_send(skb, gfp); in dn_nsp_queue_xmit()
515 void dn_send_conn_conf(struct sock *sk, gfp_t gfp) in dn_send_conn_conf() argument
522 if ((skb = dn_alloc_skb(sk, 50 + len, gfp)) == NULL) in dn_send_conn_conf()
547 unsigned short reason, gfp_t gfp, in dn_nsp_do_disc() argument
561 if ((skb = dn_alloc_skb(sk, size, gfp)) == NULL) in dn_nsp_do_disc()
590 unsigned short reason, gfp_t gfp) in dn_nsp_send_disc() argument
601 dn_nsp_do_disc(sk, msgflg, reason, gfp, __sk_dst_get(sk), ddl, in dn_nsp_send_disc()
[all …]
/linux-4.1.27/net/rds/
Dtcp_recv.c151 gfp_t gfp; member
174 arg->gfp); in rds_tcp_data_recv()
210 clone = skb_clone(skb, arg->gfp); in rds_tcp_data_recv()
244 arg->gfp); in rds_tcp_data_recv()
261 static int rds_tcp_read_sock(struct rds_connection *conn, gfp_t gfp) in rds_tcp_read_sock() argument
270 arg.gfp = gfp; in rds_tcp_read_sock()
276 rdsdebug("tcp_read_sock for tc %p gfp 0x%x returned %d\n", tc, gfp, in rds_tcp_read_sock()
Dpage.c98 gfp_t gfp) in rds_page_remainder_alloc() argument
105 gfp |= __GFP_HIGHMEM; in rds_page_remainder_alloc()
109 page = alloc_page(gfp); in rds_page_remainder_alloc()
151 page = alloc_page(gfp); in rds_page_remainder_alloc()
Dconnection.c121 struct rds_transport *trans, gfp_t gfp, in __rds_conn_create() argument
149 conn = kmem_cache_zalloc(rds_conn_slab, gfp); in __rds_conn_create()
192 ret = trans->conn_alloc(conn, gfp); in __rds_conn_create()
264 struct rds_transport *trans, gfp_t gfp) in rds_conn_create() argument
266 return __rds_conn_create(laddr, faddr, trans, gfp, 0); in rds_conn_create()
271 struct rds_transport *trans, gfp_t gfp) in rds_conn_create_outgoing() argument
273 return __rds_conn_create(laddr, faddr, trans, gfp, 1); in rds_conn_create_outgoing()
Dloop.c119 static int rds_loop_conn_alloc(struct rds_connection *conn, gfp_t gfp) in rds_loop_conn_alloc() argument
124 lc = kzalloc(sizeof(struct rds_loop_connection), gfp); in rds_loop_conn_alloc()
Drds.h424 int (*conn_alloc)(struct rds_connection *conn, gfp_t gfp);
616 struct rds_transport *trans, gfp_t gfp);
618 struct rds_transport *trans, gfp_t gfp);
658 struct rds_message *rds_message_alloc(unsigned int nents, gfp_t gfp);
690 gfp_t gfp);
705 struct rds_incoming *inc, gfp_t gfp);
Dtcp.c199 static int rds_tcp_conn_alloc(struct rds_connection *conn, gfp_t gfp) in rds_tcp_conn_alloc() argument
203 tc = kmem_cache_alloc(rds_tcp_conn_slab, gfp); in rds_tcp_conn_alloc()
Dmessage.c193 struct rds_message *rds_message_alloc(unsigned int extra_len, gfp_t gfp) in rds_message_alloc() argument
200 rm = kzalloc(sizeof(struct rds_message) + extra_len, gfp); in rds_message_alloc()
/linux-4.1.27/block/
Dblk-mq.h101 gfp_t gfp; member
110 struct request_queue *q, gfp_t gfp, bool reserved, in blk_mq_set_alloc_data() argument
115 data->gfp = gfp; in blk_mq_set_alloc_data()
/linux-4.1.27/arch/openrisc/include/asm/
Ddma-mapping.h44 dma_addr_t *dma_handle, gfp_t gfp, in dma_alloc_attrs() argument
50 memory = ops->alloc(dev, size, dma_handle, gfp, attrs); in dma_alloc_attrs()
71 dma_addr_t *dma_handle, gfp_t gfp) in dma_alloc_noncoherent() argument
77 return dma_alloc_attrs(dev, size, dma_handle, gfp, &attrs); in dma_alloc_noncoherent()
/linux-4.1.27/scripts/
Dgfp-translate66 grep -q ___GFP $SOURCE/include/linux/gfp.h
68 grep "^#define ___GFP" $SOURCE/include/linux/gfp.h | sed -e 's/u$//' | grep -v GFP_BITS > $TMPFILE
70 …grep "^#define __GFP" $SOURCE/include/linux/gfp.h | sed -e 's/(__force gfp_t)//' | sed -e 's/u)/)/…
/linux-4.1.27/fs/9p/
Dcache.h44 extern int __v9fs_fscache_release_page(struct page *page, gfp_t gfp);
57 gfp_t gfp) in v9fs_fscache_release_page() argument
59 return __v9fs_fscache_release_page(page, gfp); in v9fs_fscache_release_page()
117 gfp_t gfp) { in v9fs_fscache_release_page() argument
Dvfs_addr.c135 static int v9fs_release_page(struct page *page, gfp_t gfp) in v9fs_release_page() argument
139 return v9fs_fscache_release_page(page, gfp); in v9fs_release_page()
Dcache.c280 int __v9fs_fscache_release_page(struct page *page, gfp_t gfp) in __v9fs_fscache_release_page() argument
287 return fscache_maybe_release_page(v9inode->fscache, page, gfp); in __v9fs_fscache_release_page()
/linux-4.1.27/arch/ia64/kernel/
Dpci-swiotlb.c18 dma_addr_t *dma_handle, gfp_t gfp, in ia64_swiotlb_alloc_coherent() argument
22 gfp |= GFP_DMA; in ia64_swiotlb_alloc_coherent()
23 return swiotlb_alloc_coherent(dev, size, dma_handle, gfp); in ia64_swiotlb_alloc_coherent()
/linux-4.1.27/tools/virtio/linux/
Dvirtio.h35 gfp_t gfp);
40 gfp_t gfp);
45 gfp_t gfp);
Dkernel.h44 static inline void *kmalloc(size_t s, gfp_t gfp) in kmalloc() argument
58 static inline void *krealloc(void *p, size_t s, gfp_t gfp) in krealloc() argument
64 static inline unsigned long __get_free_page(gfp_t gfp) in __get_free_page() argument
/linux-4.1.27/arch/nios2/mm/
Ddma-mapping.c25 dma_addr_t *dma_handle, gfp_t gfp) in dma_alloc_coherent() argument
30 gfp &= ~(__GFP_DMA | __GFP_HIGHMEM); in dma_alloc_coherent()
33 gfp |= __GFP_ZERO; in dma_alloc_coherent()
36 gfp |= GFP_DMA; in dma_alloc_coherent()
38 ret = (void *) __get_free_pages(gfp, get_order(size)); in dma_alloc_coherent()
/linux-4.1.27/net/ceph/
Dbuffer.c11 struct ceph_buffer *ceph_buffer_new(size_t len, gfp_t gfp) in ceph_buffer_new() argument
15 b = kmalloc(sizeof(*b), gfp); in ceph_buffer_new()
19 b->vec.iov_base = ceph_kvmalloc(len, gfp); in ceph_buffer_new()
/linux-4.1.27/arch/arm/mm/
Ddma-mapping.c148 dma_addr_t *handle, gfp_t gfp, struct dma_attrs *attrs);
254 static struct page *__dma_alloc_buffer(struct device *dev, size_t size, gfp_t gfp) in __dma_alloc_buffer() argument
259 page = alloc_pages(gfp, order); in __dma_alloc_buffer()
294 static void *__alloc_remap_buffer(struct device *dev, size_t size, gfp_t gfp,
299 __dma_alloc_remap(struct page *page, size_t size, gfp_t gfp, pgprot_t prot, in __dma_alloc_remap() argument
350 gfp_t gfp = GFP_KERNEL | GFP_DMA; in atomic_pool_init() local
362 ptr = __alloc_remap_buffer(NULL, atomic_pool_size, gfp, prot, in atomic_pool_init()
468 static void *__alloc_remap_buffer(struct device *dev, size_t size, gfp_t gfp, in __alloc_remap_buffer() argument
474 page = __dma_alloc_buffer(dev, size, gfp); in __alloc_remap_buffer()
480 ptr = __dma_alloc_remap(page, size, gfp, prot, caller); in __alloc_remap_buffer()
[all …]
/linux-4.1.27/arch/mips/cavium-octeon/
Ddma-octeon.c160 dma_addr_t *dma_handle, gfp_t gfp, struct dma_attrs *attrs) in octeon_dma_alloc_coherent() argument
168 gfp &= ~(__GFP_DMA | __GFP_DMA32 | __GFP_HIGHMEM); in octeon_dma_alloc_coherent()
172 gfp |= __GFP_DMA; in octeon_dma_alloc_coherent()
174 gfp |= __GFP_DMA; in octeon_dma_alloc_coherent()
179 gfp |= __GFP_DMA32; in octeon_dma_alloc_coherent()
185 gfp |= __GFP_NORETRY; in octeon_dma_alloc_coherent()
187 ret = swiotlb_alloc_coherent(dev, size, dma_handle, gfp); in octeon_dma_alloc_coherent()
/linux-4.1.27/net/netfilter/
Dnf_conntrack_extend.c48 size_t var_alloc_len, gfp_t gfp) in nf_ct_ext_create() argument
62 *ext = kzalloc(alloc_size, gfp); in nf_ct_ext_create()
73 size_t var_alloc_len, gfp_t gfp) in __nf_ct_ext_add_length() argument
84 return nf_ct_ext_create(&ct->ext, id, var_alloc_len, gfp); in __nf_ct_ext_add_length()
97 new = __krealloc(old, newlen, gfp); in __nf_ct_ext_add_length()
/linux-4.1.27/drivers/usb/storage/
Duas.c77 struct uas_dev_info *devinfo, gfp_t gfp);
399 static struct urb *uas_alloc_data_urb(struct uas_dev_info *devinfo, gfp_t gfp, in uas_alloc_data_urb() argument
405 struct urb *urb = usb_alloc_urb(0, gfp); in uas_alloc_data_urb()
423 static struct urb *uas_alloc_sense_urb(struct uas_dev_info *devinfo, gfp_t gfp, in uas_alloc_sense_urb() argument
428 struct urb *urb = usb_alloc_urb(0, gfp); in uas_alloc_sense_urb()
434 iu = kzalloc(sizeof(*iu), gfp); in uas_alloc_sense_urb()
450 static struct urb *uas_alloc_cmd_urb(struct uas_dev_info *devinfo, gfp_t gfp, in uas_alloc_cmd_urb() argument
456 struct urb *urb = usb_alloc_urb(0, gfp); in uas_alloc_cmd_urb()
467 iu = kzalloc(sizeof(*iu) + len, gfp); in uas_alloc_cmd_urb()
494 static struct urb *uas_submit_sense_urb(struct scsi_cmnd *cmnd, gfp_t gfp) in uas_submit_sense_urb() argument
[all …]
/linux-4.1.27/net/mac80211/
Devent.c20 gfp_t gfp) in mac80211_ev_michael_mic_failure() argument
26 keyidx, tsc, gfp); in mac80211_ev_michael_mic_failure()
Dpm.c173 gfp_t gfp) in ieee80211_report_wowlan_wakeup() argument
177 cfg80211_report_wowlan_wakeup(&sdata->wdev, wakeup, gfp); in ieee80211_report_wowlan_wakeup()
Drate.h127 gfp_t gfp) in rate_control_alloc_sta() argument
129 return ref->ops->alloc_sta(ref->priv, sta, gfp); in rate_control_alloc_sta()
Drc80211_minstrel.c569 minstrel_alloc_sta(void *priv, struct ieee80211_sta *sta, gfp_t gfp) in minstrel_alloc_sta() argument
578 mi = kzalloc(sizeof(struct minstrel_sta_info), gfp); in minstrel_alloc_sta()
588 mi->r = kzalloc(sizeof(struct minstrel_rate) * max_rates, gfp); in minstrel_alloc_sta()
592 mi->sample_table = kmalloc(SAMPLE_COLUMNS * max_rates, gfp); in minstrel_alloc_sta()
/linux-4.1.27/arch/powerpc/mm/
Ddma-noncoherent.c97 ppc_vm_region_alloc(struct ppc_vm_region *head, size_t size, gfp_t gfp) in ppc_vm_region_alloc() argument
103 new = kmalloc(sizeof(struct ppc_vm_region), gfp); in ppc_vm_region_alloc()
155 __dma_alloc_coherent(struct device *dev, size_t size, dma_addr_t *handle, gfp_t gfp) in __dma_alloc_coherent() argument
196 gfp |= GFP_DMA; in __dma_alloc_coherent()
198 page = alloc_pages(gfp, order); in __dma_alloc_coherent()
216 gfp & ~(__GFP_DMA | __GFP_HIGHMEM)); in __dma_alloc_coherent()
/linux-4.1.27/arch/tile/kernel/
Dpci-dma.c36 dma_addr_t *dma_handle, gfp_t gfp, in tile_dma_alloc_coherent() argument
46 gfp |= __GFP_ZERO; in tile_dma_alloc_coherent()
57 gfp |= GFP_DMA; in tile_dma_alloc_coherent()
61 pg = homecache_alloc_pages_node(node, gfp, order, PAGE_HOME_DMA); in tile_dma_alloc_coherent()
345 dma_addr_t *dma_handle, gfp_t gfp, in tile_pci_dma_alloc_coherent() argument
353 gfp |= __GFP_ZERO; in tile_pci_dma_alloc_coherent()
355 pg = homecache_alloc_pages_node(node, gfp, order, PAGE_HOME_DMA); in tile_pci_dma_alloc_coherent()
533 dma_addr_t *dma_handle, gfp_t gfp, in tile_swiotlb_alloc_coherent() argument
536 gfp |= GFP_DMA; in tile_swiotlb_alloc_coherent()
537 return swiotlb_alloc_coherent(dev, size, dma_handle, gfp); in tile_swiotlb_alloc_coherent()
/linux-4.1.27/arch/m68k/kernel/
Ddma.c76 dma_addr_t *dma_handle, gfp_t gfp) in dma_alloc_coherent() argument
80 gfp &= ~(__GFP_DMA | __GFP_HIGHMEM); in dma_alloc_coherent()
83 gfp |= GFP_DMA; in dma_alloc_coherent()
84 ret = (void *)__get_free_pages(gfp, get_order(size)); in dma_alloc_coherent()
/linux-4.1.27/fs/fscache/
Dpage.c65 gfp_t gfp) in __fscache_maybe_release_page() argument
70 _enter("%p,%p,%x", cookie, page, gfp); in __fscache_maybe_release_page()
125 if (!(gfp & __GFP_WAIT) || !(gfp & __GFP_FS)) { in __fscache_maybe_release_page()
135 gfp &= ~__GFP_WAIT; in __fscache_maybe_release_page()
403 gfp_t gfp) in __fscache_read_or_alloc_page() argument
473 ret = object->cache->ops->allocate_page(op, page, gfp); in __fscache_read_or_alloc_page()
479 ret = object->cache->ops->read_or_alloc_page(op, page, gfp); in __fscache_read_or_alloc_page()
538 gfp_t gfp) in __fscache_read_or_alloc_pages() argument
605 op, pages, nr_pages, gfp); in __fscache_read_or_alloc_pages()
610 op, pages, nr_pages, gfp); in __fscache_read_or_alloc_pages()
[all …]
/linux-4.1.27/arch/metag/kernel/
Ddma.c115 gfp_t gfp) in metag_vm_region_alloc() argument
121 new = kmalloc(sizeof(struct metag_vm_region), gfp); in metag_vm_region_alloc()
175 dma_addr_t *handle, gfp_t gfp) in dma_alloc_coherent() argument
203 gfp |= GFP_DMA; in dma_alloc_coherent()
205 page = alloc_pages(gfp, order); in dma_alloc_coherent()
223 gfp & ~(__GFP_DMA | __GFP_HIGHMEM)); in dma_alloc_coherent()
/linux-4.1.27/drivers/md/bcache/
Dutil.h39 #define init_heap(heap, _size, gfp) \ argument
47 (heap)->data = kmalloc(_bytes, (gfp)); \
48 if ((!(heap)->data) && ((gfp) & GFP_KERNEL)) \
131 #define __init_fifo(fifo, gfp) \ argument
144 (fifo)->data = kmalloc(_bytes, (gfp)); \
145 if ((!(fifo)->data) && ((gfp) & GFP_KERNEL)) \
150 #define init_fifo_exact(fifo, _size, gfp) \ argument
153 __init_fifo(fifo, gfp); \
156 #define init_fifo(fifo, _size, gfp) \ argument
161 __init_fifo(fifo, gfp); \
/linux-4.1.27/drivers/virtio/
Dvirtio_ring.c102 unsigned int total_sg, gfp_t gfp) in alloc_indirect() argument
112 gfp &= ~(__GFP_HIGHMEM | __GFP_HIGH); in alloc_indirect()
114 desc = kmalloc(total_sg * sizeof(struct vring_desc), gfp); in alloc_indirect()
129 gfp_t gfp) in virtqueue_add() argument
168 desc = alloc_indirect(_vq, total_sg, gfp); in virtqueue_add()
277 gfp_t gfp) in virtqueue_add_sgs() argument
287 return virtqueue_add(_vq, sgs, total_sg, out_sgs, in_sgs, data, gfp); in virtqueue_add_sgs()
307 gfp_t gfp) in virtqueue_add_outbuf() argument
309 return virtqueue_add(vq, &sg, num, 1, 0, data, gfp); in virtqueue_add_outbuf()
329 gfp_t gfp) in virtqueue_add_inbuf() argument
[all …]
/linux-4.1.27/net/rxrpc/
Dar-connection.c35 static struct rxrpc_conn_bundle *rxrpc_alloc_bundle(gfp_t gfp) in rxrpc_alloc_bundle() argument
41 bundle = kzalloc(sizeof(struct rxrpc_conn_bundle), gfp); in rxrpc_alloc_bundle()
73 gfp_t gfp) in rxrpc_get_bundle() argument
106 candidate = rxrpc_alloc_bundle(gfp); in rxrpc_get_bundle()
198 static struct rxrpc_connection *rxrpc_alloc_connection(gfp_t gfp) in rxrpc_alloc_connection() argument
204 conn = kzalloc(sizeof(struct rxrpc_connection), gfp); in rxrpc_alloc_connection()
339 gfp_t gfp) in rxrpc_connect_exclusive() argument
350 conn = rxrpc_alloc_connection(gfp); in rxrpc_connect_exclusive()
433 gfp_t gfp) in rxrpc_connect_call() argument
444 call, gfp); in rxrpc_connect_call()
[all …]
Dar-transport.c36 gfp_t gfp) in rxrpc_alloc_transport() argument
42 trans = kzalloc(sizeof(struct rxrpc_transport), gfp); in rxrpc_alloc_transport()
80 gfp_t gfp) in rxrpc_get_transport() argument
102 candidate = rxrpc_alloc_transport(local, peer, gfp); in rxrpc_get_transport()
Dar-peer.c62 gfp_t gfp) in rxrpc_alloc_peer() argument
68 peer = kzalloc(sizeof(struct rxrpc_peer), gfp); in rxrpc_alloc_peer()
106 struct rxrpc_peer *rxrpc_get_peer(struct sockaddr_rxrpc *srx, gfp_t gfp) in rxrpc_get_peer() argument
139 candidate = rxrpc_alloc_peer(srx, gfp); in rxrpc_get_peer()
Dar-call.c202 static struct rxrpc_call *rxrpc_alloc_call(gfp_t gfp) in rxrpc_alloc_call() argument
206 call = kmem_cache_zalloc(rxrpc_call_jar, gfp); in rxrpc_alloc_call()
212 gfp); in rxrpc_alloc_call()
255 gfp_t gfp) in rxrpc_alloc_client_call() argument
266 call = rxrpc_alloc_call(gfp); in rxrpc_alloc_client_call()
274 ret = rxrpc_connect_call(rx, trans, bundle, call, gfp); in rxrpc_alloc_client_call()
320 gfp_t gfp) in rxrpc_get_client_call() argument
352 candidate = rxrpc_alloc_client_call(rx, trans, bundle, gfp); in rxrpc_get_client_call()
418 gfp_t gfp) in rxrpc_incoming_call() argument
424 _enter(",%d,,%x", conn->debug_id, gfp); in rxrpc_incoming_call()
[all …]
Daf_rxrpc.c226 gfp_t gfp) in rxrpc_name_to_transport() argument
244 peer = rxrpc_get_peer(srx, gfp); in rxrpc_name_to_transport()
249 trans = rxrpc_get_transport(rx->local, peer, gfp); in rxrpc_name_to_transport()
273 gfp_t gfp) in rxrpc_kernel_begin_call() argument
287 sizeof(*srx), 0, gfp); in rxrpc_kernel_begin_call()
311 bundle = rxrpc_get_bundle(rx, trans, key, service_id, gfp); in rxrpc_kernel_begin_call()
318 gfp); in rxrpc_kernel_begin_call()
/linux-4.1.27/drivers/usb/wusbcore/
Dwa-rpipe.c194 gfp_t gfp) in rpipe_get_idle() argument
201 rpipe = kzalloc(sizeof(*rpipe), gfp); in rpipe_get_idle()
314 struct usb_host_endpoint *ep, struct urb *urb, gfp_t gfp) in rpipe_aim() argument
398 const struct urb *urb, gfp_t gfp) in rpipe_check_aim() argument
438 struct urb *urb, gfp_t gfp) in rpipe_get_by_ep() argument
449 result = rpipe_check_aim(rpipe, wa, ep, urb, gfp); in rpipe_get_by_ep()
461 result = rpipe_get_idle(&rpipe, wa, 1 << eptype, gfp); in rpipe_get_by_ep()
464 result = rpipe_aim(rpipe, wa, ep, urb, gfp); in rpipe_get_by_ep()
/linux-4.1.27/fs/ceph/
Dcache.h69 static inline int ceph_release_fscache_page(struct page *page, gfp_t gfp) in ceph_release_fscache_page() argument
73 return fscache_maybe_release_page(ci->fscache, page, gfp); in ceph_release_fscache_page()
161 static inline int ceph_release_fscache_page(struct page *page, gfp_t gfp) in ceph_release_fscache_page() argument
/linux-4.1.27/arch/sh/mm/
Dconsistent.c36 dma_addr_t *dma_handle, gfp_t gfp, in dma_generic_alloc_coherent() argument
42 gfp |= __GFP_ZERO; in dma_generic_alloc_coherent()
44 ret = (void *)__get_free_pages(gfp, order); in dma_generic_alloc_coherent()
/linux-4.1.27/security/selinux/
Dxfrm.c82 gfp_t gfp) in selinux_xfrm_alloc_user() argument
98 ctx = kmalloc(sizeof(*ctx) + str_len + 1, gfp); in selinux_xfrm_alloc_user()
107 rc = security_context_to_sid(ctx->ctx_str, str_len, &ctx->ctx_sid, gfp); in selinux_xfrm_alloc_user()
287 gfp_t gfp) in selinux_xfrm_policy_alloc() argument
289 return selinux_xfrm_alloc_user(ctxp, uctx, gfp); in selinux_xfrm_policy_alloc()
/linux-4.1.27/kernel/irq/
Dirqdesc.c39 static int alloc_masks(struct irq_desc *desc, gfp_t gfp, int node) in alloc_masks() argument
41 if (!zalloc_cpumask_var_node(&desc->irq_data.affinity, gfp, node)) in alloc_masks()
45 if (!zalloc_cpumask_var_node(&desc->pending_mask, gfp, node)) { in alloc_masks()
69 alloc_masks(struct irq_desc *desc, gfp_t gfp, int node) { return 0; } in alloc_masks() argument
148 gfp_t gfp = GFP_KERNEL; in alloc_desc() local
150 desc = kzalloc_node(sizeof(*desc), gfp, node); in alloc_desc()
158 if (alloc_masks(desc, gfp, node)) in alloc_desc()
/linux-4.1.27/arch/alpha/kernel/
Dpci-noop.c111 dma_addr_t *dma_handle, gfp_t gfp, in alpha_noop_alloc_coherent() argument
117 gfp &= ~GFP_DMA; in alpha_noop_alloc_coherent()
118 ret = (void *)__get_free_pages(gfp, get_order(size)); in alpha_noop_alloc_coherent()
Dpci_iommu.c435 dma_addr_t *dma_addrp, gfp_t gfp, in alpha_pci_alloc_coherent() argument
442 gfp &= ~GFP_DMA; in alpha_pci_alloc_coherent()
445 cpu_addr = (void *)__get_free_pages(gfp, order); in alpha_pci_alloc_coherent()
459 if (alpha_mv.mv_pci_tbi || (gfp & GFP_DMA)) in alpha_pci_alloc_coherent()
463 gfp |= GFP_DMA; in alpha_pci_alloc_coherent()
/linux-4.1.27/arch/frv/mb93090-mb00/
Dpci-dma.c21 void *dma_alloc_coherent(struct device *hwdev, size_t size, dma_addr_t *dma_handle, gfp_t gfp) in dma_alloc_coherent() argument
25 ret = consistent_alloc(gfp, size, dma_handle); in dma_alloc_coherent()
Dpci-dma-nommu.c37 void *dma_alloc_coherent(struct device *hwdev, size_t size, dma_addr_t *dma_handle, gfp_t gfp) in dma_alloc_coherent() argument
/linux-4.1.27/arch/alpha/include/asm/
Ddma-mapping.h18 dma_addr_t *dma_handle, gfp_t gfp, in dma_alloc_attrs() argument
21 return get_dma_ops(dev)->alloc(dev, size, dma_handle, gfp, attrs); in dma_alloc_attrs()
/linux-4.1.27/include/crypto/
Dskcipher.h69 struct crypto_ablkcipher *tfm, gfp_t gfp) in skcipher_givcrypt_alloc() argument
74 crypto_ablkcipher_reqsize(tfm), gfp); in skcipher_givcrypt_alloc()
Daead.h58 struct crypto_aead *tfm, gfp_t gfp) in aead_givcrypt_alloc() argument
63 crypto_aead_reqsize(tfm), gfp); in aead_givcrypt_alloc()
Dhash.h512 struct crypto_ahash *tfm, gfp_t gfp) in ahash_request_alloc() argument
517 crypto_ahash_reqsize(tfm), gfp); in ahash_request_alloc()
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx4/
Dalloc.c587 struct mlx4_buf *buf, gfp_t gfp) in mlx4_buf_alloc() argument
596 size, &t, gfp); in mlx4_buf_alloc()
616 gfp); in mlx4_buf_alloc()
624 &t, gfp); in mlx4_buf_alloc()
635 pages = kmalloc(sizeof *pages * buf->nbufs, gfp); in mlx4_buf_alloc()
680 gfp_t gfp) in mlx4_alloc_db_pgdir() argument
684 pgdir = kzalloc(sizeof *pgdir, gfp); in mlx4_alloc_db_pgdir()
692 &pgdir->db_dma, gfp); in mlx4_alloc_db_pgdir()
732 int mlx4_db_alloc(struct mlx4_dev *dev, struct mlx4_db *db, int order, gfp_t gfp) in mlx4_db_alloc() argument
744 pgdir = mlx4_alloc_db_pgdir(&dev->persist->pdev->dev, gfp); in mlx4_db_alloc()
Dqp.c298 int __mlx4_qp_alloc_icm(struct mlx4_dev *dev, int qpn, gfp_t gfp) in __mlx4_qp_alloc_icm() argument
304 err = mlx4_table_get(dev, &qp_table->qp_table, qpn, gfp); in __mlx4_qp_alloc_icm()
308 err = mlx4_table_get(dev, &qp_table->auxc_table, qpn, gfp); in __mlx4_qp_alloc_icm()
312 err = mlx4_table_get(dev, &qp_table->altc_table, qpn, gfp); in __mlx4_qp_alloc_icm()
316 err = mlx4_table_get(dev, &qp_table->rdmarc_table, qpn, gfp); in __mlx4_qp_alloc_icm()
320 err = mlx4_table_get(dev, &qp_table->cmpt_table, qpn, gfp); in __mlx4_qp_alloc_icm()
342 static int mlx4_qp_alloc_icm(struct mlx4_dev *dev, int qpn, gfp_t gfp) in mlx4_qp_alloc_icm() argument
352 return __mlx4_qp_alloc_icm(dev, qpn, gfp); in mlx4_qp_alloc_icm()
381 int mlx4_qp_alloc(struct mlx4_dev *dev, int qpn, struct mlx4_qp *qp, gfp_t gfp) in mlx4_qp_alloc() argument
392 err = mlx4_qp_alloc_icm(dev, qpn, gfp); in mlx4_qp_alloc()
Dmr.c485 int __mlx4_mpt_alloc_icm(struct mlx4_dev *dev, u32 index, gfp_t gfp) in __mlx4_mpt_alloc_icm() argument
489 return mlx4_table_get(dev, &mr_table->dmpt_table, index, gfp); in __mlx4_mpt_alloc_icm()
492 static int mlx4_mpt_alloc_icm(struct mlx4_dev *dev, u32 index, gfp_t gfp) in mlx4_mpt_alloc_icm() argument
503 return __mlx4_mpt_alloc_icm(dev, index, gfp); in mlx4_mpt_alloc_icm()
793 struct mlx4_buf *buf, gfp_t gfp) in mlx4_buf_write_mtt() argument
800 gfp); in mlx4_buf_write_mtt()
Den_rx.c61 gfp_t gfp = _gfp; in mlx4_alloc_pages() local
64 gfp |= __GFP_COMP | __GFP_NOWARN; in mlx4_alloc_pages()
65 page = alloc_pages(gfp, order); in mlx4_alloc_pages()
94 gfp_t gfp) in mlx4_en_alloc_frags() argument
111 if (mlx4_alloc_pages(priv, &page_alloc[i], frag_info, gfp)) in mlx4_en_alloc_frags()
238 gfp_t gfp) in mlx4_en_prepare_rx_desc() argument
244 return mlx4_en_alloc_frags(priv, rx_desc, frags, ring->page_alloc, gfp); in mlx4_en_prepare_rx_desc()
Dicm.h75 gfp_t gfp);
/linux-4.1.27/security/apparmor/include/
Dcontext.h44 static inline struct aa_file_cxt *aa_alloc_file_context(gfp_t gfp) in aa_alloc_file_context() argument
46 return kzalloc(sizeof(struct aa_file_cxt), gfp); in aa_alloc_file_context()
Daudit.h136 int aa_audit(int type, struct aa_profile *profile, gfp_t gfp,
/linux-4.1.27/arch/frv/mm/
Ddma-alloc.c81 void *consistent_alloc(gfp_t gfp, size_t size, dma_addr_t *dma_handle) in consistent_alloc() argument
95 page = __get_free_pages(gfp, order); in consistent_alloc()
/linux-4.1.27/fs/cifs/
Dfscache.h51 extern int cifs_fscache_release_page(struct page *page, gfp_t gfp);
118 static inline int cifs_fscache_release_page(struct page *page, gfp_t gfp) in cifs_fscache_release_page() argument
Dfscache.c128 int cifs_fscache_release_page(struct page *page, gfp_t gfp) in cifs_fscache_release_page() argument
136 if (!fscache_maybe_release_page(cifsi->fscache, page, gfp)) in cifs_fscache_release_page()
/linux-4.1.27/arch/ia64/include/asm/
Ddma-mapping.h29 dma_addr_t *daddr, gfp_t gfp, in dma_alloc_attrs() argument
35 caddr = ops->alloc(dev, size, daddr, gfp, attrs); in dma_alloc_attrs()
/linux-4.1.27/arch/mips/include/asm/
Ddma-mapping.h71 dma_addr_t *dma_handle, gfp_t gfp, in dma_alloc_attrs() argument
77 ret = ops->alloc(dev, size, dma_handle, gfp, attrs); in dma_alloc_attrs()
/linux-4.1.27/Documentation/fault-injection/
Dfault-injection.txt98 - /sys/kernel/debug/fail_page_alloc/ignore-gfp-highmem:
104 - /sys/kernel/debug/failslab/ignore-gfp-wait:
105 - /sys/kernel/debug/fail_page_alloc/ignore-gfp-wait:
180 echo 1 > /sys/kernel/debug/$FAILTYPE/ignore-gfp-wait
234 echo 1 > /sys/kernel/debug/$FAILTYPE/ignore-gfp-wait
235 echo 1 > /sys/kernel/debug/$FAILTYPE/ignore-gfp-highmem
/linux-4.1.27/arch/sh/include/asm/
Ddma-mapping.h59 dma_addr_t *dma_handle, gfp_t gfp, in dma_alloc_attrs() argument
70 memory = ops->alloc(dev, size, dma_handle, gfp, attrs); in dma_alloc_attrs()
/linux-4.1.27/drivers/infiniband/hw/qib/
Dqib_qp.c104 gfp_t gfp) in get_map_page() argument
106 unsigned long page = get_zeroed_page(gfp); in get_map_page()
125 enum ib_qp_type type, u8 port, gfp_t gfp) in alloc_qpn() argument
155 get_map_page(qpt, map, gfp); in alloc_qpn()
987 gfp_t gfp; in qib_create_qp() local
1000 gfp = init_attr->create_flags & IB_QP_CREATE_USE_GFP_NOIO ? in qib_create_qp()
1034 gfp, PAGE_KERNEL); in qib_create_qp()
1050 qp = kzalloc(sz + sg_list_sz, gfp); in qib_create_qp()
1056 qp->s_hdr = kzalloc(sizeof(*qp->s_hdr), gfp); in qib_create_qp()
1071 if (gfp != GFP_NOIO) in qib_create_qp()
[all …]
/linux-4.1.27/arch/microblaze/mm/
Dconsistent.c62 void *consistent_alloc(gfp_t gfp, size_t size, dma_addr_t *dma_handle) in consistent_alloc() argument
82 vaddr = __get_free_pages(gfp, order); in consistent_alloc()
/linux-4.1.27/drivers/crypto/ccp/
Dccp-crypto-main.c272 gfp_t gfp; in ccp_crypto_enqueue_request() local
274 gfp = req->flags & CRYPTO_TFM_REQ_MAY_SLEEP ? GFP_KERNEL : GFP_ATOMIC; in ccp_crypto_enqueue_request()
276 crypto_cmd = kzalloc(sizeof(*crypto_cmd), gfp); in ccp_crypto_enqueue_request()
Dccp-crypto-aes-cmac.c68 gfp_t gfp; in ccp_do_cmac_update() local
111 gfp = req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP ? in ccp_do_cmac_update()
113 ret = sg_alloc_table(&rctx->data_sg, sg_count, gfp); in ccp_do_cmac_update()
Dccp-crypto-sha.c67 gfp_t gfp; in ccp_do_sha_update() local
101 gfp = req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP ? in ccp_do_sha_update()
104 ret = sg_alloc_table(&rctx->data_sg, sg_count, gfp); in ccp_do_sha_update()
/linux-4.1.27/arch/x86/kernel/cpu/
Dperf_event_intel_pt.c285 static struct topa *topa_alloc(int cpu, gfp_t gfp) in topa_alloc() argument
291 p = alloc_pages_node(node, gfp | __GFP_ZERO, 0); in topa_alloc()
375 static int topa_insert_pages(struct pt_buffer *buf, gfp_t gfp) in topa_insert_pages() argument
386 topa = topa_alloc(buf->cpu, gfp); in topa_insert_pages()
753 gfp_t gfp) in pt_buffer_init_topa() argument
758 topa = topa_alloc(buf->cpu, gfp); in pt_buffer_init_topa()
765 err = topa_insert_pages(buf, gfp); in pt_buffer_init_topa()
/linux-4.1.27/include/linux/ceph/
Dbuffer.h21 extern struct ceph_buffer *ceph_buffer_new(size_t len, gfp_t gfp);
Ddecode.h106 size_t *lenp, gfp_t gfp) in ceph_extract_encoded_string() argument
116 buf = kmalloc(len + 1, gfp); in ceph_extract_encoded_string()
/linux-4.1.27/drivers/net/
Dvirtio_net.c530 gfp_t gfp) in add_recvbuf_small() argument
536 skb = __netdev_alloc_skb_ip_align(vi->dev, GOOD_PACKET_LEN, gfp); in add_recvbuf_small()
547 err = virtqueue_add_inbuf(rq->vq, rq->sg, 2, skb, gfp); in add_recvbuf_small()
555 gfp_t gfp) in add_recvbuf_big() argument
565 first = get_a_page(rq, gfp); in add_recvbuf_big()
578 first = get_a_page(rq, gfp); in add_recvbuf_big()
596 first, gfp); in add_recvbuf_big()
613 static int add_recvbuf_mergeable(struct receive_queue *rq, gfp_t gfp) in add_recvbuf_mergeable() argument
622 if (unlikely(!skb_page_frag_refill(len, alloc_frag, gfp))) in add_recvbuf_mergeable()
641 err = virtqueue_add_inbuf(rq->vq, rq->sg, 1, (void *)ctx, gfp); in add_recvbuf_mergeable()
[all …]
/linux-4.1.27/arch/openrisc/kernel/
Ddma.c85 dma_addr_t *dma_handle, gfp_t gfp, in or1k_dma_alloc() argument
95 page = alloc_pages_exact(size, gfp); in or1k_dma_alloc()
/linux-4.1.27/drivers/vhost/
Dvringh.c181 static int resize_iovec(struct vringh_kiov *iov, gfp_t gfp) in resize_iovec() argument
191 new = krealloc(iov->iov, new_num * sizeof(struct iovec), gfp); in resize_iovec()
193 new = kmalloc(new_num * sizeof(struct iovec), gfp); in resize_iovec()
262 gfp_t gfp, in __vringh_iov() argument
358 err = resize_iovec(iov, gfp); in __vringh_iov()
912 gfp_t gfp) in vringh_getdesc_kern() argument
926 gfp, copydesc_kern); in vringh_getdesc_kern()
/linux-4.1.27/arch/frv/include/asm/
Dpci.h28 extern void *consistent_alloc(gfp_t gfp, size_t size, dma_addr_t *dma_handle);
/linux-4.1.27/arch/avr32/include/asm/
Ddma-mapping.h55 dma_addr_t *handle, gfp_t gfp);
85 dma_addr_t *handle, gfp_t gfp);
/linux-4.1.27/include/scsi/
Dscsi_dh.h85 gfp_t gfp) in scsi_dh_attached_handler_name() argument
/linux-4.1.27/arch/sparc/kernel/
Dioport.c262 dma_addr_t *dma_addrp, gfp_t gfp, in sbus_alloc_coherent() argument
281 va = __get_free_pages(gfp, order); in sbus_alloc_coherent()
431 dma_addr_t *pba, gfp_t gfp, in pci32_alloc_coherent() argument
447 va = (void *) __get_free_pages(gfp, order); in pci32_alloc_coherent()
/linux-4.1.27/arch/arc/include/asm/
Ddma-mapping.h35 dma_addr_t *dma_handle, gfp_t gfp);
41 dma_addr_t *dma_handle, gfp_t gfp);
/linux-4.1.27/arch/powerpc/include/asm/
Ddma-mapping.h45 dma_addr_t *handle, gfp_t gfp);
57 #define __dma_alloc_coherent(dev, gfp, size, handle) NULL argument
/linux-4.1.27/include/net/iucv/
Diucv.h223 static inline struct iucv_path *iucv_path_alloc(u16 msglim, u8 flags, gfp_t gfp) in iucv_path_alloc() argument
227 path = kzalloc(sizeof(struct iucv_path), gfp); in iucv_path_alloc()
/linux-4.1.27/drivers/staging/lustre/lustre/libcfs/
Dtracefile.c69 static struct cfs_trace_page *cfs_tage_alloc(gfp_t gfp) in cfs_tage_alloc() argument
82 gfp |= __GFP_NOWARN; in cfs_tage_alloc()
83 page = alloc_page(gfp); in cfs_tage_alloc()
87 tage = kmalloc(sizeof(*tage), gfp); in cfs_tage_alloc()
117 int cfs_trace_refill_stock(struct cfs_trace_cpu_data *tcd, gfp_t gfp, in cfs_trace_refill_stock() argument
130 tage = cfs_tage_alloc(gfp); in cfs_trace_refill_stock()
/linux-4.1.27/drivers/scsi/cxgbi/
Dlibcxgbi.h418 static inline struct sk_buff *alloc_wr(int wrlen, int dlen, gfp_t gfp) in alloc_wr() argument
420 struct sk_buff *skb = alloc_skb(wrlen + dlen, gfp); in alloc_wr()
676 gfp_t gfp) in cxgbi_alloc_big_mem() argument
678 void *p = kzalloc(size, gfp | __GFP_NOWARN); in cxgbi_alloc_big_mem()
/linux-4.1.27/drivers/md/
Ddm-bio-prison.h60 gfp_t gfp);
/linux-4.1.27/arch/arm/include/asm/
Ddma-mapping.h184 dma_addr_t *handle, gfp_t gfp) in dma_alloc_noncoherent() argument
210 gfp_t gfp, struct dma_attrs *attrs);
/linux-4.1.27/drivers/infiniband/hw/cxgb3/
Diwch_cm.c109 static struct sk_buff *get_skb(struct sk_buff *skb, int len, gfp_t gfp);
266 static void *alloc_ep(int size, gfp_t gfp) in alloc_ep() argument
270 epc = kzalloc(size, gfp); in alloc_ep()
324 static struct sk_buff *get_skb(struct sk_buff *skb, int len, gfp_t gfp) in get_skb() argument
330 skb = alloc_skb(len, gfp); in get_skb()
387 static int send_halfclose(struct iwch_ep *ep, gfp_t gfp) in send_halfclose() argument
393 skb = get_skb(NULL, sizeof(*req), gfp); in send_halfclose()
407 static int send_abort(struct iwch_ep *ep, struct sk_buff *skb, gfp_t gfp) in send_abort() argument
412 skb = get_skb(skb, sizeof(*req), gfp); in send_abort()
661 static void abort_connection(struct iwch_ep *ep, struct sk_buff *skb, gfp_t gfp) in abort_connection() argument
[all …]
/linux-4.1.27/drivers/crypto/qce/
Dablkcipher.c72 gfp_t gfp; in qce_ablkcipher_async_req_handle() local
95 gfp = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ? in qce_ablkcipher_async_req_handle()
98 ret = sg_alloc_table(&rctx->dst_tbl, rctx->dst_nents, gfp); in qce_ablkcipher_async_req_handle()
/linux-4.1.27/drivers/net/wireless/rtlwifi/
Drc.c258 struct ieee80211_sta *sta, gfp_t gfp) in rtl_rate_alloc_sta() argument
263 rate_priv = kzalloc(sizeof(struct rtl_rate_priv), gfp); in rtl_rate_alloc_sta()
/linux-4.1.27/security/selinux/include/
Dxfrm.h14 gfp_t gfp);
/linux-4.1.27/drivers/infiniband/core/
Diwpm_util.h104 u8 nl_client, gfp_t gfp);
/linux-4.1.27/drivers/net/wireless/ath/carl9170/
Dusb.c314 static int carl9170_usb_submit_rx_urb(struct ar9170 *ar, gfp_t gfp) in carl9170_usb_submit_rx_urb() argument
325 err = usb_submit_urb(urb, gfp); in carl9170_usb_submit_rx_urb()
457 static struct urb *carl9170_usb_alloc_rx_urb(struct ar9170 *ar, gfp_t gfp) in carl9170_usb_alloc_rx_urb() argument
462 buf = kmalloc(ar->fw.rx_size, gfp); in carl9170_usb_alloc_rx_urb()
466 urb = usb_alloc_urb(0, gfp); in carl9170_usb_alloc_rx_urb()
/linux-4.1.27/drivers/power/
Dpower_supply_sysfs.c256 static char *kstruprdup(const char *str, gfp_t gfp) in kstruprdup() argument
260 ustr = ret = kmalloc(strlen(str) + 1, gfp); in kstruprdup()
/linux-4.1.27/security/tomoyo/
Dtomoyo.c18 static int tomoyo_cred_alloc_blank(struct cred *new, gfp_t gfp) in tomoyo_cred_alloc_blank() argument
34 gfp_t gfp) in tomoyo_cred_prepare() argument
/linux-4.1.27/arch/c6x/mm/
Ddma-coherent.c77 dma_addr_t *handle, gfp_t gfp) in dma_alloc_coherent() argument
/linux-4.1.27/drivers/mfd/
Ddln2.c306 int *obuf_len, gfp_t gfp) in dln2_prep_buf() argument
313 buf = kmalloc(len, gfp); in dln2_prep_buf()
633 static int dln2_start_rx_urbs(struct dln2_dev *dln2, gfp_t gfp) in dln2_start_rx_urbs() argument
640 ret = usb_submit_urb(dln2->rx_urb[i], gfp); in dln2_start_rx_urbs()
/linux-4.1.27/fs/
Dposix_acl.c483 __posix_acl_create(struct posix_acl **acl, gfp_t gfp, umode_t *mode_p) in __posix_acl_create() argument
485 struct posix_acl *clone = posix_acl_clone(*acl, gfp); in __posix_acl_create()
501 __posix_acl_chmod(struct posix_acl **acl, gfp_t gfp, umode_t mode) in __posix_acl_chmod() argument
503 struct posix_acl *clone = posix_acl_clone(*acl, gfp); in __posix_acl_chmod()
/linux-4.1.27/security/apparmor/
Daudit.c172 int aa_audit(int type, struct aa_profile *profile, gfp_t gfp, in aa_audit() argument
Dlsm.c58 static int apparmor_cred_alloc_blank(struct cred *cred, gfp_t gfp) in apparmor_cred_alloc_blank() argument
61 struct aa_task_cxt *cxt = aa_alloc_task_context(gfp); in apparmor_cred_alloc_blank()
73 gfp_t gfp) in apparmor_cred_prepare() argument
76 struct aa_task_cxt *cxt = aa_alloc_task_context(gfp); in apparmor_cred_prepare()
/linux-4.1.27/drivers/net/ethernet/amd/xgbe/
Dxgbe-desc.c259 struct xgbe_page_alloc *pa, gfp_t gfp, int order) in xgbe_alloc_pages() argument
266 gfp |= __GFP_COLD | __GFP_COMP | __GFP_NOWARN; in xgbe_alloc_pages()
268 pages = alloc_pages(gfp, order); in xgbe_alloc_pages()
/linux-4.1.27/drivers/dma/sh/
Dusb-dmac.c267 gfp_t gfp) in usb_dmac_desc_alloc() argument
272 desc = kzalloc(sizeof(*desc) + sg_len * sizeof(desc->sg[0]), gfp); in usb_dmac_desc_alloc()
302 unsigned int sg_len, gfp_t gfp) in usb_dmac_desc_get() argument
319 if (!usb_dmac_desc_alloc(chan, sg_len, gfp)) { in usb_dmac_desc_get()
/linux-4.1.27/fs/cachefiles/
Drdwr.c392 gfp_t gfp) in cachefiles_read_or_alloc_page() argument
681 gfp_t gfp) in cachefiles_read_or_alloc_pages() argument
798 gfp_t gfp) in cachefiles_allocate_page() argument
836 gfp_t gfp) in cachefiles_allocate_pages() argument
/linux-4.1.27/arch/blackfin/kernel/
Ddma-mapping.c82 dma_addr_t *dma_handle, gfp_t gfp) in dma_alloc_coherent() argument
/linux-4.1.27/fs/nfs/
Dfscache.c260 int nfs_fscache_release_page(struct page *page, gfp_t gfp) in nfs_fscache_release_page() argument
269 if (!fscache_maybe_release_page(cookie, page, gfp)) in nfs_fscache_release_page()

12