Lines Matching refs:gfp_mask
160 struct sk_buff *__alloc_skb_head(gfp_t gfp_mask, int node) in __alloc_skb_head() argument
166 gfp_mask & ~__GFP_DMA, node); in __alloc_skb_head()
202 struct sk_buff *__alloc_skb(unsigned int size, gfp_t gfp_mask, in __alloc_skb() argument
215 gfp_mask |= __GFP_MEMALLOC; in __alloc_skb()
218 skb = kmem_cache_alloc_node(cache, gfp_mask & ~__GFP_DMA, node); in __alloc_skb()
230 data = kmalloc_reserve(size, gfp_mask, node, &pfmemalloc); in __alloc_skb()
363 gfp_t gfp_mask) in __page_frag_refill() argument
367 gfp_t gfp = gfp_mask; in __page_frag_refill()
370 gfp_mask |= __GFP_COMP | __GFP_NOWARN | __GFP_NORETRY | in __page_frag_refill()
372 page = alloc_pages_node(NUMA_NO_NODE, gfp_mask, order); in __page_frag_refill()
385 unsigned int fragsz, gfp_t gfp_mask) in __alloc_page_frag() argument
394 page = __page_frag_refill(nc, gfp_mask); in __alloc_page_frag()
433 static void *__netdev_alloc_frag(unsigned int fragsz, gfp_t gfp_mask) in __netdev_alloc_frag() argument
439 data = __alloc_page_frag(&netdev_alloc_cache, fragsz, gfp_mask); in __netdev_alloc_frag()
457 static void *__napi_alloc_frag(unsigned int fragsz, gfp_t gfp_mask) in __napi_alloc_frag() argument
459 return __alloc_page_frag(&napi_alloc_cache, fragsz, gfp_mask); in __napi_alloc_frag()
484 static struct sk_buff *__alloc_rx_skb(unsigned int length, gfp_t gfp_mask, in __alloc_rx_skb() argument
491 if (fragsz <= PAGE_SIZE && !(gfp_mask & (__GFP_WAIT | GFP_DMA))) { in __alloc_rx_skb()
495 gfp_mask |= __GFP_MEMALLOC; in __alloc_rx_skb()
498 __napi_alloc_frag(fragsz, gfp_mask) : in __alloc_rx_skb()
499 __netdev_alloc_frag(fragsz, gfp_mask); in __alloc_rx_skb()
507 skb = __alloc_skb(length, gfp_mask, in __alloc_rx_skb()
527 unsigned int length, gfp_t gfp_mask) in __netdev_alloc_skb() argument
532 skb = __alloc_rx_skb(length, gfp_mask, 0); in __netdev_alloc_skb()
557 unsigned int length, gfp_t gfp_mask) in __napi_alloc_skb() argument
562 skb = __alloc_rx_skb(length, gfp_mask, SKB_ALLOC_NAPI); in __napi_alloc_skb()
926 int skb_copy_ubufs(struct sk_buff *skb, gfp_t gfp_mask) in skb_copy_ubufs() argument
937 page = alloc_page(gfp_mask); in skb_copy_ubufs()
986 struct sk_buff *skb_clone(struct sk_buff *skb, gfp_t gfp_mask) in skb_clone() argument
993 if (skb_orphan_frags(skb, gfp_mask)) in skb_clone()
1002 gfp_mask |= __GFP_MEMALLOC; in skb_clone()
1004 n = kmem_cache_alloc(skbuff_head_cache, gfp_mask); in skb_clone()
1064 struct sk_buff *skb_copy(const struct sk_buff *skb, gfp_t gfp_mask) in skb_copy() argument
1068 struct sk_buff *n = __alloc_skb(size, gfp_mask, in skb_copy()
1105 gfp_t gfp_mask, bool fclone) in __pskb_copy_fclone() argument
1109 struct sk_buff *n = __alloc_skb(size, gfp_mask, flags, NUMA_NO_NODE); in __pskb_copy_fclone()
1128 if (skb_orphan_frags(skb, gfp_mask)) { in __pskb_copy_fclone()
1168 gfp_t gfp_mask) in pskb_expand_head() argument
1183 gfp_mask |= __GFP_MEMALLOC; in pskb_expand_head()
1185 gfp_mask, NUMA_NO_NODE, NULL); in pskb_expand_head()
1206 if (skb_orphan_frags(skb, gfp_mask)) in pskb_expand_head()
1285 gfp_t gfp_mask) in skb_copy_expand() argument
1291 gfp_mask, skb_alloc_rx_flag(skb), in skb_copy_expand()
4374 gfp_t gfp_mask) in alloc_skb_with_frags() argument
4390 gfp_head = gfp_mask; in alloc_skb_with_frags()
4406 page = alloc_pages((gfp_mask & ~__GFP_WAIT) | in alloc_skb_with_frags()
4419 page = alloc_page(gfp_mask); in alloc_skb_with_frags()