Lines Matching refs:gfp_mask
160 struct sk_buff *__alloc_skb_head(gfp_t gfp_mask, int node) in __alloc_skb_head() argument
166 gfp_mask & ~__GFP_DMA, node); in __alloc_skb_head()
202 struct sk_buff *__alloc_skb(unsigned int size, gfp_t gfp_mask, in __alloc_skb() argument
215 gfp_mask |= __GFP_MEMALLOC; in __alloc_skb()
218 skb = kmem_cache_alloc_node(cache, gfp_mask & ~__GFP_DMA, node); in __alloc_skb()
230 data = kmalloc_reserve(size, gfp_mask, node, &pfmemalloc); in __alloc_skb()
355 static void *__netdev_alloc_frag(unsigned int fragsz, gfp_t gfp_mask) in __netdev_alloc_frag() argument
363 data = __alloc_page_frag(nc, fragsz, gfp_mask); in __netdev_alloc_frag()
381 static void *__napi_alloc_frag(unsigned int fragsz, gfp_t gfp_mask) in __napi_alloc_frag() argument
385 return __alloc_page_frag(nc, fragsz, gfp_mask); in __napi_alloc_frag()
408 gfp_t gfp_mask) in __netdev_alloc_skb() argument
419 (gfp_mask & (__GFP_DIRECT_RECLAIM | GFP_DMA))) { in __netdev_alloc_skb()
420 skb = __alloc_skb(len, gfp_mask, SKB_ALLOC_RX, NUMA_NO_NODE); in __netdev_alloc_skb()
430 gfp_mask |= __GFP_MEMALLOC; in __netdev_alloc_skb()
435 data = __alloc_page_frag(nc, len, gfp_mask); in __netdev_alloc_skb()
477 gfp_t gfp_mask) in __napi_alloc_skb() argument
486 (gfp_mask & (__GFP_DIRECT_RECLAIM | GFP_DMA))) { in __napi_alloc_skb()
487 skb = __alloc_skb(len, gfp_mask, SKB_ALLOC_RX, NUMA_NO_NODE); in __napi_alloc_skb()
497 gfp_mask |= __GFP_MEMALLOC; in __napi_alloc_skb()
499 data = __alloc_page_frag(nc, len, gfp_mask); in __napi_alloc_skb()
878 int skb_copy_ubufs(struct sk_buff *skb, gfp_t gfp_mask) in skb_copy_ubufs() argument
889 page = alloc_page(gfp_mask); in skb_copy_ubufs()
938 struct sk_buff *skb_clone(struct sk_buff *skb, gfp_t gfp_mask) in skb_clone() argument
945 if (skb_orphan_frags(skb, gfp_mask)) in skb_clone()
954 gfp_mask |= __GFP_MEMALLOC; in skb_clone()
956 n = kmem_cache_alloc(skbuff_head_cache, gfp_mask); in skb_clone()
1016 struct sk_buff *skb_copy(const struct sk_buff *skb, gfp_t gfp_mask) in skb_copy() argument
1020 struct sk_buff *n = __alloc_skb(size, gfp_mask, in skb_copy()
1057 gfp_t gfp_mask, bool fclone) in __pskb_copy_fclone() argument
1061 struct sk_buff *n = __alloc_skb(size, gfp_mask, flags, NUMA_NO_NODE); in __pskb_copy_fclone()
1080 if (skb_orphan_frags(skb, gfp_mask)) { in __pskb_copy_fclone()
1120 gfp_t gfp_mask) in pskb_expand_head() argument
1135 gfp_mask |= __GFP_MEMALLOC; in pskb_expand_head()
1137 gfp_mask, NUMA_NO_NODE, NULL); in pskb_expand_head()
1158 if (skb_orphan_frags(skb, gfp_mask)) in pskb_expand_head()
1237 gfp_t gfp_mask) in skb_copy_expand() argument
1243 gfp_mask, skb_alloc_rx_flag(skb), in skb_copy_expand()
4461 gfp_t gfp_mask) in alloc_skb_with_frags() argument
4477 gfp_head = gfp_mask; in alloc_skb_with_frags()
4493 page = alloc_pages((gfp_mask & ~__GFP_DIRECT_RECLAIM) | in alloc_skb_with_frags()
4506 page = alloc_page(gfp_mask); in alloc_skb_with_frags()