Lines Matching refs:buff
1139 struct sk_buff *buff; in tcp_fragment() local
1155 buff = sk_stream_alloc_skb(sk, nsize, gfp, true); in tcp_fragment()
1156 if (!buff) in tcp_fragment()
1159 sk->sk_wmem_queued += buff->truesize; in tcp_fragment()
1160 sk_mem_charge(sk, buff->truesize); in tcp_fragment()
1162 buff->truesize += nlen; in tcp_fragment()
1166 TCP_SKB_CB(buff)->seq = TCP_SKB_CB(skb)->seq + len; in tcp_fragment()
1167 TCP_SKB_CB(buff)->end_seq = TCP_SKB_CB(skb)->end_seq; in tcp_fragment()
1168 TCP_SKB_CB(skb)->end_seq = TCP_SKB_CB(buff)->seq; in tcp_fragment()
1173 TCP_SKB_CB(buff)->tcp_flags = flags; in tcp_fragment()
1174 TCP_SKB_CB(buff)->sacked = TCP_SKB_CB(skb)->sacked; in tcp_fragment()
1178 buff->csum = csum_partial_copy_nocheck(skb->data + len, in tcp_fragment()
1179 skb_put(buff, nsize), in tcp_fragment()
1184 skb->csum = csum_block_sub(skb->csum, buff->csum, len); in tcp_fragment()
1187 skb_split(skb, buff, len); in tcp_fragment()
1190 buff->ip_summed = skb->ip_summed; in tcp_fragment()
1192 buff->tstamp = skb->tstamp; in tcp_fragment()
1193 tcp_fragment_tstamp(skb, buff); in tcp_fragment()
1199 tcp_set_skb_tso_segs(buff, mss_now); in tcp_fragment()
1204 if (!before(tp->snd_nxt, TCP_SKB_CB(buff)->end_seq)) { in tcp_fragment()
1206 tcp_skb_pcount(buff); in tcp_fragment()
1213 __skb_header_release(buff); in tcp_fragment()
1214 tcp_insert_write_queue_after(skb, buff, sk); in tcp_fragment()
1705 struct sk_buff *buff; in tso_fragment() local
1713 buff = sk_stream_alloc_skb(sk, 0, gfp, true); in tso_fragment()
1714 if (unlikely(!buff)) in tso_fragment()
1717 sk->sk_wmem_queued += buff->truesize; in tso_fragment()
1718 sk_mem_charge(sk, buff->truesize); in tso_fragment()
1719 buff->truesize += nlen; in tso_fragment()
1723 TCP_SKB_CB(buff)->seq = TCP_SKB_CB(skb)->seq + len; in tso_fragment()
1724 TCP_SKB_CB(buff)->end_seq = TCP_SKB_CB(skb)->end_seq; in tso_fragment()
1725 TCP_SKB_CB(skb)->end_seq = TCP_SKB_CB(buff)->seq; in tso_fragment()
1730 TCP_SKB_CB(buff)->tcp_flags = flags; in tso_fragment()
1733 TCP_SKB_CB(buff)->sacked = 0; in tso_fragment()
1735 buff->ip_summed = skb->ip_summed = CHECKSUM_PARTIAL; in tso_fragment()
1736 skb_split(skb, buff, len); in tso_fragment()
1737 tcp_fragment_tstamp(skb, buff); in tso_fragment()
1741 tcp_set_skb_tso_segs(buff, mss_now); in tso_fragment()
1744 __skb_header_release(buff); in tso_fragment()
1745 tcp_insert_write_queue_after(skb, buff, sk); in tso_fragment()
3245 struct sk_buff *buff; in tcp_connect() local
3255 buff = sk_stream_alloc_skb(sk, 0, sk->sk_allocation, true); in tcp_connect()
3256 if (unlikely(!buff)) in tcp_connect()
3259 tcp_init_nondata_skb(buff, tp->write_seq++, TCPHDR_SYN); in tcp_connect()
3261 tcp_connect_queue_skb(sk, buff); in tcp_connect()
3262 tcp_ecn_send_syn(sk, buff); in tcp_connect()
3265 err = tp->fastopen_req ? tcp_send_syn_data(sk, buff) : in tcp_connect()
3266 tcp_transmit_skb(sk, buff, 1, sk->sk_allocation); in tcp_connect()
3346 struct sk_buff *buff; in tcp_send_ack() local
3358 buff = alloc_skb(MAX_TCP_HEADER, sk_gfp_atomic(sk, GFP_ATOMIC)); in tcp_send_ack()
3359 if (!buff) { in tcp_send_ack()
3368 skb_reserve(buff, MAX_TCP_HEADER); in tcp_send_ack()
3369 tcp_init_nondata_skb(buff, tcp_acceptable_seq(sk), TCPHDR_ACK); in tcp_send_ack()
3377 skb_set_tcp_pure_ack(buff); in tcp_send_ack()
3380 skb_mstamp_get(&buff->skb_mstamp); in tcp_send_ack()
3381 tcp_transmit_skb(sk, buff, 0, sk_gfp_atomic(sk, GFP_ATOMIC)); in tcp_send_ack()