uh 175 arch/sparc/include/asm/uaccess_32.h __get_user_asm(__gu_val, uh, addr, __gu_ret); \ uh 201 arch/sparc/include/asm/uaccess_32.h case 2: __get_user_asm(__gu_val, uh, addr, __gu_ret); break; \ uh 143 arch/sparc/include/asm/uaccess_64.h case 2: __get_user_asm(__gu_val, uh, addr, __gu_ret); break; \ uh 1297 drivers/net/ethernet/broadcom/bnxt/bnxt.c struct udphdr *uh = NULL; uh 1303 drivers/net/ethernet/broadcom/bnxt/bnxt.c uh = (struct udphdr *)(iph + 1); uh 1308 drivers/net/ethernet/broadcom/bnxt/bnxt.c uh = (struct udphdr *)(iph + 1); uh 1310 drivers/net/ethernet/broadcom/bnxt/bnxt.c if (uh) { uh 1311 drivers/net/ethernet/broadcom/bnxt/bnxt.c if (uh->check) uh 60 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ptp.c struct udphdr *uh; uh 62 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ptp.c uh = udp_hdr(skb); uh 67 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ptp.c uh->dest == htons(PTP_EVENT_PORT); uh 86 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ptp.c struct udphdr *uh = (struct udphdr *)(skb->data + ETH_HLEN + uh 89 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ptp.c return uh->dest == htons(PTP_EVENT_PORT) && uh 90 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ptp.c uh->source == htons(PTP_EVENT_PORT); uh 1432 fs/ntfs/super.c USN_HEADER *uh; uh 1532 fs/ntfs/super.c uh = (USN_HEADER*)page_address(page); uh 1534 fs/ntfs/super.c if (unlikely(sle64_to_cpu(uh->allocation_delta) > uh 1535 fs/ntfs/super.c sle64_to_cpu(uh->maximum_size))) { uh 1538 fs/ntfs/super.c (long long)sle64_to_cpu(uh->allocation_delta), uh 1539 fs/ntfs/super.c (long long)sle64_to_cpu(uh->maximum_size)); uh 1547 fs/ntfs/super.c if (unlikely(sle64_to_cpu(uh->lowest_valid_usn) >= uh 1549 fs/ntfs/super.c if (likely(sle64_to_cpu(uh->lowest_valid_usn) == uh 1562 fs/ntfs/super.c (long long)sle64_to_cpu(uh->lowest_valid_usn), uh 39 fs/ntfs/usnjrnl.c USN_HEADER *uh; uh 47 fs/ntfs/usnjrnl.c uh = (USN_HEADER*)page_address(page); uh 53 fs/ntfs/usnjrnl.c (long long)sle64_to_cpu(uh->journal_id), uh 54 fs/ntfs/usnjrnl.c (long long)sle64_to_cpu(uh->lowest_valid_usn), uh 57 fs/ntfs/usnjrnl.c uh->lowest_valid_usn = uh 59 fs/ntfs/usnjrnl.c uh->journal_id = stamp; uh 418 include/linux/sctp.h struct sctp_chunkhdr uh; uh 100 include/net/ip6_checksum.h int udp6_csum_init(struct sk_buff *skb, struct udphdr *uh, int proto); uh 170 include/net/udp.h struct udphdr *uh, udp_lookup_t lookup); uh 178 include/net/udp.h struct udphdr *uh; uh 182 include/net/udp.h hlen = off + sizeof(*uh); uh 183 include/net/udp.h uh = skb_gro_header_fast(skb, off); uh 185 include/net/udp.h uh = skb_gro_header_slow(skb, hlen, off); uh 187 include/net/udp.h return uh; uh 38 include/net/udplite.h static inline int udplite_checksum_init(struct sk_buff *skb, struct udphdr *uh) uh 45 include/net/udplite.h if (uh->check == 0) { uh 50 include/net/udplite.h cscov = ntohs(uh->len); uh 33 lib/test_blackhole_dev.c struct udphdr *uh; uh 50 lib/test_blackhole_dev.c uh = (struct udphdr *)skb_push(skb, sizeof(struct udphdr)); uh 52 lib/test_blackhole_dev.c uh->source = uh->dest = htons(UDP_PORT); uh 53 lib/test_blackhole_dev.c uh->len = htons(data_len); uh 54 lib/test_blackhole_dev.c uh->check = 0; uh 40 net/6lowpan/nhc_udp.c struct udphdr uh; uh 49 net/6lowpan/nhc_udp.c fail |= lowpan_fetch_skb(skb, &uh.source, sizeof(uh.source)); uh 50 net/6lowpan/nhc_udp.c fail |= lowpan_fetch_skb(skb, &uh.dest, sizeof(uh.dest)); uh 53 net/6lowpan/nhc_udp.c fail |= lowpan_fetch_skb(skb, &uh.source, sizeof(uh.source)); uh 55 net/6lowpan/nhc_udp.c uh.dest = htons(val + LOWPAN_NHC_UDP_8BIT_PORT); uh 59 net/6lowpan/nhc_udp.c uh.source = htons(val + LOWPAN_NHC_UDP_8BIT_PORT); uh 60 net/6lowpan/nhc_udp.c fail |= lowpan_fetch_skb(skb, &uh.dest, sizeof(uh.dest)); uh 64 net/6lowpan/nhc_udp.c uh.source = htons(LOWPAN_NHC_UDP_4BIT_PORT + (val >> 4)); uh 65 net/6lowpan/nhc_udp.c uh.dest = htons(LOWPAN_NHC_UDP_4BIT_PORT + (val & 0x0f)); uh 72 net/6lowpan/nhc_udp.c ntohs(uh.source), ntohs(uh.dest)); uh 79 net/6lowpan/nhc_udp.c fail |= lowpan_fetch_skb(skb, &uh.check, sizeof(uh.check)); uh 92 net/6lowpan/nhc_udp.c uh.len = htons(lowpan_802154_cb(skb)->d_size - uh 95 net/6lowpan/nhc_udp.c uh.len = htons(skb->len + sizeof(struct udphdr)); uh 98 net/6lowpan/nhc_udp.c uh.len = htons(skb->len + sizeof(struct udphdr)); uh 101 net/6lowpan/nhc_udp.c pr_debug("uncompressed UDP length: src = %d", ntohs(uh.len)); uh 111 net/6lowpan/nhc_udp.c skb_copy_to_linear_data(skb, &uh, sizeof(struct udphdr)); uh 118 net/6lowpan/nhc_udp.c const struct udphdr *uh = udp_hdr(skb); uh 121 net/6lowpan/nhc_udp.c if (((ntohs(uh->source) & LOWPAN_NHC_UDP_4BIT_MASK) == uh 123 net/6lowpan/nhc_udp.c ((ntohs(uh->dest) & LOWPAN_NHC_UDP_4BIT_MASK) == uh 130 net/6lowpan/nhc_udp.c tmp = ntohs(uh->dest) - LOWPAN_NHC_UDP_4BIT_PORT + uh 131 net/6lowpan/nhc_udp.c ((ntohs(uh->source) - LOWPAN_NHC_UDP_4BIT_PORT) << 4); uh 133 net/6lowpan/nhc_udp.c } else if ((ntohs(uh->dest) & LOWPAN_NHC_UDP_8BIT_MASK) == uh 140 net/6lowpan/nhc_udp.c lowpan_push_hc_data(hc_ptr, &uh->source, sizeof(uh->source)); uh 142 net/6lowpan/nhc_udp.c tmp = ntohs(uh->dest) - LOWPAN_NHC_UDP_8BIT_PORT; uh 144 net/6lowpan/nhc_udp.c } else if ((ntohs(uh->source) & LOWPAN_NHC_UDP_8BIT_MASK) == uh 151 net/6lowpan/nhc_udp.c tmp = ntohs(uh->source) - LOWPAN_NHC_UDP_8BIT_PORT; uh 154 net/6lowpan/nhc_udp.c lowpan_push_hc_data(hc_ptr, &uh->dest, sizeof(uh->dest)); uh 161 net/6lowpan/nhc_udp.c lowpan_push_hc_data(hc_ptr, &uh->source, sizeof(uh->source)); uh 163 net/6lowpan/nhc_udp.c lowpan_push_hc_data(hc_ptr, &uh->dest, sizeof(uh->dest)); uh 167 net/6lowpan/nhc_udp.c lowpan_push_hc_data(hc_ptr, &uh->check, sizeof(uh->check)); uh 231 net/ipv4/esp4.c struct udphdr *uh; uh 248 net/ipv4/esp4.c uh = (struct udphdr *)esph; uh 249 net/ipv4/esp4.c uh->source = sport; uh 250 net/ipv4/esp4.c uh->dest = dport; uh 251 net/ipv4/esp4.c uh->len = htons(len); uh 252 net/ipv4/esp4.c uh->check = 0; uh 257 net/ipv4/esp4.c esph = (struct ip_esp_hdr *)(uh + 1); uh 260 net/ipv4/esp4.c udpdata32 = (__be32 *)(uh + 1); uh 603 net/ipv4/esp4.c struct udphdr *uh = (void *)(skb_network_header(skb) + ihl); uh 612 net/ipv4/esp4.c uh->source != encap->encap_sport) { uh 616 net/ipv4/esp4.c km_new_mapping(x, &ipaddr, uh->source); uh 1065 net/ipv4/fou.c struct udphdr *uh; uh 1070 net/ipv4/fou.c uh = udp_hdr(skb); uh 1072 net/ipv4/fou.c uh->dest = e->dport; uh 1073 net/ipv4/fou.c uh->source = sport; uh 1074 net/ipv4/fou.c uh->len = htons(skb->len); uh 584 net/ipv4/udp.c struct udphdr *uh, uh 600 net/ipv4/udp.c sk = __udp4_lib_lookup(net, iph->daddr, uh->source, uh 601 net/ipv4/udp.c iph->saddr, uh->dest, skb->dev->ifindex, 0, uh 636 net/ipv4/udp.c struct udphdr *uh = (struct udphdr *)(skb->data+(iph->ihl<<2)); uh 645 net/ipv4/udp.c sk = __udp4_lib_lookup(net, iph->daddr, uh->dest, uh 646 net/ipv4/udp.c iph->saddr, uh->source, skb->dev->ifindex, uh 652 net/ipv4/udp.c sk = __udp4_lib_err_encap(net, iph, uh, udptable, skb, uh 714 net/ipv4/udp.c ip_icmp_error(sk, skb, err, uh->dest, info, (u8 *)(uh+1)); uh 751 net/ipv4/udp.c struct udphdr *uh = udp_hdr(skb); uh 763 net/ipv4/udp.c uh->check = ~csum_tcpudp_magic(src, dst, len, uh 781 net/ipv4/udp.c uh->check = csum_tcpudp_magic(src, dst, len, IPPROTO_UDP, csum); uh 782 net/ipv4/udp.c if (uh->check == 0) uh 783 net/ipv4/udp.c uh->check = CSUM_MANGLED_0; uh 794 net/ipv4/udp.c struct udphdr *uh = udp_hdr(skb); uh 797 net/ipv4/udp.c uh->check = 0; uh 799 net/ipv4/udp.c uh->check = ~udp_v4_check(len, saddr, daddr, 0); uh 801 net/ipv4/udp.c uh->check = 0; uh 802 net/ipv4/udp.c uh->check = udp_v4_check(len, saddr, daddr, lco_csum(skb)); uh 803 net/ipv4/udp.c if (uh->check == 0) uh 804 net/ipv4/udp.c uh->check = CSUM_MANGLED_0; uh 809 net/ipv4/udp.c uh->check = ~udp_v4_check(len, saddr, daddr, 0); uh 819 net/ipv4/udp.c struct udphdr *uh; uh 824 net/ipv4/udp.c int datalen = len - sizeof(*uh); uh 830 net/ipv4/udp.c uh = udp_hdr(skb); uh 831 net/ipv4/udp.c uh->source = inet->inet_sport; uh 832 net/ipv4/udp.c uh->dest = fl4->fl4_dport; uh 833 net/ipv4/udp.c uh->len = htons(len); uh 834 net/ipv4/udp.c uh->check = 0; uh 885 net/ipv4/udp.c uh->check = csum_tcpudp_magic(fl4->saddr, fl4->daddr, len, uh 887 net/ipv4/udp.c if (uh->check == 0) uh 888 net/ipv4/udp.c uh->check = CSUM_MANGLED_0; uh 2144 net/ipv4/udp.c struct udphdr *uh, uh 2150 net/ipv4/udp.c unsigned short hnum = ntohs(uh->dest); uh 2169 net/ipv4/udp.c if (!__udp_is_mcast_sock(net, sk, uh->dest, daddr, uh 2170 net/ipv4/udp.c uh->source, saddr, dif, sdif, hnum)) uh 2213 net/ipv4/udp.c static inline int udp4_csum_init(struct sk_buff *skb, struct udphdr *uh, uh 2222 net/ipv4/udp.c err = udplite_checksum_init(skb, uh); uh 2235 net/ipv4/udp.c err = (__force int)skb_checksum_init_zero_check(skb, proto, uh->check, uh 2259 net/ipv4/udp.c struct udphdr *uh) uh 2263 net/ipv4/udp.c if (inet_get_convert_csum(sk) && uh->check && !IS_UDPLITE(sk)) uh 2284 net/ipv4/udp.c struct udphdr *uh; uh 2296 net/ipv4/udp.c uh = udp_hdr(skb); uh 2297 net/ipv4/udp.c ulen = ntohs(uh->len); uh 2306 net/ipv4/udp.c if (ulen < sizeof(*uh) || pskb_trim_rcsum(skb, ulen)) uh 2308 net/ipv4/udp.c uh = udp_hdr(skb); uh 2311 net/ipv4/udp.c if (udp4_csum_init(skb, uh, proto)) uh 2322 net/ipv4/udp.c ret = udp_unicast_rcv_skb(sk, skb, uh); uh 2328 net/ipv4/udp.c return __udp4_lib_mcast_deliver(net, skb, uh, uh 2331 net/ipv4/udp.c sk = __udp4_lib_lookup_skb(skb, uh->source, uh->dest, udptable); uh 2333 net/ipv4/udp.c return udp_unicast_rcv_skb(sk, skb, uh); uh 2356 net/ipv4/udp.c &saddr, ntohs(uh->source), uh 2358 net/ipv4/udp.c &daddr, ntohs(uh->dest)); uh 2368 net/ipv4/udp.c &saddr, ntohs(uh->source), &daddr, ntohs(uh->dest), uh 2439 net/ipv4/udp.c const struct udphdr *uh; uh 2451 net/ipv4/udp.c uh = udp_hdr(skb); uh 2464 net/ipv4/udp.c sk = __udp4_lib_mcast_demux_lookup(net, uh->dest, iph->daddr, uh 2465 net/ipv4/udp.c uh->source, iph->saddr, uh 2468 net/ipv4/udp.c sk = __udp4_lib_demux_lookup(net, uh->dest, iph->daddr, uh 2469 net/ipv4/udp.c uh->source, iph->saddr, dif, sdif); uh 23 net/ipv4/udp_offload.c struct udphdr *uh = udp_hdr(skb); uh 41 net/ipv4/udp_offload.c partial = (__force __wsum)uh->len; uh 44 net/ipv4/udp_offload.c partial = csum_sub(csum_unfold(uh->check), partial); uh 114 net/ipv4/udp_offload.c uh = udp_hdr(skb); uh 121 net/ipv4/udp_offload.c uh->len = htons(skb_shinfo(skb)->gso_size + uh 123 net/ipv4/udp_offload.c skb->head - (unsigned char *)uh); uh 125 net/ipv4/udp_offload.c uh->len = htons(len); uh 131 net/ipv4/udp_offload.c uh->check = ~csum_fold(csum_add(partial, uh 135 net/ipv4/udp_offload.c uh->check = gso_make_checksum(skb, ~uh->check); uh 136 net/ipv4/udp_offload.c if (uh->check == 0) uh 137 net/ipv4/udp_offload.c uh->check = CSUM_MANGLED_0; uh 193 net/ipv4/udp_offload.c struct udphdr *uh; uh 200 net/ipv4/udp_offload.c if (gso_skb->len <= sizeof(*uh) + mss) uh 203 net/ipv4/udp_offload.c skb_pull(gso_skb, sizeof(*uh)); uh 225 net/ipv4/udp_offload.c uh = udp_hdr(seg); uh 233 net/ipv4/udp_offload.c newlen = htons(sizeof(*uh) + mss); uh 234 net/ipv4/udp_offload.c check = csum16_add(csum16_sub(uh->check, uh->len), newlen); uh 246 net/ipv4/udp_offload.c uh->len = newlen; uh 247 net/ipv4/udp_offload.c uh->check = check; uh 252 net/ipv4/udp_offload.c uh->check = gso_make_checksum(seg, ~check) ? : uh 256 net/ipv4/udp_offload.c uh = udp_hdr(seg); uh 262 net/ipv4/udp_offload.c check = csum16_add(csum16_sub(uh->check, uh->len), newlen); uh 264 net/ipv4/udp_offload.c uh->len = newlen; uh 265 net/ipv4/udp_offload.c uh->check = check; uh 270 net/ipv4/udp_offload.c uh->check = gso_make_checksum(seg, ~check) ? : CSUM_MANGLED_0; uh 294 net/ipv4/udp_offload.c struct udphdr *uh; uh 322 net/ipv4/udp_offload.c uh = udp_hdr(skb); uh 325 net/ipv4/udp_offload.c uh->check = 0; uh 327 net/ipv4/udp_offload.c uh->check = udp_v4_check(skb->len, iph->saddr, iph->daddr, csum); uh 328 net/ipv4/udp_offload.c if (uh->check == 0) uh 329 net/ipv4/udp_offload.c uh->check = CSUM_MANGLED_0; uh 352 net/ipv4/udp_offload.c struct udphdr *uh = udp_hdr(skb); uh 359 net/ipv4/udp_offload.c if (!uh->check) { uh 365 net/ipv4/udp_offload.c ulen = ntohs(uh->len); uh 366 net/ipv4/udp_offload.c if (ulen <= sizeof(*uh) || ulen != skb_gro_len(skb)) { uh 372 net/ipv4/udp_offload.c skb_gro_postpull_rcsum(skb, uh, sizeof(struct udphdr)); uh 381 net/ipv4/udp_offload.c if ((*(u32 *)&uh->source != *(u32 *)&uh2->source)) { uh 407 net/ipv4/udp_offload.c struct udphdr *uh, udp_lookup_t lookup) uh 418 net/ipv4/udp_offload.c udp4_lib_lookup_skb, skb, uh->source, uh->dest); uh 449 net/ipv4/udp_offload.c if ((*(u32 *)&uh->source != *(u32 *)&uh2->source) || uh 450 net/ipv4/udp_offload.c (!uh->check ^ !uh2->check)) { uh 457 net/ipv4/udp_offload.c skb_gro_postpull_rcsum(skb, uh, sizeof(struct udphdr)); uh 470 net/ipv4/udp_offload.c struct udphdr *uh = udp_gro_udphdr(skb); uh 472 net/ipv4/udp_offload.c if (unlikely(!uh) || !static_branch_unlikely(&udp_encap_needed_key)) uh 479 net/ipv4/udp_offload.c if (skb_gro_checksum_validate_zero_check(skb, IPPROTO_UDP, uh->check, uh 482 net/ipv4/udp_offload.c else if (uh->check) uh 483 net/ipv4/udp_offload.c skb_gro_checksum_try_convert(skb, IPPROTO_UDP, uh->check, uh 487 net/ipv4/udp_offload.c return udp_gro_receive(head, skb, uh, udp4_lib_lookup_skb); uh 496 net/ipv4/udp_offload.c struct udphdr *uh = udp_hdr(skb); uh 498 net/ipv4/udp_offload.c skb->csum_start = (unsigned char *)uh - skb->head; uh 511 net/ipv4/udp_offload.c struct udphdr *uh = (struct udphdr *)(skb->data + nhoff); uh 515 net/ipv4/udp_offload.c uh->len = newlen; uh 519 net/ipv4/udp_offload.c udp4_lib_lookup_skb, skb, uh->source, uh->dest); uh 523 net/ipv4/udp_offload.c skb_shinfo(skb)->gso_type = uh->check ? SKB_GSO_UDP_TUNNEL_CSUM uh 545 net/ipv4/udp_offload.c struct udphdr *uh = (struct udphdr *)(skb->data + nhoff); uh 547 net/ipv4/udp_offload.c if (uh->check) uh 548 net/ipv4/udp_offload.c uh->check = ~udp_v4_check(skb->len - nhoff, iph->saddr, uh 178 net/ipv4/udp_tunnel.c struct udphdr *uh; uh 180 net/ipv4/udp_tunnel.c __skb_push(skb, sizeof(*uh)); uh 182 net/ipv4/udp_tunnel.c uh = udp_hdr(skb); uh 184 net/ipv4/udp_tunnel.c uh->dest = dst_port; uh 185 net/ipv4/udp_tunnel.c uh->source = src_port; uh 186 net/ipv4/udp_tunnel.c uh->len = htons(skb->len); uh 90 net/ipv4/xfrm4_input.c struct udphdr *uh; uh 109 net/ipv4/xfrm4_input.c uh = udp_hdr(skb); uh 110 net/ipv4/xfrm4_input.c udpdata = (__u8 *)uh + sizeof(struct udphdr); uh 24 net/ipv6/fou6.c struct udphdr *uh; uh 29 net/ipv6/fou6.c uh = udp_hdr(skb); uh 31 net/ipv6/fou6.c uh->dest = e->dport; uh 32 net/ipv6/fou6.c uh->source = sport; uh 33 net/ipv6/fou6.c uh->len = htons(skb->len); uh 94 net/ipv6/ila/ila_common.c struct udphdr *uh = (struct udphdr *) uh 97 net/ipv6/ila/ila_common.c if (uh->check || skb->ip_summed == CHECKSUM_PARTIAL) { uh 99 net/ipv6/ila/ila_common.c inet_proto_csum_replace_by_diff(&uh->check, skb, uh 101 net/ipv6/ila/ila_common.c if (!uh->check) uh 102 net/ipv6/ila/ila_common.c uh->check = CSUM_MANGLED_0; uh 65 net/ipv6/ip6_checksum.c int udp6_csum_init(struct sk_buff *skb, struct udphdr *uh, int proto) uh 73 net/ipv6/ip6_checksum.c err = udplite_checksum_init(skb, uh); uh 91 net/ipv6/ip6_checksum.c err = (__force int)skb_checksum_init_zero_check(skb, proto, uh->check, uh 119 net/ipv6/ip6_checksum.c struct udphdr *uh = udp_hdr(skb); uh 122 net/ipv6/ip6_checksum.c uh->check = 0; uh 124 net/ipv6/ip6_checksum.c uh->check = ~udp_v6_check(len, saddr, daddr, 0); uh 126 net/ipv6/ip6_checksum.c uh->check = 0; uh 127 net/ipv6/ip6_checksum.c uh->check = udp_v6_check(len, saddr, daddr, lco_csum(skb)); uh 128 net/ipv6/ip6_checksum.c if (uh->check == 0) uh 129 net/ipv6/ip6_checksum.c uh->check = CSUM_MANGLED_0; uh 134 net/ipv6/ip6_checksum.c uh->check = ~udp_v6_check(len, saddr, daddr, 0); uh 88 net/ipv6/ip6_udp_tunnel.c struct udphdr *uh; uh 91 net/ipv6/ip6_udp_tunnel.c __skb_push(skb, sizeof(*uh)); uh 93 net/ipv6/ip6_udp_tunnel.c uh = udp_hdr(skb); uh 95 net/ipv6/ip6_udp_tunnel.c uh->dest = dst_port; uh 96 net/ipv6/ip6_udp_tunnel.c uh->source = src_port; uh 98 net/ipv6/ip6_udp_tunnel.c uh->len = htons(skb->len); uh 450 net/ipv6/udp.c struct udphdr *uh, uh 468 net/ipv6/udp.c sk = __udp6_lib_lookup(net, &hdr->daddr, uh->source, uh 469 net/ipv6/udp.c &hdr->saddr, uh->dest, uh 499 net/ipv6/udp.c struct udphdr *uh = (struct udphdr *)(skb->data+offset); uh 506 net/ipv6/udp.c sk = __udp6_lib_lookup(net, daddr, uh->dest, saddr, uh->source, uh 512 net/ipv6/udp.c sk = __udp6_lib_err_encap(net, hdr, offset, uh, uh 556 net/ipv6/udp.c ipv6_icmp_error(sk, skb, err, uh->dest, ntohl(info), (u8 *)(uh+1)); uh 748 net/ipv6/udp.c const struct udphdr *uh = udp_hdr(skb); uh 749 net/ipv6/udp.c unsigned short hnum = ntohs(uh->dest); uh 768 net/ipv6/udp.c if (!__udp_v6_is_mcast_sock(net, sk, uh->dest, daddr, uh 769 net/ipv6/udp.c uh->source, saddr, dif, sdif, uh 775 net/ipv6/udp.c if (!uh->check && !udp_sk(sk)->no_check6_rx) uh 825 net/ipv6/udp.c struct udphdr *uh) uh 829 net/ipv6/udp.c if (inet_get_convert_csum(sk) && uh->check && !IS_UDPLITE(sk)) uh 845 net/ipv6/udp.c struct udphdr *uh; uh 854 net/ipv6/udp.c uh = udp_hdr(skb); uh 856 net/ipv6/udp.c ulen = ntohs(uh->len); uh 867 net/ipv6/udp.c if (ulen < sizeof(*uh)) uh 875 net/ipv6/udp.c uh = udp_hdr(skb); uh 879 net/ipv6/udp.c if (udp6_csum_init(skb, uh, proto)) uh 891 net/ipv6/udp.c if (!uh->check && !udp_sk(sk)->no_check6_rx) { uh 896 net/ipv6/udp.c ret = udp6_unicast_rcv_skb(sk, skb, uh); uh 909 net/ipv6/udp.c sk = __udp6_lib_lookup_skb(skb, uh->source, uh->dest, udptable); uh 911 net/ipv6/udp.c if (!uh->check && !udp_sk(sk)->no_check6_rx) uh 913 net/ipv6/udp.c return udp6_unicast_rcv_skb(sk, skb, uh); uh 916 net/ipv6/udp.c if (!uh->check) uh 934 net/ipv6/udp.c saddr, ntohs(uh->source), uh 936 net/ipv6/udp.c daddr, ntohs(uh->dest)); uh 975 net/ipv6/udp.c const struct udphdr *uh; uh 985 net/ipv6/udp.c uh = udp_hdr(skb); uh 988 net/ipv6/udp.c sk = __udp6_lib_demux_lookup(net, uh->dest, uh 990 net/ipv6/udp.c uh->source, &ipv6_hdr(skb)->saddr, uh 1066 net/ipv6/udp.c struct udphdr *uh = udp_hdr(skb); uh 1074 net/ipv6/udp.c uh->check = ~csum_ipv6_magic(saddr, daddr, len, IPPROTO_UDP, 0); uh 1091 net/ipv6/udp.c uh->check = csum_ipv6_magic(saddr, daddr, len, IPPROTO_UDP, uh 1093 net/ipv6/udp.c if (uh->check == 0) uh 1094 net/ipv6/udp.c uh->check = CSUM_MANGLED_0; uh 1106 net/ipv6/udp.c struct udphdr *uh; uh 1112 net/ipv6/udp.c int datalen = len - sizeof(*uh); uh 1117 net/ipv6/udp.c uh = udp_hdr(skb); uh 1118 net/ipv6/udp.c uh->source = fl6->fl6_sport; uh 1119 net/ipv6/udp.c uh->dest = fl6->fl6_dport; uh 1120 net/ipv6/udp.c uh->len = htons(len); uh 1121 net/ipv6/udp.c uh->check = 0; uh 1167 net/ipv6/udp.c uh->check = csum_ipv6_magic(&fl6->saddr, &fl6->daddr, uh 1169 net/ipv6/udp.c if (uh->check == 0) uh 1170 net/ipv6/udp.c uh->check = CSUM_MANGLED_0; uh 40 net/ipv6/udp_offload.c struct udphdr *uh; uh 55 net/ipv6/udp_offload.c uh = udp_hdr(skb); uh 58 net/ipv6/udp_offload.c uh->check = 0; uh 60 net/ipv6/udp_offload.c uh->check = udp_v6_check(skb->len, &ipv6h->saddr, uh 62 net/ipv6/udp_offload.c if (uh->check == 0) uh 63 net/ipv6/udp_offload.c uh->check = CSUM_MANGLED_0; uh 117 net/ipv6/udp_offload.c struct udphdr *uh = udp_gro_udphdr(skb); uh 119 net/ipv6/udp_offload.c if (unlikely(!uh) || !static_branch_unlikely(&udpv6_encap_needed_key)) uh 126 net/ipv6/udp_offload.c if (skb_gro_checksum_validate_zero_check(skb, IPPROTO_UDP, uh->check, uh 129 net/ipv6/udp_offload.c else if (uh->check) uh 130 net/ipv6/udp_offload.c skb_gro_checksum_try_convert(skb, IPPROTO_UDP, uh->check, uh 135 net/ipv6/udp_offload.c return udp_gro_receive(head, skb, uh, udp6_lib_lookup_skb); uh 145 net/ipv6/udp_offload.c struct udphdr *uh = (struct udphdr *)(skb->data + nhoff); uh 147 net/ipv6/udp_offload.c if (uh->check) uh 148 net/ipv6/udp_offload.c uh->check = ~udp_v6_check(skb->len - nhoff, &ipv6h->saddr, uh 1061 net/l2tp/l2tp_core.c struct udphdr *uh; uh 1113 net/l2tp/l2tp_core.c __skb_push(skb, sizeof(*uh)); uh 1115 net/l2tp/l2tp_core.c uh = udp_hdr(skb); uh 1116 net/l2tp/l2tp_core.c uh->source = inet->inet_sport; uh 1117 net/l2tp/l2tp_core.c uh->dest = inet->inet_dport; uh 1119 net/l2tp/l2tp_core.c uh->len = htons(udp_len); uh 56 net/netfilter/ipset/ip_set_getport.c const struct udphdr *uh; uh 58 net/netfilter/ipset/ip_set_getport.c uh = skb_header_pointer(skb, protooff, sizeof(_udph), &_udph); uh 59 net/netfilter/ipset/ip_set_getport.c if (!uh) uh 63 net/netfilter/ipset/ip_set_getport.c *port = src ? uh->source : uh->dest; uh 37 net/netfilter/ipvs/ip_vs_proto_udp.c struct udphdr _udph, *uh; uh 42 net/netfilter/ipvs/ip_vs_proto_udp.c uh = skb_header_pointer(skb, iph->len, sizeof(_udph), &_udph); uh 43 net/netfilter/ipvs/ip_vs_proto_udp.c if (uh) uh 44 net/netfilter/ipvs/ip_vs_proto_udp.c ports = &uh->source; uh 302 net/netfilter/ipvs/ip_vs_proto_udp.c struct udphdr _udph, *uh; uh 312 net/netfilter/ipvs/ip_vs_proto_udp.c uh = skb_header_pointer(skb, udphoff, sizeof(_udph), &_udph); uh 313 net/netfilter/ipvs/ip_vs_proto_udp.c if (uh == NULL) uh 316 net/netfilter/ipvs/ip_vs_proto_udp.c if (uh->check != 0) { uh 1211 net/netfilter/nf_conntrack_h323_main.c const struct udphdr *uh; uh 1215 net/netfilter/nf_conntrack_h323_main.c uh = skb_header_pointer(skb, protoff, sizeof(_uh), &_uh); uh 1216 net/netfilter/nf_conntrack_h323_main.c if (uh == NULL) uh 25 net/netfilter/nf_log_common.c const struct udphdr *uh; uh 37 net/netfilter/nf_log_common.c uh = skb_header_pointer(skb, offset, sizeof(_udph), &_udph); uh 38 net/netfilter/nf_log_common.c if (uh == NULL) { uh 46 net/netfilter/nf_log_common.c ntohs(uh->source), ntohs(uh->dest), ntohs(uh->len)); uh 283 net/netfilter/nf_nat_sip.c struct udphdr *uh; uh 290 net/netfilter/nf_nat_sip.c uh = (void *)skb->data + protoff; uh 291 net/netfilter/nf_nat_sip.c uh->dest = ct_sip_info->forced_dport; uh 404 net/netfilter/nft_payload.c struct udphdr *uh, _uh; uh 406 net/netfilter/nft_payload.c uh = skb_header_pointer(skb, thoff, sizeof(_uh), &_uh); uh 407 net/netfilter/nft_payload.c if (!uh) uh 410 net/netfilter/nft_payload.c return (__force bool)uh->check; uh 131 net/netfilter/xt_tcpudp.c const struct udphdr *uh; uh 139 net/netfilter/xt_tcpudp.c uh = skb_header_pointer(skb, par->thoff, sizeof(_udph), &_udph); uh 140 net/netfilter/xt_tcpudp.c if (uh == NULL) { uh 149 net/netfilter/xt_tcpudp.c ntohs(uh->source), uh 152 net/netfilter/xt_tcpudp.c ntohs(uh->dest), uh 359 net/openvswitch/actions.c struct udphdr *uh = udp_hdr(skb); uh 361 net/openvswitch/actions.c if (uh->check || skb->ip_summed == CHECKSUM_PARTIAL) { uh 362 net/openvswitch/actions.c inet_proto_csum_replace4(&uh->check, skb, uh 364 net/openvswitch/actions.c if (!uh->check) uh 365 net/openvswitch/actions.c uh->check = CSUM_MANGLED_0; uh 391 net/openvswitch/actions.c struct udphdr *uh = udp_hdr(skb); uh 393 net/openvswitch/actions.c if (uh->check || skb->ip_summed == CHECKSUM_PARTIAL) { uh 394 net/openvswitch/actions.c inet_proto_csum_replace16(&uh->check, skb, uh 396 net/openvswitch/actions.c if (!uh->check) uh 397 net/openvswitch/actions.c uh->check = CSUM_MANGLED_0; uh 642 net/openvswitch/actions.c struct udphdr *uh; uh 651 net/openvswitch/actions.c uh = udp_hdr(skb); uh 653 net/openvswitch/actions.c src = OVS_MASKED(uh->source, key->udp_src, mask->udp_src); uh 654 net/openvswitch/actions.c dst = OVS_MASKED(uh->dest, key->udp_dst, mask->udp_dst); uh 656 net/openvswitch/actions.c if (uh->check && skb->ip_summed != CHECKSUM_PARTIAL) { uh 657 net/openvswitch/actions.c if (likely(src != uh->source)) { uh 658 net/openvswitch/actions.c set_tp_port(skb, &uh->source, src, &uh->check); uh 661 net/openvswitch/actions.c if (likely(dst != uh->dest)) { uh 662 net/openvswitch/actions.c set_tp_port(skb, &uh->dest, dst, &uh->check); uh 666 net/openvswitch/actions.c if (unlikely(!uh->check)) uh 667 net/openvswitch/actions.c uh->check = CSUM_MANGLED_0; uh 669 net/openvswitch/actions.c uh->source = src; uh 670 net/openvswitch/actions.c uh->dest = dst; uh 49 samples/bpf/xdp_tx_iptunnel_kern.c struct udphdr *uh; uh 58 samples/bpf/xdp_tx_iptunnel_kern.c uh = (struct udphdr *)trans_data; uh 59 samples/bpf/xdp_tx_iptunnel_kern.c if (uh + 1 > data_end) uh 61 samples/bpf/xdp_tx_iptunnel_kern.c return uh->dest; uh 69 security/lsm_audit.c struct udphdr *uh = udp_hdr(skb); uh 70 security/lsm_audit.c if (uh == NULL) uh 73 security/lsm_audit.c ad->u.net->sport = uh->source; uh 74 security/lsm_audit.c ad->u.net->dport = uh->dest; uh 145 security/lsm_audit.c struct udphdr _udph, *uh; uh 147 security/lsm_audit.c uh = skb_header_pointer(skb, offset, sizeof(_udph), &_udph); uh 148 security/lsm_audit.c if (uh == NULL) uh 151 security/lsm_audit.c ad->u.net->sport = uh->source; uh 152 security/lsm_audit.c ad->u.net->dport = uh->dest; uh 4215 security/selinux/hooks.c struct udphdr _udph, *uh; uh 4221 security/selinux/hooks.c uh = skb_header_pointer(skb, offset, sizeof(_udph), &_udph); uh 4222 security/selinux/hooks.c if (uh == NULL) uh 4225 security/selinux/hooks.c ad->u.net->sport = uh->source; uh 4226 security/selinux/hooks.c ad->u.net->dport = uh->dest; uh 4313 security/selinux/hooks.c struct udphdr _udph, *uh; uh 4315 security/selinux/hooks.c uh = skb_header_pointer(skb, offset, sizeof(_udph), &_udph); uh 4316 security/selinux/hooks.c if (uh == NULL) uh 4319 security/selinux/hooks.c ad->u.net->sport = uh->source; uh 4320 security/selinux/hooks.c ad->u.net->dport = uh->dest; uh 3786 security/smack/smack_lsm.c struct udphdr _udph, *uh; uh 3812 security/smack/smack_lsm.c uh = skb_header_pointer(skb, offset, sizeof(_udph), &_udph); uh 3813 security/smack/smack_lsm.c if (uh != NULL) uh 3814 security/smack/smack_lsm.c sip->sin6_port = uh->source; uh 129 tools/testing/selftests/bpf/progs/test_select_reuseport_kern.c struct udphdr *uh = data; uh 131 tools/testing/selftests/bpf/progs/test_select_reuseport_kern.c if (uh + 1 > data_end) uh 134 tools/testing/selftests/bpf/progs/test_select_reuseport_kern.c data_check.skb_ports[0] = uh->source; uh 135 tools/testing/selftests/bpf/progs/test_select_reuseport_kern.c data_check.skb_ports[1] = uh->dest; uh 52 tools/testing/selftests/bpf/progs/test_xdp.c struct udphdr *uh; uh 61 tools/testing/selftests/bpf/progs/test_xdp.c uh = (struct udphdr *)trans_data; uh 62 tools/testing/selftests/bpf/progs/test_xdp.c if (uh + 1 > data_end) uh 64 tools/testing/selftests/bpf/progs/test_xdp.c return uh->dest; uh 48 tools/testing/selftests/bpf/progs/test_xdp_loop.c struct udphdr *uh; uh 57 tools/testing/selftests/bpf/progs/test_xdp_loop.c uh = (struct udphdr *)trans_data; uh 58 tools/testing/selftests/bpf/progs/test_xdp_loop.c if (uh + 1 > data_end) uh 60 tools/testing/selftests/bpf/progs/test_xdp_loop.c return uh->dest;