mss 132 arch/powerpc/platforms/cell/spufs/context.c if (ctx->mss) mss 133 arch/powerpc/platforms/cell/spufs/context.c unmap_mapping_range(ctx->mss, 0, SPUFS_MSS_MAP_SIZE, 1); mss 1337 arch/powerpc/platforms/cell/spufs/file.c ctx->mss = inode->i_mapping; mss 1350 arch/powerpc/platforms/cell/spufs/file.c ctx->mss = NULL; mss 76 arch/powerpc/platforms/cell/spufs/spufs.h struct address_space *mss; /* 'mss' area mappings. */ mss 164 arch/sparc/kernel/ldc.c u8 mss; mss 253 arch/sparc/kernel/ldc.c unsigned int mss; mss 267 arch/sparc/kernel/ldc.c mss = lp->mss; mss 269 arch/sparc/kernel/ldc.c if (diff * mss < size) mss 1127 arch/sparc/kernel/ldc.c u8 mss, *mssbuf; mss 1143 arch/sparc/kernel/ldc.c mss = LDC_PACKET_SIZE; mss 1148 arch/sparc/kernel/ldc.c mss = LDC_PACKET_SIZE - 8; mss 1153 arch/sparc/kernel/ldc.c mss = LDC_PACKET_SIZE - 8 - 8; mss 1188 arch/sparc/kernel/ldc.c lp->mss = mss; mss 1607 arch/sparc/kernel/ldc.c if (data_len > lp->mss) mss 1608 arch/sparc/kernel/ldc.c data_len = lp->mss; mss 207 drivers/crypto/chelsio/chtls/chtls.h u32 mss; mss 68 drivers/crypto/chelsio/chtls/chtls_cm.c csk->mss = MAX_MSS; mss 870 drivers/crypto/chelsio/chtls/chtls_cm.c unsigned int mss; mss 873 drivers/crypto/chelsio/chtls/chtls_cm.c mss = ntohs(req->tcpopt.mss); mss 889 drivers/crypto/chelsio/chtls/chtls_cm.c if (mss && tp->advmss > mss) mss 890 drivers/crypto/chelsio/chtls/chtls_cm.c tp->advmss = mss; mss 1203 drivers/crypto/chelsio/chtls/chtls_cm.c oreq->mss = 0; mss 990 drivers/crypto/chelsio/chtls/chtls_io.c int mss, flags, err; mss 1010 drivers/crypto/chelsio/chtls/chtls_io.c mss = csk->mss; mss 1018 drivers/crypto/chelsio/chtls/chtls_io.c copy = mss - skb->len; mss 1060 drivers/crypto/chelsio/chtls/chtls_io.c copy = mss; mss 1151 drivers/crypto/chelsio/chtls/chtls_io.c if (unlikely(skb->len == mss)) mss 1205 drivers/crypto/chelsio/chtls/chtls_io.c int mss, err, copied; mss 1220 drivers/crypto/chelsio/chtls/chtls_io.c mss = csk->mss; mss 1228 drivers/crypto/chelsio/chtls/chtls_io.c (copy = mss - skb->len) <= 0) { mss 1244 drivers/crypto/chelsio/chtls/chtls_io.c copy = mss; mss 1262 drivers/crypto/chelsio/chtls/chtls_io.c if (skb->len == mss) mss 254 drivers/infiniband/hw/cxgb4/cm.c ep->mss = ep->emss; mss 261 drivers/infiniband/hw/cxgb4/cm.c TCPOPT_MSS_G(opt), ep->mss, ep->emss); mss 262 drivers/infiniband/hw/cxgb4/cm.c pr_debug("mss_idx %u mss %u emss=%u\n", TCPOPT_MSS_G(opt), ep->mss, mss 2525 drivers/infiniband/hw/cxgb4/cm.c u16 peer_mss = ntohs(req->tcpopt.mss); mss 3982 drivers/infiniband/hw/cxgb4/cm.c req->tcpopt.mss = htons(tmp_opt.mss_clamp); mss 899 drivers/infiniband/hw/cxgb4/iw_cxgb4.h u16 mss; mss 667 drivers/infiniband/hw/i40iw/i40iw_cm.c tmp = ntohs(all_options->as_mss.mss); mss 668 drivers/infiniband/hw/i40iw/i40iw_cm.c if (tmp > 0 && tmp < cm_node->tcp_cntxt.mss) mss 669 drivers/infiniband/hw/i40iw/i40iw_cm.c cm_node->tcp_cntxt.mss = tmp; mss 685 drivers/infiniband/hw/i40iw/i40iw_cm.c cm_node->tcp_cntxt.mss = I40IW_CM_DEFAULT_MSS; mss 1360 drivers/infiniband/hw/i40iw/i40iw_cm.c options->as_mss.mss = htons(cm_node->tcp_cntxt.mss); mss 2217 drivers/infiniband/hw/i40iw/i40iw_cm.c cm_node->tcp_cntxt.mss = iwdev->vsi.mtu - I40IW_MTU_TO_MSS_IPV4; mss 2230 drivers/infiniband/hw/i40iw/i40iw_cm.c cm_node->tcp_cntxt.mss = iwdev->vsi.mtu - I40IW_MTU_TO_MSS_IPV6; mss 3324 drivers/infiniband/hw/i40iw/i40iw_cm.c tcp_info->cwnd = cpu_to_le32(2 * cm_node->tcp_cntxt.mss); mss 3332 drivers/infiniband/hw/i40iw/i40iw_cm.c tcp_info->snd_mss = cpu_to_le32(((u32)cm_node->tcp_cntxt.mss)); mss 146 drivers/infiniband/hw/i40iw/i40iw_cm.h __be16 mss; mss 276 drivers/infiniband/hw/i40iw/i40iw_cm.h u32 mss; mss 108 drivers/infiniband/hw/i40iw/i40iw_puda.h u16 mss; mss 126 drivers/infiniband/hw/i40iw/i40iw_puda.h u16 mss; mss 3506 drivers/infiniband/hw/mlx4/qp.c *lso_hdr_sz = cpu_to_be32(wr->mss << 16 | wr->hlen); mss 4113 drivers/infiniband/hw/mlx5/qp.c eseg->mss = cpu_to_be16(ud_wr->mss); mss 590 drivers/infiniband/hw/qedr/qedr_iw_cm.c in_params.mss = dev->iwarp_max_mtu - mss 606 drivers/infiniband/hw/qedr/qedr_iw_cm.c in_params.mss = dev->iwarp_max_mtu - mss 556 drivers/infiniband/ulp/ipoib/ipoib_ib.c priv->tx_wr.mss = skb_shinfo(skb)->gso_size; mss 114 drivers/net/ethernet/amazon/ena/ena_com.h u16 mss; mss 302 drivers/net/ethernet/amazon/ena/ena_eth_com.c meta_desc->word2 |= (ena_meta->mss << mss 306 drivers/net/ethernet/amazon/ena/ena_eth_com.c meta_desc->len_ctrl |= ((ena_meta->mss >> 10) << mss 2062 drivers/net/ethernet/amazon/ena/ena_netdev.c u32 mss = skb_shinfo(skb)->gso_size; mss 2066 drivers/net/ethernet/amazon/ena/ena_netdev.c if ((skb->ip_summed == CHECKSUM_PARTIAL) || mss) { mss 2068 drivers/net/ethernet/amazon/ena/ena_netdev.c if (mss) { mss 2083 drivers/net/ethernet/amazon/ena/ena_netdev.c if (mss) mss 2100 drivers/net/ethernet/amazon/ena/ena_netdev.c ena_meta->mss = mss; mss 551 drivers/net/ethernet/amd/xgbe/xgbe-desc.c if ((tso && (packet->mss != ring->tx.cur_mss)) || mss 1695 drivers/net/ethernet/amd/xgbe/xgbe-dev.c if (tso && (packet->mss != ring->tx.cur_mss)) mss 1733 drivers/net/ethernet/amd/xgbe/xgbe-dev.c packet->mss); mss 1737 drivers/net/ethernet/amd/xgbe/xgbe-dev.c MSS, packet->mss); mss 1747 drivers/net/ethernet/amd/xgbe/xgbe-dev.c ring->tx.cur_mss = packet->mss; mss 1760 drivers/net/ethernet/amd/xgbe/xgbe-drv.c packet->mss = skb_shinfo(skb)->gso_size; mss 1765 drivers/net/ethernet/amd/xgbe/xgbe-drv.c DBGPR(" packet->mss=%u\n", packet->mss); mss 363 drivers/net/ethernet/amd/xgbe/xgbe.h unsigned short mss; mss 276 drivers/net/ethernet/apm/xgene/xgene_enet_main.c static int xgene_enet_setup_mss(struct net_device *ndev, u32 mss) mss 286 drivers/net/ethernet/apm/xgene/xgene_enet_main.c if (pdata->mss[i] == mss) { mss 296 drivers/net/ethernet/apm/xgene/xgene_enet_main.c pdata->mac_ops->set_mss(pdata, mss, i); mss 297 drivers/net/ethernet/apm/xgene/xgene_enet_main.c pdata->mss[i] = mss; mss 313 drivers/net/ethernet/apm/xgene/xgene_enet_main.c u32 hdr_len, mss = 0; mss 336 drivers/net/ethernet/apm/xgene/xgene_enet_main.c mss = skb_shinfo(skb)->gso_size; mss 353 drivers/net/ethernet/apm/xgene/xgene_enet_main.c if (!mss || ((skb->len - hdr_len) <= mss)) mss 356 drivers/net/ethernet/apm/xgene/xgene_enet_main.c mss_index = xgene_enet_setup_mss(ndev, mss); mss 153 drivers/net/ethernet/apm/xgene/xgene_enet_main.h void (*set_mss)(struct xgene_enet_pdata *pdata, u16 mss, u8 index); mss 230 drivers/net/ethernet/apm/xgene/xgene_enet_main.h u32 mss[NUM_MSS_REG]; mss 222 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c u16 mss, u8 index) mss 232 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c SET_VAL(TSO_MSS0, mss); mss 234 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c data = SET_VAL(TSO_MSS1, mss) | SET_VAL(TSO_MSS0, data); mss 438 drivers/net/ethernet/aquantia/atlantic/aq_nic.c dx_buff->mss = skb_shinfo(skb)->gso_size; mss 56 drivers/net/ethernet/aquantia/atlantic/aq_ring.h u32 mss; mss 459 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c txd->ctl2 |= (buff->mss << 16) | mss 515 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->ctl2 |= (buff->mss << 16); mss 2349 drivers/net/ethernet/atheros/atlx/atl1.c unsigned int mss = 0; mss 2367 drivers/net/ethernet/atheros/atlx/atl1.c mss = skb_shinfo(skb)->gso_size; mss 2368 drivers/net/ethernet/atheros/atlx/atl1.c if (mss) { mss 6601 drivers/net/ethernet/broadcom/bnx2.c u32 len, vlan_tag_flags, last_frag, mss; mss 6635 drivers/net/ethernet/broadcom/bnx2.c if ((mss = skb_shinfo(skb)->gso_size)) { mss 6657 drivers/net/ethernet/broadcom/bnx2.c mss |= (tcp_off & 0xc) << TX_BD_TCP6_OFF2_SHL; mss 6667 drivers/net/ethernet/broadcom/bnx2.c mss = 0; mss 6683 drivers/net/ethernet/broadcom/bnx2.c txbd->tx_bd_mss_nbytes = len | (mss << 16); mss 6707 drivers/net/ethernet/broadcom/bnx2.c txbd->tx_bd_mss_nbytes = len | (mss << 16); mss 4750 drivers/net/ethernet/broadcom/bnxt/bnxt.c u16 mss = bp->dev->mtu - 40; mss 4771 drivers/net/ethernet/broadcom/bnxt/bnxt.c if (mss <= BNXT_RX_PAGE_SIZE) { mss 4772 drivers/net/ethernet/broadcom/bnxt/bnxt.c n = BNXT_RX_PAGE_SIZE / mss; mss 4775 drivers/net/ethernet/broadcom/bnxt/bnxt.c n = mss / BNXT_RX_PAGE_SIZE; mss 4776 drivers/net/ethernet/broadcom/bnxt/bnxt.c if (mss & (BNXT_RX_PAGE_SIZE - 1)) mss 2067 drivers/net/ethernet/broadcom/cnic.c xstorm_buf->mss = 0xffff; mss 3512 drivers/net/ethernet/broadcom/cnic.c l4kwqe3->mss = l4kwqe3->pmtu - sizeof(struct ipv6hdr) - mss 3516 drivers/net/ethernet/broadcom/cnic.c l4kwqe3->mss = l4kwqe3->pmtu - sizeof(struct iphdr) - mss 359 drivers/net/ethernet/broadcom/cnic_defs.h u16 mss; mss 361 drivers/net/ethernet/broadcom/cnic_defs.h u16 mss; mss 1733 drivers/net/ethernet/broadcom/cnic_defs.h u16 mss; mss 1735 drivers/net/ethernet/broadcom/cnic_defs.h u16 mss; mss 4090 drivers/net/ethernet/broadcom/cnic_defs.h u16 mss; mss 4096 drivers/net/ethernet/broadcom/cnic_defs.h u16 mss; mss 5116 drivers/net/ethernet/broadcom/cnic_defs.h u16 mss; mss 5120 drivers/net/ethernet/broadcom/cnic_defs.h u16 mss; mss 7672 drivers/net/ethernet/broadcom/tg3.c u32 len, u32 mss) mss 7674 drivers/net/ethernet/broadcom/tg3.c if (tg3_asic_rev(tp) == ASIC_REV_5762 && mss) { mss 7677 drivers/net/ethernet/broadcom/tg3.c return ((base + len + (mss & 0x3fff)) < base); mss 7697 drivers/net/ethernet/broadcom/tg3.c u32 mss, u32 vlan) mss 7702 drivers/net/ethernet/broadcom/tg3.c txbd->vlan_tag = (mss << TXD_MSS_SHIFT) | (vlan << TXD_VLAN_TAG_SHIFT); mss 7707 drivers/net/ethernet/broadcom/tg3.c u32 mss, u32 vlan) mss 7718 drivers/net/ethernet/broadcom/tg3.c if (tg3_4g_tso_overflow_test(tp, map, len, mss)) mss 7740 drivers/net/ethernet/broadcom/tg3.c frag_len, tmp_flag, mss, vlan); mss 7751 drivers/net/ethernet/broadcom/tg3.c len, flags, mss, vlan); mss 7761 drivers/net/ethernet/broadcom/tg3.c len, flags, mss, vlan); mss 7810 drivers/net/ethernet/broadcom/tg3.c u32 base_flags, u32 mss, u32 vlan) mss 7848 drivers/net/ethernet/broadcom/tg3.c mss, vlan)) { mss 7918 drivers/net/ethernet/broadcom/tg3.c u32 len, entry, base_flags, mss, vlan = 0; mss 7956 drivers/net/ethernet/broadcom/tg3.c mss = skb_shinfo(skb)->gso_size; mss 7957 drivers/net/ethernet/broadcom/tg3.c if (mss) { mss 7988 drivers/net/ethernet/broadcom/tg3.c iph->tot_len = htons(mss + hdr_len); mss 8008 drivers/net/ethernet/broadcom/tg3.c mss |= (hdr_len & 0xc) << 12; mss 8013 drivers/net/ethernet/broadcom/tg3.c mss |= hdr_len << 9; mss 8020 drivers/net/ethernet/broadcom/tg3.c mss |= (tsflags << 11); mss 8044 drivers/net/ethernet/broadcom/tg3.c !mss && skb->len > VLAN_ETH_FRAME_LEN) mss 8075 drivers/net/ethernet/broadcom/tg3.c mss, vlan)) { mss 8078 drivers/net/ethernet/broadcom/tg3.c u32 tmp_mss = mss; mss 8116 drivers/net/ethernet/broadcom/tg3.c if (mss && tg3_tso_bug_gso_check(tnapi, skb)) { mss 8134 drivers/net/ethernet/broadcom/tg3.c base_flags, mss, vlan)) mss 13429 drivers/net/ethernet/broadcom/tg3.c u32 base_flags = 0, mss = 0, desc_idx, coal_now, data_off, val; mss 13470 drivers/net/ethernet/broadcom/tg3.c mss = TG3_TSO_MSS; mss 13476 drivers/net/ethernet/broadcom/tg3.c iph->tot_len = htons((u16)(mss + hdr_len)); mss 13492 drivers/net/ethernet/broadcom/tg3.c mss |= (hdr_len & 0xc) << 12; mss 13497 drivers/net/ethernet/broadcom/tg3.c mss |= hdr_len << 9; mss 13500 drivers/net/ethernet/broadcom/tg3.c mss |= (TG3_TSO_TCP_OPT_LEN << 9); mss 13537 drivers/net/ethernet/broadcom/tg3.c base_flags | TXD_FLAG_END, mss, 0)) { mss 179 drivers/net/ethernet/chelsio/cxgb/cpl5_cmd.h u16 mss; mss 634 drivers/net/ethernet/chelsio/cxgb/cpl5_cmd.h u32 mss; mss 1752 drivers/net/ethernet/chelsio/cxgb/sge.c #define MK_ETH_TYPE_MSS(type, mss) (((mss) & 0x3FFF) | ((type) << 14)) mss 210 drivers/net/ethernet/chelsio/cxgb3/t3_cpl.h __be16 mss; mss 1526 drivers/net/ethernet/chelsio/cxgb4/sge.c lso->mss = htons(ssi->gso_size); mss 1833 drivers/net/ethernet/chelsio/cxgb4/sge.c lso->mss = cpu_to_be16(ssi->gso_size); mss 378 drivers/net/ethernet/chelsio/cxgb4/t4_msg.h __be16 mss; mss 903 drivers/net/ethernet/chelsio/cxgb4/t4_msg.h __be16 mss; mss 1301 drivers/net/ethernet/chelsio/cxgb4vf/sge.c lso->mss = cpu_to_be16(ssi->gso_size); mss 705 drivers/net/ethernet/cisco/enic/enic_main.c struct sk_buff *skb, unsigned int mss, mss 736 drivers/net/ethernet/cisco/enic/enic_main.c enic_queue_wq_desc_tso(wq, skb, dma_addr, len, mss, hdr_len, mss 809 drivers/net/ethernet/cisco/enic/enic_main.c unsigned int mss = skb_shinfo(skb)->gso_size; mss 824 drivers/net/ethernet/cisco/enic/enic_main.c if (mss) mss 825 drivers/net/ethernet/cisco/enic/enic_main.c err = enic_queue_wq_skb_tso(enic, wq, skb, mss, mss 113 drivers/net/ethernet/cisco/enic/enic_res.h unsigned int mss, unsigned int hdr_len, int vlan_tag_insert, mss 117 drivers/net/ethernet/cisco/enic/enic_res.h mss, hdr_len, vlan_tag_insert, vlan_tag, mss 54 drivers/net/ethernet/cisco/enic/wq_enet_desc.h u64 address, u16 length, u16 mss, u16 header_length, mss 60 drivers/net/ethernet/cisco/enic/wq_enet_desc.h desc->mss_loopback = cpu_to_le16((mss & WQ_ENET_MSS_MASK) << mss 73 drivers/net/ethernet/cisco/enic/wq_enet_desc.h u64 *address, u16 *length, u16 *mss, u16 *header_length, mss 79 drivers/net/ethernet/cisco/enic/wq_enet_desc.h *mss = (le16_to_cpu(desc->mss_loopback) >> WQ_ENET_MSS_SHIFT) & mss 38 drivers/net/ethernet/google/gve/gve_desc.h __be16 mss; /* TSO MSS */ mss 65 drivers/net/ethernet/google/gve/gve_desc.h __be16 mss; mss 387 drivers/net/ethernet/google/gve/gve_tx.c seg_desc->seg.mss = cpu_to_be16(skb_shinfo(skb)->gso_size); mss 187 drivers/net/ethernet/hisilicon/hns/hnae.h __le16 mss; mss 50 drivers/net/ethernet/hisilicon/hns/hns_enet.c u16 mss = 0; mss 95 drivers/net/ethernet/hisilicon/hns/hns_enet.c mss = skb_shinfo(skb)->gso_size; mss 109 drivers/net/ethernet/hisilicon/hns/hns_enet.c mss = skb_shinfo(skb)->gso_size; mss 115 drivers/net/ethernet/hisilicon/hns/hns_enet.c desc->tx.mss = cpu_to_le16(mss); mss 215 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c dev_info(dev, "(TX)mss: %u\n", tx_desc->tx.mss); mss 674 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c u16 *mss, u32 *type_cs_vlan_tso) mss 737 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c *mss = skb_shinfo(skb)->gso_size; mss 1039 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c u16 mss = 0; mss 1084 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c ret = hns3_set_tso(skb, &paylen, &mss, mss 1099 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c desc->tx.mss = cpu_to_le16(mss); mss 276 drivers/net/ethernet/hisilicon/hns3/hns3_enet.h __le16 mss; mss 538 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c u32 mss = HINIC_MSS_DEFAULT; mss 554 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c *queue_info |= HINIC_SQ_CTRL_SET(mss, QUEUE_INFO_MSS); mss 559 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c u32 l4_len, u32 offset, u32 ip_ident, u32 mss) mss 581 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c *queue_info |= HINIC_SQ_CTRL_SET(mss, QUEUE_INFO_MSS); mss 167 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.h u32 offset, u32 ip_ident, u32 mss); mss 1612 drivers/net/ethernet/ibm/ehea/ehea_main.c swqe->mss = skb_shinfo(skb)->gso_size; mss 101 drivers/net/ethernet/ibm/ehea/ehea_qmr.h u16 mss; mss 984 drivers/net/ethernet/ibm/ibmveth.c union ibmveth_buf_desc *descs, unsigned long mss) mss 1001 drivers/net/ethernet/ibm/ibmveth.c correlator, &correlator, mss, mss 1023 drivers/net/ethernet/ibm/ibmveth.c unsigned long mss = 0; mss 1124 drivers/net/ethernet/ibm/ibmveth.c mss = (unsigned long)skb_shinfo(skb)->gso_size; mss 1138 drivers/net/ethernet/ibm/ibmveth.c if (ibmveth_send(adapter, descs, mss)) { mss 1183 drivers/net/ethernet/ibm/ibmveth.c static void ibmveth_rx_mss_helper(struct sk_buff *skb, u16 mss, int lrg_pkt) mss 1216 drivers/net/ethernet/ibm/ibmveth.c skb_shinfo(skb)->gso_size = mss; mss 1302 drivers/net/ethernet/ibm/ibmveth.c u16 mss = 0; mss 1331 drivers/net/ethernet/ibm/ibmveth.c mss = (u16)be64_to_cpu(*rxmss); mss 1362 drivers/net/ethernet/ibm/ibmveth.c ibmveth_rx_mss_helper(skb, mss, lrg_pkt); mss 53 drivers/net/ethernet/ibm/ibmveth.h unsigned long mss, unsigned long large_send_support) mss 61 drivers/net/ethernet/ibm/ibmveth.h corellator_in, mss); mss 1546 drivers/net/ethernet/ibm/ibmvnic.c tx_crq.v1.mss = cpu_to_be16(skb_shinfo(skb)->gso_size); mss 247 drivers/net/ethernet/ibm/ibmvnic.h __be16 mss; mss 681 drivers/net/ethernet/intel/e1000/e1000_hw.h __le16 mss; /* Maximum segment size */ mss 2695 drivers/net/ethernet/intel/e1000/e1000_main.c u16 ipcse = 0, tucse, mss; mss 2706 drivers/net/ethernet/intel/e1000/e1000_main.c mss = skb_shinfo(skb)->gso_size; mss 2744 drivers/net/ethernet/intel/e1000/e1000_main.c context_desc->tcp_seg_setup.fields.mss = cpu_to_le16(mss); mss 2823 drivers/net/ethernet/intel/e1000/e1000_main.c unsigned int mss) mss 2851 drivers/net/ethernet/intel/e1000/e1000_main.c if (unlikely(mss && !nr_frags && size == len && size > 8)) mss 2908 drivers/net/ethernet/intel/e1000/e1000_main.c if (unlikely(mss && f == (nr_frags-1) && mss 3107 drivers/net/ethernet/intel/e1000/e1000_main.c unsigned int mss; mss 3127 drivers/net/ethernet/intel/e1000/e1000_main.c mss = skb_shinfo(skb)->gso_size; mss 3135 drivers/net/ethernet/intel/e1000/e1000_main.c if (mss) { mss 3137 drivers/net/ethernet/intel/e1000/e1000_main.c max_per_txd = min(mss << 2, max_per_txd); mss 3173 drivers/net/ethernet/intel/e1000/e1000_main.c if ((mss) || (skb->ip_summed == CHECKSUM_PARTIAL)) mss 3242 drivers/net/ethernet/intel/e1000/e1000_main.c nr_frags, mss); mss 309 drivers/net/ethernet/intel/e1000e/hw.h __le16 mss; /* Maximum segment size */ mss 5439 drivers/net/ethernet/intel/e1000e/netdev.c u16 ipcse = 0, mss; mss 5451 drivers/net/ethernet/intel/e1000e/netdev.c mss = skb_shinfo(skb)->gso_size; mss 5485 drivers/net/ethernet/intel/e1000e/netdev.c context_desc->tcp_seg_setup.fields.mss = cpu_to_le16(mss); mss 5795 drivers/net/ethernet/intel/e1000e/netdev.c unsigned int mss; mss 5817 drivers/net/ethernet/intel/e1000e/netdev.c mss = skb_shinfo(skb)->gso_size; mss 5818 drivers/net/ethernet/intel/e1000e/netdev.c if (mss) { mss 5843 drivers/net/ethernet/intel/e1000e/netdev.c if ((mss) || (skb->ip_summed == CHECKSUM_PARTIAL)) mss 64 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c txd->mss, txd->hdrlen, txd->flags); mss 788 drivers/net/ethernet/intel/fm10k/fm10k_main.c tx_desc->mss = cpu_to_le16(skb_shinfo(skb)->gso_size); mss 882 drivers/net/ethernet/intel/fm10k/fm10k_main.c tx_desc->mss = 0; mss 668 drivers/net/ethernet/intel/fm10k/fm10k_type.h __le16 mss; /* MSS for segmentation offload */ mss 570 drivers/net/ethernet/intel/ixgb/ixgb_hw.h __le16 mss; mss 1190 drivers/net/ethernet/intel/ixgb/ixgb_main.c u16 ipcse, tucse, mss; mss 1202 drivers/net/ethernet/intel/ixgb/ixgb_main.c mss = skb_shinfo(skb)->gso_size; mss 1227 drivers/net/ethernet/intel/ixgb/ixgb_main.c context_desc->mss = cpu_to_le16(mss); mss 1272 drivers/net/ethernet/intel/ixgb/ixgb_main.c context_desc->mss = 0; mss 1298 drivers/net/ethernet/intel/ixgb/ixgb_main.c unsigned int mss = skb_shinfo(skb)->gso_size; mss 1309 drivers/net/ethernet/intel/ixgb/ixgb_main.c if (unlikely(mss && !nr_frags && size == len && size > 8)) mss 1348 drivers/net/ethernet/intel/ixgb/ixgb_main.c if (unlikely(mss && (f == (nr_frags - 1)) mss 2285 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c unsigned int mss = 0; mss 2381 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (!mss) { mss 2382 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c mss = rx_ring->netdev->mtu - mss 2386 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (mss > 512) mss 2387 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c mss &= ~511; mss 2391 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c mss); mss 2065 drivers/net/ethernet/jme.c jme_tx_tso(struct sk_buff *skb, __le16 *mss, u8 *flags) mss 2067 drivers/net/ethernet/jme.c *mss = cpu_to_le16(skb_shinfo(skb)->gso_size << TXDESC_MSS_SHIFT); mss 2068 drivers/net/ethernet/jme.c if (*mss) { mss 2164 drivers/net/ethernet/jme.c if (jme_tx_tso(skb, &txdesc->desc1.mss, &flags)) mss 180 drivers/net/ethernet/jme.h __le16 mss; mss 1836 drivers/net/ethernet/marvell/sky2.c u16 mss; mss 1862 drivers/net/ethernet/marvell/sky2.c mss = skb_shinfo(skb)->gso_size; mss 1863 drivers/net/ethernet/marvell/sky2.c if (mss != 0) { mss 1866 drivers/net/ethernet/marvell/sky2.c mss += ETH_HLEN + ip_hdrlen(skb) + tcp_hdrlen(skb); mss 1868 drivers/net/ethernet/marvell/sky2.c if (mss != sky2->tx_last_mss) { mss 1870 drivers/net/ethernet/marvell/sky2.c le->addr = cpu_to_le32(mss); mss 1876 drivers/net/ethernet/marvell/sky2.c sky2->tx_last_mss = mss; mss 1931 drivers/net/ethernet/marvell/sky2.c le->opcode = mss ? (OP_LARGESEND | HW_OWNER) : (OP_PACKET | HW_OWNER); mss 381 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c u32 mss; mss 390 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c for (mss = 2; mss < MAX_LSO_MSS; mss++) { mss 391 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c mss_inv = div_u64(1ULL << 32, mss) >> 16; mss 392 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c mlx5e_ipsec_inverse_table[mss] = htons(mss_inv); mss 151 drivers/net/ethernet/mellanox/mlx5/core/en_accel/tls_rxtx.c int data_len, mss; mss 167 drivers/net/ethernet/mellanox/mlx5/core/en_accel/tls_rxtx.c mss = nskb->dev->mtu - (headln - skb_network_offset(nskb)); mss 169 drivers/net/ethernet/mellanox/mlx5/core/en_accel/tls_rxtx.c if (data_len > mss) { mss 170 drivers/net/ethernet/mellanox/mlx5/core/en_accel/tls_rxtx.c skb_shinfo(nskb)->gso_size = mss; mss 171 drivers/net/ethernet/mellanox/mlx5/core/en_accel/tls_rxtx.c skb_shinfo(nskb)->gso_segs = DIV_ROUND_UP(data_len, mss); mss 284 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c __be16 mss; mss 290 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c mss = cpu_to_be16(skb_shinfo(skb)->gso_size); mss 298 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c mss = 0; mss 349 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c eseg->mss = mss; mss 607 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c __be16 mss; mss 613 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c mss = cpu_to_be16(skb_shinfo(skb)->gso_size); mss 621 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c mss = 0; mss 660 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c eseg->mss = mss; mss 2637 drivers/net/ethernet/myricom/myri10ge/myri10ge.c int idx, avail, frag_cnt, frag_idx, count, mss, max_segments; mss 2651 drivers/net/ethernet/myricom/myri10ge/myri10ge.c mss = 0; mss 2655 drivers/net/ethernet/myricom/myri10ge/myri10ge.c mss = skb_shinfo(skb)->gso_size; mss 2676 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if (unlikely(!mss && (cksum_offset > 255 || mss 2690 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if (mss) { /* TSO */ mss 2712 drivers/net/ethernet/myricom/myri10ge/myri10ge.c pseudo_hdr_offset = mss; mss 2780 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if (mss) { /* TSO */ mss 2786 drivers/net/ethernet/myricom/myri10ge/myri10ge.c chop = (cum_len_next > mss); mss 2787 drivers/net/ethernet/myricom/myri10ge/myri10ge.c cum_len_next = cum_len_next % mss; mss 2800 drivers/net/ethernet/myricom/myri10ge/myri10ge.c small = (mss <= MXGEFW_SEND_SMALL_SIZE); mss 2822 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if (cksum_offset != 0 && !(mss && skb_is_gso_v6(skb))) { mss 2848 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if (mss) mss 1654 drivers/net/ethernet/neterion/vxge/vxge-config.h static inline void vxge_hw_fifo_txdl_mss_set(void *txdlh, int mss) mss 1659 drivers/net/ethernet/neterion/vxge/vxge-config.h txdp->control_0 |= VXGE_HW_FIFO_TXD_LSO_MSS(mss); mss 949 drivers/net/ethernet/neterion/vxge/vxge-main.c int mss = vxge_tcp_mss(skb); mss 950 drivers/net/ethernet/neterion/vxge/vxge-main.c if (mss) { mss 952 drivers/net/ethernet/neterion/vxge/vxge-main.c dev->name, __func__, __LINE__, mss); mss 953 drivers/net/ethernet/neterion/vxge/vxge-main.c vxge_hw_fifo_txdl_mss_set(dtr, mss); mss 151 drivers/net/ethernet/netronome/nfp/nfp_net.h __le16 mss; /* MSS to be used for LSO */ mss 731 drivers/net/ethernet/netronome/nfp/nfp_net_common.c u16 mss; mss 750 drivers/net/ethernet/netronome/nfp/nfp_net_common.c mss = skb_shinfo(skb)->gso_size & PCIE_DESC_TX_MSS_MASK; mss 754 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->mss = cpu_to_le16(mss); mss 1044 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->mss = 0; mss 1783 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->mss = 0; mss 2131 drivers/net/ethernet/netronome/nfp/nfp_net_common.c txd->mss = 0; mss 707 drivers/net/ethernet/pensando/ionic/ionic_if.h __le16 mss; mss 506 drivers/net/ethernet/pensando/ionic/ionic_txrx.c unsigned int hdrlen, unsigned int mss, mss 524 drivers/net/ethernet/pensando/ionic/ionic_txrx.c desc->mss = cpu_to_le16(mss); mss 564 drivers/net/ethernet/pensando/ionic/ionic_txrx.c unsigned int mss; mss 575 drivers/net/ethernet/pensando/ionic/ionic_txrx.c mss = skb_shinfo(skb)->gso_size; mss 602 drivers/net/ethernet/pensando/ionic/ionic_txrx.c seglen = hdrlen + mss; mss 625 drivers/net/ethernet/pensando/ionic/ionic_txrx.c hdrlen, mss, mss 633 drivers/net/ethernet/pensando/ionic/ionic_txrx.c seglen = mss; mss 664 drivers/net/ethernet/pensando/ionic/ionic_txrx.c hdrlen, mss, outer_csum, mss 672 drivers/net/ethernet/pensando/ionic/ionic_txrx.c len = min(mss, left); mss 673 drivers/net/ethernet/pensando/ionic/ionic_txrx.c frag_left = mss - len; mss 687 drivers/net/ethernet/pensando/ionic/ionic_txrx.c hdrlen, mss, outer_csum, mss 323 drivers/net/ethernet/qlogic/netxen/netxen_nic.h __le16 mss; mss 1887 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c first_desc->mss = cpu_to_le16(skb_shinfo(skb)->gso_size); mss 10381 drivers/net/ethernet/qlogic/qed/qed_hsi.h __le16 mss; mss 10454 drivers/net/ethernet/qlogic/qed/qed_hsi.h __le16 mss; mss 127 drivers/net/ethernet/qlogic/qed/qed_iscsi.c u16 mss; mss 416 drivers/net/ethernet/qlogic/qed/qed_iscsi.c p_tcp->mss = cpu_to_le16(p_conn->mss); mss 453 drivers/net/ethernet/qlogic/qed/qed_iscsi.c p_tcp2->mss = cpu_to_le16(p_conn->mss); mss 1316 drivers/net/ethernet/qlogic/qed/qed_iscsi.c con->mss = conn_info->mss; mss 628 drivers/net/ethernet/qlogic/qed/qed_iwarp.c p_tcp_ramrod->tcp.mss, mss 718 drivers/net/ethernet/qlogic/qed/qed_iwarp.c tcp->mss = cpu_to_le16(ep->mss); mss 724 drivers/net/ethernet/qlogic/qed/qed_iwarp.c tcp->cwnd = QED_IWARP_DEF_CWND_FACTOR * tcp->mss; mss 1227 drivers/net/ethernet/qlogic/qed/qed_iwarp.c ep->mss = iparams->mss; mss 1228 drivers/net/ethernet/qlogic/qed/qed_iwarp.c ep->mss = min_t(u16, QED_IWARP_MAX_FW_MSS, ep->mss); mss 2430 drivers/net/ethernet/qlogic/qed/qed_iwarp.c ep->mss = p_hwfn->p_rdma_info->iwarp.max_mtu - hdr_size; mss 2431 drivers/net/ethernet/qlogic/qed/qed_iwarp.c ep->mss = min_t(u16, QED_IWARP_MAX_FW_MSS, ep->mss); mss 150 drivers/net/ethernet/qlogic/qed/qed_iwarp.h u16 mss; mss 180 drivers/net/ethernet/qlogic/qlcnic/qlcnic.h __le16 mss; mss 396 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c first_desc->mss = cpu_to_le16(skb_shinfo(skb)->gso_size); mss 501 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c first_desc->mss = cpu_to_le16(skb_shinfo(skb)->gso_size); mss 739 drivers/net/ethernet/realtek/8139cp.c int mss = 0; mss 753 drivers/net/ethernet/realtek/8139cp.c mss = skb_shinfo(skb)->gso_size; mss 755 drivers/net/ethernet/realtek/8139cp.c if (mss > MSSMask) { mss 757 drivers/net/ethernet/realtek/8139cp.c mss); mss 763 drivers/net/ethernet/realtek/8139cp.c if (mss) mss 764 drivers/net/ethernet/realtek/8139cp.c opts1 |= LargeSend | (mss << MSSShift); mss 5773 drivers/net/ethernet/realtek/r8169_main.c u32 mss = skb_shinfo(skb)->gso_size; mss 5775 drivers/net/ethernet/realtek/r8169_main.c if (mss) { mss 5777 drivers/net/ethernet/realtek/r8169_main.c opts[0] |= min(mss, TD_MSS_MAX) << TD0_MSS_SHIFT; mss 5794 drivers/net/ethernet/realtek/r8169_main.c u32 mss = skb_shinfo(skb)->gso_size; mss 5796 drivers/net/ethernet/realtek/r8169_main.c if (mss) { mss 5815 drivers/net/ethernet/realtek/r8169_main.c opts[1] |= min(mss, TD_MSS_MAX) << TD1_MSS_SHIFT; mss 133 drivers/net/ethernet/samsung/sxgbe/sxgbe_desc.c static void sxgbe_tx_ctxt_desc_set_mss(struct sxgbe_tx_ctxt_desc *p, u16 mss) mss 135 drivers/net/ethernet/samsung/sxgbe/sxgbe_desc.c p->maxseg_size = mss; mss 212 drivers/net/ethernet/samsung/sxgbe/sxgbe_desc.h void (*tx_ctxt_desc_set_mss)(struct sxgbe_tx_ctxt_desc *p, u16 mss); mss 2332 drivers/net/ethernet/sfc/ef10.c u32 mss; mss 2336 drivers/net/ethernet/sfc/ef10.c mss = skb_shinfo(skb)->gso_size; mss 2338 drivers/net/ethernet/sfc/ef10.c if (unlikely(mss < 4)) { mss 2339 drivers/net/ethernet/sfc/ef10.c WARN_ONCE(1, "MSS of %u is too small for TSO v2\n", mss); mss 2385 drivers/net/ethernet/sfc/ef10.c ESF_DZ_TX_TSO_TCP_MSS, mss mss 419 drivers/net/ethernet/stmicro/stmmac/dwmac4_descs.c static void dwmac4_set_mss_ctxt(struct dma_desc *p, unsigned int mss) mss 423 drivers/net/ethernet/stmicro/stmmac/dwmac4_descs.c p->des2 = cpu_to_le32(mss); mss 234 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_descs.c static void dwxgmac2_set_mss(struct dma_desc *p, unsigned int mss) mss 238 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_descs.c p->des2 = cpu_to_le32(mss); mss 82 drivers/net/ethernet/stmicro/stmmac/hwif.h void (*set_mss)(struct dma_desc *p, unsigned int mss); mss 56 drivers/net/ethernet/stmicro/stmmac/stmmac.h u32 mss; mss 1409 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c tx_q->mss = 0; mss 2001 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c tx_q->mss = 0; mss 2920 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c u32 pay_len, mss; mss 2947 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c mss = skb_shinfo(skb)->gso_size; mss 2950 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c if (mss != tx_q->mss) { mss 2952 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c stmmac_set_mss(priv, mss_desc, mss); mss 2953 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c tx_q->mss = mss; mss 2960 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c __func__, tcp_hdrlen(skb), proto_hdr_len, pay_len, mss); mss 4831 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c tx_q->mss = 0; mss 527 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c if ((tso && (pkt_info->mss != ring->tx.cur_mss)) || mss 727 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c if (tso && (pkt_info->mss != ring->tx.cur_mss)) mss 766 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c pkt_info->mss); mss 773 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c pkt_info->mss); mss 789 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c ring->tx.cur_mss = pkt_info->mss; mss 87 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c pkt_info->mss = skb_shinfo(skb)->gso_size; mss 92 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c XLGMAC_PR("mss=%u\n", pkt_info->mss); mss 271 drivers/net/ethernet/synopsys/dwc-xlgmac.h unsigned short mss; mss 1640 drivers/net/ethernet/tehuti/tehuti.c txdd->mss = CPU_CHIP_SWAP16(txd_mss); mss 1647 drivers/net/ethernet/tehuti/tehuti.c DBG("=== w2: mss 0x%x len 0x%x\n", txdd->mss, txdd->length); mss 329 drivers/net/ethernet/tehuti/tehuti.h u16 mss; mss 1428 drivers/net/hyperv/hyperv_net.h u32 mss:20; mss 1439 drivers/net/hyperv/hyperv_net.h u32 mss:20; mss 646 drivers/net/hyperv/netvsc_drv.c lso_info->lso_v2_transmit.mss = skb_shinfo(skb)->gso_size; mss 1422 drivers/net/usb/ax88179_178a.c int mss = skb_shinfo(skb)->gso_size; mss 1427 drivers/net/usb/ax88179_178a.c tx_hdr2 = mss; mss 2751 drivers/net/usb/lan78xx.c u16 mss = max(skb_shinfo(skb)->gso_size, TX_CMD_B_MSS_MIN_); mss 2753 drivers/net/usb/lan78xx.c tx_cmd_b = (mss << TX_CMD_B_MSS_SHIFT_) & TX_CMD_B_MSS_MASK_; mss 1792 drivers/net/usb/r8152.c u32 mss = skb_shinfo(skb)->gso_size; mss 1800 drivers/net/usb/r8152.c if (mss) { mss 1828 drivers/net/usb/r8152.c opts2 |= min(mss, MSS_MAX) << MSS_SHIFT; mss 2391 drivers/net/usb/r8152.c u32 mss = skb_shinfo(skb)->gso_size; mss 2392 drivers/net/usb/r8152.c int max_offset = mss ? GTTCPHO_MAX : TCPHO_MAX; mss 2395 drivers/net/usb/r8152.c if ((mss || skb->ip_summed == CHECKSUM_PARTIAL) && offset > max_offset) mss 2271 drivers/net/usb/smsc75xx.c u16 mss = max(skb_shinfo(skb)->gso_size, TX_MSS_MIN); mss 2272 drivers/net/usb/smsc75xx.c tx_cmd_b = (mss << TX_CMD_B_MSS_SHIFT) & TX_CMD_B_MSS; mss 307 drivers/net/vmxnet3/vmxnet3_defs.h u32 mss:16; mss 309 drivers/net/vmxnet3/vmxnet3_defs.h u32 mss:16; mss 844 drivers/net/vmxnet3/vmxnet3_drv.c if (ctx->mss) { /* TSO */ mss 999 drivers/net/vmxnet3/vmxnet3_drv.c ctx.mss = skb_shinfo(skb)->gso_size; mss 1000 drivers/net/vmxnet3/vmxnet3_drv.c if (ctx.mss) { mss 1031 drivers/net/vmxnet3/vmxnet3_drv.c if (ctx.mss) { mss 1085 drivers/net/vmxnet3/vmxnet3_drv.c if (ctx.mss) { mss 1088 drivers/net/vmxnet3/vmxnet3_drv.c gdesc->txd.msscof = ctx.mss; mss 1089 drivers/net/vmxnet3/vmxnet3_drv.c num_pkts = (skb->len - gdesc->txd.hlen + ctx.mss - 1) / ctx.mss; mss 1288 drivers/net/vmxnet3/vmxnet3_drv.c u16 segCnt = 0, mss = 0; mss 1430 drivers/net/vmxnet3/vmxnet3_drv.c mss = rcdlro->mss; mss 1504 drivers/net/vmxnet3/vmxnet3_drv.c if (segCnt != 0 && mss != 0) { mss 1507 drivers/net/vmxnet3/vmxnet3_drv.c skb_shinfo(skb)->gso_size = mss; mss 220 drivers/net/vmxnet3/vmxnet3_int.h u16 mss; mss 1754 drivers/net/wireless/ath/wil6210/txrx.c int mss = skb_shinfo(skb)->gso_size; /* payload size w/o headers */ mss 1761 drivers/net/wireless/ath/wil6210/txrx.c int rem_data = mss; mss 1951 drivers/net/wireless/ath/wil6210/txrx.c rem_data = mss; mss 1328 drivers/net/wireless/ath/wil6210/txrx_edma.c int mss) mss 1333 drivers/net/wireless/ath/wil6210/txrx_edma.c d->mac.tso_mss |= cpu_to_le16(mss >> 2); mss 1359 drivers/net/wireless/ath/wil6210/txrx_edma.c int mss, int *descs_used) mss 1387 drivers/net/wireless/ath/wil6210/txrx_edma.c skb_net_hdr_len, mss); mss 1423 drivers/net/wireless/ath/wil6210/txrx_edma.c int mss = skb_shinfo(skb)->gso_size; mss 1468 drivers/net/wireless/ath/wil6210/txrx_edma.c mss, &descs_used); mss 1480 drivers/net/wireless/ath/wil6210/txrx_edma.c mss, &descs_used); mss 1498 drivers/net/wireless/ath/wil6210/txrx_edma.c mss, &descs_used); mss 827 drivers/net/wireless/intel/iwlwifi/mvm/tx.c unsigned int mss = skb_shinfo(skb)->gso_size; mss 832 drivers/net/wireless/intel/iwlwifi/mvm/tx.c skb_shinfo(skb)->gso_size = num_subframes * mss; mss 836 drivers/net/wireless/intel/iwlwifi/mvm/tx.c skb_shinfo(skb)->gso_size = mss; mss 860 drivers/net/wireless/intel/iwlwifi/mvm/tx.c if (tcp_payload_len > mss) { mss 861 drivers/net/wireless/intel/iwlwifi/mvm/tx.c skb_shinfo(tmp)->gso_size = mss; mss 892 drivers/net/wireless/intel/iwlwifi/mvm/tx.c unsigned int mss = skb_shinfo(skb)->gso_size; mss 950 drivers/net/wireless/intel/iwlwifi/mvm/tx.c subf_len = sizeof(struct ethhdr) + snap_ip_tcp + mss; mss 980 drivers/net/wireless/intel/iwlwifi/mvm/tx.c if (num_subframes * mss >= tcp_payload_len) { mss 254 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c unsigned int mss = skb_shinfo(skb)->gso_size; mss 270 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c hdr_room = DIV_ROUND_UP(total_len, mss) * mss 300 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c unsigned int data_left = min_t(unsigned int, mss, total_len); mss 2109 drivers/net/wireless/intel/iwlwifi/pcie/tx.c unsigned int mss = skb_shinfo(skb)->gso_size; mss 2132 drivers/net/wireless/intel/iwlwifi/pcie/tx.c hdr_room = DIV_ROUND_UP(total_len, mss) * mss 2165 drivers/net/wireless/intel/iwlwifi/pcie/tx.c min_t(unsigned int, mss, total_len); mss 1181 drivers/net/xen-netback/netback.c int mss, hdrlen; mss 1190 drivers/net/xen-netback/netback.c mss = skb_shinfo(skb)->gso_size; mss 1196 drivers/net/xen-netback/netback.c DIV_ROUND_UP(skb->len - hdrlen, mss); mss 335 drivers/s390/net/qeth_core.h __u16 mss; mss 4013 drivers/s390/net/qeth_core_main.c ext->mss = skb_shinfo(skb)->gso_size; mss 6000 drivers/s390/net/qeth_core_main.c tso_data->mss = cmd->data.setassparms.data.tso.mss; mss 6029 drivers/s390/net/qeth_core_main.c if (!tso_data.mss || !(tso_data.supported & QETH_IPA_LARGE_SEND_TCP)) { mss 6058 drivers/s390/net/qeth_core_main.c tso_data.mss); mss 406 drivers/s390/net/qeth_core_mpc.h u32 mss; mss 220 drivers/scsi/fnic/fcpio.h u16 mss; /* FC vNIC only: max burst */ mss 268 drivers/scsi/fnic/fcpio.h u16 mss; /* FC vNIC only: max burst */ mss 88 drivers/scsi/fnic/fnic_res.h u32 d_id, u16 mss, mss 117 drivers/scsi/fnic/fnic_res.h desc->u.icmnd_16.mss = mss; /* FC vNIC only: max burst */ mss 52 drivers/scsi/fnic/wq_enet_desc.h u64 address, u16 length, u16 mss, u16 header_length, mss 58 drivers/scsi/fnic/wq_enet_desc.h desc->mss_loopback = cpu_to_le16((mss & WQ_ENET_MSS_MASK) << mss 71 drivers/scsi/fnic/wq_enet_desc.h u64 *address, u16 *length, u16 *mss, u16 *header_length, mss 77 drivers/scsi/fnic/wq_enet_desc.h *mss = (le16_to_cpu(desc->mss_loopback) >> WQ_ENET_MSS_SHIFT) & mss 462 drivers/scsi/qedi/qedi_iscsi.c u16 mss = 0; mss 470 drivers/scsi/qedi/qedi_iscsi.c mss = pmtu - hdrs; mss 472 drivers/scsi/qedi/qedi_iscsi.c if (!mss) mss 473 drivers/scsi/qedi/qedi_iscsi.c mss = DEF_MSS; mss 475 drivers/scsi/qedi/qedi_iscsi.c return mss; mss 545 drivers/scsi/qedi/qedi_iscsi.c conn_info->mss = qedi_calc_mss(qedi_ep->pmtu, mss 549 drivers/scsi/qedi/qedi_iscsi.c conn_info->cwnd = DEF_MAX_CWND * conn_info->mss; mss 556 drivers/scsi/qedi/qedi_iscsi.c conn_info->default_cq, conn_info->mss); mss 891 drivers/scsi/qla4xxx/ql4_fw.h uint16_t mss; /* 80-81 */ mss 3544 drivers/scsi/qla4xxx/ql4_os.c conn->max_segment_size = le16_to_cpu(fw_ddb_entry->mss); mss 3672 drivers/scsi/qla4xxx/ql4_os.c fw_ddb_entry->mss = cpu_to_le16(conn->max_segment_size); mss 3781 drivers/scsi/qla4xxx/ql4_os.c conn->max_segment_size = le16_to_cpu(fw_ddb_entry->mss); mss 52 drivers/scsi/snic/wq_enet_desc.h u64 address, u16 length, u16 mss, u16 header_length, mss 58 drivers/scsi/snic/wq_enet_desc.h desc->mss_loopback = cpu_to_le16((mss & WQ_ENET_MSS_MASK) << mss 71 drivers/scsi/snic/wq_enet_desc.h u64 *address, u16 *length, u16 *mss, u16 *header_length, mss 77 drivers/scsi/snic/wq_enet_desc.h *mss = (le16_to_cpu(desc->mss_loopback) >> WQ_ENET_MSS_SHIFT) & mss 1141 drivers/staging/qlge/qlge.h __le16 mss; mss 1895 drivers/staging/qlge/qlge_dbg.c le16_to_cpu(ob_mac_tso_iocb->mss)); mss 2581 drivers/staging/qlge/qlge_main.c mac_iocb_ptr->mss = cpu_to_le16(skb_shinfo(skb)->gso_size); mss 228 drivers/target/iscsi/cxgbit/cxgbit.h u16 mss; mss 765 drivers/target/iscsi/cxgbit/cxgbit_cm.c csk->mss = csk->emss; mss 772 drivers/target/iscsi/cxgbit/cxgbit_cm.c TCPOPT_MSS_G(opt), csk->mss, csk->emss); mss 774 drivers/target/iscsi/cxgbit/cxgbit_cm.c csk->mss, csk->emss); mss 1224 drivers/target/iscsi/cxgbit/cxgbit_cm.c u16 peer_mss = ntohs(req->tcpopt.mss); mss 433 fs/proc/task_mmu.c static void smaps_page_accumulate(struct mem_size_stats *mss, mss 437 fs/proc/task_mmu.c mss->pss += pss; mss 440 fs/proc/task_mmu.c mss->pss_anon += pss; mss 442 fs/proc/task_mmu.c mss->pss_shmem += pss; mss 444 fs/proc/task_mmu.c mss->pss_file += pss; mss 447 fs/proc/task_mmu.c mss->pss_locked += pss; mss 451 fs/proc/task_mmu.c mss->private_dirty += size; mss 453 fs/proc/task_mmu.c mss->shared_dirty += size; mss 456 fs/proc/task_mmu.c mss->private_clean += size; mss 458 fs/proc/task_mmu.c mss->shared_clean += size; mss 462 fs/proc/task_mmu.c static void smaps_account(struct mem_size_stats *mss, struct page *page, mss 473 fs/proc/task_mmu.c mss->anonymous += size; mss 475 fs/proc/task_mmu.c mss->lazyfree += size; mss 478 fs/proc/task_mmu.c mss->resident += size; mss 481 fs/proc/task_mmu.c mss->referenced += size; mss 492 fs/proc/task_mmu.c smaps_page_accumulate(mss, page, size, size << PSS_SHIFT, dirty, mss 501 fs/proc/task_mmu.c smaps_page_accumulate(mss, page, PAGE_SIZE, pss, dirty, locked, mss 510 fs/proc/task_mmu.c struct mem_size_stats *mss = walk->private; mss 512 fs/proc/task_mmu.c mss->swap += shmem_partial_swap_usage( mss 524 fs/proc/task_mmu.c struct mem_size_stats *mss = walk->private; mss 537 fs/proc/task_mmu.c mss->swap += PAGE_SIZE; mss 543 fs/proc/task_mmu.c mss->swap_pss += pss_delta; mss 545 fs/proc/task_mmu.c mss->swap_pss += (u64)PAGE_SIZE << PSS_SHIFT; mss 551 fs/proc/task_mmu.c } else if (unlikely(IS_ENABLED(CONFIG_SHMEM) && mss->check_shmem_swap mss 559 fs/proc/task_mmu.c mss->swap += PAGE_SIZE; mss 569 fs/proc/task_mmu.c smaps_account(mss, page, false, pte_young(*pte), pte_dirty(*pte), locked); mss 576 fs/proc/task_mmu.c struct mem_size_stats *mss = walk->private; mss 586 fs/proc/task_mmu.c mss->anonymous_thp += HPAGE_PMD_SIZE; mss 588 fs/proc/task_mmu.c mss->shmem_thp += HPAGE_PMD_SIZE; mss 592 fs/proc/task_mmu.c mss->file_thp += HPAGE_PMD_SIZE; mss 593 fs/proc/task_mmu.c smaps_account(mss, page, true, pmd_young(*pmd), pmd_dirty(*pmd), locked); mss 711 fs/proc/task_mmu.c struct mem_size_stats *mss = walk->private; mss 729 fs/proc/task_mmu.c mss->shared_hugetlb += huge_page_size(hstate_vma(vma)); mss 731 fs/proc/task_mmu.c mss->private_hugetlb += huge_page_size(hstate_vma(vma)); mss 751 fs/proc/task_mmu.c struct mem_size_stats *mss) mss 755 fs/proc/task_mmu.c mss->check_shmem_swap = false; mss 771 fs/proc/task_mmu.c mss->swap += shmem_swapped; mss 773 fs/proc/task_mmu.c mss->check_shmem_swap = true; mss 774 fs/proc/task_mmu.c walk_page_vma(vma, &smaps_shmem_walk_ops, mss); mss 780 fs/proc/task_mmu.c walk_page_vma(vma, &smaps_walk_ops, mss); mss 787 fs/proc/task_mmu.c static void __show_smap(struct seq_file *m, const struct mem_size_stats *mss, mss 790 fs/proc/task_mmu.c SEQ_PUT_DEC("Rss: ", mss->resident); mss 791 fs/proc/task_mmu.c SEQ_PUT_DEC(" kB\nPss: ", mss->pss >> PSS_SHIFT); mss 798 fs/proc/task_mmu.c mss->pss_anon >> PSS_SHIFT); mss 800 fs/proc/task_mmu.c mss->pss_file >> PSS_SHIFT); mss 802 fs/proc/task_mmu.c mss->pss_shmem >> PSS_SHIFT); mss 804 fs/proc/task_mmu.c SEQ_PUT_DEC(" kB\nShared_Clean: ", mss->shared_clean); mss 805 fs/proc/task_mmu.c SEQ_PUT_DEC(" kB\nShared_Dirty: ", mss->shared_dirty); mss 806 fs/proc/task_mmu.c SEQ_PUT_DEC(" kB\nPrivate_Clean: ", mss->private_clean); mss 807 fs/proc/task_mmu.c SEQ_PUT_DEC(" kB\nPrivate_Dirty: ", mss->private_dirty); mss 808 fs/proc/task_mmu.c SEQ_PUT_DEC(" kB\nReferenced: ", mss->referenced); mss 809 fs/proc/task_mmu.c SEQ_PUT_DEC(" kB\nAnonymous: ", mss->anonymous); mss 810 fs/proc/task_mmu.c SEQ_PUT_DEC(" kB\nLazyFree: ", mss->lazyfree); mss 811 fs/proc/task_mmu.c SEQ_PUT_DEC(" kB\nAnonHugePages: ", mss->anonymous_thp); mss 812 fs/proc/task_mmu.c SEQ_PUT_DEC(" kB\nShmemPmdMapped: ", mss->shmem_thp); mss 813 fs/proc/task_mmu.c SEQ_PUT_DEC(" kB\nFilePmdMapped: ", mss->file_thp); mss 814 fs/proc/task_mmu.c SEQ_PUT_DEC(" kB\nShared_Hugetlb: ", mss->shared_hugetlb); mss 816 fs/proc/task_mmu.c mss->private_hugetlb >> 10, 7); mss 817 fs/proc/task_mmu.c SEQ_PUT_DEC(" kB\nSwap: ", mss->swap); mss 819 fs/proc/task_mmu.c mss->swap_pss >> PSS_SHIFT); mss 821 fs/proc/task_mmu.c mss->pss_locked >> PSS_SHIFT); mss 828 fs/proc/task_mmu.c struct mem_size_stats mss; mss 830 fs/proc/task_mmu.c memset(&mss, 0, sizeof(mss)); mss 832 fs/proc/task_mmu.c smap_gather_stats(vma, &mss); mss 841 fs/proc/task_mmu.c __show_smap(m, &mss, false); mss 858 fs/proc/task_mmu.c struct mem_size_stats mss; mss 874 fs/proc/task_mmu.c memset(&mss, 0, sizeof(mss)); mss 883 fs/proc/task_mmu.c smap_gather_stats(vma, &mss); mss 892 fs/proc/task_mmu.c __show_smap(m, &mss, true); mss 249 include/linux/mlx5/qp.h __be16 mss; mss 684 include/linux/qed/common_hsi.h __le16 mss; mss 123 include/linux/qed/qed_iscsi_if.h u16 mss; mss 515 include/linux/qed/qed_rdma_if.h u16 mss; mss 112 include/linux/qed/tcp_common.h __le16 mss; mss 178 include/linux/qed/tcp_common.h __le16 mss; mss 243 include/linux/qed/tcp_common.h __le16 mss; mss 477 include/linux/tcp.h static inline u16 tcp_mss_clamp(const struct tcp_sock *tp, u16 mss) mss 484 include/linux/tcp.h return (user_mss && user_mss < mss) ? user_mss : mss; mss 55 include/net/request_sock.h u16 mss; mss 550 include/net/tcp.h __u32 cookie_v4_init_sequence(const struct sk_buff *skb, __u16 *mss); mss 564 include/net/tcp.h __u32 cookie_v6_init_sequence(const struct sk_buff *skb, __u16 *mss); mss 656 include/net/tcp.h int tcp_mss_to_mtu(struct sock *sk, int mss); mss 1359 include/net/tcp.h u32 tcp_default_init_rwnd(u32 mss); mss 1378 include/net/tcp.h __u32 mss, __u32 *rcv_wnd, mss 1635 include/net/tcp.h void tcp_fastopen_cache_get(struct sock *sk, u16 *mss, mss 1637 include/net/tcp.h void tcp_fastopen_cache_set(struct sock *sk, u16 mss, mss 1659 include/net/tcp.h bool tcp_fastopen_cookie_check(struct sock *sk, u16 *mss, mss 1991 include/net/tcp.h __u16 *mss); mss 2006 include/net/tcp.h __u16 *mss) mss 2010 include/net/tcp.h return ops->cookie_init_seq(skb, mss); mss 2015 include/net/tcp.h __u16 *mss) mss 1371 include/rdma/ib_verbs.h int mss; mss 20 include/uapi/linux/netfilter/nf_synproxy.h __u16 mss; mss 58 include/uapi/linux/netfilter/nfnetlink_osf.h __u16 ss, mss; mss 8 include/uapi/linux/netfilter/xt_TCPMSS.h __u16 mss; mss 183 ipc/msg.c struct msg_sender *mss, size_t msgsz) mss 185 ipc/msg.c mss->tsk = current; mss 186 ipc/msg.c mss->msgsz = msgsz; mss 188 ipc/msg.c list_add_tail(&mss->list, &msq->q_senders); mss 191 ipc/msg.c static inline void ss_del(struct msg_sender *mss) mss 193 ipc/msg.c if (mss->list.next) mss 194 ipc/msg.c list_del(&mss->list); mss 200 ipc/msg.c struct msg_sender *mss, *t; mss 204 ipc/msg.c list_for_each_entry_safe(mss, t, h, list) { mss 206 ipc/msg.c mss->list.next = NULL; mss 213 ipc/msg.c else if (stop_tsk == mss->tsk) mss 222 ipc/msg.c else if (!msg_fits_inqueue(msq, mss->msgsz)) { mss 224 ipc/msg.c stop_tsk = mss->tsk; mss 226 ipc/msg.c list_move_tail(&mss->list, &msq->q_senders); mss 230 ipc/msg.c wake_q_add(wake_q, mss->tsk); mss 142 net/batman-adv/tp_meter.c static void batadv_tp_update_cwnd(struct batadv_tp_vars *tp_vars, u32 mss) mss 149 net/batman-adv/tp_meter.c tp_vars->cwnd = batadv_tp_cwnd(tp_vars->cwnd, mss, mss); mss 156 net/batman-adv/tp_meter.c ((mss * mss) << 6) / (tp_vars->cwnd << 3)); mss 157 net/batman-adv/tp_meter.c if (tp_vars->dec_cwnd < (mss << 3)) { mss 162 net/batman-adv/tp_meter.c tp_vars->cwnd = batadv_tp_cwnd(tp_vars->cwnd, mss, mss); mss 630 net/batman-adv/tp_meter.c size_t packet_len, mss; mss 635 net/batman-adv/tp_meter.c mss = BATADV_TP_PLEN; mss 700 net/batman-adv/tp_meter.c tp_vars->cwnd = batadv_tp_cwnd(tp_vars->ss_threshold, 3 * mss, mss 701 net/batman-adv/tp_meter.c mss); mss 727 net/batman-adv/tp_meter.c mss, mss); mss 735 net/batman-adv/tp_meter.c mss); mss 741 net/batman-adv/tp_meter.c if (recv_ack - atomic_read(&tp_vars->last_acked) >= mss) mss 742 net/batman-adv/tp_meter.c batadv_tp_update_cwnd(tp_vars, mss); mss 5866 net/core/filter.c u16 mss; mss 5891 net/core/filter.c mss = tcp_v4_get_syncookie(sk, iph, th, &cookie); mss 5902 net/core/filter.c mss = tcp_v6_get_syncookie(sk, iph, th, &cookie); mss 5909 net/core/filter.c if (mss == 0) mss 5912 net/core/filter.c return cookie | ((u64)mss << 32); mss 3657 net/core/skbuff.c unsigned int mss = skb_shinfo(head_skb)->gso_size; mss 3688 net/core/skbuff.c if (mss != GSO_BY_FRAGS && mss != skb_headlen(head_skb)) mss 3700 net/core/skbuff.c if (sg && csum && (mss != GSO_BY_FRAGS)) { mss 3735 net/core/skbuff.c partial_segs = len / mss; mss 3737 net/core/skbuff.c mss *= partial_segs; mss 3752 net/core/skbuff.c if (unlikely(mss == GSO_BY_FRAGS)) { mss 3756 net/core/skbuff.c if (len > mss) mss 3757 net/core/skbuff.c len = mss; mss 3889 net/core/skbuff.c pos, mss); mss 1848 net/decnet/af_decnet.c unsigned int mss = 230 - DN_MAX_NSP_DATA_HEADER; mss 1863 net/decnet/af_decnet.c if (mtu > mss) mss 1864 net/decnet/af_decnet.c mss = mtu; mss 1865 net/decnet/af_decnet.c return mss; mss 1910 net/decnet/af_decnet.c size_t mss; mss 1959 net/decnet/af_decnet.c mss = scp->segsize_rem; mss 1962 net/decnet/af_decnet.c mss = dn_current_mss(sk, flags); mss 1966 net/decnet/af_decnet.c if (size > mss) { mss 1989 net/decnet/af_decnet.c if (len > mss) mss 1990 net/decnet/af_decnet.c len = mss; mss 276 net/decnet/dn_route.c u32 mss = mtu - DN_MAX_NSP_DATA_HEADER; mss 278 net/decnet/dn_route.c if (!existing_mss || existing_mss > mss) mss 279 net/decnet/dn_route.c dst_metric_set(dst, RTAX_ADVMSS, mss); mss 888 net/decnet/dn_route.c unsigned int mss = dn_mss_from_pmtu(dev, dst_mtu(&rt->dst)); mss 889 net/decnet/dn_route.c if (mss_metric > mss) mss 890 net/decnet/dn_route.c dst_metric_set(&rt->dst, RTAX_ADVMSS, mss); mss 40 net/ipv4/netfilter/ipt_SYNPROXY.c opts.mss_option = info->mss; mss 165 net/ipv4/syncookies.c const __u16 mss = *mssp; mss 168 net/ipv4/syncookies.c if (mss >= msstab[mssind]) mss 294 net/ipv4/syncookies.c int mss; mss 306 net/ipv4/syncookies.c mss = __cookie_v4_check(ip_hdr(skb), th, cookie); mss 307 net/ipv4/syncookies.c if (mss == 0) { mss 339 net/ipv4/syncookies.c req->mss = mss; mss 390 net/ipv4/syncookies.c tcp_select_initial_window(sk, tcp_full_space(sk), req->mss, mss 244 net/ipv4/tcp_bbr.c unsigned int mss = tcp_sk(sk)->mss_cache; mss 246 net/ipv4/tcp_bbr.c rate *= mss; mss 406 net/ipv4/tcp_fastopen.c bool tcp_fastopen_cookie_check(struct sock *sk, u16 *mss, mss 411 net/ipv4/tcp_fastopen.c tcp_fastopen_cache_get(sk, mss, cookie); mss 439 net/ipv4/tcp_fastopen.c u16 mss; mss 442 net/ipv4/tcp_fastopen.c if (tcp_fastopen_cookie_check(sk, &mss, &cookie)) { mss 888 net/ipv4/tcp_input.c const u32 mss = tp->mss_cache; mss 896 net/ipv4/tcp_input.c if ((metric > tp->reordering * mss) && mss) { mss 905 net/ipv4/tcp_input.c tp->reordering = min_t(u32, (metric + mss - 1) / mss, mss 1155 net/ipv4/tcp_input.c unsigned int mss; mss 1162 net/ipv4/tcp_input.c mss = tcp_skb_mss(skb); mss 1167 net/ipv4/tcp_input.c if (pkt_len < mss) mss 1168 net/ipv4/tcp_input.c pkt_len = mss; mss 1171 net/ipv4/tcp_input.c if (pkt_len < mss) mss 1178 net/ipv4/tcp_input.c if (pkt_len > mss) { mss 1179 net/ipv4/tcp_input.c unsigned int new_len = (pkt_len / mss) * mss; mss 1181 net/ipv4/tcp_input.c new_len += mss; mss 1189 net/ipv4/tcp_input.c pkt_len, mss, GFP_ATOMIC); mss 1285 net/ipv4/tcp_input.c unsigned int pcount, int shifted, int mss, mss 1321 net/ipv4/tcp_input.c TCP_SKB_CB(prev)->tcp_gso_size = mss; mss 1403 net/ipv4/tcp_input.c int mss; mss 1435 net/ipv4/tcp_input.c mss = tcp_skb_seglen(skb); mss 1440 net/ipv4/tcp_input.c if (mss != tcp_skb_seglen(prev)) mss 1476 net/ipv4/tcp_input.c mss = tcp_skb_mss(skb); mss 1481 net/ipv4/tcp_input.c if (mss != tcp_skb_seglen(prev)) mss 1484 net/ipv4/tcp_input.c if (len == mss) { mss 1486 net/ipv4/tcp_input.c } else if (len < mss) { mss 1489 net/ipv4/tcp_input.c pcount = len / mss; mss 1490 net/ipv4/tcp_input.c len = pcount * mss; mss 1500 net/ipv4/tcp_input.c if (!tcp_shifted_skb(sk, prev, skb, state, pcount, len, mss, dup_sack)) mss 1512 net/ipv4/tcp_input.c (mss != tcp_skb_seglen(skb))) mss 1519 net/ipv4/tcp_input.c len, mss, 0); mss 2195 net/ipv4/tcp_input.c unsigned int mss; mss 2231 net/ipv4/tcp_input.c mss = tcp_skb_mss(skb); mss 2233 net/ipv4/tcp_input.c lost = (packets - oldcnt) * mss; mss 2236 net/ipv4/tcp_input.c lost, mss, GFP_ATOMIC) < 0) mss 2604 net/ipv4/tcp_input.c unsigned int mss = tcp_current_mss(sk); mss 2607 net/ipv4/tcp_input.c if (tcp_skb_seglen(skb) > mss && mss 3799 net/ipv4/tcp_input.c u16 mss = 0; mss 3807 net/ipv4/tcp_input.c return mss; mss 3813 net/ipv4/tcp_input.c return mss; mss 3816 net/ipv4/tcp_input.c return mss; mss 3818 net/ipv4/tcp_input.c return mss; /* fail on partial options */ mss 3825 net/ipv4/tcp_input.c mss = in_mss; mss 3832 net/ipv4/tcp_input.c return mss; mss 5789 net/ipv4/tcp_input.c u16 mss = tp->rx_opt.mss_clamp, try_exp = 0; mss 5792 net/ipv4/tcp_input.c if (mss == tp->rx_opt.user_mss) { mss 5799 net/ipv4/tcp_input.c mss = opt.mss_clamp; mss 5820 net/ipv4/tcp_input.c tcp_fastopen_cache_set(sk, mss, cookie, syn_drop, try_exp); mss 6443 net/ipv4/tcp_input.c req->mss = rx_opt->mss_clamp; mss 6536 net/ipv4/tcp_input.c u16 mss; mss 6550 net/ipv4/tcp_input.c mss = tcp_parse_mss_option(th, tp->rx_opt.user_mss); mss 6551 net/ipv4/tcp_input.c if (!mss) mss 6552 net/ipv4/tcp_input.c mss = af_ops->mss_clamp; mss 6554 net/ipv4/tcp_input.c return mss; mss 6651 net/ipv4/tcp_input.c isn = cookie_init_sequence(af_ops, sk, skb, &req->mss); mss 1528 net/ipv4/tcp_ipv4.c u16 mss = 0; mss 1530 net/ipv4/tcp_ipv4.c mss = tcp_get_syncookie_mss(&tcp_request_sock_ops, mss 1532 net/ipv4/tcp_ipv4.c if (mss) { mss 1533 net/ipv4/tcp_ipv4.c *cookie = __cookie_v4_init_sequence(iph, th, &mss); mss 1537 net/ipv4/tcp_ipv4.c return mss; mss 29 net/ipv4/tcp_metrics.c u16 mss; mss 124 net/ipv4/tcp_metrics.c tm->tcpm_fastopen.mss = 0; mss 538 net/ipv4/tcp_metrics.c void tcp_fastopen_cache_get(struct sock *sk, u16 *mss, mss 551 net/ipv4/tcp_metrics.c if (tfom->mss) mss 552 net/ipv4/tcp_metrics.c *mss = tfom->mss; mss 561 net/ipv4/tcp_metrics.c void tcp_fastopen_cache_set(struct sock *sk, u16 mss, mss 576 net/ipv4/tcp_metrics.c if (mss) mss 577 net/ipv4/tcp_metrics.c tfom->mss = mss; mss 693 net/ipv4/tcp_metrics.c if (tfom->mss && mss 695 net/ipv4/tcp_metrics.c tfom->mss) < 0) mss 370 net/ipv4/tcp_minisocks.c int mss; mss 372 net/ipv4/tcp_minisocks.c mss = tcp_mss_clamp(tp, dst_metric_advmss(dst)); mss 385 net/ipv4/tcp_minisocks.c else if (full_space < rcv_wnd * mss) mss 386 net/ipv4/tcp_minisocks.c full_space = rcv_wnd * mss; mss 390 net/ipv4/tcp_minisocks.c mss - (ireq->tstamp_ok ? TCPOLEN_TSTAMP_ALIGNED : 0), mss 546 net/ipv4/tcp_minisocks.c newtp->rx_opt.mss_clamp = req->mss; mss 15 net/ipv4/tcp_offload.c unsigned int seq, unsigned int mss) mss 18 net/ipv4/tcp_offload.c if (before(ts_seq, seq + mss)) { mss 25 net/ipv4/tcp_offload.c seq += mss; mss 64 net/ipv4/tcp_offload.c unsigned int mss; mss 80 net/ipv4/tcp_offload.c mss = skb_shinfo(skb)->gso_size; mss 81 net/ipv4/tcp_offload.c if (unlikely(skb->len <= mss)) mss 87 net/ipv4/tcp_offload.c skb_shinfo(skb)->gso_segs = DIV_ROUND_UP(skb->len, mss); mss 110 net/ipv4/tcp_offload.c mss *= skb_shinfo(segs)->gso_segs; mss 112 net/ipv4/tcp_offload.c delta = htonl(oldlen + (thlen + mss)); mss 119 net/ipv4/tcp_offload.c tcp_gso_tstamp(segs, skb_shinfo(gso_skb)->tskey, seq, mss); mss 133 net/ipv4/tcp_offload.c seq += mss; mss 189 net/ipv4/tcp_offload.c unsigned int mss = 1; mss 258 net/ipv4/tcp_offload.c mss = skb_shinfo(p)->gso_size; mss 260 net/ipv4/tcp_offload.c flush |= (len - 1) >= mss; mss 267 net/ipv4/tcp_offload.c mss = 1; mss 274 net/ipv4/tcp_offload.c flush = len < mss; mss 123 net/ipv4/tcp_output.c int mss = tp->advmss; mss 128 net/ipv4/tcp_output.c if (metric < mss) { mss 129 net/ipv4/tcp_output.c mss = metric; mss 130 net/ipv4/tcp_output.c tp->advmss = mss; mss 134 net/ipv4/tcp_output.c return (__u16)mss; mss 207 net/ipv4/tcp_output.c void tcp_select_initial_window(const struct sock *sk, int __space, __u32 mss, mss 220 net/ipv4/tcp_output.c if (space > mss) mss 221 net/ipv4/tcp_output.c space = rounddown(space, mss); mss 237 net/ipv4/tcp_output.c *rcv_wnd = min(*rcv_wnd, init_rcv_wnd * mss); mss 435 net/ipv4/tcp_output.c u16 mss; /* 0 to disable */ mss 470 net/ipv4/tcp_output.c if (unlikely(opts->mss)) { mss 473 net/ipv4/tcp_output.c opts->mss); mss 619 net/ipv4/tcp_output.c opts->mss = tcp_advertise_mss(sk); mss 662 net/ipv4/tcp_output.c unsigned int mss, struct sk_buff *skb, mss 685 net/ipv4/tcp_output.c opts->mss = mss; mss 1507 net/ipv4/tcp_output.c int tcp_mss_to_mtu(struct sock *sk, int mss) mss 1513 net/ipv4/tcp_output.c mtu = mss + mss 2052 net/ipv4/tcp_output.c int mss = tcp_current_mss(sk); mss 2059 net/ipv4/tcp_output.c icsk->icsk_mtup.search_low = tcp_mss_to_mtu(sk, mss); mss 2563 net/ipv4/tcp_output.c int mss = tcp_current_mss(sk); mss 2566 net/ipv4/tcp_output.c if (skb && tcp_snd_wnd_test(tp, skb, mss)) { mss 2568 net/ipv4/tcp_output.c tcp_write_xmit(sk, mss, TCP_NAGLE_OFF, 2, GFP_ATOMIC); mss 2577 net/ipv4/tcp_output.c tp->packets_out, sk->sk_state, tp->snd_cwnd, mss); mss 2593 net/ipv4/tcp_output.c if ((pcount > 1) && (skb->len > (pcount - 1) * mss)) { mss 2595 net/ipv4/tcp_output.c (pcount - 1) * mss, mss, mss 2711 net/ipv4/tcp_output.c int mss = icsk->icsk_ack.rcv_mss; mss 2717 net/ipv4/tcp_output.c if (unlikely(mss > full_space)) { mss 2718 net/ipv4/tcp_output.c mss = full_space; mss 2719 net/ipv4/tcp_output.c if (mss <= 0) mss 2741 net/ipv4/tcp_output.c if (free_space < (allowed_space >> 4) || free_space < mss) mss 2769 net/ipv4/tcp_output.c if (window <= free_space - mss || window > free_space) mss 2770 net/ipv4/tcp_output.c window = rounddown(free_space, mss); mss 2771 net/ipv4/tcp_output.c else if (mss == full_space && mss 3278 net/ipv4/tcp_output.c int mss; mss 3308 net/ipv4/tcp_output.c mss = tcp_mss_clamp(tp, dst_metric_advmss(dst)); mss 3328 net/ipv4/tcp_output.c tcp_header_size = tcp_synack_options(sk, req, mss, skb, &opts, md5, mss 3781 net/ipv4/tcp_output.c unsigned int mss = tcp_current_mss(sk); mss 3792 net/ipv4/tcp_output.c skb->len > mss) { mss 3793 net/ipv4/tcp_output.c seg_size = min(seg_size, mss); mss 3796 net/ipv4/tcp_output.c skb, seg_size, mss, GFP_ATOMIC)) mss 3799 net/ipv4/tcp_output.c tcp_set_skb_tso_segs(skb, mss); mss 239 net/ipv4/tcp_recovery.c u32 mss; mss 244 net/ipv4/tcp_recovery.c mss = tcp_skb_mss(skb); mss 245 net/ipv4/tcp_recovery.c if (tcp_skb_pcount(skb) > 1 && skb->len > mss) mss 247 net/ipv4/tcp_recovery.c mss, mss, GFP_ATOMIC); mss 145 net/ipv4/tcp_timer.c int mss; mss 155 net/ipv4/tcp_timer.c mss = tcp_mtu_to_mss(sk, icsk->icsk_mtup.search_low) >> 1; mss 156 net/ipv4/tcp_timer.c mss = min(net->ipv4.sysctl_tcp_base_mss, mss); mss 157 net/ipv4/tcp_timer.c mss = max(mss, net->ipv4.sysctl_tcp_mtu_probe_floor); mss 158 net/ipv4/tcp_timer.c mss = max(mss, net->ipv4.sysctl_tcp_min_snd_mss); mss 159 net/ipv4/tcp_timer.c icsk->icsk_mtup.search_low = tcp_mss_to_mtu(sk, mss); mss 194 net/ipv4/udp_offload.c unsigned int mss; mss 199 net/ipv4/udp_offload.c mss = skb_shinfo(gso_skb)->gso_size; mss 200 net/ipv4/udp_offload.c if (gso_skb->len <= sizeof(*uh) + mss) mss 222 net/ipv4/udp_offload.c mss *= skb_shinfo(segs)->gso_segs; mss 233 net/ipv4/udp_offload.c newlen = htons(sizeof(*uh) + mss); mss 292 net/ipv4/udp_offload.c unsigned int mss; mss 313 net/ipv4/udp_offload.c mss = skb_shinfo(skb)->gso_size; mss 314 net/ipv4/udp_offload.c if (unlikely(skb->len <= mss)) mss 40 net/ipv6/netfilter/ip6t_SYNPROXY.c opts.mss_option = info->mss; mss 96 net/ipv6/syncookies.c const __u16 mss = *mssp; mss 99 net/ipv6/syncookies.c if (mss >= msstab[mssind]) mss 139 net/ipv6/syncookies.c int mss; mss 150 net/ipv6/syncookies.c mss = __cookie_v6_check(ipv6_hdr(skb), th, cookie); mss 151 net/ipv6/syncookies.c if (mss == 0) { mss 184 net/ipv6/syncookies.c req->mss = mss; mss 244 net/ipv6/syncookies.c tcp_select_initial_window(sk, tcp_full_space(sk), req->mss, mss 1076 net/ipv6/tcp_ipv6.c u16 mss = 0; mss 1078 net/ipv6/tcp_ipv6.c mss = tcp_get_syncookie_mss(&tcp6_request_sock_ops, mss 1080 net/ipv6/tcp_ipv6.c if (mss) { mss 1081 net/ipv6/tcp_ipv6.c *cookie = __cookie_v6_init_sequence(iph, th, &mss); mss 1085 net/ipv6/tcp_ipv6.c return mss; mss 21 net/ipv6/udp_offload.c unsigned int mss; mss 31 net/ipv6/udp_offload.c mss = skb_shinfo(skb)->gso_size; mss 32 net/ipv6/udp_offload.c if (unlikely(skb->len <= mss)) mss 473 net/netfilter/nf_synproxy_core.c u16 mss = opts->mss_encode; mss 490 net/netfilter/nf_synproxy_core.c nth->seq = htonl(__cookie_v4_init_sequence(iph, th, &mss)); mss 636 net/netfilter/nf_synproxy_core.c int mss; mss 638 net/netfilter/nf_synproxy_core.c mss = __cookie_v4_check(ip_hdr(skb), th, ntohl(th->ack_seq) - 1); mss 639 net/netfilter/nf_synproxy_core.c if (mss == 0) { mss 645 net/netfilter/nf_synproxy_core.c opts->mss_option = mss; mss 887 net/netfilter/nf_synproxy_core.c u16 mss = opts->mss_encode; mss 904 net/netfilter/nf_synproxy_core.c nth->seq = htonl(nf_ipv6_cookie_init_sequence(iph, th, &mss)); mss 1054 net/netfilter/nf_synproxy_core.c int mss; mss 1056 net/netfilter/nf_synproxy_core.c mss = nf_cookie_v6_check(ipv6_hdr(skb), th, ntohl(th->ack_seq) - 1); mss 1057 net/netfilter/nf_synproxy_core.c if (mss == 0) { mss 1063 net/netfilter/nf_synproxy_core.c opts->mss_option = mss; mss 73 net/netfilter/nfnetlink_osf.c u16 mss = 0; mss 106 net/netfilter/nfnetlink_osf.c mss = ctx->optp[3]; mss 107 net/netfilter/nfnetlink_osf.c mss <<= 8; mss 108 net/netfilter/nfnetlink_osf.c mss |= ctx->optp[2]; mss 110 net/netfilter/nfnetlink_osf.c mss = ntohs((__force __be16)mss); mss 141 net/netfilter/nfnetlink_osf.c if (ctx->window == f->wss.val * mss || mss 147 net/netfilter/nfnetlink_osf.c if (ctx->window == f->wss.val * (mss + 40) || mss 35 net/netfilter/nft_synproxy.c opts->mss_option = info->mss; mss 161 net/netfilter/nft_synproxy.c priv->info.mss = ntohs(nla_get_be16(tb[NFTA_SYNPROXY_MSS])); mss 230 net/netfilter/nft_synproxy.c if (nla_put_be16(skb, NFTA_SYNPROXY_MSS, htons(priv->info.mss)) || mss 102 net/netfilter/xt_TCPMSS.c if (info->mss == XT_TCPMSS_CLAMP_PMTU) { mss 114 net/netfilter/xt_TCPMSS.c newmss = info->mss; mss 269 net/netfilter/xt_TCPMSS.c if (info->mss == XT_TCPMSS_CLAMP_PMTU && mss 293 net/netfilter/xt_TCPMSS.c if (info->mss == XT_TCPMSS_CLAMP_PMTU && mss 135 tools/testing/selftests/bpf/progs/test_tcp_check_syncookie_kern.c __u32 mss = seq_mss >> 32; mss 138 tools/testing/selftests/bpf/progs/test_tcp_check_syncookie_kern.c bpf_map_update_elem(&results, &key_mss, &mss, 0); mss 310 tools/testing/selftests/net/tcp_mmap.c int mss = 0; mss 341 tools/testing/selftests/net/tcp_mmap.c mss = atoi(optarg); mss 368 tools/testing/selftests/net/tcp_mmap.c if (mss && mss 370 tools/testing/selftests/net/tcp_mmap.c &mss, sizeof(mss)) == -1) { mss 400 tools/testing/selftests/net/tcp_mmap.c if (mss && mss 401 tools/testing/selftests/net/tcp_mmap.c setsockopt(fd, IPPROTO_TCP, TCP_MAXSEG, &mss, sizeof(mss)) == -1) { mss 508 tools/testing/selftests/net/udpgso.c int i, ret, val, mss; mss 531 tools/testing/selftests/net/udpgso.c mss = test->gso_len; mss 533 tools/testing/selftests/net/udpgso.c mss = addr->sa_family == AF_INET ? CONST_MSS_V4 : CONST_MSS_V6; mss 539 tools/testing/selftests/net/udpgso.c if (ret != mss) mss 540 tools/testing/selftests/net/udpgso.c error(1, 0, "recv.%d: %d != %d", i, ret, mss);