mss               132 arch/powerpc/platforms/cell/spufs/context.c 	if (ctx->mss)
mss               133 arch/powerpc/platforms/cell/spufs/context.c 		unmap_mapping_range(ctx->mss, 0, SPUFS_MSS_MAP_SIZE, 1);
mss              1337 arch/powerpc/platforms/cell/spufs/file.c 		ctx->mss = inode->i_mapping;
mss              1350 arch/powerpc/platforms/cell/spufs/file.c 		ctx->mss = NULL;
mss                76 arch/powerpc/platforms/cell/spufs/spufs.h 	struct address_space *mss;	   /* 'mss' area mappings. */
mss               164 arch/sparc/kernel/ldc.c 	u8				mss;
mss               253 arch/sparc/kernel/ldc.c 	unsigned int mss;
mss               267 arch/sparc/kernel/ldc.c 	mss = lp->mss;
mss               269 arch/sparc/kernel/ldc.c 	if (diff * mss < size)
mss              1127 arch/sparc/kernel/ldc.c 	u8 mss, *mssbuf;
mss              1143 arch/sparc/kernel/ldc.c 		mss = LDC_PACKET_SIZE;
mss              1148 arch/sparc/kernel/ldc.c 		mss = LDC_PACKET_SIZE - 8;
mss              1153 arch/sparc/kernel/ldc.c 		mss = LDC_PACKET_SIZE - 8 - 8;
mss              1188 arch/sparc/kernel/ldc.c 	lp->mss = mss;
mss              1607 arch/sparc/kernel/ldc.c 		if (data_len > lp->mss)
mss              1608 arch/sparc/kernel/ldc.c 			data_len = lp->mss;
mss               207 drivers/crypto/chelsio/chtls/chtls.h 	u32 mss;
mss                68 drivers/crypto/chelsio/chtls/chtls_cm.c 	csk->mss = MAX_MSS;
mss               870 drivers/crypto/chelsio/chtls/chtls_cm.c 	unsigned int mss;
mss               873 drivers/crypto/chelsio/chtls/chtls_cm.c 	mss = ntohs(req->tcpopt.mss);
mss               889 drivers/crypto/chelsio/chtls/chtls_cm.c 	if (mss && tp->advmss > mss)
mss               890 drivers/crypto/chelsio/chtls/chtls_cm.c 		tp->advmss = mss;
mss              1203 drivers/crypto/chelsio/chtls/chtls_cm.c 	oreq->mss = 0;
mss               990 drivers/crypto/chelsio/chtls/chtls_io.c 	int mss, flags, err;
mss              1010 drivers/crypto/chelsio/chtls/chtls_io.c 	mss = csk->mss;
mss              1018 drivers/crypto/chelsio/chtls/chtls_io.c 			copy = mss - skb->len;
mss              1060 drivers/crypto/chelsio/chtls/chtls_io.c 			copy = mss;
mss              1151 drivers/crypto/chelsio/chtls/chtls_io.c 		if (unlikely(skb->len == mss))
mss              1205 drivers/crypto/chelsio/chtls/chtls_io.c 	int mss, err, copied;
mss              1220 drivers/crypto/chelsio/chtls/chtls_io.c 	mss = csk->mss;
mss              1228 drivers/crypto/chelsio/chtls/chtls_io.c 		    (copy = mss - skb->len) <= 0) {
mss              1244 drivers/crypto/chelsio/chtls/chtls_io.c 			copy = mss;
mss              1262 drivers/crypto/chelsio/chtls/chtls_io.c 		if (skb->len == mss)
mss               254 drivers/infiniband/hw/cxgb4/cm.c 	ep->mss = ep->emss;
mss               261 drivers/infiniband/hw/cxgb4/cm.c 			 TCPOPT_MSS_G(opt), ep->mss, ep->emss);
mss               262 drivers/infiniband/hw/cxgb4/cm.c 	pr_debug("mss_idx %u mss %u emss=%u\n", TCPOPT_MSS_G(opt), ep->mss,
mss              2525 drivers/infiniband/hw/cxgb4/cm.c 	u16 peer_mss = ntohs(req->tcpopt.mss);
mss              3982 drivers/infiniband/hw/cxgb4/cm.c 	req->tcpopt.mss = htons(tmp_opt.mss_clamp);
mss               899 drivers/infiniband/hw/cxgb4/iw_cxgb4.h 	u16 mss;
mss               667 drivers/infiniband/hw/i40iw/i40iw_cm.c 			tmp = ntohs(all_options->as_mss.mss);
mss               668 drivers/infiniband/hw/i40iw/i40iw_cm.c 			if (tmp > 0 && tmp < cm_node->tcp_cntxt.mss)
mss               669 drivers/infiniband/hw/i40iw/i40iw_cm.c 				cm_node->tcp_cntxt.mss = tmp;
mss               685 drivers/infiniband/hw/i40iw/i40iw_cm.c 		cm_node->tcp_cntxt.mss = I40IW_CM_DEFAULT_MSS;
mss              1360 drivers/infiniband/hw/i40iw/i40iw_cm.c 	options->as_mss.mss = htons(cm_node->tcp_cntxt.mss);
mss              2217 drivers/infiniband/hw/i40iw/i40iw_cm.c 		cm_node->tcp_cntxt.mss = iwdev->vsi.mtu - I40IW_MTU_TO_MSS_IPV4;
mss              2230 drivers/infiniband/hw/i40iw/i40iw_cm.c 		cm_node->tcp_cntxt.mss = iwdev->vsi.mtu - I40IW_MTU_TO_MSS_IPV6;
mss              3324 drivers/infiniband/hw/i40iw/i40iw_cm.c 	tcp_info->cwnd = cpu_to_le32(2 * cm_node->tcp_cntxt.mss);
mss              3332 drivers/infiniband/hw/i40iw/i40iw_cm.c 	tcp_info->snd_mss = cpu_to_le32(((u32)cm_node->tcp_cntxt.mss));
mss               146 drivers/infiniband/hw/i40iw/i40iw_cm.h 	__be16 mss;
mss               276 drivers/infiniband/hw/i40iw/i40iw_cm.h 	u32 mss;
mss               108 drivers/infiniband/hw/i40iw/i40iw_puda.h 	u16 mss;
mss               126 drivers/infiniband/hw/i40iw/i40iw_puda.h 	u16 mss;
mss              3506 drivers/infiniband/hw/mlx4/qp.c 	*lso_hdr_sz  = cpu_to_be32(wr->mss << 16 | wr->hlen);
mss              4113 drivers/infiniband/hw/mlx5/qp.c 		eseg->mss = cpu_to_be16(ud_wr->mss);
mss               590 drivers/infiniband/hw/qedr/qedr_iw_cm.c 		in_params.mss = dev->iwarp_max_mtu -
mss               606 drivers/infiniband/hw/qedr/qedr_iw_cm.c 		in_params.mss = dev->iwarp_max_mtu -
mss               556 drivers/infiniband/ulp/ipoib/ipoib_ib.c 		priv->tx_wr.mss		= skb_shinfo(skb)->gso_size;
mss               114 drivers/net/ethernet/amazon/ena/ena_com.h 	u16 mss;
mss               302 drivers/net/ethernet/amazon/ena/ena_eth_com.c 	meta_desc->word2 |= (ena_meta->mss <<
mss               306 drivers/net/ethernet/amazon/ena/ena_eth_com.c 	meta_desc->len_ctrl |= ((ena_meta->mss >> 10) <<
mss              2062 drivers/net/ethernet/amazon/ena/ena_netdev.c 	u32 mss = skb_shinfo(skb)->gso_size;
mss              2066 drivers/net/ethernet/amazon/ena/ena_netdev.c 	if ((skb->ip_summed == CHECKSUM_PARTIAL) || mss) {
mss              2068 drivers/net/ethernet/amazon/ena/ena_netdev.c 		if (mss) {
mss              2083 drivers/net/ethernet/amazon/ena/ena_netdev.c 			if (mss)
mss              2100 drivers/net/ethernet/amazon/ena/ena_netdev.c 		ena_meta->mss = mss;
mss               551 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 	if ((tso && (packet->mss != ring->tx.cur_mss)) ||
mss              1695 drivers/net/ethernet/amd/xgbe/xgbe-dev.c 	if (tso && (packet->mss != ring->tx.cur_mss))
mss              1733 drivers/net/ethernet/amd/xgbe/xgbe-dev.c 				  packet->mss);
mss              1737 drivers/net/ethernet/amd/xgbe/xgbe-dev.c 					  MSS, packet->mss);
mss              1747 drivers/net/ethernet/amd/xgbe/xgbe-dev.c 			ring->tx.cur_mss = packet->mss;
mss              1760 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 	packet->mss = skb_shinfo(skb)->gso_size;
mss              1765 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 	DBGPR("  packet->mss=%u\n", packet->mss);
mss               363 drivers/net/ethernet/amd/xgbe/xgbe.h 	unsigned short mss;
mss               276 drivers/net/ethernet/apm/xgene/xgene_enet_main.c static int xgene_enet_setup_mss(struct net_device *ndev, u32 mss)
mss               286 drivers/net/ethernet/apm/xgene/xgene_enet_main.c 		if (pdata->mss[i] == mss) {
mss               296 drivers/net/ethernet/apm/xgene/xgene_enet_main.c 			pdata->mac_ops->set_mss(pdata, mss, i);
mss               297 drivers/net/ethernet/apm/xgene/xgene_enet_main.c 			pdata->mss[i] = mss;
mss               313 drivers/net/ethernet/apm/xgene/xgene_enet_main.c 	u32 hdr_len, mss = 0;
mss               336 drivers/net/ethernet/apm/xgene/xgene_enet_main.c 			mss = skb_shinfo(skb)->gso_size;
mss               353 drivers/net/ethernet/apm/xgene/xgene_enet_main.c 			if (!mss || ((skb->len - hdr_len) <= mss))
mss               356 drivers/net/ethernet/apm/xgene/xgene_enet_main.c 			mss_index = xgene_enet_setup_mss(ndev, mss);
mss               153 drivers/net/ethernet/apm/xgene/xgene_enet_main.h 	void (*set_mss)(struct xgene_enet_pdata *pdata, u16 mss, u8 index);
mss               230 drivers/net/ethernet/apm/xgene/xgene_enet_main.h 	u32 mss[NUM_MSS_REG];
mss               222 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c 				u16 mss, u8 index)
mss               232 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c 			SET_VAL(TSO_MSS0, mss);
mss               234 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c 		data = SET_VAL(TSO_MSS1, mss) | SET_VAL(TSO_MSS0, data);
mss               438 drivers/net/ethernet/aquantia/atlantic/aq_nic.c 		dx_buff->mss = skb_shinfo(skb)->gso_size;
mss                56 drivers/net/ethernet/aquantia/atlantic/aq_ring.h 			u32 mss;
mss               459 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c 			txd->ctl2 |= (buff->mss << 16) |
mss               515 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 			txd->ctl2 |= (buff->mss << 16);
mss              2349 drivers/net/ethernet/atheros/atlx/atl1.c 	unsigned int mss = 0;
mss              2367 drivers/net/ethernet/atheros/atlx/atl1.c 	mss = skb_shinfo(skb)->gso_size;
mss              2368 drivers/net/ethernet/atheros/atlx/atl1.c 	if (mss) {
mss              6601 drivers/net/ethernet/broadcom/bnx2.c 	u32 len, vlan_tag_flags, last_frag, mss;
mss              6635 drivers/net/ethernet/broadcom/bnx2.c 	if ((mss = skb_shinfo(skb)->gso_size)) {
mss              6657 drivers/net/ethernet/broadcom/bnx2.c 				mss |= (tcp_off & 0xc) << TX_BD_TCP6_OFF2_SHL;
mss              6667 drivers/net/ethernet/broadcom/bnx2.c 		mss = 0;
mss              6683 drivers/net/ethernet/broadcom/bnx2.c 	txbd->tx_bd_mss_nbytes = len | (mss << 16);
mss              6707 drivers/net/ethernet/broadcom/bnx2.c 		txbd->tx_bd_mss_nbytes = len | (mss << 16);
mss              4750 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		u16 mss = bp->dev->mtu - 40;
mss              4771 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		if (mss <= BNXT_RX_PAGE_SIZE) {
mss              4772 drivers/net/ethernet/broadcom/bnxt/bnxt.c 			n = BNXT_RX_PAGE_SIZE / mss;
mss              4775 drivers/net/ethernet/broadcom/bnxt/bnxt.c 			n = mss / BNXT_RX_PAGE_SIZE;
mss              4776 drivers/net/ethernet/broadcom/bnxt/bnxt.c 			if (mss & (BNXT_RX_PAGE_SIZE - 1))
mss              2067 drivers/net/ethernet/broadcom/cnic.c 	xstorm_buf->mss = 0xffff;
mss              3512 drivers/net/ethernet/broadcom/cnic.c 		l4kwqe3->mss = l4kwqe3->pmtu - sizeof(struct ipv6hdr) -
mss              3516 drivers/net/ethernet/broadcom/cnic.c 		l4kwqe3->mss = l4kwqe3->pmtu - sizeof(struct iphdr) -
mss               359 drivers/net/ethernet/broadcom/cnic_defs.h 	u16 mss;
mss               361 drivers/net/ethernet/broadcom/cnic_defs.h 	u16 mss;
mss              1733 drivers/net/ethernet/broadcom/cnic_defs.h 	u16 mss;
mss              1735 drivers/net/ethernet/broadcom/cnic_defs.h 	u16 mss;
mss              4090 drivers/net/ethernet/broadcom/cnic_defs.h 	u16 mss;
mss              4096 drivers/net/ethernet/broadcom/cnic_defs.h 	u16 mss;
mss              5116 drivers/net/ethernet/broadcom/cnic_defs.h 	u16 mss;
mss              5120 drivers/net/ethernet/broadcom/cnic_defs.h 	u16 mss;
mss              7672 drivers/net/ethernet/broadcom/tg3.c 					   u32 len, u32 mss)
mss              7674 drivers/net/ethernet/broadcom/tg3.c 	if (tg3_asic_rev(tp) == ASIC_REV_5762 && mss) {
mss              7677 drivers/net/ethernet/broadcom/tg3.c 		return ((base + len + (mss & 0x3fff)) < base);
mss              7697 drivers/net/ethernet/broadcom/tg3.c 				 u32 mss, u32 vlan)
mss              7702 drivers/net/ethernet/broadcom/tg3.c 	txbd->vlan_tag = (mss << TXD_MSS_SHIFT) | (vlan << TXD_VLAN_TAG_SHIFT);
mss              7707 drivers/net/ethernet/broadcom/tg3.c 			    u32 mss, u32 vlan)
mss              7718 drivers/net/ethernet/broadcom/tg3.c 	if (tg3_4g_tso_overflow_test(tp, map, len, mss))
mss              7740 drivers/net/ethernet/broadcom/tg3.c 				      frag_len, tmp_flag, mss, vlan);
mss              7751 drivers/net/ethernet/broadcom/tg3.c 					      len, flags, mss, vlan);
mss              7761 drivers/net/ethernet/broadcom/tg3.c 			      len, flags, mss, vlan);
mss              7810 drivers/net/ethernet/broadcom/tg3.c 				       u32 base_flags, u32 mss, u32 vlan)
mss              7848 drivers/net/ethernet/broadcom/tg3.c 					    mss, vlan)) {
mss              7918 drivers/net/ethernet/broadcom/tg3.c 	u32 len, entry, base_flags, mss, vlan = 0;
mss              7956 drivers/net/ethernet/broadcom/tg3.c 	mss = skb_shinfo(skb)->gso_size;
mss              7957 drivers/net/ethernet/broadcom/tg3.c 	if (mss) {
mss              7988 drivers/net/ethernet/broadcom/tg3.c 			iph->tot_len = htons(mss + hdr_len);
mss              8008 drivers/net/ethernet/broadcom/tg3.c 			mss |= (hdr_len & 0xc) << 12;
mss              8013 drivers/net/ethernet/broadcom/tg3.c 			mss |= hdr_len << 9;
mss              8020 drivers/net/ethernet/broadcom/tg3.c 				mss |= (tsflags << 11);
mss              8044 drivers/net/ethernet/broadcom/tg3.c 	    !mss && skb->len > VLAN_ETH_FRAME_LEN)
mss              8075 drivers/net/ethernet/broadcom/tg3.c 			    mss, vlan)) {
mss              8078 drivers/net/ethernet/broadcom/tg3.c 		u32 tmp_mss = mss;
mss              8116 drivers/net/ethernet/broadcom/tg3.c 		if (mss && tg3_tso_bug_gso_check(tnapi, skb)) {
mss              8134 drivers/net/ethernet/broadcom/tg3.c 						base_flags, mss, vlan))
mss              13429 drivers/net/ethernet/broadcom/tg3.c 	u32 base_flags = 0, mss = 0, desc_idx, coal_now, data_off, val;
mss              13470 drivers/net/ethernet/broadcom/tg3.c 		mss = TG3_TSO_MSS;
mss              13476 drivers/net/ethernet/broadcom/tg3.c 		iph->tot_len = htons((u16)(mss + hdr_len));
mss              13492 drivers/net/ethernet/broadcom/tg3.c 			mss |= (hdr_len & 0xc) << 12;
mss              13497 drivers/net/ethernet/broadcom/tg3.c 			mss |= hdr_len << 9;
mss              13500 drivers/net/ethernet/broadcom/tg3.c 			mss |= (TG3_TSO_TCP_OPT_LEN << 9);
mss              13537 drivers/net/ethernet/broadcom/tg3.c 			    base_flags | TXD_FLAG_END, mss, 0)) {
mss               179 drivers/net/ethernet/chelsio/cxgb/cpl5_cmd.h 	u16 mss;
mss               634 drivers/net/ethernet/chelsio/cxgb/cpl5_cmd.h 	u32 mss;
mss              1752 drivers/net/ethernet/chelsio/cxgb/sge.c #define MK_ETH_TYPE_MSS(type, mss) (((mss) & 0x3FFF) | ((type) << 14))
mss               210 drivers/net/ethernet/chelsio/cxgb3/t3_cpl.h 	__be16 mss;
mss              1526 drivers/net/ethernet/chelsio/cxgb4/sge.c 			lso->mss = htons(ssi->gso_size);
mss              1833 drivers/net/ethernet/chelsio/cxgb4/sge.c 		lso->mss = cpu_to_be16(ssi->gso_size);
mss               378 drivers/net/ethernet/chelsio/cxgb4/t4_msg.h 	__be16 mss;
mss               903 drivers/net/ethernet/chelsio/cxgb4/t4_msg.h 	__be16 mss;
mss              1301 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		lso->mss = cpu_to_be16(ssi->gso_size);
mss               705 drivers/net/ethernet/cisco/enic/enic_main.c 				 struct sk_buff *skb, unsigned int mss,
mss               736 drivers/net/ethernet/cisco/enic/enic_main.c 		enic_queue_wq_desc_tso(wq, skb, dma_addr, len, mss, hdr_len,
mss               809 drivers/net/ethernet/cisco/enic/enic_main.c 	unsigned int mss = skb_shinfo(skb)->gso_size;
mss               824 drivers/net/ethernet/cisco/enic/enic_main.c 	if (mss)
mss               825 drivers/net/ethernet/cisco/enic/enic_main.c 		err = enic_queue_wq_skb_tso(enic, wq, skb, mss,
mss               113 drivers/net/ethernet/cisco/enic/enic_res.h 	unsigned int mss, unsigned int hdr_len, int vlan_tag_insert,
mss               117 drivers/net/ethernet/cisco/enic/enic_res.h 		mss, hdr_len, vlan_tag_insert, vlan_tag,
mss                54 drivers/net/ethernet/cisco/enic/wq_enet_desc.h 	u64 address, u16 length, u16 mss, u16 header_length,
mss                60 drivers/net/ethernet/cisco/enic/wq_enet_desc.h 	desc->mss_loopback = cpu_to_le16((mss & WQ_ENET_MSS_MASK) <<
mss                73 drivers/net/ethernet/cisco/enic/wq_enet_desc.h 	u64 *address, u16 *length, u16 *mss, u16 *header_length,
mss                79 drivers/net/ethernet/cisco/enic/wq_enet_desc.h 	*mss = (le16_to_cpu(desc->mss_loopback) >> WQ_ENET_MSS_SHIFT) &
mss                38 drivers/net/ethernet/google/gve/gve_desc.h 	__be16	mss;		/* TSO MSS				*/
mss                65 drivers/net/ethernet/google/gve/gve_desc.h 	__be16	mss;
mss               387 drivers/net/ethernet/google/gve/gve_tx.c 		seg_desc->seg.mss = cpu_to_be16(skb_shinfo(skb)->gso_size);
mss               187 drivers/net/ethernet/hisilicon/hns/hnae.h 			__le16 mss;
mss                50 drivers/net/ethernet/hisilicon/hns/hns_enet.c 	u16 mss = 0;
mss                95 drivers/net/ethernet/hisilicon/hns/hns_enet.c 					mss = skb_shinfo(skb)->gso_size;
mss               109 drivers/net/ethernet/hisilicon/hns/hns_enet.c 					mss = skb_shinfo(skb)->gso_size;
mss               115 drivers/net/ethernet/hisilicon/hns/hns_enet.c 			desc->tx.mss = cpu_to_le16(mss);
mss               215 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c 	dev_info(dev, "(TX)mss: %u\n", tx_desc->tx.mss);
mss               674 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 			u16 *mss, u32 *type_cs_vlan_tso)
mss               737 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	*mss = skb_shinfo(skb)->gso_size;
mss              1039 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	u16 mss = 0;
mss              1084 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 		ret = hns3_set_tso(skb, &paylen, &mss,
mss              1099 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	desc->tx.mss = cpu_to_le16(mss);
mss               276 drivers/net/ethernet/hisilicon/hns3/hns3_enet.h 			__le16 mss;
mss               538 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 	u32 mss = HINIC_MSS_DEFAULT;
mss               554 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 	*queue_info |= HINIC_SQ_CTRL_SET(mss, QUEUE_INFO_MSS);
mss               559 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 			    u32 l4_len, u32 offset, u32 ip_ident, u32 mss)
mss               581 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c 	*queue_info |= HINIC_SQ_CTRL_SET(mss, QUEUE_INFO_MSS);
mss               167 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.h 			    u32 offset, u32 ip_ident, u32 mss);
mss              1612 drivers/net/ethernet/ibm/ehea/ehea_main.c 		swqe->mss = skb_shinfo(skb)->gso_size;
mss               101 drivers/net/ethernet/ibm/ehea/ehea_qmr.h 	u16 mss;
mss               984 drivers/net/ethernet/ibm/ibmveth.c 			union ibmveth_buf_desc *descs, unsigned long mss)
mss              1001 drivers/net/ethernet/ibm/ibmveth.c 					     correlator, &correlator, mss,
mss              1023 drivers/net/ethernet/ibm/ibmveth.c 	unsigned long mss = 0;
mss              1124 drivers/net/ethernet/ibm/ibmveth.c 			mss = (unsigned long)skb_shinfo(skb)->gso_size;
mss              1138 drivers/net/ethernet/ibm/ibmveth.c 	if (ibmveth_send(adapter, descs, mss)) {
mss              1183 drivers/net/ethernet/ibm/ibmveth.c static void ibmveth_rx_mss_helper(struct sk_buff *skb, u16 mss, int lrg_pkt)
mss              1216 drivers/net/ethernet/ibm/ibmveth.c 		skb_shinfo(skb)->gso_size = mss;
mss              1302 drivers/net/ethernet/ibm/ibmveth.c 	u16 mss = 0;
mss              1331 drivers/net/ethernet/ibm/ibmveth.c 				mss = (u16)be64_to_cpu(*rxmss);
mss              1362 drivers/net/ethernet/ibm/ibmveth.c 				ibmveth_rx_mss_helper(skb, mss, lrg_pkt);
mss                53 drivers/net/ethernet/ibm/ibmveth.h 		unsigned long mss, unsigned long large_send_support)
mss                61 drivers/net/ethernet/ibm/ibmveth.h 				  corellator_in, mss);
mss              1546 drivers/net/ethernet/ibm/ibmvnic.c 		tx_crq.v1.mss = cpu_to_be16(skb_shinfo(skb)->gso_size);
mss               247 drivers/net/ethernet/ibm/ibmvnic.h 	__be16 mss;
mss               681 drivers/net/ethernet/intel/e1000/e1000_hw.h 			__le16 mss;	/* Maximum segment size */
mss              2695 drivers/net/ethernet/intel/e1000/e1000_main.c 	u16 ipcse = 0, tucse, mss;
mss              2706 drivers/net/ethernet/intel/e1000/e1000_main.c 		mss = skb_shinfo(skb)->gso_size;
mss              2744 drivers/net/ethernet/intel/e1000/e1000_main.c 		context_desc->tcp_seg_setup.fields.mss     = cpu_to_le16(mss);
mss              2823 drivers/net/ethernet/intel/e1000/e1000_main.c 			unsigned int mss)
mss              2851 drivers/net/ethernet/intel/e1000/e1000_main.c 		if (unlikely(mss && !nr_frags && size == len && size > 8))
mss              2908 drivers/net/ethernet/intel/e1000/e1000_main.c 			if (unlikely(mss && f == (nr_frags-1) &&
mss              3107 drivers/net/ethernet/intel/e1000/e1000_main.c 	unsigned int mss;
mss              3127 drivers/net/ethernet/intel/e1000/e1000_main.c 	mss = skb_shinfo(skb)->gso_size;
mss              3135 drivers/net/ethernet/intel/e1000/e1000_main.c 	if (mss) {
mss              3137 drivers/net/ethernet/intel/e1000/e1000_main.c 		max_per_txd = min(mss << 2, max_per_txd);
mss              3173 drivers/net/ethernet/intel/e1000/e1000_main.c 	if ((mss) || (skb->ip_summed == CHECKSUM_PARTIAL))
mss              3242 drivers/net/ethernet/intel/e1000/e1000_main.c 			     nr_frags, mss);
mss               309 drivers/net/ethernet/intel/e1000e/hw.h 			__le16 mss;       /* Maximum segment size */
mss              5439 drivers/net/ethernet/intel/e1000e/netdev.c 	u16 ipcse = 0, mss;
mss              5451 drivers/net/ethernet/intel/e1000e/netdev.c 	mss = skb_shinfo(skb)->gso_size;
mss              5485 drivers/net/ethernet/intel/e1000e/netdev.c 	context_desc->tcp_seg_setup.fields.mss = cpu_to_le16(mss);
mss              5795 drivers/net/ethernet/intel/e1000e/netdev.c 	unsigned int mss;
mss              5817 drivers/net/ethernet/intel/e1000e/netdev.c 	mss = skb_shinfo(skb)->gso_size;
mss              5818 drivers/net/ethernet/intel/e1000e/netdev.c 	if (mss) {
mss              5843 drivers/net/ethernet/intel/e1000e/netdev.c 	if ((mss) || (skb->ip_summed == CHECKSUM_PARTIAL))
mss                64 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c 			   txd->mss, txd->hdrlen, txd->flags);
mss               788 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	tx_desc->mss = cpu_to_le16(skb_shinfo(skb)->gso_size);
mss               882 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	tx_desc->mss = 0;
mss               668 drivers/net/ethernet/intel/fm10k/fm10k_type.h 	__le16 mss;		/* MSS for segmentation offload */
mss               570 drivers/net/ethernet/intel/ixgb/ixgb_hw.h 	__le16 mss;
mss              1190 drivers/net/ethernet/intel/ixgb/ixgb_main.c 	u16 ipcse, tucse, mss;
mss              1202 drivers/net/ethernet/intel/ixgb/ixgb_main.c 		mss = skb_shinfo(skb)->gso_size;
mss              1227 drivers/net/ethernet/intel/ixgb/ixgb_main.c 		context_desc->mss = cpu_to_le16(mss);
mss              1272 drivers/net/ethernet/intel/ixgb/ixgb_main.c 		context_desc->mss = 0;
mss              1298 drivers/net/ethernet/intel/ixgb/ixgb_main.c 	unsigned int mss = skb_shinfo(skb)->gso_size;
mss              1309 drivers/net/ethernet/intel/ixgb/ixgb_main.c 		if (unlikely(mss && !nr_frags && size == len && size > 8))
mss              1348 drivers/net/ethernet/intel/ixgb/ixgb_main.c 			if (unlikely(mss && (f == (nr_frags - 1))
mss              2285 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	unsigned int mss = 0;
mss              2381 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 				if (!mss) {
mss              2382 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 					mss = rx_ring->netdev->mtu -
mss              2386 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 					if (mss > 512)
mss              2387 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 						mss &= ~511;
mss              2391 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 								 mss);
mss              2065 drivers/net/ethernet/jme.c jme_tx_tso(struct sk_buff *skb, __le16 *mss, u8 *flags)
mss              2067 drivers/net/ethernet/jme.c 	*mss = cpu_to_le16(skb_shinfo(skb)->gso_size << TXDESC_MSS_SHIFT);
mss              2068 drivers/net/ethernet/jme.c 	if (*mss) {
mss              2164 drivers/net/ethernet/jme.c 	if (jme_tx_tso(skb, &txdesc->desc1.mss, &flags))
mss               180 drivers/net/ethernet/jme.h 			__le16	mss;
mss              1836 drivers/net/ethernet/marvell/sky2.c 	u16 mss;
mss              1862 drivers/net/ethernet/marvell/sky2.c 	mss = skb_shinfo(skb)->gso_size;
mss              1863 drivers/net/ethernet/marvell/sky2.c 	if (mss != 0) {
mss              1866 drivers/net/ethernet/marvell/sky2.c 			mss += ETH_HLEN + ip_hdrlen(skb) + tcp_hdrlen(skb);
mss              1868 drivers/net/ethernet/marvell/sky2.c   		if (mss != sky2->tx_last_mss) {
mss              1870 drivers/net/ethernet/marvell/sky2.c   			le->addr = cpu_to_le32(mss);
mss              1876 drivers/net/ethernet/marvell/sky2.c 			sky2->tx_last_mss = mss;
mss              1931 drivers/net/ethernet/marvell/sky2.c 	le->opcode = mss ? (OP_LARGESEND | HW_OWNER) : (OP_PACKET | HW_OWNER);
mss               381 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c 	u32 mss;
mss               390 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c 	for (mss = 2; mss < MAX_LSO_MSS; mss++) {
mss               391 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c 		mss_inv = div_u64(1ULL << 32, mss) >> 16;
mss               392 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c 		mlx5e_ipsec_inverse_table[mss] = htons(mss_inv);
mss               151 drivers/net/ethernet/mellanox/mlx5/core/en_accel/tls_rxtx.c 	int data_len, mss;
mss               167 drivers/net/ethernet/mellanox/mlx5/core/en_accel/tls_rxtx.c 	mss = nskb->dev->mtu - (headln - skb_network_offset(nskb));
mss               169 drivers/net/ethernet/mellanox/mlx5/core/en_accel/tls_rxtx.c 	if (data_len > mss) {
mss               170 drivers/net/ethernet/mellanox/mlx5/core/en_accel/tls_rxtx.c 		skb_shinfo(nskb)->gso_size = mss;
mss               171 drivers/net/ethernet/mellanox/mlx5/core/en_accel/tls_rxtx.c 		skb_shinfo(nskb)->gso_segs = DIV_ROUND_UP(data_len, mss);
mss               284 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c 	__be16 mss;
mss               290 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c 		mss       = cpu_to_be16(skb_shinfo(skb)->gso_size);
mss               298 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c 		mss       = 0;
mss               349 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c 	eseg->mss = mss;
mss               607 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c 	__be16 mss;
mss               613 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c 		mss       = cpu_to_be16(skb_shinfo(skb)->gso_size);
mss               621 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c 		mss       = 0;
mss               660 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c 	eseg->mss = mss;
mss              2637 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	int idx, avail, frag_cnt, frag_idx, count, mss, max_segments;
mss              2651 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	mss = 0;
mss              2655 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		mss = skb_shinfo(skb)->gso_size;
mss              2676 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		if (unlikely(!mss && (cksum_offset > 255 ||
mss              2690 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	if (mss) {		/* TSO */
mss              2712 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		pseudo_hdr_offset = mss;
mss              2780 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 			if (mss) {	/* TSO */
mss              2786 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 					chop = (cum_len_next > mss);
mss              2787 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 					cum_len_next = cum_len_next % mss;
mss              2800 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 					small = (mss <= MXGEFW_SEND_SMALL_SIZE);
mss              2822 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 			if (cksum_offset != 0 && !(mss && skb_is_gso_v6(skb))) {
mss              2848 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	if (mss)
mss              1654 drivers/net/ethernet/neterion/vxge/vxge-config.h static inline void vxge_hw_fifo_txdl_mss_set(void *txdlh, int mss)
mss              1659 drivers/net/ethernet/neterion/vxge/vxge-config.h 	txdp->control_0 |= VXGE_HW_FIFO_TXD_LSO_MSS(mss);
mss               949 drivers/net/ethernet/neterion/vxge/vxge-main.c 		int mss = vxge_tcp_mss(skb);
mss               950 drivers/net/ethernet/neterion/vxge/vxge-main.c 		if (mss) {
mss               952 drivers/net/ethernet/neterion/vxge/vxge-main.c 				dev->name, __func__, __LINE__, mss);
mss               953 drivers/net/ethernet/neterion/vxge/vxge-main.c 			vxge_hw_fifo_txdl_mss_set(dtr, mss);
mss               151 drivers/net/ethernet/netronome/nfp/nfp_net.h 			__le16 mss;	/* MSS to be used for LSO */
mss               731 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	u16 mss;
mss               750 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	mss = skb_shinfo(skb)->gso_size & PCIE_DESC_TX_MSS_MASK;
mss               754 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->mss = cpu_to_le16(mss);
mss              1044 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->mss = 0;
mss              1783 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->mss = 0;
mss              2131 drivers/net/ethernet/netronome/nfp/nfp_net_common.c 	txd->mss = 0;
mss               707 drivers/net/ethernet/pensando/ionic/ionic_if.h 		__le16  mss;
mss               506 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 			      unsigned int hdrlen, unsigned int mss,
mss               524 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 	desc->mss = cpu_to_le16(mss);
mss               564 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 	unsigned int mss;
mss               575 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 	mss = skb_shinfo(skb)->gso_size;
mss               602 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 	seglen = hdrlen + mss;
mss               625 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 				  hdrlen, mss,
mss               633 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 		seglen = mss;
mss               664 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 						  hdrlen, mss, outer_csum,
mss               672 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 				len = min(mss, left);
mss               673 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 				frag_left = mss - len;
mss               687 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 						  hdrlen, mss, outer_csum,
mss               323 drivers/net/ethernet/qlogic/netxen/netxen_nic.h 	__le16 mss;
mss              1887 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c 		first_desc->mss = cpu_to_le16(skb_shinfo(skb)->gso_size);
mss              10381 drivers/net/ethernet/qlogic/qed/qed_hsi.h 	__le16 mss;
mss              10454 drivers/net/ethernet/qlogic/qed/qed_hsi.h 	__le16 mss;
mss               127 drivers/net/ethernet/qlogic/qed/qed_iscsi.c 	u16 mss;
mss               416 drivers/net/ethernet/qlogic/qed/qed_iscsi.c 		p_tcp->mss = cpu_to_le16(p_conn->mss);
mss               453 drivers/net/ethernet/qlogic/qed/qed_iscsi.c 		p_tcp2->mss = cpu_to_le16(p_conn->mss);
mss              1316 drivers/net/ethernet/qlogic/qed/qed_iscsi.c 	con->mss = conn_info->mss;
mss               628 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 		   p_tcp_ramrod->tcp.mss,
mss               718 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	tcp->mss = cpu_to_le16(ep->mss);
mss               724 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	tcp->cwnd = QED_IWARP_DEF_CWND_FACTOR *  tcp->mss;
mss              1227 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	ep->mss = iparams->mss;
mss              1228 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	ep->mss = min_t(u16, QED_IWARP_MAX_FW_MSS, ep->mss);
mss              2430 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	ep->mss = p_hwfn->p_rdma_info->iwarp.max_mtu - hdr_size;
mss              2431 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	ep->mss = min_t(u16, QED_IWARP_MAX_FW_MSS, ep->mss);
mss               150 drivers/net/ethernet/qlogic/qed/qed_iwarp.h 	u16 mss;
mss               180 drivers/net/ethernet/qlogic/qlcnic/qlcnic.h 	__le16 mss;
mss               396 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c 		first_desc->mss = cpu_to_le16(skb_shinfo(skb)->gso_size);
mss               501 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c 		first_desc->mss = cpu_to_le16(skb_shinfo(skb)->gso_size);
mss               739 drivers/net/ethernet/realtek/8139cp.c 	int mss = 0;
mss               753 drivers/net/ethernet/realtek/8139cp.c 	mss = skb_shinfo(skb)->gso_size;
mss               755 drivers/net/ethernet/realtek/8139cp.c 	if (mss > MSSMask) {
mss               757 drivers/net/ethernet/realtek/8139cp.c 				 mss);
mss               763 drivers/net/ethernet/realtek/8139cp.c 	if (mss)
mss               764 drivers/net/ethernet/realtek/8139cp.c 		opts1 |= LargeSend | (mss << MSSShift);
mss              5773 drivers/net/ethernet/realtek/r8169_main.c 	u32 mss = skb_shinfo(skb)->gso_size;
mss              5775 drivers/net/ethernet/realtek/r8169_main.c 	if (mss) {
mss              5777 drivers/net/ethernet/realtek/r8169_main.c 		opts[0] |= min(mss, TD_MSS_MAX) << TD0_MSS_SHIFT;
mss              5794 drivers/net/ethernet/realtek/r8169_main.c 	u32 mss = skb_shinfo(skb)->gso_size;
mss              5796 drivers/net/ethernet/realtek/r8169_main.c 	if (mss) {
mss              5815 drivers/net/ethernet/realtek/r8169_main.c 		opts[1] |= min(mss, TD_MSS_MAX) << TD1_MSS_SHIFT;
mss               133 drivers/net/ethernet/samsung/sxgbe/sxgbe_desc.c static void sxgbe_tx_ctxt_desc_set_mss(struct sxgbe_tx_ctxt_desc *p, u16 mss)
mss               135 drivers/net/ethernet/samsung/sxgbe/sxgbe_desc.c 	p->maxseg_size = mss;
mss               212 drivers/net/ethernet/samsung/sxgbe/sxgbe_desc.h 	void (*tx_ctxt_desc_set_mss)(struct sxgbe_tx_ctxt_desc *p, u16 mss);
mss              2332 drivers/net/ethernet/sfc/ef10.c 	u32 mss;
mss              2336 drivers/net/ethernet/sfc/ef10.c 	mss = skb_shinfo(skb)->gso_size;
mss              2338 drivers/net/ethernet/sfc/ef10.c 	if (unlikely(mss < 4)) {
mss              2339 drivers/net/ethernet/sfc/ef10.c 		WARN_ONCE(1, "MSS of %u is too small for TSO v2\n", mss);
mss              2385 drivers/net/ethernet/sfc/ef10.c 			ESF_DZ_TX_TSO_TCP_MSS, mss
mss               419 drivers/net/ethernet/stmicro/stmmac/dwmac4_descs.c static void dwmac4_set_mss_ctxt(struct dma_desc *p, unsigned int mss)
mss               423 drivers/net/ethernet/stmicro/stmmac/dwmac4_descs.c 	p->des2 = cpu_to_le32(mss);
mss               234 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_descs.c static void dwxgmac2_set_mss(struct dma_desc *p, unsigned int mss)
mss               238 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_descs.c 	p->des2 = cpu_to_le32(mss);
mss                82 drivers/net/ethernet/stmicro/stmmac/hwif.h 	void (*set_mss)(struct dma_desc *p, unsigned int mss);
mss                56 drivers/net/ethernet/stmicro/stmmac/stmmac.h 	u32 mss;
mss              1409 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		tx_q->mss = 0;
mss              2001 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	tx_q->mss = 0;
mss              2920 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	u32 pay_len, mss;
mss              2947 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	mss = skb_shinfo(skb)->gso_size;
mss              2950 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	if (mss != tx_q->mss) {
mss              2952 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		stmmac_set_mss(priv, mss_desc, mss);
mss              2953 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		tx_q->mss = mss;
mss              2960 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 			__func__, tcp_hdrlen(skb), proto_hdr_len, pay_len, mss);
mss              4831 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		tx_q->mss = 0;
mss               527 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 	if ((tso && (pkt_info->mss != ring->tx.cur_mss)) ||
mss               727 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c 	if (tso && (pkt_info->mss != ring->tx.cur_mss))
mss               766 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c 				  pkt_info->mss);
mss               773 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c 						pkt_info->mss);
mss               789 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c 			ring->tx.cur_mss = pkt_info->mss;
mss                87 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 	pkt_info->mss = skb_shinfo(skb)->gso_size;
mss                92 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 	XLGMAC_PR("mss=%u\n", pkt_info->mss);
mss               271 drivers/net/ethernet/synopsys/dwc-xlgmac.h 	unsigned short mss;
mss              1640 drivers/net/ethernet/tehuti/tehuti.c 	txdd->mss = CPU_CHIP_SWAP16(txd_mss);
mss              1647 drivers/net/ethernet/tehuti/tehuti.c 	DBG("=== w2: mss 0x%x len 0x%x\n", txdd->mss, txdd->length);
mss               329 drivers/net/ethernet/tehuti/tehuti.h 	u16 mss;
mss              1428 drivers/net/hyperv/hyperv_net.h 			u32 mss:20;
mss              1439 drivers/net/hyperv/hyperv_net.h 			u32 mss:20;
mss               646 drivers/net/hyperv/netvsc_drv.c 		lso_info->lso_v2_transmit.mss = skb_shinfo(skb)->gso_size;
mss              1422 drivers/net/usb/ax88179_178a.c 	int mss = skb_shinfo(skb)->gso_size;
mss              1427 drivers/net/usb/ax88179_178a.c 	tx_hdr2 = mss;
mss              2751 drivers/net/usb/lan78xx.c 		u16 mss = max(skb_shinfo(skb)->gso_size, TX_CMD_B_MSS_MIN_);
mss              2753 drivers/net/usb/lan78xx.c 		tx_cmd_b = (mss << TX_CMD_B_MSS_SHIFT_) & TX_CMD_B_MSS_MASK_;
mss              1792 drivers/net/usb/r8152.c 	u32 mss = skb_shinfo(skb)->gso_size;
mss              1800 drivers/net/usb/r8152.c 	if (mss) {
mss              1828 drivers/net/usb/r8152.c 		opts2 |= min(mss, MSS_MAX) << MSS_SHIFT;
mss              2391 drivers/net/usb/r8152.c 	u32 mss = skb_shinfo(skb)->gso_size;
mss              2392 drivers/net/usb/r8152.c 	int max_offset = mss ? GTTCPHO_MAX : TCPHO_MAX;
mss              2395 drivers/net/usb/r8152.c 	if ((mss || skb->ip_summed == CHECKSUM_PARTIAL) && offset > max_offset)
mss              2271 drivers/net/usb/smsc75xx.c 		u16 mss = max(skb_shinfo(skb)->gso_size, TX_MSS_MIN);
mss              2272 drivers/net/usb/smsc75xx.c 		tx_cmd_b = (mss << TX_CMD_B_MSS_SHIFT) & TX_CMD_B_MSS;
mss               307 drivers/net/vmxnet3/vmxnet3_defs.h 	u32		mss:16;
mss               309 drivers/net/vmxnet3/vmxnet3_defs.h 	u32		mss:16;
mss               844 drivers/net/vmxnet3/vmxnet3_drv.c 	if (ctx->mss) {	/* TSO */
mss               999 drivers/net/vmxnet3/vmxnet3_drv.c 	ctx.mss = skb_shinfo(skb)->gso_size;
mss              1000 drivers/net/vmxnet3/vmxnet3_drv.c 	if (ctx.mss) {
mss              1031 drivers/net/vmxnet3/vmxnet3_drv.c 		if (ctx.mss) {
mss              1085 drivers/net/vmxnet3/vmxnet3_drv.c 	if (ctx.mss) {
mss              1088 drivers/net/vmxnet3/vmxnet3_drv.c 		gdesc->txd.msscof = ctx.mss;
mss              1089 drivers/net/vmxnet3/vmxnet3_drv.c 		num_pkts = (skb->len - gdesc->txd.hlen + ctx.mss - 1) / ctx.mss;
mss              1288 drivers/net/vmxnet3/vmxnet3_drv.c 	u16 segCnt = 0, mss = 0;
mss              1430 drivers/net/vmxnet3/vmxnet3_drv.c 				mss = rcdlro->mss;
mss              1504 drivers/net/vmxnet3/vmxnet3_drv.c 			if (segCnt != 0 && mss != 0) {
mss              1507 drivers/net/vmxnet3/vmxnet3_drv.c 				skb_shinfo(skb)->gso_size = mss;
mss               220 drivers/net/vmxnet3/vmxnet3_int.h 	u16 mss;
mss              1754 drivers/net/wireless/ath/wil6210/txrx.c 	int mss = skb_shinfo(skb)->gso_size;	/* payload size w/o headers */
mss              1761 drivers/net/wireless/ath/wil6210/txrx.c 	int rem_data = mss;
mss              1951 drivers/net/wireless/ath/wil6210/txrx.c 				rem_data = mss;
mss              1328 drivers/net/wireless/ath/wil6210/txrx_edma.c 					       int mss)
mss              1333 drivers/net/wireless/ath/wil6210/txrx_edma.c 	d->mac.tso_mss |= cpu_to_le16(mss >> 2);
mss              1359 drivers/net/wireless/ath/wil6210/txrx_edma.c 			       int mss, int *descs_used)
mss              1387 drivers/net/wireless/ath/wil6210/txrx_edma.c 					   skb_net_hdr_len, mss);
mss              1423 drivers/net/wireless/ath/wil6210/txrx_edma.c 	int mss = skb_shinfo(skb)->gso_size;
mss              1468 drivers/net/wireless/ath/wil6210/txrx_edma.c 				 mss, &descs_used);
mss              1480 drivers/net/wireless/ath/wil6210/txrx_edma.c 				 mss, &descs_used);
mss              1498 drivers/net/wireless/ath/wil6210/txrx_edma.c 					 mss, &descs_used);
mss               827 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 	unsigned int mss = skb_shinfo(skb)->gso_size;
mss               832 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 	skb_shinfo(skb)->gso_size = num_subframes * mss;
mss               836 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 	skb_shinfo(skb)->gso_size = mss;
mss               860 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 		if (tcp_payload_len > mss) {
mss               861 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 			skb_shinfo(tmp)->gso_size = mss;
mss               892 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 	unsigned int mss = skb_shinfo(skb)->gso_size;
mss               950 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 	subf_len = sizeof(struct ethhdr) + snap_ip_tcp + mss;
mss               980 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 	if (num_subframes * mss >= tcp_payload_len) {
mss               254 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	unsigned int mss = skb_shinfo(skb)->gso_size;
mss               270 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	hdr_room = DIV_ROUND_UP(total_len, mss) *
mss               300 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 		unsigned int data_left = min_t(unsigned int, mss, total_len);
mss              2109 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	unsigned int mss = skb_shinfo(skb)->gso_size;
mss              2132 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	hdr_room = DIV_ROUND_UP(total_len, mss) *
mss              2165 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			min_t(unsigned int, mss, total_len);
mss              1181 drivers/net/xen-netback/netback.c 			int mss, hdrlen;
mss              1190 drivers/net/xen-netback/netback.c 			mss = skb_shinfo(skb)->gso_size;
mss              1196 drivers/net/xen-netback/netback.c 				DIV_ROUND_UP(skb->len - hdrlen, mss);
mss               335 drivers/s390/net/qeth_core.h 	__u16 mss;
mss              4013 drivers/s390/net/qeth_core_main.c 	ext->mss = skb_shinfo(skb)->gso_size;
mss              6000 drivers/s390/net/qeth_core_main.c 	tso_data->mss = cmd->data.setassparms.data.tso.mss;
mss              6029 drivers/s390/net/qeth_core_main.c 	if (!tso_data.mss || !(tso_data.supported & QETH_IPA_LARGE_SEND_TCP)) {
mss              6058 drivers/s390/net/qeth_core_main.c 		 tso_data.mss);
mss               406 drivers/s390/net/qeth_core_mpc.h 	u32 mss;
mss               220 drivers/scsi/fnic/fcpio.h 	u16	  mss;			/* FC vNIC only: max burst */
mss               268 drivers/scsi/fnic/fcpio.h 	u16   mss;                    /* FC vNIC only: max burst */
mss                88 drivers/scsi/fnic/fnic_res.h 						    u32 d_id, u16 mss,
mss               117 drivers/scsi/fnic/fnic_res.h 	desc->u.icmnd_16.mss = mss;            	/* FC vNIC only: max burst */
mss                52 drivers/scsi/fnic/wq_enet_desc.h 	u64 address, u16 length, u16 mss, u16 header_length,
mss                58 drivers/scsi/fnic/wq_enet_desc.h 	desc->mss_loopback = cpu_to_le16((mss & WQ_ENET_MSS_MASK) <<
mss                71 drivers/scsi/fnic/wq_enet_desc.h 	u64 *address, u16 *length, u16 *mss, u16 *header_length,
mss                77 drivers/scsi/fnic/wq_enet_desc.h 	*mss = (le16_to_cpu(desc->mss_loopback) >> WQ_ENET_MSS_SHIFT) &
mss               462 drivers/scsi/qedi/qedi_iscsi.c 	u16 mss = 0;
mss               470 drivers/scsi/qedi/qedi_iscsi.c 	mss = pmtu - hdrs;
mss               472 drivers/scsi/qedi/qedi_iscsi.c 	if (!mss)
mss               473 drivers/scsi/qedi/qedi_iscsi.c 		mss = DEF_MSS;
mss               475 drivers/scsi/qedi/qedi_iscsi.c 	return mss;
mss               545 drivers/scsi/qedi/qedi_iscsi.c 	conn_info->mss = qedi_calc_mss(qedi_ep->pmtu,
mss               549 drivers/scsi/qedi/qedi_iscsi.c 	conn_info->cwnd = DEF_MAX_CWND * conn_info->mss;
mss               556 drivers/scsi/qedi/qedi_iscsi.c 		  conn_info->default_cq, conn_info->mss);
mss               891 drivers/scsi/qla4xxx/ql4_fw.h 	uint16_t mss;	/* 80-81 */
mss              3544 drivers/scsi/qla4xxx/ql4_os.c 	conn->max_segment_size = le16_to_cpu(fw_ddb_entry->mss);
mss              3672 drivers/scsi/qla4xxx/ql4_os.c 	fw_ddb_entry->mss = cpu_to_le16(conn->max_segment_size);
mss              3781 drivers/scsi/qla4xxx/ql4_os.c 	conn->max_segment_size = le16_to_cpu(fw_ddb_entry->mss);
mss                52 drivers/scsi/snic/wq_enet_desc.h 	u64 address, u16 length, u16 mss, u16 header_length,
mss                58 drivers/scsi/snic/wq_enet_desc.h 	desc->mss_loopback = cpu_to_le16((mss & WQ_ENET_MSS_MASK) <<
mss                71 drivers/scsi/snic/wq_enet_desc.h 	u64 *address, u16 *length, u16 *mss, u16 *header_length,
mss                77 drivers/scsi/snic/wq_enet_desc.h 	*mss = (le16_to_cpu(desc->mss_loopback) >> WQ_ENET_MSS_SHIFT) &
mss              1141 drivers/staging/qlge/qlge.h 	__le16 mss;
mss              1895 drivers/staging/qlge/qlge_dbg.c 		       le16_to_cpu(ob_mac_tso_iocb->mss));
mss              2581 drivers/staging/qlge/qlge_main.c 		mac_iocb_ptr->mss = cpu_to_le16(skb_shinfo(skb)->gso_size);
mss               228 drivers/target/iscsi/cxgbit/cxgbit.h 	u16 mss;
mss               765 drivers/target/iscsi/cxgbit/cxgbit_cm.c 	csk->mss = csk->emss;
mss               772 drivers/target/iscsi/cxgbit/cxgbit_cm.c 			TCPOPT_MSS_G(opt), csk->mss, csk->emss);
mss               774 drivers/target/iscsi/cxgbit/cxgbit_cm.c 		 csk->mss, csk->emss);
mss              1224 drivers/target/iscsi/cxgbit/cxgbit_cm.c 	u16 peer_mss = ntohs(req->tcpopt.mss);
mss               433 fs/proc/task_mmu.c static void smaps_page_accumulate(struct mem_size_stats *mss,
mss               437 fs/proc/task_mmu.c 	mss->pss += pss;
mss               440 fs/proc/task_mmu.c 		mss->pss_anon += pss;
mss               442 fs/proc/task_mmu.c 		mss->pss_shmem += pss;
mss               444 fs/proc/task_mmu.c 		mss->pss_file += pss;
mss               447 fs/proc/task_mmu.c 		mss->pss_locked += pss;
mss               451 fs/proc/task_mmu.c 			mss->private_dirty += size;
mss               453 fs/proc/task_mmu.c 			mss->shared_dirty += size;
mss               456 fs/proc/task_mmu.c 			mss->private_clean += size;
mss               458 fs/proc/task_mmu.c 			mss->shared_clean += size;
mss               462 fs/proc/task_mmu.c static void smaps_account(struct mem_size_stats *mss, struct page *page,
mss               473 fs/proc/task_mmu.c 		mss->anonymous += size;
mss               475 fs/proc/task_mmu.c 			mss->lazyfree += size;
mss               478 fs/proc/task_mmu.c 	mss->resident += size;
mss               481 fs/proc/task_mmu.c 		mss->referenced += size;
mss               492 fs/proc/task_mmu.c 		smaps_page_accumulate(mss, page, size, size << PSS_SHIFT, dirty,
mss               501 fs/proc/task_mmu.c 		smaps_page_accumulate(mss, page, PAGE_SIZE, pss, dirty, locked,
mss               510 fs/proc/task_mmu.c 	struct mem_size_stats *mss = walk->private;
mss               512 fs/proc/task_mmu.c 	mss->swap += shmem_partial_swap_usage(
mss               524 fs/proc/task_mmu.c 	struct mem_size_stats *mss = walk->private;
mss               537 fs/proc/task_mmu.c 			mss->swap += PAGE_SIZE;
mss               543 fs/proc/task_mmu.c 				mss->swap_pss += pss_delta;
mss               545 fs/proc/task_mmu.c 				mss->swap_pss += (u64)PAGE_SIZE << PSS_SHIFT;
mss               551 fs/proc/task_mmu.c 	} else if (unlikely(IS_ENABLED(CONFIG_SHMEM) && mss->check_shmem_swap
mss               559 fs/proc/task_mmu.c 			mss->swap += PAGE_SIZE;
mss               569 fs/proc/task_mmu.c 	smaps_account(mss, page, false, pte_young(*pte), pte_dirty(*pte), locked);
mss               576 fs/proc/task_mmu.c 	struct mem_size_stats *mss = walk->private;
mss               586 fs/proc/task_mmu.c 		mss->anonymous_thp += HPAGE_PMD_SIZE;
mss               588 fs/proc/task_mmu.c 		mss->shmem_thp += HPAGE_PMD_SIZE;
mss               592 fs/proc/task_mmu.c 		mss->file_thp += HPAGE_PMD_SIZE;
mss               593 fs/proc/task_mmu.c 	smaps_account(mss, page, true, pmd_young(*pmd), pmd_dirty(*pmd), locked);
mss               711 fs/proc/task_mmu.c 	struct mem_size_stats *mss = walk->private;
mss               729 fs/proc/task_mmu.c 			mss->shared_hugetlb += huge_page_size(hstate_vma(vma));
mss               731 fs/proc/task_mmu.c 			mss->private_hugetlb += huge_page_size(hstate_vma(vma));
mss               751 fs/proc/task_mmu.c 			     struct mem_size_stats *mss)
mss               755 fs/proc/task_mmu.c 	mss->check_shmem_swap = false;
mss               771 fs/proc/task_mmu.c 			mss->swap += shmem_swapped;
mss               773 fs/proc/task_mmu.c 			mss->check_shmem_swap = true;
mss               774 fs/proc/task_mmu.c 			walk_page_vma(vma, &smaps_shmem_walk_ops, mss);
mss               780 fs/proc/task_mmu.c 	walk_page_vma(vma, &smaps_walk_ops, mss);
mss               787 fs/proc/task_mmu.c static void __show_smap(struct seq_file *m, const struct mem_size_stats *mss,
mss               790 fs/proc/task_mmu.c 	SEQ_PUT_DEC("Rss:            ", mss->resident);
mss               791 fs/proc/task_mmu.c 	SEQ_PUT_DEC(" kB\nPss:            ", mss->pss >> PSS_SHIFT);
mss               798 fs/proc/task_mmu.c 			mss->pss_anon >> PSS_SHIFT);
mss               800 fs/proc/task_mmu.c 			mss->pss_file >> PSS_SHIFT);
mss               802 fs/proc/task_mmu.c 			mss->pss_shmem >> PSS_SHIFT);
mss               804 fs/proc/task_mmu.c 	SEQ_PUT_DEC(" kB\nShared_Clean:   ", mss->shared_clean);
mss               805 fs/proc/task_mmu.c 	SEQ_PUT_DEC(" kB\nShared_Dirty:   ", mss->shared_dirty);
mss               806 fs/proc/task_mmu.c 	SEQ_PUT_DEC(" kB\nPrivate_Clean:  ", mss->private_clean);
mss               807 fs/proc/task_mmu.c 	SEQ_PUT_DEC(" kB\nPrivate_Dirty:  ", mss->private_dirty);
mss               808 fs/proc/task_mmu.c 	SEQ_PUT_DEC(" kB\nReferenced:     ", mss->referenced);
mss               809 fs/proc/task_mmu.c 	SEQ_PUT_DEC(" kB\nAnonymous:      ", mss->anonymous);
mss               810 fs/proc/task_mmu.c 	SEQ_PUT_DEC(" kB\nLazyFree:       ", mss->lazyfree);
mss               811 fs/proc/task_mmu.c 	SEQ_PUT_DEC(" kB\nAnonHugePages:  ", mss->anonymous_thp);
mss               812 fs/proc/task_mmu.c 	SEQ_PUT_DEC(" kB\nShmemPmdMapped: ", mss->shmem_thp);
mss               813 fs/proc/task_mmu.c 	SEQ_PUT_DEC(" kB\nFilePmdMapped: ", mss->file_thp);
mss               814 fs/proc/task_mmu.c 	SEQ_PUT_DEC(" kB\nShared_Hugetlb: ", mss->shared_hugetlb);
mss               816 fs/proc/task_mmu.c 				  mss->private_hugetlb >> 10, 7);
mss               817 fs/proc/task_mmu.c 	SEQ_PUT_DEC(" kB\nSwap:           ", mss->swap);
mss               819 fs/proc/task_mmu.c 					mss->swap_pss >> PSS_SHIFT);
mss               821 fs/proc/task_mmu.c 					mss->pss_locked >> PSS_SHIFT);
mss               828 fs/proc/task_mmu.c 	struct mem_size_stats mss;
mss               830 fs/proc/task_mmu.c 	memset(&mss, 0, sizeof(mss));
mss               832 fs/proc/task_mmu.c 	smap_gather_stats(vma, &mss);
mss               841 fs/proc/task_mmu.c 	__show_smap(m, &mss, false);
mss               858 fs/proc/task_mmu.c 	struct mem_size_stats mss;
mss               874 fs/proc/task_mmu.c 	memset(&mss, 0, sizeof(mss));
mss               883 fs/proc/task_mmu.c 		smap_gather_stats(vma, &mss);
mss               892 fs/proc/task_mmu.c 	__show_smap(m, &mss, true);
mss               249 include/linux/mlx5/qp.h 	__be16          mss;
mss               684 include/linux/qed/common_hsi.h 	__le16 mss;
mss               123 include/linux/qed/qed_iscsi_if.h 	u16 mss;
mss               515 include/linux/qed/qed_rdma_if.h 	u16 mss;
mss               112 include/linux/qed/tcp_common.h 	__le16 mss;
mss               178 include/linux/qed/tcp_common.h 	__le16 mss;
mss               243 include/linux/qed/tcp_common.h 	__le16 mss;
mss               477 include/linux/tcp.h static inline u16 tcp_mss_clamp(const struct tcp_sock *tp, u16 mss)
mss               484 include/linux/tcp.h 	return (user_mss && user_mss < mss) ? user_mss : mss;
mss                55 include/net/request_sock.h 	u16				mss;
mss               550 include/net/tcp.h __u32 cookie_v4_init_sequence(const struct sk_buff *skb, __u16 *mss);
mss               564 include/net/tcp.h __u32 cookie_v6_init_sequence(const struct sk_buff *skb, __u16 *mss);
mss               656 include/net/tcp.h int tcp_mss_to_mtu(struct sock *sk, int mss);
mss              1359 include/net/tcp.h u32 tcp_default_init_rwnd(u32 mss);
mss              1378 include/net/tcp.h 			       __u32 mss, __u32 *rcv_wnd,
mss              1635 include/net/tcp.h void tcp_fastopen_cache_get(struct sock *sk, u16 *mss,
mss              1637 include/net/tcp.h void tcp_fastopen_cache_set(struct sock *sk, u16 mss,
mss              1659 include/net/tcp.h bool tcp_fastopen_cookie_check(struct sock *sk, u16 *mss,
mss              1991 include/net/tcp.h 				 __u16 *mss);
mss              2006 include/net/tcp.h 					 __u16 *mss)
mss              2010 include/net/tcp.h 	return ops->cookie_init_seq(skb, mss);
mss              2015 include/net/tcp.h 					 __u16 *mss)
mss              1371 include/rdma/ib_verbs.h 	int			mss;
mss                20 include/uapi/linux/netfilter/nf_synproxy.h 	__u16	mss;
mss                58 include/uapi/linux/netfilter/nfnetlink_osf.h 	__u16	ss, mss;
mss                 8 include/uapi/linux/netfilter/xt_TCPMSS.h 	__u16 mss;
mss               183 ipc/msg.c      			  struct msg_sender *mss, size_t msgsz)
mss               185 ipc/msg.c      	mss->tsk = current;
mss               186 ipc/msg.c      	mss->msgsz = msgsz;
mss               188 ipc/msg.c      	list_add_tail(&mss->list, &msq->q_senders);
mss               191 ipc/msg.c      static inline void ss_del(struct msg_sender *mss)
mss               193 ipc/msg.c      	if (mss->list.next)
mss               194 ipc/msg.c      		list_del(&mss->list);
mss               200 ipc/msg.c      	struct msg_sender *mss, *t;
mss               204 ipc/msg.c      	list_for_each_entry_safe(mss, t, h, list) {
mss               206 ipc/msg.c      			mss->list.next = NULL;
mss               213 ipc/msg.c      		else if (stop_tsk == mss->tsk)
mss               222 ipc/msg.c      		else if (!msg_fits_inqueue(msq, mss->msgsz)) {
mss               224 ipc/msg.c      				stop_tsk = mss->tsk;
mss               226 ipc/msg.c      			list_move_tail(&mss->list, &msq->q_senders);
mss               230 ipc/msg.c      		wake_q_add(wake_q, mss->tsk);
mss               142 net/batman-adv/tp_meter.c static void batadv_tp_update_cwnd(struct batadv_tp_vars *tp_vars, u32 mss)
mss               149 net/batman-adv/tp_meter.c 		tp_vars->cwnd = batadv_tp_cwnd(tp_vars->cwnd, mss, mss);
mss               156 net/batman-adv/tp_meter.c 				   ((mss * mss) << 6) / (tp_vars->cwnd << 3));
mss               157 net/batman-adv/tp_meter.c 	if (tp_vars->dec_cwnd < (mss << 3)) {
mss               162 net/batman-adv/tp_meter.c 	tp_vars->cwnd = batadv_tp_cwnd(tp_vars->cwnd, mss, mss);
mss               630 net/batman-adv/tp_meter.c 	size_t packet_len, mss;
mss               635 net/batman-adv/tp_meter.c 	mss = BATADV_TP_PLEN;
mss               700 net/batman-adv/tp_meter.c 		tp_vars->cwnd = batadv_tp_cwnd(tp_vars->ss_threshold, 3 * mss,
mss               701 net/batman-adv/tp_meter.c 					       mss);
mss               727 net/batman-adv/tp_meter.c 							       mss, mss);
mss               735 net/batman-adv/tp_meter.c 						      mss);
mss               741 net/batman-adv/tp_meter.c 		if (recv_ack - atomic_read(&tp_vars->last_acked) >= mss)
mss               742 net/batman-adv/tp_meter.c 			batadv_tp_update_cwnd(tp_vars, mss);
mss              5866 net/core/filter.c 	u16 mss;
mss              5891 net/core/filter.c 		mss = tcp_v4_get_syncookie(sk, iph, th, &cookie);
mss              5902 net/core/filter.c 		mss = tcp_v6_get_syncookie(sk, iph, th, &cookie);
mss              5909 net/core/filter.c 	if (mss == 0)
mss              5912 net/core/filter.c 	return cookie | ((u64)mss << 32);
mss              3657 net/core/skbuff.c 	unsigned int mss = skb_shinfo(head_skb)->gso_size;
mss              3688 net/core/skbuff.c 		if (mss != GSO_BY_FRAGS && mss != skb_headlen(head_skb))
mss              3700 net/core/skbuff.c 	if (sg && csum && (mss != GSO_BY_FRAGS))  {
mss              3735 net/core/skbuff.c 		partial_segs = len / mss;
mss              3737 net/core/skbuff.c 			mss *= partial_segs;
mss              3752 net/core/skbuff.c 		if (unlikely(mss == GSO_BY_FRAGS)) {
mss              3756 net/core/skbuff.c 			if (len > mss)
mss              3757 net/core/skbuff.c 				len = mss;
mss              3889 net/core/skbuff.c 					pos, mss);
mss              1848 net/decnet/af_decnet.c 	unsigned int mss = 230 - DN_MAX_NSP_DATA_HEADER;
mss              1863 net/decnet/af_decnet.c 	if (mtu > mss)
mss              1864 net/decnet/af_decnet.c 		mss = mtu;
mss              1865 net/decnet/af_decnet.c 	return mss;
mss              1910 net/decnet/af_decnet.c 	size_t mss;
mss              1959 net/decnet/af_decnet.c 	mss = scp->segsize_rem;
mss              1962 net/decnet/af_decnet.c 	mss = dn_current_mss(sk, flags);
mss              1966 net/decnet/af_decnet.c 		if (size > mss) {
mss              1989 net/decnet/af_decnet.c 		if (len > mss)
mss              1990 net/decnet/af_decnet.c 			len = mss;
mss               276 net/decnet/dn_route.c 			u32 mss = mtu - DN_MAX_NSP_DATA_HEADER;
mss               278 net/decnet/dn_route.c 			if (!existing_mss || existing_mss > mss)
mss               279 net/decnet/dn_route.c 				dst_metric_set(dst, RTAX_ADVMSS, mss);
mss               888 net/decnet/dn_route.c 		unsigned int mss = dn_mss_from_pmtu(dev, dst_mtu(&rt->dst));
mss               889 net/decnet/dn_route.c 		if (mss_metric > mss)
mss               890 net/decnet/dn_route.c 			dst_metric_set(&rt->dst, RTAX_ADVMSS, mss);
mss                40 net/ipv4/netfilter/ipt_SYNPROXY.c 		opts.mss_option = info->mss;
mss               165 net/ipv4/syncookies.c 	const __u16 mss = *mssp;
mss               168 net/ipv4/syncookies.c 		if (mss >= msstab[mssind])
mss               294 net/ipv4/syncookies.c 	int mss;
mss               306 net/ipv4/syncookies.c 	mss = __cookie_v4_check(ip_hdr(skb), th, cookie);
mss               307 net/ipv4/syncookies.c 	if (mss == 0) {
mss               339 net/ipv4/syncookies.c 	req->mss		= mss;
mss               390 net/ipv4/syncookies.c 	tcp_select_initial_window(sk, tcp_full_space(sk), req->mss,
mss               244 net/ipv4/tcp_bbr.c 	unsigned int mss = tcp_sk(sk)->mss_cache;
mss               246 net/ipv4/tcp_bbr.c 	rate *= mss;
mss               406 net/ipv4/tcp_fastopen.c bool tcp_fastopen_cookie_check(struct sock *sk, u16 *mss,
mss               411 net/ipv4/tcp_fastopen.c 	tcp_fastopen_cache_get(sk, mss, cookie);
mss               439 net/ipv4/tcp_fastopen.c 	u16 mss;
mss               442 net/ipv4/tcp_fastopen.c 		if (tcp_fastopen_cookie_check(sk, &mss, &cookie)) {
mss               888 net/ipv4/tcp_input.c 	const u32 mss = tp->mss_cache;
mss               896 net/ipv4/tcp_input.c 	if ((metric > tp->reordering * mss) && mss) {
mss               905 net/ipv4/tcp_input.c 		tp->reordering = min_t(u32, (metric + mss - 1) / mss,
mss              1155 net/ipv4/tcp_input.c 	unsigned int mss;
mss              1162 net/ipv4/tcp_input.c 		mss = tcp_skb_mss(skb);
mss              1167 net/ipv4/tcp_input.c 			if (pkt_len < mss)
mss              1168 net/ipv4/tcp_input.c 				pkt_len = mss;
mss              1171 net/ipv4/tcp_input.c 			if (pkt_len < mss)
mss              1178 net/ipv4/tcp_input.c 		if (pkt_len > mss) {
mss              1179 net/ipv4/tcp_input.c 			unsigned int new_len = (pkt_len / mss) * mss;
mss              1181 net/ipv4/tcp_input.c 				new_len += mss;
mss              1189 net/ipv4/tcp_input.c 				   pkt_len, mss, GFP_ATOMIC);
mss              1285 net/ipv4/tcp_input.c 			    unsigned int pcount, int shifted, int mss,
mss              1321 net/ipv4/tcp_input.c 		TCP_SKB_CB(prev)->tcp_gso_size = mss;
mss              1403 net/ipv4/tcp_input.c 	int mss;
mss              1435 net/ipv4/tcp_input.c 		mss = tcp_skb_seglen(skb);
mss              1440 net/ipv4/tcp_input.c 		if (mss != tcp_skb_seglen(prev))
mss              1476 net/ipv4/tcp_input.c 		mss = tcp_skb_mss(skb);
mss              1481 net/ipv4/tcp_input.c 		if (mss != tcp_skb_seglen(prev))
mss              1484 net/ipv4/tcp_input.c 		if (len == mss) {
mss              1486 net/ipv4/tcp_input.c 		} else if (len < mss) {
mss              1489 net/ipv4/tcp_input.c 			pcount = len / mss;
mss              1490 net/ipv4/tcp_input.c 			len = pcount * mss;
mss              1500 net/ipv4/tcp_input.c 	if (!tcp_shifted_skb(sk, prev, skb, state, pcount, len, mss, dup_sack))
mss              1512 net/ipv4/tcp_input.c 	    (mss != tcp_skb_seglen(skb)))
mss              1519 net/ipv4/tcp_input.c 				len, mss, 0);
mss              2195 net/ipv4/tcp_input.c 	unsigned int mss;
mss              2231 net/ipv4/tcp_input.c 			mss = tcp_skb_mss(skb);
mss              2233 net/ipv4/tcp_input.c 			lost = (packets - oldcnt) * mss;
mss              2236 net/ipv4/tcp_input.c 					 lost, mss, GFP_ATOMIC) < 0)
mss              2604 net/ipv4/tcp_input.c 	unsigned int mss = tcp_current_mss(sk);
mss              2607 net/ipv4/tcp_input.c 		if (tcp_skb_seglen(skb) > mss &&
mss              3799 net/ipv4/tcp_input.c 	u16 mss = 0;
mss              3807 net/ipv4/tcp_input.c 			return mss;
mss              3813 net/ipv4/tcp_input.c 				return mss;
mss              3816 net/ipv4/tcp_input.c 				return mss;
mss              3818 net/ipv4/tcp_input.c 				return mss;	/* fail on partial options */
mss              3825 net/ipv4/tcp_input.c 					mss = in_mss;
mss              3832 net/ipv4/tcp_input.c 	return mss;
mss              5789 net/ipv4/tcp_input.c 	u16 mss = tp->rx_opt.mss_clamp, try_exp = 0;
mss              5792 net/ipv4/tcp_input.c 	if (mss == tp->rx_opt.user_mss) {
mss              5799 net/ipv4/tcp_input.c 		mss = opt.mss_clamp;
mss              5820 net/ipv4/tcp_input.c 	tcp_fastopen_cache_set(sk, mss, cookie, syn_drop, try_exp);
mss              6443 net/ipv4/tcp_input.c 	req->mss = rx_opt->mss_clamp;
mss              6536 net/ipv4/tcp_input.c 	u16 mss;
mss              6550 net/ipv4/tcp_input.c 	mss = tcp_parse_mss_option(th, tp->rx_opt.user_mss);
mss              6551 net/ipv4/tcp_input.c 	if (!mss)
mss              6552 net/ipv4/tcp_input.c 		mss = af_ops->mss_clamp;
mss              6554 net/ipv4/tcp_input.c 	return mss;
mss              6651 net/ipv4/tcp_input.c 		isn = cookie_init_sequence(af_ops, sk, skb, &req->mss);
mss              1528 net/ipv4/tcp_ipv4.c 	u16 mss = 0;
mss              1530 net/ipv4/tcp_ipv4.c 	mss = tcp_get_syncookie_mss(&tcp_request_sock_ops,
mss              1532 net/ipv4/tcp_ipv4.c 	if (mss) {
mss              1533 net/ipv4/tcp_ipv4.c 		*cookie = __cookie_v4_init_sequence(iph, th, &mss);
mss              1537 net/ipv4/tcp_ipv4.c 	return mss;
mss                29 net/ipv4/tcp_metrics.c 	u16	mss;
mss               124 net/ipv4/tcp_metrics.c 		tm->tcpm_fastopen.mss = 0;
mss               538 net/ipv4/tcp_metrics.c void tcp_fastopen_cache_get(struct sock *sk, u16 *mss,
mss               551 net/ipv4/tcp_metrics.c 			if (tfom->mss)
mss               552 net/ipv4/tcp_metrics.c 				*mss = tfom->mss;
mss               561 net/ipv4/tcp_metrics.c void tcp_fastopen_cache_set(struct sock *sk, u16 mss,
mss               576 net/ipv4/tcp_metrics.c 		if (mss)
mss               577 net/ipv4/tcp_metrics.c 			tfom->mss = mss;
mss               693 net/ipv4/tcp_metrics.c 		if (tfom->mss &&
mss               695 net/ipv4/tcp_metrics.c 				tfom->mss) < 0)
mss               370 net/ipv4/tcp_minisocks.c 	int mss;
mss               372 net/ipv4/tcp_minisocks.c 	mss = tcp_mss_clamp(tp, dst_metric_advmss(dst));
mss               385 net/ipv4/tcp_minisocks.c 	else if (full_space < rcv_wnd * mss)
mss               386 net/ipv4/tcp_minisocks.c 		full_space = rcv_wnd * mss;
mss               390 net/ipv4/tcp_minisocks.c 		mss - (ireq->tstamp_ok ? TCPOLEN_TSTAMP_ALIGNED : 0),
mss               546 net/ipv4/tcp_minisocks.c 	newtp->rx_opt.mss_clamp = req->mss;
mss                15 net/ipv4/tcp_offload.c 			   unsigned int seq, unsigned int mss)
mss                18 net/ipv4/tcp_offload.c 		if (before(ts_seq, seq + mss)) {
mss                25 net/ipv4/tcp_offload.c 		seq += mss;
mss                64 net/ipv4/tcp_offload.c 	unsigned int mss;
mss                80 net/ipv4/tcp_offload.c 	mss = skb_shinfo(skb)->gso_size;
mss                81 net/ipv4/tcp_offload.c 	if (unlikely(skb->len <= mss))
mss                87 net/ipv4/tcp_offload.c 		skb_shinfo(skb)->gso_segs = DIV_ROUND_UP(skb->len, mss);
mss               110 net/ipv4/tcp_offload.c 		mss *= skb_shinfo(segs)->gso_segs;
mss               112 net/ipv4/tcp_offload.c 	delta = htonl(oldlen + (thlen + mss));
mss               119 net/ipv4/tcp_offload.c 		tcp_gso_tstamp(segs, skb_shinfo(gso_skb)->tskey, seq, mss);
mss               133 net/ipv4/tcp_offload.c 		seq += mss;
mss               189 net/ipv4/tcp_offload.c 	unsigned int mss = 1;
mss               258 net/ipv4/tcp_offload.c 	mss = skb_shinfo(p)->gso_size;
mss               260 net/ipv4/tcp_offload.c 	flush |= (len - 1) >= mss;
mss               267 net/ipv4/tcp_offload.c 		mss = 1;
mss               274 net/ipv4/tcp_offload.c 	flush = len < mss;
mss               123 net/ipv4/tcp_output.c 	int mss = tp->advmss;
mss               128 net/ipv4/tcp_output.c 		if (metric < mss) {
mss               129 net/ipv4/tcp_output.c 			mss = metric;
mss               130 net/ipv4/tcp_output.c 			tp->advmss = mss;
mss               134 net/ipv4/tcp_output.c 	return (__u16)mss;
mss               207 net/ipv4/tcp_output.c void tcp_select_initial_window(const struct sock *sk, int __space, __u32 mss,
mss               220 net/ipv4/tcp_output.c 	if (space > mss)
mss               221 net/ipv4/tcp_output.c 		space = rounddown(space, mss);
mss               237 net/ipv4/tcp_output.c 		*rcv_wnd = min(*rcv_wnd, init_rcv_wnd * mss);
mss               435 net/ipv4/tcp_output.c 	u16 mss;		/* 0 to disable */
mss               470 net/ipv4/tcp_output.c 	if (unlikely(opts->mss)) {
mss               473 net/ipv4/tcp_output.c 			       opts->mss);
mss               619 net/ipv4/tcp_output.c 	opts->mss = tcp_advertise_mss(sk);
mss               662 net/ipv4/tcp_output.c 				       unsigned int mss, struct sk_buff *skb,
mss               685 net/ipv4/tcp_output.c 	opts->mss = mss;
mss              1507 net/ipv4/tcp_output.c int tcp_mss_to_mtu(struct sock *sk, int mss)
mss              1513 net/ipv4/tcp_output.c 	mtu = mss +
mss              2052 net/ipv4/tcp_output.c 		int mss = tcp_current_mss(sk);
mss              2059 net/ipv4/tcp_output.c 		icsk->icsk_mtup.search_low = tcp_mss_to_mtu(sk, mss);
mss              2563 net/ipv4/tcp_output.c 	int mss = tcp_current_mss(sk);
mss              2566 net/ipv4/tcp_output.c 	if (skb && tcp_snd_wnd_test(tp, skb, mss)) {
mss              2568 net/ipv4/tcp_output.c 		tcp_write_xmit(sk, mss, TCP_NAGLE_OFF, 2, GFP_ATOMIC);
mss              2577 net/ipv4/tcp_output.c 			  tp->packets_out, sk->sk_state, tp->snd_cwnd, mss);
mss              2593 net/ipv4/tcp_output.c 	if ((pcount > 1) && (skb->len > (pcount - 1) * mss)) {
mss              2595 net/ipv4/tcp_output.c 					  (pcount - 1) * mss, mss,
mss              2711 net/ipv4/tcp_output.c 	int mss = icsk->icsk_ack.rcv_mss;
mss              2717 net/ipv4/tcp_output.c 	if (unlikely(mss > full_space)) {
mss              2718 net/ipv4/tcp_output.c 		mss = full_space;
mss              2719 net/ipv4/tcp_output.c 		if (mss <= 0)
mss              2741 net/ipv4/tcp_output.c 		if (free_space < (allowed_space >> 4) || free_space < mss)
mss              2769 net/ipv4/tcp_output.c 		if (window <= free_space - mss || window > free_space)
mss              2770 net/ipv4/tcp_output.c 			window = rounddown(free_space, mss);
mss              2771 net/ipv4/tcp_output.c 		else if (mss == full_space &&
mss              3278 net/ipv4/tcp_output.c 	int mss;
mss              3308 net/ipv4/tcp_output.c 	mss = tcp_mss_clamp(tp, dst_metric_advmss(dst));
mss              3328 net/ipv4/tcp_output.c 	tcp_header_size = tcp_synack_options(sk, req, mss, skb, &opts, md5,
mss              3781 net/ipv4/tcp_output.c 		unsigned int mss = tcp_current_mss(sk);
mss              3792 net/ipv4/tcp_output.c 		    skb->len > mss) {
mss              3793 net/ipv4/tcp_output.c 			seg_size = min(seg_size, mss);
mss              3796 net/ipv4/tcp_output.c 					 skb, seg_size, mss, GFP_ATOMIC))
mss              3799 net/ipv4/tcp_output.c 			tcp_set_skb_tso_segs(skb, mss);
mss               239 net/ipv4/tcp_recovery.c 		u32 mss;
mss               244 net/ipv4/tcp_recovery.c 		mss = tcp_skb_mss(skb);
mss               245 net/ipv4/tcp_recovery.c 		if (tcp_skb_pcount(skb) > 1 && skb->len > mss)
mss               247 net/ipv4/tcp_recovery.c 				     mss, mss, GFP_ATOMIC);
mss               145 net/ipv4/tcp_timer.c 	int mss;
mss               155 net/ipv4/tcp_timer.c 		mss = tcp_mtu_to_mss(sk, icsk->icsk_mtup.search_low) >> 1;
mss               156 net/ipv4/tcp_timer.c 		mss = min(net->ipv4.sysctl_tcp_base_mss, mss);
mss               157 net/ipv4/tcp_timer.c 		mss = max(mss, net->ipv4.sysctl_tcp_mtu_probe_floor);
mss               158 net/ipv4/tcp_timer.c 		mss = max(mss, net->ipv4.sysctl_tcp_min_snd_mss);
mss               159 net/ipv4/tcp_timer.c 		icsk->icsk_mtup.search_low = tcp_mss_to_mtu(sk, mss);
mss               194 net/ipv4/udp_offload.c 	unsigned int mss;
mss               199 net/ipv4/udp_offload.c 	mss = skb_shinfo(gso_skb)->gso_size;
mss               200 net/ipv4/udp_offload.c 	if (gso_skb->len <= sizeof(*uh) + mss)
mss               222 net/ipv4/udp_offload.c 		mss *= skb_shinfo(segs)->gso_segs;
mss               233 net/ipv4/udp_offload.c 	newlen = htons(sizeof(*uh) + mss);
mss               292 net/ipv4/udp_offload.c 	unsigned int mss;
mss               313 net/ipv4/udp_offload.c 	mss = skb_shinfo(skb)->gso_size;
mss               314 net/ipv4/udp_offload.c 	if (unlikely(skb->len <= mss))
mss                40 net/ipv6/netfilter/ip6t_SYNPROXY.c 		opts.mss_option = info->mss;
mss                96 net/ipv6/syncookies.c 	const __u16 mss = *mssp;
mss                99 net/ipv6/syncookies.c 		if (mss >= msstab[mssind])
mss               139 net/ipv6/syncookies.c 	int mss;
mss               150 net/ipv6/syncookies.c 	mss = __cookie_v6_check(ipv6_hdr(skb), th, cookie);
mss               151 net/ipv6/syncookies.c 	if (mss == 0) {
mss               184 net/ipv6/syncookies.c 	req->mss = mss;
mss               244 net/ipv6/syncookies.c 	tcp_select_initial_window(sk, tcp_full_space(sk), req->mss,
mss              1076 net/ipv6/tcp_ipv6.c 	u16 mss = 0;
mss              1078 net/ipv6/tcp_ipv6.c 	mss = tcp_get_syncookie_mss(&tcp6_request_sock_ops,
mss              1080 net/ipv6/tcp_ipv6.c 	if (mss) {
mss              1081 net/ipv6/tcp_ipv6.c 		*cookie = __cookie_v6_init_sequence(iph, th, &mss);
mss              1085 net/ipv6/tcp_ipv6.c 	return mss;
mss                21 net/ipv6/udp_offload.c 	unsigned int mss;
mss                31 net/ipv6/udp_offload.c 	mss = skb_shinfo(skb)->gso_size;
mss                32 net/ipv6/udp_offload.c 	if (unlikely(skb->len <= mss))
mss               473 net/netfilter/nf_synproxy_core.c 	u16 mss = opts->mss_encode;
mss               490 net/netfilter/nf_synproxy_core.c 	nth->seq	= htonl(__cookie_v4_init_sequence(iph, th, &mss));
mss               636 net/netfilter/nf_synproxy_core.c 	int mss;
mss               638 net/netfilter/nf_synproxy_core.c 	mss = __cookie_v4_check(ip_hdr(skb), th, ntohl(th->ack_seq) - 1);
mss               639 net/netfilter/nf_synproxy_core.c 	if (mss == 0) {
mss               645 net/netfilter/nf_synproxy_core.c 	opts->mss_option = mss;
mss               887 net/netfilter/nf_synproxy_core.c 	u16 mss = opts->mss_encode;
mss               904 net/netfilter/nf_synproxy_core.c 	nth->seq	= htonl(nf_ipv6_cookie_init_sequence(iph, th, &mss));
mss              1054 net/netfilter/nf_synproxy_core.c 	int mss;
mss              1056 net/netfilter/nf_synproxy_core.c 	mss = nf_cookie_v6_check(ipv6_hdr(skb), th, ntohl(th->ack_seq) - 1);
mss              1057 net/netfilter/nf_synproxy_core.c 	if (mss == 0) {
mss              1063 net/netfilter/nf_synproxy_core.c 	opts->mss_option = mss;
mss                73 net/netfilter/nfnetlink_osf.c 	u16 mss = 0;
mss               106 net/netfilter/nfnetlink_osf.c 				mss = ctx->optp[3];
mss               107 net/netfilter/nfnetlink_osf.c 				mss <<= 8;
mss               108 net/netfilter/nfnetlink_osf.c 				mss |= ctx->optp[2];
mss               110 net/netfilter/nfnetlink_osf.c 				mss = ntohs((__force __be16)mss);
mss               141 net/netfilter/nfnetlink_osf.c 			if (ctx->window == f->wss.val * mss ||
mss               147 net/netfilter/nfnetlink_osf.c 			if (ctx->window == f->wss.val * (mss + 40) ||
mss                35 net/netfilter/nft_synproxy.c 	opts->mss_option = info->mss;
mss               161 net/netfilter/nft_synproxy.c 		priv->info.mss = ntohs(nla_get_be16(tb[NFTA_SYNPROXY_MSS]));
mss               230 net/netfilter/nft_synproxy.c 	if (nla_put_be16(skb, NFTA_SYNPROXY_MSS, htons(priv->info.mss)) ||
mss               102 net/netfilter/xt_TCPMSS.c 	if (info->mss == XT_TCPMSS_CLAMP_PMTU) {
mss               114 net/netfilter/xt_TCPMSS.c 		newmss = info->mss;
mss               269 net/netfilter/xt_TCPMSS.c 	if (info->mss == XT_TCPMSS_CLAMP_PMTU &&
mss               293 net/netfilter/xt_TCPMSS.c 	if (info->mss == XT_TCPMSS_CLAMP_PMTU &&
mss               135 tools/testing/selftests/bpf/progs/test_tcp_check_syncookie_kern.c 		__u32 mss = seq_mss >> 32;
mss               138 tools/testing/selftests/bpf/progs/test_tcp_check_syncookie_kern.c 		bpf_map_update_elem(&results, &key_mss, &mss, 0);
mss               310 tools/testing/selftests/net/tcp_mmap.c 	int mss = 0;
mss               341 tools/testing/selftests/net/tcp_mmap.c 			mss = atoi(optarg);
mss               368 tools/testing/selftests/net/tcp_mmap.c 		if (mss &&
mss               370 tools/testing/selftests/net/tcp_mmap.c 			       &mss, sizeof(mss)) == -1) {
mss               400 tools/testing/selftests/net/tcp_mmap.c 	if (mss &&
mss               401 tools/testing/selftests/net/tcp_mmap.c 	    setsockopt(fd, IPPROTO_TCP, TCP_MAXSEG, &mss, sizeof(mss)) == -1) {
mss               508 tools/testing/selftests/net/udpgso.c 	int i, ret, val, mss;
mss               531 tools/testing/selftests/net/udpgso.c 		mss = test->gso_len;
mss               533 tools/testing/selftests/net/udpgso.c 		mss = addr->sa_family == AF_INET ? CONST_MSS_V4 : CONST_MSS_V6;
mss               539 tools/testing/selftests/net/udpgso.c 		if (ret != mss)
mss               540 tools/testing/selftests/net/udpgso.c 			error(1, 0, "recv.%d: %d != %d", i, ret, mss);