rx_buf_sz 768 drivers/atm/iphase.c iadev->rx_buf_sz = IA_RX_BUF_SZ; rx_buf_sz 779 drivers/atm/iphase.c iadev->rx_buf_sz = IA_RX_BUF_SZ; rx_buf_sz 791 drivers/atm/iphase.c iadev->rx_buf_sz = IA_RX_BUF_SZ; rx_buf_sz 796 drivers/atm/iphase.c iadev->rx_buf_sz, iadev->rx_pkt_ram);) rx_buf_sz 1171 drivers/atm/iphase.c if (len > iadev->rx_buf_sz) { rx_buf_sz 1172 drivers/atm/iphase.c printk("Over %d bytes sdu received, dropped!!!\n", iadev->rx_buf_sz); rx_buf_sz 1332 drivers/atm/iphase.c if ((length > iadev->rx_buf_sz) || (length > rx_buf_sz 1479 drivers/atm/iphase.c writew(iadev->rx_buf_sz, iadev->reass_reg+BUF_SIZE); rx_buf_sz 1493 drivers/atm/iphase.c rx_pkt_start += iadev->rx_buf_sz; rx_buf_sz 3155 drivers/atm/iphase.c iadev->num_rx_desc, iadev->rx_buf_sz, rx_buf_sz 1017 drivers/atm/iphase.h u16 num_rx_desc, rx_buf_sz, rxing; rx_buf_sz 363 drivers/iio/adc/at91-sama5d2_adc.c int rx_buf_sz; rx_buf_sz 806 drivers/iio/adc/at91-sama5d2_adc.c i = st->dma_st.rx_buf_sz - state.residue; rx_buf_sz 812 drivers/iio/adc/at91-sama5d2_adc.c size = st->dma_st.rx_buf_sz + i - st->dma_st.buf_idx; rx_buf_sz 841 drivers/iio/adc/at91-sama5d2_adc.c st->dma_st.rx_buf_sz = 0; rx_buf_sz 851 drivers/iio/adc/at91-sama5d2_adc.c st->dma_st.rx_buf_sz += chan->scan_type.storagebits / 8; rx_buf_sz 853 drivers/iio/adc/at91-sama5d2_adc.c st->dma_st.rx_buf_sz *= st->dma_st.watermark; rx_buf_sz 858 drivers/iio/adc/at91-sama5d2_adc.c st->dma_st.rx_buf_sz, rx_buf_sz 859 drivers/iio/adc/at91-sama5d2_adc.c st->dma_st.rx_buf_sz / 2, rx_buf_sz 1064 drivers/iio/adc/at91-sama5d2_adc.c sample_size = div_s64(st->dma_st.rx_buf_sz, st->dma_st.watermark); rx_buf_sz 1091 drivers/iio/adc/at91-sama5d2_adc.c if (st->dma_st.buf_idx >= st->dma_st.rx_buf_sz) rx_buf_sz 208 drivers/iio/adc/stm32-adc.c unsigned int rx_buf_sz; rx_buf_sz 1250 drivers/iio/adc/stm32-adc.c unsigned int rx_buf_sz = STM32_DMA_BUFFER_SIZE; rx_buf_sz 1259 drivers/iio/adc/stm32-adc.c adc->rx_buf_sz = min(rx_buf_sz, watermark * 2 * adc->num_conv); rx_buf_sz 1352 drivers/iio/adc/stm32-adc.c unsigned int i = adc->rx_buf_sz - state.residue; rx_buf_sz 1359 drivers/iio/adc/stm32-adc.c size = adc->rx_buf_sz + i - adc->bufi; rx_buf_sz 1391 drivers/iio/adc/stm32-adc.c if (adc->bufi >= adc->rx_buf_sz) rx_buf_sz 1407 drivers/iio/adc/stm32-adc.c adc->rx_buf_sz, adc->rx_buf_sz / 2); rx_buf_sz 1412 drivers/iio/adc/stm32-adc.c adc->rx_buf_sz, adc->rx_buf_sz / 2, rx_buf_sz 1552 drivers/iio/adc/stm32-adc.c if (adc->bufi >= adc->rx_buf_sz) rx_buf_sz 782 drivers/iio/adc/stm32-dfsdm-adc.c unsigned int rx_buf_sz = DFSDM_DMA_BUFFER_SIZE; rx_buf_sz 791 drivers/iio/adc/stm32-dfsdm-adc.c adc->buf_sz = min(rx_buf_sz, watermark * 2 * adc->nconv); rx_buf_sz 561 drivers/net/ethernet/adaptec/starfire.c unsigned int rx_buf_sz; /* Based on MTU+slack. */ rx_buf_sz 924 drivers/net/ethernet/adaptec/starfire.c writel((np->rx_buf_sz << RxBufferLenShift) | rx_buf_sz 1146 drivers/net/ethernet/adaptec/starfire.c np->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); rx_buf_sz 1150 drivers/net/ethernet/adaptec/starfire.c struct sk_buff *skb = netdev_alloc_skb(dev, np->rx_buf_sz); rx_buf_sz 1154 drivers/net/ethernet/adaptec/starfire.c np->rx_info[i].mapping = pci_map_single(np->pci_dev, skb->data, np->rx_buf_sz, PCI_DMA_FROMDEVICE); rx_buf_sz 1488 drivers/net/ethernet/adaptec/starfire.c pci_unmap_single(np->pci_dev, np->rx_info[entry].mapping, np->rx_buf_sz, PCI_DMA_FROMDEVICE); rx_buf_sz 1601 drivers/net/ethernet/adaptec/starfire.c skb = netdev_alloc_skb(dev, np->rx_buf_sz); rx_buf_sz 1606 drivers/net/ethernet/adaptec/starfire.c pci_map_single(np->pci_dev, skb->data, np->rx_buf_sz, PCI_DMA_FROMDEVICE); rx_buf_sz 1982 drivers/net/ethernet/adaptec/starfire.c pci_unmap_single(np->pci_dev, np->rx_info[i].mapping, np->rx_buf_sz, PCI_DMA_FROMDEVICE); rx_buf_sz 310 drivers/net/ethernet/dec/tulip/de2104x.c unsigned rx_buf_sz; rx_buf_sz 439 drivers/net/ethernet/dec/tulip/de2104x.c buflen = copying_skb ? (len + RX_OFFSET) : de->rx_buf_sz; rx_buf_sz 480 drivers/net/ethernet/dec/tulip/de2104x.c cpu_to_le32(RingEnd | de->rx_buf_sz); rx_buf_sz 482 drivers/net/ethernet/dec/tulip/de2104x.c de->rx_ring[rx_tail].opts2 = cpu_to_le32(de->rx_buf_sz); rx_buf_sz 1281 drivers/net/ethernet/dec/tulip/de2104x.c skb = netdev_alloc_skb(de->dev, de->rx_buf_sz); rx_buf_sz 1286 drivers/net/ethernet/dec/tulip/de2104x.c skb->data, de->rx_buf_sz, PCI_DMA_FROMDEVICE); rx_buf_sz 1292 drivers/net/ethernet/dec/tulip/de2104x.c cpu_to_le32(RingEnd | de->rx_buf_sz); rx_buf_sz 1294 drivers/net/ethernet/dec/tulip/de2104x.c de->rx_ring[i].opts2 = cpu_to_le32(de->rx_buf_sz); rx_buf_sz 1340 drivers/net/ethernet/dec/tulip/de2104x.c de->rx_buf_sz, PCI_DMA_FROMDEVICE); rx_buf_sz 1383 drivers/net/ethernet/dec/tulip/de2104x.c de->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); rx_buf_sz 313 drivers/net/ethernet/dec/tulip/winbond-840.c unsigned int rx_buf_sz; /* Based on MTU+slack. */ rx_buf_sz 802 drivers/net/ethernet/dec/tulip/winbond-840.c np->rx_ring[i].length = np->rx_buf_sz; rx_buf_sz 811 drivers/net/ethernet/dec/tulip/winbond-840.c struct sk_buff *skb = netdev_alloc_skb(dev, np->rx_buf_sz); rx_buf_sz 816 drivers/net/ethernet/dec/tulip/winbond-840.c np->rx_buf_sz,PCI_DMA_FROMDEVICE); rx_buf_sz 975 drivers/net/ethernet/dec/tulip/winbond-840.c np->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); rx_buf_sz 1267 drivers/net/ethernet/dec/tulip/winbond-840.c skb = netdev_alloc_skb(dev, np->rx_buf_sz); rx_buf_sz 1273 drivers/net/ethernet/dec/tulip/winbond-840.c np->rx_buf_sz, PCI_DMA_FROMDEVICE); rx_buf_sz 489 drivers/net/ethernet/dlink/dl2k.c np->rx_buf_sz = (dev->mtu <= 1500 ? PACKET_SIZE : dev->mtu + 32); rx_buf_sz 504 drivers/net/ethernet/dlink/dl2k.c skb = netdev_alloc_skb_ip_align(dev, np->rx_buf_sz); rx_buf_sz 517 drivers/net/ethernet/dlink/dl2k.c np->pdev, skb->data, np->rx_buf_sz, rx_buf_sz 519 drivers/net/ethernet/dlink/dl2k.c np->rx_ring[i].fraginfo |= cpu_to_le64((u64)np->rx_buf_sz << 48); rx_buf_sz 674 drivers/net/ethernet/dlink/dl2k.c np->rx_buf_sz); rx_buf_sz 685 drivers/net/ethernet/dlink/dl2k.c (np->pdev, skb->data, np->rx_buf_sz, rx_buf_sz 689 drivers/net/ethernet/dlink/dl2k.c cpu_to_le64((u64)np->rx_buf_sz << 48); rx_buf_sz 963 drivers/net/ethernet/dlink/dl2k.c np->rx_buf_sz, rx_buf_sz 970 drivers/net/ethernet/dlink/dl2k.c np->rx_buf_sz, rx_buf_sz 978 drivers/net/ethernet/dlink/dl2k.c np->rx_buf_sz, rx_buf_sz 1001 drivers/net/ethernet/dlink/dl2k.c skb = netdev_alloc_skb_ip_align(dev, np->rx_buf_sz); rx_buf_sz 1013 drivers/net/ethernet/dlink/dl2k.c (np->pdev, skb->data, np->rx_buf_sz, rx_buf_sz 1017 drivers/net/ethernet/dlink/dl2k.c cpu_to_le64((u64)np->rx_buf_sz << 48); rx_buf_sz 377 drivers/net/ethernet/dlink/dl2k.h unsigned int rx_buf_sz; /* Based on MTU+slack. */ rx_buf_sz 396 drivers/net/ethernet/dlink/sundance.c unsigned int rx_buf_sz; /* Based on MTU+slack. */ rx_buf_sz 1034 drivers/net/ethernet/dlink/sundance.c np->rx_buf_sz = (dev->mtu <= 1520 ? PKT_BUF_SZ : dev->mtu + 16); rx_buf_sz 1048 drivers/net/ethernet/dlink/sundance.c netdev_alloc_skb(dev, np->rx_buf_sz + 2); rx_buf_sz 1055 drivers/net/ethernet/dlink/sundance.c np->rx_buf_sz, DMA_FROM_DEVICE)); rx_buf_sz 1062 drivers/net/ethernet/dlink/sundance.c np->rx_ring[i].frag[0].length = cpu_to_le32(np->rx_buf_sz | LastFrag); rx_buf_sz 1388 drivers/net/ethernet/dlink/sundance.c np->rx_buf_sz, DMA_FROM_DEVICE); rx_buf_sz 1392 drivers/net/ethernet/dlink/sundance.c np->rx_buf_sz, DMA_FROM_DEVICE); rx_buf_sz 1397 drivers/net/ethernet/dlink/sundance.c np->rx_buf_sz, DMA_FROM_DEVICE); rx_buf_sz 1437 drivers/net/ethernet/dlink/sundance.c skb = netdev_alloc_skb(dev, np->rx_buf_sz + 2); rx_buf_sz 1444 drivers/net/ethernet/dlink/sundance.c np->rx_buf_sz, DMA_FROM_DEVICE)); rx_buf_sz 1454 drivers/net/ethernet/dlink/sundance.c cpu_to_le32(np->rx_buf_sz | LastFrag); rx_buf_sz 1909 drivers/net/ethernet/dlink/sundance.c np->rx_buf_sz, DMA_FROM_DEVICE); rx_buf_sz 407 drivers/net/ethernet/fealnx.c unsigned int rx_buf_sz; /* Based on MTU+slack. */ rx_buf_sz 1065 drivers/net/ethernet/fealnx.c skb = netdev_alloc_skb(dev, np->rx_buf_sz); rx_buf_sz 1074 drivers/net/ethernet/fealnx.c np->rx_buf_sz, PCI_DMA_FROMDEVICE); rx_buf_sz 1238 drivers/net/ethernet/fealnx.c np->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); rx_buf_sz 1246 drivers/net/ethernet/fealnx.c np->rx_ring[i].control = np->rx_buf_sz << RBSShift; rx_buf_sz 1259 drivers/net/ethernet/fealnx.c struct sk_buff *skb = netdev_alloc_skb(dev, np->rx_buf_sz); rx_buf_sz 1269 drivers/net/ethernet/fealnx.c np->rx_buf_sz, PCI_DMA_FROMDEVICE); rx_buf_sz 1701 drivers/net/ethernet/fealnx.c np->rx_buf_sz, rx_buf_sz 1715 drivers/net/ethernet/fealnx.c np->rx_buf_sz, rx_buf_sz 1720 drivers/net/ethernet/fealnx.c np->rx_buf_sz, rx_buf_sz 1915 drivers/net/ethernet/fealnx.c np->rx_buf_sz, PCI_DMA_FROMDEVICE); rx_buf_sz 264 drivers/net/ethernet/huawei/hinic/hinic_hw_dev.c fw_ctxt.rx_buf_sz = HINIC_RX_BUF_SZ; rx_buf_sz 135 drivers/net/ethernet/huawei/hinic/hinic_hw_dev.h u16 rx_buf_sz; rx_buf_sz 561 drivers/net/ethernet/natsemi/natsemi.c unsigned int rx_buf_sz; rx_buf_sz 1748 drivers/net/ethernet/natsemi/natsemi.c if (np->rx_buf_sz > NATSEMI_LONGPKT) rx_buf_sz 1937 drivers/net/ethernet/natsemi/natsemi.c unsigned int buflen = np->rx_buf_sz+NATSEMI_PADDING; rx_buf_sz 1952 drivers/net/ethernet/natsemi/natsemi.c np->rx_ring[entry].cmd_status = cpu_to_le32(np->rx_buf_sz); rx_buf_sz 1965 drivers/net/ethernet/natsemi/natsemi.c np->rx_buf_sz = ETH_DATA_LEN + NATSEMI_HEADERS; rx_buf_sz 1967 drivers/net/ethernet/natsemi/natsemi.c np->rx_buf_sz = dev->mtu + NATSEMI_HEADERS; rx_buf_sz 2029 drivers/net/ethernet/natsemi/natsemi.c unsigned int buflen = np->rx_buf_sz; rx_buf_sz 2288 drivers/net/ethernet/natsemi/natsemi.c unsigned int buflen = np->rx_buf_sz; rx_buf_sz 2348 drivers/net/ethernet/natsemi/natsemi.c } else if (pkt_len > np->rx_buf_sz) { rx_buf_sz 805 drivers/net/ethernet/nvidia/forcedeth.c unsigned int rx_buf_sz; rx_buf_sz 1837 drivers/net/ethernet/nvidia/forcedeth.c struct sk_buff *skb = netdev_alloc_skb(dev, np->rx_buf_sz + NV_RX_ALLOC_PAD); rx_buf_sz 1852 drivers/net/ethernet/nvidia/forcedeth.c np->put_rx.orig->flaglen = cpu_to_le32(np->rx_buf_sz | NV_RX_AVAIL); rx_buf_sz 1878 drivers/net/ethernet/nvidia/forcedeth.c struct sk_buff *skb = netdev_alloc_skb(dev, np->rx_buf_sz + NV_RX_ALLOC_PAD); rx_buf_sz 1894 drivers/net/ethernet/nvidia/forcedeth.c np->put_rx.ex->flaglen = cpu_to_le32(np->rx_buf_sz | NV_RX2_AVAIL); rx_buf_sz 3052 drivers/net/ethernet/nvidia/forcedeth.c np->rx_buf_sz = ETH_DATA_LEN + NV_RX_HEADERS; rx_buf_sz 3054 drivers/net/ethernet/nvidia/forcedeth.c np->rx_buf_sz = dev->mtu + NV_RX_HEADERS; rx_buf_sz 3099 drivers/net/ethernet/nvidia/forcedeth.c writel(np->rx_buf_sz, base + NvRegOffloadConfig); rx_buf_sz 4181 drivers/net/ethernet/nvidia/forcedeth.c writel(np->rx_buf_sz, base + NvRegOffloadConfig); rx_buf_sz 4734 drivers/net/ethernet/nvidia/forcedeth.c writel(np->rx_buf_sz, base + NvRegOffloadConfig); rx_buf_sz 5128 drivers/net/ethernet/nvidia/forcedeth.c writel(np->rx_buf_sz, base + NvRegOffloadConfig); rx_buf_sz 5287 drivers/net/ethernet/nvidia/forcedeth.c writel(np->rx_buf_sz, base + NvRegOffloadConfig); rx_buf_sz 5485 drivers/net/ethernet/nvidia/forcedeth.c writel(np->rx_buf_sz, base + NvRegOffloadConfig); rx_buf_sz 496 drivers/net/ethernet/packetengines/hamachi.c unsigned int rx_buf_sz; /* Based on MTU+slack. */ rx_buf_sz 1120 drivers/net/ethernet/packetengines/hamachi.c hmp->rx_buf_sz, PCI_DMA_FROMDEVICE); rx_buf_sz 1129 drivers/net/ethernet/packetengines/hamachi.c skb = netdev_alloc_skb_ip_align(dev, hmp->rx_buf_sz); rx_buf_sz 1135 drivers/net/ethernet/packetengines/hamachi.c skb->data, hmp->rx_buf_sz, PCI_DMA_FROMDEVICE)); rx_buf_sz 1137 drivers/net/ethernet/packetengines/hamachi.c DescEndPacket | DescIntr | (hmp->rx_buf_sz - 2)); rx_buf_sz 1171 drivers/net/ethernet/packetengines/hamachi.c hmp->rx_buf_sz = (dev->mtu <= 1492 ? PKT_BUF_SZ : rx_buf_sz 1181 drivers/net/ethernet/packetengines/hamachi.c struct sk_buff *skb = netdev_alloc_skb(dev, hmp->rx_buf_sz + 2); rx_buf_sz 1187 drivers/net/ethernet/packetengines/hamachi.c skb->data, hmp->rx_buf_sz, PCI_DMA_FROMDEVICE)); rx_buf_sz 1190 drivers/net/ethernet/packetengines/hamachi.c DescEndPacket | DescIntr | (hmp->rx_buf_sz -2)); rx_buf_sz 1418 drivers/net/ethernet/packetengines/hamachi.c hmp->rx_buf_sz, rx_buf_sz 1488 drivers/net/ethernet/packetengines/hamachi.c hmp->rx_buf_sz, rx_buf_sz 1501 drivers/net/ethernet/packetengines/hamachi.c hmp->rx_buf_sz, rx_buf_sz 1506 drivers/net/ethernet/packetengines/hamachi.c hmp->rx_buf_sz, PCI_DMA_FROMDEVICE); rx_buf_sz 1583 drivers/net/ethernet/packetengines/hamachi.c struct sk_buff *skb = netdev_alloc_skb(dev, hmp->rx_buf_sz + 2); rx_buf_sz 1590 drivers/net/ethernet/packetengines/hamachi.c skb->data, hmp->rx_buf_sz, PCI_DMA_FROMDEVICE)); rx_buf_sz 1592 drivers/net/ethernet/packetengines/hamachi.c desc->status_n_length = cpu_to_le32(hmp->rx_buf_sz); rx_buf_sz 1709 drivers/net/ethernet/packetengines/hamachi.c hmp->rx_buf_sz, PCI_DMA_FROMDEVICE); rx_buf_sz 324 drivers/net/ethernet/packetengines/yellowfin.c unsigned int rx_buf_sz; /* Based on MTU+slack. */ rx_buf_sz 728 drivers/net/ethernet/packetengines/yellowfin.c yp->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); rx_buf_sz 732 drivers/net/ethernet/packetengines/yellowfin.c cpu_to_le32(CMD_RX_BUF | INTR_ALWAYS | yp->rx_buf_sz); rx_buf_sz 738 drivers/net/ethernet/packetengines/yellowfin.c struct sk_buff *skb = netdev_alloc_skb(dev, yp->rx_buf_sz + 2); rx_buf_sz 744 drivers/net/ethernet/packetengines/yellowfin.c skb->data, yp->rx_buf_sz, PCI_DMA_FROMDEVICE)); rx_buf_sz 1059 drivers/net/ethernet/packetengines/yellowfin.c yp->rx_buf_sz, PCI_DMA_FROMDEVICE); rx_buf_sz 1126 drivers/net/ethernet/packetengines/yellowfin.c yp->rx_buf_sz, rx_buf_sz 1138 drivers/net/ethernet/packetengines/yellowfin.c yp->rx_buf_sz, rx_buf_sz 1153 drivers/net/ethernet/packetengines/yellowfin.c struct sk_buff *skb = netdev_alloc_skb(dev, yp->rx_buf_sz + 2); rx_buf_sz 1159 drivers/net/ethernet/packetengines/yellowfin.c skb->data, yp->rx_buf_sz, PCI_DMA_FROMDEVICE)); rx_buf_sz 1165 drivers/net/ethernet/packetengines/yellowfin.c cpu_to_le32(CMD_RX_BUF | INTR_ALWAYS | yp->rx_buf_sz); rx_buf_sz 1169 drivers/net/ethernet/packetengines/yellowfin.c | yp->rx_buf_sz); rx_buf_sz 347 drivers/net/ethernet/realtek/8139cp.c unsigned rx_buf_sz; rx_buf_sz 413 drivers/net/ethernet/realtek/8139cp.c cp->rx_buf_sz = mtu + ETH_HLEN + 8; rx_buf_sz 415 drivers/net/ethernet/realtek/8139cp.c cp->rx_buf_sz = PKT_BUF_SZ; rx_buf_sz 477 drivers/net/ethernet/realtek/8139cp.c const unsigned buflen = cp->rx_buf_sz; rx_buf_sz 546 drivers/net/ethernet/realtek/8139cp.c cp->rx_buf_sz); rx_buf_sz 548 drivers/net/ethernet/realtek/8139cp.c desc->opts1 = cpu_to_le32(DescOwn | cp->rx_buf_sz); rx_buf_sz 1067 drivers/net/ethernet/realtek/8139cp.c skb = netdev_alloc_skb_ip_align(dev, cp->rx_buf_sz); rx_buf_sz 1072 drivers/net/ethernet/realtek/8139cp.c cp->rx_buf_sz, PCI_DMA_FROMDEVICE); rx_buf_sz 1083 drivers/net/ethernet/realtek/8139cp.c cpu_to_le32(DescOwn | RingEnd | cp->rx_buf_sz); rx_buf_sz 1086 drivers/net/ethernet/realtek/8139cp.c cpu_to_le32(DescOwn | cp->rx_buf_sz); rx_buf_sz 1142 drivers/net/ethernet/realtek/8139cp.c cp->rx_buf_sz, PCI_DMA_FROMDEVICE); rx_buf_sz 1320 drivers/net/ethernet/renesas/sh_eth.c ALIGN(mdp->rx_buf_sz, 32), rx_buf_sz 1362 drivers/net/ethernet/renesas/sh_eth.c int skbuff_size = mdp->rx_buf_sz + SH_ETH_RX_ALIGN + 32 - 1; rx_buf_sz 1383 drivers/net/ethernet/renesas/sh_eth.c buf_len = ALIGN(mdp->rx_buf_sz, 32); rx_buf_sz 1442 drivers/net/ethernet/renesas/sh_eth.c mdp->rx_buf_sz = (ndev->mtu <= 1492 ? PKT_BUF_SZ : rx_buf_sz 1445 drivers/net/ethernet/renesas/sh_eth.c mdp->rx_buf_sz += NET_IP_ALIGN; rx_buf_sz 1629 drivers/net/ethernet/renesas/sh_eth.c int skbuff_size = mdp->rx_buf_sz + SH_ETH_RX_ALIGN + 32 - 1; rx_buf_sz 1688 drivers/net/ethernet/renesas/sh_eth.c ALIGN(mdp->rx_buf_sz, 32), rx_buf_sz 1709 drivers/net/ethernet/renesas/sh_eth.c buf_len = ALIGN(mdp->rx_buf_sz, 32); rx_buf_sz 531 drivers/net/ethernet/renesas/sh_eth.h u32 rx_buf_sz; /* Based on MTU+slack. */ rx_buf_sz 270 drivers/net/ethernet/sis/sis190.c u32 rx_buf_sz; rx_buf_sz 461 drivers/net/ethernet/sis/sis190.c static inline void sis190_give_to_asic(struct RxDesc *desc, u32 rx_buf_sz) rx_buf_sz 466 drivers/net/ethernet/sis/sis190.c desc->size = cpu_to_le32((rx_buf_sz & RX_BUF_MASK) | eor); rx_buf_sz 472 drivers/net/ethernet/sis/sis190.c u32 rx_buf_sz) rx_buf_sz 475 drivers/net/ethernet/sis/sis190.c sis190_give_to_asic(desc, rx_buf_sz); rx_buf_sz 490 drivers/net/ethernet/sis/sis190.c u32 rx_buf_sz = tp->rx_buf_sz; rx_buf_sz 494 drivers/net/ethernet/sis/sis190.c skb = netdev_alloc_skb(tp->dev, rx_buf_sz); rx_buf_sz 497 drivers/net/ethernet/sis/sis190.c mapping = pci_map_single(tp->pci_dev, skb->data, tp->rx_buf_sz, rx_buf_sz 501 drivers/net/ethernet/sis/sis190.c sis190_map_to_asic(desc, mapping, rx_buf_sz); rx_buf_sz 545 drivers/net/ethernet/sis/sis190.c pci_dma_sync_single_for_cpu(tp->pci_dev, addr, tp->rx_buf_sz, rx_buf_sz 597 drivers/net/ethernet/sis/sis190.c sis190_give_to_asic(desc, tp->rx_buf_sz); rx_buf_sz 604 drivers/net/ethernet/sis/sis190.c if (unlikely(pkt_size > tp->rx_buf_sz)) { rx_buf_sz 609 drivers/net/ethernet/sis/sis190.c sis190_give_to_asic(desc, tp->rx_buf_sz); rx_buf_sz 616 drivers/net/ethernet/sis/sis190.c tp->rx_buf_sz, PCI_DMA_FROMDEVICE); rx_buf_sz 617 drivers/net/ethernet/sis/sis190.c sis190_give_to_asic(desc, tp->rx_buf_sz); rx_buf_sz 619 drivers/net/ethernet/sis/sis190.c pci_unmap_single(pdev, addr, tp->rx_buf_sz, rx_buf_sz 788 drivers/net/ethernet/sis/sis190.c pci_unmap_single(pdev, le32_to_cpu(desc->addr), tp->rx_buf_sz, rx_buf_sz 1052 drivers/net/ethernet/sis/sis190.c tp->rx_buf_sz = (mtu > RX_BUF_SIZE) ? mtu + ETH_HLEN + 8 : RX_BUF_SIZE; rx_buf_sz 1054 drivers/net/ethernet/sis/sis190.c if (tp->rx_buf_sz & 0x07) { rx_buf_sz 1055 drivers/net/ethernet/sis/sis190.c tp->rx_buf_sz += 8; rx_buf_sz 1056 drivers/net/ethernet/sis/sis190.c tp->rx_buf_sz &= RX_BUF_MASK; rx_buf_sz 271 drivers/net/ethernet/smsc/epic100.c unsigned int rx_buf_sz; /* Based on MTU+slack. */ rx_buf_sz 900 drivers/net/ethernet/smsc/epic100.c ep->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); rx_buf_sz 905 drivers/net/ethernet/smsc/epic100.c ep->rx_ring[i].buflength = ep->rx_buf_sz; rx_buf_sz 915 drivers/net/ethernet/smsc/epic100.c struct sk_buff *skb = netdev_alloc_skb(dev, ep->rx_buf_sz + 2); rx_buf_sz 921 drivers/net/ethernet/smsc/epic100.c skb->data, ep->rx_buf_sz, PCI_DMA_FROMDEVICE); rx_buf_sz 1182 drivers/net/ethernet/smsc/epic100.c ep->rx_buf_sz, rx_buf_sz 1188 drivers/net/ethernet/smsc/epic100.c ep->rx_buf_sz, rx_buf_sz 1193 drivers/net/ethernet/smsc/epic100.c ep->rx_buf_sz, PCI_DMA_FROMDEVICE); rx_buf_sz 1211 drivers/net/ethernet/smsc/epic100.c skb = ep->rx_skbuff[entry] = netdev_alloc_skb(dev, ep->rx_buf_sz + 2); rx_buf_sz 1216 drivers/net/ethernet/smsc/epic100.c skb->data, ep->rx_buf_sz, PCI_DMA_FROMDEVICE); rx_buf_sz 1297 drivers/net/ethernet/smsc/epic100.c ep->rx_buf_sz, PCI_DMA_FROMDEVICE); rx_buf_sz 828 drivers/net/ethernet/sun/sungem.c skb_put(new_skb, (gp->rx_buf_sz + RX_OFFSET)); rx_buf_sz 1636 drivers/net/ethernet/sun/sungem.c gp->rx_buf_sz = max(dev->mtu + ETH_HLEN + VLAN_HLEN, rx_buf_sz 1651 drivers/net/ethernet/sun/sungem.c skb_put(skb, (gp->rx_buf_sz + RX_OFFSET)); rx_buf_sz 1833 drivers/net/ethernet/sun/sungem.c writel(0x20000000 | (gp->rx_buf_sz + 4), gp->regs + MAC_MAXFSZ); rx_buf_sz 1914 drivers/net/ethernet/sun/sungem.c int max_frame = (gp->rx_buf_sz + 4 + 64) & ~63; rx_buf_sz 939 drivers/net/ethernet/sun/sungem.h #define RX_BUF_ALLOC_SIZE(gp) ((gp)->rx_buf_sz + 28 + RX_OFFSET + 64) rx_buf_sz 994 drivers/net/ethernet/sun/sungem.h int rx_buf_sz; rx_buf_sz 476 drivers/net/ethernet/via/via-rhine.c unsigned int rx_buf_sz; /* Based on MTU+slack. */ rx_buf_sz 1219 drivers/net/ethernet/via/via-rhine.c const int size = rp->rx_buf_sz; rx_buf_sz 1263 drivers/net/ethernet/via/via-rhine.c rp->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); rx_buf_sz 1269 drivers/net/ethernet/via/via-rhine.c rp->rx_ring[i].desc_length = cpu_to_le32(rp->rx_buf_sz); rx_buf_sz 1308 drivers/net/ethernet/via/via-rhine.c rp->rx_buf_sz, DMA_FROM_DEVICE); rx_buf_sz 2104 drivers/net/ethernet/via/via-rhine.c rp->rx_buf_sz, rx_buf_sz 2113 drivers/net/ethernet/via/via-rhine.c rp->rx_buf_sz, rx_buf_sz 2125 drivers/net/ethernet/via/via-rhine.c rp->rx_buf_sz, rx_buf_sz 792 drivers/net/usb/r8152.c u32 rx_buf_sz; rx_buf_sz 1540 drivers/net/usb/r8152.c unsigned int order = get_order(tp->rx_buf_sz); rx_buf_sz 2278 drivers/net/usb/r8152.c agg->buffer, tp->rx_buf_sz, rx_buf_sz 2648 drivers/net/usb/r8152.c u32 ocp_data = tp->rx_buf_sz - rx_reserved_size(tp->netdev->mtu); rx_buf_sz 5535 drivers/net/usb/r8152.c tp->rx_buf_sz = 16 * 1024; rx_buf_sz 5555 drivers/net/usb/r8152.c tp->rx_buf_sz = 32 * 1024; rx_buf_sz 5573 drivers/net/usb/r8152.c tp->rx_buf_sz = 32 * 1024; rx_buf_sz 560 drivers/staging/vt6655/card.c pDesc->rd0.res_count = cpu_to_le16(priv->rx_buf_sz); rx_buf_sz 562 drivers/staging/vt6655/card.c pDesc->rd1.req_count = cpu_to_le16(priv->rx_buf_sz); rx_buf_sz 568 drivers/staging/vt6655/card.c pDesc->rd0.res_count = cpu_to_le16(priv->rx_buf_sz); rx_buf_sz 570 drivers/staging/vt6655/card.c pDesc->rd1.req_count = cpu_to_le16(priv->rx_buf_sz); rx_buf_sz 153 drivers/staging/vt6655/device.h u32 rx_buf_sz; rx_buf_sz 783 drivers/staging/vt6655/device_main.c rd_info->skb = dev_alloc_skb((int)priv->rx_buf_sz); rx_buf_sz 790 drivers/staging/vt6655/device_main.c priv->rx_buf_sz, DMA_FROM_DEVICE); rx_buf_sz 799 drivers/staging/vt6655/device_main.c rd->rd0.res_count = cpu_to_le16(priv->rx_buf_sz); rx_buf_sz 801 drivers/staging/vt6655/device_main.c rd->rd1.req_count = cpu_to_le16(priv->rx_buf_sz); rx_buf_sz 813 drivers/staging/vt6655/device_main.c priv->rx_buf_sz, DMA_FROM_DEVICE); rx_buf_sz 1224 drivers/staging/vt6655/device_main.c priv->rx_buf_sz = PKT_BUF_SZ; rx_buf_sz 129 drivers/staging/vt6655/dpc.c priv->rx_buf_sz, DMA_FROM_DEVICE); rx_buf_sz 269 drivers/staging/vt6656/device.h u32 rx_buf_sz; rx_buf_sz 483 drivers/staging/vt6656/main_usb.c rcb->skb = dev_alloc_skb(priv->rx_buf_sz); rx_buf_sz 535 drivers/staging/vt6656/main_usb.c priv->rx_buf_sz = MAX_TOTAL_SIZE_WITH_ALL_HEADERS; rx_buf_sz 229 drivers/staging/vt6656/usbpipe.c rcb->skb = dev_alloc_skb(priv->rx_buf_sz);