Lines Matching refs:np
483 struct netdev_private *np; in fealnx_init_one() local
549 np = netdev_priv(dev); in fealnx_init_one()
550 np->mem = ioaddr; in fealnx_init_one()
551 spin_lock_init(&np->lock); in fealnx_init_one()
552 np->pci_dev = pdev; in fealnx_init_one()
553 np->flags = skel_netdrv_tbl[chip_id].flags; in fealnx_init_one()
555 np->mii.dev = dev; in fealnx_init_one()
556 np->mii.mdio_read = mdio_read; in fealnx_init_one()
557 np->mii.mdio_write = mdio_write; in fealnx_init_one()
558 np->mii.phy_id_mask = 0x1f; in fealnx_init_one()
559 np->mii.reg_num_mask = 0x1f; in fealnx_init_one()
566 np->rx_ring = ring_space; in fealnx_init_one()
567 np->rx_ring_dma = ring_dma; in fealnx_init_one()
574 np->tx_ring = ring_space; in fealnx_init_one()
575 np->tx_ring_dma = ring_dma; in fealnx_init_one()
578 if (np->flags == HAS_MII_XCVR) { in fealnx_init_one()
581 for (phy = 1; phy < 32 && phy_idx < ARRAY_SIZE(np->phys); in fealnx_init_one()
586 np->phys[phy_idx++] = phy; in fealnx_init_one()
594 data = mdio_read(dev, np->phys[0], 2); in fealnx_init_one()
596 np->PHYType = SeeqPHY; in fealnx_init_one()
598 np->PHYType = AhdocPHY; in fealnx_init_one()
600 np->PHYType = MarvellPHY; in fealnx_init_one()
602 np->PHYType = Myson981; in fealnx_init_one()
604 np->PHYType = LevelOnePHY; in fealnx_init_one()
606 np->PHYType = OtherPHY; in fealnx_init_one()
611 np->mii_cnt = phy_idx; in fealnx_init_one()
617 np->phys[0] = 32; in fealnx_init_one()
621 np->PHYType = MysonPHY; in fealnx_init_one()
623 np->PHYType = OtherPHY; in fealnx_init_one()
625 np->mii.phy_id = np->phys[0]; in fealnx_init_one()
633 np->mii.full_duplex = 1; in fealnx_init_one()
634 np->default_port = option & 15; in fealnx_init_one()
638 np->mii.full_duplex = full_duplex[card_idx]; in fealnx_init_one()
640 if (np->mii.full_duplex) { in fealnx_init_one()
644 if ((np->PHYType == MarvellPHY) || (np->PHYType == LevelOnePHY)) { in fealnx_init_one()
647 data = mdio_read(dev, np->phys[0], 9); in fealnx_init_one()
649 mdio_write(dev, np->phys[0], 9, data); in fealnx_init_one()
652 if (np->flags == HAS_MII_XCVR) in fealnx_init_one()
653 mdio_write(dev, np->phys[0], MII_ADVERTISE, ADVERTISE_FULL); in fealnx_init_one()
656 np->mii.force_media = 1; in fealnx_init_one()
674 pci_free_consistent(pdev, TX_TOTAL_SIZE, np->tx_ring, np->tx_ring_dma); in fealnx_init_one()
676 pci_free_consistent(pdev, RX_TOTAL_SIZE, np->rx_ring, np->rx_ring_dma); in fealnx_init_one()
692 struct netdev_private *np = netdev_priv(dev); in fealnx_remove_one() local
694 pci_free_consistent(pdev, TX_TOTAL_SIZE, np->tx_ring, in fealnx_remove_one()
695 np->tx_ring_dma); in fealnx_remove_one()
696 pci_free_consistent(pdev, RX_TOTAL_SIZE, np->rx_ring, in fealnx_remove_one()
697 np->rx_ring_dma); in fealnx_remove_one()
699 pci_iounmap(pdev, np->mem); in fealnx_remove_one()
758 struct netdev_private *np = netdev_priv(dev); in mdio_read() local
759 void __iomem *miiport = np->mem + MANAGEMENT; in mdio_read()
797 struct netdev_private *np = netdev_priv(dev); in mdio_write() local
798 void __iomem *miiport = np->mem + MANAGEMENT; in mdio_write()
829 struct netdev_private *np = netdev_priv(dev); in netdev_open() local
830 void __iomem *ioaddr = np->mem; in netdev_open()
831 const int irq = np->pci_dev->irq; in netdev_open()
846 iowrite32(np->rx_ring_dma, ioaddr + RXLBA); in netdev_open()
847 iowrite32(np->tx_ring_dma, ioaddr + TXLBA); in netdev_open()
866 np->bcrvalue = 0x10; /* little-endian, 8 burst length */ in netdev_open()
868 np->bcrvalue |= 0x04; /* big-endian */ in netdev_open()
873 np->crvalue = 0xa00; in netdev_open()
876 np->crvalue = 0xe00; /* rx 128 burst length */ in netdev_open()
882 np->imrvalue = TUNF | CNTOVF | RBU | TI | RI; in netdev_open()
883 if (np->pci_dev->device == 0x891) { in netdev_open()
884 np->bcrvalue |= 0x200; /* set PROG bit */ in netdev_open()
885 np->crvalue |= CR_W_ENH; /* set enhanced bit */ in netdev_open()
886 np->imrvalue |= ETI; in netdev_open()
888 iowrite32(np->bcrvalue, ioaddr + BCR); in netdev_open()
891 dev->if_port = np->default_port; in netdev_open()
896 np->crvalue |= 0x00e40001; /* tx store and forward, tx/rx enable */ in netdev_open()
897 np->mii.full_duplex = np->mii.force_media; in netdev_open()
899 if (np->linkok) in netdev_open()
907 iowrite32(np->imrvalue, ioaddr + IMR); in netdev_open()
913 init_timer(&np->timer); in netdev_open()
914 np->timer.expires = RUN_AT(3 * HZ); in netdev_open()
915 np->timer.data = (unsigned long) dev; in netdev_open()
916 np->timer.function = netdev_timer; in netdev_open()
919 add_timer(&np->timer); in netdev_open()
921 init_timer(&np->reset_timer); in netdev_open()
922 np->reset_timer.data = (unsigned long) dev; in netdev_open()
923 np->reset_timer.function = reset_timer; in netdev_open()
924 np->reset_timer_armed = 0; in netdev_open()
934 struct netdev_private *np = netdev_priv(dev); in getlinkstatus() local
937 np->linkok = 0; in getlinkstatus()
939 if (np->PHYType == MysonPHY) { in getlinkstatus()
941 if (ioread32(np->mem + BMCRSR) & LinkIsUp2) { in getlinkstatus()
942 np->linkok = 1; in getlinkstatus()
949 if (mdio_read(dev, np->phys[0], MII_BMSR) & BMSR_LSTATUS) { in getlinkstatus()
950 np->linkok = 1; in getlinkstatus()
961 struct netdev_private *np = netdev_priv(dev); in getlinktype() local
963 if (np->PHYType == MysonPHY) { /* 3-in-1 case */ in getlinktype()
964 if (ioread32(np->mem + TCRRCR) & CR_R_FD) in getlinktype()
965 np->duplexmode = 2; /* full duplex */ in getlinktype()
967 np->duplexmode = 1; /* half duplex */ in getlinktype()
968 if (ioread32(np->mem + TCRRCR) & CR_R_PS10) in getlinktype()
969 np->line_speed = 1; /* 10M */ in getlinktype()
971 np->line_speed = 2; /* 100M */ in getlinktype()
973 if (np->PHYType == SeeqPHY) { /* this PHY is SEEQ 80225 */ in getlinktype()
976 data = mdio_read(dev, np->phys[0], MIIRegister18); in getlinktype()
978 np->line_speed = 2; /* 100M */ in getlinktype()
980 np->line_speed = 1; /* 10M */ in getlinktype()
982 np->duplexmode = 2; /* full duplex mode */ in getlinktype()
984 np->duplexmode = 1; /* half duplex mode */ in getlinktype()
985 } else if (np->PHYType == AhdocPHY) { in getlinktype()
988 data = mdio_read(dev, np->phys[0], DiagnosticReg); in getlinktype()
990 np->line_speed = 2; /* 100M */ in getlinktype()
992 np->line_speed = 1; /* 10M */ in getlinktype()
994 np->duplexmode = 2; /* full duplex mode */ in getlinktype()
996 np->duplexmode = 1; /* half duplex mode */ in getlinktype()
999 else if (np->PHYType == MarvellPHY) { in getlinktype()
1002 data = mdio_read(dev, np->phys[0], SpecificReg); in getlinktype()
1004 np->duplexmode = 2; /* full duplex mode */ in getlinktype()
1006 np->duplexmode = 1; /* half duplex mode */ in getlinktype()
1009 np->line_speed = 3; /* 1000M */ in getlinktype()
1011 np->line_speed = 2; /* 100M */ in getlinktype()
1013 np->line_speed = 1; /* 10M */ in getlinktype()
1017 else if (np->PHYType == Myson981) { in getlinktype()
1020 data = mdio_read(dev, np->phys[0], StatusRegister); in getlinktype()
1023 np->line_speed = 2; in getlinktype()
1025 np->line_speed = 1; in getlinktype()
1028 np->duplexmode = 2; in getlinktype()
1030 np->duplexmode = 1; in getlinktype()
1034 else if (np->PHYType == LevelOnePHY) { in getlinktype()
1037 data = mdio_read(dev, np->phys[0], SpecificReg); in getlinktype()
1039 np->duplexmode = 2; /* full duplex mode */ in getlinktype()
1041 np->duplexmode = 1; /* half duplex mode */ in getlinktype()
1044 np->line_speed = 3; /* 1000M */ in getlinktype()
1046 np->line_speed = 2; /* 100M */ in getlinktype()
1048 np->line_speed = 1; /* 10M */ in getlinktype()
1050 np->crvalue &= (~CR_W_PS10) & (~CR_W_FD) & (~CR_W_PS1000); in getlinktype()
1051 if (np->line_speed == 1) in getlinktype()
1052 np->crvalue |= CR_W_PS10; in getlinktype()
1053 else if (np->line_speed == 3) in getlinktype()
1054 np->crvalue |= CR_W_PS1000; in getlinktype()
1055 if (np->duplexmode == 2) in getlinktype()
1056 np->crvalue |= CR_W_FD; in getlinktype()
1064 struct netdev_private *np = netdev_priv(dev); in allocate_rx_buffers() local
1067 while (np->really_rx_count != RX_RING_SIZE) { in allocate_rx_buffers()
1070 skb = netdev_alloc_skb(dev, np->rx_buf_sz); in allocate_rx_buffers()
1074 while (np->lack_rxbuf->skbuff) in allocate_rx_buffers()
1075 np->lack_rxbuf = np->lack_rxbuf->next_desc_logical; in allocate_rx_buffers()
1077 np->lack_rxbuf->skbuff = skb; in allocate_rx_buffers()
1078 np->lack_rxbuf->buffer = pci_map_single(np->pci_dev, skb->data, in allocate_rx_buffers()
1079 np->rx_buf_sz, PCI_DMA_FROMDEVICE); in allocate_rx_buffers()
1080 np->lack_rxbuf->status = RXOWN; in allocate_rx_buffers()
1081 ++np->really_rx_count; in allocate_rx_buffers()
1089 struct netdev_private *np = netdev_priv(dev); in netdev_timer() local
1090 void __iomem *ioaddr = np->mem; in netdev_timer()
1091 int old_crvalue = np->crvalue; in netdev_timer()
1092 unsigned int old_linkok = np->linkok; in netdev_timer()
1100 spin_lock_irqsave(&np->lock, flags); in netdev_timer()
1102 if (np->flags == HAS_MII_XCVR) { in netdev_timer()
1104 if ((old_linkok == 0) && (np->linkok == 1)) { /* we need to detect the media type again */ in netdev_timer()
1106 if (np->crvalue != old_crvalue) { in netdev_timer()
1107 stop_nic_rxtx(ioaddr, np->crvalue); in netdev_timer()
1108 iowrite32(np->crvalue, ioaddr + TCRRCR); in netdev_timer()
1115 spin_unlock_irqrestore(&np->lock, flags); in netdev_timer()
1117 np->timer.expires = RUN_AT(10 * HZ); in netdev_timer()
1118 add_timer(&np->timer); in netdev_timer()
1126 struct netdev_private *np = netdev_priv(dev); in reset_and_disable_rxtx() local
1127 void __iomem *ioaddr = np->mem; in reset_and_disable_rxtx()
1152 struct netdev_private *np = netdev_priv(dev); in enable_rxtx() local
1153 void __iomem *ioaddr = np->mem; in enable_rxtx()
1157 iowrite32(np->tx_ring_dma + ((char*)np->cur_tx - (char*)np->tx_ring), in enable_rxtx()
1159 iowrite32(np->rx_ring_dma + ((char*)np->cur_rx - (char*)np->rx_ring), in enable_rxtx()
1162 iowrite32(np->bcrvalue, ioaddr + BCR); in enable_rxtx()
1169 iowrite32(np->imrvalue, ioaddr + IMR); in enable_rxtx()
1178 struct netdev_private *np = netdev_priv(dev); in reset_timer() local
1183 spin_lock_irqsave(&np->lock, flags); in reset_timer()
1184 np->crvalue = np->crvalue_sv; in reset_timer()
1185 np->imrvalue = np->imrvalue_sv; in reset_timer()
1193 np->reset_timer_armed = 0; in reset_timer()
1195 spin_unlock_irqrestore(&np->lock, flags); in reset_timer()
1201 struct netdev_private *np = netdev_priv(dev); in fealnx_tx_timeout() local
1202 void __iomem *ioaddr = np->mem; in fealnx_tx_timeout()
1211 printk(KERN_DEBUG " Rx ring %p: ", np->rx_ring); in fealnx_tx_timeout()
1214 (unsigned int) np->rx_ring[i].status); in fealnx_tx_timeout()
1216 printk(KERN_DEBUG " Tx ring %p: ", np->tx_ring); in fealnx_tx_timeout()
1218 printk(KERN_CONT " %4.4x", np->tx_ring[i].status); in fealnx_tx_timeout()
1222 spin_lock_irqsave(&np->lock, flags); in fealnx_tx_timeout()
1228 spin_unlock_irqrestore(&np->lock, flags); in fealnx_tx_timeout()
1239 struct netdev_private *np = netdev_priv(dev); in init_ring() local
1243 np->rx_buf_sz = (dev->mtu <= 1500 ? PKT_BUF_SZ : dev->mtu + 32); in init_ring()
1244 np->cur_rx = &np->rx_ring[0]; in init_ring()
1245 np->lack_rxbuf = np->rx_ring; in init_ring()
1246 np->really_rx_count = 0; in init_ring()
1250 np->rx_ring[i].status = 0; in init_ring()
1251 np->rx_ring[i].control = np->rx_buf_sz << RBSShift; in init_ring()
1252 np->rx_ring[i].next_desc = np->rx_ring_dma + in init_ring()
1254 np->rx_ring[i].next_desc_logical = &np->rx_ring[i + 1]; in init_ring()
1255 np->rx_ring[i].skbuff = NULL; in init_ring()
1259 np->rx_ring[i - 1].next_desc = np->rx_ring_dma; in init_ring()
1260 np->rx_ring[i - 1].next_desc_logical = np->rx_ring; in init_ring()
1264 struct sk_buff *skb = netdev_alloc_skb(dev, np->rx_buf_sz); in init_ring()
1267 np->lack_rxbuf = &np->rx_ring[i]; in init_ring()
1271 ++np->really_rx_count; in init_ring()
1272 np->rx_ring[i].skbuff = skb; in init_ring()
1273 np->rx_ring[i].buffer = pci_map_single(np->pci_dev, skb->data, in init_ring()
1274 np->rx_buf_sz, PCI_DMA_FROMDEVICE); in init_ring()
1275 np->rx_ring[i].status = RXOWN; in init_ring()
1276 np->rx_ring[i].control |= RXIC; in init_ring()
1280 np->cur_tx = &np->tx_ring[0]; in init_ring()
1281 np->cur_tx_copy = &np->tx_ring[0]; in init_ring()
1282 np->really_tx_count = 0; in init_ring()
1283 np->free_tx_count = TX_RING_SIZE; in init_ring()
1286 np->tx_ring[i].status = 0; in init_ring()
1288 np->tx_ring[i].next_desc = np->tx_ring_dma + in init_ring()
1290 np->tx_ring[i].next_desc_logical = &np->tx_ring[i + 1]; in init_ring()
1291 np->tx_ring[i].skbuff = NULL; in init_ring()
1295 np->tx_ring[i - 1].next_desc = np->tx_ring_dma; in init_ring()
1296 np->tx_ring[i - 1].next_desc_logical = &np->tx_ring[0]; in init_ring()
1302 struct netdev_private *np = netdev_priv(dev); in start_tx() local
1305 spin_lock_irqsave(&np->lock, flags); in start_tx()
1307 np->cur_tx_copy->skbuff = skb; in start_tx()
1312 np->cur_tx_copy->buffer = pci_map_single(np->pci_dev, skb->data, in start_tx()
1314 np->cur_tx_copy->control = TXIC | TXLD | TXFD | CRCEnable | PADEnable; in start_tx()
1315 np->cur_tx_copy->control |= (skb->len << PKTSShift); /* pkt size */ in start_tx()
1316 np->cur_tx_copy->control |= (skb->len << TBSShift); /* buffer size */ in start_tx()
1318 if (np->pci_dev->device == 0x891) in start_tx()
1319 np->cur_tx_copy->control |= ETIControl | RetryTxLC; in start_tx()
1320 np->cur_tx_copy->status = TXOWN; in start_tx()
1321 np->cur_tx_copy = np->cur_tx_copy->next_desc_logical; in start_tx()
1322 --np->free_tx_count; in start_tx()
1328 np->cur_tx_copy->buffer = pci_map_single(np->pci_dev, skb->data, in start_tx()
1330 np->cur_tx_copy->control = TXIC | TXFD | CRCEnable | PADEnable; in start_tx()
1331 np->cur_tx_copy->control |= (skb->len << PKTSShift); /* pkt size */ in start_tx()
1332 np->cur_tx_copy->control |= (BPT << TBSShift); /* buffer size */ in start_tx()
1335 next = np->cur_tx_copy->next_desc_logical; in start_tx()
1341 if (np->pci_dev->device == 0x891) in start_tx()
1342 np->cur_tx_copy->control |= ETIControl | RetryTxLC; in start_tx()
1347 np->cur_tx_copy->status = TXOWN; in start_tx()
1349 np->cur_tx_copy = next->next_desc_logical; in start_tx()
1350 np->free_tx_count -= 2; in start_tx()
1352 np->cur_tx_copy->buffer = pci_map_single(np->pci_dev, skb->data, in start_tx()
1354 np->cur_tx_copy->control = TXIC | TXLD | TXFD | CRCEnable | PADEnable; in start_tx()
1355 np->cur_tx_copy->control |= (skb->len << PKTSShift); /* pkt size */ in start_tx()
1356 np->cur_tx_copy->control |= (skb->len << TBSShift); /* buffer size */ in start_tx()
1358 if (np->pci_dev->device == 0x891) in start_tx()
1359 np->cur_tx_copy->control |= ETIControl | RetryTxLC; in start_tx()
1360 np->cur_tx_copy->status = TXOWN; in start_tx()
1361 np->cur_tx_copy = np->cur_tx_copy->next_desc_logical; in start_tx()
1362 --np->free_tx_count; in start_tx()
1366 if (np->free_tx_count < 2) in start_tx()
1368 ++np->really_tx_count; in start_tx()
1369 iowrite32(0, np->mem + TXPDR); in start_tx()
1371 spin_unlock_irqrestore(&np->lock, flags); in start_tx()
1380 struct netdev_private *np = netdev_priv(dev); in reset_tx_descriptors() local
1385 np->cur_tx = &np->tx_ring[0]; in reset_tx_descriptors()
1386 np->cur_tx_copy = &np->tx_ring[0]; in reset_tx_descriptors()
1387 np->really_tx_count = 0; in reset_tx_descriptors()
1388 np->free_tx_count = TX_RING_SIZE; in reset_tx_descriptors()
1391 cur = &np->tx_ring[i]; in reset_tx_descriptors()
1393 pci_unmap_single(np->pci_dev, cur->buffer, in reset_tx_descriptors()
1401 cur->next_desc = np->tx_ring_dma + in reset_tx_descriptors()
1403 cur->next_desc_logical = &np->tx_ring[i + 1]; in reset_tx_descriptors()
1406 np->tx_ring[TX_RING_SIZE - 1].next_desc = np->tx_ring_dma; in reset_tx_descriptors()
1407 np->tx_ring[TX_RING_SIZE - 1].next_desc_logical = &np->tx_ring[0]; in reset_tx_descriptors()
1414 struct netdev_private *np = netdev_priv(dev); in reset_rx_descriptors() local
1415 struct fealnx_desc *cur = np->cur_rx; in reset_rx_descriptors()
1426 iowrite32(np->rx_ring_dma + ((char*)np->cur_rx - (char*)np->rx_ring), in reset_rx_descriptors()
1427 np->mem + RXLBA); in reset_rx_descriptors()
1436 struct netdev_private *np = netdev_priv(dev); in intr_handler() local
1437 void __iomem *ioaddr = np->mem; in intr_handler()
1442 spin_lock(&np->lock); in intr_handler()
1456 if (!(intr_status & np->imrvalue)) in intr_handler()
1487 stop_nic_rx(ioaddr, np->crvalue); in intr_handler()
1489 iowrite32(np->crvalue, ioaddr + TCRRCR); in intr_handler()
1493 while (np->really_tx_count) { in intr_handler()
1494 long tx_status = np->cur_tx->status; in intr_handler()
1495 long tx_control = np->cur_tx->control; in intr_handler()
1500 next = np->cur_tx->next_desc_logical; in intr_handler()
1508 if (!(np->crvalue & CR_W_ENH)) { in intr_handler()
1519 if ((tx_status & HF) && np->mii.full_duplex == 0) in intr_handler()
1537 pci_unmap_single(np->pci_dev, np->cur_tx->buffer, in intr_handler()
1538 np->cur_tx->skbuff->len, PCI_DMA_TODEVICE); in intr_handler()
1539 dev_kfree_skb_irq(np->cur_tx->skbuff); in intr_handler()
1540 np->cur_tx->skbuff = NULL; in intr_handler()
1541 --np->really_tx_count; in intr_handler()
1542 if (np->cur_tx->control & TXLD) { in intr_handler()
1543 np->cur_tx = np->cur_tx->next_desc_logical; in intr_handler()
1544 ++np->free_tx_count; in intr_handler()
1546 np->cur_tx = np->cur_tx->next_desc_logical; in intr_handler()
1547 np->cur_tx = np->cur_tx->next_desc_logical; in intr_handler()
1548 np->free_tx_count += 2; in intr_handler()
1553 if (num_tx && np->free_tx_count >= 2) in intr_handler()
1557 if (np->crvalue & CR_W_ENH) { in intr_handler()
1572 if (!np->reset_timer_armed) { in intr_handler()
1573 np->reset_timer_armed = 1; in intr_handler()
1574 np->reset_timer.expires = RUN_AT(HZ/2); in intr_handler()
1575 add_timer(&np->reset_timer); in intr_handler()
1580 np->crvalue_sv = np->crvalue; in intr_handler()
1581 np->imrvalue_sv = np->imrvalue; in intr_handler()
1582 np->crvalue &= ~(CR_W_TXEN | CR_W_RXEN); /* or simply = 0? */ in intr_handler()
1583 np->imrvalue = 0; in intr_handler()
1602 iowrite32(np->imrvalue, ioaddr + IMR); in intr_handler()
1604 spin_unlock(&np->lock); in intr_handler()
1614 struct netdev_private *np = netdev_priv(dev); in netdev_rx() local
1615 void __iomem *ioaddr = np->mem; in netdev_rx()
1618 while (!(np->cur_rx->status & RXOWN) && np->cur_rx->skbuff) { in netdev_rx()
1619 s32 rx_status = np->cur_rx->status; in netdev_rx()
1621 if (np->really_rx_count == 0) in netdev_rx()
1650 cur = np->cur_rx; in netdev_rx()
1651 while (desno <= np->really_rx_count) { in netdev_rx()
1659 if (desno > np->really_rx_count) in netdev_rx()
1671 if (!np->cur_rx->skbuff) { in netdev_rx()
1676 np->cur_rx->status = RXOWN; in netdev_rx()
1677 np->cur_rx = np->cur_rx->next_desc_logical; in netdev_rx()
1681 stop_nic_rx(ioaddr, np->crvalue); in netdev_rx()
1683 iowrite32(np->crvalue, ioaddr + TCRRCR); in netdev_rx()
1704 pci_dma_sync_single_for_cpu(np->pci_dev, in netdev_rx()
1705 np->cur_rx->buffer, in netdev_rx()
1706 np->rx_buf_sz, in netdev_rx()
1712 np->cur_rx->skbuff->data, pkt_len); in netdev_rx()
1716 np->cur_rx->skbuff->data, pkt_len); in netdev_rx()
1718 pci_dma_sync_single_for_device(np->pci_dev, in netdev_rx()
1719 np->cur_rx->buffer, in netdev_rx()
1720 np->rx_buf_sz, in netdev_rx()
1723 pci_unmap_single(np->pci_dev, in netdev_rx()
1724 np->cur_rx->buffer, in netdev_rx()
1725 np->rx_buf_sz, in netdev_rx()
1727 skb_put(skb = np->cur_rx->skbuff, pkt_len); in netdev_rx()
1728 np->cur_rx->skbuff = NULL; in netdev_rx()
1729 --np->really_rx_count; in netdev_rx()
1737 np->cur_rx = np->cur_rx->next_desc_logical; in netdev_rx()
1749 struct netdev_private *np = netdev_priv(dev); in get_stats() local
1750 void __iomem *ioaddr = np->mem; in get_stats()
1778 struct netdev_private *np = netdev_priv(dev); in __set_rx_mode() local
1779 void __iomem *ioaddr = np->mem; in __set_rx_mode()
1803 stop_nic_rxtx(ioaddr, np->crvalue); in __set_rx_mode()
1807 np->crvalue &= ~CR_W_RXMODEMASK; in __set_rx_mode()
1808 np->crvalue |= rx_mode; in __set_rx_mode()
1809 iowrite32(np->crvalue, ioaddr + TCRRCR); in __set_rx_mode()
1814 struct netdev_private *np = netdev_priv(dev); in netdev_get_drvinfo() local
1818 strlcpy(info->bus_info, pci_name(np->pci_dev), sizeof(info->bus_info)); in netdev_get_drvinfo()
1823 struct netdev_private *np = netdev_priv(dev); in netdev_get_settings() local
1826 spin_lock_irq(&np->lock); in netdev_get_settings()
1827 rc = mii_ethtool_gset(&np->mii, cmd); in netdev_get_settings()
1828 spin_unlock_irq(&np->lock); in netdev_get_settings()
1835 struct netdev_private *np = netdev_priv(dev); in netdev_set_settings() local
1838 spin_lock_irq(&np->lock); in netdev_set_settings()
1839 rc = mii_ethtool_sset(&np->mii, cmd); in netdev_set_settings()
1840 spin_unlock_irq(&np->lock); in netdev_set_settings()
1847 struct netdev_private *np = netdev_priv(dev); in netdev_nway_reset() local
1848 return mii_nway_restart(&np->mii); in netdev_nway_reset()
1853 struct netdev_private *np = netdev_priv(dev); in netdev_get_link() local
1854 return mii_link_ok(&np->mii); in netdev_get_link()
1879 struct netdev_private *np = netdev_priv(dev); in mii_ioctl() local
1885 spin_lock_irq(&np->lock); in mii_ioctl()
1886 rc = generic_mii_ioctl(&np->mii, if_mii(rq), cmd, NULL); in mii_ioctl()
1887 spin_unlock_irq(&np->lock); in mii_ioctl()
1895 struct netdev_private *np = netdev_priv(dev); in netdev_close() local
1896 void __iomem *ioaddr = np->mem; in netdev_close()
1907 del_timer_sync(&np->timer); in netdev_close()
1908 del_timer_sync(&np->reset_timer); in netdev_close()
1910 free_irq(np->pci_dev->irq, dev); in netdev_close()
1914 struct sk_buff *skb = np->rx_ring[i].skbuff; in netdev_close()
1916 np->rx_ring[i].status = 0; in netdev_close()
1918 pci_unmap_single(np->pci_dev, np->rx_ring[i].buffer, in netdev_close()
1919 np->rx_buf_sz, PCI_DMA_FROMDEVICE); in netdev_close()
1921 np->rx_ring[i].skbuff = NULL; in netdev_close()
1926 struct sk_buff *skb = np->tx_ring[i].skbuff; in netdev_close()
1929 pci_unmap_single(np->pci_dev, np->tx_ring[i].buffer, in netdev_close()
1932 np->tx_ring[i].skbuff = NULL; in netdev_close()