Lines Matching refs:np
60 static void dl2k_enable_int(struct netdev_private *np) in dl2k_enable_int() argument
62 void __iomem *ioaddr = np->ioaddr; in dl2k_enable_int()
117 struct netdev_private *np; in rio_probe1() local
142 dev = alloc_etherdev (sizeof (*np)); in rio_probe1()
147 np = netdev_priv(dev); in rio_probe1()
153 np->eeprom_addr = ioaddr; in rio_probe1()
161 np->ioaddr = ioaddr; in rio_probe1()
162 np->chip_id = chip_idx; in rio_probe1()
163 np->pdev = pdev; in rio_probe1()
164 spin_lock_init (&np->tx_lock); in rio_probe1()
165 spin_lock_init (&np->rx_lock); in rio_probe1()
168 np->an_enable = 1; in rio_probe1()
169 np->tx_coalesce = 1; in rio_probe1()
172 np->an_enable = 0; in rio_probe1()
176 np->an_enable = 2; in rio_probe1()
179 np->speed = 100; in rio_probe1()
180 np->full_duplex = 1; in rio_probe1()
183 np->speed = 100; in rio_probe1()
184 np->full_duplex = 0; in rio_probe1()
187 np->speed = 10; in rio_probe1()
188 np->full_duplex = 1; in rio_probe1()
191 np->speed = 10; in rio_probe1()
192 np->full_duplex = 0; in rio_probe1()
195 np->speed=1000; in rio_probe1()
196 np->full_duplex=1; in rio_probe1()
199 np->speed = 1000; in rio_probe1()
200 np->full_duplex = 0; in rio_probe1()
202 np->an_enable = 1; in rio_probe1()
206 np->jumbo = 1; in rio_probe1()
209 np->jumbo = 0; in rio_probe1()
213 np->vlan = (vlan[card_idx] > 0 && vlan[card_idx] < 4096) ? in rio_probe1()
216 np->rx_coalesce = rx_coalesce; in rio_probe1()
217 np->rx_timeout = rx_timeout; in rio_probe1()
218 np->coalesce = 1; in rio_probe1()
220 np->tx_flow = (tx_flow == 0) ? 0 : 1; in rio_probe1()
221 np->rx_flow = (rx_flow == 0) ? 0 : 1; in rio_probe1()
239 np->tx_ring = ring_space; in rio_probe1()
240 np->tx_ring_dma = ring_dma; in rio_probe1()
245 np->rx_ring = ring_space; in rio_probe1()
246 np->rx_ring_dma = ring_dma; in rio_probe1()
257 np->phy_media = (dr16(ASICCtrl) & PhyMedia) ? 1 : 0; in rio_probe1()
258 np->link_status = 0; in rio_probe1()
260 if (np->phy_media) { in rio_probe1()
262 if (np->an_enable == 2) { in rio_probe1()
263 np->an_enable = 1; in rio_probe1()
269 if (np->speed == 1000) in rio_probe1()
270 np->an_enable = 1; in rio_probe1()
281 dev->name, np->name, dev->dev_addr, irq); in rio_probe1()
285 if (np->coalesce) in rio_probe1()
289 np->rx_coalesce, np->rx_timeout*640); in rio_probe1()
290 if (np->vlan) in rio_probe1()
291 printk(KERN_INFO "vlan(id):\t%d\n", np->vlan); in rio_probe1()
295 pci_free_consistent (pdev, RX_TOTAL_SIZE, np->rx_ring, np->rx_ring_dma); in rio_probe1()
297 pci_free_consistent (pdev, TX_TOTAL_SIZE, np->tx_ring, np->tx_ring_dma); in rio_probe1()
300 pci_iounmap(pdev, np->ioaddr); in rio_probe1()
302 pci_iounmap(pdev, np->eeprom_addr); in rio_probe1()
315 struct netdev_private *np = netdev_priv(dev); in find_miiphy() local
317 np = netdev_priv(dev); in find_miiphy()
318 np->phy_addr = 1; in find_miiphy()
323 np->phy_addr = i; in find_miiphy()
337 struct netdev_private *np = netdev_priv(dev); in parse_eeprom() local
338 void __iomem *ioaddr = np->ioaddr; in parse_eeprom()
348 ((__le16 *) sromdata)[i] = cpu_to_le16(read_eeprom(np, i)); in parse_eeprom()
350 if (np->pdev->vendor == PCI_VENDOR_ID_DLINK) { /* D-Link Only */ in parse_eeprom()
364 if (np->pdev->vendor != PCI_VENDOR_ID_DLINK) { in parse_eeprom()
384 np->duplex_polarity = psib[i]; in parse_eeprom()
388 np->wake_polarity = psib[i]; in parse_eeprom()
392 memcpy (np->name, &(psib[i]), j); in parse_eeprom()
412 struct netdev_private *np = netdev_priv(dev); in rio_open() local
413 void __iomem *ioaddr = np->ioaddr; in rio_open()
414 const int irq = np->pdev->irq; in rio_open()
431 if (np->jumbo != 0) in rio_open()
441 if (np->coalesce) { in rio_open()
442 dw32(RxDMAIntCtrl, np->rx_coalesce | np->rx_timeout << 16); in rio_open()
454 if (np->vlan) { in rio_open()
458 dw16(VLANId, np->vlan); in rio_open()
460 dw32(VLANTag, 0x8100 << 16 | np->vlan); in rio_open()
466 init_timer (&np->timer); in rio_open()
467 np->timer.expires = jiffies + 1*HZ; in rio_open()
468 np->timer.data = (unsigned long) dev; in rio_open()
469 np->timer.function = rio_timer; in rio_open()
470 add_timer (&np->timer); in rio_open()
476 macctrl |= (np->vlan) ? AutoVLANuntagging : 0; in rio_open()
477 macctrl |= (np->full_duplex) ? DuplexSelect : 0; in rio_open()
478 macctrl |= (np->tx_flow) ? TxFlowControlEnable : 0; in rio_open()
479 macctrl |= (np->rx_flow) ? RxFlowControlEnable : 0; in rio_open()
484 dl2k_enable_int(np); in rio_open()
492 struct netdev_private *np = netdev_priv(dev); in rio_timer() local
497 spin_lock_irqsave(&np->rx_lock, flags); in rio_timer()
499 if (np->cur_rx - np->old_rx >= RX_RING_SIZE) { in rio_timer()
502 for (; np->cur_rx - np->old_rx > 0; np->old_rx++) { in rio_timer()
504 entry = np->old_rx % RX_RING_SIZE; in rio_timer()
506 if (np->rx_skbuff[entry] == NULL) { in rio_timer()
508 np->rx_buf_sz); in rio_timer()
510 np->rx_ring[entry].fraginfo = 0; in rio_timer()
516 np->rx_skbuff[entry] = skb; in rio_timer()
517 np->rx_ring[entry].fraginfo = in rio_timer()
519 (np->pdev, skb->data, np->rx_buf_sz, in rio_timer()
522 np->rx_ring[entry].fraginfo |= in rio_timer()
523 cpu_to_le64((u64)np->rx_buf_sz << 48); in rio_timer()
524 np->rx_ring[entry].status = 0; in rio_timer()
527 spin_unlock_irqrestore (&np->rx_lock, flags); in rio_timer()
528 np->timer.expires = jiffies + next_tick; in rio_timer()
529 add_timer(&np->timer); in rio_timer()
535 struct netdev_private *np = netdev_priv(dev); in rio_tx_timeout() local
536 void __iomem *ioaddr = np->ioaddr; in rio_tx_timeout()
549 struct netdev_private *np = netdev_priv(dev); in alloc_list() local
550 void __iomem *ioaddr = np->ioaddr; in alloc_list()
553 np->cur_rx = np->cur_tx = 0; in alloc_list()
554 np->old_rx = np->old_tx = 0; in alloc_list()
555 np->rx_buf_sz = (dev->mtu <= 1500 ? PACKET_SIZE : dev->mtu + 32); in alloc_list()
559 np->tx_skbuff[i] = NULL; in alloc_list()
560 np->tx_ring[i].status = cpu_to_le64 (TFDDone); in alloc_list()
561 np->tx_ring[i].next_desc = cpu_to_le64 (np->tx_ring_dma + in alloc_list()
568 np->rx_ring[i].next_desc = cpu_to_le64 (np->rx_ring_dma + in alloc_list()
571 np->rx_ring[i].status = 0; in alloc_list()
572 np->rx_ring[i].fraginfo = 0; in alloc_list()
573 np->rx_skbuff[i] = NULL; in alloc_list()
581 skb = netdev_alloc_skb_ip_align(dev, np->rx_buf_sz); in alloc_list()
582 np->rx_skbuff[i] = skb; in alloc_list()
587 np->rx_ring[i].fraginfo = in alloc_list()
589 np->pdev, skb->data, np->rx_buf_sz, in alloc_list()
591 np->rx_ring[i].fraginfo |= cpu_to_le64((u64)np->rx_buf_sz << 48); in alloc_list()
595 dw32(RFDListPtr0, np->rx_ring_dma); in alloc_list()
602 struct netdev_private *np = netdev_priv(dev); in start_xmit() local
603 void __iomem *ioaddr = np->ioaddr; in start_xmit()
608 if (np->link_status == 0) { /* Link Down */ in start_xmit()
612 entry = np->cur_tx % TX_RING_SIZE; in start_xmit()
613 np->tx_skbuff[entry] = skb; in start_xmit()
614 txdesc = &np->tx_ring[entry]; in start_xmit()
623 if (np->vlan) { in start_xmit()
625 ((u64)np->vlan << 32) | in start_xmit()
628 txdesc->fraginfo = cpu_to_le64 (pci_map_single (np->pdev, skb->data, in start_xmit()
635 if (entry % np->tx_coalesce == 0 || np->speed == 10) in start_xmit()
649 np->cur_tx = (np->cur_tx + 1) % TX_RING_SIZE; in start_xmit()
650 if ((np->cur_tx - np->old_tx + TX_RING_SIZE) % TX_RING_SIZE in start_xmit()
651 < TX_QUEUE_LEN - 1 && np->speed != 10) { in start_xmit()
659 dw32(TFDListPtr0, np->tx_ring_dma + in start_xmit()
671 struct netdev_private *np = netdev_priv(dev); in rio_interrupt() local
672 void __iomem *ioaddr = np->ioaddr; in rio_interrupt()
702 if (np->cur_tx != np->old_tx) in rio_interrupt()
715 struct netdev_private *np = netdev_priv(dev); in rio_free_tx() local
716 int entry = np->old_tx % TX_RING_SIZE; in rio_free_tx()
721 spin_lock(&np->tx_lock); in rio_free_tx()
723 spin_lock_irqsave(&np->tx_lock, flag); in rio_free_tx()
726 while (entry != np->cur_tx) { in rio_free_tx()
729 if (!(np->tx_ring[entry].status & cpu_to_le64(TFDDone))) in rio_free_tx()
731 skb = np->tx_skbuff[entry]; in rio_free_tx()
732 pci_unmap_single (np->pdev, in rio_free_tx()
733 desc_to_dma(&np->tx_ring[entry]), in rio_free_tx()
740 np->tx_skbuff[entry] = NULL; in rio_free_tx()
745 spin_unlock(&np->tx_lock); in rio_free_tx()
747 spin_unlock_irqrestore(&np->tx_lock, flag); in rio_free_tx()
748 np->old_tx = entry; in rio_free_tx()
754 ((np->cur_tx - np->old_tx + TX_RING_SIZE) % TX_RING_SIZE in rio_free_tx()
755 < TX_QUEUE_LEN - 1 || np->speed == 10)) { in rio_free_tx()
763 struct netdev_private *np = netdev_priv(dev); in tx_error() local
764 void __iomem *ioaddr = np->ioaddr; in tx_error()
771 np->stats.tx_errors++; in tx_error()
774 np->stats.tx_fifo_errors++; in tx_error()
787 dw32(TFDListPtr0, np->tx_ring_dma + in tx_error()
788 np->old_tx * sizeof (struct netdev_desc)); in tx_error()
795 np->stats.tx_fifo_errors++; in tx_error()
809 np->stats.collisions16++; in tx_error()
812 np->stats.collisions++; in tx_error()
821 struct netdev_private *np = netdev_priv(dev); in receive_packet() local
822 int entry = np->cur_rx % RX_RING_SIZE; in receive_packet()
827 struct netdev_desc *desc = &np->rx_ring[entry]; in receive_packet()
843 np->stats.rx_errors++; in receive_packet()
845 np->stats.rx_length_errors++; in receive_packet()
847 np->stats.rx_crc_errors++; in receive_packet()
848 if (frame_status & RxAlignmentError && np->speed != 1000) in receive_packet()
849 np->stats.rx_frame_errors++; in receive_packet()
851 np->stats.rx_fifo_errors++; in receive_packet()
857 pci_unmap_single (np->pdev, in receive_packet()
859 np->rx_buf_sz, in receive_packet()
861 skb_put (skb = np->rx_skbuff[entry], pkt_len); in receive_packet()
862 np->rx_skbuff[entry] = NULL; in receive_packet()
864 pci_dma_sync_single_for_cpu(np->pdev, in receive_packet()
866 np->rx_buf_sz, in receive_packet()
869 np->rx_skbuff[entry]->data, in receive_packet()
872 pci_dma_sync_single_for_device(np->pdev, in receive_packet()
874 np->rx_buf_sz, in receive_packet()
880 if (np->pdev->pci_rev_id >= 0x0c && in receive_packet()
889 spin_lock(&np->rx_lock); in receive_packet()
890 np->cur_rx = entry; in receive_packet()
892 entry = np->old_rx; in receive_packet()
893 while (entry != np->cur_rx) { in receive_packet()
896 if (np->rx_skbuff[entry] == NULL) { in receive_packet()
897 skb = netdev_alloc_skb_ip_align(dev, np->rx_buf_sz); in receive_packet()
899 np->rx_ring[entry].fraginfo = 0; in receive_packet()
906 np->rx_skbuff[entry] = skb; in receive_packet()
907 np->rx_ring[entry].fraginfo = in receive_packet()
909 (np->pdev, skb->data, np->rx_buf_sz, in receive_packet()
912 np->rx_ring[entry].fraginfo |= in receive_packet()
913 cpu_to_le64((u64)np->rx_buf_sz << 48); in receive_packet()
914 np->rx_ring[entry].status = 0; in receive_packet()
917 np->old_rx = entry; in receive_packet()
918 spin_unlock(&np->rx_lock); in receive_packet()
925 struct netdev_private *np = netdev_priv(dev); in rio_error() local
926 void __iomem *ioaddr = np->ioaddr; in rio_error()
933 if (np->phy_media) in rio_error()
937 if (np->speed == 1000) in rio_error()
938 np->tx_coalesce = tx_coalesce; in rio_error()
940 np->tx_coalesce = 1; in rio_error()
942 macctrl |= (np->vlan) ? AutoVLANuntagging : 0; in rio_error()
943 macctrl |= (np->full_duplex) ? DuplexSelect : 0; in rio_error()
944 macctrl |= (np->tx_flow) ? in rio_error()
946 macctrl |= (np->rx_flow) ? in rio_error()
949 np->link_status = 1; in rio_error()
953 np->link_status = 0; in rio_error()
976 struct netdev_private *np = netdev_priv(dev); in get_stats() local
977 void __iomem *ioaddr = np->ioaddr; in get_stats()
986 np->stats.rx_packets += dr32(FramesRcvOk); in get_stats()
987 np->stats.tx_packets += dr32(FramesXmtOk); in get_stats()
988 np->stats.rx_bytes += dr32(OctetRcvOk); in get_stats()
989 np->stats.tx_bytes += dr32(OctetXmtOk); in get_stats()
991 np->stats.multicast = dr32(McstFramesRcvdOk); in get_stats()
992 np->stats.collisions += dr32(SingleColFrames) in get_stats()
997 np->stats.tx_aborted_errors += stat_reg; in get_stats()
998 np->stats.tx_errors += stat_reg; in get_stats()
1001 np->stats.tx_carrier_errors += stat_reg; in get_stats()
1002 np->stats.tx_errors += stat_reg; in get_stats()
1032 return &np->stats; in get_stats()
1038 struct netdev_private *np = netdev_priv(dev); in clear_stats() local
1039 void __iomem *ioaddr = np->ioaddr; in clear_stats()
1094 struct netdev_private *np = netdev_priv(dev); in change_mtu() local
1095 int max = (np->jumbo) ? MAX_JUMBO : 1536; in change_mtu()
1109 struct netdev_private *np = netdev_priv(dev); in set_multicast() local
1110 void __iomem *ioaddr = np->ioaddr; in set_multicast()
1143 if (np->vlan) { in set_multicast()
1155 struct netdev_private *np = netdev_priv(dev); in rio_get_drvinfo() local
1159 strlcpy(info->bus_info, pci_name(np->pdev), sizeof(info->bus_info)); in rio_get_drvinfo()
1164 struct netdev_private *np = netdev_priv(dev); in rio_get_settings() local
1165 if (np->phy_media) { in rio_get_settings()
1184 if ( np->link_status ) { in rio_get_settings()
1185 ethtool_cmd_speed_set(cmd, np->speed); in rio_get_settings()
1186 cmd->duplex = np->full_duplex ? DUPLEX_FULL : DUPLEX_HALF; in rio_get_settings()
1191 if ( np->an_enable) in rio_get_settings()
1196 cmd->phy_address = np->phy_addr; in rio_get_settings()
1202 struct netdev_private *np = netdev_priv(dev); in rio_set_settings() local
1205 if (np->an_enable) in rio_set_settings()
1208 np->an_enable = 1; in rio_set_settings()
1213 np->an_enable = 0; in rio_set_settings()
1214 if (np->speed == 1000) { in rio_set_settings()
1221 np->speed = 10; in rio_set_settings()
1222 np->full_duplex = (cmd->duplex == DUPLEX_FULL); in rio_set_settings()
1225 np->speed = 100; in rio_set_settings()
1226 np->full_duplex = (cmd->duplex == DUPLEX_FULL); in rio_set_settings()
1239 struct netdev_private *np = netdev_priv(dev); in rio_get_link() local
1240 return np->link_status; in rio_get_link()
1254 struct netdev_private *np = netdev_priv(dev); in rio_ioctl() local
1257 phy_addr = np->phy_addr; in rio_ioctl()
1280 static int read_eeprom(struct netdev_private *np, int eep_addr) in read_eeprom() argument
1282 void __iomem *ioaddr = np->eeprom_addr; in read_eeprom()
1302 struct netdev_private *np = netdev_priv(dev); in mii_sendbit() local
1303 void __iomem *ioaddr = np->ioaddr; in mii_sendbit()
1315 struct netdev_private *np = netdev_priv(dev); in mii_getbit() local
1316 void __iomem *ioaddr = np->ioaddr; in mii_getbit()
1385 struct netdev_private *np; in mii_wait_link() local
1387 np = netdev_priv(dev); in mii_wait_link()
1388 phy_addr = np->phy_addr; in mii_wait_link()
1406 struct netdev_private *np; in mii_get_media() local
1408 np = netdev_priv(dev); in mii_get_media()
1409 phy_addr = np->phy_addr; in mii_get_media()
1412 if (np->an_enable) { in mii_get_media()
1422 np->speed = 1000; in mii_get_media()
1423 np->full_duplex = 1; in mii_get_media()
1426 np->speed = 1000; in mii_get_media()
1427 np->full_duplex = 0; in mii_get_media()
1430 np->speed = 100; in mii_get_media()
1431 np->full_duplex = 1; in mii_get_media()
1434 np->speed = 100; in mii_get_media()
1435 np->full_duplex = 0; in mii_get_media()
1438 np->speed = 10; in mii_get_media()
1439 np->full_duplex = 1; in mii_get_media()
1442 np->speed = 10; in mii_get_media()
1443 np->full_duplex = 0; in mii_get_media()
1447 np->tx_flow &= 1; in mii_get_media()
1448 np->rx_flow &= 1; in mii_get_media()
1450 np->tx_flow = 0; in mii_get_media()
1451 np->rx_flow &= 1; in mii_get_media()
1472 if (np->tx_flow) in mii_get_media()
1476 if (np->rx_flow) in mii_get_media()
1492 struct netdev_private *np; in mii_set_media() local
1493 np = netdev_priv(dev); in mii_set_media()
1494 phy_addr = np->phy_addr; in mii_set_media()
1497 if (np->an_enable) { in mii_set_media()
1549 if (np->speed == 100) { in mii_set_media()
1552 } else if (np->speed == 10) { in mii_set_media()
1555 if (np->full_duplex) { in mii_set_media()
1579 struct netdev_private *np; in mii_get_media_pcs() local
1581 np = netdev_priv(dev); in mii_get_media_pcs()
1582 phy_addr = np->phy_addr; in mii_get_media_pcs()
1585 if (np->an_enable) { in mii_get_media_pcs()
1592 np->speed = 1000; in mii_get_media_pcs()
1595 np->full_duplex = 1; in mii_get_media_pcs()
1598 np->full_duplex = 0; in mii_get_media_pcs()
1601 np->tx_flow &= 1; in mii_get_media_pcs()
1602 np->rx_flow &= 1; in mii_get_media_pcs()
1604 np->tx_flow = 0; in mii_get_media_pcs()
1605 np->rx_flow &= 1; in mii_get_media_pcs()
1617 if (np->tx_flow) in mii_get_media_pcs()
1621 if (np->rx_flow) in mii_get_media_pcs()
1636 struct netdev_private *np; in mii_set_media_pcs() local
1637 np = netdev_priv(dev); in mii_set_media_pcs()
1638 phy_addr = np->phy_addr; in mii_set_media_pcs()
1641 if (np->an_enable) { in mii_set_media_pcs()
1665 if (np->full_duplex) { in mii_set_media_pcs()
1685 struct netdev_private *np = netdev_priv(dev); in rio_close() local
1686 void __iomem *ioaddr = np->ioaddr; in rio_close()
1688 struct pci_dev *pdev = np->pdev; in rio_close()
1701 del_timer_sync (&np->timer); in rio_close()
1705 skb = np->rx_skbuff[i]; in rio_close()
1707 pci_unmap_single(pdev, desc_to_dma(&np->rx_ring[i]), in rio_close()
1710 np->rx_skbuff[i] = NULL; in rio_close()
1712 np->rx_ring[i].status = 0; in rio_close()
1713 np->rx_ring[i].fraginfo = 0; in rio_close()
1716 skb = np->tx_skbuff[i]; in rio_close()
1718 pci_unmap_single(pdev, desc_to_dma(&np->tx_ring[i]), in rio_close()
1721 np->tx_skbuff[i] = NULL; in rio_close()
1734 struct netdev_private *np = netdev_priv(dev); in rio_remove1() local
1737 pci_free_consistent (pdev, RX_TOTAL_SIZE, np->rx_ring, in rio_remove1()
1738 np->rx_ring_dma); in rio_remove1()
1739 pci_free_consistent (pdev, TX_TOTAL_SIZE, np->tx_ring, in rio_remove1()
1740 np->tx_ring_dma); in rio_remove1()
1742 pci_iounmap(pdev, np->ioaddr); in rio_remove1()
1744 pci_iounmap(pdev, np->eeprom_addr); in rio_remove1()