Lines Matching refs:ep

182 #define rdb(ep, off)		__raw_readb((ep)->base_addr + (off))  argument
183 #define rdw(ep, off) __raw_readw((ep)->base_addr + (off)) argument
184 #define rdl(ep, off) __raw_readl((ep)->base_addr + (off)) argument
185 #define wrb(ep, off, val) __raw_writeb((val), (ep)->base_addr + (off)) argument
186 #define wrw(ep, off, val) __raw_writew((val), (ep)->base_addr + (off)) argument
187 #define wrl(ep, off, val) __raw_writel((val), (ep)->base_addr + (off)) argument
191 struct ep93xx_priv *ep = netdev_priv(dev); in ep93xx_mdio_read() local
195 wrl(ep, REG_MIICMD, REG_MIICMD_READ | (phy_id << 5) | reg); in ep93xx_mdio_read()
198 if ((rdl(ep, REG_MIISTS) & REG_MIISTS_BUSY) == 0) in ep93xx_mdio_read()
207 data = rdl(ep, REG_MIIDATA); in ep93xx_mdio_read()
215 struct ep93xx_priv *ep = netdev_priv(dev); in ep93xx_mdio_write() local
218 wrl(ep, REG_MIIDATA, data); in ep93xx_mdio_write()
219 wrl(ep, REG_MIICMD, REG_MIICMD_WRITE | (phy_id << 5) | reg); in ep93xx_mdio_write()
222 if ((rdl(ep, REG_MIISTS) & REG_MIISTS_BUSY) == 0) in ep93xx_mdio_write()
233 struct ep93xx_priv *ep = netdev_priv(dev); in ep93xx_rx() local
243 entry = ep->rx_pointer; in ep93xx_rx()
244 rstat = ep->descs->rstat + entry; in ep93xx_rx()
286 struct ep93xx_rdesc *rxd = &ep->descs->rdesc[entry]; in ep93xx_rx()
290 skb_copy_to_linear_data(skb, ep->rx_buf[entry], length); in ep93xx_rx()
306 ep->rx_pointer = (entry + 1) & (RX_QUEUE_ENTRIES - 1); in ep93xx_rx()
313 static int ep93xx_have_more_rx(struct ep93xx_priv *ep) in ep93xx_have_more_rx() argument
315 struct ep93xx_rstat *rstat = ep->descs->rstat + ep->rx_pointer; in ep93xx_have_more_rx()
321 struct ep93xx_priv *ep = container_of(napi, struct ep93xx_priv, napi); in ep93xx_poll() local
322 struct net_device *dev = ep->dev; in ep93xx_poll()
330 spin_lock_irq(&ep->rx_lock); in ep93xx_poll()
332 wrl(ep, REG_INTEN, REG_INTEN_TX | REG_INTEN_RX); in ep93xx_poll()
333 if (ep93xx_have_more_rx(ep)) { in ep93xx_poll()
334 wrl(ep, REG_INTEN, REG_INTEN_TX); in ep93xx_poll()
335 wrl(ep, REG_INTSTSP, REG_INTSTS_RX); in ep93xx_poll()
338 spin_unlock_irq(&ep->rx_lock); in ep93xx_poll()
345 wrw(ep, REG_RXDENQ, rx); in ep93xx_poll()
346 wrw(ep, REG_RXSTSENQ, rx); in ep93xx_poll()
354 struct ep93xx_priv *ep = netdev_priv(dev); in ep93xx_xmit() local
364 entry = ep->tx_pointer; in ep93xx_xmit()
365 ep->tx_pointer = (ep->tx_pointer + 1) & (TX_QUEUE_ENTRIES - 1); in ep93xx_xmit()
367 txd = &ep->descs->tdesc[entry]; in ep93xx_xmit()
372 skb_copy_and_csum_dev(skb, ep->tx_buf[entry]); in ep93xx_xmit()
377 spin_lock_irq(&ep->tx_pending_lock); in ep93xx_xmit()
378 ep->tx_pending++; in ep93xx_xmit()
379 if (ep->tx_pending == TX_QUEUE_ENTRIES) in ep93xx_xmit()
381 spin_unlock_irq(&ep->tx_pending_lock); in ep93xx_xmit()
383 wrl(ep, REG_TXDENQ, 1); in ep93xx_xmit()
390 struct ep93xx_priv *ep = netdev_priv(dev); in ep93xx_tx_complete() local
395 spin_lock(&ep->tx_pending_lock); in ep93xx_tx_complete()
401 entry = ep->tx_clean_pointer; in ep93xx_tx_complete()
402 tstat = ep->descs->tstat + entry; in ep93xx_tx_complete()
416 int length = ep->descs->tdesc[entry].tdesc1 & 0xfff; in ep93xx_tx_complete()
430 ep->tx_clean_pointer = (entry + 1) & (TX_QUEUE_ENTRIES - 1); in ep93xx_tx_complete()
431 if (ep->tx_pending == TX_QUEUE_ENTRIES) in ep93xx_tx_complete()
433 ep->tx_pending--; in ep93xx_tx_complete()
435 spin_unlock(&ep->tx_pending_lock); in ep93xx_tx_complete()
444 struct ep93xx_priv *ep = netdev_priv(dev); in ep93xx_irq() local
447 status = rdl(ep, REG_INTSTSC); in ep93xx_irq()
452 spin_lock(&ep->rx_lock); in ep93xx_irq()
453 if (likely(napi_schedule_prep(&ep->napi))) { in ep93xx_irq()
454 wrl(ep, REG_INTEN, REG_INTEN_TX); in ep93xx_irq()
455 __napi_schedule(&ep->napi); in ep93xx_irq()
457 spin_unlock(&ep->rx_lock); in ep93xx_irq()
466 static void ep93xx_free_buffers(struct ep93xx_priv *ep) in ep93xx_free_buffers() argument
468 struct device *dev = ep->dev->dev.parent; in ep93xx_free_buffers()
474 d = ep->descs->rdesc[i].buf_addr; in ep93xx_free_buffers()
478 kfree(ep->rx_buf[i]); in ep93xx_free_buffers()
484 d = ep->descs->tdesc[i].buf_addr; in ep93xx_free_buffers()
488 kfree(ep->tx_buf[i]); in ep93xx_free_buffers()
491 dma_free_coherent(dev, sizeof(struct ep93xx_descs), ep->descs, in ep93xx_free_buffers()
492 ep->descs_dma_addr); in ep93xx_free_buffers()
495 static int ep93xx_alloc_buffers(struct ep93xx_priv *ep) in ep93xx_alloc_buffers() argument
497 struct device *dev = ep->dev->dev.parent; in ep93xx_alloc_buffers()
500 ep->descs = dma_alloc_coherent(dev, sizeof(struct ep93xx_descs), in ep93xx_alloc_buffers()
501 &ep->descs_dma_addr, GFP_KERNEL); in ep93xx_alloc_buffers()
502 if (ep->descs == NULL) in ep93xx_alloc_buffers()
519 ep->rx_buf[i] = buf; in ep93xx_alloc_buffers()
520 ep->descs->rdesc[i].buf_addr = d; in ep93xx_alloc_buffers()
521 ep->descs->rdesc[i].rdesc1 = (i << 16) | PKT_BUF_SIZE; in ep93xx_alloc_buffers()
538 ep->tx_buf[i] = buf; in ep93xx_alloc_buffers()
539 ep->descs->tdesc[i].buf_addr = d; in ep93xx_alloc_buffers()
545 ep93xx_free_buffers(ep); in ep93xx_alloc_buffers()
551 struct ep93xx_priv *ep = netdev_priv(dev); in ep93xx_start_hw() local
555 wrl(ep, REG_SELFCTL, REG_SELFCTL_RESET); in ep93xx_start_hw()
557 if ((rdl(ep, REG_SELFCTL) & REG_SELFCTL_RESET) == 0) in ep93xx_start_hw()
567 wrl(ep, REG_SELFCTL, ((ep->mdc_divisor - 1) << 9)); in ep93xx_start_hw()
570 if ((ep93xx_mdio_read(dev, ep->mii.phy_id, MII_BMSR) & 0x0040) != 0) in ep93xx_start_hw()
571 wrl(ep, REG_SELFCTL, ((ep->mdc_divisor - 1) << 9) | (1 << 8)); in ep93xx_start_hw()
574 addr = ep->descs_dma_addr + offsetof(struct ep93xx_descs, rdesc); in ep93xx_start_hw()
575 wrl(ep, REG_RXDQBADD, addr); in ep93xx_start_hw()
576 wrl(ep, REG_RXDCURADD, addr); in ep93xx_start_hw()
577 wrw(ep, REG_RXDQBLEN, RX_QUEUE_ENTRIES * sizeof(struct ep93xx_rdesc)); in ep93xx_start_hw()
580 addr = ep->descs_dma_addr + offsetof(struct ep93xx_descs, rstat); in ep93xx_start_hw()
581 wrl(ep, REG_RXSTSQBADD, addr); in ep93xx_start_hw()
582 wrl(ep, REG_RXSTSQCURADD, addr); in ep93xx_start_hw()
583 wrw(ep, REG_RXSTSQBLEN, RX_QUEUE_ENTRIES * sizeof(struct ep93xx_rstat)); in ep93xx_start_hw()
586 addr = ep->descs_dma_addr + offsetof(struct ep93xx_descs, tdesc); in ep93xx_start_hw()
587 wrl(ep, REG_TXDQBADD, addr); in ep93xx_start_hw()
588 wrl(ep, REG_TXDQCURADD, addr); in ep93xx_start_hw()
589 wrw(ep, REG_TXDQBLEN, TX_QUEUE_ENTRIES * sizeof(struct ep93xx_tdesc)); in ep93xx_start_hw()
592 addr = ep->descs_dma_addr + offsetof(struct ep93xx_descs, tstat); in ep93xx_start_hw()
593 wrl(ep, REG_TXSTSQBADD, addr); in ep93xx_start_hw()
594 wrl(ep, REG_TXSTSQCURADD, addr); in ep93xx_start_hw()
595 wrw(ep, REG_TXSTSQBLEN, TX_QUEUE_ENTRIES * sizeof(struct ep93xx_tstat)); in ep93xx_start_hw()
597 wrl(ep, REG_BMCTL, REG_BMCTL_ENABLE_TX | REG_BMCTL_ENABLE_RX); in ep93xx_start_hw()
598 wrl(ep, REG_INTEN, REG_INTEN_TX | REG_INTEN_RX); in ep93xx_start_hw()
599 wrl(ep, REG_GIINTMSK, 0); in ep93xx_start_hw()
602 if ((rdl(ep, REG_BMSTS) & REG_BMSTS_RX_ACTIVE) != 0) in ep93xx_start_hw()
612 wrl(ep, REG_RXDENQ, RX_QUEUE_ENTRIES); in ep93xx_start_hw()
613 wrl(ep, REG_RXSTSENQ, RX_QUEUE_ENTRIES); in ep93xx_start_hw()
615 wrb(ep, REG_INDAD0, dev->dev_addr[0]); in ep93xx_start_hw()
616 wrb(ep, REG_INDAD1, dev->dev_addr[1]); in ep93xx_start_hw()
617 wrb(ep, REG_INDAD2, dev->dev_addr[2]); in ep93xx_start_hw()
618 wrb(ep, REG_INDAD3, dev->dev_addr[3]); in ep93xx_start_hw()
619 wrb(ep, REG_INDAD4, dev->dev_addr[4]); in ep93xx_start_hw()
620 wrb(ep, REG_INDAD5, dev->dev_addr[5]); in ep93xx_start_hw()
621 wrl(ep, REG_AFP, 0); in ep93xx_start_hw()
623 wrl(ep, REG_MAXFRMLEN, (MAX_PKT_SIZE << 16) | MAX_PKT_SIZE); in ep93xx_start_hw()
625 wrl(ep, REG_RXCTL, REG_RXCTL_DEFAULT); in ep93xx_start_hw()
626 wrl(ep, REG_TXCTL, REG_TXCTL_ENABLE); in ep93xx_start_hw()
633 struct ep93xx_priv *ep = netdev_priv(dev); in ep93xx_stop_hw() local
636 wrl(ep, REG_SELFCTL, REG_SELFCTL_RESET); in ep93xx_stop_hw()
638 if ((rdl(ep, REG_SELFCTL) & REG_SELFCTL_RESET) == 0) in ep93xx_stop_hw()
649 struct ep93xx_priv *ep = netdev_priv(dev); in ep93xx_open() local
652 if (ep93xx_alloc_buffers(ep)) in ep93xx_open()
655 napi_enable(&ep->napi); in ep93xx_open()
658 napi_disable(&ep->napi); in ep93xx_open()
659 ep93xx_free_buffers(ep); in ep93xx_open()
663 spin_lock_init(&ep->rx_lock); in ep93xx_open()
664 ep->rx_pointer = 0; in ep93xx_open()
665 ep->tx_clean_pointer = 0; in ep93xx_open()
666 ep->tx_pointer = 0; in ep93xx_open()
667 spin_lock_init(&ep->tx_pending_lock); in ep93xx_open()
668 ep->tx_pending = 0; in ep93xx_open()
670 err = request_irq(ep->irq, ep93xx_irq, IRQF_SHARED, dev->name, dev); in ep93xx_open()
672 napi_disable(&ep->napi); in ep93xx_open()
674 ep93xx_free_buffers(ep); in ep93xx_open()
678 wrl(ep, REG_GIINTMSK, REG_GIINTMSK_ENABLE); in ep93xx_open()
687 struct ep93xx_priv *ep = netdev_priv(dev); in ep93xx_close() local
689 napi_disable(&ep->napi); in ep93xx_close()
692 wrl(ep, REG_GIINTMSK, 0); in ep93xx_close()
693 free_irq(ep->irq, dev); in ep93xx_close()
695 ep93xx_free_buffers(ep); in ep93xx_close()
702 struct ep93xx_priv *ep = netdev_priv(dev); in ep93xx_ioctl() local
705 return generic_mii_ioctl(&ep->mii, data, cmd, NULL); in ep93xx_ioctl()
716 struct ep93xx_priv *ep = netdev_priv(dev); in ep93xx_get_settings() local
717 return mii_ethtool_gset(&ep->mii, cmd); in ep93xx_get_settings()
722 struct ep93xx_priv *ep = netdev_priv(dev); in ep93xx_set_settings() local
723 return mii_ethtool_sset(&ep->mii, cmd); in ep93xx_set_settings()
728 struct ep93xx_priv *ep = netdev_priv(dev); in ep93xx_nway_reset() local
729 return mii_nway_restart(&ep->mii); in ep93xx_nway_reset()
734 struct ep93xx_priv *ep = netdev_priv(dev); in ep93xx_get_link() local
735 return mii_link_ok(&ep->mii); in ep93xx_get_link()
778 struct ep93xx_priv *ep; in ep93xx_eth_remove() local
784 ep = netdev_priv(dev); in ep93xx_eth_remove()
788 ep93xx_free_buffers(ep); in ep93xx_eth_remove()
790 if (ep->base_addr != NULL) in ep93xx_eth_remove()
791 iounmap(ep->base_addr); in ep93xx_eth_remove()
793 if (ep->res != NULL) { in ep93xx_eth_remove()
794 release_resource(ep->res); in ep93xx_eth_remove()
795 kfree(ep->res); in ep93xx_eth_remove()
807 struct ep93xx_priv *ep; in ep93xx_eth_probe() local
826 ep = netdev_priv(dev); in ep93xx_eth_probe()
827 ep->dev = dev; in ep93xx_eth_probe()
829 netif_napi_add(dev, &ep->napi, ep93xx_poll, 64); in ep93xx_eth_probe()
833 ep->res = request_mem_region(mem->start, resource_size(mem), in ep93xx_eth_probe()
835 if (ep->res == NULL) { in ep93xx_eth_probe()
841 ep->base_addr = ioremap(mem->start, resource_size(mem)); in ep93xx_eth_probe()
842 if (ep->base_addr == NULL) { in ep93xx_eth_probe()
847 ep->irq = irq; in ep93xx_eth_probe()
849 ep->mii.phy_id = data->phy_id; in ep93xx_eth_probe()
850 ep->mii.phy_id_mask = 0x1f; in ep93xx_eth_probe()
851 ep->mii.reg_num_mask = 0x1f; in ep93xx_eth_probe()
852 ep->mii.dev = dev; in ep93xx_eth_probe()
853 ep->mii.mdio_read = ep93xx_mdio_read; in ep93xx_eth_probe()
854 ep->mii.mdio_write = ep93xx_mdio_write; in ep93xx_eth_probe()
855 ep->mdc_divisor = 40; /* Max HCLK 100 MHz, min MDIO clk 2.5 MHz. */ in ep93xx_eth_probe()
867 dev->name, ep->irq, dev->dev_addr); in ep93xx_eth_probe()