Lines Matching refs:qdev

104 static int ql_sem_spinlock(struct ql3_adapter *qdev,  in ql_sem_spinlock()  argument
108 qdev->mem_map_registers; in ql_sem_spinlock()
123 static void ql_sem_unlock(struct ql3_adapter *qdev, u32 sem_mask) in ql_sem_unlock() argument
126 qdev->mem_map_registers; in ql_sem_unlock()
131 static int ql_sem_lock(struct ql3_adapter *qdev, u32 sem_mask, u32 sem_bits) in ql_sem_lock() argument
134 qdev->mem_map_registers; in ql_sem_lock()
145 static int ql_wait_for_drvr_lock(struct ql3_adapter *qdev) in ql_wait_for_drvr_lock() argument
150 if (ql_sem_lock(qdev, in ql_wait_for_drvr_lock()
152 (QL_RESOURCE_BITS_BASE_CODE | (qdev->mac_index) in ql_wait_for_drvr_lock()
154 netdev_printk(KERN_DEBUG, qdev->ndev, in ql_wait_for_drvr_lock()
161 netdev_err(qdev->ndev, "Timed out waiting for driver lock...\n"); in ql_wait_for_drvr_lock()
165 static void ql_set_register_page(struct ql3_adapter *qdev, u32 page) in ql_set_register_page() argument
168 qdev->mem_map_registers; in ql_set_register_page()
173 qdev->current_page = page; in ql_set_register_page()
176 static u32 ql_read_common_reg_l(struct ql3_adapter *qdev, u32 __iomem *reg) in ql_read_common_reg_l() argument
181 spin_lock_irqsave(&qdev->hw_lock, hw_flags); in ql_read_common_reg_l()
183 spin_unlock_irqrestore(&qdev->hw_lock, hw_flags); in ql_read_common_reg_l()
188 static u32 ql_read_common_reg(struct ql3_adapter *qdev, u32 __iomem *reg) in ql_read_common_reg() argument
193 static u32 ql_read_page0_reg_l(struct ql3_adapter *qdev, u32 __iomem *reg) in ql_read_page0_reg_l() argument
198 spin_lock_irqsave(&qdev->hw_lock, hw_flags); in ql_read_page0_reg_l()
200 if (qdev->current_page != 0) in ql_read_page0_reg_l()
201 ql_set_register_page(qdev, 0); in ql_read_page0_reg_l()
204 spin_unlock_irqrestore(&qdev->hw_lock, hw_flags); in ql_read_page0_reg_l()
208 static u32 ql_read_page0_reg(struct ql3_adapter *qdev, u32 __iomem *reg) in ql_read_page0_reg() argument
210 if (qdev->current_page != 0) in ql_read_page0_reg()
211 ql_set_register_page(qdev, 0); in ql_read_page0_reg()
215 static void ql_write_common_reg_l(struct ql3_adapter *qdev, in ql_write_common_reg_l() argument
220 spin_lock_irqsave(&qdev->hw_lock, hw_flags); in ql_write_common_reg_l()
223 spin_unlock_irqrestore(&qdev->hw_lock, hw_flags); in ql_write_common_reg_l()
226 static void ql_write_common_reg(struct ql3_adapter *qdev, in ql_write_common_reg() argument
233 static void ql_write_nvram_reg(struct ql3_adapter *qdev, in ql_write_nvram_reg() argument
241 static void ql_write_page0_reg(struct ql3_adapter *qdev, in ql_write_page0_reg() argument
244 if (qdev->current_page != 0) in ql_write_page0_reg()
245 ql_set_register_page(qdev, 0); in ql_write_page0_reg()
253 static void ql_write_page1_reg(struct ql3_adapter *qdev, in ql_write_page1_reg() argument
256 if (qdev->current_page != 1) in ql_write_page1_reg()
257 ql_set_register_page(qdev, 1); in ql_write_page1_reg()
265 static void ql_write_page2_reg(struct ql3_adapter *qdev, in ql_write_page2_reg() argument
268 if (qdev->current_page != 2) in ql_write_page2_reg()
269 ql_set_register_page(qdev, 2); in ql_write_page2_reg()
274 static void ql_disable_interrupts(struct ql3_adapter *qdev) in ql_disable_interrupts() argument
277 qdev->mem_map_registers; in ql_disable_interrupts()
279 ql_write_common_reg_l(qdev, &port_regs->CommonRegs.ispInterruptMaskReg, in ql_disable_interrupts()
284 static void ql_enable_interrupts(struct ql3_adapter *qdev) in ql_enable_interrupts() argument
287 qdev->mem_map_registers; in ql_enable_interrupts()
289 ql_write_common_reg_l(qdev, &port_regs->CommonRegs.ispInterruptMaskReg, in ql_enable_interrupts()
294 static void ql_release_to_lrg_buf_free_list(struct ql3_adapter *qdev, in ql_release_to_lrg_buf_free_list() argument
301 if (qdev->lrg_buf_free_tail == NULL) { /* The list is empty */ in ql_release_to_lrg_buf_free_list()
302 qdev->lrg_buf_free_head = qdev->lrg_buf_free_tail = lrg_buf_cb; in ql_release_to_lrg_buf_free_list()
304 qdev->lrg_buf_free_tail->next = lrg_buf_cb; in ql_release_to_lrg_buf_free_list()
305 qdev->lrg_buf_free_tail = lrg_buf_cb; in ql_release_to_lrg_buf_free_list()
309 lrg_buf_cb->skb = netdev_alloc_skb(qdev->ndev, in ql_release_to_lrg_buf_free_list()
310 qdev->lrg_buffer_len); in ql_release_to_lrg_buf_free_list()
312 qdev->lrg_buf_skb_check++; in ql_release_to_lrg_buf_free_list()
319 map = pci_map_single(qdev->pdev, in ql_release_to_lrg_buf_free_list()
321 qdev->lrg_buffer_len - in ql_release_to_lrg_buf_free_list()
324 err = pci_dma_mapping_error(qdev->pdev, map); in ql_release_to_lrg_buf_free_list()
326 netdev_err(qdev->ndev, in ql_release_to_lrg_buf_free_list()
332 qdev->lrg_buf_skb_check++; in ql_release_to_lrg_buf_free_list()
342 qdev->lrg_buffer_len - in ql_release_to_lrg_buf_free_list()
347 qdev->lrg_buf_free_count++; in ql_release_to_lrg_buf_free_list()
351 *qdev) in ql_get_from_lrg_buf_free_list()
353 struct ql_rcv_buf_cb *lrg_buf_cb = qdev->lrg_buf_free_head; in ql_get_from_lrg_buf_free_list()
356 qdev->lrg_buf_free_head = lrg_buf_cb->next; in ql_get_from_lrg_buf_free_list()
357 if (qdev->lrg_buf_free_head == NULL) in ql_get_from_lrg_buf_free_list()
358 qdev->lrg_buf_free_tail = NULL; in ql_get_from_lrg_buf_free_list()
359 qdev->lrg_buf_free_count--; in ql_get_from_lrg_buf_free_list()
368 static void fm93c56a_deselect(struct ql3_adapter *qdev);
369 static void eeprom_readword(struct ql3_adapter *qdev, u32 eepromAddr,
375 static void fm93c56a_select(struct ql3_adapter *qdev) in fm93c56a_select() argument
378 qdev->mem_map_registers; in fm93c56a_select()
381 qdev->eeprom_cmd_data = AUBURN_EEPROM_CS_1; in fm93c56a_select()
382 ql_write_nvram_reg(qdev, spir, ISP_NVRAM_MASK | qdev->eeprom_cmd_data); in fm93c56a_select()
383 ql_write_nvram_reg(qdev, spir, in fm93c56a_select()
384 ((ISP_NVRAM_MASK << 16) | qdev->eeprom_cmd_data)); in fm93c56a_select()
390 static void fm93c56a_cmd(struct ql3_adapter *qdev, u32 cmd, u32 eepromAddr) in fm93c56a_cmd() argument
397 qdev->mem_map_registers; in fm93c56a_cmd()
401 ql_write_nvram_reg(qdev, spir, in fm93c56a_cmd()
402 (ISP_NVRAM_MASK | qdev->eeprom_cmd_data | in fm93c56a_cmd()
404 ql_write_nvram_reg(qdev, spir, in fm93c56a_cmd()
405 (ISP_NVRAM_MASK | qdev->eeprom_cmd_data | in fm93c56a_cmd()
407 ql_write_nvram_reg(qdev, spir, in fm93c56a_cmd()
408 (ISP_NVRAM_MASK | qdev->eeprom_cmd_data | in fm93c56a_cmd()
420 ql_write_nvram_reg(qdev, spir, in fm93c56a_cmd()
422 qdev->eeprom_cmd_data | dataBit)); in fm93c56a_cmd()
425 ql_write_nvram_reg(qdev, spir, in fm93c56a_cmd()
426 (ISP_NVRAM_MASK | qdev->eeprom_cmd_data | in fm93c56a_cmd()
428 ql_write_nvram_reg(qdev, spir, in fm93c56a_cmd()
429 (ISP_NVRAM_MASK | qdev->eeprom_cmd_data | in fm93c56a_cmd()
445 ql_write_nvram_reg(qdev, spir, in fm93c56a_cmd()
447 qdev->eeprom_cmd_data | dataBit)); in fm93c56a_cmd()
450 ql_write_nvram_reg(qdev, spir, in fm93c56a_cmd()
451 (ISP_NVRAM_MASK | qdev->eeprom_cmd_data | in fm93c56a_cmd()
453 ql_write_nvram_reg(qdev, spir, in fm93c56a_cmd()
454 (ISP_NVRAM_MASK | qdev->eeprom_cmd_data | in fm93c56a_cmd()
463 static void fm93c56a_deselect(struct ql3_adapter *qdev) in fm93c56a_deselect() argument
466 qdev->mem_map_registers; in fm93c56a_deselect()
469 qdev->eeprom_cmd_data = AUBURN_EEPROM_CS_0; in fm93c56a_deselect()
470 ql_write_nvram_reg(qdev, spir, ISP_NVRAM_MASK | qdev->eeprom_cmd_data); in fm93c56a_deselect()
476 static void fm93c56a_datain(struct ql3_adapter *qdev, unsigned short *value) in fm93c56a_datain() argument
482 qdev->mem_map_registers; in fm93c56a_datain()
488 ql_write_nvram_reg(qdev, spir, in fm93c56a_datain()
489 ISP_NVRAM_MASK | qdev->eeprom_cmd_data | in fm93c56a_datain()
491 ql_write_nvram_reg(qdev, spir, in fm93c56a_datain()
492 ISP_NVRAM_MASK | qdev->eeprom_cmd_data | in fm93c56a_datain()
494 dataBit = (ql_read_common_reg(qdev, spir) & in fm93c56a_datain()
504 static void eeprom_readword(struct ql3_adapter *qdev, in eeprom_readword() argument
507 fm93c56a_select(qdev); in eeprom_readword()
508 fm93c56a_cmd(qdev, (int)FM93C56A_READ, eepromAddr); in eeprom_readword()
509 fm93c56a_datain(qdev, value); in eeprom_readword()
510 fm93c56a_deselect(qdev); in eeprom_readword()
521 static int ql_get_nvram_params(struct ql3_adapter *qdev) in ql_get_nvram_params() argument
528 spin_lock_irqsave(&qdev->hw_lock, hw_flags); in ql_get_nvram_params()
530 pEEPROMData = (u16 *)&qdev->nvram_data; in ql_get_nvram_params()
531 qdev->eeprom_cmd_data = 0; in ql_get_nvram_params()
532 if (ql_sem_spinlock(qdev, QL_NVRAM_SEM_MASK, in ql_get_nvram_params()
533 (QL_RESOURCE_BITS_BASE_CODE | (qdev->mac_index) * in ql_get_nvram_params()
536 spin_unlock_irqrestore(&qdev->hw_lock, hw_flags); in ql_get_nvram_params()
541 eeprom_readword(qdev, index, pEEPROMData); in ql_get_nvram_params()
545 ql_sem_unlock(qdev, QL_NVRAM_SEM_MASK); in ql_get_nvram_params()
548 netdev_err(qdev->ndev, "checksum should be zero, is %x!!\n", in ql_get_nvram_params()
550 spin_unlock_irqrestore(&qdev->hw_lock, hw_flags); in ql_get_nvram_params()
554 spin_unlock_irqrestore(&qdev->hw_lock, hw_flags); in ql_get_nvram_params()
562 static int ql_wait_for_mii_ready(struct ql3_adapter *qdev) in ql_wait_for_mii_ready() argument
565 qdev->mem_map_registers; in ql_wait_for_mii_ready()
570 temp = ql_read_page0_reg(qdev, &port_regs->macMIIStatusReg); in ql_wait_for_mii_ready()
579 static void ql_mii_enable_scan_mode(struct ql3_adapter *qdev) in ql_mii_enable_scan_mode() argument
582 qdev->mem_map_registers; in ql_mii_enable_scan_mode()
585 if (qdev->numPorts > 1) { in ql_mii_enable_scan_mode()
598 ql_write_page0_reg(qdev, &port_regs->macMIIMgmtAddrReg, in ql_mii_enable_scan_mode()
601 ql_write_page0_reg(qdev, &port_regs->macMIIMgmtControlReg, in ql_mii_enable_scan_mode()
606 static u8 ql_mii_disable_scan_mode(struct ql3_adapter *qdev) in ql_mii_disable_scan_mode() argument
610 qdev->mem_map_registers; in ql_mii_disable_scan_mode()
613 if (ql_read_page0_reg(qdev, &port_regs->macMIIMgmtControlReg) & in ql_mii_disable_scan_mode()
626 ql_write_page0_reg(qdev, &port_regs->macMIIMgmtAddrReg, in ql_mii_disable_scan_mode()
629 ql_write_page0_reg(qdev, &port_regs->macMIIMgmtControlReg, in ql_mii_disable_scan_mode()
636 static int ql_mii_write_reg_ex(struct ql3_adapter *qdev, in ql_mii_write_reg_ex() argument
640 qdev->mem_map_registers; in ql_mii_write_reg_ex()
643 scanWasEnabled = ql_mii_disable_scan_mode(qdev); in ql_mii_write_reg_ex()
645 if (ql_wait_for_mii_ready(qdev)) { in ql_mii_write_reg_ex()
646 netif_warn(qdev, link, qdev->ndev, TIMED_OUT_MSG); in ql_mii_write_reg_ex()
650 ql_write_page0_reg(qdev, &port_regs->macMIIMgmtAddrReg, in ql_mii_write_reg_ex()
653 ql_write_page0_reg(qdev, &port_regs->macMIIMgmtDataReg, value); in ql_mii_write_reg_ex()
656 if (ql_wait_for_mii_ready(qdev)) { in ql_mii_write_reg_ex()
657 netif_warn(qdev, link, qdev->ndev, TIMED_OUT_MSG); in ql_mii_write_reg_ex()
662 ql_mii_enable_scan_mode(qdev); in ql_mii_write_reg_ex()
667 static int ql_mii_read_reg_ex(struct ql3_adapter *qdev, u16 regAddr, in ql_mii_read_reg_ex() argument
671 qdev->mem_map_registers; in ql_mii_read_reg_ex()
675 scanWasEnabled = ql_mii_disable_scan_mode(qdev); in ql_mii_read_reg_ex()
677 if (ql_wait_for_mii_ready(qdev)) { in ql_mii_read_reg_ex()
678 netif_warn(qdev, link, qdev->ndev, TIMED_OUT_MSG); in ql_mii_read_reg_ex()
682 ql_write_page0_reg(qdev, &port_regs->macMIIMgmtAddrReg, in ql_mii_read_reg_ex()
685 ql_write_page0_reg(qdev, &port_regs->macMIIMgmtControlReg, in ql_mii_read_reg_ex()
688 ql_write_page0_reg(qdev, &port_regs->macMIIMgmtControlReg, in ql_mii_read_reg_ex()
692 if (ql_wait_for_mii_ready(qdev)) { in ql_mii_read_reg_ex()
693 netif_warn(qdev, link, qdev->ndev, TIMED_OUT_MSG); in ql_mii_read_reg_ex()
697 temp = ql_read_page0_reg(qdev, &port_regs->macMIIMgmtDataReg); in ql_mii_read_reg_ex()
701 ql_mii_enable_scan_mode(qdev); in ql_mii_read_reg_ex()
706 static int ql_mii_write_reg(struct ql3_adapter *qdev, u16 regAddr, u16 value) in ql_mii_write_reg() argument
709 qdev->mem_map_registers; in ql_mii_write_reg()
711 ql_mii_disable_scan_mode(qdev); in ql_mii_write_reg()
713 if (ql_wait_for_mii_ready(qdev)) { in ql_mii_write_reg()
714 netif_warn(qdev, link, qdev->ndev, TIMED_OUT_MSG); in ql_mii_write_reg()
718 ql_write_page0_reg(qdev, &port_regs->macMIIMgmtAddrReg, in ql_mii_write_reg()
719 qdev->PHYAddr | regAddr); in ql_mii_write_reg()
721 ql_write_page0_reg(qdev, &port_regs->macMIIMgmtDataReg, value); in ql_mii_write_reg()
724 if (ql_wait_for_mii_ready(qdev)) { in ql_mii_write_reg()
725 netif_warn(qdev, link, qdev->ndev, TIMED_OUT_MSG); in ql_mii_write_reg()
729 ql_mii_enable_scan_mode(qdev); in ql_mii_write_reg()
734 static int ql_mii_read_reg(struct ql3_adapter *qdev, u16 regAddr, u16 *value) in ql_mii_read_reg() argument
738 qdev->mem_map_registers; in ql_mii_read_reg()
740 ql_mii_disable_scan_mode(qdev); in ql_mii_read_reg()
742 if (ql_wait_for_mii_ready(qdev)) { in ql_mii_read_reg()
743 netif_warn(qdev, link, qdev->ndev, TIMED_OUT_MSG); in ql_mii_read_reg()
747 ql_write_page0_reg(qdev, &port_regs->macMIIMgmtAddrReg, in ql_mii_read_reg()
748 qdev->PHYAddr | regAddr); in ql_mii_read_reg()
750 ql_write_page0_reg(qdev, &port_regs->macMIIMgmtControlReg, in ql_mii_read_reg()
753 ql_write_page0_reg(qdev, &port_regs->macMIIMgmtControlReg, in ql_mii_read_reg()
757 if (ql_wait_for_mii_ready(qdev)) { in ql_mii_read_reg()
758 netif_warn(qdev, link, qdev->ndev, TIMED_OUT_MSG); in ql_mii_read_reg()
762 temp = ql_read_page0_reg(qdev, &port_regs->macMIIMgmtDataReg); in ql_mii_read_reg()
765 ql_mii_enable_scan_mode(qdev); in ql_mii_read_reg()
770 static void ql_petbi_reset(struct ql3_adapter *qdev) in ql_petbi_reset() argument
772 ql_mii_write_reg(qdev, PETBI_CONTROL_REG, PETBI_CTRL_SOFT_RESET); in ql_petbi_reset()
775 static void ql_petbi_start_neg(struct ql3_adapter *qdev) in ql_petbi_start_neg() argument
780 ql_mii_read_reg(qdev, PETBI_TBI_CTRL, &reg); in ql_petbi_start_neg()
782 ql_mii_write_reg(qdev, PETBI_TBI_CTRL, reg); in ql_petbi_start_neg()
784 ql_mii_write_reg(qdev, PETBI_NEG_ADVER, in ql_petbi_start_neg()
787 ql_mii_write_reg(qdev, PETBI_CONTROL_REG, in ql_petbi_start_neg()
793 static void ql_petbi_reset_ex(struct ql3_adapter *qdev) in ql_petbi_reset_ex() argument
795 ql_mii_write_reg_ex(qdev, PETBI_CONTROL_REG, PETBI_CTRL_SOFT_RESET, in ql_petbi_reset_ex()
796 PHYAddr[qdev->mac_index]); in ql_petbi_reset_ex()
799 static void ql_petbi_start_neg_ex(struct ql3_adapter *qdev) in ql_petbi_start_neg_ex() argument
804 ql_mii_read_reg_ex(qdev, PETBI_TBI_CTRL, &reg, in ql_petbi_start_neg_ex()
805 PHYAddr[qdev->mac_index]); in ql_petbi_start_neg_ex()
807 ql_mii_write_reg_ex(qdev, PETBI_TBI_CTRL, reg, in ql_petbi_start_neg_ex()
808 PHYAddr[qdev->mac_index]); in ql_petbi_start_neg_ex()
810 ql_mii_write_reg_ex(qdev, PETBI_NEG_ADVER, in ql_petbi_start_neg_ex()
812 PHYAddr[qdev->mac_index]); in ql_petbi_start_neg_ex()
814 ql_mii_write_reg_ex(qdev, PETBI_CONTROL_REG, in ql_petbi_start_neg_ex()
817 PHYAddr[qdev->mac_index]); in ql_petbi_start_neg_ex()
820 static void ql_petbi_init(struct ql3_adapter *qdev) in ql_petbi_init() argument
822 ql_petbi_reset(qdev); in ql_petbi_init()
823 ql_petbi_start_neg(qdev); in ql_petbi_init()
826 static void ql_petbi_init_ex(struct ql3_adapter *qdev) in ql_petbi_init_ex() argument
828 ql_petbi_reset_ex(qdev); in ql_petbi_init_ex()
829 ql_petbi_start_neg_ex(qdev); in ql_petbi_init_ex()
832 static int ql_is_petbi_neg_pause(struct ql3_adapter *qdev) in ql_is_petbi_neg_pause() argument
836 if (ql_mii_read_reg(qdev, PETBI_NEG_PARTNER, &reg) < 0) in ql_is_petbi_neg_pause()
842 static void phyAgereSpecificInit(struct ql3_adapter *qdev, u32 miiAddr) in phyAgereSpecificInit() argument
844 netdev_info(qdev->ndev, "enabling Agere specific PHY\n"); in phyAgereSpecificInit()
846 ql_mii_write_reg_ex(qdev, 0x00, 0x1940, miiAddr); in phyAgereSpecificInit()
848 ql_mii_write_reg_ex(qdev, 0x12, 0x840e, miiAddr); in phyAgereSpecificInit()
850 ql_mii_write_reg_ex(qdev, 0x10, 0x8805, miiAddr); in phyAgereSpecificInit()
852 ql_mii_write_reg_ex(qdev, 0x11, 0xf03e, miiAddr); in phyAgereSpecificInit()
854 ql_mii_write_reg_ex(qdev, 0x10, 0x8806, miiAddr); in phyAgereSpecificInit()
856 ql_mii_write_reg_ex(qdev, 0x11, 0x003e, miiAddr); in phyAgereSpecificInit()
858 ql_mii_write_reg_ex(qdev, 0x10, 0x8807, miiAddr); in phyAgereSpecificInit()
860 ql_mii_write_reg_ex(qdev, 0x11, 0x1f00, miiAddr); in phyAgereSpecificInit()
862 ql_mii_write_reg_ex(qdev, 0x10, 0x2806, miiAddr); in phyAgereSpecificInit()
864 ql_mii_write_reg_ex(qdev, 0x11, in phyAgereSpecificInit()
865 0x0020 | (PHYAddr[qdev->mac_index] >> 8), miiAddr); in phyAgereSpecificInit()
871 ql_mii_write_reg(qdev, 0x12, 0x840a); in phyAgereSpecificInit()
872 ql_mii_write_reg(qdev, 0x00, 0x1140); in phyAgereSpecificInit()
873 ql_mii_write_reg(qdev, 0x1c, 0xfaf0); in phyAgereSpecificInit()
876 static enum PHY_DEVICE_TYPE getPhyType(struct ql3_adapter *qdev, in getPhyType() argument
899 netdev_info(qdev->ndev, "Phy: %s\n", in getPhyType()
909 static int ql_phy_get_speed(struct ql3_adapter *qdev) in ql_phy_get_speed() argument
913 switch (qdev->phyType) { in ql_phy_get_speed()
915 if (ql_mii_read_reg(qdev, 0x1A, &reg) < 0) in ql_phy_get_speed()
922 if (ql_mii_read_reg(qdev, AUX_CONTROL_STATUS, &reg) < 0) in ql_phy_get_speed()
940 static int ql_is_full_dup(struct ql3_adapter *qdev) in ql_is_full_dup() argument
944 switch (qdev->phyType) { in ql_is_full_dup()
946 if (ql_mii_read_reg(qdev, 0x1A, &reg)) in ql_is_full_dup()
953 if (ql_mii_read_reg(qdev, AUX_CONTROL_STATUS, &reg) < 0) in ql_is_full_dup()
960 static int ql_is_phy_neg_pause(struct ql3_adapter *qdev) in ql_is_phy_neg_pause() argument
964 if (ql_mii_read_reg(qdev, PHY_NEG_PARTNER, &reg) < 0) in ql_is_phy_neg_pause()
970 static int PHY_Setup(struct ql3_adapter *qdev) in PHY_Setup() argument
979 err = ql_mii_read_reg(qdev, PHY_ID_0_REG, &reg1); in PHY_Setup()
981 netdev_err(qdev->ndev, "Could not read from reg PHY_ID_0_REG\n"); in PHY_Setup()
985 err = ql_mii_read_reg(qdev, PHY_ID_1_REG, &reg2); in PHY_Setup()
987 netdev_err(qdev->ndev, "Could not read from reg PHY_ID_1_REG\n"); in PHY_Setup()
996 if (qdev->mac_index == 0) in PHY_Setup()
1001 err = ql_mii_read_reg_ex(qdev, PHY_ID_0_REG, &reg1, miiAddr); in PHY_Setup()
1003 netdev_err(qdev->ndev, in PHY_Setup()
1008 err = ql_mii_read_reg_ex(qdev, PHY_ID_1_REG, &reg2, miiAddr); in PHY_Setup()
1010 netdev_err(qdev->ndev, "Could not read from reg PHY_ID_1_REG after Agere detected\n"); in PHY_Setup()
1020 qdev->phyType = getPhyType(qdev, reg1, reg2); in PHY_Setup()
1022 if ((qdev->phyType == PHY_AGERE_ET1011C) && agereAddrChangeNeeded) { in PHY_Setup()
1024 phyAgereSpecificInit(qdev, miiAddr); in PHY_Setup()
1025 } else if (qdev->phyType == PHY_TYPE_UNKNOWN) { in PHY_Setup()
1026 netdev_err(qdev->ndev, "PHY is unknown\n"); in PHY_Setup()
1036 static void ql_mac_enable(struct ql3_adapter *qdev, u32 enable) in ql_mac_enable() argument
1039 qdev->mem_map_registers; in ql_mac_enable()
1047 if (qdev->mac_index) in ql_mac_enable()
1048 ql_write_page0_reg(qdev, &port_regs->mac1ConfigReg, value); in ql_mac_enable()
1050 ql_write_page0_reg(qdev, &port_regs->mac0ConfigReg, value); in ql_mac_enable()
1056 static void ql_mac_cfg_soft_reset(struct ql3_adapter *qdev, u32 enable) in ql_mac_cfg_soft_reset() argument
1059 qdev->mem_map_registers; in ql_mac_cfg_soft_reset()
1067 if (qdev->mac_index) in ql_mac_cfg_soft_reset()
1068 ql_write_page0_reg(qdev, &port_regs->mac1ConfigReg, value); in ql_mac_cfg_soft_reset()
1070 ql_write_page0_reg(qdev, &port_regs->mac0ConfigReg, value); in ql_mac_cfg_soft_reset()
1076 static void ql_mac_cfg_gig(struct ql3_adapter *qdev, u32 enable) in ql_mac_cfg_gig() argument
1079 qdev->mem_map_registers; in ql_mac_cfg_gig()
1087 if (qdev->mac_index) in ql_mac_cfg_gig()
1088 ql_write_page0_reg(qdev, &port_regs->mac1ConfigReg, value); in ql_mac_cfg_gig()
1090 ql_write_page0_reg(qdev, &port_regs->mac0ConfigReg, value); in ql_mac_cfg_gig()
1096 static void ql_mac_cfg_full_dup(struct ql3_adapter *qdev, u32 enable) in ql_mac_cfg_full_dup() argument
1099 qdev->mem_map_registers; in ql_mac_cfg_full_dup()
1107 if (qdev->mac_index) in ql_mac_cfg_full_dup()
1108 ql_write_page0_reg(qdev, &port_regs->mac1ConfigReg, value); in ql_mac_cfg_full_dup()
1110 ql_write_page0_reg(qdev, &port_regs->mac0ConfigReg, value); in ql_mac_cfg_full_dup()
1116 static void ql_mac_cfg_pause(struct ql3_adapter *qdev, u32 enable) in ql_mac_cfg_pause() argument
1119 qdev->mem_map_registers; in ql_mac_cfg_pause()
1129 if (qdev->mac_index) in ql_mac_cfg_pause()
1130 ql_write_page0_reg(qdev, &port_regs->mac1ConfigReg, value); in ql_mac_cfg_pause()
1132 ql_write_page0_reg(qdev, &port_regs->mac0ConfigReg, value); in ql_mac_cfg_pause()
1138 static int ql_is_fiber(struct ql3_adapter *qdev) in ql_is_fiber() argument
1141 qdev->mem_map_registers; in ql_is_fiber()
1145 switch (qdev->mac_index) { in ql_is_fiber()
1154 temp = ql_read_page0_reg(qdev, &port_regs->portStatus); in ql_is_fiber()
1158 static int ql_is_auto_cfg(struct ql3_adapter *qdev) in ql_is_auto_cfg() argument
1161 ql_mii_read_reg(qdev, 0x00, &reg); in ql_is_auto_cfg()
1168 static int ql_is_auto_neg_complete(struct ql3_adapter *qdev) in ql_is_auto_neg_complete() argument
1171 qdev->mem_map_registers; in ql_is_auto_neg_complete()
1175 switch (qdev->mac_index) { in ql_is_auto_neg_complete()
1184 temp = ql_read_page0_reg(qdev, &port_regs->portStatus); in ql_is_auto_neg_complete()
1186 netif_info(qdev, link, qdev->ndev, "Auto-Negotiate complete\n"); in ql_is_auto_neg_complete()
1189 netif_info(qdev, link, qdev->ndev, "Auto-Negotiate incomplete\n"); in ql_is_auto_neg_complete()
1196 static int ql_is_neg_pause(struct ql3_adapter *qdev) in ql_is_neg_pause() argument
1198 if (ql_is_fiber(qdev)) in ql_is_neg_pause()
1199 return ql_is_petbi_neg_pause(qdev); in ql_is_neg_pause()
1201 return ql_is_phy_neg_pause(qdev); in ql_is_neg_pause()
1204 static int ql_auto_neg_error(struct ql3_adapter *qdev) in ql_auto_neg_error() argument
1207 qdev->mem_map_registers; in ql_auto_neg_error()
1211 switch (qdev->mac_index) { in ql_auto_neg_error()
1219 temp = ql_read_page0_reg(qdev, &port_regs->portStatus); in ql_auto_neg_error()
1223 static u32 ql_get_link_speed(struct ql3_adapter *qdev) in ql_get_link_speed() argument
1225 if (ql_is_fiber(qdev)) in ql_get_link_speed()
1228 return ql_phy_get_speed(qdev); in ql_get_link_speed()
1231 static int ql_is_link_full_dup(struct ql3_adapter *qdev) in ql_is_link_full_dup() argument
1233 if (ql_is_fiber(qdev)) in ql_is_link_full_dup()
1236 return ql_is_full_dup(qdev); in ql_is_link_full_dup()
1242 static int ql_link_down_detect(struct ql3_adapter *qdev) in ql_link_down_detect() argument
1245 qdev->mem_map_registers; in ql_link_down_detect()
1249 switch (qdev->mac_index) { in ql_link_down_detect()
1259 ql_read_common_reg(qdev, &port_regs->CommonRegs.ispControlStatus); in ql_link_down_detect()
1266 static int ql_link_down_detect_clear(struct ql3_adapter *qdev) in ql_link_down_detect_clear() argument
1269 qdev->mem_map_registers; in ql_link_down_detect_clear()
1271 switch (qdev->mac_index) { in ql_link_down_detect_clear()
1273 ql_write_common_reg(qdev, in ql_link_down_detect_clear()
1280 ql_write_common_reg(qdev, in ql_link_down_detect_clear()
1296 static int ql_this_adapter_controls_port(struct ql3_adapter *qdev) in ql_this_adapter_controls_port() argument
1299 qdev->mem_map_registers; in ql_this_adapter_controls_port()
1303 switch (qdev->mac_index) { in ql_this_adapter_controls_port()
1314 temp = ql_read_page0_reg(qdev, &port_regs->portStatus); in ql_this_adapter_controls_port()
1316 netif_printk(qdev, link, KERN_DEBUG, qdev->ndev, in ql_this_adapter_controls_port()
1321 netif_printk(qdev, link, KERN_DEBUG, qdev->ndev, "link master\n"); in ql_this_adapter_controls_port()
1325 static void ql_phy_reset_ex(struct ql3_adapter *qdev) in ql_phy_reset_ex() argument
1327 ql_mii_write_reg_ex(qdev, CONTROL_REG, PHY_CTRL_SOFT_RESET, in ql_phy_reset_ex()
1328 PHYAddr[qdev->mac_index]); in ql_phy_reset_ex()
1331 static void ql_phy_start_neg_ex(struct ql3_adapter *qdev) in ql_phy_start_neg_ex() argument
1336 if (qdev->phyType == PHY_AGERE_ET1011C) in ql_phy_start_neg_ex()
1337 ql_mii_write_reg(qdev, 0x13, 0x0000); in ql_phy_start_neg_ex()
1340 if (qdev->mac_index == 0) in ql_phy_start_neg_ex()
1342 qdev->nvram_data.macCfg_port0.portConfiguration; in ql_phy_start_neg_ex()
1345 qdev->nvram_data.macCfg_port1.portConfiguration; in ql_phy_start_neg_ex()
1353 ql_mii_read_reg_ex(qdev, PHY_GIG_CONTROL, &reg, in ql_phy_start_neg_ex()
1354 PHYAddr[qdev->mac_index]); in ql_phy_start_neg_ex()
1364 ql_mii_write_reg_ex(qdev, PHY_GIG_CONTROL, reg, in ql_phy_start_neg_ex()
1365 PHYAddr[qdev->mac_index]); in ql_phy_start_neg_ex()
1368 ql_mii_read_reg_ex(qdev, PHY_NEG_ADVER, &reg, in ql_phy_start_neg_ex()
1369 PHYAddr[qdev->mac_index]); in ql_phy_start_neg_ex()
1394 ql_mii_write_reg_ex(qdev, PHY_NEG_ADVER, reg, in ql_phy_start_neg_ex()
1395 PHYAddr[qdev->mac_index]); in ql_phy_start_neg_ex()
1397 ql_mii_read_reg_ex(qdev, CONTROL_REG, &reg, PHYAddr[qdev->mac_index]); in ql_phy_start_neg_ex()
1399 ql_mii_write_reg_ex(qdev, CONTROL_REG, in ql_phy_start_neg_ex()
1401 PHYAddr[qdev->mac_index]); in ql_phy_start_neg_ex()
1404 static void ql_phy_init_ex(struct ql3_adapter *qdev) in ql_phy_init_ex() argument
1406 ql_phy_reset_ex(qdev); in ql_phy_init_ex()
1407 PHY_Setup(qdev); in ql_phy_init_ex()
1408 ql_phy_start_neg_ex(qdev); in ql_phy_init_ex()
1414 static u32 ql_get_link_state(struct ql3_adapter *qdev) in ql_get_link_state() argument
1417 qdev->mem_map_registers; in ql_get_link_state()
1421 switch (qdev->mac_index) { in ql_get_link_state()
1430 temp = ql_read_page0_reg(qdev, &port_regs->portStatus); in ql_get_link_state()
1439 static int ql_port_start(struct ql3_adapter *qdev) in ql_port_start() argument
1441 if (ql_sem_spinlock(qdev, QL_PHY_GIO_SEM_MASK, in ql_port_start()
1442 (QL_RESOURCE_BITS_BASE_CODE | (qdev->mac_index) * in ql_port_start()
1444 netdev_err(qdev->ndev, "Could not get hw lock for GIO\n"); in ql_port_start()
1448 if (ql_is_fiber(qdev)) { in ql_port_start()
1449 ql_petbi_init(qdev); in ql_port_start()
1452 ql_phy_init_ex(qdev); in ql_port_start()
1455 ql_sem_unlock(qdev, QL_PHY_GIO_SEM_MASK); in ql_port_start()
1459 static int ql_finish_auto_neg(struct ql3_adapter *qdev) in ql_finish_auto_neg() argument
1462 if (ql_sem_spinlock(qdev, QL_PHY_GIO_SEM_MASK, in ql_finish_auto_neg()
1463 (QL_RESOURCE_BITS_BASE_CODE | (qdev->mac_index) * in ql_finish_auto_neg()
1467 if (!ql_auto_neg_error(qdev)) { in ql_finish_auto_neg()
1468 if (test_bit(QL_LINK_MASTER, &qdev->flags)) { in ql_finish_auto_neg()
1470 netif_printk(qdev, link, KERN_DEBUG, qdev->ndev, in ql_finish_auto_neg()
1472 ql_mac_cfg_soft_reset(qdev, 1); in ql_finish_auto_neg()
1473 ql_mac_cfg_gig(qdev, in ql_finish_auto_neg()
1475 (qdev) == in ql_finish_auto_neg()
1477 ql_mac_cfg_full_dup(qdev, in ql_finish_auto_neg()
1479 (qdev)); in ql_finish_auto_neg()
1480 ql_mac_cfg_pause(qdev, in ql_finish_auto_neg()
1482 (qdev)); in ql_finish_auto_neg()
1483 ql_mac_cfg_soft_reset(qdev, 0); in ql_finish_auto_neg()
1486 netif_printk(qdev, link, KERN_DEBUG, qdev->ndev, in ql_finish_auto_neg()
1488 ql_mac_enable(qdev, 1); in ql_finish_auto_neg()
1491 qdev->port_link_state = LS_UP; in ql_finish_auto_neg()
1492 netif_start_queue(qdev->ndev); in ql_finish_auto_neg()
1493 netif_carrier_on(qdev->ndev); in ql_finish_auto_neg()
1494 netif_info(qdev, link, qdev->ndev, in ql_finish_auto_neg()
1496 ql_get_link_speed(qdev), in ql_finish_auto_neg()
1497 ql_is_link_full_dup(qdev) ? "full" : "half"); in ql_finish_auto_neg()
1501 if (test_bit(QL_LINK_MASTER, &qdev->flags)) { in ql_finish_auto_neg()
1502 netif_printk(qdev, link, KERN_DEBUG, qdev->ndev, in ql_finish_auto_neg()
1508 ql_sem_unlock(qdev, QL_PHY_GIO_SEM_MASK); in ql_finish_auto_neg()
1509 if (ql_port_start(qdev)) /* Restart port */ in ql_finish_auto_neg()
1514 ql_sem_unlock(qdev, QL_PHY_GIO_SEM_MASK); in ql_finish_auto_neg()
1520 struct ql3_adapter *qdev = in ql_link_state_machine_work() local
1526 spin_lock_irqsave(&qdev->hw_lock, hw_flags); in ql_link_state_machine_work()
1528 curr_link_state = ql_get_link_state(qdev); in ql_link_state_machine_work()
1530 if (test_bit(QL_RESET_ACTIVE, &qdev->flags)) { in ql_link_state_machine_work()
1531 netif_info(qdev, link, qdev->ndev, in ql_link_state_machine_work()
1534 spin_unlock_irqrestore(&qdev->hw_lock, hw_flags); in ql_link_state_machine_work()
1537 mod_timer(&qdev->adapter_timer, jiffies + HZ * 1); in ql_link_state_machine_work()
1542 switch (qdev->port_link_state) { in ql_link_state_machine_work()
1544 if (test_bit(QL_LINK_MASTER, &qdev->flags)) in ql_link_state_machine_work()
1545 ql_port_start(qdev); in ql_link_state_machine_work()
1546 qdev->port_link_state = LS_DOWN; in ql_link_state_machine_work()
1551 netif_info(qdev, link, qdev->ndev, "Link is up\n"); in ql_link_state_machine_work()
1552 if (ql_is_auto_neg_complete(qdev)) in ql_link_state_machine_work()
1553 ql_finish_auto_neg(qdev); in ql_link_state_machine_work()
1555 if (qdev->port_link_state == LS_UP) in ql_link_state_machine_work()
1556 ql_link_down_detect_clear(qdev); in ql_link_state_machine_work()
1558 qdev->port_link_state = LS_UP; in ql_link_state_machine_work()
1568 netif_info(qdev, link, qdev->ndev, "Link is down\n"); in ql_link_state_machine_work()
1569 qdev->port_link_state = LS_DOWN; in ql_link_state_machine_work()
1571 if (ql_link_down_detect(qdev)) in ql_link_state_machine_work()
1572 qdev->port_link_state = LS_DOWN; in ql_link_state_machine_work()
1575 spin_unlock_irqrestore(&qdev->hw_lock, hw_flags); in ql_link_state_machine_work()
1578 mod_timer(&qdev->adapter_timer, jiffies + HZ * 1); in ql_link_state_machine_work()
1584 static void ql_get_phy_owner(struct ql3_adapter *qdev) in ql_get_phy_owner() argument
1586 if (ql_this_adapter_controls_port(qdev)) in ql_get_phy_owner()
1587 set_bit(QL_LINK_MASTER, &qdev->flags); in ql_get_phy_owner()
1589 clear_bit(QL_LINK_MASTER, &qdev->flags); in ql_get_phy_owner()
1595 static void ql_init_scan_mode(struct ql3_adapter *qdev) in ql_init_scan_mode() argument
1597 ql_mii_enable_scan_mode(qdev); in ql_init_scan_mode()
1599 if (test_bit(QL_LINK_OPTICAL, &qdev->flags)) { in ql_init_scan_mode()
1600 if (ql_this_adapter_controls_port(qdev)) in ql_init_scan_mode()
1601 ql_petbi_init_ex(qdev); in ql_init_scan_mode()
1603 if (ql_this_adapter_controls_port(qdev)) in ql_init_scan_mode()
1604 ql_phy_init_ex(qdev); in ql_init_scan_mode()
1614 static int ql_mii_setup(struct ql3_adapter *qdev) in ql_mii_setup() argument
1618 qdev->mem_map_registers; in ql_mii_setup()
1620 if (ql_sem_spinlock(qdev, QL_PHY_GIO_SEM_MASK, in ql_mii_setup()
1621 (QL_RESOURCE_BITS_BASE_CODE | (qdev->mac_index) * in ql_mii_setup()
1625 if (qdev->device_id == QL3032_DEVICE_ID) in ql_mii_setup()
1626 ql_write_page0_reg(qdev, in ql_mii_setup()
1632 ql_write_page0_reg(qdev, &port_regs->macMIIMgmtControlReg, in ql_mii_setup()
1635 ql_sem_unlock(qdev, QL_PHY_GIO_SEM_MASK); in ql_mii_setup()
1651 static u32 ql_supported_modes(struct ql3_adapter *qdev) in ql_supported_modes() argument
1653 if (test_bit(QL_LINK_OPTICAL, &qdev->flags)) in ql_supported_modes()
1659 static int ql_get_auto_cfg_status(struct ql3_adapter *qdev) in ql_get_auto_cfg_status() argument
1663 spin_lock_irqsave(&qdev->hw_lock, hw_flags); in ql_get_auto_cfg_status()
1664 if (ql_sem_spinlock(qdev, QL_PHY_GIO_SEM_MASK, in ql_get_auto_cfg_status()
1666 (qdev->mac_index) * 2) << 7)) { in ql_get_auto_cfg_status()
1667 spin_unlock_irqrestore(&qdev->hw_lock, hw_flags); in ql_get_auto_cfg_status()
1670 status = ql_is_auto_cfg(qdev); in ql_get_auto_cfg_status()
1671 ql_sem_unlock(qdev, QL_PHY_GIO_SEM_MASK); in ql_get_auto_cfg_status()
1672 spin_unlock_irqrestore(&qdev->hw_lock, hw_flags); in ql_get_auto_cfg_status()
1676 static u32 ql_get_speed(struct ql3_adapter *qdev) in ql_get_speed() argument
1680 spin_lock_irqsave(&qdev->hw_lock, hw_flags); in ql_get_speed()
1681 if (ql_sem_spinlock(qdev, QL_PHY_GIO_SEM_MASK, in ql_get_speed()
1683 (qdev->mac_index) * 2) << 7)) { in ql_get_speed()
1684 spin_unlock_irqrestore(&qdev->hw_lock, hw_flags); in ql_get_speed()
1687 status = ql_get_link_speed(qdev); in ql_get_speed()
1688 ql_sem_unlock(qdev, QL_PHY_GIO_SEM_MASK); in ql_get_speed()
1689 spin_unlock_irqrestore(&qdev->hw_lock, hw_flags); in ql_get_speed()
1693 static int ql_get_full_dup(struct ql3_adapter *qdev) in ql_get_full_dup() argument
1697 spin_lock_irqsave(&qdev->hw_lock, hw_flags); in ql_get_full_dup()
1698 if (ql_sem_spinlock(qdev, QL_PHY_GIO_SEM_MASK, in ql_get_full_dup()
1700 (qdev->mac_index) * 2) << 7)) { in ql_get_full_dup()
1701 spin_unlock_irqrestore(&qdev->hw_lock, hw_flags); in ql_get_full_dup()
1704 status = ql_is_link_full_dup(qdev); in ql_get_full_dup()
1705 ql_sem_unlock(qdev, QL_PHY_GIO_SEM_MASK); in ql_get_full_dup()
1706 spin_unlock_irqrestore(&qdev->hw_lock, hw_flags); in ql_get_full_dup()
1712 struct ql3_adapter *qdev = netdev_priv(ndev); in ql_get_settings() local
1715 ecmd->supported = ql_supported_modes(qdev); in ql_get_settings()
1717 if (test_bit(QL_LINK_OPTICAL, &qdev->flags)) { in ql_get_settings()
1721 ecmd->phy_address = qdev->PHYAddr; in ql_get_settings()
1723 ecmd->advertising = ql_supported_modes(qdev); in ql_get_settings()
1724 ecmd->autoneg = ql_get_auto_cfg_status(qdev); in ql_get_settings()
1725 ethtool_cmd_speed_set(ecmd, ql_get_speed(qdev)); in ql_get_settings()
1726 ecmd->duplex = ql_get_full_dup(qdev); in ql_get_settings()
1733 struct ql3_adapter *qdev = netdev_priv(ndev); in ql_get_drvinfo() local
1737 strlcpy(drvinfo->bus_info, pci_name(qdev->pdev), in ql_get_drvinfo()
1743 struct ql3_adapter *qdev = netdev_priv(ndev); in ql_get_msglevel() local
1744 return qdev->msg_enable; in ql_get_msglevel()
1749 struct ql3_adapter *qdev = netdev_priv(ndev); in ql_set_msglevel() local
1750 qdev->msg_enable = value; in ql_set_msglevel()
1756 struct ql3_adapter *qdev = netdev_priv(ndev); in ql_get_pauseparam() local
1758 qdev->mem_map_registers; in ql_get_pauseparam()
1761 if (qdev->mac_index == 0) in ql_get_pauseparam()
1762 reg = ql_read_page0_reg(qdev, &port_regs->mac0ConfigReg); in ql_get_pauseparam()
1764 reg = ql_read_page0_reg(qdev, &port_regs->mac1ConfigReg); in ql_get_pauseparam()
1766 pause->autoneg = ql_get_auto_cfg_status(qdev); in ql_get_pauseparam()
1780 static int ql_populate_free_queue(struct ql3_adapter *qdev) in ql_populate_free_queue() argument
1782 struct ql_rcv_buf_cb *lrg_buf_cb = qdev->lrg_buf_free_head; in ql_populate_free_queue()
1789 netdev_alloc_skb(qdev->ndev, in ql_populate_free_queue()
1790 qdev->lrg_buffer_len); in ql_populate_free_queue()
1792 netdev_printk(KERN_DEBUG, qdev->ndev, in ql_populate_free_queue()
1801 map = pci_map_single(qdev->pdev, in ql_populate_free_queue()
1803 qdev->lrg_buffer_len - in ql_populate_free_queue()
1807 err = pci_dma_mapping_error(qdev->pdev, map); in ql_populate_free_queue()
1809 netdev_err(qdev->ndev, in ql_populate_free_queue()
1824 qdev->lrg_buffer_len - in ql_populate_free_queue()
1826 --qdev->lrg_buf_skb_check; in ql_populate_free_queue()
1827 if (!qdev->lrg_buf_skb_check) in ql_populate_free_queue()
1839 static void ql_update_small_bufq_prod_index(struct ql3_adapter *qdev) in ql_update_small_bufq_prod_index() argument
1842 qdev->mem_map_registers; in ql_update_small_bufq_prod_index()
1844 if (qdev->small_buf_release_cnt >= 16) { in ql_update_small_bufq_prod_index()
1845 while (qdev->small_buf_release_cnt >= 16) { in ql_update_small_bufq_prod_index()
1846 qdev->small_buf_q_producer_index++; in ql_update_small_bufq_prod_index()
1848 if (qdev->small_buf_q_producer_index == in ql_update_small_bufq_prod_index()
1850 qdev->small_buf_q_producer_index = 0; in ql_update_small_bufq_prod_index()
1851 qdev->small_buf_release_cnt -= 8; in ql_update_small_bufq_prod_index()
1854 writel(qdev->small_buf_q_producer_index, in ql_update_small_bufq_prod_index()
1862 static void ql_update_lrg_bufq_prod_index(struct ql3_adapter *qdev) in ql_update_lrg_bufq_prod_index() argument
1868 qdev->mem_map_registers; in ql_update_lrg_bufq_prod_index()
1870 if ((qdev->lrg_buf_free_count >= 8) && in ql_update_lrg_bufq_prod_index()
1871 (qdev->lrg_buf_release_cnt >= 16)) { in ql_update_lrg_bufq_prod_index()
1873 if (qdev->lrg_buf_skb_check) in ql_update_lrg_bufq_prod_index()
1874 if (!ql_populate_free_queue(qdev)) in ql_update_lrg_bufq_prod_index()
1877 lrg_buf_q_ele = qdev->lrg_buf_next_free; in ql_update_lrg_bufq_prod_index()
1879 while ((qdev->lrg_buf_release_cnt >= 16) && in ql_update_lrg_bufq_prod_index()
1880 (qdev->lrg_buf_free_count >= 8)) { in ql_update_lrg_bufq_prod_index()
1884 ql_get_from_lrg_buf_free_list(qdev); in ql_update_lrg_bufq_prod_index()
1891 qdev->lrg_buf_release_cnt--; in ql_update_lrg_bufq_prod_index()
1894 qdev->lrg_buf_q_producer_index++; in ql_update_lrg_bufq_prod_index()
1896 if (qdev->lrg_buf_q_producer_index == in ql_update_lrg_bufq_prod_index()
1897 qdev->num_lbufq_entries) in ql_update_lrg_bufq_prod_index()
1898 qdev->lrg_buf_q_producer_index = 0; in ql_update_lrg_bufq_prod_index()
1900 if (qdev->lrg_buf_q_producer_index == in ql_update_lrg_bufq_prod_index()
1901 (qdev->num_lbufq_entries - 1)) { in ql_update_lrg_bufq_prod_index()
1902 lrg_buf_q_ele = qdev->lrg_buf_q_virt_addr; in ql_update_lrg_bufq_prod_index()
1906 qdev->lrg_buf_next_free = lrg_buf_q_ele; in ql_update_lrg_bufq_prod_index()
1907 writel(qdev->lrg_buf_q_producer_index, in ql_update_lrg_bufq_prod_index()
1912 static void ql_process_mac_tx_intr(struct ql3_adapter *qdev, in ql_process_mac_tx_intr() argument
1919 netdev_warn(qdev->ndev, in ql_process_mac_tx_intr()
1923 tx_cb = &qdev->tx_buf[mac_rsp->transaction_id]; in ql_process_mac_tx_intr()
1927 netdev_err(qdev->ndev, in ql_process_mac_tx_intr()
1930 qdev->ndev->stats.tx_errors++; in ql_process_mac_tx_intr()
1935 netdev_err(qdev->ndev, "tx_cb->seg_count == 0: %d\n", in ql_process_mac_tx_intr()
1938 qdev->ndev->stats.tx_errors++; in ql_process_mac_tx_intr()
1942 pci_unmap_single(qdev->pdev, in ql_process_mac_tx_intr()
1949 pci_unmap_page(qdev->pdev, in ql_process_mac_tx_intr()
1956 qdev->ndev->stats.tx_packets++; in ql_process_mac_tx_intr()
1957 qdev->ndev->stats.tx_bytes += tx_cb->skb->len; in ql_process_mac_tx_intr()
1964 atomic_inc(&qdev->tx_count); in ql_process_mac_tx_intr()
1967 static void ql_get_sbuf(struct ql3_adapter *qdev) in ql_get_sbuf() argument
1969 if (++qdev->small_buf_index == NUM_SMALL_BUFFERS) in ql_get_sbuf()
1970 qdev->small_buf_index = 0; in ql_get_sbuf()
1971 qdev->small_buf_release_cnt++; in ql_get_sbuf()
1974 static struct ql_rcv_buf_cb *ql_get_lbuf(struct ql3_adapter *qdev) in ql_get_lbuf() argument
1977 lrg_buf_cb = &qdev->lrg_buf[qdev->lrg_buf_index]; in ql_get_lbuf()
1978 qdev->lrg_buf_release_cnt++; in ql_get_lbuf()
1979 if (++qdev->lrg_buf_index == qdev->num_large_buffers) in ql_get_lbuf()
1980 qdev->lrg_buf_index = 0; in ql_get_lbuf()
1996 static void ql_process_mac_rx_intr(struct ql3_adapter *qdev, in ql_process_mac_rx_intr() argument
2007 ql_get_sbuf(qdev); in ql_process_mac_rx_intr()
2009 if (qdev->device_id == QL3022_DEVICE_ID) in ql_process_mac_rx_intr()
2010 lrg_buf_cb1 = ql_get_lbuf(qdev); in ql_process_mac_rx_intr()
2013 lrg_buf_cb2 = ql_get_lbuf(qdev); in ql_process_mac_rx_intr()
2016 qdev->ndev->stats.rx_packets++; in ql_process_mac_rx_intr()
2017 qdev->ndev->stats.rx_bytes += length; in ql_process_mac_rx_intr()
2020 pci_unmap_single(qdev->pdev, in ql_process_mac_rx_intr()
2026 skb->protocol = eth_type_trans(skb, qdev->ndev); in ql_process_mac_rx_intr()
2031 if (qdev->device_id == QL3022_DEVICE_ID) in ql_process_mac_rx_intr()
2032 ql_release_to_lrg_buf_free_list(qdev, lrg_buf_cb1); in ql_process_mac_rx_intr()
2033 ql_release_to_lrg_buf_free_list(qdev, lrg_buf_cb2); in ql_process_mac_rx_intr()
2036 static void ql_process_macip_rx_intr(struct ql3_adapter *qdev, in ql_process_macip_rx_intr() argument
2042 struct net_device *ndev = qdev->ndev; in ql_process_macip_rx_intr()
2050 ql_get_sbuf(qdev); in ql_process_macip_rx_intr()
2052 if (qdev->device_id == QL3022_DEVICE_ID) { in ql_process_macip_rx_intr()
2054 lrg_buf_cb1 = ql_get_lbuf(qdev); in ql_process_macip_rx_intr()
2062 lrg_buf_cb2 = ql_get_lbuf(qdev); in ql_process_macip_rx_intr()
2066 pci_unmap_single(qdev->pdev, in ql_process_macip_rx_intr()
2073 if (qdev->device_id == QL3022_DEVICE_ID) { in ql_process_macip_rx_intr()
2096 skb2->protocol = eth_type_trans(skb2, qdev->ndev); in ql_process_macip_rx_intr()
2103 if (qdev->device_id == QL3022_DEVICE_ID) in ql_process_macip_rx_intr()
2104 ql_release_to_lrg_buf_free_list(qdev, lrg_buf_cb1); in ql_process_macip_rx_intr()
2105 ql_release_to_lrg_buf_free_list(qdev, lrg_buf_cb2); in ql_process_macip_rx_intr()
2108 static int ql_tx_rx_clean(struct ql3_adapter *qdev, in ql_tx_rx_clean() argument
2112 struct net_device *ndev = qdev->ndev; in ql_tx_rx_clean()
2116 while ((le32_to_cpu(*(qdev->prsp_producer_index)) != in ql_tx_rx_clean()
2117 qdev->rsp_consumer_index) && (work_done < work_to_do)) { in ql_tx_rx_clean()
2119 net_rsp = qdev->rsp_current; in ql_tx_rx_clean()
2125 if (qdev->device_id == QL3032_DEVICE_ID) in ql_tx_rx_clean()
2131 ql_process_mac_tx_intr(qdev, (struct ob_mac_iocb_rsp *) in ql_tx_rx_clean()
2138 ql_process_mac_rx_intr(qdev, (struct ib_mac_iocb_rsp *) in ql_tx_rx_clean()
2145 ql_process_macip_rx_intr(qdev, (struct ib_ip_iocb_rsp *) in ql_tx_rx_clean()
2163 qdev->rsp_consumer_index++; in ql_tx_rx_clean()
2165 if (qdev->rsp_consumer_index == NUM_RSP_Q_ENTRIES) { in ql_tx_rx_clean()
2166 qdev->rsp_consumer_index = 0; in ql_tx_rx_clean()
2167 qdev->rsp_current = qdev->rsp_q_virt_addr; in ql_tx_rx_clean()
2169 qdev->rsp_current++; in ql_tx_rx_clean()
2180 struct ql3_adapter *qdev = container_of(napi, struct ql3_adapter, napi); in ql_poll() local
2184 qdev->mem_map_registers; in ql_poll()
2186 ql_tx_rx_clean(qdev, &tx_cleaned, &rx_cleaned, budget); in ql_poll()
2189 spin_lock_irqsave(&qdev->hw_lock, hw_flags); in ql_poll()
2191 ql_update_small_bufq_prod_index(qdev); in ql_poll()
2192 ql_update_lrg_bufq_prod_index(qdev); in ql_poll()
2193 writel(qdev->rsp_consumer_index, in ql_poll()
2195 spin_unlock_irqrestore(&qdev->hw_lock, hw_flags); in ql_poll()
2197 ql_enable_interrupts(qdev); in ql_poll()
2206 struct ql3_adapter *qdev = netdev_priv(ndev); in ql3xxx_isr() local
2208 qdev->mem_map_registers; in ql3xxx_isr()
2213 value = ql_read_common_reg_l(qdev, in ql3xxx_isr()
2217 spin_lock(&qdev->adapter_lock); in ql3xxx_isr()
2218 netif_stop_queue(qdev->ndev); in ql3xxx_isr()
2219 netif_carrier_off(qdev->ndev); in ql3xxx_isr()
2220 ql_disable_interrupts(qdev); in ql3xxx_isr()
2221 qdev->port_link_state = LS_DOWN; in ql3xxx_isr()
2222 set_bit(QL_RESET_ACTIVE, &qdev->flags) ; in ql3xxx_isr()
2229 ql_read_page0_reg_l(qdev, in ql3xxx_isr()
2234 set_bit(QL_RESET_START, &qdev->flags) ; in ql3xxx_isr()
2239 set_bit(QL_RESET_PER_SCSI, &qdev->flags) ; in ql3xxx_isr()
2244 queue_delayed_work(qdev->workqueue, &qdev->reset_work, 0); in ql3xxx_isr()
2245 spin_unlock(&qdev->adapter_lock); in ql3xxx_isr()
2247 ql_disable_interrupts(qdev); in ql3xxx_isr()
2248 if (likely(napi_schedule_prep(&qdev->napi))) in ql3xxx_isr()
2249 __napi_schedule(&qdev->napi); in ql3xxx_isr()
2263 static int ql_get_seg_count(struct ql3_adapter *qdev, unsigned short frags) in ql_get_seg_count() argument
2265 if (qdev->device_id == QL3022_DEVICE_ID) in ql_get_seg_count()
2303 static int ql_send_map(struct ql3_adapter *qdev, in ql_send_map() argument
2321 map = pci_map_single(qdev->pdev, skb->data, len, PCI_DMA_TODEVICE); in ql_send_map()
2323 err = pci_dma_mapping_error(qdev->pdev, map); in ql_send_map()
2325 netdev_err(qdev->ndev, "PCI mapping failed with error: %d\n", in ql_send_map()
2359 map = pci_map_single(qdev->pdev, oal, in ql_send_map()
2363 err = pci_dma_mapping_error(qdev->pdev, map); in ql_send_map()
2365 netdev_err(qdev->ndev, in ql_send_map()
2383 map = skb_frag_dma_map(&qdev->pdev->dev, frag, 0, skb_frag_size(frag), in ql_send_map()
2386 err = dma_mapping_error(&qdev->pdev->dev, map); in ql_send_map()
2388 netdev_err(qdev->ndev, in ql_send_map()
2425 pci_unmap_single(qdev->pdev, in ql_send_map()
2433 pci_unmap_page(qdev->pdev, in ql_send_map()
2439 pci_unmap_single(qdev->pdev, in ql_send_map()
2462 struct ql3_adapter *qdev = netdev_priv(ndev); in ql3xxx_send() local
2464 qdev->mem_map_registers; in ql3xxx_send()
2469 if (unlikely(atomic_read(&qdev->tx_count) < 2)) in ql3xxx_send()
2472 tx_cb = &qdev->tx_buf[qdev->req_producer_index]; in ql3xxx_send()
2473 tx_cb->seg_count = ql_get_seg_count(qdev, in ql3xxx_send()
2482 mac_iocb_ptr->opcode = qdev->mac_ob_opcode; in ql3xxx_send()
2484 mac_iocb_ptr->flags |= qdev->mb_bit_mask; in ql3xxx_send()
2485 mac_iocb_ptr->transaction_id = qdev->req_producer_index; in ql3xxx_send()
2488 if (qdev->device_id == QL3032_DEVICE_ID && in ql3xxx_send()
2492 if (ql_send_map(qdev, mac_iocb_ptr, tx_cb, skb) != NETDEV_TX_OK) { in ql3xxx_send()
2498 qdev->req_producer_index++; in ql3xxx_send()
2499 if (qdev->req_producer_index == NUM_REQ_Q_ENTRIES) in ql3xxx_send()
2500 qdev->req_producer_index = 0; in ql3xxx_send()
2502 ql_write_common_reg_l(qdev, in ql3xxx_send()
2504 qdev->req_producer_index); in ql3xxx_send()
2506 netif_printk(qdev, tx_queued, KERN_DEBUG, ndev, in ql3xxx_send()
2508 qdev->req_producer_index, skb->len); in ql3xxx_send()
2510 atomic_dec(&qdev->tx_count); in ql3xxx_send()
2514 static int ql_alloc_net_req_rsp_queues(struct ql3_adapter *qdev) in ql_alloc_net_req_rsp_queues() argument
2516 qdev->req_q_size = in ql_alloc_net_req_rsp_queues()
2519 qdev->rsp_q_size = NUM_RSP_Q_ENTRIES * sizeof(struct net_rsp_iocb); in ql_alloc_net_req_rsp_queues()
2526 qdev->req_q_virt_addr = in ql_alloc_net_req_rsp_queues()
2527 pci_alloc_consistent(qdev->pdev, in ql_alloc_net_req_rsp_queues()
2528 (size_t) qdev->req_q_size, in ql_alloc_net_req_rsp_queues()
2529 &qdev->req_q_phy_addr); in ql_alloc_net_req_rsp_queues()
2531 if ((qdev->req_q_virt_addr == NULL) || in ql_alloc_net_req_rsp_queues()
2532 LS_64BITS(qdev->req_q_phy_addr) & (qdev->req_q_size - 1)) { in ql_alloc_net_req_rsp_queues()
2533 netdev_err(qdev->ndev, "reqQ failed\n"); in ql_alloc_net_req_rsp_queues()
2537 qdev->rsp_q_virt_addr = in ql_alloc_net_req_rsp_queues()
2538 pci_alloc_consistent(qdev->pdev, in ql_alloc_net_req_rsp_queues()
2539 (size_t) qdev->rsp_q_size, in ql_alloc_net_req_rsp_queues()
2540 &qdev->rsp_q_phy_addr); in ql_alloc_net_req_rsp_queues()
2542 if ((qdev->rsp_q_virt_addr == NULL) || in ql_alloc_net_req_rsp_queues()
2543 LS_64BITS(qdev->rsp_q_phy_addr) & (qdev->rsp_q_size - 1)) { in ql_alloc_net_req_rsp_queues()
2544 netdev_err(qdev->ndev, "rspQ allocation failed\n"); in ql_alloc_net_req_rsp_queues()
2545 pci_free_consistent(qdev->pdev, (size_t) qdev->req_q_size, in ql_alloc_net_req_rsp_queues()
2546 qdev->req_q_virt_addr, in ql_alloc_net_req_rsp_queues()
2547 qdev->req_q_phy_addr); in ql_alloc_net_req_rsp_queues()
2551 set_bit(QL_ALLOC_REQ_RSP_Q_DONE, &qdev->flags); in ql_alloc_net_req_rsp_queues()
2556 static void ql_free_net_req_rsp_queues(struct ql3_adapter *qdev) in ql_free_net_req_rsp_queues() argument
2558 if (!test_bit(QL_ALLOC_REQ_RSP_Q_DONE, &qdev->flags)) { in ql_free_net_req_rsp_queues()
2559 netdev_info(qdev->ndev, "Already done\n"); in ql_free_net_req_rsp_queues()
2563 pci_free_consistent(qdev->pdev, in ql_free_net_req_rsp_queues()
2564 qdev->req_q_size, in ql_free_net_req_rsp_queues()
2565 qdev->req_q_virt_addr, qdev->req_q_phy_addr); in ql_free_net_req_rsp_queues()
2567 qdev->req_q_virt_addr = NULL; in ql_free_net_req_rsp_queues()
2569 pci_free_consistent(qdev->pdev, in ql_free_net_req_rsp_queues()
2570 qdev->rsp_q_size, in ql_free_net_req_rsp_queues()
2571 qdev->rsp_q_virt_addr, qdev->rsp_q_phy_addr); in ql_free_net_req_rsp_queues()
2573 qdev->rsp_q_virt_addr = NULL; in ql_free_net_req_rsp_queues()
2575 clear_bit(QL_ALLOC_REQ_RSP_Q_DONE, &qdev->flags); in ql_free_net_req_rsp_queues()
2578 static int ql_alloc_buffer_queues(struct ql3_adapter *qdev) in ql_alloc_buffer_queues() argument
2581 qdev->lrg_buf_q_size = in ql_alloc_buffer_queues()
2582 qdev->num_lbufq_entries * sizeof(struct lrg_buf_q_entry); in ql_alloc_buffer_queues()
2583 if (qdev->lrg_buf_q_size < PAGE_SIZE) in ql_alloc_buffer_queues()
2584 qdev->lrg_buf_q_alloc_size = PAGE_SIZE; in ql_alloc_buffer_queues()
2586 qdev->lrg_buf_q_alloc_size = qdev->lrg_buf_q_size * 2; in ql_alloc_buffer_queues()
2588 qdev->lrg_buf = kmalloc_array(qdev->num_large_buffers, in ql_alloc_buffer_queues()
2591 if (qdev->lrg_buf == NULL) in ql_alloc_buffer_queues()
2594 qdev->lrg_buf_q_alloc_virt_addr = in ql_alloc_buffer_queues()
2595 pci_alloc_consistent(qdev->pdev, in ql_alloc_buffer_queues()
2596 qdev->lrg_buf_q_alloc_size, in ql_alloc_buffer_queues()
2597 &qdev->lrg_buf_q_alloc_phy_addr); in ql_alloc_buffer_queues()
2599 if (qdev->lrg_buf_q_alloc_virt_addr == NULL) { in ql_alloc_buffer_queues()
2600 netdev_err(qdev->ndev, "lBufQ failed\n"); in ql_alloc_buffer_queues()
2603 qdev->lrg_buf_q_virt_addr = qdev->lrg_buf_q_alloc_virt_addr; in ql_alloc_buffer_queues()
2604 qdev->lrg_buf_q_phy_addr = qdev->lrg_buf_q_alloc_phy_addr; in ql_alloc_buffer_queues()
2607 qdev->small_buf_q_size = in ql_alloc_buffer_queues()
2609 if (qdev->small_buf_q_size < PAGE_SIZE) in ql_alloc_buffer_queues()
2610 qdev->small_buf_q_alloc_size = PAGE_SIZE; in ql_alloc_buffer_queues()
2612 qdev->small_buf_q_alloc_size = qdev->small_buf_q_size * 2; in ql_alloc_buffer_queues()
2614 qdev->small_buf_q_alloc_virt_addr = in ql_alloc_buffer_queues()
2615 pci_alloc_consistent(qdev->pdev, in ql_alloc_buffer_queues()
2616 qdev->small_buf_q_alloc_size, in ql_alloc_buffer_queues()
2617 &qdev->small_buf_q_alloc_phy_addr); in ql_alloc_buffer_queues()
2619 if (qdev->small_buf_q_alloc_virt_addr == NULL) { in ql_alloc_buffer_queues()
2620 netdev_err(qdev->ndev, "Small Buffer Queue allocation failed\n"); in ql_alloc_buffer_queues()
2621 pci_free_consistent(qdev->pdev, qdev->lrg_buf_q_alloc_size, in ql_alloc_buffer_queues()
2622 qdev->lrg_buf_q_alloc_virt_addr, in ql_alloc_buffer_queues()
2623 qdev->lrg_buf_q_alloc_phy_addr); in ql_alloc_buffer_queues()
2627 qdev->small_buf_q_virt_addr = qdev->small_buf_q_alloc_virt_addr; in ql_alloc_buffer_queues()
2628 qdev->small_buf_q_phy_addr = qdev->small_buf_q_alloc_phy_addr; in ql_alloc_buffer_queues()
2629 set_bit(QL_ALLOC_BUFQS_DONE, &qdev->flags); in ql_alloc_buffer_queues()
2633 static void ql_free_buffer_queues(struct ql3_adapter *qdev) in ql_free_buffer_queues() argument
2635 if (!test_bit(QL_ALLOC_BUFQS_DONE, &qdev->flags)) { in ql_free_buffer_queues()
2636 netdev_info(qdev->ndev, "Already done\n"); in ql_free_buffer_queues()
2639 kfree(qdev->lrg_buf); in ql_free_buffer_queues()
2640 pci_free_consistent(qdev->pdev, in ql_free_buffer_queues()
2641 qdev->lrg_buf_q_alloc_size, in ql_free_buffer_queues()
2642 qdev->lrg_buf_q_alloc_virt_addr, in ql_free_buffer_queues()
2643 qdev->lrg_buf_q_alloc_phy_addr); in ql_free_buffer_queues()
2645 qdev->lrg_buf_q_virt_addr = NULL; in ql_free_buffer_queues()
2647 pci_free_consistent(qdev->pdev, in ql_free_buffer_queues()
2648 qdev->small_buf_q_alloc_size, in ql_free_buffer_queues()
2649 qdev->small_buf_q_alloc_virt_addr, in ql_free_buffer_queues()
2650 qdev->small_buf_q_alloc_phy_addr); in ql_free_buffer_queues()
2652 qdev->small_buf_q_virt_addr = NULL; in ql_free_buffer_queues()
2654 clear_bit(QL_ALLOC_BUFQS_DONE, &qdev->flags); in ql_free_buffer_queues()
2657 static int ql_alloc_small_buffers(struct ql3_adapter *qdev) in ql_alloc_small_buffers() argument
2663 qdev->small_buf_total_size = in ql_alloc_small_buffers()
2667 qdev->small_buf_virt_addr = in ql_alloc_small_buffers()
2668 pci_alloc_consistent(qdev->pdev, in ql_alloc_small_buffers()
2669 qdev->small_buf_total_size, in ql_alloc_small_buffers()
2670 &qdev->small_buf_phy_addr); in ql_alloc_small_buffers()
2672 if (qdev->small_buf_virt_addr == NULL) { in ql_alloc_small_buffers()
2673 netdev_err(qdev->ndev, "Failed to get small buffer memory\n"); in ql_alloc_small_buffers()
2677 qdev->small_buf_phy_addr_low = LS_64BITS(qdev->small_buf_phy_addr); in ql_alloc_small_buffers()
2678 qdev->small_buf_phy_addr_high = MS_64BITS(qdev->small_buf_phy_addr); in ql_alloc_small_buffers()
2680 small_buf_q_entry = qdev->small_buf_q_virt_addr; in ql_alloc_small_buffers()
2685 cpu_to_le32(qdev->small_buf_phy_addr_high); in ql_alloc_small_buffers()
2687 cpu_to_le32(qdev->small_buf_phy_addr_low + in ql_alloc_small_buffers()
2691 qdev->small_buf_index = 0; in ql_alloc_small_buffers()
2692 set_bit(QL_ALLOC_SMALL_BUF_DONE, &qdev->flags); in ql_alloc_small_buffers()
2696 static void ql_free_small_buffers(struct ql3_adapter *qdev) in ql_free_small_buffers() argument
2698 if (!test_bit(QL_ALLOC_SMALL_BUF_DONE, &qdev->flags)) { in ql_free_small_buffers()
2699 netdev_info(qdev->ndev, "Already done\n"); in ql_free_small_buffers()
2702 if (qdev->small_buf_virt_addr != NULL) { in ql_free_small_buffers()
2703 pci_free_consistent(qdev->pdev, in ql_free_small_buffers()
2704 qdev->small_buf_total_size, in ql_free_small_buffers()
2705 qdev->small_buf_virt_addr, in ql_free_small_buffers()
2706 qdev->small_buf_phy_addr); in ql_free_small_buffers()
2708 qdev->small_buf_virt_addr = NULL; in ql_free_small_buffers()
2712 static void ql_free_large_buffers(struct ql3_adapter *qdev) in ql_free_large_buffers() argument
2717 for (i = 0; i < qdev->num_large_buffers; i++) { in ql_free_large_buffers()
2718 lrg_buf_cb = &qdev->lrg_buf[i]; in ql_free_large_buffers()
2721 pci_unmap_single(qdev->pdev, in ql_free_large_buffers()
2732 static void ql_init_large_buffers(struct ql3_adapter *qdev) in ql_init_large_buffers() argument
2736 struct bufq_addr_element *buf_addr_ele = qdev->lrg_buf_q_virt_addr; in ql_init_large_buffers()
2738 for (i = 0; i < qdev->num_large_buffers; i++) { in ql_init_large_buffers()
2739 lrg_buf_cb = &qdev->lrg_buf[i]; in ql_init_large_buffers()
2744 qdev->lrg_buf_index = 0; in ql_init_large_buffers()
2745 qdev->lrg_buf_skb_check = 0; in ql_init_large_buffers()
2748 static int ql_alloc_large_buffers(struct ql3_adapter *qdev) in ql_alloc_large_buffers() argument
2756 for (i = 0; i < qdev->num_large_buffers; i++) { in ql_alloc_large_buffers()
2757 skb = netdev_alloc_skb(qdev->ndev, in ql_alloc_large_buffers()
2758 qdev->lrg_buffer_len); in ql_alloc_large_buffers()
2761 netdev_err(qdev->ndev, in ql_alloc_large_buffers()
2763 qdev->lrg_buffer_len * 2, i); in ql_alloc_large_buffers()
2764 ql_free_large_buffers(qdev); in ql_alloc_large_buffers()
2768 lrg_buf_cb = &qdev->lrg_buf[i]; in ql_alloc_large_buffers()
2777 map = pci_map_single(qdev->pdev, in ql_alloc_large_buffers()
2779 qdev->lrg_buffer_len - in ql_alloc_large_buffers()
2783 err = pci_dma_mapping_error(qdev->pdev, map); in ql_alloc_large_buffers()
2785 netdev_err(qdev->ndev, in ql_alloc_large_buffers()
2788 ql_free_large_buffers(qdev); in ql_alloc_large_buffers()
2794 qdev->lrg_buffer_len - in ql_alloc_large_buffers()
2805 static void ql_free_send_free_list(struct ql3_adapter *qdev) in ql_free_send_free_list() argument
2810 tx_cb = &qdev->tx_buf[0]; in ql_free_send_free_list()
2818 static int ql_create_send_free_list(struct ql3_adapter *qdev) in ql_create_send_free_list() argument
2822 struct ob_mac_iocb_req *req_q_curr = qdev->req_q_virt_addr; in ql_create_send_free_list()
2827 tx_cb = &qdev->tx_buf[i]; in ql_create_send_free_list()
2838 static int ql_alloc_mem_resources(struct ql3_adapter *qdev) in ql_alloc_mem_resources() argument
2840 if (qdev->ndev->mtu == NORMAL_MTU_SIZE) { in ql_alloc_mem_resources()
2841 qdev->num_lbufq_entries = NUM_LBUFQ_ENTRIES; in ql_alloc_mem_resources()
2842 qdev->lrg_buffer_len = NORMAL_MTU_SIZE; in ql_alloc_mem_resources()
2843 } else if (qdev->ndev->mtu == JUMBO_MTU_SIZE) { in ql_alloc_mem_resources()
2847 qdev->num_lbufq_entries = JUMBO_NUM_LBUFQ_ENTRIES; in ql_alloc_mem_resources()
2848 qdev->lrg_buffer_len = JUMBO_MTU_SIZE; in ql_alloc_mem_resources()
2850 netdev_err(qdev->ndev, "Invalid mtu size: %d. Only %d and %d are accepted.\n", in ql_alloc_mem_resources()
2851 qdev->ndev->mtu, NORMAL_MTU_SIZE, JUMBO_MTU_SIZE); in ql_alloc_mem_resources()
2854 qdev->num_large_buffers = in ql_alloc_mem_resources()
2855 qdev->num_lbufq_entries * QL_ADDR_ELE_PER_BUFQ_ENTRY; in ql_alloc_mem_resources()
2856 qdev->lrg_buffer_len += VLAN_ETH_HLEN + VLAN_ID_LEN + QL_HEADER_SPACE; in ql_alloc_mem_resources()
2857 qdev->max_frame_size = in ql_alloc_mem_resources()
2858 (qdev->lrg_buffer_len - QL_HEADER_SPACE) + ETHERNET_CRC_SIZE; in ql_alloc_mem_resources()
2865 qdev->shadow_reg_virt_addr = in ql_alloc_mem_resources()
2866 pci_alloc_consistent(qdev->pdev, in ql_alloc_mem_resources()
2867 PAGE_SIZE, &qdev->shadow_reg_phy_addr); in ql_alloc_mem_resources()
2869 if (qdev->shadow_reg_virt_addr != NULL) { in ql_alloc_mem_resources()
2870 qdev->preq_consumer_index = qdev->shadow_reg_virt_addr; in ql_alloc_mem_resources()
2871 qdev->req_consumer_index_phy_addr_high = in ql_alloc_mem_resources()
2872 MS_64BITS(qdev->shadow_reg_phy_addr); in ql_alloc_mem_resources()
2873 qdev->req_consumer_index_phy_addr_low = in ql_alloc_mem_resources()
2874 LS_64BITS(qdev->shadow_reg_phy_addr); in ql_alloc_mem_resources()
2876 qdev->prsp_producer_index = in ql_alloc_mem_resources()
2877 (__le32 *) (((u8 *) qdev->preq_consumer_index) + 8); in ql_alloc_mem_resources()
2878 qdev->rsp_producer_index_phy_addr_high = in ql_alloc_mem_resources()
2879 qdev->req_consumer_index_phy_addr_high; in ql_alloc_mem_resources()
2880 qdev->rsp_producer_index_phy_addr_low = in ql_alloc_mem_resources()
2881 qdev->req_consumer_index_phy_addr_low + 8; in ql_alloc_mem_resources()
2883 netdev_err(qdev->ndev, "shadowReg Alloc failed\n"); in ql_alloc_mem_resources()
2887 if (ql_alloc_net_req_rsp_queues(qdev) != 0) { in ql_alloc_mem_resources()
2888 netdev_err(qdev->ndev, "ql_alloc_net_req_rsp_queues failed\n"); in ql_alloc_mem_resources()
2892 if (ql_alloc_buffer_queues(qdev) != 0) { in ql_alloc_mem_resources()
2893 netdev_err(qdev->ndev, "ql_alloc_buffer_queues failed\n"); in ql_alloc_mem_resources()
2897 if (ql_alloc_small_buffers(qdev) != 0) { in ql_alloc_mem_resources()
2898 netdev_err(qdev->ndev, "ql_alloc_small_buffers failed\n"); in ql_alloc_mem_resources()
2902 if (ql_alloc_large_buffers(qdev) != 0) { in ql_alloc_mem_resources()
2903 netdev_err(qdev->ndev, "ql_alloc_large_buffers failed\n"); in ql_alloc_mem_resources()
2908 ql_init_large_buffers(qdev); in ql_alloc_mem_resources()
2909 if (ql_create_send_free_list(qdev)) in ql_alloc_mem_resources()
2912 qdev->rsp_current = qdev->rsp_q_virt_addr; in ql_alloc_mem_resources()
2916 ql_free_send_free_list(qdev); in ql_alloc_mem_resources()
2918 ql_free_buffer_queues(qdev); in ql_alloc_mem_resources()
2920 ql_free_net_req_rsp_queues(qdev); in ql_alloc_mem_resources()
2922 pci_free_consistent(qdev->pdev, in ql_alloc_mem_resources()
2924 qdev->shadow_reg_virt_addr, in ql_alloc_mem_resources()
2925 qdev->shadow_reg_phy_addr); in ql_alloc_mem_resources()
2930 static void ql_free_mem_resources(struct ql3_adapter *qdev) in ql_free_mem_resources() argument
2932 ql_free_send_free_list(qdev); in ql_free_mem_resources()
2933 ql_free_large_buffers(qdev); in ql_free_mem_resources()
2934 ql_free_small_buffers(qdev); in ql_free_mem_resources()
2935 ql_free_buffer_queues(qdev); in ql_free_mem_resources()
2936 ql_free_net_req_rsp_queues(qdev); in ql_free_mem_resources()
2937 if (qdev->shadow_reg_virt_addr != NULL) { in ql_free_mem_resources()
2938 pci_free_consistent(qdev->pdev, in ql_free_mem_resources()
2940 qdev->shadow_reg_virt_addr, in ql_free_mem_resources()
2941 qdev->shadow_reg_phy_addr); in ql_free_mem_resources()
2942 qdev->shadow_reg_virt_addr = NULL; in ql_free_mem_resources()
2946 static int ql_init_misc_registers(struct ql3_adapter *qdev) in ql_init_misc_registers() argument
2949 (void __iomem *)qdev->mem_map_registers; in ql_init_misc_registers()
2951 if (ql_sem_spinlock(qdev, QL_DDR_RAM_SEM_MASK, in ql_init_misc_registers()
2952 (QL_RESOURCE_BITS_BASE_CODE | (qdev->mac_index) * in ql_init_misc_registers()
2956 ql_write_page2_reg(qdev, in ql_init_misc_registers()
2957 &local_ram->bufletSize, qdev->nvram_data.bufletSize); in ql_init_misc_registers()
2959 ql_write_page2_reg(qdev, in ql_init_misc_registers()
2961 qdev->nvram_data.bufletCount); in ql_init_misc_registers()
2963 ql_write_page2_reg(qdev, in ql_init_misc_registers()
2965 (qdev->nvram_data.tcpWindowThreshold25 << 16) | in ql_init_misc_registers()
2966 (qdev->nvram_data.tcpWindowThreshold0)); in ql_init_misc_registers()
2968 ql_write_page2_reg(qdev, in ql_init_misc_registers()
2970 qdev->nvram_data.tcpWindowThreshold50); in ql_init_misc_registers()
2972 ql_write_page2_reg(qdev, in ql_init_misc_registers()
2974 (qdev->nvram_data.ipHashTableBaseHi << 16) | in ql_init_misc_registers()
2975 qdev->nvram_data.ipHashTableBaseLo); in ql_init_misc_registers()
2976 ql_write_page2_reg(qdev, in ql_init_misc_registers()
2978 qdev->nvram_data.ipHashTableSize); in ql_init_misc_registers()
2979 ql_write_page2_reg(qdev, in ql_init_misc_registers()
2981 (qdev->nvram_data.tcpHashTableBaseHi << 16) | in ql_init_misc_registers()
2982 qdev->nvram_data.tcpHashTableBaseLo); in ql_init_misc_registers()
2983 ql_write_page2_reg(qdev, in ql_init_misc_registers()
2985 qdev->nvram_data.tcpHashTableSize); in ql_init_misc_registers()
2986 ql_write_page2_reg(qdev, in ql_init_misc_registers()
2988 (qdev->nvram_data.ncbTableBaseHi << 16) | in ql_init_misc_registers()
2989 qdev->nvram_data.ncbTableBaseLo); in ql_init_misc_registers()
2990 ql_write_page2_reg(qdev, in ql_init_misc_registers()
2992 qdev->nvram_data.ncbTableSize); in ql_init_misc_registers()
2993 ql_write_page2_reg(qdev, in ql_init_misc_registers()
2995 (qdev->nvram_data.drbTableBaseHi << 16) | in ql_init_misc_registers()
2996 qdev->nvram_data.drbTableBaseLo); in ql_init_misc_registers()
2997 ql_write_page2_reg(qdev, in ql_init_misc_registers()
2999 qdev->nvram_data.drbTableSize); in ql_init_misc_registers()
3000 ql_sem_unlock(qdev, QL_DDR_RAM_SEM_MASK); in ql_init_misc_registers()
3004 static int ql_adapter_initialize(struct ql3_adapter *qdev) in ql_adapter_initialize() argument
3008 qdev->mem_map_registers; in ql_adapter_initialize()
3015 if (ql_mii_setup(qdev)) in ql_adapter_initialize()
3019 ql_write_common_reg(qdev, spir, in ql_adapter_initialize()
3024 qdev->port_link_state = LS_DOWN; in ql_adapter_initialize()
3025 netif_carrier_off(qdev->ndev); in ql_adapter_initialize()
3028 ql_write_common_reg(qdev, spir, in ql_adapter_initialize()
3033 *((u32 *)(qdev->preq_consumer_index)) = 0; in ql_adapter_initialize()
3034 atomic_set(&qdev->tx_count, NUM_REQ_Q_ENTRIES); in ql_adapter_initialize()
3035 qdev->req_producer_index = 0; in ql_adapter_initialize()
3037 ql_write_page1_reg(qdev, in ql_adapter_initialize()
3039 qdev->req_consumer_index_phy_addr_high); in ql_adapter_initialize()
3040 ql_write_page1_reg(qdev, in ql_adapter_initialize()
3042 qdev->req_consumer_index_phy_addr_low); in ql_adapter_initialize()
3044 ql_write_page1_reg(qdev, in ql_adapter_initialize()
3046 MS_64BITS(qdev->req_q_phy_addr)); in ql_adapter_initialize()
3047 ql_write_page1_reg(qdev, in ql_adapter_initialize()
3049 LS_64BITS(qdev->req_q_phy_addr)); in ql_adapter_initialize()
3050 ql_write_page1_reg(qdev, &hmem_regs->reqLength, NUM_REQ_Q_ENTRIES); in ql_adapter_initialize()
3053 *((__le16 *) (qdev->prsp_producer_index)) = 0; in ql_adapter_initialize()
3054 qdev->rsp_consumer_index = 0; in ql_adapter_initialize()
3055 qdev->rsp_current = qdev->rsp_q_virt_addr; in ql_adapter_initialize()
3057 ql_write_page1_reg(qdev, in ql_adapter_initialize()
3059 qdev->rsp_producer_index_phy_addr_high); in ql_adapter_initialize()
3061 ql_write_page1_reg(qdev, in ql_adapter_initialize()
3063 qdev->rsp_producer_index_phy_addr_low); in ql_adapter_initialize()
3065 ql_write_page1_reg(qdev, in ql_adapter_initialize()
3067 MS_64BITS(qdev->rsp_q_phy_addr)); in ql_adapter_initialize()
3069 ql_write_page1_reg(qdev, in ql_adapter_initialize()
3071 LS_64BITS(qdev->rsp_q_phy_addr)); in ql_adapter_initialize()
3073 ql_write_page1_reg(qdev, &hmem_regs->rspLength, NUM_RSP_Q_ENTRIES); in ql_adapter_initialize()
3076 ql_write_page1_reg(qdev, in ql_adapter_initialize()
3078 MS_64BITS(qdev->lrg_buf_q_phy_addr)); in ql_adapter_initialize()
3080 ql_write_page1_reg(qdev, in ql_adapter_initialize()
3082 LS_64BITS(qdev->lrg_buf_q_phy_addr)); in ql_adapter_initialize()
3084 ql_write_page1_reg(qdev, in ql_adapter_initialize()
3086 qdev->num_lbufq_entries); in ql_adapter_initialize()
3088 ql_write_page1_reg(qdev, in ql_adapter_initialize()
3090 qdev->lrg_buffer_len); in ql_adapter_initialize()
3093 ql_write_page1_reg(qdev, in ql_adapter_initialize()
3095 MS_64BITS(qdev->small_buf_q_phy_addr)); in ql_adapter_initialize()
3097 ql_write_page1_reg(qdev, in ql_adapter_initialize()
3099 LS_64BITS(qdev->small_buf_q_phy_addr)); in ql_adapter_initialize()
3101 ql_write_page1_reg(qdev, &hmem_regs->rxSmallQLength, NUM_SBUFQ_ENTRIES); in ql_adapter_initialize()
3102 ql_write_page1_reg(qdev, in ql_adapter_initialize()
3106 qdev->small_buf_q_producer_index = NUM_SBUFQ_ENTRIES - 1; in ql_adapter_initialize()
3107 qdev->small_buf_release_cnt = 8; in ql_adapter_initialize()
3108 qdev->lrg_buf_q_producer_index = qdev->num_lbufq_entries - 1; in ql_adapter_initialize()
3109 qdev->lrg_buf_release_cnt = 8; in ql_adapter_initialize()
3110 qdev->lrg_buf_next_free = qdev->lrg_buf_q_virt_addr; in ql_adapter_initialize()
3111 qdev->small_buf_index = 0; in ql_adapter_initialize()
3112 qdev->lrg_buf_index = 0; in ql_adapter_initialize()
3113 qdev->lrg_buf_free_count = 0; in ql_adapter_initialize()
3114 qdev->lrg_buf_free_head = NULL; in ql_adapter_initialize()
3115 qdev->lrg_buf_free_tail = NULL; in ql_adapter_initialize()
3117 ql_write_common_reg(qdev, in ql_adapter_initialize()
3120 qdev->small_buf_q_producer_index); in ql_adapter_initialize()
3121 ql_write_common_reg(qdev, in ql_adapter_initialize()
3124 qdev->lrg_buf_q_producer_index); in ql_adapter_initialize()
3130 clear_bit(QL_LINK_MASTER, &qdev->flags); in ql_adapter_initialize()
3131 value = ql_read_page0_reg(qdev, &port_regs->portStatus); in ql_adapter_initialize()
3135 if (ql_init_misc_registers(qdev)) { in ql_adapter_initialize()
3140 value = qdev->nvram_data.tcpMaxWindowSize; in ql_adapter_initialize()
3141 ql_write_page0_reg(qdev, &port_regs->tcpMaxWindow, value); in ql_adapter_initialize()
3143 value = (0xFFFF << 16) | qdev->nvram_data.extHwConfig; in ql_adapter_initialize()
3145 if (ql_sem_spinlock(qdev, QL_FLASH_SEM_MASK, in ql_adapter_initialize()
3146 (QL_RESOURCE_BITS_BASE_CODE | (qdev->mac_index) in ql_adapter_initialize()
3151 ql_write_page0_reg(qdev, &port_regs->ExternalHWConfig, value); in ql_adapter_initialize()
3152 ql_write_page0_reg(qdev, &port_regs->InternalChipConfig, in ql_adapter_initialize()
3156 ql_sem_unlock(qdev, QL_FLASH_SEM_MASK); in ql_adapter_initialize()
3159 if (qdev->mac_index) in ql_adapter_initialize()
3160 ql_write_page0_reg(qdev, in ql_adapter_initialize()
3162 qdev->max_frame_size); in ql_adapter_initialize()
3164 ql_write_page0_reg(qdev, in ql_adapter_initialize()
3166 qdev->max_frame_size); in ql_adapter_initialize()
3168 if (ql_sem_spinlock(qdev, QL_PHY_GIO_SEM_MASK, in ql_adapter_initialize()
3169 (QL_RESOURCE_BITS_BASE_CODE | (qdev->mac_index) * in ql_adapter_initialize()
3175 PHY_Setup(qdev); in ql_adapter_initialize()
3176 ql_init_scan_mode(qdev); in ql_adapter_initialize()
3177 ql_get_phy_owner(qdev); in ql_adapter_initialize()
3182 ql_write_page0_reg(qdev, &port_regs->macAddrIndirectPtrReg, in ql_adapter_initialize()
3184 ql_write_page0_reg(qdev, &port_regs->macAddrDataReg, in ql_adapter_initialize()
3185 ((qdev->ndev->dev_addr[2] << 24) in ql_adapter_initialize()
3186 | (qdev->ndev->dev_addr[3] << 16) in ql_adapter_initialize()
3187 | (qdev->ndev->dev_addr[4] << 8) in ql_adapter_initialize()
3188 | qdev->ndev->dev_addr[5])); in ql_adapter_initialize()
3191 ql_write_page0_reg(qdev, &port_regs->macAddrIndirectPtrReg, in ql_adapter_initialize()
3193 ql_write_page0_reg(qdev, &port_regs->macAddrDataReg, in ql_adapter_initialize()
3194 ((qdev->ndev->dev_addr[0] << 8) in ql_adapter_initialize()
3195 | qdev->ndev->dev_addr[1])); in ql_adapter_initialize()
3198 ql_write_page0_reg(qdev, &port_regs->macAddrIndirectPtrReg, in ql_adapter_initialize()
3203 ql_write_page0_reg(qdev, &port_regs->ipAddrIndexReg, in ql_adapter_initialize()
3205 (qdev->mac_index << 2))); in ql_adapter_initialize()
3206 ql_write_page0_reg(qdev, &port_regs->ipAddrDataReg, 0); in ql_adapter_initialize()
3208 ql_write_page0_reg(qdev, &port_regs->ipAddrIndexReg, in ql_adapter_initialize()
3210 ((qdev->mac_index << 2) + 1))); in ql_adapter_initialize()
3211 ql_write_page0_reg(qdev, &port_regs->ipAddrDataReg, 0); in ql_adapter_initialize()
3213 ql_sem_unlock(qdev, QL_PHY_GIO_SEM_MASK); in ql_adapter_initialize()
3216 ql_write_page0_reg(qdev, in ql_adapter_initialize()
3221 value = ql_read_page0_reg(qdev, &port_regs->portStatus); in ql_adapter_initialize()
3224 spin_unlock_irq(&qdev->hw_lock); in ql_adapter_initialize()
3226 spin_lock_irq(&qdev->hw_lock); in ql_adapter_initialize()
3230 netdev_err(qdev->ndev, "Hw Initialization timeout\n"); in ql_adapter_initialize()
3236 if (qdev->device_id == QL3032_DEVICE_ID) { in ql_adapter_initialize()
3241 ql_write_page0_reg(qdev, &port_regs->functionControl, in ql_adapter_initialize()
3247 ql_write_page0_reg(qdev, &port_regs->portControl, in ql_adapter_initialize()
3259 static int ql_adapter_reset(struct ql3_adapter *qdev) in ql_adapter_reset() argument
3262 qdev->mem_map_registers; in ql_adapter_reset()
3267 set_bit(QL_RESET_ACTIVE, &qdev->flags); in ql_adapter_reset()
3268 clear_bit(QL_RESET_DONE, &qdev->flags); in ql_adapter_reset()
3273 netdev_printk(KERN_DEBUG, qdev->ndev, "Issue soft reset to chip\n"); in ql_adapter_reset()
3274 ql_write_common_reg(qdev, in ql_adapter_reset()
3279 netdev_printk(KERN_DEBUG, qdev->ndev, in ql_adapter_reset()
3286 ql_read_common_reg(qdev, in ql_adapter_reset()
3299 ql_read_common_reg(qdev, &port_regs->CommonRegs.ispControlStatus); in ql_adapter_reset()
3301 netdev_printk(KERN_DEBUG, qdev->ndev, in ql_adapter_reset()
3303 ql_write_common_reg(qdev, in ql_adapter_reset()
3311 ql_write_common_reg(qdev, in ql_adapter_reset()
3322 value = ql_read_common_reg(qdev, in ql_adapter_reset()
3333 clear_bit(QL_RESET_ACTIVE, &qdev->flags); in ql_adapter_reset()
3334 set_bit(QL_RESET_DONE, &qdev->flags); in ql_adapter_reset()
3338 static void ql_set_mac_info(struct ql3_adapter *qdev) in ql_set_mac_info() argument
3341 qdev->mem_map_registers; in ql_set_mac_info()
3347 ql_read_common_reg_l(qdev, &port_regs->CommonRegs.ispControlStatus); in ql_set_mac_info()
3349 port_status = ql_read_page0_reg(qdev, &port_regs->portStatus); in ql_set_mac_info()
3352 qdev->mac_index = 0; in ql_set_mac_info()
3353 qdev->mac_ob_opcode = OUTBOUND_MAC_IOCB | func_number; in ql_set_mac_info()
3354 qdev->mb_bit_mask = FN0_MA_BITS_MASK; in ql_set_mac_info()
3355 qdev->PHYAddr = PORT0_PHY_ADDRESS; in ql_set_mac_info()
3357 set_bit(QL_LINK_OPTICAL, &qdev->flags); in ql_set_mac_info()
3359 clear_bit(QL_LINK_OPTICAL, &qdev->flags); in ql_set_mac_info()
3363 qdev->mac_index = 1; in ql_set_mac_info()
3364 qdev->mac_ob_opcode = OUTBOUND_MAC_IOCB | func_number; in ql_set_mac_info()
3365 qdev->mb_bit_mask = FN1_MA_BITS_MASK; in ql_set_mac_info()
3366 qdev->PHYAddr = PORT1_PHY_ADDRESS; in ql_set_mac_info()
3368 set_bit(QL_LINK_OPTICAL, &qdev->flags); in ql_set_mac_info()
3370 clear_bit(QL_LINK_OPTICAL, &qdev->flags); in ql_set_mac_info()
3376 netdev_printk(KERN_DEBUG, qdev->ndev, in ql_set_mac_info()
3381 qdev->numPorts = qdev->nvram_data.version_and_numPorts >> 8; in ql_set_mac_info()
3386 struct ql3_adapter *qdev = netdev_priv(ndev); in ql_display_dev_info() local
3387 struct pci_dev *pdev = qdev->pdev; in ql_display_dev_info()
3391 DRV_NAME, qdev->index, qdev->chip_rev_id, in ql_display_dev_info()
3392 qdev->device_id == QL3032_DEVICE_ID ? "QLA3032" : "QLA3022", in ql_display_dev_info()
3393 qdev->pci_slot); in ql_display_dev_info()
3395 test_bit(QL_LINK_OPTICAL, &qdev->flags) ? "OPTICAL" : "COPPER"); in ql_display_dev_info()
3401 ((qdev->pci_width == 64) ? "64-bit" : "32-bit"), in ql_display_dev_info()
3402 ((qdev->pci_x) ? "PCI-X" : "PCI")); in ql_display_dev_info()
3405 qdev->mem_map_registers); in ql_display_dev_info()
3408 netif_info(qdev, probe, ndev, "MAC address %pM\n", ndev->dev_addr); in ql_display_dev_info()
3411 static int ql_adapter_down(struct ql3_adapter *qdev, int do_reset) in ql_adapter_down() argument
3413 struct net_device *ndev = qdev->ndev; in ql_adapter_down()
3419 clear_bit(QL_ADAPTER_UP, &qdev->flags); in ql_adapter_down()
3420 clear_bit(QL_LINK_MASTER, &qdev->flags); in ql_adapter_down()
3422 ql_disable_interrupts(qdev); in ql_adapter_down()
3424 free_irq(qdev->pdev->irq, ndev); in ql_adapter_down()
3426 if (qdev->msi && test_bit(QL_MSI_ENABLED, &qdev->flags)) { in ql_adapter_down()
3427 netdev_info(qdev->ndev, "calling pci_disable_msi()\n"); in ql_adapter_down()
3428 clear_bit(QL_MSI_ENABLED, &qdev->flags); in ql_adapter_down()
3429 pci_disable_msi(qdev->pdev); in ql_adapter_down()
3432 del_timer_sync(&qdev->adapter_timer); in ql_adapter_down()
3434 napi_disable(&qdev->napi); in ql_adapter_down()
3440 spin_lock_irqsave(&qdev->hw_lock, hw_flags); in ql_adapter_down()
3441 if (ql_wait_for_drvr_lock(qdev)) { in ql_adapter_down()
3442 soft_reset = ql_adapter_reset(qdev); in ql_adapter_down()
3445 qdev->index); in ql_adapter_down()
3454 spin_unlock_irqrestore(&qdev->hw_lock, hw_flags); in ql_adapter_down()
3456 ql_free_mem_resources(qdev); in ql_adapter_down()
3460 static int ql_adapter_up(struct ql3_adapter *qdev) in ql_adapter_up() argument
3462 struct net_device *ndev = qdev->ndev; in ql_adapter_up()
3467 if (ql_alloc_mem_resources(qdev)) { in ql_adapter_up()
3472 if (qdev->msi) { in ql_adapter_up()
3473 if (pci_enable_msi(qdev->pdev)) { in ql_adapter_up()
3476 qdev->msi = 0; in ql_adapter_up()
3479 set_bit(QL_MSI_ENABLED, &qdev->flags); in ql_adapter_up()
3484 err = request_irq(qdev->pdev->irq, ql3xxx_isr, in ql_adapter_up()
3489 qdev->pdev->irq); in ql_adapter_up()
3493 spin_lock_irqsave(&qdev->hw_lock, hw_flags); in ql_adapter_up()
3495 err = ql_wait_for_drvr_lock(qdev); in ql_adapter_up()
3497 err = ql_adapter_initialize(qdev); in ql_adapter_up()
3503 ql_sem_unlock(qdev, QL_DRVR_SEM_MASK); in ql_adapter_up()
3509 spin_unlock_irqrestore(&qdev->hw_lock, hw_flags); in ql_adapter_up()
3511 set_bit(QL_ADAPTER_UP, &qdev->flags); in ql_adapter_up()
3513 mod_timer(&qdev->adapter_timer, jiffies + HZ * 1); in ql_adapter_up()
3515 napi_enable(&qdev->napi); in ql_adapter_up()
3516 ql_enable_interrupts(qdev); in ql_adapter_up()
3520 ql_sem_unlock(qdev, QL_DRVR_SEM_MASK); in ql_adapter_up()
3522 spin_unlock_irqrestore(&qdev->hw_lock, hw_flags); in ql_adapter_up()
3523 free_irq(qdev->pdev->irq, ndev); in ql_adapter_up()
3525 if (qdev->msi && test_bit(QL_MSI_ENABLED, &qdev->flags)) { in ql_adapter_up()
3527 clear_bit(QL_MSI_ENABLED, &qdev->flags); in ql_adapter_up()
3528 pci_disable_msi(qdev->pdev); in ql_adapter_up()
3533 static int ql_cycle_adapter(struct ql3_adapter *qdev, int reset) in ql_cycle_adapter() argument
3535 if (ql_adapter_down(qdev, reset) || ql_adapter_up(qdev)) { in ql_cycle_adapter()
3536 netdev_err(qdev->ndev, in ql_cycle_adapter()
3539 dev_close(qdev->ndev); in ql_cycle_adapter()
3548 struct ql3_adapter *qdev = netdev_priv(ndev); in ql3xxx_close() local
3554 while (!test_bit(QL_ADAPTER_UP, &qdev->flags)) in ql3xxx_close()
3557 ql_adapter_down(qdev, QL_DO_RESET); in ql3xxx_close()
3563 struct ql3_adapter *qdev = netdev_priv(ndev); in ql3xxx_open() local
3564 return ql_adapter_up(qdev); in ql3xxx_open()
3569 struct ql3_adapter *qdev = netdev_priv(ndev); in ql3xxx_set_mac_address() local
3571 qdev->mem_map_registers; in ql3xxx_set_mac_address()
3583 spin_lock_irqsave(&qdev->hw_lock, hw_flags); in ql3xxx_set_mac_address()
3585 ql_write_page0_reg(qdev, &port_regs->macAddrIndirectPtrReg, in ql3xxx_set_mac_address()
3587 ql_write_page0_reg(qdev, &port_regs->macAddrDataReg, in ql3xxx_set_mac_address()
3593 ql_write_page0_reg(qdev, &port_regs->macAddrIndirectPtrReg, in ql3xxx_set_mac_address()
3595 ql_write_page0_reg(qdev, &port_regs->macAddrDataReg, in ql3xxx_set_mac_address()
3597 spin_unlock_irqrestore(&qdev->hw_lock, hw_flags); in ql3xxx_set_mac_address()
3604 struct ql3_adapter *qdev = netdev_priv(ndev); in ql3xxx_tx_timeout() local
3615 queue_delayed_work(qdev->workqueue, &qdev->tx_timeout_work, 0); in ql3xxx_tx_timeout()
3620 struct ql3_adapter *qdev = in ql_reset_work() local
3622 struct net_device *ndev = qdev->ndev; in ql_reset_work()
3627 qdev->mem_map_registers; in ql_reset_work()
3630 if (test_bit((QL_RESET_PER_SCSI | QL_RESET_START), &qdev->flags)) { in ql_reset_work()
3631 clear_bit(QL_LINK_MASTER, &qdev->flags); in ql_reset_work()
3638 tx_cb = &qdev->tx_buf[i]; in ql_reset_work()
3642 pci_unmap_single(qdev->pdev, in ql_reset_work()
3648 pci_unmap_page(qdev->pdev, in ql_reset_work()
3661 spin_lock_irqsave(&qdev->hw_lock, hw_flags); in ql_reset_work()
3662 ql_write_common_reg(qdev, in ql_reset_work()
3671 value = ql_read_common_reg(qdev, in ql_reset_work()
3684 ql_write_common_reg(qdev, in ql_reset_work()
3692 spin_unlock_irqrestore(&qdev->hw_lock, hw_flags); in ql_reset_work()
3694 spin_lock_irqsave(&qdev->hw_lock, hw_flags); in ql_reset_work()
3696 spin_unlock_irqrestore(&qdev->hw_lock, hw_flags); in ql_reset_work()
3707 clear_bit(QL_RESET_PER_SCSI, &qdev->flags); in ql_reset_work()
3708 clear_bit(QL_RESET_START, &qdev->flags); in ql_reset_work()
3709 ql_cycle_adapter(qdev, QL_DO_RESET); in ql_reset_work()
3713 clear_bit(QL_RESET_ACTIVE, &qdev->flags); in ql_reset_work()
3714 clear_bit(QL_RESET_PER_SCSI, &qdev->flags); in ql_reset_work()
3715 clear_bit(QL_RESET_START, &qdev->flags); in ql_reset_work()
3716 ql_cycle_adapter(qdev, QL_NO_RESET); in ql_reset_work()
3722 struct ql3_adapter *qdev = in ql_tx_timeout_work() local
3725 ql_cycle_adapter(qdev, QL_DO_RESET); in ql_tx_timeout_work()
3728 static void ql_get_board_info(struct ql3_adapter *qdev) in ql_get_board_info() argument
3731 qdev->mem_map_registers; in ql_get_board_info()
3734 value = ql_read_page0_reg_l(qdev, &port_regs->portStatus); in ql_get_board_info()
3736 qdev->chip_rev_id = ((value & PORT_STATUS_REV_ID_MASK) >> 12); in ql_get_board_info()
3738 qdev->pci_width = 64; in ql_get_board_info()
3740 qdev->pci_width = 32; in ql_get_board_info()
3742 qdev->pci_x = 1; in ql_get_board_info()
3744 qdev->pci_x = 0; in ql_get_board_info()
3745 qdev->pci_slot = (u8) PCI_SLOT(qdev->pdev->devfn); in ql_get_board_info()
3750 struct ql3_adapter *qdev = (struct ql3_adapter *)ptr; in ql3xxx_timer() local
3751 queue_delayed_work(qdev->workqueue, &qdev->link_state_work, 0); in ql3xxx_timer()
3768 struct ql3_adapter *qdev = NULL; in ql3xxx_probe() local
3809 qdev = netdev_priv(ndev); in ql3xxx_probe()
3810 qdev->index = cards_found; in ql3xxx_probe()
3811 qdev->ndev = ndev; in ql3xxx_probe()
3812 qdev->pdev = pdev; in ql3xxx_probe()
3813 qdev->device_id = pci_entry->device; in ql3xxx_probe()
3814 qdev->port_link_state = LS_DOWN; in ql3xxx_probe()
3816 qdev->msi = 1; in ql3xxx_probe()
3818 qdev->msg_enable = netif_msg_init(debug, default_msg); in ql3xxx_probe()
3822 if (qdev->device_id == QL3032_DEVICE_ID) in ql3xxx_probe()
3825 qdev->mem_map_registers = pci_ioremap_bar(pdev, 1); in ql3xxx_probe()
3826 if (!qdev->mem_map_registers) { in ql3xxx_probe()
3832 spin_lock_init(&qdev->adapter_lock); in ql3xxx_probe()
3833 spin_lock_init(&qdev->hw_lock); in ql3xxx_probe()
3840 netif_napi_add(ndev, &qdev->napi, ql_poll, 64); in ql3xxx_probe()
3845 if (ql_get_nvram_params(qdev)) { in ql3xxx_probe()
3847 __func__, qdev->index); in ql3xxx_probe()
3852 ql_set_mac_info(qdev); in ql3xxx_probe()
3855 if (qdev->mac_index) { in ql3xxx_probe()
3856 ndev->mtu = qdev->nvram_data.macCfg_port1.etherMtu_mac ; in ql3xxx_probe()
3857 ql_set_mac_addr(ndev, qdev->nvram_data.funcCfg_fn2.macAddress); in ql3xxx_probe()
3859 ndev->mtu = qdev->nvram_data.macCfg_port0.etherMtu_mac ; in ql3xxx_probe()
3860 ql_set_mac_addr(ndev, qdev->nvram_data.funcCfg_fn0.macAddress); in ql3xxx_probe()
3866 ql_get_board_info(qdev); in ql3xxx_probe()
3872 if (qdev->pci_x) in ql3xxx_probe()
3886 qdev->workqueue = create_singlethread_workqueue(ndev->name); in ql3xxx_probe()
3887 INIT_DELAYED_WORK(&qdev->reset_work, ql_reset_work); in ql3xxx_probe()
3888 INIT_DELAYED_WORK(&qdev->tx_timeout_work, ql_tx_timeout_work); in ql3xxx_probe()
3889 INIT_DELAYED_WORK(&qdev->link_state_work, ql_link_state_machine_work); in ql3xxx_probe()
3891 init_timer(&qdev->adapter_timer); in ql3xxx_probe()
3892 qdev->adapter_timer.function = ql3xxx_timer; in ql3xxx_probe()
3893 qdev->adapter_timer.expires = jiffies + HZ * 2; /* two second delay */ in ql3xxx_probe()
3894 qdev->adapter_timer.data = (unsigned long)qdev; in ql3xxx_probe()
3907 iounmap(qdev->mem_map_registers); in ql3xxx_probe()
3921 struct ql3_adapter *qdev = netdev_priv(ndev); in ql3xxx_remove() local
3925 ql_disable_interrupts(qdev); in ql3xxx_remove()
3927 if (qdev->workqueue) { in ql3xxx_remove()
3928 cancel_delayed_work(&qdev->reset_work); in ql3xxx_remove()
3929 cancel_delayed_work(&qdev->tx_timeout_work); in ql3xxx_remove()
3930 destroy_workqueue(qdev->workqueue); in ql3xxx_remove()
3931 qdev->workqueue = NULL; in ql3xxx_remove()
3934 iounmap(qdev->mem_map_registers); in ql3xxx_remove()