Lines Matching refs:rx_ring

286 struct rx_ring {  struct
491 struct rx_ring rx_ring; member
739 struct rx_ring *rx_ring = &adapter->rx_ring; in et131x_rx_dma_enable() local
741 if (rx_ring->fbr[1]->buffsize == 4096) in et131x_rx_dma_enable()
743 else if (rx_ring->fbr[1]->buffsize == 8192) in et131x_rx_dma_enable()
745 else if (rx_ring->fbr[1]->buffsize == 16384) in et131x_rx_dma_enable()
749 if (rx_ring->fbr[0]->buffsize == 256) in et131x_rx_dma_enable()
751 else if (rx_ring->fbr[0]->buffsize == 512) in et131x_rx_dma_enable()
753 else if (rx_ring->fbr[0]->buffsize == 1024) in et131x_rx_dma_enable()
1542 struct rx_ring *rx_local = &adapter->rx_ring; in et131x_config_rx_dma_regs()
1868 struct rx_ring *rx_ring = &adapter->rx_ring; in et131x_rx_dma_memory_alloc() local
1872 rx_ring->fbr[0] = kzalloc(sizeof(*fbr), GFP_KERNEL); in et131x_rx_dma_memory_alloc()
1873 if (rx_ring->fbr[0] == NULL) in et131x_rx_dma_memory_alloc()
1875 rx_ring->fbr[1] = kzalloc(sizeof(*fbr), GFP_KERNEL); in et131x_rx_dma_memory_alloc()
1876 if (rx_ring->fbr[1] == NULL) in et131x_rx_dma_memory_alloc()
1897 rx_ring->fbr[0]->buffsize = 256; in et131x_rx_dma_memory_alloc()
1898 rx_ring->fbr[0]->num_entries = 512; in et131x_rx_dma_memory_alloc()
1899 rx_ring->fbr[1]->buffsize = 2048; in et131x_rx_dma_memory_alloc()
1900 rx_ring->fbr[1]->num_entries = 512; in et131x_rx_dma_memory_alloc()
1902 rx_ring->fbr[0]->buffsize = 512; in et131x_rx_dma_memory_alloc()
1903 rx_ring->fbr[0]->num_entries = 1024; in et131x_rx_dma_memory_alloc()
1904 rx_ring->fbr[1]->buffsize = 4096; in et131x_rx_dma_memory_alloc()
1905 rx_ring->fbr[1]->num_entries = 512; in et131x_rx_dma_memory_alloc()
1907 rx_ring->fbr[0]->buffsize = 1024; in et131x_rx_dma_memory_alloc()
1908 rx_ring->fbr[0]->num_entries = 768; in et131x_rx_dma_memory_alloc()
1909 rx_ring->fbr[1]->buffsize = 16384; in et131x_rx_dma_memory_alloc()
1910 rx_ring->fbr[1]->num_entries = 128; in et131x_rx_dma_memory_alloc()
1913 rx_ring->psr_entries = rx_ring->fbr[0]->num_entries + in et131x_rx_dma_memory_alloc()
1914 rx_ring->fbr[1]->num_entries; in et131x_rx_dma_memory_alloc()
1917 fbr = rx_ring->fbr[id]; in et131x_rx_dma_memory_alloc()
1933 fbr = rx_ring->fbr[id]; in et131x_rx_dma_memory_alloc()
1973 psr_size = sizeof(struct pkt_stat_desc) * rx_ring->psr_entries; in et131x_rx_dma_memory_alloc()
1975 rx_ring->ps_ring_virtaddr = dma_alloc_coherent(&adapter->pdev->dev, in et131x_rx_dma_memory_alloc()
1977 &rx_ring->ps_ring_physaddr, in et131x_rx_dma_memory_alloc()
1980 if (!rx_ring->ps_ring_virtaddr) { in et131x_rx_dma_memory_alloc()
1987 rx_ring->rx_status_block = dma_alloc_coherent(&adapter->pdev->dev, in et131x_rx_dma_memory_alloc()
1989 &rx_ring->rx_status_bus, in et131x_rx_dma_memory_alloc()
1991 if (!rx_ring->rx_status_block) { in et131x_rx_dma_memory_alloc()
1996 rx_ring->num_rfd = NIC_DEFAULT_NUM_RFD; in et131x_rx_dma_memory_alloc()
2001 INIT_LIST_HEAD(&rx_ring->recv_list); in et131x_rx_dma_memory_alloc()
2012 struct rx_ring *rx_ring = &adapter->rx_ring; in et131x_rx_dma_memory_free() local
2016 WARN_ON(rx_ring->num_ready_recv != rx_ring->num_rfd); in et131x_rx_dma_memory_free()
2018 while (!list_empty(&rx_ring->recv_list)) { in et131x_rx_dma_memory_free()
2019 rfd = list_entry(rx_ring->recv_list.next, in et131x_rx_dma_memory_free()
2029 fbr = rx_ring->fbr[id]; in et131x_rx_dma_memory_free()
2059 if (rx_ring->ps_ring_virtaddr) { in et131x_rx_dma_memory_free()
2060 psr_size = sizeof(struct pkt_stat_desc) * rx_ring->psr_entries; in et131x_rx_dma_memory_free()
2063 rx_ring->ps_ring_virtaddr, in et131x_rx_dma_memory_free()
2064 rx_ring->ps_ring_physaddr); in et131x_rx_dma_memory_free()
2066 rx_ring->ps_ring_virtaddr = NULL; in et131x_rx_dma_memory_free()
2070 if (rx_ring->rx_status_block) { in et131x_rx_dma_memory_free()
2073 rx_ring->rx_status_block, in et131x_rx_dma_memory_free()
2074 rx_ring->rx_status_bus); in et131x_rx_dma_memory_free()
2075 rx_ring->rx_status_block = NULL; in et131x_rx_dma_memory_free()
2079 kfree(rx_ring->fbr[0]); in et131x_rx_dma_memory_free()
2080 kfree(rx_ring->fbr[1]); in et131x_rx_dma_memory_free()
2083 rx_ring->num_ready_recv = 0; in et131x_rx_dma_memory_free()
2091 struct rx_ring *rx_ring = &adapter->rx_ring; in et131x_init_recv() local
2094 for (rfdct = 0; rfdct < rx_ring->num_rfd; rfdct++) { in et131x_init_recv()
2102 list_add_tail(&rfd->list_node, &rx_ring->recv_list); in et131x_init_recv()
2105 rx_ring->num_ready_recv++; in et131x_init_recv()
2128 struct rx_ring *rx_local = &adapter->rx_ring; in nic_return_rfd()
2187 struct rx_ring *rx_local = &adapter->rx_ring; in nic_rx_pkts()
2301 struct rx_ring *rx_ring = &adapter->rx_ring; in et131x_handle_recv_pkts() local
2308 if (list_empty(&rx_ring->recv_list)) { in et131x_handle_recv_pkts()
2309 WARN_ON(rx_ring->num_ready_recv != 0); in et131x_handle_recv_pkts()
2331 if (rx_ring->num_ready_recv < RFD_LOW_WATER_MARK) in et131x_handle_recv_pkts()
2338 rx_ring->unfinished_receives = true; in et131x_handle_recv_pkts()
2343 rx_ring->unfinished_receives = false; in et131x_handle_recv_pkts()
3400 struct rx_ring *rx_ring = &adapter->rx_ring; in et131x_isr() local
3434 if (rx_ring->unfinished_receives) in et131x_isr()