Lines Matching refs:fbr
287 struct fbr_lookup *fbr[NUM_FBRS]; member
741 if (rx_ring->fbr[1]->buffsize == 4096) in et131x_rx_dma_enable()
743 else if (rx_ring->fbr[1]->buffsize == 8192) in et131x_rx_dma_enable()
745 else if (rx_ring->fbr[1]->buffsize == 16384) in et131x_rx_dma_enable()
749 if (rx_ring->fbr[0]->buffsize == 256) in et131x_rx_dma_enable()
751 else if (rx_ring->fbr[0]->buffsize == 512) in et131x_rx_dma_enable()
753 else if (rx_ring->fbr[0]->buffsize == 1024) in et131x_rx_dma_enable()
1578 struct fbr_lookup *fbr = rx_local->fbr[id]; in et131x_config_rx_dma_regs() local
1595 fbr_entry = fbr->ring_virtaddr; in et131x_config_rx_dma_regs()
1596 for (entry = 0; entry < fbr->num_entries; entry++) { in et131x_config_rx_dma_regs()
1597 fbr_entry->addr_hi = fbr->bus_high[entry]; in et131x_config_rx_dma_regs()
1598 fbr_entry->addr_lo = fbr->bus_low[entry]; in et131x_config_rx_dma_regs()
1604 writel(upper_32_bits(fbr->ring_physaddr), base_hi); in et131x_config_rx_dma_regs()
1605 writel(lower_32_bits(fbr->ring_physaddr), base_lo); in et131x_config_rx_dma_regs()
1606 writel(fbr->num_entries - 1, num_des); in et131x_config_rx_dma_regs()
1612 fbr->local_full = ET_DMA10_WRAP; in et131x_config_rx_dma_regs()
1613 writel(((fbr->num_entries * LO_MARK_PERCENT_FOR_RX) / 100) - 1, in et131x_config_rx_dma_regs()
1869 struct fbr_lookup *fbr; in et131x_rx_dma_memory_alloc() local
1872 rx_ring->fbr[0] = kzalloc(sizeof(*fbr), GFP_KERNEL); in et131x_rx_dma_memory_alloc()
1873 if (rx_ring->fbr[0] == NULL) in et131x_rx_dma_memory_alloc()
1875 rx_ring->fbr[1] = kzalloc(sizeof(*fbr), GFP_KERNEL); in et131x_rx_dma_memory_alloc()
1876 if (rx_ring->fbr[1] == NULL) in et131x_rx_dma_memory_alloc()
1897 rx_ring->fbr[0]->buffsize = 256; in et131x_rx_dma_memory_alloc()
1898 rx_ring->fbr[0]->num_entries = 512; in et131x_rx_dma_memory_alloc()
1899 rx_ring->fbr[1]->buffsize = 2048; in et131x_rx_dma_memory_alloc()
1900 rx_ring->fbr[1]->num_entries = 512; in et131x_rx_dma_memory_alloc()
1902 rx_ring->fbr[0]->buffsize = 512; in et131x_rx_dma_memory_alloc()
1903 rx_ring->fbr[0]->num_entries = 1024; in et131x_rx_dma_memory_alloc()
1904 rx_ring->fbr[1]->buffsize = 4096; in et131x_rx_dma_memory_alloc()
1905 rx_ring->fbr[1]->num_entries = 512; in et131x_rx_dma_memory_alloc()
1907 rx_ring->fbr[0]->buffsize = 1024; in et131x_rx_dma_memory_alloc()
1908 rx_ring->fbr[0]->num_entries = 768; in et131x_rx_dma_memory_alloc()
1909 rx_ring->fbr[1]->buffsize = 16384; in et131x_rx_dma_memory_alloc()
1910 rx_ring->fbr[1]->num_entries = 128; in et131x_rx_dma_memory_alloc()
1913 rx_ring->psr_entries = rx_ring->fbr[0]->num_entries + in et131x_rx_dma_memory_alloc()
1914 rx_ring->fbr[1]->num_entries; in et131x_rx_dma_memory_alloc()
1917 fbr = rx_ring->fbr[id]; in et131x_rx_dma_memory_alloc()
1919 bufsize = sizeof(struct fbr_desc) * fbr->num_entries; in et131x_rx_dma_memory_alloc()
1920 fbr->ring_virtaddr = dma_alloc_coherent(&adapter->pdev->dev, in et131x_rx_dma_memory_alloc()
1922 &fbr->ring_physaddr, in et131x_rx_dma_memory_alloc()
1924 if (!fbr->ring_virtaddr) { in et131x_rx_dma_memory_alloc()
1933 fbr = rx_ring->fbr[id]; in et131x_rx_dma_memory_alloc()
1934 fbr_chunksize = (FBR_CHUNKS * fbr->buffsize); in et131x_rx_dma_memory_alloc()
1936 for (i = 0; i < fbr->num_entries / FBR_CHUNKS; i++) { in et131x_rx_dma_memory_alloc()
1939 fbr->mem_virtaddrs[i] = dma_alloc_coherent( in et131x_rx_dma_memory_alloc()
1941 &fbr->mem_physaddrs[i], in et131x_rx_dma_memory_alloc()
1944 if (!fbr->mem_virtaddrs[i]) { in et131x_rx_dma_memory_alloc()
1951 fbr_physaddr = fbr->mem_physaddrs[i]; in et131x_rx_dma_memory_alloc()
1959 fbr->virt[k] = (u8 *)fbr->mem_virtaddrs[i] + in et131x_rx_dma_memory_alloc()
1960 (j * fbr->buffsize); in et131x_rx_dma_memory_alloc()
1965 fbr->bus_high[k] = upper_32_bits(fbr_physaddr); in et131x_rx_dma_memory_alloc()
1966 fbr->bus_low[k] = lower_32_bits(fbr_physaddr); in et131x_rx_dma_memory_alloc()
1967 fbr_physaddr += fbr->buffsize; in et131x_rx_dma_memory_alloc()
2013 struct fbr_lookup *fbr; in et131x_rx_dma_memory_free() local
2029 fbr = rx_ring->fbr[id]; in et131x_rx_dma_memory_free()
2031 if (!fbr || !fbr->ring_virtaddr) in et131x_rx_dma_memory_free()
2035 for (ii = 0; ii < fbr->num_entries / FBR_CHUNKS; ii++) { in et131x_rx_dma_memory_free()
2036 if (fbr->mem_virtaddrs[ii]) { in et131x_rx_dma_memory_free()
2037 bufsize = fbr->buffsize * FBR_CHUNKS; in et131x_rx_dma_memory_free()
2041 fbr->mem_virtaddrs[ii], in et131x_rx_dma_memory_free()
2042 fbr->mem_physaddrs[ii]); in et131x_rx_dma_memory_free()
2044 fbr->mem_virtaddrs[ii] = NULL; in et131x_rx_dma_memory_free()
2048 bufsize = sizeof(struct fbr_desc) * fbr->num_entries; in et131x_rx_dma_memory_free()
2052 fbr->ring_virtaddr, in et131x_rx_dma_memory_free()
2053 fbr->ring_physaddr); in et131x_rx_dma_memory_free()
2055 fbr->ring_virtaddr = NULL; in et131x_rx_dma_memory_free()
2079 kfree(rx_ring->fbr[0]); in et131x_rx_dma_memory_free()
2080 kfree(rx_ring->fbr[1]); in et131x_rx_dma_memory_free()
2133 struct fbr_lookup *fbr = rx_local->fbr[ring_index]; in nic_return_rfd() local
2138 if (buff_index < fbr->num_entries) { in nic_return_rfd()
2148 next = (struct fbr_desc *)(fbr->ring_virtaddr) + in nic_return_rfd()
2149 INDEX10(fbr->local_full); in nic_return_rfd()
2155 next->addr_hi = fbr->bus_high[buff_index]; in nic_return_rfd()
2156 next->addr_lo = fbr->bus_low[buff_index]; in nic_return_rfd()
2159 free_buff_ring = bump_free_buff_ring(&fbr->local_full, in nic_return_rfd()
2160 fbr->num_entries - 1); in nic_return_rfd()
2199 struct fbr_lookup *fbr; in nic_rx_pkts() local
2221 fbr = rx_local->fbr[ring_index]; in nic_rx_pkts()
2236 if (ring_index > 1 || buff_index > fbr->num_entries - 1) { in nic_rx_pkts()
2284 memcpy(skb_put(skb, rfd->len), fbr->virt[buff_index], rfd->len); in nic_rx_pkts()