Lines Matching refs:rx_bi
470 struct i40e_rx_buffer *rx_bi; in i40evf_clean_rx_ring() local
475 if (!rx_ring->rx_bi) in i40evf_clean_rx_ring()
481 rx_bi = &rx_ring->rx_bi[0]; in i40evf_clean_rx_ring()
482 if (rx_bi->hdr_buf) { in i40evf_clean_rx_ring()
485 rx_bi->hdr_buf, in i40evf_clean_rx_ring()
486 rx_bi->dma); in i40evf_clean_rx_ring()
488 rx_bi = &rx_ring->rx_bi[i]; in i40evf_clean_rx_ring()
489 rx_bi->dma = 0; in i40evf_clean_rx_ring()
490 rx_bi->hdr_buf = NULL; in i40evf_clean_rx_ring()
496 rx_bi = &rx_ring->rx_bi[i]; in i40evf_clean_rx_ring()
497 if (rx_bi->dma) { in i40evf_clean_rx_ring()
499 rx_bi->dma, in i40evf_clean_rx_ring()
502 rx_bi->dma = 0; in i40evf_clean_rx_ring()
504 if (rx_bi->skb) { in i40evf_clean_rx_ring()
505 dev_kfree_skb(rx_bi->skb); in i40evf_clean_rx_ring()
506 rx_bi->skb = NULL; in i40evf_clean_rx_ring()
508 if (rx_bi->page) { in i40evf_clean_rx_ring()
509 if (rx_bi->page_dma) { in i40evf_clean_rx_ring()
511 rx_bi->page_dma, in i40evf_clean_rx_ring()
514 rx_bi->page_dma = 0; in i40evf_clean_rx_ring()
516 __free_page(rx_bi->page); in i40evf_clean_rx_ring()
517 rx_bi->page = NULL; in i40evf_clean_rx_ring()
518 rx_bi->page_offset = 0; in i40evf_clean_rx_ring()
523 memset(rx_ring->rx_bi, 0, bi_size); in i40evf_clean_rx_ring()
541 kfree(rx_ring->rx_bi); in i40evf_free_rx_resources()
542 rx_ring->rx_bi = NULL; in i40evf_free_rx_resources()
561 struct i40e_rx_buffer *rx_bi; in i40evf_alloc_rx_headers() local
567 if (rx_ring->rx_bi[0].hdr_buf) in i40evf_alloc_rx_headers()
576 rx_bi = &rx_ring->rx_bi[i]; in i40evf_alloc_rx_headers()
577 rx_bi->dma = dma + (i * buf_size); in i40evf_alloc_rx_headers()
578 rx_bi->hdr_buf = buffer + (i * buf_size); in i40evf_alloc_rx_headers()
594 WARN_ON(rx_ring->rx_bi); in i40evf_setup_rx_descriptors()
596 rx_ring->rx_bi = kzalloc(bi_size, GFP_KERNEL); in i40evf_setup_rx_descriptors()
597 if (!rx_ring->rx_bi) in i40evf_setup_rx_descriptors()
621 kfree(rx_ring->rx_bi); in i40evf_setup_rx_descriptors()
622 rx_ring->rx_bi = NULL; in i40evf_setup_rx_descriptors()
660 bi = &rx_ring->rx_bi[i]; in i40evf_alloc_rx_buffers_ps()
726 bi = &rx_ring->rx_bi[i]; in i40evf_alloc_rx_buffers_1buf()
954 struct i40e_rx_buffer *rx_bi; in i40e_clean_rx_irq_ps() local
977 rx_bi = &rx_ring->rx_bi[i]; in i40e_clean_rx_irq_ps()
978 skb = rx_bi->skb; in i40e_clean_rx_irq_ps()
991 rx_bi->dma, in i40e_clean_rx_irq_ps()
1010 prefetch(rx_bi->page); in i40e_clean_rx_irq_ps()
1011 rx_bi->skb = NULL; in i40e_clean_rx_irq_ps()
1020 memcpy(__skb_put(skb, len), rx_bi->hdr_buf, len); in i40e_clean_rx_irq_ps()
1027 rx_bi->page + rx_bi->page_offset, in i40e_clean_rx_irq_ps()
1029 rx_bi->page_offset += len; in i40e_clean_rx_irq_ps()
1036 rx_bi->page, in i40e_clean_rx_irq_ps()
1037 rx_bi->page_offset, in i40e_clean_rx_irq_ps()
1044 if ((page_count(rx_bi->page) == 1) && in i40e_clean_rx_irq_ps()
1045 (page_to_nid(rx_bi->page) == current_node)) in i40e_clean_rx_irq_ps()
1046 get_page(rx_bi->page); in i40e_clean_rx_irq_ps()
1048 rx_bi->page = NULL; in i40e_clean_rx_irq_ps()
1051 rx_bi->page_dma, in i40e_clean_rx_irq_ps()
1054 rx_bi->page_dma = 0; in i40e_clean_rx_irq_ps()
1062 next_buffer = &rx_ring->rx_bi[i]; in i40e_clean_rx_irq_ps()
1130 struct i40e_rx_buffer *rx_bi; in i40e_clean_rx_irq_1buf() local
1154 rx_bi = &rx_ring->rx_bi[i]; in i40e_clean_rx_irq_1buf()
1155 skb = rx_bi->skb; in i40e_clean_rx_irq_1buf()
1167 rx_bi->skb = NULL; in i40e_clean_rx_irq_1buf()
1174 dma_unmap_single(rx_ring->dev, rx_bi->dma, rx_ring->rx_buf_len, in i40e_clean_rx_irq_1buf()
1176 rx_bi->dma = 0; in i40e_clean_rx_irq_1buf()