rx_buffer 368 drivers/input/keyboard/applespi.c u8 *rx_buffer; rx_buffer 550 drivers/input/keyboard/applespi.c rd_t->rx_buf = applespi->rx_buffer; rx_buffer 1391 drivers/input/keyboard/applespi.c applespi_get_trace_fun(evt_type)(evt_type, PT_READ, applespi->rx_buffer, rx_buffer 1405 drivers/input/keyboard/applespi.c if (!applespi_verify_crc(applespi, applespi->rx_buffer, rx_buffer 1423 drivers/input/keyboard/applespi.c packet = (struct spi_packet *)applespi->rx_buffer; rx_buffer 1657 drivers/input/keyboard/applespi.c applespi->rx_buffer = devm_kmalloc(&spi->dev, APPLESPI_PACKET_SIZE, rx_buffer 1664 drivers/input/keyboard/applespi.c !applespi->rx_buffer || !applespi->msg_buf) rx_buffer 37 drivers/mailbox/mailbox-test.c char *rx_buffer; rx_buffer 207 drivers/mailbox/mailbox-test.c ptr = tdev->rx_buffer; rx_buffer 220 drivers/mailbox/mailbox-test.c memset(tdev->rx_buffer, 0, MBOX_MAX_MSG_LEN); rx_buffer 283 drivers/mailbox/mailbox-test.c memcpy_fromio(tdev->rx_buffer, tdev->rx_mmio, MBOX_MAX_MSG_LEN); rx_buffer 285 drivers/mailbox/mailbox-test.c tdev->rx_buffer, MBOX_MAX_MSG_LEN); rx_buffer 289 drivers/mailbox/mailbox-test.c memcpy(tdev->rx_buffer, message, MBOX_MAX_MSG_LEN); rx_buffer 397 drivers/mailbox/mailbox-test.c tdev->rx_buffer = devm_kzalloc(&pdev->dev, rx_buffer 399 drivers/mailbox/mailbox-test.c if (!tdev->rx_buffer) rx_buffer 106 drivers/media/dvb-core/dvb_ca_en50221.c struct dvb_ringbuffer rx_buffer; rx_buffer 162 drivers/media/dvb-core/dvb_ca_en50221.c vfree(ca->slot_info[i].rx_buffer.data); rx_buffer 646 drivers/media/dvb-core/dvb_ca_en50221.c if (!sl->rx_buffer.data) { rx_buffer 650 drivers/media/dvb-core/dvb_ca_en50221.c buf_free = dvb_ringbuffer_free(&sl->rx_buffer); rx_buffer 750 drivers/media/dvb-core/dvb_ca_en50221.c if (!sl->rx_buffer.data) { rx_buffer 754 drivers/media/dvb-core/dvb_ca_en50221.c dvb_ringbuffer_pkt_write(&sl->rx_buffer, buf, bytes_read); rx_buffer 1231 drivers/media/dvb-core/dvb_ca_en50221.c if (!sl->rx_buffer.data) { rx_buffer 1240 drivers/media/dvb-core/dvb_ca_en50221.c dvb_ringbuffer_init(&sl->rx_buffer, rxbuf, rx_buffer 1553 drivers/media/dvb-core/dvb_ca_en50221.c if (!sl->rx_buffer.data) rx_buffer 1556 drivers/media/dvb-core/dvb_ca_en50221.c idx = dvb_ringbuffer_pkt_next(&sl->rx_buffer, -1, &fraglen); rx_buffer 1558 drivers/media/dvb-core/dvb_ca_en50221.c dvb_ringbuffer_pkt_read(&sl->rx_buffer, idx, 0, hdr, 2); rx_buffer 1568 drivers/media/dvb-core/dvb_ca_en50221.c idx = dvb_ringbuffer_pkt_next(&sl->rx_buffer, idx, rx_buffer 1636 drivers/media/dvb-core/dvb_ca_en50221.c idx = dvb_ringbuffer_pkt_next(&sl->rx_buffer, -1, &fraglen); rx_buffer 1646 drivers/media/dvb-core/dvb_ca_en50221.c dvb_ringbuffer_pkt_read(&sl->rx_buffer, idx, 0, hdr, 2); rx_buffer 1657 drivers/media/dvb-core/dvb_ca_en50221.c dvb_ringbuffer_pkt_read_user(&sl->rx_buffer, rx_buffer 1672 drivers/media/dvb-core/dvb_ca_en50221.c idx2 = dvb_ringbuffer_pkt_next(&sl->rx_buffer, idx, &fraglen); rx_buffer 1674 drivers/media/dvb-core/dvb_ca_en50221.c dvb_ringbuffer_pkt_dispose(&sl->rx_buffer, idx); rx_buffer 1722 drivers/media/dvb-core/dvb_ca_en50221.c if (!sl->rx_buffer.data) { rx_buffer 1728 drivers/media/dvb-core/dvb_ca_en50221.c dvb_ringbuffer_flush(&sl->rx_buffer); rx_buffer 112 drivers/net/ethernet/brocade/bna/bfi_enet.h union bfi_addr_be_u rx_buffer; rx_buffer 79 drivers/net/ethernet/dec/tulip/xircom_cb.c __le32 *rx_buffer; rx_buffer 234 drivers/net/ethernet/dec/tulip/xircom_cb.c private->rx_buffer = dma_alloc_coherent(d, 8192, rx_buffer 237 drivers/net/ethernet/dec/tulip/xircom_cb.c if (private->rx_buffer == NULL) rx_buffer 292 drivers/net/ethernet/dec/tulip/xircom_cb.c dma_free_coherent(d, 8192, private->rx_buffer, private->rx_dma_handle); rx_buffer 318 drivers/net/ethernet/dec/tulip/xircom_cb.c dma_free_coherent(d, 8192, card->rx_buffer, card->rx_dma_handle); rx_buffer 340 drivers/net/ethernet/dec/tulip/xircom_cb.c card->rx_buffer[0], card->rx_buffer[4]); rx_buffer 566 drivers/net/ethernet/dec/tulip/xircom_cb.c BUG_ON(card->rx_buffer == NULL); rx_buffer 570 drivers/net/ethernet/dec/tulip/xircom_cb.c memset(card->rx_buffer, 0, 128); /* clear the descriptors */ rx_buffer 574 drivers/net/ethernet/dec/tulip/xircom_cb.c card->rx_buffer[i*4 + 0] = cpu_to_le32(0x80000000); rx_buffer 576 drivers/net/ethernet/dec/tulip/xircom_cb.c card->rx_buffer[i*4 + 1] = cpu_to_le32(1536); rx_buffer 578 drivers/net/ethernet/dec/tulip/xircom_cb.c card->rx_buffer[i*4 + 1] |= cpu_to_le32(1 << 25); rx_buffer 584 drivers/net/ethernet/dec/tulip/xircom_cb.c card->rx_buffer[i*4 + 2] = cpu_to_le32(address + bufferoffsets[i]); rx_buffer 586 drivers/net/ethernet/dec/tulip/xircom_cb.c card->rx_buffer[i*4 + 3] = 0; rx_buffer 1101 drivers/net/ethernet/dec/tulip/xircom_cb.c status = le32_to_cpu(card->rx_buffer[4*descnr]); rx_buffer 1123 drivers/net/ethernet/dec/tulip/xircom_cb.c &card->rx_buffer[bufferoffset / 4], rx_buffer 1133 drivers/net/ethernet/dec/tulip/xircom_cb.c card->rx_buffer[4*descnr] = cpu_to_le32(0x80000000); rx_buffer 111 drivers/net/ethernet/intel/fm10k/fm10k.h struct fm10k_rx_buffer *rx_buffer; rx_buffer 123 drivers/net/ethernet/intel/fm10k/fm10k_main.c bi = &rx_ring->rx_buffer[i]; rx_buffer 140 drivers/net/ethernet/intel/fm10k/fm10k_main.c bi = rx_ring->rx_buffer; rx_buffer 184 drivers/net/ethernet/intel/fm10k/fm10k_main.c new_buff = &rx_ring->rx_buffer[nta]; rx_buffer 205 drivers/net/ethernet/intel/fm10k/fm10k_main.c static bool fm10k_can_reuse_rx_page(struct fm10k_rx_buffer *rx_buffer, rx_buffer 219 drivers/net/ethernet/intel/fm10k/fm10k_main.c rx_buffer->page_offset ^= FM10K_RX_BUFSZ; rx_buffer 222 drivers/net/ethernet/intel/fm10k/fm10k_main.c rx_buffer->page_offset += truesize; rx_buffer 224 drivers/net/ethernet/intel/fm10k/fm10k_main.c if (rx_buffer->page_offset > (PAGE_SIZE - FM10K_RX_BUFSZ)) rx_buffer 251 drivers/net/ethernet/intel/fm10k/fm10k_main.c static bool fm10k_add_rx_frag(struct fm10k_rx_buffer *rx_buffer, rx_buffer 256 drivers/net/ethernet/intel/fm10k/fm10k_main.c struct page *page = rx_buffer->page; rx_buffer 257 drivers/net/ethernet/intel/fm10k/fm10k_main.c unsigned char *va = page_address(page) + rx_buffer->page_offset; rx_buffer 296 drivers/net/ethernet/intel/fm10k/fm10k_main.c return fm10k_can_reuse_rx_page(rx_buffer, page, truesize); rx_buffer 304 drivers/net/ethernet/intel/fm10k/fm10k_main.c struct fm10k_rx_buffer *rx_buffer; rx_buffer 307 drivers/net/ethernet/intel/fm10k/fm10k_main.c rx_buffer = &rx_ring->rx_buffer[rx_ring->next_to_clean]; rx_buffer 308 drivers/net/ethernet/intel/fm10k/fm10k_main.c page = rx_buffer->page; rx_buffer 313 drivers/net/ethernet/intel/fm10k/fm10k_main.c rx_buffer->page_offset; rx_buffer 338 drivers/net/ethernet/intel/fm10k/fm10k_main.c rx_buffer->dma, rx_buffer 339 drivers/net/ethernet/intel/fm10k/fm10k_main.c rx_buffer->page_offset, rx_buffer 344 drivers/net/ethernet/intel/fm10k/fm10k_main.c if (fm10k_add_rx_frag(rx_buffer, size, rx_desc, skb)) { rx_buffer 346 drivers/net/ethernet/intel/fm10k/fm10k_main.c fm10k_reuse_rx_page(rx_ring, rx_buffer); rx_buffer 349 drivers/net/ethernet/intel/fm10k/fm10k_main.c dma_unmap_page(rx_ring->dev, rx_buffer->dma, rx_buffer 354 drivers/net/ethernet/intel/fm10k/fm10k_main.c rx_buffer->page = NULL; rx_buffer 90 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c rx_ring->rx_buffer = vzalloc(size); rx_buffer 91 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c if (!rx_ring->rx_buffer) rx_buffer 107 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c vfree(rx_ring->rx_buffer); rx_buffer 108 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c rx_ring->rx_buffer = NULL; rx_buffer 253 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c if (!rx_ring->rx_buffer) rx_buffer 261 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c struct fm10k_rx_buffer *buffer = &rx_ring->rx_buffer[i]; rx_buffer 274 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c memset(rx_ring->rx_buffer, 0, size); rx_buffer 294 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c vfree(rx_ring->rx_buffer); rx_buffer 295 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c rx_ring->rx_buffer = NULL; rx_buffer 1265 drivers/net/ethernet/intel/i40e/i40e_txrx.c struct i40e_rx_buffer *rx_buffer; rx_buffer 1275 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer = &rx_ring->rx_bi[ntc++]; rx_buffer 1287 drivers/net/ethernet/intel/i40e/i40e_txrx.c return rx_buffer; rx_buffer 1889 drivers/net/ethernet/intel/i40e/i40e_txrx.c static bool i40e_can_reuse_rx_page(struct i40e_rx_buffer *rx_buffer) rx_buffer 1891 drivers/net/ethernet/intel/i40e/i40e_txrx.c unsigned int pagecnt_bias = rx_buffer->pagecnt_bias; rx_buffer 1892 drivers/net/ethernet/intel/i40e/i40e_txrx.c struct page *page = rx_buffer->page; rx_buffer 1905 drivers/net/ethernet/intel/i40e/i40e_txrx.c if (rx_buffer->page_offset > I40E_LAST_OFFSET) rx_buffer 1915 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer->pagecnt_bias = USHRT_MAX; rx_buffer 1934 drivers/net/ethernet/intel/i40e/i40e_txrx.c struct i40e_rx_buffer *rx_buffer, rx_buffer 1944 drivers/net/ethernet/intel/i40e/i40e_txrx.c skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, rx_buffer->page, rx_buffer 1945 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer->page_offset, size, truesize); rx_buffer 1949 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer->page_offset ^= truesize; rx_buffer 1951 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer->page_offset += truesize; rx_buffer 1966 drivers/net/ethernet/intel/i40e/i40e_txrx.c struct i40e_rx_buffer *rx_buffer; rx_buffer 1968 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer = &rx_ring->rx_bi[rx_ring->next_to_clean]; rx_buffer 1969 drivers/net/ethernet/intel/i40e/i40e_txrx.c prefetchw(rx_buffer->page); rx_buffer 1973 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer->dma, rx_buffer 1974 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer->page_offset, rx_buffer 1979 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer->pagecnt_bias--; rx_buffer 1981 drivers/net/ethernet/intel/i40e/i40e_txrx.c return rx_buffer; rx_buffer 1995 drivers/net/ethernet/intel/i40e/i40e_txrx.c struct i40e_rx_buffer *rx_buffer, rx_buffer 2048 drivers/net/ethernet/intel/i40e/i40e_txrx.c skb_add_rx_frag(skb, 0, rx_buffer->page, rx_buffer 2049 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer->page_offset + headlen, rx_buffer 2054 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer->page_offset ^= truesize; rx_buffer 2056 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer->page_offset += truesize; rx_buffer 2060 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer->pagecnt_bias++; rx_buffer 2076 drivers/net/ethernet/intel/i40e/i40e_txrx.c struct i40e_rx_buffer *rx_buffer, rx_buffer 2111 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer->page_offset ^= truesize; rx_buffer 2113 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer->page_offset += truesize; rx_buffer 2128 drivers/net/ethernet/intel/i40e/i40e_txrx.c struct i40e_rx_buffer *rx_buffer) rx_buffer 2130 drivers/net/ethernet/intel/i40e/i40e_txrx.c if (i40e_can_reuse_rx_page(rx_buffer)) { rx_buffer 2132 drivers/net/ethernet/intel/i40e/i40e_txrx.c i40e_reuse_rx_page(rx_ring, rx_buffer); rx_buffer 2135 drivers/net/ethernet/intel/i40e/i40e_txrx.c dma_unmap_page_attrs(rx_ring->dev, rx_buffer->dma, rx_buffer 2138 drivers/net/ethernet/intel/i40e/i40e_txrx.c __page_frag_cache_drain(rx_buffer->page, rx_buffer 2139 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer->pagecnt_bias); rx_buffer 2141 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer->page = NULL; rx_buffer 2246 drivers/net/ethernet/intel/i40e/i40e_txrx.c struct i40e_rx_buffer *rx_buffer, rx_buffer 2252 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer->page_offset ^= truesize; rx_buffer 2256 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer->page_offset += truesize; rx_buffer 2341 drivers/net/ethernet/intel/i40e/i40e_txrx.c struct i40e_rx_buffer *rx_buffer; rx_buffer 2368 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer = i40e_clean_programming_status(rx_ring, rx_desc, rx_buffer 2370 drivers/net/ethernet/intel/i40e/i40e_txrx.c if (unlikely(rx_buffer)) { rx_buffer 2371 drivers/net/ethernet/intel/i40e/i40e_txrx.c i40e_reuse_rx_page(rx_ring, rx_buffer); rx_buffer 2382 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer = i40e_get_rx_buffer(rx_ring, size); rx_buffer 2386 drivers/net/ethernet/intel/i40e/i40e_txrx.c xdp.data = page_address(rx_buffer->page) + rx_buffer 2387 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer->page_offset; rx_buffer 2401 drivers/net/ethernet/intel/i40e/i40e_txrx.c i40e_rx_buffer_flip(rx_ring, rx_buffer, size); rx_buffer 2403 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer->pagecnt_bias++; rx_buffer 2408 drivers/net/ethernet/intel/i40e/i40e_txrx.c i40e_add_rx_frag(rx_ring, rx_buffer, skb, size); rx_buffer 2410 drivers/net/ethernet/intel/i40e/i40e_txrx.c skb = i40e_build_skb(rx_ring, rx_buffer, &xdp); rx_buffer 2412 drivers/net/ethernet/intel/i40e/i40e_txrx.c skb = i40e_construct_skb(rx_ring, rx_buffer, &xdp); rx_buffer 2418 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_buffer->pagecnt_bias++; rx_buffer 2422 drivers/net/ethernet/intel/i40e/i40e_txrx.c i40e_put_rx_buffer(rx_ring, rx_buffer); rx_buffer 1184 drivers/net/ethernet/intel/iavf/iavf_txrx.c static bool iavf_can_reuse_rx_page(struct iavf_rx_buffer *rx_buffer) rx_buffer 1186 drivers/net/ethernet/intel/iavf/iavf_txrx.c unsigned int pagecnt_bias = rx_buffer->pagecnt_bias; rx_buffer 1187 drivers/net/ethernet/intel/iavf/iavf_txrx.c struct page *page = rx_buffer->page; rx_buffer 1200 drivers/net/ethernet/intel/iavf/iavf_txrx.c if (rx_buffer->page_offset > IAVF_LAST_OFFSET) rx_buffer 1210 drivers/net/ethernet/intel/iavf/iavf_txrx.c rx_buffer->pagecnt_bias = USHRT_MAX; rx_buffer 1229 drivers/net/ethernet/intel/iavf/iavf_txrx.c struct iavf_rx_buffer *rx_buffer, rx_buffer 1242 drivers/net/ethernet/intel/iavf/iavf_txrx.c skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, rx_buffer->page, rx_buffer 1243 drivers/net/ethernet/intel/iavf/iavf_txrx.c rx_buffer->page_offset, size, truesize); rx_buffer 1247 drivers/net/ethernet/intel/iavf/iavf_txrx.c rx_buffer->page_offset ^= truesize; rx_buffer 1249 drivers/net/ethernet/intel/iavf/iavf_txrx.c rx_buffer->page_offset += truesize; rx_buffer 1264 drivers/net/ethernet/intel/iavf/iavf_txrx.c struct iavf_rx_buffer *rx_buffer; rx_buffer 1269 drivers/net/ethernet/intel/iavf/iavf_txrx.c rx_buffer = &rx_ring->rx_bi[rx_ring->next_to_clean]; rx_buffer 1270 drivers/net/ethernet/intel/iavf/iavf_txrx.c prefetchw(rx_buffer->page); rx_buffer 1274 drivers/net/ethernet/intel/iavf/iavf_txrx.c rx_buffer->dma, rx_buffer 1275 drivers/net/ethernet/intel/iavf/iavf_txrx.c rx_buffer->page_offset, rx_buffer 1280 drivers/net/ethernet/intel/iavf/iavf_txrx.c rx_buffer->pagecnt_bias--; rx_buffer 1282 drivers/net/ethernet/intel/iavf/iavf_txrx.c return rx_buffer; rx_buffer 1296 drivers/net/ethernet/intel/iavf/iavf_txrx.c struct iavf_rx_buffer *rx_buffer, rx_buffer 1308 drivers/net/ethernet/intel/iavf/iavf_txrx.c if (!rx_buffer) rx_buffer 1311 drivers/net/ethernet/intel/iavf/iavf_txrx.c va = page_address(rx_buffer->page) + rx_buffer->page_offset; rx_buffer 1335 drivers/net/ethernet/intel/iavf/iavf_txrx.c skb_add_rx_frag(skb, 0, rx_buffer->page, rx_buffer 1336 drivers/net/ethernet/intel/iavf/iavf_txrx.c rx_buffer->page_offset + headlen, rx_buffer 1341 drivers/net/ethernet/intel/iavf/iavf_txrx.c rx_buffer->page_offset ^= truesize; rx_buffer 1343 drivers/net/ethernet/intel/iavf/iavf_txrx.c rx_buffer->page_offset += truesize; rx_buffer 1347 drivers/net/ethernet/intel/iavf/iavf_txrx.c rx_buffer->pagecnt_bias++; rx_buffer 1363 drivers/net/ethernet/intel/iavf/iavf_txrx.c struct iavf_rx_buffer *rx_buffer, rx_buffer 1375 drivers/net/ethernet/intel/iavf/iavf_txrx.c if (!rx_buffer) rx_buffer 1378 drivers/net/ethernet/intel/iavf/iavf_txrx.c va = page_address(rx_buffer->page) + rx_buffer->page_offset; rx_buffer 1394 drivers/net/ethernet/intel/iavf/iavf_txrx.c rx_buffer->page_offset ^= truesize; rx_buffer 1396 drivers/net/ethernet/intel/iavf/iavf_txrx.c rx_buffer->page_offset += truesize; rx_buffer 1411 drivers/net/ethernet/intel/iavf/iavf_txrx.c struct iavf_rx_buffer *rx_buffer) rx_buffer 1413 drivers/net/ethernet/intel/iavf/iavf_txrx.c if (!rx_buffer) rx_buffer 1416 drivers/net/ethernet/intel/iavf/iavf_txrx.c if (iavf_can_reuse_rx_page(rx_buffer)) { rx_buffer 1418 drivers/net/ethernet/intel/iavf/iavf_txrx.c iavf_reuse_rx_page(rx_ring, rx_buffer); rx_buffer 1422 drivers/net/ethernet/intel/iavf/iavf_txrx.c dma_unmap_page_attrs(rx_ring->dev, rx_buffer->dma, rx_buffer 1425 drivers/net/ethernet/intel/iavf/iavf_txrx.c __page_frag_cache_drain(rx_buffer->page, rx_buffer 1426 drivers/net/ethernet/intel/iavf/iavf_txrx.c rx_buffer->pagecnt_bias); rx_buffer 1430 drivers/net/ethernet/intel/iavf/iavf_txrx.c rx_buffer->page = NULL; rx_buffer 1486 drivers/net/ethernet/intel/iavf/iavf_txrx.c struct iavf_rx_buffer *rx_buffer; rx_buffer 1522 drivers/net/ethernet/intel/iavf/iavf_txrx.c rx_buffer = iavf_get_rx_buffer(rx_ring, size); rx_buffer 1526 drivers/net/ethernet/intel/iavf/iavf_txrx.c iavf_add_rx_frag(rx_ring, rx_buffer, skb, size); rx_buffer 1528 drivers/net/ethernet/intel/iavf/iavf_txrx.c skb = iavf_build_skb(rx_ring, rx_buffer, size); rx_buffer 1530 drivers/net/ethernet/intel/iavf/iavf_txrx.c skb = iavf_construct_skb(rx_ring, rx_buffer, size); rx_buffer 1535 drivers/net/ethernet/intel/iavf/iavf_txrx.c if (rx_buffer) rx_buffer 1536 drivers/net/ethernet/intel/iavf/iavf_txrx.c rx_buffer->pagecnt_bias++; rx_buffer 1540 drivers/net/ethernet/intel/iavf/iavf_txrx.c iavf_put_rx_buffer(rx_ring, rx_buffer); rx_buffer 1781 drivers/net/ethernet/intel/igb/igb_ethtool.c static int igb_check_lbtest_frame(struct igb_rx_buffer *rx_buffer, rx_buffer 1789 drivers/net/ethernet/intel/igb/igb_ethtool.c data = kmap(rx_buffer->page); rx_buffer 1796 drivers/net/ethernet/intel/igb/igb_ethtool.c kunmap(rx_buffer->page); rx_buffer 7943 drivers/net/ethernet/intel/igb/igb_main.c static bool igb_can_reuse_rx_page(struct igb_rx_buffer *rx_buffer) rx_buffer 7945 drivers/net/ethernet/intel/igb/igb_main.c unsigned int pagecnt_bias = rx_buffer->pagecnt_bias; rx_buffer 7946 drivers/net/ethernet/intel/igb/igb_main.c struct page *page = rx_buffer->page; rx_buffer 7960 drivers/net/ethernet/intel/igb/igb_main.c if (rx_buffer->page_offset > IGB_LAST_OFFSET) rx_buffer 7970 drivers/net/ethernet/intel/igb/igb_main.c rx_buffer->pagecnt_bias = USHRT_MAX; rx_buffer 7986 drivers/net/ethernet/intel/igb/igb_main.c struct igb_rx_buffer *rx_buffer, rx_buffer 7997 drivers/net/ethernet/intel/igb/igb_main.c skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, rx_buffer->page, rx_buffer 7998 drivers/net/ethernet/intel/igb/igb_main.c rx_buffer->page_offset, size, truesize); rx_buffer 8000 drivers/net/ethernet/intel/igb/igb_main.c rx_buffer->page_offset ^= truesize; rx_buffer 8002 drivers/net/ethernet/intel/igb/igb_main.c rx_buffer->page_offset += truesize; rx_buffer 8007 drivers/net/ethernet/intel/igb/igb_main.c struct igb_rx_buffer *rx_buffer, rx_buffer 8011 drivers/net/ethernet/intel/igb/igb_main.c void *va = page_address(rx_buffer->page) + rx_buffer->page_offset; rx_buffer 8048 drivers/net/ethernet/intel/igb/igb_main.c skb_add_rx_frag(skb, 0, rx_buffer->page, rx_buffer 8049 drivers/net/ethernet/intel/igb/igb_main.c (va + headlen) - page_address(rx_buffer->page), rx_buffer 8052 drivers/net/ethernet/intel/igb/igb_main.c rx_buffer->page_offset ^= truesize; rx_buffer 8054 drivers/net/ethernet/intel/igb/igb_main.c rx_buffer->page_offset += truesize; rx_buffer 8057 drivers/net/ethernet/intel/igb/igb_main.c rx_buffer->pagecnt_bias++; rx_buffer 8064 drivers/net/ethernet/intel/igb/igb_main.c struct igb_rx_buffer *rx_buffer, rx_buffer 8068 drivers/net/ethernet/intel/igb/igb_main.c void *va = page_address(rx_buffer->page) + rx_buffer->page_offset; rx_buffer 8100 drivers/net/ethernet/intel/igb/igb_main.c rx_buffer->page_offset ^= truesize; rx_buffer 8102 drivers/net/ethernet/intel/igb/igb_main.c rx_buffer->page_offset += truesize; rx_buffer 8265 drivers/net/ethernet/intel/igb/igb_main.c struct igb_rx_buffer *rx_buffer; rx_buffer 8267 drivers/net/ethernet/intel/igb/igb_main.c rx_buffer = &rx_ring->rx_buffer_info[rx_ring->next_to_clean]; rx_buffer 8268 drivers/net/ethernet/intel/igb/igb_main.c prefetchw(rx_buffer->page); rx_buffer 8272 drivers/net/ethernet/intel/igb/igb_main.c rx_buffer->dma, rx_buffer 8273 drivers/net/ethernet/intel/igb/igb_main.c rx_buffer->page_offset, rx_buffer 8277 drivers/net/ethernet/intel/igb/igb_main.c rx_buffer->pagecnt_bias--; rx_buffer 8279 drivers/net/ethernet/intel/igb/igb_main.c return rx_buffer; rx_buffer 8283 drivers/net/ethernet/intel/igb/igb_main.c struct igb_rx_buffer *rx_buffer) rx_buffer 8285 drivers/net/ethernet/intel/igb/igb_main.c if (igb_can_reuse_rx_page(rx_buffer)) { rx_buffer 8287 drivers/net/ethernet/intel/igb/igb_main.c igb_reuse_rx_page(rx_ring, rx_buffer); rx_buffer 8292 drivers/net/ethernet/intel/igb/igb_main.c dma_unmap_page_attrs(rx_ring->dev, rx_buffer->dma, rx_buffer 8295 drivers/net/ethernet/intel/igb/igb_main.c __page_frag_cache_drain(rx_buffer->page, rx_buffer 8296 drivers/net/ethernet/intel/igb/igb_main.c rx_buffer->pagecnt_bias); rx_buffer 8300 drivers/net/ethernet/intel/igb/igb_main.c rx_buffer->page = NULL; rx_buffer 8312 drivers/net/ethernet/intel/igb/igb_main.c struct igb_rx_buffer *rx_buffer; rx_buffer 8332 drivers/net/ethernet/intel/igb/igb_main.c rx_buffer = igb_get_rx_buffer(rx_ring, size); rx_buffer 8336 drivers/net/ethernet/intel/igb/igb_main.c igb_add_rx_frag(rx_ring, rx_buffer, skb, size); rx_buffer 8338 drivers/net/ethernet/intel/igb/igb_main.c skb = igb_build_skb(rx_ring, rx_buffer, rx_desc, size); rx_buffer 8340 drivers/net/ethernet/intel/igb/igb_main.c skb = igb_construct_skb(rx_ring, rx_buffer, rx_buffer 8346 drivers/net/ethernet/intel/igb/igb_main.c rx_buffer->pagecnt_bias++; rx_buffer 8350 drivers/net/ethernet/intel/igb/igb_main.c igb_put_rx_buffer(rx_ring, rx_buffer); rx_buffer 1200 drivers/net/ethernet/intel/igc/igc_main.c struct igc_rx_buffer *rx_buffer; rx_buffer 1202 drivers/net/ethernet/intel/igc/igc_main.c rx_buffer = &rx_ring->rx_buffer_info[rx_ring->next_to_clean]; rx_buffer 1203 drivers/net/ethernet/intel/igc/igc_main.c prefetchw(rx_buffer->page); rx_buffer 1207 drivers/net/ethernet/intel/igc/igc_main.c rx_buffer->dma, rx_buffer 1208 drivers/net/ethernet/intel/igc/igc_main.c rx_buffer->page_offset, rx_buffer 1212 drivers/net/ethernet/intel/igc/igc_main.c rx_buffer->pagecnt_bias--; rx_buffer 1214 drivers/net/ethernet/intel/igc/igc_main.c return rx_buffer; rx_buffer 1227 drivers/net/ethernet/intel/igc/igc_main.c struct igc_rx_buffer *rx_buffer, rx_buffer 1234 drivers/net/ethernet/intel/igc/igc_main.c skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, rx_buffer->page, rx_buffer 1235 drivers/net/ethernet/intel/igc/igc_main.c rx_buffer->page_offset, size, truesize); rx_buffer 1236 drivers/net/ethernet/intel/igc/igc_main.c rx_buffer->page_offset ^= truesize; rx_buffer 1241 drivers/net/ethernet/intel/igc/igc_main.c skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, rx_buffer->page, rx_buffer 1242 drivers/net/ethernet/intel/igc/igc_main.c rx_buffer->page_offset, size, truesize); rx_buffer 1243 drivers/net/ethernet/intel/igc/igc_main.c rx_buffer->page_offset += truesize; rx_buffer 1248 drivers/net/ethernet/intel/igc/igc_main.c struct igc_rx_buffer *rx_buffer, rx_buffer 1252 drivers/net/ethernet/intel/igc/igc_main.c void *va = page_address(rx_buffer->page) + rx_buffer->page_offset; rx_buffer 1278 drivers/net/ethernet/intel/igc/igc_main.c rx_buffer->page_offset ^= truesize; rx_buffer 1280 drivers/net/ethernet/intel/igc/igc_main.c rx_buffer->page_offset += truesize; rx_buffer 1287 drivers/net/ethernet/intel/igc/igc_main.c struct igc_rx_buffer *rx_buffer, rx_buffer 1291 drivers/net/ethernet/intel/igc/igc_main.c void *va = page_address(rx_buffer->page) + rx_buffer->page_offset; rx_buffer 1322 drivers/net/ethernet/intel/igc/igc_main.c skb_add_rx_frag(skb, 0, rx_buffer->page, rx_buffer 1323 drivers/net/ethernet/intel/igc/igc_main.c (va + headlen) - page_address(rx_buffer->page), rx_buffer 1326 drivers/net/ethernet/intel/igc/igc_main.c rx_buffer->page_offset ^= truesize; rx_buffer 1328 drivers/net/ethernet/intel/igc/igc_main.c rx_buffer->page_offset += truesize; rx_buffer 1331 drivers/net/ethernet/intel/igc/igc_main.c rx_buffer->pagecnt_bias++; rx_buffer 1371 drivers/net/ethernet/intel/igc/igc_main.c static bool igc_can_reuse_rx_page(struct igc_rx_buffer *rx_buffer) rx_buffer 1373 drivers/net/ethernet/intel/igc/igc_main.c unsigned int pagecnt_bias = rx_buffer->pagecnt_bias; rx_buffer 1374 drivers/net/ethernet/intel/igc/igc_main.c struct page *page = rx_buffer->page; rx_buffer 1388 drivers/net/ethernet/intel/igc/igc_main.c if (rx_buffer->page_offset > IGC_LAST_OFFSET) rx_buffer 1398 drivers/net/ethernet/intel/igc/igc_main.c rx_buffer->pagecnt_bias = USHRT_MAX; rx_buffer 1468 drivers/net/ethernet/intel/igc/igc_main.c struct igc_rx_buffer *rx_buffer) rx_buffer 1470 drivers/net/ethernet/intel/igc/igc_main.c if (igc_can_reuse_rx_page(rx_buffer)) { rx_buffer 1472 drivers/net/ethernet/intel/igc/igc_main.c igc_reuse_rx_page(rx_ring, rx_buffer); rx_buffer 1477 drivers/net/ethernet/intel/igc/igc_main.c dma_unmap_page_attrs(rx_ring->dev, rx_buffer->dma, rx_buffer 1480 drivers/net/ethernet/intel/igc/igc_main.c __page_frag_cache_drain(rx_buffer->page, rx_buffer 1481 drivers/net/ethernet/intel/igc/igc_main.c rx_buffer->pagecnt_bias); rx_buffer 1485 drivers/net/ethernet/intel/igc/igc_main.c rx_buffer->page = NULL; rx_buffer 1566 drivers/net/ethernet/intel/igc/igc_main.c struct igc_rx_buffer *rx_buffer; rx_buffer 1586 drivers/net/ethernet/intel/igc/igc_main.c rx_buffer = igc_get_rx_buffer(rx_ring, size); rx_buffer 1590 drivers/net/ethernet/intel/igc/igc_main.c igc_add_rx_frag(rx_ring, rx_buffer, skb, size); rx_buffer 1592 drivers/net/ethernet/intel/igc/igc_main.c skb = igc_build_skb(rx_ring, rx_buffer, rx_desc, size); rx_buffer 1594 drivers/net/ethernet/intel/igc/igc_main.c skb = igc_construct_skb(rx_ring, rx_buffer, rx_buffer 1600 drivers/net/ethernet/intel/igc/igc_main.c rx_buffer->pagecnt_bias++; rx_buffer 1604 drivers/net/ethernet/intel/igc/igc_main.c igc_put_rx_buffer(rx_ring, rx_buffer); rx_buffer 1866 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c static bool ixgbe_check_lbtest_frame(struct ixgbe_rx_buffer *rx_buffer, rx_buffer 1874 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c data = kmap(rx_buffer->page) + rx_buffer->page_offset; rx_buffer 1881 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c kunmap(rx_buffer->page); rx_buffer 1928 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c struct ixgbe_rx_buffer *rx_buffer; rx_buffer 1931 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c rx_buffer = &rx_ring->rx_buffer_info[rx_ntc]; rx_buffer 1935 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c rx_buffer->dma, rx_buffer 1940 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c if (ixgbe_check_lbtest_frame(rx_buffer, size)) rx_buffer 1947 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c rx_buffer->dma, rx_buffer 1950 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c static bool ixgbe_can_reuse_rx_page(struct ixgbe_rx_buffer *rx_buffer) rx_buffer 1952 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c unsigned int pagecnt_bias = rx_buffer->pagecnt_bias; rx_buffer 1953 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct page *page = rx_buffer->page; rx_buffer 1971 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (rx_buffer->page_offset > IXGBE_LAST_OFFSET) rx_buffer 1981 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer->pagecnt_bias = USHRT_MAX; rx_buffer 2003 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_rx_buffer *rx_buffer, rx_buffer 2014 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, rx_buffer->page, rx_buffer 2015 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer->page_offset, size, truesize); rx_buffer 2017 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer->page_offset ^= truesize; rx_buffer 2019 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer->page_offset += truesize; rx_buffer 2028 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_rx_buffer *rx_buffer; rx_buffer 2030 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer = &rx_ring->rx_buffer_info[rx_ring->next_to_clean]; rx_buffer 2031 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c prefetchw(rx_buffer->page); rx_buffer 2032 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c *skb = rx_buffer->skb; rx_buffer 2048 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer->dma, rx_buffer 2049 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer->page_offset, rx_buffer 2053 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer->pagecnt_bias--; rx_buffer 2055 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c return rx_buffer; rx_buffer 2059 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_rx_buffer *rx_buffer, rx_buffer 2062 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (ixgbe_can_reuse_rx_page(rx_buffer)) { rx_buffer 2064 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_reuse_rx_page(rx_ring, rx_buffer); rx_buffer 2066 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (!IS_ERR(skb) && IXGBE_CB(skb)->dma == rx_buffer->dma) { rx_buffer 2071 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c dma_unmap_page_attrs(rx_ring->dev, rx_buffer->dma, rx_buffer 2076 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c __page_frag_cache_drain(rx_buffer->page, rx_buffer 2077 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer->pagecnt_bias); rx_buffer 2081 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer->page = NULL; rx_buffer 2082 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer->skb = NULL; rx_buffer 2086 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_rx_buffer *rx_buffer, rx_buffer 2127 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c IXGBE_CB(skb)->dma = rx_buffer->dma; rx_buffer 2129 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c skb_add_rx_frag(skb, 0, rx_buffer->page, rx_buffer 2130 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c xdp->data - page_address(rx_buffer->page), rx_buffer 2133 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer->page_offset ^= truesize; rx_buffer 2135 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer->page_offset += truesize; rx_buffer 2140 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer->pagecnt_bias++; rx_buffer 2147 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_rx_buffer *rx_buffer, rx_buffer 2184 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c IXGBE_CB(skb)->dma = rx_buffer->dma; rx_buffer 2188 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer->page_offset ^= truesize; rx_buffer 2190 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer->page_offset += truesize; rx_buffer 2248 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_rx_buffer *rx_buffer, rx_buffer 2254 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer->page_offset ^= truesize; rx_buffer 2260 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer->page_offset += truesize; rx_buffer 2295 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_rx_buffer *rx_buffer; rx_buffer 2316 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer = ixgbe_get_rx_buffer(rx_ring, rx_desc, &skb, size); rx_buffer 2320 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c xdp.data = page_address(rx_buffer->page) + rx_buffer 2321 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer->page_offset; rx_buffer 2335 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_rx_buffer_flip(rx_ring, rx_buffer, size); rx_buffer 2337 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer->pagecnt_bias++; rx_buffer 2342 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_add_rx_frag(rx_ring, rx_buffer, skb, size); rx_buffer 2344 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c skb = ixgbe_build_skb(rx_ring, rx_buffer, rx_buffer 2347 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c skb = ixgbe_construct_skb(rx_ring, rx_buffer, rx_buffer 2354 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer->pagecnt_bias++; rx_buffer 2358 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_put_rx_buffer(rx_ring, rx_buffer, skb); rx_buffer 5291 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_rx_buffer *rx_buffer = &rx_ring->rx_buffer_info[i]; rx_buffer 5300 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (rx_buffer->skb) { rx_buffer 5301 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct sk_buff *skb = rx_buffer->skb; rx_buffer 5315 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer->dma, rx_buffer 5316 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer->page_offset, rx_buffer 5321 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c dma_unmap_page_attrs(rx_ring->dev, rx_buffer->dma, rx_buffer 5325 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c __page_frag_cache_drain(rx_buffer->page, rx_buffer 5326 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer->pagecnt_bias); rx_buffer 5329 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer++; rx_buffer 5332 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c rx_buffer = rx_ring->rx_buffer_info; rx_buffer 114 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c static bool ixgbevf_can_reuse_rx_page(struct ixgbevf_rx_buffer *rx_buffer); rx_buffer 533 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c struct ixgbevf_rx_buffer *rx_buffer; rx_buffer 535 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer = &rx_ring->rx_buffer_info[rx_ring->next_to_clean]; rx_buffer 536 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c prefetchw(rx_buffer->page); rx_buffer 540 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer->dma, rx_buffer 541 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer->page_offset, rx_buffer 545 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer->pagecnt_bias--; rx_buffer 547 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c return rx_buffer; rx_buffer 551 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c struct ixgbevf_rx_buffer *rx_buffer, rx_buffer 554 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c if (ixgbevf_can_reuse_rx_page(rx_buffer)) { rx_buffer 556 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_reuse_rx_page(rx_ring, rx_buffer); rx_buffer 562 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c dma_unmap_page_attrs(rx_ring->dev, rx_buffer->dma, rx_buffer 566 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c __page_frag_cache_drain(rx_buffer->page, rx_buffer 567 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer->pagecnt_bias); rx_buffer 571 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer->page = NULL; rx_buffer 791 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c static bool ixgbevf_can_reuse_rx_page(struct ixgbevf_rx_buffer *rx_buffer) rx_buffer 793 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c unsigned int pagecnt_bias = rx_buffer->pagecnt_bias; rx_buffer 794 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c struct page *page = rx_buffer->page; rx_buffer 808 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c if (rx_buffer->page_offset > IXGBEVF_LAST_OFFSET) rx_buffer 819 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer->pagecnt_bias = USHRT_MAX; rx_buffer 835 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c struct ixgbevf_rx_buffer *rx_buffer, rx_buffer 846 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, rx_buffer->page, rx_buffer 847 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer->page_offset, size, truesize); rx_buffer 849 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer->page_offset ^= truesize; rx_buffer 851 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer->page_offset += truesize; rx_buffer 857 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c struct ixgbevf_rx_buffer *rx_buffer, rx_buffer 909 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c skb_add_rx_frag(skb, 0, rx_buffer->page, rx_buffer 911 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c page_address(rx_buffer->page), rx_buffer 914 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer->page_offset ^= truesize; rx_buffer 916 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer->page_offset += truesize; rx_buffer 919 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer->pagecnt_bias++; rx_buffer 934 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c struct ixgbevf_rx_buffer *rx_buffer, rx_buffer 971 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer->page_offset ^= truesize; rx_buffer 973 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer->page_offset += truesize; rx_buffer 1099 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c struct ixgbevf_rx_buffer *rx_buffer, rx_buffer 1105 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer->page_offset ^= truesize; rx_buffer 1111 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer->page_offset += truesize; rx_buffer 1129 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c struct ixgbevf_rx_buffer *rx_buffer; rx_buffer 1150 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer = ixgbevf_get_rx_buffer(rx_ring, size); rx_buffer 1154 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c xdp.data = page_address(rx_buffer->page) + rx_buffer 1155 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer->page_offset; rx_buffer 1167 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_rx_buffer_flip(rx_ring, rx_buffer, rx_buffer 1170 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer->pagecnt_bias++; rx_buffer 1175 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_add_rx_frag(rx_ring, rx_buffer, skb, size); rx_buffer 1177 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c skb = ixgbevf_build_skb(rx_ring, rx_buffer, rx_buffer 1180 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c skb = ixgbevf_construct_skb(rx_ring, rx_buffer, rx_buffer 1187 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer->pagecnt_bias++; rx_buffer 1191 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_put_rx_buffer(rx_ring, rx_buffer, skb); rx_buffer 2339 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c struct ixgbevf_rx_buffer *rx_buffer; rx_buffer 2341 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer = &rx_ring->rx_buffer_info[i]; rx_buffer 2347 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer->dma, rx_buffer 2348 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer->page_offset, rx_buffer 2354 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer->dma, rx_buffer 2359 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c __page_frag_cache_drain(rx_buffer->page, rx_buffer 2360 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c rx_buffer->pagecnt_bias); rx_buffer 426 drivers/net/ethernet/oki-semi/pch_gbe/pch_gbe.h unsigned char *rx_buffer; rx_buffer 1381 drivers/net/ethernet/oki-semi/pch_gbe/pch_gbe_main.c buffer_info->rx_buffer, rx_buffer 1437 drivers/net/ethernet/oki-semi/pch_gbe/pch_gbe_main.c buffer_info->rx_buffer = rx_ring->rx_buff_pool + bufsz * i; rx_buffer 1682 drivers/net/ethernet/oki-semi/pch_gbe/pch_gbe_main.c memcpy(skb->data, buffer_info->rx_buffer, length); rx_buffer 389 drivers/net/ethernet/qualcomm/qca_spi.c bytes_read = qcaspi_read_legacy(qca, qca->rx_buffer, rx_buffer 392 drivers/net/ethernet/qualcomm/qca_spi.c bytes_read = qcaspi_read_burst(qca, qca->rx_buffer, rx_buffer 406 drivers/net/ethernet/qualcomm/qca_spi.c cp = qca->rx_buffer; rx_buffer 817 drivers/net/ethernet/qualcomm/qca_spi.c qca->rx_buffer = kmalloc(qca->buffer_size, GFP_KERNEL); rx_buffer 818 drivers/net/ethernet/qualcomm/qca_spi.c if (!qca->rx_buffer) rx_buffer 824 drivers/net/ethernet/qualcomm/qca_spi.c kfree(qca->rx_buffer); rx_buffer 837 drivers/net/ethernet/qualcomm/qca_spi.c kfree(qca->rx_buffer); rx_buffer 88 drivers/net/ethernet/qualcomm/qca_spi.h u8 *rx_buffer; rx_buffer 181 drivers/net/ethernet/ti/tlan.h u8 *rx_buffer; rx_buffer 293 drivers/net/wan/fsl_ucc_hdlc.c priv->rx_buffer = bd_buffer; rx_buffer 546 drivers/net/wan/fsl_ucc_hdlc.c bdbuffer = priv->rx_buffer + rx_buffer 783 drivers/net/wan/fsl_ucc_hdlc.c if (priv->rx_buffer) { rx_buffer 786 drivers/net/wan/fsl_ucc_hdlc.c priv->rx_buffer, priv->dma_rx_addr); rx_buffer 787 drivers/net/wan/fsl_ucc_hdlc.c priv->rx_buffer = NULL; rx_buffer 80 drivers/net/wan/fsl_ucc_hdlc.h u8 *rx_buffer; rx_buffer 138 drivers/net/wireless/ti/wl1251/rx.c u8 *rx_buffer, beacon = 0; rx_buffer 165 drivers/net/wireless/ti/wl1251/rx.c rx_buffer = skb_put(skb, length); rx_buffer 166 drivers/net/wireless/ti/wl1251/rx.c wl1251_mem_read(wl, rx_packet_ring_addr, rx_buffer, length); rx_buffer 77 drivers/spi/spidev.c u8 *rx_buffer; rx_buffer 130 drivers/spi/spidev.c .rx_buf = spidev->rx_buffer, rx_buffer 161 drivers/spi/spidev.c missing = copy_to_user(buf, spidev->rx_buffer, status); rx_buffer 219 drivers/spi/spidev.c rx_buf = spidev->rx_buffer; rx_buffer 293 drivers/spi/spidev.c rx_buf = spidev->rx_buffer; rx_buffer 581 drivers/spi/spidev.c if (!spidev->rx_buffer) { rx_buffer 582 drivers/spi/spidev.c spidev->rx_buffer = kmalloc(bufsiz, GFP_KERNEL); rx_buffer 583 drivers/spi/spidev.c if (!spidev->rx_buffer) { rx_buffer 621 drivers/spi/spidev.c kfree(spidev->rx_buffer); rx_buffer 622 drivers/spi/spidev.c spidev->rx_buffer = NULL; rx_buffer 426 drivers/staging/ks7010/ks7010_sdio.c struct rx_device_buffer *rx_buffer; rx_buffer 435 drivers/staging/ks7010/ks7010_sdio.c rx_buffer = &priv->rx_dev.rx_dev_buff[priv->rx_dev.qtail]; rx_buffer 437 drivers/staging/ks7010/ks7010_sdio.c ret = ks7010_sdio_read(priv, DATA_WINDOW, &rx_buffer->data[0], rx_buffer 447 drivers/staging/ks7010/ks7010_sdio.c rx_buffer->data, 32); rx_buffer 457 drivers/staging/ks7010/ks7010_sdio.c hdr = (struct hostif_hdr *)&rx_buffer->data[0]; rx_buffer 458 drivers/staging/ks7010/ks7010_sdio.c rx_buffer->size = le16_to_cpu(hdr->size) + sizeof(hdr->size); rx_buffer 90 drivers/staging/pi433/pi433_if.c u8 *rx_buffer; rx_buffer 528 drivers/staging/pi433/pi433_if.c &dev->rx_buffer[dev->rx_position], rx_buffer 799 drivers/staging/pi433/pi433_if.c retval = copy_to_user(buf, device->rx_buffer, bytes_received); rx_buffer 1154 drivers/staging/pi433/pi433_if.c device->rx_buffer = kmalloc(MAX_MSG_SIZE, GFP_KERNEL); rx_buffer 1155 drivers/staging/pi433/pi433_if.c if (!device->rx_buffer) { rx_buffer 1269 drivers/staging/pi433/pi433_if.c kfree(device->rx_buffer); rx_buffer 1294 drivers/staging/pi433/pi433_if.c kfree(device->rx_buffer); rx_buffer 251 drivers/staging/wilc1000/wilc_wfi_netdevice.h u8 *rx_buffer; rx_buffer 783 drivers/staging/wilc1000/wilc_wlan.c buffer = &wilc->rx_buffer[offset]; rx_buffer 1010 drivers/staging/wilc1000/wilc_wlan.c kfree(wilc->rx_buffer); rx_buffer 1011 drivers/staging/wilc1000/wilc_wlan.c wilc->rx_buffer = NULL; rx_buffer 1245 drivers/staging/wilc1000/wilc_wlan.c if (!wilc->rx_buffer) rx_buffer 1246 drivers/staging/wilc1000/wilc_wlan.c wilc->rx_buffer = kmalloc(WILC_RX_BUFF_SIZE, GFP_KERNEL); rx_buffer 1248 drivers/staging/wilc1000/wilc_wlan.c if (!wilc->rx_buffer) { rx_buffer 1262 drivers/staging/wilc1000/wilc_wlan.c kfree(wilc->rx_buffer); rx_buffer 1263 drivers/staging/wilc1000/wilc_wlan.c wilc->rx_buffer = NULL; rx_buffer 652 drivers/tty/serial/ifx6x60.c ifx_dev->swap_buf(ifx_dev->rx_buffer, IFX_SPI_HEADER_OVERHEAD, rx_buffer 653 drivers/tty/serial/ifx6x60.c &ifx_dev->rx_buffer[IFX_SPI_HEADER_OVERHEAD]); rx_buffer 654 drivers/tty/serial/ifx6x60.c decode_result = ifx_spi_decode_spi_header(ifx_dev->rx_buffer, rx_buffer 672 drivers/tty/serial/ifx6x60.c (ifx_dev->rx_buffer + IFX_SPI_HEADER_OVERHEAD), rx_buffer 674 drivers/tty/serial/ifx6x60.c &ifx_dev->rx_buffer[IFX_SPI_TRANSFER_SIZE]); rx_buffer 677 drivers/tty/serial/ifx6x60.c ifx_dev->rx_buffer + IFX_SPI_HEADER_OVERHEAD, rx_buffer 757 drivers/tty/serial/ifx6x60.c ifx_dev->spi_xfer.rx_buf = ifx_dev->rx_buffer; rx_buffer 937 drivers/tty/serial/ifx6x60.c ifx_dev->rx_buffer, rx_buffer 1056 drivers/tty/serial/ifx6x60.c ifx_dev->rx_buffer = dma_alloc_coherent(ifx_dev->spi_dev->dev.parent, rx_buffer 1060 drivers/tty/serial/ifx6x60.c if (!ifx_dev->rx_buffer) { rx_buffer 83 drivers/tty/serial/ifx6x60.h unsigned char *rx_buffer;