rx_buffer         368 drivers/input/keyboard/applespi.c 	u8				*rx_buffer;
rx_buffer         550 drivers/input/keyboard/applespi.c 	rd_t->rx_buf = applespi->rx_buffer;
rx_buffer        1391 drivers/input/keyboard/applespi.c 	applespi_get_trace_fun(evt_type)(evt_type, PT_READ, applespi->rx_buffer,
rx_buffer        1405 drivers/input/keyboard/applespi.c 	if (!applespi_verify_crc(applespi, applespi->rx_buffer,
rx_buffer        1423 drivers/input/keyboard/applespi.c 	packet = (struct spi_packet *)applespi->rx_buffer;
rx_buffer        1657 drivers/input/keyboard/applespi.c 	applespi->rx_buffer = devm_kmalloc(&spi->dev, APPLESPI_PACKET_SIZE,
rx_buffer        1664 drivers/input/keyboard/applespi.c 	    !applespi->rx_buffer || !applespi->msg_buf)
rx_buffer          37 drivers/mailbox/mailbox-test.c 	char			*rx_buffer;
rx_buffer         207 drivers/mailbox/mailbox-test.c 	ptr = tdev->rx_buffer;
rx_buffer         220 drivers/mailbox/mailbox-test.c 	memset(tdev->rx_buffer, 0, MBOX_MAX_MSG_LEN);
rx_buffer         283 drivers/mailbox/mailbox-test.c 		memcpy_fromio(tdev->rx_buffer, tdev->rx_mmio, MBOX_MAX_MSG_LEN);
rx_buffer         285 drivers/mailbox/mailbox-test.c 				     tdev->rx_buffer, MBOX_MAX_MSG_LEN);
rx_buffer         289 drivers/mailbox/mailbox-test.c 		memcpy(tdev->rx_buffer, message, MBOX_MAX_MSG_LEN);
rx_buffer         397 drivers/mailbox/mailbox-test.c 		tdev->rx_buffer = devm_kzalloc(&pdev->dev,
rx_buffer         399 drivers/mailbox/mailbox-test.c 		if (!tdev->rx_buffer)
rx_buffer         106 drivers/media/dvb-core/dvb_ca_en50221.c 	struct dvb_ringbuffer rx_buffer;
rx_buffer         162 drivers/media/dvb-core/dvb_ca_en50221.c 		vfree(ca->slot_info[i].rx_buffer.data);
rx_buffer         646 drivers/media/dvb-core/dvb_ca_en50221.c 		if (!sl->rx_buffer.data) {
rx_buffer         650 drivers/media/dvb-core/dvb_ca_en50221.c 		buf_free = dvb_ringbuffer_free(&sl->rx_buffer);
rx_buffer         750 drivers/media/dvb-core/dvb_ca_en50221.c 		if (!sl->rx_buffer.data) {
rx_buffer         754 drivers/media/dvb-core/dvb_ca_en50221.c 		dvb_ringbuffer_pkt_write(&sl->rx_buffer, buf, bytes_read);
rx_buffer        1231 drivers/media/dvb-core/dvb_ca_en50221.c 		if (!sl->rx_buffer.data) {
rx_buffer        1240 drivers/media/dvb-core/dvb_ca_en50221.c 			dvb_ringbuffer_init(&sl->rx_buffer, rxbuf,
rx_buffer        1553 drivers/media/dvb-core/dvb_ca_en50221.c 		if (!sl->rx_buffer.data)
rx_buffer        1556 drivers/media/dvb-core/dvb_ca_en50221.c 		idx = dvb_ringbuffer_pkt_next(&sl->rx_buffer, -1, &fraglen);
rx_buffer        1558 drivers/media/dvb-core/dvb_ca_en50221.c 			dvb_ringbuffer_pkt_read(&sl->rx_buffer, idx, 0, hdr, 2);
rx_buffer        1568 drivers/media/dvb-core/dvb_ca_en50221.c 			idx = dvb_ringbuffer_pkt_next(&sl->rx_buffer, idx,
rx_buffer        1636 drivers/media/dvb-core/dvb_ca_en50221.c 	idx = dvb_ringbuffer_pkt_next(&sl->rx_buffer, -1, &fraglen);
rx_buffer        1646 drivers/media/dvb-core/dvb_ca_en50221.c 		dvb_ringbuffer_pkt_read(&sl->rx_buffer, idx, 0, hdr, 2);
rx_buffer        1657 drivers/media/dvb-core/dvb_ca_en50221.c 				   dvb_ringbuffer_pkt_read_user(&sl->rx_buffer,
rx_buffer        1672 drivers/media/dvb-core/dvb_ca_en50221.c 		idx2 = dvb_ringbuffer_pkt_next(&sl->rx_buffer, idx, &fraglen);
rx_buffer        1674 drivers/media/dvb-core/dvb_ca_en50221.c 			dvb_ringbuffer_pkt_dispose(&sl->rx_buffer, idx);
rx_buffer        1722 drivers/media/dvb-core/dvb_ca_en50221.c 			if (!sl->rx_buffer.data) {
rx_buffer        1728 drivers/media/dvb-core/dvb_ca_en50221.c 				dvb_ringbuffer_flush(&sl->rx_buffer);
rx_buffer         112 drivers/net/ethernet/brocade/bna/bfi_enet.h 	union bfi_addr_be_u  rx_buffer;
rx_buffer          79 drivers/net/ethernet/dec/tulip/xircom_cb.c 	__le32 *rx_buffer;
rx_buffer         234 drivers/net/ethernet/dec/tulip/xircom_cb.c 	private->rx_buffer = dma_alloc_coherent(d, 8192,
rx_buffer         237 drivers/net/ethernet/dec/tulip/xircom_cb.c 	if (private->rx_buffer == NULL)
rx_buffer         292 drivers/net/ethernet/dec/tulip/xircom_cb.c 	dma_free_coherent(d, 8192, private->rx_buffer, private->rx_dma_handle);
rx_buffer         318 drivers/net/ethernet/dec/tulip/xircom_cb.c 	dma_free_coherent(d, 8192, card->rx_buffer, card->rx_dma_handle);
rx_buffer         340 drivers/net/ethernet/dec/tulip/xircom_cb.c 		 card->rx_buffer[0], card->rx_buffer[4]);
rx_buffer         566 drivers/net/ethernet/dec/tulip/xircom_cb.c 	BUG_ON(card->rx_buffer == NULL);
rx_buffer         570 drivers/net/ethernet/dec/tulip/xircom_cb.c 	memset(card->rx_buffer, 0, 128);	/* clear the descriptors */
rx_buffer         574 drivers/net/ethernet/dec/tulip/xircom_cb.c 		card->rx_buffer[i*4 + 0] = cpu_to_le32(0x80000000);
rx_buffer         576 drivers/net/ethernet/dec/tulip/xircom_cb.c 		card->rx_buffer[i*4 + 1] = cpu_to_le32(1536);
rx_buffer         578 drivers/net/ethernet/dec/tulip/xircom_cb.c 			card->rx_buffer[i*4 + 1] |= cpu_to_le32(1 << 25);
rx_buffer         584 drivers/net/ethernet/dec/tulip/xircom_cb.c 		card->rx_buffer[i*4 + 2] = cpu_to_le32(address + bufferoffsets[i]);
rx_buffer         586 drivers/net/ethernet/dec/tulip/xircom_cb.c 		card->rx_buffer[i*4 + 3] = 0;
rx_buffer        1101 drivers/net/ethernet/dec/tulip/xircom_cb.c 	status = le32_to_cpu(card->rx_buffer[4*descnr]);
rx_buffer        1123 drivers/net/ethernet/dec/tulip/xircom_cb.c 					&card->rx_buffer[bufferoffset / 4],
rx_buffer        1133 drivers/net/ethernet/dec/tulip/xircom_cb.c 		card->rx_buffer[4*descnr] = cpu_to_le32(0x80000000);
rx_buffer         111 drivers/net/ethernet/intel/fm10k/fm10k.h 		struct fm10k_rx_buffer *rx_buffer;
rx_buffer         123 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	bi = &rx_ring->rx_buffer[i];
rx_buffer         140 drivers/net/ethernet/intel/fm10k/fm10k_main.c 			bi = rx_ring->rx_buffer;
rx_buffer         184 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	new_buff = &rx_ring->rx_buffer[nta];
rx_buffer         205 drivers/net/ethernet/intel/fm10k/fm10k_main.c static bool fm10k_can_reuse_rx_page(struct fm10k_rx_buffer *rx_buffer,
rx_buffer         219 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	rx_buffer->page_offset ^= FM10K_RX_BUFSZ;
rx_buffer         222 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	rx_buffer->page_offset += truesize;
rx_buffer         224 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	if (rx_buffer->page_offset > (PAGE_SIZE - FM10K_RX_BUFSZ))
rx_buffer         251 drivers/net/ethernet/intel/fm10k/fm10k_main.c static bool fm10k_add_rx_frag(struct fm10k_rx_buffer *rx_buffer,
rx_buffer         256 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	struct page *page = rx_buffer->page;
rx_buffer         257 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	unsigned char *va = page_address(page) + rx_buffer->page_offset;
rx_buffer         296 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	return fm10k_can_reuse_rx_page(rx_buffer, page, truesize);
rx_buffer         304 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	struct fm10k_rx_buffer *rx_buffer;
rx_buffer         307 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	rx_buffer = &rx_ring->rx_buffer[rx_ring->next_to_clean];
rx_buffer         308 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	page = rx_buffer->page;
rx_buffer         313 drivers/net/ethernet/intel/fm10k/fm10k_main.c 				  rx_buffer->page_offset;
rx_buffer         338 drivers/net/ethernet/intel/fm10k/fm10k_main.c 				      rx_buffer->dma,
rx_buffer         339 drivers/net/ethernet/intel/fm10k/fm10k_main.c 				      rx_buffer->page_offset,
rx_buffer         344 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	if (fm10k_add_rx_frag(rx_buffer, size, rx_desc, skb)) {
rx_buffer         346 drivers/net/ethernet/intel/fm10k/fm10k_main.c 		fm10k_reuse_rx_page(rx_ring, rx_buffer);
rx_buffer         349 drivers/net/ethernet/intel/fm10k/fm10k_main.c 		dma_unmap_page(rx_ring->dev, rx_buffer->dma,
rx_buffer         354 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	rx_buffer->page = NULL;
rx_buffer          90 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c 	rx_ring->rx_buffer = vzalloc(size);
rx_buffer          91 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c 	if (!rx_ring->rx_buffer)
rx_buffer         107 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c 	vfree(rx_ring->rx_buffer);
rx_buffer         108 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c 	rx_ring->rx_buffer = NULL;
rx_buffer         253 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c 	if (!rx_ring->rx_buffer)
rx_buffer         261 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c 		struct fm10k_rx_buffer *buffer = &rx_ring->rx_buffer[i];
rx_buffer         274 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c 	memset(rx_ring->rx_buffer, 0, size);
rx_buffer         294 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c 	vfree(rx_ring->rx_buffer);
rx_buffer         295 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c 	rx_ring->rx_buffer = NULL;
rx_buffer        1265 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	struct i40e_rx_buffer *rx_buffer;
rx_buffer        1275 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	rx_buffer = &rx_ring->rx_bi[ntc++];
rx_buffer        1287 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	return rx_buffer;
rx_buffer        1889 drivers/net/ethernet/intel/i40e/i40e_txrx.c static bool i40e_can_reuse_rx_page(struct i40e_rx_buffer *rx_buffer)
rx_buffer        1891 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	unsigned int pagecnt_bias = rx_buffer->pagecnt_bias;
rx_buffer        1892 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	struct page *page = rx_buffer->page;
rx_buffer        1905 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	if (rx_buffer->page_offset > I40E_LAST_OFFSET)
rx_buffer        1915 drivers/net/ethernet/intel/i40e/i40e_txrx.c 		rx_buffer->pagecnt_bias = USHRT_MAX;
rx_buffer        1934 drivers/net/ethernet/intel/i40e/i40e_txrx.c 			     struct i40e_rx_buffer *rx_buffer,
rx_buffer        1944 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, rx_buffer->page,
rx_buffer        1945 drivers/net/ethernet/intel/i40e/i40e_txrx.c 			rx_buffer->page_offset, size, truesize);
rx_buffer        1949 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	rx_buffer->page_offset ^= truesize;
rx_buffer        1951 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	rx_buffer->page_offset += truesize;
rx_buffer        1966 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	struct i40e_rx_buffer *rx_buffer;
rx_buffer        1968 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	rx_buffer = &rx_ring->rx_bi[rx_ring->next_to_clean];
rx_buffer        1969 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	prefetchw(rx_buffer->page);
rx_buffer        1973 drivers/net/ethernet/intel/i40e/i40e_txrx.c 				      rx_buffer->dma,
rx_buffer        1974 drivers/net/ethernet/intel/i40e/i40e_txrx.c 				      rx_buffer->page_offset,
rx_buffer        1979 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	rx_buffer->pagecnt_bias--;
rx_buffer        1981 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	return rx_buffer;
rx_buffer        1995 drivers/net/ethernet/intel/i40e/i40e_txrx.c 					  struct i40e_rx_buffer *rx_buffer,
rx_buffer        2048 drivers/net/ethernet/intel/i40e/i40e_txrx.c 		skb_add_rx_frag(skb, 0, rx_buffer->page,
rx_buffer        2049 drivers/net/ethernet/intel/i40e/i40e_txrx.c 				rx_buffer->page_offset + headlen,
rx_buffer        2054 drivers/net/ethernet/intel/i40e/i40e_txrx.c 		rx_buffer->page_offset ^= truesize;
rx_buffer        2056 drivers/net/ethernet/intel/i40e/i40e_txrx.c 		rx_buffer->page_offset += truesize;
rx_buffer        2060 drivers/net/ethernet/intel/i40e/i40e_txrx.c 		rx_buffer->pagecnt_bias++;
rx_buffer        2076 drivers/net/ethernet/intel/i40e/i40e_txrx.c 				      struct i40e_rx_buffer *rx_buffer,
rx_buffer        2111 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	rx_buffer->page_offset ^= truesize;
rx_buffer        2113 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	rx_buffer->page_offset += truesize;
rx_buffer        2128 drivers/net/ethernet/intel/i40e/i40e_txrx.c 			       struct i40e_rx_buffer *rx_buffer)
rx_buffer        2130 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	if (i40e_can_reuse_rx_page(rx_buffer)) {
rx_buffer        2132 drivers/net/ethernet/intel/i40e/i40e_txrx.c 		i40e_reuse_rx_page(rx_ring, rx_buffer);
rx_buffer        2135 drivers/net/ethernet/intel/i40e/i40e_txrx.c 		dma_unmap_page_attrs(rx_ring->dev, rx_buffer->dma,
rx_buffer        2138 drivers/net/ethernet/intel/i40e/i40e_txrx.c 		__page_frag_cache_drain(rx_buffer->page,
rx_buffer        2139 drivers/net/ethernet/intel/i40e/i40e_txrx.c 					rx_buffer->pagecnt_bias);
rx_buffer        2141 drivers/net/ethernet/intel/i40e/i40e_txrx.c 		rx_buffer->page = NULL;
rx_buffer        2246 drivers/net/ethernet/intel/i40e/i40e_txrx.c 				struct i40e_rx_buffer *rx_buffer,
rx_buffer        2252 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	rx_buffer->page_offset ^= truesize;
rx_buffer        2256 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	rx_buffer->page_offset += truesize;
rx_buffer        2341 drivers/net/ethernet/intel/i40e/i40e_txrx.c 		struct i40e_rx_buffer *rx_buffer;
rx_buffer        2368 drivers/net/ethernet/intel/i40e/i40e_txrx.c 		rx_buffer = i40e_clean_programming_status(rx_ring, rx_desc,
rx_buffer        2370 drivers/net/ethernet/intel/i40e/i40e_txrx.c 		if (unlikely(rx_buffer)) {
rx_buffer        2371 drivers/net/ethernet/intel/i40e/i40e_txrx.c 			i40e_reuse_rx_page(rx_ring, rx_buffer);
rx_buffer        2382 drivers/net/ethernet/intel/i40e/i40e_txrx.c 		rx_buffer = i40e_get_rx_buffer(rx_ring, size);
rx_buffer        2386 drivers/net/ethernet/intel/i40e/i40e_txrx.c 			xdp.data = page_address(rx_buffer->page) +
rx_buffer        2387 drivers/net/ethernet/intel/i40e/i40e_txrx.c 				   rx_buffer->page_offset;
rx_buffer        2401 drivers/net/ethernet/intel/i40e/i40e_txrx.c 				i40e_rx_buffer_flip(rx_ring, rx_buffer, size);
rx_buffer        2403 drivers/net/ethernet/intel/i40e/i40e_txrx.c 				rx_buffer->pagecnt_bias++;
rx_buffer        2408 drivers/net/ethernet/intel/i40e/i40e_txrx.c 			i40e_add_rx_frag(rx_ring, rx_buffer, skb, size);
rx_buffer        2410 drivers/net/ethernet/intel/i40e/i40e_txrx.c 			skb = i40e_build_skb(rx_ring, rx_buffer, &xdp);
rx_buffer        2412 drivers/net/ethernet/intel/i40e/i40e_txrx.c 			skb = i40e_construct_skb(rx_ring, rx_buffer, &xdp);
rx_buffer        2418 drivers/net/ethernet/intel/i40e/i40e_txrx.c 			rx_buffer->pagecnt_bias++;
rx_buffer        2422 drivers/net/ethernet/intel/i40e/i40e_txrx.c 		i40e_put_rx_buffer(rx_ring, rx_buffer);
rx_buffer        1184 drivers/net/ethernet/intel/iavf/iavf_txrx.c static bool iavf_can_reuse_rx_page(struct iavf_rx_buffer *rx_buffer)
rx_buffer        1186 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	unsigned int pagecnt_bias = rx_buffer->pagecnt_bias;
rx_buffer        1187 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	struct page *page = rx_buffer->page;
rx_buffer        1200 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	if (rx_buffer->page_offset > IAVF_LAST_OFFSET)
rx_buffer        1210 drivers/net/ethernet/intel/iavf/iavf_txrx.c 		rx_buffer->pagecnt_bias = USHRT_MAX;
rx_buffer        1229 drivers/net/ethernet/intel/iavf/iavf_txrx.c 			     struct iavf_rx_buffer *rx_buffer,
rx_buffer        1242 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, rx_buffer->page,
rx_buffer        1243 drivers/net/ethernet/intel/iavf/iavf_txrx.c 			rx_buffer->page_offset, size, truesize);
rx_buffer        1247 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	rx_buffer->page_offset ^= truesize;
rx_buffer        1249 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	rx_buffer->page_offset += truesize;
rx_buffer        1264 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	struct iavf_rx_buffer *rx_buffer;
rx_buffer        1269 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	rx_buffer = &rx_ring->rx_bi[rx_ring->next_to_clean];
rx_buffer        1270 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	prefetchw(rx_buffer->page);
rx_buffer        1274 drivers/net/ethernet/intel/iavf/iavf_txrx.c 				      rx_buffer->dma,
rx_buffer        1275 drivers/net/ethernet/intel/iavf/iavf_txrx.c 				      rx_buffer->page_offset,
rx_buffer        1280 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	rx_buffer->pagecnt_bias--;
rx_buffer        1282 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	return rx_buffer;
rx_buffer        1296 drivers/net/ethernet/intel/iavf/iavf_txrx.c 					  struct iavf_rx_buffer *rx_buffer,
rx_buffer        1308 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	if (!rx_buffer)
rx_buffer        1311 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	va = page_address(rx_buffer->page) + rx_buffer->page_offset;
rx_buffer        1335 drivers/net/ethernet/intel/iavf/iavf_txrx.c 		skb_add_rx_frag(skb, 0, rx_buffer->page,
rx_buffer        1336 drivers/net/ethernet/intel/iavf/iavf_txrx.c 				rx_buffer->page_offset + headlen,
rx_buffer        1341 drivers/net/ethernet/intel/iavf/iavf_txrx.c 		rx_buffer->page_offset ^= truesize;
rx_buffer        1343 drivers/net/ethernet/intel/iavf/iavf_txrx.c 		rx_buffer->page_offset += truesize;
rx_buffer        1347 drivers/net/ethernet/intel/iavf/iavf_txrx.c 		rx_buffer->pagecnt_bias++;
rx_buffer        1363 drivers/net/ethernet/intel/iavf/iavf_txrx.c 				      struct iavf_rx_buffer *rx_buffer,
rx_buffer        1375 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	if (!rx_buffer)
rx_buffer        1378 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	va = page_address(rx_buffer->page) + rx_buffer->page_offset;
rx_buffer        1394 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	rx_buffer->page_offset ^= truesize;
rx_buffer        1396 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	rx_buffer->page_offset += truesize;
rx_buffer        1411 drivers/net/ethernet/intel/iavf/iavf_txrx.c 			       struct iavf_rx_buffer *rx_buffer)
rx_buffer        1413 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	if (!rx_buffer)
rx_buffer        1416 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	if (iavf_can_reuse_rx_page(rx_buffer)) {
rx_buffer        1418 drivers/net/ethernet/intel/iavf/iavf_txrx.c 		iavf_reuse_rx_page(rx_ring, rx_buffer);
rx_buffer        1422 drivers/net/ethernet/intel/iavf/iavf_txrx.c 		dma_unmap_page_attrs(rx_ring->dev, rx_buffer->dma,
rx_buffer        1425 drivers/net/ethernet/intel/iavf/iavf_txrx.c 		__page_frag_cache_drain(rx_buffer->page,
rx_buffer        1426 drivers/net/ethernet/intel/iavf/iavf_txrx.c 					rx_buffer->pagecnt_bias);
rx_buffer        1430 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	rx_buffer->page = NULL;
rx_buffer        1486 drivers/net/ethernet/intel/iavf/iavf_txrx.c 		struct iavf_rx_buffer *rx_buffer;
rx_buffer        1522 drivers/net/ethernet/intel/iavf/iavf_txrx.c 		rx_buffer = iavf_get_rx_buffer(rx_ring, size);
rx_buffer        1526 drivers/net/ethernet/intel/iavf/iavf_txrx.c 			iavf_add_rx_frag(rx_ring, rx_buffer, skb, size);
rx_buffer        1528 drivers/net/ethernet/intel/iavf/iavf_txrx.c 			skb = iavf_build_skb(rx_ring, rx_buffer, size);
rx_buffer        1530 drivers/net/ethernet/intel/iavf/iavf_txrx.c 			skb = iavf_construct_skb(rx_ring, rx_buffer, size);
rx_buffer        1535 drivers/net/ethernet/intel/iavf/iavf_txrx.c 			if (rx_buffer)
rx_buffer        1536 drivers/net/ethernet/intel/iavf/iavf_txrx.c 				rx_buffer->pagecnt_bias++;
rx_buffer        1540 drivers/net/ethernet/intel/iavf/iavf_txrx.c 		iavf_put_rx_buffer(rx_ring, rx_buffer);
rx_buffer        1781 drivers/net/ethernet/intel/igb/igb_ethtool.c static int igb_check_lbtest_frame(struct igb_rx_buffer *rx_buffer,
rx_buffer        1789 drivers/net/ethernet/intel/igb/igb_ethtool.c 	data = kmap(rx_buffer->page);
rx_buffer        1796 drivers/net/ethernet/intel/igb/igb_ethtool.c 	kunmap(rx_buffer->page);
rx_buffer        7943 drivers/net/ethernet/intel/igb/igb_main.c static bool igb_can_reuse_rx_page(struct igb_rx_buffer *rx_buffer)
rx_buffer        7945 drivers/net/ethernet/intel/igb/igb_main.c 	unsigned int pagecnt_bias = rx_buffer->pagecnt_bias;
rx_buffer        7946 drivers/net/ethernet/intel/igb/igb_main.c 	struct page *page = rx_buffer->page;
rx_buffer        7960 drivers/net/ethernet/intel/igb/igb_main.c 	if (rx_buffer->page_offset > IGB_LAST_OFFSET)
rx_buffer        7970 drivers/net/ethernet/intel/igb/igb_main.c 		rx_buffer->pagecnt_bias = USHRT_MAX;
rx_buffer        7986 drivers/net/ethernet/intel/igb/igb_main.c 			    struct igb_rx_buffer *rx_buffer,
rx_buffer        7997 drivers/net/ethernet/intel/igb/igb_main.c 	skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, rx_buffer->page,
rx_buffer        7998 drivers/net/ethernet/intel/igb/igb_main.c 			rx_buffer->page_offset, size, truesize);
rx_buffer        8000 drivers/net/ethernet/intel/igb/igb_main.c 	rx_buffer->page_offset ^= truesize;
rx_buffer        8002 drivers/net/ethernet/intel/igb/igb_main.c 	rx_buffer->page_offset += truesize;
rx_buffer        8007 drivers/net/ethernet/intel/igb/igb_main.c 					 struct igb_rx_buffer *rx_buffer,
rx_buffer        8011 drivers/net/ethernet/intel/igb/igb_main.c 	void *va = page_address(rx_buffer->page) + rx_buffer->page_offset;
rx_buffer        8048 drivers/net/ethernet/intel/igb/igb_main.c 		skb_add_rx_frag(skb, 0, rx_buffer->page,
rx_buffer        8049 drivers/net/ethernet/intel/igb/igb_main.c 				(va + headlen) - page_address(rx_buffer->page),
rx_buffer        8052 drivers/net/ethernet/intel/igb/igb_main.c 		rx_buffer->page_offset ^= truesize;
rx_buffer        8054 drivers/net/ethernet/intel/igb/igb_main.c 		rx_buffer->page_offset += truesize;
rx_buffer        8057 drivers/net/ethernet/intel/igb/igb_main.c 		rx_buffer->pagecnt_bias++;
rx_buffer        8064 drivers/net/ethernet/intel/igb/igb_main.c 				     struct igb_rx_buffer *rx_buffer,
rx_buffer        8068 drivers/net/ethernet/intel/igb/igb_main.c 	void *va = page_address(rx_buffer->page) + rx_buffer->page_offset;
rx_buffer        8100 drivers/net/ethernet/intel/igb/igb_main.c 	rx_buffer->page_offset ^= truesize;
rx_buffer        8102 drivers/net/ethernet/intel/igb/igb_main.c 	rx_buffer->page_offset += truesize;
rx_buffer        8265 drivers/net/ethernet/intel/igb/igb_main.c 	struct igb_rx_buffer *rx_buffer;
rx_buffer        8267 drivers/net/ethernet/intel/igb/igb_main.c 	rx_buffer = &rx_ring->rx_buffer_info[rx_ring->next_to_clean];
rx_buffer        8268 drivers/net/ethernet/intel/igb/igb_main.c 	prefetchw(rx_buffer->page);
rx_buffer        8272 drivers/net/ethernet/intel/igb/igb_main.c 				      rx_buffer->dma,
rx_buffer        8273 drivers/net/ethernet/intel/igb/igb_main.c 				      rx_buffer->page_offset,
rx_buffer        8277 drivers/net/ethernet/intel/igb/igb_main.c 	rx_buffer->pagecnt_bias--;
rx_buffer        8279 drivers/net/ethernet/intel/igb/igb_main.c 	return rx_buffer;
rx_buffer        8283 drivers/net/ethernet/intel/igb/igb_main.c 			      struct igb_rx_buffer *rx_buffer)
rx_buffer        8285 drivers/net/ethernet/intel/igb/igb_main.c 	if (igb_can_reuse_rx_page(rx_buffer)) {
rx_buffer        8287 drivers/net/ethernet/intel/igb/igb_main.c 		igb_reuse_rx_page(rx_ring, rx_buffer);
rx_buffer        8292 drivers/net/ethernet/intel/igb/igb_main.c 		dma_unmap_page_attrs(rx_ring->dev, rx_buffer->dma,
rx_buffer        8295 drivers/net/ethernet/intel/igb/igb_main.c 		__page_frag_cache_drain(rx_buffer->page,
rx_buffer        8296 drivers/net/ethernet/intel/igb/igb_main.c 					rx_buffer->pagecnt_bias);
rx_buffer        8300 drivers/net/ethernet/intel/igb/igb_main.c 	rx_buffer->page = NULL;
rx_buffer        8312 drivers/net/ethernet/intel/igb/igb_main.c 		struct igb_rx_buffer *rx_buffer;
rx_buffer        8332 drivers/net/ethernet/intel/igb/igb_main.c 		rx_buffer = igb_get_rx_buffer(rx_ring, size);
rx_buffer        8336 drivers/net/ethernet/intel/igb/igb_main.c 			igb_add_rx_frag(rx_ring, rx_buffer, skb, size);
rx_buffer        8338 drivers/net/ethernet/intel/igb/igb_main.c 			skb = igb_build_skb(rx_ring, rx_buffer, rx_desc, size);
rx_buffer        8340 drivers/net/ethernet/intel/igb/igb_main.c 			skb = igb_construct_skb(rx_ring, rx_buffer,
rx_buffer        8346 drivers/net/ethernet/intel/igb/igb_main.c 			rx_buffer->pagecnt_bias++;
rx_buffer        8350 drivers/net/ethernet/intel/igb/igb_main.c 		igb_put_rx_buffer(rx_ring, rx_buffer);
rx_buffer        1200 drivers/net/ethernet/intel/igc/igc_main.c 	struct igc_rx_buffer *rx_buffer;
rx_buffer        1202 drivers/net/ethernet/intel/igc/igc_main.c 	rx_buffer = &rx_ring->rx_buffer_info[rx_ring->next_to_clean];
rx_buffer        1203 drivers/net/ethernet/intel/igc/igc_main.c 	prefetchw(rx_buffer->page);
rx_buffer        1207 drivers/net/ethernet/intel/igc/igc_main.c 				      rx_buffer->dma,
rx_buffer        1208 drivers/net/ethernet/intel/igc/igc_main.c 				      rx_buffer->page_offset,
rx_buffer        1212 drivers/net/ethernet/intel/igc/igc_main.c 	rx_buffer->pagecnt_bias--;
rx_buffer        1214 drivers/net/ethernet/intel/igc/igc_main.c 	return rx_buffer;
rx_buffer        1227 drivers/net/ethernet/intel/igc/igc_main.c 			    struct igc_rx_buffer *rx_buffer,
rx_buffer        1234 drivers/net/ethernet/intel/igc/igc_main.c 	skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, rx_buffer->page,
rx_buffer        1235 drivers/net/ethernet/intel/igc/igc_main.c 			rx_buffer->page_offset, size, truesize);
rx_buffer        1236 drivers/net/ethernet/intel/igc/igc_main.c 	rx_buffer->page_offset ^= truesize;
rx_buffer        1241 drivers/net/ethernet/intel/igc/igc_main.c 	skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, rx_buffer->page,
rx_buffer        1242 drivers/net/ethernet/intel/igc/igc_main.c 			rx_buffer->page_offset, size, truesize);
rx_buffer        1243 drivers/net/ethernet/intel/igc/igc_main.c 	rx_buffer->page_offset += truesize;
rx_buffer        1248 drivers/net/ethernet/intel/igc/igc_main.c 				     struct igc_rx_buffer *rx_buffer,
rx_buffer        1252 drivers/net/ethernet/intel/igc/igc_main.c 	void *va = page_address(rx_buffer->page) + rx_buffer->page_offset;
rx_buffer        1278 drivers/net/ethernet/intel/igc/igc_main.c 	rx_buffer->page_offset ^= truesize;
rx_buffer        1280 drivers/net/ethernet/intel/igc/igc_main.c 	rx_buffer->page_offset += truesize;
rx_buffer        1287 drivers/net/ethernet/intel/igc/igc_main.c 					 struct igc_rx_buffer *rx_buffer,
rx_buffer        1291 drivers/net/ethernet/intel/igc/igc_main.c 	void *va = page_address(rx_buffer->page) + rx_buffer->page_offset;
rx_buffer        1322 drivers/net/ethernet/intel/igc/igc_main.c 		skb_add_rx_frag(skb, 0, rx_buffer->page,
rx_buffer        1323 drivers/net/ethernet/intel/igc/igc_main.c 				(va + headlen) - page_address(rx_buffer->page),
rx_buffer        1326 drivers/net/ethernet/intel/igc/igc_main.c 		rx_buffer->page_offset ^= truesize;
rx_buffer        1328 drivers/net/ethernet/intel/igc/igc_main.c 		rx_buffer->page_offset += truesize;
rx_buffer        1331 drivers/net/ethernet/intel/igc/igc_main.c 		rx_buffer->pagecnt_bias++;
rx_buffer        1371 drivers/net/ethernet/intel/igc/igc_main.c static bool igc_can_reuse_rx_page(struct igc_rx_buffer *rx_buffer)
rx_buffer        1373 drivers/net/ethernet/intel/igc/igc_main.c 	unsigned int pagecnt_bias = rx_buffer->pagecnt_bias;
rx_buffer        1374 drivers/net/ethernet/intel/igc/igc_main.c 	struct page *page = rx_buffer->page;
rx_buffer        1388 drivers/net/ethernet/intel/igc/igc_main.c 	if (rx_buffer->page_offset > IGC_LAST_OFFSET)
rx_buffer        1398 drivers/net/ethernet/intel/igc/igc_main.c 		rx_buffer->pagecnt_bias = USHRT_MAX;
rx_buffer        1468 drivers/net/ethernet/intel/igc/igc_main.c 			      struct igc_rx_buffer *rx_buffer)
rx_buffer        1470 drivers/net/ethernet/intel/igc/igc_main.c 	if (igc_can_reuse_rx_page(rx_buffer)) {
rx_buffer        1472 drivers/net/ethernet/intel/igc/igc_main.c 		igc_reuse_rx_page(rx_ring, rx_buffer);
rx_buffer        1477 drivers/net/ethernet/intel/igc/igc_main.c 		dma_unmap_page_attrs(rx_ring->dev, rx_buffer->dma,
rx_buffer        1480 drivers/net/ethernet/intel/igc/igc_main.c 		__page_frag_cache_drain(rx_buffer->page,
rx_buffer        1481 drivers/net/ethernet/intel/igc/igc_main.c 					rx_buffer->pagecnt_bias);
rx_buffer        1485 drivers/net/ethernet/intel/igc/igc_main.c 	rx_buffer->page = NULL;
rx_buffer        1566 drivers/net/ethernet/intel/igc/igc_main.c 		struct igc_rx_buffer *rx_buffer;
rx_buffer        1586 drivers/net/ethernet/intel/igc/igc_main.c 		rx_buffer = igc_get_rx_buffer(rx_ring, size);
rx_buffer        1590 drivers/net/ethernet/intel/igc/igc_main.c 			igc_add_rx_frag(rx_ring, rx_buffer, skb, size);
rx_buffer        1592 drivers/net/ethernet/intel/igc/igc_main.c 			skb = igc_build_skb(rx_ring, rx_buffer, rx_desc, size);
rx_buffer        1594 drivers/net/ethernet/intel/igc/igc_main.c 			skb = igc_construct_skb(rx_ring, rx_buffer,
rx_buffer        1600 drivers/net/ethernet/intel/igc/igc_main.c 			rx_buffer->pagecnt_bias++;
rx_buffer        1604 drivers/net/ethernet/intel/igc/igc_main.c 		igc_put_rx_buffer(rx_ring, rx_buffer);
rx_buffer        1866 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c static bool ixgbe_check_lbtest_frame(struct ixgbe_rx_buffer *rx_buffer,
rx_buffer        1874 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c 	data = kmap(rx_buffer->page) + rx_buffer->page_offset;
rx_buffer        1881 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c 	kunmap(rx_buffer->page);
rx_buffer        1928 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c 		struct ixgbe_rx_buffer *rx_buffer;
rx_buffer        1931 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c 		rx_buffer = &rx_ring->rx_buffer_info[rx_ntc];
rx_buffer        1935 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c 					rx_buffer->dma,
rx_buffer        1940 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c 		if (ixgbe_check_lbtest_frame(rx_buffer, size))
rx_buffer        1947 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c 					   rx_buffer->dma,
rx_buffer        1950 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c static bool ixgbe_can_reuse_rx_page(struct ixgbe_rx_buffer *rx_buffer)
rx_buffer        1952 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	unsigned int pagecnt_bias = rx_buffer->pagecnt_bias;
rx_buffer        1953 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	struct page *page = rx_buffer->page;
rx_buffer        1971 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	if (rx_buffer->page_offset > IXGBE_LAST_OFFSET)
rx_buffer        1981 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		rx_buffer->pagecnt_bias = USHRT_MAX;
rx_buffer        2003 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 			      struct ixgbe_rx_buffer *rx_buffer,
rx_buffer        2014 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, rx_buffer->page,
rx_buffer        2015 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 			rx_buffer->page_offset, size, truesize);
rx_buffer        2017 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	rx_buffer->page_offset ^= truesize;
rx_buffer        2019 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	rx_buffer->page_offset += truesize;
rx_buffer        2028 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	struct ixgbe_rx_buffer *rx_buffer;
rx_buffer        2030 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	rx_buffer = &rx_ring->rx_buffer_info[rx_ring->next_to_clean];
rx_buffer        2031 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	prefetchw(rx_buffer->page);
rx_buffer        2032 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	*skb = rx_buffer->skb;
rx_buffer        2048 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 				      rx_buffer->dma,
rx_buffer        2049 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 				      rx_buffer->page_offset,
rx_buffer        2053 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	rx_buffer->pagecnt_bias--;
rx_buffer        2055 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	return rx_buffer;
rx_buffer        2059 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 				struct ixgbe_rx_buffer *rx_buffer,
rx_buffer        2062 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	if (ixgbe_can_reuse_rx_page(rx_buffer)) {
rx_buffer        2064 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		ixgbe_reuse_rx_page(rx_ring, rx_buffer);
rx_buffer        2066 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		if (!IS_ERR(skb) && IXGBE_CB(skb)->dma == rx_buffer->dma) {
rx_buffer        2071 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 			dma_unmap_page_attrs(rx_ring->dev, rx_buffer->dma,
rx_buffer        2076 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		__page_frag_cache_drain(rx_buffer->page,
rx_buffer        2077 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 					rx_buffer->pagecnt_bias);
rx_buffer        2081 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	rx_buffer->page = NULL;
rx_buffer        2082 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	rx_buffer->skb = NULL;
rx_buffer        2086 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 					   struct ixgbe_rx_buffer *rx_buffer,
rx_buffer        2127 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 			IXGBE_CB(skb)->dma = rx_buffer->dma;
rx_buffer        2129 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		skb_add_rx_frag(skb, 0, rx_buffer->page,
rx_buffer        2130 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 				xdp->data - page_address(rx_buffer->page),
rx_buffer        2133 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		rx_buffer->page_offset ^= truesize;
rx_buffer        2135 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		rx_buffer->page_offset += truesize;
rx_buffer        2140 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		rx_buffer->pagecnt_bias++;
rx_buffer        2147 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 				       struct ixgbe_rx_buffer *rx_buffer,
rx_buffer        2184 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		IXGBE_CB(skb)->dma = rx_buffer->dma;
rx_buffer        2188 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	rx_buffer->page_offset ^= truesize;
rx_buffer        2190 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	rx_buffer->page_offset += truesize;
rx_buffer        2248 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 				 struct ixgbe_rx_buffer *rx_buffer,
rx_buffer        2254 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	rx_buffer->page_offset ^= truesize;
rx_buffer        2260 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	rx_buffer->page_offset += truesize;
rx_buffer        2295 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		struct ixgbe_rx_buffer *rx_buffer;
rx_buffer        2316 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		rx_buffer = ixgbe_get_rx_buffer(rx_ring, rx_desc, &skb, size);
rx_buffer        2320 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 			xdp.data = page_address(rx_buffer->page) +
rx_buffer        2321 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 				   rx_buffer->page_offset;
rx_buffer        2335 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 				ixgbe_rx_buffer_flip(rx_ring, rx_buffer, size);
rx_buffer        2337 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 				rx_buffer->pagecnt_bias++;
rx_buffer        2342 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 			ixgbe_add_rx_frag(rx_ring, rx_buffer, skb, size);
rx_buffer        2344 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 			skb = ixgbe_build_skb(rx_ring, rx_buffer,
rx_buffer        2347 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 			skb = ixgbe_construct_skb(rx_ring, rx_buffer,
rx_buffer        2354 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 			rx_buffer->pagecnt_bias++;
rx_buffer        2358 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		ixgbe_put_rx_buffer(rx_ring, rx_buffer, skb);
rx_buffer        5291 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	struct ixgbe_rx_buffer *rx_buffer = &rx_ring->rx_buffer_info[i];
rx_buffer        5300 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		if (rx_buffer->skb) {
rx_buffer        5301 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 			struct sk_buff *skb = rx_buffer->skb;
rx_buffer        5315 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 					      rx_buffer->dma,
rx_buffer        5316 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 					      rx_buffer->page_offset,
rx_buffer        5321 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		dma_unmap_page_attrs(rx_ring->dev, rx_buffer->dma,
rx_buffer        5325 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		__page_frag_cache_drain(rx_buffer->page,
rx_buffer        5326 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 					rx_buffer->pagecnt_bias);
rx_buffer        5329 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		rx_buffer++;
rx_buffer        5332 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 			rx_buffer = rx_ring->rx_buffer_info;
rx_buffer         114 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c static bool ixgbevf_can_reuse_rx_page(struct ixgbevf_rx_buffer *rx_buffer);
rx_buffer         533 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	struct ixgbevf_rx_buffer *rx_buffer;
rx_buffer         535 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	rx_buffer = &rx_ring->rx_buffer_info[rx_ring->next_to_clean];
rx_buffer         536 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	prefetchw(rx_buffer->page);
rx_buffer         540 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 				      rx_buffer->dma,
rx_buffer         541 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 				      rx_buffer->page_offset,
rx_buffer         545 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	rx_buffer->pagecnt_bias--;
rx_buffer         547 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	return rx_buffer;
rx_buffer         551 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 				  struct ixgbevf_rx_buffer *rx_buffer,
rx_buffer         554 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	if (ixgbevf_can_reuse_rx_page(rx_buffer)) {
rx_buffer         556 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		ixgbevf_reuse_rx_page(rx_ring, rx_buffer);
rx_buffer         562 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 			dma_unmap_page_attrs(rx_ring->dev, rx_buffer->dma,
rx_buffer         566 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		__page_frag_cache_drain(rx_buffer->page,
rx_buffer         567 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 					rx_buffer->pagecnt_bias);
rx_buffer         571 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	rx_buffer->page = NULL;
rx_buffer         791 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c static bool ixgbevf_can_reuse_rx_page(struct ixgbevf_rx_buffer *rx_buffer)
rx_buffer         793 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	unsigned int pagecnt_bias = rx_buffer->pagecnt_bias;
rx_buffer         794 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	struct page *page = rx_buffer->page;
rx_buffer         808 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	if (rx_buffer->page_offset > IXGBEVF_LAST_OFFSET)
rx_buffer         819 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		rx_buffer->pagecnt_bias = USHRT_MAX;
rx_buffer         835 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 				struct ixgbevf_rx_buffer *rx_buffer,
rx_buffer         846 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, rx_buffer->page,
rx_buffer         847 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 			rx_buffer->page_offset, size, truesize);
rx_buffer         849 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	rx_buffer->page_offset ^= truesize;
rx_buffer         851 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	rx_buffer->page_offset += truesize;
rx_buffer         857 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 				      struct ixgbevf_rx_buffer *rx_buffer,
rx_buffer         909 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		skb_add_rx_frag(skb, 0, rx_buffer->page,
rx_buffer         911 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 					page_address(rx_buffer->page),
rx_buffer         914 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		rx_buffer->page_offset ^= truesize;
rx_buffer         916 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		rx_buffer->page_offset += truesize;
rx_buffer         919 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		rx_buffer->pagecnt_bias++;
rx_buffer         934 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 					 struct ixgbevf_rx_buffer *rx_buffer,
rx_buffer         971 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	rx_buffer->page_offset ^= truesize;
rx_buffer         973 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	rx_buffer->page_offset += truesize;
rx_buffer        1099 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 				   struct ixgbevf_rx_buffer *rx_buffer,
rx_buffer        1105 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	rx_buffer->page_offset ^= truesize;
rx_buffer        1111 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	rx_buffer->page_offset += truesize;
rx_buffer        1129 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		struct ixgbevf_rx_buffer *rx_buffer;
rx_buffer        1150 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		rx_buffer = ixgbevf_get_rx_buffer(rx_ring, size);
rx_buffer        1154 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 			xdp.data = page_address(rx_buffer->page) +
rx_buffer        1155 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 				   rx_buffer->page_offset;
rx_buffer        1167 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 				ixgbevf_rx_buffer_flip(rx_ring, rx_buffer,
rx_buffer        1170 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 				rx_buffer->pagecnt_bias++;
rx_buffer        1175 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 			ixgbevf_add_rx_frag(rx_ring, rx_buffer, skb, size);
rx_buffer        1177 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 			skb = ixgbevf_build_skb(rx_ring, rx_buffer,
rx_buffer        1180 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 			skb = ixgbevf_construct_skb(rx_ring, rx_buffer,
rx_buffer        1187 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 			rx_buffer->pagecnt_bias++;
rx_buffer        1191 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		ixgbevf_put_rx_buffer(rx_ring, rx_buffer, skb);
rx_buffer        2339 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		struct ixgbevf_rx_buffer *rx_buffer;
rx_buffer        2341 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		rx_buffer = &rx_ring->rx_buffer_info[i];
rx_buffer        2347 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 					      rx_buffer->dma,
rx_buffer        2348 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 					      rx_buffer->page_offset,
rx_buffer        2354 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 				     rx_buffer->dma,
rx_buffer        2359 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		__page_frag_cache_drain(rx_buffer->page,
rx_buffer        2360 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 					rx_buffer->pagecnt_bias);
rx_buffer         426 drivers/net/ethernet/oki-semi/pch_gbe/pch_gbe.h 	unsigned char *rx_buffer;
rx_buffer        1381 drivers/net/ethernet/oki-semi/pch_gbe/pch_gbe_main.c 						  buffer_info->rx_buffer,
rx_buffer        1437 drivers/net/ethernet/oki-semi/pch_gbe/pch_gbe_main.c 		buffer_info->rx_buffer = rx_ring->rx_buff_pool + bufsz * i;
rx_buffer        1682 drivers/net/ethernet/oki-semi/pch_gbe/pch_gbe_main.c 			memcpy(skb->data, buffer_info->rx_buffer, length);
rx_buffer         389 drivers/net/ethernet/qualcomm/qca_spi.c 			bytes_read = qcaspi_read_legacy(qca, qca->rx_buffer,
rx_buffer         392 drivers/net/ethernet/qualcomm/qca_spi.c 			bytes_read = qcaspi_read_burst(qca, qca->rx_buffer,
rx_buffer         406 drivers/net/ethernet/qualcomm/qca_spi.c 		cp = qca->rx_buffer;
rx_buffer         817 drivers/net/ethernet/qualcomm/qca_spi.c 	qca->rx_buffer = kmalloc(qca->buffer_size, GFP_KERNEL);
rx_buffer         818 drivers/net/ethernet/qualcomm/qca_spi.c 	if (!qca->rx_buffer)
rx_buffer         824 drivers/net/ethernet/qualcomm/qca_spi.c 		kfree(qca->rx_buffer);
rx_buffer         837 drivers/net/ethernet/qualcomm/qca_spi.c 	kfree(qca->rx_buffer);
rx_buffer          88 drivers/net/ethernet/qualcomm/qca_spi.h 	u8 *rx_buffer;
rx_buffer         181 drivers/net/ethernet/ti/tlan.h 	u8			*rx_buffer;
rx_buffer         293 drivers/net/wan/fsl_ucc_hdlc.c 	priv->rx_buffer = bd_buffer;
rx_buffer         546 drivers/net/wan/fsl_ucc_hdlc.c 		bdbuffer = priv->rx_buffer +
rx_buffer         783 drivers/net/wan/fsl_ucc_hdlc.c 	if (priv->rx_buffer) {
rx_buffer         786 drivers/net/wan/fsl_ucc_hdlc.c 				  priv->rx_buffer, priv->dma_rx_addr);
rx_buffer         787 drivers/net/wan/fsl_ucc_hdlc.c 		priv->rx_buffer = NULL;
rx_buffer          80 drivers/net/wan/fsl_ucc_hdlc.h 	u8 *rx_buffer;
rx_buffer         138 drivers/net/wireless/ti/wl1251/rx.c 	u8 *rx_buffer, beacon = 0;
rx_buffer         165 drivers/net/wireless/ti/wl1251/rx.c 	rx_buffer = skb_put(skb, length);
rx_buffer         166 drivers/net/wireless/ti/wl1251/rx.c 	wl1251_mem_read(wl, rx_packet_ring_addr, rx_buffer, length);
rx_buffer          77 drivers/spi/spidev.c 	u8			*rx_buffer;
rx_buffer         130 drivers/spi/spidev.c 			.rx_buf		= spidev->rx_buffer,
rx_buffer         161 drivers/spi/spidev.c 		missing = copy_to_user(buf, spidev->rx_buffer, status);
rx_buffer         219 drivers/spi/spidev.c 	rx_buf = spidev->rx_buffer;
rx_buffer         293 drivers/spi/spidev.c 	rx_buf = spidev->rx_buffer;
rx_buffer         581 drivers/spi/spidev.c 	if (!spidev->rx_buffer) {
rx_buffer         582 drivers/spi/spidev.c 		spidev->rx_buffer = kmalloc(bufsiz, GFP_KERNEL);
rx_buffer         583 drivers/spi/spidev.c 		if (!spidev->rx_buffer) {
rx_buffer         621 drivers/spi/spidev.c 		kfree(spidev->rx_buffer);
rx_buffer         622 drivers/spi/spidev.c 		spidev->rx_buffer = NULL;
rx_buffer         426 drivers/staging/ks7010/ks7010_sdio.c 	struct rx_device_buffer *rx_buffer;
rx_buffer         435 drivers/staging/ks7010/ks7010_sdio.c 	rx_buffer = &priv->rx_dev.rx_dev_buff[priv->rx_dev.qtail];
rx_buffer         437 drivers/staging/ks7010/ks7010_sdio.c 	ret = ks7010_sdio_read(priv, DATA_WINDOW, &rx_buffer->data[0],
rx_buffer         447 drivers/staging/ks7010/ks7010_sdio.c 				     rx_buffer->data, 32);
rx_buffer         457 drivers/staging/ks7010/ks7010_sdio.c 	hdr = (struct hostif_hdr *)&rx_buffer->data[0];
rx_buffer         458 drivers/staging/ks7010/ks7010_sdio.c 	rx_buffer->size = le16_to_cpu(hdr->size) + sizeof(hdr->size);
rx_buffer          90 drivers/staging/pi433/pi433_if.c 	u8			*rx_buffer;
rx_buffer         528 drivers/staging/pi433/pi433_if.c 					&dev->rx_buffer[dev->rx_position],
rx_buffer         799 drivers/staging/pi433/pi433_if.c 		retval = copy_to_user(buf, device->rx_buffer, bytes_received);
rx_buffer        1154 drivers/staging/pi433/pi433_if.c 	device->rx_buffer = kmalloc(MAX_MSG_SIZE, GFP_KERNEL);
rx_buffer        1155 drivers/staging/pi433/pi433_if.c 	if (!device->rx_buffer) {
rx_buffer        1269 drivers/staging/pi433/pi433_if.c 	kfree(device->rx_buffer);
rx_buffer        1294 drivers/staging/pi433/pi433_if.c 	kfree(device->rx_buffer);
rx_buffer         251 drivers/staging/wilc1000/wilc_wfi_netdevice.h 	u8 *rx_buffer;
rx_buffer         783 drivers/staging/wilc1000/wilc_wlan.c 	buffer = &wilc->rx_buffer[offset];
rx_buffer        1010 drivers/staging/wilc1000/wilc_wlan.c 	kfree(wilc->rx_buffer);
rx_buffer        1011 drivers/staging/wilc1000/wilc_wlan.c 	wilc->rx_buffer = NULL;
rx_buffer        1245 drivers/staging/wilc1000/wilc_wlan.c 	if (!wilc->rx_buffer)
rx_buffer        1246 drivers/staging/wilc1000/wilc_wlan.c 		wilc->rx_buffer = kmalloc(WILC_RX_BUFF_SIZE, GFP_KERNEL);
rx_buffer        1248 drivers/staging/wilc1000/wilc_wlan.c 	if (!wilc->rx_buffer) {
rx_buffer        1262 drivers/staging/wilc1000/wilc_wlan.c 	kfree(wilc->rx_buffer);
rx_buffer        1263 drivers/staging/wilc1000/wilc_wlan.c 	wilc->rx_buffer = NULL;
rx_buffer         652 drivers/tty/serial/ifx6x60.c 		ifx_dev->swap_buf(ifx_dev->rx_buffer, IFX_SPI_HEADER_OVERHEAD,
rx_buffer         653 drivers/tty/serial/ifx6x60.c 			&ifx_dev->rx_buffer[IFX_SPI_HEADER_OVERHEAD]);
rx_buffer         654 drivers/tty/serial/ifx6x60.c 		decode_result = ifx_spi_decode_spi_header(ifx_dev->rx_buffer,
rx_buffer         672 drivers/tty/serial/ifx6x60.c 			(ifx_dev->rx_buffer + IFX_SPI_HEADER_OVERHEAD),
rx_buffer         674 drivers/tty/serial/ifx6x60.c 			 &ifx_dev->rx_buffer[IFX_SPI_TRANSFER_SIZE]);
rx_buffer         677 drivers/tty/serial/ifx6x60.c 			ifx_dev->rx_buffer + IFX_SPI_HEADER_OVERHEAD,
rx_buffer         757 drivers/tty/serial/ifx6x60.c 		ifx_dev->spi_xfer.rx_buf = ifx_dev->rx_buffer;
rx_buffer         937 drivers/tty/serial/ifx6x60.c 				ifx_dev->rx_buffer,
rx_buffer        1056 drivers/tty/serial/ifx6x60.c 	ifx_dev->rx_buffer = dma_alloc_coherent(ifx_dev->spi_dev->dev.parent,
rx_buffer        1060 drivers/tty/serial/ifx6x60.c 	if (!ifx_dev->rx_buffer) {
rx_buffer          83 drivers/tty/serial/ifx6x60.h 	unsigned char *rx_buffer;