rx_queue          937 arch/um/drivers/vector_kern.c 	struct vector_queue *qi = vp->rx_queue;
rx_queue         1154 arch/um/drivers/vector_kern.c 	if (vp->rx_queue != NULL)
rx_queue         1155 arch/um/drivers/vector_kern.c 		destroy_queue(vp->rx_queue);
rx_queue         1208 arch/um/drivers/vector_kern.c 		vp->rx_queue = create_queue(
rx_queue         1214 arch/um/drivers/vector_kern.c 		vp->rx_queue->queue_depth = get_depth(vp->parsed);
rx_queue         1355 arch/um/drivers/vector_kern.c 	ring->rx_max_pending = vp->rx_queue->max_depth;
rx_queue         1357 arch/um/drivers/vector_kern.c 	ring->rx_pending = vp->rx_queue->max_depth;
rx_queue           83 arch/um/drivers/vector_kern.h 	struct vector_queue *rx_queue;
rx_queue          340 drivers/atm/eni.c 	skb_queue_head_init(&ENI_DEV(dev)->rx_queue);
rx_queue          476 drivers/atm/eni.c 		skb_queue_tail(&eni_dev->rx_queue,skb);
rx_queue          738 drivers/atm/eni.c 		skb = skb_dequeue(&eni_dev->rx_queue);
rx_queue          758 drivers/atm/eni.c 			skb_queue_head(&eni_dev->rx_queue,skb);
rx_queue          921 drivers/atm/eni.c 	skb_queue_head_init(&eni_dev->rx_queue);
rx_queue          101 drivers/atm/eni.h 	struct sk_buff_head rx_queue;	/* PDUs currently being RX-DMAed */
rx_queue          181 drivers/net/dsa/mv88e6xxx/chip.h 	struct sk_buff_head rx_queue;
rx_queue          348 drivers/net/dsa/mv88e6xxx/hwtstamp.c 	skb = skb_dequeue(&ps->rx_queue);
rx_queue          352 drivers/net/dsa/mv88e6xxx/hwtstamp.c 				   &ps->rx_queue);
rx_queue          387 drivers/net/dsa/mv88e6xxx/hwtstamp.c 		skb_queue_tail(&ps->rx_queue, skb);
rx_queue          576 drivers/net/dsa/mv88e6xxx/hwtstamp.c 	skb_queue_head_init(&ps->rx_queue);
rx_queue          585 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h 	u8			rx_queue;	/* index for skb_record */
rx_queue          749 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	skb_record_rx_queue(skb, fp->rx_queue);
rx_queue         1079 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		skb_record_rx_queue(skb, fp->rx_queue);
rx_queue         6346 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	fp->rx_queue = fp_idx;
rx_queue         6449 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	bnx2x_fcoe(bp, rx_queue) = BNX2X_NUM_ETH_QUEUES(bp);
rx_queue          144 drivers/net/ethernet/chelsio/cxgb3/adapter.h 	struct sk_buff_head rx_queue; /* offload packet receive queue */
rx_queue         1854 drivers/net/ethernet/chelsio/cxgb3/sge.c 	int was_empty = skb_queue_empty(&q->rx_queue);
rx_queue         1856 drivers/net/ethernet/chelsio/cxgb3/sge.c 	__skb_queue_tail(&q->rx_queue, skb);
rx_queue         1909 drivers/net/ethernet/chelsio/cxgb3/sge.c 		skb_queue_splice_init(&q->rx_queue, &queue);
rx_queue         1936 drivers/net/ethernet/chelsio/cxgb3/sge.c 			skb_queue_splice(&queue, &q->rx_queue);
rx_queue         3099 drivers/net/ethernet/chelsio/cxgb3/sge.c 	skb_queue_head_init(&q->rspq.rx_queue);
rx_queue          524 drivers/net/ethernet/freescale/fec.h 	struct fec_enet_priv_rx_q *rx_queue[FEC_ENET_MAX_RX_QS];
rx_queue          838 drivers/net/ethernet/freescale/fec_main.c 		rxq = fep->rx_queue[q];
rx_queue          894 drivers/net/ethernet/freescale/fec_main.c 		writel(0, fep->rx_queue[i]->bd.reg_desc_active);
rx_queue          905 drivers/net/ethernet/freescale/fec_main.c 		rxq = fep->rx_queue[i];
rx_queue         1434 drivers/net/ethernet/freescale/fec_main.c 	rxq = fep->rx_queue[queue_id];
rx_queue         2753 drivers/net/ethernet/freescale/fec_main.c 		rxq = fep->rx_queue[q];
rx_queue         2798 drivers/net/ethernet/freescale/fec_main.c 		kfree(fep->rx_queue[i]);
rx_queue         2836 drivers/net/ethernet/freescale/fec_main.c 		fep->rx_queue[i] = kzalloc(sizeof(*fep->rx_queue[i]),
rx_queue         2838 drivers/net/ethernet/freescale/fec_main.c 		if (!fep->rx_queue[i]) {
rx_queue         2843 drivers/net/ethernet/freescale/fec_main.c 		fep->rx_queue[i]->bd.ring_size = RX_RING_SIZE;
rx_queue         2844 drivers/net/ethernet/freescale/fec_main.c 		fep->total_rx_ring_size += fep->rx_queue[i]->bd.ring_size;
rx_queue         2862 drivers/net/ethernet/freescale/fec_main.c 	rxq = fep->rx_queue[queue];
rx_queue         3274 drivers/net/ethernet/freescale/fec_main.c 		struct fec_enet_priv_rx_q *rxq = fep->rx_queue[i];
rx_queue          112 drivers/net/ethernet/freescale/gianfar.c static void gfar_init_rxbdp(struct gfar_priv_rx_q *rx_queue, struct rxbd8 *bdp,
rx_queue          120 drivers/net/ethernet/freescale/gianfar.c 	if (bdp == rx_queue->rx_bd_base + rx_queue->rx_ring_size - 1)
rx_queue          142 drivers/net/ethernet/freescale/gianfar.c 		gfar_write(baddr, priv->rx_queue[i]->rx_bd_dma_base);
rx_queue          155 drivers/net/ethernet/freescale/gianfar.c 		gfar_write(baddr, priv->rx_queue[i]->rx_ring_size |
rx_queue          261 drivers/net/ethernet/freescale/gianfar.c 			if (likely(priv->rx_queue[i]->rxcoalescing))
rx_queue          262 drivers/net/ethernet/freescale/gianfar.c 				gfar_write(baddr + i, priv->rx_queue[i]->rxic);
rx_queue          273 drivers/net/ethernet/freescale/gianfar.c 		if (unlikely(priv->rx_queue[0]->rxcoalescing))
rx_queue          274 drivers/net/ethernet/freescale/gianfar.c 			gfar_write(&regs->rxic, priv->rx_queue[0]->rxic);
rx_queue          291 drivers/net/ethernet/freescale/gianfar.c 		rx_packets += priv->rx_queue[i]->stats.rx_packets;
rx_queue          292 drivers/net/ethernet/freescale/gianfar.c 		rx_bytes   += priv->rx_queue[i]->stats.rx_bytes;
rx_queue          293 drivers/net/ethernet/freescale/gianfar.c 		rx_dropped += priv->rx_queue[i]->stats.rx_dropped;
rx_queue          422 drivers/net/ethernet/freescale/gianfar.c 		priv->rx_queue[i] = kzalloc(sizeof(struct gfar_priv_rx_q),
rx_queue          424 drivers/net/ethernet/freescale/gianfar.c 		if (!priv->rx_queue[i])
rx_queue          427 drivers/net/ethernet/freescale/gianfar.c 		priv->rx_queue[i]->qindex = i;
rx_queue          428 drivers/net/ethernet/freescale/gianfar.c 		priv->rx_queue[i]->ndev = priv->ndev;
rx_queue          446 drivers/net/ethernet/freescale/gianfar.c 		kfree(priv->rx_queue[i]);
rx_queue          561 drivers/net/ethernet/freescale/gianfar.c 		if (!grp->rx_queue)
rx_queue          562 drivers/net/ethernet/freescale/gianfar.c 			grp->rx_queue = priv->rx_queue[i];
rx_queue          566 drivers/net/ethernet/freescale/gianfar.c 		priv->rx_queue[i]->grp = grp;
rx_queue         1116 drivers/net/ethernet/freescale/gianfar.c static void free_skb_rx_queue(struct gfar_priv_rx_q *rx_queue)
rx_queue         1120 drivers/net/ethernet/freescale/gianfar.c 	struct rxbd8 *rxbdp = rx_queue->rx_bd_base;
rx_queue         1122 drivers/net/ethernet/freescale/gianfar.c 	dev_kfree_skb(rx_queue->skb);
rx_queue         1124 drivers/net/ethernet/freescale/gianfar.c 	for (i = 0; i < rx_queue->rx_ring_size; i++) {
rx_queue         1125 drivers/net/ethernet/freescale/gianfar.c 		struct	gfar_rx_buff *rxb = &rx_queue->rx_buff[i];
rx_queue         1134 drivers/net/ethernet/freescale/gianfar.c 		dma_unmap_page(rx_queue->dev, rxb->dma,
rx_queue         1141 drivers/net/ethernet/freescale/gianfar.c 	kfree(rx_queue->rx_buff);
rx_queue         1142 drivers/net/ethernet/freescale/gianfar.c 	rx_queue->rx_buff = NULL;
rx_queue         1151 drivers/net/ethernet/freescale/gianfar.c 	struct gfar_priv_rx_q *rx_queue = NULL;
rx_queue         1166 drivers/net/ethernet/freescale/gianfar.c 		rx_queue = priv->rx_queue[i];
rx_queue         1167 drivers/net/ethernet/freescale/gianfar.c 		if (rx_queue->rx_buff)
rx_queue         1168 drivers/net/ethernet/freescale/gianfar.c 			free_skb_rx_queue(rx_queue);
rx_queue         1258 drivers/net/ethernet/freescale/gianfar.c static void gfar_rx_alloc_err(struct gfar_priv_rx_q *rx_queue)
rx_queue         1260 drivers/net/ethernet/freescale/gianfar.c 	struct gfar_private *priv = netdev_priv(rx_queue->ndev);
rx_queue         1263 drivers/net/ethernet/freescale/gianfar.c 	netdev_err(rx_queue->ndev, "Can't alloc RX buffers\n");
rx_queue         1267 drivers/net/ethernet/freescale/gianfar.c static void gfar_alloc_rx_buffs(struct gfar_priv_rx_q *rx_queue,
rx_queue         1274 drivers/net/ethernet/freescale/gianfar.c 	i = rx_queue->next_to_use;
rx_queue         1275 drivers/net/ethernet/freescale/gianfar.c 	bdp = &rx_queue->rx_bd_base[i];
rx_queue         1276 drivers/net/ethernet/freescale/gianfar.c 	rxb = &rx_queue->rx_buff[i];
rx_queue         1281 drivers/net/ethernet/freescale/gianfar.c 			if (unlikely(!gfar_new_page(rx_queue, rxb))) {
rx_queue         1282 drivers/net/ethernet/freescale/gianfar.c 				gfar_rx_alloc_err(rx_queue);
rx_queue         1288 drivers/net/ethernet/freescale/gianfar.c 		gfar_init_rxbdp(rx_queue, bdp,
rx_queue         1295 drivers/net/ethernet/freescale/gianfar.c 		if (unlikely(++i == rx_queue->rx_ring_size)) {
rx_queue         1297 drivers/net/ethernet/freescale/gianfar.c 			bdp = rx_queue->rx_bd_base;
rx_queue         1298 drivers/net/ethernet/freescale/gianfar.c 			rxb = rx_queue->rx_buff;
rx_queue         1302 drivers/net/ethernet/freescale/gianfar.c 	rx_queue->next_to_use = i;
rx_queue         1303 drivers/net/ethernet/freescale/gianfar.c 	rx_queue->next_to_alloc = i;
rx_queue         1311 drivers/net/ethernet/freescale/gianfar.c 	struct gfar_priv_rx_q *rx_queue = NULL;
rx_queue         1341 drivers/net/ethernet/freescale/gianfar.c 		rx_queue = priv->rx_queue[i];
rx_queue         1343 drivers/net/ethernet/freescale/gianfar.c 		rx_queue->next_to_clean = 0;
rx_queue         1344 drivers/net/ethernet/freescale/gianfar.c 		rx_queue->next_to_use = 0;
rx_queue         1345 drivers/net/ethernet/freescale/gianfar.c 		rx_queue->next_to_alloc = 0;
rx_queue         1350 drivers/net/ethernet/freescale/gianfar.c 		gfar_alloc_rx_buffs(rx_queue, gfar_rxbd_unused(rx_queue));
rx_queue         1352 drivers/net/ethernet/freescale/gianfar.c 		rx_queue->rfbptr = rfbptr;
rx_queue         1365 drivers/net/ethernet/freescale/gianfar.c 	struct gfar_priv_rx_q *rx_queue = NULL;
rx_queue         1373 drivers/net/ethernet/freescale/gianfar.c 		priv->total_rx_ring_size += priv->rx_queue[i]->rx_ring_size;
rx_queue         1397 drivers/net/ethernet/freescale/gianfar.c 		rx_queue = priv->rx_queue[i];
rx_queue         1398 drivers/net/ethernet/freescale/gianfar.c 		rx_queue->rx_bd_base = vaddr;
rx_queue         1399 drivers/net/ethernet/freescale/gianfar.c 		rx_queue->rx_bd_dma_base = addr;
rx_queue         1400 drivers/net/ethernet/freescale/gianfar.c 		rx_queue->ndev = ndev;
rx_queue         1401 drivers/net/ethernet/freescale/gianfar.c 		rx_queue->dev = dev;
rx_queue         1402 drivers/net/ethernet/freescale/gianfar.c 		addr  += sizeof(struct rxbd8) * rx_queue->rx_ring_size;
rx_queue         1403 drivers/net/ethernet/freescale/gianfar.c 		vaddr += sizeof(struct rxbd8) * rx_queue->rx_ring_size;
rx_queue         1421 drivers/net/ethernet/freescale/gianfar.c 		rx_queue = priv->rx_queue[i];
rx_queue         1422 drivers/net/ethernet/freescale/gianfar.c 		rx_queue->rx_buff = kcalloc(rx_queue->rx_ring_size,
rx_queue         1423 drivers/net/ethernet/freescale/gianfar.c 					    sizeof(*rx_queue->rx_buff),
rx_queue         1425 drivers/net/ethernet/freescale/gianfar.c 		if (!rx_queue->rx_buff)
rx_queue         1513 drivers/net/ethernet/freescale/gianfar.c 	struct gfar_priv_rx_q *rx_queue = NULL;
rx_queue         1573 drivers/net/ethernet/freescale/gianfar.c 				rx_queue = priv->rx_queue[i];
rx_queue         1574 drivers/net/ethernet/freescale/gianfar.c 				bdp_dma = gfar_rxbd_dma_lastfree(rx_queue);
rx_queue         1575 drivers/net/ethernet/freescale/gianfar.c 				gfar_write(rx_queue->rfbptr, bdp_dma);
rx_queue         2435 drivers/net/ethernet/freescale/gianfar.c static struct sk_buff *gfar_get_next_rxbuff(struct gfar_priv_rx_q *rx_queue,
rx_queue         2438 drivers/net/ethernet/freescale/gianfar.c 	struct gfar_rx_buff *rxb = &rx_queue->rx_buff[rx_queue->next_to_clean];
rx_queue         2447 drivers/net/ethernet/freescale/gianfar.c 			gfar_rx_alloc_err(rx_queue);
rx_queue         2454 drivers/net/ethernet/freescale/gianfar.c 	dma_sync_single_range_for_cpu(rx_queue->dev, rxb->dma, rxb->page_offset,
rx_queue         2459 drivers/net/ethernet/freescale/gianfar.c 		gfar_reuse_rx_page(rx_queue, rxb);
rx_queue         2462 drivers/net/ethernet/freescale/gianfar.c 		dma_unmap_page(rx_queue->dev, rxb->dma,
rx_queue         2532 drivers/net/ethernet/freescale/gianfar.c static int gfar_clean_rx_ring(struct gfar_priv_rx_q *rx_queue,
rx_queue         2535 drivers/net/ethernet/freescale/gianfar.c 	struct net_device *ndev = rx_queue->ndev;
rx_queue         2539 drivers/net/ethernet/freescale/gianfar.c 	struct sk_buff *skb = rx_queue->skb;
rx_queue         2540 drivers/net/ethernet/freescale/gianfar.c 	int cleaned_cnt = gfar_rxbd_unused(rx_queue);
rx_queue         2544 drivers/net/ethernet/freescale/gianfar.c 	i = rx_queue->next_to_clean;
rx_queue         2550 drivers/net/ethernet/freescale/gianfar.c 			gfar_alloc_rx_buffs(rx_queue, cleaned_cnt);
rx_queue         2554 drivers/net/ethernet/freescale/gianfar.c 		bdp = &rx_queue->rx_bd_base[i];
rx_queue         2563 drivers/net/ethernet/freescale/gianfar.c 		skb = gfar_get_next_rxbuff(rx_queue, lstatus, skb);
rx_queue         2570 drivers/net/ethernet/freescale/gianfar.c 		if (unlikely(++i == rx_queue->rx_ring_size))
rx_queue         2573 drivers/net/ethernet/freescale/gianfar.c 		rx_queue->next_to_clean = i;
rx_queue         2585 drivers/net/ethernet/freescale/gianfar.c 			rx_queue->stats.rx_dropped++;
rx_queue         2595 drivers/net/ethernet/freescale/gianfar.c 		skb_record_rx_queue(skb, rx_queue->qindex);
rx_queue         2600 drivers/net/ethernet/freescale/gianfar.c 		napi_gro_receive(&rx_queue->grp->napi_rx, skb);
rx_queue         2606 drivers/net/ethernet/freescale/gianfar.c 	rx_queue->skb = skb;
rx_queue         2608 drivers/net/ethernet/freescale/gianfar.c 	rx_queue->stats.rx_packets += total_pkts;
rx_queue         2609 drivers/net/ethernet/freescale/gianfar.c 	rx_queue->stats.rx_bytes += total_bytes;
rx_queue         2612 drivers/net/ethernet/freescale/gianfar.c 		gfar_alloc_rx_buffs(rx_queue, cleaned_cnt);
rx_queue         2616 drivers/net/ethernet/freescale/gianfar.c 		u32 bdp_dma = gfar_rxbd_dma_lastfree(rx_queue);
rx_queue         2618 drivers/net/ethernet/freescale/gianfar.c 		gfar_write(rx_queue->rfbptr, bdp_dma);
rx_queue         2629 drivers/net/ethernet/freescale/gianfar.c 	struct gfar_priv_rx_q *rx_queue = gfargrp->rx_queue;
rx_queue         2637 drivers/net/ethernet/freescale/gianfar.c 	work_done = gfar_clean_rx_ring(rx_queue, budget);
rx_queue         2689 drivers/net/ethernet/freescale/gianfar.c 	struct gfar_priv_rx_q *rx_queue = NULL;
rx_queue         2711 drivers/net/ethernet/freescale/gianfar.c 		rx_queue = priv->rx_queue[i];
rx_queue         2713 drivers/net/ethernet/freescale/gianfar.c 			gfar_clean_rx_ring(rx_queue, budget_per_q);
rx_queue         3389 drivers/net/ethernet/freescale/gianfar.c 		priv->rx_queue[i]->rx_ring_size = DEFAULT_RX_RING_SIZE;
rx_queue         3390 drivers/net/ethernet/freescale/gianfar.c 		priv->rx_queue[i]->rxcoalescing = DEFAULT_RX_COALESCE;
rx_queue         3391 drivers/net/ethernet/freescale/gianfar.c 		priv->rx_queue[i]->rxic = DEFAULT_RXIC;
rx_queue         3452 drivers/net/ethernet/freescale/gianfar.c 			    i, priv->rx_queue[i]->rx_ring_size);
rx_queue         3532 drivers/net/ethernet/freescale/gianfar.c 		u8 qindex = (u8)priv->gfargrp[0].rx_queue->qindex;
rx_queue         1052 drivers/net/ethernet/freescale/gianfar.h 	struct gfar_priv_rx_q *rx_queue;
rx_queue         1103 drivers/net/ethernet/freescale/gianfar.h 	struct gfar_priv_rx_q *rx_queue[MAX_RX_QS];
rx_queue          253 drivers/net/ethernet/freescale/gianfar_ethtool.c 	struct gfar_priv_rx_q *rx_queue = NULL;
rx_queue          266 drivers/net/ethernet/freescale/gianfar_ethtool.c 	rx_queue = priv->rx_queue[0];
rx_queue          269 drivers/net/ethernet/freescale/gianfar_ethtool.c 	rxtime  = get_ictt_value(rx_queue->rxic);
rx_queue          270 drivers/net/ethernet/freescale/gianfar_ethtool.c 	rxcount = get_icft_value(rx_queue->rxic);
rx_queue          360 drivers/net/ethernet/freescale/gianfar_ethtool.c 			priv->rx_queue[i]->rxcoalescing = 0;
rx_queue          363 drivers/net/ethernet/freescale/gianfar_ethtool.c 			priv->rx_queue[i]->rxcoalescing = 1;
rx_queue          367 drivers/net/ethernet/freescale/gianfar_ethtool.c 		priv->rx_queue[i]->rxic = mk_ic_value(
rx_queue          408 drivers/net/ethernet/freescale/gianfar_ethtool.c 	struct gfar_priv_rx_q *rx_queue = NULL;
rx_queue          411 drivers/net/ethernet/freescale/gianfar_ethtool.c 	rx_queue = priv->rx_queue[0];
rx_queue          421 drivers/net/ethernet/freescale/gianfar_ethtool.c 	rvals->rx_pending = rx_queue->rx_ring_size;
rx_queue          422 drivers/net/ethernet/freescale/gianfar_ethtool.c 	rvals->rx_mini_pending = rx_queue->rx_ring_size;
rx_queue          423 drivers/net/ethernet/freescale/gianfar_ethtool.c 	rvals->rx_jumbo_pending = rx_queue->rx_ring_size;
rx_queue          460 drivers/net/ethernet/freescale/gianfar_ethtool.c 		priv->rx_queue[i]->rx_ring_size = rvals->rx_pending;
rx_queue          105 drivers/net/ethernet/ibm/ibmveth.c 	return be32_to_cpu(adapter->rx_queue.queue_addr[adapter->rx_queue.index].flags_off);
rx_queue          116 drivers/net/ethernet/ibm/ibmveth.c 	return ibmveth_rxq_toggle(adapter) == adapter->rx_queue.toggle;
rx_queue          136 drivers/net/ethernet/ibm/ibmveth.c 	return be32_to_cpu(adapter->rx_queue.queue_addr[adapter->rx_queue.index].length);
rx_queue          397 drivers/net/ethernet/ibm/ibmveth.c 	u64 correlator = adapter->rx_queue.queue_addr[adapter->rx_queue.index].correlator;
rx_queue          410 drivers/net/ethernet/ibm/ibmveth.c 	u32 q_index = adapter->rx_queue.index;
rx_queue          411 drivers/net/ethernet/ibm/ibmveth.c 	u64 correlator = adapter->rx_queue.queue_addr[q_index].correlator;
rx_queue          436 drivers/net/ethernet/ibm/ibmveth.c 		ibmveth_remove_buffer_from_pool(adapter, adapter->rx_queue.queue_addr[adapter->rx_queue.index].correlator);
rx_queue          440 drivers/net/ethernet/ibm/ibmveth.c 	if (++adapter->rx_queue.index == adapter->rx_queue.num_slots) {
rx_queue          441 drivers/net/ethernet/ibm/ibmveth.c 		adapter->rx_queue.index = 0;
rx_queue          442 drivers/net/ethernet/ibm/ibmveth.c 		adapter->rx_queue.toggle = !adapter->rx_queue.toggle;
rx_queue          451 drivers/net/ethernet/ibm/ibmveth.c 	ibmveth_remove_buffer_from_pool(adapter, adapter->rx_queue.queue_addr[adapter->rx_queue.index].correlator);
rx_queue          453 drivers/net/ethernet/ibm/ibmveth.c 	if (++adapter->rx_queue.index == adapter->rx_queue.num_slots) {
rx_queue          454 drivers/net/ethernet/ibm/ibmveth.c 		adapter->rx_queue.index = 0;
rx_queue          455 drivers/net/ethernet/ibm/ibmveth.c 		adapter->rx_queue.toggle = !adapter->rx_queue.toggle;
rx_queue          530 drivers/net/ethernet/ibm/ibmveth.c 	adapter->rx_queue.queue_len = sizeof(struct ibmveth_rx_q_entry) *
rx_queue          532 drivers/net/ethernet/ibm/ibmveth.c 	adapter->rx_queue.queue_addr =
rx_queue          533 drivers/net/ethernet/ibm/ibmveth.c 		dma_alloc_coherent(dev, adapter->rx_queue.queue_len,
rx_queue          534 drivers/net/ethernet/ibm/ibmveth.c 				   &adapter->rx_queue.queue_dma, GFP_KERNEL);
rx_queue          535 drivers/net/ethernet/ibm/ibmveth.c 	if (!adapter->rx_queue.queue_addr)
rx_queue          552 drivers/net/ethernet/ibm/ibmveth.c 	adapter->rx_queue.index = 0;
rx_queue          553 drivers/net/ethernet/ibm/ibmveth.c 	adapter->rx_queue.num_slots = rxq_entries;
rx_queue          554 drivers/net/ethernet/ibm/ibmveth.c 	adapter->rx_queue.toggle = 1;
rx_queue          559 drivers/net/ethernet/ibm/ibmveth.c 					adapter->rx_queue.queue_len;
rx_queue          560 drivers/net/ethernet/ibm/ibmveth.c 	rxq_desc.fields.address = adapter->rx_queue.queue_dma;
rx_queue          564 drivers/net/ethernet/ibm/ibmveth.c 	netdev_dbg(netdev, "receive q   @ 0x%p\n", adapter->rx_queue.queue_addr);
rx_queue          647 drivers/net/ethernet/ibm/ibmveth.c 	dma_free_coherent(dev, adapter->rx_queue.queue_len,
rx_queue          648 drivers/net/ethernet/ibm/ibmveth.c 			  adapter->rx_queue.queue_addr,
rx_queue          649 drivers/net/ethernet/ibm/ibmveth.c 			  adapter->rx_queue.queue_dma);
rx_queue          696 drivers/net/ethernet/ibm/ibmveth.c 	dma_free_coherent(dev, adapter->rx_queue.queue_len,
rx_queue          697 drivers/net/ethernet/ibm/ibmveth.c 			  adapter->rx_queue.queue_addr,
rx_queue          698 drivers/net/ethernet/ibm/ibmveth.c 			  adapter->rx_queue.queue_dma);
rx_queue          143 drivers/net/ethernet/ibm/ibmveth.h     struct ibmveth_rx_q rx_queue;
rx_queue          794 drivers/net/ethernet/intel/igb/igb_main.c 	int rx_queue = IGB_N0_QUEUE;
rx_queue          799 drivers/net/ethernet/intel/igb/igb_main.c 		rx_queue = q_vector->rx.ring->reg_idx;
rx_queue          810 drivers/net/ethernet/intel/igb/igb_main.c 		if (rx_queue > IGB_N0_QUEUE)
rx_queue          811 drivers/net/ethernet/intel/igb/igb_main.c 			msixbm = E1000_EICR_RX_QUEUE0 << rx_queue;
rx_queue          825 drivers/net/ethernet/intel/igb/igb_main.c 		if (rx_queue > IGB_N0_QUEUE)
rx_queue          827 drivers/net/ethernet/intel/igb/igb_main.c 				       rx_queue & 0x7,
rx_queue          828 drivers/net/ethernet/intel/igb/igb_main.c 				       (rx_queue & 0x8) << 1);
rx_queue          846 drivers/net/ethernet/intel/igb/igb_main.c 		if (rx_queue > IGB_N0_QUEUE)
rx_queue          848 drivers/net/ethernet/intel/igb/igb_main.c 				       rx_queue >> 1,
rx_queue          849 drivers/net/ethernet/intel/igb/igb_main.c 				       (rx_queue & 0x1) << 4);
rx_queue          921 drivers/net/ethernet/intel/igbvf/netdev.c static void igbvf_assign_vector(struct igbvf_adapter *adapter, int rx_queue,
rx_queue          932 drivers/net/ethernet/intel/igbvf/netdev.c 	if (rx_queue > IGBVF_NO_QUEUE) {
rx_queue          933 drivers/net/ethernet/intel/igbvf/netdev.c 		index = (rx_queue >> 1);
rx_queue          935 drivers/net/ethernet/intel/igbvf/netdev.c 		if (rx_queue & 0x1) {
rx_queue          944 drivers/net/ethernet/intel/igbvf/netdev.c 		adapter->rx_ring[rx_queue].eims_value = BIT(msix_vector);
rx_queue         2595 drivers/net/ethernet/intel/igc/igc_main.c 	int rx_queue = IGC_N0_QUEUE;
rx_queue         2599 drivers/net/ethernet/intel/igc/igc_main.c 		rx_queue = q_vector->rx.ring->reg_idx;
rx_queue         2605 drivers/net/ethernet/intel/igc/igc_main.c 		if (rx_queue > IGC_N0_QUEUE)
rx_queue         2607 drivers/net/ethernet/intel/igc/igc_main.c 				       rx_queue >> 1,
rx_queue         2608 drivers/net/ethernet/intel/igc/igc_main.c 				       (rx_queue & 0x1) << 4);
rx_queue          398 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct rx_queue rxq[8];
rx_queue          440 drivers/net/ethernet/marvell/mv643xx_eth.c static struct mv643xx_eth_private *rxq_to_mp(struct rx_queue *rxq)
rx_queue          450 drivers/net/ethernet/marvell/mv643xx_eth.c static void rxq_enable(struct rx_queue *rxq)
rx_queue          456 drivers/net/ethernet/marvell/mv643xx_eth.c static void rxq_disable(struct rx_queue *rxq)
rx_queue          505 drivers/net/ethernet/marvell/mv643xx_eth.c static int rxq_process(struct rx_queue *rxq, int budget)
rx_queue          597 drivers/net/ethernet/marvell/mv643xx_eth.c static int rxq_refill(struct rx_queue *rxq, int budget)
rx_queue         1934 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct rx_queue *rxq = mp->rxq + index;
rx_queue         1999 drivers/net/ethernet/marvell/mv643xx_eth.c static void rxq_deinit(struct rx_queue *rxq)
rx_queue         2377 drivers/net/ethernet/marvell/mv643xx_eth.c 		struct rx_queue *rxq = mp->rxq + i;
rx_queue         2777 drivers/net/ethernet/marvell/mvneta.c 	int rx_queue;
rx_queue         2810 drivers/net/ethernet/marvell/mvneta.c 	rx_queue = fls(((cause_rx_tx >> 8) & 0xff));
rx_queue         2811 drivers/net/ethernet/marvell/mvneta.c 	if (rx_queue) {
rx_queue         2812 drivers/net/ethernet/marvell/mvneta.c 		rx_queue = rx_queue - 1;
rx_queue         2815 drivers/net/ethernet/marvell/mvneta.c 						 &pp->rxqs[rx_queue]);
rx_queue         2818 drivers/net/ethernet/marvell/mvneta.c 						 &pp->rxqs[rx_queue]);
rx_queue           66 drivers/net/ethernet/qlogic/qed/qed_ll2.c #define QED_LL2_RX_REGISTERED(ll2)	((ll2)->rx_queue.b_cb_registered)
rx_queue          446 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	struct qed_ll2_rx_queue *p_rx = &p_ll2_conn->rx_queue;
rx_queue          481 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	struct qed_ll2_rx_queue *p_rx = &p_ll2_conn->rx_queue;
rx_queue          523 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	struct qed_ll2_rx_queue *p_rx = &p_ll2_conn->rx_queue;
rx_queue          577 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	p_rx = &p_ll2_conn->rx_queue;
rx_queue          634 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	struct qed_ll2_rx_queue *p_rx = &p_ll2_conn->rx_queue;
rx_queue          945 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	struct qed_ll2_rx_queue *p_rx = &p_ll2_conn->rx_queue;
rx_queue         1165 drivers/net/ethernet/qlogic/qed/qed_ll2.c 			     &p_ll2_info->rx_queue.rxq_chain, NULL);
rx_queue         1171 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	capacity = qed_chain_get_capacity(&p_ll2_info->rx_queue.rxq_chain);
rx_queue         1179 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	p_ll2_info->rx_queue.descq_array = p_descq;
rx_queue         1187 drivers/net/ethernet/qlogic/qed/qed_ll2.c 			     &p_ll2_info->rx_queue.rcq_chain, NULL);
rx_queue         1423 drivers/net/ethernet/qlogic/qed/qed_ll2.c 				    &p_ll2_info->rx_queue.rx_sb_index,
rx_queue         1424 drivers/net/ethernet/qlogic/qed/qed_ll2.c 				    &p_ll2_info->rx_queue.p_fw_cons);
rx_queue         1425 drivers/net/ethernet/qlogic/qed/qed_ll2.c 		p_ll2_info->rx_queue.b_cb_registered = true;
rx_queue         1455 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	DIRECT_REG_WR(p_ll2_conn->rx_queue.set_prod_addr, 0x0);
rx_queue         1501 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	p_rx = &p_ll2_conn->rx_queue;
rx_queue         1644 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	p_rx = &p_ll2_conn->rx_queue;
rx_queue         1966 drivers/net/ethernet/qlogic/qed/qed_ll2.c 		p_ll2_conn->rx_queue.b_cb_registered = false;
rx_queue         1973 drivers/net/ethernet/qlogic/qed/qed_ll2.c 		qed_int_unregister_cb(p_hwfn, p_ll2_conn->rx_queue.rx_sb_index);
rx_queue         2025 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	kfree(p_ll2_conn->rx_queue.descq_array);
rx_queue         2026 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	qed_chain_free(p_hwfn->cdev, &p_ll2_conn->rx_queue.rxq_chain);
rx_queue         2027 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	qed_chain_free(p_hwfn->cdev, &p_ll2_conn->rx_queue.rcq_chain);
rx_queue          126 drivers/net/ethernet/qlogic/qed/qed_ll2.h 	struct qed_ll2_rx_queue rx_queue;
rx_queue         3072 drivers/net/ethernet/sfc/ef10.c static int efx_ef10_rx_probe(struct efx_rx_queue *rx_queue)
rx_queue         3074 drivers/net/ethernet/sfc/ef10.c 	return efx_nic_alloc_buffer(rx_queue->efx, &rx_queue->rxd.buf,
rx_queue         3075 drivers/net/ethernet/sfc/ef10.c 				    (rx_queue->ptr_mask + 1) *
rx_queue         3080 drivers/net/ethernet/sfc/ef10.c static void efx_ef10_rx_init(struct efx_rx_queue *rx_queue)
rx_queue         3085 drivers/net/ethernet/sfc/ef10.c 	struct efx_channel *channel = efx_rx_queue_channel(rx_queue);
rx_queue         3086 drivers/net/ethernet/sfc/ef10.c 	size_t entries = rx_queue->rxd.buf.len / EFX_BUF_SIZE;
rx_queue         3087 drivers/net/ethernet/sfc/ef10.c 	struct efx_nic *efx = rx_queue->efx;
rx_queue         3095 drivers/net/ethernet/sfc/ef10.c 	rx_queue->scatter_n = 0;
rx_queue         3096 drivers/net/ethernet/sfc/ef10.c 	rx_queue->scatter_len = 0;
rx_queue         3098 drivers/net/ethernet/sfc/ef10.c 	MCDI_SET_DWORD(inbuf, INIT_RXQ_IN_SIZE, rx_queue->ptr_mask + 1);
rx_queue         3100 drivers/net/ethernet/sfc/ef10.c 	MCDI_SET_DWORD(inbuf, INIT_RXQ_IN_LABEL, efx_rx_queue_index(rx_queue));
rx_queue         3102 drivers/net/ethernet/sfc/ef10.c 		       efx_rx_queue_index(rx_queue));
rx_queue         3109 drivers/net/ethernet/sfc/ef10.c 	dma_addr = rx_queue->rxd.buf.dma_addr;
rx_queue         3112 drivers/net/ethernet/sfc/ef10.c 		  efx_rx_queue_index(rx_queue), entries, (u64)dma_addr);
rx_queue         3125 drivers/net/ethernet/sfc/ef10.c 			    efx_rx_queue_index(rx_queue));
rx_queue         3128 drivers/net/ethernet/sfc/ef10.c static void efx_ef10_rx_fini(struct efx_rx_queue *rx_queue)
rx_queue         3132 drivers/net/ethernet/sfc/ef10.c 	struct efx_nic *efx = rx_queue->efx;
rx_queue         3137 drivers/net/ethernet/sfc/ef10.c 		       efx_rx_queue_index(rx_queue));
rx_queue         3152 drivers/net/ethernet/sfc/ef10.c static void efx_ef10_rx_remove(struct efx_rx_queue *rx_queue)
rx_queue         3154 drivers/net/ethernet/sfc/ef10.c 	efx_nic_free_buffer(rx_queue->efx, &rx_queue->rxd.buf);
rx_queue         3159 drivers/net/ethernet/sfc/ef10.c efx_ef10_build_rx_desc(struct efx_rx_queue *rx_queue, unsigned int index)
rx_queue         3164 drivers/net/ethernet/sfc/ef10.c 	rxd = efx_rx_desc(rx_queue, index);
rx_queue         3165 drivers/net/ethernet/sfc/ef10.c 	rx_buf = efx_rx_buffer(rx_queue, index);
rx_queue         3171 drivers/net/ethernet/sfc/ef10.c static void efx_ef10_rx_write(struct efx_rx_queue *rx_queue)
rx_queue         3173 drivers/net/ethernet/sfc/ef10.c 	struct efx_nic *efx = rx_queue->efx;
rx_queue         3178 drivers/net/ethernet/sfc/ef10.c 	write_count = rx_queue->added_count & ~7;
rx_queue         3179 drivers/net/ethernet/sfc/ef10.c 	if (rx_queue->notified_count == write_count)
rx_queue         3184 drivers/net/ethernet/sfc/ef10.c 			rx_queue,
rx_queue         3185 drivers/net/ethernet/sfc/ef10.c 			rx_queue->notified_count & rx_queue->ptr_mask);
rx_queue         3186 drivers/net/ethernet/sfc/ef10.c 	while (++rx_queue->notified_count != write_count);
rx_queue         3190 drivers/net/ethernet/sfc/ef10.c 			     write_count & rx_queue->ptr_mask);
rx_queue         3192 drivers/net/ethernet/sfc/ef10.c 			efx_rx_queue_index(rx_queue));
rx_queue         3197 drivers/net/ethernet/sfc/ef10.c static void efx_ef10_rx_defer_refill(struct efx_rx_queue *rx_queue)
rx_queue         3199 drivers/net/ethernet/sfc/ef10.c 	struct efx_channel *channel = efx_rx_queue_channel(rx_queue);
rx_queue         3394 drivers/net/ethernet/sfc/ef10.c static void efx_ef10_handle_rx_wrong_queue(struct efx_rx_queue *rx_queue,
rx_queue         3397 drivers/net/ethernet/sfc/ef10.c 	struct efx_nic *efx = rx_queue->efx;
rx_queue         3401 drivers/net/ethernet/sfc/ef10.c 		   efx_rx_queue_index(rx_queue), rx_queue_label);
rx_queue         3407 drivers/net/ethernet/sfc/ef10.c efx_ef10_handle_rx_bad_lbits(struct efx_rx_queue *rx_queue,
rx_queue         3410 drivers/net/ethernet/sfc/ef10.c 	unsigned int dropped = (actual - expected) & rx_queue->ptr_mask;
rx_queue         3411 drivers/net/ethernet/sfc/ef10.c 	struct efx_nic *efx = rx_queue->efx;
rx_queue         3421 drivers/net/ethernet/sfc/ef10.c static void efx_ef10_handle_rx_abort(struct efx_rx_queue *rx_queue)
rx_queue         3425 drivers/net/ethernet/sfc/ef10.c 	netif_dbg(rx_queue->efx, hw, rx_queue->efx->net_dev,
rx_queue         3427 drivers/net/ethernet/sfc/ef10.c 		  rx_queue->scatter_n);
rx_queue         3429 drivers/net/ethernet/sfc/ef10.c 	rx_desc_ptr = rx_queue->removed_count & rx_queue->ptr_mask;
rx_queue         3431 drivers/net/ethernet/sfc/ef10.c 	efx_rx_packet(rx_queue, rx_desc_ptr, rx_queue->scatter_n,
rx_queue         3434 drivers/net/ethernet/sfc/ef10.c 	rx_queue->removed_count += rx_queue->scatter_n;
rx_queue         3435 drivers/net/ethernet/sfc/ef10.c 	rx_queue->scatter_n = 0;
rx_queue         3436 drivers/net/ethernet/sfc/ef10.c 	rx_queue->scatter_len = 0;
rx_queue         3437 drivers/net/ethernet/sfc/ef10.c 	++efx_rx_queue_channel(rx_queue)->n_rx_nodesc_trunc;
rx_queue         3539 drivers/net/ethernet/sfc/ef10.c 	struct efx_rx_queue *rx_queue;
rx_queue         3565 drivers/net/ethernet/sfc/ef10.c 	rx_queue = efx_channel_get_rx_queue(channel);
rx_queue         3567 drivers/net/ethernet/sfc/ef10.c 	if (unlikely(rx_queue_label != efx_rx_queue_index(rx_queue)))
rx_queue         3568 drivers/net/ethernet/sfc/ef10.c 		efx_ef10_handle_rx_wrong_queue(rx_queue, rx_queue_label);
rx_queue         3570 drivers/net/ethernet/sfc/ef10.c 	n_descs = ((next_ptr_lbits - rx_queue->removed_count) &
rx_queue         3573 drivers/net/ethernet/sfc/ef10.c 	if (n_descs != rx_queue->scatter_n + 1) {
rx_queue         3577 drivers/net/ethernet/sfc/ef10.c 		if (unlikely(n_descs == rx_queue->scatter_n)) {
rx_queue         3578 drivers/net/ethernet/sfc/ef10.c 			if (rx_queue->scatter_n == 0 || rx_bytes != 0)
rx_queue         3582 drivers/net/ethernet/sfc/ef10.c 					    rx_queue->scatter_n,
rx_queue         3584 drivers/net/ethernet/sfc/ef10.c 			efx_ef10_handle_rx_abort(rx_queue);
rx_queue         3594 drivers/net/ethernet/sfc/ef10.c 		    rx_queue->scatter_n != 0 || rx_cont) {
rx_queue         3596 drivers/net/ethernet/sfc/ef10.c 				rx_queue, next_ptr_lbits,
rx_queue         3597 drivers/net/ethernet/sfc/ef10.c 				(rx_queue->removed_count +
rx_queue         3598 drivers/net/ethernet/sfc/ef10.c 				 rx_queue->scatter_n + 1) &
rx_queue         3604 drivers/net/ethernet/sfc/ef10.c 		rx_queue->scatter_n = 1;
rx_queue         3605 drivers/net/ethernet/sfc/ef10.c 		rx_queue->scatter_len = 0;
rx_queue         3611 drivers/net/ethernet/sfc/ef10.c 		++rx_queue->scatter_n;
rx_queue         3612 drivers/net/ethernet/sfc/ef10.c 		rx_queue->scatter_len += rx_bytes;
rx_queue         3659 drivers/net/ethernet/sfc/ef10.c 		efx_rx_packet(rx_queue,
rx_queue         3660 drivers/net/ethernet/sfc/ef10.c 			      rx_queue->removed_count & rx_queue->ptr_mask,
rx_queue         3661 drivers/net/ethernet/sfc/ef10.c 			      rx_queue->scatter_n, rx_queue->scatter_len,
rx_queue         3663 drivers/net/ethernet/sfc/ef10.c 		rx_queue->removed_count += rx_queue->scatter_n;
rx_queue         3666 drivers/net/ethernet/sfc/ef10.c 	rx_queue->scatter_n = 0;
rx_queue         3667 drivers/net/ethernet/sfc/ef10.c 	rx_queue->scatter_len = 0;
rx_queue         3802 drivers/net/ethernet/sfc/ef10.c 		efx_fast_push_rx_descriptors(&channel->rx_queue, true);
rx_queue         3960 drivers/net/ethernet/sfc/ef10.c 	struct efx_rx_queue *rx_queue;
rx_queue         3974 drivers/net/ethernet/sfc/ef10.c 			efx_for_each_channel_rx_queue(rx_queue, channel)
rx_queue         3975 drivers/net/ethernet/sfc/ef10.c 				efx_ef10_rx_fini(rx_queue);
rx_queue          282 drivers/net/ethernet/sfc/efx.c 		struct efx_rx_queue *rx_queue =
rx_queue          286 drivers/net/ethernet/sfc/efx.c 		efx_fast_push_rx_descriptors(rx_queue, true);
rx_queue          464 drivers/net/ethernet/sfc/efx.c 	struct efx_rx_queue *rx_queue;
rx_queue          487 drivers/net/ethernet/sfc/efx.c 	rx_queue = &channel->rx_queue;
rx_queue          488 drivers/net/ethernet/sfc/efx.c 	rx_queue->efx = efx;
rx_queue          489 drivers/net/ethernet/sfc/efx.c 	timer_setup(&rx_queue->slow_fill, efx_rx_slow_fill, 0);
rx_queue          501 drivers/net/ethernet/sfc/efx.c 	struct efx_rx_queue *rx_queue;
rx_queue          526 drivers/net/ethernet/sfc/efx.c 	rx_queue = &channel->rx_queue;
rx_queue          527 drivers/net/ethernet/sfc/efx.c 	rx_queue->buffer = NULL;
rx_queue          528 drivers/net/ethernet/sfc/efx.c 	memset(&rx_queue->rxd, 0, sizeof(rx_queue->rxd));
rx_queue          529 drivers/net/ethernet/sfc/efx.c 	timer_setup(&rx_queue->slow_fill, efx_rx_slow_fill, 0);
rx_queue          540 drivers/net/ethernet/sfc/efx.c 	struct efx_rx_queue *rx_queue;
rx_queue          560 drivers/net/ethernet/sfc/efx.c 	efx_for_each_channel_rx_queue(rx_queue, channel) {
rx_queue          561 drivers/net/ethernet/sfc/efx.c 		rc = efx_probe_rx_queue(rx_queue);
rx_queue          644 drivers/net/ethernet/sfc/efx.c 	struct efx_rx_queue *rx_queue;
rx_queue          716 drivers/net/ethernet/sfc/efx.c 		efx_for_each_channel_rx_queue(rx_queue, channel) {
rx_queue          717 drivers/net/ethernet/sfc/efx.c 			efx_init_rx_queue(rx_queue);
rx_queue          720 drivers/net/ethernet/sfc/efx.c 			efx_fast_push_rx_descriptors(rx_queue, false);
rx_queue          737 drivers/net/ethernet/sfc/efx.c 	struct efx_rx_queue *rx_queue;
rx_queue          747 drivers/net/ethernet/sfc/efx.c 		efx_for_each_channel_rx_queue(rx_queue, channel)
rx_queue          748 drivers/net/ethernet/sfc/efx.c 			rx_queue->refill_enabled = false;
rx_queue          773 drivers/net/ethernet/sfc/efx.c 		efx_for_each_channel_rx_queue(rx_queue, channel)
rx_queue          774 drivers/net/ethernet/sfc/efx.c 			efx_fini_rx_queue(rx_queue);
rx_queue          783 drivers/net/ethernet/sfc/efx.c 	struct efx_rx_queue *rx_queue;
rx_queue          788 drivers/net/ethernet/sfc/efx.c 	efx_for_each_channel_rx_queue(rx_queue, channel)
rx_queue          789 drivers/net/ethernet/sfc/efx.c 		efx_remove_rx_queue(rx_queue);
rx_queue          820 drivers/net/ethernet/sfc/efx.c 		struct efx_rx_queue *rx_queue;
rx_queue          828 drivers/net/ethernet/sfc/efx.c 		efx_for_each_channel_rx_queue(rx_queue, channel)
rx_queue          830 drivers/net/ethernet/sfc/efx.c 						rx_queue->rxd.index +
rx_queue          831 drivers/net/ethernet/sfc/efx.c 						rx_queue->rxd.entries);
rx_queue          914 drivers/net/ethernet/sfc/efx.c void efx_schedule_slow_fill(struct efx_rx_queue *rx_queue)
rx_queue          916 drivers/net/ethernet/sfc/efx.c 	mod_timer(&rx_queue->slow_fill, jiffies + msecs_to_jiffies(10));
rx_queue         1743 drivers/net/ethernet/sfc/efx.c 			channel->rx_queue.core_index = channel->channel;
rx_queue         1745 drivers/net/ethernet/sfc/efx.c 			channel->rx_queue.core_index = -1;
rx_queue           37 drivers/net/ethernet/sfc/efx.h int efx_probe_rx_queue(struct efx_rx_queue *rx_queue);
rx_queue           38 drivers/net/ethernet/sfc/efx.h void efx_remove_rx_queue(struct efx_rx_queue *rx_queue);
rx_queue           39 drivers/net/ethernet/sfc/efx.h void efx_init_rx_queue(struct efx_rx_queue *rx_queue);
rx_queue           40 drivers/net/ethernet/sfc/efx.h void efx_fini_rx_queue(struct efx_rx_queue *rx_queue);
rx_queue           41 drivers/net/ethernet/sfc/efx.h void efx_fast_push_rx_descriptors(struct efx_rx_queue *rx_queue, bool atomic);
rx_queue           44 drivers/net/ethernet/sfc/efx.h void efx_rx_packet(struct efx_rx_queue *rx_queue, unsigned int index,
rx_queue           51 drivers/net/ethernet/sfc/efx.h void efx_schedule_slow_fill(struct efx_rx_queue *rx_queue);
rx_queue          458 drivers/net/ethernet/sfc/ethtool.c 	struct efx_rx_queue *rx_queue;
rx_queue          506 drivers/net/ethernet/sfc/ethtool.c 			efx_for_each_channel_rx_queue(rx_queue, channel) {
rx_queue          507 drivers/net/ethernet/sfc/ethtool.c 				*data += rx_queue->rx_packets;
rx_queue          253 drivers/net/ethernet/sfc/falcon/efx.c 		struct ef4_rx_queue *rx_queue =
rx_queue          257 drivers/net/ethernet/sfc/falcon/efx.c 		ef4_fast_push_rx_descriptors(rx_queue, true);
rx_queue          428 drivers/net/ethernet/sfc/falcon/efx.c 	struct ef4_rx_queue *rx_queue;
rx_queue          447 drivers/net/ethernet/sfc/falcon/efx.c 	rx_queue = &channel->rx_queue;
rx_queue          448 drivers/net/ethernet/sfc/falcon/efx.c 	rx_queue->efx = efx;
rx_queue          449 drivers/net/ethernet/sfc/falcon/efx.c 	timer_setup(&rx_queue->slow_fill, ef4_rx_slow_fill, 0);
rx_queue          461 drivers/net/ethernet/sfc/falcon/efx.c 	struct ef4_rx_queue *rx_queue;
rx_queue          485 drivers/net/ethernet/sfc/falcon/efx.c 	rx_queue = &channel->rx_queue;
rx_queue          486 drivers/net/ethernet/sfc/falcon/efx.c 	rx_queue->buffer = NULL;
rx_queue          487 drivers/net/ethernet/sfc/falcon/efx.c 	memset(&rx_queue->rxd, 0, sizeof(rx_queue->rxd));
rx_queue          488 drivers/net/ethernet/sfc/falcon/efx.c 	timer_setup(&rx_queue->slow_fill, ef4_rx_slow_fill, 0);
rx_queue          496 drivers/net/ethernet/sfc/falcon/efx.c 	struct ef4_rx_queue *rx_queue;
rx_queue          516 drivers/net/ethernet/sfc/falcon/efx.c 	ef4_for_each_channel_rx_queue(rx_queue, channel) {
rx_queue          517 drivers/net/ethernet/sfc/falcon/efx.c 		rc = ef4_probe_rx_queue(rx_queue);
rx_queue          598 drivers/net/ethernet/sfc/falcon/efx.c 	struct ef4_rx_queue *rx_queue;
rx_queue          670 drivers/net/ethernet/sfc/falcon/efx.c 		ef4_for_each_channel_rx_queue(rx_queue, channel) {
rx_queue          671 drivers/net/ethernet/sfc/falcon/efx.c 			ef4_init_rx_queue(rx_queue);
rx_queue          674 drivers/net/ethernet/sfc/falcon/efx.c 			ef4_fast_push_rx_descriptors(rx_queue, false);
rx_queue          689 drivers/net/ethernet/sfc/falcon/efx.c 	struct ef4_rx_queue *rx_queue;
rx_queue          697 drivers/net/ethernet/sfc/falcon/efx.c 		ef4_for_each_channel_rx_queue(rx_queue, channel)
rx_queue          698 drivers/net/ethernet/sfc/falcon/efx.c 			rx_queue->refill_enabled = false;
rx_queue          732 drivers/net/ethernet/sfc/falcon/efx.c 		ef4_for_each_channel_rx_queue(rx_queue, channel)
rx_queue          733 drivers/net/ethernet/sfc/falcon/efx.c 			ef4_fini_rx_queue(rx_queue);
rx_queue          742 drivers/net/ethernet/sfc/falcon/efx.c 	struct ef4_rx_queue *rx_queue;
rx_queue          747 drivers/net/ethernet/sfc/falcon/efx.c 	ef4_for_each_channel_rx_queue(rx_queue, channel)
rx_queue          748 drivers/net/ethernet/sfc/falcon/efx.c 		ef4_remove_rx_queue(rx_queue);
rx_queue          779 drivers/net/ethernet/sfc/falcon/efx.c 		struct ef4_rx_queue *rx_queue;
rx_queue          787 drivers/net/ethernet/sfc/falcon/efx.c 		ef4_for_each_channel_rx_queue(rx_queue, channel)
rx_queue          789 drivers/net/ethernet/sfc/falcon/efx.c 						rx_queue->rxd.index +
rx_queue          790 drivers/net/ethernet/sfc/falcon/efx.c 						rx_queue->rxd.entries);
rx_queue          873 drivers/net/ethernet/sfc/falcon/efx.c void ef4_schedule_slow_fill(struct ef4_rx_queue *rx_queue)
rx_queue          875 drivers/net/ethernet/sfc/falcon/efx.c 	mod_timer(&rx_queue->slow_fill, jiffies + msecs_to_jiffies(100));
rx_queue         1608 drivers/net/ethernet/sfc/falcon/efx.c 			channel->rx_queue.core_index = channel->channel;
rx_queue         1610 drivers/net/ethernet/sfc/falcon/efx.c 			channel->rx_queue.core_index = -1;
rx_queue           40 drivers/net/ethernet/sfc/falcon/efx.h int ef4_probe_rx_queue(struct ef4_rx_queue *rx_queue);
rx_queue           41 drivers/net/ethernet/sfc/falcon/efx.h void ef4_remove_rx_queue(struct ef4_rx_queue *rx_queue);
rx_queue           42 drivers/net/ethernet/sfc/falcon/efx.h void ef4_init_rx_queue(struct ef4_rx_queue *rx_queue);
rx_queue           43 drivers/net/ethernet/sfc/falcon/efx.h void ef4_fini_rx_queue(struct ef4_rx_queue *rx_queue);
rx_queue           44 drivers/net/ethernet/sfc/falcon/efx.h void ef4_fast_push_rx_descriptors(struct ef4_rx_queue *rx_queue, bool atomic);
rx_queue           47 drivers/net/ethernet/sfc/falcon/efx.h void ef4_rx_packet(struct ef4_rx_queue *rx_queue, unsigned int index,
rx_queue           54 drivers/net/ethernet/sfc/falcon/efx.h void ef4_schedule_slow_fill(struct ef4_rx_queue *rx_queue);
rx_queue          438 drivers/net/ethernet/sfc/falcon/ethtool.c 	struct ef4_rx_queue *rx_queue;
rx_queue          486 drivers/net/ethernet/sfc/falcon/ethtool.c 			ef4_for_each_channel_rx_queue(rx_queue, channel) {
rx_queue          487 drivers/net/ethernet/sfc/falcon/ethtool.c 				*data += rx_queue->rx_packets;
rx_queue          471 drivers/net/ethernet/sfc/falcon/farch.c ef4_farch_build_rx_desc(struct ef4_rx_queue *rx_queue, unsigned index)
rx_queue          476 drivers/net/ethernet/sfc/falcon/farch.c 	rxd = ef4_rx_desc(rx_queue, index);
rx_queue          477 drivers/net/ethernet/sfc/falcon/farch.c 	rx_buf = ef4_rx_buffer(rx_queue, index);
rx_queue          481 drivers/net/ethernet/sfc/falcon/farch.c 			     rx_queue->efx->type->rx_buffer_padding,
rx_queue          489 drivers/net/ethernet/sfc/falcon/farch.c void ef4_farch_rx_write(struct ef4_rx_queue *rx_queue)
rx_queue          491 drivers/net/ethernet/sfc/falcon/farch.c 	struct ef4_nic *efx = rx_queue->efx;
rx_queue          495 drivers/net/ethernet/sfc/falcon/farch.c 	while (rx_queue->notified_count != rx_queue->added_count) {
rx_queue          497 drivers/net/ethernet/sfc/falcon/farch.c 			rx_queue,
rx_queue          498 drivers/net/ethernet/sfc/falcon/farch.c 			rx_queue->notified_count & rx_queue->ptr_mask);
rx_queue          499 drivers/net/ethernet/sfc/falcon/farch.c 		++rx_queue->notified_count;
rx_queue          503 drivers/net/ethernet/sfc/falcon/farch.c 	write_ptr = rx_queue->added_count & rx_queue->ptr_mask;
rx_queue          506 drivers/net/ethernet/sfc/falcon/farch.c 			ef4_rx_queue_index(rx_queue));
rx_queue          509 drivers/net/ethernet/sfc/falcon/farch.c int ef4_farch_rx_probe(struct ef4_rx_queue *rx_queue)
rx_queue          511 drivers/net/ethernet/sfc/falcon/farch.c 	struct ef4_nic *efx = rx_queue->efx;
rx_queue          514 drivers/net/ethernet/sfc/falcon/farch.c 	entries = rx_queue->ptr_mask + 1;
rx_queue          515 drivers/net/ethernet/sfc/falcon/farch.c 	return ef4_alloc_special_buffer(efx, &rx_queue->rxd,
rx_queue          519 drivers/net/ethernet/sfc/falcon/farch.c void ef4_farch_rx_init(struct ef4_rx_queue *rx_queue)
rx_queue          522 drivers/net/ethernet/sfc/falcon/farch.c 	struct ef4_nic *efx = rx_queue->efx;
rx_queue          536 drivers/net/ethernet/sfc/falcon/farch.c 		  ef4_rx_queue_index(rx_queue), rx_queue->rxd.index,
rx_queue          537 drivers/net/ethernet/sfc/falcon/farch.c 		  rx_queue->rxd.index + rx_queue->rxd.entries - 1);
rx_queue          539 drivers/net/ethernet/sfc/falcon/farch.c 	rx_queue->scatter_n = 0;
rx_queue          542 drivers/net/ethernet/sfc/falcon/farch.c 	ef4_init_special_buffer(efx, &rx_queue->rxd);
rx_queue          548 drivers/net/ethernet/sfc/falcon/farch.c 			      FRF_AZ_RX_DESCQ_BUF_BASE_ID, rx_queue->rxd.index,
rx_queue          550 drivers/net/ethernet/sfc/falcon/farch.c 			      ef4_rx_queue_channel(rx_queue)->channel,
rx_queue          553 drivers/net/ethernet/sfc/falcon/farch.c 			      ef4_rx_queue_index(rx_queue),
rx_queue          555 drivers/net/ethernet/sfc/falcon/farch.c 			      __ffs(rx_queue->rxd.entries),
rx_queue          560 drivers/net/ethernet/sfc/falcon/farch.c 			 ef4_rx_queue_index(rx_queue));
rx_queue          563 drivers/net/ethernet/sfc/falcon/farch.c static void ef4_farch_flush_rx_queue(struct ef4_rx_queue *rx_queue)
rx_queue          565 drivers/net/ethernet/sfc/falcon/farch.c 	struct ef4_nic *efx = rx_queue->efx;
rx_queue          571 drivers/net/ethernet/sfc/falcon/farch.c 			     ef4_rx_queue_index(rx_queue));
rx_queue          575 drivers/net/ethernet/sfc/falcon/farch.c void ef4_farch_rx_fini(struct ef4_rx_queue *rx_queue)
rx_queue          578 drivers/net/ethernet/sfc/falcon/farch.c 	struct ef4_nic *efx = rx_queue->efx;
rx_queue          583 drivers/net/ethernet/sfc/falcon/farch.c 			 ef4_rx_queue_index(rx_queue));
rx_queue          586 drivers/net/ethernet/sfc/falcon/farch.c 	ef4_fini_special_buffer(efx, &rx_queue->rxd);
rx_queue          590 drivers/net/ethernet/sfc/falcon/farch.c void ef4_farch_rx_remove(struct ef4_rx_queue *rx_queue)
rx_queue          592 drivers/net/ethernet/sfc/falcon/farch.c 	ef4_free_special_buffer(rx_queue->efx, &rx_queue->rxd);
rx_queue          662 drivers/net/ethernet/sfc/falcon/farch.c 	struct ef4_rx_queue *rx_queue;
rx_queue          670 drivers/net/ethernet/sfc/falcon/farch.c 		ef4_for_each_channel_rx_queue(rx_queue, channel) {
rx_queue          671 drivers/net/ethernet/sfc/falcon/farch.c 			rx_queue->flush_pending = true;
rx_queue          682 drivers/net/ethernet/sfc/falcon/farch.c 			ef4_for_each_channel_rx_queue(rx_queue, channel) {
rx_queue          687 drivers/net/ethernet/sfc/falcon/farch.c 				if (rx_queue->flush_pending) {
rx_queue          688 drivers/net/ethernet/sfc/falcon/farch.c 					rx_queue->flush_pending = false;
rx_queue          691 drivers/net/ethernet/sfc/falcon/farch.c 					ef4_farch_flush_rx_queue(rx_queue);
rx_queue          721 drivers/net/ethernet/sfc/falcon/farch.c 	struct ef4_rx_queue *rx_queue;
rx_queue          734 drivers/net/ethernet/sfc/falcon/farch.c 			ef4_for_each_channel_rx_queue(rx_queue, channel)
rx_queue          735 drivers/net/ethernet/sfc/falcon/farch.c 				ef4_farch_rx_fini(rx_queue);
rx_queue          868 drivers/net/ethernet/sfc/falcon/farch.c static u16 ef4_farch_handle_rx_not_ok(struct ef4_rx_queue *rx_queue,
rx_queue          871 drivers/net/ethernet/sfc/falcon/farch.c 	struct ef4_channel *channel = ef4_rx_queue_channel(rx_queue);
rx_queue          872 drivers/net/ethernet/sfc/falcon/farch.c 	struct ef4_nic *efx = rx_queue->efx;
rx_queue          923 drivers/net/ethernet/sfc/falcon/farch.c 			  ef4_rx_queue_index(rx_queue), EF4_QWORD_VAL(*event),
rx_queue          948 drivers/net/ethernet/sfc/falcon/farch.c ef4_farch_handle_rx_bad_index(struct ef4_rx_queue *rx_queue, unsigned index)
rx_queue          950 drivers/net/ethernet/sfc/falcon/farch.c 	struct ef4_channel *channel = ef4_rx_queue_channel(rx_queue);
rx_queue          951 drivers/net/ethernet/sfc/falcon/farch.c 	struct ef4_nic *efx = rx_queue->efx;
rx_queue          954 drivers/net/ethernet/sfc/falcon/farch.c 	if (rx_queue->scatter_n &&
rx_queue          955 drivers/net/ethernet/sfc/falcon/farch.c 	    index == ((rx_queue->removed_count + rx_queue->scatter_n - 1) &
rx_queue          956 drivers/net/ethernet/sfc/falcon/farch.c 		      rx_queue->ptr_mask)) {
rx_queue          961 drivers/net/ethernet/sfc/falcon/farch.c 	expected = rx_queue->removed_count & rx_queue->ptr_mask;
rx_queue          962 drivers/net/ethernet/sfc/falcon/farch.c 	dropped = (index - expected) & rx_queue->ptr_mask;
rx_queue          987 drivers/net/ethernet/sfc/falcon/farch.c 	struct ef4_rx_queue *rx_queue;
rx_queue          998 drivers/net/ethernet/sfc/falcon/farch.c 	rx_queue = ef4_channel_get_rx_queue(channel);
rx_queue         1001 drivers/net/ethernet/sfc/falcon/farch.c 	expected_ptr = ((rx_queue->removed_count + rx_queue->scatter_n) &
rx_queue         1002 drivers/net/ethernet/sfc/falcon/farch.c 			rx_queue->ptr_mask);
rx_queue         1006 drivers/net/ethernet/sfc/falcon/farch.c 	    unlikely(rx_ev_sop != (rx_queue->scatter_n == 0))) {
rx_queue         1008 drivers/net/ethernet/sfc/falcon/farch.c 		    !ef4_farch_handle_rx_bad_index(rx_queue, rx_ev_desc_ptr))
rx_queue         1012 drivers/net/ethernet/sfc/falcon/farch.c 		if (rx_queue->scatter_n) {
rx_queue         1014 drivers/net/ethernet/sfc/falcon/farch.c 				rx_queue,
rx_queue         1015 drivers/net/ethernet/sfc/falcon/farch.c 				rx_queue->removed_count & rx_queue->ptr_mask,
rx_queue         1016 drivers/net/ethernet/sfc/falcon/farch.c 				rx_queue->scatter_n, 0, EF4_RX_PKT_DISCARD);
rx_queue         1017 drivers/net/ethernet/sfc/falcon/farch.c 			rx_queue->removed_count += rx_queue->scatter_n;
rx_queue         1018 drivers/net/ethernet/sfc/falcon/farch.c 			rx_queue->scatter_n = 0;
rx_queue         1028 drivers/net/ethernet/sfc/falcon/farch.c 				rx_queue,
rx_queue         1029 drivers/net/ethernet/sfc/falcon/farch.c 				rx_queue->removed_count & rx_queue->ptr_mask,
rx_queue         1031 drivers/net/ethernet/sfc/falcon/farch.c 			++rx_queue->removed_count;
rx_queue         1036 drivers/net/ethernet/sfc/falcon/farch.c 	++rx_queue->scatter_n;
rx_queue         1061 drivers/net/ethernet/sfc/falcon/farch.c 		flags = ef4_farch_handle_rx_not_ok(rx_queue, event);
rx_queue         1079 drivers/net/ethernet/sfc/falcon/farch.c 	ef4_rx_packet(rx_queue,
rx_queue         1080 drivers/net/ethernet/sfc/falcon/farch.c 		      rx_queue->removed_count & rx_queue->ptr_mask,
rx_queue         1081 drivers/net/ethernet/sfc/falcon/farch.c 		      rx_queue->scatter_n, rx_ev_byte_cnt, flags);
rx_queue         1082 drivers/net/ethernet/sfc/falcon/farch.c 	rx_queue->removed_count += rx_queue->scatter_n;
rx_queue         1083 drivers/net/ethernet/sfc/falcon/farch.c 	rx_queue->scatter_n = 0;
rx_queue         1115 drivers/net/ethernet/sfc/falcon/farch.c 	struct ef4_rx_queue *rx_queue;
rx_queue         1126 drivers/net/ethernet/sfc/falcon/farch.c 	rx_queue = ef4_channel_get_rx_queue(channel);
rx_queue         1131 drivers/net/ethernet/sfc/falcon/farch.c 		rx_queue->flush_pending = true;
rx_queue         1134 drivers/net/ethernet/sfc/falcon/farch.c 		ef4_farch_magic_event(ef4_rx_queue_channel(rx_queue),
rx_queue         1135 drivers/net/ethernet/sfc/falcon/farch.c 				      EF4_CHANNEL_MAGIC_RX_DRAIN(rx_queue));
rx_queue         1157 drivers/net/ethernet/sfc/falcon/farch.c 	struct ef4_rx_queue *rx_queue =
rx_queue         1167 drivers/net/ethernet/sfc/falcon/farch.c 	} else if (rx_queue && magic == EF4_CHANNEL_MAGIC_FILL(rx_queue)) {
rx_queue         1171 drivers/net/ethernet/sfc/falcon/farch.c 		ef4_fast_push_rx_descriptors(rx_queue, true);
rx_queue         1172 drivers/net/ethernet/sfc/falcon/farch.c 	} else if (rx_queue && magic == EF4_CHANNEL_MAGIC_RX_DRAIN(rx_queue)) {
rx_queue         1391 drivers/net/ethernet/sfc/falcon/farch.c void ef4_farch_rx_defer_refill(struct ef4_rx_queue *rx_queue)
rx_queue         1393 drivers/net/ethernet/sfc/falcon/farch.c 	ef4_farch_magic_event(ef4_rx_queue_channel(rx_queue),
rx_queue         1394 drivers/net/ethernet/sfc/falcon/farch.c 			      EF4_CHANNEL_MAGIC_FILL(rx_queue));
rx_queue          444 drivers/net/ethernet/sfc/falcon/net_driver.h 	struct ef4_rx_queue rx_queue;
rx_queue         1090 drivers/net/ethernet/sfc/falcon/net_driver.h 	int (*rx_probe)(struct ef4_rx_queue *rx_queue);
rx_queue         1091 drivers/net/ethernet/sfc/falcon/net_driver.h 	void (*rx_init)(struct ef4_rx_queue *rx_queue);
rx_queue         1092 drivers/net/ethernet/sfc/falcon/net_driver.h 	void (*rx_remove)(struct ef4_rx_queue *rx_queue);
rx_queue         1093 drivers/net/ethernet/sfc/falcon/net_driver.h 	void (*rx_write)(struct ef4_rx_queue *rx_queue);
rx_queue         1094 drivers/net/ethernet/sfc/falcon/net_driver.h 	void (*rx_defer_refill)(struct ef4_rx_queue *rx_queue);
rx_queue         1236 drivers/net/ethernet/sfc/falcon/net_driver.h 	return channel->rx_queue.core_index >= 0;
rx_queue         1243 drivers/net/ethernet/sfc/falcon/net_driver.h 	return &channel->rx_queue;
rx_queue         1251 drivers/net/ethernet/sfc/falcon/net_driver.h 		for (_rx_queue = &(_channel)->rx_queue;			\
rx_queue         1256 drivers/net/ethernet/sfc/falcon/net_driver.h ef4_rx_queue_channel(struct ef4_rx_queue *rx_queue)
rx_queue         1258 drivers/net/ethernet/sfc/falcon/net_driver.h 	return container_of(rx_queue, struct ef4_channel, rx_queue);
rx_queue         1261 drivers/net/ethernet/sfc/falcon/net_driver.h static inline int ef4_rx_queue_index(struct ef4_rx_queue *rx_queue)
rx_queue         1263 drivers/net/ethernet/sfc/falcon/net_driver.h 	return ef4_rx_queue_channel(rx_queue)->channel;
rx_queue         1269 drivers/net/ethernet/sfc/falcon/net_driver.h static inline struct ef4_rx_buffer *ef4_rx_buffer(struct ef4_rx_queue *rx_queue,
rx_queue         1272 drivers/net/ethernet/sfc/falcon/net_driver.h 	return &rx_queue->buffer[index];
rx_queue          110 drivers/net/ethernet/sfc/falcon/nic.h ef4_rx_desc(struct ef4_rx_queue *rx_queue, unsigned int index)
rx_queue          112 drivers/net/ethernet/sfc/falcon/nic.h 	return ((ef4_qword_t *) (rx_queue->rxd.buf.addr)) + index;
rx_queue          333 drivers/net/ethernet/sfc/falcon/nic.h static inline int ef4_nic_probe_rx(struct ef4_rx_queue *rx_queue)
rx_queue          335 drivers/net/ethernet/sfc/falcon/nic.h 	return rx_queue->efx->type->rx_probe(rx_queue);
rx_queue          337 drivers/net/ethernet/sfc/falcon/nic.h static inline void ef4_nic_init_rx(struct ef4_rx_queue *rx_queue)
rx_queue          339 drivers/net/ethernet/sfc/falcon/nic.h 	rx_queue->efx->type->rx_init(rx_queue);
rx_queue          341 drivers/net/ethernet/sfc/falcon/nic.h static inline void ef4_nic_remove_rx(struct ef4_rx_queue *rx_queue)
rx_queue          343 drivers/net/ethernet/sfc/falcon/nic.h 	rx_queue->efx->type->rx_remove(rx_queue);
rx_queue          345 drivers/net/ethernet/sfc/falcon/nic.h static inline void ef4_nic_notify_rx_desc(struct ef4_rx_queue *rx_queue)
rx_queue          347 drivers/net/ethernet/sfc/falcon/nic.h 	rx_queue->efx->type->rx_write(rx_queue);
rx_queue          349 drivers/net/ethernet/sfc/falcon/nic.h static inline void ef4_nic_generate_fill_event(struct ef4_rx_queue *rx_queue)
rx_queue          351 drivers/net/ethernet/sfc/falcon/nic.h 	rx_queue->efx->type->rx_defer_refill(rx_queue);
rx_queue          390 drivers/net/ethernet/sfc/falcon/nic.h int ef4_farch_rx_probe(struct ef4_rx_queue *rx_queue);
rx_queue          391 drivers/net/ethernet/sfc/falcon/nic.h void ef4_farch_rx_init(struct ef4_rx_queue *rx_queue);
rx_queue          392 drivers/net/ethernet/sfc/falcon/nic.h void ef4_farch_rx_fini(struct ef4_rx_queue *rx_queue);
rx_queue          393 drivers/net/ethernet/sfc/falcon/nic.h void ef4_farch_rx_remove(struct ef4_rx_queue *rx_queue);
rx_queue          394 drivers/net/ethernet/sfc/falcon/nic.h void ef4_farch_rx_write(struct ef4_rx_queue *rx_queue);
rx_queue          395 drivers/net/ethernet/sfc/falcon/nic.h void ef4_farch_rx_defer_refill(struct ef4_rx_queue *rx_queue);
rx_queue           76 drivers/net/ethernet/sfc/falcon/rx.c ef4_rx_buf_next(struct ef4_rx_queue *rx_queue, struct ef4_rx_buffer *rx_buf)
rx_queue           78 drivers/net/ethernet/sfc/falcon/rx.c 	if (unlikely(rx_buf == ef4_rx_buffer(rx_queue, rx_queue->ptr_mask)))
rx_queue           79 drivers/net/ethernet/sfc/falcon/rx.c 		return ef4_rx_buffer(rx_queue, 0);
rx_queue          106 drivers/net/ethernet/sfc/falcon/rx.c static struct page *ef4_reuse_page(struct ef4_rx_queue *rx_queue)
rx_queue          108 drivers/net/ethernet/sfc/falcon/rx.c 	struct ef4_nic *efx = rx_queue->efx;
rx_queue          113 drivers/net/ethernet/sfc/falcon/rx.c 	index = rx_queue->page_remove & rx_queue->page_ptr_mask;
rx_queue          114 drivers/net/ethernet/sfc/falcon/rx.c 	page = rx_queue->page_ring[index];
rx_queue          118 drivers/net/ethernet/sfc/falcon/rx.c 	rx_queue->page_ring[index] = NULL;
rx_queue          120 drivers/net/ethernet/sfc/falcon/rx.c 	if (rx_queue->page_remove != rx_queue->page_add)
rx_queue          121 drivers/net/ethernet/sfc/falcon/rx.c 		++rx_queue->page_remove;
rx_queue          125 drivers/net/ethernet/sfc/falcon/rx.c 		++rx_queue->page_recycle_count;
rx_queue          133 drivers/net/ethernet/sfc/falcon/rx.c 		++rx_queue->page_recycle_failed;
rx_queue          149 drivers/net/ethernet/sfc/falcon/rx.c static int ef4_init_rx_buffers(struct ef4_rx_queue *rx_queue, bool atomic)
rx_queue          151 drivers/net/ethernet/sfc/falcon/rx.c 	struct ef4_nic *efx = rx_queue->efx;
rx_queue          161 drivers/net/ethernet/sfc/falcon/rx.c 		page = ef4_reuse_page(rx_queue);
rx_queue          188 drivers/net/ethernet/sfc/falcon/rx.c 			index = rx_queue->added_count & rx_queue->ptr_mask;
rx_queue          189 drivers/net/ethernet/sfc/falcon/rx.c 			rx_buf = ef4_rx_buffer(rx_queue, index);
rx_queue          195 drivers/net/ethernet/sfc/falcon/rx.c 			++rx_queue->added_count;
rx_queue          224 drivers/net/ethernet/sfc/falcon/rx.c static void ef4_free_rx_buffers(struct ef4_rx_queue *rx_queue,
rx_queue          233 drivers/net/ethernet/sfc/falcon/rx.c 		rx_buf = ef4_rx_buf_next(rx_queue, rx_buf);
rx_queue          245 drivers/net/ethernet/sfc/falcon/rx.c 	struct ef4_rx_queue *rx_queue = ef4_channel_get_rx_queue(channel);
rx_queue          246 drivers/net/ethernet/sfc/falcon/rx.c 	struct ef4_nic *efx = rx_queue->efx;
rx_queue          253 drivers/net/ethernet/sfc/falcon/rx.c 	index = rx_queue->page_add & rx_queue->page_ptr_mask;
rx_queue          254 drivers/net/ethernet/sfc/falcon/rx.c 	if (rx_queue->page_ring[index] == NULL) {
rx_queue          255 drivers/net/ethernet/sfc/falcon/rx.c 		unsigned read_index = rx_queue->page_remove &
rx_queue          256 drivers/net/ethernet/sfc/falcon/rx.c 			rx_queue->page_ptr_mask;
rx_queue          263 drivers/net/ethernet/sfc/falcon/rx.c 			++rx_queue->page_remove;
rx_queue          264 drivers/net/ethernet/sfc/falcon/rx.c 		rx_queue->page_ring[index] = page;
rx_queue          265 drivers/net/ethernet/sfc/falcon/rx.c 		++rx_queue->page_add;
rx_queue          268 drivers/net/ethernet/sfc/falcon/rx.c 	++rx_queue->page_recycle_full;
rx_queue          273 drivers/net/ethernet/sfc/falcon/rx.c static void ef4_fini_rx_buffer(struct ef4_rx_queue *rx_queue,
rx_queue          282 drivers/net/ethernet/sfc/falcon/rx.c 		ef4_unmap_rx_buffer(rx_queue->efx, rx_buf);
rx_queue          283 drivers/net/ethernet/sfc/falcon/rx.c 		ef4_free_rx_buffers(rx_queue, rx_buf, 1);
rx_queue          293 drivers/net/ethernet/sfc/falcon/rx.c 	struct ef4_rx_queue *rx_queue = ef4_channel_get_rx_queue(channel);
rx_queue          297 drivers/net/ethernet/sfc/falcon/rx.c 		rx_buf = ef4_rx_buf_next(rx_queue, rx_buf);
rx_queue          305 drivers/net/ethernet/sfc/falcon/rx.c 	struct ef4_rx_queue *rx_queue = ef4_channel_get_rx_queue(channel);
rx_queue          309 drivers/net/ethernet/sfc/falcon/rx.c 	ef4_free_rx_buffers(rx_queue, rx_buf, n_frags);
rx_queue          324 drivers/net/ethernet/sfc/falcon/rx.c void ef4_fast_push_rx_descriptors(struct ef4_rx_queue *rx_queue, bool atomic)
rx_queue          326 drivers/net/ethernet/sfc/falcon/rx.c 	struct ef4_nic *efx = rx_queue->efx;
rx_queue          330 drivers/net/ethernet/sfc/falcon/rx.c 	if (!rx_queue->refill_enabled)
rx_queue          334 drivers/net/ethernet/sfc/falcon/rx.c 	fill_level = (rx_queue->added_count - rx_queue->removed_count);
rx_queue          335 drivers/net/ethernet/sfc/falcon/rx.c 	EF4_BUG_ON_PARANOID(fill_level > rx_queue->efx->rxq_entries);
rx_queue          336 drivers/net/ethernet/sfc/falcon/rx.c 	if (fill_level >= rx_queue->fast_fill_trigger)
rx_queue          340 drivers/net/ethernet/sfc/falcon/rx.c 	if (unlikely(fill_level < rx_queue->min_fill)) {
rx_queue          342 drivers/net/ethernet/sfc/falcon/rx.c 			rx_queue->min_fill = fill_level;
rx_queue          346 drivers/net/ethernet/sfc/falcon/rx.c 	space = rx_queue->max_fill - fill_level;
rx_queue          349 drivers/net/ethernet/sfc/falcon/rx.c 	netif_vdbg(rx_queue->efx, rx_status, rx_queue->efx->net_dev,
rx_queue          352 drivers/net/ethernet/sfc/falcon/rx.c 		   ef4_rx_queue_index(rx_queue), fill_level,
rx_queue          353 drivers/net/ethernet/sfc/falcon/rx.c 		   rx_queue->max_fill);
rx_queue          357 drivers/net/ethernet/sfc/falcon/rx.c 		rc = ef4_init_rx_buffers(rx_queue, atomic);
rx_queue          360 drivers/net/ethernet/sfc/falcon/rx.c 			if (rx_queue->added_count == rx_queue->removed_count)
rx_queue          361 drivers/net/ethernet/sfc/falcon/rx.c 				ef4_schedule_slow_fill(rx_queue);
rx_queue          366 drivers/net/ethernet/sfc/falcon/rx.c 	netif_vdbg(rx_queue->efx, rx_status, rx_queue->efx->net_dev,
rx_queue          368 drivers/net/ethernet/sfc/falcon/rx.c 		   "to level %d\n", ef4_rx_queue_index(rx_queue),
rx_queue          369 drivers/net/ethernet/sfc/falcon/rx.c 		   rx_queue->added_count - rx_queue->removed_count);
rx_queue          372 drivers/net/ethernet/sfc/falcon/rx.c 	if (rx_queue->notified_count != rx_queue->added_count)
rx_queue          373 drivers/net/ethernet/sfc/falcon/rx.c 		ef4_nic_notify_rx_desc(rx_queue);
rx_queue          378 drivers/net/ethernet/sfc/falcon/rx.c 	struct ef4_rx_queue *rx_queue = from_timer(rx_queue, t, slow_fill);
rx_queue          381 drivers/net/ethernet/sfc/falcon/rx.c 	ef4_nic_generate_fill_event(rx_queue);
rx_queue          382 drivers/net/ethernet/sfc/falcon/rx.c 	++rx_queue->slow_fill_count;
rx_queue          385 drivers/net/ethernet/sfc/falcon/rx.c static void ef4_rx_packet__check_len(struct ef4_rx_queue *rx_queue,
rx_queue          389 drivers/net/ethernet/sfc/falcon/rx.c 	struct ef4_nic *efx = rx_queue->efx;
rx_queue          405 drivers/net/ethernet/sfc/falcon/rx.c 				  ef4_rx_queue_index(rx_queue), len, max_len,
rx_queue          413 drivers/net/ethernet/sfc/falcon/rx.c 				  ef4_rx_queue_index(rx_queue), len, max_len);
rx_queue          416 drivers/net/ethernet/sfc/falcon/rx.c 	ef4_rx_queue_channel(rx_queue)->n_rx_overlength++;
rx_queue          432 drivers/net/ethernet/sfc/falcon/rx.c 		struct ef4_rx_queue *rx_queue;
rx_queue          434 drivers/net/ethernet/sfc/falcon/rx.c 		rx_queue = ef4_channel_get_rx_queue(channel);
rx_queue          435 drivers/net/ethernet/sfc/falcon/rx.c 		ef4_free_rx_buffers(rx_queue, rx_buf, n_frags);
rx_queue          454 drivers/net/ethernet/sfc/falcon/rx.c 		rx_buf = ef4_rx_buf_next(&channel->rx_queue, rx_buf);
rx_queue          460 drivers/net/ethernet/sfc/falcon/rx.c 	skb_record_rx_queue(skb, channel->rx_queue.core_index);
rx_queue          505 drivers/net/ethernet/sfc/falcon/rx.c 			rx_buf = ef4_rx_buf_next(&channel->rx_queue, rx_buf);
rx_queue          523 drivers/net/ethernet/sfc/falcon/rx.c void ef4_rx_packet(struct ef4_rx_queue *rx_queue, unsigned int index,
rx_queue          526 drivers/net/ethernet/sfc/falcon/rx.c 	struct ef4_nic *efx = rx_queue->efx;
rx_queue          527 drivers/net/ethernet/sfc/falcon/rx.c 	struct ef4_channel *channel = ef4_rx_queue_channel(rx_queue);
rx_queue          530 drivers/net/ethernet/sfc/falcon/rx.c 	rx_queue->rx_packets++;
rx_queue          532 drivers/net/ethernet/sfc/falcon/rx.c 	rx_buf = ef4_rx_buffer(rx_queue, index);
rx_queue          538 drivers/net/ethernet/sfc/falcon/rx.c 			ef4_rx_packet__check_len(rx_queue, rx_buf, len);
rx_queue          552 drivers/net/ethernet/sfc/falcon/rx.c 		   ef4_rx_queue_index(rx_queue), index,
rx_queue          553 drivers/net/ethernet/sfc/falcon/rx.c 		   (index + n_frags - 1) & rx_queue->ptr_mask, len,
rx_queue          589 drivers/net/ethernet/sfc/falcon/rx.c 			rx_buf = ef4_rx_buf_next(rx_queue, rx_buf);
rx_queue          599 drivers/net/ethernet/sfc/falcon/rx.c 	rx_buf = ef4_rx_buffer(rx_queue, index);
rx_queue          619 drivers/net/ethernet/sfc/falcon/rx.c 		struct ef4_rx_queue *rx_queue;
rx_queue          621 drivers/net/ethernet/sfc/falcon/rx.c 		rx_queue = ef4_channel_get_rx_queue(channel);
rx_queue          622 drivers/net/ethernet/sfc/falcon/rx.c 		ef4_free_rx_buffers(rx_queue, rx_buf, n_frags);
rx_queue          625 drivers/net/ethernet/sfc/falcon/rx.c 	skb_record_rx_queue(skb, channel->rx_queue.core_index);
rx_queue          645 drivers/net/ethernet/sfc/falcon/rx.c 		ef4_rx_buffer(&channel->rx_queue, channel->rx_pkt_index);
rx_queue          659 drivers/net/ethernet/sfc/falcon/rx.c 		struct ef4_rx_queue *rx_queue;
rx_queue          662 drivers/net/ethernet/sfc/falcon/rx.c 		rx_queue = ef4_channel_get_rx_queue(channel);
rx_queue          663 drivers/net/ethernet/sfc/falcon/rx.c 		ef4_free_rx_buffers(rx_queue, rx_buf,
rx_queue          679 drivers/net/ethernet/sfc/falcon/rx.c int ef4_probe_rx_queue(struct ef4_rx_queue *rx_queue)
rx_queue          681 drivers/net/ethernet/sfc/falcon/rx.c 	struct ef4_nic *efx = rx_queue->efx;
rx_queue          688 drivers/net/ethernet/sfc/falcon/rx.c 	rx_queue->ptr_mask = entries - 1;
rx_queue          692 drivers/net/ethernet/sfc/falcon/rx.c 		  ef4_rx_queue_index(rx_queue), efx->rxq_entries,
rx_queue          693 drivers/net/ethernet/sfc/falcon/rx.c 		  rx_queue->ptr_mask);
rx_queue          696 drivers/net/ethernet/sfc/falcon/rx.c 	rx_queue->buffer = kcalloc(entries, sizeof(*rx_queue->buffer),
rx_queue          698 drivers/net/ethernet/sfc/falcon/rx.c 	if (!rx_queue->buffer)
rx_queue          701 drivers/net/ethernet/sfc/falcon/rx.c 	rc = ef4_nic_probe_rx(rx_queue);
rx_queue          703 drivers/net/ethernet/sfc/falcon/rx.c 		kfree(rx_queue->buffer);
rx_queue          704 drivers/net/ethernet/sfc/falcon/rx.c 		rx_queue->buffer = NULL;
rx_queue          711 drivers/net/ethernet/sfc/falcon/rx.c 				     struct ef4_rx_queue *rx_queue)
rx_queue          727 drivers/net/ethernet/sfc/falcon/rx.c 	rx_queue->page_ring = kcalloc(page_ring_size,
rx_queue          728 drivers/net/ethernet/sfc/falcon/rx.c 				      sizeof(*rx_queue->page_ring), GFP_KERNEL);
rx_queue          729 drivers/net/ethernet/sfc/falcon/rx.c 	rx_queue->page_ptr_mask = page_ring_size - 1;
rx_queue          732 drivers/net/ethernet/sfc/falcon/rx.c void ef4_init_rx_queue(struct ef4_rx_queue *rx_queue)
rx_queue          734 drivers/net/ethernet/sfc/falcon/rx.c 	struct ef4_nic *efx = rx_queue->efx;
rx_queue          737 drivers/net/ethernet/sfc/falcon/rx.c 	netif_dbg(rx_queue->efx, drv, rx_queue->efx->net_dev,
rx_queue          738 drivers/net/ethernet/sfc/falcon/rx.c 		  "initialising RX queue %d\n", ef4_rx_queue_index(rx_queue));
rx_queue          741 drivers/net/ethernet/sfc/falcon/rx.c 	rx_queue->added_count = 0;
rx_queue          742 drivers/net/ethernet/sfc/falcon/rx.c 	rx_queue->notified_count = 0;
rx_queue          743 drivers/net/ethernet/sfc/falcon/rx.c 	rx_queue->removed_count = 0;
rx_queue          744 drivers/net/ethernet/sfc/falcon/rx.c 	rx_queue->min_fill = -1U;
rx_queue          745 drivers/net/ethernet/sfc/falcon/rx.c 	ef4_init_rx_recycle_ring(efx, rx_queue);
rx_queue          747 drivers/net/ethernet/sfc/falcon/rx.c 	rx_queue->page_remove = 0;
rx_queue          748 drivers/net/ethernet/sfc/falcon/rx.c 	rx_queue->page_add = rx_queue->page_ptr_mask + 1;
rx_queue          749 drivers/net/ethernet/sfc/falcon/rx.c 	rx_queue->page_recycle_count = 0;
rx_queue          750 drivers/net/ethernet/sfc/falcon/rx.c 	rx_queue->page_recycle_failed = 0;
rx_queue          751 drivers/net/ethernet/sfc/falcon/rx.c 	rx_queue->page_recycle_full = 0;
rx_queue          765 drivers/net/ethernet/sfc/falcon/rx.c 	rx_queue->max_fill = max_fill;
rx_queue          766 drivers/net/ethernet/sfc/falcon/rx.c 	rx_queue->fast_fill_trigger = trigger;
rx_queue          767 drivers/net/ethernet/sfc/falcon/rx.c 	rx_queue->refill_enabled = true;
rx_queue          770 drivers/net/ethernet/sfc/falcon/rx.c 	ef4_nic_init_rx(rx_queue);
rx_queue          773 drivers/net/ethernet/sfc/falcon/rx.c void ef4_fini_rx_queue(struct ef4_rx_queue *rx_queue)
rx_queue          776 drivers/net/ethernet/sfc/falcon/rx.c 	struct ef4_nic *efx = rx_queue->efx;
rx_queue          779 drivers/net/ethernet/sfc/falcon/rx.c 	netif_dbg(rx_queue->efx, drv, rx_queue->efx->net_dev,
rx_queue          780 drivers/net/ethernet/sfc/falcon/rx.c 		  "shutting down RX queue %d\n", ef4_rx_queue_index(rx_queue));
rx_queue          782 drivers/net/ethernet/sfc/falcon/rx.c 	del_timer_sync(&rx_queue->slow_fill);
rx_queue          785 drivers/net/ethernet/sfc/falcon/rx.c 	if (rx_queue->buffer) {
rx_queue          786 drivers/net/ethernet/sfc/falcon/rx.c 		for (i = rx_queue->removed_count; i < rx_queue->added_count;
rx_queue          788 drivers/net/ethernet/sfc/falcon/rx.c 			unsigned index = i & rx_queue->ptr_mask;
rx_queue          789 drivers/net/ethernet/sfc/falcon/rx.c 			rx_buf = ef4_rx_buffer(rx_queue, index);
rx_queue          790 drivers/net/ethernet/sfc/falcon/rx.c 			ef4_fini_rx_buffer(rx_queue, rx_buf);
rx_queue          795 drivers/net/ethernet/sfc/falcon/rx.c 	for (i = 0; i <= rx_queue->page_ptr_mask; i++) {
rx_queue          796 drivers/net/ethernet/sfc/falcon/rx.c 		struct page *page = rx_queue->page_ring[i];
rx_queue          808 drivers/net/ethernet/sfc/falcon/rx.c 	kfree(rx_queue->page_ring);
rx_queue          809 drivers/net/ethernet/sfc/falcon/rx.c 	rx_queue->page_ring = NULL;
rx_queue          812 drivers/net/ethernet/sfc/falcon/rx.c void ef4_remove_rx_queue(struct ef4_rx_queue *rx_queue)
rx_queue          814 drivers/net/ethernet/sfc/falcon/rx.c 	netif_dbg(rx_queue->efx, drv, rx_queue->efx->net_dev,
rx_queue          815 drivers/net/ethernet/sfc/falcon/rx.c 		  "destroying RX queue %d\n", ef4_rx_queue_index(rx_queue));
rx_queue          817 drivers/net/ethernet/sfc/falcon/rx.c 	ef4_nic_remove_rx(rx_queue);
rx_queue          819 drivers/net/ethernet/sfc/falcon/rx.c 	kfree(rx_queue->buffer);
rx_queue          820 drivers/net/ethernet/sfc/falcon/rx.c 	rx_queue->buffer = NULL;
rx_queue          459 drivers/net/ethernet/sfc/farch.c efx_farch_build_rx_desc(struct efx_rx_queue *rx_queue, unsigned index)
rx_queue          464 drivers/net/ethernet/sfc/farch.c 	rxd = efx_rx_desc(rx_queue, index);
rx_queue          465 drivers/net/ethernet/sfc/farch.c 	rx_buf = efx_rx_buffer(rx_queue, index);
rx_queue          469 drivers/net/ethernet/sfc/farch.c 			     rx_queue->efx->type->rx_buffer_padding,
rx_queue          477 drivers/net/ethernet/sfc/farch.c void efx_farch_rx_write(struct efx_rx_queue *rx_queue)
rx_queue          479 drivers/net/ethernet/sfc/farch.c 	struct efx_nic *efx = rx_queue->efx;
rx_queue          483 drivers/net/ethernet/sfc/farch.c 	while (rx_queue->notified_count != rx_queue->added_count) {
rx_queue          485 drivers/net/ethernet/sfc/farch.c 			rx_queue,
rx_queue          486 drivers/net/ethernet/sfc/farch.c 			rx_queue->notified_count & rx_queue->ptr_mask);
rx_queue          487 drivers/net/ethernet/sfc/farch.c 		++rx_queue->notified_count;
rx_queue          491 drivers/net/ethernet/sfc/farch.c 	write_ptr = rx_queue->added_count & rx_queue->ptr_mask;
rx_queue          494 drivers/net/ethernet/sfc/farch.c 			efx_rx_queue_index(rx_queue));
rx_queue          497 drivers/net/ethernet/sfc/farch.c int efx_farch_rx_probe(struct efx_rx_queue *rx_queue)
rx_queue          499 drivers/net/ethernet/sfc/farch.c 	struct efx_nic *efx = rx_queue->efx;
rx_queue          502 drivers/net/ethernet/sfc/farch.c 	entries = rx_queue->ptr_mask + 1;
rx_queue          503 drivers/net/ethernet/sfc/farch.c 	return efx_alloc_special_buffer(efx, &rx_queue->rxd,
rx_queue          507 drivers/net/ethernet/sfc/farch.c void efx_farch_rx_init(struct efx_rx_queue *rx_queue)
rx_queue          510 drivers/net/ethernet/sfc/farch.c 	struct efx_nic *efx = rx_queue->efx;
rx_queue          518 drivers/net/ethernet/sfc/farch.c 		  efx_rx_queue_index(rx_queue), rx_queue->rxd.index,
rx_queue          519 drivers/net/ethernet/sfc/farch.c 		  rx_queue->rxd.index + rx_queue->rxd.entries - 1);
rx_queue          521 drivers/net/ethernet/sfc/farch.c 	rx_queue->scatter_n = 0;
rx_queue          524 drivers/net/ethernet/sfc/farch.c 	efx_init_special_buffer(efx, &rx_queue->rxd);
rx_queue          530 drivers/net/ethernet/sfc/farch.c 			      FRF_AZ_RX_DESCQ_BUF_BASE_ID, rx_queue->rxd.index,
rx_queue          532 drivers/net/ethernet/sfc/farch.c 			      efx_rx_queue_channel(rx_queue)->channel,
rx_queue          535 drivers/net/ethernet/sfc/farch.c 			      efx_rx_queue_index(rx_queue),
rx_queue          537 drivers/net/ethernet/sfc/farch.c 			      __ffs(rx_queue->rxd.entries),
rx_queue          542 drivers/net/ethernet/sfc/farch.c 			 efx_rx_queue_index(rx_queue));
rx_queue          545 drivers/net/ethernet/sfc/farch.c static void efx_farch_flush_rx_queue(struct efx_rx_queue *rx_queue)
rx_queue          547 drivers/net/ethernet/sfc/farch.c 	struct efx_nic *efx = rx_queue->efx;
rx_queue          553 drivers/net/ethernet/sfc/farch.c 			     efx_rx_queue_index(rx_queue));
rx_queue          557 drivers/net/ethernet/sfc/farch.c void efx_farch_rx_fini(struct efx_rx_queue *rx_queue)
rx_queue          560 drivers/net/ethernet/sfc/farch.c 	struct efx_nic *efx = rx_queue->efx;
rx_queue          565 drivers/net/ethernet/sfc/farch.c 			 efx_rx_queue_index(rx_queue));
rx_queue          568 drivers/net/ethernet/sfc/farch.c 	efx_fini_special_buffer(efx, &rx_queue->rxd);
rx_queue          572 drivers/net/ethernet/sfc/farch.c void efx_farch_rx_remove(struct efx_rx_queue *rx_queue)
rx_queue          574 drivers/net/ethernet/sfc/farch.c 	efx_free_special_buffer(rx_queue->efx, &rx_queue->rxd);
rx_queue          644 drivers/net/ethernet/sfc/farch.c 	struct efx_rx_queue *rx_queue;
rx_queue          652 drivers/net/ethernet/sfc/farch.c 		efx_for_each_channel_rx_queue(rx_queue, channel) {
rx_queue          653 drivers/net/ethernet/sfc/farch.c 			rx_queue->flush_pending = true;
rx_queue          674 drivers/net/ethernet/sfc/farch.c 			efx_for_each_channel_rx_queue(rx_queue, channel) {
rx_queue          679 drivers/net/ethernet/sfc/farch.c 				if (rx_queue->flush_pending) {
rx_queue          680 drivers/net/ethernet/sfc/farch.c 					rx_queue->flush_pending = false;
rx_queue          683 drivers/net/ethernet/sfc/farch.c 					efx_farch_flush_rx_queue(rx_queue);
rx_queue          714 drivers/net/ethernet/sfc/farch.c 	struct efx_rx_queue *rx_queue;
rx_queue          727 drivers/net/ethernet/sfc/farch.c 			efx_for_each_channel_rx_queue(rx_queue, channel)
rx_queue          728 drivers/net/ethernet/sfc/farch.c 				efx_farch_rx_fini(rx_queue);
rx_queue          856 drivers/net/ethernet/sfc/farch.c static u16 efx_farch_handle_rx_not_ok(struct efx_rx_queue *rx_queue,
rx_queue          859 drivers/net/ethernet/sfc/farch.c 	struct efx_channel *channel = efx_rx_queue_channel(rx_queue);
rx_queue          860 drivers/net/ethernet/sfc/farch.c 	struct efx_nic *efx = rx_queue->efx;
rx_queue          909 drivers/net/ethernet/sfc/farch.c 			  efx_rx_queue_index(rx_queue), EFX_QWORD_VAL(*event),
rx_queue          937 drivers/net/ethernet/sfc/farch.c efx_farch_handle_rx_bad_index(struct efx_rx_queue *rx_queue, unsigned index)
rx_queue          939 drivers/net/ethernet/sfc/farch.c 	struct efx_channel *channel = efx_rx_queue_channel(rx_queue);
rx_queue          940 drivers/net/ethernet/sfc/farch.c 	struct efx_nic *efx = rx_queue->efx;
rx_queue          943 drivers/net/ethernet/sfc/farch.c 	if (rx_queue->scatter_n &&
rx_queue          944 drivers/net/ethernet/sfc/farch.c 	    index == ((rx_queue->removed_count + rx_queue->scatter_n - 1) &
rx_queue          945 drivers/net/ethernet/sfc/farch.c 		      rx_queue->ptr_mask)) {
rx_queue          950 drivers/net/ethernet/sfc/farch.c 	expected = rx_queue->removed_count & rx_queue->ptr_mask;
rx_queue          951 drivers/net/ethernet/sfc/farch.c 	dropped = (index - expected) & rx_queue->ptr_mask;
rx_queue          975 drivers/net/ethernet/sfc/farch.c 	struct efx_rx_queue *rx_queue;
rx_queue          986 drivers/net/ethernet/sfc/farch.c 	rx_queue = efx_channel_get_rx_queue(channel);
rx_queue          989 drivers/net/ethernet/sfc/farch.c 	expected_ptr = ((rx_queue->removed_count + rx_queue->scatter_n) &
rx_queue          990 drivers/net/ethernet/sfc/farch.c 			rx_queue->ptr_mask);
rx_queue          994 drivers/net/ethernet/sfc/farch.c 	    unlikely(rx_ev_sop != (rx_queue->scatter_n == 0))) {
rx_queue          996 drivers/net/ethernet/sfc/farch.c 		    !efx_farch_handle_rx_bad_index(rx_queue, rx_ev_desc_ptr))
rx_queue         1000 drivers/net/ethernet/sfc/farch.c 		if (rx_queue->scatter_n) {
rx_queue         1002 drivers/net/ethernet/sfc/farch.c 				rx_queue,
rx_queue         1003 drivers/net/ethernet/sfc/farch.c 				rx_queue->removed_count & rx_queue->ptr_mask,
rx_queue         1004 drivers/net/ethernet/sfc/farch.c 				rx_queue->scatter_n, 0, EFX_RX_PKT_DISCARD);
rx_queue         1005 drivers/net/ethernet/sfc/farch.c 			rx_queue->removed_count += rx_queue->scatter_n;
rx_queue         1006 drivers/net/ethernet/sfc/farch.c 			rx_queue->scatter_n = 0;
rx_queue         1016 drivers/net/ethernet/sfc/farch.c 				rx_queue,
rx_queue         1017 drivers/net/ethernet/sfc/farch.c 				rx_queue->removed_count & rx_queue->ptr_mask,
rx_queue         1019 drivers/net/ethernet/sfc/farch.c 			++rx_queue->removed_count;
rx_queue         1024 drivers/net/ethernet/sfc/farch.c 	++rx_queue->scatter_n;
rx_queue         1049 drivers/net/ethernet/sfc/farch.c 		flags = efx_farch_handle_rx_not_ok(rx_queue, event);
rx_queue         1067 drivers/net/ethernet/sfc/farch.c 	efx_rx_packet(rx_queue,
rx_queue         1068 drivers/net/ethernet/sfc/farch.c 		      rx_queue->removed_count & rx_queue->ptr_mask,
rx_queue         1069 drivers/net/ethernet/sfc/farch.c 		      rx_queue->scatter_n, rx_ev_byte_cnt, flags);
rx_queue         1070 drivers/net/ethernet/sfc/farch.c 	rx_queue->removed_count += rx_queue->scatter_n;
rx_queue         1071 drivers/net/ethernet/sfc/farch.c 	rx_queue->scatter_n = 0;
rx_queue         1103 drivers/net/ethernet/sfc/farch.c 	struct efx_rx_queue *rx_queue;
rx_queue         1114 drivers/net/ethernet/sfc/farch.c 	rx_queue = efx_channel_get_rx_queue(channel);
rx_queue         1119 drivers/net/ethernet/sfc/farch.c 		rx_queue->flush_pending = true;
rx_queue         1122 drivers/net/ethernet/sfc/farch.c 		efx_farch_magic_event(efx_rx_queue_channel(rx_queue),
rx_queue         1123 drivers/net/ethernet/sfc/farch.c 				      EFX_CHANNEL_MAGIC_RX_DRAIN(rx_queue));
rx_queue         1145 drivers/net/ethernet/sfc/farch.c 	struct efx_rx_queue *rx_queue =
rx_queue         1155 drivers/net/ethernet/sfc/farch.c 	} else if (rx_queue && magic == EFX_CHANNEL_MAGIC_FILL(rx_queue)) {
rx_queue         1159 drivers/net/ethernet/sfc/farch.c 		efx_fast_push_rx_descriptors(rx_queue, true);
rx_queue         1160 drivers/net/ethernet/sfc/farch.c 	} else if (rx_queue && magic == EFX_CHANNEL_MAGIC_RX_DRAIN(rx_queue)) {
rx_queue         1403 drivers/net/ethernet/sfc/farch.c void efx_farch_rx_defer_refill(struct efx_rx_queue *rx_queue)
rx_queue         1405 drivers/net/ethernet/sfc/farch.c 	efx_farch_magic_event(efx_rx_queue_channel(rx_queue),
rx_queue         1406 drivers/net/ethernet/sfc/farch.c 			      EFX_CHANNEL_MAGIC_FILL(rx_queue));
rx_queue         1970 drivers/net/ethernet/sfc/mcdi.c 	struct efx_rx_queue *rx_queue;
rx_queue         1980 drivers/net/ethernet/sfc/mcdi.c 		efx_for_each_channel_rx_queue(rx_queue, channel) {
rx_queue         1981 drivers/net/ethernet/sfc/mcdi.c 			if (rx_queue->flush_pending) {
rx_queue         1982 drivers/net/ethernet/sfc/mcdi.c 				rx_queue->flush_pending = false;
rx_queue         1986 drivers/net/ethernet/sfc/mcdi.c 					count, efx_rx_queue_index(rx_queue));
rx_queue          503 drivers/net/ethernet/sfc/net_driver.h 	struct efx_rx_queue rx_queue;
rx_queue         1337 drivers/net/ethernet/sfc/net_driver.h 	int (*rx_probe)(struct efx_rx_queue *rx_queue);
rx_queue         1338 drivers/net/ethernet/sfc/net_driver.h 	void (*rx_init)(struct efx_rx_queue *rx_queue);
rx_queue         1339 drivers/net/ethernet/sfc/net_driver.h 	void (*rx_remove)(struct efx_rx_queue *rx_queue);
rx_queue         1340 drivers/net/ethernet/sfc/net_driver.h 	void (*rx_write)(struct efx_rx_queue *rx_queue);
rx_queue         1341 drivers/net/ethernet/sfc/net_driver.h 	void (*rx_defer_refill)(struct efx_rx_queue *rx_queue);
rx_queue         1517 drivers/net/ethernet/sfc/net_driver.h 	return channel->rx_queue.core_index >= 0;
rx_queue         1524 drivers/net/ethernet/sfc/net_driver.h 	return &channel->rx_queue;
rx_queue         1532 drivers/net/ethernet/sfc/net_driver.h 		for (_rx_queue = &(_channel)->rx_queue;			\
rx_queue         1537 drivers/net/ethernet/sfc/net_driver.h efx_rx_queue_channel(struct efx_rx_queue *rx_queue)
rx_queue         1539 drivers/net/ethernet/sfc/net_driver.h 	return container_of(rx_queue, struct efx_channel, rx_queue);
rx_queue         1542 drivers/net/ethernet/sfc/net_driver.h static inline int efx_rx_queue_index(struct efx_rx_queue *rx_queue)
rx_queue         1544 drivers/net/ethernet/sfc/net_driver.h 	return efx_rx_queue_channel(rx_queue)->channel;
rx_queue         1550 drivers/net/ethernet/sfc/net_driver.h static inline struct efx_rx_buffer *efx_rx_buffer(struct efx_rx_queue *rx_queue,
rx_queue         1553 drivers/net/ethernet/sfc/net_driver.h 	return &rx_queue->buffer[index];
rx_queue          133 drivers/net/ethernet/sfc/nic.h efx_rx_desc(struct efx_rx_queue *rx_queue, unsigned int index)
rx_queue          135 drivers/net/ethernet/sfc/nic.h 	return ((efx_qword_t *) (rx_queue->rxd.buf.addr)) + index;
rx_queue          510 drivers/net/ethernet/sfc/nic.h static inline int efx_nic_probe_rx(struct efx_rx_queue *rx_queue)
rx_queue          512 drivers/net/ethernet/sfc/nic.h 	return rx_queue->efx->type->rx_probe(rx_queue);
rx_queue          514 drivers/net/ethernet/sfc/nic.h static inline void efx_nic_init_rx(struct efx_rx_queue *rx_queue)
rx_queue          516 drivers/net/ethernet/sfc/nic.h 	rx_queue->efx->type->rx_init(rx_queue);
rx_queue          518 drivers/net/ethernet/sfc/nic.h static inline void efx_nic_remove_rx(struct efx_rx_queue *rx_queue)
rx_queue          520 drivers/net/ethernet/sfc/nic.h 	rx_queue->efx->type->rx_remove(rx_queue);
rx_queue          522 drivers/net/ethernet/sfc/nic.h static inline void efx_nic_notify_rx_desc(struct efx_rx_queue *rx_queue)
rx_queue          524 drivers/net/ethernet/sfc/nic.h 	rx_queue->efx->type->rx_write(rx_queue);
rx_queue          526 drivers/net/ethernet/sfc/nic.h static inline void efx_nic_generate_fill_event(struct efx_rx_queue *rx_queue)
rx_queue          528 drivers/net/ethernet/sfc/nic.h 	rx_queue->efx->type->rx_defer_refill(rx_queue);
rx_queue          567 drivers/net/ethernet/sfc/nic.h int efx_farch_rx_probe(struct efx_rx_queue *rx_queue);
rx_queue          568 drivers/net/ethernet/sfc/nic.h void efx_farch_rx_init(struct efx_rx_queue *rx_queue);
rx_queue          569 drivers/net/ethernet/sfc/nic.h void efx_farch_rx_fini(struct efx_rx_queue *rx_queue);
rx_queue          570 drivers/net/ethernet/sfc/nic.h void efx_farch_rx_remove(struct efx_rx_queue *rx_queue);
rx_queue          571 drivers/net/ethernet/sfc/nic.h void efx_farch_rx_write(struct efx_rx_queue *rx_queue);
rx_queue          572 drivers/net/ethernet/sfc/nic.h void efx_farch_rx_defer_refill(struct efx_rx_queue *rx_queue);
rx_queue         1544 drivers/net/ethernet/sfc/ptp.c 	channel->rx_queue.core_index = 0;
rx_queue           76 drivers/net/ethernet/sfc/rx.c efx_rx_buf_next(struct efx_rx_queue *rx_queue, struct efx_rx_buffer *rx_buf)
rx_queue           78 drivers/net/ethernet/sfc/rx.c 	if (unlikely(rx_buf == efx_rx_buffer(rx_queue, rx_queue->ptr_mask)))
rx_queue           79 drivers/net/ethernet/sfc/rx.c 		return efx_rx_buffer(rx_queue, 0);
rx_queue          106 drivers/net/ethernet/sfc/rx.c static struct page *efx_reuse_page(struct efx_rx_queue *rx_queue)
rx_queue          108 drivers/net/ethernet/sfc/rx.c 	struct efx_nic *efx = rx_queue->efx;
rx_queue          113 drivers/net/ethernet/sfc/rx.c 	index = rx_queue->page_remove & rx_queue->page_ptr_mask;
rx_queue          114 drivers/net/ethernet/sfc/rx.c 	page = rx_queue->page_ring[index];
rx_queue          118 drivers/net/ethernet/sfc/rx.c 	rx_queue->page_ring[index] = NULL;
rx_queue          120 drivers/net/ethernet/sfc/rx.c 	if (rx_queue->page_remove != rx_queue->page_add)
rx_queue          121 drivers/net/ethernet/sfc/rx.c 		++rx_queue->page_remove;
rx_queue          125 drivers/net/ethernet/sfc/rx.c 		++rx_queue->page_recycle_count;
rx_queue          133 drivers/net/ethernet/sfc/rx.c 		++rx_queue->page_recycle_failed;
rx_queue          149 drivers/net/ethernet/sfc/rx.c static int efx_init_rx_buffers(struct efx_rx_queue *rx_queue, bool atomic)
rx_queue          151 drivers/net/ethernet/sfc/rx.c 	struct efx_nic *efx = rx_queue->efx;
rx_queue          161 drivers/net/ethernet/sfc/rx.c 		page = efx_reuse_page(rx_queue);
rx_queue          188 drivers/net/ethernet/sfc/rx.c 			index = rx_queue->added_count & rx_queue->ptr_mask;
rx_queue          189 drivers/net/ethernet/sfc/rx.c 			rx_buf = efx_rx_buffer(rx_queue, index);
rx_queue          195 drivers/net/ethernet/sfc/rx.c 			++rx_queue->added_count;
rx_queue          224 drivers/net/ethernet/sfc/rx.c static void efx_free_rx_buffers(struct efx_rx_queue *rx_queue,
rx_queue          233 drivers/net/ethernet/sfc/rx.c 		rx_buf = efx_rx_buf_next(rx_queue, rx_buf);
rx_queue          245 drivers/net/ethernet/sfc/rx.c 	struct efx_rx_queue *rx_queue = efx_channel_get_rx_queue(channel);
rx_queue          246 drivers/net/ethernet/sfc/rx.c 	struct efx_nic *efx = rx_queue->efx;
rx_queue          253 drivers/net/ethernet/sfc/rx.c 	index = rx_queue->page_add & rx_queue->page_ptr_mask;
rx_queue          254 drivers/net/ethernet/sfc/rx.c 	if (rx_queue->page_ring[index] == NULL) {
rx_queue          255 drivers/net/ethernet/sfc/rx.c 		unsigned read_index = rx_queue->page_remove &
rx_queue          256 drivers/net/ethernet/sfc/rx.c 			rx_queue->page_ptr_mask;
rx_queue          263 drivers/net/ethernet/sfc/rx.c 			++rx_queue->page_remove;
rx_queue          264 drivers/net/ethernet/sfc/rx.c 		rx_queue->page_ring[index] = page;
rx_queue          265 drivers/net/ethernet/sfc/rx.c 		++rx_queue->page_add;
rx_queue          268 drivers/net/ethernet/sfc/rx.c 	++rx_queue->page_recycle_full;
rx_queue          273 drivers/net/ethernet/sfc/rx.c static void efx_fini_rx_buffer(struct efx_rx_queue *rx_queue,
rx_queue          282 drivers/net/ethernet/sfc/rx.c 		efx_unmap_rx_buffer(rx_queue->efx, rx_buf);
rx_queue          283 drivers/net/ethernet/sfc/rx.c 		efx_free_rx_buffers(rx_queue, rx_buf, 1);
rx_queue          293 drivers/net/ethernet/sfc/rx.c 	struct efx_rx_queue *rx_queue = efx_channel_get_rx_queue(channel);
rx_queue          297 drivers/net/ethernet/sfc/rx.c 		rx_buf = efx_rx_buf_next(rx_queue, rx_buf);
rx_queue          305 drivers/net/ethernet/sfc/rx.c 	struct efx_rx_queue *rx_queue = efx_channel_get_rx_queue(channel);
rx_queue          309 drivers/net/ethernet/sfc/rx.c 	efx_free_rx_buffers(rx_queue, rx_buf, n_frags);
rx_queue          324 drivers/net/ethernet/sfc/rx.c void efx_fast_push_rx_descriptors(struct efx_rx_queue *rx_queue, bool atomic)
rx_queue          326 drivers/net/ethernet/sfc/rx.c 	struct efx_nic *efx = rx_queue->efx;
rx_queue          330 drivers/net/ethernet/sfc/rx.c 	if (!rx_queue->refill_enabled)
rx_queue          334 drivers/net/ethernet/sfc/rx.c 	fill_level = (rx_queue->added_count - rx_queue->removed_count);
rx_queue          335 drivers/net/ethernet/sfc/rx.c 	EFX_WARN_ON_ONCE_PARANOID(fill_level > rx_queue->efx->rxq_entries);
rx_queue          336 drivers/net/ethernet/sfc/rx.c 	if (fill_level >= rx_queue->fast_fill_trigger)
rx_queue          340 drivers/net/ethernet/sfc/rx.c 	if (unlikely(fill_level < rx_queue->min_fill)) {
rx_queue          342 drivers/net/ethernet/sfc/rx.c 			rx_queue->min_fill = fill_level;
rx_queue          346 drivers/net/ethernet/sfc/rx.c 	space = rx_queue->max_fill - fill_level;
rx_queue          349 drivers/net/ethernet/sfc/rx.c 	netif_vdbg(rx_queue->efx, rx_status, rx_queue->efx->net_dev,
rx_queue          352 drivers/net/ethernet/sfc/rx.c 		   efx_rx_queue_index(rx_queue), fill_level,
rx_queue          353 drivers/net/ethernet/sfc/rx.c 		   rx_queue->max_fill);
rx_queue          357 drivers/net/ethernet/sfc/rx.c 		rc = efx_init_rx_buffers(rx_queue, atomic);
rx_queue          360 drivers/net/ethernet/sfc/rx.c 			efx_schedule_slow_fill(rx_queue);
rx_queue          365 drivers/net/ethernet/sfc/rx.c 	netif_vdbg(rx_queue->efx, rx_status, rx_queue->efx->net_dev,
rx_queue          367 drivers/net/ethernet/sfc/rx.c 		   "to level %d\n", efx_rx_queue_index(rx_queue),
rx_queue          368 drivers/net/ethernet/sfc/rx.c 		   rx_queue->added_count - rx_queue->removed_count);
rx_queue          371 drivers/net/ethernet/sfc/rx.c 	if (rx_queue->notified_count != rx_queue->added_count)
rx_queue          372 drivers/net/ethernet/sfc/rx.c 		efx_nic_notify_rx_desc(rx_queue);
rx_queue          377 drivers/net/ethernet/sfc/rx.c 	struct efx_rx_queue *rx_queue = from_timer(rx_queue, t, slow_fill);
rx_queue          380 drivers/net/ethernet/sfc/rx.c 	efx_nic_generate_fill_event(rx_queue);
rx_queue          381 drivers/net/ethernet/sfc/rx.c 	++rx_queue->slow_fill_count;
rx_queue          384 drivers/net/ethernet/sfc/rx.c static void efx_rx_packet__check_len(struct efx_rx_queue *rx_queue,
rx_queue          388 drivers/net/ethernet/sfc/rx.c 	struct efx_nic *efx = rx_queue->efx;
rx_queue          402 drivers/net/ethernet/sfc/rx.c 			  efx_rx_queue_index(rx_queue), len, max_len);
rx_queue          404 drivers/net/ethernet/sfc/rx.c 	efx_rx_queue_channel(rx_queue)->n_rx_overlength++;
rx_queue          420 drivers/net/ethernet/sfc/rx.c 		struct efx_rx_queue *rx_queue;
rx_queue          422 drivers/net/ethernet/sfc/rx.c 		rx_queue = efx_channel_get_rx_queue(channel);
rx_queue          423 drivers/net/ethernet/sfc/rx.c 		efx_free_rx_buffers(rx_queue, rx_buf, n_frags);
rx_queue          443 drivers/net/ethernet/sfc/rx.c 		rx_buf = efx_rx_buf_next(&channel->rx_queue, rx_buf);
rx_queue          449 drivers/net/ethernet/sfc/rx.c 	skb_record_rx_queue(skb, channel->rx_queue.core_index);
rx_queue          494 drivers/net/ethernet/sfc/rx.c 			rx_buf = efx_rx_buf_next(&channel->rx_queue, rx_buf);
rx_queue          512 drivers/net/ethernet/sfc/rx.c void efx_rx_packet(struct efx_rx_queue *rx_queue, unsigned int index,
rx_queue          515 drivers/net/ethernet/sfc/rx.c 	struct efx_nic *efx = rx_queue->efx;
rx_queue          516 drivers/net/ethernet/sfc/rx.c 	struct efx_channel *channel = efx_rx_queue_channel(rx_queue);
rx_queue          519 drivers/net/ethernet/sfc/rx.c 	rx_queue->rx_packets++;
rx_queue          521 drivers/net/ethernet/sfc/rx.c 	rx_buf = efx_rx_buffer(rx_queue, index);
rx_queue          527 drivers/net/ethernet/sfc/rx.c 			efx_rx_packet__check_len(rx_queue, rx_buf, len);
rx_queue          541 drivers/net/ethernet/sfc/rx.c 		   efx_rx_queue_index(rx_queue), index,
rx_queue          542 drivers/net/ethernet/sfc/rx.c 		   (index + n_frags - 1) & rx_queue->ptr_mask, len,
rx_queue          578 drivers/net/ethernet/sfc/rx.c 			rx_buf = efx_rx_buf_next(rx_queue, rx_buf);
rx_queue          588 drivers/net/ethernet/sfc/rx.c 	rx_buf = efx_rx_buffer(rx_queue, index);
rx_queue          608 drivers/net/ethernet/sfc/rx.c 		struct efx_rx_queue *rx_queue;
rx_queue          610 drivers/net/ethernet/sfc/rx.c 		rx_queue = efx_channel_get_rx_queue(channel);
rx_queue          611 drivers/net/ethernet/sfc/rx.c 		efx_free_rx_buffers(rx_queue, rx_buf, n_frags);
rx_queue          614 drivers/net/ethernet/sfc/rx.c 	skb_record_rx_queue(skb, channel->rx_queue.core_index);
rx_queue          643 drivers/net/ethernet/sfc/rx.c 		efx_rx_buffer(&channel->rx_queue, channel->rx_pkt_index);
rx_queue          657 drivers/net/ethernet/sfc/rx.c 		struct efx_rx_queue *rx_queue;
rx_queue          660 drivers/net/ethernet/sfc/rx.c 		rx_queue = efx_channel_get_rx_queue(channel);
rx_queue          661 drivers/net/ethernet/sfc/rx.c 		efx_free_rx_buffers(rx_queue, rx_buf,
rx_queue          677 drivers/net/ethernet/sfc/rx.c int efx_probe_rx_queue(struct efx_rx_queue *rx_queue)
rx_queue          679 drivers/net/ethernet/sfc/rx.c 	struct efx_nic *efx = rx_queue->efx;
rx_queue          686 drivers/net/ethernet/sfc/rx.c 	rx_queue->ptr_mask = entries - 1;
rx_queue          690 drivers/net/ethernet/sfc/rx.c 		  efx_rx_queue_index(rx_queue), efx->rxq_entries,
rx_queue          691 drivers/net/ethernet/sfc/rx.c 		  rx_queue->ptr_mask);
rx_queue          694 drivers/net/ethernet/sfc/rx.c 	rx_queue->buffer = kcalloc(entries, sizeof(*rx_queue->buffer),
rx_queue          696 drivers/net/ethernet/sfc/rx.c 	if (!rx_queue->buffer)
rx_queue          699 drivers/net/ethernet/sfc/rx.c 	rc = efx_nic_probe_rx(rx_queue);
rx_queue          701 drivers/net/ethernet/sfc/rx.c 		kfree(rx_queue->buffer);
rx_queue          702 drivers/net/ethernet/sfc/rx.c 		rx_queue->buffer = NULL;
rx_queue          709 drivers/net/ethernet/sfc/rx.c 				     struct efx_rx_queue *rx_queue)
rx_queue          725 drivers/net/ethernet/sfc/rx.c 	rx_queue->page_ring = kcalloc(page_ring_size,
rx_queue          726 drivers/net/ethernet/sfc/rx.c 				      sizeof(*rx_queue->page_ring), GFP_KERNEL);
rx_queue          727 drivers/net/ethernet/sfc/rx.c 	rx_queue->page_ptr_mask = page_ring_size - 1;
rx_queue          730 drivers/net/ethernet/sfc/rx.c void efx_init_rx_queue(struct efx_rx_queue *rx_queue)
rx_queue          732 drivers/net/ethernet/sfc/rx.c 	struct efx_nic *efx = rx_queue->efx;
rx_queue          735 drivers/net/ethernet/sfc/rx.c 	netif_dbg(rx_queue->efx, drv, rx_queue->efx->net_dev,
rx_queue          736 drivers/net/ethernet/sfc/rx.c 		  "initialising RX queue %d\n", efx_rx_queue_index(rx_queue));
rx_queue          739 drivers/net/ethernet/sfc/rx.c 	rx_queue->added_count = 0;
rx_queue          740 drivers/net/ethernet/sfc/rx.c 	rx_queue->notified_count = 0;
rx_queue          741 drivers/net/ethernet/sfc/rx.c 	rx_queue->removed_count = 0;
rx_queue          742 drivers/net/ethernet/sfc/rx.c 	rx_queue->min_fill = -1U;
rx_queue          743 drivers/net/ethernet/sfc/rx.c 	efx_init_rx_recycle_ring(efx, rx_queue);
rx_queue          745 drivers/net/ethernet/sfc/rx.c 	rx_queue->page_remove = 0;
rx_queue          746 drivers/net/ethernet/sfc/rx.c 	rx_queue->page_add = rx_queue->page_ptr_mask + 1;
rx_queue          747 drivers/net/ethernet/sfc/rx.c 	rx_queue->page_recycle_count = 0;
rx_queue          748 drivers/net/ethernet/sfc/rx.c 	rx_queue->page_recycle_failed = 0;
rx_queue          749 drivers/net/ethernet/sfc/rx.c 	rx_queue->page_recycle_full = 0;
rx_queue          763 drivers/net/ethernet/sfc/rx.c 	rx_queue->max_fill = max_fill;
rx_queue          764 drivers/net/ethernet/sfc/rx.c 	rx_queue->fast_fill_trigger = trigger;
rx_queue          765 drivers/net/ethernet/sfc/rx.c 	rx_queue->refill_enabled = true;
rx_queue          768 drivers/net/ethernet/sfc/rx.c 	efx_nic_init_rx(rx_queue);
rx_queue          771 drivers/net/ethernet/sfc/rx.c void efx_fini_rx_queue(struct efx_rx_queue *rx_queue)
rx_queue          774 drivers/net/ethernet/sfc/rx.c 	struct efx_nic *efx = rx_queue->efx;
rx_queue          777 drivers/net/ethernet/sfc/rx.c 	netif_dbg(rx_queue->efx, drv, rx_queue->efx->net_dev,
rx_queue          778 drivers/net/ethernet/sfc/rx.c 		  "shutting down RX queue %d\n", efx_rx_queue_index(rx_queue));
rx_queue          780 drivers/net/ethernet/sfc/rx.c 	del_timer_sync(&rx_queue->slow_fill);
rx_queue          783 drivers/net/ethernet/sfc/rx.c 	if (rx_queue->buffer) {
rx_queue          784 drivers/net/ethernet/sfc/rx.c 		for (i = rx_queue->removed_count; i < rx_queue->added_count;
rx_queue          786 drivers/net/ethernet/sfc/rx.c 			unsigned index = i & rx_queue->ptr_mask;
rx_queue          787 drivers/net/ethernet/sfc/rx.c 			rx_buf = efx_rx_buffer(rx_queue, index);
rx_queue          788 drivers/net/ethernet/sfc/rx.c 			efx_fini_rx_buffer(rx_queue, rx_buf);
rx_queue          793 drivers/net/ethernet/sfc/rx.c 	for (i = 0; i <= rx_queue->page_ptr_mask; i++) {
rx_queue          794 drivers/net/ethernet/sfc/rx.c 		struct page *page = rx_queue->page_ring[i];
rx_queue          806 drivers/net/ethernet/sfc/rx.c 	kfree(rx_queue->page_ring);
rx_queue          807 drivers/net/ethernet/sfc/rx.c 	rx_queue->page_ring = NULL;
rx_queue          810 drivers/net/ethernet/sfc/rx.c void efx_remove_rx_queue(struct efx_rx_queue *rx_queue)
rx_queue          812 drivers/net/ethernet/sfc/rx.c 	netif_dbg(rx_queue->efx, drv, rx_queue->efx->net_dev,
rx_queue          813 drivers/net/ethernet/sfc/rx.c 		  "destroying RX queue %d\n", efx_rx_queue_index(rx_queue));
rx_queue          815 drivers/net/ethernet/sfc/rx.c 	efx_nic_remove_rx(rx_queue);
rx_queue          817 drivers/net/ethernet/sfc/rx.c 	kfree(rx_queue->buffer);
rx_queue          818 drivers/net/ethernet/sfc/rx.c 	rx_queue->buffer = NULL;
rx_queue          167 drivers/net/ethernet/stmicro/stmmac/stmmac.h 	struct stmmac_rx_queue rx_queue[MTL_MAX_RX_QUEUES];
rx_queue          311 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue];
rx_queue         1065 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue];
rx_queue         1136 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue];
rx_queue         1209 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue];
rx_queue         1243 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue];
rx_queue         1307 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue];
rx_queue         1481 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue];
rx_queue         1549 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue];
rx_queue         2219 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		rx_q = &priv->rx_queue[chan];
rx_queue         3386 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue];
rx_queue         3458 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 	struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue];
rx_queue         3881 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 			struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue];
rx_queue         4075 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue];
rx_queue         4820 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 		struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue];
rx_queue          794 drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c 		tail = priv->rx_queue[i].dma_rx_phy +
rx_queue           92 drivers/net/ethernet/ti/netcp.h 	void			*rx_queue;
rx_queue          624 drivers/net/ethernet/ti/netcp_core.c 		dma = knav_queue_pop(netcp->rx_queue, &dma_sz);
rx_queue          653 drivers/net/ethernet/ti/netcp_core.c 	dma_desc = knav_queue_pop(netcp->rx_queue, &dma_sz);
rx_queue          959 drivers/net/ethernet/ti/netcp_core.c 		knav_queue_enable_notify(netcp->rx_queue);
rx_queue          969 drivers/net/ethernet/ti/netcp_core.c 	knav_queue_disable_notify(netcp->rx_queue);
rx_queue         1561 drivers/net/ethernet/ti/netcp_core.c 	if (!IS_ERR_OR_NULL(netcp->rx_queue)) {
rx_queue         1562 drivers/net/ethernet/ti/netcp_core.c 		knav_queue_close(netcp->rx_queue);
rx_queue         1563 drivers/net/ethernet/ti/netcp_core.c 		netcp->rx_queue = NULL;
rx_queue         1634 drivers/net/ethernet/ti/netcp_core.c 	netcp->rx_queue = knav_queue_open(name, netcp->rx_queue_id, 0);
rx_queue         1635 drivers/net/ethernet/ti/netcp_core.c 	if (IS_ERR(netcp->rx_queue)) {
rx_queue         1636 drivers/net/ethernet/ti/netcp_core.c 		ret = PTR_ERR(netcp->rx_queue);
rx_queue         1639 drivers/net/ethernet/ti/netcp_core.c 	netcp->rx_queue_id = knav_queue_get_id(netcp->rx_queue);
rx_queue         1644 drivers/net/ethernet/ti/netcp_core.c 	ret = knav_queue_device_control(netcp->rx_queue,
rx_queue         1650 drivers/net/ethernet/ti/netcp_core.c 	knav_queue_disable_notify(netcp->rx_queue);
rx_queue         1726 drivers/net/ethernet/ti/netcp_core.c 	knav_queue_enable_notify(netcp->rx_queue);
rx_queue         1756 drivers/net/ethernet/ti/netcp_core.c 	knav_queue_disable_notify(netcp->rx_queue);
rx_queue          117 drivers/net/phy/dp83640.c 	struct sk_buff_head rx_queue;
rx_queue          875 drivers/net/phy/dp83640.c 	spin_lock(&dp83640->rx_queue.lock);
rx_queue          876 drivers/net/phy/dp83640.c 	skb_queue_walk(&dp83640->rx_queue, skb) {
rx_queue          881 drivers/net/phy/dp83640.c 			__skb_unlink(skb, &dp83640->rx_queue);
rx_queue          889 drivers/net/phy/dp83640.c 	spin_unlock(&dp83640->rx_queue.lock);
rx_queue         1162 drivers/net/phy/dp83640.c 	skb_queue_head_init(&dp83640->rx_queue);
rx_queue         1202 drivers/net/phy/dp83640.c 	skb_queue_purge(&dp83640->rx_queue);
rx_queue         1415 drivers/net/phy/dp83640.c 	while ((skb = skb_dequeue(&dp83640->rx_queue))) {
rx_queue         1420 drivers/net/phy/dp83640.c 			skb_queue_head(&dp83640->rx_queue, skb);
rx_queue         1427 drivers/net/phy/dp83640.c 	if (!skb_queue_empty(&dp83640->rx_queue))
rx_queue         1471 drivers/net/phy/dp83640.c 		skb_queue_tail(&dp83640->rx_queue, skb);
rx_queue          742 drivers/net/usb/r8152.c 	struct sk_buff_head tx_queue, rx_queue;
rx_queue         1624 drivers/net/usb/r8152.c 	skb_queue_head_init(&tp->rx_queue);
rx_queue         2044 drivers/net/usb/r8152.c 	struct list_head *cursor, *next, rx_queue;
rx_queue         2048 drivers/net/usb/r8152.c 	if (!skb_queue_empty(&tp->rx_queue)) {
rx_queue         2050 drivers/net/usb/r8152.c 			struct sk_buff *skb = __skb_dequeue(&tp->rx_queue);
rx_queue         2069 drivers/net/usb/r8152.c 	INIT_LIST_HEAD(&rx_queue);
rx_queue         2071 drivers/net/usb/r8152.c 	list_splice_init(&tp->rx_done, &rx_queue);
rx_queue         2074 drivers/net/usb/r8152.c 	list_for_each_safe(cursor, next, &rx_queue) {
rx_queue         2101 drivers/net/usb/r8152.c 			if (unlikely(skb_queue_len(&tp->rx_queue) >= 1000))
rx_queue         2147 drivers/net/usb/r8152.c 				__skb_queue_tail(&tp->rx_queue, skb);
rx_queue         2180 drivers/net/usb/r8152.c 	if (!list_empty(&rx_queue)) {
rx_queue         2182 drivers/net/usb/r8152.c 		list_splice_tail(&rx_queue, &tp->rx_done);
rx_queue         2570 drivers/net/usb/r8152.c 	while (!skb_queue_empty(&tp->rx_queue))
rx_queue         2571 drivers/net/usb/r8152.c 		dev_kfree_skb(__skb_dequeue(&tp->rx_queue));
rx_queue         1625 drivers/net/vmxnet3/vmxnet3_drv.c 		vmxnet3_rq_cleanup(&adapter->rx_queue[i], adapter);
rx_queue         1684 drivers/net/vmxnet3/vmxnet3_drv.c 		struct vmxnet3_rx_queue *rq = &adapter->rx_queue[i];
rx_queue         1756 drivers/net/vmxnet3/vmxnet3_drv.c 		err = vmxnet3_rq_init(&adapter->rx_queue[i], adapter);
rx_queue         1841 drivers/net/vmxnet3/vmxnet3_drv.c 		err = vmxnet3_rq_create(&adapter->rx_queue[i], adapter);
rx_queue         1872 drivers/net/vmxnet3/vmxnet3_drv.c 		rcd_done += vmxnet3_rq_rx_complete(&adapter->rx_queue[i],
rx_queue         1881 drivers/net/vmxnet3/vmxnet3_drv.c 	struct vmxnet3_rx_queue *rx_queue = container_of(napi,
rx_queue         1885 drivers/net/vmxnet3/vmxnet3_drv.c 	rxd_done = vmxnet3_do_poll(rx_queue->adapter, budget);
rx_queue         1889 drivers/net/vmxnet3/vmxnet3_drv.c 		vmxnet3_enable_all_intrs(rx_queue->adapter);
rx_queue         1912 drivers/net/vmxnet3/vmxnet3_drv.c 				&adapter->tx_queue[rq - adapter->rx_queue];
rx_queue         2030 drivers/net/vmxnet3/vmxnet3_drv.c 	napi_schedule(&adapter->rx_queue[0].napi);
rx_queue         2048 drivers/net/vmxnet3/vmxnet3_drv.c 			vmxnet3_msix_rx(0, &adapter->rx_queue[i]);
rx_queue         2109 drivers/net/vmxnet3/vmxnet3_drv.c 				sprintf(adapter->rx_queue[i].name, "%s-rx-%d",
rx_queue         2112 drivers/net/vmxnet3/vmxnet3_drv.c 				sprintf(adapter->rx_queue[i].name, "%s-rxtx-%d",
rx_queue         2116 drivers/net/vmxnet3/vmxnet3_drv.c 					  adapter->rx_queue[i].name,
rx_queue         2117 drivers/net/vmxnet3/vmxnet3_drv.c 					  &(adapter->rx_queue[i]));
rx_queue         2122 drivers/net/vmxnet3/vmxnet3_drv.c 					   adapter->rx_queue[i].name, err);
rx_queue         2126 drivers/net/vmxnet3/vmxnet3_drv.c 			adapter->rx_queue[i].comp_ring.intr_idx = vector++;
rx_queue         2157 drivers/net/vmxnet3/vmxnet3_drv.c 			struct vmxnet3_rx_queue *rq = &adapter->rx_queue[i];
rx_queue         2170 drivers/net/vmxnet3/vmxnet3_drv.c 			adapter->rx_queue[0].comp_ring.intr_idx = 0;
rx_queue         2205 drivers/net/vmxnet3/vmxnet3_drv.c 				 &(adapter->rx_queue[i]));
rx_queue         2390 drivers/net/vmxnet3/vmxnet3_drv.c 		vmxnet3_rq_destroy(&adapter->rx_queue[i], adapter);
rx_queue         2464 drivers/net/vmxnet3/vmxnet3_drv.c 		struct vmxnet3_rx_queue	*rq = &adapter->rx_queue[i];
rx_queue         2571 drivers/net/vmxnet3/vmxnet3_drv.c 		adapter->rx_queue[0].rx_ring[0].size,
rx_queue         2572 drivers/net/vmxnet3/vmxnet3_drv.c 		adapter->rx_queue[0].rx_ring[1].size);
rx_queue         2613 drivers/net/vmxnet3/vmxnet3_drv.c 				adapter->rx_queue[i].rx_ring[0].next2fill);
rx_queue         2616 drivers/net/vmxnet3/vmxnet3_drv.c 				adapter->rx_queue[i].rx_ring[1].next2fill);
rx_queue         2628 drivers/net/vmxnet3/vmxnet3_drv.c 		napi_enable(&adapter->rx_queue[i].napi);
rx_queue         2671 drivers/net/vmxnet3/vmxnet3_drv.c 		napi_disable(&adapter->rx_queue[i].napi);
rx_queue         2799 drivers/net/vmxnet3/vmxnet3_drv.c 	ring0_size = adapter->rx_queue[0].rx_ring[0].size;
rx_queue         2803 drivers/net/vmxnet3/vmxnet3_drv.c 	ring1_size = adapter->rx_queue[0].rx_ring[1].size;
rx_queue         2810 drivers/net/vmxnet3/vmxnet3_drv.c 		struct vmxnet3_rx_queue	*rq = &adapter->rx_queue[i];
rx_queue         2845 drivers/net/vmxnet3/vmxnet3_drv.c 	adapter->rx_queue[0].rx_ring[0].size = rx_ring_size;
rx_queue         2846 drivers/net/vmxnet3/vmxnet3_drv.c 	adapter->rx_queue[0].rx_ring[1].size = rx_ring2_size;
rx_queue         2851 drivers/net/vmxnet3/vmxnet3_drv.c 		struct vmxnet3_rx_queue *rq = &adapter->rx_queue[i];
rx_queue         2978 drivers/net/vmxnet3/vmxnet3_drv.c 		napi_enable(&adapter->rx_queue[i].napi);
rx_queue         3479 drivers/net/vmxnet3/vmxnet3_drv.c 				       &adapter->rx_queue[i].napi,
rx_queue         3483 drivers/net/vmxnet3/vmxnet3_drv.c 		netif_napi_add(adapter->netdev, &adapter->rx_queue[0].napi,
rx_queue         3629 drivers/net/vmxnet3/vmxnet3_drv.c 		napi_disable(&adapter->rx_queue[i].napi);
rx_queue          150 drivers/net/vmxnet3/vmxnet3_ethtool.c 		drvRxStats = &adapter->rx_queue[i].stats;
rx_queue          343 drivers/net/vmxnet3/vmxnet3_ethtool.c 		base = (u8 *)&adapter->rx_queue[j].stats;
rx_queue          420 drivers/net/vmxnet3/vmxnet3_ethtool.c 		struct vmxnet3_rx_queue *rq = &adapter->rx_queue[i];
rx_queue          332 drivers/net/vmxnet3/vmxnet3_int.h 	struct vmxnet3_rx_queue		rx_queue[VMXNET3_DEVICE_MAX_RX_QUEUES];
rx_queue          684 drivers/net/wireless/broadcom/b43/b43.h 	struct b43_pio_rxqueue *rx_queue;
rx_queue         2029 drivers/net/wireless/broadcom/b43/main.c 			b43_pio_rx(dev->pio.rx_queue);
rx_queue          227 drivers/net/wireless/broadcom/b43/pio.c 	destroy_queue_rx(pio, rx_queue);
rx_queue          264 drivers/net/wireless/broadcom/b43/pio.c 	pio->rx_queue = b43_setup_pioqueue_rx(dev, 0);
rx_queue          265 drivers/net/wireless/broadcom/b43/pio.c 	if (!pio->rx_queue)
rx_queue         2518 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	priv->rx_queue.drv[i].host_addr = packet->dma_addr;
rx_queue         2600 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	priv->rx_queue.drv[i].host_addr = packet->dma_addr;
rx_queue         2663 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	struct ipw2100_bd_queue *rxq = &priv->rx_queue;
rx_queue         4562 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	err = bd_queue_allocate(priv, &priv->rx_queue, RX_QUEUE_LENGTH);
rx_queue         4571 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		bd_queue_free(priv, &priv->rx_queue);
rx_queue         4584 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		bd_queue_free(priv, &priv->rx_queue);
rx_queue         4601 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		priv->rx_queue.drv[i].host_addr = packet->dma_addr;
rx_queue         4602 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		priv->rx_queue.drv[i].buf_length = IPW_RX_NIC_BUFFER_LENGTH;
rx_queue         4619 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	bd_queue_free(priv, &priv->rx_queue);
rx_queue         4630 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	priv->rx_queue.oldest = 0;
rx_queue         4631 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	priv->rx_queue.available = priv->rx_queue.entries - 1;
rx_queue         4632 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	priv->rx_queue.next = priv->rx_queue.entries - 1;
rx_queue         4635 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	SET_STAT(&priv->rxq_stat, priv->rx_queue.available);
rx_queue         4637 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	bd_queue_initialize(priv, &priv->rx_queue,
rx_queue         4656 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	bd_queue_free(priv, &priv->rx_queue);
rx_queue          529 drivers/net/wireless/intel/ipw2x00/ipw2100.h 	struct ipw2100_bd_queue rx_queue;
rx_queue         1294 drivers/net/wireless/intel/iwlegacy/debug.c DEBUGFS_READ_FILE_OPS(rx_queue);
rx_queue         1338 drivers/net/wireless/intel/iwlegacy/debug.c 	DEBUGFS_ADD_FILE(rx_queue, dir_debug, 0400);
rx_queue         2914 drivers/net/wireless/intel/iwlwifi/pcie/trans.c DEBUGFS_READ_FILE_OPS(rx_queue);
rx_queue         2930 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 	DEBUGFS_ADD_FILE(rx_queue, dir, 0400);
rx_queue          147 drivers/net/wireless/intersil/p54/p54usb.c 	skb_unlink(skb, &priv->rx_queue);
rx_queue          186 drivers/net/wireless/intersil/p54/p54usb.c 	skb_queue_tail(&priv->rx_queue, skb);
rx_queue          189 drivers/net/wireless/intersil/p54/p54usb.c 		skb_unlink(skb, &priv->rx_queue);
rx_queue          230 drivers/net/wireless/intersil/p54/p54usb.c 	while (skb_queue_len(&priv->rx_queue) < 32) {
rx_queue          249 drivers/net/wireless/intersil/p54/p54usb.c 		skb_queue_tail(&priv->rx_queue, skb);
rx_queue          254 drivers/net/wireless/intersil/p54/p54usb.c 			skb_unlink(skb, &priv->rx_queue);
rx_queue         1007 drivers/net/wireless/intersil/p54/p54usb.c 	skb_queue_head_init(&priv->rx_queue);
rx_queue          151 drivers/net/wireless/intersil/p54/p54usb.h 	struct sk_buff_head rx_queue;
rx_queue          336 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 	spin_lock_irqsave(&priv->rx_queue.lock, f);
rx_queue          337 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 	__skb_unlink(skb, &priv->rx_queue);
rx_queue          338 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 	spin_unlock_irqrestore(&priv->rx_queue.lock, f);
rx_queue          404 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 	skb_queue_tail(&priv->rx_queue, skb);
rx_queue          409 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 		skb_unlink(skb, &priv->rx_queue);
rx_queue          422 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 	while (skb_queue_len(&priv->rx_queue) < 32) {
rx_queue          441 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 		skb_queue_tail(&priv->rx_queue, skb);
rx_queue          445 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 			skb_unlink(skb, &priv->rx_queue);
rx_queue         1462 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 	skb_queue_head_init(&priv->rx_queue);
rx_queue          134 drivers/net/wireless/realtek/rtl818x/rtl8187/rtl8187.h 	struct sk_buff_head rx_queue;
rx_queue          312 drivers/net/wireless/realtek/rtlwifi/usb.c 	skb_queue_head_init(&rtlusb->rx_queue);
rx_queue          515 drivers/net/wireless/realtek/rtlwifi/usb.c 	struct sk_buff_head rx_queue;
rx_queue          518 drivers/net/wireless/realtek/rtlwifi/usb.c 	skb_queue_head_init(&rx_queue);
rx_queue          520 drivers/net/wireless/realtek/rtlwifi/usb.c 		rtlusb->usb_rx_segregate_hdl(hw, skb, &rx_queue);
rx_queue          521 drivers/net/wireless/realtek/rtlwifi/usb.c 	WARN_ON(skb_queue_empty(&rx_queue));
rx_queue          522 drivers/net/wireless/realtek/rtlwifi/usb.c 	while (!skb_queue_empty(&rx_queue)) {
rx_queue          523 drivers/net/wireless/realtek/rtlwifi/usb.c 		_skb = skb_dequeue(&rx_queue);
rx_queue          537 drivers/net/wireless/realtek/rtlwifi/usb.c 	while ((skb = skb_dequeue(&rtlusb->rx_queue))) {
rx_queue          611 drivers/net/wireless/realtek/rtlwifi/usb.c 		qlen = skb_queue_len(&rtlusb->rx_queue);
rx_queue          637 drivers/net/wireless/realtek/rtlwifi/usb.c 		skb_queue_tail(&rtlusb->rx_queue, skb);
rx_queue          686 drivers/net/wireless/realtek/rtlwifi/usb.c 	skb_queue_purge(&rtlusb->rx_queue);
rx_queue          808 drivers/net/wireless/realtek/rtlwifi/usb.c 	skb_queue_purge(&rtlusb->rx_queue);
rx_queue          121 drivers/net/wireless/realtek/rtlwifi/usb.h 	struct sk_buff_head	rx_queue;
rx_queue           88 drivers/net/wireless/st/cw1200/cw1200.h 	struct sk_buff_head		rx_queue;
rx_queue         1564 drivers/net/wireless/st/cw1200/sta.c 	while ((skb = skb_dequeue(&entry->rx_queue)))
rx_queue         1210 drivers/net/wireless/st/cw1200/txrx.c 			skb_queue_tail(&entry->rx_queue, skb);
rx_queue         1351 drivers/net/wireless/st/cw1200/txrx.c 		skb_queue_head_init(&entry->rx_queue);
rx_queue         1459 drivers/net/wireless/st/cw1200/txrx.c 			skb_queue_purge(&priv->link_id_db[i].rx_queue);
rx_queue          187 drivers/net/xen-netback/common.h 	struct sk_buff_head rx_queue;
rx_queue          537 drivers/net/xen-netback/interface.c 	skb_queue_head_init(&queue->rx_queue);
rx_queue           42 drivers/net/xen-netback/rx.c 	skb = skb_peek(&queue->rx_queue);
rx_queue           74 drivers/net/xen-netback/rx.c 	spin_lock_irqsave(&queue->rx_queue.lock, flags);
rx_queue           76 drivers/net/xen-netback/rx.c 	__skb_queue_tail(&queue->rx_queue, skb);
rx_queue           85 drivers/net/xen-netback/rx.c 	spin_unlock_irqrestore(&queue->rx_queue.lock, flags);
rx_queue           92 drivers/net/xen-netback/rx.c 	spin_lock_irq(&queue->rx_queue.lock);
rx_queue           94 drivers/net/xen-netback/rx.c 	skb = __skb_dequeue(&queue->rx_queue);
rx_queue          105 drivers/net/xen-netback/rx.c 	spin_unlock_irq(&queue->rx_queue.lock);
rx_queue          123 drivers/net/xen-netback/rx.c 		skb = skb_peek(&queue->rx_queue);
rx_queue          508 drivers/net/xen-netback/rx.c 	skb = skb_peek(&queue->rx_queue);
rx_queue          603 drivers/net/xen-netback/rx.c 		if (!skb_queue_empty(&queue->rx_queue))
rx_queue           96 drivers/net/xen-netback/xenbus.c 		   skb_queue_len(&queue->rx_queue),
rx_queue          111 drivers/rpmsg/qcom_glink_native.c 	struct list_head rx_queue;
rx_queue          777 drivers/rpmsg/qcom_glink_native.c 	list_add_tail(&dcmd->node, &glink->rx_queue);
rx_queue         1523 drivers/rpmsg/qcom_glink_native.c 		if (list_empty(&glink->rx_queue)) {
rx_queue         1527 drivers/rpmsg/qcom_glink_native.c 		dcmd = list_first_entry(&glink->rx_queue,
rx_queue         1573 drivers/rpmsg/qcom_glink_native.c 	list_for_each_entry_safe(dcmd, tmp, &glink->rx_queue, node)
rx_queue         1600 drivers/rpmsg/qcom_glink_native.c 	INIT_LIST_HEAD(&glink->rx_queue);
rx_queue          880 drivers/staging/rtl8192u/r8192U.h 	struct sk_buff_head rx_queue;
rx_queue          724 drivers/staging/rtl8192u/r8192U_core.c 	while (skb_queue_len(&priv->rx_queue) < MAX_RX_URB) {
rx_queue          741 drivers/staging/rtl8192u/r8192U_core.c 		skb_queue_tail(&priv->rx_queue, skb);
rx_queue          746 drivers/staging/rtl8192u/r8192U_core.c 	while (skb_queue_len(&priv->rx_queue) < MAX_RX_URB + 3) {
rx_queue          763 drivers/staging/rtl8192u/r8192U_core.c 		skb_queue_tail(&priv->rx_queue, skb);
rx_queue          823 drivers/staging/rtl8192u/r8192U_core.c 	while ((skb = __skb_dequeue(&priv->rx_queue))) {
rx_queue          858 drivers/staging/rtl8192u/r8192U_core.c 	skb_unlink(skb, &priv->rx_queue);
rx_queue          884 drivers/staging/rtl8192u/r8192U_core.c 	skb_queue_tail(&priv->rx_queue, skb);
rx_queue         2195 drivers/staging/rtl8192u/r8192U_core.c 	skb_queue_head_init(&priv->rx_queue);
rx_queue          246 drivers/tty/ipwireless/hardware.c 	struct list_head rx_queue;
rx_queue          690 drivers/tty/ipwireless/hardware.c 		list_add_tail(&packet->queue, &hw->rx_queue);
rx_queue          710 drivers/tty/ipwireless/hardware.c 	while (!list_empty(&hw->rx_queue)) {
rx_queue          712 drivers/tty/ipwireless/hardware.c 			list_first_entry(&hw->rx_queue,
rx_queue         1635 drivers/tty/ipwireless/hardware.c 	INIT_LIST_HEAD(&hw->rx_queue);
rx_queue         1750 drivers/tty/ipwireless/hardware.c 	list_for_each_entry_safe(rp, rq, &hw->rx_queue, queue) {
rx_queue         4286 include/linux/skbuff.h static inline void skb_record_rx_queue(struct sk_buff *skb, u16 rx_queue)
rx_queue         4288 include/linux/skbuff.h 	skb->queue_mapping = rx_queue + 1;
rx_queue           46 include/linux/virtio_vsock.h 	struct list_head rx_queue;
rx_queue         1777 include/net/sock.h 		u16 rx_queue = skb_get_rx_queue(skb);
rx_queue         1779 include/net/sock.h 		if (WARN_ON_ONCE(rx_queue == NO_QUEUE_MAPPING))
rx_queue         1782 include/net/sock.h 		sk->sk_rx_queue_mapping = rx_queue;
rx_queue         2436 net/ipv4/tcp_ipv4.c 	int rx_queue;
rx_queue         2457 net/ipv4/tcp_ipv4.c 		rx_queue = sk->sk_ack_backlog;
rx_queue         2462 net/ipv4/tcp_ipv4.c 		rx_queue = max_t(int, READ_ONCE(tp->rcv_nxt) -
rx_queue         2469 net/ipv4/tcp_ipv4.c 		rx_queue,
rx_queue         1868 net/ipv6/tcp_ipv6.c 	int rx_queue;
rx_queue         1894 net/ipv6/tcp_ipv6.c 		rx_queue = sp->sk_ack_backlog;
rx_queue         1899 net/ipv6/tcp_ipv6.c 		rx_queue = max_t(int, READ_ONCE(tp->rcv_nxt) -
rx_queue         1912 net/ipv6/tcp_ipv6.c 		   rx_queue,
rx_queue           97 net/mac80211/ieee80211_i.h 	u8 rx_queue;
rx_queue         2087 net/mac80211/rx.c 			 unsigned int frag, unsigned int seq, int rx_queue,
rx_queue         2103 net/mac80211/rx.c 	entry->rx_queue = rx_queue;
rx_queue         2114 net/mac80211/rx.c 			  int rx_queue, struct ieee80211_hdr *hdr)
rx_queue         2130 net/mac80211/rx.c 		    entry->rx_queue != rx_queue ||
rx_queue          126 net/qrtr/qrtr.c 	struct sk_buff_head rx_queue;
rx_queue          154 net/qrtr/qrtr.c 	skb_queue_purge(&node->rx_queue);
rx_queue          319 net/qrtr/qrtr.c 	skb_queue_tail(&node->rx_queue, skb);
rx_queue          370 net/qrtr/qrtr.c 	while ((skb = skb_dequeue(&node->rx_queue)) != NULL) {
rx_queue          432 net/qrtr/qrtr.c 	skb_queue_head_init(&node->rx_queue);
rx_queue          446 net/rxrpc/ar-internal.h 	struct sk_buff_head	rx_queue;	/* received conn-level packets */
rx_queue          456 net/rxrpc/conn_event.c 	while ((skb = skb_dequeue(&conn->rx_queue))) {
rx_queue          477 net/rxrpc/conn_event.c 	skb_queue_head(&conn->rx_queue, skb);
rx_queue           50 net/rxrpc/conn_object.c 		skb_queue_head_init(&conn->rx_queue);
rx_queue          251 net/rxrpc/conn_object.c 	rxrpc_purge_queue(&conn->rx_queue);
rx_queue          361 net/rxrpc/conn_object.c 	rxrpc_purge_queue(&conn->rx_queue);
rx_queue         1118 net/rxrpc/input.c 	skb_queue_tail(&conn->rx_queue, skb);
rx_queue          282 net/vmw_vsock/virtio_transport_common.c 	while (total < len && !list_empty(&vvs->rx_queue)) {
rx_queue          283 net/vmw_vsock/virtio_transport_common.c 		pkt = list_first_entry(&vvs->rx_queue,
rx_queue          423 net/vmw_vsock/virtio_transport_common.c 	INIT_LIST_HEAD(&vvs->rx_queue);
rx_queue          834 net/vmw_vsock/virtio_transport_common.c 	list_for_each_entry_safe(pkt, tmp, &vvs->rx_queue, list) {
rx_queue          903 net/vmw_vsock/virtio_transport_common.c 	if (pkt->len <= GOOD_COPY_LEN && !list_empty(&vvs->rx_queue)) {
rx_queue          906 net/vmw_vsock/virtio_transport_common.c 		last_pkt = list_last_entry(&vvs->rx_queue,
rx_queue          921 net/vmw_vsock/virtio_transport_common.c 	list_add_tail(&pkt->list, &vvs->rx_queue);