rx_queue 937 arch/um/drivers/vector_kern.c struct vector_queue *qi = vp->rx_queue; rx_queue 1154 arch/um/drivers/vector_kern.c if (vp->rx_queue != NULL) rx_queue 1155 arch/um/drivers/vector_kern.c destroy_queue(vp->rx_queue); rx_queue 1208 arch/um/drivers/vector_kern.c vp->rx_queue = create_queue( rx_queue 1214 arch/um/drivers/vector_kern.c vp->rx_queue->queue_depth = get_depth(vp->parsed); rx_queue 1355 arch/um/drivers/vector_kern.c ring->rx_max_pending = vp->rx_queue->max_depth; rx_queue 1357 arch/um/drivers/vector_kern.c ring->rx_pending = vp->rx_queue->max_depth; rx_queue 83 arch/um/drivers/vector_kern.h struct vector_queue *rx_queue; rx_queue 340 drivers/atm/eni.c skb_queue_head_init(&ENI_DEV(dev)->rx_queue); rx_queue 476 drivers/atm/eni.c skb_queue_tail(&eni_dev->rx_queue,skb); rx_queue 738 drivers/atm/eni.c skb = skb_dequeue(&eni_dev->rx_queue); rx_queue 758 drivers/atm/eni.c skb_queue_head(&eni_dev->rx_queue,skb); rx_queue 921 drivers/atm/eni.c skb_queue_head_init(&eni_dev->rx_queue); rx_queue 101 drivers/atm/eni.h struct sk_buff_head rx_queue; /* PDUs currently being RX-DMAed */ rx_queue 181 drivers/net/dsa/mv88e6xxx/chip.h struct sk_buff_head rx_queue; rx_queue 348 drivers/net/dsa/mv88e6xxx/hwtstamp.c skb = skb_dequeue(&ps->rx_queue); rx_queue 352 drivers/net/dsa/mv88e6xxx/hwtstamp.c &ps->rx_queue); rx_queue 387 drivers/net/dsa/mv88e6xxx/hwtstamp.c skb_queue_tail(&ps->rx_queue, skb); rx_queue 576 drivers/net/dsa/mv88e6xxx/hwtstamp.c skb_queue_head_init(&ps->rx_queue); rx_queue 585 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h u8 rx_queue; /* index for skb_record */ rx_queue 749 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c skb_record_rx_queue(skb, fp->rx_queue); rx_queue 1079 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c skb_record_rx_queue(skb, fp->rx_queue); rx_queue 6346 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c fp->rx_queue = fp_idx; rx_queue 6449 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bnx2x_fcoe(bp, rx_queue) = BNX2X_NUM_ETH_QUEUES(bp); rx_queue 144 drivers/net/ethernet/chelsio/cxgb3/adapter.h struct sk_buff_head rx_queue; /* offload packet receive queue */ rx_queue 1854 drivers/net/ethernet/chelsio/cxgb3/sge.c int was_empty = skb_queue_empty(&q->rx_queue); rx_queue 1856 drivers/net/ethernet/chelsio/cxgb3/sge.c __skb_queue_tail(&q->rx_queue, skb); rx_queue 1909 drivers/net/ethernet/chelsio/cxgb3/sge.c skb_queue_splice_init(&q->rx_queue, &queue); rx_queue 1936 drivers/net/ethernet/chelsio/cxgb3/sge.c skb_queue_splice(&queue, &q->rx_queue); rx_queue 3099 drivers/net/ethernet/chelsio/cxgb3/sge.c skb_queue_head_init(&q->rspq.rx_queue); rx_queue 524 drivers/net/ethernet/freescale/fec.h struct fec_enet_priv_rx_q *rx_queue[FEC_ENET_MAX_RX_QS]; rx_queue 838 drivers/net/ethernet/freescale/fec_main.c rxq = fep->rx_queue[q]; rx_queue 894 drivers/net/ethernet/freescale/fec_main.c writel(0, fep->rx_queue[i]->bd.reg_desc_active); rx_queue 905 drivers/net/ethernet/freescale/fec_main.c rxq = fep->rx_queue[i]; rx_queue 1434 drivers/net/ethernet/freescale/fec_main.c rxq = fep->rx_queue[queue_id]; rx_queue 2753 drivers/net/ethernet/freescale/fec_main.c rxq = fep->rx_queue[q]; rx_queue 2798 drivers/net/ethernet/freescale/fec_main.c kfree(fep->rx_queue[i]); rx_queue 2836 drivers/net/ethernet/freescale/fec_main.c fep->rx_queue[i] = kzalloc(sizeof(*fep->rx_queue[i]), rx_queue 2838 drivers/net/ethernet/freescale/fec_main.c if (!fep->rx_queue[i]) { rx_queue 2843 drivers/net/ethernet/freescale/fec_main.c fep->rx_queue[i]->bd.ring_size = RX_RING_SIZE; rx_queue 2844 drivers/net/ethernet/freescale/fec_main.c fep->total_rx_ring_size += fep->rx_queue[i]->bd.ring_size; rx_queue 2862 drivers/net/ethernet/freescale/fec_main.c rxq = fep->rx_queue[queue]; rx_queue 3274 drivers/net/ethernet/freescale/fec_main.c struct fec_enet_priv_rx_q *rxq = fep->rx_queue[i]; rx_queue 112 drivers/net/ethernet/freescale/gianfar.c static void gfar_init_rxbdp(struct gfar_priv_rx_q *rx_queue, struct rxbd8 *bdp, rx_queue 120 drivers/net/ethernet/freescale/gianfar.c if (bdp == rx_queue->rx_bd_base + rx_queue->rx_ring_size - 1) rx_queue 142 drivers/net/ethernet/freescale/gianfar.c gfar_write(baddr, priv->rx_queue[i]->rx_bd_dma_base); rx_queue 155 drivers/net/ethernet/freescale/gianfar.c gfar_write(baddr, priv->rx_queue[i]->rx_ring_size | rx_queue 261 drivers/net/ethernet/freescale/gianfar.c if (likely(priv->rx_queue[i]->rxcoalescing)) rx_queue 262 drivers/net/ethernet/freescale/gianfar.c gfar_write(baddr + i, priv->rx_queue[i]->rxic); rx_queue 273 drivers/net/ethernet/freescale/gianfar.c if (unlikely(priv->rx_queue[0]->rxcoalescing)) rx_queue 274 drivers/net/ethernet/freescale/gianfar.c gfar_write(®s->rxic, priv->rx_queue[0]->rxic); rx_queue 291 drivers/net/ethernet/freescale/gianfar.c rx_packets += priv->rx_queue[i]->stats.rx_packets; rx_queue 292 drivers/net/ethernet/freescale/gianfar.c rx_bytes += priv->rx_queue[i]->stats.rx_bytes; rx_queue 293 drivers/net/ethernet/freescale/gianfar.c rx_dropped += priv->rx_queue[i]->stats.rx_dropped; rx_queue 422 drivers/net/ethernet/freescale/gianfar.c priv->rx_queue[i] = kzalloc(sizeof(struct gfar_priv_rx_q), rx_queue 424 drivers/net/ethernet/freescale/gianfar.c if (!priv->rx_queue[i]) rx_queue 427 drivers/net/ethernet/freescale/gianfar.c priv->rx_queue[i]->qindex = i; rx_queue 428 drivers/net/ethernet/freescale/gianfar.c priv->rx_queue[i]->ndev = priv->ndev; rx_queue 446 drivers/net/ethernet/freescale/gianfar.c kfree(priv->rx_queue[i]); rx_queue 561 drivers/net/ethernet/freescale/gianfar.c if (!grp->rx_queue) rx_queue 562 drivers/net/ethernet/freescale/gianfar.c grp->rx_queue = priv->rx_queue[i]; rx_queue 566 drivers/net/ethernet/freescale/gianfar.c priv->rx_queue[i]->grp = grp; rx_queue 1116 drivers/net/ethernet/freescale/gianfar.c static void free_skb_rx_queue(struct gfar_priv_rx_q *rx_queue) rx_queue 1120 drivers/net/ethernet/freescale/gianfar.c struct rxbd8 *rxbdp = rx_queue->rx_bd_base; rx_queue 1122 drivers/net/ethernet/freescale/gianfar.c dev_kfree_skb(rx_queue->skb); rx_queue 1124 drivers/net/ethernet/freescale/gianfar.c for (i = 0; i < rx_queue->rx_ring_size; i++) { rx_queue 1125 drivers/net/ethernet/freescale/gianfar.c struct gfar_rx_buff *rxb = &rx_queue->rx_buff[i]; rx_queue 1134 drivers/net/ethernet/freescale/gianfar.c dma_unmap_page(rx_queue->dev, rxb->dma, rx_queue 1141 drivers/net/ethernet/freescale/gianfar.c kfree(rx_queue->rx_buff); rx_queue 1142 drivers/net/ethernet/freescale/gianfar.c rx_queue->rx_buff = NULL; rx_queue 1151 drivers/net/ethernet/freescale/gianfar.c struct gfar_priv_rx_q *rx_queue = NULL; rx_queue 1166 drivers/net/ethernet/freescale/gianfar.c rx_queue = priv->rx_queue[i]; rx_queue 1167 drivers/net/ethernet/freescale/gianfar.c if (rx_queue->rx_buff) rx_queue 1168 drivers/net/ethernet/freescale/gianfar.c free_skb_rx_queue(rx_queue); rx_queue 1258 drivers/net/ethernet/freescale/gianfar.c static void gfar_rx_alloc_err(struct gfar_priv_rx_q *rx_queue) rx_queue 1260 drivers/net/ethernet/freescale/gianfar.c struct gfar_private *priv = netdev_priv(rx_queue->ndev); rx_queue 1263 drivers/net/ethernet/freescale/gianfar.c netdev_err(rx_queue->ndev, "Can't alloc RX buffers\n"); rx_queue 1267 drivers/net/ethernet/freescale/gianfar.c static void gfar_alloc_rx_buffs(struct gfar_priv_rx_q *rx_queue, rx_queue 1274 drivers/net/ethernet/freescale/gianfar.c i = rx_queue->next_to_use; rx_queue 1275 drivers/net/ethernet/freescale/gianfar.c bdp = &rx_queue->rx_bd_base[i]; rx_queue 1276 drivers/net/ethernet/freescale/gianfar.c rxb = &rx_queue->rx_buff[i]; rx_queue 1281 drivers/net/ethernet/freescale/gianfar.c if (unlikely(!gfar_new_page(rx_queue, rxb))) { rx_queue 1282 drivers/net/ethernet/freescale/gianfar.c gfar_rx_alloc_err(rx_queue); rx_queue 1288 drivers/net/ethernet/freescale/gianfar.c gfar_init_rxbdp(rx_queue, bdp, rx_queue 1295 drivers/net/ethernet/freescale/gianfar.c if (unlikely(++i == rx_queue->rx_ring_size)) { rx_queue 1297 drivers/net/ethernet/freescale/gianfar.c bdp = rx_queue->rx_bd_base; rx_queue 1298 drivers/net/ethernet/freescale/gianfar.c rxb = rx_queue->rx_buff; rx_queue 1302 drivers/net/ethernet/freescale/gianfar.c rx_queue->next_to_use = i; rx_queue 1303 drivers/net/ethernet/freescale/gianfar.c rx_queue->next_to_alloc = i; rx_queue 1311 drivers/net/ethernet/freescale/gianfar.c struct gfar_priv_rx_q *rx_queue = NULL; rx_queue 1341 drivers/net/ethernet/freescale/gianfar.c rx_queue = priv->rx_queue[i]; rx_queue 1343 drivers/net/ethernet/freescale/gianfar.c rx_queue->next_to_clean = 0; rx_queue 1344 drivers/net/ethernet/freescale/gianfar.c rx_queue->next_to_use = 0; rx_queue 1345 drivers/net/ethernet/freescale/gianfar.c rx_queue->next_to_alloc = 0; rx_queue 1350 drivers/net/ethernet/freescale/gianfar.c gfar_alloc_rx_buffs(rx_queue, gfar_rxbd_unused(rx_queue)); rx_queue 1352 drivers/net/ethernet/freescale/gianfar.c rx_queue->rfbptr = rfbptr; rx_queue 1365 drivers/net/ethernet/freescale/gianfar.c struct gfar_priv_rx_q *rx_queue = NULL; rx_queue 1373 drivers/net/ethernet/freescale/gianfar.c priv->total_rx_ring_size += priv->rx_queue[i]->rx_ring_size; rx_queue 1397 drivers/net/ethernet/freescale/gianfar.c rx_queue = priv->rx_queue[i]; rx_queue 1398 drivers/net/ethernet/freescale/gianfar.c rx_queue->rx_bd_base = vaddr; rx_queue 1399 drivers/net/ethernet/freescale/gianfar.c rx_queue->rx_bd_dma_base = addr; rx_queue 1400 drivers/net/ethernet/freescale/gianfar.c rx_queue->ndev = ndev; rx_queue 1401 drivers/net/ethernet/freescale/gianfar.c rx_queue->dev = dev; rx_queue 1402 drivers/net/ethernet/freescale/gianfar.c addr += sizeof(struct rxbd8) * rx_queue->rx_ring_size; rx_queue 1403 drivers/net/ethernet/freescale/gianfar.c vaddr += sizeof(struct rxbd8) * rx_queue->rx_ring_size; rx_queue 1421 drivers/net/ethernet/freescale/gianfar.c rx_queue = priv->rx_queue[i]; rx_queue 1422 drivers/net/ethernet/freescale/gianfar.c rx_queue->rx_buff = kcalloc(rx_queue->rx_ring_size, rx_queue 1423 drivers/net/ethernet/freescale/gianfar.c sizeof(*rx_queue->rx_buff), rx_queue 1425 drivers/net/ethernet/freescale/gianfar.c if (!rx_queue->rx_buff) rx_queue 1513 drivers/net/ethernet/freescale/gianfar.c struct gfar_priv_rx_q *rx_queue = NULL; rx_queue 1573 drivers/net/ethernet/freescale/gianfar.c rx_queue = priv->rx_queue[i]; rx_queue 1574 drivers/net/ethernet/freescale/gianfar.c bdp_dma = gfar_rxbd_dma_lastfree(rx_queue); rx_queue 1575 drivers/net/ethernet/freescale/gianfar.c gfar_write(rx_queue->rfbptr, bdp_dma); rx_queue 2435 drivers/net/ethernet/freescale/gianfar.c static struct sk_buff *gfar_get_next_rxbuff(struct gfar_priv_rx_q *rx_queue, rx_queue 2438 drivers/net/ethernet/freescale/gianfar.c struct gfar_rx_buff *rxb = &rx_queue->rx_buff[rx_queue->next_to_clean]; rx_queue 2447 drivers/net/ethernet/freescale/gianfar.c gfar_rx_alloc_err(rx_queue); rx_queue 2454 drivers/net/ethernet/freescale/gianfar.c dma_sync_single_range_for_cpu(rx_queue->dev, rxb->dma, rxb->page_offset, rx_queue 2459 drivers/net/ethernet/freescale/gianfar.c gfar_reuse_rx_page(rx_queue, rxb); rx_queue 2462 drivers/net/ethernet/freescale/gianfar.c dma_unmap_page(rx_queue->dev, rxb->dma, rx_queue 2532 drivers/net/ethernet/freescale/gianfar.c static int gfar_clean_rx_ring(struct gfar_priv_rx_q *rx_queue, rx_queue 2535 drivers/net/ethernet/freescale/gianfar.c struct net_device *ndev = rx_queue->ndev; rx_queue 2539 drivers/net/ethernet/freescale/gianfar.c struct sk_buff *skb = rx_queue->skb; rx_queue 2540 drivers/net/ethernet/freescale/gianfar.c int cleaned_cnt = gfar_rxbd_unused(rx_queue); rx_queue 2544 drivers/net/ethernet/freescale/gianfar.c i = rx_queue->next_to_clean; rx_queue 2550 drivers/net/ethernet/freescale/gianfar.c gfar_alloc_rx_buffs(rx_queue, cleaned_cnt); rx_queue 2554 drivers/net/ethernet/freescale/gianfar.c bdp = &rx_queue->rx_bd_base[i]; rx_queue 2563 drivers/net/ethernet/freescale/gianfar.c skb = gfar_get_next_rxbuff(rx_queue, lstatus, skb); rx_queue 2570 drivers/net/ethernet/freescale/gianfar.c if (unlikely(++i == rx_queue->rx_ring_size)) rx_queue 2573 drivers/net/ethernet/freescale/gianfar.c rx_queue->next_to_clean = i; rx_queue 2585 drivers/net/ethernet/freescale/gianfar.c rx_queue->stats.rx_dropped++; rx_queue 2595 drivers/net/ethernet/freescale/gianfar.c skb_record_rx_queue(skb, rx_queue->qindex); rx_queue 2600 drivers/net/ethernet/freescale/gianfar.c napi_gro_receive(&rx_queue->grp->napi_rx, skb); rx_queue 2606 drivers/net/ethernet/freescale/gianfar.c rx_queue->skb = skb; rx_queue 2608 drivers/net/ethernet/freescale/gianfar.c rx_queue->stats.rx_packets += total_pkts; rx_queue 2609 drivers/net/ethernet/freescale/gianfar.c rx_queue->stats.rx_bytes += total_bytes; rx_queue 2612 drivers/net/ethernet/freescale/gianfar.c gfar_alloc_rx_buffs(rx_queue, cleaned_cnt); rx_queue 2616 drivers/net/ethernet/freescale/gianfar.c u32 bdp_dma = gfar_rxbd_dma_lastfree(rx_queue); rx_queue 2618 drivers/net/ethernet/freescale/gianfar.c gfar_write(rx_queue->rfbptr, bdp_dma); rx_queue 2629 drivers/net/ethernet/freescale/gianfar.c struct gfar_priv_rx_q *rx_queue = gfargrp->rx_queue; rx_queue 2637 drivers/net/ethernet/freescale/gianfar.c work_done = gfar_clean_rx_ring(rx_queue, budget); rx_queue 2689 drivers/net/ethernet/freescale/gianfar.c struct gfar_priv_rx_q *rx_queue = NULL; rx_queue 2711 drivers/net/ethernet/freescale/gianfar.c rx_queue = priv->rx_queue[i]; rx_queue 2713 drivers/net/ethernet/freescale/gianfar.c gfar_clean_rx_ring(rx_queue, budget_per_q); rx_queue 3389 drivers/net/ethernet/freescale/gianfar.c priv->rx_queue[i]->rx_ring_size = DEFAULT_RX_RING_SIZE; rx_queue 3390 drivers/net/ethernet/freescale/gianfar.c priv->rx_queue[i]->rxcoalescing = DEFAULT_RX_COALESCE; rx_queue 3391 drivers/net/ethernet/freescale/gianfar.c priv->rx_queue[i]->rxic = DEFAULT_RXIC; rx_queue 3452 drivers/net/ethernet/freescale/gianfar.c i, priv->rx_queue[i]->rx_ring_size); rx_queue 3532 drivers/net/ethernet/freescale/gianfar.c u8 qindex = (u8)priv->gfargrp[0].rx_queue->qindex; rx_queue 1052 drivers/net/ethernet/freescale/gianfar.h struct gfar_priv_rx_q *rx_queue; rx_queue 1103 drivers/net/ethernet/freescale/gianfar.h struct gfar_priv_rx_q *rx_queue[MAX_RX_QS]; rx_queue 253 drivers/net/ethernet/freescale/gianfar_ethtool.c struct gfar_priv_rx_q *rx_queue = NULL; rx_queue 266 drivers/net/ethernet/freescale/gianfar_ethtool.c rx_queue = priv->rx_queue[0]; rx_queue 269 drivers/net/ethernet/freescale/gianfar_ethtool.c rxtime = get_ictt_value(rx_queue->rxic); rx_queue 270 drivers/net/ethernet/freescale/gianfar_ethtool.c rxcount = get_icft_value(rx_queue->rxic); rx_queue 360 drivers/net/ethernet/freescale/gianfar_ethtool.c priv->rx_queue[i]->rxcoalescing = 0; rx_queue 363 drivers/net/ethernet/freescale/gianfar_ethtool.c priv->rx_queue[i]->rxcoalescing = 1; rx_queue 367 drivers/net/ethernet/freescale/gianfar_ethtool.c priv->rx_queue[i]->rxic = mk_ic_value( rx_queue 408 drivers/net/ethernet/freescale/gianfar_ethtool.c struct gfar_priv_rx_q *rx_queue = NULL; rx_queue 411 drivers/net/ethernet/freescale/gianfar_ethtool.c rx_queue = priv->rx_queue[0]; rx_queue 421 drivers/net/ethernet/freescale/gianfar_ethtool.c rvals->rx_pending = rx_queue->rx_ring_size; rx_queue 422 drivers/net/ethernet/freescale/gianfar_ethtool.c rvals->rx_mini_pending = rx_queue->rx_ring_size; rx_queue 423 drivers/net/ethernet/freescale/gianfar_ethtool.c rvals->rx_jumbo_pending = rx_queue->rx_ring_size; rx_queue 460 drivers/net/ethernet/freescale/gianfar_ethtool.c priv->rx_queue[i]->rx_ring_size = rvals->rx_pending; rx_queue 105 drivers/net/ethernet/ibm/ibmveth.c return be32_to_cpu(adapter->rx_queue.queue_addr[adapter->rx_queue.index].flags_off); rx_queue 116 drivers/net/ethernet/ibm/ibmveth.c return ibmveth_rxq_toggle(adapter) == adapter->rx_queue.toggle; rx_queue 136 drivers/net/ethernet/ibm/ibmveth.c return be32_to_cpu(adapter->rx_queue.queue_addr[adapter->rx_queue.index].length); rx_queue 397 drivers/net/ethernet/ibm/ibmveth.c u64 correlator = adapter->rx_queue.queue_addr[adapter->rx_queue.index].correlator; rx_queue 410 drivers/net/ethernet/ibm/ibmveth.c u32 q_index = adapter->rx_queue.index; rx_queue 411 drivers/net/ethernet/ibm/ibmveth.c u64 correlator = adapter->rx_queue.queue_addr[q_index].correlator; rx_queue 436 drivers/net/ethernet/ibm/ibmveth.c ibmveth_remove_buffer_from_pool(adapter, adapter->rx_queue.queue_addr[adapter->rx_queue.index].correlator); rx_queue 440 drivers/net/ethernet/ibm/ibmveth.c if (++adapter->rx_queue.index == adapter->rx_queue.num_slots) { rx_queue 441 drivers/net/ethernet/ibm/ibmveth.c adapter->rx_queue.index = 0; rx_queue 442 drivers/net/ethernet/ibm/ibmveth.c adapter->rx_queue.toggle = !adapter->rx_queue.toggle; rx_queue 451 drivers/net/ethernet/ibm/ibmveth.c ibmveth_remove_buffer_from_pool(adapter, adapter->rx_queue.queue_addr[adapter->rx_queue.index].correlator); rx_queue 453 drivers/net/ethernet/ibm/ibmveth.c if (++adapter->rx_queue.index == adapter->rx_queue.num_slots) { rx_queue 454 drivers/net/ethernet/ibm/ibmveth.c adapter->rx_queue.index = 0; rx_queue 455 drivers/net/ethernet/ibm/ibmveth.c adapter->rx_queue.toggle = !adapter->rx_queue.toggle; rx_queue 530 drivers/net/ethernet/ibm/ibmveth.c adapter->rx_queue.queue_len = sizeof(struct ibmveth_rx_q_entry) * rx_queue 532 drivers/net/ethernet/ibm/ibmveth.c adapter->rx_queue.queue_addr = rx_queue 533 drivers/net/ethernet/ibm/ibmveth.c dma_alloc_coherent(dev, adapter->rx_queue.queue_len, rx_queue 534 drivers/net/ethernet/ibm/ibmveth.c &adapter->rx_queue.queue_dma, GFP_KERNEL); rx_queue 535 drivers/net/ethernet/ibm/ibmveth.c if (!adapter->rx_queue.queue_addr) rx_queue 552 drivers/net/ethernet/ibm/ibmveth.c adapter->rx_queue.index = 0; rx_queue 553 drivers/net/ethernet/ibm/ibmveth.c adapter->rx_queue.num_slots = rxq_entries; rx_queue 554 drivers/net/ethernet/ibm/ibmveth.c adapter->rx_queue.toggle = 1; rx_queue 559 drivers/net/ethernet/ibm/ibmveth.c adapter->rx_queue.queue_len; rx_queue 560 drivers/net/ethernet/ibm/ibmveth.c rxq_desc.fields.address = adapter->rx_queue.queue_dma; rx_queue 564 drivers/net/ethernet/ibm/ibmveth.c netdev_dbg(netdev, "receive q @ 0x%p\n", adapter->rx_queue.queue_addr); rx_queue 647 drivers/net/ethernet/ibm/ibmveth.c dma_free_coherent(dev, adapter->rx_queue.queue_len, rx_queue 648 drivers/net/ethernet/ibm/ibmveth.c adapter->rx_queue.queue_addr, rx_queue 649 drivers/net/ethernet/ibm/ibmveth.c adapter->rx_queue.queue_dma); rx_queue 696 drivers/net/ethernet/ibm/ibmveth.c dma_free_coherent(dev, adapter->rx_queue.queue_len, rx_queue 697 drivers/net/ethernet/ibm/ibmveth.c adapter->rx_queue.queue_addr, rx_queue 698 drivers/net/ethernet/ibm/ibmveth.c adapter->rx_queue.queue_dma); rx_queue 143 drivers/net/ethernet/ibm/ibmveth.h struct ibmveth_rx_q rx_queue; rx_queue 794 drivers/net/ethernet/intel/igb/igb_main.c int rx_queue = IGB_N0_QUEUE; rx_queue 799 drivers/net/ethernet/intel/igb/igb_main.c rx_queue = q_vector->rx.ring->reg_idx; rx_queue 810 drivers/net/ethernet/intel/igb/igb_main.c if (rx_queue > IGB_N0_QUEUE) rx_queue 811 drivers/net/ethernet/intel/igb/igb_main.c msixbm = E1000_EICR_RX_QUEUE0 << rx_queue; rx_queue 825 drivers/net/ethernet/intel/igb/igb_main.c if (rx_queue > IGB_N0_QUEUE) rx_queue 827 drivers/net/ethernet/intel/igb/igb_main.c rx_queue & 0x7, rx_queue 828 drivers/net/ethernet/intel/igb/igb_main.c (rx_queue & 0x8) << 1); rx_queue 846 drivers/net/ethernet/intel/igb/igb_main.c if (rx_queue > IGB_N0_QUEUE) rx_queue 848 drivers/net/ethernet/intel/igb/igb_main.c rx_queue >> 1, rx_queue 849 drivers/net/ethernet/intel/igb/igb_main.c (rx_queue & 0x1) << 4); rx_queue 921 drivers/net/ethernet/intel/igbvf/netdev.c static void igbvf_assign_vector(struct igbvf_adapter *adapter, int rx_queue, rx_queue 932 drivers/net/ethernet/intel/igbvf/netdev.c if (rx_queue > IGBVF_NO_QUEUE) { rx_queue 933 drivers/net/ethernet/intel/igbvf/netdev.c index = (rx_queue >> 1); rx_queue 935 drivers/net/ethernet/intel/igbvf/netdev.c if (rx_queue & 0x1) { rx_queue 944 drivers/net/ethernet/intel/igbvf/netdev.c adapter->rx_ring[rx_queue].eims_value = BIT(msix_vector); rx_queue 2595 drivers/net/ethernet/intel/igc/igc_main.c int rx_queue = IGC_N0_QUEUE; rx_queue 2599 drivers/net/ethernet/intel/igc/igc_main.c rx_queue = q_vector->rx.ring->reg_idx; rx_queue 2605 drivers/net/ethernet/intel/igc/igc_main.c if (rx_queue > IGC_N0_QUEUE) rx_queue 2607 drivers/net/ethernet/intel/igc/igc_main.c rx_queue >> 1, rx_queue 2608 drivers/net/ethernet/intel/igc/igc_main.c (rx_queue & 0x1) << 4); rx_queue 398 drivers/net/ethernet/marvell/mv643xx_eth.c struct rx_queue rxq[8]; rx_queue 440 drivers/net/ethernet/marvell/mv643xx_eth.c static struct mv643xx_eth_private *rxq_to_mp(struct rx_queue *rxq) rx_queue 450 drivers/net/ethernet/marvell/mv643xx_eth.c static void rxq_enable(struct rx_queue *rxq) rx_queue 456 drivers/net/ethernet/marvell/mv643xx_eth.c static void rxq_disable(struct rx_queue *rxq) rx_queue 505 drivers/net/ethernet/marvell/mv643xx_eth.c static int rxq_process(struct rx_queue *rxq, int budget) rx_queue 597 drivers/net/ethernet/marvell/mv643xx_eth.c static int rxq_refill(struct rx_queue *rxq, int budget) rx_queue 1934 drivers/net/ethernet/marvell/mv643xx_eth.c struct rx_queue *rxq = mp->rxq + index; rx_queue 1999 drivers/net/ethernet/marvell/mv643xx_eth.c static void rxq_deinit(struct rx_queue *rxq) rx_queue 2377 drivers/net/ethernet/marvell/mv643xx_eth.c struct rx_queue *rxq = mp->rxq + i; rx_queue 2777 drivers/net/ethernet/marvell/mvneta.c int rx_queue; rx_queue 2810 drivers/net/ethernet/marvell/mvneta.c rx_queue = fls(((cause_rx_tx >> 8) & 0xff)); rx_queue 2811 drivers/net/ethernet/marvell/mvneta.c if (rx_queue) { rx_queue 2812 drivers/net/ethernet/marvell/mvneta.c rx_queue = rx_queue - 1; rx_queue 2815 drivers/net/ethernet/marvell/mvneta.c &pp->rxqs[rx_queue]); rx_queue 2818 drivers/net/ethernet/marvell/mvneta.c &pp->rxqs[rx_queue]); rx_queue 66 drivers/net/ethernet/qlogic/qed/qed_ll2.c #define QED_LL2_RX_REGISTERED(ll2) ((ll2)->rx_queue.b_cb_registered) rx_queue 446 drivers/net/ethernet/qlogic/qed/qed_ll2.c struct qed_ll2_rx_queue *p_rx = &p_ll2_conn->rx_queue; rx_queue 481 drivers/net/ethernet/qlogic/qed/qed_ll2.c struct qed_ll2_rx_queue *p_rx = &p_ll2_conn->rx_queue; rx_queue 523 drivers/net/ethernet/qlogic/qed/qed_ll2.c struct qed_ll2_rx_queue *p_rx = &p_ll2_conn->rx_queue; rx_queue 577 drivers/net/ethernet/qlogic/qed/qed_ll2.c p_rx = &p_ll2_conn->rx_queue; rx_queue 634 drivers/net/ethernet/qlogic/qed/qed_ll2.c struct qed_ll2_rx_queue *p_rx = &p_ll2_conn->rx_queue; rx_queue 945 drivers/net/ethernet/qlogic/qed/qed_ll2.c struct qed_ll2_rx_queue *p_rx = &p_ll2_conn->rx_queue; rx_queue 1165 drivers/net/ethernet/qlogic/qed/qed_ll2.c &p_ll2_info->rx_queue.rxq_chain, NULL); rx_queue 1171 drivers/net/ethernet/qlogic/qed/qed_ll2.c capacity = qed_chain_get_capacity(&p_ll2_info->rx_queue.rxq_chain); rx_queue 1179 drivers/net/ethernet/qlogic/qed/qed_ll2.c p_ll2_info->rx_queue.descq_array = p_descq; rx_queue 1187 drivers/net/ethernet/qlogic/qed/qed_ll2.c &p_ll2_info->rx_queue.rcq_chain, NULL); rx_queue 1423 drivers/net/ethernet/qlogic/qed/qed_ll2.c &p_ll2_info->rx_queue.rx_sb_index, rx_queue 1424 drivers/net/ethernet/qlogic/qed/qed_ll2.c &p_ll2_info->rx_queue.p_fw_cons); rx_queue 1425 drivers/net/ethernet/qlogic/qed/qed_ll2.c p_ll2_info->rx_queue.b_cb_registered = true; rx_queue 1455 drivers/net/ethernet/qlogic/qed/qed_ll2.c DIRECT_REG_WR(p_ll2_conn->rx_queue.set_prod_addr, 0x0); rx_queue 1501 drivers/net/ethernet/qlogic/qed/qed_ll2.c p_rx = &p_ll2_conn->rx_queue; rx_queue 1644 drivers/net/ethernet/qlogic/qed/qed_ll2.c p_rx = &p_ll2_conn->rx_queue; rx_queue 1966 drivers/net/ethernet/qlogic/qed/qed_ll2.c p_ll2_conn->rx_queue.b_cb_registered = false; rx_queue 1973 drivers/net/ethernet/qlogic/qed/qed_ll2.c qed_int_unregister_cb(p_hwfn, p_ll2_conn->rx_queue.rx_sb_index); rx_queue 2025 drivers/net/ethernet/qlogic/qed/qed_ll2.c kfree(p_ll2_conn->rx_queue.descq_array); rx_queue 2026 drivers/net/ethernet/qlogic/qed/qed_ll2.c qed_chain_free(p_hwfn->cdev, &p_ll2_conn->rx_queue.rxq_chain); rx_queue 2027 drivers/net/ethernet/qlogic/qed/qed_ll2.c qed_chain_free(p_hwfn->cdev, &p_ll2_conn->rx_queue.rcq_chain); rx_queue 126 drivers/net/ethernet/qlogic/qed/qed_ll2.h struct qed_ll2_rx_queue rx_queue; rx_queue 3072 drivers/net/ethernet/sfc/ef10.c static int efx_ef10_rx_probe(struct efx_rx_queue *rx_queue) rx_queue 3074 drivers/net/ethernet/sfc/ef10.c return efx_nic_alloc_buffer(rx_queue->efx, &rx_queue->rxd.buf, rx_queue 3075 drivers/net/ethernet/sfc/ef10.c (rx_queue->ptr_mask + 1) * rx_queue 3080 drivers/net/ethernet/sfc/ef10.c static void efx_ef10_rx_init(struct efx_rx_queue *rx_queue) rx_queue 3085 drivers/net/ethernet/sfc/ef10.c struct efx_channel *channel = efx_rx_queue_channel(rx_queue); rx_queue 3086 drivers/net/ethernet/sfc/ef10.c size_t entries = rx_queue->rxd.buf.len / EFX_BUF_SIZE; rx_queue 3087 drivers/net/ethernet/sfc/ef10.c struct efx_nic *efx = rx_queue->efx; rx_queue 3095 drivers/net/ethernet/sfc/ef10.c rx_queue->scatter_n = 0; rx_queue 3096 drivers/net/ethernet/sfc/ef10.c rx_queue->scatter_len = 0; rx_queue 3098 drivers/net/ethernet/sfc/ef10.c MCDI_SET_DWORD(inbuf, INIT_RXQ_IN_SIZE, rx_queue->ptr_mask + 1); rx_queue 3100 drivers/net/ethernet/sfc/ef10.c MCDI_SET_DWORD(inbuf, INIT_RXQ_IN_LABEL, efx_rx_queue_index(rx_queue)); rx_queue 3102 drivers/net/ethernet/sfc/ef10.c efx_rx_queue_index(rx_queue)); rx_queue 3109 drivers/net/ethernet/sfc/ef10.c dma_addr = rx_queue->rxd.buf.dma_addr; rx_queue 3112 drivers/net/ethernet/sfc/ef10.c efx_rx_queue_index(rx_queue), entries, (u64)dma_addr); rx_queue 3125 drivers/net/ethernet/sfc/ef10.c efx_rx_queue_index(rx_queue)); rx_queue 3128 drivers/net/ethernet/sfc/ef10.c static void efx_ef10_rx_fini(struct efx_rx_queue *rx_queue) rx_queue 3132 drivers/net/ethernet/sfc/ef10.c struct efx_nic *efx = rx_queue->efx; rx_queue 3137 drivers/net/ethernet/sfc/ef10.c efx_rx_queue_index(rx_queue)); rx_queue 3152 drivers/net/ethernet/sfc/ef10.c static void efx_ef10_rx_remove(struct efx_rx_queue *rx_queue) rx_queue 3154 drivers/net/ethernet/sfc/ef10.c efx_nic_free_buffer(rx_queue->efx, &rx_queue->rxd.buf); rx_queue 3159 drivers/net/ethernet/sfc/ef10.c efx_ef10_build_rx_desc(struct efx_rx_queue *rx_queue, unsigned int index) rx_queue 3164 drivers/net/ethernet/sfc/ef10.c rxd = efx_rx_desc(rx_queue, index); rx_queue 3165 drivers/net/ethernet/sfc/ef10.c rx_buf = efx_rx_buffer(rx_queue, index); rx_queue 3171 drivers/net/ethernet/sfc/ef10.c static void efx_ef10_rx_write(struct efx_rx_queue *rx_queue) rx_queue 3173 drivers/net/ethernet/sfc/ef10.c struct efx_nic *efx = rx_queue->efx; rx_queue 3178 drivers/net/ethernet/sfc/ef10.c write_count = rx_queue->added_count & ~7; rx_queue 3179 drivers/net/ethernet/sfc/ef10.c if (rx_queue->notified_count == write_count) rx_queue 3184 drivers/net/ethernet/sfc/ef10.c rx_queue, rx_queue 3185 drivers/net/ethernet/sfc/ef10.c rx_queue->notified_count & rx_queue->ptr_mask); rx_queue 3186 drivers/net/ethernet/sfc/ef10.c while (++rx_queue->notified_count != write_count); rx_queue 3190 drivers/net/ethernet/sfc/ef10.c write_count & rx_queue->ptr_mask); rx_queue 3192 drivers/net/ethernet/sfc/ef10.c efx_rx_queue_index(rx_queue)); rx_queue 3197 drivers/net/ethernet/sfc/ef10.c static void efx_ef10_rx_defer_refill(struct efx_rx_queue *rx_queue) rx_queue 3199 drivers/net/ethernet/sfc/ef10.c struct efx_channel *channel = efx_rx_queue_channel(rx_queue); rx_queue 3394 drivers/net/ethernet/sfc/ef10.c static void efx_ef10_handle_rx_wrong_queue(struct efx_rx_queue *rx_queue, rx_queue 3397 drivers/net/ethernet/sfc/ef10.c struct efx_nic *efx = rx_queue->efx; rx_queue 3401 drivers/net/ethernet/sfc/ef10.c efx_rx_queue_index(rx_queue), rx_queue_label); rx_queue 3407 drivers/net/ethernet/sfc/ef10.c efx_ef10_handle_rx_bad_lbits(struct efx_rx_queue *rx_queue, rx_queue 3410 drivers/net/ethernet/sfc/ef10.c unsigned int dropped = (actual - expected) & rx_queue->ptr_mask; rx_queue 3411 drivers/net/ethernet/sfc/ef10.c struct efx_nic *efx = rx_queue->efx; rx_queue 3421 drivers/net/ethernet/sfc/ef10.c static void efx_ef10_handle_rx_abort(struct efx_rx_queue *rx_queue) rx_queue 3425 drivers/net/ethernet/sfc/ef10.c netif_dbg(rx_queue->efx, hw, rx_queue->efx->net_dev, rx_queue 3427 drivers/net/ethernet/sfc/ef10.c rx_queue->scatter_n); rx_queue 3429 drivers/net/ethernet/sfc/ef10.c rx_desc_ptr = rx_queue->removed_count & rx_queue->ptr_mask; rx_queue 3431 drivers/net/ethernet/sfc/ef10.c efx_rx_packet(rx_queue, rx_desc_ptr, rx_queue->scatter_n, rx_queue 3434 drivers/net/ethernet/sfc/ef10.c rx_queue->removed_count += rx_queue->scatter_n; rx_queue 3435 drivers/net/ethernet/sfc/ef10.c rx_queue->scatter_n = 0; rx_queue 3436 drivers/net/ethernet/sfc/ef10.c rx_queue->scatter_len = 0; rx_queue 3437 drivers/net/ethernet/sfc/ef10.c ++efx_rx_queue_channel(rx_queue)->n_rx_nodesc_trunc; rx_queue 3539 drivers/net/ethernet/sfc/ef10.c struct efx_rx_queue *rx_queue; rx_queue 3565 drivers/net/ethernet/sfc/ef10.c rx_queue = efx_channel_get_rx_queue(channel); rx_queue 3567 drivers/net/ethernet/sfc/ef10.c if (unlikely(rx_queue_label != efx_rx_queue_index(rx_queue))) rx_queue 3568 drivers/net/ethernet/sfc/ef10.c efx_ef10_handle_rx_wrong_queue(rx_queue, rx_queue_label); rx_queue 3570 drivers/net/ethernet/sfc/ef10.c n_descs = ((next_ptr_lbits - rx_queue->removed_count) & rx_queue 3573 drivers/net/ethernet/sfc/ef10.c if (n_descs != rx_queue->scatter_n + 1) { rx_queue 3577 drivers/net/ethernet/sfc/ef10.c if (unlikely(n_descs == rx_queue->scatter_n)) { rx_queue 3578 drivers/net/ethernet/sfc/ef10.c if (rx_queue->scatter_n == 0 || rx_bytes != 0) rx_queue 3582 drivers/net/ethernet/sfc/ef10.c rx_queue->scatter_n, rx_queue 3584 drivers/net/ethernet/sfc/ef10.c efx_ef10_handle_rx_abort(rx_queue); rx_queue 3594 drivers/net/ethernet/sfc/ef10.c rx_queue->scatter_n != 0 || rx_cont) { rx_queue 3596 drivers/net/ethernet/sfc/ef10.c rx_queue, next_ptr_lbits, rx_queue 3597 drivers/net/ethernet/sfc/ef10.c (rx_queue->removed_count + rx_queue 3598 drivers/net/ethernet/sfc/ef10.c rx_queue->scatter_n + 1) & rx_queue 3604 drivers/net/ethernet/sfc/ef10.c rx_queue->scatter_n = 1; rx_queue 3605 drivers/net/ethernet/sfc/ef10.c rx_queue->scatter_len = 0; rx_queue 3611 drivers/net/ethernet/sfc/ef10.c ++rx_queue->scatter_n; rx_queue 3612 drivers/net/ethernet/sfc/ef10.c rx_queue->scatter_len += rx_bytes; rx_queue 3659 drivers/net/ethernet/sfc/ef10.c efx_rx_packet(rx_queue, rx_queue 3660 drivers/net/ethernet/sfc/ef10.c rx_queue->removed_count & rx_queue->ptr_mask, rx_queue 3661 drivers/net/ethernet/sfc/ef10.c rx_queue->scatter_n, rx_queue->scatter_len, rx_queue 3663 drivers/net/ethernet/sfc/ef10.c rx_queue->removed_count += rx_queue->scatter_n; rx_queue 3666 drivers/net/ethernet/sfc/ef10.c rx_queue->scatter_n = 0; rx_queue 3667 drivers/net/ethernet/sfc/ef10.c rx_queue->scatter_len = 0; rx_queue 3802 drivers/net/ethernet/sfc/ef10.c efx_fast_push_rx_descriptors(&channel->rx_queue, true); rx_queue 3960 drivers/net/ethernet/sfc/ef10.c struct efx_rx_queue *rx_queue; rx_queue 3974 drivers/net/ethernet/sfc/ef10.c efx_for_each_channel_rx_queue(rx_queue, channel) rx_queue 3975 drivers/net/ethernet/sfc/ef10.c efx_ef10_rx_fini(rx_queue); rx_queue 282 drivers/net/ethernet/sfc/efx.c struct efx_rx_queue *rx_queue = rx_queue 286 drivers/net/ethernet/sfc/efx.c efx_fast_push_rx_descriptors(rx_queue, true); rx_queue 464 drivers/net/ethernet/sfc/efx.c struct efx_rx_queue *rx_queue; rx_queue 487 drivers/net/ethernet/sfc/efx.c rx_queue = &channel->rx_queue; rx_queue 488 drivers/net/ethernet/sfc/efx.c rx_queue->efx = efx; rx_queue 489 drivers/net/ethernet/sfc/efx.c timer_setup(&rx_queue->slow_fill, efx_rx_slow_fill, 0); rx_queue 501 drivers/net/ethernet/sfc/efx.c struct efx_rx_queue *rx_queue; rx_queue 526 drivers/net/ethernet/sfc/efx.c rx_queue = &channel->rx_queue; rx_queue 527 drivers/net/ethernet/sfc/efx.c rx_queue->buffer = NULL; rx_queue 528 drivers/net/ethernet/sfc/efx.c memset(&rx_queue->rxd, 0, sizeof(rx_queue->rxd)); rx_queue 529 drivers/net/ethernet/sfc/efx.c timer_setup(&rx_queue->slow_fill, efx_rx_slow_fill, 0); rx_queue 540 drivers/net/ethernet/sfc/efx.c struct efx_rx_queue *rx_queue; rx_queue 560 drivers/net/ethernet/sfc/efx.c efx_for_each_channel_rx_queue(rx_queue, channel) { rx_queue 561 drivers/net/ethernet/sfc/efx.c rc = efx_probe_rx_queue(rx_queue); rx_queue 644 drivers/net/ethernet/sfc/efx.c struct efx_rx_queue *rx_queue; rx_queue 716 drivers/net/ethernet/sfc/efx.c efx_for_each_channel_rx_queue(rx_queue, channel) { rx_queue 717 drivers/net/ethernet/sfc/efx.c efx_init_rx_queue(rx_queue); rx_queue 720 drivers/net/ethernet/sfc/efx.c efx_fast_push_rx_descriptors(rx_queue, false); rx_queue 737 drivers/net/ethernet/sfc/efx.c struct efx_rx_queue *rx_queue; rx_queue 747 drivers/net/ethernet/sfc/efx.c efx_for_each_channel_rx_queue(rx_queue, channel) rx_queue 748 drivers/net/ethernet/sfc/efx.c rx_queue->refill_enabled = false; rx_queue 773 drivers/net/ethernet/sfc/efx.c efx_for_each_channel_rx_queue(rx_queue, channel) rx_queue 774 drivers/net/ethernet/sfc/efx.c efx_fini_rx_queue(rx_queue); rx_queue 783 drivers/net/ethernet/sfc/efx.c struct efx_rx_queue *rx_queue; rx_queue 788 drivers/net/ethernet/sfc/efx.c efx_for_each_channel_rx_queue(rx_queue, channel) rx_queue 789 drivers/net/ethernet/sfc/efx.c efx_remove_rx_queue(rx_queue); rx_queue 820 drivers/net/ethernet/sfc/efx.c struct efx_rx_queue *rx_queue; rx_queue 828 drivers/net/ethernet/sfc/efx.c efx_for_each_channel_rx_queue(rx_queue, channel) rx_queue 830 drivers/net/ethernet/sfc/efx.c rx_queue->rxd.index + rx_queue 831 drivers/net/ethernet/sfc/efx.c rx_queue->rxd.entries); rx_queue 914 drivers/net/ethernet/sfc/efx.c void efx_schedule_slow_fill(struct efx_rx_queue *rx_queue) rx_queue 916 drivers/net/ethernet/sfc/efx.c mod_timer(&rx_queue->slow_fill, jiffies + msecs_to_jiffies(10)); rx_queue 1743 drivers/net/ethernet/sfc/efx.c channel->rx_queue.core_index = channel->channel; rx_queue 1745 drivers/net/ethernet/sfc/efx.c channel->rx_queue.core_index = -1; rx_queue 37 drivers/net/ethernet/sfc/efx.h int efx_probe_rx_queue(struct efx_rx_queue *rx_queue); rx_queue 38 drivers/net/ethernet/sfc/efx.h void efx_remove_rx_queue(struct efx_rx_queue *rx_queue); rx_queue 39 drivers/net/ethernet/sfc/efx.h void efx_init_rx_queue(struct efx_rx_queue *rx_queue); rx_queue 40 drivers/net/ethernet/sfc/efx.h void efx_fini_rx_queue(struct efx_rx_queue *rx_queue); rx_queue 41 drivers/net/ethernet/sfc/efx.h void efx_fast_push_rx_descriptors(struct efx_rx_queue *rx_queue, bool atomic); rx_queue 44 drivers/net/ethernet/sfc/efx.h void efx_rx_packet(struct efx_rx_queue *rx_queue, unsigned int index, rx_queue 51 drivers/net/ethernet/sfc/efx.h void efx_schedule_slow_fill(struct efx_rx_queue *rx_queue); rx_queue 458 drivers/net/ethernet/sfc/ethtool.c struct efx_rx_queue *rx_queue; rx_queue 506 drivers/net/ethernet/sfc/ethtool.c efx_for_each_channel_rx_queue(rx_queue, channel) { rx_queue 507 drivers/net/ethernet/sfc/ethtool.c *data += rx_queue->rx_packets; rx_queue 253 drivers/net/ethernet/sfc/falcon/efx.c struct ef4_rx_queue *rx_queue = rx_queue 257 drivers/net/ethernet/sfc/falcon/efx.c ef4_fast_push_rx_descriptors(rx_queue, true); rx_queue 428 drivers/net/ethernet/sfc/falcon/efx.c struct ef4_rx_queue *rx_queue; rx_queue 447 drivers/net/ethernet/sfc/falcon/efx.c rx_queue = &channel->rx_queue; rx_queue 448 drivers/net/ethernet/sfc/falcon/efx.c rx_queue->efx = efx; rx_queue 449 drivers/net/ethernet/sfc/falcon/efx.c timer_setup(&rx_queue->slow_fill, ef4_rx_slow_fill, 0); rx_queue 461 drivers/net/ethernet/sfc/falcon/efx.c struct ef4_rx_queue *rx_queue; rx_queue 485 drivers/net/ethernet/sfc/falcon/efx.c rx_queue = &channel->rx_queue; rx_queue 486 drivers/net/ethernet/sfc/falcon/efx.c rx_queue->buffer = NULL; rx_queue 487 drivers/net/ethernet/sfc/falcon/efx.c memset(&rx_queue->rxd, 0, sizeof(rx_queue->rxd)); rx_queue 488 drivers/net/ethernet/sfc/falcon/efx.c timer_setup(&rx_queue->slow_fill, ef4_rx_slow_fill, 0); rx_queue 496 drivers/net/ethernet/sfc/falcon/efx.c struct ef4_rx_queue *rx_queue; rx_queue 516 drivers/net/ethernet/sfc/falcon/efx.c ef4_for_each_channel_rx_queue(rx_queue, channel) { rx_queue 517 drivers/net/ethernet/sfc/falcon/efx.c rc = ef4_probe_rx_queue(rx_queue); rx_queue 598 drivers/net/ethernet/sfc/falcon/efx.c struct ef4_rx_queue *rx_queue; rx_queue 670 drivers/net/ethernet/sfc/falcon/efx.c ef4_for_each_channel_rx_queue(rx_queue, channel) { rx_queue 671 drivers/net/ethernet/sfc/falcon/efx.c ef4_init_rx_queue(rx_queue); rx_queue 674 drivers/net/ethernet/sfc/falcon/efx.c ef4_fast_push_rx_descriptors(rx_queue, false); rx_queue 689 drivers/net/ethernet/sfc/falcon/efx.c struct ef4_rx_queue *rx_queue; rx_queue 697 drivers/net/ethernet/sfc/falcon/efx.c ef4_for_each_channel_rx_queue(rx_queue, channel) rx_queue 698 drivers/net/ethernet/sfc/falcon/efx.c rx_queue->refill_enabled = false; rx_queue 732 drivers/net/ethernet/sfc/falcon/efx.c ef4_for_each_channel_rx_queue(rx_queue, channel) rx_queue 733 drivers/net/ethernet/sfc/falcon/efx.c ef4_fini_rx_queue(rx_queue); rx_queue 742 drivers/net/ethernet/sfc/falcon/efx.c struct ef4_rx_queue *rx_queue; rx_queue 747 drivers/net/ethernet/sfc/falcon/efx.c ef4_for_each_channel_rx_queue(rx_queue, channel) rx_queue 748 drivers/net/ethernet/sfc/falcon/efx.c ef4_remove_rx_queue(rx_queue); rx_queue 779 drivers/net/ethernet/sfc/falcon/efx.c struct ef4_rx_queue *rx_queue; rx_queue 787 drivers/net/ethernet/sfc/falcon/efx.c ef4_for_each_channel_rx_queue(rx_queue, channel) rx_queue 789 drivers/net/ethernet/sfc/falcon/efx.c rx_queue->rxd.index + rx_queue 790 drivers/net/ethernet/sfc/falcon/efx.c rx_queue->rxd.entries); rx_queue 873 drivers/net/ethernet/sfc/falcon/efx.c void ef4_schedule_slow_fill(struct ef4_rx_queue *rx_queue) rx_queue 875 drivers/net/ethernet/sfc/falcon/efx.c mod_timer(&rx_queue->slow_fill, jiffies + msecs_to_jiffies(100)); rx_queue 1608 drivers/net/ethernet/sfc/falcon/efx.c channel->rx_queue.core_index = channel->channel; rx_queue 1610 drivers/net/ethernet/sfc/falcon/efx.c channel->rx_queue.core_index = -1; rx_queue 40 drivers/net/ethernet/sfc/falcon/efx.h int ef4_probe_rx_queue(struct ef4_rx_queue *rx_queue); rx_queue 41 drivers/net/ethernet/sfc/falcon/efx.h void ef4_remove_rx_queue(struct ef4_rx_queue *rx_queue); rx_queue 42 drivers/net/ethernet/sfc/falcon/efx.h void ef4_init_rx_queue(struct ef4_rx_queue *rx_queue); rx_queue 43 drivers/net/ethernet/sfc/falcon/efx.h void ef4_fini_rx_queue(struct ef4_rx_queue *rx_queue); rx_queue 44 drivers/net/ethernet/sfc/falcon/efx.h void ef4_fast_push_rx_descriptors(struct ef4_rx_queue *rx_queue, bool atomic); rx_queue 47 drivers/net/ethernet/sfc/falcon/efx.h void ef4_rx_packet(struct ef4_rx_queue *rx_queue, unsigned int index, rx_queue 54 drivers/net/ethernet/sfc/falcon/efx.h void ef4_schedule_slow_fill(struct ef4_rx_queue *rx_queue); rx_queue 438 drivers/net/ethernet/sfc/falcon/ethtool.c struct ef4_rx_queue *rx_queue; rx_queue 486 drivers/net/ethernet/sfc/falcon/ethtool.c ef4_for_each_channel_rx_queue(rx_queue, channel) { rx_queue 487 drivers/net/ethernet/sfc/falcon/ethtool.c *data += rx_queue->rx_packets; rx_queue 471 drivers/net/ethernet/sfc/falcon/farch.c ef4_farch_build_rx_desc(struct ef4_rx_queue *rx_queue, unsigned index) rx_queue 476 drivers/net/ethernet/sfc/falcon/farch.c rxd = ef4_rx_desc(rx_queue, index); rx_queue 477 drivers/net/ethernet/sfc/falcon/farch.c rx_buf = ef4_rx_buffer(rx_queue, index); rx_queue 481 drivers/net/ethernet/sfc/falcon/farch.c rx_queue->efx->type->rx_buffer_padding, rx_queue 489 drivers/net/ethernet/sfc/falcon/farch.c void ef4_farch_rx_write(struct ef4_rx_queue *rx_queue) rx_queue 491 drivers/net/ethernet/sfc/falcon/farch.c struct ef4_nic *efx = rx_queue->efx; rx_queue 495 drivers/net/ethernet/sfc/falcon/farch.c while (rx_queue->notified_count != rx_queue->added_count) { rx_queue 497 drivers/net/ethernet/sfc/falcon/farch.c rx_queue, rx_queue 498 drivers/net/ethernet/sfc/falcon/farch.c rx_queue->notified_count & rx_queue->ptr_mask); rx_queue 499 drivers/net/ethernet/sfc/falcon/farch.c ++rx_queue->notified_count; rx_queue 503 drivers/net/ethernet/sfc/falcon/farch.c write_ptr = rx_queue->added_count & rx_queue->ptr_mask; rx_queue 506 drivers/net/ethernet/sfc/falcon/farch.c ef4_rx_queue_index(rx_queue)); rx_queue 509 drivers/net/ethernet/sfc/falcon/farch.c int ef4_farch_rx_probe(struct ef4_rx_queue *rx_queue) rx_queue 511 drivers/net/ethernet/sfc/falcon/farch.c struct ef4_nic *efx = rx_queue->efx; rx_queue 514 drivers/net/ethernet/sfc/falcon/farch.c entries = rx_queue->ptr_mask + 1; rx_queue 515 drivers/net/ethernet/sfc/falcon/farch.c return ef4_alloc_special_buffer(efx, &rx_queue->rxd, rx_queue 519 drivers/net/ethernet/sfc/falcon/farch.c void ef4_farch_rx_init(struct ef4_rx_queue *rx_queue) rx_queue 522 drivers/net/ethernet/sfc/falcon/farch.c struct ef4_nic *efx = rx_queue->efx; rx_queue 536 drivers/net/ethernet/sfc/falcon/farch.c ef4_rx_queue_index(rx_queue), rx_queue->rxd.index, rx_queue 537 drivers/net/ethernet/sfc/falcon/farch.c rx_queue->rxd.index + rx_queue->rxd.entries - 1); rx_queue 539 drivers/net/ethernet/sfc/falcon/farch.c rx_queue->scatter_n = 0; rx_queue 542 drivers/net/ethernet/sfc/falcon/farch.c ef4_init_special_buffer(efx, &rx_queue->rxd); rx_queue 548 drivers/net/ethernet/sfc/falcon/farch.c FRF_AZ_RX_DESCQ_BUF_BASE_ID, rx_queue->rxd.index, rx_queue 550 drivers/net/ethernet/sfc/falcon/farch.c ef4_rx_queue_channel(rx_queue)->channel, rx_queue 553 drivers/net/ethernet/sfc/falcon/farch.c ef4_rx_queue_index(rx_queue), rx_queue 555 drivers/net/ethernet/sfc/falcon/farch.c __ffs(rx_queue->rxd.entries), rx_queue 560 drivers/net/ethernet/sfc/falcon/farch.c ef4_rx_queue_index(rx_queue)); rx_queue 563 drivers/net/ethernet/sfc/falcon/farch.c static void ef4_farch_flush_rx_queue(struct ef4_rx_queue *rx_queue) rx_queue 565 drivers/net/ethernet/sfc/falcon/farch.c struct ef4_nic *efx = rx_queue->efx; rx_queue 571 drivers/net/ethernet/sfc/falcon/farch.c ef4_rx_queue_index(rx_queue)); rx_queue 575 drivers/net/ethernet/sfc/falcon/farch.c void ef4_farch_rx_fini(struct ef4_rx_queue *rx_queue) rx_queue 578 drivers/net/ethernet/sfc/falcon/farch.c struct ef4_nic *efx = rx_queue->efx; rx_queue 583 drivers/net/ethernet/sfc/falcon/farch.c ef4_rx_queue_index(rx_queue)); rx_queue 586 drivers/net/ethernet/sfc/falcon/farch.c ef4_fini_special_buffer(efx, &rx_queue->rxd); rx_queue 590 drivers/net/ethernet/sfc/falcon/farch.c void ef4_farch_rx_remove(struct ef4_rx_queue *rx_queue) rx_queue 592 drivers/net/ethernet/sfc/falcon/farch.c ef4_free_special_buffer(rx_queue->efx, &rx_queue->rxd); rx_queue 662 drivers/net/ethernet/sfc/falcon/farch.c struct ef4_rx_queue *rx_queue; rx_queue 670 drivers/net/ethernet/sfc/falcon/farch.c ef4_for_each_channel_rx_queue(rx_queue, channel) { rx_queue 671 drivers/net/ethernet/sfc/falcon/farch.c rx_queue->flush_pending = true; rx_queue 682 drivers/net/ethernet/sfc/falcon/farch.c ef4_for_each_channel_rx_queue(rx_queue, channel) { rx_queue 687 drivers/net/ethernet/sfc/falcon/farch.c if (rx_queue->flush_pending) { rx_queue 688 drivers/net/ethernet/sfc/falcon/farch.c rx_queue->flush_pending = false; rx_queue 691 drivers/net/ethernet/sfc/falcon/farch.c ef4_farch_flush_rx_queue(rx_queue); rx_queue 721 drivers/net/ethernet/sfc/falcon/farch.c struct ef4_rx_queue *rx_queue; rx_queue 734 drivers/net/ethernet/sfc/falcon/farch.c ef4_for_each_channel_rx_queue(rx_queue, channel) rx_queue 735 drivers/net/ethernet/sfc/falcon/farch.c ef4_farch_rx_fini(rx_queue); rx_queue 868 drivers/net/ethernet/sfc/falcon/farch.c static u16 ef4_farch_handle_rx_not_ok(struct ef4_rx_queue *rx_queue, rx_queue 871 drivers/net/ethernet/sfc/falcon/farch.c struct ef4_channel *channel = ef4_rx_queue_channel(rx_queue); rx_queue 872 drivers/net/ethernet/sfc/falcon/farch.c struct ef4_nic *efx = rx_queue->efx; rx_queue 923 drivers/net/ethernet/sfc/falcon/farch.c ef4_rx_queue_index(rx_queue), EF4_QWORD_VAL(*event), rx_queue 948 drivers/net/ethernet/sfc/falcon/farch.c ef4_farch_handle_rx_bad_index(struct ef4_rx_queue *rx_queue, unsigned index) rx_queue 950 drivers/net/ethernet/sfc/falcon/farch.c struct ef4_channel *channel = ef4_rx_queue_channel(rx_queue); rx_queue 951 drivers/net/ethernet/sfc/falcon/farch.c struct ef4_nic *efx = rx_queue->efx; rx_queue 954 drivers/net/ethernet/sfc/falcon/farch.c if (rx_queue->scatter_n && rx_queue 955 drivers/net/ethernet/sfc/falcon/farch.c index == ((rx_queue->removed_count + rx_queue->scatter_n - 1) & rx_queue 956 drivers/net/ethernet/sfc/falcon/farch.c rx_queue->ptr_mask)) { rx_queue 961 drivers/net/ethernet/sfc/falcon/farch.c expected = rx_queue->removed_count & rx_queue->ptr_mask; rx_queue 962 drivers/net/ethernet/sfc/falcon/farch.c dropped = (index - expected) & rx_queue->ptr_mask; rx_queue 987 drivers/net/ethernet/sfc/falcon/farch.c struct ef4_rx_queue *rx_queue; rx_queue 998 drivers/net/ethernet/sfc/falcon/farch.c rx_queue = ef4_channel_get_rx_queue(channel); rx_queue 1001 drivers/net/ethernet/sfc/falcon/farch.c expected_ptr = ((rx_queue->removed_count + rx_queue->scatter_n) & rx_queue 1002 drivers/net/ethernet/sfc/falcon/farch.c rx_queue->ptr_mask); rx_queue 1006 drivers/net/ethernet/sfc/falcon/farch.c unlikely(rx_ev_sop != (rx_queue->scatter_n == 0))) { rx_queue 1008 drivers/net/ethernet/sfc/falcon/farch.c !ef4_farch_handle_rx_bad_index(rx_queue, rx_ev_desc_ptr)) rx_queue 1012 drivers/net/ethernet/sfc/falcon/farch.c if (rx_queue->scatter_n) { rx_queue 1014 drivers/net/ethernet/sfc/falcon/farch.c rx_queue, rx_queue 1015 drivers/net/ethernet/sfc/falcon/farch.c rx_queue->removed_count & rx_queue->ptr_mask, rx_queue 1016 drivers/net/ethernet/sfc/falcon/farch.c rx_queue->scatter_n, 0, EF4_RX_PKT_DISCARD); rx_queue 1017 drivers/net/ethernet/sfc/falcon/farch.c rx_queue->removed_count += rx_queue->scatter_n; rx_queue 1018 drivers/net/ethernet/sfc/falcon/farch.c rx_queue->scatter_n = 0; rx_queue 1028 drivers/net/ethernet/sfc/falcon/farch.c rx_queue, rx_queue 1029 drivers/net/ethernet/sfc/falcon/farch.c rx_queue->removed_count & rx_queue->ptr_mask, rx_queue 1031 drivers/net/ethernet/sfc/falcon/farch.c ++rx_queue->removed_count; rx_queue 1036 drivers/net/ethernet/sfc/falcon/farch.c ++rx_queue->scatter_n; rx_queue 1061 drivers/net/ethernet/sfc/falcon/farch.c flags = ef4_farch_handle_rx_not_ok(rx_queue, event); rx_queue 1079 drivers/net/ethernet/sfc/falcon/farch.c ef4_rx_packet(rx_queue, rx_queue 1080 drivers/net/ethernet/sfc/falcon/farch.c rx_queue->removed_count & rx_queue->ptr_mask, rx_queue 1081 drivers/net/ethernet/sfc/falcon/farch.c rx_queue->scatter_n, rx_ev_byte_cnt, flags); rx_queue 1082 drivers/net/ethernet/sfc/falcon/farch.c rx_queue->removed_count += rx_queue->scatter_n; rx_queue 1083 drivers/net/ethernet/sfc/falcon/farch.c rx_queue->scatter_n = 0; rx_queue 1115 drivers/net/ethernet/sfc/falcon/farch.c struct ef4_rx_queue *rx_queue; rx_queue 1126 drivers/net/ethernet/sfc/falcon/farch.c rx_queue = ef4_channel_get_rx_queue(channel); rx_queue 1131 drivers/net/ethernet/sfc/falcon/farch.c rx_queue->flush_pending = true; rx_queue 1134 drivers/net/ethernet/sfc/falcon/farch.c ef4_farch_magic_event(ef4_rx_queue_channel(rx_queue), rx_queue 1135 drivers/net/ethernet/sfc/falcon/farch.c EF4_CHANNEL_MAGIC_RX_DRAIN(rx_queue)); rx_queue 1157 drivers/net/ethernet/sfc/falcon/farch.c struct ef4_rx_queue *rx_queue = rx_queue 1167 drivers/net/ethernet/sfc/falcon/farch.c } else if (rx_queue && magic == EF4_CHANNEL_MAGIC_FILL(rx_queue)) { rx_queue 1171 drivers/net/ethernet/sfc/falcon/farch.c ef4_fast_push_rx_descriptors(rx_queue, true); rx_queue 1172 drivers/net/ethernet/sfc/falcon/farch.c } else if (rx_queue && magic == EF4_CHANNEL_MAGIC_RX_DRAIN(rx_queue)) { rx_queue 1391 drivers/net/ethernet/sfc/falcon/farch.c void ef4_farch_rx_defer_refill(struct ef4_rx_queue *rx_queue) rx_queue 1393 drivers/net/ethernet/sfc/falcon/farch.c ef4_farch_magic_event(ef4_rx_queue_channel(rx_queue), rx_queue 1394 drivers/net/ethernet/sfc/falcon/farch.c EF4_CHANNEL_MAGIC_FILL(rx_queue)); rx_queue 444 drivers/net/ethernet/sfc/falcon/net_driver.h struct ef4_rx_queue rx_queue; rx_queue 1090 drivers/net/ethernet/sfc/falcon/net_driver.h int (*rx_probe)(struct ef4_rx_queue *rx_queue); rx_queue 1091 drivers/net/ethernet/sfc/falcon/net_driver.h void (*rx_init)(struct ef4_rx_queue *rx_queue); rx_queue 1092 drivers/net/ethernet/sfc/falcon/net_driver.h void (*rx_remove)(struct ef4_rx_queue *rx_queue); rx_queue 1093 drivers/net/ethernet/sfc/falcon/net_driver.h void (*rx_write)(struct ef4_rx_queue *rx_queue); rx_queue 1094 drivers/net/ethernet/sfc/falcon/net_driver.h void (*rx_defer_refill)(struct ef4_rx_queue *rx_queue); rx_queue 1236 drivers/net/ethernet/sfc/falcon/net_driver.h return channel->rx_queue.core_index >= 0; rx_queue 1243 drivers/net/ethernet/sfc/falcon/net_driver.h return &channel->rx_queue; rx_queue 1251 drivers/net/ethernet/sfc/falcon/net_driver.h for (_rx_queue = &(_channel)->rx_queue; \ rx_queue 1256 drivers/net/ethernet/sfc/falcon/net_driver.h ef4_rx_queue_channel(struct ef4_rx_queue *rx_queue) rx_queue 1258 drivers/net/ethernet/sfc/falcon/net_driver.h return container_of(rx_queue, struct ef4_channel, rx_queue); rx_queue 1261 drivers/net/ethernet/sfc/falcon/net_driver.h static inline int ef4_rx_queue_index(struct ef4_rx_queue *rx_queue) rx_queue 1263 drivers/net/ethernet/sfc/falcon/net_driver.h return ef4_rx_queue_channel(rx_queue)->channel; rx_queue 1269 drivers/net/ethernet/sfc/falcon/net_driver.h static inline struct ef4_rx_buffer *ef4_rx_buffer(struct ef4_rx_queue *rx_queue, rx_queue 1272 drivers/net/ethernet/sfc/falcon/net_driver.h return &rx_queue->buffer[index]; rx_queue 110 drivers/net/ethernet/sfc/falcon/nic.h ef4_rx_desc(struct ef4_rx_queue *rx_queue, unsigned int index) rx_queue 112 drivers/net/ethernet/sfc/falcon/nic.h return ((ef4_qword_t *) (rx_queue->rxd.buf.addr)) + index; rx_queue 333 drivers/net/ethernet/sfc/falcon/nic.h static inline int ef4_nic_probe_rx(struct ef4_rx_queue *rx_queue) rx_queue 335 drivers/net/ethernet/sfc/falcon/nic.h return rx_queue->efx->type->rx_probe(rx_queue); rx_queue 337 drivers/net/ethernet/sfc/falcon/nic.h static inline void ef4_nic_init_rx(struct ef4_rx_queue *rx_queue) rx_queue 339 drivers/net/ethernet/sfc/falcon/nic.h rx_queue->efx->type->rx_init(rx_queue); rx_queue 341 drivers/net/ethernet/sfc/falcon/nic.h static inline void ef4_nic_remove_rx(struct ef4_rx_queue *rx_queue) rx_queue 343 drivers/net/ethernet/sfc/falcon/nic.h rx_queue->efx->type->rx_remove(rx_queue); rx_queue 345 drivers/net/ethernet/sfc/falcon/nic.h static inline void ef4_nic_notify_rx_desc(struct ef4_rx_queue *rx_queue) rx_queue 347 drivers/net/ethernet/sfc/falcon/nic.h rx_queue->efx->type->rx_write(rx_queue); rx_queue 349 drivers/net/ethernet/sfc/falcon/nic.h static inline void ef4_nic_generate_fill_event(struct ef4_rx_queue *rx_queue) rx_queue 351 drivers/net/ethernet/sfc/falcon/nic.h rx_queue->efx->type->rx_defer_refill(rx_queue); rx_queue 390 drivers/net/ethernet/sfc/falcon/nic.h int ef4_farch_rx_probe(struct ef4_rx_queue *rx_queue); rx_queue 391 drivers/net/ethernet/sfc/falcon/nic.h void ef4_farch_rx_init(struct ef4_rx_queue *rx_queue); rx_queue 392 drivers/net/ethernet/sfc/falcon/nic.h void ef4_farch_rx_fini(struct ef4_rx_queue *rx_queue); rx_queue 393 drivers/net/ethernet/sfc/falcon/nic.h void ef4_farch_rx_remove(struct ef4_rx_queue *rx_queue); rx_queue 394 drivers/net/ethernet/sfc/falcon/nic.h void ef4_farch_rx_write(struct ef4_rx_queue *rx_queue); rx_queue 395 drivers/net/ethernet/sfc/falcon/nic.h void ef4_farch_rx_defer_refill(struct ef4_rx_queue *rx_queue); rx_queue 76 drivers/net/ethernet/sfc/falcon/rx.c ef4_rx_buf_next(struct ef4_rx_queue *rx_queue, struct ef4_rx_buffer *rx_buf) rx_queue 78 drivers/net/ethernet/sfc/falcon/rx.c if (unlikely(rx_buf == ef4_rx_buffer(rx_queue, rx_queue->ptr_mask))) rx_queue 79 drivers/net/ethernet/sfc/falcon/rx.c return ef4_rx_buffer(rx_queue, 0); rx_queue 106 drivers/net/ethernet/sfc/falcon/rx.c static struct page *ef4_reuse_page(struct ef4_rx_queue *rx_queue) rx_queue 108 drivers/net/ethernet/sfc/falcon/rx.c struct ef4_nic *efx = rx_queue->efx; rx_queue 113 drivers/net/ethernet/sfc/falcon/rx.c index = rx_queue->page_remove & rx_queue->page_ptr_mask; rx_queue 114 drivers/net/ethernet/sfc/falcon/rx.c page = rx_queue->page_ring[index]; rx_queue 118 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->page_ring[index] = NULL; rx_queue 120 drivers/net/ethernet/sfc/falcon/rx.c if (rx_queue->page_remove != rx_queue->page_add) rx_queue 121 drivers/net/ethernet/sfc/falcon/rx.c ++rx_queue->page_remove; rx_queue 125 drivers/net/ethernet/sfc/falcon/rx.c ++rx_queue->page_recycle_count; rx_queue 133 drivers/net/ethernet/sfc/falcon/rx.c ++rx_queue->page_recycle_failed; rx_queue 149 drivers/net/ethernet/sfc/falcon/rx.c static int ef4_init_rx_buffers(struct ef4_rx_queue *rx_queue, bool atomic) rx_queue 151 drivers/net/ethernet/sfc/falcon/rx.c struct ef4_nic *efx = rx_queue->efx; rx_queue 161 drivers/net/ethernet/sfc/falcon/rx.c page = ef4_reuse_page(rx_queue); rx_queue 188 drivers/net/ethernet/sfc/falcon/rx.c index = rx_queue->added_count & rx_queue->ptr_mask; rx_queue 189 drivers/net/ethernet/sfc/falcon/rx.c rx_buf = ef4_rx_buffer(rx_queue, index); rx_queue 195 drivers/net/ethernet/sfc/falcon/rx.c ++rx_queue->added_count; rx_queue 224 drivers/net/ethernet/sfc/falcon/rx.c static void ef4_free_rx_buffers(struct ef4_rx_queue *rx_queue, rx_queue 233 drivers/net/ethernet/sfc/falcon/rx.c rx_buf = ef4_rx_buf_next(rx_queue, rx_buf); rx_queue 245 drivers/net/ethernet/sfc/falcon/rx.c struct ef4_rx_queue *rx_queue = ef4_channel_get_rx_queue(channel); rx_queue 246 drivers/net/ethernet/sfc/falcon/rx.c struct ef4_nic *efx = rx_queue->efx; rx_queue 253 drivers/net/ethernet/sfc/falcon/rx.c index = rx_queue->page_add & rx_queue->page_ptr_mask; rx_queue 254 drivers/net/ethernet/sfc/falcon/rx.c if (rx_queue->page_ring[index] == NULL) { rx_queue 255 drivers/net/ethernet/sfc/falcon/rx.c unsigned read_index = rx_queue->page_remove & rx_queue 256 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->page_ptr_mask; rx_queue 263 drivers/net/ethernet/sfc/falcon/rx.c ++rx_queue->page_remove; rx_queue 264 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->page_ring[index] = page; rx_queue 265 drivers/net/ethernet/sfc/falcon/rx.c ++rx_queue->page_add; rx_queue 268 drivers/net/ethernet/sfc/falcon/rx.c ++rx_queue->page_recycle_full; rx_queue 273 drivers/net/ethernet/sfc/falcon/rx.c static void ef4_fini_rx_buffer(struct ef4_rx_queue *rx_queue, rx_queue 282 drivers/net/ethernet/sfc/falcon/rx.c ef4_unmap_rx_buffer(rx_queue->efx, rx_buf); rx_queue 283 drivers/net/ethernet/sfc/falcon/rx.c ef4_free_rx_buffers(rx_queue, rx_buf, 1); rx_queue 293 drivers/net/ethernet/sfc/falcon/rx.c struct ef4_rx_queue *rx_queue = ef4_channel_get_rx_queue(channel); rx_queue 297 drivers/net/ethernet/sfc/falcon/rx.c rx_buf = ef4_rx_buf_next(rx_queue, rx_buf); rx_queue 305 drivers/net/ethernet/sfc/falcon/rx.c struct ef4_rx_queue *rx_queue = ef4_channel_get_rx_queue(channel); rx_queue 309 drivers/net/ethernet/sfc/falcon/rx.c ef4_free_rx_buffers(rx_queue, rx_buf, n_frags); rx_queue 324 drivers/net/ethernet/sfc/falcon/rx.c void ef4_fast_push_rx_descriptors(struct ef4_rx_queue *rx_queue, bool atomic) rx_queue 326 drivers/net/ethernet/sfc/falcon/rx.c struct ef4_nic *efx = rx_queue->efx; rx_queue 330 drivers/net/ethernet/sfc/falcon/rx.c if (!rx_queue->refill_enabled) rx_queue 334 drivers/net/ethernet/sfc/falcon/rx.c fill_level = (rx_queue->added_count - rx_queue->removed_count); rx_queue 335 drivers/net/ethernet/sfc/falcon/rx.c EF4_BUG_ON_PARANOID(fill_level > rx_queue->efx->rxq_entries); rx_queue 336 drivers/net/ethernet/sfc/falcon/rx.c if (fill_level >= rx_queue->fast_fill_trigger) rx_queue 340 drivers/net/ethernet/sfc/falcon/rx.c if (unlikely(fill_level < rx_queue->min_fill)) { rx_queue 342 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->min_fill = fill_level; rx_queue 346 drivers/net/ethernet/sfc/falcon/rx.c space = rx_queue->max_fill - fill_level; rx_queue 349 drivers/net/ethernet/sfc/falcon/rx.c netif_vdbg(rx_queue->efx, rx_status, rx_queue->efx->net_dev, rx_queue 352 drivers/net/ethernet/sfc/falcon/rx.c ef4_rx_queue_index(rx_queue), fill_level, rx_queue 353 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->max_fill); rx_queue 357 drivers/net/ethernet/sfc/falcon/rx.c rc = ef4_init_rx_buffers(rx_queue, atomic); rx_queue 360 drivers/net/ethernet/sfc/falcon/rx.c if (rx_queue->added_count == rx_queue->removed_count) rx_queue 361 drivers/net/ethernet/sfc/falcon/rx.c ef4_schedule_slow_fill(rx_queue); rx_queue 366 drivers/net/ethernet/sfc/falcon/rx.c netif_vdbg(rx_queue->efx, rx_status, rx_queue->efx->net_dev, rx_queue 368 drivers/net/ethernet/sfc/falcon/rx.c "to level %d\n", ef4_rx_queue_index(rx_queue), rx_queue 369 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->added_count - rx_queue->removed_count); rx_queue 372 drivers/net/ethernet/sfc/falcon/rx.c if (rx_queue->notified_count != rx_queue->added_count) rx_queue 373 drivers/net/ethernet/sfc/falcon/rx.c ef4_nic_notify_rx_desc(rx_queue); rx_queue 378 drivers/net/ethernet/sfc/falcon/rx.c struct ef4_rx_queue *rx_queue = from_timer(rx_queue, t, slow_fill); rx_queue 381 drivers/net/ethernet/sfc/falcon/rx.c ef4_nic_generate_fill_event(rx_queue); rx_queue 382 drivers/net/ethernet/sfc/falcon/rx.c ++rx_queue->slow_fill_count; rx_queue 385 drivers/net/ethernet/sfc/falcon/rx.c static void ef4_rx_packet__check_len(struct ef4_rx_queue *rx_queue, rx_queue 389 drivers/net/ethernet/sfc/falcon/rx.c struct ef4_nic *efx = rx_queue->efx; rx_queue 405 drivers/net/ethernet/sfc/falcon/rx.c ef4_rx_queue_index(rx_queue), len, max_len, rx_queue 413 drivers/net/ethernet/sfc/falcon/rx.c ef4_rx_queue_index(rx_queue), len, max_len); rx_queue 416 drivers/net/ethernet/sfc/falcon/rx.c ef4_rx_queue_channel(rx_queue)->n_rx_overlength++; rx_queue 432 drivers/net/ethernet/sfc/falcon/rx.c struct ef4_rx_queue *rx_queue; rx_queue 434 drivers/net/ethernet/sfc/falcon/rx.c rx_queue = ef4_channel_get_rx_queue(channel); rx_queue 435 drivers/net/ethernet/sfc/falcon/rx.c ef4_free_rx_buffers(rx_queue, rx_buf, n_frags); rx_queue 454 drivers/net/ethernet/sfc/falcon/rx.c rx_buf = ef4_rx_buf_next(&channel->rx_queue, rx_buf); rx_queue 460 drivers/net/ethernet/sfc/falcon/rx.c skb_record_rx_queue(skb, channel->rx_queue.core_index); rx_queue 505 drivers/net/ethernet/sfc/falcon/rx.c rx_buf = ef4_rx_buf_next(&channel->rx_queue, rx_buf); rx_queue 523 drivers/net/ethernet/sfc/falcon/rx.c void ef4_rx_packet(struct ef4_rx_queue *rx_queue, unsigned int index, rx_queue 526 drivers/net/ethernet/sfc/falcon/rx.c struct ef4_nic *efx = rx_queue->efx; rx_queue 527 drivers/net/ethernet/sfc/falcon/rx.c struct ef4_channel *channel = ef4_rx_queue_channel(rx_queue); rx_queue 530 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->rx_packets++; rx_queue 532 drivers/net/ethernet/sfc/falcon/rx.c rx_buf = ef4_rx_buffer(rx_queue, index); rx_queue 538 drivers/net/ethernet/sfc/falcon/rx.c ef4_rx_packet__check_len(rx_queue, rx_buf, len); rx_queue 552 drivers/net/ethernet/sfc/falcon/rx.c ef4_rx_queue_index(rx_queue), index, rx_queue 553 drivers/net/ethernet/sfc/falcon/rx.c (index + n_frags - 1) & rx_queue->ptr_mask, len, rx_queue 589 drivers/net/ethernet/sfc/falcon/rx.c rx_buf = ef4_rx_buf_next(rx_queue, rx_buf); rx_queue 599 drivers/net/ethernet/sfc/falcon/rx.c rx_buf = ef4_rx_buffer(rx_queue, index); rx_queue 619 drivers/net/ethernet/sfc/falcon/rx.c struct ef4_rx_queue *rx_queue; rx_queue 621 drivers/net/ethernet/sfc/falcon/rx.c rx_queue = ef4_channel_get_rx_queue(channel); rx_queue 622 drivers/net/ethernet/sfc/falcon/rx.c ef4_free_rx_buffers(rx_queue, rx_buf, n_frags); rx_queue 625 drivers/net/ethernet/sfc/falcon/rx.c skb_record_rx_queue(skb, channel->rx_queue.core_index); rx_queue 645 drivers/net/ethernet/sfc/falcon/rx.c ef4_rx_buffer(&channel->rx_queue, channel->rx_pkt_index); rx_queue 659 drivers/net/ethernet/sfc/falcon/rx.c struct ef4_rx_queue *rx_queue; rx_queue 662 drivers/net/ethernet/sfc/falcon/rx.c rx_queue = ef4_channel_get_rx_queue(channel); rx_queue 663 drivers/net/ethernet/sfc/falcon/rx.c ef4_free_rx_buffers(rx_queue, rx_buf, rx_queue 679 drivers/net/ethernet/sfc/falcon/rx.c int ef4_probe_rx_queue(struct ef4_rx_queue *rx_queue) rx_queue 681 drivers/net/ethernet/sfc/falcon/rx.c struct ef4_nic *efx = rx_queue->efx; rx_queue 688 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->ptr_mask = entries - 1; rx_queue 692 drivers/net/ethernet/sfc/falcon/rx.c ef4_rx_queue_index(rx_queue), efx->rxq_entries, rx_queue 693 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->ptr_mask); rx_queue 696 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->buffer = kcalloc(entries, sizeof(*rx_queue->buffer), rx_queue 698 drivers/net/ethernet/sfc/falcon/rx.c if (!rx_queue->buffer) rx_queue 701 drivers/net/ethernet/sfc/falcon/rx.c rc = ef4_nic_probe_rx(rx_queue); rx_queue 703 drivers/net/ethernet/sfc/falcon/rx.c kfree(rx_queue->buffer); rx_queue 704 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->buffer = NULL; rx_queue 711 drivers/net/ethernet/sfc/falcon/rx.c struct ef4_rx_queue *rx_queue) rx_queue 727 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->page_ring = kcalloc(page_ring_size, rx_queue 728 drivers/net/ethernet/sfc/falcon/rx.c sizeof(*rx_queue->page_ring), GFP_KERNEL); rx_queue 729 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->page_ptr_mask = page_ring_size - 1; rx_queue 732 drivers/net/ethernet/sfc/falcon/rx.c void ef4_init_rx_queue(struct ef4_rx_queue *rx_queue) rx_queue 734 drivers/net/ethernet/sfc/falcon/rx.c struct ef4_nic *efx = rx_queue->efx; rx_queue 737 drivers/net/ethernet/sfc/falcon/rx.c netif_dbg(rx_queue->efx, drv, rx_queue->efx->net_dev, rx_queue 738 drivers/net/ethernet/sfc/falcon/rx.c "initialising RX queue %d\n", ef4_rx_queue_index(rx_queue)); rx_queue 741 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->added_count = 0; rx_queue 742 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->notified_count = 0; rx_queue 743 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->removed_count = 0; rx_queue 744 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->min_fill = -1U; rx_queue 745 drivers/net/ethernet/sfc/falcon/rx.c ef4_init_rx_recycle_ring(efx, rx_queue); rx_queue 747 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->page_remove = 0; rx_queue 748 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->page_add = rx_queue->page_ptr_mask + 1; rx_queue 749 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->page_recycle_count = 0; rx_queue 750 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->page_recycle_failed = 0; rx_queue 751 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->page_recycle_full = 0; rx_queue 765 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->max_fill = max_fill; rx_queue 766 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->fast_fill_trigger = trigger; rx_queue 767 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->refill_enabled = true; rx_queue 770 drivers/net/ethernet/sfc/falcon/rx.c ef4_nic_init_rx(rx_queue); rx_queue 773 drivers/net/ethernet/sfc/falcon/rx.c void ef4_fini_rx_queue(struct ef4_rx_queue *rx_queue) rx_queue 776 drivers/net/ethernet/sfc/falcon/rx.c struct ef4_nic *efx = rx_queue->efx; rx_queue 779 drivers/net/ethernet/sfc/falcon/rx.c netif_dbg(rx_queue->efx, drv, rx_queue->efx->net_dev, rx_queue 780 drivers/net/ethernet/sfc/falcon/rx.c "shutting down RX queue %d\n", ef4_rx_queue_index(rx_queue)); rx_queue 782 drivers/net/ethernet/sfc/falcon/rx.c del_timer_sync(&rx_queue->slow_fill); rx_queue 785 drivers/net/ethernet/sfc/falcon/rx.c if (rx_queue->buffer) { rx_queue 786 drivers/net/ethernet/sfc/falcon/rx.c for (i = rx_queue->removed_count; i < rx_queue->added_count; rx_queue 788 drivers/net/ethernet/sfc/falcon/rx.c unsigned index = i & rx_queue->ptr_mask; rx_queue 789 drivers/net/ethernet/sfc/falcon/rx.c rx_buf = ef4_rx_buffer(rx_queue, index); rx_queue 790 drivers/net/ethernet/sfc/falcon/rx.c ef4_fini_rx_buffer(rx_queue, rx_buf); rx_queue 795 drivers/net/ethernet/sfc/falcon/rx.c for (i = 0; i <= rx_queue->page_ptr_mask; i++) { rx_queue 796 drivers/net/ethernet/sfc/falcon/rx.c struct page *page = rx_queue->page_ring[i]; rx_queue 808 drivers/net/ethernet/sfc/falcon/rx.c kfree(rx_queue->page_ring); rx_queue 809 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->page_ring = NULL; rx_queue 812 drivers/net/ethernet/sfc/falcon/rx.c void ef4_remove_rx_queue(struct ef4_rx_queue *rx_queue) rx_queue 814 drivers/net/ethernet/sfc/falcon/rx.c netif_dbg(rx_queue->efx, drv, rx_queue->efx->net_dev, rx_queue 815 drivers/net/ethernet/sfc/falcon/rx.c "destroying RX queue %d\n", ef4_rx_queue_index(rx_queue)); rx_queue 817 drivers/net/ethernet/sfc/falcon/rx.c ef4_nic_remove_rx(rx_queue); rx_queue 819 drivers/net/ethernet/sfc/falcon/rx.c kfree(rx_queue->buffer); rx_queue 820 drivers/net/ethernet/sfc/falcon/rx.c rx_queue->buffer = NULL; rx_queue 459 drivers/net/ethernet/sfc/farch.c efx_farch_build_rx_desc(struct efx_rx_queue *rx_queue, unsigned index) rx_queue 464 drivers/net/ethernet/sfc/farch.c rxd = efx_rx_desc(rx_queue, index); rx_queue 465 drivers/net/ethernet/sfc/farch.c rx_buf = efx_rx_buffer(rx_queue, index); rx_queue 469 drivers/net/ethernet/sfc/farch.c rx_queue->efx->type->rx_buffer_padding, rx_queue 477 drivers/net/ethernet/sfc/farch.c void efx_farch_rx_write(struct efx_rx_queue *rx_queue) rx_queue 479 drivers/net/ethernet/sfc/farch.c struct efx_nic *efx = rx_queue->efx; rx_queue 483 drivers/net/ethernet/sfc/farch.c while (rx_queue->notified_count != rx_queue->added_count) { rx_queue 485 drivers/net/ethernet/sfc/farch.c rx_queue, rx_queue 486 drivers/net/ethernet/sfc/farch.c rx_queue->notified_count & rx_queue->ptr_mask); rx_queue 487 drivers/net/ethernet/sfc/farch.c ++rx_queue->notified_count; rx_queue 491 drivers/net/ethernet/sfc/farch.c write_ptr = rx_queue->added_count & rx_queue->ptr_mask; rx_queue 494 drivers/net/ethernet/sfc/farch.c efx_rx_queue_index(rx_queue)); rx_queue 497 drivers/net/ethernet/sfc/farch.c int efx_farch_rx_probe(struct efx_rx_queue *rx_queue) rx_queue 499 drivers/net/ethernet/sfc/farch.c struct efx_nic *efx = rx_queue->efx; rx_queue 502 drivers/net/ethernet/sfc/farch.c entries = rx_queue->ptr_mask + 1; rx_queue 503 drivers/net/ethernet/sfc/farch.c return efx_alloc_special_buffer(efx, &rx_queue->rxd, rx_queue 507 drivers/net/ethernet/sfc/farch.c void efx_farch_rx_init(struct efx_rx_queue *rx_queue) rx_queue 510 drivers/net/ethernet/sfc/farch.c struct efx_nic *efx = rx_queue->efx; rx_queue 518 drivers/net/ethernet/sfc/farch.c efx_rx_queue_index(rx_queue), rx_queue->rxd.index, rx_queue 519 drivers/net/ethernet/sfc/farch.c rx_queue->rxd.index + rx_queue->rxd.entries - 1); rx_queue 521 drivers/net/ethernet/sfc/farch.c rx_queue->scatter_n = 0; rx_queue 524 drivers/net/ethernet/sfc/farch.c efx_init_special_buffer(efx, &rx_queue->rxd); rx_queue 530 drivers/net/ethernet/sfc/farch.c FRF_AZ_RX_DESCQ_BUF_BASE_ID, rx_queue->rxd.index, rx_queue 532 drivers/net/ethernet/sfc/farch.c efx_rx_queue_channel(rx_queue)->channel, rx_queue 535 drivers/net/ethernet/sfc/farch.c efx_rx_queue_index(rx_queue), rx_queue 537 drivers/net/ethernet/sfc/farch.c __ffs(rx_queue->rxd.entries), rx_queue 542 drivers/net/ethernet/sfc/farch.c efx_rx_queue_index(rx_queue)); rx_queue 545 drivers/net/ethernet/sfc/farch.c static void efx_farch_flush_rx_queue(struct efx_rx_queue *rx_queue) rx_queue 547 drivers/net/ethernet/sfc/farch.c struct efx_nic *efx = rx_queue->efx; rx_queue 553 drivers/net/ethernet/sfc/farch.c efx_rx_queue_index(rx_queue)); rx_queue 557 drivers/net/ethernet/sfc/farch.c void efx_farch_rx_fini(struct efx_rx_queue *rx_queue) rx_queue 560 drivers/net/ethernet/sfc/farch.c struct efx_nic *efx = rx_queue->efx; rx_queue 565 drivers/net/ethernet/sfc/farch.c efx_rx_queue_index(rx_queue)); rx_queue 568 drivers/net/ethernet/sfc/farch.c efx_fini_special_buffer(efx, &rx_queue->rxd); rx_queue 572 drivers/net/ethernet/sfc/farch.c void efx_farch_rx_remove(struct efx_rx_queue *rx_queue) rx_queue 574 drivers/net/ethernet/sfc/farch.c efx_free_special_buffer(rx_queue->efx, &rx_queue->rxd); rx_queue 644 drivers/net/ethernet/sfc/farch.c struct efx_rx_queue *rx_queue; rx_queue 652 drivers/net/ethernet/sfc/farch.c efx_for_each_channel_rx_queue(rx_queue, channel) { rx_queue 653 drivers/net/ethernet/sfc/farch.c rx_queue->flush_pending = true; rx_queue 674 drivers/net/ethernet/sfc/farch.c efx_for_each_channel_rx_queue(rx_queue, channel) { rx_queue 679 drivers/net/ethernet/sfc/farch.c if (rx_queue->flush_pending) { rx_queue 680 drivers/net/ethernet/sfc/farch.c rx_queue->flush_pending = false; rx_queue 683 drivers/net/ethernet/sfc/farch.c efx_farch_flush_rx_queue(rx_queue); rx_queue 714 drivers/net/ethernet/sfc/farch.c struct efx_rx_queue *rx_queue; rx_queue 727 drivers/net/ethernet/sfc/farch.c efx_for_each_channel_rx_queue(rx_queue, channel) rx_queue 728 drivers/net/ethernet/sfc/farch.c efx_farch_rx_fini(rx_queue); rx_queue 856 drivers/net/ethernet/sfc/farch.c static u16 efx_farch_handle_rx_not_ok(struct efx_rx_queue *rx_queue, rx_queue 859 drivers/net/ethernet/sfc/farch.c struct efx_channel *channel = efx_rx_queue_channel(rx_queue); rx_queue 860 drivers/net/ethernet/sfc/farch.c struct efx_nic *efx = rx_queue->efx; rx_queue 909 drivers/net/ethernet/sfc/farch.c efx_rx_queue_index(rx_queue), EFX_QWORD_VAL(*event), rx_queue 937 drivers/net/ethernet/sfc/farch.c efx_farch_handle_rx_bad_index(struct efx_rx_queue *rx_queue, unsigned index) rx_queue 939 drivers/net/ethernet/sfc/farch.c struct efx_channel *channel = efx_rx_queue_channel(rx_queue); rx_queue 940 drivers/net/ethernet/sfc/farch.c struct efx_nic *efx = rx_queue->efx; rx_queue 943 drivers/net/ethernet/sfc/farch.c if (rx_queue->scatter_n && rx_queue 944 drivers/net/ethernet/sfc/farch.c index == ((rx_queue->removed_count + rx_queue->scatter_n - 1) & rx_queue 945 drivers/net/ethernet/sfc/farch.c rx_queue->ptr_mask)) { rx_queue 950 drivers/net/ethernet/sfc/farch.c expected = rx_queue->removed_count & rx_queue->ptr_mask; rx_queue 951 drivers/net/ethernet/sfc/farch.c dropped = (index - expected) & rx_queue->ptr_mask; rx_queue 975 drivers/net/ethernet/sfc/farch.c struct efx_rx_queue *rx_queue; rx_queue 986 drivers/net/ethernet/sfc/farch.c rx_queue = efx_channel_get_rx_queue(channel); rx_queue 989 drivers/net/ethernet/sfc/farch.c expected_ptr = ((rx_queue->removed_count + rx_queue->scatter_n) & rx_queue 990 drivers/net/ethernet/sfc/farch.c rx_queue->ptr_mask); rx_queue 994 drivers/net/ethernet/sfc/farch.c unlikely(rx_ev_sop != (rx_queue->scatter_n == 0))) { rx_queue 996 drivers/net/ethernet/sfc/farch.c !efx_farch_handle_rx_bad_index(rx_queue, rx_ev_desc_ptr)) rx_queue 1000 drivers/net/ethernet/sfc/farch.c if (rx_queue->scatter_n) { rx_queue 1002 drivers/net/ethernet/sfc/farch.c rx_queue, rx_queue 1003 drivers/net/ethernet/sfc/farch.c rx_queue->removed_count & rx_queue->ptr_mask, rx_queue 1004 drivers/net/ethernet/sfc/farch.c rx_queue->scatter_n, 0, EFX_RX_PKT_DISCARD); rx_queue 1005 drivers/net/ethernet/sfc/farch.c rx_queue->removed_count += rx_queue->scatter_n; rx_queue 1006 drivers/net/ethernet/sfc/farch.c rx_queue->scatter_n = 0; rx_queue 1016 drivers/net/ethernet/sfc/farch.c rx_queue, rx_queue 1017 drivers/net/ethernet/sfc/farch.c rx_queue->removed_count & rx_queue->ptr_mask, rx_queue 1019 drivers/net/ethernet/sfc/farch.c ++rx_queue->removed_count; rx_queue 1024 drivers/net/ethernet/sfc/farch.c ++rx_queue->scatter_n; rx_queue 1049 drivers/net/ethernet/sfc/farch.c flags = efx_farch_handle_rx_not_ok(rx_queue, event); rx_queue 1067 drivers/net/ethernet/sfc/farch.c efx_rx_packet(rx_queue, rx_queue 1068 drivers/net/ethernet/sfc/farch.c rx_queue->removed_count & rx_queue->ptr_mask, rx_queue 1069 drivers/net/ethernet/sfc/farch.c rx_queue->scatter_n, rx_ev_byte_cnt, flags); rx_queue 1070 drivers/net/ethernet/sfc/farch.c rx_queue->removed_count += rx_queue->scatter_n; rx_queue 1071 drivers/net/ethernet/sfc/farch.c rx_queue->scatter_n = 0; rx_queue 1103 drivers/net/ethernet/sfc/farch.c struct efx_rx_queue *rx_queue; rx_queue 1114 drivers/net/ethernet/sfc/farch.c rx_queue = efx_channel_get_rx_queue(channel); rx_queue 1119 drivers/net/ethernet/sfc/farch.c rx_queue->flush_pending = true; rx_queue 1122 drivers/net/ethernet/sfc/farch.c efx_farch_magic_event(efx_rx_queue_channel(rx_queue), rx_queue 1123 drivers/net/ethernet/sfc/farch.c EFX_CHANNEL_MAGIC_RX_DRAIN(rx_queue)); rx_queue 1145 drivers/net/ethernet/sfc/farch.c struct efx_rx_queue *rx_queue = rx_queue 1155 drivers/net/ethernet/sfc/farch.c } else if (rx_queue && magic == EFX_CHANNEL_MAGIC_FILL(rx_queue)) { rx_queue 1159 drivers/net/ethernet/sfc/farch.c efx_fast_push_rx_descriptors(rx_queue, true); rx_queue 1160 drivers/net/ethernet/sfc/farch.c } else if (rx_queue && magic == EFX_CHANNEL_MAGIC_RX_DRAIN(rx_queue)) { rx_queue 1403 drivers/net/ethernet/sfc/farch.c void efx_farch_rx_defer_refill(struct efx_rx_queue *rx_queue) rx_queue 1405 drivers/net/ethernet/sfc/farch.c efx_farch_magic_event(efx_rx_queue_channel(rx_queue), rx_queue 1406 drivers/net/ethernet/sfc/farch.c EFX_CHANNEL_MAGIC_FILL(rx_queue)); rx_queue 1970 drivers/net/ethernet/sfc/mcdi.c struct efx_rx_queue *rx_queue; rx_queue 1980 drivers/net/ethernet/sfc/mcdi.c efx_for_each_channel_rx_queue(rx_queue, channel) { rx_queue 1981 drivers/net/ethernet/sfc/mcdi.c if (rx_queue->flush_pending) { rx_queue 1982 drivers/net/ethernet/sfc/mcdi.c rx_queue->flush_pending = false; rx_queue 1986 drivers/net/ethernet/sfc/mcdi.c count, efx_rx_queue_index(rx_queue)); rx_queue 503 drivers/net/ethernet/sfc/net_driver.h struct efx_rx_queue rx_queue; rx_queue 1337 drivers/net/ethernet/sfc/net_driver.h int (*rx_probe)(struct efx_rx_queue *rx_queue); rx_queue 1338 drivers/net/ethernet/sfc/net_driver.h void (*rx_init)(struct efx_rx_queue *rx_queue); rx_queue 1339 drivers/net/ethernet/sfc/net_driver.h void (*rx_remove)(struct efx_rx_queue *rx_queue); rx_queue 1340 drivers/net/ethernet/sfc/net_driver.h void (*rx_write)(struct efx_rx_queue *rx_queue); rx_queue 1341 drivers/net/ethernet/sfc/net_driver.h void (*rx_defer_refill)(struct efx_rx_queue *rx_queue); rx_queue 1517 drivers/net/ethernet/sfc/net_driver.h return channel->rx_queue.core_index >= 0; rx_queue 1524 drivers/net/ethernet/sfc/net_driver.h return &channel->rx_queue; rx_queue 1532 drivers/net/ethernet/sfc/net_driver.h for (_rx_queue = &(_channel)->rx_queue; \ rx_queue 1537 drivers/net/ethernet/sfc/net_driver.h efx_rx_queue_channel(struct efx_rx_queue *rx_queue) rx_queue 1539 drivers/net/ethernet/sfc/net_driver.h return container_of(rx_queue, struct efx_channel, rx_queue); rx_queue 1542 drivers/net/ethernet/sfc/net_driver.h static inline int efx_rx_queue_index(struct efx_rx_queue *rx_queue) rx_queue 1544 drivers/net/ethernet/sfc/net_driver.h return efx_rx_queue_channel(rx_queue)->channel; rx_queue 1550 drivers/net/ethernet/sfc/net_driver.h static inline struct efx_rx_buffer *efx_rx_buffer(struct efx_rx_queue *rx_queue, rx_queue 1553 drivers/net/ethernet/sfc/net_driver.h return &rx_queue->buffer[index]; rx_queue 133 drivers/net/ethernet/sfc/nic.h efx_rx_desc(struct efx_rx_queue *rx_queue, unsigned int index) rx_queue 135 drivers/net/ethernet/sfc/nic.h return ((efx_qword_t *) (rx_queue->rxd.buf.addr)) + index; rx_queue 510 drivers/net/ethernet/sfc/nic.h static inline int efx_nic_probe_rx(struct efx_rx_queue *rx_queue) rx_queue 512 drivers/net/ethernet/sfc/nic.h return rx_queue->efx->type->rx_probe(rx_queue); rx_queue 514 drivers/net/ethernet/sfc/nic.h static inline void efx_nic_init_rx(struct efx_rx_queue *rx_queue) rx_queue 516 drivers/net/ethernet/sfc/nic.h rx_queue->efx->type->rx_init(rx_queue); rx_queue 518 drivers/net/ethernet/sfc/nic.h static inline void efx_nic_remove_rx(struct efx_rx_queue *rx_queue) rx_queue 520 drivers/net/ethernet/sfc/nic.h rx_queue->efx->type->rx_remove(rx_queue); rx_queue 522 drivers/net/ethernet/sfc/nic.h static inline void efx_nic_notify_rx_desc(struct efx_rx_queue *rx_queue) rx_queue 524 drivers/net/ethernet/sfc/nic.h rx_queue->efx->type->rx_write(rx_queue); rx_queue 526 drivers/net/ethernet/sfc/nic.h static inline void efx_nic_generate_fill_event(struct efx_rx_queue *rx_queue) rx_queue 528 drivers/net/ethernet/sfc/nic.h rx_queue->efx->type->rx_defer_refill(rx_queue); rx_queue 567 drivers/net/ethernet/sfc/nic.h int efx_farch_rx_probe(struct efx_rx_queue *rx_queue); rx_queue 568 drivers/net/ethernet/sfc/nic.h void efx_farch_rx_init(struct efx_rx_queue *rx_queue); rx_queue 569 drivers/net/ethernet/sfc/nic.h void efx_farch_rx_fini(struct efx_rx_queue *rx_queue); rx_queue 570 drivers/net/ethernet/sfc/nic.h void efx_farch_rx_remove(struct efx_rx_queue *rx_queue); rx_queue 571 drivers/net/ethernet/sfc/nic.h void efx_farch_rx_write(struct efx_rx_queue *rx_queue); rx_queue 572 drivers/net/ethernet/sfc/nic.h void efx_farch_rx_defer_refill(struct efx_rx_queue *rx_queue); rx_queue 1544 drivers/net/ethernet/sfc/ptp.c channel->rx_queue.core_index = 0; rx_queue 76 drivers/net/ethernet/sfc/rx.c efx_rx_buf_next(struct efx_rx_queue *rx_queue, struct efx_rx_buffer *rx_buf) rx_queue 78 drivers/net/ethernet/sfc/rx.c if (unlikely(rx_buf == efx_rx_buffer(rx_queue, rx_queue->ptr_mask))) rx_queue 79 drivers/net/ethernet/sfc/rx.c return efx_rx_buffer(rx_queue, 0); rx_queue 106 drivers/net/ethernet/sfc/rx.c static struct page *efx_reuse_page(struct efx_rx_queue *rx_queue) rx_queue 108 drivers/net/ethernet/sfc/rx.c struct efx_nic *efx = rx_queue->efx; rx_queue 113 drivers/net/ethernet/sfc/rx.c index = rx_queue->page_remove & rx_queue->page_ptr_mask; rx_queue 114 drivers/net/ethernet/sfc/rx.c page = rx_queue->page_ring[index]; rx_queue 118 drivers/net/ethernet/sfc/rx.c rx_queue->page_ring[index] = NULL; rx_queue 120 drivers/net/ethernet/sfc/rx.c if (rx_queue->page_remove != rx_queue->page_add) rx_queue 121 drivers/net/ethernet/sfc/rx.c ++rx_queue->page_remove; rx_queue 125 drivers/net/ethernet/sfc/rx.c ++rx_queue->page_recycle_count; rx_queue 133 drivers/net/ethernet/sfc/rx.c ++rx_queue->page_recycle_failed; rx_queue 149 drivers/net/ethernet/sfc/rx.c static int efx_init_rx_buffers(struct efx_rx_queue *rx_queue, bool atomic) rx_queue 151 drivers/net/ethernet/sfc/rx.c struct efx_nic *efx = rx_queue->efx; rx_queue 161 drivers/net/ethernet/sfc/rx.c page = efx_reuse_page(rx_queue); rx_queue 188 drivers/net/ethernet/sfc/rx.c index = rx_queue->added_count & rx_queue->ptr_mask; rx_queue 189 drivers/net/ethernet/sfc/rx.c rx_buf = efx_rx_buffer(rx_queue, index); rx_queue 195 drivers/net/ethernet/sfc/rx.c ++rx_queue->added_count; rx_queue 224 drivers/net/ethernet/sfc/rx.c static void efx_free_rx_buffers(struct efx_rx_queue *rx_queue, rx_queue 233 drivers/net/ethernet/sfc/rx.c rx_buf = efx_rx_buf_next(rx_queue, rx_buf); rx_queue 245 drivers/net/ethernet/sfc/rx.c struct efx_rx_queue *rx_queue = efx_channel_get_rx_queue(channel); rx_queue 246 drivers/net/ethernet/sfc/rx.c struct efx_nic *efx = rx_queue->efx; rx_queue 253 drivers/net/ethernet/sfc/rx.c index = rx_queue->page_add & rx_queue->page_ptr_mask; rx_queue 254 drivers/net/ethernet/sfc/rx.c if (rx_queue->page_ring[index] == NULL) { rx_queue 255 drivers/net/ethernet/sfc/rx.c unsigned read_index = rx_queue->page_remove & rx_queue 256 drivers/net/ethernet/sfc/rx.c rx_queue->page_ptr_mask; rx_queue 263 drivers/net/ethernet/sfc/rx.c ++rx_queue->page_remove; rx_queue 264 drivers/net/ethernet/sfc/rx.c rx_queue->page_ring[index] = page; rx_queue 265 drivers/net/ethernet/sfc/rx.c ++rx_queue->page_add; rx_queue 268 drivers/net/ethernet/sfc/rx.c ++rx_queue->page_recycle_full; rx_queue 273 drivers/net/ethernet/sfc/rx.c static void efx_fini_rx_buffer(struct efx_rx_queue *rx_queue, rx_queue 282 drivers/net/ethernet/sfc/rx.c efx_unmap_rx_buffer(rx_queue->efx, rx_buf); rx_queue 283 drivers/net/ethernet/sfc/rx.c efx_free_rx_buffers(rx_queue, rx_buf, 1); rx_queue 293 drivers/net/ethernet/sfc/rx.c struct efx_rx_queue *rx_queue = efx_channel_get_rx_queue(channel); rx_queue 297 drivers/net/ethernet/sfc/rx.c rx_buf = efx_rx_buf_next(rx_queue, rx_buf); rx_queue 305 drivers/net/ethernet/sfc/rx.c struct efx_rx_queue *rx_queue = efx_channel_get_rx_queue(channel); rx_queue 309 drivers/net/ethernet/sfc/rx.c efx_free_rx_buffers(rx_queue, rx_buf, n_frags); rx_queue 324 drivers/net/ethernet/sfc/rx.c void efx_fast_push_rx_descriptors(struct efx_rx_queue *rx_queue, bool atomic) rx_queue 326 drivers/net/ethernet/sfc/rx.c struct efx_nic *efx = rx_queue->efx; rx_queue 330 drivers/net/ethernet/sfc/rx.c if (!rx_queue->refill_enabled) rx_queue 334 drivers/net/ethernet/sfc/rx.c fill_level = (rx_queue->added_count - rx_queue->removed_count); rx_queue 335 drivers/net/ethernet/sfc/rx.c EFX_WARN_ON_ONCE_PARANOID(fill_level > rx_queue->efx->rxq_entries); rx_queue 336 drivers/net/ethernet/sfc/rx.c if (fill_level >= rx_queue->fast_fill_trigger) rx_queue 340 drivers/net/ethernet/sfc/rx.c if (unlikely(fill_level < rx_queue->min_fill)) { rx_queue 342 drivers/net/ethernet/sfc/rx.c rx_queue->min_fill = fill_level; rx_queue 346 drivers/net/ethernet/sfc/rx.c space = rx_queue->max_fill - fill_level; rx_queue 349 drivers/net/ethernet/sfc/rx.c netif_vdbg(rx_queue->efx, rx_status, rx_queue->efx->net_dev, rx_queue 352 drivers/net/ethernet/sfc/rx.c efx_rx_queue_index(rx_queue), fill_level, rx_queue 353 drivers/net/ethernet/sfc/rx.c rx_queue->max_fill); rx_queue 357 drivers/net/ethernet/sfc/rx.c rc = efx_init_rx_buffers(rx_queue, atomic); rx_queue 360 drivers/net/ethernet/sfc/rx.c efx_schedule_slow_fill(rx_queue); rx_queue 365 drivers/net/ethernet/sfc/rx.c netif_vdbg(rx_queue->efx, rx_status, rx_queue->efx->net_dev, rx_queue 367 drivers/net/ethernet/sfc/rx.c "to level %d\n", efx_rx_queue_index(rx_queue), rx_queue 368 drivers/net/ethernet/sfc/rx.c rx_queue->added_count - rx_queue->removed_count); rx_queue 371 drivers/net/ethernet/sfc/rx.c if (rx_queue->notified_count != rx_queue->added_count) rx_queue 372 drivers/net/ethernet/sfc/rx.c efx_nic_notify_rx_desc(rx_queue); rx_queue 377 drivers/net/ethernet/sfc/rx.c struct efx_rx_queue *rx_queue = from_timer(rx_queue, t, slow_fill); rx_queue 380 drivers/net/ethernet/sfc/rx.c efx_nic_generate_fill_event(rx_queue); rx_queue 381 drivers/net/ethernet/sfc/rx.c ++rx_queue->slow_fill_count; rx_queue 384 drivers/net/ethernet/sfc/rx.c static void efx_rx_packet__check_len(struct efx_rx_queue *rx_queue, rx_queue 388 drivers/net/ethernet/sfc/rx.c struct efx_nic *efx = rx_queue->efx; rx_queue 402 drivers/net/ethernet/sfc/rx.c efx_rx_queue_index(rx_queue), len, max_len); rx_queue 404 drivers/net/ethernet/sfc/rx.c efx_rx_queue_channel(rx_queue)->n_rx_overlength++; rx_queue 420 drivers/net/ethernet/sfc/rx.c struct efx_rx_queue *rx_queue; rx_queue 422 drivers/net/ethernet/sfc/rx.c rx_queue = efx_channel_get_rx_queue(channel); rx_queue 423 drivers/net/ethernet/sfc/rx.c efx_free_rx_buffers(rx_queue, rx_buf, n_frags); rx_queue 443 drivers/net/ethernet/sfc/rx.c rx_buf = efx_rx_buf_next(&channel->rx_queue, rx_buf); rx_queue 449 drivers/net/ethernet/sfc/rx.c skb_record_rx_queue(skb, channel->rx_queue.core_index); rx_queue 494 drivers/net/ethernet/sfc/rx.c rx_buf = efx_rx_buf_next(&channel->rx_queue, rx_buf); rx_queue 512 drivers/net/ethernet/sfc/rx.c void efx_rx_packet(struct efx_rx_queue *rx_queue, unsigned int index, rx_queue 515 drivers/net/ethernet/sfc/rx.c struct efx_nic *efx = rx_queue->efx; rx_queue 516 drivers/net/ethernet/sfc/rx.c struct efx_channel *channel = efx_rx_queue_channel(rx_queue); rx_queue 519 drivers/net/ethernet/sfc/rx.c rx_queue->rx_packets++; rx_queue 521 drivers/net/ethernet/sfc/rx.c rx_buf = efx_rx_buffer(rx_queue, index); rx_queue 527 drivers/net/ethernet/sfc/rx.c efx_rx_packet__check_len(rx_queue, rx_buf, len); rx_queue 541 drivers/net/ethernet/sfc/rx.c efx_rx_queue_index(rx_queue), index, rx_queue 542 drivers/net/ethernet/sfc/rx.c (index + n_frags - 1) & rx_queue->ptr_mask, len, rx_queue 578 drivers/net/ethernet/sfc/rx.c rx_buf = efx_rx_buf_next(rx_queue, rx_buf); rx_queue 588 drivers/net/ethernet/sfc/rx.c rx_buf = efx_rx_buffer(rx_queue, index); rx_queue 608 drivers/net/ethernet/sfc/rx.c struct efx_rx_queue *rx_queue; rx_queue 610 drivers/net/ethernet/sfc/rx.c rx_queue = efx_channel_get_rx_queue(channel); rx_queue 611 drivers/net/ethernet/sfc/rx.c efx_free_rx_buffers(rx_queue, rx_buf, n_frags); rx_queue 614 drivers/net/ethernet/sfc/rx.c skb_record_rx_queue(skb, channel->rx_queue.core_index); rx_queue 643 drivers/net/ethernet/sfc/rx.c efx_rx_buffer(&channel->rx_queue, channel->rx_pkt_index); rx_queue 657 drivers/net/ethernet/sfc/rx.c struct efx_rx_queue *rx_queue; rx_queue 660 drivers/net/ethernet/sfc/rx.c rx_queue = efx_channel_get_rx_queue(channel); rx_queue 661 drivers/net/ethernet/sfc/rx.c efx_free_rx_buffers(rx_queue, rx_buf, rx_queue 677 drivers/net/ethernet/sfc/rx.c int efx_probe_rx_queue(struct efx_rx_queue *rx_queue) rx_queue 679 drivers/net/ethernet/sfc/rx.c struct efx_nic *efx = rx_queue->efx; rx_queue 686 drivers/net/ethernet/sfc/rx.c rx_queue->ptr_mask = entries - 1; rx_queue 690 drivers/net/ethernet/sfc/rx.c efx_rx_queue_index(rx_queue), efx->rxq_entries, rx_queue 691 drivers/net/ethernet/sfc/rx.c rx_queue->ptr_mask); rx_queue 694 drivers/net/ethernet/sfc/rx.c rx_queue->buffer = kcalloc(entries, sizeof(*rx_queue->buffer), rx_queue 696 drivers/net/ethernet/sfc/rx.c if (!rx_queue->buffer) rx_queue 699 drivers/net/ethernet/sfc/rx.c rc = efx_nic_probe_rx(rx_queue); rx_queue 701 drivers/net/ethernet/sfc/rx.c kfree(rx_queue->buffer); rx_queue 702 drivers/net/ethernet/sfc/rx.c rx_queue->buffer = NULL; rx_queue 709 drivers/net/ethernet/sfc/rx.c struct efx_rx_queue *rx_queue) rx_queue 725 drivers/net/ethernet/sfc/rx.c rx_queue->page_ring = kcalloc(page_ring_size, rx_queue 726 drivers/net/ethernet/sfc/rx.c sizeof(*rx_queue->page_ring), GFP_KERNEL); rx_queue 727 drivers/net/ethernet/sfc/rx.c rx_queue->page_ptr_mask = page_ring_size - 1; rx_queue 730 drivers/net/ethernet/sfc/rx.c void efx_init_rx_queue(struct efx_rx_queue *rx_queue) rx_queue 732 drivers/net/ethernet/sfc/rx.c struct efx_nic *efx = rx_queue->efx; rx_queue 735 drivers/net/ethernet/sfc/rx.c netif_dbg(rx_queue->efx, drv, rx_queue->efx->net_dev, rx_queue 736 drivers/net/ethernet/sfc/rx.c "initialising RX queue %d\n", efx_rx_queue_index(rx_queue)); rx_queue 739 drivers/net/ethernet/sfc/rx.c rx_queue->added_count = 0; rx_queue 740 drivers/net/ethernet/sfc/rx.c rx_queue->notified_count = 0; rx_queue 741 drivers/net/ethernet/sfc/rx.c rx_queue->removed_count = 0; rx_queue 742 drivers/net/ethernet/sfc/rx.c rx_queue->min_fill = -1U; rx_queue 743 drivers/net/ethernet/sfc/rx.c efx_init_rx_recycle_ring(efx, rx_queue); rx_queue 745 drivers/net/ethernet/sfc/rx.c rx_queue->page_remove = 0; rx_queue 746 drivers/net/ethernet/sfc/rx.c rx_queue->page_add = rx_queue->page_ptr_mask + 1; rx_queue 747 drivers/net/ethernet/sfc/rx.c rx_queue->page_recycle_count = 0; rx_queue 748 drivers/net/ethernet/sfc/rx.c rx_queue->page_recycle_failed = 0; rx_queue 749 drivers/net/ethernet/sfc/rx.c rx_queue->page_recycle_full = 0; rx_queue 763 drivers/net/ethernet/sfc/rx.c rx_queue->max_fill = max_fill; rx_queue 764 drivers/net/ethernet/sfc/rx.c rx_queue->fast_fill_trigger = trigger; rx_queue 765 drivers/net/ethernet/sfc/rx.c rx_queue->refill_enabled = true; rx_queue 768 drivers/net/ethernet/sfc/rx.c efx_nic_init_rx(rx_queue); rx_queue 771 drivers/net/ethernet/sfc/rx.c void efx_fini_rx_queue(struct efx_rx_queue *rx_queue) rx_queue 774 drivers/net/ethernet/sfc/rx.c struct efx_nic *efx = rx_queue->efx; rx_queue 777 drivers/net/ethernet/sfc/rx.c netif_dbg(rx_queue->efx, drv, rx_queue->efx->net_dev, rx_queue 778 drivers/net/ethernet/sfc/rx.c "shutting down RX queue %d\n", efx_rx_queue_index(rx_queue)); rx_queue 780 drivers/net/ethernet/sfc/rx.c del_timer_sync(&rx_queue->slow_fill); rx_queue 783 drivers/net/ethernet/sfc/rx.c if (rx_queue->buffer) { rx_queue 784 drivers/net/ethernet/sfc/rx.c for (i = rx_queue->removed_count; i < rx_queue->added_count; rx_queue 786 drivers/net/ethernet/sfc/rx.c unsigned index = i & rx_queue->ptr_mask; rx_queue 787 drivers/net/ethernet/sfc/rx.c rx_buf = efx_rx_buffer(rx_queue, index); rx_queue 788 drivers/net/ethernet/sfc/rx.c efx_fini_rx_buffer(rx_queue, rx_buf); rx_queue 793 drivers/net/ethernet/sfc/rx.c for (i = 0; i <= rx_queue->page_ptr_mask; i++) { rx_queue 794 drivers/net/ethernet/sfc/rx.c struct page *page = rx_queue->page_ring[i]; rx_queue 806 drivers/net/ethernet/sfc/rx.c kfree(rx_queue->page_ring); rx_queue 807 drivers/net/ethernet/sfc/rx.c rx_queue->page_ring = NULL; rx_queue 810 drivers/net/ethernet/sfc/rx.c void efx_remove_rx_queue(struct efx_rx_queue *rx_queue) rx_queue 812 drivers/net/ethernet/sfc/rx.c netif_dbg(rx_queue->efx, drv, rx_queue->efx->net_dev, rx_queue 813 drivers/net/ethernet/sfc/rx.c "destroying RX queue %d\n", efx_rx_queue_index(rx_queue)); rx_queue 815 drivers/net/ethernet/sfc/rx.c efx_nic_remove_rx(rx_queue); rx_queue 817 drivers/net/ethernet/sfc/rx.c kfree(rx_queue->buffer); rx_queue 818 drivers/net/ethernet/sfc/rx.c rx_queue->buffer = NULL; rx_queue 167 drivers/net/ethernet/stmicro/stmmac/stmmac.h struct stmmac_rx_queue rx_queue[MTL_MAX_RX_QUEUES]; rx_queue 311 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue]; rx_queue 1065 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue]; rx_queue 1136 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue]; rx_queue 1209 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue]; rx_queue 1243 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue]; rx_queue 1307 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue]; rx_queue 1481 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue]; rx_queue 1549 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue]; rx_queue 2219 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c rx_q = &priv->rx_queue[chan]; rx_queue 3386 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue]; rx_queue 3458 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue]; rx_queue 3881 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue]; rx_queue 4075 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue]; rx_queue 4820 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c struct stmmac_rx_queue *rx_q = &priv->rx_queue[queue]; rx_queue 794 drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c tail = priv->rx_queue[i].dma_rx_phy + rx_queue 92 drivers/net/ethernet/ti/netcp.h void *rx_queue; rx_queue 624 drivers/net/ethernet/ti/netcp_core.c dma = knav_queue_pop(netcp->rx_queue, &dma_sz); rx_queue 653 drivers/net/ethernet/ti/netcp_core.c dma_desc = knav_queue_pop(netcp->rx_queue, &dma_sz); rx_queue 959 drivers/net/ethernet/ti/netcp_core.c knav_queue_enable_notify(netcp->rx_queue); rx_queue 969 drivers/net/ethernet/ti/netcp_core.c knav_queue_disable_notify(netcp->rx_queue); rx_queue 1561 drivers/net/ethernet/ti/netcp_core.c if (!IS_ERR_OR_NULL(netcp->rx_queue)) { rx_queue 1562 drivers/net/ethernet/ti/netcp_core.c knav_queue_close(netcp->rx_queue); rx_queue 1563 drivers/net/ethernet/ti/netcp_core.c netcp->rx_queue = NULL; rx_queue 1634 drivers/net/ethernet/ti/netcp_core.c netcp->rx_queue = knav_queue_open(name, netcp->rx_queue_id, 0); rx_queue 1635 drivers/net/ethernet/ti/netcp_core.c if (IS_ERR(netcp->rx_queue)) { rx_queue 1636 drivers/net/ethernet/ti/netcp_core.c ret = PTR_ERR(netcp->rx_queue); rx_queue 1639 drivers/net/ethernet/ti/netcp_core.c netcp->rx_queue_id = knav_queue_get_id(netcp->rx_queue); rx_queue 1644 drivers/net/ethernet/ti/netcp_core.c ret = knav_queue_device_control(netcp->rx_queue, rx_queue 1650 drivers/net/ethernet/ti/netcp_core.c knav_queue_disable_notify(netcp->rx_queue); rx_queue 1726 drivers/net/ethernet/ti/netcp_core.c knav_queue_enable_notify(netcp->rx_queue); rx_queue 1756 drivers/net/ethernet/ti/netcp_core.c knav_queue_disable_notify(netcp->rx_queue); rx_queue 117 drivers/net/phy/dp83640.c struct sk_buff_head rx_queue; rx_queue 875 drivers/net/phy/dp83640.c spin_lock(&dp83640->rx_queue.lock); rx_queue 876 drivers/net/phy/dp83640.c skb_queue_walk(&dp83640->rx_queue, skb) { rx_queue 881 drivers/net/phy/dp83640.c __skb_unlink(skb, &dp83640->rx_queue); rx_queue 889 drivers/net/phy/dp83640.c spin_unlock(&dp83640->rx_queue.lock); rx_queue 1162 drivers/net/phy/dp83640.c skb_queue_head_init(&dp83640->rx_queue); rx_queue 1202 drivers/net/phy/dp83640.c skb_queue_purge(&dp83640->rx_queue); rx_queue 1415 drivers/net/phy/dp83640.c while ((skb = skb_dequeue(&dp83640->rx_queue))) { rx_queue 1420 drivers/net/phy/dp83640.c skb_queue_head(&dp83640->rx_queue, skb); rx_queue 1427 drivers/net/phy/dp83640.c if (!skb_queue_empty(&dp83640->rx_queue)) rx_queue 1471 drivers/net/phy/dp83640.c skb_queue_tail(&dp83640->rx_queue, skb); rx_queue 742 drivers/net/usb/r8152.c struct sk_buff_head tx_queue, rx_queue; rx_queue 1624 drivers/net/usb/r8152.c skb_queue_head_init(&tp->rx_queue); rx_queue 2044 drivers/net/usb/r8152.c struct list_head *cursor, *next, rx_queue; rx_queue 2048 drivers/net/usb/r8152.c if (!skb_queue_empty(&tp->rx_queue)) { rx_queue 2050 drivers/net/usb/r8152.c struct sk_buff *skb = __skb_dequeue(&tp->rx_queue); rx_queue 2069 drivers/net/usb/r8152.c INIT_LIST_HEAD(&rx_queue); rx_queue 2071 drivers/net/usb/r8152.c list_splice_init(&tp->rx_done, &rx_queue); rx_queue 2074 drivers/net/usb/r8152.c list_for_each_safe(cursor, next, &rx_queue) { rx_queue 2101 drivers/net/usb/r8152.c if (unlikely(skb_queue_len(&tp->rx_queue) >= 1000)) rx_queue 2147 drivers/net/usb/r8152.c __skb_queue_tail(&tp->rx_queue, skb); rx_queue 2180 drivers/net/usb/r8152.c if (!list_empty(&rx_queue)) { rx_queue 2182 drivers/net/usb/r8152.c list_splice_tail(&rx_queue, &tp->rx_done); rx_queue 2570 drivers/net/usb/r8152.c while (!skb_queue_empty(&tp->rx_queue)) rx_queue 2571 drivers/net/usb/r8152.c dev_kfree_skb(__skb_dequeue(&tp->rx_queue)); rx_queue 1625 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_rq_cleanup(&adapter->rx_queue[i], adapter); rx_queue 1684 drivers/net/vmxnet3/vmxnet3_drv.c struct vmxnet3_rx_queue *rq = &adapter->rx_queue[i]; rx_queue 1756 drivers/net/vmxnet3/vmxnet3_drv.c err = vmxnet3_rq_init(&adapter->rx_queue[i], adapter); rx_queue 1841 drivers/net/vmxnet3/vmxnet3_drv.c err = vmxnet3_rq_create(&adapter->rx_queue[i], adapter); rx_queue 1872 drivers/net/vmxnet3/vmxnet3_drv.c rcd_done += vmxnet3_rq_rx_complete(&adapter->rx_queue[i], rx_queue 1881 drivers/net/vmxnet3/vmxnet3_drv.c struct vmxnet3_rx_queue *rx_queue = container_of(napi, rx_queue 1885 drivers/net/vmxnet3/vmxnet3_drv.c rxd_done = vmxnet3_do_poll(rx_queue->adapter, budget); rx_queue 1889 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_enable_all_intrs(rx_queue->adapter); rx_queue 1912 drivers/net/vmxnet3/vmxnet3_drv.c &adapter->tx_queue[rq - adapter->rx_queue]; rx_queue 2030 drivers/net/vmxnet3/vmxnet3_drv.c napi_schedule(&adapter->rx_queue[0].napi); rx_queue 2048 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_msix_rx(0, &adapter->rx_queue[i]); rx_queue 2109 drivers/net/vmxnet3/vmxnet3_drv.c sprintf(adapter->rx_queue[i].name, "%s-rx-%d", rx_queue 2112 drivers/net/vmxnet3/vmxnet3_drv.c sprintf(adapter->rx_queue[i].name, "%s-rxtx-%d", rx_queue 2116 drivers/net/vmxnet3/vmxnet3_drv.c adapter->rx_queue[i].name, rx_queue 2117 drivers/net/vmxnet3/vmxnet3_drv.c &(adapter->rx_queue[i])); rx_queue 2122 drivers/net/vmxnet3/vmxnet3_drv.c adapter->rx_queue[i].name, err); rx_queue 2126 drivers/net/vmxnet3/vmxnet3_drv.c adapter->rx_queue[i].comp_ring.intr_idx = vector++; rx_queue 2157 drivers/net/vmxnet3/vmxnet3_drv.c struct vmxnet3_rx_queue *rq = &adapter->rx_queue[i]; rx_queue 2170 drivers/net/vmxnet3/vmxnet3_drv.c adapter->rx_queue[0].comp_ring.intr_idx = 0; rx_queue 2205 drivers/net/vmxnet3/vmxnet3_drv.c &(adapter->rx_queue[i])); rx_queue 2390 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_rq_destroy(&adapter->rx_queue[i], adapter); rx_queue 2464 drivers/net/vmxnet3/vmxnet3_drv.c struct vmxnet3_rx_queue *rq = &adapter->rx_queue[i]; rx_queue 2571 drivers/net/vmxnet3/vmxnet3_drv.c adapter->rx_queue[0].rx_ring[0].size, rx_queue 2572 drivers/net/vmxnet3/vmxnet3_drv.c adapter->rx_queue[0].rx_ring[1].size); rx_queue 2613 drivers/net/vmxnet3/vmxnet3_drv.c adapter->rx_queue[i].rx_ring[0].next2fill); rx_queue 2616 drivers/net/vmxnet3/vmxnet3_drv.c adapter->rx_queue[i].rx_ring[1].next2fill); rx_queue 2628 drivers/net/vmxnet3/vmxnet3_drv.c napi_enable(&adapter->rx_queue[i].napi); rx_queue 2671 drivers/net/vmxnet3/vmxnet3_drv.c napi_disable(&adapter->rx_queue[i].napi); rx_queue 2799 drivers/net/vmxnet3/vmxnet3_drv.c ring0_size = adapter->rx_queue[0].rx_ring[0].size; rx_queue 2803 drivers/net/vmxnet3/vmxnet3_drv.c ring1_size = adapter->rx_queue[0].rx_ring[1].size; rx_queue 2810 drivers/net/vmxnet3/vmxnet3_drv.c struct vmxnet3_rx_queue *rq = &adapter->rx_queue[i]; rx_queue 2845 drivers/net/vmxnet3/vmxnet3_drv.c adapter->rx_queue[0].rx_ring[0].size = rx_ring_size; rx_queue 2846 drivers/net/vmxnet3/vmxnet3_drv.c adapter->rx_queue[0].rx_ring[1].size = rx_ring2_size; rx_queue 2851 drivers/net/vmxnet3/vmxnet3_drv.c struct vmxnet3_rx_queue *rq = &adapter->rx_queue[i]; rx_queue 2978 drivers/net/vmxnet3/vmxnet3_drv.c napi_enable(&adapter->rx_queue[i].napi); rx_queue 3479 drivers/net/vmxnet3/vmxnet3_drv.c &adapter->rx_queue[i].napi, rx_queue 3483 drivers/net/vmxnet3/vmxnet3_drv.c netif_napi_add(adapter->netdev, &adapter->rx_queue[0].napi, rx_queue 3629 drivers/net/vmxnet3/vmxnet3_drv.c napi_disable(&adapter->rx_queue[i].napi); rx_queue 150 drivers/net/vmxnet3/vmxnet3_ethtool.c drvRxStats = &adapter->rx_queue[i].stats; rx_queue 343 drivers/net/vmxnet3/vmxnet3_ethtool.c base = (u8 *)&adapter->rx_queue[j].stats; rx_queue 420 drivers/net/vmxnet3/vmxnet3_ethtool.c struct vmxnet3_rx_queue *rq = &adapter->rx_queue[i]; rx_queue 332 drivers/net/vmxnet3/vmxnet3_int.h struct vmxnet3_rx_queue rx_queue[VMXNET3_DEVICE_MAX_RX_QUEUES]; rx_queue 684 drivers/net/wireless/broadcom/b43/b43.h struct b43_pio_rxqueue *rx_queue; rx_queue 2029 drivers/net/wireless/broadcom/b43/main.c b43_pio_rx(dev->pio.rx_queue); rx_queue 227 drivers/net/wireless/broadcom/b43/pio.c destroy_queue_rx(pio, rx_queue); rx_queue 264 drivers/net/wireless/broadcom/b43/pio.c pio->rx_queue = b43_setup_pioqueue_rx(dev, 0); rx_queue 265 drivers/net/wireless/broadcom/b43/pio.c if (!pio->rx_queue) rx_queue 2518 drivers/net/wireless/intel/ipw2x00/ipw2100.c priv->rx_queue.drv[i].host_addr = packet->dma_addr; rx_queue 2600 drivers/net/wireless/intel/ipw2x00/ipw2100.c priv->rx_queue.drv[i].host_addr = packet->dma_addr; rx_queue 2663 drivers/net/wireless/intel/ipw2x00/ipw2100.c struct ipw2100_bd_queue *rxq = &priv->rx_queue; rx_queue 4562 drivers/net/wireless/intel/ipw2x00/ipw2100.c err = bd_queue_allocate(priv, &priv->rx_queue, RX_QUEUE_LENGTH); rx_queue 4571 drivers/net/wireless/intel/ipw2x00/ipw2100.c bd_queue_free(priv, &priv->rx_queue); rx_queue 4584 drivers/net/wireless/intel/ipw2x00/ipw2100.c bd_queue_free(priv, &priv->rx_queue); rx_queue 4601 drivers/net/wireless/intel/ipw2x00/ipw2100.c priv->rx_queue.drv[i].host_addr = packet->dma_addr; rx_queue 4602 drivers/net/wireless/intel/ipw2x00/ipw2100.c priv->rx_queue.drv[i].buf_length = IPW_RX_NIC_BUFFER_LENGTH; rx_queue 4619 drivers/net/wireless/intel/ipw2x00/ipw2100.c bd_queue_free(priv, &priv->rx_queue); rx_queue 4630 drivers/net/wireless/intel/ipw2x00/ipw2100.c priv->rx_queue.oldest = 0; rx_queue 4631 drivers/net/wireless/intel/ipw2x00/ipw2100.c priv->rx_queue.available = priv->rx_queue.entries - 1; rx_queue 4632 drivers/net/wireless/intel/ipw2x00/ipw2100.c priv->rx_queue.next = priv->rx_queue.entries - 1; rx_queue 4635 drivers/net/wireless/intel/ipw2x00/ipw2100.c SET_STAT(&priv->rxq_stat, priv->rx_queue.available); rx_queue 4637 drivers/net/wireless/intel/ipw2x00/ipw2100.c bd_queue_initialize(priv, &priv->rx_queue, rx_queue 4656 drivers/net/wireless/intel/ipw2x00/ipw2100.c bd_queue_free(priv, &priv->rx_queue); rx_queue 529 drivers/net/wireless/intel/ipw2x00/ipw2100.h struct ipw2100_bd_queue rx_queue; rx_queue 1294 drivers/net/wireless/intel/iwlegacy/debug.c DEBUGFS_READ_FILE_OPS(rx_queue); rx_queue 1338 drivers/net/wireless/intel/iwlegacy/debug.c DEBUGFS_ADD_FILE(rx_queue, dir_debug, 0400); rx_queue 2914 drivers/net/wireless/intel/iwlwifi/pcie/trans.c DEBUGFS_READ_FILE_OPS(rx_queue); rx_queue 2930 drivers/net/wireless/intel/iwlwifi/pcie/trans.c DEBUGFS_ADD_FILE(rx_queue, dir, 0400); rx_queue 147 drivers/net/wireless/intersil/p54/p54usb.c skb_unlink(skb, &priv->rx_queue); rx_queue 186 drivers/net/wireless/intersil/p54/p54usb.c skb_queue_tail(&priv->rx_queue, skb); rx_queue 189 drivers/net/wireless/intersil/p54/p54usb.c skb_unlink(skb, &priv->rx_queue); rx_queue 230 drivers/net/wireless/intersil/p54/p54usb.c while (skb_queue_len(&priv->rx_queue) < 32) { rx_queue 249 drivers/net/wireless/intersil/p54/p54usb.c skb_queue_tail(&priv->rx_queue, skb); rx_queue 254 drivers/net/wireless/intersil/p54/p54usb.c skb_unlink(skb, &priv->rx_queue); rx_queue 1007 drivers/net/wireless/intersil/p54/p54usb.c skb_queue_head_init(&priv->rx_queue); rx_queue 151 drivers/net/wireless/intersil/p54/p54usb.h struct sk_buff_head rx_queue; rx_queue 336 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c spin_lock_irqsave(&priv->rx_queue.lock, f); rx_queue 337 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c __skb_unlink(skb, &priv->rx_queue); rx_queue 338 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c spin_unlock_irqrestore(&priv->rx_queue.lock, f); rx_queue 404 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c skb_queue_tail(&priv->rx_queue, skb); rx_queue 409 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c skb_unlink(skb, &priv->rx_queue); rx_queue 422 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c while (skb_queue_len(&priv->rx_queue) < 32) { rx_queue 441 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c skb_queue_tail(&priv->rx_queue, skb); rx_queue 445 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c skb_unlink(skb, &priv->rx_queue); rx_queue 1462 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c skb_queue_head_init(&priv->rx_queue); rx_queue 134 drivers/net/wireless/realtek/rtl818x/rtl8187/rtl8187.h struct sk_buff_head rx_queue; rx_queue 312 drivers/net/wireless/realtek/rtlwifi/usb.c skb_queue_head_init(&rtlusb->rx_queue); rx_queue 515 drivers/net/wireless/realtek/rtlwifi/usb.c struct sk_buff_head rx_queue; rx_queue 518 drivers/net/wireless/realtek/rtlwifi/usb.c skb_queue_head_init(&rx_queue); rx_queue 520 drivers/net/wireless/realtek/rtlwifi/usb.c rtlusb->usb_rx_segregate_hdl(hw, skb, &rx_queue); rx_queue 521 drivers/net/wireless/realtek/rtlwifi/usb.c WARN_ON(skb_queue_empty(&rx_queue)); rx_queue 522 drivers/net/wireless/realtek/rtlwifi/usb.c while (!skb_queue_empty(&rx_queue)) { rx_queue 523 drivers/net/wireless/realtek/rtlwifi/usb.c _skb = skb_dequeue(&rx_queue); rx_queue 537 drivers/net/wireless/realtek/rtlwifi/usb.c while ((skb = skb_dequeue(&rtlusb->rx_queue))) { rx_queue 611 drivers/net/wireless/realtek/rtlwifi/usb.c qlen = skb_queue_len(&rtlusb->rx_queue); rx_queue 637 drivers/net/wireless/realtek/rtlwifi/usb.c skb_queue_tail(&rtlusb->rx_queue, skb); rx_queue 686 drivers/net/wireless/realtek/rtlwifi/usb.c skb_queue_purge(&rtlusb->rx_queue); rx_queue 808 drivers/net/wireless/realtek/rtlwifi/usb.c skb_queue_purge(&rtlusb->rx_queue); rx_queue 121 drivers/net/wireless/realtek/rtlwifi/usb.h struct sk_buff_head rx_queue; rx_queue 88 drivers/net/wireless/st/cw1200/cw1200.h struct sk_buff_head rx_queue; rx_queue 1564 drivers/net/wireless/st/cw1200/sta.c while ((skb = skb_dequeue(&entry->rx_queue))) rx_queue 1210 drivers/net/wireless/st/cw1200/txrx.c skb_queue_tail(&entry->rx_queue, skb); rx_queue 1351 drivers/net/wireless/st/cw1200/txrx.c skb_queue_head_init(&entry->rx_queue); rx_queue 1459 drivers/net/wireless/st/cw1200/txrx.c skb_queue_purge(&priv->link_id_db[i].rx_queue); rx_queue 187 drivers/net/xen-netback/common.h struct sk_buff_head rx_queue; rx_queue 537 drivers/net/xen-netback/interface.c skb_queue_head_init(&queue->rx_queue); rx_queue 42 drivers/net/xen-netback/rx.c skb = skb_peek(&queue->rx_queue); rx_queue 74 drivers/net/xen-netback/rx.c spin_lock_irqsave(&queue->rx_queue.lock, flags); rx_queue 76 drivers/net/xen-netback/rx.c __skb_queue_tail(&queue->rx_queue, skb); rx_queue 85 drivers/net/xen-netback/rx.c spin_unlock_irqrestore(&queue->rx_queue.lock, flags); rx_queue 92 drivers/net/xen-netback/rx.c spin_lock_irq(&queue->rx_queue.lock); rx_queue 94 drivers/net/xen-netback/rx.c skb = __skb_dequeue(&queue->rx_queue); rx_queue 105 drivers/net/xen-netback/rx.c spin_unlock_irq(&queue->rx_queue.lock); rx_queue 123 drivers/net/xen-netback/rx.c skb = skb_peek(&queue->rx_queue); rx_queue 508 drivers/net/xen-netback/rx.c skb = skb_peek(&queue->rx_queue); rx_queue 603 drivers/net/xen-netback/rx.c if (!skb_queue_empty(&queue->rx_queue)) rx_queue 96 drivers/net/xen-netback/xenbus.c skb_queue_len(&queue->rx_queue), rx_queue 111 drivers/rpmsg/qcom_glink_native.c struct list_head rx_queue; rx_queue 777 drivers/rpmsg/qcom_glink_native.c list_add_tail(&dcmd->node, &glink->rx_queue); rx_queue 1523 drivers/rpmsg/qcom_glink_native.c if (list_empty(&glink->rx_queue)) { rx_queue 1527 drivers/rpmsg/qcom_glink_native.c dcmd = list_first_entry(&glink->rx_queue, rx_queue 1573 drivers/rpmsg/qcom_glink_native.c list_for_each_entry_safe(dcmd, tmp, &glink->rx_queue, node) rx_queue 1600 drivers/rpmsg/qcom_glink_native.c INIT_LIST_HEAD(&glink->rx_queue); rx_queue 880 drivers/staging/rtl8192u/r8192U.h struct sk_buff_head rx_queue; rx_queue 724 drivers/staging/rtl8192u/r8192U_core.c while (skb_queue_len(&priv->rx_queue) < MAX_RX_URB) { rx_queue 741 drivers/staging/rtl8192u/r8192U_core.c skb_queue_tail(&priv->rx_queue, skb); rx_queue 746 drivers/staging/rtl8192u/r8192U_core.c while (skb_queue_len(&priv->rx_queue) < MAX_RX_URB + 3) { rx_queue 763 drivers/staging/rtl8192u/r8192U_core.c skb_queue_tail(&priv->rx_queue, skb); rx_queue 823 drivers/staging/rtl8192u/r8192U_core.c while ((skb = __skb_dequeue(&priv->rx_queue))) { rx_queue 858 drivers/staging/rtl8192u/r8192U_core.c skb_unlink(skb, &priv->rx_queue); rx_queue 884 drivers/staging/rtl8192u/r8192U_core.c skb_queue_tail(&priv->rx_queue, skb); rx_queue 2195 drivers/staging/rtl8192u/r8192U_core.c skb_queue_head_init(&priv->rx_queue); rx_queue 246 drivers/tty/ipwireless/hardware.c struct list_head rx_queue; rx_queue 690 drivers/tty/ipwireless/hardware.c list_add_tail(&packet->queue, &hw->rx_queue); rx_queue 710 drivers/tty/ipwireless/hardware.c while (!list_empty(&hw->rx_queue)) { rx_queue 712 drivers/tty/ipwireless/hardware.c list_first_entry(&hw->rx_queue, rx_queue 1635 drivers/tty/ipwireless/hardware.c INIT_LIST_HEAD(&hw->rx_queue); rx_queue 1750 drivers/tty/ipwireless/hardware.c list_for_each_entry_safe(rp, rq, &hw->rx_queue, queue) { rx_queue 4286 include/linux/skbuff.h static inline void skb_record_rx_queue(struct sk_buff *skb, u16 rx_queue) rx_queue 4288 include/linux/skbuff.h skb->queue_mapping = rx_queue + 1; rx_queue 46 include/linux/virtio_vsock.h struct list_head rx_queue; rx_queue 1777 include/net/sock.h u16 rx_queue = skb_get_rx_queue(skb); rx_queue 1779 include/net/sock.h if (WARN_ON_ONCE(rx_queue == NO_QUEUE_MAPPING)) rx_queue 1782 include/net/sock.h sk->sk_rx_queue_mapping = rx_queue; rx_queue 2436 net/ipv4/tcp_ipv4.c int rx_queue; rx_queue 2457 net/ipv4/tcp_ipv4.c rx_queue = sk->sk_ack_backlog; rx_queue 2462 net/ipv4/tcp_ipv4.c rx_queue = max_t(int, READ_ONCE(tp->rcv_nxt) - rx_queue 2469 net/ipv4/tcp_ipv4.c rx_queue, rx_queue 1868 net/ipv6/tcp_ipv6.c int rx_queue; rx_queue 1894 net/ipv6/tcp_ipv6.c rx_queue = sp->sk_ack_backlog; rx_queue 1899 net/ipv6/tcp_ipv6.c rx_queue = max_t(int, READ_ONCE(tp->rcv_nxt) - rx_queue 1912 net/ipv6/tcp_ipv6.c rx_queue, rx_queue 97 net/mac80211/ieee80211_i.h u8 rx_queue; rx_queue 2087 net/mac80211/rx.c unsigned int frag, unsigned int seq, int rx_queue, rx_queue 2103 net/mac80211/rx.c entry->rx_queue = rx_queue; rx_queue 2114 net/mac80211/rx.c int rx_queue, struct ieee80211_hdr *hdr) rx_queue 2130 net/mac80211/rx.c entry->rx_queue != rx_queue || rx_queue 126 net/qrtr/qrtr.c struct sk_buff_head rx_queue; rx_queue 154 net/qrtr/qrtr.c skb_queue_purge(&node->rx_queue); rx_queue 319 net/qrtr/qrtr.c skb_queue_tail(&node->rx_queue, skb); rx_queue 370 net/qrtr/qrtr.c while ((skb = skb_dequeue(&node->rx_queue)) != NULL) { rx_queue 432 net/qrtr/qrtr.c skb_queue_head_init(&node->rx_queue); rx_queue 446 net/rxrpc/ar-internal.h struct sk_buff_head rx_queue; /* received conn-level packets */ rx_queue 456 net/rxrpc/conn_event.c while ((skb = skb_dequeue(&conn->rx_queue))) { rx_queue 477 net/rxrpc/conn_event.c skb_queue_head(&conn->rx_queue, skb); rx_queue 50 net/rxrpc/conn_object.c skb_queue_head_init(&conn->rx_queue); rx_queue 251 net/rxrpc/conn_object.c rxrpc_purge_queue(&conn->rx_queue); rx_queue 361 net/rxrpc/conn_object.c rxrpc_purge_queue(&conn->rx_queue); rx_queue 1118 net/rxrpc/input.c skb_queue_tail(&conn->rx_queue, skb); rx_queue 282 net/vmw_vsock/virtio_transport_common.c while (total < len && !list_empty(&vvs->rx_queue)) { rx_queue 283 net/vmw_vsock/virtio_transport_common.c pkt = list_first_entry(&vvs->rx_queue, rx_queue 423 net/vmw_vsock/virtio_transport_common.c INIT_LIST_HEAD(&vvs->rx_queue); rx_queue 834 net/vmw_vsock/virtio_transport_common.c list_for_each_entry_safe(pkt, tmp, &vvs->rx_queue, list) { rx_queue 903 net/vmw_vsock/virtio_transport_common.c if (pkt->len <= GOOD_COPY_LEN && !list_empty(&vvs->rx_queue)) { rx_queue 906 net/vmw_vsock/virtio_transport_common.c last_pkt = list_last_entry(&vvs->rx_queue, rx_queue 921 net/vmw_vsock/virtio_transport_common.c list_add_tail(&pkt->list, &vvs->rx_queue);