Home
last modified time | relevance | path

Searched refs:rxq (Results 1 – 80 of 80) sorted by relevance

/linux-4.1.27/drivers/net/wireless/iwlwifi/pcie/
Drx.c114 static int iwl_rxq_space(const struct iwl_rxq *rxq) in iwl_rxq_space() argument
125 return (rxq->read - rxq->write - 1) & (RX_QUEUE_SIZE - 1); in iwl_rxq_space()
152 struct iwl_rxq *rxq = &trans_pcie->rxq; in iwl_pcie_rxq_inc_wr_ptr() local
155 lockdep_assert_held(&rxq->lock); in iwl_pcie_rxq_inc_wr_ptr()
171 rxq->need_update = true; in iwl_pcie_rxq_inc_wr_ptr()
176 rxq->write_actual = round_down(rxq->write, 8); in iwl_pcie_rxq_inc_wr_ptr()
177 iwl_write32(trans, FH_RSCSR_CHNL0_WPTR, rxq->write_actual); in iwl_pcie_rxq_inc_wr_ptr()
183 struct iwl_rxq *rxq = &trans_pcie->rxq; in iwl_pcie_rxq_check_wrptr() local
185 spin_lock(&rxq->lock); in iwl_pcie_rxq_check_wrptr()
187 if (!rxq->need_update) in iwl_pcie_rxq_check_wrptr()
[all …]
Dtrans.c1849 struct iwl_rxq *rxq = &trans_pcie->rxq; in iwl_dbgfs_rx_queue_read() local
1855 rxq->read); in iwl_dbgfs_rx_queue_read()
1857 rxq->write); in iwl_dbgfs_rx_queue_read()
1859 rxq->write_actual); in iwl_dbgfs_rx_queue_read()
1861 rxq->need_update); in iwl_dbgfs_rx_queue_read()
1863 rxq->free_count); in iwl_dbgfs_rx_queue_read()
1864 if (rxq->rb_stts) { in iwl_dbgfs_rx_queue_read()
1866 le16_to_cpu(rxq->rb_stts->closed_rb_num) & 0x0FFF); in iwl_dbgfs_rx_queue_read()
Dinternal.h275 struct iwl_rxq rxq; member
/linux-4.1.27/drivers/atm/
Dambassador.c688 amb_rxq * rxq = &dev->rxq[pool]; in rx_give() local
693 spin_lock_irqsave (&rxq->lock, flags); in rx_give()
695 if (rxq->pending < rxq->maximum) { in rx_give()
696 PRINTD (DBG_RX, "RX in slot %p", rxq->in.ptr); in rx_give()
698 *rxq->in.ptr = *rx; in rx_give()
699 rxq->pending++; in rx_give()
700 rxq->in.ptr = NEXTQ (rxq->in.ptr, rxq->in.start, rxq->in.limit); in rx_give()
702 wr_mem (dev, offsetof(amb_mem, mb.adapter.rx_address[pool]), virt_to_bus (rxq->in.ptr)); in rx_give()
704 spin_unlock_irqrestore (&rxq->lock, flags); in rx_give()
707 spin_unlock_irqrestore (&rxq->lock, flags); in rx_give()
[all …]
Dfore200e.c1180 struct host_rxq* rxq = &fore200e->host_rxq; in fore200e_rx_irq() local
1187 entry = &rxq->host_entry[ rxq->head ]; in fore200e_rx_irq()
1218 FORE200E_NEXT_ENTRY(rxq->head, QUEUE_SIZE_RX); in fore200e_rx_irq()
2209 struct host_rxq* rxq = &fore200e->host_rxq; in fore200e_init_rx_queue() local
2217 &rxq->status, in fore200e_init_rx_queue()
2226 &rxq->rpd, in fore200e_init_rx_queue()
2231 fore200e->bus->dma_chunk_free(fore200e, &rxq->status); in fore200e_init_rx_queue()
2241 rxq->host_entry[ i ].status = in fore200e_init_rx_queue()
2242 FORE200E_INDEX(rxq->status.align_addr, enum status, i); in fore200e_init_rx_queue()
2243 rxq->host_entry[ i ].rpd = in fore200e_init_rx_queue()
[all …]
Dambassador.h635 amb_rxq rxq[NUM_RX_POOLS]; member
/linux-4.1.27/drivers/net/ethernet/marvell/
Dmvneta.c576 struct mvneta_rx_queue *rxq, in mvneta_rxq_non_occup_desc_add() argument
583 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), in mvneta_rxq_non_occup_desc_add()
589 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), in mvneta_rxq_non_occup_desc_add()
595 struct mvneta_rx_queue *rxq) in mvneta_rxq_busy_desc_num_get() argument
599 val = mvreg_read(pp, MVNETA_RXQ_STATUS_REG(rxq->id)); in mvneta_rxq_busy_desc_num_get()
607 struct mvneta_rx_queue *rxq, in mvneta_rxq_desc_num_update() argument
615 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), val); in mvneta_rxq_desc_num_update()
635 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), val); in mvneta_rxq_desc_num_update()
641 mvneta_rxq_next_desc_get(struct mvneta_rx_queue *rxq) in mvneta_rxq_next_desc_get() argument
643 int rx_desc = rxq->next_desc_to_proc; in mvneta_rxq_next_desc_get()
[all …]
Dmv643xx_eth.c414 struct rx_queue rxq[8]; member
456 static struct mv643xx_eth_private *rxq_to_mp(struct rx_queue *rxq) in rxq_to_mp() argument
458 return container_of(rxq, struct mv643xx_eth_private, rxq[rxq->index]); in rxq_to_mp()
466 static void rxq_enable(struct rx_queue *rxq) in rxq_enable() argument
468 struct mv643xx_eth_private *mp = rxq_to_mp(rxq); in rxq_enable()
469 wrlp(mp, RXQ_COMMAND, 1 << rxq->index); in rxq_enable()
472 static void rxq_disable(struct rx_queue *rxq) in rxq_disable() argument
474 struct mv643xx_eth_private *mp = rxq_to_mp(rxq); in rxq_disable()
475 u8 mask = 1 << rxq->index; in rxq_disable()
521 static int rxq_process(struct rx_queue *rxq, int budget) in rxq_process() argument
[all …]
Dmvpp2.c46 #define MVPP2_RXQ_CONFIG_REG(rxq) (0x800 + 4 * (rxq)) argument
102 #define MVPP2_RXQ_STATUS_UPDATE_REG(rxq) (0x3000 + 4 * (rxq)) argument
105 #define MVPP2_RXQ_STATUS_REG(rxq) (0x3400 + 4 * (rxq)) argument
154 #define MVPP2_ISR_RX_THRESHOLD_REG(rxq) (0x5200 + 4 * (rxq)) argument
155 #define MVPP2_ISR_RXQ_GROUP_REG(rxq) (0x5400 + 4 * (rxq)) argument
3707 int rxq; in mvpp2_swf_bm_pool_init() local
3721 for (rxq = 0; rxq < rxq_number; rxq++) in mvpp2_swf_bm_pool_init()
3722 mvpp2_rxq_long_pool_set(port, rxq, port->pool_long->id); in mvpp2_swf_bm_pool_init()
3737 for (rxq = 0; rxq < rxq_number; rxq++) in mvpp2_swf_bm_pool_init()
3738 mvpp2_rxq_short_pool_set(port, rxq, in mvpp2_swf_bm_pool_init()
[all …]
Dsky2.c1339 unsigned rxq = rxqaddr[sky2->port]; in sky2_rx_stop() local
1343 sky2_write8(hw, RB_ADDR(rxq, RB_CTRL), RB_DIS_OP_MD); in sky2_rx_stop()
1346 if (sky2_read8(hw, RB_ADDR(rxq, Q_RSL)) in sky2_rx_stop()
1347 == sky2_read8(hw, RB_ADDR(rxq, Q_RL))) in sky2_rx_stop()
1352 sky2_write32(hw, Q_ADDR(rxq, Q_CSR), BMU_RST_SET | BMU_FIFO_RST); in sky2_rx_stop()
1355 sky2_write32(hw, Y2_QADDR(rxq, PREF_UNIT_CTRL), PREF_UNIT_RST_SET); in sky2_rx_stop()
1493 static inline void sky2_rx_update(struct sky2_port *sky2, unsigned rxq) in sky2_rx_update() argument
1495 sky2_put_idx(sky2->hw, rxq, sky2->rx_put); in sky2_rx_update()
1535 unsigned rxq = rxqaddr[sky2->port]; in sky2_rx_start() local
1539 sky2_qset(hw, rxq); in sky2_rx_start()
[all …]
/linux-4.1.27/drivers/net/wireless/iwlegacy/
D3945-mac.c956 struct il_rx_queue *rxq = &il->rxq; in il3945_rx_queue_restock() local
962 spin_lock_irqsave(&rxq->lock, flags); in il3945_rx_queue_restock()
963 write = rxq->write & ~0x7; in il3945_rx_queue_restock()
964 while (il_rx_queue_space(rxq) > 0 && rxq->free_count) { in il3945_rx_queue_restock()
966 element = rxq->rx_free.next; in il3945_rx_queue_restock()
971 rxq->bd[rxq->write] = in il3945_rx_queue_restock()
973 rxq->queue[rxq->write] = rxb; in il3945_rx_queue_restock()
974 rxq->write = (rxq->write + 1) & RX_QUEUE_MASK; in il3945_rx_queue_restock()
975 rxq->free_count--; in il3945_rx_queue_restock()
977 spin_unlock_irqrestore(&rxq->lock, flags); in il3945_rx_queue_restock()
[all …]
D4965-mac.c100 il4965_rx_queue_reset(struct il_priv *il, struct il_rx_queue *rxq) in il4965_rx_queue_reset() argument
104 spin_lock_irqsave(&rxq->lock, flags); in il4965_rx_queue_reset()
105 INIT_LIST_HEAD(&rxq->rx_free); in il4965_rx_queue_reset()
106 INIT_LIST_HEAD(&rxq->rx_used); in il4965_rx_queue_reset()
111 if (rxq->pool[i].page != NULL) { in il4965_rx_queue_reset()
112 pci_unmap_page(il->pci_dev, rxq->pool[i].page_dma, in il4965_rx_queue_reset()
115 __il_free_pages(il, rxq->pool[i].page); in il4965_rx_queue_reset()
116 rxq->pool[i].page = NULL; in il4965_rx_queue_reset()
118 list_add_tail(&rxq->pool[i].list, &rxq->rx_used); in il4965_rx_queue_reset()
122 rxq->queue[i] = NULL; in il4965_rx_queue_reset()
[all …]
D4965.h58 void il4965_rx_queue_reset(struct il_priv *il, struct il_rx_queue *rxq);
59 int il4965_rx_init(struct il_priv *il, struct il_rx_queue *rxq);
69 void il4965_rx_queue_free(struct il_priv *il, struct il_rx_queue *rxq);
D3945.c793 il3945_rx_init(struct il_priv *il, struct il_rx_queue *rxq) in il3945_rx_init() argument
795 il_wr(il, FH39_RCSR_RBD_BASE(0), rxq->bd_dma); in il3945_rx_init()
796 il_wr(il, FH39_RCSR_RPTR_ADDR(0), rxq->rb_stts_dma); in il3945_rx_init()
972 struct il_rx_queue *rxq = &il->rxq; in il3945_hw_nic_init() local
982 if (!rxq->bd) { in il3945_hw_nic_init()
989 il3945_rx_queue_reset(il, rxq); in il3945_hw_nic_init()
993 il3945_rx_init(il, rxq); in il3945_hw_nic_init()
1000 il_wr(il, FH39_RCSR_WPTR(0), rxq->write & ~7); in il3945_hw_nic_init()
Dcommon.c2614 struct il_rx_queue *rxq = &il->rxq; in il_rx_queue_alloc() local
2618 spin_lock_init(&rxq->lock); in il_rx_queue_alloc()
2619 INIT_LIST_HEAD(&rxq->rx_free); in il_rx_queue_alloc()
2620 INIT_LIST_HEAD(&rxq->rx_used); in il_rx_queue_alloc()
2623 rxq->bd = dma_alloc_coherent(dev, 4 * RX_QUEUE_SIZE, &rxq->bd_dma, in il_rx_queue_alloc()
2625 if (!rxq->bd) in il_rx_queue_alloc()
2628 rxq->rb_stts = dma_alloc_coherent(dev, sizeof(struct il_rb_status), in il_rx_queue_alloc()
2629 &rxq->rb_stts_dma, GFP_KERNEL); in il_rx_queue_alloc()
2630 if (!rxq->rb_stts) in il_rx_queue_alloc()
2635 list_add_tail(&rxq->pool[i].list, &rxq->rx_used); in il_rx_queue_alloc()
[all …]
Ddebug.c899 struct il_rx_queue *rxq = &il->rxq; in il_dbgfs_rx_queue_read() local
904 pos += scnprintf(buf + pos, bufsz - pos, "read: %u\n", rxq->read); in il_dbgfs_rx_queue_read()
905 pos += scnprintf(buf + pos, bufsz - pos, "write: %u\n", rxq->write); in il_dbgfs_rx_queue_read()
908 rxq->free_count); in il_dbgfs_rx_queue_read()
909 if (rxq->rb_stts) { in il_dbgfs_rx_queue_read()
912 le16_to_cpu(rxq->rb_stts-> in il_dbgfs_rx_queue_read()
D3945.h194 void il3945_rx_queue_reset(struct il_priv *il, struct il_rx_queue *rxq);
Dcommon.h1270 struct il_rx_queue rxq; member
/linux-4.1.27/drivers/net/ethernet/atheros/alx/
Dmain.c75 struct alx_rx_queue *rxq = &alx->rxq; in alx_refill_rx_ring() local
81 next = cur = rxq->write_idx; in alx_refill_rx_ring()
84 cur_buf = &rxq->bufs[cur]; in alx_refill_rx_ring()
86 while (!cur_buf->skb && next != rxq->read_idx) { in alx_refill_rx_ring()
87 struct alx_rfd *rfd = &rxq->rfd[cur]; in alx_refill_rx_ring()
116 cur_buf = &rxq->bufs[cur]; in alx_refill_rx_ring()
123 rxq->write_idx = cur; in alx_refill_rx_ring()
189 struct alx_rx_queue *rxq = &alx->rxq; in alx_clean_rx_irq() local
197 rrd = &rxq->rrd[rxq->rrd_read_idx]; in alx_clean_rx_irq()
203 RRD_SI) != rxq->read_idx || in alx_clean_rx_irq()
[all …]
Dalx.h101 struct alx_rx_queue rxq; member
Dhw.c378 u32 rxq, txq, val; in alx_stop_mac() local
381 rxq = alx_read_mem32(hw, ALX_RXQ0); in alx_stop_mac()
382 alx_write_mem32(hw, ALX_RXQ0, rxq & ~ALX_RXQ0_EN); in alx_stop_mac()
609 u32 mac, txq, rxq; in alx_start_mac() local
611 rxq = alx_read_mem32(hw, ALX_RXQ0); in alx_start_mac()
612 alx_write_mem32(hw, ALX_RXQ0, rxq | ALX_RXQ0_EN); in alx_start_mac()
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4vf/
Dcxgb4vf_main.c324 int rxq, msi, err; in request_msix_queue_irqs() local
338 for_each_ethrxq(s, rxq) { in request_msix_queue_irqs()
342 &s->ethrxq[rxq].rspq); in request_msix_queue_irqs()
350 while (--rxq >= 0) in request_msix_queue_irqs()
351 free_irq(adapter->msix_info[--msi].vec, &s->ethrxq[rxq].rspq); in request_msix_queue_irqs()
362 int rxq, msi; in free_msix_queue_irqs() local
366 for_each_ethrxq(s, rxq) in free_msix_queue_irqs()
368 &s->ethrxq[rxq].rspq); in free_msix_queue_irqs()
393 int rxq; in enable_rx() local
396 for_each_ethrxq(s, rxq) in enable_rx()
[all …]
Dsge.c1551 static void do_gro(struct sge_eth_rxq *rxq, const struct pkt_gl *gl, in do_gro() argument
1554 struct adapter *adapter = rxq->rspq.adapter; in do_gro()
1559 skb = napi_get_frags(&rxq->rspq.napi); in do_gro()
1562 rxq->stats.rx_drops++; in do_gro()
1571 skb_record_rx_queue(skb, rxq->rspq.idx); in do_gro()
1576 rxq->stats.vlan_ex++; in do_gro()
1578 ret = napi_gro_frags(&rxq->rspq.napi); in do_gro()
1581 rxq->stats.lro_pkts++; in do_gro()
1583 rxq->stats.lro_merged++; in do_gro()
1584 rxq->stats.pkts++; in do_gro()
[all …]
/linux-4.1.27/drivers/net/ethernet/broadcom/bnx2x/
Dbnx2x_vfpf.c618 req->rxq.rcq_addr = fp->rx_comp_mapping; in bnx2x_vfpf_setup_q()
619 req->rxq.rcq_np_addr = fp->rx_comp_mapping + BCM_PAGE_SIZE; in bnx2x_vfpf_setup_q()
620 req->rxq.rxq_addr = fp->rx_desc_mapping; in bnx2x_vfpf_setup_q()
621 req->rxq.sge_addr = fp->rx_sge_mapping; in bnx2x_vfpf_setup_q()
622 req->rxq.vf_sb = fp_idx; in bnx2x_vfpf_setup_q()
623 req->rxq.sb_index = HC_INDEX_ETH_RX_CQ_CONS; in bnx2x_vfpf_setup_q()
624 req->rxq.hc_rate = bp->rx_ticks ? 1000000/bp->rx_ticks : 0; in bnx2x_vfpf_setup_q()
625 req->rxq.mtu = bp->dev->mtu; in bnx2x_vfpf_setup_q()
626 req->rxq.buf_sz = fp->rx_buf_size; in bnx2x_vfpf_setup_q()
627 req->rxq.sge_buf_sz = BCM_PAGE_SIZE * PAGES_PER_SGE; in bnx2x_vfpf_setup_q()
[all …]
Dbnx2x_vfpf.h262 } rxq; member
Dbnx2x_sriov.c1954 struct bnx2x_vf_queue *rxq = vfq_get(vf, j); in bnx2x_iov_adjust_stats_req() local
1960 if (bnx2x_get_q_logical_state(bp, &rxq->sp_obj) == in bnx2x_iov_adjust_stats_req()
1966 cur_query_entry->index = vfq_stat_id(vf, rxq); in bnx2x_iov_adjust_stats_req()
/linux-4.1.27/drivers/net/ethernet/samsung/sxgbe/
Dsxgbe_main.c330 priv->hw->desc->init_rx_desc(&priv->rxq[j]->dma_rx[i], in sxgbe_clear_descriptors()
588 priv->rxq[queue_num], rx_rsize); in init_dma_desc_rings()
597 priv->rxq[queue_num]->priv_ptr = priv; in init_dma_desc_rings()
611 free_rx_ring(priv->device, priv->rxq[queue_num], rx_rsize); in init_dma_desc_rings()
663 free_rx_ring(priv->device, priv->rxq[queue_num], rx_rsize); in free_dma_desc_resources()
686 priv->rxq[queue_num] = devm_kmalloc(priv->device, in rxring_mem_alloc()
688 if (!priv->rxq[queue_num]) in rxring_mem_alloc()
973 (priv->rxq[queue_num])->dma_rx_phy, in sxgbe_init_dma_engine()
1148 (priv->rxq[queue_num])->irq_no, in sxgbe_open()
1150 dev->name, priv->rxq[queue_num]); in sxgbe_open()
[all …]
Dsxgbe_platform.c138 priv->rxq[i]->irq_no = irq_of_parse_and_map(node, chan++); in sxgbe_platform_probe()
139 if (priv->rxq[i]->irq_no <= 0) { in sxgbe_platform_probe()
159 irq_dispose_mapping(priv->rxq[i]->irq_no); in sxgbe_platform_probe()
Dsxgbe_common.h459 struct sxgbe_rx_queue *rxq[SXGBE_RX_QUEUES]; member
/linux-4.1.27/drivers/net/ethernet/freescale/
Dfec_main.c227 struct fec_enet_priv_rx_q *rxq = fep->rx_queue[queue_id]; in fec_enet_get_nextdesc() local
237 base = rxq->rx_bd_base; in fec_enet_get_nextdesc()
238 ring_size = rxq->rx_ring_size; in fec_enet_get_nextdesc()
239 ex_base = (struct bufdesc_ex *)rxq->rx_bd_base; in fec_enet_get_nextdesc()
258 struct fec_enet_priv_rx_q *rxq = fep->rx_queue[queue_id]; in fec_enet_get_prevdesc() local
268 base = rxq->rx_bd_base; in fec_enet_get_prevdesc()
269 ring_size = rxq->rx_ring_size; in fec_enet_get_prevdesc()
270 ex_base = (struct bufdesc_ex *)rxq->rx_bd_base; in fec_enet_get_prevdesc()
807 struct fec_enet_priv_rx_q *rxq; in fec_enet_bd_init() local
814 rxq = fep->rx_queue[q]; in fec_enet_bd_init()
[all …]
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4/
Dsge.c1796 static void do_gro(struct sge_eth_rxq *rxq, const struct pkt_gl *gl, in do_gro() argument
1799 struct adapter *adapter = rxq->rspq.adap; in do_gro()
1804 skb = napi_get_frags(&rxq->rspq.napi); in do_gro()
1807 rxq->stats.rx_drops++; in do_gro()
1816 skb_record_rx_queue(skb, rxq->rspq.idx); in do_gro()
1817 skb_mark_napi_id(skb, &rxq->rspq.napi); in do_gro()
1818 if (rxq->rspq.netdev->features & NETIF_F_RXHASH) in do_gro()
1824 rxq->stats.vlan_ex++; in do_gro()
1826 ret = napi_gro_frags(&rxq->rspq.napi); in do_gro()
1828 rxq->stats.lro_pkts++; in do_gro()
[all …]
Dcxgb4_main.c683 struct sge_ofld_rxq *rxq = container_of(q, struct sge_ofld_rxq, rspq); in uldrx_handler() local
692 rxq->stats.nomem++; in uldrx_handler()
696 rxq->stats.imm++; in uldrx_handler()
698 rxq->stats.an++; in uldrx_handler()
700 rxq->stats.pkts++; in uldrx_handler()
Dcxgb4.h1271 unsigned int rxqi, unsigned int rxq, unsigned int tc,
Dt4_hw.c4461 unsigned int rxqi, unsigned int rxq, unsigned int tc, in t4_cfg_pfvf() argument
4473 FW_PFVF_CMD_NIQ_V(rxq)); in t4_cfg_pfvf()
/linux-4.1.27/drivers/net/ethernet/brocade/bna/
Dbnad_ethtool.c668 rcb[1]->rxq) { in bnad_get_strings()
750 bnad->rx_info[i].rx_ctrl[j].ccb->rcb[1]->rxq) in bnad_get_stats_count_locked()
775 bnad->rx_info[i].rx_ctrl[j].ccb->rcb[0]->rxq) { in bnad_per_q_stats_fill()
801 rcb[0]->rxq) { in bnad_per_q_stats_fill()
804 buf[bi++] = rcb->rxq->rx_packets; in bnad_per_q_stats_fill()
805 buf[bi++] = rcb->rxq->rx_bytes; in bnad_per_q_stats_fill()
806 buf[bi++] = rcb->rxq-> in bnad_per_q_stats_fill()
808 buf[bi++] = rcb->rxq-> in bnad_per_q_stats_fill()
815 rcb[1]->rxq) { in bnad_per_q_stats_fill()
818 buf[bi++] = rcb->rxq->rx_packets; in bnad_per_q_stats_fill()
[all …]
Dbna_tx_rx.c1959 struct bna_rxq *rxq = NULL; in bna_rxq_get() local
1964 rxq = (struct bna_rxq *)qe; in bna_rxq_get()
1965 bfa_q_qe_init(&rxq->qe); in bna_rxq_get()
1967 return rxq; in bna_rxq_get()
1971 bna_rxq_put(struct bna_rx_mod *rx_mod, struct bna_rxq *rxq) in bna_rxq_put() argument
1973 bfa_q_qe_init(&rxq->qe); in bna_rxq_put()
1974 list_add_tail(&rxq->qe, &rx_mod->rxq_free_q); in bna_rxq_put()
2058 rxp->rxq.single.only = q0; in bna_rxp_add_rxqs()
2059 rxp->rxq.single.reserved = NULL; in bna_rxp_add_rxqs()
2062 rxp->rxq.slr.large = q0; in bna_rxp_add_rxqs()
[all …]
Dbna.h310 (q0) = rxp->rxq.single.only; \
314 (q0) = rxp->rxq.slr.large; \
315 (q1) = rxp->rxq.slr.small; \
318 (q0) = rxp->rxq.hds.data; \
319 (q1) = rxp->rxq.hds.hdr; \
Dbna_types.h574 struct bna_rxq *rxq; member
711 union bna_rxq_u rxq; member
858 struct bna_rxq *rxq; /* BFI_MAX_RXQ entries */ member
Dbnad.c291 order = get_order(rcb->rxq->buffer_size); in bnad_rxq_alloc_init()
297 unmap_q->map_size = rcb->rxq->buffer_size; in bnad_rxq_alloc_init()
299 if (rcb->rxq->multi_buffer) { in bnad_rxq_alloc_init()
301 unmap_q->map_size = rcb->rxq->buffer_size; in bnad_rxq_alloc_init()
306 (rcb->rxq->buffer_size > 2048) ? in bnad_rxq_alloc_init()
396 rcb->rxq->rxbuf_alloc_failed++; in bnad_rxq_refill_page()
441 buff_sz = rcb->rxq->buffer_size; in bnad_rxq_refill_skb()
453 rcb->rxq->rxbuf_alloc_failed++; in bnad_rxq_refill_skb()
685 rcb->rxq->rx_packets_with_error++; in bnad_cq_process()
695 rcb->rxq->rx_packets++; in bnad_cq_process()
[all …]
/linux-4.1.27/drivers/net/wireless/
Dmwl8k.c247 struct mwl8k_rx_queue rxq[MWL8K_RX_QUEUES]; member
1152 struct mwl8k_rx_queue *rxq = priv->rxq + index; in mwl8k_rxq_init() local
1156 rxq->rxd_count = 0; in mwl8k_rxq_init()
1157 rxq->head = 0; in mwl8k_rxq_init()
1158 rxq->tail = 0; in mwl8k_rxq_init()
1162 rxq->rxd = pci_zalloc_consistent(priv->pdev, size, &rxq->rxd_dma); in mwl8k_rxq_init()
1163 if (rxq->rxd == NULL) { in mwl8k_rxq_init()
1168 rxq->buf = kcalloc(MWL8K_RX_DESCS, sizeof(*rxq->buf), GFP_KERNEL); in mwl8k_rxq_init()
1169 if (rxq->buf == NULL) { in mwl8k_rxq_init()
1170 pci_free_consistent(priv->pdev, size, rxq->rxd, rxq->rxd_dma); in mwl8k_rxq_init()
[all …]
/linux-4.1.27/drivers/net/ethernet/xscale/
Dixp4xx_eth.c678 qmgr_disable_irq(port->plat->rxq); in eth_rx_irq()
686 unsigned int rxq = port->plat->rxq, rxfreeq = RXFREE_QUEUE(port->id); in eth_poll() local
702 if ((n = queue_get_desc(rxq, port, 0)) < 0) { in eth_poll()
708 qmgr_enable_irq(rxq); in eth_poll()
709 if (!qmgr_stat_below_low_watermark(rxq) && in eth_poll()
716 qmgr_disable_irq(rxq); in eth_poll()
1074 err = qmgr_request_queue(port->plat->rxq, RX_DESCS, 0, 0, in request_queues()
1103 qmgr_release_queue(port->plat->rxq); in request_queues()
1114 qmgr_release_queue(port->plat->rxq); in release_queues()
1227 msg.byte5 = port->plat->rxq | 0x80; in eth_open()
[all …]
/linux-4.1.27/drivers/net/ethernet/intel/i40evf/
Di40evf_virtchnl.c243 vqpi->rxq.vsi_id = vqci->vsi_id; in i40evf_configure_queues()
244 vqpi->rxq.queue_id = i; in i40evf_configure_queues()
245 vqpi->rxq.ring_len = adapter->rx_rings[i]->count; in i40evf_configure_queues()
246 vqpi->rxq.dma_ring_addr = adapter->rx_rings[i]->dma; in i40evf_configure_queues()
247 vqpi->rxq.max_pkt_size = adapter->netdev->mtu in i40evf_configure_queues()
249 vqpi->rxq.databuffer_size = adapter->rx_rings[i]->rx_buf_len; in i40evf_configure_queues()
Di40e_virtchnl.h208 struct i40e_virtchnl_rxq_info rxq; member
/linux-4.1.27/drivers/net/ethernet/ibm/
Dibmveth.h50 #define h_register_logical_lan(ua, buflst, rxq, fltlst, mac) \ argument
51 plpar_hcall_norets(H_REGISTER_LOGICAL_LAN, ua, buflst, rxq, fltlst, mac)
/linux-4.1.27/drivers/net/usb/
Dusbnet.c491 spin_lock_irqsave (&dev->rxq.lock, lockflags); in rx_submit()
517 __usbnet_queue_skb(&dev->rxq, skb, rx_start); in rx_submit()
523 spin_unlock_irqrestore (&dev->rxq.lock, lockflags); in rx_submit()
640 state = defer_bh(dev, skb, &dev->rxq, state); in rx_complete()
744 (void) unlink_urbs (dev, &dev->rxq); in usbnet_unlink_rx_urbs()
762 unlink_urbs(dev, &dev->rxq); in usbnet_terminate_urbs()
765 while (!skb_queue_empty(&dev->rxq) in usbnet_terminate_urbs()
1038 unlink_urbs(dev, &dev->rxq); in __handle_link_change()
1104 unlink_urbs (dev, &dev->rxq); in usbnet_deferred_kevent()
1432 for (i = 0; i < 10 && dev->rxq.qlen < RX_QLEN(dev); i++) { in rx_alloc_submit()
[all …]
/linux-4.1.27/arch/arm/mach-ixp4xx/
Dvulcan-setup.c126 .rxq = 3,
131 .rxq = 4,
Domixp-setup.c177 .rxq = 3,
181 .rxq = 4,
Dfsg-setup.c130 .rxq = 3,
134 .rxq = 4,
Dixdp425-setup.c185 .rxq = 3,
189 .rxq = 4,
Dgoramo_mlr.c284 .rxq = 3,
288 .rxq = 4,
Dnslu2-setup.c175 .rxq = 3,
Dnas100d-setup.c163 .rxq = 3,
/linux-4.1.27/drivers/mailbox/
Domap-mailbox.c114 struct omap_mbox_queue *rxq; member
336 struct omap_mbox_queue *mq = mbox->rxq; in __mbox_rx_interrupt()
356 schedule_work(&mbox->rxq->work); in __mbox_rx_interrupt()
411 mbox->rxq = mq; in omap_mbox_startup()
426 mbox_queue_free(mbox->rxq); in omap_mbox_startup()
434 flush_work(&mbox->rxq->work); in omap_mbox_fini()
435 mbox_queue_free(mbox->rxq); in omap_mbox_fini()
/linux-4.1.27/drivers/net/wireless/ipw2x00/
Dipw2200.c3451 struct ipw_rx_queue *rxq) in ipw_rx_queue_reset() argument
3456 spin_lock_irqsave(&rxq->lock, flags); in ipw_rx_queue_reset()
3458 INIT_LIST_HEAD(&rxq->rx_free); in ipw_rx_queue_reset()
3459 INIT_LIST_HEAD(&rxq->rx_used); in ipw_rx_queue_reset()
3465 if (rxq->pool[i].skb != NULL) { in ipw_rx_queue_reset()
3466 pci_unmap_single(priv->pci_dev, rxq->pool[i].dma_addr, in ipw_rx_queue_reset()
3468 dev_kfree_skb(rxq->pool[i].skb); in ipw_rx_queue_reset()
3469 rxq->pool[i].skb = NULL; in ipw_rx_queue_reset()
3471 list_add_tail(&rxq->pool[i].list, &rxq->rx_used); in ipw_rx_queue_reset()
3476 rxq->read = rxq->write = 0; in ipw_rx_queue_reset()
[all …]
Dipw2100.c2675 struct ipw2100_bd_queue *rxq = &priv->rx_queue; in __ipw2100_rx_process() local
2688 if (r >= rxq->entries) { in __ipw2100_rx_process()
2693 i = (rxq->next + 1) % rxq->entries; in __ipw2100_rx_process()
2765 rxq->drv[i].status.info.field = 0; in __ipw2100_rx_process()
2767 i = (i + 1) % rxq->entries; in __ipw2100_rx_process()
2772 rxq->next = (i ? i : rxq->entries) - 1; in __ipw2100_rx_process()
2775 IPW_MEM_HOST_SHARED_RX_WRITE_INDEX, rxq->next); in __ipw2100_rx_process()
Dipw2200.h1205 struct ipw_rx_queue *rxq; member
/linux-4.1.27/drivers/net/
Dmacvtap.c231 __u32 rxq; in macvtap_get_queue() local
237 rxq = skb_get_hash(skb); in macvtap_get_queue()
238 if (rxq) { in macvtap_get_queue()
239 tap = rcu_dereference(vlan->taps[rxq % numvtaps]); in macvtap_get_queue()
244 rxq = skb_get_rx_queue(skb); in macvtap_get_queue()
246 while (unlikely(rxq >= numvtaps)) in macvtap_get_queue()
247 rxq -= numvtaps; in macvtap_get_queue()
249 tap = rcu_dereference(vlan->taps[rxq]); in macvtap_get_queue()
Dxen-netfront.c877 struct sk_buff_head *rxq) in handle_incoming_queue() argument
883 while ((skb = __skb_dequeue(rxq)) != NULL) { in handle_incoming_queue()
922 struct sk_buff_head rxq; in xennet_poll() local
929 skb_queue_head_init(&rxq); in xennet_poll()
982 __skb_queue_tail(&rxq, skb); in xennet_poll()
990 work_done -= handle_incoming_queue(queue, &rxq); in xennet_poll()
Dvirtio_net.c173 static int rxq2vq(int rxq) in rxq2vq() argument
175 return rxq * 2; in rxq2vq()
/linux-4.1.27/arch/arm/mach-ixp4xx/include/mach/
Dplatform.h101 u8 rxq; /* configurable, currently 0 - 31 only */ member
/linux-4.1.27/drivers/net/ethernet/sfc/
Dvfdi.h198 u32 rxq; member
Dptp.c272 struct sk_buff_head rxq; member
998 while ((skb = skb_dequeue(&ptp->rxq))) { in efx_ptp_process_events()
1013 skb_queue_head(&ptp->rxq, skb); in efx_ptp_process_events()
1129 efx_ptp_deliver_rx_queue(&efx->ptp_data->rxq); in efx_ptp_stop()
1226 skb_queue_head_init(&ptp->rxq); in efx_ptp_probe()
1315 skb_queue_purge(&efx->ptp_data->rxq); in efx_ptp_remove()
1439 skb_queue_tail(&ptp->rxq, skb); in efx_ptp_rx()
Dsiena_sriov.c765 unsigned vf_rxq = req->u.mac_filter.rxq; in efx_vfdi_insert_filter()
/linux-4.1.27/Documentation/devicetree/bindings/net/
Dbrcm,systemport.txt19 - systemport,num-rxq: number of HW receive queues, an integer
/linux-4.1.27/drivers/net/ethernet/intel/i40e/
Di40e_virtchnl.h208 struct i40e_virtchnl_rxq_info rxq; member
Di40e_virtchnl_pf.c1281 (qpi->rxq.vsi_id != vsi_id) || in i40e_vc_config_queues_msg()
1282 (qpi->rxq.queue_id != vsi_queue_id) || in i40e_vc_config_queues_msg()
1289 &qpi->rxq) || in i40e_vc_config_queues_msg()
/linux-4.1.27/drivers/net/wan/
Dixp4xx_hss.c658 unsigned int rxq = queue_ids[port->id].rx; in hss_hdlc_poll() local
675 if ((n = queue_get_desc(rxq, port, 0)) < 0) { in hss_hdlc_poll()
681 qmgr_enable_irq(rxq); in hss_hdlc_poll()
682 if (!qmgr_stat_empty(rxq) && in hss_hdlc_poll()
689 qmgr_disable_irq(rxq); in hss_hdlc_poll()
/linux-4.1.27/include/linux/usb/
Dusbnet.h57 struct sk_buff_head rxq; member
/linux-4.1.27/drivers/net/ethernet/emulex/benet/
Dbe_main.c1748 struct be_queue_info *rxq = &rxo->q; in get_rx_page_info() local
1749 u16 frag_idx = rxq->tail; in get_rx_page_info()
1765 queue_tail_inc(rxq); in get_rx_page_info()
1766 atomic_dec(&rxq->used); in get_rx_page_info()
2060 struct be_queue_info *rxq = &rxo->q; in be_post_rx_frags() local
2067 page_info = &rxo->page_info_tbl[rxq->head]; in be_post_rx_frags()
2092 rxd = queue_head_node(rxq); in be_post_rx_frags()
2108 queue_head_inc(rxq); in be_post_rx_frags()
2109 page_info = &rxo->page_info_tbl[rxq->head]; in be_post_rx_frags()
2121 atomic_add(posted, &rxq->used); in be_post_rx_frags()
[all …]
Dbe_cmds.c1311 struct be_queue_info *rxq, u16 cq_id, u16 frag_size, in be_cmd_rxq_create() argument
1316 struct be_dma_mem *q_mem = &rxq->dma_mem; in be_cmd_rxq_create()
1343 rxq->id = le16_to_cpu(resp->id); in be_cmd_rxq_create()
1344 rxq->created = true; in be_cmd_rxq_create()
Dbe_cmds.h2255 int be_cmd_rxq_create(struct be_adapter *adapter, struct be_queue_info *rxq,
/linux-4.1.27/drivers/net/wireless/ath/ath6kl/
Dhtc_mbox.c1913 struct list_head *rxq, in ath6kl_htc_rx_bundle() argument
1922 n_scat_pkt = get_queue_depth(rxq); in ath6kl_htc_rx_bundle()
1925 if ((get_queue_depth(rxq) - n_scat_pkt) > 0) { in ath6kl_htc_rx_bundle()
1938 __func__, get_queue_depth(rxq), n_scat_pkt); in ath6kl_htc_rx_bundle()
1945 get_queue_depth(rxq), n_scat_pkt); in ath6kl_htc_rx_bundle()
1955 packet = list_first_entry(rxq, struct htc_packet, list); in ath6kl_htc_rx_bundle()
1962 list_add(&packet->list, rxq); in ath6kl_htc_rx_bundle()
/linux-4.1.27/drivers/net/ethernet/broadcom/
Dbcmsysport.c1665 u32 txq, rxq; in bcm_sysport_probe() local
1674 if (of_property_read_u32(dn, "systemport,num-rxq", &rxq)) in bcm_sysport_probe()
1675 rxq = 1; in bcm_sysport_probe()
1677 dev = alloc_etherdev_mqs(sizeof(*priv), txq, rxq); in bcm_sysport_probe()
1764 priv->base, priv->irq0, priv->irq1, txq, rxq); in bcm_sysport_probe()
/linux-4.1.27/drivers/net/xen-netback/
Dnetback.c502 struct sk_buff_head rxq; in xenvif_rx_action() local
514 skb_queue_head_init(&rxq); in xenvif_rx_action()
527 __skb_queue_tail(&rxq, skb); in xenvif_rx_action()
538 while ((skb = __skb_dequeue(&rxq)) != NULL) { in xenvif_rx_action()
/linux-4.1.27/net/core/
Dnet-sysfs.c1321 int error = 0, txq = 0, rxq = 0, real_rx = 0, real_tx = 0; in register_queue_kobjects() local
1335 rxq = real_rx; in register_queue_kobjects()
1346 net_rx_queue_update_kobjects(dev, rxq, 0); in register_queue_kobjects()
Ddev.c2171 int netif_set_real_num_rx_queues(struct net_device *dev, unsigned int rxq) in netif_set_real_num_rx_queues() argument
2175 if (rxq < 1 || rxq > dev->num_rx_queues) in netif_set_real_num_rx_queues()
2182 rxq); in netif_set_real_num_rx_queues()
2187 dev->real_num_rx_queues = rxq; in netif_set_real_num_rx_queues()
3529 static int ing_filter(struct sk_buff *skb, struct netdev_queue *rxq) in ing_filter() argument
3545 q = rcu_dereference(rxq->qdisc); in ing_filter()
3560 struct netdev_queue *rxq = rcu_dereference(skb->dev->ingress_queue); in handle_ing() local
3562 if (!rxq || rcu_access_pointer(rxq->qdisc) == &noop_qdisc) in handle_ing()
3570 switch (ing_filter(skb, rxq)) { in handle_ing()
/linux-4.1.27/drivers/infiniband/hw/cxgb3/
Diwch_cm.c107 static struct sk_buff_head rxq; variable
2187 while ((skb = skb_dequeue(&rxq))) { in process_work()
2218 skb_queue_tail(&rxq, skb); in sched()
2259 skb_queue_head_init(&rxq); in iwch_cm_init()
/linux-4.1.27/drivers/net/ethernet/atheros/atl1c/
Datl1c_main.c1198 u32 mac, txq, rxq; in atl1c_start_mac() local
1205 AT_READ_REG(hw, REG_RXQ_CTRL, &rxq); in atl1c_start_mac()
1209 rxq |= RXQ_CTRL_EN; in atl1c_start_mac()
1223 AT_WRITE_REG(hw, REG_RXQ_CTRL, rxq); in atl1c_start_mac()
/linux-4.1.27/drivers/infiniband/hw/cxgb4/
Dcm.c142 static struct sk_buff_head rxq; variable
3876 while ((skb = skb_dequeue(&rxq))) { in process_work()
3925 skb_queue_tail(&rxq, skb); in sched()
4039 skb_queue_head_init(&rxq); in c4iw_cm_init()
/linux-4.1.27/include/linux/
Dnetdevice.h2867 int netif_set_real_num_rx_queues(struct net_device *dev, unsigned int rxq);
2870 unsigned int rxq) in netif_set_real_num_rx_queues() argument