Home
last modified time | relevance | path

Searched refs:rxq (Results 1 – 92 of 92) sorted by relevance

/linux-4.4.14/drivers/net/wireless/iwlwifi/pcie/
Drx.c141 static int iwl_rxq_space(const struct iwl_rxq *rxq) in iwl_rxq_space() argument
152 return (rxq->read - rxq->write - 1) & (RX_QUEUE_SIZE - 1); in iwl_rxq_space()
179 struct iwl_rxq *rxq = &trans_pcie->rxq; in iwl_pcie_rxq_inc_wr_ptr() local
182 lockdep_assert_held(&rxq->lock); in iwl_pcie_rxq_inc_wr_ptr()
198 rxq->need_update = true; in iwl_pcie_rxq_inc_wr_ptr()
203 rxq->write_actual = round_down(rxq->write, 8); in iwl_pcie_rxq_inc_wr_ptr()
204 iwl_write32(trans, FH_RSCSR_CHNL0_WPTR, rxq->write_actual); in iwl_pcie_rxq_inc_wr_ptr()
210 struct iwl_rxq *rxq = &trans_pcie->rxq; in iwl_pcie_rxq_check_wrptr() local
212 spin_lock(&rxq->lock); in iwl_pcie_rxq_check_wrptr()
214 if (!rxq->need_update) in iwl_pcie_rxq_check_wrptr()
[all …]
Dtrans.c1969 struct iwl_rxq *rxq = &trans_pcie->rxq; in iwl_dbgfs_rx_queue_read() local
1975 rxq->read); in iwl_dbgfs_rx_queue_read()
1977 rxq->write); in iwl_dbgfs_rx_queue_read()
1979 rxq->write_actual); in iwl_dbgfs_rx_queue_read()
1981 rxq->need_update); in iwl_dbgfs_rx_queue_read()
1983 rxq->free_count); in iwl_dbgfs_rx_queue_read()
1984 if (rxq->rb_stts) { in iwl_dbgfs_rx_queue_read()
1986 le16_to_cpu(rxq->rb_stts->closed_rb_num) & 0x0FFF); in iwl_dbgfs_rx_queue_read()
2302 struct iwl_rxq *rxq = &trans_pcie->rxq; in iwl_trans_pcie_dump_rbs() local
2305 spin_lock(&rxq->lock); in iwl_trans_pcie_dump_rbs()
[all …]
Dinternal.h318 struct iwl_rxq rxq; member
/linux-4.4.14/drivers/atm/
Dambassador.c688 amb_rxq * rxq = &dev->rxq[pool]; in rx_give() local
693 spin_lock_irqsave (&rxq->lock, flags); in rx_give()
695 if (rxq->pending < rxq->maximum) { in rx_give()
696 PRINTD (DBG_RX, "RX in slot %p", rxq->in.ptr); in rx_give()
698 *rxq->in.ptr = *rx; in rx_give()
699 rxq->pending++; in rx_give()
700 rxq->in.ptr = NEXTQ (rxq->in.ptr, rxq->in.start, rxq->in.limit); in rx_give()
702 wr_mem (dev, offsetof(amb_mem, mb.adapter.rx_address[pool]), virt_to_bus (rxq->in.ptr)); in rx_give()
704 spin_unlock_irqrestore (&rxq->lock, flags); in rx_give()
707 spin_unlock_irqrestore (&rxq->lock, flags); in rx_give()
[all …]
Dfore200e.c1180 struct host_rxq* rxq = &fore200e->host_rxq; in fore200e_rx_irq() local
1187 entry = &rxq->host_entry[ rxq->head ]; in fore200e_rx_irq()
1218 FORE200E_NEXT_ENTRY(rxq->head, QUEUE_SIZE_RX); in fore200e_rx_irq()
2209 struct host_rxq* rxq = &fore200e->host_rxq; in fore200e_init_rx_queue() local
2217 &rxq->status, in fore200e_init_rx_queue()
2226 &rxq->rpd, in fore200e_init_rx_queue()
2231 fore200e->bus->dma_chunk_free(fore200e, &rxq->status); in fore200e_init_rx_queue()
2241 rxq->host_entry[ i ].status = in fore200e_init_rx_queue()
2242 FORE200E_INDEX(rxq->status.align_addr, enum status, i); in fore200e_init_rx_queue()
2243 rxq->host_entry[ i ].rpd = in fore200e_init_rx_queue()
[all …]
Dambassador.h635 amb_rxq rxq[NUM_RX_POOLS]; member
/linux-4.4.14/drivers/net/ethernet/qlogic/qede/
Dqede_main.c89 struct qede_rx_queue *rxq);
699 static bool qede_has_rx_work(struct qede_rx_queue *rxq) in qede_has_rx_work() argument
706 hw_comp_cons = le16_to_cpu(*rxq->hw_cons_ptr); in qede_has_rx_work()
707 sw_comp_cons = qed_chain_get_cons_idx(&rxq->rx_comp_ring); in qede_has_rx_work()
725 static void qede_reuse_rx_data(struct qede_rx_queue *rxq) in qede_reuse_rx_data() argument
727 struct eth_rx_bd *rx_bd_cons = qed_chain_consume(&rxq->rx_bd_ring); in qede_reuse_rx_data()
728 struct eth_rx_bd *rx_bd_prod = qed_chain_produce(&rxq->rx_bd_ring); in qede_reuse_rx_data()
730 &rxq->sw_rx_ring[rxq->sw_rx_cons & NUM_RX_BDS_MAX]; in qede_reuse_rx_data()
732 &rxq->sw_rx_ring[rxq->sw_rx_prod & NUM_RX_BDS_MAX]; in qede_reuse_rx_data()
740 rxq->sw_rx_cons++; in qede_reuse_rx_data()
[all …]
Dqede.h238 struct qede_rx_queue *rxq; member
Dqede_ethtool.c40 (*((u64 *)(((char *)(dev->fp_array[(rqindex)].rxq)) +\
/linux-4.4.14/drivers/net/ethernet/marvell/
Dmv643xx_eth.c414 struct rx_queue rxq[8]; member
456 static struct mv643xx_eth_private *rxq_to_mp(struct rx_queue *rxq) in rxq_to_mp() argument
458 return container_of(rxq, struct mv643xx_eth_private, rxq[rxq->index]); in rxq_to_mp()
466 static void rxq_enable(struct rx_queue *rxq) in rxq_enable() argument
468 struct mv643xx_eth_private *mp = rxq_to_mp(rxq); in rxq_enable()
469 wrlp(mp, RXQ_COMMAND, 1 << rxq->index); in rxq_enable()
472 static void rxq_disable(struct rx_queue *rxq) in rxq_disable() argument
474 struct mv643xx_eth_private *mp = rxq_to_mp(rxq); in rxq_disable()
475 u8 mask = 1 << rxq->index; in rxq_disable()
521 static int rxq_process(struct rx_queue *rxq, int budget) in rxq_process() argument
[all …]
Dmvneta.c640 struct mvneta_rx_queue *rxq, in mvneta_rxq_non_occup_desc_add() argument
647 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), in mvneta_rxq_non_occup_desc_add()
653 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), in mvneta_rxq_non_occup_desc_add()
659 struct mvneta_rx_queue *rxq) in mvneta_rxq_busy_desc_num_get() argument
663 val = mvreg_read(pp, MVNETA_RXQ_STATUS_REG(rxq->id)); in mvneta_rxq_busy_desc_num_get()
671 struct mvneta_rx_queue *rxq, in mvneta_rxq_desc_num_update() argument
679 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), val); in mvneta_rxq_desc_num_update()
699 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), val); in mvneta_rxq_desc_num_update()
705 mvneta_rxq_next_desc_get(struct mvneta_rx_queue *rxq) in mvneta_rxq_next_desc_get() argument
707 int rx_desc = rxq->next_desc_to_proc; in mvneta_rxq_next_desc_get()
[all …]
Dmvpp2.c48 #define MVPP2_RXQ_CONFIG_REG(rxq) (0x800 + 4 * (rxq)) argument
104 #define MVPP2_RXQ_STATUS_UPDATE_REG(rxq) (0x3000 + 4 * (rxq)) argument
107 #define MVPP2_RXQ_STATUS_REG(rxq) (0x3400 + 4 * (rxq)) argument
156 #define MVPP2_ISR_RX_THRESHOLD_REG(rxq) (0x5200 + 4 * (rxq)) argument
157 #define MVPP2_ISR_RXQ_GROUP_REG(rxq) (0x5400 + 4 * (rxq)) argument
3725 int rxq; in mvpp2_swf_bm_pool_init() local
3737 for (rxq = 0; rxq < rxq_number; rxq++) in mvpp2_swf_bm_pool_init()
3738 mvpp2_rxq_long_pool_set(port, rxq, port->pool_long->id); in mvpp2_swf_bm_pool_init()
3751 for (rxq = 0; rxq < rxq_number; rxq++) in mvpp2_swf_bm_pool_init()
3752 mvpp2_rxq_short_pool_set(port, rxq, in mvpp2_swf_bm_pool_init()
[all …]
Dsky2.c1339 unsigned rxq = rxqaddr[sky2->port]; in sky2_rx_stop() local
1343 sky2_write8(hw, RB_ADDR(rxq, RB_CTRL), RB_DIS_OP_MD); in sky2_rx_stop()
1346 if (sky2_read8(hw, RB_ADDR(rxq, Q_RSL)) in sky2_rx_stop()
1347 == sky2_read8(hw, RB_ADDR(rxq, Q_RL))) in sky2_rx_stop()
1352 sky2_write32(hw, Q_ADDR(rxq, Q_CSR), BMU_RST_SET | BMU_FIFO_RST); in sky2_rx_stop()
1355 sky2_write32(hw, Y2_QADDR(rxq, PREF_UNIT_CTRL), PREF_UNIT_RST_SET); in sky2_rx_stop()
1493 static inline void sky2_rx_update(struct sky2_port *sky2, unsigned rxq) in sky2_rx_update() argument
1495 sky2_put_idx(sky2->hw, rxq, sky2->rx_put); in sky2_rx_update()
1535 unsigned rxq = rxqaddr[sky2->port]; in sky2_rx_start() local
1539 sky2_qset(hw, rxq); in sky2_rx_start()
[all …]
/linux-4.4.14/drivers/net/wireless/iwlegacy/
D3945-mac.c956 struct il_rx_queue *rxq = &il->rxq; in il3945_rx_queue_restock() local
962 spin_lock_irqsave(&rxq->lock, flags); in il3945_rx_queue_restock()
963 write = rxq->write & ~0x7; in il3945_rx_queue_restock()
964 while (il_rx_queue_space(rxq) > 0 && rxq->free_count) { in il3945_rx_queue_restock()
966 element = rxq->rx_free.next; in il3945_rx_queue_restock()
971 rxq->bd[rxq->write] = in il3945_rx_queue_restock()
973 rxq->queue[rxq->write] = rxb; in il3945_rx_queue_restock()
974 rxq->write = (rxq->write + 1) & RX_QUEUE_MASK; in il3945_rx_queue_restock()
975 rxq->free_count--; in il3945_rx_queue_restock()
977 spin_unlock_irqrestore(&rxq->lock, flags); in il3945_rx_queue_restock()
[all …]
D4965-mac.c100 il4965_rx_queue_reset(struct il_priv *il, struct il_rx_queue *rxq) in il4965_rx_queue_reset() argument
104 spin_lock_irqsave(&rxq->lock, flags); in il4965_rx_queue_reset()
105 INIT_LIST_HEAD(&rxq->rx_free); in il4965_rx_queue_reset()
106 INIT_LIST_HEAD(&rxq->rx_used); in il4965_rx_queue_reset()
111 if (rxq->pool[i].page != NULL) { in il4965_rx_queue_reset()
112 pci_unmap_page(il->pci_dev, rxq->pool[i].page_dma, in il4965_rx_queue_reset()
115 __il_free_pages(il, rxq->pool[i].page); in il4965_rx_queue_reset()
116 rxq->pool[i].page = NULL; in il4965_rx_queue_reset()
118 list_add_tail(&rxq->pool[i].list, &rxq->rx_used); in il4965_rx_queue_reset()
122 rxq->queue[i] = NULL; in il4965_rx_queue_reset()
[all …]
D4965.h58 void il4965_rx_queue_reset(struct il_priv *il, struct il_rx_queue *rxq);
59 int il4965_rx_init(struct il_priv *il, struct il_rx_queue *rxq);
69 void il4965_rx_queue_free(struct il_priv *il, struct il_rx_queue *rxq);
D3945.c793 il3945_rx_init(struct il_priv *il, struct il_rx_queue *rxq) in il3945_rx_init() argument
795 il_wr(il, FH39_RCSR_RBD_BASE(0), rxq->bd_dma); in il3945_rx_init()
796 il_wr(il, FH39_RCSR_RPTR_ADDR(0), rxq->rb_stts_dma); in il3945_rx_init()
972 struct il_rx_queue *rxq = &il->rxq; in il3945_hw_nic_init() local
982 if (!rxq->bd) { in il3945_hw_nic_init()
989 il3945_rx_queue_reset(il, rxq); in il3945_hw_nic_init()
993 il3945_rx_init(il, rxq); in il3945_hw_nic_init()
1000 il_wr(il, FH39_RCSR_WPTR(0), rxq->write & ~7); in il3945_hw_nic_init()
Dcommon.c2614 struct il_rx_queue *rxq = &il->rxq; in il_rx_queue_alloc() local
2618 spin_lock_init(&rxq->lock); in il_rx_queue_alloc()
2619 INIT_LIST_HEAD(&rxq->rx_free); in il_rx_queue_alloc()
2620 INIT_LIST_HEAD(&rxq->rx_used); in il_rx_queue_alloc()
2623 rxq->bd = dma_alloc_coherent(dev, 4 * RX_QUEUE_SIZE, &rxq->bd_dma, in il_rx_queue_alloc()
2625 if (!rxq->bd) in il_rx_queue_alloc()
2628 rxq->rb_stts = dma_alloc_coherent(dev, sizeof(struct il_rb_status), in il_rx_queue_alloc()
2629 &rxq->rb_stts_dma, GFP_KERNEL); in il_rx_queue_alloc()
2630 if (!rxq->rb_stts) in il_rx_queue_alloc()
2635 list_add_tail(&rxq->pool[i].list, &rxq->rx_used); in il_rx_queue_alloc()
[all …]
Ddebug.c895 struct il_rx_queue *rxq = &il->rxq; in il_dbgfs_rx_queue_read() local
900 pos += scnprintf(buf + pos, bufsz - pos, "read: %u\n", rxq->read); in il_dbgfs_rx_queue_read()
901 pos += scnprintf(buf + pos, bufsz - pos, "write: %u\n", rxq->write); in il_dbgfs_rx_queue_read()
904 rxq->free_count); in il_dbgfs_rx_queue_read()
905 if (rxq->rb_stts) { in il_dbgfs_rx_queue_read()
908 le16_to_cpu(rxq->rb_stts-> in il_dbgfs_rx_queue_read()
D3945.h194 void il3945_rx_queue_reset(struct il_priv *il, struct il_rx_queue *rxq);
Dcommon.h1270 struct il_rx_queue rxq; member
/linux-4.4.14/drivers/net/ethernet/atheros/alx/
Dmain.c75 struct alx_rx_queue *rxq = &alx->rxq; in alx_refill_rx_ring() local
81 next = cur = rxq->write_idx; in alx_refill_rx_ring()
84 cur_buf = &rxq->bufs[cur]; in alx_refill_rx_ring()
86 while (!cur_buf->skb && next != rxq->read_idx) { in alx_refill_rx_ring()
87 struct alx_rfd *rfd = &rxq->rfd[cur]; in alx_refill_rx_ring()
116 cur_buf = &rxq->bufs[cur]; in alx_refill_rx_ring()
123 rxq->write_idx = cur; in alx_refill_rx_ring()
189 struct alx_rx_queue *rxq = &alx->rxq; in alx_clean_rx_irq() local
197 rrd = &rxq->rrd[rxq->rrd_read_idx]; in alx_clean_rx_irq()
203 RRD_SI) != rxq->read_idx || in alx_clean_rx_irq()
[all …]
Dalx.h101 struct alx_rx_queue rxq; member
Dhw.c378 u32 rxq, txq, val; in alx_stop_mac() local
381 rxq = alx_read_mem32(hw, ALX_RXQ0); in alx_stop_mac()
382 alx_write_mem32(hw, ALX_RXQ0, rxq & ~ALX_RXQ0_EN); in alx_stop_mac()
609 u32 mac, txq, rxq; in alx_start_mac() local
611 rxq = alx_read_mem32(hw, ALX_RXQ0); in alx_start_mac()
612 alx_write_mem32(hw, ALX_RXQ0, rxq | ALX_RXQ0_EN); in alx_start_mac()
/linux-4.4.14/drivers/net/ethernet/chelsio/cxgb4vf/
Dcxgb4vf_main.c324 int rxq, msi, err; in request_msix_queue_irqs() local
338 for_each_ethrxq(s, rxq) { in request_msix_queue_irqs()
342 &s->ethrxq[rxq].rspq); in request_msix_queue_irqs()
350 while (--rxq >= 0) in request_msix_queue_irqs()
351 free_irq(adapter->msix_info[--msi].vec, &s->ethrxq[rxq].rspq); in request_msix_queue_irqs()
362 int rxq, msi; in free_msix_queue_irqs() local
366 for_each_ethrxq(s, rxq) in free_msix_queue_irqs()
368 &s->ethrxq[rxq].rspq); in free_msix_queue_irqs()
393 int rxq; in enable_rx() local
396 for_each_ethrxq(s, rxq) in enable_rx()
[all …]
Dsge.c1568 static void do_gro(struct sge_eth_rxq *rxq, const struct pkt_gl *gl, in do_gro() argument
1571 struct adapter *adapter = rxq->rspq.adapter; in do_gro()
1576 skb = napi_get_frags(&rxq->rspq.napi); in do_gro()
1579 rxq->stats.rx_drops++; in do_gro()
1588 skb_record_rx_queue(skb, rxq->rspq.idx); in do_gro()
1593 rxq->stats.vlan_ex++; in do_gro()
1595 ret = napi_gro_frags(&rxq->rspq.napi); in do_gro()
1598 rxq->stats.lro_pkts++; in do_gro()
1600 rxq->stats.lro_merged++; in do_gro()
1601 rxq->stats.pkts++; in do_gro()
[all …]
/linux-4.4.14/drivers/net/ethernet/broadcom/bnx2x/
Dbnx2x_vfpf.c634 req->rxq.rcq_addr = fp->rx_comp_mapping; in bnx2x_vfpf_setup_q()
635 req->rxq.rcq_np_addr = fp->rx_comp_mapping + BCM_PAGE_SIZE; in bnx2x_vfpf_setup_q()
636 req->rxq.rxq_addr = fp->rx_desc_mapping; in bnx2x_vfpf_setup_q()
637 req->rxq.sge_addr = fp->rx_sge_mapping; in bnx2x_vfpf_setup_q()
638 req->rxq.vf_sb = fp_idx; in bnx2x_vfpf_setup_q()
639 req->rxq.sb_index = HC_INDEX_ETH_RX_CQ_CONS; in bnx2x_vfpf_setup_q()
640 req->rxq.hc_rate = bp->rx_ticks ? 1000000/bp->rx_ticks : 0; in bnx2x_vfpf_setup_q()
641 req->rxq.mtu = bp->dev->mtu; in bnx2x_vfpf_setup_q()
642 req->rxq.buf_sz = fp->rx_buf_size; in bnx2x_vfpf_setup_q()
643 req->rxq.sge_buf_sz = BCM_PAGE_SIZE * PAGES_PER_SGE; in bnx2x_vfpf_setup_q()
[all …]
Dbnx2x_vfpf.h273 } rxq; member
Dbnx2x_sriov.c1912 struct bnx2x_vf_queue *rxq = vfq_get(vf, j); in bnx2x_iov_adjust_stats_req() local
1918 if (bnx2x_get_q_logical_state(bp, &rxq->sp_obj) == in bnx2x_iov_adjust_stats_req()
1924 cur_query_entry->index = vfq_stat_id(vf, rxq); in bnx2x_iov_adjust_stats_req()
/linux-4.4.14/drivers/net/ethernet/freescale/
Dgianfar.h1311 static inline int gfar_rxbd_unused(struct gfar_priv_rx_q *rxq) in gfar_rxbd_unused() argument
1313 if (rxq->next_to_clean > rxq->next_to_use) in gfar_rxbd_unused()
1314 return rxq->next_to_clean - rxq->next_to_use - 1; in gfar_rxbd_unused()
1316 return rxq->rx_ring_size + rxq->next_to_clean - rxq->next_to_use - 1; in gfar_rxbd_unused()
1319 static inline u32 gfar_rxbd_dma_lastfree(struct gfar_priv_rx_q *rxq) in gfar_rxbd_dma_lastfree() argument
1325 i = rxq->next_to_use ? rxq->next_to_use - 1 : rxq->rx_ring_size - 1; in gfar_rxbd_dma_lastfree()
1326 bdp = &rxq->rx_bd_base[i]; in gfar_rxbd_dma_lastfree()
1327 bdp_dma = lower_32_bits(rxq->rx_bd_dma_base); in gfar_rxbd_dma_lastfree()
1328 bdp_dma += (uintptr_t)bdp - (uintptr_t)rxq->rx_bd_base; in gfar_rxbd_dma_lastfree()
Dfec_main.c231 struct fec_enet_priv_rx_q *rxq = fep->rx_queue[queue_id]; in fec_enet_get_nextdesc() local
241 base = rxq->rx_bd_base; in fec_enet_get_nextdesc()
242 ring_size = rxq->rx_ring_size; in fec_enet_get_nextdesc()
243 ex_base = (struct bufdesc_ex *)rxq->rx_bd_base; in fec_enet_get_nextdesc()
262 struct fec_enet_priv_rx_q *rxq = fep->rx_queue[queue_id]; in fec_enet_get_prevdesc() local
272 base = rxq->rx_bd_base; in fec_enet_get_prevdesc()
273 ring_size = rxq->rx_ring_size; in fec_enet_get_prevdesc()
274 ex_base = (struct bufdesc_ex *)rxq->rx_bd_base; in fec_enet_get_prevdesc()
811 struct fec_enet_priv_rx_q *rxq; in fec_enet_bd_init() local
818 rxq = fep->rx_queue[q]; in fec_enet_bd_init()
[all …]
Dgianfar.c2762 static bool gfar_new_page(struct gfar_priv_rx_q *rxq, struct gfar_rx_buff *rxb) in gfar_new_page() argument
2771 addr = dma_map_page(rxq->dev, page, 0, PAGE_SIZE, DMA_FROM_DEVICE); in gfar_new_page()
2772 if (unlikely(dma_mapping_error(rxq->dev, addr))) { in gfar_new_page()
2953 static void gfar_reuse_rx_page(struct gfar_priv_rx_q *rxq, in gfar_reuse_rx_page() argument
2957 u16 nta = rxq->next_to_alloc; in gfar_reuse_rx_page()
2959 new_rxb = &rxq->rx_buff[nta]; in gfar_reuse_rx_page()
2963 rxq->next_to_alloc = (nta < rxq->rx_ring_size) ? nta : 0; in gfar_reuse_rx_page()
2969 dma_sync_single_range_for_device(rxq->dev, old_rxb->dma, in gfar_reuse_rx_page()
/linux-4.4.14/drivers/net/ethernet/samsung/sxgbe/
Dsxgbe_main.c330 priv->hw->desc->init_rx_desc(&priv->rxq[j]->dma_rx[i], in sxgbe_clear_descriptors()
588 priv->rxq[queue_num], rx_rsize); in init_dma_desc_rings()
597 priv->rxq[queue_num]->priv_ptr = priv; in init_dma_desc_rings()
611 free_rx_ring(priv->device, priv->rxq[queue_num], rx_rsize); in init_dma_desc_rings()
663 free_rx_ring(priv->device, priv->rxq[queue_num], rx_rsize); in free_dma_desc_resources()
686 priv->rxq[queue_num] = devm_kmalloc(priv->device, in rxring_mem_alloc()
688 if (!priv->rxq[queue_num]) in rxring_mem_alloc()
973 (priv->rxq[queue_num])->dma_rx_phy, in sxgbe_init_dma_engine()
1148 (priv->rxq[queue_num])->irq_no, in sxgbe_open()
1150 dev->name, priv->rxq[queue_num]); in sxgbe_open()
[all …]
Dsxgbe_platform.c138 priv->rxq[i]->irq_no = irq_of_parse_and_map(node, chan++); in sxgbe_platform_probe()
139 if (priv->rxq[i]->irq_no <= 0) { in sxgbe_platform_probe()
159 irq_dispose_mapping(priv->rxq[i]->irq_no); in sxgbe_platform_probe()
Dsxgbe_common.h459 struct sxgbe_rx_queue *rxq[SXGBE_RX_QUEUES]; member
/linux-4.4.14/drivers/net/ethernet/brocade/bna/
Dbnad_ethtool.c670 rcb[1]->rxq) { in bnad_get_strings()
752 bnad->rx_info[i].rx_ctrl[j].ccb->rcb[1]->rxq) in bnad_get_stats_count_locked()
777 bnad->rx_info[i].rx_ctrl[j].ccb->rcb[0]->rxq) { in bnad_per_q_stats_fill()
803 rcb[0]->rxq) { in bnad_per_q_stats_fill()
806 buf[bi++] = rcb->rxq->rx_packets; in bnad_per_q_stats_fill()
807 buf[bi++] = rcb->rxq->rx_bytes; in bnad_per_q_stats_fill()
808 buf[bi++] = rcb->rxq-> in bnad_per_q_stats_fill()
810 buf[bi++] = rcb->rxq-> in bnad_per_q_stats_fill()
812 buf[bi++] = rcb->rxq->rxbuf_map_failed; in bnad_per_q_stats_fill()
818 rcb[1]->rxq) { in bnad_per_q_stats_fill()
[all …]
Dbna.h192 (q0) = rxp->rxq.single.only; \
196 (q0) = rxp->rxq.slr.large; \
197 (q1) = rxp->rxq.slr.small; \
200 (q0) = rxp->rxq.hds.data; \
201 (q1) = rxp->rxq.hds.hdr; \
Dbna_tx_rx.c1772 struct bna_rxq *rxq = NULL; in bna_rxq_get() local
1774 rxq = list_first_entry(&rx_mod->rxq_free_q, struct bna_rxq, qe); in bna_rxq_get()
1775 list_del(&rxq->qe); in bna_rxq_get()
1778 return rxq; in bna_rxq_get()
1782 bna_rxq_put(struct bna_rx_mod *rx_mod, struct bna_rxq *rxq) in bna_rxq_put() argument
1784 list_add_tail(&rxq->qe, &rx_mod->rxq_free_q); in bna_rxq_put()
1844 rxp->rxq.single.only = q0; in bna_rxp_add_rxqs()
1845 rxp->rxq.single.reserved = NULL; in bna_rxp_add_rxqs()
1848 rxp->rxq.slr.large = q0; in bna_rxp_add_rxqs()
1849 rxp->rxq.slr.small = q1; in bna_rxp_add_rxqs()
[all …]
Dbna_types.h561 struct bna_rxq *rxq; member
698 union bna_rxq_u rxq; member
840 struct bna_rxq *rxq; /* BFI_MAX_RXQ entries */ member
Dbnad.c292 order = get_order(rcb->rxq->buffer_size); in bnad_rxq_alloc_init()
298 unmap_q->map_size = rcb->rxq->buffer_size; in bnad_rxq_alloc_init()
300 if (rcb->rxq->multi_buffer) { in bnad_rxq_alloc_init()
302 unmap_q->map_size = rcb->rxq->buffer_size; in bnad_rxq_alloc_init()
307 (rcb->rxq->buffer_size > 2048) ? in bnad_rxq_alloc_init()
397 rcb->rxq->rxbuf_alloc_failed++; in bnad_rxq_refill_page()
406 rcb->rxq->rxbuf_map_failed++; in bnad_rxq_refill_page()
448 buff_sz = rcb->rxq->buffer_size; in bnad_rxq_refill_skb()
460 rcb->rxq->rxbuf_alloc_failed++; in bnad_rxq_refill_skb()
469 rcb->rxq->rxbuf_map_failed++; in bnad_rxq_refill_skb()
[all …]
/linux-4.4.14/drivers/net/wireless/
Dmwl8k.c247 struct mwl8k_rx_queue rxq[MWL8K_RX_QUEUES]; member
1152 struct mwl8k_rx_queue *rxq = priv->rxq + index; in mwl8k_rxq_init() local
1156 rxq->rxd_count = 0; in mwl8k_rxq_init()
1157 rxq->head = 0; in mwl8k_rxq_init()
1158 rxq->tail = 0; in mwl8k_rxq_init()
1162 rxq->rxd = pci_zalloc_consistent(priv->pdev, size, &rxq->rxd_dma); in mwl8k_rxq_init()
1163 if (rxq->rxd == NULL) { in mwl8k_rxq_init()
1168 rxq->buf = kcalloc(MWL8K_RX_DESCS, sizeof(*rxq->buf), GFP_KERNEL); in mwl8k_rxq_init()
1169 if (rxq->buf == NULL) { in mwl8k_rxq_init()
1170 pci_free_consistent(priv->pdev, size, rxq->rxd, rxq->rxd_dma); in mwl8k_rxq_init()
[all …]
/linux-4.4.14/drivers/net/ethernet/chelsio/cxgb4/
Dsge.c1845 static void do_gro(struct sge_eth_rxq *rxq, const struct pkt_gl *gl, in do_gro() argument
1848 struct adapter *adapter = rxq->rspq.adap; in do_gro()
1854 skb = napi_get_frags(&rxq->rspq.napi); in do_gro()
1857 rxq->stats.rx_drops++; in do_gro()
1866 skb_record_rx_queue(skb, rxq->rspq.idx); in do_gro()
1867 skb_mark_napi_id(skb, &rxq->rspq.napi); in do_gro()
1872 if (rxq->rspq.netdev->features & NETIF_F_RXHASH) in do_gro()
1878 rxq->stats.vlan_ex++; in do_gro()
1880 ret = napi_gro_frags(&rxq->rspq.napi); in do_gro()
1882 rxq->stats.lro_pkts++; in do_gro()
[all …]
Dcxgb4_main.c695 struct sge_ofld_rxq *rxq = container_of(q, struct sge_ofld_rxq, rspq); in uldrx_handler() local
704 rxq->stats.nomem++; in uldrx_handler()
708 rxq->stats.imm++; in uldrx_handler()
710 rxq->stats.an++; in uldrx_handler()
712 rxq->stats.pkts++; in uldrx_handler()
878 const struct sge_eth_rxq *rxq; in cxgb4_write_rss() local
880 rxq = &adapter->sge.ethrxq[pi->first_qset]; in cxgb4_write_rss()
887 rss[i] = rxq[*queues].rspq.abs_id; in cxgb4_write_rss()
Dcxgb4.h1372 unsigned int rxqi, unsigned int rxq, unsigned int tc,
Dt4_hw.c6362 unsigned int rxqi, unsigned int rxq, unsigned int tc, in t4_cfg_pfvf() argument
6374 FW_PFVF_CMD_NIQ_V(rxq)); in t4_cfg_pfvf()
/linux-4.4.14/drivers/net/ethernet/xscale/
Dixp4xx_eth.c678 qmgr_disable_irq(port->plat->rxq); in eth_rx_irq()
686 unsigned int rxq = port->plat->rxq, rxfreeq = RXFREE_QUEUE(port->id); in eth_poll() local
702 if ((n = queue_get_desc(rxq, port, 0)) < 0) { in eth_poll()
708 qmgr_enable_irq(rxq); in eth_poll()
709 if (!qmgr_stat_below_low_watermark(rxq) && in eth_poll()
716 qmgr_disable_irq(rxq); in eth_poll()
1074 err = qmgr_request_queue(port->plat->rxq, RX_DESCS, 0, 0, in request_queues()
1103 qmgr_release_queue(port->plat->rxq); in request_queues()
1114 qmgr_release_queue(port->plat->rxq); in release_queues()
1227 msg.byte5 = port->plat->rxq | 0x80; in eth_open()
[all …]
/linux-4.4.14/drivers/net/ethernet/intel/i40evf/
Di40evf_virtchnl.c264 vqpi->rxq.vsi_id = vqci->vsi_id; in i40evf_configure_queues()
265 vqpi->rxq.queue_id = i; in i40evf_configure_queues()
266 vqpi->rxq.ring_len = adapter->rx_rings[i]->count; in i40evf_configure_queues()
267 vqpi->rxq.dma_ring_addr = adapter->rx_rings[i]->dma; in i40evf_configure_queues()
268 vqpi->rxq.max_pkt_size = adapter->netdev->mtu in i40evf_configure_queues()
270 vqpi->rxq.databuffer_size = adapter->rx_rings[i]->rx_buf_len; in i40evf_configure_queues()
Di40e_virtchnl.h215 struct i40e_virtchnl_rxq_info rxq; member
/linux-4.4.14/arch/arm/mach-ixp4xx/
Dvulcan-setup.c126 .rxq = 3,
131 .rxq = 4,
Domixp-setup.c177 .rxq = 3,
181 .rxq = 4,
Dfsg-setup.c130 .rxq = 3,
134 .rxq = 4,
Dixdp425-setup.c185 .rxq = 3,
189 .rxq = 4,
Dgoramo_mlr.c284 .rxq = 3,
288 .rxq = 4,
Dnslu2-setup.c175 .rxq = 3,
Dnas100d-setup.c163 .rxq = 3,
/linux-4.4.14/drivers/net/ethernet/ibm/
Dibmveth.h52 #define h_register_logical_lan(ua, buflst, rxq, fltlst, mac) \ argument
53 plpar_hcall_norets(H_REGISTER_LOGICAL_LAN, ua, buflst, rxq, fltlst, mac)
/linux-4.4.14/drivers/net/usb/
Dusbnet.c498 spin_lock_irqsave (&dev->rxq.lock, lockflags); in rx_submit()
524 __usbnet_queue_skb(&dev->rxq, skb, rx_start); in rx_submit()
530 spin_unlock_irqrestore (&dev->rxq.lock, lockflags); in rx_submit()
647 state = defer_bh(dev, skb, &dev->rxq, state); in rx_complete()
751 (void) unlink_urbs (dev, &dev->rxq); in usbnet_unlink_rx_urbs()
783 unlink_urbs(dev, &dev->rxq); in usbnet_terminate_urbs()
786 wait_skb_queue_empty(&dev->rxq); in usbnet_terminate_urbs()
1056 unlink_urbs(dev, &dev->rxq); in __handle_link_change()
1122 unlink_urbs (dev, &dev->rxq); in usbnet_deferred_kevent()
1450 for (i = 0; i < 10 && dev->rxq.qlen < RX_QLEN(dev); i++) { in rx_alloc_submit()
[all …]
Dlan78xx.c244 struct sk_buff_head rxq; member
1654 unlink_urbs(dev, &dev->rxq); in lan78xx_change_mtu()
1992 temp = unlink_urbs(dev, &dev->txq) + unlink_urbs(dev, &dev->rxq); in lan78xx_terminate_urbs()
1995 while (!skb_queue_empty(&dev->rxq) && in lan78xx_terminate_urbs()
2472 spin_lock_irqsave(&dev->rxq.lock, lockflags); in rx_submit()
2481 lan78xx_queue_skb(&dev->rxq, skb, rx_start); in rx_submit()
2502 spin_unlock_irqrestore(&dev->rxq.lock, lockflags); in rx_submit()
2566 state = defer_bh(dev, skb, &dev->rxq, state); in rx_complete()
2709 if (skb_queue_len(&dev->rxq) < dev->rx_qlen) { in lan78xx_rx_bh()
2711 if (skb_queue_len(&dev->rxq) >= dev->rx_qlen) in lan78xx_rx_bh()
[all …]
/linux-4.4.14/drivers/net/wireless/ipw2x00/
Dipw2200.c3451 struct ipw_rx_queue *rxq) in ipw_rx_queue_reset() argument
3456 spin_lock_irqsave(&rxq->lock, flags); in ipw_rx_queue_reset()
3458 INIT_LIST_HEAD(&rxq->rx_free); in ipw_rx_queue_reset()
3459 INIT_LIST_HEAD(&rxq->rx_used); in ipw_rx_queue_reset()
3465 if (rxq->pool[i].skb != NULL) { in ipw_rx_queue_reset()
3466 pci_unmap_single(priv->pci_dev, rxq->pool[i].dma_addr, in ipw_rx_queue_reset()
3468 dev_kfree_skb(rxq->pool[i].skb); in ipw_rx_queue_reset()
3469 rxq->pool[i].skb = NULL; in ipw_rx_queue_reset()
3471 list_add_tail(&rxq->pool[i].list, &rxq->rx_used); in ipw_rx_queue_reset()
3476 rxq->read = rxq->write = 0; in ipw_rx_queue_reset()
[all …]
Dipw2100.c2675 struct ipw2100_bd_queue *rxq = &priv->rx_queue; in __ipw2100_rx_process() local
2688 if (r >= rxq->entries) { in __ipw2100_rx_process()
2693 i = (rxq->next + 1) % rxq->entries; in __ipw2100_rx_process()
2765 rxq->drv[i].status.info.field = 0; in __ipw2100_rx_process()
2767 i = (i + 1) % rxq->entries; in __ipw2100_rx_process()
2772 rxq->next = (i ? i : rxq->entries) - 1; in __ipw2100_rx_process()
2775 IPW_MEM_HOST_SHARED_RX_WRITE_INDEX, rxq->next); in __ipw2100_rx_process()
Dipw2200.h1205 struct ipw_rx_queue *rxq; member
/linux-4.4.14/drivers/mailbox/
Domap-mailbox.c117 struct omap_mbox_queue *rxq; member
340 struct omap_mbox_queue *mq = mbox->rxq; in __mbox_rx_interrupt()
360 schedule_work(&mbox->rxq->work); in __mbox_rx_interrupt()
415 mbox->rxq = mq; in omap_mbox_startup()
433 mbox_queue_free(mbox->rxq); in omap_mbox_startup()
441 flush_work(&mbox->rxq->work); in omap_mbox_fini()
442 mbox_queue_free(mbox->rxq); in omap_mbox_fini()
/linux-4.4.14/drivers/net/
Dmacvtap.c286 __u32 rxq; in macvtap_get_queue() local
292 rxq = skb_get_hash(skb); in macvtap_get_queue()
293 if (rxq) { in macvtap_get_queue()
294 tap = rcu_dereference(vlan->taps[rxq % numvtaps]); in macvtap_get_queue()
299 rxq = skb_get_rx_queue(skb); in macvtap_get_queue()
301 while (unlikely(rxq >= numvtaps)) in macvtap_get_queue()
302 rxq -= numvtaps; in macvtap_get_queue()
304 tap = rcu_dereference(vlan->taps[rxq]); in macvtap_get_queue()
Dxen-netfront.c928 struct sk_buff_head *rxq) in handle_incoming_queue() argument
934 while ((skb = __skb_dequeue(rxq)) != NULL) { in handle_incoming_queue()
973 struct sk_buff_head rxq; in xennet_poll() local
980 skb_queue_head_init(&rxq); in xennet_poll()
1033 __skb_queue_tail(&rxq, skb); in xennet_poll()
1041 work_done -= handle_incoming_queue(queue, &rxq); in xennet_poll()
Dvirtio_net.c179 static int rxq2vq(int rxq) in rxq2vq() argument
181 return rxq * 2; in rxq2vq()
/linux-4.4.14/arch/arm/mach-ixp4xx/include/mach/
Dplatform.h101 u8 rxq; /* configurable, currently 0 - 31 only */ member
/linux-4.4.14/drivers/net/ethernet/sfc/
Dvfdi.h198 u32 rxq; member
Dptp.c272 struct sk_buff_head rxq; member
1008 while ((skb = skb_dequeue(&ptp->rxq))) { in efx_ptp_process_events()
1023 skb_queue_head(&ptp->rxq, skb); in efx_ptp_process_events()
1139 efx_ptp_deliver_rx_queue(&efx->ptp_data->rxq); in efx_ptp_stop()
1236 skb_queue_head_init(&ptp->rxq); in efx_ptp_probe()
1325 skb_queue_purge(&efx->ptp_data->rxq); in efx_ptp_remove()
1449 skb_queue_tail(&ptp->rxq, skb); in efx_ptp_rx()
Dsiena_sriov.c767 unsigned vf_rxq = req->u.mac_filter.rxq; in efx_vfdi_insert_filter()
/linux-4.4.14/drivers/net/wireless/iwlwifi/
Diwl-fw-error-dump.h240 __le32 rxq; member
/linux-4.4.14/Documentation/devicetree/bindings/net/
Dbrcm,systemport.txt19 - systemport,num-rxq: number of HW receive queues, an integer
/linux-4.4.14/drivers/net/ethernet/intel/i40e/
Di40e_virtchnl.h215 struct i40e_virtchnl_rxq_info rxq; member
Di40e_virtchnl_pf.c1330 (qpi->rxq.vsi_id != vsi_id) || in i40e_vc_config_queues_msg()
1331 (qpi->rxq.queue_id != vsi_queue_id) || in i40e_vc_config_queues_msg()
1338 &qpi->rxq) || in i40e_vc_config_queues_msg()
/linux-4.4.14/drivers/net/ethernet/cavium/liquidio/
Docteon_network.h49 int rxq; member
Dlio_main.c3322 lio->rxq = lio->linfo.rxpciq[0]; in setup_nic_devices()
3325 lio->rx_qsize = octeon_get_rx_qsize(octeon_dev, lio->rxq); in setup_nic_devices()
/linux-4.4.14/drivers/net/wan/
Dixp4xx_hss.c658 unsigned int rxq = queue_ids[port->id].rx; in hss_hdlc_poll() local
675 if ((n = queue_get_desc(rxq, port, 0)) < 0) { in hss_hdlc_poll()
681 qmgr_enable_irq(rxq); in hss_hdlc_poll()
682 if (!qmgr_stat_empty(rxq) && in hss_hdlc_poll()
689 qmgr_disable_irq(rxq); in hss_hdlc_poll()
/linux-4.4.14/include/linux/usb/
Dusbnet.h57 struct sk_buff_head rxq; member
/linux-4.4.14/drivers/net/ethernet/emulex/benet/
Dbe_main.c1990 struct be_queue_info *rxq = &rxo->q; in get_rx_page_info() local
1991 u16 frag_idx = rxq->tail; in get_rx_page_info()
2007 queue_tail_inc(rxq); in get_rx_page_info()
2008 atomic_dec(&rxq->used); in get_rx_page_info()
2302 struct be_queue_info *rxq = &rxo->q; in be_post_rx_frags() local
2309 page_info = &rxo->page_info_tbl[rxq->head]; in be_post_rx_frags()
2334 rxd = queue_head_node(rxq); in be_post_rx_frags()
2350 queue_head_inc(rxq); in be_post_rx_frags()
2351 page_info = &rxo->page_info_tbl[rxq->head]; in be_post_rx_frags()
2363 atomic_add(posted, &rxq->used); in be_post_rx_frags()
[all …]
Dbe_cmds.c1346 struct be_queue_info *rxq, u16 cq_id, u16 frag_size, in be_cmd_rxq_create() argument
1351 struct be_dma_mem *q_mem = &rxq->dma_mem; in be_cmd_rxq_create()
1378 rxq->id = le16_to_cpu(resp->id); in be_cmd_rxq_create()
1379 rxq->created = true; in be_cmd_rxq_create()
Dbe_cmds.h2293 int be_cmd_rxq_create(struct be_adapter *adapter, struct be_queue_info *rxq,
/linux-4.4.14/drivers/net/wireless/ath/ath6kl/
Dhtc_mbox.c1911 struct list_head *rxq, in ath6kl_htc_rx_bundle() argument
1920 n_scat_pkt = get_queue_depth(rxq); in ath6kl_htc_rx_bundle()
1923 if ((get_queue_depth(rxq) - n_scat_pkt) > 0) { in ath6kl_htc_rx_bundle()
1936 __func__, get_queue_depth(rxq), n_scat_pkt); in ath6kl_htc_rx_bundle()
1943 get_queue_depth(rxq), n_scat_pkt); in ath6kl_htc_rx_bundle()
1953 packet = list_first_entry(rxq, struct htc_packet, list); in ath6kl_htc_rx_bundle()
1960 list_add(&packet->list, rxq); in ath6kl_htc_rx_bundle()
/linux-4.4.14/drivers/net/ethernet/broadcom/
Dbcmsysport.c1754 u32 txq, rxq; in bcm_sysport_probe() local
1763 if (of_property_read_u32(dn, "systemport,num-rxq", &rxq)) in bcm_sysport_probe()
1764 rxq = 1; in bcm_sysport_probe()
1766 dev = alloc_etherdev_mqs(sizeof(*priv), txq, rxq); in bcm_sysport_probe()
1853 priv->base, priv->irq0, priv->irq1, txq, rxq); in bcm_sysport_probe()
/linux-4.4.14/net/core/
Dnet-sysfs.c1338 int error = 0, txq = 0, rxq = 0, real_rx = 0, real_tx = 0; in register_queue_kobjects() local
1352 rxq = real_rx; in register_queue_kobjects()
1363 net_rx_queue_update_kobjects(dev, rxq, 0); in register_queue_kobjects()
Ddev.c2203 int netif_set_real_num_rx_queues(struct net_device *dev, unsigned int rxq) in netif_set_real_num_rx_queues() argument
2207 if (rxq < 1 || rxq > dev->num_rx_queues) in netif_set_real_num_rx_queues()
2214 rxq); in netif_set_real_num_rx_queues()
2219 dev->real_num_rx_queues = rxq; in netif_set_real_num_rx_queues()
/linux-4.4.14/drivers/net/xen-netback/
Dnetback.c556 struct sk_buff_head rxq; in xenvif_rx_action() local
568 skb_queue_head_init(&rxq); in xenvif_rx_action()
576 __skb_queue_tail(&rxq, skb); in xenvif_rx_action()
587 while ((skb = __skb_dequeue(&rxq)) != NULL) { in xenvif_rx_action()
/linux-4.4.14/drivers/infiniband/hw/cxgb3/
Diwch_cm.c107 static struct sk_buff_head rxq; variable
2187 while ((skb = skb_dequeue(&rxq))) { in process_work()
2218 skb_queue_tail(&rxq, skb); in sched()
2259 skb_queue_head_init(&rxq); in iwch_cm_init()
/linux-4.4.14/drivers/net/ethernet/atheros/atl1c/
Datl1c_main.c1200 u32 mac, txq, rxq; in atl1c_start_mac() local
1207 AT_READ_REG(hw, REG_RXQ_CTRL, &rxq); in atl1c_start_mac()
1211 rxq |= RXQ_CTRL_EN; in atl1c_start_mac()
1225 AT_WRITE_REG(hw, REG_RXQ_CTRL, rxq); in atl1c_start_mac()
/linux-4.4.14/drivers/net/ethernet/broadcom/bnxt/
Dbnxt.h754 u16 rxq; member
Dbnxt_ethtool.c352 fs->ring_cookie = fltr->rxq; in bnxt_grxclsrule()
Dbnxt.c2818 struct bnxt_vnic_info *vnic = &bp->vnic_info[fltr->rxq + 1]; in bnxt_hwrm_cfa_ntuple_filter_alloc()
5397 new_fltr->rxq = rxq_index; in bnxt_rx_flow_steer()
5427 if (rps_may_expire_flow(bp->dev, fltr->rxq, in bnxt_cfg_ntp_filters()
/linux-4.4.14/drivers/infiniband/hw/cxgb4/
Dcm.c143 static struct sk_buff_head rxq; variable
4017 while ((skb = skb_dequeue(&rxq))) { in process_work()
4066 skb_queue_tail(&rxq, skb); in sched()
4180 skb_queue_head_init(&rxq); in c4iw_cm_init()
/linux-4.4.14/include/linux/
Dnetdevice.h2953 int netif_set_real_num_rx_queues(struct net_device *dev, unsigned int rxq);
2956 unsigned int rxq) in netif_set_real_num_rx_queues() argument