/linux-4.1.27/drivers/net/wireless/iwlwifi/pcie/ |
D | rx.c | 114 static int iwl_rxq_space(const struct iwl_rxq *rxq) in iwl_rxq_space() argument 125 return (rxq->read - rxq->write - 1) & (RX_QUEUE_SIZE - 1); in iwl_rxq_space() 152 struct iwl_rxq *rxq = &trans_pcie->rxq; in iwl_pcie_rxq_inc_wr_ptr() local 155 lockdep_assert_held(&rxq->lock); in iwl_pcie_rxq_inc_wr_ptr() 171 rxq->need_update = true; in iwl_pcie_rxq_inc_wr_ptr() 176 rxq->write_actual = round_down(rxq->write, 8); in iwl_pcie_rxq_inc_wr_ptr() 177 iwl_write32(trans, FH_RSCSR_CHNL0_WPTR, rxq->write_actual); in iwl_pcie_rxq_inc_wr_ptr() 183 struct iwl_rxq *rxq = &trans_pcie->rxq; in iwl_pcie_rxq_check_wrptr() local 185 spin_lock(&rxq->lock); in iwl_pcie_rxq_check_wrptr() 187 if (!rxq->need_update) in iwl_pcie_rxq_check_wrptr() [all …]
|
D | trans.c | 1849 struct iwl_rxq *rxq = &trans_pcie->rxq; in iwl_dbgfs_rx_queue_read() local 1855 rxq->read); in iwl_dbgfs_rx_queue_read() 1857 rxq->write); in iwl_dbgfs_rx_queue_read() 1859 rxq->write_actual); in iwl_dbgfs_rx_queue_read() 1861 rxq->need_update); in iwl_dbgfs_rx_queue_read() 1863 rxq->free_count); in iwl_dbgfs_rx_queue_read() 1864 if (rxq->rb_stts) { in iwl_dbgfs_rx_queue_read() 1866 le16_to_cpu(rxq->rb_stts->closed_rb_num) & 0x0FFF); in iwl_dbgfs_rx_queue_read()
|
D | internal.h | 275 struct iwl_rxq rxq; member
|
/linux-4.1.27/drivers/atm/ |
D | ambassador.c | 688 amb_rxq * rxq = &dev->rxq[pool]; in rx_give() local 693 spin_lock_irqsave (&rxq->lock, flags); in rx_give() 695 if (rxq->pending < rxq->maximum) { in rx_give() 696 PRINTD (DBG_RX, "RX in slot %p", rxq->in.ptr); in rx_give() 698 *rxq->in.ptr = *rx; in rx_give() 699 rxq->pending++; in rx_give() 700 rxq->in.ptr = NEXTQ (rxq->in.ptr, rxq->in.start, rxq->in.limit); in rx_give() 702 wr_mem (dev, offsetof(amb_mem, mb.adapter.rx_address[pool]), virt_to_bus (rxq->in.ptr)); in rx_give() 704 spin_unlock_irqrestore (&rxq->lock, flags); in rx_give() 707 spin_unlock_irqrestore (&rxq->lock, flags); in rx_give() [all …]
|
D | fore200e.c | 1180 struct host_rxq* rxq = &fore200e->host_rxq; in fore200e_rx_irq() local 1187 entry = &rxq->host_entry[ rxq->head ]; in fore200e_rx_irq() 1218 FORE200E_NEXT_ENTRY(rxq->head, QUEUE_SIZE_RX); in fore200e_rx_irq() 2209 struct host_rxq* rxq = &fore200e->host_rxq; in fore200e_init_rx_queue() local 2217 &rxq->status, in fore200e_init_rx_queue() 2226 &rxq->rpd, in fore200e_init_rx_queue() 2231 fore200e->bus->dma_chunk_free(fore200e, &rxq->status); in fore200e_init_rx_queue() 2241 rxq->host_entry[ i ].status = in fore200e_init_rx_queue() 2242 FORE200E_INDEX(rxq->status.align_addr, enum status, i); in fore200e_init_rx_queue() 2243 rxq->host_entry[ i ].rpd = in fore200e_init_rx_queue() [all …]
|
D | ambassador.h | 635 amb_rxq rxq[NUM_RX_POOLS]; member
|
/linux-4.1.27/drivers/net/ethernet/marvell/ |
D | mvneta.c | 576 struct mvneta_rx_queue *rxq, in mvneta_rxq_non_occup_desc_add() argument 583 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), in mvneta_rxq_non_occup_desc_add() 589 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), in mvneta_rxq_non_occup_desc_add() 595 struct mvneta_rx_queue *rxq) in mvneta_rxq_busy_desc_num_get() argument 599 val = mvreg_read(pp, MVNETA_RXQ_STATUS_REG(rxq->id)); in mvneta_rxq_busy_desc_num_get() 607 struct mvneta_rx_queue *rxq, in mvneta_rxq_desc_num_update() argument 615 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), val); in mvneta_rxq_desc_num_update() 635 mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id), val); in mvneta_rxq_desc_num_update() 641 mvneta_rxq_next_desc_get(struct mvneta_rx_queue *rxq) in mvneta_rxq_next_desc_get() argument 643 int rx_desc = rxq->next_desc_to_proc; in mvneta_rxq_next_desc_get() [all …]
|
D | mv643xx_eth.c | 414 struct rx_queue rxq[8]; member 456 static struct mv643xx_eth_private *rxq_to_mp(struct rx_queue *rxq) in rxq_to_mp() argument 458 return container_of(rxq, struct mv643xx_eth_private, rxq[rxq->index]); in rxq_to_mp() 466 static void rxq_enable(struct rx_queue *rxq) in rxq_enable() argument 468 struct mv643xx_eth_private *mp = rxq_to_mp(rxq); in rxq_enable() 469 wrlp(mp, RXQ_COMMAND, 1 << rxq->index); in rxq_enable() 472 static void rxq_disable(struct rx_queue *rxq) in rxq_disable() argument 474 struct mv643xx_eth_private *mp = rxq_to_mp(rxq); in rxq_disable() 475 u8 mask = 1 << rxq->index; in rxq_disable() 521 static int rxq_process(struct rx_queue *rxq, int budget) in rxq_process() argument [all …]
|
D | mvpp2.c | 46 #define MVPP2_RXQ_CONFIG_REG(rxq) (0x800 + 4 * (rxq)) argument 102 #define MVPP2_RXQ_STATUS_UPDATE_REG(rxq) (0x3000 + 4 * (rxq)) argument 105 #define MVPP2_RXQ_STATUS_REG(rxq) (0x3400 + 4 * (rxq)) argument 154 #define MVPP2_ISR_RX_THRESHOLD_REG(rxq) (0x5200 + 4 * (rxq)) argument 155 #define MVPP2_ISR_RXQ_GROUP_REG(rxq) (0x5400 + 4 * (rxq)) argument 3707 int rxq; in mvpp2_swf_bm_pool_init() local 3721 for (rxq = 0; rxq < rxq_number; rxq++) in mvpp2_swf_bm_pool_init() 3722 mvpp2_rxq_long_pool_set(port, rxq, port->pool_long->id); in mvpp2_swf_bm_pool_init() 3737 for (rxq = 0; rxq < rxq_number; rxq++) in mvpp2_swf_bm_pool_init() 3738 mvpp2_rxq_short_pool_set(port, rxq, in mvpp2_swf_bm_pool_init() [all …]
|
D | sky2.c | 1339 unsigned rxq = rxqaddr[sky2->port]; in sky2_rx_stop() local 1343 sky2_write8(hw, RB_ADDR(rxq, RB_CTRL), RB_DIS_OP_MD); in sky2_rx_stop() 1346 if (sky2_read8(hw, RB_ADDR(rxq, Q_RSL)) in sky2_rx_stop() 1347 == sky2_read8(hw, RB_ADDR(rxq, Q_RL))) in sky2_rx_stop() 1352 sky2_write32(hw, Q_ADDR(rxq, Q_CSR), BMU_RST_SET | BMU_FIFO_RST); in sky2_rx_stop() 1355 sky2_write32(hw, Y2_QADDR(rxq, PREF_UNIT_CTRL), PREF_UNIT_RST_SET); in sky2_rx_stop() 1493 static inline void sky2_rx_update(struct sky2_port *sky2, unsigned rxq) in sky2_rx_update() argument 1495 sky2_put_idx(sky2->hw, rxq, sky2->rx_put); in sky2_rx_update() 1535 unsigned rxq = rxqaddr[sky2->port]; in sky2_rx_start() local 1539 sky2_qset(hw, rxq); in sky2_rx_start() [all …]
|
/linux-4.1.27/drivers/net/wireless/iwlegacy/ |
D | 3945-mac.c | 956 struct il_rx_queue *rxq = &il->rxq; in il3945_rx_queue_restock() local 962 spin_lock_irqsave(&rxq->lock, flags); in il3945_rx_queue_restock() 963 write = rxq->write & ~0x7; in il3945_rx_queue_restock() 964 while (il_rx_queue_space(rxq) > 0 && rxq->free_count) { in il3945_rx_queue_restock() 966 element = rxq->rx_free.next; in il3945_rx_queue_restock() 971 rxq->bd[rxq->write] = in il3945_rx_queue_restock() 973 rxq->queue[rxq->write] = rxb; in il3945_rx_queue_restock() 974 rxq->write = (rxq->write + 1) & RX_QUEUE_MASK; in il3945_rx_queue_restock() 975 rxq->free_count--; in il3945_rx_queue_restock() 977 spin_unlock_irqrestore(&rxq->lock, flags); in il3945_rx_queue_restock() [all …]
|
D | 4965-mac.c | 100 il4965_rx_queue_reset(struct il_priv *il, struct il_rx_queue *rxq) in il4965_rx_queue_reset() argument 104 spin_lock_irqsave(&rxq->lock, flags); in il4965_rx_queue_reset() 105 INIT_LIST_HEAD(&rxq->rx_free); in il4965_rx_queue_reset() 106 INIT_LIST_HEAD(&rxq->rx_used); in il4965_rx_queue_reset() 111 if (rxq->pool[i].page != NULL) { in il4965_rx_queue_reset() 112 pci_unmap_page(il->pci_dev, rxq->pool[i].page_dma, in il4965_rx_queue_reset() 115 __il_free_pages(il, rxq->pool[i].page); in il4965_rx_queue_reset() 116 rxq->pool[i].page = NULL; in il4965_rx_queue_reset() 118 list_add_tail(&rxq->pool[i].list, &rxq->rx_used); in il4965_rx_queue_reset() 122 rxq->queue[i] = NULL; in il4965_rx_queue_reset() [all …]
|
D | 4965.h | 58 void il4965_rx_queue_reset(struct il_priv *il, struct il_rx_queue *rxq); 59 int il4965_rx_init(struct il_priv *il, struct il_rx_queue *rxq); 69 void il4965_rx_queue_free(struct il_priv *il, struct il_rx_queue *rxq);
|
D | 3945.c | 793 il3945_rx_init(struct il_priv *il, struct il_rx_queue *rxq) in il3945_rx_init() argument 795 il_wr(il, FH39_RCSR_RBD_BASE(0), rxq->bd_dma); in il3945_rx_init() 796 il_wr(il, FH39_RCSR_RPTR_ADDR(0), rxq->rb_stts_dma); in il3945_rx_init() 972 struct il_rx_queue *rxq = &il->rxq; in il3945_hw_nic_init() local 982 if (!rxq->bd) { in il3945_hw_nic_init() 989 il3945_rx_queue_reset(il, rxq); in il3945_hw_nic_init() 993 il3945_rx_init(il, rxq); in il3945_hw_nic_init() 1000 il_wr(il, FH39_RCSR_WPTR(0), rxq->write & ~7); in il3945_hw_nic_init()
|
D | common.c | 2614 struct il_rx_queue *rxq = &il->rxq; in il_rx_queue_alloc() local 2618 spin_lock_init(&rxq->lock); in il_rx_queue_alloc() 2619 INIT_LIST_HEAD(&rxq->rx_free); in il_rx_queue_alloc() 2620 INIT_LIST_HEAD(&rxq->rx_used); in il_rx_queue_alloc() 2623 rxq->bd = dma_alloc_coherent(dev, 4 * RX_QUEUE_SIZE, &rxq->bd_dma, in il_rx_queue_alloc() 2625 if (!rxq->bd) in il_rx_queue_alloc() 2628 rxq->rb_stts = dma_alloc_coherent(dev, sizeof(struct il_rb_status), in il_rx_queue_alloc() 2629 &rxq->rb_stts_dma, GFP_KERNEL); in il_rx_queue_alloc() 2630 if (!rxq->rb_stts) in il_rx_queue_alloc() 2635 list_add_tail(&rxq->pool[i].list, &rxq->rx_used); in il_rx_queue_alloc() [all …]
|
D | debug.c | 899 struct il_rx_queue *rxq = &il->rxq; in il_dbgfs_rx_queue_read() local 904 pos += scnprintf(buf + pos, bufsz - pos, "read: %u\n", rxq->read); in il_dbgfs_rx_queue_read() 905 pos += scnprintf(buf + pos, bufsz - pos, "write: %u\n", rxq->write); in il_dbgfs_rx_queue_read() 908 rxq->free_count); in il_dbgfs_rx_queue_read() 909 if (rxq->rb_stts) { in il_dbgfs_rx_queue_read() 912 le16_to_cpu(rxq->rb_stts-> in il_dbgfs_rx_queue_read()
|
D | 3945.h | 194 void il3945_rx_queue_reset(struct il_priv *il, struct il_rx_queue *rxq);
|
D | common.h | 1270 struct il_rx_queue rxq; member
|
/linux-4.1.27/drivers/net/ethernet/atheros/alx/ |
D | main.c | 75 struct alx_rx_queue *rxq = &alx->rxq; in alx_refill_rx_ring() local 81 next = cur = rxq->write_idx; in alx_refill_rx_ring() 84 cur_buf = &rxq->bufs[cur]; in alx_refill_rx_ring() 86 while (!cur_buf->skb && next != rxq->read_idx) { in alx_refill_rx_ring() 87 struct alx_rfd *rfd = &rxq->rfd[cur]; in alx_refill_rx_ring() 116 cur_buf = &rxq->bufs[cur]; in alx_refill_rx_ring() 123 rxq->write_idx = cur; in alx_refill_rx_ring() 189 struct alx_rx_queue *rxq = &alx->rxq; in alx_clean_rx_irq() local 197 rrd = &rxq->rrd[rxq->rrd_read_idx]; in alx_clean_rx_irq() 203 RRD_SI) != rxq->read_idx || in alx_clean_rx_irq() [all …]
|
D | alx.h | 101 struct alx_rx_queue rxq; member
|
D | hw.c | 378 u32 rxq, txq, val; in alx_stop_mac() local 381 rxq = alx_read_mem32(hw, ALX_RXQ0); in alx_stop_mac() 382 alx_write_mem32(hw, ALX_RXQ0, rxq & ~ALX_RXQ0_EN); in alx_stop_mac() 609 u32 mac, txq, rxq; in alx_start_mac() local 611 rxq = alx_read_mem32(hw, ALX_RXQ0); in alx_start_mac() 612 alx_write_mem32(hw, ALX_RXQ0, rxq | ALX_RXQ0_EN); in alx_start_mac()
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | cxgb4vf_main.c | 324 int rxq, msi, err; in request_msix_queue_irqs() local 338 for_each_ethrxq(s, rxq) { in request_msix_queue_irqs() 342 &s->ethrxq[rxq].rspq); in request_msix_queue_irqs() 350 while (--rxq >= 0) in request_msix_queue_irqs() 351 free_irq(adapter->msix_info[--msi].vec, &s->ethrxq[rxq].rspq); in request_msix_queue_irqs() 362 int rxq, msi; in free_msix_queue_irqs() local 366 for_each_ethrxq(s, rxq) in free_msix_queue_irqs() 368 &s->ethrxq[rxq].rspq); in free_msix_queue_irqs() 393 int rxq; in enable_rx() local 396 for_each_ethrxq(s, rxq) in enable_rx() [all …]
|
D | sge.c | 1551 static void do_gro(struct sge_eth_rxq *rxq, const struct pkt_gl *gl, in do_gro() argument 1554 struct adapter *adapter = rxq->rspq.adapter; in do_gro() 1559 skb = napi_get_frags(&rxq->rspq.napi); in do_gro() 1562 rxq->stats.rx_drops++; in do_gro() 1571 skb_record_rx_queue(skb, rxq->rspq.idx); in do_gro() 1576 rxq->stats.vlan_ex++; in do_gro() 1578 ret = napi_gro_frags(&rxq->rspq.napi); in do_gro() 1581 rxq->stats.lro_pkts++; in do_gro() 1583 rxq->stats.lro_merged++; in do_gro() 1584 rxq->stats.pkts++; in do_gro() [all …]
|
/linux-4.1.27/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_vfpf.c | 618 req->rxq.rcq_addr = fp->rx_comp_mapping; in bnx2x_vfpf_setup_q() 619 req->rxq.rcq_np_addr = fp->rx_comp_mapping + BCM_PAGE_SIZE; in bnx2x_vfpf_setup_q() 620 req->rxq.rxq_addr = fp->rx_desc_mapping; in bnx2x_vfpf_setup_q() 621 req->rxq.sge_addr = fp->rx_sge_mapping; in bnx2x_vfpf_setup_q() 622 req->rxq.vf_sb = fp_idx; in bnx2x_vfpf_setup_q() 623 req->rxq.sb_index = HC_INDEX_ETH_RX_CQ_CONS; in bnx2x_vfpf_setup_q() 624 req->rxq.hc_rate = bp->rx_ticks ? 1000000/bp->rx_ticks : 0; in bnx2x_vfpf_setup_q() 625 req->rxq.mtu = bp->dev->mtu; in bnx2x_vfpf_setup_q() 626 req->rxq.buf_sz = fp->rx_buf_size; in bnx2x_vfpf_setup_q() 627 req->rxq.sge_buf_sz = BCM_PAGE_SIZE * PAGES_PER_SGE; in bnx2x_vfpf_setup_q() [all …]
|
D | bnx2x_vfpf.h | 262 } rxq; member
|
D | bnx2x_sriov.c | 1954 struct bnx2x_vf_queue *rxq = vfq_get(vf, j); in bnx2x_iov_adjust_stats_req() local 1960 if (bnx2x_get_q_logical_state(bp, &rxq->sp_obj) == in bnx2x_iov_adjust_stats_req() 1966 cur_query_entry->index = vfq_stat_id(vf, rxq); in bnx2x_iov_adjust_stats_req()
|
/linux-4.1.27/drivers/net/ethernet/samsung/sxgbe/ |
D | sxgbe_main.c | 330 priv->hw->desc->init_rx_desc(&priv->rxq[j]->dma_rx[i], in sxgbe_clear_descriptors() 588 priv->rxq[queue_num], rx_rsize); in init_dma_desc_rings() 597 priv->rxq[queue_num]->priv_ptr = priv; in init_dma_desc_rings() 611 free_rx_ring(priv->device, priv->rxq[queue_num], rx_rsize); in init_dma_desc_rings() 663 free_rx_ring(priv->device, priv->rxq[queue_num], rx_rsize); in free_dma_desc_resources() 686 priv->rxq[queue_num] = devm_kmalloc(priv->device, in rxring_mem_alloc() 688 if (!priv->rxq[queue_num]) in rxring_mem_alloc() 973 (priv->rxq[queue_num])->dma_rx_phy, in sxgbe_init_dma_engine() 1148 (priv->rxq[queue_num])->irq_no, in sxgbe_open() 1150 dev->name, priv->rxq[queue_num]); in sxgbe_open() [all …]
|
D | sxgbe_platform.c | 138 priv->rxq[i]->irq_no = irq_of_parse_and_map(node, chan++); in sxgbe_platform_probe() 139 if (priv->rxq[i]->irq_no <= 0) { in sxgbe_platform_probe() 159 irq_dispose_mapping(priv->rxq[i]->irq_no); in sxgbe_platform_probe()
|
D | sxgbe_common.h | 459 struct sxgbe_rx_queue *rxq[SXGBE_RX_QUEUES]; member
|
/linux-4.1.27/drivers/net/ethernet/freescale/ |
D | fec_main.c | 227 struct fec_enet_priv_rx_q *rxq = fep->rx_queue[queue_id]; in fec_enet_get_nextdesc() local 237 base = rxq->rx_bd_base; in fec_enet_get_nextdesc() 238 ring_size = rxq->rx_ring_size; in fec_enet_get_nextdesc() 239 ex_base = (struct bufdesc_ex *)rxq->rx_bd_base; in fec_enet_get_nextdesc() 258 struct fec_enet_priv_rx_q *rxq = fep->rx_queue[queue_id]; in fec_enet_get_prevdesc() local 268 base = rxq->rx_bd_base; in fec_enet_get_prevdesc() 269 ring_size = rxq->rx_ring_size; in fec_enet_get_prevdesc() 270 ex_base = (struct bufdesc_ex *)rxq->rx_bd_base; in fec_enet_get_prevdesc() 807 struct fec_enet_priv_rx_q *rxq; in fec_enet_bd_init() local 814 rxq = fep->rx_queue[q]; in fec_enet_bd_init() [all …]
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4/ |
D | sge.c | 1796 static void do_gro(struct sge_eth_rxq *rxq, const struct pkt_gl *gl, in do_gro() argument 1799 struct adapter *adapter = rxq->rspq.adap; in do_gro() 1804 skb = napi_get_frags(&rxq->rspq.napi); in do_gro() 1807 rxq->stats.rx_drops++; in do_gro() 1816 skb_record_rx_queue(skb, rxq->rspq.idx); in do_gro() 1817 skb_mark_napi_id(skb, &rxq->rspq.napi); in do_gro() 1818 if (rxq->rspq.netdev->features & NETIF_F_RXHASH) in do_gro() 1824 rxq->stats.vlan_ex++; in do_gro() 1826 ret = napi_gro_frags(&rxq->rspq.napi); in do_gro() 1828 rxq->stats.lro_pkts++; in do_gro() [all …]
|
D | cxgb4_main.c | 683 struct sge_ofld_rxq *rxq = container_of(q, struct sge_ofld_rxq, rspq); in uldrx_handler() local 692 rxq->stats.nomem++; in uldrx_handler() 696 rxq->stats.imm++; in uldrx_handler() 698 rxq->stats.an++; in uldrx_handler() 700 rxq->stats.pkts++; in uldrx_handler()
|
D | cxgb4.h | 1271 unsigned int rxqi, unsigned int rxq, unsigned int tc,
|
D | t4_hw.c | 4461 unsigned int rxqi, unsigned int rxq, unsigned int tc, in t4_cfg_pfvf() argument 4473 FW_PFVF_CMD_NIQ_V(rxq)); in t4_cfg_pfvf()
|
/linux-4.1.27/drivers/net/ethernet/brocade/bna/ |
D | bnad_ethtool.c | 668 rcb[1]->rxq) { in bnad_get_strings() 750 bnad->rx_info[i].rx_ctrl[j].ccb->rcb[1]->rxq) in bnad_get_stats_count_locked() 775 bnad->rx_info[i].rx_ctrl[j].ccb->rcb[0]->rxq) { in bnad_per_q_stats_fill() 801 rcb[0]->rxq) { in bnad_per_q_stats_fill() 804 buf[bi++] = rcb->rxq->rx_packets; in bnad_per_q_stats_fill() 805 buf[bi++] = rcb->rxq->rx_bytes; in bnad_per_q_stats_fill() 806 buf[bi++] = rcb->rxq-> in bnad_per_q_stats_fill() 808 buf[bi++] = rcb->rxq-> in bnad_per_q_stats_fill() 815 rcb[1]->rxq) { in bnad_per_q_stats_fill() 818 buf[bi++] = rcb->rxq->rx_packets; in bnad_per_q_stats_fill() [all …]
|
D | bna_tx_rx.c | 1959 struct bna_rxq *rxq = NULL; in bna_rxq_get() local 1964 rxq = (struct bna_rxq *)qe; in bna_rxq_get() 1965 bfa_q_qe_init(&rxq->qe); in bna_rxq_get() 1967 return rxq; in bna_rxq_get() 1971 bna_rxq_put(struct bna_rx_mod *rx_mod, struct bna_rxq *rxq) in bna_rxq_put() argument 1973 bfa_q_qe_init(&rxq->qe); in bna_rxq_put() 1974 list_add_tail(&rxq->qe, &rx_mod->rxq_free_q); in bna_rxq_put() 2058 rxp->rxq.single.only = q0; in bna_rxp_add_rxqs() 2059 rxp->rxq.single.reserved = NULL; in bna_rxp_add_rxqs() 2062 rxp->rxq.slr.large = q0; in bna_rxp_add_rxqs() [all …]
|
D | bna.h | 310 (q0) = rxp->rxq.single.only; \ 314 (q0) = rxp->rxq.slr.large; \ 315 (q1) = rxp->rxq.slr.small; \ 318 (q0) = rxp->rxq.hds.data; \ 319 (q1) = rxp->rxq.hds.hdr; \
|
D | bna_types.h | 574 struct bna_rxq *rxq; member 711 union bna_rxq_u rxq; member 858 struct bna_rxq *rxq; /* BFI_MAX_RXQ entries */ member
|
D | bnad.c | 291 order = get_order(rcb->rxq->buffer_size); in bnad_rxq_alloc_init() 297 unmap_q->map_size = rcb->rxq->buffer_size; in bnad_rxq_alloc_init() 299 if (rcb->rxq->multi_buffer) { in bnad_rxq_alloc_init() 301 unmap_q->map_size = rcb->rxq->buffer_size; in bnad_rxq_alloc_init() 306 (rcb->rxq->buffer_size > 2048) ? in bnad_rxq_alloc_init() 396 rcb->rxq->rxbuf_alloc_failed++; in bnad_rxq_refill_page() 441 buff_sz = rcb->rxq->buffer_size; in bnad_rxq_refill_skb() 453 rcb->rxq->rxbuf_alloc_failed++; in bnad_rxq_refill_skb() 685 rcb->rxq->rx_packets_with_error++; in bnad_cq_process() 695 rcb->rxq->rx_packets++; in bnad_cq_process() [all …]
|
/linux-4.1.27/drivers/net/wireless/ |
D | mwl8k.c | 247 struct mwl8k_rx_queue rxq[MWL8K_RX_QUEUES]; member 1152 struct mwl8k_rx_queue *rxq = priv->rxq + index; in mwl8k_rxq_init() local 1156 rxq->rxd_count = 0; in mwl8k_rxq_init() 1157 rxq->head = 0; in mwl8k_rxq_init() 1158 rxq->tail = 0; in mwl8k_rxq_init() 1162 rxq->rxd = pci_zalloc_consistent(priv->pdev, size, &rxq->rxd_dma); in mwl8k_rxq_init() 1163 if (rxq->rxd == NULL) { in mwl8k_rxq_init() 1168 rxq->buf = kcalloc(MWL8K_RX_DESCS, sizeof(*rxq->buf), GFP_KERNEL); in mwl8k_rxq_init() 1169 if (rxq->buf == NULL) { in mwl8k_rxq_init() 1170 pci_free_consistent(priv->pdev, size, rxq->rxd, rxq->rxd_dma); in mwl8k_rxq_init() [all …]
|
/linux-4.1.27/drivers/net/ethernet/xscale/ |
D | ixp4xx_eth.c | 678 qmgr_disable_irq(port->plat->rxq); in eth_rx_irq() 686 unsigned int rxq = port->plat->rxq, rxfreeq = RXFREE_QUEUE(port->id); in eth_poll() local 702 if ((n = queue_get_desc(rxq, port, 0)) < 0) { in eth_poll() 708 qmgr_enable_irq(rxq); in eth_poll() 709 if (!qmgr_stat_below_low_watermark(rxq) && in eth_poll() 716 qmgr_disable_irq(rxq); in eth_poll() 1074 err = qmgr_request_queue(port->plat->rxq, RX_DESCS, 0, 0, in request_queues() 1103 qmgr_release_queue(port->plat->rxq); in request_queues() 1114 qmgr_release_queue(port->plat->rxq); in release_queues() 1227 msg.byte5 = port->plat->rxq | 0x80; in eth_open() [all …]
|
/linux-4.1.27/drivers/net/ethernet/intel/i40evf/ |
D | i40evf_virtchnl.c | 243 vqpi->rxq.vsi_id = vqci->vsi_id; in i40evf_configure_queues() 244 vqpi->rxq.queue_id = i; in i40evf_configure_queues() 245 vqpi->rxq.ring_len = adapter->rx_rings[i]->count; in i40evf_configure_queues() 246 vqpi->rxq.dma_ring_addr = adapter->rx_rings[i]->dma; in i40evf_configure_queues() 247 vqpi->rxq.max_pkt_size = adapter->netdev->mtu in i40evf_configure_queues() 249 vqpi->rxq.databuffer_size = adapter->rx_rings[i]->rx_buf_len; in i40evf_configure_queues()
|
D | i40e_virtchnl.h | 208 struct i40e_virtchnl_rxq_info rxq; member
|
/linux-4.1.27/drivers/net/ethernet/ibm/ |
D | ibmveth.h | 50 #define h_register_logical_lan(ua, buflst, rxq, fltlst, mac) \ argument 51 plpar_hcall_norets(H_REGISTER_LOGICAL_LAN, ua, buflst, rxq, fltlst, mac)
|
/linux-4.1.27/drivers/net/usb/ |
D | usbnet.c | 491 spin_lock_irqsave (&dev->rxq.lock, lockflags); in rx_submit() 517 __usbnet_queue_skb(&dev->rxq, skb, rx_start); in rx_submit() 523 spin_unlock_irqrestore (&dev->rxq.lock, lockflags); in rx_submit() 640 state = defer_bh(dev, skb, &dev->rxq, state); in rx_complete() 744 (void) unlink_urbs (dev, &dev->rxq); in usbnet_unlink_rx_urbs() 762 unlink_urbs(dev, &dev->rxq); in usbnet_terminate_urbs() 765 while (!skb_queue_empty(&dev->rxq) in usbnet_terminate_urbs() 1038 unlink_urbs(dev, &dev->rxq); in __handle_link_change() 1104 unlink_urbs (dev, &dev->rxq); in usbnet_deferred_kevent() 1432 for (i = 0; i < 10 && dev->rxq.qlen < RX_QLEN(dev); i++) { in rx_alloc_submit() [all …]
|
/linux-4.1.27/arch/arm/mach-ixp4xx/ |
D | vulcan-setup.c | 126 .rxq = 3, 131 .rxq = 4,
|
D | omixp-setup.c | 177 .rxq = 3, 181 .rxq = 4,
|
D | fsg-setup.c | 130 .rxq = 3, 134 .rxq = 4,
|
D | ixdp425-setup.c | 185 .rxq = 3, 189 .rxq = 4,
|
D | goramo_mlr.c | 284 .rxq = 3, 288 .rxq = 4,
|
D | nslu2-setup.c | 175 .rxq = 3,
|
D | nas100d-setup.c | 163 .rxq = 3,
|
/linux-4.1.27/drivers/mailbox/ |
D | omap-mailbox.c | 114 struct omap_mbox_queue *rxq; member 336 struct omap_mbox_queue *mq = mbox->rxq; in __mbox_rx_interrupt() 356 schedule_work(&mbox->rxq->work); in __mbox_rx_interrupt() 411 mbox->rxq = mq; in omap_mbox_startup() 426 mbox_queue_free(mbox->rxq); in omap_mbox_startup() 434 flush_work(&mbox->rxq->work); in omap_mbox_fini() 435 mbox_queue_free(mbox->rxq); in omap_mbox_fini()
|
/linux-4.1.27/drivers/net/wireless/ipw2x00/ |
D | ipw2200.c | 3451 struct ipw_rx_queue *rxq) in ipw_rx_queue_reset() argument 3456 spin_lock_irqsave(&rxq->lock, flags); in ipw_rx_queue_reset() 3458 INIT_LIST_HEAD(&rxq->rx_free); in ipw_rx_queue_reset() 3459 INIT_LIST_HEAD(&rxq->rx_used); in ipw_rx_queue_reset() 3465 if (rxq->pool[i].skb != NULL) { in ipw_rx_queue_reset() 3466 pci_unmap_single(priv->pci_dev, rxq->pool[i].dma_addr, in ipw_rx_queue_reset() 3468 dev_kfree_skb(rxq->pool[i].skb); in ipw_rx_queue_reset() 3469 rxq->pool[i].skb = NULL; in ipw_rx_queue_reset() 3471 list_add_tail(&rxq->pool[i].list, &rxq->rx_used); in ipw_rx_queue_reset() 3476 rxq->read = rxq->write = 0; in ipw_rx_queue_reset() [all …]
|
D | ipw2100.c | 2675 struct ipw2100_bd_queue *rxq = &priv->rx_queue; in __ipw2100_rx_process() local 2688 if (r >= rxq->entries) { in __ipw2100_rx_process() 2693 i = (rxq->next + 1) % rxq->entries; in __ipw2100_rx_process() 2765 rxq->drv[i].status.info.field = 0; in __ipw2100_rx_process() 2767 i = (i + 1) % rxq->entries; in __ipw2100_rx_process() 2772 rxq->next = (i ? i : rxq->entries) - 1; in __ipw2100_rx_process() 2775 IPW_MEM_HOST_SHARED_RX_WRITE_INDEX, rxq->next); in __ipw2100_rx_process()
|
D | ipw2200.h | 1205 struct ipw_rx_queue *rxq; member
|
/linux-4.1.27/drivers/net/ |
D | macvtap.c | 231 __u32 rxq; in macvtap_get_queue() local 237 rxq = skb_get_hash(skb); in macvtap_get_queue() 238 if (rxq) { in macvtap_get_queue() 239 tap = rcu_dereference(vlan->taps[rxq % numvtaps]); in macvtap_get_queue() 244 rxq = skb_get_rx_queue(skb); in macvtap_get_queue() 246 while (unlikely(rxq >= numvtaps)) in macvtap_get_queue() 247 rxq -= numvtaps; in macvtap_get_queue() 249 tap = rcu_dereference(vlan->taps[rxq]); in macvtap_get_queue()
|
D | xen-netfront.c | 877 struct sk_buff_head *rxq) in handle_incoming_queue() argument 883 while ((skb = __skb_dequeue(rxq)) != NULL) { in handle_incoming_queue() 922 struct sk_buff_head rxq; in xennet_poll() local 929 skb_queue_head_init(&rxq); in xennet_poll() 982 __skb_queue_tail(&rxq, skb); in xennet_poll() 990 work_done -= handle_incoming_queue(queue, &rxq); in xennet_poll()
|
D | virtio_net.c | 173 static int rxq2vq(int rxq) in rxq2vq() argument 175 return rxq * 2; in rxq2vq()
|
/linux-4.1.27/arch/arm/mach-ixp4xx/include/mach/ |
D | platform.h | 101 u8 rxq; /* configurable, currently 0 - 31 only */ member
|
/linux-4.1.27/drivers/net/ethernet/sfc/ |
D | vfdi.h | 198 u32 rxq; member
|
D | ptp.c | 272 struct sk_buff_head rxq; member 998 while ((skb = skb_dequeue(&ptp->rxq))) { in efx_ptp_process_events() 1013 skb_queue_head(&ptp->rxq, skb); in efx_ptp_process_events() 1129 efx_ptp_deliver_rx_queue(&efx->ptp_data->rxq); in efx_ptp_stop() 1226 skb_queue_head_init(&ptp->rxq); in efx_ptp_probe() 1315 skb_queue_purge(&efx->ptp_data->rxq); in efx_ptp_remove() 1439 skb_queue_tail(&ptp->rxq, skb); in efx_ptp_rx()
|
D | siena_sriov.c | 765 unsigned vf_rxq = req->u.mac_filter.rxq; in efx_vfdi_insert_filter()
|
/linux-4.1.27/Documentation/devicetree/bindings/net/ |
D | brcm,systemport.txt | 19 - systemport,num-rxq: number of HW receive queues, an integer
|
/linux-4.1.27/drivers/net/ethernet/intel/i40e/ |
D | i40e_virtchnl.h | 208 struct i40e_virtchnl_rxq_info rxq; member
|
D | i40e_virtchnl_pf.c | 1281 (qpi->rxq.vsi_id != vsi_id) || in i40e_vc_config_queues_msg() 1282 (qpi->rxq.queue_id != vsi_queue_id) || in i40e_vc_config_queues_msg() 1289 &qpi->rxq) || in i40e_vc_config_queues_msg()
|
/linux-4.1.27/drivers/net/wan/ |
D | ixp4xx_hss.c | 658 unsigned int rxq = queue_ids[port->id].rx; in hss_hdlc_poll() local 675 if ((n = queue_get_desc(rxq, port, 0)) < 0) { in hss_hdlc_poll() 681 qmgr_enable_irq(rxq); in hss_hdlc_poll() 682 if (!qmgr_stat_empty(rxq) && in hss_hdlc_poll() 689 qmgr_disable_irq(rxq); in hss_hdlc_poll()
|
/linux-4.1.27/include/linux/usb/ |
D | usbnet.h | 57 struct sk_buff_head rxq; member
|
/linux-4.1.27/drivers/net/ethernet/emulex/benet/ |
D | be_main.c | 1748 struct be_queue_info *rxq = &rxo->q; in get_rx_page_info() local 1749 u16 frag_idx = rxq->tail; in get_rx_page_info() 1765 queue_tail_inc(rxq); in get_rx_page_info() 1766 atomic_dec(&rxq->used); in get_rx_page_info() 2060 struct be_queue_info *rxq = &rxo->q; in be_post_rx_frags() local 2067 page_info = &rxo->page_info_tbl[rxq->head]; in be_post_rx_frags() 2092 rxd = queue_head_node(rxq); in be_post_rx_frags() 2108 queue_head_inc(rxq); in be_post_rx_frags() 2109 page_info = &rxo->page_info_tbl[rxq->head]; in be_post_rx_frags() 2121 atomic_add(posted, &rxq->used); in be_post_rx_frags() [all …]
|
D | be_cmds.c | 1311 struct be_queue_info *rxq, u16 cq_id, u16 frag_size, in be_cmd_rxq_create() argument 1316 struct be_dma_mem *q_mem = &rxq->dma_mem; in be_cmd_rxq_create() 1343 rxq->id = le16_to_cpu(resp->id); in be_cmd_rxq_create() 1344 rxq->created = true; in be_cmd_rxq_create()
|
D | be_cmds.h | 2255 int be_cmd_rxq_create(struct be_adapter *adapter, struct be_queue_info *rxq,
|
/linux-4.1.27/drivers/net/wireless/ath/ath6kl/ |
D | htc_mbox.c | 1913 struct list_head *rxq, in ath6kl_htc_rx_bundle() argument 1922 n_scat_pkt = get_queue_depth(rxq); in ath6kl_htc_rx_bundle() 1925 if ((get_queue_depth(rxq) - n_scat_pkt) > 0) { in ath6kl_htc_rx_bundle() 1938 __func__, get_queue_depth(rxq), n_scat_pkt); in ath6kl_htc_rx_bundle() 1945 get_queue_depth(rxq), n_scat_pkt); in ath6kl_htc_rx_bundle() 1955 packet = list_first_entry(rxq, struct htc_packet, list); in ath6kl_htc_rx_bundle() 1962 list_add(&packet->list, rxq); in ath6kl_htc_rx_bundle()
|
/linux-4.1.27/drivers/net/ethernet/broadcom/ |
D | bcmsysport.c | 1665 u32 txq, rxq; in bcm_sysport_probe() local 1674 if (of_property_read_u32(dn, "systemport,num-rxq", &rxq)) in bcm_sysport_probe() 1675 rxq = 1; in bcm_sysport_probe() 1677 dev = alloc_etherdev_mqs(sizeof(*priv), txq, rxq); in bcm_sysport_probe() 1764 priv->base, priv->irq0, priv->irq1, txq, rxq); in bcm_sysport_probe()
|
/linux-4.1.27/drivers/net/xen-netback/ |
D | netback.c | 502 struct sk_buff_head rxq; in xenvif_rx_action() local 514 skb_queue_head_init(&rxq); in xenvif_rx_action() 527 __skb_queue_tail(&rxq, skb); in xenvif_rx_action() 538 while ((skb = __skb_dequeue(&rxq)) != NULL) { in xenvif_rx_action()
|
/linux-4.1.27/net/core/ |
D | net-sysfs.c | 1321 int error = 0, txq = 0, rxq = 0, real_rx = 0, real_tx = 0; in register_queue_kobjects() local 1335 rxq = real_rx; in register_queue_kobjects() 1346 net_rx_queue_update_kobjects(dev, rxq, 0); in register_queue_kobjects()
|
D | dev.c | 2171 int netif_set_real_num_rx_queues(struct net_device *dev, unsigned int rxq) in netif_set_real_num_rx_queues() argument 2175 if (rxq < 1 || rxq > dev->num_rx_queues) in netif_set_real_num_rx_queues() 2182 rxq); in netif_set_real_num_rx_queues() 2187 dev->real_num_rx_queues = rxq; in netif_set_real_num_rx_queues() 3529 static int ing_filter(struct sk_buff *skb, struct netdev_queue *rxq) in ing_filter() argument 3545 q = rcu_dereference(rxq->qdisc); in ing_filter() 3560 struct netdev_queue *rxq = rcu_dereference(skb->dev->ingress_queue); in handle_ing() local 3562 if (!rxq || rcu_access_pointer(rxq->qdisc) == &noop_qdisc) in handle_ing() 3570 switch (ing_filter(skb, rxq)) { in handle_ing()
|
/linux-4.1.27/drivers/infiniband/hw/cxgb3/ |
D | iwch_cm.c | 107 static struct sk_buff_head rxq; variable 2187 while ((skb = skb_dequeue(&rxq))) { in process_work() 2218 skb_queue_tail(&rxq, skb); in sched() 2259 skb_queue_head_init(&rxq); in iwch_cm_init()
|
/linux-4.1.27/drivers/net/ethernet/atheros/atl1c/ |
D | atl1c_main.c | 1198 u32 mac, txq, rxq; in atl1c_start_mac() local 1205 AT_READ_REG(hw, REG_RXQ_CTRL, &rxq); in atl1c_start_mac() 1209 rxq |= RXQ_CTRL_EN; in atl1c_start_mac() 1223 AT_WRITE_REG(hw, REG_RXQ_CTRL, rxq); in atl1c_start_mac()
|
/linux-4.1.27/drivers/infiniband/hw/cxgb4/ |
D | cm.c | 142 static struct sk_buff_head rxq; variable 3876 while ((skb = skb_dequeue(&rxq))) { in process_work() 3925 skb_queue_tail(&rxq, skb); in sched() 4039 skb_queue_head_init(&rxq); in c4iw_cm_init()
|
/linux-4.1.27/include/linux/ |
D | netdevice.h | 2867 int netif_set_real_num_rx_queues(struct net_device *dev, unsigned int rxq); 2870 unsigned int rxq) in netif_set_real_num_rx_queues() argument
|