rxq_index 11035 drivers/net/ethernet/broadcom/bnxt/bnxt.c u16 rxq_index, u32 flow_id) rxq_index 11115 drivers/net/ethernet/broadcom/bnxt/bnxt.c new_fltr->rxq = rxq_index; rxq_index 172 drivers/net/ethernet/cisco/enic/enic_clsf.c u16 rxq_index, u32 flow_id) rxq_index 192 drivers/net/ethernet/cisco/enic/enic_clsf.c if (rxq_index == n->rq_id) { rxq_index 214 drivers/net/ethernet/cisco/enic/enic_clsf.c res = enic_addfltr_5t(enic, &keys, rxq_index); rxq_index 224 drivers/net/ethernet/cisco/enic/enic_clsf.c res = enic_addfltr_5t(enic, &keys, rxq_index); rxq_index 248 drivers/net/ethernet/cisco/enic/enic_clsf.c n->rq_id = rxq_index; rxq_index 267 drivers/net/ethernet/cisco/enic/enic_clsf.c res = enic_addfltr_5t(enic, &keys, rxq_index); rxq_index 273 drivers/net/ethernet/cisco/enic/enic_clsf.c n->rq_id = rxq_index; rxq_index 18 drivers/net/ethernet/cisco/enic/enic_clsf.h u16 rxq_index, u32 flow_id); rxq_index 1793 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c u16 rxq_index, u32 flow_id) rxq_index 1813 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c return h->ae_algo->ops->add_arfs_entry(h, rxq_index, flow_id, &fkeys); rxq_index 1085 drivers/net/ethernet/intel/ice/ice_common.c ice_copy_rxq_ctx_to_hw(struct ice_hw *hw, u8 *ice_rxq_ctx, u32 rxq_index) rxq_index 1092 drivers/net/ethernet/intel/ice/ice_common.c if (rxq_index > QRX_CTRL_MAX_INDEX) rxq_index 1097 drivers/net/ethernet/intel/ice/ice_common.c wr32(hw, QRX_CONTEXT(i, rxq_index), rxq_index 1145 drivers/net/ethernet/intel/ice/ice_common.c u32 rxq_index) rxq_index 1155 drivers/net/ethernet/intel/ice/ice_common.c return ice_copy_rxq_ctx_to_hw(hw, ctx_buf, rxq_index); rxq_index 51 drivers/net/ethernet/intel/ice/ice_common.h u32 rxq_index); rxq_index 169 drivers/net/ethernet/mellanox/mlx4/en_netdev.c int rxq_index; rxq_index 249 drivers/net/ethernet/mellanox/mlx4/en_netdev.c rule.qpn = priv->rss_map.qps[filter->rxq_index].qpn; rxq_index 288 drivers/net/ethernet/mellanox/mlx4/en_netdev.c mlx4_en_filter_alloc(struct mlx4_en_priv *priv, int rxq_index, __be32 src_ip, rxq_index 299 drivers/net/ethernet/mellanox/mlx4/en_netdev.c filter->rxq_index = rxq_index; rxq_index 360 drivers/net/ethernet/mellanox/mlx4/en_netdev.c u16 rxq_index, u32 flow_id) rxq_index 395 drivers/net/ethernet/mellanox/mlx4/en_netdev.c if (filter->rxq_index == rxq_index) rxq_index 398 drivers/net/ethernet/mellanox/mlx4/en_netdev.c filter->rxq_index = rxq_index; rxq_index 400 drivers/net/ethernet/mellanox/mlx4/en_netdev.c filter = mlx4_en_filter_alloc(priv, rxq_index, rxq_index 451 drivers/net/ethernet/mellanox/mlx4/en_netdev.c filter->rxq_index, filter->flow_id, rxq_index 205 drivers/net/ethernet/mellanox/mlx5/core/en/fs.h u16 rxq_index, u32 flow_id); rxq_index 672 drivers/net/ethernet/mellanox/mlx5/core/en_arfs.c u16 rxq_index, u32 flow_id) rxq_index 697 drivers/net/ethernet/mellanox/mlx5/core/en_arfs.c if (arfs_rule->rxq == rxq_index) { rxq_index 701 drivers/net/ethernet/mellanox/mlx5/core/en_arfs.c arfs_rule->rxq = rxq_index; rxq_index 703 drivers/net/ethernet/mellanox/mlx5/core/en_arfs.c arfs_rule = arfs_alloc_rule(priv, arfs_t, &fk, rxq_index, flow_id); rxq_index 61 drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c int ionic_rx_filter_save(struct ionic_lif *lif, u32 flow_id, u16 rxq_index, rxq_index 92 drivers/net/ethernet/pensando/ionic/ionic_rx_filter.c f->rxq_index = rxq_index; rxq_index 11 drivers/net/ethernet/pensando/ionic/ionic_rx_filter.h u16 rxq_index; rxq_index 30 drivers/net/ethernet/pensando/ionic/ionic_rx_filter.h int ionic_rx_filter_save(struct ionic_lif *lif, u32 flow_id, u16 rxq_index, rxq_index 491 drivers/net/ethernet/qlogic/qede/qede.h u16 rxq_index, u32 flow_id); rxq_index 454 drivers/net/ethernet/qlogic/qede/qede_filter.c u16 rxq_index, u32 flow_id) rxq_index 491 drivers/net/ethernet/qlogic/qede/qede_filter.c n->next_rxq_id = rxq_index; rxq_index 494 drivers/net/ethernet/qlogic/qede/qede_filter.c if (n->rxq_id != rxq_index) rxq_index 499 drivers/net/ethernet/qlogic/qede/qede_filter.c n->rxq_id = rxq_index; rxq_index 518 drivers/net/ethernet/qlogic/qede/qede_filter.c n->rxq_id = rxq_index; rxq_index 519 drivers/net/ethernet/qlogic/qede/qede_filter.c n->next_rxq_id = rxq_index; rxq_index 1654 drivers/net/ethernet/qlogic/qede/qede_main.c int queue_id, rxq_index = 0, txq_index = 0; rxq_index 1665 drivers/net/ethernet/qlogic/qede/qede_main.c rxq_index); rxq_index 1670 drivers/net/ethernet/qlogic/qede/qede_main.c fp->rxq->rxq_id = rxq_index++; rxq_index 168 drivers/net/ethernet/sfc/efx.h u16 rxq_index, u32 flow_id); rxq_index 165 drivers/net/ethernet/sfc/falcon/efx.h u16 rxq_index, u32 flow_id); rxq_index 831 drivers/net/ethernet/sfc/falcon/rx.c u16 rxq_index, u32 flow_id) rxq_index 852 drivers/net/ethernet/sfc/falcon/rx.c rxq_index); rxq_index 876 drivers/net/ethernet/sfc/falcon/rx.c channel = ef4_get_channel(efx, rxq_index); rxq_index 885 drivers/net/ethernet/sfc/falcon/rx.c ntohs(spec.loc_port), rxq_index, flow_id, rc); rxq_index 891 drivers/net/ethernet/sfc/falcon/rx.c ntohs(spec.loc_port), rxq_index, flow_id, rc); rxq_index 758 drivers/net/ethernet/sfc/net_driver.h u16 rxq_index; rxq_index 778 drivers/net/ethernet/sfc/net_driver.h u16 rxq_index; rxq_index 833 drivers/net/ethernet/sfc/rx.c struct efx_channel *channel = efx_get_channel(efx, req->rxq_index); rxq_index 875 drivers/net/ethernet/sfc/rx.c req->rxq_index, req->flow_id, rc, arfs_id); rxq_index 882 drivers/net/ethernet/sfc/rx.c req->rxq_index, req->flow_id, rc, arfs_id); rxq_index 891 drivers/net/ethernet/sfc/rx.c u16 rxq_index, u32 flow_id) rxq_index 930 drivers/net/ethernet/sfc/rx.c rxq_index); rxq_index 963 drivers/net/ethernet/sfc/rx.c if (!new && rule->rxq_index == rxq_index && rxq_index 966 drivers/net/ethernet/sfc/rx.c rule->rxq_index = rxq_index; rxq_index 981 drivers/net/ethernet/sfc/rx.c req->rxq_index = rxq_index; rxq_index 732 include/linux/netdevice.h bool rps_may_expire_flow(struct net_device *dev, u16 rxq_index, u32 flow_id, rxq_index 1358 include/linux/netdevice.h u16 rxq_index, rxq_index 3907 net/core/dev.c u16 rxq_index; rxq_index 3914 net/core/dev.c rxq_index = cpu_rmap_lookup_index(dev->rx_cpu_rmap, next_cpu); rxq_index 3915 net/core/dev.c if (rxq_index == skb_get_rx_queue(skb)) rxq_index 3918 net/core/dev.c rxqueue = dev->_rx + rxq_index; rxq_index 3924 net/core/dev.c rxq_index, flow_id); rxq_index 4055 net/core/dev.c bool rps_may_expire_flow(struct net_device *dev, u16 rxq_index, rxq_index 4058 net/core/dev.c struct netdev_rx_queue *rxqueue = dev->_rx + rxq_index;