rxqs 2797 drivers/infiniband/core/verbs.c setup, params.txqs, params.rxqs); rxqs 711 drivers/net/can/dev.c unsigned int txqs, unsigned int rxqs) rxqs 737 drivers/net/can/dev.c txqs, rxqs); rxqs 69 drivers/net/ethernet/huawei/hinic/hinic_dev.h struct hinic_rxq *rxqs; rxqs 622 drivers/net/ethernet/huawei/hinic/hinic_ethtool.c if (!nic_dev->rxqs) rxqs 625 drivers/net/ethernet/huawei/hinic/hinic_ethtool.c hinic_rxq_get_stats(&nic_dev->rxqs[qid], &rxq_stats); rxqs 120 drivers/net/ethernet/huawei/hinic/hinic_main.c update_rx_stats(nic_dev, &nic_dev->rxqs[i]); rxqs 198 drivers/net/ethernet/huawei/hinic/hinic_main.c if (nic_dev->rxqs) rxqs 201 drivers/net/ethernet/huawei/hinic/hinic_main.c rxq_size = num_rxqs * sizeof(*nic_dev->rxqs); rxqs 202 drivers/net/ethernet/huawei/hinic/hinic_main.c nic_dev->rxqs = devm_kzalloc(&netdev->dev, rxq_size, GFP_KERNEL); rxqs 203 drivers/net/ethernet/huawei/hinic/hinic_main.c if (!nic_dev->rxqs) rxqs 209 drivers/net/ethernet/huawei/hinic/hinic_main.c err = hinic_init_rxq(&nic_dev->rxqs[i], rq, netdev); rxqs 221 drivers/net/ethernet/huawei/hinic/hinic_main.c hinic_clean_rxq(&nic_dev->rxqs[j]); rxqs 223 drivers/net/ethernet/huawei/hinic/hinic_main.c devm_kfree(&netdev->dev, nic_dev->rxqs); rxqs 236 drivers/net/ethernet/huawei/hinic/hinic_main.c if (!nic_dev->rxqs) rxqs 240 drivers/net/ethernet/huawei/hinic/hinic_main.c hinic_clean_rxq(&nic_dev->rxqs[i]); rxqs 242 drivers/net/ethernet/huawei/hinic/hinic_main.c devm_kfree(&netdev->dev, nic_dev->rxqs); rxqs 243 drivers/net/ethernet/huawei/hinic/hinic_main.c nic_dev->rxqs = NULL; rxqs 969 drivers/net/ethernet/huawei/hinic/hinic_main.c nic_dev->rxqs = NULL; rxqs 415 drivers/net/ethernet/marvell/mvneta.c struct mvneta_rx_queue *rxqs; rxqs 1164 drivers/net/ethernet/marvell/mvneta.c struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; rxqs 2815 drivers/net/ethernet/marvell/mvneta.c &pp->rxqs[rx_queue]); rxqs 2818 drivers/net/ethernet/marvell/mvneta.c &pp->rxqs[rx_queue]); rxqs 3126 drivers/net/ethernet/marvell/mvneta.c mvneta_rxq_deinit(pp, &pp->rxqs[queue]); rxqs 3136 drivers/net/ethernet/marvell/mvneta.c int err = mvneta_rxq_init(pp, &pp->rxqs[queue]); rxqs 3986 drivers/net/ethernet/marvell/mvneta.c struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; rxqs 4008 drivers/net/ethernet/marvell/mvneta.c c->rx_coalesce_usecs = pp->rxqs[0].time_coal; rxqs 4009 drivers/net/ethernet/marvell/mvneta.c c->rx_max_coalesced_frames = pp->rxqs[0].pkts_coal; rxqs 4124 drivers/net/ethernet/marvell/mvneta.c val = pp->rxqs[0].skb_alloc_err; rxqs 4127 drivers/net/ethernet/marvell/mvneta.c val = pp->rxqs[0].refill_err; rxqs 4395 drivers/net/ethernet/marvell/mvneta.c pp->rxqs = devm_kcalloc(dev, rxq_number, sizeof(*pp->rxqs), GFP_KERNEL); rxqs 4396 drivers/net/ethernet/marvell/mvneta.c if (!pp->rxqs) rxqs 4401 drivers/net/ethernet/marvell/mvneta.c struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; rxqs 4782 drivers/net/ethernet/marvell/mvneta.c struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; rxqs 4833 drivers/net/ethernet/marvell/mvneta.c struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; rxqs 906 drivers/net/ethernet/marvell/mvpp2/mvpp2.h struct mvpp2_rx_queue **rxqs; rxqs 596 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c prxq = port->rxqs[lrxq]->id; rxqs 617 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c prxq = port->rxqs[lrxq]->id; rxqs 1741 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c queue = port->rxqs[lrxq]->id; rxqs 1759 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c queue = port->rxqs[lrxq]->id; rxqs 1772 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c queue = port->rxqs[lrxq]->id; rxqs 2255 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c return port->rxqs[queue]; rxqs 2665 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_rxq_deinit(port, port->rxqs[queue]); rxqs 2674 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c err = mvpp2_rxq_init(port, port->rxqs[queue]); rxqs 4046 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_rx_queue *rxq = port->rxqs[queue]; rxqs 4077 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c c->rx_coalesce_usecs = port->rxqs[0]->time_coal; rxqs 4078 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c c->rx_max_coalesced_frames = port->rxqs[0]->pkts_coal; rxqs 4597 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c port->rxqs = devm_kcalloc(dev, port->nrxqs, sizeof(*port->rxqs), rxqs 4599 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c if (!port->rxqs) { rxqs 4619 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c port->rxqs[queue] = rxq; rxqs 4626 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_rx_queue *rxq = port->rxqs[queue]; rxqs 758 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib.c .rxqs = nch, rxqs 406 drivers/net/ethernet/netronome/nfp/nfp_net_repr.c nfp_repr_alloc_mqs(struct nfp_app *app, unsigned int txqs, unsigned int rxqs) rxqs 411 drivers/net/ethernet/netronome/nfp/nfp_net_repr.c netdev = alloc_etherdev_mqs(sizeof(*repr), txqs, rxqs); rxqs 102 drivers/net/ethernet/netronome/nfp/nfp_net_repr.h nfp_repr_alloc_mqs(struct nfp_app *app, unsigned int txqs, unsigned int rxqs); rxqs 181 include/linux/can/dev.h unsigned int txqs, unsigned int rxqs); rxqs 50 include/linux/etherdevice.h unsigned int rxqs); rxqs 56 include/linux/etherdevice.h unsigned int rxqs); rxqs 3553 include/linux/netdevice.h unsigned int rxqs) rxqs 3555 include/linux/netdevice.h dev->real_num_rx_queues = rxqs; rxqs 4092 include/linux/netdevice.h unsigned int txqs, unsigned int rxqs); rxqs 2213 include/rdma/ib_verbs.h unsigned int rxqs; rxqs 9471 net/core/dev.c unsigned int txqs, unsigned int rxqs) rxqs 9484 net/core/dev.c if (rxqs < 1) { rxqs 9548 net/core/dev.c dev->num_rx_queues = rxqs; rxqs 9549 net/core/dev.c dev->real_num_rx_queues = rxqs; rxqs 412 net/ethernet/eth.c unsigned int rxqs) rxqs 415 net/ethernet/eth.c ether_setup, txqs, rxqs); rxqs 425 net/ethernet/eth.c unsigned int txqs, unsigned int rxqs) rxqs 434 net/ethernet/eth.c netdev = alloc_etherdev_mqs(sizeof_priv, txqs, rxqs);