txqs 2797 drivers/infiniband/core/verbs.c setup, params.txqs, params.rxqs); txqs 711 drivers/net/can/dev.c unsigned int txqs, unsigned int rxqs) txqs 737 drivers/net/can/dev.c txqs, rxqs); txqs 68 drivers/net/ethernet/huawei/hinic/hinic_dev.h struct hinic_txq *txqs; txqs 609 drivers/net/ethernet/huawei/hinic/hinic_ethtool.c if (!nic_dev->txqs) txqs 612 drivers/net/ethernet/huawei/hinic/hinic_ethtool.c hinic_txq_get_stats(&nic_dev->txqs[qid], &txq_stats); txqs 123 drivers/net/ethernet/huawei/hinic/hinic_main.c update_tx_stats(nic_dev, &nic_dev->txqs[i]); txqs 138 drivers/net/ethernet/huawei/hinic/hinic_main.c if (nic_dev->txqs) txqs 141 drivers/net/ethernet/huawei/hinic/hinic_main.c txq_size = num_txqs * sizeof(*nic_dev->txqs); txqs 142 drivers/net/ethernet/huawei/hinic/hinic_main.c nic_dev->txqs = devm_kzalloc(&netdev->dev, txq_size, GFP_KERNEL); txqs 143 drivers/net/ethernet/huawei/hinic/hinic_main.c if (!nic_dev->txqs) txqs 149 drivers/net/ethernet/huawei/hinic/hinic_main.c err = hinic_init_txq(&nic_dev->txqs[i], sq, netdev); txqs 161 drivers/net/ethernet/huawei/hinic/hinic_main.c hinic_clean_txq(&nic_dev->txqs[j]); txqs 163 drivers/net/ethernet/huawei/hinic/hinic_main.c devm_kfree(&netdev->dev, nic_dev->txqs); txqs 176 drivers/net/ethernet/huawei/hinic/hinic_main.c if (!nic_dev->txqs) txqs 180 drivers/net/ethernet/huawei/hinic/hinic_main.c hinic_clean_txq(&nic_dev->txqs[i]); txqs 182 drivers/net/ethernet/huawei/hinic/hinic_main.c devm_kfree(&netdev->dev, nic_dev->txqs); txqs 183 drivers/net/ethernet/huawei/hinic/hinic_main.c nic_dev->txqs = NULL; txqs 968 drivers/net/ethernet/huawei/hinic/hinic_main.c nic_dev->txqs = NULL; txqs 473 drivers/net/ethernet/huawei/hinic/hinic_tx.c txq = &nic_dev->txqs[q_id]; txqs 1404 drivers/net/ethernet/intel/ice/ice_adminq_cmd.h struct ice_aqc_add_txqs_perq txqs[1]; txqs 2812 drivers/net/ethernet/intel/ice/ice_common.c (sizeof(*qg_list) - sizeof(*qg_list->txqs)); txqs 2816 drivers/net/ethernet/intel/ice/ice_common.c struct ice_aqc_add_txqs_perq *q = list->txqs; txqs 3241 drivers/net/ethernet/intel/ice/ice_common.c buf->txqs[0].info.valid_sections = ICE_AQC_ELEM_VALID_GENERIC; txqs 3247 drivers/net/ethernet/intel/ice/ice_common.c le16_to_cpu(buf->txqs[0].txq_id), txqs 3252 drivers/net/ethernet/intel/ice/ice_common.c node.node_teid = buf->txqs[0].q_teid; txqs 1736 drivers/net/ethernet/intel/ice/ice_lib.c qg_buf->txqs[0].txq_id = cpu_to_le16(pf_q); txqs 1737 drivers/net/ethernet/intel/ice/ice_lib.c ice_set_ctx((u8 *)&tlan_ctx, qg_buf->txqs[0].txq_ctx, txqs 1763 drivers/net/ethernet/intel/ice/ice_lib.c txq = &qg_buf->txqs[0]; txqs 416 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txqs; txqs 1155 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq = &pp->txqs[queue]; txqs 1763 drivers/net/ethernet/marvell/mvneta.c return &pp->txqs[queue]; txqs 2418 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq = &pp->txqs[txq_id]; txqs 2879 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_done_force(pp, &pp->txqs[queue]); txqs 3117 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_deinit(pp, &pp->txqs[queue]); txqs 3155 drivers/net/ethernet/marvell/mvneta.c int err = mvneta_txq_init(pp, &pp->txqs[queue]); txqs 3994 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq = &pp->txqs[queue]; txqs 4011 drivers/net/ethernet/marvell/mvneta.c c->tx_max_coalesced_frames = pp->txqs[0].done_pkts_coal; txqs 4383 drivers/net/ethernet/marvell/mvneta.c pp->txqs = devm_kcalloc(dev, txq_number, sizeof(*pp->txqs), GFP_KERNEL); txqs 4384 drivers/net/ethernet/marvell/mvneta.c if (!pp->txqs) txqs 4389 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq = &pp->txqs[queue]; txqs 4788 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq = &pp->txqs[queue]; txqs 4840 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq = &pp->txqs[queue]; txqs 908 drivers/net/ethernet/marvell/mvpp2/mvpp2.h struct mvpp2_tx_queue **txqs; txqs 1791 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_tx_queue *txq = port->txqs[queue]; txqs 2085 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c int id = port->txqs[queue]->id; txqs 2263 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c return port->txqs[queue]; txqs 2648 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq = port->txqs[queue]; txqs 2692 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq = port->txqs[queue]; txqs 2705 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq = port->txqs[queue]; txqs 3216 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq = port->txqs[txq_id]; txqs 4060 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_tx_queue *txq = port->txqs[queue]; txqs 4079 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c c->tx_max_coalesced_frames = port->txqs[0]->done_pkts_coal; txqs 4562 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c port->txqs = devm_kcalloc(dev, port->ntxqs, sizeof(*port->txqs), txqs 4564 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c if (!port->txqs) txqs 4594 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c port->txqs[queue] = txq; txqs 4662 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c if (!port->txqs[queue]) txqs 4664 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c free_percpu(port->txqs[queue]->pcpu); txqs 5429 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c free_percpu(port->txqs[i]->pcpu); txqs 5453 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c free_percpu(port->txqs[i]->pcpu); txqs 757 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib.c .txqs = nch * MLX5E_MAX_NUM_TC, txqs 85 drivers/net/ethernet/netronome/nfp/abm/main.c unsigned int txqs; txqs 90 drivers/net/ethernet/netronome/nfp/abm/main.c txqs = 1; txqs 93 drivers/net/ethernet/netronome/nfp/abm/main.c txqs = alink->vnic->max_rx_rings; txqs 96 drivers/net/ethernet/netronome/nfp/abm/main.c netdev = nfp_repr_alloc_mqs(app, txqs, 1); txqs 406 drivers/net/ethernet/netronome/nfp/nfp_net_repr.c nfp_repr_alloc_mqs(struct nfp_app *app, unsigned int txqs, unsigned int rxqs) txqs 411 drivers/net/ethernet/netronome/nfp/nfp_net_repr.c netdev = alloc_etherdev_mqs(sizeof(*repr), txqs, rxqs); txqs 102 drivers/net/ethernet/netronome/nfp/nfp_net_repr.h nfp_repr_alloc_mqs(struct nfp_app *app, unsigned int txqs, unsigned int rxqs); txqs 437 drivers/net/wan/farsync.c int txqs; /* index to get next buffer to tx */ txqs 1349 drivers/net/wan/farsync.c if ((txq_length = port->txqe - port->txqs) < 0) { txqs 1362 drivers/net/wan/farsync.c skb = port->txq[port->txqs]; txqs 1363 drivers/net/wan/farsync.c port->txqs++; txqs 1364 drivers/net/wan/farsync.c if (port->txqs == FST_TXQ_DEPTH) { txqs 1365 drivers/net/wan/farsync.c port->txqs = 0; txqs 2160 drivers/net/wan/farsync.c port->txqs = 0; txqs 2298 drivers/net/wan/farsync.c if ((txq_length = port->txqe - port->txqs) < 0) { txqs 1318 drivers/net/wireless/ath/ath5k/ath5k.h struct ath5k_txq txqs[AR5K_NUM_TX_QUEUES]; /* tx queues */ txqs 1010 drivers/net/wireless/ath/ath5k/base.c txq = &ah->txqs[qnum]; txqs 1022 drivers/net/wireless/ath/ath5k/base.c return &ah->txqs[qnum]; txqs 1116 drivers/net/wireless/ath/ath5k/base.c for (i = 0; i < ARRAY_SIZE(ah->txqs); i++) { txqs 1117 drivers/net/wireless/ath/ath5k/base.c if (ah->txqs[i].setup) { txqs 1118 drivers/net/wireless/ath/ath5k/base.c txq = &ah->txqs[i]; txqs 1141 drivers/net/wireless/ath/ath5k/base.c struct ath5k_txq *txq = ah->txqs; txqs 1144 drivers/net/wireless/ath/ath5k/base.c for (i = 0; i < ARRAY_SIZE(ah->txqs); i++, txq++) txqs 1793 drivers/net/wireless/ath/ath5k/base.c if (ah->txqs[i].setup && (ah->ah_txq_isr_txok_all & BIT(i))) txqs 1794 drivers/net/wireless/ath/ath5k/base.c ath5k_tx_processq(ah, &ah->txqs[i]); txqs 2009 drivers/net/wireless/ath/ath5k/base.c trace_ath5k_tx(ah, bf->skb, &ah->txqs[ah->bhalq]); txqs 2474 drivers/net/wireless/ath/ath5k/base.c for (i = 0; i < ARRAY_SIZE(ah->txqs); i++) { txqs 2475 drivers/net/wireless/ath/ath5k/base.c if (ah->txqs[i].setup) { txqs 2476 drivers/net/wireless/ath/ath5k/base.c txq = &ah->txqs[i]; txqs 842 drivers/net/wireless/ath/ath5k/debug.c for (i = 0; i < ARRAY_SIZE(ah->txqs); i++) { txqs 843 drivers/net/wireless/ath/ath5k/debug.c txq = &ah->txqs[i]; txqs 69 drivers/net/wireless/ath/ath5k/mac80211-ops.c ath5k_tx_queue(hw, skb, &ah->txqs[qnum], control); txqs 744 drivers/net/wireless/ath/ath5k/mac80211-ops.c *tx = ah->txqs[AR5K_TX_QUEUE_ID_DATA_MIN].txq_max; txqs 764 drivers/net/wireless/ath/ath5k/mac80211-ops.c for (qnum = 0; qnum < ARRAY_SIZE(ah->txqs); qnum++) { txqs 765 drivers/net/wireless/ath/ath5k/mac80211-ops.c if (!ah->txqs[qnum].setup) txqs 767 drivers/net/wireless/ath/ath5k/mac80211-ops.c if (ah->txqs[qnum].qnum < AR5K_TX_QUEUE_ID_DATA_MIN || txqs 768 drivers/net/wireless/ath/ath5k/mac80211-ops.c ah->txqs[qnum].qnum > AR5K_TX_QUEUE_ID_DATA_MAX) txqs 771 drivers/net/wireless/ath/ath5k/mac80211-ops.c ah->txqs[qnum].txq_max = tx; txqs 772 drivers/net/wireless/ath/ath5k/mac80211-ops.c if (ah->txqs[qnum].txq_len >= ah->txqs[qnum].txq_max) txqs 773 drivers/net/wireless/ath/ath5k/mac80211-ops.c ieee80211_stop_queue(hw, ah->txqs[qnum].qnum); txqs 584 drivers/net/wireless/intel/iwlwifi/iwl-trans.h void (*freeze_txq_timer)(struct iwl_trans *trans, unsigned long txqs, txqs 1086 drivers/net/wireless/intel/iwlwifi/iwl-trans.h unsigned long txqs, txqs 1095 drivers/net/wireless/intel/iwlwifi/iwl-trans.h trans->ops->freeze_txq_timer(trans, txqs, freeze); txqs 1111 drivers/net/wireless/intel/iwlwifi/iwl-trans.h u32 txqs) txqs 1121 drivers/net/wireless/intel/iwlwifi/iwl-trans.h return trans->ops->wait_tx_queues_empty(trans, txqs); txqs 2764 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c unsigned long txqs = 0, tids = 0; txqs 2782 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c __set_bit(tid_data->txq_id, &txqs); txqs 2795 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c if (txqs) txqs 2796 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c iwl_trans_freeze_txq_timer(mvm->trans, txqs, true); txqs 2807 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c if (txqs) txqs 2808 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c iwl_trans_freeze_txq_timer(mvm->trans, txqs, false); txqs 2215 drivers/net/wireless/intel/iwlwifi/pcie/trans.c unsigned long txqs, txqs 2221 drivers/net/wireless/intel/iwlwifi/pcie/trans.c for_each_set_bit(queue, &txqs, BITS_PER_LONG) { txqs 181 include/linux/can/dev.h unsigned int txqs, unsigned int rxqs); txqs 49 include/linux/etherdevice.h struct net_device *alloc_etherdev_mqs(int sizeof_priv, unsigned int txqs, txqs 55 include/linux/etherdevice.h unsigned int txqs, txqs 4092 include/linux/netdevice.h unsigned int txqs, unsigned int rxqs); txqs 2212 include/rdma/ib_verbs.h unsigned int txqs; txqs 9471 net/core/dev.c unsigned int txqs, unsigned int rxqs) txqs 9479 net/core/dev.c if (txqs < 1) { txqs 9543 net/core/dev.c dev->num_tx_queues = txqs; txqs 9544 net/core/dev.c dev->real_num_tx_queues = txqs; txqs 411 net/ethernet/eth.c struct net_device *alloc_etherdev_mqs(int sizeof_priv, unsigned int txqs, txqs 415 net/ethernet/eth.c ether_setup, txqs, rxqs); txqs 425 net/ethernet/eth.c unsigned int txqs, unsigned int rxqs) txqs 434 net/ethernet/eth.c netdev = alloc_etherdev_mqs(sizeof_priv, txqs, rxqs); txqs 1733 net/mac80211/iface.c int txqs = 1; txqs 1767 net/mac80211/iface.c txqs = IEEE80211_NUM_ACS; txqs 1772 net/mac80211/iface.c if_setup, txqs, 1);