q_idx            1355 drivers/infiniband/hw/hfi1/hfi.h 	int (*process_vnic_dma_send)(struct hfi1_devdata *dd, u8 q_idx,
q_idx              90 drivers/infiniband/hw/hfi1/vnic.h 	u8 q_idx;
q_idx             156 drivers/infiniband/hw/hfi1/vnic.h 				u8 q_idx);
q_idx             165 drivers/infiniband/hw/hfi1/vnic.h int hfi1_vnic_send_dma(struct hfi1_devdata *dd, u8 q_idx,
q_idx             270 drivers/infiniband/hw/hfi1/vnic_main.c 					 u8 q_idx, struct sk_buff *skb, int err)
q_idx             273 drivers/infiniband/hw/hfi1/vnic_main.c 	struct opa_vnic_stats *stats = &vinfo->stats[q_idx];
q_idx             299 drivers/infiniband/hw/hfi1/vnic_main.c 					 u8 q_idx, struct sk_buff *skb, int err)
q_idx             302 drivers/infiniband/hw/hfi1/vnic_main.c 	struct opa_vnic_stats *stats = &vinfo->stats[q_idx];
q_idx             351 drivers/infiniband/hw/hfi1/vnic_main.c 				    u8 q_idx)
q_idx             353 drivers/infiniband/hw/hfi1/vnic_main.c 	netif_stop_subqueue(vinfo->netdev, q_idx);
q_idx             354 drivers/infiniband/hw/hfi1/vnic_main.c 	if (!hfi1_vnic_sdma_write_avail(vinfo, q_idx))
q_idx             357 drivers/infiniband/hw/hfi1/vnic_main.c 	netif_start_subqueue(vinfo->netdev, q_idx);
q_idx             364 drivers/infiniband/hw/hfi1/vnic_main.c 	u8 pad_len, q_idx = skb->queue_mapping;
q_idx             371 drivers/infiniband/hw/hfi1/vnic_main.c 	v_dbg("xmit: queue %d skb len %d\n", q_idx, skb->len);
q_idx             373 drivers/infiniband/hw/hfi1/vnic_main.c 		vinfo->stats[q_idx].tx_drop_state++;
q_idx             381 drivers/infiniband/hw/hfi1/vnic_main.c 		vinfo->stats[q_idx].tx_dlid_zero++;
q_idx             401 drivers/infiniband/hw/hfi1/vnic_main.c 	err = dd->process_vnic_dma_send(dd, q_idx, vinfo, skb, pbc, pad_len);
q_idx             404 drivers/infiniband/hw/hfi1/vnic_main.c 			vinfo->stats[q_idx].netstats.tx_fifo_errors++;
q_idx             406 drivers/infiniband/hw/hfi1/vnic_main.c 			vinfo->stats[q_idx].netstats.tx_carrier_errors++;
q_idx             412 drivers/infiniband/hw/hfi1/vnic_main.c 		hfi1_vnic_maybe_stop_tx(vinfo, q_idx);
q_idx             419 drivers/infiniband/hw/hfi1/vnic_main.c 	hfi1_vnic_update_tx_counters(vinfo, q_idx, skb, err);
q_idx             531 drivers/infiniband/hw/hfi1/vnic_main.c 	u8 q_idx;
q_idx             562 drivers/infiniband/hw/hfi1/vnic_main.c 	q_idx = packet->rcd->vnic_q_idx;
q_idx             563 drivers/infiniband/hw/hfi1/vnic_main.c 	rxq = &vinfo->rxq[q_idx];
q_idx             565 drivers/infiniband/hw/hfi1/vnic_main.c 		vinfo->stats[q_idx].rx_drop_state++;
q_idx             571 drivers/infiniband/hw/hfi1/vnic_main.c 		vinfo->stats[q_idx].netstats.rx_fifo_errors++;
q_idx             577 drivers/infiniband/hw/hfi1/vnic_main.c 		vinfo->stats[q_idx].netstats.rx_fifo_errors++;
q_idx             586 drivers/infiniband/hw/hfi1/vnic_main.c 		v_dbg("napi %d scheduling\n", q_idx);
q_idx             167 drivers/infiniband/hw/hfi1/vnic_sdma.c int hfi1_vnic_send_dma(struct hfi1_devdata *dd, u8 q_idx,
q_idx             171 drivers/infiniband/hw/hfi1/vnic_sdma.c 	struct hfi1_vnic_sdma *vnic_sdma = &vinfo->sdma[q_idx];
q_idx             265 drivers/infiniband/hw/hfi1/vnic_sdma.c 	if (__netif_subqueue_stopped(vinfo->netdev, vnic_sdma->q_idx))
q_idx             266 drivers/infiniband/hw/hfi1/vnic_sdma.c 		netif_wake_subqueue(vinfo->netdev, vnic_sdma->q_idx);
q_idx             270 drivers/infiniband/hw/hfi1/vnic_sdma.c 				       u8 q_idx)
q_idx             272 drivers/infiniband/hw/hfi1/vnic_sdma.c 	struct hfi1_vnic_sdma *vnic_sdma = &vinfo->sdma[q_idx];
q_idx             290 drivers/infiniband/hw/hfi1/vnic_sdma.c 		vnic_sdma->q_idx = i;
q_idx             129 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 			      u16 q_idx, u16 sb_idx)
q_idx             134 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	   q_idx,
q_idx             145 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 			    u16 q_idx, u16 sb_idx)
q_idx             152 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	   q_idx,
q_idx             438 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h 			    u16 q_idx, u16 sb_idx);
q_idx             443 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h 			    u16 q_idx, u16 sb_idx);
q_idx            2331 drivers/net/ethernet/cavium/liquidio/lio_main.c 	int q_idx = 0, iq_no = 0;
q_idx            2339 drivers/net/ethernet/cavium/liquidio/lio_main.c 	q_idx = skb_iq(oct, skb);
q_idx            2340 drivers/net/ethernet/cavium/liquidio/lio_main.c 	tag = q_idx;
q_idx            2341 drivers/net/ethernet/cavium/liquidio/lio_main.c 	iq_no = lio->linfo.txpciq[q_idx].s.q_no;
q_idx            2430 drivers/net/ethernet/cavium/liquidio/lio_main.c 		spin_lock(&lio->glist_lock[q_idx]);
q_idx            2432 drivers/net/ethernet/cavium/liquidio/lio_main.c 			lio_list_delete_head(&lio->glist[q_idx]);
q_idx            2433 drivers/net/ethernet/cavium/liquidio/lio_main.c 		spin_unlock(&lio->glist_lock[q_idx]);
q_idx            2536 drivers/net/ethernet/cavium/liquidio/lio_main.c 		netif_stop_subqueue(netdev, q_idx);
q_idx            1412 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 	int q_idx = 0, iq_no = 0;
q_idx            1424 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 	q_idx = skb_iq(lio->oct_dev, skb);
q_idx            1425 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 	tag = q_idx;
q_idx            1426 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 	iq_no = lio->linfo.txpciq[q_idx].s.q_no;
q_idx            1504 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 		spin_lock(&lio->glist_lock[q_idx]);
q_idx            1506 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 			lio_list_delete_head(&lio->glist[q_idx]);
q_idx            1507 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 		spin_unlock(&lio->glist_lock[q_idx]);
q_idx            1601 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 		netif_stop_subqueue(netdev, q_idx);
q_idx            1722 drivers/net/ethernet/cavium/thunder/nicvf_queues.c static u64 nicvf_int_type_to_mask(int int_type, int q_idx)
q_idx            1728 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 		reg_val = ((1ULL << q_idx) << NICVF_INTR_CQ_SHIFT);
q_idx            1731 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 		reg_val = ((1ULL << q_idx) << NICVF_INTR_SQ_SHIFT);
q_idx            1734 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 		reg_val = ((1ULL << q_idx) << NICVF_INTR_RBDR_SHIFT);
q_idx            1756 drivers/net/ethernet/cavium/thunder/nicvf_queues.c void nicvf_enable_intr(struct nicvf *nic, int int_type, int q_idx)
q_idx            1758 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 	u64 mask = nicvf_int_type_to_mask(int_type, q_idx);
q_idx            1770 drivers/net/ethernet/cavium/thunder/nicvf_queues.c void nicvf_disable_intr(struct nicvf *nic, int int_type, int q_idx)
q_idx            1772 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 	u64 mask = nicvf_int_type_to_mask(int_type, q_idx);
q_idx            1784 drivers/net/ethernet/cavium/thunder/nicvf_queues.c void nicvf_clear_intr(struct nicvf *nic, int int_type, int q_idx)
q_idx            1786 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 	u64 mask = nicvf_int_type_to_mask(int_type, q_idx);
q_idx            1798 drivers/net/ethernet/cavium/thunder/nicvf_queues.c int nicvf_is_intr_enabled(struct nicvf *nic, int int_type, int q_idx)
q_idx            1800 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 	u64 mask = nicvf_int_type_to_mask(int_type, q_idx);
q_idx             354 drivers/net/ethernet/cavium/thunder/nicvf_queues.h void nicvf_enable_intr(struct nicvf *nic, int int_type, int q_idx);
q_idx             355 drivers/net/ethernet/cavium/thunder/nicvf_queues.h void nicvf_disable_intr(struct nicvf *nic, int int_type, int q_idx);
q_idx             356 drivers/net/ethernet/cavium/thunder/nicvf_queues.h void nicvf_clear_intr(struct nicvf *nic, int int_type, int q_idx);
q_idx             357 drivers/net/ethernet/cavium/thunder/nicvf_queues.h int nicvf_is_intr_enabled(struct nicvf *nic, int int_type, int q_idx);
q_idx            2537 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c 	int i, j, cpu_idx = 0, q_idx = 0;
q_idx            2554 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c 				     (i << 16) | rspq[q_idx++]);
q_idx            2555 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c 			if (rspq[q_idx] == 0xffff)
q_idx            2556 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c 				q_idx = 0;
q_idx            1372 drivers/net/ethernet/emulex/benet/be_main.c 	u16 q_idx = skb_get_queue_mapping(skb);
q_idx            1373 drivers/net/ethernet/emulex/benet/be_main.c 	struct be_tx_obj *txo = &adapter->tx_obj[q_idx];
q_idx            1403 drivers/net/ethernet/emulex/benet/be_main.c 		netif_stop_subqueue(netdev, q_idx);
q_idx            1407 drivers/net/ethernet/emulex/benet/be_main.c 	if (flush || __netif_subqueue_stopped(netdev, q_idx))
q_idx            1885 drivers/net/ethernet/intel/fm10k/fm10k_main.c 		int q_idx = pc;
q_idx            1888 drivers/net/ethernet/intel/fm10k/fm10k_main.c 			interface->tx_ring[offset + i]->reg_idx = q_idx;
q_idx            1890 drivers/net/ethernet/intel/fm10k/fm10k_main.c 			interface->rx_ring[offset + i]->reg_idx = q_idx;
q_idx            1892 drivers/net/ethernet/intel/fm10k/fm10k_main.c 			q_idx += pc_stride;
q_idx            1175 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 	int q_idx;
q_idx            1177 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 	for (q_idx = 0; q_idx < interface->num_q_vectors; q_idx++) {
q_idx            1178 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 		q_vector = interface->q_vector[q_idx];
q_idx            1872 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 	int q_idx;
q_idx            1874 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 	for (q_idx = 0; q_idx < interface->num_q_vectors; q_idx++) {
q_idx            1875 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 		q_vector = interface->q_vector[q_idx];
q_idx             502 drivers/net/ethernet/intel/fm10k/fm10k_pf.c 	u16 vsi, queue, pc, q_idx;
q_idx             519 drivers/net/ethernet/intel/fm10k/fm10k_pf.c 	q_idx = dglort->queue_b;
q_idx             523 drivers/net/ethernet/intel/fm10k/fm10k_pf.c 		for (queue = 0; queue < queue_count; queue++, q_idx++) {
q_idx             524 drivers/net/ethernet/intel/fm10k/fm10k_pf.c 			if (q_idx >= FM10K_MAX_QUEUES)
q_idx             527 drivers/net/ethernet/intel/fm10k/fm10k_pf.c 			fm10k_write_reg(hw, FM10K_TX_SGLORT(q_idx), glort);
q_idx             528 drivers/net/ethernet/intel/fm10k/fm10k_pf.c 			fm10k_write_reg(hw, FM10K_RX_SGLORT(q_idx), glort);
q_idx             538 drivers/net/ethernet/intel/fm10k/fm10k_pf.c 		q_idx = pc + dglort->queue_b;
q_idx             540 drivers/net/ethernet/intel/fm10k/fm10k_pf.c 			if (q_idx >= FM10K_MAX_QUEUES)
q_idx             543 drivers/net/ethernet/intel/fm10k/fm10k_pf.c 			txqctl = fm10k_read_reg(hw, FM10K_TXQCTL(q_idx));
q_idx             546 drivers/net/ethernet/intel/fm10k/fm10k_pf.c 			fm10k_write_reg(hw, FM10K_TXQCTL(q_idx), txqctl);
q_idx             548 drivers/net/ethernet/intel/fm10k/fm10k_pf.c 			q_idx += pc_count;
q_idx            4811 drivers/net/ethernet/intel/i40e/i40e_main.c 	int q_idx;
q_idx            4816 drivers/net/ethernet/intel/i40e/i40e_main.c 	for (q_idx = 0; q_idx < vsi->num_q_vectors; q_idx++) {
q_idx            4817 drivers/net/ethernet/intel/i40e/i40e_main.c 		struct i40e_q_vector *q_vector = vsi->q_vectors[q_idx];
q_idx            4830 drivers/net/ethernet/intel/i40e/i40e_main.c 	int q_idx;
q_idx            4835 drivers/net/ethernet/intel/i40e/i40e_main.c 	for (q_idx = 0; q_idx < vsi->num_q_vectors; q_idx++) {
q_idx            4836 drivers/net/ethernet/intel/i40e/i40e_main.c 		struct i40e_q_vector *q_vector = vsi->q_vectors[q_idx];
q_idx             892 drivers/net/ethernet/intel/iavf/iavf_main.c 	int q_idx;
q_idx             896 drivers/net/ethernet/intel/iavf/iavf_main.c 	for (q_idx = 0; q_idx < q_vectors; q_idx++) {
q_idx             899 drivers/net/ethernet/intel/iavf/iavf_main.c 		q_vector = &adapter->q_vectors[q_idx];
q_idx             911 drivers/net/ethernet/intel/iavf/iavf_main.c 	int q_idx;
q_idx             915 drivers/net/ethernet/intel/iavf/iavf_main.c 	for (q_idx = 0; q_idx < q_vectors; q_idx++) {
q_idx             916 drivers/net/ethernet/intel/iavf/iavf_main.c 		q_vector = &adapter->q_vectors[q_idx];
q_idx            1336 drivers/net/ethernet/intel/iavf/iavf_main.c 	int q_idx = 0, num_q_vectors;
q_idx            1345 drivers/net/ethernet/intel/iavf/iavf_main.c 	for (q_idx = 0; q_idx < num_q_vectors; q_idx++) {
q_idx            1346 drivers/net/ethernet/intel/iavf/iavf_main.c 		q_vector = &adapter->q_vectors[q_idx];
q_idx            1349 drivers/net/ethernet/intel/iavf/iavf_main.c 		q_vector->v_idx = q_idx;
q_idx            1350 drivers/net/ethernet/intel/iavf/iavf_main.c 		q_vector->reg_idx = q_idx;
q_idx            1369 drivers/net/ethernet/intel/iavf/iavf_main.c 	int q_idx, num_q_vectors;
q_idx            1378 drivers/net/ethernet/intel/iavf/iavf_main.c 	for (q_idx = 0; q_idx < num_q_vectors; q_idx++) {
q_idx            1379 drivers/net/ethernet/intel/iavf/iavf_main.c 		struct iavf_q_vector *q_vector = &adapter->q_vectors[q_idx];
q_idx            1381 drivers/net/ethernet/intel/iavf/iavf_main.c 		if (q_idx < napi_vectors)
q_idx            1784 drivers/net/ethernet/intel/ice/ice_lib.c 	u16 q_idx = 0, i;
q_idx            1800 drivers/net/ethernet/intel/ice/ice_lib.c 			err = ice_vsi_cfg_txq(vsi, rings[q_idx], i + offset,
q_idx            1805 drivers/net/ethernet/intel/ice/ice_lib.c 			q_idx++;
q_idx            2250 drivers/net/ethernet/intel/ice/ice_lib.c 	u16 i, q_idx = 0;
q_idx            2265 drivers/net/ethernet/intel/ice/ice_lib.c 			if (!rings || !rings[q_idx])
q_idx            2268 drivers/net/ethernet/intel/ice/ice_lib.c 			ice_fill_txq_meta(vsi, rings[q_idx], &txq_meta);
q_idx            2271 drivers/net/ethernet/intel/ice/ice_lib.c 						      rings[q_idx], &txq_meta);
q_idx            2276 drivers/net/ethernet/intel/ice/ice_lib.c 			q_idx++;
q_idx            3520 drivers/net/ethernet/intel/ice/ice_main.c 	int q_idx;
q_idx            3525 drivers/net/ethernet/intel/ice/ice_main.c 	ice_for_each_q_vector(vsi, q_idx) {
q_idx            3526 drivers/net/ethernet/intel/ice/ice_main.c 		struct ice_q_vector *q_vector = vsi->q_vectors[q_idx];
q_idx            3887 drivers/net/ethernet/intel/ice/ice_main.c 	int q_idx;
q_idx            3892 drivers/net/ethernet/intel/ice/ice_main.c 	ice_for_each_q_vector(vsi, q_idx) {
q_idx            3893 drivers/net/ethernet/intel/ice/ice_main.c 		struct ice_q_vector *q_vector = vsi->q_vectors[q_idx];
q_idx            1046 drivers/net/ethernet/intel/igb/e1000_defines.h #define E1000_VLAPQF_QUEUE_SEL(_n, q_idx) (q_idx << ((_n) * 4))
q_idx             396 drivers/net/ethernet/intel/igc/igc_defines.h #define IGC_VLAPQF_QUEUE_SEL(_n, q_idx) ((q_idx) << ((_n) * 4))
q_idx            4998 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	int q_idx;
q_idx            5000 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	for (q_idx = 0; q_idx < adapter->num_q_vectors; q_idx++)
q_idx            5001 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		napi_enable(&adapter->q_vector[q_idx]->napi);
q_idx            5006 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	int q_idx;
q_idx            5008 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	for (q_idx = 0; q_idx < adapter->num_q_vectors; q_idx++)
q_idx            5009 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		napi_disable(&adapter->q_vector[q_idx]->napi);
q_idx            2141 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	int q_idx;
q_idx            2145 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	for (q_idx = 0; q_idx < q_vectors; q_idx++) {
q_idx            2146 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		q_vector = adapter->q_vector[q_idx];
q_idx            2153 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	int q_idx;
q_idx            2157 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	for (q_idx = 0; q_idx < q_vectors; q_idx++) {
q_idx            2158 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		q_vector = adapter->q_vector[q_idx];
q_idx             293 drivers/net/ethernet/qlogic/qed/qed_hw.h 		u8 q_idx;
q_idx            2895 drivers/net/ethernet/qlogic/qed/qed_sriov.c 	u16 i, q_idx;
q_idx            2927 drivers/net/ethernet/qlogic/qed/qed_sriov.c 		q_idx = p_rss_tlv->rss_ind_table[i];
q_idx            2928 drivers/net/ethernet/qlogic/qed/qed_sriov.c 		if (!qed_iov_validate_rxq(p_hwfn, vf, q_idx,
q_idx            2933 drivers/net/ethernet/qlogic/qed/qed_sriov.c 				   vf->relative_vf_id, q_idx);
q_idx            2938 drivers/net/ethernet/qlogic/qed/qed_sriov.c 		p_cid = qed_iov_get_vf_rx_queue_cid(&vf->vf_queues[q_idx]);
q_idx            1803 drivers/net/ethernet/ti/cpsw.c 	int ret, q_idx;
q_idx            1815 drivers/net/ethernet/ti/cpsw.c 	q_idx = skb_get_queue_mapping(skb);
q_idx            1816 drivers/net/ethernet/ti/cpsw.c 	if (q_idx >= cpsw->tx_ch_num)
q_idx            1817 drivers/net/ethernet/ti/cpsw.c 		q_idx = q_idx % cpsw->tx_ch_num;
q_idx            1819 drivers/net/ethernet/ti/cpsw.c 	txch = cpsw->txv[q_idx].ch;
q_idx            1820 drivers/net/ethernet/ti/cpsw.c 	txq = netdev_get_tx_queue(ndev, q_idx);
q_idx             127 drivers/net/hyperv/hyperv_net.h 	u16 q_idx;
q_idx             186 drivers/net/hyperv/hyperv_net.h int netvsc_alloc_recv_comp_ring(struct netvsc_device *net_device, u32 q_idx);
q_idx             281 drivers/net/hyperv/netvsc.c int netvsc_alloc_recv_comp_ring(struct netvsc_device *net_device, u32 q_idx)
q_idx             283 drivers/net/hyperv/netvsc.c 	struct netvsc_channel *nvchan = &net_device->chan_table[q_idx];
q_idx             679 drivers/net/hyperv/netvsc.c 	u16 q_idx = 0;
q_idx             691 drivers/net/hyperv/netvsc.c 		q_idx = packet->q_idx;
q_idx             693 drivers/net/hyperv/netvsc.c 		tx_stats = &net_device->chan_table[q_idx].tx_stats;
q_idx             704 drivers/net/hyperv/netvsc.c 		atomic_dec_return(&net_device->chan_table[q_idx].queue_sends);
q_idx             710 drivers/net/hyperv/netvsc.c 		struct netdev_queue *txq = netdev_get_tx_queue(ndev, q_idx);
q_idx             814 drivers/net/hyperv/netvsc.c 		&net_device->chan_table[packet->q_idx];
q_idx             818 drivers/net/hyperv/netvsc.c 	struct netdev_queue *txq = netdev_get_tx_queue(ndev, packet->q_idx);
q_idx             922 drivers/net/hyperv/netvsc.c 	nvchan = &net_device->chan_table[packet->q_idx];
q_idx             963 drivers/net/hyperv/netvsc.c 		!netif_xmit_stopped(netdev_get_tx_queue(ndev, packet->q_idx));
q_idx            1080 drivers/net/hyperv/netvsc.c 				 struct netvsc_device *nvdev, u16 q_idx,
q_idx            1083 drivers/net/hyperv/netvsc.c 	struct netvsc_channel *nvchan = &nvdev->chan_table[q_idx];
q_idx            1097 drivers/net/hyperv/netvsc.c 			   q_idx, tid);
q_idx            1119 drivers/net/hyperv/netvsc.c 	u16 q_idx = channel->offermsg.offer.sub_channel_index;
q_idx            1164 drivers/net/hyperv/netvsc.c 		trace_rndis_recv(ndev, q_idx, data);
q_idx            1174 drivers/net/hyperv/netvsc.c 	enq_receive_complete(ndev, net_device, q_idx,
q_idx             301 drivers/net/hyperv/netvsc_drv.c 	int q_idx;
q_idx             303 drivers/net/hyperv/netvsc_drv.c 	q_idx = ndc->tx_table[netvsc_get_hash(skb, ndc) &
q_idx             307 drivers/net/hyperv/netvsc_drv.c 	if (q_idx != old_idx &&
q_idx             309 drivers/net/hyperv/netvsc_drv.c 		sk_tx_queue_set(sk, q_idx);
q_idx             311 drivers/net/hyperv/netvsc_drv.c 	return q_idx;
q_idx             327 drivers/net/hyperv/netvsc_drv.c 	int q_idx = sk_tx_queue_get(skb->sk);
q_idx             329 drivers/net/hyperv/netvsc_drv.c 	if (q_idx < 0 || skb->ooo_okay || q_idx >= ndev->real_num_tx_queues) {
q_idx             334 drivers/net/hyperv/netvsc_drv.c 			q_idx = skb_get_rx_queue(skb);
q_idx             336 drivers/net/hyperv/netvsc_drv.c 			q_idx = netvsc_get_tx_queue(ndev, skb, q_idx);
q_idx             339 drivers/net/hyperv/netvsc_drv.c 	return q_idx;
q_idx             577 drivers/net/hyperv/netvsc_drv.c 	packet->q_idx = skb_get_queue_mapping(skb);
q_idx             826 drivers/net/hyperv/netvsc_drv.c 	u16 q_idx = channel->offermsg.offer.sub_channel_index;
q_idx             841 drivers/net/hyperv/netvsc_drv.c 	skb_record_rx_queue(skb, q_idx);
q_idx              98 drivers/net/hyperv/rndis_filter.c 	request->pkt.q_idx = 0;
q_idx             316 drivers/net/wireless/mediatek/mt76/mt7615/mac.c 	u8 fc_type, fc_stype, p_fmt, q_idx, omac_idx = 0, wmm_idx = 0;
q_idx             338 drivers/net/wireless/mediatek/mt76/mt7615/mac.c 		q_idx = wmm_idx * MT7615_MAX_WMM_SETS +
q_idx             342 drivers/net/wireless/mediatek/mt76/mt7615/mac.c 		q_idx = MT_LMAC_BCN0;
q_idx             345 drivers/net/wireless/mediatek/mt76/mt7615/mac.c 		q_idx = MT_LMAC_ALTX0;
q_idx             351 drivers/net/wireless/mediatek/mt76/mt7615/mac.c 	      FIELD_PREP(MT_TXD0_Q_IDX, q_idx);
q_idx              55 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c 	u8 seq, q_idx, pkt_fmt;
q_idx              69 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c 		q_idx = MT_TX_MCU_PORT_RX_Q0;
q_idx              72 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c 		q_idx = MT_TX_MCU_PORT_RX_FWDL;
q_idx              80 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c 	      FIELD_PREP(MT_TXD0_Q_IDX, q_idx);
q_idx              89 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c 	mcu_txd->pq_id = cpu_to_le16(MCU_PQ_ID(MT_TX_PORT_IDX_MCU, q_idx));
q_idx             534 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/trx.c u16 rtl92ee_get_available_desc(struct ieee80211_hw *hw, u8 q_idx)
q_idx             542 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/trx.c 				   get_desc_addr_fr_q_idx(q_idx));
q_idx             898 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/trx.c 	u8 q_idx = *val;
q_idx             909 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/trx.c 			struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[q_idx];
q_idx             912 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/trx.c 			if (q_idx == BEACON_QUEUE) {
q_idx             923 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/trx.c 				       get_desc_addr_fr_q_idx(q_idx),
q_idx            2331 drivers/net/wireless/realtek/rtlwifi/wifi.h 	u16 (*get_available_desc)(struct ieee80211_hw *hw, u8 q_idx);
q_idx            1678 net/sched/sch_api.c 	int ret = 0, q_idx = *q_idx_p;
q_idx            1686 net/sched/sch_api.c 	if (q_idx < s_q_idx) {
q_idx            1687 net/sched/sch_api.c 		q_idx++;
q_idx            1694 net/sched/sch_api.c 		q_idx++;
q_idx            1707 net/sched/sch_api.c 		if (q_idx < s_q_idx) {
q_idx            1708 net/sched/sch_api.c 			q_idx++;
q_idx            1716 net/sched/sch_api.c 		q_idx++;
q_idx            1720 net/sched/sch_api.c 	*q_idx_p = q_idx;
q_idx            1730 net/sched/sch_api.c 	int idx, q_idx;
q_idx            1738 net/sched/sch_api.c 	s_q_idx = q_idx = cb->args[1];
q_idx            1755 net/sched/sch_api.c 		q_idx = 0;
q_idx            1757 net/sched/sch_api.c 		if (tc_dump_qdisc_root(dev->qdisc, skb, cb, &q_idx, s_q_idx,
q_idx            1764 net/sched/sch_api.c 				       &q_idx, s_q_idx, false,
q_idx            1774 net/sched/sch_api.c 	cb->args[1] = q_idx;