q_idx 1355 drivers/infiniband/hw/hfi1/hfi.h int (*process_vnic_dma_send)(struct hfi1_devdata *dd, u8 q_idx, q_idx 90 drivers/infiniband/hw/hfi1/vnic.h u8 q_idx; q_idx 156 drivers/infiniband/hw/hfi1/vnic.h u8 q_idx); q_idx 165 drivers/infiniband/hw/hfi1/vnic.h int hfi1_vnic_send_dma(struct hfi1_devdata *dd, u8 q_idx, q_idx 270 drivers/infiniband/hw/hfi1/vnic_main.c u8 q_idx, struct sk_buff *skb, int err) q_idx 273 drivers/infiniband/hw/hfi1/vnic_main.c struct opa_vnic_stats *stats = &vinfo->stats[q_idx]; q_idx 299 drivers/infiniband/hw/hfi1/vnic_main.c u8 q_idx, struct sk_buff *skb, int err) q_idx 302 drivers/infiniband/hw/hfi1/vnic_main.c struct opa_vnic_stats *stats = &vinfo->stats[q_idx]; q_idx 351 drivers/infiniband/hw/hfi1/vnic_main.c u8 q_idx) q_idx 353 drivers/infiniband/hw/hfi1/vnic_main.c netif_stop_subqueue(vinfo->netdev, q_idx); q_idx 354 drivers/infiniband/hw/hfi1/vnic_main.c if (!hfi1_vnic_sdma_write_avail(vinfo, q_idx)) q_idx 357 drivers/infiniband/hw/hfi1/vnic_main.c netif_start_subqueue(vinfo->netdev, q_idx); q_idx 364 drivers/infiniband/hw/hfi1/vnic_main.c u8 pad_len, q_idx = skb->queue_mapping; q_idx 371 drivers/infiniband/hw/hfi1/vnic_main.c v_dbg("xmit: queue %d skb len %d\n", q_idx, skb->len); q_idx 373 drivers/infiniband/hw/hfi1/vnic_main.c vinfo->stats[q_idx].tx_drop_state++; q_idx 381 drivers/infiniband/hw/hfi1/vnic_main.c vinfo->stats[q_idx].tx_dlid_zero++; q_idx 401 drivers/infiniband/hw/hfi1/vnic_main.c err = dd->process_vnic_dma_send(dd, q_idx, vinfo, skb, pbc, pad_len); q_idx 404 drivers/infiniband/hw/hfi1/vnic_main.c vinfo->stats[q_idx].netstats.tx_fifo_errors++; q_idx 406 drivers/infiniband/hw/hfi1/vnic_main.c vinfo->stats[q_idx].netstats.tx_carrier_errors++; q_idx 412 drivers/infiniband/hw/hfi1/vnic_main.c hfi1_vnic_maybe_stop_tx(vinfo, q_idx); q_idx 419 drivers/infiniband/hw/hfi1/vnic_main.c hfi1_vnic_update_tx_counters(vinfo, q_idx, skb, err); q_idx 531 drivers/infiniband/hw/hfi1/vnic_main.c u8 q_idx; q_idx 562 drivers/infiniband/hw/hfi1/vnic_main.c q_idx = packet->rcd->vnic_q_idx; q_idx 563 drivers/infiniband/hw/hfi1/vnic_main.c rxq = &vinfo->rxq[q_idx]; q_idx 565 drivers/infiniband/hw/hfi1/vnic_main.c vinfo->stats[q_idx].rx_drop_state++; q_idx 571 drivers/infiniband/hw/hfi1/vnic_main.c vinfo->stats[q_idx].netstats.rx_fifo_errors++; q_idx 577 drivers/infiniband/hw/hfi1/vnic_main.c vinfo->stats[q_idx].netstats.rx_fifo_errors++; q_idx 586 drivers/infiniband/hw/hfi1/vnic_main.c v_dbg("napi %d scheduling\n", q_idx); q_idx 167 drivers/infiniband/hw/hfi1/vnic_sdma.c int hfi1_vnic_send_dma(struct hfi1_devdata *dd, u8 q_idx, q_idx 171 drivers/infiniband/hw/hfi1/vnic_sdma.c struct hfi1_vnic_sdma *vnic_sdma = &vinfo->sdma[q_idx]; q_idx 265 drivers/infiniband/hw/hfi1/vnic_sdma.c if (__netif_subqueue_stopped(vinfo->netdev, vnic_sdma->q_idx)) q_idx 266 drivers/infiniband/hw/hfi1/vnic_sdma.c netif_wake_subqueue(vinfo->netdev, vnic_sdma->q_idx); q_idx 270 drivers/infiniband/hw/hfi1/vnic_sdma.c u8 q_idx) q_idx 272 drivers/infiniband/hw/hfi1/vnic_sdma.c struct hfi1_vnic_sdma *vnic_sdma = &vinfo->sdma[q_idx]; q_idx 290 drivers/infiniband/hw/hfi1/vnic_sdma.c vnic_sdma->q_idx = i; q_idx 129 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c u16 q_idx, u16 sb_idx) q_idx 134 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c q_idx, q_idx 145 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c u16 q_idx, u16 sb_idx) q_idx 152 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c q_idx, q_idx 438 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h u16 q_idx, u16 sb_idx); q_idx 443 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h u16 q_idx, u16 sb_idx); q_idx 2331 drivers/net/ethernet/cavium/liquidio/lio_main.c int q_idx = 0, iq_no = 0; q_idx 2339 drivers/net/ethernet/cavium/liquidio/lio_main.c q_idx = skb_iq(oct, skb); q_idx 2340 drivers/net/ethernet/cavium/liquidio/lio_main.c tag = q_idx; q_idx 2341 drivers/net/ethernet/cavium/liquidio/lio_main.c iq_no = lio->linfo.txpciq[q_idx].s.q_no; q_idx 2430 drivers/net/ethernet/cavium/liquidio/lio_main.c spin_lock(&lio->glist_lock[q_idx]); q_idx 2432 drivers/net/ethernet/cavium/liquidio/lio_main.c lio_list_delete_head(&lio->glist[q_idx]); q_idx 2433 drivers/net/ethernet/cavium/liquidio/lio_main.c spin_unlock(&lio->glist_lock[q_idx]); q_idx 2536 drivers/net/ethernet/cavium/liquidio/lio_main.c netif_stop_subqueue(netdev, q_idx); q_idx 1412 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c int q_idx = 0, iq_no = 0; q_idx 1424 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c q_idx = skb_iq(lio->oct_dev, skb); q_idx 1425 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c tag = q_idx; q_idx 1426 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c iq_no = lio->linfo.txpciq[q_idx].s.q_no; q_idx 1504 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c spin_lock(&lio->glist_lock[q_idx]); q_idx 1506 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c lio_list_delete_head(&lio->glist[q_idx]); q_idx 1507 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c spin_unlock(&lio->glist_lock[q_idx]); q_idx 1601 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c netif_stop_subqueue(netdev, q_idx); q_idx 1722 drivers/net/ethernet/cavium/thunder/nicvf_queues.c static u64 nicvf_int_type_to_mask(int int_type, int q_idx) q_idx 1728 drivers/net/ethernet/cavium/thunder/nicvf_queues.c reg_val = ((1ULL << q_idx) << NICVF_INTR_CQ_SHIFT); q_idx 1731 drivers/net/ethernet/cavium/thunder/nicvf_queues.c reg_val = ((1ULL << q_idx) << NICVF_INTR_SQ_SHIFT); q_idx 1734 drivers/net/ethernet/cavium/thunder/nicvf_queues.c reg_val = ((1ULL << q_idx) << NICVF_INTR_RBDR_SHIFT); q_idx 1756 drivers/net/ethernet/cavium/thunder/nicvf_queues.c void nicvf_enable_intr(struct nicvf *nic, int int_type, int q_idx) q_idx 1758 drivers/net/ethernet/cavium/thunder/nicvf_queues.c u64 mask = nicvf_int_type_to_mask(int_type, q_idx); q_idx 1770 drivers/net/ethernet/cavium/thunder/nicvf_queues.c void nicvf_disable_intr(struct nicvf *nic, int int_type, int q_idx) q_idx 1772 drivers/net/ethernet/cavium/thunder/nicvf_queues.c u64 mask = nicvf_int_type_to_mask(int_type, q_idx); q_idx 1784 drivers/net/ethernet/cavium/thunder/nicvf_queues.c void nicvf_clear_intr(struct nicvf *nic, int int_type, int q_idx) q_idx 1786 drivers/net/ethernet/cavium/thunder/nicvf_queues.c u64 mask = nicvf_int_type_to_mask(int_type, q_idx); q_idx 1798 drivers/net/ethernet/cavium/thunder/nicvf_queues.c int nicvf_is_intr_enabled(struct nicvf *nic, int int_type, int q_idx) q_idx 1800 drivers/net/ethernet/cavium/thunder/nicvf_queues.c u64 mask = nicvf_int_type_to_mask(int_type, q_idx); q_idx 354 drivers/net/ethernet/cavium/thunder/nicvf_queues.h void nicvf_enable_intr(struct nicvf *nic, int int_type, int q_idx); q_idx 355 drivers/net/ethernet/cavium/thunder/nicvf_queues.h void nicvf_disable_intr(struct nicvf *nic, int int_type, int q_idx); q_idx 356 drivers/net/ethernet/cavium/thunder/nicvf_queues.h void nicvf_clear_intr(struct nicvf *nic, int int_type, int q_idx); q_idx 357 drivers/net/ethernet/cavium/thunder/nicvf_queues.h int nicvf_is_intr_enabled(struct nicvf *nic, int int_type, int q_idx); q_idx 2537 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c int i, j, cpu_idx = 0, q_idx = 0; q_idx 2554 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c (i << 16) | rspq[q_idx++]); q_idx 2555 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c if (rspq[q_idx] == 0xffff) q_idx 2556 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c q_idx = 0; q_idx 1372 drivers/net/ethernet/emulex/benet/be_main.c u16 q_idx = skb_get_queue_mapping(skb); q_idx 1373 drivers/net/ethernet/emulex/benet/be_main.c struct be_tx_obj *txo = &adapter->tx_obj[q_idx]; q_idx 1403 drivers/net/ethernet/emulex/benet/be_main.c netif_stop_subqueue(netdev, q_idx); q_idx 1407 drivers/net/ethernet/emulex/benet/be_main.c if (flush || __netif_subqueue_stopped(netdev, q_idx)) q_idx 1885 drivers/net/ethernet/intel/fm10k/fm10k_main.c int q_idx = pc; q_idx 1888 drivers/net/ethernet/intel/fm10k/fm10k_main.c interface->tx_ring[offset + i]->reg_idx = q_idx; q_idx 1890 drivers/net/ethernet/intel/fm10k/fm10k_main.c interface->rx_ring[offset + i]->reg_idx = q_idx; q_idx 1892 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_idx += pc_stride; q_idx 1175 drivers/net/ethernet/intel/fm10k/fm10k_pci.c int q_idx; q_idx 1177 drivers/net/ethernet/intel/fm10k/fm10k_pci.c for (q_idx = 0; q_idx < interface->num_q_vectors; q_idx++) { q_idx 1178 drivers/net/ethernet/intel/fm10k/fm10k_pci.c q_vector = interface->q_vector[q_idx]; q_idx 1872 drivers/net/ethernet/intel/fm10k/fm10k_pci.c int q_idx; q_idx 1874 drivers/net/ethernet/intel/fm10k/fm10k_pci.c for (q_idx = 0; q_idx < interface->num_q_vectors; q_idx++) { q_idx 1875 drivers/net/ethernet/intel/fm10k/fm10k_pci.c q_vector = interface->q_vector[q_idx]; q_idx 502 drivers/net/ethernet/intel/fm10k/fm10k_pf.c u16 vsi, queue, pc, q_idx; q_idx 519 drivers/net/ethernet/intel/fm10k/fm10k_pf.c q_idx = dglort->queue_b; q_idx 523 drivers/net/ethernet/intel/fm10k/fm10k_pf.c for (queue = 0; queue < queue_count; queue++, q_idx++) { q_idx 524 drivers/net/ethernet/intel/fm10k/fm10k_pf.c if (q_idx >= FM10K_MAX_QUEUES) q_idx 527 drivers/net/ethernet/intel/fm10k/fm10k_pf.c fm10k_write_reg(hw, FM10K_TX_SGLORT(q_idx), glort); q_idx 528 drivers/net/ethernet/intel/fm10k/fm10k_pf.c fm10k_write_reg(hw, FM10K_RX_SGLORT(q_idx), glort); q_idx 538 drivers/net/ethernet/intel/fm10k/fm10k_pf.c q_idx = pc + dglort->queue_b; q_idx 540 drivers/net/ethernet/intel/fm10k/fm10k_pf.c if (q_idx >= FM10K_MAX_QUEUES) q_idx 543 drivers/net/ethernet/intel/fm10k/fm10k_pf.c txqctl = fm10k_read_reg(hw, FM10K_TXQCTL(q_idx)); q_idx 546 drivers/net/ethernet/intel/fm10k/fm10k_pf.c fm10k_write_reg(hw, FM10K_TXQCTL(q_idx), txqctl); q_idx 548 drivers/net/ethernet/intel/fm10k/fm10k_pf.c q_idx += pc_count; q_idx 4811 drivers/net/ethernet/intel/i40e/i40e_main.c int q_idx; q_idx 4816 drivers/net/ethernet/intel/i40e/i40e_main.c for (q_idx = 0; q_idx < vsi->num_q_vectors; q_idx++) { q_idx 4817 drivers/net/ethernet/intel/i40e/i40e_main.c struct i40e_q_vector *q_vector = vsi->q_vectors[q_idx]; q_idx 4830 drivers/net/ethernet/intel/i40e/i40e_main.c int q_idx; q_idx 4835 drivers/net/ethernet/intel/i40e/i40e_main.c for (q_idx = 0; q_idx < vsi->num_q_vectors; q_idx++) { q_idx 4836 drivers/net/ethernet/intel/i40e/i40e_main.c struct i40e_q_vector *q_vector = vsi->q_vectors[q_idx]; q_idx 892 drivers/net/ethernet/intel/iavf/iavf_main.c int q_idx; q_idx 896 drivers/net/ethernet/intel/iavf/iavf_main.c for (q_idx = 0; q_idx < q_vectors; q_idx++) { q_idx 899 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector = &adapter->q_vectors[q_idx]; q_idx 911 drivers/net/ethernet/intel/iavf/iavf_main.c int q_idx; q_idx 915 drivers/net/ethernet/intel/iavf/iavf_main.c for (q_idx = 0; q_idx < q_vectors; q_idx++) { q_idx 916 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector = &adapter->q_vectors[q_idx]; q_idx 1336 drivers/net/ethernet/intel/iavf/iavf_main.c int q_idx = 0, num_q_vectors; q_idx 1345 drivers/net/ethernet/intel/iavf/iavf_main.c for (q_idx = 0; q_idx < num_q_vectors; q_idx++) { q_idx 1346 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector = &adapter->q_vectors[q_idx]; q_idx 1349 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->v_idx = q_idx; q_idx 1350 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->reg_idx = q_idx; q_idx 1369 drivers/net/ethernet/intel/iavf/iavf_main.c int q_idx, num_q_vectors; q_idx 1378 drivers/net/ethernet/intel/iavf/iavf_main.c for (q_idx = 0; q_idx < num_q_vectors; q_idx++) { q_idx 1379 drivers/net/ethernet/intel/iavf/iavf_main.c struct iavf_q_vector *q_vector = &adapter->q_vectors[q_idx]; q_idx 1381 drivers/net/ethernet/intel/iavf/iavf_main.c if (q_idx < napi_vectors) q_idx 1784 drivers/net/ethernet/intel/ice/ice_lib.c u16 q_idx = 0, i; q_idx 1800 drivers/net/ethernet/intel/ice/ice_lib.c err = ice_vsi_cfg_txq(vsi, rings[q_idx], i + offset, q_idx 1805 drivers/net/ethernet/intel/ice/ice_lib.c q_idx++; q_idx 2250 drivers/net/ethernet/intel/ice/ice_lib.c u16 i, q_idx = 0; q_idx 2265 drivers/net/ethernet/intel/ice/ice_lib.c if (!rings || !rings[q_idx]) q_idx 2268 drivers/net/ethernet/intel/ice/ice_lib.c ice_fill_txq_meta(vsi, rings[q_idx], &txq_meta); q_idx 2271 drivers/net/ethernet/intel/ice/ice_lib.c rings[q_idx], &txq_meta); q_idx 2276 drivers/net/ethernet/intel/ice/ice_lib.c q_idx++; q_idx 3520 drivers/net/ethernet/intel/ice/ice_main.c int q_idx; q_idx 3525 drivers/net/ethernet/intel/ice/ice_main.c ice_for_each_q_vector(vsi, q_idx) { q_idx 3526 drivers/net/ethernet/intel/ice/ice_main.c struct ice_q_vector *q_vector = vsi->q_vectors[q_idx]; q_idx 3887 drivers/net/ethernet/intel/ice/ice_main.c int q_idx; q_idx 3892 drivers/net/ethernet/intel/ice/ice_main.c ice_for_each_q_vector(vsi, q_idx) { q_idx 3893 drivers/net/ethernet/intel/ice/ice_main.c struct ice_q_vector *q_vector = vsi->q_vectors[q_idx]; q_idx 1046 drivers/net/ethernet/intel/igb/e1000_defines.h #define E1000_VLAPQF_QUEUE_SEL(_n, q_idx) (q_idx << ((_n) * 4)) q_idx 396 drivers/net/ethernet/intel/igc/igc_defines.h #define IGC_VLAPQF_QUEUE_SEL(_n, q_idx) ((q_idx) << ((_n) * 4)) q_idx 4998 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c int q_idx; q_idx 5000 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c for (q_idx = 0; q_idx < adapter->num_q_vectors; q_idx++) q_idx 5001 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c napi_enable(&adapter->q_vector[q_idx]->napi); q_idx 5006 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c int q_idx; q_idx 5008 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c for (q_idx = 0; q_idx < adapter->num_q_vectors; q_idx++) q_idx 5009 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c napi_disable(&adapter->q_vector[q_idx]->napi); q_idx 2141 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c int q_idx; q_idx 2145 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c for (q_idx = 0; q_idx < q_vectors; q_idx++) { q_idx 2146 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector = adapter->q_vector[q_idx]; q_idx 2153 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c int q_idx; q_idx 2157 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c for (q_idx = 0; q_idx < q_vectors; q_idx++) { q_idx 2158 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector = adapter->q_vector[q_idx]; q_idx 293 drivers/net/ethernet/qlogic/qed/qed_hw.h u8 q_idx; q_idx 2895 drivers/net/ethernet/qlogic/qed/qed_sriov.c u16 i, q_idx; q_idx 2927 drivers/net/ethernet/qlogic/qed/qed_sriov.c q_idx = p_rss_tlv->rss_ind_table[i]; q_idx 2928 drivers/net/ethernet/qlogic/qed/qed_sriov.c if (!qed_iov_validate_rxq(p_hwfn, vf, q_idx, q_idx 2933 drivers/net/ethernet/qlogic/qed/qed_sriov.c vf->relative_vf_id, q_idx); q_idx 2938 drivers/net/ethernet/qlogic/qed/qed_sriov.c p_cid = qed_iov_get_vf_rx_queue_cid(&vf->vf_queues[q_idx]); q_idx 1803 drivers/net/ethernet/ti/cpsw.c int ret, q_idx; q_idx 1815 drivers/net/ethernet/ti/cpsw.c q_idx = skb_get_queue_mapping(skb); q_idx 1816 drivers/net/ethernet/ti/cpsw.c if (q_idx >= cpsw->tx_ch_num) q_idx 1817 drivers/net/ethernet/ti/cpsw.c q_idx = q_idx % cpsw->tx_ch_num; q_idx 1819 drivers/net/ethernet/ti/cpsw.c txch = cpsw->txv[q_idx].ch; q_idx 1820 drivers/net/ethernet/ti/cpsw.c txq = netdev_get_tx_queue(ndev, q_idx); q_idx 127 drivers/net/hyperv/hyperv_net.h u16 q_idx; q_idx 186 drivers/net/hyperv/hyperv_net.h int netvsc_alloc_recv_comp_ring(struct netvsc_device *net_device, u32 q_idx); q_idx 281 drivers/net/hyperv/netvsc.c int netvsc_alloc_recv_comp_ring(struct netvsc_device *net_device, u32 q_idx) q_idx 283 drivers/net/hyperv/netvsc.c struct netvsc_channel *nvchan = &net_device->chan_table[q_idx]; q_idx 679 drivers/net/hyperv/netvsc.c u16 q_idx = 0; q_idx 691 drivers/net/hyperv/netvsc.c q_idx = packet->q_idx; q_idx 693 drivers/net/hyperv/netvsc.c tx_stats = &net_device->chan_table[q_idx].tx_stats; q_idx 704 drivers/net/hyperv/netvsc.c atomic_dec_return(&net_device->chan_table[q_idx].queue_sends); q_idx 710 drivers/net/hyperv/netvsc.c struct netdev_queue *txq = netdev_get_tx_queue(ndev, q_idx); q_idx 814 drivers/net/hyperv/netvsc.c &net_device->chan_table[packet->q_idx]; q_idx 818 drivers/net/hyperv/netvsc.c struct netdev_queue *txq = netdev_get_tx_queue(ndev, packet->q_idx); q_idx 922 drivers/net/hyperv/netvsc.c nvchan = &net_device->chan_table[packet->q_idx]; q_idx 963 drivers/net/hyperv/netvsc.c !netif_xmit_stopped(netdev_get_tx_queue(ndev, packet->q_idx)); q_idx 1080 drivers/net/hyperv/netvsc.c struct netvsc_device *nvdev, u16 q_idx, q_idx 1083 drivers/net/hyperv/netvsc.c struct netvsc_channel *nvchan = &nvdev->chan_table[q_idx]; q_idx 1097 drivers/net/hyperv/netvsc.c q_idx, tid); q_idx 1119 drivers/net/hyperv/netvsc.c u16 q_idx = channel->offermsg.offer.sub_channel_index; q_idx 1164 drivers/net/hyperv/netvsc.c trace_rndis_recv(ndev, q_idx, data); q_idx 1174 drivers/net/hyperv/netvsc.c enq_receive_complete(ndev, net_device, q_idx, q_idx 301 drivers/net/hyperv/netvsc_drv.c int q_idx; q_idx 303 drivers/net/hyperv/netvsc_drv.c q_idx = ndc->tx_table[netvsc_get_hash(skb, ndc) & q_idx 307 drivers/net/hyperv/netvsc_drv.c if (q_idx != old_idx && q_idx 309 drivers/net/hyperv/netvsc_drv.c sk_tx_queue_set(sk, q_idx); q_idx 311 drivers/net/hyperv/netvsc_drv.c return q_idx; q_idx 327 drivers/net/hyperv/netvsc_drv.c int q_idx = sk_tx_queue_get(skb->sk); q_idx 329 drivers/net/hyperv/netvsc_drv.c if (q_idx < 0 || skb->ooo_okay || q_idx >= ndev->real_num_tx_queues) { q_idx 334 drivers/net/hyperv/netvsc_drv.c q_idx = skb_get_rx_queue(skb); q_idx 336 drivers/net/hyperv/netvsc_drv.c q_idx = netvsc_get_tx_queue(ndev, skb, q_idx); q_idx 339 drivers/net/hyperv/netvsc_drv.c return q_idx; q_idx 577 drivers/net/hyperv/netvsc_drv.c packet->q_idx = skb_get_queue_mapping(skb); q_idx 826 drivers/net/hyperv/netvsc_drv.c u16 q_idx = channel->offermsg.offer.sub_channel_index; q_idx 841 drivers/net/hyperv/netvsc_drv.c skb_record_rx_queue(skb, q_idx); q_idx 98 drivers/net/hyperv/rndis_filter.c request->pkt.q_idx = 0; q_idx 316 drivers/net/wireless/mediatek/mt76/mt7615/mac.c u8 fc_type, fc_stype, p_fmt, q_idx, omac_idx = 0, wmm_idx = 0; q_idx 338 drivers/net/wireless/mediatek/mt76/mt7615/mac.c q_idx = wmm_idx * MT7615_MAX_WMM_SETS + q_idx 342 drivers/net/wireless/mediatek/mt76/mt7615/mac.c q_idx = MT_LMAC_BCN0; q_idx 345 drivers/net/wireless/mediatek/mt76/mt7615/mac.c q_idx = MT_LMAC_ALTX0; q_idx 351 drivers/net/wireless/mediatek/mt76/mt7615/mac.c FIELD_PREP(MT_TXD0_Q_IDX, q_idx); q_idx 55 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c u8 seq, q_idx, pkt_fmt; q_idx 69 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c q_idx = MT_TX_MCU_PORT_RX_Q0; q_idx 72 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c q_idx = MT_TX_MCU_PORT_RX_FWDL; q_idx 80 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c FIELD_PREP(MT_TXD0_Q_IDX, q_idx); q_idx 89 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c mcu_txd->pq_id = cpu_to_le16(MCU_PQ_ID(MT_TX_PORT_IDX_MCU, q_idx)); q_idx 534 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/trx.c u16 rtl92ee_get_available_desc(struct ieee80211_hw *hw, u8 q_idx) q_idx 542 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/trx.c get_desc_addr_fr_q_idx(q_idx)); q_idx 898 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/trx.c u8 q_idx = *val; q_idx 909 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/trx.c struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[q_idx]; q_idx 912 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/trx.c if (q_idx == BEACON_QUEUE) { q_idx 923 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/trx.c get_desc_addr_fr_q_idx(q_idx), q_idx 2331 drivers/net/wireless/realtek/rtlwifi/wifi.h u16 (*get_available_desc)(struct ieee80211_hw *hw, u8 q_idx); q_idx 1678 net/sched/sch_api.c int ret = 0, q_idx = *q_idx_p; q_idx 1686 net/sched/sch_api.c if (q_idx < s_q_idx) { q_idx 1687 net/sched/sch_api.c q_idx++; q_idx 1694 net/sched/sch_api.c q_idx++; q_idx 1707 net/sched/sch_api.c if (q_idx < s_q_idx) { q_idx 1708 net/sched/sch_api.c q_idx++; q_idx 1716 net/sched/sch_api.c q_idx++; q_idx 1720 net/sched/sch_api.c *q_idx_p = q_idx; q_idx 1730 net/sched/sch_api.c int idx, q_idx; q_idx 1738 net/sched/sch_api.c s_q_idx = q_idx = cb->args[1]; q_idx 1755 net/sched/sch_api.c q_idx = 0; q_idx 1757 net/sched/sch_api.c if (tc_dump_qdisc_root(dev->qdisc, skb, cb, &q_idx, s_q_idx, q_idx 1764 net/sched/sch_api.c &q_idx, s_q_idx, false, q_idx 1774 net/sched/sch_api.c cb->args[1] = q_idx;