queues 179 arch/powerpc/kvm/book3s_xive.c struct xive_q *q = &xc->queues[prio]; queues 249 arch/powerpc/kvm/book3s_xive.c struct xive_q *q = &xc->queues[prio]; queues 329 arch/powerpc/kvm/book3s_xive.c q = &xc->queues[prio]; queues 344 arch/powerpc/kvm/book3s_xive.c q = &xc->queues[prio]; queues 1191 arch/powerpc/kvm/book3s_xive.c struct xive_q *q = &xc->queues[i]; queues 1300 arch/powerpc/kvm/book3s_xive.c struct xive_q *q = &xc->queues[i]; queues 1459 arch/powerpc/kvm/book3s_xive.c if (xc->queues[j].qpage) queues 1460 arch/powerpc/kvm/book3s_xive.c xive_pre_save_queue(xive, &xc->queues[j]); queues 2050 arch/powerpc/kvm/book3s_xive.c struct xive_q *q = &xc->queues[i]; queues 173 arch/powerpc/kvm/book3s_xive.h struct xive_q queues[KVMPPC_XIVE_Q_COUNT]; queues 44 arch/powerpc/kvm/book3s_xive_native.c struct xive_q *q = &xc->queues[prio]; queues 592 arch/powerpc/kvm/book3s_xive_native.c q = &xc->queues[priority]; queues 737 arch/powerpc/kvm/book3s_xive_native.c q = &xc->queues[priority]; queues 893 arch/powerpc/kvm/book3s_xive_native.c struct xive_q *q = &xc->queues[prio]; queues 140 arch/powerpc/kvm/book3s_xive_template.c q = &xc->queues[prio]; queues 380 arch/powerpc/kvm/book3s_xive_template.c struct xive_q *q = &xc->queues[prio]; queues 86 drivers/block/null_blk.h struct nullb_queue *queues; queues 1281 drivers/block/null_blk_main.c return &nullb->queues[index]; queues 1380 drivers/block/null_blk_main.c cleanup_queue(&nullb->queues[i]); queues 1382 drivers/block/null_blk_main.c kfree(nullb->queues); queues 1464 drivers/block/null_blk_main.c nq = &nullb->queues[i]; queues 1499 drivers/block/null_blk_main.c nullb->queues = kcalloc(nullb->dev->submit_queues, queues 1502 drivers/block/null_blk_main.c if (!nullb->queues) queues 1516 drivers/block/null_blk_main.c nq = &nullb->queues[i]; queues 590 drivers/crypto/ccp/ccp-dev-v5.c int queues = 0; queues 601 drivers/crypto/ccp/ccp-dev-v5.c queues = bitmap_weight(cmd_q->lsbmask, MAX_LSB_CNT); queues 603 drivers/crypto/ccp/ccp-dev-v5.c cmd_q->id, queues); queues 605 drivers/crypto/ccp/ccp-dev-v5.c return queues ? 0 : -EINVAL; queues 639 drivers/crypto/hisilicon/sec/sec_drv.c if (!info->queues[i].in_use) { queues 640 drivers/crypto/hisilicon/sec/sec_drv.c info->queues[i].in_use = true; queues 644 drivers/crypto/hisilicon/sec/sec_drv.c return &info->queues[i]; queues 1235 drivers/crypto/hisilicon/sec/sec_drv.c ret = sec_queue_config(info, &info->queues[i], i); queues 1239 drivers/crypto/hisilicon/sec/sec_drv.c ret = sec_queue_irq_init(&info->queues[i]); queues 1241 drivers/crypto/hisilicon/sec/sec_drv.c sec_queue_unconfig(info, &info->queues[i]); queues 1265 drivers/crypto/hisilicon/sec/sec_drv.c sec_queue_irq_uninit(&info->queues[j]); queues 1266 drivers/crypto/hisilicon/sec/sec_drv.c sec_queue_unconfig(info, &info->queues[j]); queues 1284 drivers/crypto/hisilicon/sec/sec_drv.c sec_queue_irq_uninit(&info->queues[i]); queues 1285 drivers/crypto/hisilicon/sec/sec_drv.c sec_queue_unconfig(info, &info->queues[i]); queues 412 drivers/crypto/hisilicon/sec/sec_drv.h struct sec_queue queues[SEC_Q_NUM]; queues 1070 drivers/dma/fsl-qdma.c u32 len, chans, queues; queues 1116 drivers/dma/fsl-qdma.c ret = of_property_read_u32(np, "fsl,dma-queues", &queues); queues 1124 drivers/dma/fsl-qdma.c fsl_qdma->n_queues = queues; queues 950 drivers/dma/ti/cppi41.c const struct chan_queues *queues; queues 965 drivers/dma/ti/cppi41.c queues = cdd->queues_tx; queues 967 drivers/dma/ti/cppi41.c queues = cdd->queues_rx; queues 974 drivers/dma/ti/cppi41.c cchan->q_num = queues[cchan->port_num].submit; queues 975 drivers/dma/ti/cppi41.c cchan->q_comp_num = queues[cchan->port_num].complete; queues 785 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c list_add(&n->list, &dqm->queues); queues 817 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c list_for_each_entry_safe(cur, next, &dqm->queues, list) { queues 868 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c INIT_LIST_HEAD(&dqm->queues); queues 1017 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c INIT_LIST_HEAD(&dqm->queues); queues 1270 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c retval = pm_send_runlist(&dqm->packets, &dqm->queues); queues 1515 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c list_for_each_entry_safe(cur, next_dpn, &dqm->queues, list) { queues 1610 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c list_for_each_entry_safe(cur, next_dpn, &dqm->queues, list) { queues 177 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.h struct list_head queues; queues 120 drivers/gpu/drm/amd/amdkfd/kfd_packet_manager.c struct list_head *queues, queues 147 drivers/gpu/drm/amd/amdkfd/kfd_packet_manager.c list_for_each_entry(cur, queues, list) { queues 539 drivers/gpu/drm/amd/amdkfd/kfd_priv.h struct list_head queues; queues 36 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c list_for_each_entry(pqn, &pqm->queues, process_queue_list) { queues 137 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c INIT_LIST_HEAD(&pqm->queues); queues 152 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c list_for_each_entry_safe(pqn, next, &pqm->queues, process_queue_list) { queues 317 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c list_add(&pqn->process_queue_list, &pqm->queues); queues 509 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c list_for_each_entry(pqn, &pqm->queues, process_queue_list) { queues 76 drivers/gpu/drm/msm/adreno/a6xx_gmu.h struct a6xx_hfi_queue queues[2]; queues 85 drivers/gpu/drm/msm/adreno/a6xx_hfi.c struct a6xx_hfi_queue *queue = &gmu->queues[HFI_RESPONSE_QUEUE]; queues 154 drivers/gpu/drm/msm/adreno/a6xx_hfi.c struct a6xx_hfi_queue *queue = &gmu->queues[HFI_COMMAND_QUEUE]; queues 313 drivers/gpu/drm/msm/adreno/a6xx_hfi.c for (i = 0; i < ARRAY_SIZE(gmu->queues); i++) { queues 314 drivers/gpu/drm/msm/adreno/a6xx_hfi.c struct a6xx_hfi_queue *queue = &gmu->queues[i]; queues 364 drivers/gpu/drm/msm/adreno/a6xx_hfi.c table_size += (ARRAY_SIZE(gmu->queues) * queues 372 drivers/gpu/drm/msm/adreno/a6xx_hfi.c table->num_queues = ARRAY_SIZE(gmu->queues); queues 373 drivers/gpu/drm/msm/adreno/a6xx_hfi.c table->active_queues = ARRAY_SIZE(gmu->queues); queues 377 drivers/gpu/drm/msm/adreno/a6xx_hfi.c a6xx_hfi_queue_init(&gmu->queues[0], &headers[0], hfi->virt + offset, queues 382 drivers/gpu/drm/msm/adreno/a6xx_hfi.c a6xx_hfi_queue_init(&gmu->queues[1], &headers[1], hfi->virt + offset, queues 127 drivers/media/platform/qcom/venus/hfi_venus.c struct iface_queue queues[IFACEQ_NUM]; queues 389 drivers/media/platform/qcom/venus/hfi_venus.c queue = &hdev->queues[IFACEQ_CMD_IDX]; queues 613 drivers/media/platform/qcom/venus/hfi_venus.c queue = &hdev->queues[IFACEQ_MSG_IDX]; queues 647 drivers/media/platform/qcom/venus/hfi_venus.c queue = &hdev->queues[IFACEQ_DBG_IDX]; queues 696 drivers/media/platform/qcom/venus/hfi_venus.c memset(hdev->queues, 0, sizeof(hdev->queues)); queues 721 drivers/media/platform/qcom/venus/hfi_venus.c queue = &hdev->queues[i]; queues 753 drivers/media/platform/qcom/venus/hfi_venus.c queue = &hdev->queues[IFACEQ_DBG_IDX]; queues 852 drivers/media/platform/qcom/venus/hfi_venus.c qhdr = hdev->queues[index].qhdr; queues 1170 drivers/net/ethernet/cadence/macb.h struct macb_queue queues[MACB_MAX_QUEUES]; queues 737 drivers/net/ethernet/cadence/macb_main.c (unsigned int)(queue - bp->queues), queues 841 drivers/net/ethernet/cadence/macb_main.c u16 queue_index = queue - bp->queues; queues 1306 drivers/net/ethernet/cadence/macb_main.c struct macb_queue *queue = bp->queues; queues 1310 drivers/net/ethernet/cadence/macb_main.c for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) { queues 1325 drivers/net/ethernet/cadence/macb_main.c for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) { queues 1392 drivers/net/ethernet/cadence/macb_main.c (unsigned int)(queue - bp->queues), queues 1487 drivers/net/ethernet/cadence/macb_main.c for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) queues 1780 drivers/net/ethernet/cadence/macb_main.c struct macb_queue *queue = &bp->queues[queue_index]; queues 1902 drivers/net/ethernet/cadence/macb_main.c for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) { queues 1928 drivers/net/ethernet/cadence/macb_main.c struct macb_queue *queue = &bp->queues[0]; queues 1946 drivers/net/ethernet/cadence/macb_main.c for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) { queues 1970 drivers/net/ethernet/cadence/macb_main.c for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) { queues 1985 drivers/net/ethernet/cadence/macb_main.c struct macb_queue *queue = &bp->queues[0]; queues 2006 drivers/net/ethernet/cadence/macb_main.c for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) { queues 2049 drivers/net/ethernet/cadence/macb_main.c for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) { queues 2072 drivers/net/ethernet/cadence/macb_main.c macb_init_rx_ring(&bp->queues[0]); queues 2075 drivers/net/ethernet/cadence/macb_main.c desc = macb_tx_desc(&bp->queues[0], i); queues 2079 drivers/net/ethernet/cadence/macb_main.c bp->queues[0].tx_head = 0; queues 2080 drivers/net/ethernet/cadence/macb_main.c bp->queues[0].tx_tail = 0; queues 2105 drivers/net/ethernet/cadence/macb_main.c for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) { queues 2192 drivers/net/ethernet/cadence/macb_main.c for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) { queues 2267 drivers/net/ethernet/cadence/macb_main.c for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) { queues 2442 drivers/net/ethernet/cadence/macb_main.c for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) queues 2473 drivers/net/ethernet/cadence/macb_main.c for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) queues 2528 drivers/net/ethernet/cadence/macb_main.c for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) queues 2608 drivers/net/ethernet/cadence/macb_main.c for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) { queues 2681 drivers/net/ethernet/cadence/macb_main.c tail = macb_tx_ring_wrap(bp, bp->queues[0].tx_tail); queues 2682 drivers/net/ethernet/cadence/macb_main.c head = macb_tx_ring_wrap(bp, bp->queues[0].tx_head); queues 2695 drivers/net/ethernet/cadence/macb_main.c regs_buff[10] = macb_tx_dma(&bp->queues[0], tail); queues 2696 drivers/net/ethernet/cadence/macb_main.c regs_buff[11] = macb_tx_dma(&bp->queues[0], head); queues 3493 drivers/net/ethernet/cadence/macb_main.c queue = &bp->queues[q]; queues 3640 drivers/net/ethernet/cadence/macb_main.c struct macb_queue *q = &lp->queues[0]; queues 3734 drivers/net/ethernet/cadence/macb_main.c struct macb_queue *q = &lp->queues[0]; queues 3806 drivers/net/ethernet/cadence/macb_main.c struct macb_queue *q = &lp->queues[0]; queues 3948 drivers/net/ethernet/cadence/macb_main.c bp->queues[0].bp = bp; queues 4422 drivers/net/ethernet/cadence/macb_main.c struct macb_queue *queue = bp->queues; queues 4433 drivers/net/ethernet/cadence/macb_main.c enable_irq_wake(bp->queues[0].irq); queues 4437 drivers/net/ethernet/cadence/macb_main.c for (q = 0, queue = bp->queues; q < bp->num_queues; queues 4465 drivers/net/ethernet/cadence/macb_main.c struct macb_queue *queue = bp->queues; queues 4476 drivers/net/ethernet/cadence/macb_main.c disable_irq_wake(bp->queues[0].irq); queues 4486 drivers/net/ethernet/cadence/macb_main.c for (q = 0, queue = bp->queues; q < bp->num_queues; queues 373 drivers/net/ethernet/cadence/macb_ptp.c for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) { queues 1437 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h int cxgb4_write_rss(const struct port_info *pi, const u16 *queues); queues 834 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c int cxgb4_write_rss(const struct port_info *pi, const u16 *queues) queues 847 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c for (i = 0; i < pi->rss_size; i++, queues++) queues 848 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c rss[i] = rxq[*queues].rspq.abs_id; queues 2992 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c static inline void ixgbe_irq_enable(struct ixgbe_adapter *adapter, bool queues, queues 3045 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (queues) queues 6772 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c int err, queues; queues 6797 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c queues = adapter->num_tx_queues; queues 6798 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c err = netif_set_real_num_tx_queues(netdev, queues); queues 6802 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c queues = adapter->num_rx_queues; queues 6803 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c err = netif_set_real_num_rx_queues(netdev, queues); queues 106 drivers/net/ethernet/mellanox/mlxsw/pci.c struct mlxsw_pci_queue_type_group queues[MLXSW_PCI_QUEUE_TYPE_COUNT]; queues 180 drivers/net/ethernet/mellanox/mlxsw/pci.c return &mlxsw_pci->queues[q_type]; queues 206 drivers/net/ethernet/mellanox/mlxsw/pci.c return &mlxsw_pci->queues[q_type].q[q_num]; queues 150 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c struct dentry *queues, *tx, *rx, *xdp; queues 164 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c queues = debugfs_create_dir("queue", nn->debugfs_dir); queues 166 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c rx = debugfs_create_dir("rx", queues); queues 167 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c tx = debugfs_create_dir("tx", queues); queues 168 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c xdp = debugfs_create_dir("xdp", queues); queues 69 drivers/net/ethernet/qlogic/qed/qed_l2.c u32 queues; queues 91 drivers/net/ethernet/qlogic/qed/qed_l2.c p_l2_info->queues = RESC_NUM(p_hwfn, QED_L2_QUEUE); queues 98 drivers/net/ethernet/qlogic/qed/qed_l2.c p_l2_info->queues = max_t(u8, rx, tx); queues 101 drivers/net/ethernet/qlogic/qed/qed_l2.c pp_qids = kcalloc(p_l2_info->queues, sizeof(unsigned long *), queues 107 drivers/net/ethernet/qlogic/qed/qed_l2.c for (i = 0; i < p_l2_info->queues; i++) { queues 138 drivers/net/ethernet/qlogic/qed/qed_l2.c for (i = 0; i < p_hwfn->p_l2_info->queues; i++) { queues 161 drivers/net/ethernet/qlogic/qed/qed_l2.c if (queue_id >= p_l2_info->queues) { queues 164 drivers/net/ethernet/qlogic/qed/qed_l2.c queue_id, p_l2_info->queues); queues 2269 drivers/net/ethernet/qlogic/qed/qed_l2.c u8 queues, cids; queues 2272 drivers/net/ethernet/qlogic/qed/qed_l2.c qed_vf_get_num_rxqs(p_hwfn, &queues); queues 2273 drivers/net/ethernet/qlogic/qed/qed_l2.c info->num_queues += queues; queues 2255 drivers/net/ethernet/sfc/ef10.c u32 queues; queues 2259 drivers/net/ethernet/sfc/ef10.c queues = EFX_DWORD_FIELD(reg, ERF_DZ_ISR_REG); queues 2261 drivers/net/ethernet/sfc/ef10.c if (queues == 0) queues 2266 drivers/net/ethernet/sfc/ef10.c if (queues & (1U << efx->irq_level)) queues 2270 drivers/net/ethernet/sfc/ef10.c if (queues & 1) queues 2272 drivers/net/ethernet/sfc/ef10.c queues >>= 1; queues 436 drivers/net/ethernet/sfc/falcon/falcon.c int queues; queues 464 drivers/net/ethernet/sfc/falcon/falcon.c queues = EF4_OWORD_FIELD(*int_ker, FSF_AZ_NET_IVEC_INT_Q); queues 469 drivers/net/ethernet/sfc/falcon/falcon.c if (queues & 1) queues 471 drivers/net/ethernet/sfc/falcon/falcon.c if (queues & 2) queues 1509 drivers/net/ethernet/sfc/falcon/farch.c u32 queues; queues 1514 drivers/net/ethernet/sfc/falcon/farch.c queues = EF4_EXTRACT_DWORD(reg, 0, 31); queues 1527 drivers/net/ethernet/sfc/falcon/farch.c if (queues & (1U << efx->irq_level) && soft_enabled) { queues 1534 drivers/net/ethernet/sfc/falcon/farch.c if (queues != 0) { queues 1540 drivers/net/ethernet/sfc/falcon/farch.c if (queues & 1) queues 1542 drivers/net/ethernet/sfc/falcon/farch.c queues >>= 1; queues 1518 drivers/net/ethernet/sfc/farch.c u32 queues; queues 1523 drivers/net/ethernet/sfc/farch.c queues = EFX_EXTRACT_DWORD(reg, 0, 31); queues 1536 drivers/net/ethernet/sfc/farch.c if (queues & (1U << efx->irq_level) && soft_enabled) { queues 1543 drivers/net/ethernet/sfc/farch.c if (queues != 0) { queues 1549 drivers/net/ethernet/sfc/farch.c if (queues & 1) queues 1551 drivers/net/ethernet/sfc/farch.c queues >>= 1; queues 2768 drivers/net/tun.c int queues = ifr->ifr_flags & IFF_MULTI_QUEUE ? queues 2793 drivers/net/tun.c NET_NAME_UNKNOWN, tun_setup, queues, queues 2794 drivers/net/tun.c queues); queues 1887 drivers/net/wireless/admtek/adm8211.c dev->queues = 1; /* ADM8211C supports more, maybe ADM8211B too */ queues 1091 drivers/net/wireless/ath/ar5523/ar5523.c u32 queues, bool drop) queues 1686 drivers/net/wireless/ath/ar5523/ar5523.c hw->queues = 1; queues 7071 drivers/net/wireless/ath/ath10k/mac.c u32 queues, bool drop) queues 8895 drivers/net/wireless/ath/ath10k/mac.c ar->hw->queues = IEEE80211_MAX_QUEUES; queues 3060 drivers/net/wireless/ath/ath5k/base.c hw->queues = 4; queues 3069 drivers/net/wireless/ath/ath5k/base.c hw->queues = 1; queues 754 drivers/net/wireless/ath/ath9k/ath9k.h void __ath9k_flush(struct ieee80211_hw *hw, u32 queues, bool drop, queues 1400 drivers/net/wireless/ath/ath9k/channel.c ieee80211_stop_queue(sc->hw, sc->hw->queues - 2); queues 1419 drivers/net/wireless/ath/ath9k/channel.c ieee80211_wake_queue(sc->hw, sc->hw->queues - 2); queues 755 drivers/net/wireless/ath/ath9k/htc_drv_init.c hw->queues = 4; queues 899 drivers/net/wireless/ath/ath9k/init.c hw->queues = ATH9K_NUM_TX_QUEUES; queues 900 drivers/net/wireless/ath/ath9k/init.c hw->offchannel_tx_hw_queue = hw->queues - 1; queues 979 drivers/net/wireless/ath/ath9k/init.c hw->queues = 4; queues 1240 drivers/net/wireless/ath/ath9k/main.c vif->cab_queue = hw->queues - 2; queues 2058 drivers/net/wireless/ath/ath9k/main.c u32 queues, bool drop) queues 2077 drivers/net/wireless/ath/ath9k/main.c __ath9k_flush(hw, queues, drop, true, true); queues 2082 drivers/net/wireless/ath/ath9k/main.c __ath9k_flush(hw, queues, drop, true, false); queues 2086 drivers/net/wireless/ath/ath9k/main.c void __ath9k_flush(struct ieee80211_hw *hw, u32 queues, bool drop, queues 403 drivers/net/wireless/ath/carl9170/debug.c for (i = 0; i < ar->hw->queues; i++) { queues 229 drivers/net/wireless/ath/carl9170/main.c for (i = 0; i < ar->hw->queues; i++) { queues 286 drivers/net/wireless/ath/carl9170/main.c for (i = 0; i < ar->hw->queues; i++) { queues 307 drivers/net/wireless/ath/carl9170/main.c for (i = 0; i < ar->hw->queues; i++) queues 368 drivers/net/wireless/ath/carl9170/main.c for (i = 0; i < ar->hw->queues; i++) { queues 1699 drivers/net/wireless/ath/carl9170/main.c u32 queues, bool drop) queues 1809 drivers/net/wireless/ath/carl9170/main.c hw->queues = __AR9170_NUM_TXQ; queues 1820 drivers/net/wireless/ath/carl9170/main.c for (i = 0; i < ar->hw->queues; i++) { queues 96 drivers/net/wireless/ath/carl9170/tx.c for (i = 0; i < ar->hw->queues; i++) { queues 168 drivers/net/wireless/ath/carl9170/tx.c for (i = 0; i < ar->hw->queues; i++) { queues 566 drivers/net/wireless/ath/carl9170/tx.c for (i = 0; i < ar->hw->queues; i++) { queues 1352 drivers/net/wireless/ath/carl9170/tx.c for (i = 0; i < ar->hw->queues; i++) { queues 1200 drivers/net/wireless/ath/wcn36xx/main.c wcn->hw->queues = 4; queues 2587 drivers/net/wireless/broadcom/b43/main.c wl->hw->queues = B43_QOS_QUEUE_NUM; queues 2589 drivers/net/wireless/broadcom/b43/main.c wl->hw->queues = 1; queues 2695 drivers/net/wireless/broadcom/b43/main.c dev->qos_enabled = dev->wl->hw->queues > 1; queues 3812 drivers/net/wireless/broadcom/b43legacy/main.c hw->queues = 1; /* FIXME: hardware has more queues */ queues 909 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c u32 queues, bool drop) queues 1096 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c hw->queues = N_TX_QUEUES; queues 3555 drivers/net/wireless/intel/iwlegacy/3945-mac.c hw->queues = 4; queues 5767 drivers/net/wireless/intel/iwlegacy/4965-mac.c hw->queues = 4; queues 4754 drivers/net/wireless/intel/iwlegacy/common.c u32 queues, bool drop) queues 1713 drivers/net/wireless/intel/iwlegacy/common.h u32 queues, bool drop); queues 187 drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c hw->queues = IWLAGN_FIRST_AMPDU_QUEUE; queues 1085 drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c u32 queues, bool drop) queues 142 drivers/net/wireless/intel/iwlwifi/mvm/d3.c int tid, int queues) queues 151 drivers/net/wireless/intel/iwlwifi/mvm/d3.c for (i = 1; i < queues; i++) { queues 432 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c hw->queues = IEEE80211_MAX_QUEUES; queues 4602 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c static void iwl_mvm_flush_no_vif(struct iwl_mvm *mvm, u32 queues, bool drop) queues 4610 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c iwl_mvm_flushable_queues(mvm) & queues, 0); queues 4613 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c iwl_trans_wait_tx_queues_empty(mvm->trans, queues); queues 4637 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c struct ieee80211_vif *vif, u32 queues, bool drop) queues 4647 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c iwl_mvm_flush_no_vif(mvm, queues, drop); queues 152 drivers/net/wireless/intersil/p54/fwio.c priv->hw->queues = P54_QUEUE_AC_NUM; queues 667 drivers/net/wireless/intersil/p54/main.c u32 queues, bool drop) queues 761 drivers/net/wireless/intersil/p54/main.c dev->queues = 1; queues 177 drivers/net/wireless/intersil/p54/txrx.c for (i = 0; i < priv->hw->queues; i++) { queues 2003 drivers/net/wireless/mac80211_hwsim.c u32 queues, bool drop) queues 2891 drivers/net/wireless/mac80211_hwsim.c hw->queues = 5; queues 574 drivers/net/wireless/marvell/libertas_tf/main.c hw->queues = 1; queues 6092 drivers/net/wireless/marvell/mwl8k.c hw->queues = MWL8K_TX_WMM_QUEUES; queues 544 drivers/net/wireless/mediatek/mt76/mt7603/init.c hw->queues = 4; queues 541 drivers/net/wireless/mediatek/mt76/mt7603/main.c u32 queues, bool drop) queues 258 drivers/net/wireless/mediatek/mt76/mt7615/init.c hw->queues = 4; queues 69 drivers/net/wireless/mediatek/mt76/mt76x0/pci.c u32 queues, bool drop) queues 150 drivers/net/wireless/mediatek/mt76/mt76x02_util.c hw->queues = 4; queues 118 drivers/net/wireless/mediatek/mt76/mt76x2/pci_main.c u32 queues, bool drop) queues 595 drivers/net/wireless/mediatek/mt7601u/init.c hw->queues = 4; queues 1476 drivers/net/wireless/ralink/rt2x00/rt2x00.h u32 queues, bool drop); queues 1119 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00dev->hw->queues = rt2x00dev->ops->tx_queues; queues 694 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c u32 queues, bool drop) queues 1127 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c for (i = 0; i < (dev->queues + 1); i++) queues 1237 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c for (i = 0; i < (dev->queues + 1); i++) queues 1267 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c for (i = 0; i < (dev->queues + 1); i++) queues 1864 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c dev->queues = RTL8187SE_NR_TX_QUEUES - 1; queues 1866 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c dev->queues = RTL8180_NR_TX_QUEUES - 1; queues 1621 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c dev->queues = 1; queues 1623 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c dev->queues = 4; queues 6098 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_core.c hw->queues = 4; queues 404 drivers/net/wireless/realtek/rtlwifi/base.c hw->queues = AC_MAX; queues 1739 drivers/net/wireless/realtek/rtlwifi/core.c u32 queues, queues 1745 drivers/net/wireless/realtek/rtlwifi/core.c rtlpriv->intf_ops->flush(hw, queues, drop); queues 1704 drivers/net/wireless/realtek/rtlwifi/pci.c static void rtl_pci_flush(struct ieee80211_hw *hw, u32 queues, bool drop) queues 1720 drivers/net/wireless/realtek/rtlwifi/pci.c if (((queues >> queue_id) & 0x1) == 0) { queues 2348 drivers/net/wireless/realtek/rtlwifi/wifi.h void (*flush)(struct ieee80211_hw *hw, u32 queues, bool drop); queues 1231 drivers/net/wireless/realtek/rtw88/main.c hw->queues = IEEE80211_NUM_ACS; queues 2038 drivers/net/wireless/rsi/rsi_91x_mac80211.c hw->queues = MAX_HW_QUEUES; queues 300 drivers/net/wireless/st/cw1200/main.c hw->queues = 4; queues 618 drivers/net/wireless/st/cw1200/sta.c if (queue < dev->queues) { queues 929 drivers/net/wireless/st/cw1200/sta.c u32 queues, bool drop) queues 41 drivers/net/wireless/st/cw1200/sta.h u32 queues, bool drop); queues 1525 drivers/net/wireless/ti/wl1251/main.c wl->hw->queues = 4; queues 318 drivers/net/wireless/ti/wl12xx/main.c .queues = 0x1, queues 456 drivers/net/wireless/ti/wl18xx/main.c .queues = 0x1, queues 1492 drivers/net/wireless/ti/wlcore/acx.c conf_queues = wl->conf.rx_streaming.queues; queues 1031 drivers/net/wireless/ti/wlcore/conf.h u8 queues; queues 5630 drivers/net/wireless/ti/wlcore/main.c u32 queues, bool drop) queues 6319 drivers/net/wireless/ti/wlcore/main.c wl->hw->queues = (NUM_TX_QUEUES + 1) * WLCORE_NUM_MAC_ADDRESSES + 1; queues 6322 drivers/net/wireless/ti/wlcore/main.c wl->hw->offchannel_tx_hw_queue = wl->hw->queues - 1; queues 1400 drivers/net/wireless/zydas/zd1211rw/zd_mac.c hw->queues = 1; queues 293 drivers/net/xen-netback/common.h struct xenvif_queue *queues; queues 202 drivers/net/xen-netback/interface.c queue = &vif->queues[index]; queues 254 drivers/net/xen-netback/interface.c queue = &vif->queues[index]; queues 278 drivers/net/xen-netback/interface.c queue = &vif->queues[queue_index]; queues 294 drivers/net/xen-netback/interface.c queue = &vif->queues[queue_index]; queues 407 drivers/net/xen-netback/interface.c void *vif_stats = &vif->queues[queue_index].stats; queues 483 drivers/net/xen-netback/interface.c vif->queues = NULL; queues 747 drivers/net/xen-netback/interface.c queue = &vif->queues[queue_index]; queues 804 drivers/net/xen-netback/interface.c struct xenvif_queue *queues = vif->queues; queues 812 drivers/net/xen-netback/interface.c xenvif_deinit_queue(&queues[queue_index]); queues 813 drivers/net/xen-netback/interface.c vfree(queues); queues 219 drivers/net/xen-netback/netback.c xenvif_kick_thread(&vif->queues[0]); queues 182 drivers/net/xen-netback/xenbus.c &vif->queues[i], queues 450 drivers/net/xen-netback/xenbus.c xenvif_deinit_queue(&vif->queues[queue_index]); queues 452 drivers/net/xen-netback/xenbus.c vfree(vif->queues); queues 453 drivers/net/xen-netback/xenbus.c vif->queues = NULL; queues 690 drivers/net/xen-netback/xenbus.c struct xenvif_queue *queue = &vif->queues[queue_index]; queues 916 drivers/net/xen-netback/xenbus.c be->vif->queues = vzalloc(array_size(requested_num_queues, queues 918 drivers/net/xen-netback/xenbus.c if (!be->vif->queues) { queues 928 drivers/net/xen-netback/xenbus.c queue = &be->vif->queues[queue_index]; queues 992 drivers/net/xen-netback/xenbus.c xenvif_deinit_queue(&be->vif->queues[queue_index]); queues 993 drivers/net/xen-netback/xenbus.c vfree(be->vif->queues); queues 994 drivers/net/xen-netback/xenbus.c be->vif->queues = NULL; queues 156 drivers/net/xen-netfront.c struct netfront_queue *queues; queues 352 drivers/net/xen-netfront.c if (!np->queues) queues 356 drivers/net/xen-netfront.c queue = &np->queues[i]; queues 587 drivers/net/xen-netfront.c queue = &np->queues[queue_index]; queues 718 drivers/net/xen-netfront.c queue = &np->queues[i]; queues 1260 drivers/net/xen-netfront.c xennet_interrupt(0, &info->queues[i]); queues 1302 drivers/net/xen-netfront.c np->queues = NULL; queues 1392 drivers/net/xen-netfront.c for (i = 0; i < num_queues && info->queues; ++i) { queues 1393 drivers/net/xen-netfront.c struct netfront_queue *queue = &info->queues[i]; queues 1746 drivers/net/xen-netfront.c struct netfront_queue *queue = &info->queues[i]; queues 1753 drivers/net/xen-netfront.c kfree(info->queues); queues 1754 drivers/net/xen-netfront.c info->queues = NULL; queues 1763 drivers/net/xen-netfront.c info->queues = kcalloc(*num_queues, sizeof(struct netfront_queue), queues 1765 drivers/net/xen-netfront.c if (!info->queues) queues 1769 drivers/net/xen-netfront.c struct netfront_queue *queue = &info->queues[i]; queues 1829 drivers/net/xen-netfront.c if (info->queues) queues 1835 drivers/net/xen-netfront.c kfree(info->queues); queues 1836 drivers/net/xen-netfront.c info->queues = NULL; queues 1843 drivers/net/xen-netfront.c queue = &info->queues[i]; queues 1868 drivers/net/xen-netfront.c err = write_queue_xenstore_keys(&info->queues[0], &xbt, 0); /* flat */ queues 1874 drivers/net/xen-netfront.c queue = &info->queues[i]; queues 1987 drivers/net/xen-netfront.c queue = &np->queues[j]; queues 2169 drivers/net/xen-netfront.c if (info->queues) { queues 135 drivers/nvme/host/fc.c struct nvme_fc_queue *queues; queues 1768 drivers/nvme/host/fc.c struct nvme_fc_queue *queue = &ctrl->queues[queue_idx]; queues 1798 drivers/nvme/host/fc.c ret = __nvme_fc_init_request(ctrl, &ctrl->queues[0], queues 1839 drivers/nvme/host/fc.c struct nvme_fc_queue *queue = &ctrl->queues[qidx]; queues 1872 drivers/nvme/host/fc.c queue = &ctrl->queues[idx]; queues 1937 drivers/nvme/host/fc.c nvme_fc_free_queue(&ctrl->queues[i]); queues 1957 drivers/nvme/host/fc.c struct nvme_fc_queue *queue = &ctrl->queues[ctrl->ctrl.queue_count - 1]; queues 1967 drivers/nvme/host/fc.c struct nvme_fc_queue *queue = &ctrl->queues[1]; queues 1980 drivers/nvme/host/fc.c __nvme_fc_delete_hw_queue(ctrl, &ctrl->queues[i], i); queues 1990 drivers/nvme/host/fc.c ret = nvme_fc_connect_queue(ctrl, &ctrl->queues[i], qsize, queues 1998 drivers/nvme/host/fc.c set_bit(NVME_FC_Q_LIVE, &ctrl->queues[i].flags); queues 2035 drivers/nvme/host/fc.c kfree(ctrl->queues); queues 2647 drivers/nvme/host/fc.c ret = __nvme_fc_create_hw_queue(ctrl, &ctrl->queues[0], 0, queues 2652 drivers/nvme/host/fc.c ret = nvme_fc_connect_admin_queue(ctrl, &ctrl->queues[0], queues 2661 drivers/nvme/host/fc.c set_bit(NVME_FC_Q_LIVE, &ctrl->queues[0].flags); queues 2743 drivers/nvme/host/fc.c __nvme_fc_delete_hw_queue(ctrl, &ctrl->queues[0], 0); queues 2745 drivers/nvme/host/fc.c nvme_fc_free_queue(&ctrl->queues[0]); queues 2838 drivers/nvme/host/fc.c __nvme_fc_delete_hw_queue(ctrl, &ctrl->queues[0], 0); queues 2839 drivers/nvme/host/fc.c nvme_fc_free_queue(&ctrl->queues[0]); queues 3112 drivers/nvme/host/fc.c ctrl->queues = kcalloc(ctrl->ctrl.queue_count, queues 3114 drivers/nvme/host/fc.c if (!ctrl->queues) queues 3223 drivers/nvme/host/fc.c kfree(ctrl->queues); queues 91 drivers/nvme/host/pci.c struct nvme_queue *queues; queues 375 drivers/nvme/host/pci.c struct nvme_queue *nvmeq = &dev->queues[0]; queues 388 drivers/nvme/host/pci.c struct nvme_queue *nvmeq = &dev->queues[hctx_idx + 1]; queues 401 drivers/nvme/host/pci.c struct nvme_queue *nvmeq = &dev->queues[queue_idx]; queues 1096 drivers/nvme/host/pci.c struct nvme_queue *nvmeq = &dev->queues[0]; queues 1366 drivers/nvme/host/pci.c nvme_free_queue(&dev->queues[i]); queues 1395 drivers/nvme/host/pci.c nvme_suspend_queue(&dev->queues[i]); queues 1400 drivers/nvme/host/pci.c struct nvme_queue *nvmeq = &dev->queues[0]; queues 1422 drivers/nvme/host/pci.c nvme_process_cq(&dev->queues[i], &start, &end, -1); queues 1423 drivers/nvme/host/pci.c nvme_complete_cqes(&dev->queues[i], start, end); queues 1479 drivers/nvme/host/pci.c struct nvme_queue *nvmeq = &dev->queues[qid]; queues 1705 drivers/nvme/host/pci.c nvmeq = &dev->queues[0]; queues 1752 drivers/nvme/host/pci.c ret = nvme_create_queue(&dev->queues[i], i, polled); queues 2110 drivers/nvme/host/pci.c struct nvme_queue *adminq = &dev->queues[0]; queues 2248 drivers/nvme/host/pci.c if (nvme_delete_queue(&dev->queues[nr_queues], opcode)) queues 2254 drivers/nvme/host/pci.c struct nvme_queue *nvmeq = &dev->queues[nr_queues + sent]; queues 2447 drivers/nvme/host/pci.c nvme_suspend_queue(&dev->queues[0]); queues 2516 drivers/nvme/host/pci.c kfree(dev->queues); queues 2797 drivers/nvme/host/pci.c dev->queues = kcalloc_node(max_queue_count(), sizeof(struct nvme_queue), queues 2799 drivers/nvme/host/pci.c if (!dev->queues) queues 2858 drivers/nvme/host/pci.c kfree(dev->queues); queues 92 drivers/nvme/host/rdma.c struct nvme_rdma_queue *queues; queues 155 drivers/nvme/host/rdma.c return queue - queue->ctrl->queues; queues 294 drivers/nvme/host/rdma.c struct nvme_rdma_queue *queue = &ctrl->queues[queue_idx]; queues 310 drivers/nvme/host/rdma.c struct nvme_rdma_queue *queue = &ctrl->queues[hctx_idx + 1]; queues 322 drivers/nvme/host/rdma.c struct nvme_rdma_queue *queue = &ctrl->queues[0]; queues 522 drivers/nvme/host/rdma.c queue = &ctrl->queues[idx]; queues 598 drivers/nvme/host/rdma.c nvme_rdma_free_queue(&ctrl->queues[i]); queues 606 drivers/nvme/host/rdma.c nvme_rdma_stop_queue(&ctrl->queues[i]); queues 611 drivers/nvme/host/rdma.c struct nvme_rdma_queue *queue = &ctrl->queues[idx]; queues 645 drivers/nvme/host/rdma.c nvme_rdma_stop_queue(&ctrl->queues[i]); queues 714 drivers/nvme/host/rdma.c nvme_rdma_free_queue(&ctrl->queues[i]); queues 775 drivers/nvme/host/rdma.c nvme_rdma_free_queue(&ctrl->queues[0]); queues 787 drivers/nvme/host/rdma.c ctrl->device = ctrl->queues[0].device; queues 842 drivers/nvme/host/rdma.c nvme_rdma_stop_queue(&ctrl->queues[0]); queues 859 drivers/nvme/host/rdma.c nvme_rdma_free_queue(&ctrl->queues[0]); queues 919 drivers/nvme/host/rdma.c nvme_rdma_stop_queue(&ctrl->queues[0]); queues 960 drivers/nvme/host/rdma.c kfree(ctrl->queues); queues 1040 drivers/nvme/host/rdma.c nvme_rdma_stop_queue(&ctrl->queues[0]); queues 1417 drivers/nvme/host/rdma.c struct nvme_rdma_queue *queue = &ctrl->queues[0]; queues 2030 drivers/nvme/host/rdma.c ctrl->queues = kcalloc(ctrl->ctrl.queue_count, sizeof(*ctrl->queues), queues 2032 drivers/nvme/host/rdma.c if (!ctrl->queues) queues 2065 drivers/nvme/host/rdma.c kfree(ctrl->queues); queues 103 drivers/nvme/host/tcp.c struct nvme_tcp_queue *queues; queues 132 drivers/nvme/host/tcp.c return queue - queue->ctrl->queues; queues 366 drivers/nvme/host/tcp.c struct nvme_tcp_queue *queue = &ctrl->queues[queue_idx]; queues 385 drivers/nvme/host/tcp.c struct nvme_tcp_queue *queue = &ctrl->queues[hctx_idx + 1]; queues 395 drivers/nvme/host/tcp.c struct nvme_tcp_queue *queue = &ctrl->queues[0]; queues 1122 drivers/nvme/host/tcp.c struct nvme_tcp_queue *queue = &ctrl->queues[0]; queues 1132 drivers/nvme/host/tcp.c async->queue = &ctrl->queues[0]; queues 1139 drivers/nvme/host/tcp.c struct nvme_tcp_queue *queue = &ctrl->queues[qid]; queues 1253 drivers/nvme/host/tcp.c struct nvme_tcp_queue *queue = &ctrl->queues[qid]; queues 1434 drivers/nvme/host/tcp.c struct nvme_tcp_queue *queue = &ctrl->queues[qid]; queues 1453 drivers/nvme/host/tcp.c set_bit(NVME_TCP_Q_LIVE, &ctrl->queues[idx].flags); queues 1455 drivers/nvme/host/tcp.c if (test_bit(NVME_TCP_Q_ALLOCATED, &ctrl->queues[idx].flags)) queues 1456 drivers/nvme/host/tcp.c __nvme_tcp_stop_queue(&ctrl->queues[idx]); queues 1982 drivers/nvme/host/tcp.c kfree(ctrl->queues); queues 2020 drivers/nvme/host/tcp.c struct nvme_tcp_queue *queue = &ctrl->queues[0]; queues 2339 drivers/nvme/host/tcp.c ctrl->queues = kcalloc(ctrl->ctrl.queue_count, sizeof(*ctrl->queues), queues 2341 drivers/nvme/host/tcp.c if (!ctrl->queues) { queues 2378 drivers/nvme/host/tcp.c kfree(ctrl->queues); queues 144 drivers/nvme/target/fc.c struct nvmet_fc_tgt_queue *queues[NVMET_NR_QUEUES + 1]; queues 623 drivers/nvme/target/fc.c WARN_ON(assoc->queues[qid]); queues 625 drivers/nvme/target/fc.c assoc->queues[qid] = queue; queues 649 drivers/nvme/target/fc.c queue->assoc->queues[queue->qid] = NULL; queues 767 drivers/nvme/target/fc.c queue = assoc->queues[qid]; queues 881 drivers/nvme/target/fc.c queue = assoc->queues[i]; queues 1171 drivers/nvme/target/fc.c queue = assoc->queues[0]; queues 30 drivers/nvme/target/loop.c struct nvme_loop_queue *queues; queues 72 drivers/nvme/target/loop.c return queue - queue->ctrl->queues; queues 177 drivers/nvme/target/loop.c struct nvme_loop_queue *queue = &ctrl->queues[0]; queues 199 drivers/nvme/target/loop.c iod->queue = &ctrl->queues[queue_idx]; queues 219 drivers/nvme/target/loop.c struct nvme_loop_queue *queue = &ctrl->queues[hctx_idx + 1]; queues 231 drivers/nvme/target/loop.c struct nvme_loop_queue *queue = &ctrl->queues[0]; queues 255 drivers/nvme/target/loop.c clear_bit(NVME_LOOP_Q_LIVE, &ctrl->queues[0].flags); queues 256 drivers/nvme/target/loop.c nvmet_sq_destroy(&ctrl->queues[0].nvme_sq); queues 277 drivers/nvme/target/loop.c kfree(ctrl->queues); queues 288 drivers/nvme/target/loop.c clear_bit(NVME_LOOP_Q_LIVE, &ctrl->queues[i].flags); queues 289 drivers/nvme/target/loop.c nvmet_sq_destroy(&ctrl->queues[i].nvme_sq); queues 307 drivers/nvme/target/loop.c ctrl->queues[i].ctrl = ctrl; queues 308 drivers/nvme/target/loop.c ret = nvmet_sq_init(&ctrl->queues[i].nvme_sq); queues 330 drivers/nvme/target/loop.c set_bit(NVME_LOOP_Q_LIVE, &ctrl->queues[i].flags); queues 352 drivers/nvme/target/loop.c ctrl->queues[0].ctrl = ctrl; queues 353 drivers/nvme/target/loop.c error = nvmet_sq_init(&ctrl->queues[0].nvme_sq); queues 379 drivers/nvme/target/loop.c set_bit(NVME_LOOP_Q_LIVE, &ctrl->queues[0].flags); queues 403 drivers/nvme/target/loop.c nvmet_sq_destroy(&ctrl->queues[0].nvme_sq); queues 594 drivers/nvme/target/loop.c ctrl->queues = kcalloc(opts->nr_io_queues + 1, sizeof(*ctrl->queues), queues 596 drivers/nvme/target/loop.c if (!ctrl->queues) queues 638 drivers/nvme/target/loop.c kfree(ctrl->queues); queues 804 drivers/s390/crypto/ap_bus.c &to_ap_queue(dev)->card->queues); queues 159 drivers/s390/crypto/ap_bus.h struct list_head queues; /* List of assoc. AP queues */ queues 182 drivers/s390/crypto/ap_card.c INIT_LIST_HEAD(&ac->queues); queues 1558 drivers/scsi/aacraid/aacraid.h struct aac_queue_block *queues; queues 313 drivers/scsi/aacraid/commctrl.c if (status && !dev->in_reset && dev->queues && dev->fsa_dev) { queues 374 drivers/scsi/aacraid/comminit.c struct aac_entry * queues; queues 376 drivers/scsi/aacraid/comminit.c struct aac_queue_block * comm = dev->queues; queues 395 drivers/scsi/aacraid/comminit.c queues = (struct aac_entry *)(((ulong)headers) + hdrsize); queues 398 drivers/scsi/aacraid/comminit.c comm->queue[HostNormCmdQueue].base = queues; queues 400 drivers/scsi/aacraid/comminit.c queues += HOST_NORM_CMD_ENTRIES; queues 404 drivers/scsi/aacraid/comminit.c comm->queue[HostHighCmdQueue].base = queues; queues 407 drivers/scsi/aacraid/comminit.c queues += HOST_HIGH_CMD_ENTRIES; queues 411 drivers/scsi/aacraid/comminit.c comm->queue[AdapNormCmdQueue].base = queues; queues 414 drivers/scsi/aacraid/comminit.c queues += ADAP_NORM_CMD_ENTRIES; queues 418 drivers/scsi/aacraid/comminit.c comm->queue[AdapHighCmdQueue].base = queues; queues 421 drivers/scsi/aacraid/comminit.c queues += ADAP_HIGH_CMD_ENTRIES; queues 425 drivers/scsi/aacraid/comminit.c comm->queue[HostNormRespQueue].base = queues; queues 427 drivers/scsi/aacraid/comminit.c queues += HOST_NORM_RESP_ENTRIES; queues 431 drivers/scsi/aacraid/comminit.c comm->queue[HostHighRespQueue].base = queues; queues 434 drivers/scsi/aacraid/comminit.c queues += HOST_HIGH_RESP_ENTRIES; queues 438 drivers/scsi/aacraid/comminit.c comm->queue[AdapNormRespQueue].base = queues; queues 441 drivers/scsi/aacraid/comminit.c queues += ADAP_NORM_RESP_ENTRIES; queues 445 drivers/scsi/aacraid/comminit.c comm->queue[AdapHighRespQueue].base = queues; queues 633 drivers/scsi/aacraid/comminit.c dev->queues = kzalloc(sizeof(struct aac_queue_block), GFP_KERNEL); queues 634 drivers/scsi/aacraid/comminit.c if (dev->queues == NULL) { queues 640 drivers/scsi/aacraid/comminit.c kfree(dev->queues); queues 647 drivers/scsi/aacraid/comminit.c kfree(dev->queues); queues 369 drivers/scsi/aacraid/commsup.c q = &dev->queues->queue[qid]; queues 578 drivers/scsi/aacraid/commsup.c if (!dev->queues) queues 650 drivers/scsi/aacraid/commsup.c struct aac_queue * q = &dev->queues->queue[AdapNormCmdQueue]; queues 918 drivers/scsi/aacraid/commsup.c q = &dev->queues->queue[AdapNormRespQueue]; queues 1552 drivers/scsi/aacraid/commsup.c kfree(aac->queues); queues 1553 drivers/scsi/aacraid/commsup.c aac->queues = NULL; queues 2175 drivers/scsi/aacraid/commsup.c t_lock = dev->queues->queue[HostNormCmdQueue].lock; queues 2178 drivers/scsi/aacraid/commsup.c while (!list_empty(&(dev->queues->queue[HostNormCmdQueue].cmdq))) { queues 2187 drivers/scsi/aacraid/commsup.c entry = dev->queues->queue[HostNormCmdQueue].cmdq.next; queues 2190 drivers/scsi/aacraid/commsup.c t_lock = dev->queues->queue[HostNormCmdQueue].lock; queues 2282 drivers/scsi/aacraid/commsup.c t_lock = dev->queues->queue[HostNormCmdQueue].lock; queues 2288 drivers/scsi/aacraid/commsup.c t_lock = dev->queues->queue[HostNormCmdQueue].lock; queues 2457 drivers/scsi/aacraid/commsup.c add_wait_queue(&dev->queues->queue[HostNormCmdQueue].cmdready, &wait); queues 2474 drivers/scsi/aacraid/commsup.c } else if (!dev->queues) queues 2484 drivers/scsi/aacraid/commsup.c if (ret || !dev->queues) queues 2528 drivers/scsi/aacraid/commsup.c if (dev->queues) queues 2529 drivers/scsi/aacraid/commsup.c remove_wait_queue(&dev->queues->queue[HostNormCmdQueue].cmdready, &wait); queues 72 drivers/scsi/aacraid/dpcsup.c atomic_dec(&dev->queues->queue[AdapNormCmdQueue].numpending); queues 278 drivers/scsi/aacraid/dpcsup.c struct aac_queue *q = &dev->queues->queue[HostNormCmdQueue]; queues 344 drivers/scsi/aacraid/dpcsup.c atomic_dec(&dev->queues->queue[AdapNormCmdQueue].numpending); queues 1804 drivers/scsi/aacraid/linit.c kfree(aac->queues); queues 1941 drivers/scsi/aacraid/linit.c kfree(aac->queues); queues 55 drivers/scsi/aacraid/rx.c aac_command_normal(&dev->queues->queue[HostNormCmdQueue]); queues 59 drivers/scsi/aacraid/rx.c aac_response_normal(&dev->queues->queue[HostNormRespQueue]); queues 391 drivers/scsi/aacraid/rx.c struct aac_queue *q = &dev->queues->queue[AdapNormCmdQueue]; queues 414 drivers/scsi/aacraid/rx.c struct aac_queue *q = &dev->queues->queue[AdapNormCmdQueue]; queues 55 drivers/scsi/aacraid/sa.c aac_command_normal(&dev->queues->queue[HostNormCmdQueue]); queues 58 drivers/scsi/aacraid/sa.c aac_response_normal(&dev->queues->queue[HostNormRespQueue]); queues 477 drivers/scsi/aacraid/src.c struct aac_queue *q = &dev->queues->queue[AdapNormCmdQueue]; queues 706 drivers/scsi/arm/acornscsi.c SCpnt = queue_remove_exclude(&host->queues.issue, host->busyluns); queues 714 drivers/scsi/arm/acornscsi.c queue_add_cmd_tail(&host->queues.disconnected, host->SCpnt); queues 1795 drivers/scsi/arm/acornscsi.c if (!ok && queue_probetgtlun(&host->queues.disconnected, target, lun)) queues 1810 drivers/scsi/arm/acornscsi.c queue_add_cmd_tail(&host->queues.disconnected, host->SCpnt); queues 1837 drivers/scsi/arm/acornscsi.c queue_add_cmd_tail(&host->queues.disconnected, host->SCpnt); queues 1847 drivers/scsi/arm/acornscsi.c host->SCpnt = queue_remove_tgtluntag(&host->queues.disconnected, queues 2499 drivers/scsi/arm/acornscsi.c host->stats.queues += 1; queues 2504 drivers/scsi/arm/acornscsi.c if (!queue_add_cmd_ordered(&host->queues.issue, SCpnt)) { queues 2556 drivers/scsi/arm/acornscsi.c if (queue_remove_cmd(&host->queues.issue, SCpnt)) { queues 2567 drivers/scsi/arm/acornscsi.c } else if (queue_remove_cmd(&host->queues.disconnected, SCpnt)) { queues 2754 drivers/scsi/arm/acornscsi.c while ((SCptr = queue_remove(&host->queues.disconnected)) != NULL) queues 2825 drivers/scsi/arm/acornscsi.c host->stats.queues, host->stats.removes, queues 2932 drivers/scsi/arm/acornscsi.c queue_initialise(&ashost->queues.issue); queues 2933 drivers/scsi/arm/acornscsi.c queue_initialise(&ashost->queues.disconnected); queues 2948 drivers/scsi/arm/acornscsi.c queue_free(&ashost->queues.disconnected); queues 2949 drivers/scsi/arm/acornscsi.c queue_free(&ashost->queues.issue); queues 2976 drivers/scsi/arm/acornscsi.c queue_free(&ashost->queues.disconnected); queues 2977 drivers/scsi/arm/acornscsi.c queue_free(&ashost->queues.issue); queues 302 drivers/scsi/arm/acornscsi.h unsigned int queues; queues 317 drivers/scsi/arm/acornscsi.h } queues; queues 206 drivers/scsi/arm/fas216.c info->stats.queues, info->stats.removes, info->stats.fins, queues 996 drivers/scsi/arm/fas216.c queue_add_cmd_tail(&info->queues.disconnected, info->SCpnt); queues 1002 drivers/scsi/arm/fas216.c info->SCpnt = queue_remove_tgtluntag(&info->queues.disconnected, queues 1932 drivers/scsi/arm/fas216.c SCpnt = queue_remove_exclude(&info->queues.issue, queues 1956 drivers/scsi/arm/fas216.c queue_add_cmd_tail(&info->queues.disconnected, info->SCpnt); queues 2223 drivers/scsi/arm/fas216.c info->stats.queues += 1; queues 2232 drivers/scsi/arm/fas216.c result = !queue_add_cmd_ordered(&info->queues.issue, SCpnt); queues 2360 drivers/scsi/arm/fas216.c if (queue_remove_cmd(&info->queues.issue, SCpnt)) { queues 2370 drivers/scsi/arm/fas216.c } else if (queue_remove_cmd(&info->queues.disconnected, SCpnt)) { queues 2504 drivers/scsi/arm/fas216.c queue_remove_all_target(&info->queues.issue, target); queues 2505 drivers/scsi/arm/fas216.c queue_remove_all_target(&info->queues.disconnected, target); queues 2596 drivers/scsi/arm/fas216.c queue_remove_all_target(&info->queues.issue, SDpnt->id); queues 2597 drivers/scsi/arm/fas216.c queue_remove_all_target(&info->queues.disconnected, SDpnt->id); queues 2823 drivers/scsi/arm/fas216.c while (queue_remove(&info->queues.disconnected) != NULL); queues 2865 drivers/scsi/arm/fas216.c if (!queue_initialise(&info->queues.issue)) queues 2868 drivers/scsi/arm/fas216.c if (!queue_initialise(&info->queues.disconnected)) { queues 2869 drivers/scsi/arm/fas216.c queue_free(&info->queues.issue); queues 2956 drivers/scsi/arm/fas216.c queue_free(&info->queues.disconnected); queues 2957 drivers/scsi/arm/fas216.c queue_free(&info->queues.issue); queues 2986 drivers/scsi/arm/fas216.c info->stats.queues, info->stats.removes, queues 255 drivers/scsi/arm/fas216.h unsigned int queues; queues 283 drivers/scsi/arm/fas216.h } queues; queues 679 drivers/scsi/qla4xxx/ql4_def.h void *queues; queues 4144 drivers/scsi/qla4xxx/ql4_os.c if (ha->queues) queues 4145 drivers/scsi/qla4xxx/ql4_os.c dma_free_coherent(&ha->pdev->dev, ha->queues_len, ha->queues, queues 4152 drivers/scsi/qla4xxx/ql4_os.c ha->queues = NULL; queues 4211 drivers/scsi/qla4xxx/ql4_os.c ha->queues = dma_alloc_coherent(&ha->pdev->dev, ha->queues_len, queues 4213 drivers/scsi/qla4xxx/ql4_os.c if (ha->queues == NULL) { queues 4231 drivers/scsi/qla4xxx/ql4_os.c ha->request_ring = (struct queue_entry *) (ha->queues + align); queues 4234 drivers/scsi/qla4xxx/ql4_os.c ha->response_ring = (struct queue_entry *) (ha->queues + align + queues 4240 drivers/scsi/qla4xxx/ql4_os.c ha->shadow_regs = (struct shadow_regs *) (ha->queues + align + queues 76 drivers/staging/media/ipu3/ipu3.c &imgu_pipe->queues[i].dmap); queues 97 drivers/staging/media/ipu3/ipu3.c &imgu_pipe->queues[i].dmap, size)) { queues 137 drivers/staging/media/ipu3/ipu3.c &imgu_pipe->queues[i].dmap, queues 144 drivers/staging/media/ipu3/ipu3.c imgu_css_buf_init(&imgu_pipe->queues[i].dummybufs[k], i, queues 145 drivers/staging/media/ipu3/ipu3.c imgu_pipe->queues[i].dmap.daddr); queues 162 drivers/staging/media/ipu3/ipu3.c if (WARN_ON(!imgu_pipe->queues[queue].dmap.vaddr)) queues 167 drivers/staging/media/ipu3/ipu3.c if (imgu_css_buf_state(&imgu_pipe->queues[queue].dummybufs[i]) != queues 174 drivers/staging/media/ipu3/ipu3.c imgu_css_buf_init(&imgu_pipe->queues[queue].dummybufs[i], queue, queues 175 drivers/staging/media/ipu3/ipu3.c imgu_pipe->queues[queue].dmap.daddr); queues 177 drivers/staging/media/ipu3/ipu3.c return &imgu_pipe->queues[queue].dummybufs[i]; queues 189 drivers/staging/media/ipu3/ipu3.c if (buf == &imgu_pipe->queues[buf->queue].dummybufs[i]) queues 112 drivers/staging/media/ipu3/ipu3.h } queues[IPU3_CSS_QUEUES]; queues 1895 drivers/usb/gadget/udc/net2280.c static DEVICE_ATTR_RO(queues); queues 123 fs/autofs/autofs_i.h struct autofs_wait_queue *queues; /* Wait queue pointer */ queues 252 fs/autofs/inode.c sbi->queues = NULL; queues 28 fs/autofs/waitq.c wq = sbi->queues; queues 29 fs/autofs/waitq.c sbi->queues = NULL; /* Erase all wait queues */ queues 228 fs/autofs/waitq.c for (wq = sbi->queues; wq; wq = wq->next) { queues 429 fs/autofs/waitq.c wq->next = sbi->queues; queues 430 fs/autofs/waitq.c sbi->queues = wq; queues 532 fs/autofs/waitq.c for (wql = &sbi->queues; (wq = *wql) != NULL; wql = &wq->next) { queues 770 include/linux/netdevice.h u16 queues[0]; queues 773 include/linux/netdevice.h #define XPS_MIN_MAP_ALLOC ((L1_CACHE_ALIGN(offsetof(struct xps_map, queues[1])) \ queues 624 include/linux/ptr_ring.h void ***queues; queues 627 include/linux/ptr_ring.h queues = kmalloc_array(nrings, sizeof(*queues), gfp); queues 628 include/linux/ptr_ring.h if (!queues) queues 632 include/linux/ptr_ring.h queues[i] = __ptr_ring_init_queue_alloc(size, gfp); queues 633 include/linux/ptr_ring.h if (!queues[i]) queues 640 include/linux/ptr_ring.h queues[i] = __ptr_ring_swap_queue(rings[i], queues[i], queues 647 include/linux/ptr_ring.h kvfree(queues[i]); queues 649 include/linux/ptr_ring.h kfree(queues); queues 655 include/linux/ptr_ring.h kvfree(queues[i]); queues 657 include/linux/ptr_ring.h kfree(queues); queues 2481 include/net/mac80211.h u16 queues; queues 3917 include/net/mac80211.h u32 queues, bool drop); queues 313 include/uapi/sound/asequencer.h int queues; /* maximum queues count */ queues 265 kernel/futex.c struct futex_hash_bucket *queues; queues 268 kernel/futex.c #define futex_queues (__futex_data.queues) queues 2082 net/core/dev.c if (map->queues[pos] != index) queues 2086 net/core/dev.c map->queues[pos] = map->queues[--map->len]; queues 2207 net/core/dev.c if (map->queues[pos] != index) queues 2232 net/core/dev.c new_map->queues[i] = map->queues[i]; queues 2334 net/core/dev.c while ((pos < map->len) && (map->queues[pos] != index)) queues 2338 net/core/dev.c map->queues[map->len++] = index; queues 3552 net/core/dev.c queue_index = map->queues[0]; queues 3554 net/core/dev.c queue_index = map->queues[reciprocal_scale( queues 1278 net/core/net-sysfs.c if (map->queues[i] == index) { queues 1361 net/core/net-sysfs.c if (map->queues[i] == index) { queues 1369 net/mac80211/cfg.c local->hw.queues >= IEEE80211_NUM_ACS) queues 2244 net/mac80211/cfg.c if (local->hw.queues < IEEE80211_NUM_ACS) queues 348 net/mac80211/debugfs.c for (q = 0; q < local->hw.queues; q++) queues 358 net/mac80211/debugfs.c DEBUGFS_READONLY_FILE_OPS(queues); queues 428 net/mac80211/debugfs.c DEBUGFS_ADD(queues); queues 628 net/mac80211/driver-ops.h u32 queues, bool drop) queues 637 net/mac80211/driver-ops.h trace_drv_flush(local, queues, drop); queues 639 net/mac80211/driver-ops.h local->ops->flush(&local->hw, vif, queues, drop); queues 205 net/mac80211/ibss.c if (local->hw.queues >= IEEE80211_NUM_ACS) queues 1042 net/mac80211/ibss.c elems->wmm_info && local->hw.queues >= IEEE80211_NUM_ACS) { queues 2016 net/mac80211/ieee80211_i.h unsigned long queues, queues 2026 net/mac80211/ieee80211_i.h unsigned long queues, queues 2044 net/mac80211/ieee80211_i.h unsigned int queues, bool drop); queues 317 net/mac80211/iface.c int n_queues = sdata->local->hw.queues; queues 379 net/mac80211/iface.c else if (local->hw.queues >= IEEE80211_NUM_ACS) queues 741 net/mac80211/iface.c if (local->hw.queues < IEEE80211_NUM_ACS) queues 1185 net/mac80211/iface.c if (local->hw.queues < IEEE80211_NUM_ACS) queues 1766 net/mac80211/iface.c if (local->hw.queues >= IEEE80211_NUM_ACS) queues 627 net/mac80211/main.c local->hw.queues = 1; queues 874 net/mac80211/main.c local->hw.offchannel_tx_hw_queue >= local->hw.queues)) queues 1140 net/mac80211/main.c if (hw->queues > IEEE80211_MAX_QUEUES) queues 1141 net/mac80211/main.c hw->queues = IEEE80211_MAX_QUEUES; queues 1772 net/mac80211/mlme.c for (q = 0; q < local->hw.queues; q++) { queues 1840 net/mac80211/mlme.c if (local->hw.queues < IEEE80211_NUM_ACS) queues 1942 net/mac80211/mlme.c if (local->hw.queues < IEEE80211_NUM_ACS) queues 3439 net/mac80211/mlme.c sta->sta.wme = elems.wmm_param && local->hw.queues >= IEEE80211_NUM_ACS; queues 5280 net/mac80211/mlme.c (local->hw.queues >= IEEE80211_NUM_ACS); queues 5304 net/mac80211/mlme.c local->hw.queues < IEEE80211_NUM_ACS || !bss->wmm_used || queues 5315 net/mac80211/mlme.c local->hw.queues < IEEE80211_NUM_ACS || !bss->wmm_used || queues 403 net/mac80211/tdls.c if (local->hw.queues >= IEEE80211_NUM_ACS && queues 584 net/mac80211/tdls.c if (local->hw.queues >= IEEE80211_NUM_ACS && sta->sta.wme) queues 1108 net/mac80211/trace.h u32 queues, bool drop), queues 1110 net/mac80211/trace.h TP_ARGS(local, queues, drop), queues 1115 net/mac80211/trace.h __field(u32, queues) queues 1121 net/mac80211/trace.h __entry->queues = queues; queues 1126 net/mac80211/trace.h LOCAL_PR_ARG, __entry->queues, __entry->drop queues 1639 net/mac80211/tx.c if (WARN_ON_ONCE(q >= local->hw.queues)) { queues 4112 net/mac80211/tx.c for (i = 0; i < local->hw.queues; i++) { queues 4171 net/mac80211/tx.c for (i = 0; i < local->hw.queues; i++) { queues 4972 net/mac80211/tx.c u32 queues; queues 5013 net/mac80211/tx.c queues = BIT(sdata->vif.hw_queue[ieee802_1d_to_ac[tid]]); queues 5014 net/mac80211/tx.c __ieee80211_flush_queues(local, sdata, queues, false); queues 313 net/mac80211/util.c if (local->hw.queues < IEEE80211_NUM_ACS) queues 316 net/mac80211/util.c for (i = 0; i < local->hw.queues; i++) { queues 356 net/mac80211/util.c if (local->hw.queues < IEEE80211_NUM_ACS) queues 390 net/mac80211/util.c if (WARN_ON(queue >= hw->queues)) queues 463 net/mac80211/util.c if (WARN_ON(queue >= hw->queues)) queues 474 net/mac80211/util.c if (local->hw.queues < IEEE80211_NUM_ACS) queues 568 net/mac80211/util.c for (i = 0; i < hw->queues; i++) queues 576 net/mac80211/util.c unsigned long queues, queues 586 net/mac80211/util.c for_each_set_bit(i, &queues, hw->queues) queues 606 net/mac80211/util.c if (WARN_ON(queue >= hw->queues)) queues 618 net/mac80211/util.c unsigned long queues, queues 628 net/mac80211/util.c for_each_set_bit(i, &queues, hw->queues) queues 646 net/mac80211/util.c unsigned int queues; queues 651 net/mac80211/util.c queues = 0; queues 654 net/mac80211/util.c queues |= BIT(sdata->vif.hw_queue[ac]); queues 656 net/mac80211/util.c queues |= BIT(sdata->vif.cab_queue); queues 659 net/mac80211/util.c queues = BIT(local->hw.queues) - 1; queues 662 net/mac80211/util.c return queues; queues 667 net/mac80211/util.c unsigned int queues, bool drop) queues 676 net/mac80211/util.c if (!queues || !ieee80211_hw_check(&local->hw, QUEUE_CONTROL)) queues 677 net/mac80211/util.c queues = ieee80211_get_vif_queues(local, sdata); queues 679 net/mac80211/util.c ieee80211_stop_queues_by_reason(&local->hw, queues, queues 683 net/mac80211/util.c drv_flush(local, sdata, queues, drop); queues 685 net/mac80211/util.c ieee80211_wake_queues_by_reason(&local->hw, queues, queues 1455 net/mac80211/util.c if (local->hw.queues < IEEE80211_NUM_ACS) queues 123 net/mac80211/wme.c if (local->hw.queues < IEEE80211_NUM_ACS) queues 190 net/mac80211/wme.c if (local->hw.queues < IEEE80211_NUM_ACS || skb->len < 6) { queues 25 net/sched/sch_multiq.c struct Qdisc **queues; queues 54 net/sched/sch_multiq.c return q->queues[0]; queues 56 net/sched/sch_multiq.c return q->queues[band]; queues 105 net/sched/sch_multiq.c qdisc = q->queues[q->curband]; queues 137 net/sched/sch_multiq.c qdisc = q->queues[curband]; queues 154 net/sched/sch_multiq.c qdisc_reset(q->queues[band]); queues 167 net/sched/sch_multiq.c qdisc_put(q->queues[band]); queues 169 net/sched/sch_multiq.c kfree(q->queues); queues 197 net/sched/sch_multiq.c if (q->queues[i] != &noop_qdisc) { queues 198 net/sched/sch_multiq.c struct Qdisc *child = q->queues[i]; queues 200 net/sched/sch_multiq.c q->queues[i] = &noop_qdisc; queues 213 net/sched/sch_multiq.c if (q->queues[i] == &noop_qdisc) { queues 221 net/sched/sch_multiq.c old = q->queues[i]; queues 222 net/sched/sch_multiq.c q->queues[i] = child; queues 242 net/sched/sch_multiq.c q->queues = NULL; queues 253 net/sched/sch_multiq.c q->queues = kcalloc(q->max_bands, sizeof(struct Qdisc *), GFP_KERNEL); queues 254 net/sched/sch_multiq.c if (!q->queues) queues 257 net/sched/sch_multiq.c q->queues[i] = &noop_qdisc; queues 290 net/sched/sch_multiq.c *old = qdisc_replace(sch, new, &q->queues[band]); queues 300 net/sched/sch_multiq.c return q->queues[band]; queues 330 net/sched/sch_multiq.c tcm->tcm_info = q->queues[cl - 1]->handle; queues 340 net/sched/sch_multiq.c cl_q = q->queues[cl - 1]; queues 26 net/sched/sch_prio.c struct Qdisc *queues[TCQ_PRIO_BANDS]; queues 57 net/sched/sch_prio.c return q->queues[q->prio2band[band & TC_PRIO_MAX]]; queues 63 net/sched/sch_prio.c return q->queues[q->prio2band[0]]; queues 65 net/sched/sch_prio.c return q->queues[band]; queues 103 net/sched/sch_prio.c struct Qdisc *qdisc = q->queues[prio]; queues 117 net/sched/sch_prio.c struct Qdisc *qdisc = q->queues[prio]; queues 137 net/sched/sch_prio.c qdisc_reset(q->queues[prio]); queues 175 net/sched/sch_prio.c qdisc_put(q->queues[prio]); queues 182 net/sched/sch_prio.c struct Qdisc *queues[TCQ_PRIO_BANDS]; queues 200 net/sched/sch_prio.c queues[i] = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, queues 203 net/sched/sch_prio.c if (!queues[i]) { queues 205 net/sched/sch_prio.c qdisc_put(queues[--i]); queues 216 net/sched/sch_prio.c qdisc_tree_flush_backlog(q->queues[i]); queues 219 net/sched/sch_prio.c q->queues[i] = queues[i]; queues 220 net/sched/sch_prio.c if (q->queues[i] != &noop_qdisc) queues 221 net/sched/sch_prio.c qdisc_hash_add(q->queues[i], true); queues 227 net/sched/sch_prio.c qdisc_put(q->queues[i]); queues 304 net/sched/sch_prio.c *old = qdisc_replace(sch, new, &q->queues[band]); queues 324 net/sched/sch_prio.c return q->queues[band]; queues 353 net/sched/sch_prio.c tcm->tcm_info = q->queues[cl-1]->handle; queues 363 net/sched/sch_prio.c cl_q = q->queues[cl - 1]; queues 1176 sound/core/seq/seq_clientmgr.c info->queues = SNDRV_SEQ_MAX_QUEUES; queues 30 tools/perf/util/arm-spe.c struct auxtrace_queues queues; queues 122 tools/perf/util/arm-spe.c err = auxtrace_queues__add_event(&spe->queues, session, event, queues 158 tools/perf/util/arm-spe.c struct auxtrace_queues *queues = &spe->queues; queues 161 tools/perf/util/arm-spe.c for (i = 0; i < queues->nr_queues; i++) { queues 162 tools/perf/util/arm-spe.c arm_spe_free_queue(queues->queue_array[i].priv); queues 163 tools/perf/util/arm-spe.c queues->queue_array[i].priv = NULL; queues 165 tools/perf/util/arm-spe.c auxtrace_queues__free(queues); queues 207 tools/perf/util/arm-spe.c err = auxtrace_queues__init(&spe->queues); queues 170 tools/perf/util/auxtrace.c int auxtrace_queues__init(struct auxtrace_queues *queues) queues 172 tools/perf/util/auxtrace.c queues->nr_queues = AUXTRACE_INIT_NR_QUEUES; queues 173 tools/perf/util/auxtrace.c queues->queue_array = auxtrace_alloc_queue_array(queues->nr_queues); queues 174 tools/perf/util/auxtrace.c if (!queues->queue_array) queues 179 tools/perf/util/auxtrace.c static int auxtrace_queues__grow(struct auxtrace_queues *queues, queues 182 tools/perf/util/auxtrace.c unsigned int nr_queues = queues->nr_queues; queues 192 tools/perf/util/auxtrace.c if (nr_queues < queues->nr_queues || nr_queues < new_nr_queues) queues 199 tools/perf/util/auxtrace.c for (i = 0; i < queues->nr_queues; i++) { queues 200 tools/perf/util/auxtrace.c list_splice_tail(&queues->queue_array[i].head, queues 202 tools/perf/util/auxtrace.c queue_array[i].tid = queues->queue_array[i].tid; queues 203 tools/perf/util/auxtrace.c queue_array[i].cpu = queues->queue_array[i].cpu; queues 204 tools/perf/util/auxtrace.c queue_array[i].set = queues->queue_array[i].set; queues 205 tools/perf/util/auxtrace.c queue_array[i].priv = queues->queue_array[i].priv; queues 208 tools/perf/util/auxtrace.c queues->nr_queues = nr_queues; queues 209 tools/perf/util/auxtrace.c queues->queue_array = queue_array; queues 236 tools/perf/util/auxtrace.c static int auxtrace_queues__queue_buffer(struct auxtrace_queues *queues, queues 243 tools/perf/util/auxtrace.c if (idx >= queues->nr_queues) { queues 244 tools/perf/util/auxtrace.c err = auxtrace_queues__grow(queues, idx + 1); queues 249 tools/perf/util/auxtrace.c queue = &queues->queue_array[idx]; queues 261 tools/perf/util/auxtrace.c buffer->buffer_nr = queues->next_buffer_nr++; queues 265 tools/perf/util/auxtrace.c queues->new_data = true; queues 266 tools/perf/util/auxtrace.c queues->populated = true; queues 274 tools/perf/util/auxtrace.c static int auxtrace_queues__split_buffer(struct auxtrace_queues *queues, queues 289 tools/perf/util/auxtrace.c err = auxtrace_queues__queue_buffer(queues, idx, b); queues 312 tools/perf/util/auxtrace.c static int auxtrace_queues__add_buffer(struct auxtrace_queues *queues, queues 337 tools/perf/util/auxtrace.c err = auxtrace_queues__split_buffer(queues, idx, buffer); queues 342 tools/perf/util/auxtrace.c err = auxtrace_queues__queue_buffer(queues, idx, buffer); queues 357 tools/perf/util/auxtrace.c int auxtrace_queues__add_event(struct auxtrace_queues *queues, queues 373 tools/perf/util/auxtrace.c return auxtrace_queues__add_buffer(queues, session, idx, &buffer, queues 377 tools/perf/util/auxtrace.c static int auxtrace_queues__add_indexed_event(struct auxtrace_queues *queues, queues 397 tools/perf/util/auxtrace.c err = auxtrace_queues__add_event(queues, session, event, queues 404 tools/perf/util/auxtrace.c void auxtrace_queues__free(struct auxtrace_queues *queues) queues 408 tools/perf/util/auxtrace.c for (i = 0; i < queues->nr_queues; i++) { queues 409 tools/perf/util/auxtrace.c while (!list_empty(&queues->queue_array[i].head)) { queues 412 tools/perf/util/auxtrace.c buffer = list_entry(queues->queue_array[i].head.next, queues 419 tools/perf/util/auxtrace.c zfree(&queues->queue_array); queues 420 tools/perf/util/auxtrace.c queues->nr_queues = 0; queues 766 tools/perf/util/auxtrace.c static int auxtrace_queues__process_index_entry(struct auxtrace_queues *queues, queues 770 tools/perf/util/auxtrace.c return auxtrace_queues__add_indexed_event(queues, session, queues 774 tools/perf/util/auxtrace.c int auxtrace_queues__process_index(struct auxtrace_queues *queues, queues 788 tools/perf/util/auxtrace.c err = auxtrace_queues__process_index_entry(queues, queues 460 tools/perf/util/auxtrace.h int auxtrace_queues__init(struct auxtrace_queues *queues); queues 461 tools/perf/util/auxtrace.h int auxtrace_queues__add_event(struct auxtrace_queues *queues, queues 465 tools/perf/util/auxtrace.h void auxtrace_queues__free(struct auxtrace_queues *queues); queues 466 tools/perf/util/auxtrace.h int auxtrace_queues__process_index(struct auxtrace_queues *queues, queues 44 tools/perf/util/cs-etm.c struct auxtrace_queues queues; queues 240 tools/perf/util/cs-etm.c queue = &etmq->etm->queues.queue_array[etmq->queue_nr]; queues 583 tools/perf/util/cs-etm.c struct auxtrace_queues *queues = &aux->queues; queues 585 tools/perf/util/cs-etm.c for (i = 0; i < queues->nr_queues; i++) { queues 586 tools/perf/util/cs-etm.c cs_etm__free_queue(queues->queue_array[i].priv); queues 587 tools/perf/util/cs-etm.c queues->queue_array[i].priv = NULL; queues 590 tools/perf/util/cs-etm.c auxtrace_queues__free(queues); queues 836 tools/perf/util/cs-etm.c for (i = 0; i < etm->queues.nr_queues; i++) { queues 837 tools/perf/util/cs-etm.c ret = cs_etm__setup_queue(etm, &etm->queues.queue_array[i], i); queues 847 tools/perf/util/cs-etm.c if (etm->queues.new_data) { queues 848 tools/perf/util/cs-etm.c etm->queues.new_data = false; queues 1007 tools/perf/util/cs-etm.c queue = &etmq->etm->queues.queue_array[etmq->queue_nr]; queues 2065 tools/perf/util/cs-etm.c struct auxtrace_queues *queues = &etm->queues; queues 2067 tools/perf/util/cs-etm.c for (i = 0; i < queues->nr_queues; i++) { queues 2068 tools/perf/util/cs-etm.c struct auxtrace_queue *queue = &etm->queues.queue_array[i]; queues 2108 tools/perf/util/cs-etm.c queue = &etm->queues.queue_array[queue_nr]; queues 2317 tools/perf/util/cs-etm.c err = auxtrace_queues__add_event(&etm->queues, session, queues 2535 tools/perf/util/cs-etm.c err = auxtrace_queues__init(&etm->queues); queues 2594 tools/perf/util/cs-etm.c err = auxtrace_queues__process_index(&etm->queues, session); queues 2598 tools/perf/util/cs-etm.c etm->data_queued = etm->queues.populated; queues 2605 tools/perf/util/cs-etm.c auxtrace_queues__free(&etm->queues); queues 46 tools/perf/util/intel-bts.c struct auxtrace_queues queues; queues 211 tools/perf/util/intel-bts.c for (i = 0; i < bts->queues.nr_queues; i++) { queues 212 tools/perf/util/intel-bts.c ret = intel_bts_setup_queue(bts, &bts->queues.queue_array[i], queues 222 tools/perf/util/intel-bts.c if (bts->queues.new_data) { queues 223 tools/perf/util/intel-bts.c bts->queues.new_data = false; queues 465 tools/perf/util/intel-bts.c queue = &btsq->bts->queues.queue_array[btsq->queue_nr]; queues 539 tools/perf/util/intel-bts.c struct auxtrace_queues *queues = &bts->queues; queues 542 tools/perf/util/intel-bts.c for (i = 0; i < queues->nr_queues; i++) { queues 543 tools/perf/util/intel-bts.c struct auxtrace_queue *queue = &bts->queues.queue_array[i]; queues 568 tools/perf/util/intel-bts.c queue = &bts->queues.queue_array[queue_nr]; queues 659 tools/perf/util/intel-bts.c err = auxtrace_queues__add_event(&bts->queues, session, event, queues 710 tools/perf/util/intel-bts.c struct auxtrace_queues *queues = &bts->queues; queues 713 tools/perf/util/intel-bts.c for (i = 0; i < queues->nr_queues; i++) { queues 714 tools/perf/util/intel-bts.c intel_bts_free_queue(queues->queue_array[i].priv); queues 715 tools/perf/util/intel-bts.c queues->queue_array[i].priv = NULL; queues 717 tools/perf/util/intel-bts.c auxtrace_queues__free(queues); queues 864 tools/perf/util/intel-bts.c err = auxtrace_queues__init(&bts->queues); queues 914 tools/perf/util/intel-bts.c err = auxtrace_queues__process_index(&bts->queues, session); queues 918 tools/perf/util/intel-bts.c if (bts->queues.populated) queues 924 tools/perf/util/intel-bts.c auxtrace_queues__free(&bts->queues); queues 55 tools/perf/util/intel-pt.c struct auxtrace_queues queues; queues 312 tools/perf/util/intel-pt.c queue = &ptq->pt->queues.queue_array[ptq->queue_nr]; queues 362 tools/perf/util/intel-pt.c queue = &ptq->pt->queues.queue_array[ptq->queue_nr]; queues 1125 tools/perf/util/intel-pt.c for (i = 0; i < pt->queues.nr_queues; i++) { queues 1126 tools/perf/util/intel-pt.c ret = intel_pt_setup_queue(pt, &pt->queues.queue_array[i], i); queues 1869 tools/perf/util/intel-pt.c queue = &pt->queues.queue_array[ptq->queue_nr]; queues 2075 tools/perf/util/intel-pt.c for (i = 0; i < pt->queues.nr_queues; i++) { queues 2076 tools/perf/util/intel-pt.c struct auxtrace_queue *queue = &pt->queues.queue_array[i]; queues 2241 tools/perf/util/intel-pt.c if (pt->queues.new_data) { queues 2242 tools/perf/util/intel-pt.c pt->queues.new_data = false; queues 2265 tools/perf/util/intel-pt.c queue = &pt->queues.queue_array[queue_nr]; queues 2306 tools/perf/util/intel-pt.c struct auxtrace_queues *queues = &pt->queues; queues 2310 tools/perf/util/intel-pt.c for (i = 0; i < queues->nr_queues; i++) { queues 2311 tools/perf/util/intel-pt.c struct auxtrace_queue *queue = &pt->queues.queue_array[i]; queues 2333 tools/perf/util/intel-pt.c if (cpu < 0 || !pt->queues.nr_queues) queues 2336 tools/perf/util/intel-pt.c if ((unsigned)cpu >= pt->queues.nr_queues) queues 2337 tools/perf/util/intel-pt.c i = pt->queues.nr_queues - 1; queues 2341 tools/perf/util/intel-pt.c if (pt->queues.queue_array[i].cpu == cpu) queues 2342 tools/perf/util/intel-pt.c return pt->queues.queue_array[i].priv; queues 2345 tools/perf/util/intel-pt.c if (pt->queues.queue_array[--i].cpu == cpu) queues 2346 tools/perf/util/intel-pt.c return pt->queues.queue_array[i].priv; queues 2349 tools/perf/util/intel-pt.c for (; j < pt->queues.nr_queues; j++) { queues 2350 tools/perf/util/intel-pt.c if (pt->queues.queue_array[j].cpu == cpu) queues 2351 tools/perf/util/intel-pt.c return pt->queues.queue_array[j].priv; queues 2614 tools/perf/util/intel-pt.c struct auxtrace_queues *queues = &pt->queues; queues 2617 tools/perf/util/intel-pt.c for (i = 0; i < queues->nr_queues; i++) { queues 2618 tools/perf/util/intel-pt.c intel_pt_free_queue(queues->queue_array[i].priv); queues 2619 tools/perf/util/intel-pt.c queues->queue_array[i].priv = NULL; queues 2622 tools/perf/util/intel-pt.c auxtrace_queues__free(queues); queues 2661 tools/perf/util/intel-pt.c err = auxtrace_queues__add_event(&pt->queues, session, event, queues 3097 tools/perf/util/intel-pt.c err = auxtrace_queues__init(&pt->queues); queues 3285 tools/perf/util/intel-pt.c err = auxtrace_queues__process_index(&pt->queues, session); queues 3289 tools/perf/util/intel-pt.c if (pt->queues.populated) queues 3301 tools/perf/util/intel-pt.c auxtrace_queues__free(&pt->queues); queues 169 tools/perf/util/s390-cpumsf.c struct auxtrace_queues queues; queues 202 tools/perf/util/s390-cpumsf.c if (!sf->use_logfile || sf->queues.nr_queues <= sample->cpu) queues 205 tools/perf/util/s390-cpumsf.c q = &sf->queues.queue_array[sample->cpu]; queues 700 tools/perf/util/s390-cpumsf.c queue = &sfq->sf->queues.queue_array[sfq->queue_nr]; queues 824 tools/perf/util/s390-cpumsf.c for (i = 0; i < sf->queues.nr_queues; i++) { queues 825 tools/perf/util/s390-cpumsf.c ret = s390_cpumsf_setup_queue(sf, &sf->queues.queue_array[i], queues 835 tools/perf/util/s390-cpumsf.c if (!sf->queues.new_data) queues 838 tools/perf/util/s390-cpumsf.c sf->queues.new_data = false; queues 859 tools/perf/util/s390-cpumsf.c queue = &sf->queues.queue_array[queue_nr]; queues 984 tools/perf/util/s390-cpumsf.c err = auxtrace_queues__add_event(&sf->queues, session, event, queues 1015 tools/perf/util/s390-cpumsf.c struct auxtrace_queues *queues = &sf->queues; queues 1018 tools/perf/util/s390-cpumsf.c for (i = 0; i < queues->nr_queues; i++) { queues 1020 tools/perf/util/s390-cpumsf.c queues->queue_array[i].priv; queues 1032 tools/perf/util/s390-cpumsf.c zfree(&queues->queue_array[i].priv); queues 1034 tools/perf/util/s390-cpumsf.c auxtrace_queues__free(queues); queues 1130 tools/perf/util/s390-cpumsf.c err = auxtrace_queues__init(&sf->queues); queues 1150 tools/perf/util/s390-cpumsf.c err = auxtrace_queues__process_index(&sf->queues, session); queues 1154 tools/perf/util/s390-cpumsf.c if (sf->queues.populated) queues 1160 tools/perf/util/s390-cpumsf.c auxtrace_queues__free(&sf->queues);