qmap 290 arch/powerpc/kvm/book3s_xive.c if (xive->qmap & (1 << prio)) qmap 309 arch/powerpc/kvm/book3s_xive.c xive->qmap |= (1 << prio); qmap 1307 arch/powerpc/kvm/book3s_xive.c if (xive->qmap & (1 << i)) { qmap 129 arch/powerpc/kvm/book3s_xive.h u8 qmap; qmap 18 block/blk-mq-cpumap.c static int queue_index(struct blk_mq_queue_map *qmap, qmap 21 block/blk-mq-cpumap.c return qmap->queue_offset + (q % nr_queues); qmap 35 block/blk-mq-cpumap.c int blk_mq_map_queues(struct blk_mq_queue_map *qmap) qmap 37 block/blk-mq-cpumap.c unsigned int *map = qmap->mq_map; qmap 38 block/blk-mq-cpumap.c unsigned int nr_queues = qmap->nr_queues; qmap 51 block/blk-mq-cpumap.c map[cpu] = queue_index(qmap, nr_queues, q++); qmap 64 block/blk-mq-cpumap.c map[cpu] = queue_index(qmap, nr_queues, q++); qmap 68 block/blk-mq-cpumap.c map[cpu] = queue_index(qmap, nr_queues, q++); qmap 86 block/blk-mq-cpumap.c int blk_mq_hw_queue_to_node(struct blk_mq_queue_map *qmap, unsigned int index) qmap 91 block/blk-mq-cpumap.c if (index == qmap->mq_map[i]) qmap 26 block/blk-mq-pci.c int blk_mq_pci_map_queues(struct blk_mq_queue_map *qmap, struct pci_dev *pdev, qmap 32 block/blk-mq-pci.c for (queue = 0; queue < qmap->nr_queues; queue++) { qmap 38 block/blk-mq-pci.c qmap->mq_map[cpu] = qmap->queue_offset + queue; qmap 44 block/blk-mq-pci.c WARN_ON_ONCE(qmap->nr_queues > 1); qmap 45 block/blk-mq-pci.c blk_mq_clear_mq_map(qmap); qmap 24 block/blk-mq-virtio.c int blk_mq_virtio_map_queues(struct blk_mq_queue_map *qmap, qmap 33 block/blk-mq-virtio.c for (queue = 0; queue < qmap->nr_queues; queue++) { qmap 39 block/blk-mq-virtio.c qmap->mq_map[cpu] = qmap->queue_offset + queue; qmap 44 block/blk-mq-virtio.c return blk_mq_map_queues(qmap); qmap 82 block/blk-mq.h extern int blk_mq_hw_queue_to_node(struct blk_mq_queue_map *qmap, unsigned int); qmap 229 block/blk-mq.h static inline void blk_mq_clear_mq_map(struct blk_mq_queue_map *qmap) qmap 234 block/blk-mq.h qmap->mq_map[cpu] = 0; qmap 178 drivers/net/ethernet/broadcom/bnxt/bnxt_dcb.c unsigned long qmap = 0; qmap 187 drivers/net/ethernet/broadcom/bnxt/bnxt_dcb.c __set_bit(j, &qmap); qmap 199 drivers/net/ethernet/broadcom/bnxt/bnxt_dcb.c j = find_next_zero_bit(&qmap, max, j); qmap 201 drivers/net/ethernet/broadcom/bnxt/bnxt_dcb.c __set_bit(j, &qmap); qmap 1678 drivers/net/ethernet/intel/i40e/i40e_main.c u16 qcount = 0, max_qcount, qmap, sections = 0; qmap 1694 drivers/net/ethernet/intel/i40e/i40e_main.c qmap = (offset << I40E_AQ_VSI_TC_QUE_OFFSET_SHIFT) | qmap 1724 drivers/net/ethernet/intel/i40e/i40e_main.c ctxt->info.tc_mapping[0] = cpu_to_le16(qmap); qmap 1773 drivers/net/ethernet/intel/i40e/i40e_main.c u16 qmap; qmap 1841 drivers/net/ethernet/intel/i40e/i40e_main.c qmap = qmap 1855 drivers/net/ethernet/intel/i40e/i40e_main.c qmap = 0; qmap 1857 drivers/net/ethernet/intel/i40e/i40e_main.c ctxt->info.tc_mapping[i] = cpu_to_le16(qmap); qmap 5800 drivers/net/ethernet/intel/i40e/i40e_main.c u16 qcount, qmap, sections = 0; qmap 5815 drivers/net/ethernet/intel/i40e/i40e_main.c qmap = (offset << I40E_AQ_VSI_TC_QUE_OFFSET_SHIFT) | qmap 5819 drivers/net/ethernet/intel/i40e/i40e_main.c ctxt->info.tc_mapping[0] = cpu_to_le16(qmap); qmap 7099 drivers/net/ethernet/intel/i40e/i40e_main.c u16 sections, qmap, num_qps; qmap 7112 drivers/net/ethernet/intel/i40e/i40e_main.c qmap = (offset << I40E_AQ_VSI_TC_QUE_OFFSET_SHIFT) | qmap 7124 drivers/net/ethernet/intel/i40e/i40e_main.c ctxt.info.tc_mapping[0] = cpu_to_le16(qmap); qmap 870 drivers/net/ethernet/intel/ice/ice_lib.c u16 offset = 0, qmap = 0, tx_count = 0; qmap 945 drivers/net/ethernet/intel/ice/ice_lib.c qmap = ((offset << ICE_AQ_VSI_TC_Q_OFFSET_S) & qmap 951 drivers/net/ethernet/intel/ice/ice_lib.c ctxt->info.tc_mapping[i] = cpu_to_le16(qmap); qmap 2135 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c unsigned long qmap; qmap 2187 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c qmap = map->rxq_map; qmap 2189 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c for_each_set_bit(vsi_q_id, &qmap, ICE_MAX_BASE_QS_PER_VF) { qmap 2201 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c qmap = map->txq_map; qmap 2203 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c for_each_set_bit(vsi_q_id, &qmap, ICE_MAX_BASE_QS_PER_VF) { qmap 1784 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c u32 qmap; qmap 1789 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c qmap = 0; qmap 1794 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c qmap |= (1 << queue); qmap 1798 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_write(port->priv, MVPP2_TXP_SCHED_Q_CMD_REG, qmap); qmap 7123 drivers/scsi/qla2xxx/qla_os.c struct blk_mq_queue_map *qmap = &shost->tag_set.map[HCTX_TYPE_DEFAULT]; qmap 7126 drivers/scsi/qla2xxx/qla_os.c rc = blk_mq_map_queues(qmap); qmap 7128 drivers/scsi/qla2xxx/qla_os.c rc = blk_mq_pci_map_queues(qmap, vha->hw->pdev, vha->irq_offset); qmap 706 drivers/scsi/virtio_scsi.c struct blk_mq_queue_map *qmap = &shost->tag_set.map[HCTX_TYPE_DEFAULT]; qmap 708 drivers/scsi/virtio_scsi.c return blk_mq_virtio_map_queues(qmap, vscsi->vdev, 2); qmap 8 include/linux/blk-mq-pci.h int blk_mq_pci_map_queues(struct blk_mq_queue_map *qmap, struct pci_dev *pdev, qmap 8 include/linux/blk-mq-virtio.h int blk_mq_virtio_map_queues(struct blk_mq_queue_map *qmap,