queues_per_pool 831 drivers/net/ethernet/intel/fm10k/fm10k_pf.c u16 qmap_stride, queues_per_pool, vf_q_idx, timeout, qmap_idx, i; queues_per_pool 842 drivers/net/ethernet/intel/fm10k/fm10k_pf.c queues_per_pool = fm10k_queues_per_pool(hw); queues_per_pool 873 drivers/net/ethernet/intel/fm10k/fm10k_pf.c for (i = 0; i < queues_per_pool; i++) queues_per_pool 946 drivers/net/ethernet/intel/fm10k/fm10k_pf.c u16 qmap_stride, queues_per_pool, vf_q_idx, qmap_idx; queues_per_pool 970 drivers/net/ethernet/intel/fm10k/fm10k_pf.c queues_per_pool = fm10k_queues_per_pool(hw); queues_per_pool 995 drivers/net/ethernet/intel/fm10k/fm10k_pf.c for (i = vf_q_idx; i < (queues_per_pool + vf_q_idx); i++) { queues_per_pool 1049 drivers/net/ethernet/intel/fm10k/fm10k_pf.c for (i = queues_per_pool; i--;) { queues_per_pool 1063 drivers/net/ethernet/intel/fm10k/fm10k_pf.c for (i = queues_per_pool; i < qmap_stride; i++) { queues_per_pool 66 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c u16 queues_per_pool = __ALIGN_MASK(1, ~vmdq->mask); queues_per_pool 69 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c reg_idx = (vmdq->offset + vmdq->indices) * queues_per_pool; queues_per_pool 77 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c reg_idx = (vmdq->offset + vmdq->indices) * queues_per_pool; queues_per_pool 1520 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c u16 queue, queues_per_pool; queues_per_pool 1556 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c queues_per_pool = __ALIGN_MASK(1, ~vmdq->mask); queues_per_pool 1559 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c for (queue = 0; queue < queues_per_pool; queue++) { queues_per_pool 1560 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c unsigned int reg_idx = (vf * queues_per_pool) + queue;