queues_per_pool   831 drivers/net/ethernet/intel/fm10k/fm10k_pf.c 	u16 qmap_stride, queues_per_pool, vf_q_idx, timeout, qmap_idx, i;
queues_per_pool   842 drivers/net/ethernet/intel/fm10k/fm10k_pf.c 	queues_per_pool = fm10k_queues_per_pool(hw);
queues_per_pool   873 drivers/net/ethernet/intel/fm10k/fm10k_pf.c 	for (i = 0; i < queues_per_pool; i++)
queues_per_pool   946 drivers/net/ethernet/intel/fm10k/fm10k_pf.c 	u16 qmap_stride, queues_per_pool, vf_q_idx, qmap_idx;
queues_per_pool   970 drivers/net/ethernet/intel/fm10k/fm10k_pf.c 	queues_per_pool = fm10k_queues_per_pool(hw);
queues_per_pool   995 drivers/net/ethernet/intel/fm10k/fm10k_pf.c 	for (i = vf_q_idx; i < (queues_per_pool + vf_q_idx); i++) {
queues_per_pool  1049 drivers/net/ethernet/intel/fm10k/fm10k_pf.c 	for (i = queues_per_pool; i--;) {
queues_per_pool  1063 drivers/net/ethernet/intel/fm10k/fm10k_pf.c 	for (i = queues_per_pool; i < qmap_stride; i++) {
queues_per_pool    66 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c 		u16 queues_per_pool = __ALIGN_MASK(1, ~vmdq->mask);
queues_per_pool    69 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c 		reg_idx = (vmdq->offset + vmdq->indices) * queues_per_pool;
queues_per_pool    77 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c 		reg_idx = (vmdq->offset + vmdq->indices) * queues_per_pool;
queues_per_pool  1520 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c 	u16 queue, queues_per_pool;
queues_per_pool  1556 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c 	queues_per_pool = __ALIGN_MASK(1, ~vmdq->mask);
queues_per_pool  1559 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c 	for (queue = 0; queue < queues_per_pool; queue++) {
queues_per_pool  1560 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c 		unsigned int reg_idx = (vf * queues_per_pool) + queue;