qs_cfg            957 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 	struct qs_cfg *qs_cfg;
qs_cfg            974 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 	qs_cfg = (struct qs_cfg *)&mbx.qs.cfg;
qs_cfg            976 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 		qs_cfg->ena = 1;
qs_cfg            978 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 		qs_cfg->be = 1;
qs_cfg            980 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 		qs_cfg->vnic = qs->vnic_id;
qs_cfg            983 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 			qs_cfg->send_tstmp_ena = 1;
qs_cfg            614 drivers/net/ethernet/intel/ice/ice_lib.c static int __ice_vsi_get_qs_contig(struct ice_qs_cfg *qs_cfg)
qs_cfg            618 drivers/net/ethernet/intel/ice/ice_lib.c 	mutex_lock(qs_cfg->qs_mutex);
qs_cfg            619 drivers/net/ethernet/intel/ice/ice_lib.c 	offset = bitmap_find_next_zero_area(qs_cfg->pf_map, qs_cfg->pf_map_size,
qs_cfg            620 drivers/net/ethernet/intel/ice/ice_lib.c 					    0, qs_cfg->q_count, 0);
qs_cfg            621 drivers/net/ethernet/intel/ice/ice_lib.c 	if (offset >= qs_cfg->pf_map_size) {
qs_cfg            622 drivers/net/ethernet/intel/ice/ice_lib.c 		mutex_unlock(qs_cfg->qs_mutex);
qs_cfg            626 drivers/net/ethernet/intel/ice/ice_lib.c 	bitmap_set(qs_cfg->pf_map, offset, qs_cfg->q_count);
qs_cfg            627 drivers/net/ethernet/intel/ice/ice_lib.c 	for (i = 0; i < qs_cfg->q_count; i++)
qs_cfg            628 drivers/net/ethernet/intel/ice/ice_lib.c 		qs_cfg->vsi_map[i + qs_cfg->vsi_map_offset] = i + offset;
qs_cfg            629 drivers/net/ethernet/intel/ice/ice_lib.c 	mutex_unlock(qs_cfg->qs_mutex);
qs_cfg            640 drivers/net/ethernet/intel/ice/ice_lib.c static int __ice_vsi_get_qs_sc(struct ice_qs_cfg *qs_cfg)
qs_cfg            644 drivers/net/ethernet/intel/ice/ice_lib.c 	mutex_lock(qs_cfg->qs_mutex);
qs_cfg            645 drivers/net/ethernet/intel/ice/ice_lib.c 	for (i = 0; i < qs_cfg->q_count; i++) {
qs_cfg            646 drivers/net/ethernet/intel/ice/ice_lib.c 		index = find_next_zero_bit(qs_cfg->pf_map,
qs_cfg            647 drivers/net/ethernet/intel/ice/ice_lib.c 					   qs_cfg->pf_map_size, index);
qs_cfg            648 drivers/net/ethernet/intel/ice/ice_lib.c 		if (index >= qs_cfg->pf_map_size)
qs_cfg            650 drivers/net/ethernet/intel/ice/ice_lib.c 		set_bit(index, qs_cfg->pf_map);
qs_cfg            651 drivers/net/ethernet/intel/ice/ice_lib.c 		qs_cfg->vsi_map[i + qs_cfg->vsi_map_offset] = index;
qs_cfg            653 drivers/net/ethernet/intel/ice/ice_lib.c 	mutex_unlock(qs_cfg->qs_mutex);
qs_cfg            658 drivers/net/ethernet/intel/ice/ice_lib.c 		clear_bit(qs_cfg->vsi_map[index], qs_cfg->pf_map);
qs_cfg            659 drivers/net/ethernet/intel/ice/ice_lib.c 		qs_cfg->vsi_map[index + qs_cfg->vsi_map_offset] = 0;
qs_cfg            661 drivers/net/ethernet/intel/ice/ice_lib.c 	mutex_unlock(qs_cfg->qs_mutex);
qs_cfg            675 drivers/net/ethernet/intel/ice/ice_lib.c static int __ice_vsi_get_qs(struct ice_qs_cfg *qs_cfg)
qs_cfg            679 drivers/net/ethernet/intel/ice/ice_lib.c 	ret = __ice_vsi_get_qs_contig(qs_cfg);
qs_cfg            682 drivers/net/ethernet/intel/ice/ice_lib.c 		qs_cfg->mapping_mode = ICE_VSI_MAP_SCATTER;
qs_cfg            683 drivers/net/ethernet/intel/ice/ice_lib.c 		qs_cfg->q_count = min_t(u16, qs_cfg->q_count,
qs_cfg            684 drivers/net/ethernet/intel/ice/ice_lib.c 					qs_cfg->scatter_count);
qs_cfg            685 drivers/net/ethernet/intel/ice/ice_lib.c 		ret = __ice_vsi_get_qs_sc(qs_cfg);