sriov              86 drivers/infiniband/hw/mlx4/alias_GUID.c 	guid_indexes = be64_to_cpu((__force __be64) dev->sriov.alias_guid.
sriov             102 drivers/infiniband/hw/mlx4/alias_GUID.c 			memcpy(&dev->sriov.demux[port_index].guid_cache[slave_id],
sriov             117 drivers/infiniband/hw/mlx4/alias_GUID.c 	return *(__be64 *)&dev->sriov.demux[port - 1].guid_cache[index];
sriov             136 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags);
sriov             137 drivers/infiniband/hw/mlx4/alias_GUID.c 	if (dev->sriov.alias_guid.ports_guid[port_index].state_flags &
sriov             141 drivers/infiniband/hw/mlx4/alias_GUID.c 		curr_guid = *(__be64 *)&dev->sriov.
sriov             154 drivers/infiniband/hw/mlx4/alias_GUID.c 	*(__be64 *)&dev->sriov.alias_guid.ports_guid[port_index].
sriov             157 drivers/infiniband/hw/mlx4/alias_GUID.c 	dev->sriov.alias_guid.ports_guid[port_index].
sriov             160 drivers/infiniband/hw/mlx4/alias_GUID.c 	dev->sriov.alias_guid.ports_guid[port_index].
sriov             164 drivers/infiniband/hw/mlx4/alias_GUID.c 	dev->sriov.alias_guid.ports_guid[port_index].
sriov             166 drivers/infiniband/hw/mlx4/alias_GUID.c 	dev->sriov.alias_guid.ports_guid[port_index].
sriov             171 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags);
sriov             204 drivers/infiniband/hw/mlx4/alias_GUID.c 	rec = &dev->sriov.alias_guid.ports_guid[port_num - 1].
sriov             206 drivers/infiniband/hw/mlx4/alias_GUID.c 	guid_indexes = be64_to_cpu((__force __be64) dev->sriov.alias_guid.
sriov             236 drivers/infiniband/hw/mlx4/alias_GUID.c 		spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags);
sriov             248 drivers/infiniband/hw/mlx4/alias_GUID.c 				spin_unlock_irqrestore(&dev->sriov.
sriov             253 drivers/infiniband/hw/mlx4/alias_GUID.c 		spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock,
sriov             307 drivers/infiniband/hw/mlx4/alias_GUID.c 	rec = &dev->sriov.alias_guid.ports_guid[port_index].
sriov             327 drivers/infiniband/hw/mlx4/alias_GUID.c 	rec = &dev->sriov.alias_guid.ports_guid[port_index].
sriov             330 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags);
sriov             424 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags);
sriov             434 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_lock_irqsave(&dev->sriov.going_down_lock, flags);
sriov             435 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1);
sriov             436 drivers/infiniband/hw/mlx4/alias_GUID.c 	if (!dev->sriov.is_going_down) {
sriov             438 drivers/infiniband/hw/mlx4/alias_GUID.c 		queue_delayed_work(dev->sriov.alias_guid.ports_guid[port_index].wq,
sriov             439 drivers/infiniband/hw/mlx4/alias_GUID.c 				   &dev->sriov.alias_guid.ports_guid[port_index].
sriov             448 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1);
sriov             449 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_unlock_irqrestore(&dev->sriov.going_down_lock, flags);
sriov             458 drivers/infiniband/hw/mlx4/alias_GUID.c 	dev->sriov.alias_guid.ports_guid[port - 1].all_rec_per_port[index].status
sriov             464 drivers/infiniband/hw/mlx4/alias_GUID.c 			*(u64 *)&dev->sriov.alias_guid.ports_guid[port - 1].
sriov             477 drivers/infiniband/hw/mlx4/alias_GUID.c 	dev->sriov.alias_guid.ports_guid[port - 1].
sriov             479 drivers/infiniband/hw/mlx4/alias_GUID.c 	if (dev->sriov.alias_guid.ports_guid[port - 1].
sriov             481 drivers/infiniband/hw/mlx4/alias_GUID.c 		dev->sriov.alias_guid.ports_guid[port - 1].
sriov             500 drivers/infiniband/hw/mlx4/alias_GUID.c 		&dev->sriov.alias_guid.ports_guid[port - 1].cb_list;
sriov             540 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1);
sriov             542 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1);
sriov             545 drivers/infiniband/hw/mlx4/alias_GUID.c 		ib_sa_guid_info_rec_query(dev->sriov.alias_guid.sa_client,
sriov             555 drivers/infiniband/hw/mlx4/alias_GUID.c 		spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1);
sriov             558 drivers/infiniband/hw/mlx4/alias_GUID.c 		spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1);
sriov             567 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_lock_irqsave(&dev->sriov.going_down_lock, flags);
sriov             568 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1);
sriov             570 drivers/infiniband/hw/mlx4/alias_GUID.c 	if (!dev->sriov.is_going_down) {
sriov             571 drivers/infiniband/hw/mlx4/alias_GUID.c 		queue_delayed_work(dev->sriov.alias_guid.ports_guid[port - 1].wq,
sriov             572 drivers/infiniband/hw/mlx4/alias_GUID.c 				   &dev->sriov.alias_guid.ports_guid[port - 1].alias_guid_work,
sriov             575 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1);
sriov             576 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_unlock_irqrestore(&dev->sriov.going_down_lock, flags);
sriov             596 drivers/infiniband/hw/mlx4/alias_GUID.c 			*(__be64 *)&dev->sriov.alias_guid.ports_guid[port - 1].
sriov             613 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_lock_irqsave(&dev->sriov.going_down_lock, flags);
sriov             614 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1);
sriov             616 drivers/infiniband/hw/mlx4/alias_GUID.c 	if (dev->sriov.alias_guid.ports_guid[port - 1].state_flags &
sriov             619 drivers/infiniband/hw/mlx4/alias_GUID.c 		dev->sriov.alias_guid.ports_guid[port - 1].state_flags &=
sriov             625 drivers/infiniband/hw/mlx4/alias_GUID.c 	if (mlx4_is_master(dev->dev) && !dev->sriov.is_going_down) {
sriov             631 drivers/infiniband/hw/mlx4/alias_GUID.c 		cancel_delayed_work(&dev->sriov.alias_guid.
sriov             633 drivers/infiniband/hw/mlx4/alias_GUID.c 		queue_delayed_work(dev->sriov.alias_guid.ports_guid[port - 1].wq,
sriov             634 drivers/infiniband/hw/mlx4/alias_GUID.c 				   &dev->sriov.alias_guid.ports_guid[port - 1].alias_guid_work,
sriov             637 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1);
sriov             638 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_unlock_irqrestore(&dev->sriov.going_down_lock, flags);
sriov             651 drivers/infiniband/hw/mlx4/alias_GUID.c 			&dev->sriov.alias_guid.ports_guid[port].
sriov             700 drivers/infiniband/hw/mlx4/alias_GUID.c 		rec = dev->sriov.alias_guid.ports_guid[port].
sriov             730 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags);
sriov             740 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags);
sriov             756 drivers/infiniband/hw/mlx4/alias_GUID.c 	struct mlx4_ib_dev *dev = container_of(ib_sriov, struct mlx4_ib_dev, sriov);
sriov             781 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_lock_irqsave(&dev->sriov.going_down_lock, flags);
sriov             782 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1);
sriov             783 drivers/infiniband/hw/mlx4/alias_GUID.c 	if (!dev->sriov.is_going_down) {
sriov             788 drivers/infiniband/hw/mlx4/alias_GUID.c 		cancel_delayed_work(&dev->sriov.alias_guid.ports_guid[port].
sriov             790 drivers/infiniband/hw/mlx4/alias_GUID.c 		queue_delayed_work(dev->sriov.alias_guid.ports_guid[port].wq,
sriov             791 drivers/infiniband/hw/mlx4/alias_GUID.c 			   &dev->sriov.alias_guid.ports_guid[port].alias_guid_work, 0);
sriov             793 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1);
sriov             794 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_unlock_irqrestore(&dev->sriov.going_down_lock, flags);
sriov             800 drivers/infiniband/hw/mlx4/alias_GUID.c 	struct mlx4_ib_sriov *sriov = &dev->sriov;
sriov             807 drivers/infiniband/hw/mlx4/alias_GUID.c 		det = &sriov->alias_guid.ports_guid[i];
sriov             809 drivers/infiniband/hw/mlx4/alias_GUID.c 		spin_lock_irqsave(&sriov->alias_guid.ag_work_lock, flags);
sriov             817 drivers/infiniband/hw/mlx4/alias_GUID.c 			spin_unlock_irqrestore(&sriov->alias_guid.ag_work_lock, flags);
sriov             821 drivers/infiniband/hw/mlx4/alias_GUID.c 			spin_lock_irqsave(&sriov->alias_guid.ag_work_lock, flags);
sriov             823 drivers/infiniband/hw/mlx4/alias_GUID.c 		spin_unlock_irqrestore(&sriov->alias_guid.ag_work_lock, flags);
sriov             826 drivers/infiniband/hw/mlx4/alias_GUID.c 		flush_workqueue(dev->sriov.alias_guid.ports_guid[i].wq);
sriov             827 drivers/infiniband/hw/mlx4/alias_GUID.c 		destroy_workqueue(dev->sriov.alias_guid.ports_guid[i].wq);
sriov             829 drivers/infiniband/hw/mlx4/alias_GUID.c 	ib_sa_unregister_client(dev->sriov.alias_guid.sa_client);
sriov             830 drivers/infiniband/hw/mlx4/alias_GUID.c 	kfree(dev->sriov.alias_guid.sa_client);
sriov             842 drivers/infiniband/hw/mlx4/alias_GUID.c 	dev->sriov.alias_guid.sa_client =
sriov             843 drivers/infiniband/hw/mlx4/alias_GUID.c 		kzalloc(sizeof *dev->sriov.alias_guid.sa_client, GFP_KERNEL);
sriov             844 drivers/infiniband/hw/mlx4/alias_GUID.c 	if (!dev->sriov.alias_guid.sa_client)
sriov             847 drivers/infiniband/hw/mlx4/alias_GUID.c 	ib_sa_register_client(dev->sriov.alias_guid.sa_client);
sriov             849 drivers/infiniband/hw/mlx4/alias_GUID.c 	spin_lock_init(&dev->sriov.alias_guid.ag_work_lock);
sriov             859 drivers/infiniband/hw/mlx4/alias_GUID.c 		memset(&dev->sriov.alias_guid.ports_guid[i], 0,
sriov             861 drivers/infiniband/hw/mlx4/alias_GUID.c 		dev->sriov.alias_guid.ports_guid[i].state_flags |=
sriov             865 drivers/infiniband/hw/mlx4/alias_GUID.c 			memset(dev->sriov.alias_guid.ports_guid[i].
sriov             867 drivers/infiniband/hw/mlx4/alias_GUID.c 				sizeof(dev->sriov.alias_guid.ports_guid[i].
sriov             870 drivers/infiniband/hw/mlx4/alias_GUID.c 		INIT_LIST_HEAD(&dev->sriov.alias_guid.ports_guid[i].cb_list);
sriov             878 drivers/infiniband/hw/mlx4/alias_GUID.c 		dev->sriov.alias_guid.ports_guid[i].parent = &dev->sriov.alias_guid;
sriov             879 drivers/infiniband/hw/mlx4/alias_GUID.c 		dev->sriov.alias_guid.ports_guid[i].port  = i;
sriov             882 drivers/infiniband/hw/mlx4/alias_GUID.c 		dev->sriov.alias_guid.ports_guid[i].wq =
sriov             884 drivers/infiniband/hw/mlx4/alias_GUID.c 		if (!dev->sriov.alias_guid.ports_guid[i].wq) {
sriov             888 drivers/infiniband/hw/mlx4/alias_GUID.c 		INIT_DELAYED_WORK(&dev->sriov.alias_guid.ports_guid[i].alias_guid_work,
sriov             895 drivers/infiniband/hw/mlx4/alias_GUID.c 		destroy_workqueue(dev->sriov.alias_guid.ports_guid[i].wq);
sriov             896 drivers/infiniband/hw/mlx4/alias_GUID.c 		dev->sriov.alias_guid.ports_guid[i].wq = NULL;
sriov             900 drivers/infiniband/hw/mlx4/alias_GUID.c 	ib_sa_unregister_client(dev->sriov.alias_guid.sa_client);
sriov             901 drivers/infiniband/hw/mlx4/alias_GUID.c 	kfree(dev->sriov.alias_guid.sa_client);
sriov             902 drivers/infiniband/hw/mlx4/alias_GUID.c 	dev->sriov.alias_guid.sa_client = NULL;
sriov             146 drivers/infiniband/hw/mlx4/cm.c 	struct rb_root *sl_id_map = &to_mdev(ibdev)->sriov.sl_id_map;
sriov             173 drivers/infiniband/hw/mlx4/cm.c 	struct mlx4_ib_sriov *sriov = &dev->sriov;
sriov             174 drivers/infiniband/hw/mlx4/cm.c 	struct rb_root *sl_id_map = &sriov->sl_id_map;
sriov             176 drivers/infiniband/hw/mlx4/cm.c 	spin_lock(&sriov->id_map_lock);
sriov             177 drivers/infiniband/hw/mlx4/cm.c 	if (!xa_erase(&sriov->pv_id_table, ent->pv_cm_id))
sriov             185 drivers/infiniband/hw/mlx4/cm.c 	spin_unlock(&sriov->id_map_lock);
sriov             191 drivers/infiniband/hw/mlx4/cm.c 	struct rb_root *sl_id_map = &to_mdev(ibdev)->sriov.sl_id_map;
sriov             226 drivers/infiniband/hw/mlx4/cm.c 	struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov;
sriov             238 drivers/infiniband/hw/mlx4/cm.c 	ret = xa_alloc_cyclic(&sriov->pv_id_table, &ent->pv_cm_id, ent,
sriov             239 drivers/infiniband/hw/mlx4/cm.c 			xa_limit_32b, &sriov->pv_id_next, GFP_KERNEL);
sriov             241 drivers/infiniband/hw/mlx4/cm.c 		spin_lock(&sriov->id_map_lock);
sriov             243 drivers/infiniband/hw/mlx4/cm.c 		list_add_tail(&ent->list, &sriov->cm_list);
sriov             244 drivers/infiniband/hw/mlx4/cm.c 		spin_unlock(&sriov->id_map_lock);
sriov             258 drivers/infiniband/hw/mlx4/cm.c 	struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov;
sriov             260 drivers/infiniband/hw/mlx4/cm.c 	spin_lock(&sriov->id_map_lock);
sriov             266 drivers/infiniband/hw/mlx4/cm.c 		ent = xa_load(&sriov->pv_id_table, *pv_cm_id);
sriov             267 drivers/infiniband/hw/mlx4/cm.c 	spin_unlock(&sriov->id_map_lock);
sriov             274 drivers/infiniband/hw/mlx4/cm.c 	struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov;
sriov             277 drivers/infiniband/hw/mlx4/cm.c 	spin_lock(&sriov->id_map_lock);
sriov             278 drivers/infiniband/hw/mlx4/cm.c 	spin_lock_irqsave(&sriov->going_down_lock, flags);
sriov             280 drivers/infiniband/hw/mlx4/cm.c 	if (!sriov->is_going_down && !id->scheduled_delete) {
sriov             284 drivers/infiniband/hw/mlx4/cm.c 	spin_unlock_irqrestore(&sriov->going_down_lock, flags);
sriov             285 drivers/infiniband/hw/mlx4/cm.c 	spin_unlock(&sriov->id_map_lock);
sriov             374 drivers/infiniband/hw/mlx4/cm.c 	spin_lock_init(&dev->sriov.id_map_lock);
sriov             375 drivers/infiniband/hw/mlx4/cm.c 	INIT_LIST_HEAD(&dev->sriov.cm_list);
sriov             376 drivers/infiniband/hw/mlx4/cm.c 	dev->sriov.sl_id_map = RB_ROOT;
sriov             377 drivers/infiniband/hw/mlx4/cm.c 	xa_init_flags(&dev->sriov.pv_id_table, XA_FLAGS_ALLOC);
sriov             384 drivers/infiniband/hw/mlx4/cm.c 	struct mlx4_ib_sriov *sriov = &dev->sriov;
sriov             385 drivers/infiniband/hw/mlx4/cm.c 	struct rb_root *sl_id_map = &sriov->sl_id_map;
sriov             392 drivers/infiniband/hw/mlx4/cm.c 	spin_lock(&sriov->id_map_lock);
sriov             393 drivers/infiniband/hw/mlx4/cm.c 	list_for_each_entry_safe(map, tmp_map, &dev->sriov.cm_list, list) {
sriov             400 drivers/infiniband/hw/mlx4/cm.c 	spin_unlock(&sriov->id_map_lock);
sriov             406 drivers/infiniband/hw/mlx4/cm.c 	spin_lock(&sriov->id_map_lock);
sriov             414 drivers/infiniband/hw/mlx4/cm.c 			xa_erase(&sriov->pv_id_table, ent->pv_cm_id);
sriov             416 drivers/infiniband/hw/mlx4/cm.c 		list_splice_init(&dev->sriov.cm_list, &lh);
sriov             430 drivers/infiniband/hw/mlx4/cm.c 			xa_erase(&sriov->pv_id_table, map->pv_cm_id);
sriov             434 drivers/infiniband/hw/mlx4/cm.c 		list_for_each_entry_safe(map, tmp_map, &dev->sriov.cm_list, list) {
sriov             440 drivers/infiniband/hw/mlx4/cm.c 	spin_unlock(&sriov->id_map_lock);
sriov             284 drivers/infiniband/hw/mlx4/mad.c 				if (!dev->sriov.is_going_down)
sriov             299 drivers/infiniband/hw/mlx4/mad.c 			    !dev->sriov.is_going_down) {
sriov             438 drivers/infiniband/hw/mlx4/mad.c 		if (dev->sriov.demux[port - 1].guid_cache[i] == guid)
sriov             526 drivers/infiniband/hw/mlx4/mad.c 	tun_ctx = dev->sriov.demux[port-1].tun[slave];
sriov             739 drivers/infiniband/hw/mlx4/mad.c 			atomic64_read(&dev->sriov.demux[port - 1].subnet_prefix))) {
sriov            1090 drivers/infiniband/hw/mlx4/mad.c 	if (mlx4_is_master(dev->dev) && !dev->sriov.is_going_down)
sriov            1101 drivers/infiniband/hw/mlx4/mad.c 		if (!dev->sriov.is_going_down) {
sriov            1102 drivers/infiniband/hw/mlx4/mad.c 			mlx4_ib_mcg_port_cleanup(&dev->sriov.demux[port_num - 1], 0);
sriov            1228 drivers/infiniband/hw/mlx4/mad.c 						 (u64)atomic64_read(&dev->sriov.demux[port - 1].subnet_prefix),
sriov            1230 drivers/infiniband/hw/mlx4/mad.c 					atomic64_set(&dev->sriov.demux[port - 1].subnet_prefix,
sriov            1247 drivers/infiniband/hw/mlx4/mad.c 		if (mlx4_is_master(dev->dev) && !dev->sriov.is_going_down)
sriov            1255 drivers/infiniband/hw/mlx4/mad.c 		else if (!dev->sriov.is_going_down) {
sriov            1304 drivers/infiniband/hw/mlx4/mad.c 	spin_lock_irqsave(&dev->sriov.going_down_lock, flags);
sriov            1305 drivers/infiniband/hw/mlx4/mad.c 	if (!dev->sriov.is_going_down && ctx->state == DEMUX_PV_STATE_ACTIVE)
sriov            1307 drivers/infiniband/hw/mlx4/mad.c 	spin_unlock_irqrestore(&dev->sriov.going_down_lock, flags);
sriov            1378 drivers/infiniband/hw/mlx4/mad.c 	sqp_ctx = dev->sriov.sqps[port-1];
sriov            1972 drivers/infiniband/hw/mlx4/mad.c 	if (dev->sriov.demux[port - 1].tun[slave]) {
sriov            1973 drivers/infiniband/hw/mlx4/mad.c 		kfree(dev->sriov.demux[port - 1].tun[slave]);
sriov            1974 drivers/infiniband/hw/mlx4/mad.c 		dev->sriov.demux[port - 1].tun[slave] = NULL;
sriov            2048 drivers/infiniband/hw/mlx4/mad.c 	ctx->wq = to_mdev(ibdev)->sriov.demux[port - 1].wq;
sriov            2119 drivers/infiniband/hw/mlx4/mad.c 		clean_vf_mcast(&dev->sriov.demux[port - 1], slave);
sriov            2123 drivers/infiniband/hw/mlx4/mad.c 					     dev->sriov.sqps[port - 1], 1);
sriov            2126 drivers/infiniband/hw/mlx4/mad.c 				     dev->sriov.demux[port - 1].tun[slave], 1);
sriov            2132 drivers/infiniband/hw/mlx4/mad.c 				  dev->sriov.demux[port - 1].tun[slave]);
sriov            2137 drivers/infiniband/hw/mlx4/mad.c 					  dev->sriov.sqps[port - 1]);
sriov            2288 drivers/infiniband/hw/mlx4/mad.c 	dev->sriov.is_going_down = 0;
sriov            2289 drivers/infiniband/hw/mlx4/mad.c 	spin_lock_init(&dev->sriov.going_down_lock);
sriov            2324 drivers/infiniband/hw/mlx4/mad.c 		dev->sriov.demux[i].guid_cache[0] = gid.global.interface_id;
sriov            2325 drivers/infiniband/hw/mlx4/mad.c 		atomic64_set(&dev->sriov.demux[i].subnet_prefix,
sriov            2328 drivers/infiniband/hw/mlx4/mad.c 				      &dev->sriov.sqps[i]);
sriov            2331 drivers/infiniband/hw/mlx4/mad.c 		err = mlx4_ib_alloc_demux_ctx(dev, &dev->sriov.demux[i], i + 1);
sriov            2343 drivers/infiniband/hw/mlx4/mad.c 		mlx4_ib_free_demux_ctx(&dev->sriov.demux[i]);
sriov            2364 drivers/infiniband/hw/mlx4/mad.c 	spin_lock_irqsave(&dev->sriov.going_down_lock, flags);
sriov            2365 drivers/infiniband/hw/mlx4/mad.c 	dev->sriov.is_going_down = 1;
sriov            2366 drivers/infiniband/hw/mlx4/mad.c 	spin_unlock_irqrestore(&dev->sriov.going_down_lock, flags);
sriov            2369 drivers/infiniband/hw/mlx4/mad.c 			flush_workqueue(dev->sriov.demux[i].ud_wq);
sriov            2370 drivers/infiniband/hw/mlx4/mad.c 			mlx4_ib_free_sqp_ctx(dev->sriov.sqps[i]);
sriov            2371 drivers/infiniband/hw/mlx4/mad.c 			kfree(dev->sriov.sqps[i]);
sriov            2372 drivers/infiniband/hw/mlx4/mad.c 			dev->sriov.sqps[i] = NULL;
sriov            2373 drivers/infiniband/hw/mlx4/mad.c 			mlx4_ib_free_demux_ctx(&dev->sriov.demux[i]);
sriov            3088 drivers/infiniband/hw/mlx4/main.c 	spin_lock_irqsave(&ibdev->sriov.going_down_lock, flags);
sriov            3089 drivers/infiniband/hw/mlx4/main.c 	if (!ibdev->sriov.is_going_down) {
sriov            3091 drivers/infiniband/hw/mlx4/main.c 			queue_work(ibdev->sriov.demux[i].ud_wq, &dm[i]->work);
sriov            3092 drivers/infiniband/hw/mlx4/main.c 		spin_unlock_irqrestore(&ibdev->sriov.going_down_lock, flags);
sriov            3094 drivers/infiniband/hw/mlx4/main.c 		spin_unlock_irqrestore(&ibdev->sriov.going_down_lock, flags);
sriov             887 drivers/infiniband/hw/mlx4/mcg.c 	struct mlx4_ib_demux_ctx *ctx = &dev->sriov.demux[port - 1];
sriov             936 drivers/infiniband/hw/mlx4/mcg.c 	struct mlx4_ib_demux_ctx *ctx = &dev->sriov.demux[port - 1];
sriov             596 drivers/infiniband/hw/mlx4/mlx4_ib.h 	struct mlx4_ib_sriov	sriov;
sriov            3121 drivers/infiniband/hw/mlx4/qp.c 					cpu_to_be64(atomic64_read(&(to_mdev(ib_dev)->sriov.
sriov            3125 drivers/infiniband/hw/mlx4/qp.c 					to_mdev(ib_dev)->sriov.demux[sqp->qp.port - 1].
sriov              87 drivers/infiniband/hw/mlx4/sysfs.c 	spin_lock_irqsave(&mdev->sriov.alias_guid.ag_work_lock, flags);
sriov              89 drivers/infiniband/hw/mlx4/sysfs.c 	*(__be64 *)&mdev->sriov.alias_guid.ports_guid[port->num - 1].
sriov              95 drivers/infiniband/hw/mlx4/sysfs.c 	mdev->sriov.alias_guid.ports_guid[port->num - 1].all_rec_per_port[record_num].status
sriov             102 drivers/infiniband/hw/mlx4/sysfs.c 	mdev->sriov.alias_guid.ports_guid[port->num - 1].all_rec_per_port[record_num].guid_indexes
sriov             105 drivers/infiniband/hw/mlx4/sysfs.c 	spin_unlock_irqrestore(&mdev->sriov.alias_guid.ag_work_lock, flags);
sriov              99 drivers/infiniband/hw/mlx5/ib_virt.c 	struct mlx5_vf_context *vfs_ctx = mdev->priv.sriov.vfs_ctx;
sriov             157 drivers/infiniband/hw/mlx5/ib_virt.c 	struct mlx5_vf_context *vfs_ctx = mdev->priv.sriov.vfs_ctx;
sriov             178 drivers/infiniband/hw/mlx5/ib_virt.c 	struct mlx5_vf_context *vfs_ctx = mdev->priv.sriov.vfs_ctx;
sriov            3398 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 	ether_stat->vf_cnt = IS_SRIOV(bp) ? bp->vfdb->sriov.nr_virtfn : 0;
sriov            1055 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	struct bnx2x_sriov *iov = &bp->vfdb->sriov;
sriov            1064 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	struct bnx2x_sriov *iov = &bp->vfdb->sriov;
sriov            1073 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	struct bnx2x_sriov *iov = &bp->vfdb->sriov;
sriov            1242 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	iov = &(bp->vfdb->sriov);
sriov            1321 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	for (vf_idx = 0; vf_idx < bp->vfdb->sriov.total; vf_idx++) {
sriov            1324 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 						bp->vfdb->sriov.first_vf_in_pf +
sriov            1327 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		   bp->vfdb->sriov.first_vf_in_pf + vf_idx);
sriov            1371 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	tot_size = (BP_VFDB(bp)->sriov.first_vf_in_pf + BNX2X_NR_VIRTFN(bp)) *
sriov            1557 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	DP(BNX2X_MSG_IOV, "num of vfs: %d\n", (bp)->vfdb->sriov.nr_virtfn);
sriov            1566 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		int base_vf_cid = (BP_VFDB(bp)->sriov.first_vf_in_pf + vfid) *
sriov            2023 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	int base_vf_cid = (BP_VFDB(bp)->sriov.first_vf_in_pf + vf->index) *
sriov            2469 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	first_vf = bp->vfdb->sriov.first_vf_in_pf;
sriov             219 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h #define BNX2X_NR_VIRTFN(bp)	((bp)->vfdb->sriov.nr_virtfn)
sriov             241 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h #define GET_NUM_VFS_PER_PF(bp)		((bp)->vfdb ? (bp)->vfdb->sriov.total \
sriov             333 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h 	struct bnx2x_sriov	sriov;
sriov             541 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h 	return bp->vfdb->sriov.nr_virtfn * BNX2X_CIDS_PER_VF;
sriov            2201 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c 	if (vfpf_event->vf_id - BP_VFDB(bp)->sriov.first_vf_in_pf >
sriov            1373 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c 	mlx5_eswitch_update_num_of_vfs(esw, esw->dev->priv.sriov.num_vfs);
sriov            2594 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c 	       vport_num <= esw->dev->priv.sriov.max_vfs;
sriov              41 drivers/net/ethernet/mellanox/mlx5/core/sriov.c 	struct mlx5_core_sriov *sriov = &dev->priv.sriov;
sriov              46 drivers/net/ethernet/mellanox/mlx5/core/sriov.c 	if (sriov->vfs_ctx[vf].node_guid ||
sriov              47 drivers/net/ethernet/mellanox/mlx5/core/sriov.c 	    sriov->vfs_ctx[vf].port_guid ||
sriov              48 drivers/net/ethernet/mellanox/mlx5/core/sriov.c 	    sriov->vfs_ctx[vf].policy != MLX5_POLICY_INVALID) {
sriov              53 drivers/net/ethernet/mellanox/mlx5/core/sriov.c 		in->node_guid = sriov->vfs_ctx[vf].node_guid;
sriov              54 drivers/net/ethernet/mellanox/mlx5/core/sriov.c 		in->port_guid = sriov->vfs_ctx[vf].port_guid;
sriov              55 drivers/net/ethernet/mellanox/mlx5/core/sriov.c 		in->policy = sriov->vfs_ctx[vf].policy;
sriov              73 drivers/net/ethernet/mellanox/mlx5/core/sriov.c 	struct mlx5_core_sriov *sriov = &dev->priv.sriov;
sriov              95 drivers/net/ethernet/mellanox/mlx5/core/sriov.c 		sriov->vfs_ctx[vf].enabled = 1;
sriov             113 drivers/net/ethernet/mellanox/mlx5/core/sriov.c 	struct mlx5_core_sriov *sriov = &dev->priv.sriov;
sriov             119 drivers/net/ethernet/mellanox/mlx5/core/sriov.c 		if (!sriov->vfs_ctx[vf].enabled)
sriov             126 drivers/net/ethernet/mellanox/mlx5/core/sriov.c 		sriov->vfs_ctx[vf].enabled = 0;
sriov             166 drivers/net/ethernet/mellanox/mlx5/core/sriov.c 	struct mlx5_core_sriov *sriov = &dev->priv.sriov;
sriov             177 drivers/net/ethernet/mellanox/mlx5/core/sriov.c 		sriov->num_vfs = num_vfs;
sriov             224 drivers/net/ethernet/mellanox/mlx5/core/sriov.c 	struct mlx5_core_sriov *sriov = &dev->priv.sriov;
sriov             232 drivers/net/ethernet/mellanox/mlx5/core/sriov.c 	sriov->max_vfs = mlx5_get_max_vfs(dev);
sriov             233 drivers/net/ethernet/mellanox/mlx5/core/sriov.c 	sriov->num_vfs = pci_num_vf(pdev);
sriov             234 drivers/net/ethernet/mellanox/mlx5/core/sriov.c 	sriov->vfs_ctx = kcalloc(total_vfs, sizeof(*sriov->vfs_ctx), GFP_KERNEL);
sriov             235 drivers/net/ethernet/mellanox/mlx5/core/sriov.c 	if (!sriov->vfs_ctx)
sriov             243 drivers/net/ethernet/mellanox/mlx5/core/sriov.c 	struct mlx5_core_sriov *sriov = &dev->priv.sriov;
sriov             248 drivers/net/ethernet/mellanox/mlx5/core/sriov.c 	kfree(sriov->vfs_ctx);
sriov             530 drivers/net/ethernet/qlogic/qlcnic/qlcnic.h 	struct qlcnic_sriov *sriov;
sriov            1961 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c 		qlcnic_sriov_cleanup_async_list(&adapter->ahw->sriov->bc);
sriov             267 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov.h static inline void qlcnic_sriov_pf_handle_flr(struct qlcnic_sriov *sriov,
sriov             143 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	struct qlcnic_sriov *sriov;
sriov             153 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	sriov  = kzalloc(sizeof(struct qlcnic_sriov), GFP_KERNEL);
sriov             154 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	if (!sriov)
sriov             157 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	adapter->ahw->sriov = sriov;
sriov             158 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	sriov->num_vfs = num_vfs;
sriov             159 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	bc = &sriov->bc;
sriov             160 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	sriov->vf_info = kcalloc(num_vfs, sizeof(struct qlcnic_vf_info),
sriov             162 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	if (!sriov->vf_info) {
sriov             191 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		vf = &sriov->vf_info[i];
sriov             210 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 			sriov->vf_info[i].vp = vp;
sriov             231 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	kfree(sriov->vf_info);
sriov             234 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	kfree(adapter->ahw->sriov);
sriov             262 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	struct qlcnic_sriov *sriov = adapter->ahw->sriov;
sriov             263 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	struct qlcnic_back_channel *bc = &sriov->bc;
sriov             273 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	for (i = 0; i < sriov->num_vfs; i++) {
sriov             274 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		vf = &sriov->vf_info[i];
sriov             282 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	for (i = 0; i < sriov->num_vfs; i++)
sriov             283 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		kfree(sriov->vf_info[i].vp);
sriov             285 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	kfree(sriov->vf_info);
sriov             286 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	kfree(adapter->ahw->sriov);
sriov             435 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	struct qlcnic_sriov *sriov = adapter->ahw->sriov;
sriov             439 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	if (sriov->allowed_vlans)
sriov             442 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	sriov->any_vlan = cmd->rsp.arg[2] & 0xf;
sriov             443 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	sriov->num_allowed_vlans = cmd->rsp.arg[2] >> 16;
sriov             445 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		 sriov->num_allowed_vlans);
sriov             449 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	if (!sriov->any_vlan)
sriov             452 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	num_vlans = sriov->num_allowed_vlans;
sriov             453 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	sriov->allowed_vlans = kcalloc(num_vlans, sizeof(u16), GFP_KERNEL);
sriov             454 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	if (!sriov->allowed_vlans)
sriov             459 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		sriov->allowed_vlans[i] = vlans[i];
sriov             466 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	struct qlcnic_sriov *sriov = adapter->ahw->sriov;
sriov             480 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		sriov->vlan_mode = cmd.rsp.arg[1] & 0x3;
sriov             481 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		switch (sriov->vlan_mode) {
sriov             682 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	struct qlcnic_vf_info *vf_info = adapter->ahw->sriov->vf_info;
sriov             688 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	for (i = 0; i < adapter->ahw->sriov->num_vfs; i++) {
sriov             847 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c static void qlcnic_sriov_schedule_bc_cmd(struct qlcnic_sriov *sriov,
sriov             855 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	queue_work(sriov->bc.bc_trans_wq, &vf->trans_work);
sriov            1013 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	vf = &adapter->ahw->sriov->vf_info[index];
sriov            1080 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		qlcnic_sriov_schedule_bc_cmd(adapter->ahw->sriov, vf,
sriov            1116 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c int __qlcnic_sriov_add_act_list(struct qlcnic_sriov *sriov,
sriov            1125 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		qlcnic_sriov_schedule_bc_cmd(sriov, vf,
sriov            1130 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c static int qlcnic_sriov_add_act_list(struct qlcnic_sriov *sriov,
sriov            1138 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	__qlcnic_sriov_add_act_list(sriov, vf, trans);
sriov            1144 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c static void qlcnic_sriov_handle_pending_trans(struct qlcnic_sriov *sriov,
sriov            1182 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		if (qlcnic_sriov_add_act_list(sriov, vf, trans))
sriov            1188 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c static void qlcnic_sriov_handle_bc_cmd(struct qlcnic_sriov *sriov,
sriov            1208 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		qlcnic_sriov_handle_pending_trans(sriov, vf, hdr);
sriov            1250 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		if (qlcnic_sriov_add_act_list(sriov, vf, trans)) {
sriov            1262 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c static void qlcnic_sriov_handle_msg_event(struct qlcnic_sriov *sriov,
sriov            1275 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		qlcnic_sriov_handle_bc_cmd(sriov, &hdr, vf);
sriov            1283 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c static void qlcnic_sriov_handle_flr_event(struct qlcnic_sriov *sriov,
sriov            1289 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		qlcnic_sriov_pf_handle_flr(sriov, vf);
sriov            1298 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	struct qlcnic_sriov *sriov;
sriov            1302 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	sriov = adapter->ahw->sriov;
sriov            1309 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	vf = &sriov->vf_info[index];
sriov            1316 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		qlcnic_sriov_handle_flr_event(sriov, vf);
sriov            1321 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		qlcnic_sriov_handle_msg_event(sriov, vf);
sriov            1379 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	u16 seq = ++adapter->ahw->sriov->bc.trans_counter;
sriov            1467 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	struct qlcnic_vf_info *vf = &adapter->ahw->sriov->vf_info[0];
sriov            1499 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	struct qlcnic_sriov *sriov = adapter->ahw->sriov;
sriov            1504 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	vf = &adapter->ahw->sriov->vf_info[0];
sriov            1510 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		for (i = 0; i < sriov->num_allowed_vlans; i++) {
sriov            1662 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	struct qlcnic_back_channel *bc = &adapter->ahw->sriov->bc;
sriov            1947 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c static int qlcnic_sriov_check_vlan_id(struct qlcnic_sriov *sriov,
sriov            1957 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	for (i = 0; i < sriov->num_allowed_vlans; i++) {
sriov            1968 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c static int qlcnic_sriov_validate_num_vlans(struct qlcnic_sriov *sriov,
sriov            1975 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	if (vf->num_vlan >= sriov->num_allowed_vlans)
sriov            1985 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	struct qlcnic_sriov *sriov = adapter->ahw->sriov;
sriov            1991 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	vf = &adapter->ahw->sriov->vf_info[0];
sriov            1993 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	if (sriov->vlan_mode != QLC_GUEST_VLAN_MODE)
sriov            2000 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		if (qlcnic_sriov_validate_num_vlans(sriov, vf))
sriov            2003 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		if (sriov->any_vlan) {
sriov            2004 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 			for (i = 0; i < sriov->num_allowed_vlans; i++) {
sriov            2005 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 				if (sriov->allowed_vlans[i] == vid)
sriov            2013 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		if (!vlan_exist || qlcnic_sriov_check_vlan_id(sriov, vf, vid))
sriov            2024 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	struct qlcnic_sriov *sriov;
sriov            2026 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	sriov = adapter->ahw->sriov;
sriov            2035 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		qlcnic_sriov_add_vlan_id(sriov, vf, vlan_id);
sriov            2038 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		qlcnic_sriov_del_vlan_id(sriov, vf, vlan_id);
sriov            2051 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	struct qlcnic_sriov *sriov = adapter->ahw->sriov;
sriov            2061 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	vf = &adapter->ahw->sriov->vf_info[0];
sriov            2073 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	qlcnic_sriov_cleanup_async_list(&sriov->bc);
sriov            2165 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	struct qlcnic_sriov *sriov = adapter->ahw->sriov;
sriov            2169 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	for (i = 0; i < sriov->num_vfs; i++) {
sriov            2170 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		vf = &sriov->vf_info[i];
sriov            2171 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		vf->sriov_vlans = kcalloc(sriov->num_allowed_vlans,
sriov            2178 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	struct qlcnic_sriov *sriov = adapter->ahw->sriov;
sriov            2182 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	for (i = 0; i < sriov->num_vfs; i++) {
sriov            2183 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		vf = &sriov->vf_info[i];
sriov            2189 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c void qlcnic_sriov_add_vlan_id(struct qlcnic_sriov *sriov,
sriov            2194 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	for (i = 0; i < sriov->num_allowed_vlans; i++) {
sriov            2203 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c void qlcnic_sriov_del_vlan_id(struct qlcnic_sriov *sriov,
sriov            2208 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	for (i = 0; i < sriov->num_allowed_vlans; i++) {
sriov              71 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	struct qlcnic_sriov *sriov = adapter->ahw->sriov;
sriov              72 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	struct qlcnic_resources *res = &sriov->ff_max;
sriov              73 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	u16 num_macs = sriov->num_allowed_vlans + 1;
sriov              82 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	num_vfs = sriov->num_vfs;
sriov             104 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 		info->max_tx_ques = res->num_tx_queues - sriov->num_vfs;
sriov             109 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 		vp = sriov->vf_info[id].vp;
sriov             140 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	struct qlcnic_resources *ff_max = &adapter->ahw->sriov->ff_max;
sriov             159 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	struct qlcnic_sriov *sriov = adapter->ahw->sriov;
sriov             163 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	total_fn = sriov->num_vfs + 1;
sriov             166 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	sriov->num_allowed_vlans = temp - 1;
sriov             169 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 		sriov->num_allowed_vlans = 1;
sriov             172 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 		    sriov->num_allowed_vlans);
sriov             236 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	struct qlcnic_sriov  *sriov = adapter->ahw->sriov;
sriov             241 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 		sriov->vp_handle = 0;
sriov             246 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 		vp = sriov->vf_info[index].vp;
sriov             254 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	struct qlcnic_sriov  *sriov = adapter->ahw->sriov;
sriov             259 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 		sriov->vp_handle = vport_handle;
sriov             264 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 		vp = sriov->vf_info[index].vp;
sriov             272 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	struct qlcnic_sriov  *sriov = adapter->ahw->sriov;
sriov             277 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 		return sriov->vp_handle;
sriov             281 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 			vf_info = &sriov->vf_info[index];
sriov             409 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	struct qlcnic_sriov *sriov = adapter->ahw->sriov;
sriov             410 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	struct qlcnic_back_channel *bc = &sriov->bc;
sriov             413 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	for (i = 0; i < sriov->num_vfs; i++)
sriov             414 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 		cancel_work_sync(&sriov->vf_info[i].flr_work);
sriov             421 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	struct qlcnic_back_channel *bc = &adapter->ahw->sriov->bc;
sriov             694 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	vp = adapter->ahw->sriov->vf_info[id].vp;
sriov             745 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	struct qlcnic_sriov *sriov;
sriov             751 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	sriov = adapter->ahw->sriov;
sriov             763 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 			size = size * sriov->num_allowed_vlans;
sriov             858 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	struct qlcnic_sriov *sriov;
sriov             862 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	sriov = adapter->ahw->sriov;
sriov             866 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 		for (i = 0; i < sriov->num_allowed_vlans; i++) {
sriov            1341 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	struct qlcnic_sriov *sriov;
sriov            1344 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	sriov = adapter->ahw->sriov;
sriov            1357 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 		cmd->rsp.arg[2] = sriov->num_allowed_vlans << 16;
sriov            1371 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	struct qlcnic_sriov *sriov = adapter->ahw->sriov;
sriov            1379 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 		qlcnic_sriov_del_vlan_id(sriov, vf, vlan);
sriov            1384 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	qlcnic_sriov_del_vlan_id(sriov, vf, vlan);
sriov            1396 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	struct qlcnic_sriov *sriov = adapter->ahw->sriov;
sriov            1406 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 		qlcnic_sriov_add_vlan_id(sriov, vf, vlan);
sriov            1426 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	qlcnic_sriov_add_vlan_id(sriov, vf, vlan);
sriov            1646 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c static int qlcnic_sriov_add_act_list_irqsave(struct qlcnic_sriov *sriov,
sriov            1655 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	__qlcnic_sriov_add_act_list(sriov, vf, trans);
sriov            1678 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 		qlcnic_sriov_add_act_list_irqsave(adapter->ahw->sriov, vf,
sriov            1694 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c static void qlcnic_sriov_schedule_flr(struct qlcnic_sriov *sriov,
sriov            1702 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	queue_work(sriov->bc.bc_flr_wq, &vf->flr_work);
sriov            1709 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	struct qlcnic_sriov *sriov = adapter->ahw->sriov;
sriov            1715 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	qlcnic_sriov_schedule_flr(sriov, vf, qlcnic_sriov_pf_process_flr);
sriov            1736 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c void qlcnic_sriov_pf_handle_flr(struct qlcnic_sriov *sriov,
sriov            1755 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 		       sizeof(*vf->sriov_vlans) * sriov->num_allowed_vlans);
sriov            1757 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	qlcnic_sriov_schedule_flr(sriov, vf, qlcnic_sriov_pf_process_flr);
sriov            1764 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	struct qlcnic_sriov *sriov = ahw->sriov;
sriov            1766 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	u16 num_vfs = sriov->num_vfs;
sriov            1770 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 		vf = &sriov->vf_info[i];
sriov            1804 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	struct qlcnic_sriov *sriov = adapter->ahw->sriov;
sriov            1812 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	num_vfs = sriov->num_vfs;
sriov            1823 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 		vf_info = &sriov->vf_info[i];
sriov            1832 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	vf_info = &sriov->vf_info[vf];
sriov            1852 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	struct qlcnic_sriov *sriov = adapter->ahw->sriov;
sriov            1861 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	if (vf >= sriov->num_vfs)
sriov            1864 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	vf_info = &sriov->vf_info[vf];
sriov            1917 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	struct qlcnic_sriov *sriov = adapter->ahw->sriov;
sriov            1924 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	if (vf >= sriov->num_vfs || qos > 7)
sriov            1937 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	vf_info = &sriov->vf_info[vf];
sriov            1947 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	       sizeof(*vf_info->sriov_vlans) * sriov->num_allowed_vlans);
sriov            1959 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 		qlcnic_sriov_add_vlan_id(sriov, vf_info, vlan);
sriov            1996 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	struct qlcnic_sriov *sriov = adapter->ahw->sriov;
sriov            2002 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	if (vf >= sriov->num_vfs)
sriov            2005 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	vp = sriov->vf_info[vf].vp;
sriov            2026 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	struct qlcnic_sriov *sriov = adapter->ahw->sriov;
sriov            2033 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	if (vf >= sriov->num_vfs)
sriov            2036 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c 	vf_info = &sriov->vf_info[vf];
sriov            1243 drivers/net/ethernet/sfc/siena_sriov.c 	unsigned index, devfn, sriov, buftbl_base;
sriov            1248 drivers/net/ethernet/sfc/siena_sriov.c 	sriov = pci_find_ext_capability(pci_dev, PCI_EXT_CAP_ID_SRIOV);
sriov            1249 drivers/net/ethernet/sfc/siena_sriov.c 	if (!sriov)
sriov            1252 drivers/net/ethernet/sfc/siena_sriov.c 	pci_read_config_word(pci_dev, sriov + PCI_SRIOV_VF_OFFSET, &offset);
sriov            1253 drivers/net/ethernet/sfc/siena_sriov.c 	pci_read_config_word(pci_dev, sriov + PCI_SRIOV_VF_STRIDE, &stride);
sriov             516 drivers/net/hyperv/hyperv_net.h 			u64 sriov:1;
sriov             529 drivers/net/hyperv/netvsc.c 		init_packet->msg.v2_msg.send_ndis_config.capability.sriov = 1;
sriov              24 drivers/pci/iov.c 	return dev->bus->number + ((dev->devfn + dev->sriov->offset +
sriov              25 drivers/pci/iov.c 				    dev->sriov->stride * vf_id) >> 8);
sriov              32 drivers/pci/iov.c 	return (dev->devfn + dev->sriov->offset +
sriov              33 drivers/pci/iov.c 		dev->sriov->stride * vf_id) & 0xff;
sriov              44 drivers/pci/iov.c 	struct pci_sriov *iov = dev->sriov;
sriov              60 drivers/pci/iov.c 	struct pci_sriov *iov = dev->sriov;
sriov             111 drivers/pci/iov.c 	return dev->sriov->barsz[resno - PCI_IOV_RESOURCES];
sriov             128 drivers/pci/iov.c 			      &physfn->sriov->class);
sriov             130 drivers/pci/iov.c 			     &physfn->sriov->hdr_type);
sriov             132 drivers/pci/iov.c 			     &physfn->sriov->subsystem_vendor);
sriov             134 drivers/pci/iov.c 			     &physfn->sriov->subsystem_device);
sriov             145 drivers/pci/iov.c 	struct pci_sriov *iov = dev->sriov;
sriov             257 drivers/pci/iov.c 	return sprintf(buf, "%u\n", pdev->sriov->num_VFs);
sriov             284 drivers/pci/iov.c 	if (num_vfs == pdev->sriov->num_VFs)
sriov             301 drivers/pci/iov.c 	if (pdev->sriov->num_VFs) {
sriov             303 drivers/pci/iov.c 			 pdev->sriov->num_VFs, num_vfs);
sriov             331 drivers/pci/iov.c 	return sprintf(buf, "%u\n", pdev->sriov->offset);
sriov             340 drivers/pci/iov.c 	return sprintf(buf, "%u\n", pdev->sriov->stride);
sriov             349 drivers/pci/iov.c 	return sprintf(buf, "%x\n", pdev->sriov->vf_device);
sriov             358 drivers/pci/iov.c 	return sprintf(buf, "%u\n", pdev->sriov->drivers_autoprobe);
sriov             371 drivers/pci/iov.c 	pdev->sriov->drivers_autoprobe = drivers_autoprobe;
sriov             448 drivers/pci/iov.c 	struct pci_sriov *iov = dev->sriov;
sriov             552 drivers/pci/iov.c 	struct pci_sriov *iov = dev->sriov;
sriov             564 drivers/pci/iov.c 	struct pci_sriov *iov = dev->sriov;
sriov             676 drivers/pci/iov.c 	dev->sriov = iov;
sriov             685 drivers/pci/iov.c 	dev->sriov = NULL;
sriov             699 drivers/pci/iov.c 	BUG_ON(dev->sriov->num_VFs);
sriov             701 drivers/pci/iov.c 	if (dev != dev->sriov->dev)
sriov             702 drivers/pci/iov.c 		pci_dev_put(dev->sriov->dev);
sriov             704 drivers/pci/iov.c 	kfree(dev->sriov);
sriov             705 drivers/pci/iov.c 	dev->sriov = NULL;
sriov             712 drivers/pci/iov.c 	struct pci_sriov *iov = dev->sriov;
sriov             772 drivers/pci/iov.c 	struct pci_sriov *iov = dev->sriov;
sriov             791 drivers/pci/iov.c 	struct pci_sriov *iov = dev->is_physfn ? dev->sriov : NULL;
sriov             879 drivers/pci/iov.c 		dev->sriov->drivers_autoprobe = auto_probe;
sriov             897 drivers/pci/iov.c 		if (dev->sriov->max_VF_buses > max)
sriov             898 drivers/pci/iov.c 			max = dev->sriov->max_VF_buses;
sriov             948 drivers/pci/iov.c 	return dev->sriov->num_VFs;
sriov             973 drivers/pci/iov.c 	dev_id = dev->sriov->vf_device;
sriov            1011 drivers/pci/iov.c 	if (numvfs > dev->sriov->total_VFs)
sriov            1015 drivers/pci/iov.c 	if (dev->sriov->ctrl & PCI_SRIOV_CTRL_VFE)
sriov            1018 drivers/pci/iov.c 	dev->sriov->driver_max_VFs = numvfs;
sriov            1036 drivers/pci/iov.c 	return dev->sriov->driver_max_VFs;
sriov             402 drivers/pci/pci-driver.c 	return (!pdev->is_virtfn || pdev->physfn->sriov->drivers_autoprobe ||
sriov            1620 drivers/pci/probe.c 		return dev->physfn->sriov->class;
sriov            1630 drivers/pci/probe.c 		*vendor = dev->physfn->sriov->subsystem_vendor;
sriov            1631 drivers/pci/probe.c 		*device = dev->physfn->sriov->subsystem_device;
sriov            1645 drivers/pci/probe.c 		return dev->physfn->sriov->hdr_type;
sriov            1041 drivers/pci/quirks.c 		dev->sriov->link = dev->devfn;
sriov             769 drivers/scsi/bfa/bfa_defs.h 	u8	sriov;
sriov             583 include/linux/mlx5/driver.h 	struct mlx5_core_sriov	sriov;
sriov            1143 include/linux/mlx5/driver.h 	return dev->priv.sriov.max_vfs;
sriov             450 include/linux/pci.h 		struct pci_sriov	*sriov;		/* PF: SR-IOV info */