sriov 86 drivers/infiniband/hw/mlx4/alias_GUID.c guid_indexes = be64_to_cpu((__force __be64) dev->sriov.alias_guid. sriov 102 drivers/infiniband/hw/mlx4/alias_GUID.c memcpy(&dev->sriov.demux[port_index].guid_cache[slave_id], sriov 117 drivers/infiniband/hw/mlx4/alias_GUID.c return *(__be64 *)&dev->sriov.demux[port - 1].guid_cache[index]; sriov 136 drivers/infiniband/hw/mlx4/alias_GUID.c spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags); sriov 137 drivers/infiniband/hw/mlx4/alias_GUID.c if (dev->sriov.alias_guid.ports_guid[port_index].state_flags & sriov 141 drivers/infiniband/hw/mlx4/alias_GUID.c curr_guid = *(__be64 *)&dev->sriov. sriov 154 drivers/infiniband/hw/mlx4/alias_GUID.c *(__be64 *)&dev->sriov.alias_guid.ports_guid[port_index]. sriov 157 drivers/infiniband/hw/mlx4/alias_GUID.c dev->sriov.alias_guid.ports_guid[port_index]. sriov 160 drivers/infiniband/hw/mlx4/alias_GUID.c dev->sriov.alias_guid.ports_guid[port_index]. sriov 164 drivers/infiniband/hw/mlx4/alias_GUID.c dev->sriov.alias_guid.ports_guid[port_index]. sriov 166 drivers/infiniband/hw/mlx4/alias_GUID.c dev->sriov.alias_guid.ports_guid[port_index]. sriov 171 drivers/infiniband/hw/mlx4/alias_GUID.c spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags); sriov 204 drivers/infiniband/hw/mlx4/alias_GUID.c rec = &dev->sriov.alias_guid.ports_guid[port_num - 1]. sriov 206 drivers/infiniband/hw/mlx4/alias_GUID.c guid_indexes = be64_to_cpu((__force __be64) dev->sriov.alias_guid. sriov 236 drivers/infiniband/hw/mlx4/alias_GUID.c spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags); sriov 248 drivers/infiniband/hw/mlx4/alias_GUID.c spin_unlock_irqrestore(&dev->sriov. sriov 253 drivers/infiniband/hw/mlx4/alias_GUID.c spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, sriov 307 drivers/infiniband/hw/mlx4/alias_GUID.c rec = &dev->sriov.alias_guid.ports_guid[port_index]. sriov 327 drivers/infiniband/hw/mlx4/alias_GUID.c rec = &dev->sriov.alias_guid.ports_guid[port_index]. sriov 330 drivers/infiniband/hw/mlx4/alias_GUID.c spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags); sriov 424 drivers/infiniband/hw/mlx4/alias_GUID.c spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags); sriov 434 drivers/infiniband/hw/mlx4/alias_GUID.c spin_lock_irqsave(&dev->sriov.going_down_lock, flags); sriov 435 drivers/infiniband/hw/mlx4/alias_GUID.c spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1); sriov 436 drivers/infiniband/hw/mlx4/alias_GUID.c if (!dev->sriov.is_going_down) { sriov 438 drivers/infiniband/hw/mlx4/alias_GUID.c queue_delayed_work(dev->sriov.alias_guid.ports_guid[port_index].wq, sriov 439 drivers/infiniband/hw/mlx4/alias_GUID.c &dev->sriov.alias_guid.ports_guid[port_index]. sriov 448 drivers/infiniband/hw/mlx4/alias_GUID.c spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1); sriov 449 drivers/infiniband/hw/mlx4/alias_GUID.c spin_unlock_irqrestore(&dev->sriov.going_down_lock, flags); sriov 458 drivers/infiniband/hw/mlx4/alias_GUID.c dev->sriov.alias_guid.ports_guid[port - 1].all_rec_per_port[index].status sriov 464 drivers/infiniband/hw/mlx4/alias_GUID.c *(u64 *)&dev->sriov.alias_guid.ports_guid[port - 1]. sriov 477 drivers/infiniband/hw/mlx4/alias_GUID.c dev->sriov.alias_guid.ports_guid[port - 1]. sriov 479 drivers/infiniband/hw/mlx4/alias_GUID.c if (dev->sriov.alias_guid.ports_guid[port - 1]. sriov 481 drivers/infiniband/hw/mlx4/alias_GUID.c dev->sriov.alias_guid.ports_guid[port - 1]. sriov 500 drivers/infiniband/hw/mlx4/alias_GUID.c &dev->sriov.alias_guid.ports_guid[port - 1].cb_list; sriov 540 drivers/infiniband/hw/mlx4/alias_GUID.c spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1); sriov 542 drivers/infiniband/hw/mlx4/alias_GUID.c spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1); sriov 545 drivers/infiniband/hw/mlx4/alias_GUID.c ib_sa_guid_info_rec_query(dev->sriov.alias_guid.sa_client, sriov 555 drivers/infiniband/hw/mlx4/alias_GUID.c spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1); sriov 558 drivers/infiniband/hw/mlx4/alias_GUID.c spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1); sriov 567 drivers/infiniband/hw/mlx4/alias_GUID.c spin_lock_irqsave(&dev->sriov.going_down_lock, flags); sriov 568 drivers/infiniband/hw/mlx4/alias_GUID.c spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1); sriov 570 drivers/infiniband/hw/mlx4/alias_GUID.c if (!dev->sriov.is_going_down) { sriov 571 drivers/infiniband/hw/mlx4/alias_GUID.c queue_delayed_work(dev->sriov.alias_guid.ports_guid[port - 1].wq, sriov 572 drivers/infiniband/hw/mlx4/alias_GUID.c &dev->sriov.alias_guid.ports_guid[port - 1].alias_guid_work, sriov 575 drivers/infiniband/hw/mlx4/alias_GUID.c spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1); sriov 576 drivers/infiniband/hw/mlx4/alias_GUID.c spin_unlock_irqrestore(&dev->sriov.going_down_lock, flags); sriov 596 drivers/infiniband/hw/mlx4/alias_GUID.c *(__be64 *)&dev->sriov.alias_guid.ports_guid[port - 1]. sriov 613 drivers/infiniband/hw/mlx4/alias_GUID.c spin_lock_irqsave(&dev->sriov.going_down_lock, flags); sriov 614 drivers/infiniband/hw/mlx4/alias_GUID.c spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1); sriov 616 drivers/infiniband/hw/mlx4/alias_GUID.c if (dev->sriov.alias_guid.ports_guid[port - 1].state_flags & sriov 619 drivers/infiniband/hw/mlx4/alias_GUID.c dev->sriov.alias_guid.ports_guid[port - 1].state_flags &= sriov 625 drivers/infiniband/hw/mlx4/alias_GUID.c if (mlx4_is_master(dev->dev) && !dev->sriov.is_going_down) { sriov 631 drivers/infiniband/hw/mlx4/alias_GUID.c cancel_delayed_work(&dev->sriov.alias_guid. sriov 633 drivers/infiniband/hw/mlx4/alias_GUID.c queue_delayed_work(dev->sriov.alias_guid.ports_guid[port - 1].wq, sriov 634 drivers/infiniband/hw/mlx4/alias_GUID.c &dev->sriov.alias_guid.ports_guid[port - 1].alias_guid_work, sriov 637 drivers/infiniband/hw/mlx4/alias_GUID.c spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1); sriov 638 drivers/infiniband/hw/mlx4/alias_GUID.c spin_unlock_irqrestore(&dev->sriov.going_down_lock, flags); sriov 651 drivers/infiniband/hw/mlx4/alias_GUID.c &dev->sriov.alias_guid.ports_guid[port]. sriov 700 drivers/infiniband/hw/mlx4/alias_GUID.c rec = dev->sriov.alias_guid.ports_guid[port]. sriov 730 drivers/infiniband/hw/mlx4/alias_GUID.c spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags); sriov 740 drivers/infiniband/hw/mlx4/alias_GUID.c spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags); sriov 756 drivers/infiniband/hw/mlx4/alias_GUID.c struct mlx4_ib_dev *dev = container_of(ib_sriov, struct mlx4_ib_dev, sriov); sriov 781 drivers/infiniband/hw/mlx4/alias_GUID.c spin_lock_irqsave(&dev->sriov.going_down_lock, flags); sriov 782 drivers/infiniband/hw/mlx4/alias_GUID.c spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1); sriov 783 drivers/infiniband/hw/mlx4/alias_GUID.c if (!dev->sriov.is_going_down) { sriov 788 drivers/infiniband/hw/mlx4/alias_GUID.c cancel_delayed_work(&dev->sriov.alias_guid.ports_guid[port]. sriov 790 drivers/infiniband/hw/mlx4/alias_GUID.c queue_delayed_work(dev->sriov.alias_guid.ports_guid[port].wq, sriov 791 drivers/infiniband/hw/mlx4/alias_GUID.c &dev->sriov.alias_guid.ports_guid[port].alias_guid_work, 0); sriov 793 drivers/infiniband/hw/mlx4/alias_GUID.c spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1); sriov 794 drivers/infiniband/hw/mlx4/alias_GUID.c spin_unlock_irqrestore(&dev->sriov.going_down_lock, flags); sriov 800 drivers/infiniband/hw/mlx4/alias_GUID.c struct mlx4_ib_sriov *sriov = &dev->sriov; sriov 807 drivers/infiniband/hw/mlx4/alias_GUID.c det = &sriov->alias_guid.ports_guid[i]; sriov 809 drivers/infiniband/hw/mlx4/alias_GUID.c spin_lock_irqsave(&sriov->alias_guid.ag_work_lock, flags); sriov 817 drivers/infiniband/hw/mlx4/alias_GUID.c spin_unlock_irqrestore(&sriov->alias_guid.ag_work_lock, flags); sriov 821 drivers/infiniband/hw/mlx4/alias_GUID.c spin_lock_irqsave(&sriov->alias_guid.ag_work_lock, flags); sriov 823 drivers/infiniband/hw/mlx4/alias_GUID.c spin_unlock_irqrestore(&sriov->alias_guid.ag_work_lock, flags); sriov 826 drivers/infiniband/hw/mlx4/alias_GUID.c flush_workqueue(dev->sriov.alias_guid.ports_guid[i].wq); sriov 827 drivers/infiniband/hw/mlx4/alias_GUID.c destroy_workqueue(dev->sriov.alias_guid.ports_guid[i].wq); sriov 829 drivers/infiniband/hw/mlx4/alias_GUID.c ib_sa_unregister_client(dev->sriov.alias_guid.sa_client); sriov 830 drivers/infiniband/hw/mlx4/alias_GUID.c kfree(dev->sriov.alias_guid.sa_client); sriov 842 drivers/infiniband/hw/mlx4/alias_GUID.c dev->sriov.alias_guid.sa_client = sriov 843 drivers/infiniband/hw/mlx4/alias_GUID.c kzalloc(sizeof *dev->sriov.alias_guid.sa_client, GFP_KERNEL); sriov 844 drivers/infiniband/hw/mlx4/alias_GUID.c if (!dev->sriov.alias_guid.sa_client) sriov 847 drivers/infiniband/hw/mlx4/alias_GUID.c ib_sa_register_client(dev->sriov.alias_guid.sa_client); sriov 849 drivers/infiniband/hw/mlx4/alias_GUID.c spin_lock_init(&dev->sriov.alias_guid.ag_work_lock); sriov 859 drivers/infiniband/hw/mlx4/alias_GUID.c memset(&dev->sriov.alias_guid.ports_guid[i], 0, sriov 861 drivers/infiniband/hw/mlx4/alias_GUID.c dev->sriov.alias_guid.ports_guid[i].state_flags |= sriov 865 drivers/infiniband/hw/mlx4/alias_GUID.c memset(dev->sriov.alias_guid.ports_guid[i]. sriov 867 drivers/infiniband/hw/mlx4/alias_GUID.c sizeof(dev->sriov.alias_guid.ports_guid[i]. sriov 870 drivers/infiniband/hw/mlx4/alias_GUID.c INIT_LIST_HEAD(&dev->sriov.alias_guid.ports_guid[i].cb_list); sriov 878 drivers/infiniband/hw/mlx4/alias_GUID.c dev->sriov.alias_guid.ports_guid[i].parent = &dev->sriov.alias_guid; sriov 879 drivers/infiniband/hw/mlx4/alias_GUID.c dev->sriov.alias_guid.ports_guid[i].port = i; sriov 882 drivers/infiniband/hw/mlx4/alias_GUID.c dev->sriov.alias_guid.ports_guid[i].wq = sriov 884 drivers/infiniband/hw/mlx4/alias_GUID.c if (!dev->sriov.alias_guid.ports_guid[i].wq) { sriov 888 drivers/infiniband/hw/mlx4/alias_GUID.c INIT_DELAYED_WORK(&dev->sriov.alias_guid.ports_guid[i].alias_guid_work, sriov 895 drivers/infiniband/hw/mlx4/alias_GUID.c destroy_workqueue(dev->sriov.alias_guid.ports_guid[i].wq); sriov 896 drivers/infiniband/hw/mlx4/alias_GUID.c dev->sriov.alias_guid.ports_guid[i].wq = NULL; sriov 900 drivers/infiniband/hw/mlx4/alias_GUID.c ib_sa_unregister_client(dev->sriov.alias_guid.sa_client); sriov 901 drivers/infiniband/hw/mlx4/alias_GUID.c kfree(dev->sriov.alias_guid.sa_client); sriov 902 drivers/infiniband/hw/mlx4/alias_GUID.c dev->sriov.alias_guid.sa_client = NULL; sriov 146 drivers/infiniband/hw/mlx4/cm.c struct rb_root *sl_id_map = &to_mdev(ibdev)->sriov.sl_id_map; sriov 173 drivers/infiniband/hw/mlx4/cm.c struct mlx4_ib_sriov *sriov = &dev->sriov; sriov 174 drivers/infiniband/hw/mlx4/cm.c struct rb_root *sl_id_map = &sriov->sl_id_map; sriov 176 drivers/infiniband/hw/mlx4/cm.c spin_lock(&sriov->id_map_lock); sriov 177 drivers/infiniband/hw/mlx4/cm.c if (!xa_erase(&sriov->pv_id_table, ent->pv_cm_id)) sriov 185 drivers/infiniband/hw/mlx4/cm.c spin_unlock(&sriov->id_map_lock); sriov 191 drivers/infiniband/hw/mlx4/cm.c struct rb_root *sl_id_map = &to_mdev(ibdev)->sriov.sl_id_map; sriov 226 drivers/infiniband/hw/mlx4/cm.c struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; sriov 238 drivers/infiniband/hw/mlx4/cm.c ret = xa_alloc_cyclic(&sriov->pv_id_table, &ent->pv_cm_id, ent, sriov 239 drivers/infiniband/hw/mlx4/cm.c xa_limit_32b, &sriov->pv_id_next, GFP_KERNEL); sriov 241 drivers/infiniband/hw/mlx4/cm.c spin_lock(&sriov->id_map_lock); sriov 243 drivers/infiniband/hw/mlx4/cm.c list_add_tail(&ent->list, &sriov->cm_list); sriov 244 drivers/infiniband/hw/mlx4/cm.c spin_unlock(&sriov->id_map_lock); sriov 258 drivers/infiniband/hw/mlx4/cm.c struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; sriov 260 drivers/infiniband/hw/mlx4/cm.c spin_lock(&sriov->id_map_lock); sriov 266 drivers/infiniband/hw/mlx4/cm.c ent = xa_load(&sriov->pv_id_table, *pv_cm_id); sriov 267 drivers/infiniband/hw/mlx4/cm.c spin_unlock(&sriov->id_map_lock); sriov 274 drivers/infiniband/hw/mlx4/cm.c struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; sriov 277 drivers/infiniband/hw/mlx4/cm.c spin_lock(&sriov->id_map_lock); sriov 278 drivers/infiniband/hw/mlx4/cm.c spin_lock_irqsave(&sriov->going_down_lock, flags); sriov 280 drivers/infiniband/hw/mlx4/cm.c if (!sriov->is_going_down && !id->scheduled_delete) { sriov 284 drivers/infiniband/hw/mlx4/cm.c spin_unlock_irqrestore(&sriov->going_down_lock, flags); sriov 285 drivers/infiniband/hw/mlx4/cm.c spin_unlock(&sriov->id_map_lock); sriov 374 drivers/infiniband/hw/mlx4/cm.c spin_lock_init(&dev->sriov.id_map_lock); sriov 375 drivers/infiniband/hw/mlx4/cm.c INIT_LIST_HEAD(&dev->sriov.cm_list); sriov 376 drivers/infiniband/hw/mlx4/cm.c dev->sriov.sl_id_map = RB_ROOT; sriov 377 drivers/infiniband/hw/mlx4/cm.c xa_init_flags(&dev->sriov.pv_id_table, XA_FLAGS_ALLOC); sriov 384 drivers/infiniband/hw/mlx4/cm.c struct mlx4_ib_sriov *sriov = &dev->sriov; sriov 385 drivers/infiniband/hw/mlx4/cm.c struct rb_root *sl_id_map = &sriov->sl_id_map; sriov 392 drivers/infiniband/hw/mlx4/cm.c spin_lock(&sriov->id_map_lock); sriov 393 drivers/infiniband/hw/mlx4/cm.c list_for_each_entry_safe(map, tmp_map, &dev->sriov.cm_list, list) { sriov 400 drivers/infiniband/hw/mlx4/cm.c spin_unlock(&sriov->id_map_lock); sriov 406 drivers/infiniband/hw/mlx4/cm.c spin_lock(&sriov->id_map_lock); sriov 414 drivers/infiniband/hw/mlx4/cm.c xa_erase(&sriov->pv_id_table, ent->pv_cm_id); sriov 416 drivers/infiniband/hw/mlx4/cm.c list_splice_init(&dev->sriov.cm_list, &lh); sriov 430 drivers/infiniband/hw/mlx4/cm.c xa_erase(&sriov->pv_id_table, map->pv_cm_id); sriov 434 drivers/infiniband/hw/mlx4/cm.c list_for_each_entry_safe(map, tmp_map, &dev->sriov.cm_list, list) { sriov 440 drivers/infiniband/hw/mlx4/cm.c spin_unlock(&sriov->id_map_lock); sriov 284 drivers/infiniband/hw/mlx4/mad.c if (!dev->sriov.is_going_down) sriov 299 drivers/infiniband/hw/mlx4/mad.c !dev->sriov.is_going_down) { sriov 438 drivers/infiniband/hw/mlx4/mad.c if (dev->sriov.demux[port - 1].guid_cache[i] == guid) sriov 526 drivers/infiniband/hw/mlx4/mad.c tun_ctx = dev->sriov.demux[port-1].tun[slave]; sriov 739 drivers/infiniband/hw/mlx4/mad.c atomic64_read(&dev->sriov.demux[port - 1].subnet_prefix))) { sriov 1090 drivers/infiniband/hw/mlx4/mad.c if (mlx4_is_master(dev->dev) && !dev->sriov.is_going_down) sriov 1101 drivers/infiniband/hw/mlx4/mad.c if (!dev->sriov.is_going_down) { sriov 1102 drivers/infiniband/hw/mlx4/mad.c mlx4_ib_mcg_port_cleanup(&dev->sriov.demux[port_num - 1], 0); sriov 1228 drivers/infiniband/hw/mlx4/mad.c (u64)atomic64_read(&dev->sriov.demux[port - 1].subnet_prefix), sriov 1230 drivers/infiniband/hw/mlx4/mad.c atomic64_set(&dev->sriov.demux[port - 1].subnet_prefix, sriov 1247 drivers/infiniband/hw/mlx4/mad.c if (mlx4_is_master(dev->dev) && !dev->sriov.is_going_down) sriov 1255 drivers/infiniband/hw/mlx4/mad.c else if (!dev->sriov.is_going_down) { sriov 1304 drivers/infiniband/hw/mlx4/mad.c spin_lock_irqsave(&dev->sriov.going_down_lock, flags); sriov 1305 drivers/infiniband/hw/mlx4/mad.c if (!dev->sriov.is_going_down && ctx->state == DEMUX_PV_STATE_ACTIVE) sriov 1307 drivers/infiniband/hw/mlx4/mad.c spin_unlock_irqrestore(&dev->sriov.going_down_lock, flags); sriov 1378 drivers/infiniband/hw/mlx4/mad.c sqp_ctx = dev->sriov.sqps[port-1]; sriov 1972 drivers/infiniband/hw/mlx4/mad.c if (dev->sriov.demux[port - 1].tun[slave]) { sriov 1973 drivers/infiniband/hw/mlx4/mad.c kfree(dev->sriov.demux[port - 1].tun[slave]); sriov 1974 drivers/infiniband/hw/mlx4/mad.c dev->sriov.demux[port - 1].tun[slave] = NULL; sriov 2048 drivers/infiniband/hw/mlx4/mad.c ctx->wq = to_mdev(ibdev)->sriov.demux[port - 1].wq; sriov 2119 drivers/infiniband/hw/mlx4/mad.c clean_vf_mcast(&dev->sriov.demux[port - 1], slave); sriov 2123 drivers/infiniband/hw/mlx4/mad.c dev->sriov.sqps[port - 1], 1); sriov 2126 drivers/infiniband/hw/mlx4/mad.c dev->sriov.demux[port - 1].tun[slave], 1); sriov 2132 drivers/infiniband/hw/mlx4/mad.c dev->sriov.demux[port - 1].tun[slave]); sriov 2137 drivers/infiniband/hw/mlx4/mad.c dev->sriov.sqps[port - 1]); sriov 2288 drivers/infiniband/hw/mlx4/mad.c dev->sriov.is_going_down = 0; sriov 2289 drivers/infiniband/hw/mlx4/mad.c spin_lock_init(&dev->sriov.going_down_lock); sriov 2324 drivers/infiniband/hw/mlx4/mad.c dev->sriov.demux[i].guid_cache[0] = gid.global.interface_id; sriov 2325 drivers/infiniband/hw/mlx4/mad.c atomic64_set(&dev->sriov.demux[i].subnet_prefix, sriov 2328 drivers/infiniband/hw/mlx4/mad.c &dev->sriov.sqps[i]); sriov 2331 drivers/infiniband/hw/mlx4/mad.c err = mlx4_ib_alloc_demux_ctx(dev, &dev->sriov.demux[i], i + 1); sriov 2343 drivers/infiniband/hw/mlx4/mad.c mlx4_ib_free_demux_ctx(&dev->sriov.demux[i]); sriov 2364 drivers/infiniband/hw/mlx4/mad.c spin_lock_irqsave(&dev->sriov.going_down_lock, flags); sriov 2365 drivers/infiniband/hw/mlx4/mad.c dev->sriov.is_going_down = 1; sriov 2366 drivers/infiniband/hw/mlx4/mad.c spin_unlock_irqrestore(&dev->sriov.going_down_lock, flags); sriov 2369 drivers/infiniband/hw/mlx4/mad.c flush_workqueue(dev->sriov.demux[i].ud_wq); sriov 2370 drivers/infiniband/hw/mlx4/mad.c mlx4_ib_free_sqp_ctx(dev->sriov.sqps[i]); sriov 2371 drivers/infiniband/hw/mlx4/mad.c kfree(dev->sriov.sqps[i]); sriov 2372 drivers/infiniband/hw/mlx4/mad.c dev->sriov.sqps[i] = NULL; sriov 2373 drivers/infiniband/hw/mlx4/mad.c mlx4_ib_free_demux_ctx(&dev->sriov.demux[i]); sriov 3088 drivers/infiniband/hw/mlx4/main.c spin_lock_irqsave(&ibdev->sriov.going_down_lock, flags); sriov 3089 drivers/infiniband/hw/mlx4/main.c if (!ibdev->sriov.is_going_down) { sriov 3091 drivers/infiniband/hw/mlx4/main.c queue_work(ibdev->sriov.demux[i].ud_wq, &dm[i]->work); sriov 3092 drivers/infiniband/hw/mlx4/main.c spin_unlock_irqrestore(&ibdev->sriov.going_down_lock, flags); sriov 3094 drivers/infiniband/hw/mlx4/main.c spin_unlock_irqrestore(&ibdev->sriov.going_down_lock, flags); sriov 887 drivers/infiniband/hw/mlx4/mcg.c struct mlx4_ib_demux_ctx *ctx = &dev->sriov.demux[port - 1]; sriov 936 drivers/infiniband/hw/mlx4/mcg.c struct mlx4_ib_demux_ctx *ctx = &dev->sriov.demux[port - 1]; sriov 596 drivers/infiniband/hw/mlx4/mlx4_ib.h struct mlx4_ib_sriov sriov; sriov 3121 drivers/infiniband/hw/mlx4/qp.c cpu_to_be64(atomic64_read(&(to_mdev(ib_dev)->sriov. sriov 3125 drivers/infiniband/hw/mlx4/qp.c to_mdev(ib_dev)->sriov.demux[sqp->qp.port - 1]. sriov 87 drivers/infiniband/hw/mlx4/sysfs.c spin_lock_irqsave(&mdev->sriov.alias_guid.ag_work_lock, flags); sriov 89 drivers/infiniband/hw/mlx4/sysfs.c *(__be64 *)&mdev->sriov.alias_guid.ports_guid[port->num - 1]. sriov 95 drivers/infiniband/hw/mlx4/sysfs.c mdev->sriov.alias_guid.ports_guid[port->num - 1].all_rec_per_port[record_num].status sriov 102 drivers/infiniband/hw/mlx4/sysfs.c mdev->sriov.alias_guid.ports_guid[port->num - 1].all_rec_per_port[record_num].guid_indexes sriov 105 drivers/infiniband/hw/mlx4/sysfs.c spin_unlock_irqrestore(&mdev->sriov.alias_guid.ag_work_lock, flags); sriov 99 drivers/infiniband/hw/mlx5/ib_virt.c struct mlx5_vf_context *vfs_ctx = mdev->priv.sriov.vfs_ctx; sriov 157 drivers/infiniband/hw/mlx5/ib_virt.c struct mlx5_vf_context *vfs_ctx = mdev->priv.sriov.vfs_ctx; sriov 178 drivers/infiniband/hw/mlx5/ib_virt.c struct mlx5_vf_context *vfs_ctx = mdev->priv.sriov.vfs_ctx; sriov 3398 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c ether_stat->vf_cnt = IS_SRIOV(bp) ? bp->vfdb->sriov.nr_virtfn : 0; sriov 1055 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c struct bnx2x_sriov *iov = &bp->vfdb->sriov; sriov 1064 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c struct bnx2x_sriov *iov = &bp->vfdb->sriov; sriov 1073 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c struct bnx2x_sriov *iov = &bp->vfdb->sriov; sriov 1242 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c iov = &(bp->vfdb->sriov); sriov 1321 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c for (vf_idx = 0; vf_idx < bp->vfdb->sriov.total; vf_idx++) { sriov 1324 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c bp->vfdb->sriov.first_vf_in_pf + sriov 1327 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c bp->vfdb->sriov.first_vf_in_pf + vf_idx); sriov 1371 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c tot_size = (BP_VFDB(bp)->sriov.first_vf_in_pf + BNX2X_NR_VIRTFN(bp)) * sriov 1557 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c DP(BNX2X_MSG_IOV, "num of vfs: %d\n", (bp)->vfdb->sriov.nr_virtfn); sriov 1566 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c int base_vf_cid = (BP_VFDB(bp)->sriov.first_vf_in_pf + vfid) * sriov 2023 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c int base_vf_cid = (BP_VFDB(bp)->sriov.first_vf_in_pf + vf->index) * sriov 2469 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c first_vf = bp->vfdb->sriov.first_vf_in_pf; sriov 219 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h #define BNX2X_NR_VIRTFN(bp) ((bp)->vfdb->sriov.nr_virtfn) sriov 241 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h #define GET_NUM_VFS_PER_PF(bp) ((bp)->vfdb ? (bp)->vfdb->sriov.total \ sriov 333 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h struct bnx2x_sriov sriov; sriov 541 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h return bp->vfdb->sriov.nr_virtfn * BNX2X_CIDS_PER_VF; sriov 2201 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c if (vfpf_event->vf_id - BP_VFDB(bp)->sriov.first_vf_in_pf > sriov 1373 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c mlx5_eswitch_update_num_of_vfs(esw, esw->dev->priv.sriov.num_vfs); sriov 2594 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c vport_num <= esw->dev->priv.sriov.max_vfs; sriov 41 drivers/net/ethernet/mellanox/mlx5/core/sriov.c struct mlx5_core_sriov *sriov = &dev->priv.sriov; sriov 46 drivers/net/ethernet/mellanox/mlx5/core/sriov.c if (sriov->vfs_ctx[vf].node_guid || sriov 47 drivers/net/ethernet/mellanox/mlx5/core/sriov.c sriov->vfs_ctx[vf].port_guid || sriov 48 drivers/net/ethernet/mellanox/mlx5/core/sriov.c sriov->vfs_ctx[vf].policy != MLX5_POLICY_INVALID) { sriov 53 drivers/net/ethernet/mellanox/mlx5/core/sriov.c in->node_guid = sriov->vfs_ctx[vf].node_guid; sriov 54 drivers/net/ethernet/mellanox/mlx5/core/sriov.c in->port_guid = sriov->vfs_ctx[vf].port_guid; sriov 55 drivers/net/ethernet/mellanox/mlx5/core/sriov.c in->policy = sriov->vfs_ctx[vf].policy; sriov 73 drivers/net/ethernet/mellanox/mlx5/core/sriov.c struct mlx5_core_sriov *sriov = &dev->priv.sriov; sriov 95 drivers/net/ethernet/mellanox/mlx5/core/sriov.c sriov->vfs_ctx[vf].enabled = 1; sriov 113 drivers/net/ethernet/mellanox/mlx5/core/sriov.c struct mlx5_core_sriov *sriov = &dev->priv.sriov; sriov 119 drivers/net/ethernet/mellanox/mlx5/core/sriov.c if (!sriov->vfs_ctx[vf].enabled) sriov 126 drivers/net/ethernet/mellanox/mlx5/core/sriov.c sriov->vfs_ctx[vf].enabled = 0; sriov 166 drivers/net/ethernet/mellanox/mlx5/core/sriov.c struct mlx5_core_sriov *sriov = &dev->priv.sriov; sriov 177 drivers/net/ethernet/mellanox/mlx5/core/sriov.c sriov->num_vfs = num_vfs; sriov 224 drivers/net/ethernet/mellanox/mlx5/core/sriov.c struct mlx5_core_sriov *sriov = &dev->priv.sriov; sriov 232 drivers/net/ethernet/mellanox/mlx5/core/sriov.c sriov->max_vfs = mlx5_get_max_vfs(dev); sriov 233 drivers/net/ethernet/mellanox/mlx5/core/sriov.c sriov->num_vfs = pci_num_vf(pdev); sriov 234 drivers/net/ethernet/mellanox/mlx5/core/sriov.c sriov->vfs_ctx = kcalloc(total_vfs, sizeof(*sriov->vfs_ctx), GFP_KERNEL); sriov 235 drivers/net/ethernet/mellanox/mlx5/core/sriov.c if (!sriov->vfs_ctx) sriov 243 drivers/net/ethernet/mellanox/mlx5/core/sriov.c struct mlx5_core_sriov *sriov = &dev->priv.sriov; sriov 248 drivers/net/ethernet/mellanox/mlx5/core/sriov.c kfree(sriov->vfs_ctx); sriov 530 drivers/net/ethernet/qlogic/qlcnic/qlcnic.h struct qlcnic_sriov *sriov; sriov 1961 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c qlcnic_sriov_cleanup_async_list(&adapter->ahw->sriov->bc); sriov 267 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov.h static inline void qlcnic_sriov_pf_handle_flr(struct qlcnic_sriov *sriov, sriov 143 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c struct qlcnic_sriov *sriov; sriov 153 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c sriov = kzalloc(sizeof(struct qlcnic_sriov), GFP_KERNEL); sriov 154 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c if (!sriov) sriov 157 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c adapter->ahw->sriov = sriov; sriov 158 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c sriov->num_vfs = num_vfs; sriov 159 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c bc = &sriov->bc; sriov 160 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c sriov->vf_info = kcalloc(num_vfs, sizeof(struct qlcnic_vf_info), sriov 162 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c if (!sriov->vf_info) { sriov 191 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c vf = &sriov->vf_info[i]; sriov 210 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c sriov->vf_info[i].vp = vp; sriov 231 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c kfree(sriov->vf_info); sriov 234 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c kfree(adapter->ahw->sriov); sriov 262 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c struct qlcnic_sriov *sriov = adapter->ahw->sriov; sriov 263 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c struct qlcnic_back_channel *bc = &sriov->bc; sriov 273 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c for (i = 0; i < sriov->num_vfs; i++) { sriov 274 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c vf = &sriov->vf_info[i]; sriov 282 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c for (i = 0; i < sriov->num_vfs; i++) sriov 283 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c kfree(sriov->vf_info[i].vp); sriov 285 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c kfree(sriov->vf_info); sriov 286 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c kfree(adapter->ahw->sriov); sriov 435 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c struct qlcnic_sriov *sriov = adapter->ahw->sriov; sriov 439 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c if (sriov->allowed_vlans) sriov 442 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c sriov->any_vlan = cmd->rsp.arg[2] & 0xf; sriov 443 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c sriov->num_allowed_vlans = cmd->rsp.arg[2] >> 16; sriov 445 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c sriov->num_allowed_vlans); sriov 449 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c if (!sriov->any_vlan) sriov 452 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c num_vlans = sriov->num_allowed_vlans; sriov 453 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c sriov->allowed_vlans = kcalloc(num_vlans, sizeof(u16), GFP_KERNEL); sriov 454 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c if (!sriov->allowed_vlans) sriov 459 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c sriov->allowed_vlans[i] = vlans[i]; sriov 466 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c struct qlcnic_sriov *sriov = adapter->ahw->sriov; sriov 480 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c sriov->vlan_mode = cmd.rsp.arg[1] & 0x3; sriov 481 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c switch (sriov->vlan_mode) { sriov 682 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c struct qlcnic_vf_info *vf_info = adapter->ahw->sriov->vf_info; sriov 688 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c for (i = 0; i < adapter->ahw->sriov->num_vfs; i++) { sriov 847 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c static void qlcnic_sriov_schedule_bc_cmd(struct qlcnic_sriov *sriov, sriov 855 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c queue_work(sriov->bc.bc_trans_wq, &vf->trans_work); sriov 1013 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c vf = &adapter->ahw->sriov->vf_info[index]; sriov 1080 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c qlcnic_sriov_schedule_bc_cmd(adapter->ahw->sriov, vf, sriov 1116 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c int __qlcnic_sriov_add_act_list(struct qlcnic_sriov *sriov, sriov 1125 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c qlcnic_sriov_schedule_bc_cmd(sriov, vf, sriov 1130 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c static int qlcnic_sriov_add_act_list(struct qlcnic_sriov *sriov, sriov 1138 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c __qlcnic_sriov_add_act_list(sriov, vf, trans); sriov 1144 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c static void qlcnic_sriov_handle_pending_trans(struct qlcnic_sriov *sriov, sriov 1182 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c if (qlcnic_sriov_add_act_list(sriov, vf, trans)) sriov 1188 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c static void qlcnic_sriov_handle_bc_cmd(struct qlcnic_sriov *sriov, sriov 1208 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c qlcnic_sriov_handle_pending_trans(sriov, vf, hdr); sriov 1250 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c if (qlcnic_sriov_add_act_list(sriov, vf, trans)) { sriov 1262 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c static void qlcnic_sriov_handle_msg_event(struct qlcnic_sriov *sriov, sriov 1275 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c qlcnic_sriov_handle_bc_cmd(sriov, &hdr, vf); sriov 1283 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c static void qlcnic_sriov_handle_flr_event(struct qlcnic_sriov *sriov, sriov 1289 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c qlcnic_sriov_pf_handle_flr(sriov, vf); sriov 1298 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c struct qlcnic_sriov *sriov; sriov 1302 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c sriov = adapter->ahw->sriov; sriov 1309 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c vf = &sriov->vf_info[index]; sriov 1316 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c qlcnic_sriov_handle_flr_event(sriov, vf); sriov 1321 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c qlcnic_sriov_handle_msg_event(sriov, vf); sriov 1379 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c u16 seq = ++adapter->ahw->sriov->bc.trans_counter; sriov 1467 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c struct qlcnic_vf_info *vf = &adapter->ahw->sriov->vf_info[0]; sriov 1499 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c struct qlcnic_sriov *sriov = adapter->ahw->sriov; sriov 1504 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c vf = &adapter->ahw->sriov->vf_info[0]; sriov 1510 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c for (i = 0; i < sriov->num_allowed_vlans; i++) { sriov 1662 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c struct qlcnic_back_channel *bc = &adapter->ahw->sriov->bc; sriov 1947 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c static int qlcnic_sriov_check_vlan_id(struct qlcnic_sriov *sriov, sriov 1957 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c for (i = 0; i < sriov->num_allowed_vlans; i++) { sriov 1968 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c static int qlcnic_sriov_validate_num_vlans(struct qlcnic_sriov *sriov, sriov 1975 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c if (vf->num_vlan >= sriov->num_allowed_vlans) sriov 1985 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c struct qlcnic_sriov *sriov = adapter->ahw->sriov; sriov 1991 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c vf = &adapter->ahw->sriov->vf_info[0]; sriov 1993 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c if (sriov->vlan_mode != QLC_GUEST_VLAN_MODE) sriov 2000 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c if (qlcnic_sriov_validate_num_vlans(sriov, vf)) sriov 2003 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c if (sriov->any_vlan) { sriov 2004 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c for (i = 0; i < sriov->num_allowed_vlans; i++) { sriov 2005 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c if (sriov->allowed_vlans[i] == vid) sriov 2013 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c if (!vlan_exist || qlcnic_sriov_check_vlan_id(sriov, vf, vid)) sriov 2024 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c struct qlcnic_sriov *sriov; sriov 2026 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c sriov = adapter->ahw->sriov; sriov 2035 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c qlcnic_sriov_add_vlan_id(sriov, vf, vlan_id); sriov 2038 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c qlcnic_sriov_del_vlan_id(sriov, vf, vlan_id); sriov 2051 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c struct qlcnic_sriov *sriov = adapter->ahw->sriov; sriov 2061 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c vf = &adapter->ahw->sriov->vf_info[0]; sriov 2073 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c qlcnic_sriov_cleanup_async_list(&sriov->bc); sriov 2165 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c struct qlcnic_sriov *sriov = adapter->ahw->sriov; sriov 2169 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c for (i = 0; i < sriov->num_vfs; i++) { sriov 2170 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c vf = &sriov->vf_info[i]; sriov 2171 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c vf->sriov_vlans = kcalloc(sriov->num_allowed_vlans, sriov 2178 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c struct qlcnic_sriov *sriov = adapter->ahw->sriov; sriov 2182 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c for (i = 0; i < sriov->num_vfs; i++) { sriov 2183 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c vf = &sriov->vf_info[i]; sriov 2189 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c void qlcnic_sriov_add_vlan_id(struct qlcnic_sriov *sriov, sriov 2194 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c for (i = 0; i < sriov->num_allowed_vlans; i++) { sriov 2203 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c void qlcnic_sriov_del_vlan_id(struct qlcnic_sriov *sriov, sriov 2208 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c for (i = 0; i < sriov->num_allowed_vlans; i++) { sriov 71 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c struct qlcnic_sriov *sriov = adapter->ahw->sriov; sriov 72 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c struct qlcnic_resources *res = &sriov->ff_max; sriov 73 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c u16 num_macs = sriov->num_allowed_vlans + 1; sriov 82 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c num_vfs = sriov->num_vfs; sriov 104 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c info->max_tx_ques = res->num_tx_queues - sriov->num_vfs; sriov 109 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c vp = sriov->vf_info[id].vp; sriov 140 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c struct qlcnic_resources *ff_max = &adapter->ahw->sriov->ff_max; sriov 159 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c struct qlcnic_sriov *sriov = adapter->ahw->sriov; sriov 163 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c total_fn = sriov->num_vfs + 1; sriov 166 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c sriov->num_allowed_vlans = temp - 1; sriov 169 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c sriov->num_allowed_vlans = 1; sriov 172 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c sriov->num_allowed_vlans); sriov 236 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c struct qlcnic_sriov *sriov = adapter->ahw->sriov; sriov 241 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c sriov->vp_handle = 0; sriov 246 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c vp = sriov->vf_info[index].vp; sriov 254 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c struct qlcnic_sriov *sriov = adapter->ahw->sriov; sriov 259 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c sriov->vp_handle = vport_handle; sriov 264 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c vp = sriov->vf_info[index].vp; sriov 272 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c struct qlcnic_sriov *sriov = adapter->ahw->sriov; sriov 277 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c return sriov->vp_handle; sriov 281 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c vf_info = &sriov->vf_info[index]; sriov 409 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c struct qlcnic_sriov *sriov = adapter->ahw->sriov; sriov 410 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c struct qlcnic_back_channel *bc = &sriov->bc; sriov 413 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c for (i = 0; i < sriov->num_vfs; i++) sriov 414 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c cancel_work_sync(&sriov->vf_info[i].flr_work); sriov 421 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c struct qlcnic_back_channel *bc = &adapter->ahw->sriov->bc; sriov 694 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c vp = adapter->ahw->sriov->vf_info[id].vp; sriov 745 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c struct qlcnic_sriov *sriov; sriov 751 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c sriov = adapter->ahw->sriov; sriov 763 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c size = size * sriov->num_allowed_vlans; sriov 858 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c struct qlcnic_sriov *sriov; sriov 862 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c sriov = adapter->ahw->sriov; sriov 866 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c for (i = 0; i < sriov->num_allowed_vlans; i++) { sriov 1341 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c struct qlcnic_sriov *sriov; sriov 1344 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c sriov = adapter->ahw->sriov; sriov 1357 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c cmd->rsp.arg[2] = sriov->num_allowed_vlans << 16; sriov 1371 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c struct qlcnic_sriov *sriov = adapter->ahw->sriov; sriov 1379 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c qlcnic_sriov_del_vlan_id(sriov, vf, vlan); sriov 1384 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c qlcnic_sriov_del_vlan_id(sriov, vf, vlan); sriov 1396 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c struct qlcnic_sriov *sriov = adapter->ahw->sriov; sriov 1406 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c qlcnic_sriov_add_vlan_id(sriov, vf, vlan); sriov 1426 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c qlcnic_sriov_add_vlan_id(sriov, vf, vlan); sriov 1646 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c static int qlcnic_sriov_add_act_list_irqsave(struct qlcnic_sriov *sriov, sriov 1655 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c __qlcnic_sriov_add_act_list(sriov, vf, trans); sriov 1678 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c qlcnic_sriov_add_act_list_irqsave(adapter->ahw->sriov, vf, sriov 1694 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c static void qlcnic_sriov_schedule_flr(struct qlcnic_sriov *sriov, sriov 1702 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c queue_work(sriov->bc.bc_flr_wq, &vf->flr_work); sriov 1709 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c struct qlcnic_sriov *sriov = adapter->ahw->sriov; sriov 1715 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c qlcnic_sriov_schedule_flr(sriov, vf, qlcnic_sriov_pf_process_flr); sriov 1736 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c void qlcnic_sriov_pf_handle_flr(struct qlcnic_sriov *sriov, sriov 1755 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c sizeof(*vf->sriov_vlans) * sriov->num_allowed_vlans); sriov 1757 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c qlcnic_sriov_schedule_flr(sriov, vf, qlcnic_sriov_pf_process_flr); sriov 1764 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c struct qlcnic_sriov *sriov = ahw->sriov; sriov 1766 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c u16 num_vfs = sriov->num_vfs; sriov 1770 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c vf = &sriov->vf_info[i]; sriov 1804 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c struct qlcnic_sriov *sriov = adapter->ahw->sriov; sriov 1812 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c num_vfs = sriov->num_vfs; sriov 1823 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c vf_info = &sriov->vf_info[i]; sriov 1832 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c vf_info = &sriov->vf_info[vf]; sriov 1852 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c struct qlcnic_sriov *sriov = adapter->ahw->sriov; sriov 1861 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c if (vf >= sriov->num_vfs) sriov 1864 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c vf_info = &sriov->vf_info[vf]; sriov 1917 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c struct qlcnic_sriov *sriov = adapter->ahw->sriov; sriov 1924 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c if (vf >= sriov->num_vfs || qos > 7) sriov 1937 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c vf_info = &sriov->vf_info[vf]; sriov 1947 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c sizeof(*vf_info->sriov_vlans) * sriov->num_allowed_vlans); sriov 1959 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c qlcnic_sriov_add_vlan_id(sriov, vf_info, vlan); sriov 1996 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c struct qlcnic_sriov *sriov = adapter->ahw->sriov; sriov 2002 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c if (vf >= sriov->num_vfs) sriov 2005 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c vp = sriov->vf_info[vf].vp; sriov 2026 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c struct qlcnic_sriov *sriov = adapter->ahw->sriov; sriov 2033 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c if (vf >= sriov->num_vfs) sriov 2036 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c vf_info = &sriov->vf_info[vf]; sriov 1243 drivers/net/ethernet/sfc/siena_sriov.c unsigned index, devfn, sriov, buftbl_base; sriov 1248 drivers/net/ethernet/sfc/siena_sriov.c sriov = pci_find_ext_capability(pci_dev, PCI_EXT_CAP_ID_SRIOV); sriov 1249 drivers/net/ethernet/sfc/siena_sriov.c if (!sriov) sriov 1252 drivers/net/ethernet/sfc/siena_sriov.c pci_read_config_word(pci_dev, sriov + PCI_SRIOV_VF_OFFSET, &offset); sriov 1253 drivers/net/ethernet/sfc/siena_sriov.c pci_read_config_word(pci_dev, sriov + PCI_SRIOV_VF_STRIDE, &stride); sriov 516 drivers/net/hyperv/hyperv_net.h u64 sriov:1; sriov 529 drivers/net/hyperv/netvsc.c init_packet->msg.v2_msg.send_ndis_config.capability.sriov = 1; sriov 24 drivers/pci/iov.c return dev->bus->number + ((dev->devfn + dev->sriov->offset + sriov 25 drivers/pci/iov.c dev->sriov->stride * vf_id) >> 8); sriov 32 drivers/pci/iov.c return (dev->devfn + dev->sriov->offset + sriov 33 drivers/pci/iov.c dev->sriov->stride * vf_id) & 0xff; sriov 44 drivers/pci/iov.c struct pci_sriov *iov = dev->sriov; sriov 60 drivers/pci/iov.c struct pci_sriov *iov = dev->sriov; sriov 111 drivers/pci/iov.c return dev->sriov->barsz[resno - PCI_IOV_RESOURCES]; sriov 128 drivers/pci/iov.c &physfn->sriov->class); sriov 130 drivers/pci/iov.c &physfn->sriov->hdr_type); sriov 132 drivers/pci/iov.c &physfn->sriov->subsystem_vendor); sriov 134 drivers/pci/iov.c &physfn->sriov->subsystem_device); sriov 145 drivers/pci/iov.c struct pci_sriov *iov = dev->sriov; sriov 257 drivers/pci/iov.c return sprintf(buf, "%u\n", pdev->sriov->num_VFs); sriov 284 drivers/pci/iov.c if (num_vfs == pdev->sriov->num_VFs) sriov 301 drivers/pci/iov.c if (pdev->sriov->num_VFs) { sriov 303 drivers/pci/iov.c pdev->sriov->num_VFs, num_vfs); sriov 331 drivers/pci/iov.c return sprintf(buf, "%u\n", pdev->sriov->offset); sriov 340 drivers/pci/iov.c return sprintf(buf, "%u\n", pdev->sriov->stride); sriov 349 drivers/pci/iov.c return sprintf(buf, "%x\n", pdev->sriov->vf_device); sriov 358 drivers/pci/iov.c return sprintf(buf, "%u\n", pdev->sriov->drivers_autoprobe); sriov 371 drivers/pci/iov.c pdev->sriov->drivers_autoprobe = drivers_autoprobe; sriov 448 drivers/pci/iov.c struct pci_sriov *iov = dev->sriov; sriov 552 drivers/pci/iov.c struct pci_sriov *iov = dev->sriov; sriov 564 drivers/pci/iov.c struct pci_sriov *iov = dev->sriov; sriov 676 drivers/pci/iov.c dev->sriov = iov; sriov 685 drivers/pci/iov.c dev->sriov = NULL; sriov 699 drivers/pci/iov.c BUG_ON(dev->sriov->num_VFs); sriov 701 drivers/pci/iov.c if (dev != dev->sriov->dev) sriov 702 drivers/pci/iov.c pci_dev_put(dev->sriov->dev); sriov 704 drivers/pci/iov.c kfree(dev->sriov); sriov 705 drivers/pci/iov.c dev->sriov = NULL; sriov 712 drivers/pci/iov.c struct pci_sriov *iov = dev->sriov; sriov 772 drivers/pci/iov.c struct pci_sriov *iov = dev->sriov; sriov 791 drivers/pci/iov.c struct pci_sriov *iov = dev->is_physfn ? dev->sriov : NULL; sriov 879 drivers/pci/iov.c dev->sriov->drivers_autoprobe = auto_probe; sriov 897 drivers/pci/iov.c if (dev->sriov->max_VF_buses > max) sriov 898 drivers/pci/iov.c max = dev->sriov->max_VF_buses; sriov 948 drivers/pci/iov.c return dev->sriov->num_VFs; sriov 973 drivers/pci/iov.c dev_id = dev->sriov->vf_device; sriov 1011 drivers/pci/iov.c if (numvfs > dev->sriov->total_VFs) sriov 1015 drivers/pci/iov.c if (dev->sriov->ctrl & PCI_SRIOV_CTRL_VFE) sriov 1018 drivers/pci/iov.c dev->sriov->driver_max_VFs = numvfs; sriov 1036 drivers/pci/iov.c return dev->sriov->driver_max_VFs; sriov 402 drivers/pci/pci-driver.c return (!pdev->is_virtfn || pdev->physfn->sriov->drivers_autoprobe || sriov 1620 drivers/pci/probe.c return dev->physfn->sriov->class; sriov 1630 drivers/pci/probe.c *vendor = dev->physfn->sriov->subsystem_vendor; sriov 1631 drivers/pci/probe.c *device = dev->physfn->sriov->subsystem_device; sriov 1645 drivers/pci/probe.c return dev->physfn->sriov->hdr_type; sriov 1041 drivers/pci/quirks.c dev->sriov->link = dev->devfn; sriov 769 drivers/scsi/bfa/bfa_defs.h u8 sriov; sriov 583 include/linux/mlx5/driver.h struct mlx5_core_sriov sriov; sriov 1143 include/linux/mlx5/driver.h return dev->priv.sriov.max_vfs; sriov 450 include/linux/pci.h struct pci_sriov *sriov; /* PF: SR-IOV info */