Home
last modified time | relevance | path

Searched refs:ib_dev (Results 1 – 37 of 37) sorted by relevance

/linux-4.1.27/drivers/infiniband/hw/usnic/
Dusnic_ib_main.c64 return scnprintf(buf, buf_sz, "PF: %s ", vf->pf->ib_dev.name); in usnic_ib_dump_vf_hdr()
144 usnic_info("PF Reset on %s\n", us_ibdev->ib_dev.name); in usnic_ib_handle_usdev_event()
147 ib_event.device = &us_ibdev->ib_dev; in usnic_ib_handle_usdev_event()
157 usnic_info("Link UP on %s\n", us_ibdev->ib_dev.name); in usnic_ib_handle_usdev_event()
159 ib_event.device = &us_ibdev->ib_dev; in usnic_ib_handle_usdev_event()
165 usnic_info("Link DOWN on %s\n", us_ibdev->ib_dev.name); in usnic_ib_handle_usdev_event()
168 ib_event.device = &us_ibdev->ib_dev; in usnic_ib_handle_usdev_event()
174 us_ibdev->ib_dev.name); in usnic_ib_handle_usdev_event()
181 us_ibdev->ib_dev.name); in usnic_ib_handle_usdev_event()
184 us_ibdev->ib_dev.name, in usnic_ib_handle_usdev_event()
[all …]
Dusnic_ib_sysfs.c38 container_of(device, struct usnic_ib_dev, ib_dev.dev); in usnic_ib_show_fw_ver()
53 container_of(device, struct usnic_ib_dev, ib_dev.dev); in usnic_ib_show_board()
76 us_ibdev = container_of(device, struct usnic_ib_dev, ib_dev.dev); in usnic_ib_show_config()
96 us_ibdev->ib_dev.name, in usnic_ib_show_config()
121 us_ibdev->ib_dev.name); in usnic_ib_show_config()
135 us_ibdev = container_of(device, struct usnic_ib_dev, ib_dev.dev); in usnic_ib_show_iface()
147 us_ibdev = container_of(device, struct usnic_ib_dev, ib_dev.dev); in usnic_ib_show_max_vf()
160 us_ibdev = container_of(device, struct usnic_ib_dev, ib_dev.dev); in usnic_ib_show_qp_per_vf()
174 us_ibdev = container_of(device, struct usnic_ib_dev, ib_dev.dev); in usnic_ib_show_cq_per_vf()
285 err = device_create_file(&us_ibdev->ib_dev.dev, in usnic_ib_sysfs_register_usdev()
[all …]
Dusnic_ib.h55 struct ib_device ib_dev; member
84 return container_of(ibdev, struct usnic_ib_dev, ib_dev); in to_usdev()
Dusnic_ib_verbs.c122 usnic_err("Failed to copy udata for %s", us_ibdev->ib_dev.name); in usnic_ib_fill_create_qp_resp()
158 us_ibdev->ib_dev.name, in find_free_vf_and_create_qp_grp()
186 us_ibdev->ib_dev.name); in find_free_vf_and_create_qp_grp()
475 us_ibdev->ib_dev.name); in usnic_ib_create_qp()
482 us_ibdev->ib_dev.name); in usnic_ib_create_qp()
488 us_ibdev->ib_dev.name, init_attr->qp_type); in usnic_ib_create_qp()
Dusnic_ib_qp_grp.c477 ib_event.device = &qp_grp->vf->pf->ib_dev; in usnic_ib_qp_grp_modify()
/linux-4.1.27/drivers/target/
Dtarget_core_iblock.c77 struct iblock_dev *ib_dev = NULL; in iblock_alloc_device() local
79 ib_dev = kzalloc(sizeof(struct iblock_dev), GFP_KERNEL); in iblock_alloc_device()
80 if (!ib_dev) { in iblock_alloc_device()
87 return &ib_dev->dev; in iblock_alloc_device()
92 struct iblock_dev *ib_dev = IBLOCK_DEV(dev); in iblock_configure_device() local
99 if (!(ib_dev->ibd_flags & IBDF_HAS_UDEV_PATH)) { in iblock_configure_device()
104 ib_dev->ibd_bio_set = bioset_create(IBLOCK_BIO_POOL_SIZE, 0); in iblock_configure_device()
105 if (!ib_dev->ibd_bio_set) { in iblock_configure_device()
111 ib_dev->ibd_udev_path); in iblock_configure_device()
114 if (!ib_dev->ibd_readonly) in iblock_configure_device()
[all …]
/linux-4.1.27/drivers/infiniband/hw/mlx5/
Dmain.c750 memcpy(dev->ib_dev.node_desc, out_mad->data, 64); in init_node_data()
759 memcpy(&dev->ib_dev.node_guid, out_mad->data + 12, 8); in init_node_data()
771 container_of(device, struct mlx5_ib_dev, ib_dev.dev); in show_fw_pages()
780 container_of(device, struct mlx5_ib_dev, ib_dev.dev); in show_reg_pages()
789 container_of(device, struct mlx5_ib_dev, ib_dev.dev); in show_hca()
797 container_of(device, struct mlx5_ib_dev, ib_dev.dev); in show_fw_ver()
806 container_of(device, struct mlx5_ib_dev, ib_dev.dev); in show_rev()
814 container_of(device, struct mlx5_ib_dev, ib_dev.dev); in show_board()
884 ibev.device = &ibdev->ib_dev; in mlx5_ib_event()
923 err = mlx5_ib_query_device(&dev->ib_dev, dprops); in get_port_caps()
[all …]
Dmlx5_ib.h47 pr_debug("%s:%s:%d:(pid %d): " format, (dev)->ib_dev.name, __func__, \
51 pr_err("%s:%s:%d:(pid %d): " format, (dev)->ib_dev.name, __func__, \
55 pr_warn("%s:%s:%d:(pid %d): " format, (dev)->ib_dev.name, __func__, \
421 struct ib_device ib_dev; member
458 return container_of(ibdev, struct mlx5_ib_dev, ib_dev); in to_mdev()
Dmr.c756 struct device *ddev = dev->ib_dev.dma_device; in reg_umr()
851 struct device *ddev = dev->ib_dev.dma_device; in mlx5_ib_update_mtt()
Dcq.c60 event.device = &dev->ib_dev; in mlx5_ib_cq_event()
/linux-4.1.27/drivers/infiniband/hw/ehca/
Dehca_tools.h69 #define ehca_dbg(ib_dev, format, arg...) \ argument
72 dev_printk(KERN_DEBUG, (ib_dev)->dma_device, \
78 #define ehca_info(ib_dev, format, arg...) \ argument
79 dev_info((ib_dev)->dma_device, "PU%04x EHCA_INFO:%s " format "\n", \
82 #define ehca_warn(ib_dev, format, arg...) \ argument
83 dev_warn((ib_dev)->dma_device, "PU%04x EHCA_WARN:%s " format "\n", \
86 #define ehca_err(ib_dev, format, arg...) \ argument
87 dev_err((ib_dev)->dma_device, "PU%04x EHCA_ERR:%s " format "\n", \
Dehca_eq.c62 struct ib_device *ib_dev = &shca->ib_device; in ehca_create_eq() local
69 ehca_err(ib_dev, "Invalid EQ type %x. eq=%p", type, eq); in ehca_create_eq()
73 ehca_err(ib_dev, "EQ length must not be zero. eq=%p", eq); in ehca_create_eq()
86 ehca_err(ib_dev, "Can't allocate EQ/NEQ. eq=%p", eq); in ehca_create_eq()
93 ehca_err(ib_dev, "Can't allocate EQ pages eq=%p", eq); in ehca_create_eq()
131 ehca_err(ib_dev, "Can't map interrupt handler."); in ehca_create_eq()
139 ehca_err(ib_dev, "Can't map interrupt handler."); in ehca_create_eq()
Dehca_qp.c291 struct ib_device *ib_dev = &shca->ib_device; in init_qp_queue() local
310 ehca_err(ib_dev, "Cannot allocate page for queue. ipz_rc=%i", in init_qp_queue()
319 ehca_err(ib_dev, "ipz_qpageit_get_inc() " in init_qp_queue()
333 ehca_err(ib_dev, "hipz_qp_register_rpage() " in init_qp_queue()
340 ehca_err(ib_dev, "ipz_qpageit_get_inc() " in init_qp_queue()
347 ehca_err(ib_dev, "hipz_qp_register_rpage() " in init_qp_queue()
/linux-4.1.27/drivers/infiniband/hw/mthca/
Dmthca_provider.c1155 container_of(device, struct mthca_dev, ib_dev.dev); in show_rev()
1163 container_of(device, struct mthca_dev, ib_dev.dev); in show_fw_ver()
1173 container_of(device, struct mthca_dev, ib_dev.dev); in show_hca()
1193 container_of(device, struct mthca_dev, ib_dev.dev); in show_board()
1228 memcpy(dev->ib_dev.node_desc, out_mad->data, 64); in mthca_init_node_data()
1239 memcpy(&dev->ib_dev.node_guid, out_mad->data + 12, 8); in mthca_init_node_data()
1256 strlcpy(dev->ib_dev.name, "mthca%d", IB_DEVICE_NAME_MAX); in mthca_register_device()
1257 dev->ib_dev.owner = THIS_MODULE; in mthca_register_device()
1259 dev->ib_dev.uverbs_abi_ver = MTHCA_UVERBS_ABI_VERSION; in mthca_register_device()
1260 dev->ib_dev.uverbs_cmd_mask = in mthca_register_device()
[all …]
Dmthca_mad.c59 ret = ib_query_port(&dev->ib_dev, port_num, tprops); in mthca_update_rate()
62 ret, dev->ib_dev.name, port_num); in mthca_update_rate()
294 agent = ib_register_mad_agent(&dev->ib_dev, p + 1, in mthca_create_agents()
Dmthca_catas.c94 event.device = &dev->ib_dev; in handle_catas()
Dmthca_dev.h290 struct ib_device ib_dev; member
588 return container_of(ibdev, struct mthca_dev, ib_dev); in to_mdev()
Dmthca_av.c281 ib_get_cached_gid(&dev->ib_dev, in mthca_read_ah()
Dmthca_main.c1073 ib_dealloc_device(&mdev->ib_dev); in __mthca_init_one()
1118 ib_dealloc_device(&mdev->ib_dev); in __mthca_remove_one()
Dmthca_srq.c442 event.device = &dev->ib_dev; in mthca_srq_event()
Dmthca_qp.c258 event.device = &dev->ib_dev; in mthca_qp_event()
1521 ib_get_cached_pkey(&dev->ib_dev, sqp->qp.port, in build_mlx_header()
1524 ib_get_cached_pkey(&dev->ib_dev, sqp->qp.port, in build_mlx_header()
Dmthca_eq.c253 record.device = &dev->ib_dev; in port_change()
Dmthca_cq.c257 event.device = &dev->ib_dev; in mthca_cq_event()
/linux-4.1.27/drivers/infiniband/hw/mlx4/
Dmain.c1428 memcpy(dev->ib_dev.node_desc, out_mad->data, 64); in init_node_data()
1437 memcpy(&dev->ib_dev.node_guid, out_mad->data + 12, 8); in init_node_data()
1449 container_of(device, struct mlx4_ib_dev, ib_dev.dev); in show_hca()
1457 container_of(device, struct mlx4_ib_dev, ib_dev.dev); in show_fw_ver()
1467 container_of(device, struct mlx4_ib_dev, ib_dev.dev); in show_rev()
1475 container_of(device, struct mlx4_ib_dev, ib_dev.dev); in show_board()
1564 if (mlx4_ib_port_link_layer(&gw->dev->ib_dev, gw->port) == in reset_gids_task()
1878 if (rdma_port_get_link_layer(&ibdev->ib_dev, i) == in mlx4_ib_init_gid_table()
2092 ibdev->ib_dev.num_comp_vectors = total_eqs; in mlx4_ib_alloc_eqs()
2104 ibdev->ib_dev.num_comp_vectors = dev->caps.num_comp_vectors; in mlx4_ib_free_eqs()
[all …]
Dmad.c428 return ib_find_cached_pkey(&dev->ib_dev, port, pkey, ix); in find_slave_port_pkey_ix()
438 ret = ib_get_cached_pkey(&dev->ib_dev, port, pkey_ix, &slot_pkey); in find_slave_port_pkey_ix()
497 ret = ib_get_cached_pkey(&dev->ib_dev, port, wc->pkey_index, &cached_pkey); in mlx4_ib_send_to_slave()
540 ib_dma_sync_single_for_cpu(&dev->ib_dev, in mlx4_ib_send_to_slave()
582 ib_dma_sync_single_for_device(&dev->ib_dev, in mlx4_ib_send_to_slave()
900 ll = rdma_port_get_link_layer(&dev->ib_dev, p + 1); in mlx4_ib_mad_init()
903 agent = ib_register_mad_agent(&dev->ib_dev, p + 1, in mlx4_ib_mad_init()
991 mlx4_ib_warn(&dev->ib_dev, "failed to allocate memory for guid info mads\n"); in handle_slaves_guid_change()
1013 mlx4_ib_warn(&dev->ib_dev, "Failed in get GUID INFO MAD_IFC\n"); in handle_slaves_guid_change()
1034 struct mlx4_ib_dev *dev = ew->ib_dev; in handle_port_mgmt_change_event()
[all …]
Dsysfs.c122 ret = __mlx4_ib_query_gid(&mdev->ib_dev, port->num, in show_port_gid()
149 ret = __mlx4_ib_query_pkey(&mdev->ib_dev, port->num, in show_phys_port_pkey()
230 ret = __mlx4_ib_query_port(&device->ib_dev, port_num, &attr, 1); in add_port_entries()
593 int is_eth = rdma_port_get_link_layer(&p->dev->ib_dev, p->port_num) == in add_vf_smi_entries()
628 int is_eth = rdma_port_get_link_layer(&p->dev->ib_dev, p->port_num) == in remove_vf_smi_entries()
643 int is_eth = rdma_port_get_link_layer(&dev->ib_dev, port_num) == in add_port()
822 kobject_get(dev->ib_dev.ports_parent->parent)); in mlx4_ib_device_register_sysfs()
835 for (i = 1; i <= dev->ib_dev.phys_port_cnt; ++i) { in mlx4_ib_device_register_sysfs()
852 kobject_put(dev->ib_dev.ports_parent->parent); in mlx4_ib_device_register_sysfs()
888 kobject_put(device->ib_dev.ports_parent->parent); in mlx4_ib_device_unregister_sysfs()
Dqp.c1057 free_proxy_bufs(&dev->ib_dev, qp); in destroy_qp_common()
1277 int is_eth = rdma_port_get_link_layer(&dev->ib_dev, port) == in _mlx4_set_path()
1462 rdma_port_get_link_layer(&dev->ib_dev, qp->port) == in __mlx4_ib_modify_qp()
1693 if (rdma_port_get_link_layer(&dev->ib_dev, qp->port) == in __mlx4_ib_modify_qp()
1725 &dev->ib_dev, qp->port) == in __mlx4_ib_modify_qp()
1933 ll = rdma_port_get_link_layer(&dev->ib_dev, port); in mlx4_ib_modify_qp()
1972 (rdma_port_get_link_layer(&dev->ib_dev, attr->port_num) != in mlx4_ib_modify_qp()
2038 struct ib_device *ib_dev = &mdev->ib_dev; in build_sriov_qp0_header() local
2081 ib_get_cached_pkey(ib_dev, sqp->qp.port, 0, &pkey); in build_sriov_qp0_header()
2159 struct ib_device *ib_dev = sqp->qp.ibqp.device; in build_mlx_header() local
[all …]
Dmlx4_ib.h414 struct ib_device *ib_dev; member
424 struct ib_device *ib_dev; member
507 struct ib_device ib_dev; member
547 struct mlx4_ib_dev *ib_dev; member
560 return container_of(ibdev, struct mlx4_ib_dev, ib_dev); in to_mdev()
Dalias_GUID.c352 mlx4_ib_warn(&dev->ib_dev, in aliasguid_query_handler()
366 mlx4_ib_warn(&dev->ib_dev, "%s: Failed to set" in aliasguid_query_handler()
765 set_guid_rec(&dev->ib_dev, rec); in alias_guid_work()
848 if (dev->ib_dev.query_gid(&dev->ib_dev , i, 0, &gid)) { in mlx4_ib_init_alias_guid_service()
Dcm.c181 found_ent = id_map_find_by_sl_id(&dev->ib_dev, ent->slave_id, ent->sl_cm_id); in id_map_ent_timeout()
Dmcg.c238 if (ib_find_cached_pkey(&dev->ib_dev, ctx->port, IB_DEFAULT_PKEY_FULL, &wc.pkey_index)) in send_mad_to_slave()
/linux-4.1.27/drivers/infiniband/ulp/isert/
Dib_isert.c97 isert_query_device(struct ib_device *ib_dev, struct ib_device_attr *devattr) in isert_query_device() argument
101 ret = ib_query_device(ib_dev, devattr); in isert_query_device()
212 struct ib_device *ib_dev = device->ib_device; in isert_alloc_rx_descriptors() local
226 dma_addr = ib_dma_map_single(ib_dev, (void *)rx_desc, in isert_alloc_rx_descriptors()
228 if (ib_dma_mapping_error(ib_dev, dma_addr)) in isert_alloc_rx_descriptors()
246 ib_dma_unmap_single(ib_dev, rx_desc->dma_addr, in isert_alloc_rx_descriptors()
260 struct ib_device *ib_dev = isert_conn->device->ib_device; in isert_free_rx_descriptors() local
269 ib_dma_unmap_single(ib_dev, rx_desc->dma_addr, in isert_free_rx_descriptors()
659 struct ib_device *ib_dev = isert_conn->device->ib_device; in isert_free_login_buf() local
661 ib_dma_unmap_single(ib_dev, isert_conn->login_rsp_dma, in isert_free_login_buf()
[all …]
/linux-4.1.27/drivers/infiniband/core/
Duser_mad.c95 struct ib_device *ib_dev; member
628 if (!file->port->ib_dev) { in ib_umad_reg_agent()
677 agent = ib_register_mad_agent(file->port->ib_dev, file->port->port_num, in ib_umad_reg_agent()
730 if (!file->port->ib_dev) { in ib_umad_reg_agent2()
793 agent = ib_register_mad_agent(file->port->ib_dev, file->port->port_num, in ib_umad_reg_agent2()
932 if (!port->ib_dev) in ib_umad_open()
1038 ret = ib_modify_port(port->ib_dev, port->port_num, 0, &props); in ib_umad_sm_open()
1054 ib_modify_port(port->ib_dev, port->port_num, 0, &props); in ib_umad_sm_open()
1072 if (port->ib_dev) in ib_umad_sm_close()
1073 ret = ib_modify_port(port->ib_dev, port->port_num, 0, &props); in ib_umad_sm_close()
[all …]
Duverbs_main.c321 module_put(file->device->ib_dev->owner); in ib_uverbs_release_file()
647 if (!(file->device->ib_dev->uverbs_cmd_mask & (1ull << command))) in ib_uverbs_write()
680 if (!(file->device->ib_dev->uverbs_ex_cmd_mask & (1ull << command))) in ib_uverbs_write()
740 return file->device->ib_dev->mmap(file->ucontext, vma); in ib_uverbs_mmap()
763 if (!try_module_get(dev->ib_dev->owner)) { in ib_uverbs_open()
786 module_put(dev->ib_dev->owner); in ib_uverbs_open()
842 return sprintf(buf, "%s\n", dev->ib_dev->name); in show_ibdev()
854 return sprintf(buf, "%d\n", dev->ib_dev->uverbs_abi_ver); in show_dev_abi_version()
927 uverbs_dev->ib_dev = device; in ib_uverbs_add_one()
Duverbs_cmd.c291 struct ib_device *ibdev = file->device->ib_dev; in ib_uverbs_get_context()
372 INIT_IB_EVENT_HANDLER(&file->event_handler, file->device->ib_dev, in ib_uverbs_get_context()
408 resp->node_guid = file->device->ib_dev->node_guid; in copy_query_dev_fields()
446 resp->phys_port_cnt = file->device->ib_dev->phys_port_cnt; in copy_query_dev_fields()
464 ret = ib_query_device(file->device->ib_dev, &attr); in ib_uverbs_query_device()
493 ret = ib_query_port(file->device->ib_dev, cmd.port_num, &attr); in ib_uverbs_query_port()
518 resp.link_layer = rdma_port_get_link_layer(file->device->ib_dev, in ib_uverbs_query_port()
556 pd = file->device->ib_dev->alloc_pd(file->device->ib_dev, in ib_uverbs_alloc_pd()
563 pd->device = file->device->ib_dev; in ib_uverbs_alloc_pd()
781 xrcd = file->device->ib_dev->alloc_xrcd(file->device->ib_dev, in ib_uverbs_open_xrcd()
[all …]
Duverbs.h92 struct ib_device *ib_dev; member
Ducm.c64 struct ib_device *ib_dev; member
492 ctx->cm_id = ib_create_cm_id(file->device->ib_dev, in ib_ucm_create_id()
1228 return sprintf(buf, "%s\n", ucm_dev->ib_dev->name); in show_ibdev()
1268 ucm_dev->ib_dev = device; in ib_ucm_add_one()