ibdev 200 drivers/infiniband/core/cma_configfs.c struct ib_device *ibdev; ibdev 206 drivers/infiniband/core/cma_configfs.c ibdev = cma_get_ib_dev(cma_dev); ibdev 208 drivers/infiniband/core/cma_configfs.c if (!ibdev) ibdev 211 drivers/infiniband/core/cma_configfs.c ports_num = ibdev->phys_port_cnt; ibdev 82 drivers/infiniband/core/core_priv.h int ib_device_rename(struct ib_device *ibdev, const char *name); ibdev 83 drivers/infiniband/core/core_priv.h int ib_device_set_dim(struct ib_device *ibdev, u8 use_dim); ibdev 118 drivers/infiniband/core/core_priv.h int ib_get_client_nl_info(struct ib_device *ibdev, const char *client_name, ibdev 196 drivers/infiniband/core/device.c static void __ibdev_printk(const char *level, const struct ib_device *ibdev, ibdev 199 drivers/infiniband/core/device.c if (ibdev && ibdev->dev.parent) ibdev 201 drivers/infiniband/core/device.c ibdev->dev.parent, ibdev 203 drivers/infiniband/core/device.c dev_driver_string(ibdev->dev.parent), ibdev 204 drivers/infiniband/core/device.c dev_name(ibdev->dev.parent), ibdev 205 drivers/infiniband/core/device.c dev_name(&ibdev->dev), ibdev 207 drivers/infiniband/core/device.c else if (ibdev) ibdev 209 drivers/infiniband/core/device.c level, dev_name(&ibdev->dev), vaf); ibdev 214 drivers/infiniband/core/device.c void ibdev_printk(const char *level, const struct ib_device *ibdev, ibdev 225 drivers/infiniband/core/device.c __ibdev_printk(level, ibdev, &vaf); ibdev 232 drivers/infiniband/core/device.c void func(const struct ib_device *ibdev, const char *fmt, ...) \ ibdev 242 drivers/infiniband/core/device.c __ibdev_printk(level, ibdev, &vaf); \ ibdev 404 drivers/infiniband/core/device.c int ib_device_rename(struct ib_device *ibdev, const char *name) ibdev 411 drivers/infiniband/core/device.c if (!strcmp(name, dev_name(&ibdev->dev))) { ibdev 421 drivers/infiniband/core/device.c ret = device_rename(&ibdev->dev, name); ibdev 427 drivers/infiniband/core/device.c strlcpy(ibdev->name, name, IB_DEVICE_NAME_MAX); ibdev 428 drivers/infiniband/core/device.c ret = rename_compat_devs(ibdev); ibdev 431 drivers/infiniband/core/device.c down_read(&ibdev->client_data_rwsem); ibdev 432 drivers/infiniband/core/device.c xan_for_each_marked(&ibdev->client_data, index, client_data, ibdev 439 drivers/infiniband/core/device.c client->rename(ibdev, client_data); ibdev 441 drivers/infiniband/core/device.c up_read(&ibdev->client_data_rwsem); ibdev 446 drivers/infiniband/core/device.c int ib_device_set_dim(struct ib_device *ibdev, u8 use_dim) ibdev 450 drivers/infiniband/core/device.c ibdev->use_cq_dim = use_dim; ibdev 455 drivers/infiniband/core/device.c static int alloc_name(struct ib_device *ibdev, const char *name) ibdev 485 drivers/infiniband/core/device.c rc = dev_set_name(&ibdev->dev, name, rc); ibdev 1835 drivers/infiniband/core/device.c static int __ib_get_client_nl_info(struct ib_device *ibdev, ibdev 1843 drivers/infiniband/core/device.c down_read(&ibdev->client_data_rwsem); ibdev 1844 drivers/infiniband/core/device.c xan_for_each_marked (&ibdev->client_data, index, client_data, ibdev 1854 drivers/infiniband/core/device.c ret = client->get_nl_info(ibdev, client_data, res); ibdev 1867 drivers/infiniband/core/device.c up_read(&ibdev->client_data_rwsem); ibdev 1878 drivers/infiniband/core/device.c int ib_get_client_nl_info(struct ib_device *ibdev, const char *client_name, ibdev 1883 drivers/infiniband/core/device.c if (ibdev) ibdev 1884 drivers/infiniband/core/device.c ret = __ib_get_client_nl_info(ibdev, client_name, res); ibdev 1890 drivers/infiniband/core/device.c if (ibdev) ibdev 1891 drivers/infiniband/core/device.c ret = __ib_get_client_nl_info(ibdev, client_name, res); ibdev 1536 drivers/infiniband/core/nldev.c struct ib_device *ibdev = NULL; ibdev 1551 drivers/infiniband/core/nldev.c ibdev = ib_device_get_by_index(sock_net(skb->sk), index); ibdev 1552 drivers/infiniband/core/nldev.c if (!ibdev) ibdev 1557 drivers/infiniband/core/nldev.c if (!rdma_is_port_valid(ibdev, data.port)) { ibdev 1579 drivers/infiniband/core/nldev.c err = ib_get_client_nl_info(ibdev, client_name, &data); ibdev 1600 drivers/infiniband/core/nldev.c if (ibdev) ibdev 1601 drivers/infiniband/core/nldev.c ib_device_put(ibdev); ibdev 1609 drivers/infiniband/core/nldev.c if (ibdev) ibdev 1610 drivers/infiniband/core/nldev.c ib_device_put(ibdev); ibdev 184 drivers/infiniband/core/rdma_core.h struct uverbs_api *uverbs_alloc_api(struct ib_device *ibdev); ibdev 58 drivers/infiniband/core/sysfs.c struct ib_device *ibdev; ibdev 164 drivers/infiniband/core/sysfs.c ret = ib_query_port(p->ibdev, p->port_num, &attr); ibdev 179 drivers/infiniband/core/sysfs.c ret = ib_query_port(p->ibdev, p->port_num, &attr); ibdev 193 drivers/infiniband/core/sysfs.c ret = ib_query_port(p->ibdev, p->port_num, &attr); ibdev 206 drivers/infiniband/core/sysfs.c ret = ib_query_port(p->ibdev, p->port_num, &attr); ibdev 219 drivers/infiniband/core/sysfs.c ret = ib_query_port(p->ibdev, p->port_num, &attr); ibdev 232 drivers/infiniband/core/sysfs.c ret = ib_query_port(p->ibdev, p->port_num, &attr); ibdev 247 drivers/infiniband/core/sysfs.c ret = ib_query_port(p->ibdev, p->port_num, &attr); ibdev 317 drivers/infiniband/core/sysfs.c ret = ib_query_port(p->ibdev, p->port_num, &attr); ibdev 328 drivers/infiniband/core/sysfs.c switch (rdma_port_get_link_layer(p->ibdev, p->port_num)) { ibdev 388 drivers/infiniband/core/sysfs.c gid_attr = rdma_get_gid_attr(p->ibdev, p->port_num, tab_attr->index); ibdev 405 drivers/infiniband/core/sysfs.c gid_attr = rdma_get_gid_attr(p->ibdev, p->port_num, tab_attr->index); ibdev 447 drivers/infiniband/core/sysfs.c ret = ib_query_pkey(p->ibdev, p->port_num, tab_attr->index, &pkey); ibdev 528 drivers/infiniband/core/sysfs.c ret = get_perf_mad(p->ibdev, p->port_num, tab_attr->attr_id, &data, ibdev 838 drivers/infiniband/core/sysfs.c dev = port->ibdev; ibdev 1056 drivers/infiniband/core/sysfs.c p->ibdev = device; ibdev 237 drivers/infiniband/core/umem.c umem->ibdev = context->device; ibdev 340 drivers/infiniband/core/umem.c __ib_umem_release(umem->ibdev, umem, 1); ibdev 99 drivers/infiniband/core/umem_odp.c umem_odp->umem.ibdev->ops.invalidate_range( ibdev 112 drivers/infiniband/core/umem_odp.c item->umem.ibdev->ops.invalidate_range(item, start, end); ibdev 315 drivers/infiniband/core/umem_odp.c umem->ibdev = context->device; ibdev 357 drivers/infiniband/core/umem_odp.c umem->ibdev = root->umem.ibdev; ibdev 409 drivers/infiniband/core/umem_odp.c umem_odp->umem.ibdev = context->device; ibdev 509 drivers/infiniband/core/umem_odp.c struct ib_device *dev = umem_odp->umem.ibdev; ibdev 722 drivers/infiniband/core/umem_odp.c struct ib_device *dev = umem_odp->umem.ibdev; ibdev 1132 drivers/infiniband/core/user_mad.c static struct ib_umad_port *get_port(struct ib_device *ibdev, ibdev 1138 drivers/infiniband/core/user_mad.c if (!rdma_is_port_valid(ibdev, port)) ibdev 1140 drivers/infiniband/core/user_mad.c if (!rdma_cap_ib_mad(ibdev, port)) ibdev 1143 drivers/infiniband/core/user_mad.c return &umad_dev->ports[port - rdma_start_port(ibdev)]; ibdev 1146 drivers/infiniband/core/user_mad.c static int ib_umad_get_nl_info(struct ib_device *ibdev, void *client_data, ibdev 1149 drivers/infiniband/core/user_mad.c struct ib_umad_port *port = get_port(ibdev, client_data, res->port); ibdev 1167 drivers/infiniband/core/user_mad.c static int ib_issm_get_nl_info(struct ib_device *ibdev, void *client_data, ibdev 1170 drivers/infiniband/core/user_mad.c struct ib_umad_port *port = get_port(ibdev, client_data, res->port); ibdev 1196 drivers/infiniband/core/user_mad.c static DEVICE_ATTR_RO(ibdev); ibdev 1154 drivers/infiniband/core/uverbs_main.c static int ib_uverbs_get_nl_info(struct ib_device *ibdev, void *client_data, ibdev 1163 drivers/infiniband/core/uverbs_main.c res->abi = ibdev->ops.uverbs_abi_ver; ibdev 1172 drivers/infiniband/core/uverbs_main.c if (!ibdev->ops.uverbs_no_driver_id_binding) { ibdev 1174 drivers/infiniband/core/uverbs_main.c ibdev->ops.driver_id); ibdev 1207 drivers/infiniband/core/uverbs_main.c static DEVICE_ATTR_RO(ibdev); ibdev 58 drivers/infiniband/core/uverbs_uapi.c struct ib_device *ibdev, ibdev 83 drivers/infiniband/core/uverbs_uapi.c method_elm->disabled = !(ibdev->uverbs_ex_cmd_mask & ibdev 86 drivers/infiniband/core/uverbs_uapi.c method_elm->disabled = !(ibdev->uverbs_cmd_mask & ibdev 273 drivers/infiniband/core/uverbs_uapi.c static int uapi_merge_def(struct uverbs_api *uapi, struct ib_device *ibdev, ibdev 289 drivers/infiniband/core/uverbs_uapi.c rc = uapi_merge_def(uapi, ibdev, def->chain, is_driver); ibdev 311 drivers/infiniband/core/uverbs_uapi.c (void *)(&ibdev->ops) + def->needs_fn_offset; ibdev 323 drivers/infiniband/core/uverbs_uapi.c if (def->func_is_supported(ibdev)) ibdev 344 drivers/infiniband/core/uverbs_uapi.c uapi, ibdev, def, cur_obj_key, &cur_method_key); ibdev 640 drivers/infiniband/core/uverbs_uapi.c struct uverbs_api *uverbs_alloc_api(struct ib_device *ibdev) ibdev 650 drivers/infiniband/core/uverbs_uapi.c uapi->driver_id = ibdev->ops.driver_id; ibdev 652 drivers/infiniband/core/uverbs_uapi.c rc = uapi_merge_def(uapi, ibdev, uverbs_core_api, false); ibdev 655 drivers/infiniband/core/uverbs_uapi.c rc = uapi_merge_def(uapi, ibdev, ibdev->driver_def, true); ibdev 667 drivers/infiniband/core/uverbs_uapi.c dev_err(&ibdev->dev, ibdev 113 drivers/infiniband/hw/bnxt_re/bnxt_re.h struct ib_device ibdev; ibdev 187 drivers/infiniband/hw/bnxt_re/bnxt_re.h return &rdev->ibdev.dev; ibdev 115 drivers/infiniband/hw/bnxt_re/hw_counters.c int bnxt_re_ib_get_hw_stats(struct ib_device *ibdev, ibdev 119 drivers/infiniband/hw/bnxt_re/hw_counters.c struct bnxt_re_dev *rdev = to_bnxt_re_dev(ibdev, ibdev); ibdev 237 drivers/infiniband/hw/bnxt_re/hw_counters.c struct rdma_hw_stats *bnxt_re_ib_alloc_hw_stats(struct ib_device *ibdev, ibdev 99 drivers/infiniband/hw/bnxt_re/hw_counters.h struct rdma_hw_stats *bnxt_re_ib_alloc_hw_stats(struct ib_device *ibdev, ibdev 101 drivers/infiniband/hw/bnxt_re/hw_counters.h int bnxt_re_ib_get_hw_stats(struct ib_device *ibdev, ibdev 122 drivers/infiniband/hw/bnxt_re/ib_verbs.c int bnxt_re_query_device(struct ib_device *ibdev, ibdev 126 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_dev *rdev = to_bnxt_re_dev(ibdev, ibdev); ibdev 194 drivers/infiniband/hw/bnxt_re/ib_verbs.c int bnxt_re_modify_device(struct ib_device *ibdev, ibdev 213 drivers/infiniband/hw/bnxt_re/ib_verbs.c int bnxt_re_query_port(struct ib_device *ibdev, u8 port_num, ibdev 216 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_dev *rdev = to_bnxt_re_dev(ibdev, ibdev); ibdev 253 drivers/infiniband/hw/bnxt_re/ib_verbs.c int bnxt_re_get_port_immutable(struct ib_device *ibdev, u8 port_num, ibdev 258 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (bnxt_re_query_port(ibdev, port_num, &port_attr)) ibdev 269 drivers/infiniband/hw/bnxt_re/ib_verbs.c void bnxt_re_query_fw_str(struct ib_device *ibdev, char *str) ibdev 271 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_dev *rdev = to_bnxt_re_dev(ibdev, ibdev); ibdev 278 drivers/infiniband/hw/bnxt_re/ib_verbs.c int bnxt_re_query_pkey(struct ib_device *ibdev, u8 port_num, ibdev 281 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_dev *rdev = to_bnxt_re_dev(ibdev, ibdev); ibdev 290 drivers/infiniband/hw/bnxt_re/ib_verbs.c int bnxt_re_query_gid(struct ib_device *ibdev, u8 port_num, ibdev 293 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_dev *rdev = to_bnxt_re_dev(ibdev, ibdev); ibdev 308 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_dev *rdev = to_bnxt_re_dev(attr->device, ibdev); ibdev 363 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_dev *rdev = to_bnxt_re_dev(attr->device, ibdev); ibdev 397 drivers/infiniband/hw/bnxt_re/ib_verbs.c enum rdma_link_layer bnxt_re_get_link_layer(struct ib_device *ibdev, ibdev 570 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct ib_device *ibdev = ibpd->device; ibdev 571 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_dev *rdev = to_bnxt_re_dev(ibdev, ibdev); ibdev 908 drivers/infiniband/hw/bnxt_re/ib_verbs.c rc = bnxt_re_query_gid(&rdev->ibdev, 1, 0, &sgid); ibdev 2536 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_dev *rdev = to_bnxt_re_dev(ibcq->device, ibdev); ibdev 3605 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct ib_device *ibdev = ctx->device; ibdev 3608 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_dev *rdev = to_bnxt_re_dev(ibdev, ibdev); ibdev 3615 drivers/infiniband/hw/bnxt_re/ib_verbs.c ibdev->ops.uverbs_abi_ver); ibdev 3617 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (ibdev->ops.uverbs_abi_ver != BNXT_RE_ABI_VERSION) { ibdev 145 drivers/infiniband/hw/bnxt_re/ib_verbs.h int bnxt_re_query_device(struct ib_device *ibdev, ibdev 148 drivers/infiniband/hw/bnxt_re/ib_verbs.h int bnxt_re_modify_device(struct ib_device *ibdev, ibdev 151 drivers/infiniband/hw/bnxt_re/ib_verbs.h int bnxt_re_query_port(struct ib_device *ibdev, u8 port_num, ibdev 153 drivers/infiniband/hw/bnxt_re/ib_verbs.h int bnxt_re_get_port_immutable(struct ib_device *ibdev, u8 port_num, ibdev 155 drivers/infiniband/hw/bnxt_re/ib_verbs.h void bnxt_re_query_fw_str(struct ib_device *ibdev, char *str); ibdev 156 drivers/infiniband/hw/bnxt_re/ib_verbs.h int bnxt_re_query_pkey(struct ib_device *ibdev, u8 port_num, ibdev 160 drivers/infiniband/hw/bnxt_re/ib_verbs.h int bnxt_re_query_gid(struct ib_device *ibdev, u8 port_num, ibdev 162 drivers/infiniband/hw/bnxt_re/ib_verbs.h enum rdma_link_layer bnxt_re_get_link_layer(struct ib_device *ibdev, ibdev 568 drivers/infiniband/hw/bnxt_re/main.c rdma_device_to_drv_device(device, struct bnxt_re_dev, ibdev); ibdev 578 drivers/infiniband/hw/bnxt_re/main.c rdma_device_to_drv_device(device, struct bnxt_re_dev, ibdev); ibdev 580 drivers/infiniband/hw/bnxt_re/main.c return scnprintf(buf, PAGE_SIZE, "%s\n", rdev->ibdev.node_desc); ibdev 596 drivers/infiniband/hw/bnxt_re/main.c ib_unregister_device(&rdev->ibdev); ibdev 653 drivers/infiniband/hw/bnxt_re/main.c struct ib_device *ibdev = &rdev->ibdev; ibdev 657 drivers/infiniband/hw/bnxt_re/main.c ibdev->node_type = RDMA_NODE_IB_CA; ibdev 658 drivers/infiniband/hw/bnxt_re/main.c strlcpy(ibdev->node_desc, BNXT_RE_DESC " HCA", ibdev 660 drivers/infiniband/hw/bnxt_re/main.c ibdev->phys_port_cnt = 1; ibdev 662 drivers/infiniband/hw/bnxt_re/main.c bnxt_qplib_get_guid(rdev->netdev->dev_addr, (u8 *)&ibdev->node_guid); ibdev 664 drivers/infiniband/hw/bnxt_re/main.c ibdev->num_comp_vectors = 1; ibdev 665 drivers/infiniband/hw/bnxt_re/main.c ibdev->dev.parent = &rdev->en_dev->pdev->dev; ibdev 666 drivers/infiniband/hw/bnxt_re/main.c ibdev->local_dma_lkey = BNXT_QPLIB_RSVD_LKEY; ibdev 669 drivers/infiniband/hw/bnxt_re/main.c ibdev->uverbs_cmd_mask = ibdev 697 drivers/infiniband/hw/bnxt_re/main.c rdma_set_device_sysfs_group(ibdev, &bnxt_re_dev_attr_group); ibdev 698 drivers/infiniband/hw/bnxt_re/main.c ib_set_device_ops(ibdev, &bnxt_re_dev_ops); ibdev 699 drivers/infiniband/hw/bnxt_re/main.c ret = ib_device_set_netdev(&rdev->ibdev, rdev->netdev, 1); ibdev 703 drivers/infiniband/hw/bnxt_re/main.c return ib_register_device(ibdev, "bnxt_re%d"); ibdev 717 drivers/infiniband/hw/bnxt_re/main.c ib_dealloc_device(&rdev->ibdev); ibdev 727 drivers/infiniband/hw/bnxt_re/main.c rdev = ib_alloc_device(bnxt_re_dev, ibdev); ibdev 800 drivers/infiniband/hw/bnxt_re/main.c event.device = &qp->rdev->ibdev; ibdev 864 drivers/infiniband/hw/bnxt_re/main.c ib_event.device = &srq->rdev->ibdev; ibdev 1042 drivers/infiniband/hw/bnxt_re/main.c static void bnxt_re_dispatch_event(struct ib_device *ibdev, struct ib_qp *qp, ibdev 1047 drivers/infiniband/hw/bnxt_re/main.c ib_event.device = ibdev; ibdev 1126 drivers/infiniband/hw/bnxt_re/main.c bnxt_re_dispatch_event(&rdev->ibdev, &qp->ib_qp, ibdev 1473 drivers/infiniband/hw/bnxt_re/main.c ib_get_eth_speed(&rdev->ibdev, 1, &rdev->active_speed, ibdev 1476 drivers/infiniband/hw/bnxt_re/main.c bnxt_re_dispatch_event(&rdev->ibdev, NULL, 1, IB_EVENT_PORT_ACTIVE); ibdev 1566 drivers/infiniband/hw/bnxt_re/main.c bnxt_re_dispatch_event(&rdev->ibdev, NULL, 1, ibdev 1576 drivers/infiniband/hw/bnxt_re/main.c bnxt_re_dispatch_event(&rdev->ibdev, NULL, 1, ibdev 1578 drivers/infiniband/hw/bnxt_re/main.c ib_get_eth_speed(&rdev->ibdev, 1, &rdev->active_speed, ibdev 141 drivers/infiniband/hw/cxgb3/iwch.c rnicp = ib_alloc_device(iwch_dev, ibdev); ibdev 154 drivers/infiniband/hw/cxgb3/iwch.c ib_dealloc_device(&rnicp->ibdev); ibdev 188 drivers/infiniband/hw/cxgb3/iwch.c ib_dealloc_device(&dev->ibdev); ibdev 250 drivers/infiniband/hw/cxgb3/iwch.c event.device = &rnicp->ibdev; ibdev 105 drivers/infiniband/hw/cxgb3/iwch.h struct ib_device ibdev; ibdev 116 drivers/infiniband/hw/cxgb3/iwch.h static inline struct iwch_dev *to_iwch_dev(struct ib_device *ibdev) ibdev 118 drivers/infiniband/hw/cxgb3/iwch.h return container_of(ibdev, struct iwch_dev, ibdev); ibdev 80 drivers/infiniband/hw/cxgb3/iwch_provider.c struct ib_device *ibdev = ucontext->device; ibdev 82 drivers/infiniband/hw/cxgb3/iwch_provider.c struct iwch_dev *rhp = to_iwch_dev(ibdev); ibdev 84 drivers/infiniband/hw/cxgb3/iwch_provider.c pr_debug("%s ibdev %p\n", __func__, ibdev); ibdev 109 drivers/infiniband/hw/cxgb3/iwch_provider.c struct ib_device *ibdev = ibcq->device; ibdev 118 drivers/infiniband/hw/cxgb3/iwch_provider.c pr_debug("%s ib_dev %p entries %d\n", __func__, ibdev, entries); ibdev 307 drivers/infiniband/hw/cxgb3/iwch_provider.c struct ib_device *ibdev = pd->device; ibdev 311 drivers/infiniband/hw/cxgb3/iwch_provider.c pr_debug("%s ibdev %p\n", __func__, ibdev); ibdev 312 drivers/infiniband/hw/cxgb3/iwch_provider.c rhp = (struct iwch_dev *) ibdev; ibdev 911 drivers/infiniband/hw/cxgb3/iwch_provider.c static int iwch_query_pkey(struct ib_device *ibdev, ibdev 914 drivers/infiniband/hw/cxgb3/iwch_provider.c pr_debug("%s ibdev %p\n", __func__, ibdev); ibdev 919 drivers/infiniband/hw/cxgb3/iwch_provider.c static int iwch_query_gid(struct ib_device *ibdev, u8 port, ibdev 925 drivers/infiniband/hw/cxgb3/iwch_provider.c __func__, ibdev, port, index, gid); ibdev 926 drivers/infiniband/hw/cxgb3/iwch_provider.c dev = to_iwch_dev(ibdev); ibdev 954 drivers/infiniband/hw/cxgb3/iwch_provider.c static int iwch_query_device(struct ib_device *ibdev, struct ib_device_attr *props, ibdev 960 drivers/infiniband/hw/cxgb3/iwch_provider.c pr_debug("%s ibdev %p\n", __func__, ibdev); ibdev 965 drivers/infiniband/hw/cxgb3/iwch_provider.c dev = to_iwch_dev(ibdev); ibdev 991 drivers/infiniband/hw/cxgb3/iwch_provider.c static int iwch_query_port(struct ib_device *ibdev, ibdev 994 drivers/infiniband/hw/cxgb3/iwch_provider.c pr_debug("%s ibdev %p\n", __func__, ibdev); ibdev 1015 drivers/infiniband/hw/cxgb3/iwch_provider.c rdma_device_to_drv_device(dev, struct iwch_dev, ibdev); ibdev 1026 drivers/infiniband/hw/cxgb3/iwch_provider.c rdma_device_to_drv_device(dev, struct iwch_dev, ibdev); ibdev 1040 drivers/infiniband/hw/cxgb3/iwch_provider.c rdma_device_to_drv_device(dev, struct iwch_dev, ibdev); ibdev 1105 drivers/infiniband/hw/cxgb3/iwch_provider.c static struct rdma_hw_stats *iwch_alloc_stats(struct ib_device *ibdev, ibdev 1118 drivers/infiniband/hw/cxgb3/iwch_provider.c static int iwch_get_mib(struct ib_device *ibdev, struct rdma_hw_stats *stats, ibdev 1128 drivers/infiniband/hw/cxgb3/iwch_provider.c pr_debug("%s ibdev %p\n", __func__, ibdev); ibdev 1129 drivers/infiniband/hw/cxgb3/iwch_provider.c dev = to_iwch_dev(ibdev); ibdev 1174 drivers/infiniband/hw/cxgb3/iwch_provider.c static int iwch_port_immutable(struct ib_device *ibdev, u8 port_num, ibdev 1182 drivers/infiniband/hw/cxgb3/iwch_provider.c err = ib_query_port(ibdev, port_num, &attr); ibdev 1192 drivers/infiniband/hw/cxgb3/iwch_provider.c static void get_dev_fw_ver_str(struct ib_device *ibdev, char *str) ibdev 1194 drivers/infiniband/hw/cxgb3/iwch_provider.c struct iwch_dev *iwch_dev = to_iwch_dev(ibdev); ibdev 1270 drivers/infiniband/hw/cxgb3/iwch_provider.c memset(&dev->ibdev.node_guid, 0, sizeof(dev->ibdev.node_guid)); ibdev 1271 drivers/infiniband/hw/cxgb3/iwch_provider.c memcpy(&dev->ibdev.node_guid, dev->rdev.t3cdev_p->lldev->dev_addr, 6); ibdev 1277 drivers/infiniband/hw/cxgb3/iwch_provider.c dev->ibdev.local_dma_lkey = 0; ibdev 1279 drivers/infiniband/hw/cxgb3/iwch_provider.c dev->ibdev.uverbs_cmd_mask = ibdev 1297 drivers/infiniband/hw/cxgb3/iwch_provider.c dev->ibdev.node_type = RDMA_NODE_RNIC; ibdev 1299 drivers/infiniband/hw/cxgb3/iwch_provider.c memcpy(dev->ibdev.node_desc, IWCH_NODE_DESC, sizeof(IWCH_NODE_DESC)); ibdev 1300 drivers/infiniband/hw/cxgb3/iwch_provider.c dev->ibdev.phys_port_cnt = dev->rdev.port_info.nports; ibdev 1301 drivers/infiniband/hw/cxgb3/iwch_provider.c dev->ibdev.num_comp_vectors = 1; ibdev 1302 drivers/infiniband/hw/cxgb3/iwch_provider.c dev->ibdev.dev.parent = &dev->rdev.rnic_info.pdev->dev; ibdev 1304 drivers/infiniband/hw/cxgb3/iwch_provider.c memcpy(dev->ibdev.iw_ifname, dev->rdev.t3cdev_p->lldev->name, ibdev 1305 drivers/infiniband/hw/cxgb3/iwch_provider.c sizeof(dev->ibdev.iw_ifname)); ibdev 1307 drivers/infiniband/hw/cxgb3/iwch_provider.c rdma_set_device_sysfs_group(&dev->ibdev, &iwch_attr_group); ibdev 1308 drivers/infiniband/hw/cxgb3/iwch_provider.c ib_set_device_ops(&dev->ibdev, &iwch_dev_ops); ibdev 1309 drivers/infiniband/hw/cxgb3/iwch_provider.c err = set_netdevs(&dev->ibdev, &dev->rdev); ibdev 1313 drivers/infiniband/hw/cxgb3/iwch_provider.c return ib_register_device(&dev->ibdev, "cxgb3_%d"); ibdev 1319 drivers/infiniband/hw/cxgb3/iwch_provider.c ib_unregister_device(&dev->ibdev); ibdev 993 drivers/infiniband/hw/cxgb4/cq.c struct ib_device *ibdev = ibcq->device; ibdev 1006 drivers/infiniband/hw/cxgb4/cq.c pr_debug("ib_dev %p entries %d\n", ibdev, entries); ibdev 949 drivers/infiniband/hw/cxgb4/device.c ib_dealloc_device(&ctx->dev->ibdev); ibdev 981 drivers/infiniband/hw/cxgb4/device.c devp = ib_alloc_device(c4iw_dev, ibdev); ibdev 1020 drivers/infiniband/hw/cxgb4/device.c ib_dealloc_device(&devp->ibdev); ibdev 1032 drivers/infiniband/hw/cxgb4/device.c ib_dealloc_device(&devp->ibdev); ibdev 1044 drivers/infiniband/hw/cxgb4/device.c ib_dealloc_device(&devp->ibdev); ibdev 1253 drivers/infiniband/hw/cxgb4/device.c event.device = &ctx->dev->ibdev; ibdev 315 drivers/infiniband/hw/cxgb4/iw_cxgb4.h struct ib_device ibdev; ibdev 339 drivers/infiniband/hw/cxgb4/iw_cxgb4.h static inline struct c4iw_dev *to_c4iw_dev(struct ib_device *ibdev) ibdev 341 drivers/infiniband/hw/cxgb4/iw_cxgb4.h return container_of(ibdev, struct c4iw_dev, ibdev); ibdev 78 drivers/infiniband/hw/cxgb4/provider.c struct ib_device *ibdev = ucontext->device; ibdev 80 drivers/infiniband/hw/cxgb4/provider.c struct c4iw_dev *rhp = to_c4iw_dev(ibdev); ibdev 85 drivers/infiniband/hw/cxgb4/provider.c pr_debug("ibdev %p\n", ibdev); ibdev 210 drivers/infiniband/hw/cxgb4/provider.c struct ib_device *ibdev = pd->device; ibdev 214 drivers/infiniband/hw/cxgb4/provider.c pr_debug("ibdev %p\n", ibdev); ibdev 215 drivers/infiniband/hw/cxgb4/provider.c rhp = (struct c4iw_dev *) ibdev; ibdev 239 drivers/infiniband/hw/cxgb4/provider.c static int c4iw_query_pkey(struct ib_device *ibdev, u8 port, u16 index, ibdev 242 drivers/infiniband/hw/cxgb4/provider.c pr_debug("ibdev %p\n", ibdev); ibdev 247 drivers/infiniband/hw/cxgb4/provider.c static int c4iw_query_gid(struct ib_device *ibdev, u8 port, int index, ibdev 253 drivers/infiniband/hw/cxgb4/provider.c ibdev, port, index, gid); ibdev 256 drivers/infiniband/hw/cxgb4/provider.c dev = to_c4iw_dev(ibdev); ibdev 262 drivers/infiniband/hw/cxgb4/provider.c static int c4iw_query_device(struct ib_device *ibdev, struct ib_device_attr *props, ibdev 268 drivers/infiniband/hw/cxgb4/provider.c pr_debug("ibdev %p\n", ibdev); ibdev 273 drivers/infiniband/hw/cxgb4/provider.c dev = to_c4iw_dev(ibdev); ibdev 305 drivers/infiniband/hw/cxgb4/provider.c static int c4iw_query_port(struct ib_device *ibdev, u8 port, ibdev 308 drivers/infiniband/hw/cxgb4/provider.c pr_debug("ibdev %p\n", ibdev); ibdev 329 drivers/infiniband/hw/cxgb4/provider.c rdma_device_to_drv_device(dev, struct c4iw_dev, ibdev); ibdev 341 drivers/infiniband/hw/cxgb4/provider.c rdma_device_to_drv_device(dev, struct c4iw_dev, ibdev); ibdev 355 drivers/infiniband/hw/cxgb4/provider.c rdma_device_to_drv_device(dev, struct c4iw_dev, ibdev); ibdev 386 drivers/infiniband/hw/cxgb4/provider.c static struct rdma_hw_stats *c4iw_alloc_stats(struct ib_device *ibdev, ibdev 398 drivers/infiniband/hw/cxgb4/provider.c static int c4iw_get_mib(struct ib_device *ibdev, ibdev 403 drivers/infiniband/hw/cxgb4/provider.c struct c4iw_dev *c4iw_dev = to_c4iw_dev(ibdev); ibdev 429 drivers/infiniband/hw/cxgb4/provider.c static int c4iw_port_immutable(struct ib_device *ibdev, u8 port_num, ibdev 437 drivers/infiniband/hw/cxgb4/provider.c err = ib_query_port(ibdev, port_num, &attr); ibdev 450 drivers/infiniband/hw/cxgb4/provider.c ibdev); ibdev 542 drivers/infiniband/hw/cxgb4/provider.c memset(&dev->ibdev.node_guid, 0, sizeof(dev->ibdev.node_guid)); ibdev 543 drivers/infiniband/hw/cxgb4/provider.c memcpy(&dev->ibdev.node_guid, dev->rdev.lldi.ports[0]->dev_addr, 6); ibdev 547 drivers/infiniband/hw/cxgb4/provider.c dev->ibdev.local_dma_lkey = 0; ibdev 548 drivers/infiniband/hw/cxgb4/provider.c dev->ibdev.uverbs_cmd_mask = ibdev 570 drivers/infiniband/hw/cxgb4/provider.c dev->ibdev.node_type = RDMA_NODE_RNIC; ibdev 572 drivers/infiniband/hw/cxgb4/provider.c memcpy(dev->ibdev.node_desc, C4IW_NODE_DESC, sizeof(C4IW_NODE_DESC)); ibdev 573 drivers/infiniband/hw/cxgb4/provider.c dev->ibdev.phys_port_cnt = dev->rdev.lldi.nports; ibdev 574 drivers/infiniband/hw/cxgb4/provider.c dev->ibdev.num_comp_vectors = dev->rdev.lldi.nciq; ibdev 575 drivers/infiniband/hw/cxgb4/provider.c dev->ibdev.dev.parent = &dev->rdev.lldi.pdev->dev; ibdev 577 drivers/infiniband/hw/cxgb4/provider.c memcpy(dev->ibdev.iw_ifname, dev->rdev.lldi.ports[0]->name, ibdev 578 drivers/infiniband/hw/cxgb4/provider.c sizeof(dev->ibdev.iw_ifname)); ibdev 580 drivers/infiniband/hw/cxgb4/provider.c rdma_set_device_sysfs_group(&dev->ibdev, &c4iw_attr_group); ibdev 581 drivers/infiniband/hw/cxgb4/provider.c ib_set_device_ops(&dev->ibdev, &c4iw_dev_ops); ibdev 582 drivers/infiniband/hw/cxgb4/provider.c ret = set_netdevs(&dev->ibdev, &dev->rdev); ibdev 585 drivers/infiniband/hw/cxgb4/provider.c ret = ib_register_device(&dev->ibdev, "cxgb4_%d"); ibdev 600 drivers/infiniband/hw/cxgb4/provider.c ib_unregister_device(&dev->ibdev); ibdev 2428 drivers/infiniband/hw/cxgb4/qp.c event.device = &srq->rhp->ibdev; ibdev 52 drivers/infiniband/hw/efa/efa.h struct ib_device ibdev; ibdev 119 drivers/infiniband/hw/efa/efa.h int efa_query_device(struct ib_device *ibdev, ibdev 122 drivers/infiniband/hw/efa/efa.h int efa_query_port(struct ib_device *ibdev, u8 port, ibdev 127 drivers/infiniband/hw/efa/efa.h int efa_query_gid(struct ib_device *ibdev, u8 port, int index, ibdev 129 drivers/infiniband/hw/efa/efa.h int efa_query_pkey(struct ib_device *ibdev, u8 port, u16 index, ibdev 144 drivers/infiniband/hw/efa/efa.h int efa_get_port_immutable(struct ib_device *ibdev, u8 port_num, ibdev 157 drivers/infiniband/hw/efa/efa.h enum rdma_link_layer efa_port_link_layer(struct ib_device *ibdev, ibdev 159 drivers/infiniband/hw/efa/efa.h struct rdma_hw_stats *efa_alloc_hw_stats(struct ib_device *ibdev, u8 port_num); ibdev 160 drivers/infiniband/hw/efa/efa.h int efa_get_hw_stats(struct ib_device *ibdev, struct rdma_hw_stats *stats, ibdev 48 drivers/infiniband/hw/efa/efa_main.c ibdev_err(&dev->ibdev, ibdev 269 drivers/infiniband/hw/efa/efa_main.c dev->ibdev.node_type = RDMA_NODE_UNSPECIFIED; ibdev 270 drivers/infiniband/hw/efa/efa_main.c dev->ibdev.phys_port_cnt = 1; ibdev 271 drivers/infiniband/hw/efa/efa_main.c dev->ibdev.num_comp_vectors = 1; ibdev 272 drivers/infiniband/hw/efa/efa_main.c dev->ibdev.dev.parent = &pdev->dev; ibdev 274 drivers/infiniband/hw/efa/efa_main.c dev->ibdev.uverbs_cmd_mask = ibdev 292 drivers/infiniband/hw/efa/efa_main.c dev->ibdev.uverbs_ex_cmd_mask = ibdev 295 drivers/infiniband/hw/efa/efa_main.c ib_set_device_ops(&dev->ibdev, &efa_dev_ops); ibdev 297 drivers/infiniband/hw/efa/efa_main.c err = ib_register_device(&dev->ibdev, "efa_%d"); ibdev 301 drivers/infiniband/hw/efa/efa_main.c ibdev_info(&dev->ibdev, "IB device registered\n"); ibdev 313 drivers/infiniband/hw/efa/efa_main.c ibdev_info(&dev->ibdev, "Unregister ib device\n"); ibdev 314 drivers/infiniband/hw/efa/efa_main.c ib_unregister_device(&dev->ibdev); ibdev 403 drivers/infiniband/hw/efa/efa_main.c dev = ib_alloc_device(efa_dev, ibdev); ibdev 480 drivers/infiniband/hw/efa/efa_main.c ib_dealloc_device(&dev->ibdev); ibdev 498 drivers/infiniband/hw/efa/efa_main.c ib_dealloc_device(&dev->ibdev); ibdev 115 drivers/infiniband/hw/efa/efa_verbs.c static inline struct efa_dev *to_edev(struct ib_device *ibdev) ibdev 117 drivers/infiniband/hw/efa/efa_verbs.c return container_of(ibdev, struct efa_dev, ibdev); ibdev 167 drivers/infiniband/hw/efa/efa_verbs.c ibdev_err(&dev->ibdev, "Failed to map DMA address\n"); ibdev 192 drivers/infiniband/hw/efa/efa_verbs.c &dev->ibdev, ibdev 219 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 262 drivers/infiniband/hw/efa/efa_verbs.c &dev->ibdev, ibdev 275 drivers/infiniband/hw/efa/efa_verbs.c int efa_query_device(struct ib_device *ibdev, ibdev 281 drivers/infiniband/hw/efa/efa_verbs.c struct efa_dev *dev = to_edev(ibdev); ibdev 286 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(ibdev, ibdev 319 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(ibdev, ibdev 328 drivers/infiniband/hw/efa/efa_verbs.c int efa_query_port(struct ib_device *ibdev, u8 port, ibdev 331 drivers/infiniband/hw/efa/efa_verbs.c struct efa_dev *dev = to_edev(ibdev); ibdev 364 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 399 drivers/infiniband/hw/efa/efa_verbs.c int efa_query_gid(struct ib_device *ibdev, u8 port, int index, ibdev 402 drivers/infiniband/hw/efa/efa_verbs.c struct efa_dev *dev = to_edev(ibdev); ibdev 409 drivers/infiniband/hw/efa/efa_verbs.c int efa_query_pkey(struct ib_device *ibdev, u8 port, u16 index, ibdev 438 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 455 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 461 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, "Allocated pd[%d]\n", pd->pdn); ibdev 477 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, "Dealloc pd[%d]\n", pd->pdn); ibdev 494 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, "Destroy qp[%u]\n", ibqp->qp_num); ibdev 500 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 569 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 576 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 583 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 589 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 595 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 610 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 616 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, "SRQ is not supported\n"); ibdev 621 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, "Unsupported create flags\n"); ibdev 654 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 663 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 672 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 678 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 698 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 705 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, "Create QP: qp type %d driver qp type %#x\n", ibdev 723 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 759 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 766 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, "Created qp[%d]\n", qp->ibqp.qp_num); ibdev 796 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 804 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, "Invalid modify QP parameters\n"); ibdev 809 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, "Can't change port num\n"); ibdev 814 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, "Can't change pkey index\n"); ibdev 833 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 893 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 923 drivers/infiniband/hw/efa/efa_verbs.c struct ib_device *ibdev = ibcq->device; ibdev 924 drivers/infiniband/hw/efa/efa_verbs.c struct efa_dev *dev = to_edev(ibdev); ibdev 931 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(ibdev, "create_cq entries %d\n", entries); ibdev 934 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(ibdev, ibdev 942 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(ibdev, ibdev 951 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(ibdev, ibdev 960 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(ibdev, "Cannot copy udata for create_cq\n"); ibdev 965 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(ibdev, ibdev 972 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(ibdev, ibdev 979 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(ibdev, ibdev 1011 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(ibdev, "Could not setup cq[%u] mmap entries\n", ibdev 1022 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(ibdev, ibdev 1028 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(ibdev, "Created cq[%d], cq depth[%u]. dma[%pad] virt[0x%p]\n", ibdev 1055 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, "hp_cnt[%u], pages_in_hp[%u]\n", ibdev 1116 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 1155 drivers/infiniband/hw/efa/efa_verbs.c ibdev_err(&dev->ibdev, ibdev 1161 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 1216 drivers/infiniband/hw/efa/efa_verbs.c ibdev_err(&dev->ibdev, "Unable to map pbl to DMA address\n"); ibdev 1221 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 1255 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 1260 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 1318 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 1351 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 1367 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, "Failed to create pbl[%d]\n", err); ibdev 1406 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 1413 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 1429 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 1444 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, "Failed to find a suitable page size in page_size_cap %#llx\n", ibdev 1453 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 1481 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, "Registered mr[%d]\n", mr->ibmr.lkey); ibdev 1501 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, "Deregister mr[%d]\n", ibmr->lkey); ibdev 1514 drivers/infiniband/hw/efa/efa_verbs.c int efa_get_port_immutable(struct ib_device *ibdev, u8 port_num, ibdev 1520 drivers/infiniband/hw/efa/efa_verbs.c err = ib_query_port(ibdev, port_num, &attr); ibdev 1522 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(ibdev, "Couldn't query port err[%d]\n", err); ibdev 1602 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, "key[%#llx] does not have valid entry\n", ibdev 1607 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 1635 drivers/infiniband/hw/efa/efa_verbs.c &dev->ibdev, ibdev 1652 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 1657 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 1664 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, "Mapping executable pages is not permitted\n"); ibdev 1694 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 1702 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, "Incompatible ABI params\n"); ibdev 1723 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 1728 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, "Created ah[%d]\n", ah->ah); ibdev 1744 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, "Destroy ah[%d]\n", ah->ah); ibdev 1747 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, ibdev 1755 drivers/infiniband/hw/efa/efa_verbs.c struct rdma_hw_stats *efa_alloc_hw_stats(struct ib_device *ibdev, u8 port_num) ibdev 1762 drivers/infiniband/hw/efa/efa_verbs.c int efa_get_hw_stats(struct ib_device *ibdev, struct rdma_hw_stats *stats, ibdev 1767 drivers/infiniband/hw/efa/efa_verbs.c struct efa_dev *dev = to_edev(ibdev); ibdev 1803 drivers/infiniband/hw/efa/efa_verbs.c enum rdma_link_layer efa_port_link_layer(struct ib_device *ibdev, ibdev 10771 drivers/infiniband/hw/hfi1/chip.c event.device = &dd->verbs_dev.rdi.ibdev; ibdev 163 drivers/infiniband/hw/hfi1/driver.c struct hfi1_ibdev *ibdev = container_of(rdi, struct hfi1_ibdev, rdi); ibdev 164 drivers/infiniband/hw/hfi1/driver.c struct hfi1_devdata *dd = container_of(ibdev, ibdev 1832 drivers/infiniband/hw/hfi1/hfi.h static inline struct hfi1_devdata *dd_from_ibdev(struct ib_device *ibdev) ibdev 1834 drivers/infiniband/hw/hfi1/hfi.h return dd_from_dev(to_idev(ibdev)); ibdev 1847 drivers/infiniband/hw/hfi1/hfi.h static inline struct hfi1_ibport *to_iport(struct ib_device *ibdev, u8 port) ibdev 1849 drivers/infiniband/hw/hfi1/hfi.h struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 2047 drivers/infiniband/hw/hfi1/hfi.h int hfi1_create_port_files(struct ib_device *ibdev, u8 port_num, ibdev 2311 drivers/infiniband/hw/hfi1/hfi.h static inline void hfi1_update_ah_attr(struct ib_device *ibdev, ibdev 2322 drivers/infiniband/hw/hfi1/hfi.h ibp = to_iport(ibdev, rdma_ah_get_port_num(attr)); ibdev 135 drivers/infiniband/hw/hfi1/intr.c event.device = &dd->verbs_dev.rdi.ibdev; ibdev 116 drivers/infiniband/hw/hfi1/mad.c event.device = &dd->verbs_dev.rdi.ibdev; ibdev 303 drivers/infiniband/hw/hfi1/mad.c attr.type = rdma_ah_find_type(&dd->verbs_dev.rdi.ibdev, port_num); ibdev 583 drivers/infiniband/hw/hfi1/mad.c u8 *data, struct ib_device *ibdev, ibdev 595 drivers/infiniband/hw/hfi1/mad.c memcpy(nd->data, ibdev->node_desc, sizeof(nd->data)); ibdev 604 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 608 drivers/infiniband/hw/hfi1/mad.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 614 drivers/infiniband/hw/hfi1/mad.c if (am || pidx >= dd->num_pports || ibdev->node_guid == 0 || ibdev 616 drivers/infiniband/hw/hfi1/mad.c get_sguid(to_iport(ibdev, port), HFI1_PORT_GUID_INDEX) == 0) { ibdev 621 drivers/infiniband/hw/hfi1/mad.c ni->port_guid = get_sguid(to_iport(ibdev, port), HFI1_PORT_GUID_INDEX); ibdev 625 drivers/infiniband/hw/hfi1/mad.c ni->num_ports = ibdev->phys_port_cnt; ibdev 628 drivers/infiniband/hw/hfi1/mad.c ni->node_guid = ibdev->node_guid; ibdev 643 drivers/infiniband/hw/hfi1/mad.c static int subn_get_nodeinfo(struct ib_smp *smp, struct ib_device *ibdev, ibdev 647 drivers/infiniband/hw/hfi1/mad.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 652 drivers/infiniband/hw/hfi1/mad.c ibdev->node_guid == 0 || ibdev 653 drivers/infiniband/hw/hfi1/mad.c get_sguid(to_iport(ibdev, port), HFI1_PORT_GUID_INDEX) == 0) { ibdev 658 drivers/infiniband/hw/hfi1/mad.c nip->port_guid = get_sguid(to_iport(ibdev, port), HFI1_PORT_GUID_INDEX); ibdev 662 drivers/infiniband/hw/hfi1/mad.c nip->num_ports = ibdev->phys_port_cnt; ibdev 665 drivers/infiniband/hw/hfi1/mad.c nip->node_guid = ibdev->node_guid; ibdev 797 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 819 drivers/infiniband/hw/hfi1/mad.c dd = dd_from_ibdev(ibdev); ibdev 1022 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 1025 drivers/infiniband/hw/hfi1/mad.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 1352 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 1398 drivers/infiniband/hw/hfi1/mad.c dd = dd_from_ibdev(ibdev); ibdev 1402 drivers/infiniband/hw/hfi1/mad.c event.device = ibdev; ibdev 1642 drivers/infiniband/hw/hfi1/mad.c ret = __subn_get_opa_portinfo(smp, am, data, ibdev, port, resp_len, ibdev 1660 drivers/infiniband/hw/hfi1/mad.c return __subn_get_opa_portinfo(smp, am, data, ibdev, port, resp_len, ibdev 1721 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 1724 drivers/infiniband/hw/hfi1/mad.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 1767 drivers/infiniband/hw/hfi1/mad.c return __subn_get_opa_pkeytable(smp, am, data, ibdev, port, resp_len, ibdev 1828 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 1831 drivers/infiniband/hw/hfi1/mad.c struct hfi1_ibport *ibp = to_iport(ibdev, port); ibdev 1851 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 1854 drivers/infiniband/hw/hfi1/mad.c struct hfi1_ibport *ibp = to_iport(ibdev, port); ibdev 1875 drivers/infiniband/hw/hfi1/mad.c return __subn_get_opa_sl_to_sc(smp, am, data, ibdev, port, resp_len, ibdev 1880 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 1883 drivers/infiniband/hw/hfi1/mad.c struct hfi1_ibport *ibp = to_iport(ibdev, port); ibdev 1903 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 1906 drivers/infiniband/hw/hfi1/mad.c struct hfi1_ibport *ibp = to_iport(ibdev, port); ibdev 1919 drivers/infiniband/hw/hfi1/mad.c return __subn_get_opa_sc_to_sl(smp, am, data, ibdev, port, resp_len, ibdev 1924 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 1928 drivers/infiniband/hw/hfi1/mad.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 1946 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 1951 drivers/infiniband/hw/hfi1/mad.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 1983 drivers/infiniband/hw/hfi1/mad.c return __subn_get_opa_sc_to_vlt(smp, am, data, ibdev, port, resp_len, ibdev 1988 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 1992 drivers/infiniband/hw/hfi1/mad.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 2013 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 2017 drivers/infiniband/hw/hfi1/mad.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 2040 drivers/infiniband/hw/hfi1/mad.c return __subn_get_opa_sc_to_vlnt(smp, am, data, ibdev, port, ibdev 2045 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 2060 drivers/infiniband/hw/hfi1/mad.c ibp = to_iport(ibdev, port); ibdev 2087 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 2104 drivers/infiniband/hw/hfi1/mad.c ibp = to_iport(ibdev, port); ibdev 2130 drivers/infiniband/hw/hfi1/mad.c return __subn_get_opa_psi(smp, am, data, ibdev, port, resp_len, ibdev 2135 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 2138 drivers/infiniband/hw/hfi1/mad.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 2187 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, u32 *resp_len, ibdev 2191 drivers/infiniband/hw/hfi1/mad.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 2211 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, u32 *resp_len, ibdev 2215 drivers/infiniband/hw/hfi1/mad.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 2230 drivers/infiniband/hw/hfi1/mad.c return __subn_get_opa_bct(smp, am, data, ibdev, port, resp_len, ibdev 2235 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 2238 drivers/infiniband/hw/hfi1/mad.c struct hfi1_pportdata *ppd = ppd_from_ibp(to_iport(ibdev, port)); ibdev 2277 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 2280 drivers/infiniband/hw/hfi1/mad.c struct hfi1_pportdata *ppd = ppd_from_ibp(to_iport(ibdev, port)); ibdev 2314 drivers/infiniband/hw/hfi1/mad.c return __subn_get_opa_vl_arb(smp, am, data, ibdev, port, resp_len, ibdev 2604 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u32 *resp_len) ibdev 2725 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, ibdev 2730 drivers/infiniband/hw/hfi1/mad.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 2739 drivers/infiniband/hw/hfi1/mad.c struct hfi1_ibport *ibp = to_iport(ibdev, port); ibdev 2892 drivers/infiniband/hw/hfi1/mad.c static u64 get_error_counter_summary(struct ib_device *ibdev, u8 port, ibdev 2895 drivers/infiniband/hw/hfi1/mad.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 2896 drivers/infiniband/hw/hfi1/mad.c struct hfi1_ibport *ibp = to_iport(ibdev, port); ibdev 2954 drivers/infiniband/hw/hfi1/mad.c static void pma_get_opa_port_dctrs(struct ib_device *ibdev, ibdev 2957 drivers/infiniband/hw/hfi1/mad.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 2976 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, ibdev 2981 drivers/infiniband/hw/hfi1/mad.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 2982 drivers/infiniband/hw/hfi1/mad.c struct hfi1_ibport *ibp = to_iport(ibdev, port); ibdev 3043 drivers/infiniband/hw/hfi1/mad.c pma_get_opa_port_dctrs(ibdev, rsp); ibdev 3060 drivers/infiniband/hw/hfi1/mad.c cpu_to_be64(get_error_counter_summary(ibdev, port, ibdev 3127 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port) ibdev 3139 drivers/infiniband/hw/hfi1/mad.c pma_get_opa_port_dctrs(ibdev, &rsp); ibdev 3154 drivers/infiniband/hw/hfi1/mad.c static void pma_get_opa_port_ectrs(struct ib_device *ibdev, ibdev 3158 drivers/infiniband/hw/hfi1/mad.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 3159 drivers/infiniband/hw/hfi1/mad.c struct hfi1_ibport *ibp = to_iport(ibdev, port); ibdev 3197 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, ibdev 3204 drivers/infiniband/hw/hfi1/mad.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 3249 drivers/infiniband/hw/hfi1/mad.c ibp = to_iport(ibdev, port_num); ibdev 3255 drivers/infiniband/hw/hfi1/mad.c pma_get_opa_port_ectrs(ibdev, rsp, port_num); ibdev 3287 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port) ibdev 3297 drivers/infiniband/hw/hfi1/mad.c pma_get_opa_port_ectrs(ibdev, &rsp, port); ibdev 3372 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, ibdev 3378 drivers/infiniband/hw/hfi1/mad.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 3471 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, ibdev 3476 drivers/infiniband/hw/hfi1/mad.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 3477 drivers/infiniband/hw/hfi1/mad.c struct hfi1_ibport *ibp = to_iport(ibdev, port); ibdev 3623 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, ibdev 3628 drivers/infiniband/hw/hfi1/mad.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 3706 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 3711 drivers/infiniband/hw/hfi1/mad.c struct hfi1_ibport *ibp = to_iport(ibdev, port); ibdev 3730 drivers/infiniband/hw/hfi1/mad.c u8 *data, struct ib_device *ibdev, ibdev 3736 drivers/infiniband/hw/hfi1/mad.c struct hfi1_ibport *ibp = to_iport(ibdev, port); ibdev 3823 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 3828 drivers/infiniband/hw/hfi1/mad.c struct hfi1_ibport *ibp = to_iport(ibdev, port); ibdev 3858 drivers/infiniband/hw/hfi1/mad.c return __subn_get_opa_cong_setting(smp, am, data, ibdev, port, ibdev 3863 drivers/infiniband/hw/hfi1/mad.c u8 *data, struct ib_device *ibdev, ibdev 3866 drivers/infiniband/hw/hfi1/mad.c struct hfi1_ibport *ibp = to_iport(ibdev, port); ibdev 3929 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 3934 drivers/infiniband/hw/hfi1/mad.c struct hfi1_ibport *ibp = to_iport(ibdev, port); ibdev 3981 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 3985 drivers/infiniband/hw/hfi1/mad.c struct hfi1_ibport *ibp = to_iport(ibdev, port); ibdev 4027 drivers/infiniband/hw/hfi1/mad.c return __subn_get_opa_cc_table(smp, am, data, ibdev, port, resp_len, ibdev 4040 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 4043 drivers/infiniband/hw/hfi1/mad.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 4070 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 4073 drivers/infiniband/hw/hfi1/mad.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 4088 drivers/infiniband/hw/hfi1/mad.c return __subn_get_opa_led_info(smp, am, data, ibdev, port, resp_len, ibdev 4093 drivers/infiniband/hw/hfi1/mad.c u8 *data, struct ib_device *ibdev, u8 port, ibdev 4097 drivers/infiniband/hw/hfi1/mad.c struct hfi1_ibport *ibp = to_iport(ibdev, port); ibdev 4101 drivers/infiniband/hw/hfi1/mad.c ret = __subn_get_opa_nodedesc(smp, am, data, ibdev, port, ibdev 4105 drivers/infiniband/hw/hfi1/mad.c ret = __subn_get_opa_nodeinfo(smp, am, data, ibdev, port, ibdev 4109 drivers/infiniband/hw/hfi1/mad.c ret = __subn_get_opa_portinfo(smp, am, data, ibdev, port, ibdev 4113 drivers/infiniband/hw/hfi1/mad.c ret = __subn_get_opa_pkeytable(smp, am, data, ibdev, port, ibdev 4117 drivers/infiniband/hw/hfi1/mad.c ret = __subn_get_opa_sl_to_sc(smp, am, data, ibdev, port, ibdev 4121 drivers/infiniband/hw/hfi1/mad.c ret = __subn_get_opa_sc_to_sl(smp, am, data, ibdev, port, ibdev 4125 drivers/infiniband/hw/hfi1/mad.c ret = __subn_get_opa_sc_to_vlt(smp, am, data, ibdev, port, ibdev 4129 drivers/infiniband/hw/hfi1/mad.c ret = __subn_get_opa_sc_to_vlnt(smp, am, data, ibdev, port, ibdev 4133 drivers/infiniband/hw/hfi1/mad.c ret = __subn_get_opa_psi(smp, am, data, ibdev, port, ibdev 4137 drivers/infiniband/hw/hfi1/mad.c ret = __subn_get_opa_bct(smp, am, data, ibdev, port, ibdev 4141 drivers/infiniband/hw/hfi1/mad.c ret = __subn_get_opa_cable_info(smp, am, data, ibdev, port, ibdev 4145 drivers/infiniband/hw/hfi1/mad.c ret = __subn_get_opa_vl_arb(smp, am, data, ibdev, port, ibdev 4149 drivers/infiniband/hw/hfi1/mad.c ret = __subn_get_opa_cong_info(smp, am, data, ibdev, port, ibdev 4153 drivers/infiniband/hw/hfi1/mad.c ret = __subn_get_opa_cong_setting(smp, am, data, ibdev, ibdev 4157 drivers/infiniband/hw/hfi1/mad.c ret = __subn_get_opa_hfi1_cong_log(smp, am, data, ibdev, ibdev 4161 drivers/infiniband/hw/hfi1/mad.c ret = __subn_get_opa_cc_table(smp, am, data, ibdev, port, ibdev 4165 drivers/infiniband/hw/hfi1/mad.c ret = __subn_get_opa_led_info(smp, am, data, ibdev, port, ibdev 4183 drivers/infiniband/hw/hfi1/mad.c u8 *data, struct ib_device *ibdev, u8 port, ibdev 4187 drivers/infiniband/hw/hfi1/mad.c struct hfi1_ibport *ibp = to_iport(ibdev, port); ibdev 4191 drivers/infiniband/hw/hfi1/mad.c ret = __subn_set_opa_portinfo(smp, am, data, ibdev, port, ibdev 4195 drivers/infiniband/hw/hfi1/mad.c ret = __subn_set_opa_pkeytable(smp, am, data, ibdev, port, ibdev 4199 drivers/infiniband/hw/hfi1/mad.c ret = __subn_set_opa_sl_to_sc(smp, am, data, ibdev, port, ibdev 4203 drivers/infiniband/hw/hfi1/mad.c ret = __subn_set_opa_sc_to_sl(smp, am, data, ibdev, port, ibdev 4207 drivers/infiniband/hw/hfi1/mad.c ret = __subn_set_opa_sc_to_vlt(smp, am, data, ibdev, port, ibdev 4211 drivers/infiniband/hw/hfi1/mad.c ret = __subn_set_opa_sc_to_vlnt(smp, am, data, ibdev, port, ibdev 4215 drivers/infiniband/hw/hfi1/mad.c ret = __subn_set_opa_psi(smp, am, data, ibdev, port, ibdev 4219 drivers/infiniband/hw/hfi1/mad.c ret = __subn_set_opa_bct(smp, am, data, ibdev, port, ibdev 4223 drivers/infiniband/hw/hfi1/mad.c ret = __subn_set_opa_vl_arb(smp, am, data, ibdev, port, ibdev 4227 drivers/infiniband/hw/hfi1/mad.c ret = __subn_set_opa_cong_setting(smp, am, data, ibdev, ibdev 4231 drivers/infiniband/hw/hfi1/mad.c ret = __subn_set_opa_cc_table(smp, am, data, ibdev, port, ibdev 4235 drivers/infiniband/hw/hfi1/mad.c ret = __subn_set_opa_led_info(smp, am, data, ibdev, port, ibdev 4258 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 4292 drivers/infiniband/hw/hfi1/mad.c ibdev, port, NULL, (u32)agg_data_len); ibdev 4307 drivers/infiniband/hw/hfi1/mad.c struct ib_device *ibdev, u8 port, ibdev 4338 drivers/infiniband/hw/hfi1/mad.c ibdev, port, NULL, (u32)agg_data_len, ibdev 4512 drivers/infiniband/hw/hfi1/mad.c static int process_subn_opa(struct ib_device *ibdev, int mad_flags, ibdev 4518 drivers/infiniband/hw/hfi1/mad.c struct hfi1_ibport *ibp = to_iport(ibdev, port); ibdev 4551 drivers/infiniband/hw/hfi1/mad.c port_num && port_num <= ibdev->phys_port_cnt && ibdev 4553 drivers/infiniband/hw/hfi1/mad.c (void)check_mkey(to_iport(ibdev, port_num), ibdev 4570 drivers/infiniband/hw/hfi1/mad.c ibdev, port, resp_len, ibdev 4574 drivers/infiniband/hw/hfi1/mad.c ret = subn_get_opa_aggregate(smp, ibdev, port, ibdev 4583 drivers/infiniband/hw/hfi1/mad.c ibdev, port, resp_len, ibdev 4587 drivers/infiniband/hw/hfi1/mad.c ret = subn_set_opa_aggregate(smp, ibdev, port, ibdev 4617 drivers/infiniband/hw/hfi1/mad.c static int process_subn(struct ib_device *ibdev, int mad_flags, ibdev 4622 drivers/infiniband/hw/hfi1/mad.c struct hfi1_ibport *ibp = to_iport(ibdev, port); ibdev 4648 drivers/infiniband/hw/hfi1/mad.c port_num && port_num <= ibdev->phys_port_cnt && ibdev 4650 drivers/infiniband/hw/hfi1/mad.c (void)check_mkey(to_iport(ibdev, port_num), ibdev 4663 drivers/infiniband/hw/hfi1/mad.c ret = subn_get_nodeinfo(smp, ibdev, port); ibdev 4676 drivers/infiniband/hw/hfi1/mad.c static int process_perf(struct ib_device *ibdev, u8 port, ibdev 4696 drivers/infiniband/hw/hfi1/mad.c ret = pma_get_ib_portcounters(pmp, ibdev, port); ibdev 4699 drivers/infiniband/hw/hfi1/mad.c ret = pma_get_ib_portcounters_ext(pmp, ibdev, port); ibdev 4738 drivers/infiniband/hw/hfi1/mad.c static int process_perf_opa(struct ib_device *ibdev, u8 port, ibdev 4758 drivers/infiniband/hw/hfi1/mad.c ret = pma_get_opa_classportinfo(pmp, ibdev, resp_len); ibdev 4761 drivers/infiniband/hw/hfi1/mad.c ret = pma_get_opa_portstatus(pmp, ibdev, port, ibdev 4765 drivers/infiniband/hw/hfi1/mad.c ret = pma_get_opa_datacounters(pmp, ibdev, port, ibdev 4769 drivers/infiniband/hw/hfi1/mad.c ret = pma_get_opa_porterrors(pmp, ibdev, port, ibdev 4773 drivers/infiniband/hw/hfi1/mad.c ret = pma_get_opa_errorinfo(pmp, ibdev, port, ibdev 4786 drivers/infiniband/hw/hfi1/mad.c ret = pma_set_opa_portstatus(pmp, ibdev, port, ibdev 4790 drivers/infiniband/hw/hfi1/mad.c ret = pma_set_opa_errorinfo(pmp, ibdev, port, ibdev 4819 drivers/infiniband/hw/hfi1/mad.c static int hfi1_process_opa_mad(struct ib_device *ibdev, int mad_flags, ibdev 4830 drivers/infiniband/hw/hfi1/mad.c struct hfi1_ibport *ibp = to_iport(ibdev, port); ibdev 4849 drivers/infiniband/hw/hfi1/mad.c ret = process_subn_opa(ibdev, mad_flags, port, in_mad, ibdev 4857 drivers/infiniband/hw/hfi1/mad.c ret = process_perf_opa(ibdev, port, in_mad, out_mad, &resp_len); ibdev 4873 drivers/infiniband/hw/hfi1/mad.c static int hfi1_process_ib_mad(struct ib_device *ibdev, int mad_flags, u8 port, ibdev 4884 drivers/infiniband/hw/hfi1/mad.c ret = process_subn(ibdev, mad_flags, port, in_mad, out_mad); ibdev 4887 drivers/infiniband/hw/hfi1/mad.c ret = process_perf(ibdev, port, in_mad, out_mad); ibdev 4916 drivers/infiniband/hw/hfi1/mad.c int hfi1_process_mad(struct ib_device *ibdev, int mad_flags, u8 port, ibdev 4925 drivers/infiniband/hw/hfi1/mad.c dev_err(ibdev->dev.parent, "invalid in_mad_size\n"); ibdev 4928 drivers/infiniband/hw/hfi1/mad.c return hfi1_process_opa_mad(ibdev, mad_flags, port, ibdev 4935 drivers/infiniband/hw/hfi1/mad.c return hfi1_process_ib_mad(ibdev, mad_flags, port, ibdev 501 drivers/infiniband/hw/hfi1/sysfs.c rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); ibdev 511 drivers/infiniband/hw/hfi1/sysfs.c rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); ibdev 527 drivers/infiniband/hw/hfi1/sysfs.c rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); ibdev 539 drivers/infiniband/hw/hfi1/sysfs.c rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); ibdev 558 drivers/infiniband/hw/hfi1/sysfs.c rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); ibdev 570 drivers/infiniband/hw/hfi1/sysfs.c rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); ibdev 582 drivers/infiniband/hw/hfi1/sysfs.c rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); ibdev 612 drivers/infiniband/hw/hfi1/sysfs.c rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); ibdev 656 drivers/infiniband/hw/hfi1/sysfs.c int hfi1_create_port_files(struct ib_device *ibdev, u8 port_num, ibdev 660 drivers/infiniband/hw/hfi1/sysfs.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 838 drivers/infiniband/hw/hfi1/sysfs.c struct ib_device *dev = &dd->verbs_dev.rdi.ibdev; ibdev 654 drivers/infiniband/hw/hfi1/verbs.c struct ib_device *ibdev = ibqp->device; ibdev 655 drivers/infiniband/hw/hfi1/verbs.c struct hfi1_ibdev *dev = to_idev(ibdev); ibdev 1514 drivers/infiniband/hw/hfi1/verbs.c u8 ah_to_sc(struct ib_device *ibdev, struct rdma_ah_attr *ah) ibdev 1516 drivers/infiniband/hw/hfi1/verbs.c struct hfi1_ibport *ibp = to_iport(ibdev, rdma_ah_get_port_num(ah)); ibdev 1521 drivers/infiniband/hw/hfi1/verbs.c static int hfi1_check_ah(struct ib_device *ibdev, struct rdma_ah_attr *ah_attr) ibdev 1534 drivers/infiniband/hw/hfi1/verbs.c ibp = to_iport(ibdev, rdma_ah_get_port_num(ah_attr)); ibdev 1549 drivers/infiniband/hw/hfi1/verbs.c static void hfi1_notify_new_ah(struct ib_device *ibdev, ibdev 1564 drivers/infiniband/hw/hfi1/verbs.c ibp = to_iport(ibdev, rdma_ah_get_port_num(ah_attr)); ibdev 1567 drivers/infiniband/hw/hfi1/verbs.c hfi1_update_ah_attr(ibdev, attr); ibdev 1620 drivers/infiniband/hw/hfi1/verbs.c static void hfi1_get_dev_fw_str(struct ib_device *ibdev, char *str) ibdev 1622 drivers/infiniband/hw/hfi1/verbs.c struct rvt_dev_info *rdi = ib_to_rvt(ibdev); ibdev 1694 drivers/infiniband/hw/hfi1/verbs.c static struct rdma_hw_stats *alloc_hw_stats(struct ib_device *ibdev, ibdev 1701 drivers/infiniband/hw/hfi1/verbs.c struct hfi1_devdata *dd = dd_from_ibdev(ibdev); ibdev 1758 drivers/infiniband/hw/hfi1/verbs.c static int get_hw_stats(struct ib_device *ibdev, struct rdma_hw_stats *stats, ibdev 1768 drivers/infiniband/hw/hfi1/verbs.c hfi1_read_cntrs(dd_from_ibdev(ibdev), NULL, &values); ibdev 1774 drivers/infiniband/hw/hfi1/verbs.c struct hfi1_ibport *ibp = to_iport(ibdev, port); ibdev 1806 drivers/infiniband/hw/hfi1/verbs.c struct ib_device *ibdev = &dev->rdi.ibdev; ibdev 1829 drivers/infiniband/hw/hfi1/verbs.c ibdev->node_guid = get_sguid(ibp, HFI1_PORT_GUID_INDEX); ibdev 1837 drivers/infiniband/hw/hfi1/verbs.c ib_hfi1_sys_image_guid = ibdev->node_guid; ibdev 1838 drivers/infiniband/hw/hfi1/verbs.c ibdev->phys_port_cnt = dd->num_pports; ibdev 1839 drivers/infiniband/hw/hfi1/verbs.c ibdev->dev.parent = &dd->pcidev->dev; ibdev 1841 drivers/infiniband/hw/hfi1/verbs.c ib_set_device_ops(ibdev, &hfi1_dev_ops); ibdev 1843 drivers/infiniband/hw/hfi1/verbs.c strlcpy(ibdev->node_desc, init_utsname()->nodename, ibdev 1844 drivers/infiniband/hw/hfi1/verbs.c sizeof(ibdev->node_desc)); ibdev 1901 drivers/infiniband/hw/hfi1/verbs.c dd->verbs_dev.rdi.ibdev.num_comp_vectors = dd->comp_vect_possible_cpus; ibdev 1928 drivers/infiniband/hw/hfi1/verbs.c rdma_set_device_sysfs_group(&dd->verbs_dev.rdi.ibdev, ibdev 307 drivers/infiniband/hw/hfi1/verbs.h static inline struct hfi1_ibdev *to_idev(struct ib_device *ibdev) ibdev 311 drivers/infiniband/hw/hfi1/verbs.h rdi = container_of(ibdev, struct rvt_dev_info, ibdev); ibdev 331 drivers/infiniband/hw/hfi1/verbs.h int hfi1_process_mad(struct ib_device *ibdev, int mad_flags, u8 port, ibdev 417 drivers/infiniband/hw/hfi1/verbs.h u8 ah_to_sc(struct ib_device *ibdev, struct rdma_ah_attr *ah_attr); ibdev 707 drivers/infiniband/hw/hns/hns_roce_hw_v1.c struct ib_device *ibdev; ibdev 727 drivers/infiniband/hw/hns/hns_roce_hw_v1.c ibdev = &hr_dev->ib_dev; ibdev 728 drivers/infiniband/hw/hns/hns_roce_hw_v1.c cq = rdma_zalloc_drv_obj(ibdev, ib_cq); ibdev 745 drivers/infiniband/hw/hns/hns_roce_hw_v1.c pd = rdma_zalloc_drv_obj(ibdev, ib_pd); ibdev 751 drivers/infiniband/hw/hns/hns_roce_hw_v1.c pd->device = ibdev; ibdev 4654 drivers/infiniband/hw/hns/hns_roce_hw_v2.c struct ib_device *ibdev = &hr_dev->ib_dev; ibdev 4662 drivers/infiniband/hw/hns/hns_roce_hw_v2.c ibdev_err(ibdev, "modify QP to Reset failed.\n"); ibdev 1021 drivers/infiniband/hw/hns/hns_roce_qp.c struct ib_device *ibdev = &hr_dev->ib_dev; ibdev 1035 drivers/infiniband/hw/hns/hns_roce_qp.c ibdev_err(ibdev, "Create RC QP 0x%06lx failed(%d)\n", ibdev 1048 drivers/infiniband/hw/hns/hns_roce_qp.c ibdev_err(ibdev, "not support usr space GSI\n"); ibdev 1070 drivers/infiniband/hw/hns/hns_roce_qp.c ibdev_err(ibdev, "Create GSI QP failed!\n"); ibdev 1078 drivers/infiniband/hw/hns/hns_roce_qp.c ibdev_err(ibdev, "not support QP type %d\n", ibdev 317 drivers/infiniband/hw/i40iw/i40iw.h struct ib_device ibdev; ibdev 332 drivers/infiniband/hw/i40iw/i40iw.h static inline struct i40iw_device *to_iwdev(struct ib_device *ibdev) ibdev 334 drivers/infiniband/hw/i40iw/i40iw.h return container_of(ibdev, struct i40iw_ib_device, ibdev)->iwdev; ibdev 3488 drivers/infiniband/hw/i40iw/i40iw_cm.c iwibdev->ibdev.ops.dereg_mr(iwqp->lsmm_mr, ibdev 57 drivers/infiniband/hw/i40iw/i40iw_verbs.c static int i40iw_query_device(struct ib_device *ibdev, ibdev 61 drivers/infiniband/hw/i40iw/i40iw_verbs.c struct i40iw_device *iwdev = to_iwdev(ibdev); ibdev 96 drivers/infiniband/hw/i40iw/i40iw_verbs.c static int i40iw_query_port(struct ib_device *ibdev, ibdev 122 drivers/infiniband/hw/i40iw/i40iw_verbs.c struct ib_device *ibdev = uctx->device; ibdev 123 drivers/infiniband/hw/i40iw/i40iw_verbs.c struct i40iw_device *iwdev = to_iwdev(ibdev); ibdev 1082 drivers/infiniband/hw/i40iw/i40iw_verbs.c struct ib_device *ibdev = ibcq->device; ibdev 1083 drivers/infiniband/hw/i40iw/i40iw_verbs.c struct i40iw_device *iwdev = to_iwdev(ibdev); ibdev 2054 drivers/infiniband/hw/i40iw/i40iw_verbs.c rdma_device_to_drv_device(dev, struct i40iw_ib_device, ibdev); ibdev 2449 drivers/infiniband/hw/i40iw/i40iw_verbs.c static int i40iw_port_immutable(struct ib_device *ibdev, u8 port_num, ibdev 2457 drivers/infiniband/hw/i40iw/i40iw_verbs.c err = ib_query_port(ibdev, port_num, &attr); ibdev 2547 drivers/infiniband/hw/i40iw/i40iw_verbs.c static struct rdma_hw_stats *i40iw_alloc_hw_stats(struct ib_device *ibdev, ibdev 2550 drivers/infiniband/hw/i40iw/i40iw_verbs.c struct i40iw_device *iwdev = to_iwdev(ibdev); ibdev 2577 drivers/infiniband/hw/i40iw/i40iw_verbs.c static int i40iw_get_hw_stats(struct ib_device *ibdev, ibdev 2581 drivers/infiniband/hw/i40iw/i40iw_verbs.c struct i40iw_device *iwdev = to_iwdev(ibdev); ibdev 2605 drivers/infiniband/hw/i40iw/i40iw_verbs.c static int i40iw_query_gid(struct ib_device *ibdev, ibdev 2610 drivers/infiniband/hw/i40iw/i40iw_verbs.c struct i40iw_device *iwdev = to_iwdev(ibdev); ibdev 2624 drivers/infiniband/hw/i40iw/i40iw_verbs.c static int i40iw_query_pkey(struct ib_device *ibdev, ibdev 2692 drivers/infiniband/hw/i40iw/i40iw_verbs.c iwibdev = ib_alloc_device(i40iw_ib_device, ibdev); ibdev 2700 drivers/infiniband/hw/i40iw/i40iw_verbs.c iwibdev->ibdev.node_type = RDMA_NODE_RNIC; ibdev 2701 drivers/infiniband/hw/i40iw/i40iw_verbs.c ether_addr_copy((u8 *)&iwibdev->ibdev.node_guid, netdev->dev_addr); ibdev 2703 drivers/infiniband/hw/i40iw/i40iw_verbs.c iwibdev->ibdev.uverbs_cmd_mask = ibdev 2724 drivers/infiniband/hw/i40iw/i40iw_verbs.c iwibdev->ibdev.phys_port_cnt = 1; ibdev 2725 drivers/infiniband/hw/i40iw/i40iw_verbs.c iwibdev->ibdev.num_comp_vectors = iwdev->ceqs_count; ibdev 2726 drivers/infiniband/hw/i40iw/i40iw_verbs.c iwibdev->ibdev.dev.parent = &pcidev->dev; ibdev 2727 drivers/infiniband/hw/i40iw/i40iw_verbs.c memcpy(iwibdev->ibdev.iw_ifname, netdev->name, ibdev 2728 drivers/infiniband/hw/i40iw/i40iw_verbs.c sizeof(iwibdev->ibdev.iw_ifname)); ibdev 2729 drivers/infiniband/hw/i40iw/i40iw_verbs.c ib_set_device_ops(&iwibdev->ibdev, &i40iw_dev_ops); ibdev 2743 drivers/infiniband/hw/i40iw/i40iw_verbs.c event.device = &iwibdev->ibdev; ibdev 2755 drivers/infiniband/hw/i40iw/i40iw_verbs.c ib_unregister_device(&iwibdev->ibdev); ibdev 2759 drivers/infiniband/hw/i40iw/i40iw_verbs.c ib_dealloc_device(&iwibdev->ibdev); ibdev 2775 drivers/infiniband/hw/i40iw/i40iw_verbs.c rdma_set_device_sysfs_group(&iwibdev->ibdev, &i40iw_attr_group); ibdev 2776 drivers/infiniband/hw/i40iw/i40iw_verbs.c ret = ib_device_set_netdev(&iwibdev->ibdev, iwdev->netdev, 1); ibdev 2780 drivers/infiniband/hw/i40iw/i40iw_verbs.c ret = ib_register_device(&iwibdev->ibdev, "i40iw%d"); ibdev 2786 drivers/infiniband/hw/i40iw/i40iw_verbs.c ib_dealloc_device(&iwdev->iwibdev->ibdev); ibdev 79 drivers/infiniband/hw/mlx4/ah.c struct mlx4_ib_dev *ibdev = to_mdev(ib_ah->device); ibdev 82 drivers/infiniband/hw/mlx4/ah.c struct mlx4_dev *dev = ibdev->dev; ibdev 107 drivers/infiniband/hw/mlx4/ah.c ret = mlx4_ib_gid_index_to_real_index(ibdev, gid_attr); ibdev 486 drivers/infiniband/hw/mlx4/alias_GUID.c static int set_guid_rec(struct ib_device *ibdev, ibdev 490 drivers/infiniband/hw/mlx4/alias_GUID.c struct mlx4_ib_dev *dev = to_mdev(ibdev); ibdev 503 drivers/infiniband/hw/mlx4/alias_GUID.c err = __mlx4_ib_query_port(ibdev, port, &attr, 1); ibdev 546 drivers/infiniband/hw/mlx4/alias_GUID.c ibdev, port, &guid_info_rec, ibdev 135 drivers/infiniband/hw/mlx4/cm.c static union ib_gid gid_from_req_msg(struct ib_device *ibdev, struct ib_mad *mad) ibdev 144 drivers/infiniband/hw/mlx4/cm.c id_map_find_by_sl_id(struct ib_device *ibdev, u32 slave_id, u32 sl_cm_id) ibdev 146 drivers/infiniband/hw/mlx4/cm.c struct rb_root *sl_id_map = &to_mdev(ibdev)->sriov.sl_id_map; ibdev 189 drivers/infiniband/hw/mlx4/cm.c static void sl_id_map_add(struct ib_device *ibdev, struct id_map_entry *new) ibdev 191 drivers/infiniband/hw/mlx4/cm.c struct rb_root *sl_id_map = &to_mdev(ibdev)->sriov.sl_id_map; ibdev 197 drivers/infiniband/hw/mlx4/cm.c ent = id_map_find_by_sl_id(ibdev, slave_id, sl_cm_id); ibdev 222 drivers/infiniband/hw/mlx4/cm.c id_map_alloc(struct ib_device *ibdev, int slave_id, u32 sl_cm_id) ibdev 226 drivers/infiniband/hw/mlx4/cm.c struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; ibdev 235 drivers/infiniband/hw/mlx4/cm.c ent->dev = to_mdev(ibdev); ibdev 242 drivers/infiniband/hw/mlx4/cm.c sl_id_map_add(ibdev, ent); ibdev 250 drivers/infiniband/hw/mlx4/cm.c mlx4_ib_warn(ibdev, "Allocation failed (err:0x%x)\n", ret); ibdev 255 drivers/infiniband/hw/mlx4/cm.c id_map_get(struct ib_device *ibdev, int *pv_cm_id, int slave_id, int sl_cm_id) ibdev 258 drivers/infiniband/hw/mlx4/cm.c struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; ibdev 262 drivers/infiniband/hw/mlx4/cm.c ent = id_map_find_by_sl_id(ibdev, slave_id, sl_cm_id); ibdev 272 drivers/infiniband/hw/mlx4/cm.c static void schedule_delayed(struct ib_device *ibdev, struct id_map_entry *id) ibdev 274 drivers/infiniband/hw/mlx4/cm.c struct mlx4_ib_sriov *sriov = &to_mdev(ibdev)->sriov; ibdev 288 drivers/infiniband/hw/mlx4/cm.c int mlx4_ib_multiplex_cm_handler(struct ib_device *ibdev, int port, int slave_id, ibdev 299 drivers/infiniband/hw/mlx4/cm.c id = id_map_get(ibdev, &pv_cm_id, slave_id, sl_cm_id); ibdev 302 drivers/infiniband/hw/mlx4/cm.c id = id_map_alloc(ibdev, slave_id, sl_cm_id); ibdev 304 drivers/infiniband/hw/mlx4/cm.c mlx4_ib_warn(ibdev, "%s: id{slave: %d, sl_cm_id: 0x%x} Failed to id_map_alloc\n", ibdev 313 drivers/infiniband/hw/mlx4/cm.c id = id_map_get(ibdev, &pv_cm_id, slave_id, sl_cm_id); ibdev 326 drivers/infiniband/hw/mlx4/cm.c schedule_delayed(ibdev, id); ibdev 330 drivers/infiniband/hw/mlx4/cm.c int mlx4_ib_demux_cm_handler(struct ib_device *ibdev, int port, int *slave, ibdev 343 drivers/infiniband/hw/mlx4/cm.c gid = gid_from_req_msg(ibdev, mad); ibdev 344 drivers/infiniband/hw/mlx4/cm.c *slave = mlx4_ib_find_real_gid(ibdev, port, gid.global.interface_id); ibdev 346 drivers/infiniband/hw/mlx4/cm.c mlx4_ib_warn(ibdev, "failed matching slave_id by gid (0x%llx)\n", ibdev 354 drivers/infiniband/hw/mlx4/cm.c id = id_map_get(ibdev, (int *)&pv_cm_id, -1, -1); ibdev 367 drivers/infiniband/hw/mlx4/cm.c schedule_delayed(ibdev, id); ibdev 178 drivers/infiniband/hw/mlx4/cq.c struct ib_device *ibdev = ibcq->device; ibdev 181 drivers/infiniband/hw/mlx4/cq.c struct mlx4_ib_dev *dev = to_mdev(ibdev); ibdev 246 drivers/infiniband/hw/mlx4/cq.c vector = dev->eq_table[vector % ibdev->num_comp_vectors]; ibdev 220 drivers/infiniband/hw/mlx4/mad.c static void smp_snoop(struct ib_device *ibdev, u8 port_num, const struct ib_mad *mad, ibdev 230 drivers/infiniband/hw/mlx4/mad.c struct mlx4_ib_dev *dev = to_mdev(ibdev); ibdev 416 drivers/infiniband/hw/mlx4/mad.c static int mlx4_ib_demux_sa_handler(struct ib_device *ibdev, int port, int slave, ibdev 424 drivers/infiniband/hw/mlx4/mad.c ret = mlx4_ib_mcg_demux_handler(ibdev, port, slave, sa_mad); ibdev 432 drivers/infiniband/hw/mlx4/mad.c int mlx4_ib_find_real_gid(struct ib_device *ibdev, u8 port, __be64 guid) ibdev 434 drivers/infiniband/hw/mlx4/mad.c struct mlx4_ib_dev *dev = to_mdev(ibdev); ibdev 664 drivers/infiniband/hw/mlx4/mad.c static int mlx4_ib_demux_mad(struct ib_device *ibdev, u8 port, ibdev 668 drivers/infiniband/hw/mlx4/mad.c struct mlx4_ib_dev *dev = to_mdev(ibdev); ibdev 674 drivers/infiniband/hw/mlx4/mad.c if (rdma_port_get_link_layer(ibdev, port) == IB_LINK_LAYER_INFINIBAND) ibdev 686 drivers/infiniband/hw/mlx4/mad.c mlx4_ib_warn(ibdev, "RoCE grh not present.\n"); ibdev 690 drivers/infiniband/hw/mlx4/mad.c mlx4_ib_warn(ibdev, "RoCE mgmt class is not CM\n"); ibdev 704 drivers/infiniband/hw/mlx4/mad.c mlx4_ib_warn(ibdev, "failed matching grh\n"); ibdev 708 drivers/infiniband/hw/mlx4/mad.c mlx4_ib_warn(ibdev, "slave id: %d is bigger than allowed:%d\n", ibdev 713 drivers/infiniband/hw/mlx4/mad.c if (mlx4_ib_demux_cm_handler(ibdev, port, NULL, mad)) ibdev 742 drivers/infiniband/hw/mlx4/mad.c slave = mlx4_ib_find_real_gid(ibdev, port, ibdev 745 drivers/infiniband/hw/mlx4/mad.c mlx4_ib_warn(ibdev, "failed matching grh\n"); ibdev 760 drivers/infiniband/hw/mlx4/mad.c mlx4_ib_warn(ibdev, "demux QP0. rejecting unsolicited mad for slave %d class 0x%x, method 0x%x\n", ibdev 768 drivers/infiniband/hw/mlx4/mad.c if (mlx4_ib_demux_sa_handler(ibdev, port, slave, ibdev 773 drivers/infiniband/hw/mlx4/mad.c if (mlx4_ib_demux_cm_handler(ibdev, port, &slave, mad)) ibdev 790 drivers/infiniband/hw/mlx4/mad.c mlx4_ib_warn(ibdev, "slave id: %d is bigger than allowed:%d\n", ibdev 802 drivers/infiniband/hw/mlx4/mad.c static int ib_process_mad(struct ib_device *ibdev, int mad_flags, u8 port_num, ibdev 834 drivers/infiniband/hw/mlx4/mad.c forward_trap(to_mdev(ibdev), port_num, in_mad); ibdev 864 drivers/infiniband/hw/mlx4/mad.c !ib_query_port(ibdev, port_num, &pattr)) ibdev 867 drivers/infiniband/hw/mlx4/mad.c err = mlx4_MAD_IFC(to_mdev(ibdev), ibdev 876 drivers/infiniband/hw/mlx4/mad.c smp_snoop(ibdev, port_num, in_mad, prev_lid); ibdev 878 drivers/infiniband/hw/mlx4/mad.c if (!mlx4_is_slave(to_mdev(ibdev)->dev)) ibdev 879 drivers/infiniband/hw/mlx4/mad.c node_desc_override(ibdev, out_mad); ibdev 937 drivers/infiniband/hw/mlx4/mad.c static int iboe_process_mad(struct ib_device *ibdev, int mad_flags, u8 port_num, ibdev 942 drivers/infiniband/hw/mlx4/mad.c struct mlx4_ib_dev *dev = to_mdev(ibdev); ibdev 985 drivers/infiniband/hw/mlx4/mad.c int mlx4_ib_process_mad(struct ib_device *ibdev, int mad_flags, u8 port_num, ibdev 991 drivers/infiniband/hw/mlx4/mad.c struct mlx4_ib_dev *dev = to_mdev(ibdev); ibdev 994 drivers/infiniband/hw/mlx4/mad.c enum rdma_link_layer link = rdma_port_get_link_layer(ibdev, port_num); ibdev 1009 drivers/infiniband/hw/mlx4/mad.c return iboe_process_mad(ibdev, mad_flags, port_num, in_wc, ibdev 1012 drivers/infiniband/hw/mlx4/mad.c return ib_process_mad(ibdev, mad_flags, port_num, in_wc, ibdev 1017 drivers/infiniband/hw/mlx4/mad.c return iboe_process_mad(ibdev, mad_flags, port_num, in_wc, ibdev 1336 drivers/infiniband/hw/mlx4/mad.c static int mlx4_ib_multiplex_sa_handler(struct ib_device *ibdev, int port, ibdev 1344 drivers/infiniband/hw/mlx4/mad.c ret = mlx4_ib_mcg_multiplex_handler(ibdev, port, slave, sa_mad); ibdev 1978 drivers/infiniband/hw/mlx4/mad.c static int create_pv_resources(struct ib_device *ibdev, int slave, int port, ibdev 1989 drivers/infiniband/hw/mlx4/mad.c if (rdma_port_get_link_layer(ibdev, ctx->port) == ibdev 2048 drivers/infiniband/hw/mlx4/mad.c ctx->wq = to_mdev(ibdev)->sriov.demux[port - 1].wq; ibdev 82 drivers/infiniband/hw/mlx4/main.c static void do_slave_init(struct mlx4_ib_dev *ibdev, int slave, int do_init); ibdev 134 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_dev *ibdev = to_mdev(device); ibdev 138 drivers/infiniband/hw/mlx4/main.c dev = mlx4_get_protocol_dev(ibdev->dev, MLX4_PROT_ETH, port_num); ibdev 141 drivers/infiniband/hw/mlx4/main.c if (mlx4_is_bonded(ibdev->dev)) { ibdev 162 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_dev *ibdev, ibdev 167 drivers/infiniband/hw/mlx4/main.c struct mlx4_dev *dev = ibdev->dev; ibdev 195 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_dev *ibdev, ibdev 200 drivers/infiniband/hw/mlx4/main.c struct mlx4_dev *dev = ibdev->dev; ibdev 240 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_dev *ibdev, ibdev 243 drivers/infiniband/hw/mlx4/main.c if (ibdev->dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_ROCE_V1_V2) ibdev 244 drivers/infiniband/hw/mlx4/main.c return mlx4_ib_update_gids_v1_v2(gids, ibdev, port_num); ibdev 246 drivers/infiniband/hw/mlx4/main.c return mlx4_ib_update_gids_v1(gids, ibdev, port_num); ibdev 258 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_dev *ibdev = to_mdev(attr->device); ibdev 259 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_iboe *iboe = &ibdev->iboe; ibdev 328 drivers/infiniband/hw/mlx4/main.c ret = mlx4_ib_update_gids(gids, ibdev, attr->port_num); ibdev 344 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_dev *ibdev = to_mdev(attr->device); ibdev 345 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_iboe *iboe = &ibdev->iboe; ibdev 388 drivers/infiniband/hw/mlx4/main.c ret = mlx4_ib_update_gids(gids, ibdev, attr->port_num); ibdev 394 drivers/infiniband/hw/mlx4/main.c int mlx4_ib_gid_index_to_real_index(struct mlx4_ib_dev *ibdev, ibdev 397 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_iboe *iboe = &ibdev->iboe; ibdev 408 drivers/infiniband/hw/mlx4/main.c if (mlx4_is_bonded(ibdev->dev)) ibdev 411 drivers/infiniband/hw/mlx4/main.c if (!rdma_cap_roce_gid_table(&ibdev->ib_dev, port_num)) ibdev 433 drivers/infiniband/hw/mlx4/main.c static int mlx4_ib_query_device(struct ib_device *ibdev, ibdev 437 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_dev *dev = to_mdev(ibdev); ibdev 472 drivers/infiniband/hw/mlx4/main.c err = mlx4_MAD_IFC(to_mdev(ibdev), MLX4_MAD_IFC_IGNORE_KEYS, ibdev 562 drivers/infiniband/hw/mlx4/main.c if (mlx4_ib_port_link_layer(ibdev, 1) == IB_LINK_LAYER_ETHERNET || ibdev 563 drivers/infiniband/hw/mlx4/main.c mlx4_ib_port_link_layer(ibdev, 2) == IB_LINK_LAYER_ETHERNET) { ibdev 624 drivers/infiniband/hw/mlx4/main.c ((mlx4_ib_port_link_layer(ibdev, 1) == ibdev 626 drivers/infiniband/hw/mlx4/main.c (mlx4_ib_port_link_layer(ibdev, 2) == ibdev 657 drivers/infiniband/hw/mlx4/main.c static int ib_link_query_port(struct ib_device *ibdev, u8 port, ibdev 675 drivers/infiniband/hw/mlx4/main.c if (mlx4_is_mfunc(to_mdev(ibdev)->dev) && netw_view) ibdev 678 drivers/infiniband/hw/mlx4/main.c err = mlx4_MAD_IFC(to_mdev(ibdev), mad_ifc_flags, port, NULL, NULL, ibdev 694 drivers/infiniband/hw/mlx4/main.c props->gid_tbl_len = to_mdev(ibdev)->dev->caps.gid_table_len[port]; ibdev 695 drivers/infiniband/hw/mlx4/main.c props->max_msg_sz = to_mdev(ibdev)->dev->caps.max_msg_sz; ibdev 696 drivers/infiniband/hw/mlx4/main.c props->pkey_tbl_len = to_mdev(ibdev)->dev->caps.pkey_table_len[port]; ibdev 727 drivers/infiniband/hw/mlx4/main.c err = mlx4_MAD_IFC(to_mdev(ibdev), mad_ifc_flags, port, ibdev 753 drivers/infiniband/hw/mlx4/main.c static int eth_link_query_port(struct ib_device *ibdev, u8 port, ibdev 757 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_dev *mdev = to_mdev(ibdev); ibdev 813 drivers/infiniband/hw/mlx4/main.c int __mlx4_ib_query_port(struct ib_device *ibdev, u8 port, ibdev 820 drivers/infiniband/hw/mlx4/main.c err = mlx4_ib_port_link_layer(ibdev, port) == IB_LINK_LAYER_INFINIBAND ? ibdev 821 drivers/infiniband/hw/mlx4/main.c ib_link_query_port(ibdev, port, props, netw_view) : ibdev 822 drivers/infiniband/hw/mlx4/main.c eth_link_query_port(ibdev, port, props); ibdev 827 drivers/infiniband/hw/mlx4/main.c static int mlx4_ib_query_port(struct ib_device *ibdev, u8 port, ibdev 831 drivers/infiniband/hw/mlx4/main.c return __mlx4_ib_query_port(ibdev, port, props, 0); ibdev 834 drivers/infiniband/hw/mlx4/main.c int __mlx4_ib_query_gid(struct ib_device *ibdev, u8 port, int index, ibdev 840 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_dev *dev = to_mdev(ibdev); ibdev 890 drivers/infiniband/hw/mlx4/main.c static int mlx4_ib_query_gid(struct ib_device *ibdev, u8 port, int index, ibdev 893 drivers/infiniband/hw/mlx4/main.c if (rdma_protocol_ib(ibdev, port)) ibdev 894 drivers/infiniband/hw/mlx4/main.c return __mlx4_ib_query_gid(ibdev, port, index, gid, 0); ibdev 898 drivers/infiniband/hw/mlx4/main.c static int mlx4_ib_query_sl2vl(struct ib_device *ibdev, u8 port, u64 *sl2vl_tbl) ibdev 907 drivers/infiniband/hw/mlx4/main.c if (mlx4_is_slave(to_mdev(ibdev)->dev)) { ibdev 921 drivers/infiniband/hw/mlx4/main.c if (mlx4_is_mfunc(to_mdev(ibdev)->dev)) ibdev 924 drivers/infiniband/hw/mlx4/main.c err = mlx4_MAD_IFC(to_mdev(ibdev), mad_ifc_flags, port, NULL, NULL, ibdev 958 drivers/infiniband/hw/mlx4/main.c int __mlx4_ib_query_pkey(struct ib_device *ibdev, u8 port, u16 index, ibdev 975 drivers/infiniband/hw/mlx4/main.c if (mlx4_is_mfunc(to_mdev(ibdev)->dev) && netw_view) ibdev 978 drivers/infiniband/hw/mlx4/main.c err = mlx4_MAD_IFC(to_mdev(ibdev), mad_ifc_flags, port, NULL, NULL, ibdev 991 drivers/infiniband/hw/mlx4/main.c static int mlx4_ib_query_pkey(struct ib_device *ibdev, u8 port, u16 index, u16 *pkey) ibdev 993 drivers/infiniband/hw/mlx4/main.c return __mlx4_ib_query_pkey(ibdev, port, index, pkey, 0); ibdev 996 drivers/infiniband/hw/mlx4/main.c static int mlx4_ib_modify_device(struct ib_device *ibdev, int mask, ibdev 1008 drivers/infiniband/hw/mlx4/main.c if (mlx4_is_slave(to_mdev(ibdev)->dev)) ibdev 1011 drivers/infiniband/hw/mlx4/main.c spin_lock_irqsave(&to_mdev(ibdev)->sm_lock, flags); ibdev 1012 drivers/infiniband/hw/mlx4/main.c memcpy(ibdev->node_desc, props->node_desc, IB_DEVICE_NODE_DESC_MAX); ibdev 1013 drivers/infiniband/hw/mlx4/main.c spin_unlock_irqrestore(&to_mdev(ibdev)->sm_lock, flags); ibdev 1019 drivers/infiniband/hw/mlx4/main.c mailbox = mlx4_alloc_cmd_mailbox(to_mdev(ibdev)->dev); ibdev 1024 drivers/infiniband/hw/mlx4/main.c mlx4_cmd(to_mdev(ibdev)->dev, mailbox->dma, 1, 0, ibdev 1027 drivers/infiniband/hw/mlx4/main.c mlx4_free_cmd_mailbox(to_mdev(ibdev)->dev, mailbox); ibdev 1058 drivers/infiniband/hw/mlx4/main.c static int mlx4_ib_modify_port(struct ib_device *ibdev, u8 port, int mask, ibdev 1061 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_dev *mdev = to_mdev(ibdev); ibdev 1076 drivers/infiniband/hw/mlx4/main.c err = ib_query_port(ibdev, port, &attr); ibdev 1088 drivers/infiniband/hw/mlx4/main.c mutex_unlock(&to_mdev(ibdev)->cap_mask_mutex); ibdev 1095 drivers/infiniband/hw/mlx4/main.c struct ib_device *ibdev = uctx->device; ibdev 1096 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_dev *dev = to_mdev(ibdev); ibdev 1105 drivers/infiniband/hw/mlx4/main.c if (ibdev->ops.uverbs_abi_ver == ibdev 1118 drivers/infiniband/hw/mlx4/main.c err = mlx4_uar_alloc(to_mdev(ibdev)->dev, &context->uar); ibdev 1128 drivers/infiniband/hw/mlx4/main.c if (ibdev->ops.uverbs_abi_ver == MLX4_IB_UVERBS_NO_DEV_CAPS_ABI_VERSION) ibdev 1134 drivers/infiniband/hw/mlx4/main.c mlx4_uar_free(to_mdev(ibdev)->dev, &context->uar); ibdev 1197 drivers/infiniband/hw/mlx4/main.c struct ib_device *ibdev = ibpd->device; ibdev 1200 drivers/infiniband/hw/mlx4/main.c err = mlx4_pd_alloc(to_mdev(ibdev)->dev, &pd->pdn); ibdev 1205 drivers/infiniband/hw/mlx4/main.c mlx4_pd_free(to_mdev(ibdev)->dev, pd->pdn); ibdev 1216 drivers/infiniband/hw/mlx4/main.c static struct ib_xrcd *mlx4_ib_alloc_xrcd(struct ib_device *ibdev, ibdev 1223 drivers/infiniband/hw/mlx4/main.c if (!(to_mdev(ibdev)->dev->caps.flags & MLX4_DEV_CAP_FLAG_XRC)) ibdev 1230 drivers/infiniband/hw/mlx4/main.c err = mlx4_xrcd_alloc(to_mdev(ibdev)->dev, &xrcd->xrcdn); ibdev 1234 drivers/infiniband/hw/mlx4/main.c xrcd->pd = ib_alloc_pd(ibdev, 0); ibdev 1241 drivers/infiniband/hw/mlx4/main.c xrcd->cq = ib_create_cq(ibdev, NULL, NULL, xrcd, &cq_attr); ibdev 1252 drivers/infiniband/hw/mlx4/main.c mlx4_xrcd_free(to_mdev(ibdev)->dev, xrcd->xrcdn); ibdev 1291 drivers/infiniband/hw/mlx4/main.c static void mlx4_ib_delete_counters_table(struct mlx4_ib_dev *ibdev, ibdev 1300 drivers/infiniband/hw/mlx4/main.c mlx4_counter_free(ibdev->dev, counter->index); ibdev 2117 drivers/infiniband/hw/mlx4/main.c static struct rdma_hw_stats *mlx4_ib_alloc_hw_stats(struct ib_device *ibdev, ibdev 2120 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_dev *dev = to_mdev(ibdev); ibdev 2131 drivers/infiniband/hw/mlx4/main.c static int mlx4_ib_get_hw_stats(struct ib_device *ibdev, ibdev 2135 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_dev *dev = to_mdev(ibdev); ibdev 2156 drivers/infiniband/hw/mlx4/main.c static int __mlx4_ib_alloc_diag_counters(struct mlx4_ib_dev *ibdev, ibdev 2166 drivers/infiniband/hw/mlx4/main.c if (ibdev->dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_DIAG_PER_PORT) ibdev 2189 drivers/infiniband/hw/mlx4/main.c static void mlx4_ib_fill_diag_counters(struct mlx4_ib_dev *ibdev, ibdev 2202 drivers/infiniband/hw/mlx4/main.c if (ibdev->dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_DIAG_PER_PORT) { ibdev 2222 drivers/infiniband/hw/mlx4/main.c static int mlx4_ib_alloc_diag_counters(struct mlx4_ib_dev *ibdev) ibdev 2224 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_diag_counters *diag = ibdev->diag_counters; ibdev 2227 drivers/infiniband/hw/mlx4/main.c bool per_port = !!(ibdev->dev->caps.flags2 & ibdev 2230 drivers/infiniband/hw/mlx4/main.c if (mlx4_is_slave(ibdev->dev)) ibdev 2238 drivers/infiniband/hw/mlx4/main.c ret = __mlx4_ib_alloc_diag_counters(ibdev, &diag[i].name, ibdev 2244 drivers/infiniband/hw/mlx4/main.c mlx4_ib_fill_diag_counters(ibdev, diag[i].name, ibdev 2248 drivers/infiniband/hw/mlx4/main.c ib_set_device_ops(&ibdev->ib_dev, &mlx4_ib_hw_stats_ops); ibdev 2261 drivers/infiniband/hw/mlx4/main.c static void mlx4_ib_diag_cleanup(struct mlx4_ib_dev *ibdev) ibdev 2266 drivers/infiniband/hw/mlx4/main.c kfree(ibdev->diag_counters[i].offset); ibdev 2267 drivers/infiniband/hw/mlx4/main.c kfree(ibdev->diag_counters[i].name); ibdev 2272 drivers/infiniband/hw/mlx4/main.c static void mlx4_ib_update_qps(struct mlx4_ib_dev *ibdev, ibdev 2284 drivers/infiniband/hw/mlx4/main.c atomic64_set(&ibdev->iboe.mac[port - 1], new_smac); ibdev 2287 drivers/infiniband/hw/mlx4/main.c if (!mlx4_is_mfunc(ibdev->dev)) ibdev 2290 drivers/infiniband/hw/mlx4/main.c mutex_lock(&ibdev->qp1_proxy_lock[port - 1]); ibdev 2291 drivers/infiniband/hw/mlx4/main.c qp = ibdev->qp1_proxy[port - 1]; ibdev 2302 drivers/infiniband/hw/mlx4/main.c new_smac_index = mlx4_register_mac(ibdev->dev, port, new_smac); ibdev 2308 drivers/infiniband/hw/mlx4/main.c if (mlx4_update_qp(ibdev->dev, qp->mqp.qpn, MLX4_UPDATE_QP_SMAC, ibdev 2323 drivers/infiniband/hw/mlx4/main.c mlx4_unregister_mac(ibdev->dev, port, release_mac); ibdev 2326 drivers/infiniband/hw/mlx4/main.c mutex_unlock(&ibdev->qp1_proxy_lock[port - 1]); ibdev 2329 drivers/infiniband/hw/mlx4/main.c static void mlx4_ib_scan_netdevs(struct mlx4_ib_dev *ibdev, ibdev 2340 drivers/infiniband/hw/mlx4/main.c iboe = &ibdev->iboe; ibdev 2343 drivers/infiniband/hw/mlx4/main.c mlx4_foreach_ib_transport_port(port, ibdev->dev) { ibdev 2346 drivers/infiniband/hw/mlx4/main.c mlx4_get_protocol_dev(ibdev->dev, MLX4_PROT_ETH, port); ibdev 2358 drivers/infiniband/hw/mlx4/main.c if (ib_get_cached_port_state(&ibdev->ib_dev, port, ibdev 2372 drivers/infiniband/hw/mlx4/main.c ibev.device = &ibdev->ib_dev; ibdev 2383 drivers/infiniband/hw/mlx4/main.c mlx4_ib_update_qps(ibdev, dev, update_qps_port); ibdev 2390 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_dev *ibdev; ibdev 2395 drivers/infiniband/hw/mlx4/main.c ibdev = container_of(this, struct mlx4_ib_dev, iboe.nb); ibdev 2396 drivers/infiniband/hw/mlx4/main.c mlx4_ib_scan_netdevs(ibdev, dev, event); ibdev 2401 drivers/infiniband/hw/mlx4/main.c static void init_pkeys(struct mlx4_ib_dev *ibdev) ibdev 2407 drivers/infiniband/hw/mlx4/main.c if (mlx4_is_master(ibdev->dev)) { ibdev 2408 drivers/infiniband/hw/mlx4/main.c for (slave = 0; slave <= ibdev->dev->persist->num_vfs; ibdev 2410 drivers/infiniband/hw/mlx4/main.c for (port = 1; port <= ibdev->dev->caps.num_ports; ++port) { ibdev 2412 drivers/infiniband/hw/mlx4/main.c i < ibdev->dev->phys_caps.pkey_phys_table_len[port]; ibdev 2414 drivers/infiniband/hw/mlx4/main.c ibdev->pkeys.virt2phys_pkey[slave][port - 1][i] = ibdev 2416 drivers/infiniband/hw/mlx4/main.c (slave == mlx4_master_func_num(ibdev->dev) || !i) ? i : ibdev 2417 drivers/infiniband/hw/mlx4/main.c ibdev->dev->phys_caps.pkey_phys_table_len[port] - 1; ibdev 2418 drivers/infiniband/hw/mlx4/main.c mlx4_sync_pkey_table(ibdev->dev, slave, port, i, ibdev 2419 drivers/infiniband/hw/mlx4/main.c ibdev->pkeys.virt2phys_pkey[slave][port - 1][i]); ibdev 2424 drivers/infiniband/hw/mlx4/main.c for (port = 1; port <= ibdev->dev->caps.num_ports; ++port) { ibdev 2426 drivers/infiniband/hw/mlx4/main.c i < ibdev->dev->phys_caps.pkey_phys_table_len[port]; ibdev 2428 drivers/infiniband/hw/mlx4/main.c ibdev->pkeys.phys_pkey_cache[port-1][i] = ibdev 2434 drivers/infiniband/hw/mlx4/main.c static void mlx4_ib_alloc_eqs(struct mlx4_dev *dev, struct mlx4_ib_dev *ibdev) ibdev 2438 drivers/infiniband/hw/mlx4/main.c ibdev->eq_table = kcalloc(dev->caps.num_comp_vectors, ibdev 2439 drivers/infiniband/hw/mlx4/main.c sizeof(ibdev->eq_table[0]), GFP_KERNEL); ibdev 2440 drivers/infiniband/hw/mlx4/main.c if (!ibdev->eq_table) ibdev 2448 drivers/infiniband/hw/mlx4/main.c ibdev->eq_table[eq] = total_eqs; ibdev 2450 drivers/infiniband/hw/mlx4/main.c &ibdev->eq_table[eq])) ibdev 2453 drivers/infiniband/hw/mlx4/main.c ibdev->eq_table[eq] = -1; ibdev 2458 drivers/infiniband/hw/mlx4/main.c ibdev->eq_table[i++] = -1) ibdev 2462 drivers/infiniband/hw/mlx4/main.c ibdev->ib_dev.num_comp_vectors = eq; ibdev 2465 drivers/infiniband/hw/mlx4/main.c static void mlx4_ib_free_eqs(struct mlx4_dev *dev, struct mlx4_ib_dev *ibdev) ibdev 2468 drivers/infiniband/hw/mlx4/main.c int total_eqs = ibdev->ib_dev.num_comp_vectors; ibdev 2471 drivers/infiniband/hw/mlx4/main.c if (!ibdev->eq_table) ibdev 2475 drivers/infiniband/hw/mlx4/main.c ibdev->ib_dev.num_comp_vectors = 0; ibdev 2478 drivers/infiniband/hw/mlx4/main.c mlx4_release_eq(dev, ibdev->eq_table[i]); ibdev 2480 drivers/infiniband/hw/mlx4/main.c kfree(ibdev->eq_table); ibdev 2481 drivers/infiniband/hw/mlx4/main.c ibdev->eq_table = NULL; ibdev 2484 drivers/infiniband/hw/mlx4/main.c static int mlx4_port_immutable(struct ib_device *ibdev, u8 port_num, ibdev 2488 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_dev *mdev = to_mdev(ibdev); ibdev 2491 drivers/infiniband/hw/mlx4/main.c if (mlx4_ib_port_link_layer(ibdev, port_num) == IB_LINK_LAYER_INFINIBAND) { ibdev 2506 drivers/infiniband/hw/mlx4/main.c err = ib_query_port(ibdev, port_num, &attr); ibdev 2620 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_dev *ibdev; ibdev 2641 drivers/infiniband/hw/mlx4/main.c ibdev = ib_alloc_device(mlx4_ib_dev, ib_dev); ibdev 2642 drivers/infiniband/hw/mlx4/main.c if (!ibdev) { ibdev 2648 drivers/infiniband/hw/mlx4/main.c iboe = &ibdev->iboe; ibdev 2650 drivers/infiniband/hw/mlx4/main.c if (mlx4_pd_alloc(dev, &ibdev->priv_pdn)) ibdev 2653 drivers/infiniband/hw/mlx4/main.c if (mlx4_uar_alloc(dev, &ibdev->priv_uar)) ibdev 2656 drivers/infiniband/hw/mlx4/main.c ibdev->uar_map = ioremap((phys_addr_t) ibdev->priv_uar.pfn << PAGE_SHIFT, ibdev 2658 drivers/infiniband/hw/mlx4/main.c if (!ibdev->uar_map) ibdev 2660 drivers/infiniband/hw/mlx4/main.c MLX4_INIT_DOORBELL_LOCK(&ibdev->uar_lock); ibdev 2662 drivers/infiniband/hw/mlx4/main.c ibdev->dev = dev; ibdev 2663 drivers/infiniband/hw/mlx4/main.c ibdev->bond_next_port = 0; ibdev 2665 drivers/infiniband/hw/mlx4/main.c ibdev->ib_dev.node_type = RDMA_NODE_IB_CA; ibdev 2666 drivers/infiniband/hw/mlx4/main.c ibdev->ib_dev.local_dma_lkey = dev->caps.reserved_lkey; ibdev 2667 drivers/infiniband/hw/mlx4/main.c ibdev->num_ports = num_ports; ibdev 2668 drivers/infiniband/hw/mlx4/main.c ibdev->ib_dev.phys_port_cnt = mlx4_is_bonded(dev) ? ibdev 2669 drivers/infiniband/hw/mlx4/main.c 1 : ibdev->num_ports; ibdev 2670 drivers/infiniband/hw/mlx4/main.c ibdev->ib_dev.num_comp_vectors = dev->caps.num_comp_vectors; ibdev 2671 drivers/infiniband/hw/mlx4/main.c ibdev->ib_dev.dev.parent = &dev->persist->pdev->dev; ibdev 2673 drivers/infiniband/hw/mlx4/main.c ibdev->ib_dev.uverbs_cmd_mask = ibdev 2699 drivers/infiniband/hw/mlx4/main.c ib_set_device_ops(&ibdev->ib_dev, &mlx4_ib_dev_ops); ibdev 2700 drivers/infiniband/hw/mlx4/main.c ibdev->ib_dev.uverbs_ex_cmd_mask |= ibdev 2707 drivers/infiniband/hw/mlx4/main.c ((mlx4_ib_port_link_layer(&ibdev->ib_dev, 1) == ibdev 2709 drivers/infiniband/hw/mlx4/main.c (mlx4_ib_port_link_layer(&ibdev->ib_dev, 2) == ibdev 2711 drivers/infiniband/hw/mlx4/main.c ibdev->ib_dev.uverbs_ex_cmd_mask |= ibdev 2717 drivers/infiniband/hw/mlx4/main.c ib_set_device_ops(&ibdev->ib_dev, &mlx4_ib_dev_wq_ops); ibdev 2720 drivers/infiniband/hw/mlx4/main.c if (!mlx4_is_slave(ibdev->dev)) ibdev 2721 drivers/infiniband/hw/mlx4/main.c ib_set_device_ops(&ibdev->ib_dev, &mlx4_ib_dev_fmr_ops); ibdev 2725 drivers/infiniband/hw/mlx4/main.c ibdev->ib_dev.uverbs_cmd_mask |= ibdev 2728 drivers/infiniband/hw/mlx4/main.c ib_set_device_ops(&ibdev->ib_dev, &mlx4_ib_dev_mw_ops); ibdev 2732 drivers/infiniband/hw/mlx4/main.c ibdev->ib_dev.uverbs_cmd_mask |= ibdev 2735 drivers/infiniband/hw/mlx4/main.c ib_set_device_ops(&ibdev->ib_dev, &mlx4_ib_dev_xrc_ops); ibdev 2739 drivers/infiniband/hw/mlx4/main.c ibdev->steering_support = MLX4_STEERING_MODE_DEVICE_MANAGED; ibdev 2740 drivers/infiniband/hw/mlx4/main.c ibdev->ib_dev.uverbs_ex_cmd_mask |= ibdev 2743 drivers/infiniband/hw/mlx4/main.c ib_set_device_ops(&ibdev->ib_dev, &mlx4_ib_dev_fs_ops); ibdev 2747 drivers/infiniband/hw/mlx4/main.c ibdev->ib_dev.ops.uverbs_abi_ver = ibdev 2750 drivers/infiniband/hw/mlx4/main.c mlx4_ib_alloc_eqs(dev, ibdev); ibdev 2754 drivers/infiniband/hw/mlx4/main.c if (init_node_data(ibdev)) ibdev 2756 drivers/infiniband/hw/mlx4/main.c mlx4_init_sl2vl_tbl(ibdev); ibdev 2758 drivers/infiniband/hw/mlx4/main.c for (i = 0; i < ibdev->num_ports; ++i) { ibdev 2759 drivers/infiniband/hw/mlx4/main.c mutex_init(&ibdev->counters_table[i].mutex); ibdev 2760 drivers/infiniband/hw/mlx4/main.c INIT_LIST_HEAD(&ibdev->counters_table[i].counters_list); ibdev 2764 drivers/infiniband/hw/mlx4/main.c num_req_counters = mlx4_is_bonded(dev) ? 1 : ibdev->num_ports; ibdev 2766 drivers/infiniband/hw/mlx4/main.c mutex_init(&ibdev->qp1_proxy_lock[i]); ibdev 2768 drivers/infiniband/hw/mlx4/main.c if (mlx4_ib_port_link_layer(&ibdev->ib_dev, i + 1) == ibdev 2770 drivers/infiniband/hw/mlx4/main.c err = mlx4_counter_alloc(ibdev->dev, &counter_index, ibdev 2787 drivers/infiniband/hw/mlx4/main.c mlx4_counter_free(ibdev->dev, counter_index); ibdev 2793 drivers/infiniband/hw/mlx4/main.c &ibdev->counters_table[i].counters_list); ibdev 2794 drivers/infiniband/hw/mlx4/main.c ibdev->counters_table[i].default_counter = counter_index; ibdev 2799 drivers/infiniband/hw/mlx4/main.c for (i = 1; i < ibdev->num_ports ; ++i) { ibdev 2808 drivers/infiniband/hw/mlx4/main.c &ibdev->counters_table[i].counters_list); ibdev 2809 drivers/infiniband/hw/mlx4/main.c ibdev->counters_table[i].default_counter = ibdev 2816 drivers/infiniband/hw/mlx4/main.c spin_lock_init(&ibdev->sm_lock); ibdev 2817 drivers/infiniband/hw/mlx4/main.c mutex_init(&ibdev->cap_mask_mutex); ibdev 2818 drivers/infiniband/hw/mlx4/main.c INIT_LIST_HEAD(&ibdev->qp_list); ibdev 2819 drivers/infiniband/hw/mlx4/main.c spin_lock_init(&ibdev->reset_flow_resource_lock); ibdev 2821 drivers/infiniband/hw/mlx4/main.c if (ibdev->steering_support == MLX4_STEERING_MODE_DEVICE_MANAGED && ibdev 2823 drivers/infiniband/hw/mlx4/main.c ibdev->steer_qpn_count = MLX4_IB_UC_MAX_NUM_QPS; ibdev 2824 drivers/infiniband/hw/mlx4/main.c err = mlx4_qp_reserve_range(dev, ibdev->steer_qpn_count, ibdev 2826 drivers/infiniband/hw/mlx4/main.c &ibdev->steer_qpn_base, 0, ibdev 2831 drivers/infiniband/hw/mlx4/main.c ibdev->ib_uc_qpns_bitmap = ibdev 2832 drivers/infiniband/hw/mlx4/main.c kmalloc_array(BITS_TO_LONGS(ibdev->steer_qpn_count), ibdev 2835 drivers/infiniband/hw/mlx4/main.c if (!ibdev->ib_uc_qpns_bitmap) ibdev 2839 drivers/infiniband/hw/mlx4/main.c bitmap_zero(ibdev->ib_uc_qpns_bitmap, ibdev 2840 drivers/infiniband/hw/mlx4/main.c ibdev->steer_qpn_count); ibdev 2842 drivers/infiniband/hw/mlx4/main.c dev, ibdev->steer_qpn_base, ibdev 2843 drivers/infiniband/hw/mlx4/main.c ibdev->steer_qpn_base + ibdev 2844 drivers/infiniband/hw/mlx4/main.c ibdev->steer_qpn_count - 1); ibdev 2848 drivers/infiniband/hw/mlx4/main.c bitmap_fill(ibdev->ib_uc_qpns_bitmap, ibdev 2849 drivers/infiniband/hw/mlx4/main.c ibdev->steer_qpn_count); ibdev 2853 drivers/infiniband/hw/mlx4/main.c for (j = 1; j <= ibdev->dev->caps.num_ports; j++) ibdev 2854 drivers/infiniband/hw/mlx4/main.c atomic64_set(&iboe->mac[j - 1], ibdev->dev->caps.def_mac[j]); ibdev 2856 drivers/infiniband/hw/mlx4/main.c if (mlx4_ib_alloc_diag_counters(ibdev)) ibdev 2859 drivers/infiniband/hw/mlx4/main.c rdma_set_device_sysfs_group(&ibdev->ib_dev, &mlx4_attr_group); ibdev 2860 drivers/infiniband/hw/mlx4/main.c if (ib_register_device(&ibdev->ib_dev, "mlx4_%d")) ibdev 2863 drivers/infiniband/hw/mlx4/main.c if (mlx4_ib_mad_init(ibdev)) ibdev 2866 drivers/infiniband/hw/mlx4/main.c if (mlx4_ib_init_sriov(ibdev)) ibdev 2883 drivers/infiniband/hw/mlx4/main.c ibdev->ib_active = true; ibdev 2886 drivers/infiniband/hw/mlx4/main.c &ibdev->ib_dev); ibdev 2888 drivers/infiniband/hw/mlx4/main.c if (mlx4_is_mfunc(ibdev->dev)) ibdev 2889 drivers/infiniband/hw/mlx4/main.c init_pkeys(ibdev); ibdev 2892 drivers/infiniband/hw/mlx4/main.c if (mlx4_is_master(ibdev->dev)) { ibdev 2894 drivers/infiniband/hw/mlx4/main.c if (j == mlx4_master_func_num(ibdev->dev)) ibdev 2896 drivers/infiniband/hw/mlx4/main.c if (mlx4_is_slave_active(ibdev->dev, j)) ibdev 2897 drivers/infiniband/hw/mlx4/main.c do_slave_init(ibdev, j, 1); ibdev 2900 drivers/infiniband/hw/mlx4/main.c return ibdev; ibdev 2903 drivers/infiniband/hw/mlx4/main.c if (ibdev->iboe.nb.notifier_call) { ibdev 2904 drivers/infiniband/hw/mlx4/main.c if (unregister_netdevice_notifier(&ibdev->iboe.nb)) ibdev 2906 drivers/infiniband/hw/mlx4/main.c ibdev->iboe.nb.notifier_call = NULL; ibdev 2910 drivers/infiniband/hw/mlx4/main.c mlx4_ib_close_sriov(ibdev); ibdev 2913 drivers/infiniband/hw/mlx4/main.c mlx4_ib_mad_cleanup(ibdev); ibdev 2916 drivers/infiniband/hw/mlx4/main.c ib_unregister_device(&ibdev->ib_dev); ibdev 2919 drivers/infiniband/hw/mlx4/main.c mlx4_ib_diag_cleanup(ibdev); ibdev 2922 drivers/infiniband/hw/mlx4/main.c kfree(ibdev->ib_uc_qpns_bitmap); ibdev 2925 drivers/infiniband/hw/mlx4/main.c mlx4_qp_release_range(dev, ibdev->steer_qpn_base, ibdev 2926 drivers/infiniband/hw/mlx4/main.c ibdev->steer_qpn_count); ibdev 2928 drivers/infiniband/hw/mlx4/main.c for (i = 0; i < ibdev->num_ports; ++i) ibdev 2929 drivers/infiniband/hw/mlx4/main.c mlx4_ib_delete_counters_table(ibdev, &ibdev->counters_table[i]); ibdev 2932 drivers/infiniband/hw/mlx4/main.c mlx4_ib_free_eqs(dev, ibdev); ibdev 2933 drivers/infiniband/hw/mlx4/main.c iounmap(ibdev->uar_map); ibdev 2936 drivers/infiniband/hw/mlx4/main.c mlx4_uar_free(dev, &ibdev->priv_uar); ibdev 2939 drivers/infiniband/hw/mlx4/main.c mlx4_pd_free(dev, ibdev->priv_pdn); ibdev 2942 drivers/infiniband/hw/mlx4/main.c ib_dealloc_device(&ibdev->ib_dev); ibdev 3015 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_dev *ibdev = ibdev_ptr; ibdev 3021 drivers/infiniband/hw/mlx4/main.c ibdev->ib_active = false; ibdev 3024 drivers/infiniband/hw/mlx4/main.c if (ibdev->iboe.nb.notifier_call) { ibdev 3025 drivers/infiniband/hw/mlx4/main.c if (unregister_netdevice_notifier(&ibdev->iboe.nb)) ibdev 3027 drivers/infiniband/hw/mlx4/main.c ibdev->iboe.nb.notifier_call = NULL; ibdev 3030 drivers/infiniband/hw/mlx4/main.c mlx4_ib_close_sriov(ibdev); ibdev 3031 drivers/infiniband/hw/mlx4/main.c mlx4_ib_mad_cleanup(ibdev); ibdev 3032 drivers/infiniband/hw/mlx4/main.c ib_unregister_device(&ibdev->ib_dev); ibdev 3033 drivers/infiniband/hw/mlx4/main.c mlx4_ib_diag_cleanup(ibdev); ibdev 3035 drivers/infiniband/hw/mlx4/main.c mlx4_qp_release_range(dev, ibdev->steer_qpn_base, ibdev 3036 drivers/infiniband/hw/mlx4/main.c ibdev->steer_qpn_count); ibdev 3037 drivers/infiniband/hw/mlx4/main.c kfree(ibdev->ib_uc_qpns_bitmap); ibdev 3039 drivers/infiniband/hw/mlx4/main.c iounmap(ibdev->uar_map); ibdev 3040 drivers/infiniband/hw/mlx4/main.c for (p = 0; p < ibdev->num_ports; ++p) ibdev 3041 drivers/infiniband/hw/mlx4/main.c mlx4_ib_delete_counters_table(ibdev, &ibdev->counters_table[p]); ibdev 3046 drivers/infiniband/hw/mlx4/main.c mlx4_ib_free_eqs(dev, ibdev); ibdev 3048 drivers/infiniband/hw/mlx4/main.c mlx4_uar_free(dev, &ibdev->priv_uar); ibdev 3049 drivers/infiniband/hw/mlx4/main.c mlx4_pd_free(dev, ibdev->priv_pdn); ibdev 3050 drivers/infiniband/hw/mlx4/main.c ib_dealloc_device(&ibdev->ib_dev); ibdev 3053 drivers/infiniband/hw/mlx4/main.c static void do_slave_init(struct mlx4_ib_dev *ibdev, int slave, int do_init) ibdev 3056 drivers/infiniband/hw/mlx4/main.c struct mlx4_dev *dev = ibdev->dev; ibdev 3085 drivers/infiniband/hw/mlx4/main.c dm[i]->dev = ibdev; ibdev 3088 drivers/infiniband/hw/mlx4/main.c spin_lock_irqsave(&ibdev->sriov.going_down_lock, flags); ibdev 3089 drivers/infiniband/hw/mlx4/main.c if (!ibdev->sriov.is_going_down) { ibdev 3091 drivers/infiniband/hw/mlx4/main.c queue_work(ibdev->sriov.demux[i].ud_wq, &dm[i]->work); ibdev 3092 drivers/infiniband/hw/mlx4/main.c spin_unlock_irqrestore(&ibdev->sriov.going_down_lock, flags); ibdev 3094 drivers/infiniband/hw/mlx4/main.c spin_unlock_irqrestore(&ibdev->sriov.going_down_lock, flags); ibdev 3103 drivers/infiniband/hw/mlx4/main.c static void mlx4_ib_handle_catas_error(struct mlx4_ib_dev *ibdev) ibdev 3117 drivers/infiniband/hw/mlx4/main.c spin_lock_irqsave(&ibdev->reset_flow_resource_lock, flags); ibdev 3119 drivers/infiniband/hw/mlx4/main.c list_for_each_entry(mqp, &ibdev->qp_list, qps_list) { ibdev 3160 drivers/infiniband/hw/mlx4/main.c spin_unlock_irqrestore(&ibdev->reset_flow_resource_lock, flags); ibdev 3168 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_dev *ibdev = ew->ib_dev; ibdev 3174 drivers/infiniband/hw/mlx4/main.c spin_lock_bh(&ibdev->iboe.lock); ibdev 3176 drivers/infiniband/hw/mlx4/main.c struct net_device *curr_netdev = ibdev->iboe.netdevs[i]; ibdev 3190 drivers/infiniband/hw/mlx4/main.c spin_unlock_bh(&ibdev->iboe.lock); ibdev 3192 drivers/infiniband/hw/mlx4/main.c ibev.device = &ibdev->ib_dev; ibdev 3225 drivers/infiniband/hw/mlx4/main.c void mlx4_sched_ib_sl2vl_update_work(struct mlx4_ib_dev *ibdev, ibdev 3234 drivers/infiniband/hw/mlx4/main.c ew->ib_dev = ibdev; ibdev 3243 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_dev *ibdev = to_mdev((struct ib_device *) ibdev_ptr); ibdev 3255 drivers/infiniband/hw/mlx4/main.c ew->ib_dev = ibdev; ibdev 3267 drivers/infiniband/hw/mlx4/main.c if (p > ibdev->num_ports) ibdev 3270 drivers/infiniband/hw/mlx4/main.c rdma_port_get_link_layer(&ibdev->ib_dev, p) == ibdev 3273 drivers/infiniband/hw/mlx4/main.c mlx4_ib_invalidate_all_guid_record(ibdev, p); ibdev 3274 drivers/infiniband/hw/mlx4/main.c if (ibdev->dev->flags & MLX4_FLAG_SECURE_HOST && ibdev 3275 drivers/infiniband/hw/mlx4/main.c !(ibdev->dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_SL_TO_VL_CHANGE_EVENT)) ibdev 3276 drivers/infiniband/hw/mlx4/main.c mlx4_sched_ib_sl2vl_update_work(ibdev, p); ibdev 3282 drivers/infiniband/hw/mlx4/main.c if (p > ibdev->num_ports) ibdev 3288 drivers/infiniband/hw/mlx4/main.c ibdev->ib_active = false; ibdev 3290 drivers/infiniband/hw/mlx4/main.c mlx4_ib_handle_catas_error(ibdev); ibdev 3300 drivers/infiniband/hw/mlx4/main.c ew->ib_dev = ibdev; ibdev 3310 drivers/infiniband/hw/mlx4/main.c do_slave_init(ibdev, p, 1); ibdev 3314 drivers/infiniband/hw/mlx4/main.c for (i = 1; i <= ibdev->num_ports; i++) { ibdev 3315 drivers/infiniband/hw/mlx4/main.c if (rdma_port_get_link_layer(&ibdev->ib_dev, i) ibdev 3317 drivers/infiniband/hw/mlx4/main.c mlx4_ib_slave_alias_guid_event(ibdev, ibdev 3328 drivers/infiniband/hw/mlx4/main.c for (i = 1; i <= ibdev->num_ports; i++) { ibdev 3329 drivers/infiniband/hw/mlx4/main.c if (rdma_port_get_link_layer(&ibdev->ib_dev, i) ibdev 3331 drivers/infiniband/hw/mlx4/main.c mlx4_ib_slave_alias_guid_event(ibdev, ibdev 3337 drivers/infiniband/hw/mlx4/main.c do_slave_init(ibdev, p, 0); ibdev 3345 drivers/infiniband/hw/mlx4/main.c ibev.element.port_num = mlx4_is_bonded(ibdev->dev) ? 1 : (u8)p; ibdev 882 drivers/infiniband/hw/mlx4/mcg.c int mlx4_ib_mcg_demux_handler(struct ib_device *ibdev, int port, int slave, ibdev 885 drivers/infiniband/hw/mlx4/mcg.c struct mlx4_ib_dev *dev = to_mdev(ibdev); ibdev 931 drivers/infiniband/hw/mlx4/mcg.c int mlx4_ib_mcg_multiplex_handler(struct ib_device *ibdev, int port, ibdev 934 drivers/infiniband/hw/mlx4/mcg.c struct mlx4_ib_dev *dev = to_mdev(ibdev); ibdev 59 drivers/infiniband/hw/mlx4/mlx4_ib.h #define mlx4_ib_warn(ibdev, format, arg...) \ ibdev 60 drivers/infiniband/hw/mlx4/mlx4_ib.h dev_warn((ibdev)->dev.parent, MLX4_IB_DRV_NAME ": " format, ## arg) ibdev 641 drivers/infiniband/hw/mlx4/mlx4_ib.h static inline struct mlx4_ib_dev *to_mdev(struct ib_device *ibdev) ibdev 643 drivers/infiniband/hw/mlx4/mlx4_ib.h return container_of(ibdev, struct mlx4_ib_dev, ib_dev); ibdev 789 drivers/infiniband/hw/mlx4/mlx4_ib.h int mlx4_ib_process_mad(struct ib_device *ibdev, int mad_flags, u8 port_num, ibdev 803 drivers/infiniband/hw/mlx4/mlx4_ib.h int __mlx4_ib_query_port(struct ib_device *ibdev, u8 port, ibdev 805 drivers/infiniband/hw/mlx4/mlx4_ib.h int __mlx4_ib_query_pkey(struct ib_device *ibdev, u8 port, u16 index, ibdev 808 drivers/infiniband/hw/mlx4/mlx4_ib.h int __mlx4_ib_query_gid(struct ib_device *ibdev, u8 port, int index, ibdev 827 drivers/infiniband/hw/mlx4/mlx4_ib.h int mlx4_ib_find_real_gid(struct ib_device *ibdev, u8 port, __be64 guid); ibdev 829 drivers/infiniband/hw/mlx4/mlx4_ib.h int mlx4_ib_mcg_multiplex_handler(struct ib_device *ibdev, int port, int slave, ibdev 831 drivers/infiniband/hw/mlx4/mlx4_ib.h int mlx4_ib_mcg_demux_handler(struct ib_device *ibdev, int port, int slave, ibdev 853 drivers/infiniband/hw/mlx4/mlx4_ib.h int mlx4_ib_demux_cm_handler(struct ib_device *ibdev, int port, int *slave, ibdev 856 drivers/infiniband/hw/mlx4/mlx4_ib.h int mlx4_ib_multiplex_cm_handler(struct ib_device *ibdev, int port, int slave_id, ibdev 898 drivers/infiniband/hw/mlx4/mlx4_ib.h int mlx4_ib_gid_index_to_real_index(struct mlx4_ib_dev *ibdev, ibdev 901 drivers/infiniband/hw/mlx4/mlx4_ib.h void mlx4_sched_ib_sl2vl_update_work(struct mlx4_ib_dev *ibdev, ibdev 3012 drivers/infiniband/hw/mlx4/qp.c static int fill_gid_by_hw_index(struct mlx4_ib_dev *ibdev, u8 port_num, ibdev 3016 drivers/infiniband/hw/mlx4/qp.c struct mlx4_ib_iboe *iboe = &ibdev->iboe; ibdev 3036 drivers/infiniband/hw/mlx4/qp.c struct mlx4_ib_dev *ibdev = to_mdev(ib_dev); ibdev 3073 drivers/infiniband/hw/mlx4/qp.c err = fill_gid_by_hw_index(ibdev, sqp->qp.port, ibdev 3988 drivers/infiniband/hw/mlx4/qp.c static void to_rdma_ah_attr(struct mlx4_ib_dev *ibdev, ibdev 3992 drivers/infiniband/hw/mlx4/qp.c struct mlx4_dev *dev = ibdev->dev; ibdev 3998 drivers/infiniband/hw/mlx4/qp.c ah_attr->type = rdma_ah_find_type(&ibdev->ib_dev, port_num); ibdev 913 drivers/infiniband/hw/mlx5/cq.c struct ib_device *ibdev = ibcq->device; ibdev 916 drivers/infiniband/hw/mlx5/cq.c struct mlx5_ib_dev *dev = to_mdev(ibdev); ibdev 424 drivers/infiniband/hw/mlx5/flow.c static bool mlx5_ib_flow_action_packet_reformat_valid(struct mlx5_ib_dev *ibdev, ibdev 431 drivers/infiniband/hw/mlx5/flow.c return MLX5_CAP_FLOWTABLE(ibdev->mdev, ibdev 436 drivers/infiniband/hw/mlx5/flow.c return MLX5_CAP_FLOWTABLE_NIC_TX(ibdev->mdev, ibdev 441 drivers/infiniband/hw/mlx5/flow.c return MLX5_CAP_FLOWTABLE_NIC_RX(ibdev->mdev, ibdev 446 drivers/infiniband/hw/mlx5/flow.c return MLX5_CAP_FLOWTABLE_NIC_RX(ibdev->mdev, decap); ibdev 13 drivers/infiniband/hw/mlx5/ib_rep.c struct mlx5_ib_dev *ibdev; ibdev 16 drivers/infiniband/hw/mlx5/ib_rep.c ibdev = mlx5_ib_get_uplink_ibdev(dev->priv.eswitch); ibdev 19 drivers/infiniband/hw/mlx5/ib_rep.c ibdev->port[vport_index].rep = rep; ibdev 20 drivers/infiniband/hw/mlx5/ib_rep.c rep->rep_data[REP_IB].priv = ibdev; ibdev 21 drivers/infiniband/hw/mlx5/ib_rep.c write_lock(&ibdev->port[vport_index].roce.netdev_lock); ibdev 22 drivers/infiniband/hw/mlx5/ib_rep.c ibdev->port[vport_index].roce.netdev = ibdev 24 drivers/infiniband/hw/mlx5/ib_rep.c write_unlock(&ibdev->port[vport_index].roce.netdev_lock); ibdev 34 drivers/infiniband/hw/mlx5/ib_rep.c struct mlx5_ib_dev *ibdev; ibdev 42 drivers/infiniband/hw/mlx5/ib_rep.c ibdev = ib_alloc_device(mlx5_ib_dev, ib_dev); ibdev 43 drivers/infiniband/hw/mlx5/ib_rep.c if (!ibdev) ibdev 46 drivers/infiniband/hw/mlx5/ib_rep.c ibdev->port = kcalloc(num_ports, sizeof(*ibdev->port), ibdev 48 drivers/infiniband/hw/mlx5/ib_rep.c if (!ibdev->port) { ibdev 49 drivers/infiniband/hw/mlx5/ib_rep.c ib_dealloc_device(&ibdev->ib_dev); ibdev 53 drivers/infiniband/hw/mlx5/ib_rep.c ibdev->is_rep = true; ibdev 55 drivers/infiniband/hw/mlx5/ib_rep.c ibdev->port[vport_index].rep = rep; ibdev 56 drivers/infiniband/hw/mlx5/ib_rep.c ibdev->port[vport_index].roce.netdev = ibdev 58 drivers/infiniband/hw/mlx5/ib_rep.c ibdev->mdev = dev; ibdev 59 drivers/infiniband/hw/mlx5/ib_rep.c ibdev->num_ports = num_ports; ibdev 61 drivers/infiniband/hw/mlx5/ib_rep.c if (!__mlx5_ib_add(ibdev, profile)) ibdev 64 drivers/infiniband/hw/mlx5/ib_rep.c rep->rep_data[REP_IB].priv = ibdev; ibdev 77 drivers/infiniband/hw/mlx5/mad.c static int process_mad(struct ib_device *ibdev, int mad_flags, u8 port_num, ibdev 111 drivers/infiniband/hw/mlx5/mad.c err = mlx5_MAD_IFC(to_mdev(ibdev), ibdev 272 drivers/infiniband/hw/mlx5/mad.c int mlx5_ib_process_mad(struct ib_device *ibdev, int mad_flags, u8 port_num, ibdev 278 drivers/infiniband/hw/mlx5/mad.c struct mlx5_ib_dev *dev = to_mdev(ibdev); ibdev 294 drivers/infiniband/hw/mlx5/mad.c ret = process_mad(ibdev, mad_flags, port_num, in_wc, in_grh, ibdev 329 drivers/infiniband/hw/mlx5/mad.c int mlx5_query_mad_ifc_smp_attr_node_info(struct ib_device *ibdev, ibdev 342 drivers/infiniband/hw/mlx5/mad.c err = mlx5_MAD_IFC(to_mdev(ibdev), 1, 1, 1, NULL, NULL, in_mad, ibdev 349 drivers/infiniband/hw/mlx5/mad.c int mlx5_query_mad_ifc_system_image_guid(struct ib_device *ibdev, ibdev 359 drivers/infiniband/hw/mlx5/mad.c err = mlx5_query_mad_ifc_smp_attr_node_info(ibdev, out_mad); ibdev 371 drivers/infiniband/hw/mlx5/mad.c int mlx5_query_mad_ifc_max_pkeys(struct ib_device *ibdev, ibdev 381 drivers/infiniband/hw/mlx5/mad.c err = mlx5_query_mad_ifc_smp_attr_node_info(ibdev, out_mad); ibdev 393 drivers/infiniband/hw/mlx5/mad.c int mlx5_query_mad_ifc_vendor_id(struct ib_device *ibdev, ibdev 403 drivers/infiniband/hw/mlx5/mad.c err = mlx5_query_mad_ifc_smp_attr_node_info(ibdev, out_mad); ibdev 465 drivers/infiniband/hw/mlx5/mad.c int mlx5_query_mad_ifc_pkey(struct ib_device *ibdev, u8 port, u16 index, ibdev 481 drivers/infiniband/hw/mlx5/mad.c err = mlx5_MAD_IFC(to_mdev(ibdev), 1, 1, port, NULL, NULL, in_mad, ibdev 494 drivers/infiniband/hw/mlx5/mad.c int mlx5_query_mad_ifc_gids(struct ib_device *ibdev, u8 port, int index, ibdev 510 drivers/infiniband/hw/mlx5/mad.c err = mlx5_MAD_IFC(to_mdev(ibdev), 1, 1, port, NULL, NULL, in_mad, ibdev 521 drivers/infiniband/hw/mlx5/mad.c err = mlx5_MAD_IFC(to_mdev(ibdev), 1, 1, port, NULL, NULL, in_mad, ibdev 534 drivers/infiniband/hw/mlx5/mad.c int mlx5_query_mad_ifc_port(struct ib_device *ibdev, u8 port, ibdev 537 drivers/infiniband/hw/mlx5/mad.c struct mlx5_ib_dev *dev = to_mdev(ibdev); ibdev 119 drivers/infiniband/hw/mlx5/main.c dev = mpi->ibdev; ibdev 146 drivers/infiniband/hw/mlx5/main.c static int get_port_state(struct ib_device *ibdev, ibdev 154 drivers/infiniband/hw/mlx5/main.c ret = ibdev->ops.query_port(ibdev, port_num, &attr); ibdev 195 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_dev *ibdev; ibdev 197 drivers/infiniband/hw/mlx5/main.c ibdev = roce->dev; ibdev 198 drivers/infiniband/hw/mlx5/main.c mdev = mlx5_ib_get_native_port_mdev(ibdev, port_num, NULL); ibdev 205 drivers/infiniband/hw/mlx5/main.c if (ibdev->is_rep) ibdev 232 drivers/infiniband/hw/mlx5/main.c if (ibdev->is_rep) ibdev 233 drivers/infiniband/hw/mlx5/main.c roce = mlx5_get_rep_roce(ibdev, ndev, &port_num); ibdev 237 drivers/infiniband/hw/mlx5/main.c && ibdev->ib_active) { ibdev 241 drivers/infiniband/hw/mlx5/main.c if (get_port_state(&ibdev->ib_dev, port_num, ibdev 249 drivers/infiniband/hw/mlx5/main.c ibev.device = &ibdev->ib_dev; ibdev 267 drivers/infiniband/hw/mlx5/main.c mlx5_ib_put_native_port_mdev(ibdev, port_num); ibdev 274 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_dev *ibdev = to_mdev(device); ibdev 278 drivers/infiniband/hw/mlx5/main.c mdev = mlx5_ib_get_native_port_mdev(ibdev, port_num, NULL); ibdev 288 drivers/infiniband/hw/mlx5/main.c read_lock(&ibdev->port[port_num - 1].roce.netdev_lock); ibdev 289 drivers/infiniband/hw/mlx5/main.c ndev = ibdev->port[port_num - 1].roce.netdev; ibdev 292 drivers/infiniband/hw/mlx5/main.c read_unlock(&ibdev->port[port_num - 1].roce.netdev_lock); ibdev 295 drivers/infiniband/hw/mlx5/main.c mlx5_ib_put_native_port_mdev(ibdev, port_num); ibdev 299 drivers/infiniband/hw/mlx5/main.c struct mlx5_core_dev *mlx5_ib_get_native_port_mdev(struct mlx5_ib_dev *ibdev, ibdev 303 drivers/infiniband/hw/mlx5/main.c enum rdma_link_layer ll = mlx5_ib_port_link_layer(&ibdev->ib_dev, ibdev 309 drivers/infiniband/hw/mlx5/main.c if (!mlx5_core_mp_enabled(ibdev->mdev) || ibdev 313 drivers/infiniband/hw/mlx5/main.c return ibdev->mdev; ibdev 319 drivers/infiniband/hw/mlx5/main.c port = &ibdev->port[ib_port_num - 1]; ibdev 324 drivers/infiniband/hw/mlx5/main.c mpi = ibdev->port[ib_port_num - 1].mp.mpi; ibdev 338 drivers/infiniband/hw/mlx5/main.c void mlx5_ib_put_native_port_mdev(struct mlx5_ib_dev *ibdev, u8 port_num) ibdev 340 drivers/infiniband/hw/mlx5/main.c enum rdma_link_layer ll = mlx5_ib_port_link_layer(&ibdev->ib_dev, ibdev 345 drivers/infiniband/hw/mlx5/main.c if (!mlx5_core_mp_enabled(ibdev->mdev) || ll != IB_LINK_LAYER_ETHERNET) ibdev 348 drivers/infiniband/hw/mlx5/main.c port = &ibdev->port[port_num - 1]; ibdev 351 drivers/infiniband/hw/mlx5/main.c mpi = ibdev->port[port_num - 1].mp.mpi; ibdev 654 drivers/infiniband/hw/mlx5/main.c static int mlx5_get_vport_access_method(struct ib_device *ibdev) ibdev 656 drivers/infiniband/hw/mlx5/main.c if (mlx5_use_mad_ifc(to_mdev(ibdev))) ibdev 659 drivers/infiniband/hw/mlx5/main.c if (mlx5_ib_port_link_layer(ibdev, 1) == ibdev 711 drivers/infiniband/hw/mlx5/main.c static int mlx5_query_system_image_guid(struct ib_device *ibdev, ibdev 714 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_dev *dev = to_mdev(ibdev); ibdev 719 drivers/infiniband/hw/mlx5/main.c switch (mlx5_get_vport_access_method(ibdev)) { ibdev 721 drivers/infiniband/hw/mlx5/main.c return mlx5_query_mad_ifc_system_image_guid(ibdev, ibdev 743 drivers/infiniband/hw/mlx5/main.c static int mlx5_query_max_pkeys(struct ib_device *ibdev, ibdev 746 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_dev *dev = to_mdev(ibdev); ibdev 749 drivers/infiniband/hw/mlx5/main.c switch (mlx5_get_vport_access_method(ibdev)) { ibdev 751 drivers/infiniband/hw/mlx5/main.c return mlx5_query_mad_ifc_max_pkeys(ibdev, max_pkeys); ibdev 764 drivers/infiniband/hw/mlx5/main.c static int mlx5_query_vendor_id(struct ib_device *ibdev, ibdev 767 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_dev *dev = to_mdev(ibdev); ibdev 769 drivers/infiniband/hw/mlx5/main.c switch (mlx5_get_vport_access_method(ibdev)) { ibdev 771 drivers/infiniband/hw/mlx5/main.c return mlx5_query_mad_ifc_vendor_id(ibdev, vendor_id); ibdev 828 drivers/infiniband/hw/mlx5/main.c static int mlx5_ib_query_device(struct ib_device *ibdev, ibdev 833 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_dev *dev = to_mdev(ibdev); ibdev 855 drivers/infiniband/hw/mlx5/main.c err = mlx5_query_system_image_guid(ibdev, ibdev 860 drivers/infiniband/hw/mlx5/main.c err = mlx5_query_max_pkeys(ibdev, &props->max_pkeys); ibdev 864 drivers/infiniband/hw/mlx5/main.c err = mlx5_query_vendor_id(ibdev, &props->vendor_id); ibdev 1038 drivers/infiniband/hw/mlx5/main.c if (mlx5_ib_port_link_layer(ibdev, 1) == ibdev 1210 drivers/infiniband/hw/mlx5/main.c static void translate_active_width(struct ib_device *ibdev, u8 active_width, ibdev 1213 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_dev *dev = to_mdev(ibdev); ibdev 1268 drivers/infiniband/hw/mlx5/main.c static int translate_max_vl_num(struct ib_device *ibdev, u8 vl_hw_cap, ibdev 1295 drivers/infiniband/hw/mlx5/main.c static int mlx5_query_hca_port(struct ib_device *ibdev, u8 port, ibdev 1298 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_dev *dev = to_mdev(ibdev); ibdev 1341 drivers/infiniband/hw/mlx5/main.c translate_active_width(ibdev, ib_link_width_oper, &props->active_width); ibdev 1359 drivers/infiniband/hw/mlx5/main.c err = translate_max_vl_num(ibdev, vl_hw_cap, ibdev 1366 drivers/infiniband/hw/mlx5/main.c int mlx5_ib_query_port(struct ib_device *ibdev, u8 port, ibdev 1372 drivers/infiniband/hw/mlx5/main.c switch (mlx5_get_vport_access_method(ibdev)) { ibdev 1374 drivers/infiniband/hw/mlx5/main.c ret = mlx5_query_mad_ifc_port(ibdev, port, props); ibdev 1378 drivers/infiniband/hw/mlx5/main.c ret = mlx5_query_hca_port(ibdev, port, props); ibdev 1382 drivers/infiniband/hw/mlx5/main.c ret = mlx5_query_port_roce(ibdev, port, props); ibdev 1390 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_dev *dev = to_mdev(ibdev); ibdev 1411 drivers/infiniband/hw/mlx5/main.c static int mlx5_ib_rep_query_port(struct ib_device *ibdev, u8 port, ibdev 1419 drivers/infiniband/hw/mlx5/main.c ret = mlx5_query_port_roce(ibdev, port, props); ibdev 1429 drivers/infiniband/hw/mlx5/main.c static int mlx5_ib_query_gid(struct ib_device *ibdev, u8 port, int index, ibdev 1432 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_dev *dev = to_mdev(ibdev); ibdev 1435 drivers/infiniband/hw/mlx5/main.c switch (mlx5_get_vport_access_method(ibdev)) { ibdev 1437 drivers/infiniband/hw/mlx5/main.c return mlx5_query_mad_ifc_gids(ibdev, port, index, gid); ibdev 1448 drivers/infiniband/hw/mlx5/main.c static int mlx5_query_hca_nic_pkey(struct ib_device *ibdev, u8 port, ibdev 1451 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_dev *dev = to_mdev(ibdev); ibdev 1475 drivers/infiniband/hw/mlx5/main.c static int mlx5_ib_query_pkey(struct ib_device *ibdev, u8 port, u16 index, ibdev 1478 drivers/infiniband/hw/mlx5/main.c switch (mlx5_get_vport_access_method(ibdev)) { ibdev 1480 drivers/infiniband/hw/mlx5/main.c return mlx5_query_mad_ifc_pkey(ibdev, port, index, pkey); ibdev 1484 drivers/infiniband/hw/mlx5/main.c return mlx5_query_hca_nic_pkey(ibdev, port, index, pkey); ibdev 1490 drivers/infiniband/hw/mlx5/main.c static int mlx5_ib_modify_device(struct ib_device *ibdev, int mask, ibdev 1493 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_dev *dev = to_mdev(ibdev); ibdev 1514 drivers/infiniband/hw/mlx5/main.c memcpy(ibdev->node_desc, props->node_desc, IB_DEVICE_NODE_DESC_MAX); ibdev 1553 drivers/infiniband/hw/mlx5/main.c static int mlx5_ib_modify_port(struct ib_device *ibdev, u8 port, int mask, ibdev 1556 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_dev *dev = to_mdev(ibdev); ibdev 1562 drivers/infiniband/hw/mlx5/main.c bool is_ib = (mlx5_ib_port_link_layer(ibdev, port) == ibdev 1579 drivers/infiniband/hw/mlx5/main.c err = ib_query_port(ibdev, port, &attr); ibdev 1768 drivers/infiniband/hw/mlx5/main.c struct ib_device *ibdev = uctx->device; ibdev 1769 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_dev *dev = to_mdev(ibdev); ibdev 1901 drivers/infiniband/hw/mlx5/main.c if (mlx5_ib_port_link_layer(ibdev, 1) == IB_LINK_LAYER_ETHERNET) { ibdev 2364 drivers/infiniband/hw/mlx5/main.c struct ib_dm *mlx5_ib_alloc_dm(struct ib_device *ibdev, ibdev 2379 drivers/infiniband/hw/mlx5/main.c mlx5_ib_dbg(to_mdev(ibdev), "alloc_dm req: dm_type=%d user_length=0x%llx log_alignment=%d\n", ibdev 2382 drivers/infiniband/hw/mlx5/main.c err = check_dm_type_support(to_mdev(ibdev), type); ibdev 2470 drivers/infiniband/hw/mlx5/main.c struct ib_device *ibdev = ibpd->device; ibdev 2482 drivers/infiniband/hw/mlx5/main.c err = mlx5_cmd_exec(to_mdev(ibdev)->mdev, in, sizeof(in), ibdev 2492 drivers/infiniband/hw/mlx5/main.c mlx5_cmd_dealloc_pd(to_mdev(ibdev)->mdev, pd->pdn, uid); ibdev 3364 drivers/infiniband/hw/mlx5/main.c static int read_flow_counters(struct ib_device *ibdev, ibdev 3368 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_dev *dev = to_mdev(ibdev); ibdev 4467 drivers/infiniband/hw/mlx5/main.c static void mlx5_ib_handle_internal_error(struct mlx5_ib_dev *ibdev) ibdev 4480 drivers/infiniband/hw/mlx5/main.c spin_lock_irqsave(&ibdev->reset_flow_resource_lock, flags); ibdev 4481 drivers/infiniband/hw/mlx5/main.c list_for_each_entry(mqp, &ibdev->qp_list, qps_list) { ibdev 4523 drivers/infiniband/hw/mlx5/main.c spin_unlock_irqrestore(&ibdev->reset_flow_resource_lock, flags); ibdev 4546 drivers/infiniband/hw/mlx5/main.c static void handle_general_event(struct mlx5_ib_dev *ibdev, struct mlx5_eqe *eqe, ibdev 4553 drivers/infiniband/hw/mlx5/main.c if (mlx5_ib_port_link_layer(&ibdev->ib_dev, port) == ibdev 4555 drivers/infiniband/hw/mlx5/main.c schedule_work(&ibdev->delay_drop.delay_drop_work); ibdev 4562 drivers/infiniband/hw/mlx5/main.c static int handle_port_change(struct mlx5_ib_dev *ibdev, struct mlx5_eqe *eqe, ibdev 4576 drivers/infiniband/hw/mlx5/main.c if (mlx5_ib_port_link_layer(&ibdev->ib_dev, port) == ibdev 4590 drivers/infiniband/hw/mlx5/main.c schedule_work(&ibdev->devr.ports[port - 1].pkey_change_work); ibdev 4611 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_dev *ibdev; ibdev 4616 drivers/infiniband/hw/mlx5/main.c ibdev = mlx5_ib_get_ibdev_from_mpi(work->mpi); ibdev 4617 drivers/infiniband/hw/mlx5/main.c if (!ibdev) ibdev 4620 drivers/infiniband/hw/mlx5/main.c ibdev = work->dev; ibdev 4626 drivers/infiniband/hw/mlx5/main.c mlx5_ib_handle_internal_error(ibdev); ibdev 4631 drivers/infiniband/hw/mlx5/main.c if (handle_port_change(ibdev, work->param, &ibev)) ibdev 4635 drivers/infiniband/hw/mlx5/main.c handle_general_event(ibdev, work->param, &ibev); ibdev 4641 drivers/infiniband/hw/mlx5/main.c ibev.device = &ibdev->ib_dev; ibdev 4643 drivers/infiniband/hw/mlx5/main.c if (!rdma_is_port_valid(&ibdev->ib_dev, ibev.element.port_num)) { ibdev 4644 drivers/infiniband/hw/mlx5/main.c mlx5_ib_warn(ibdev, "warning: event on port %d\n", ibev.element.port_num); ibdev 4648 drivers/infiniband/hw/mlx5/main.c if (ibdev->ib_active) ibdev 4652 drivers/infiniband/hw/mlx5/main.c ibdev->ib_active = false; ibdev 4931 drivers/infiniband/hw/mlx5/main.c struct ib_device *ibdev; ibdev 4937 drivers/infiniband/hw/mlx5/main.c ibdev = &dev->ib_dev; ibdev 4941 drivers/infiniband/hw/mlx5/main.c devr->p0 = rdma_zalloc_drv_obj(ibdev, ib_pd); ibdev 4945 drivers/infiniband/hw/mlx5/main.c devr->p0->device = ibdev; ibdev 4953 drivers/infiniband/hw/mlx5/main.c devr->c0 = rdma_zalloc_drv_obj(ibdev, ib_cq); ibdev 4995 drivers/infiniband/hw/mlx5/main.c devr->s0 = rdma_zalloc_drv_obj(ibdev, ib_srq); ibdev 5019 drivers/infiniband/hw/mlx5/main.c devr->s1 = rdma_zalloc_drv_obj(ibdev, ib_srq); ibdev 5086 drivers/infiniband/hw/mlx5/main.c static u32 get_core_cap_flags(struct ib_device *ibdev, ibdev 5089 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_dev *dev = to_mdev(ibdev); ibdev 5090 drivers/infiniband/hw/mlx5/main.c enum rdma_link_layer ll = mlx5_ib_port_link_layer(ibdev, 1); ibdev 5120 drivers/infiniband/hw/mlx5/main.c static int mlx5_port_immutable(struct ib_device *ibdev, u8 port_num, ibdev 5124 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_dev *dev = to_mdev(ibdev); ibdev 5125 drivers/infiniband/hw/mlx5/main.c enum rdma_link_layer ll = mlx5_ib_port_link_layer(ibdev, port_num); ibdev 5129 drivers/infiniband/hw/mlx5/main.c err = ib_query_port(ibdev, port_num, &attr); ibdev 5142 drivers/infiniband/hw/mlx5/main.c immutable->core_cap_flags = get_core_cap_flags(ibdev, &rep); ibdev 5149 drivers/infiniband/hw/mlx5/main.c static int mlx5_port_rep_immutable(struct ib_device *ibdev, u8 port_num, ibdev 5157 drivers/infiniband/hw/mlx5/main.c err = ib_query_port(ibdev, port_num, &attr); ibdev 5168 drivers/infiniband/hw/mlx5/main.c static void get_dev_fw_str(struct ib_device *ibdev, char *str) ibdev 5171 drivers/infiniband/hw/mlx5/main.c container_of(ibdev, struct mlx5_ib_dev, ib_dev); ibdev 5505 drivers/infiniband/hw/mlx5/main.c static struct rdma_hw_stats *mlx5_ib_alloc_hw_stats(struct ib_device *ibdev, ibdev 5508 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_dev *dev = to_mdev(ibdev); ibdev 5578 drivers/infiniband/hw/mlx5/main.c static int mlx5_ib_get_hw_stats(struct ib_device *ibdev, ibdev 5582 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_dev *dev = to_mdev(ibdev); ibdev 5829 drivers/infiniband/hw/mlx5/main.c static void mlx5_ib_unbind_slave_port(struct mlx5_ib_dev *ibdev, ibdev 5833 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_port *port = &ibdev->port[port_num]; ibdev 5840 drivers/infiniband/hw/mlx5/main.c mlx5_ib_cleanup_cong_debugfs(ibdev, port_num); ibdev 5843 drivers/infiniband/hw/mlx5/main.c if (!mpi->ibdev) { ibdev 5848 drivers/infiniband/hw/mlx5/main.c mpi->ibdev = NULL; ibdev 5854 drivers/infiniband/hw/mlx5/main.c mlx5_remove_netdev_notifier(ibdev, port_num); ibdev 5878 drivers/infiniband/hw/mlx5/main.c mlx5_ib_dbg(ibdev, "unaffiliated port %d\n", port_num + 1); ibdev 5883 drivers/infiniband/hw/mlx5/main.c mlx5_ib_err(ibdev, "Failed to unaffiliate port %u\n", ibdev 5886 drivers/infiniband/hw/mlx5/main.c ibdev->port[port_num].roce.last_port_state = IB_PORT_DOWN; ibdev 5889 drivers/infiniband/hw/mlx5/main.c static bool mlx5_ib_bind_slave_port(struct mlx5_ib_dev *ibdev, ibdev 5897 drivers/infiniband/hw/mlx5/main.c spin_lock(&ibdev->port[port_num].mp.mpi_lock); ibdev 5898 drivers/infiniband/hw/mlx5/main.c if (ibdev->port[port_num].mp.mpi) { ibdev 5899 drivers/infiniband/hw/mlx5/main.c mlx5_ib_dbg(ibdev, "port %d already affiliated.\n", ibdev 5901 drivers/infiniband/hw/mlx5/main.c spin_unlock(&ibdev->port[port_num].mp.mpi_lock); ibdev 5905 drivers/infiniband/hw/mlx5/main.c ibdev->port[port_num].mp.mpi = mpi; ibdev 5906 drivers/infiniband/hw/mlx5/main.c mpi->ibdev = ibdev; ibdev 5908 drivers/infiniband/hw/mlx5/main.c spin_unlock(&ibdev->port[port_num].mp.mpi_lock); ibdev 5910 drivers/infiniband/hw/mlx5/main.c err = mlx5_nic_vport_affiliate_multiport(ibdev->mdev, mpi->mdev); ibdev 5914 drivers/infiniband/hw/mlx5/main.c err = get_port_caps(ibdev, mlx5_core_native_port_num(mpi->mdev)); ibdev 5918 drivers/infiniband/hw/mlx5/main.c err = mlx5_add_netdev_notifier(ibdev, port_num); ibdev 5920 drivers/infiniband/hw/mlx5/main.c mlx5_ib_err(ibdev, "failed adding netdev notifier for port %u\n", ibdev 5928 drivers/infiniband/hw/mlx5/main.c mlx5_ib_init_cong_debugfs(ibdev, port_num); ibdev 5933 drivers/infiniband/hw/mlx5/main.c mlx5_ib_unbind_slave_port(ibdev, mpi); ibdev 5975 drivers/infiniband/hw/mlx5/main.c mpi->ibdev = dev; ibdev 6948 drivers/infiniband/hw/mlx5/main.c if (mpi->ibdev) ibdev 6949 drivers/infiniband/hw/mlx5/main.c mlx5_ib_unbind_slave_port(mpi->ibdev, mpi); ibdev 852 drivers/infiniband/hw/mlx5/mlx5_ib.h struct mlx5_ib_dev *ibdev; ibdev 907 drivers/infiniband/hw/mlx5/mlx5_ib.h int (*read_counters)(struct ib_device *ibdev, ibdev 1015 drivers/infiniband/hw/mlx5/mlx5_ib.h static inline struct mlx5_ib_dev *to_mdev(struct ib_device *ibdev) ibdev 1017 drivers/infiniband/hw/mlx5/mlx5_ib.h return container_of(ibdev, struct mlx5_ib_dev, ib_dev); ibdev 1181 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_ib_process_mad(struct ib_device *ibdev, int mad_flags, u8 port_num, ibdev 1186 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_xrcd *mlx5_ib_alloc_xrcd(struct ib_device *ibdev, ibdev 1191 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_query_mad_ifc_smp_attr_node_info(struct ib_device *ibdev, ibdev 1193 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_query_mad_ifc_system_image_guid(struct ib_device *ibdev, ibdev 1195 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_query_mad_ifc_max_pkeys(struct ib_device *ibdev, ibdev 1197 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_query_mad_ifc_vendor_id(struct ib_device *ibdev, ibdev 1201 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_query_mad_ifc_pkey(struct ib_device *ibdev, u8 port, u16 index, ibdev 1203 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_query_mad_ifc_gids(struct ib_device *ibdev, u8 port, int index, ibdev 1205 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_query_mad_ifc_port(struct ib_device *ibdev, u8 port, ibdev 1207 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_ib_query_port(struct ib_device *ibdev, u8 port, ibdev 1240 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_dm *mlx5_ib_alloc_dm(struct ib_device *ibdev, ibdev 1251 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_ib_odp_init_one(struct mlx5_ib_dev *ibdev); ibdev 1252 drivers/infiniband/hw/mlx5/mlx5_ib.h void mlx5_ib_odp_cleanup_one(struct mlx5_ib_dev *ibdev); ibdev 1270 drivers/infiniband/hw/mlx5/mlx5_ib.h static inline int mlx5_ib_odp_init_one(struct mlx5_ib_dev *ibdev) { return 0; } ibdev 1271 drivers/infiniband/hw/mlx5/mlx5_ib.h static inline void mlx5_ib_odp_cleanup_one(struct mlx5_ib_dev *ibdev) {} ibdev 5480 drivers/infiniband/hw/mlx5/qp.c static void to_rdma_ah_attr(struct mlx5_ib_dev *ibdev, ibdev 5487 drivers/infiniband/hw/mlx5/qp.c if (!path->port || path->port > ibdev->num_ports) ibdev 5490 drivers/infiniband/hw/mlx5/qp.c ah_attr->type = rdma_ah_find_type(&ibdev->ib_dev, path->port); ibdev 5838 drivers/infiniband/hw/mlx5/qp.c struct ib_xrcd *mlx5_ib_alloc_xrcd(struct ib_device *ibdev, ibdev 5841 drivers/infiniband/hw/mlx5/qp.c struct mlx5_ib_dev *dev = to_mdev(ibdev); ibdev 579 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_process_mad(struct ib_device *ibdev, ibdev 590 drivers/infiniband/hw/mthca/mthca_dev.h static inline struct mthca_dev *to_mdev(struct ib_device *ibdev) ibdev 592 drivers/infiniband/hw/mthca/mthca_dev.h return container_of(ibdev, struct mthca_dev, ib_dev); ibdev 107 drivers/infiniband/hw/mthca/mthca_mad.c static void smp_snoop(struct ib_device *ibdev, ibdev 122 drivers/infiniband/hw/mthca/mthca_mad.c mthca_update_rate(to_mdev(ibdev), port_num); ibdev 123 drivers/infiniband/hw/mthca/mthca_mad.c update_sm_ah(to_mdev(ibdev), port_num, ibdev 127 drivers/infiniband/hw/mthca/mthca_mad.c event.device = ibdev; ibdev 142 drivers/infiniband/hw/mthca/mthca_mad.c event.device = ibdev; ibdev 199 drivers/infiniband/hw/mthca/mthca_mad.c int mthca_process_mad(struct ib_device *ibdev, ibdev 222 drivers/infiniband/hw/mthca/mthca_mad.c forward_trap(to_mdev(ibdev), port_num, in_mad); ibdev 259 drivers/infiniband/hw/mthca/mthca_mad.c !ib_query_port(ibdev, port_num, &pattr)) ibdev 262 drivers/infiniband/hw/mthca/mthca_mad.c err = mthca_MAD_IFC(to_mdev(ibdev), ibdev 269 drivers/infiniband/hw/mthca/mthca_mad.c mthca_err(to_mdev(ibdev), "MAD_IFC returned %d\n", err); ibdev 274 drivers/infiniband/hw/mthca/mthca_mad.c smp_snoop(ibdev, port_num, in_mad, prev_lid); ibdev 275 drivers/infiniband/hw/mthca/mthca_mad.c node_desc_override(ibdev, out_mad); ibdev 61 drivers/infiniband/hw/mthca/mthca_provider.c static int mthca_query_device(struct ib_device *ibdev, struct ib_device_attr *props, ibdev 67 drivers/infiniband/hw/mthca/mthca_provider.c struct mthca_dev *mdev = to_mdev(ibdev); ibdev 139 drivers/infiniband/hw/mthca/mthca_provider.c static int mthca_query_port(struct ib_device *ibdev, ibdev 157 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1, ibdev 169 drivers/infiniband/hw/mthca/mthca_provider.c props->gid_tbl_len = to_mdev(ibdev)->limits.gid_table_len; ibdev 171 drivers/infiniband/hw/mthca/mthca_provider.c props->pkey_tbl_len = to_mdev(ibdev)->limits.pkey_table_len; ibdev 188 drivers/infiniband/hw/mthca/mthca_provider.c static int mthca_modify_device(struct ib_device *ibdev, ibdev 196 drivers/infiniband/hw/mthca/mthca_provider.c if (mutex_lock_interruptible(&to_mdev(ibdev)->cap_mask_mutex)) ibdev 198 drivers/infiniband/hw/mthca/mthca_provider.c memcpy(ibdev->node_desc, props->node_desc, ibdev 200 drivers/infiniband/hw/mthca/mthca_provider.c mutex_unlock(&to_mdev(ibdev)->cap_mask_mutex); ibdev 206 drivers/infiniband/hw/mthca/mthca_provider.c static int mthca_modify_port(struct ib_device *ibdev, ibdev 214 drivers/infiniband/hw/mthca/mthca_provider.c if (mutex_lock_interruptible(&to_mdev(ibdev)->cap_mask_mutex)) ibdev 217 drivers/infiniband/hw/mthca/mthca_provider.c err = ib_query_port(ibdev, port, &attr); ibdev 227 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_SET_IB(to_mdev(ibdev), &set_ib, port); ibdev 231 drivers/infiniband/hw/mthca/mthca_provider.c mutex_unlock(&to_mdev(ibdev)->cap_mask_mutex); ibdev 235 drivers/infiniband/hw/mthca/mthca_provider.c static int mthca_query_pkey(struct ib_device *ibdev, ibdev 251 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1, ibdev 264 drivers/infiniband/hw/mthca/mthca_provider.c static int mthca_query_gid(struct ib_device *ibdev, u8 port, ibdev 280 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1, ibdev 291 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_MAD_IFC(to_mdev(ibdev), 1, 1, ibdev 307 drivers/infiniband/hw/mthca/mthca_provider.c struct ib_device *ibdev = uctx->device; ibdev 312 drivers/infiniband/hw/mthca/mthca_provider.c if (!(to_mdev(ibdev)->active)) ibdev 315 drivers/infiniband/hw/mthca/mthca_provider.c uresp.qp_tab_size = to_mdev(ibdev)->limits.num_qps; ibdev 316 drivers/infiniband/hw/mthca/mthca_provider.c if (mthca_is_memfree(to_mdev(ibdev))) ibdev 317 drivers/infiniband/hw/mthca/mthca_provider.c uresp.uarc_size = to_mdev(ibdev)->uar_table.uarc_size; ibdev 321 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_uar_alloc(to_mdev(ibdev), &context->uar); ibdev 325 drivers/infiniband/hw/mthca/mthca_provider.c context->db_tab = mthca_init_user_db_tab(to_mdev(ibdev)); ibdev 328 drivers/infiniband/hw/mthca/mthca_provider.c mthca_uar_free(to_mdev(ibdev), &context->uar); ibdev 333 drivers/infiniband/hw/mthca/mthca_provider.c mthca_cleanup_user_db_tab(to_mdev(ibdev), &context->uar, context->db_tab); ibdev 334 drivers/infiniband/hw/mthca/mthca_provider.c mthca_uar_free(to_mdev(ibdev), &context->uar); ibdev 368 drivers/infiniband/hw/mthca/mthca_provider.c struct ib_device *ibdev = ibpd->device; ibdev 372 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_pd_alloc(to_mdev(ibdev), !udata, pd); ibdev 378 drivers/infiniband/hw/mthca/mthca_provider.c mthca_pd_free(to_mdev(ibdev), pd); ibdev 608 drivers/infiniband/hw/mthca/mthca_provider.c struct ib_device *ibdev = ibcq->device; ibdev 620 drivers/infiniband/hw/mthca/mthca_provider.c if (entries < 1 || entries > to_mdev(ibdev)->limits.max_cqes) ibdev 627 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_map_user_db(to_mdev(ibdev), &context->uar, ibdev 633 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_map_user_db(to_mdev(ibdev), &context->uar, ibdev 651 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_init_cq(to_mdev(ibdev), nent, context, ibdev 652 drivers/infiniband/hw/mthca/mthca_provider.c udata ? ucmd.pdn : to_mdev(ibdev)->driver_pd.pd_num, ibdev 658 drivers/infiniband/hw/mthca/mthca_provider.c mthca_free_cq(to_mdev(ibdev), cq); ibdev 669 drivers/infiniband/hw/mthca/mthca_provider.c mthca_unmap_user_db(to_mdev(ibdev), &context->uar, ibdev 674 drivers/infiniband/hw/mthca/mthca_provider.c mthca_unmap_user_db(to_mdev(ibdev), &context->uar, ibdev 1116 drivers/infiniband/hw/mthca/mthca_provider.c static int mthca_port_immutable(struct ib_device *ibdev, u8 port_num, ibdev 1124 drivers/infiniband/hw/mthca/mthca_provider.c err = ib_query_port(ibdev, port_num, &attr); ibdev 241 drivers/infiniband/hw/ocrdma/ocrdma.h struct ib_device ibdev; ibdev 457 drivers/infiniband/hw/ocrdma/ocrdma.h static inline struct ocrdma_dev *get_ocrdma_dev(struct ib_device *ibdev) ibdev 459 drivers/infiniband/hw/ocrdma/ocrdma.h return container_of(ibdev, struct ocrdma_dev, ibdev); ibdev 250 drivers/infiniband/hw/ocrdma/ocrdma_ah.c int ocrdma_process_mad(struct ib_device *ibdev, ibdev 270 drivers/infiniband/hw/ocrdma/ocrdma_ah.c dev = get_ocrdma_dev(ibdev); ibdev 717 drivers/infiniband/hw/ocrdma/ocrdma_hw.c ib_evt.device = &dev->ibdev; ibdev 795 drivers/infiniband/hw/ocrdma/ocrdma_hw.c dev_err(&dev->ibdev.dev, "Fatal event received\n"); ibdev 85 drivers/infiniband/hw/ocrdma/ocrdma_main.c static int ocrdma_port_immutable(struct ib_device *ibdev, u8 port_num, ibdev 92 drivers/infiniband/hw/ocrdma/ocrdma_main.c dev = get_ocrdma_dev(ibdev); ibdev 97 drivers/infiniband/hw/ocrdma/ocrdma_main.c err = ib_query_port(ibdev, port_num, &attr); ibdev 120 drivers/infiniband/hw/ocrdma/ocrdma_main.c rdma_device_to_drv_device(device, struct ocrdma_dev, ibdev); ibdev 130 drivers/infiniband/hw/ocrdma/ocrdma_main.c rdma_device_to_drv_device(device, struct ocrdma_dev, ibdev); ibdev 204 drivers/infiniband/hw/ocrdma/ocrdma_main.c ocrdma_get_guid(dev, (u8 *)&dev->ibdev.node_guid); ibdev 206 drivers/infiniband/hw/ocrdma/ocrdma_main.c memcpy(dev->ibdev.node_desc, OCRDMA_NODE_DESC, ibdev 208 drivers/infiniband/hw/ocrdma/ocrdma_main.c dev->ibdev.uverbs_cmd_mask = ibdev 229 drivers/infiniband/hw/ocrdma/ocrdma_main.c dev->ibdev.uverbs_cmd_mask |= ibdev 235 drivers/infiniband/hw/ocrdma/ocrdma_main.c dev->ibdev.node_type = RDMA_NODE_IB_CA; ibdev 236 drivers/infiniband/hw/ocrdma/ocrdma_main.c dev->ibdev.phys_port_cnt = 1; ibdev 237 drivers/infiniband/hw/ocrdma/ocrdma_main.c dev->ibdev.num_comp_vectors = dev->eq_cnt; ibdev 240 drivers/infiniband/hw/ocrdma/ocrdma_main.c dev->ibdev.dev.parent = &dev->nic_info.pdev->dev; ibdev 242 drivers/infiniband/hw/ocrdma/ocrdma_main.c ib_set_device_ops(&dev->ibdev, &ocrdma_dev_ops); ibdev 245 drivers/infiniband/hw/ocrdma/ocrdma_main.c dev->ibdev.uverbs_cmd_mask |= ibdev 252 drivers/infiniband/hw/ocrdma/ocrdma_main.c ib_set_device_ops(&dev->ibdev, &ocrdma_dev_srq_ops); ibdev 254 drivers/infiniband/hw/ocrdma/ocrdma_main.c rdma_set_device_sysfs_group(&dev->ibdev, &ocrdma_attr_group); ibdev 255 drivers/infiniband/hw/ocrdma/ocrdma_main.c ret = ib_device_set_netdev(&dev->ibdev, dev->nic_info.netdev, 1); ibdev 259 drivers/infiniband/hw/ocrdma/ocrdma_main.c return ib_register_device(&dev->ibdev, "ocrdma%d"); ibdev 311 drivers/infiniband/hw/ocrdma/ocrdma_main.c dev = ib_alloc_device(ocrdma_dev, ibdev); ibdev 360 drivers/infiniband/hw/ocrdma/ocrdma_main.c ib_dealloc_device(&dev->ibdev); ibdev 369 drivers/infiniband/hw/ocrdma/ocrdma_main.c ib_dealloc_device(&dev->ibdev); ibdev 378 drivers/infiniband/hw/ocrdma/ocrdma_main.c ib_unregister_device(&dev->ibdev); ibdev 392 drivers/infiniband/hw/ocrdma/ocrdma_main.c port_event.device = &dev->ibdev; ibdev 403 drivers/infiniband/hw/ocrdma/ocrdma_main.c err_event.device = &dev->ibdev; ibdev 57 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c int ocrdma_query_pkey(struct ib_device *ibdev, u8 port, u16 index, u16 *pkey) ibdev 66 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c int ocrdma_query_device(struct ib_device *ibdev, struct ib_device_attr *attr, ibdev 69 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_dev *dev = get_ocrdma_dev(ibdev); ibdev 154 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c int ocrdma_query_port(struct ib_device *ibdev, ibdev 162 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c dev = get_ocrdma_dev(ibdev); ibdev 193 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c int ocrdma_modify_port(struct ib_device *ibdev, u8 port, int mask, ibdev 413 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ib_device *ibdev = &dev->ibdev; ibdev 417 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c pd = rdma_zalloc_drv_obj(ibdev, ib_pd); ibdev 421 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c pd->device = ibdev; ibdev 431 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c uctx->cntxt_pd->ibpd.device = &dev->ibdev; ibdev 473 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ib_device *ibdev = uctx->device; ibdev 477 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_dev *dev = get_ocrdma_dev(ibdev); ibdev 637 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ib_device *ibdev = ibpd->device; ibdev 638 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_dev *dev = get_ocrdma_dev(ibdev); ibdev 982 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ib_device *ibdev = ibcq->device; ibdev 985 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_dev *dev = get_ocrdma_dev(ibdev); ibdev 61 drivers/infiniband/hw/qedr/main.c ibev.device = &dev->ibdev; ibdev 74 drivers/infiniband/hw/qedr/main.c static void qedr_get_dev_fw_str(struct ib_device *ibdev, char *str) ibdev 76 drivers/infiniband/hw/qedr/main.c struct qedr_dev *qedr = get_qedr_dev(ibdev); ibdev 84 drivers/infiniband/hw/qedr/main.c static int qedr_roce_port_immutable(struct ib_device *ibdev, u8 port_num, ibdev 90 drivers/infiniband/hw/qedr/main.c err = qedr_query_port(ibdev, port_num, &attr); ibdev 103 drivers/infiniband/hw/qedr/main.c static int qedr_iw_port_immutable(struct ib_device *ibdev, u8 port_num, ibdev 109 drivers/infiniband/hw/qedr/main.c err = qedr_query_port(ibdev, port_num, &attr); ibdev 126 drivers/infiniband/hw/qedr/main.c rdma_device_to_drv_device(device, struct qedr_dev, ibdev); ibdev 136 drivers/infiniband/hw/qedr/main.c rdma_device_to_drv_device(device, struct qedr_dev, ibdev); ibdev 140 drivers/infiniband/hw/qedr/main.c rdma_protocol_iwarp(&dev->ibdev, 1) ? ibdev 170 drivers/infiniband/hw/qedr/main.c dev->ibdev.node_type = RDMA_NODE_RNIC; ibdev 172 drivers/infiniband/hw/qedr/main.c ib_set_device_ops(&dev->ibdev, &qedr_iw_dev_ops); ibdev 174 drivers/infiniband/hw/qedr/main.c memcpy(dev->ibdev.iw_ifname, ibdev 175 drivers/infiniband/hw/qedr/main.c dev->ndev->name, sizeof(dev->ibdev.iw_ifname)); ibdev 186 drivers/infiniband/hw/qedr/main.c dev->ibdev.node_type = RDMA_NODE_IB_CA; ibdev 188 drivers/infiniband/hw/qedr/main.c ib_set_device_ops(&dev->ibdev, &qedr_roce_dev_ops); ibdev 243 drivers/infiniband/hw/qedr/main.c dev->ibdev.node_guid = dev->attr.node_guid; ibdev 244 drivers/infiniband/hw/qedr/main.c memcpy(dev->ibdev.node_desc, QEDR_NODE_DESC, sizeof(QEDR_NODE_DESC)); ibdev 246 drivers/infiniband/hw/qedr/main.c dev->ibdev.uverbs_cmd_mask = QEDR_UVERBS(GET_CONTEXT) | ibdev 279 drivers/infiniband/hw/qedr/main.c dev->ibdev.phys_port_cnt = 1; ibdev 280 drivers/infiniband/hw/qedr/main.c dev->ibdev.num_comp_vectors = dev->num_cnq; ibdev 281 drivers/infiniband/hw/qedr/main.c dev->ibdev.dev.parent = &dev->pdev->dev; ibdev 283 drivers/infiniband/hw/qedr/main.c rdma_set_device_sysfs_group(&dev->ibdev, &qedr_attr_group); ibdev 284 drivers/infiniband/hw/qedr/main.c ib_set_device_ops(&dev->ibdev, &qedr_dev_ops); ibdev 286 drivers/infiniband/hw/qedr/main.c rc = ib_device_set_netdev(&dev->ibdev, dev->ndev, 1); ibdev 290 drivers/infiniband/hw/qedr/main.c return ib_register_device(&dev->ibdev, "qedr%d"); ibdev 857 drivers/infiniband/hw/qedr/main.c dev = ib_alloc_device(qedr_dev, ibdev); ibdev 937 drivers/infiniband/hw/qedr/main.c ib_dealloc_device(&dev->ibdev); ibdev 948 drivers/infiniband/hw/qedr/main.c ib_unregister_device(&dev->ibdev); ibdev 957 drivers/infiniband/hw/qedr/main.c ib_dealloc_device(&dev->ibdev); ibdev 47 drivers/infiniband/hw/qedr/qedr.h #define DP_NAME(_dev) dev_name(&(_dev)->ibdev.dev) ibdev 128 drivers/infiniband/hw/qedr/qedr.h struct ib_device ibdev; ibdev 558 drivers/infiniband/hw/qedr/qedr.h static inline struct qedr_dev *get_qedr_dev(struct ib_device *ibdev) ibdev 560 drivers/infiniband/hw/qedr/qedr.h return container_of(ibdev, struct qedr_dev, ibdev); ibdev 799 drivers/infiniband/hw/qedr/qedr_iw_cm.c struct ib_qp *qedr_iw_get_qp(struct ib_device *ibdev, int qpn) ibdev 801 drivers/infiniband/hw/qedr/qedr_iw_cm.c struct qedr_dev *dev = get_qedr_dev(ibdev); ibdev 70 drivers/infiniband/hw/qedr/verbs.c int qedr_query_pkey(struct ib_device *ibdev, u8 port, u16 index, u16 *pkey) ibdev 79 drivers/infiniband/hw/qedr/verbs.c int qedr_iw_query_gid(struct ib_device *ibdev, u8 port, ibdev 82 drivers/infiniband/hw/qedr/verbs.c struct qedr_dev *dev = get_qedr_dev(ibdev); ibdev 106 drivers/infiniband/hw/qedr/verbs.c int qedr_query_device(struct ib_device *ibdev, ibdev 109 drivers/infiniband/hw/qedr/verbs.c struct qedr_dev *dev = get_qedr_dev(ibdev); ibdev 208 drivers/infiniband/hw/qedr/verbs.c int qedr_query_port(struct ib_device *ibdev, u8 port, struct ib_port_attr *attr) ibdev 213 drivers/infiniband/hw/qedr/verbs.c dev = get_qedr_dev(ibdev); ibdev 237 drivers/infiniband/hw/qedr/verbs.c if (rdma_protocol_iwarp(&dev->ibdev, 1)) { ibdev 254 drivers/infiniband/hw/qedr/verbs.c int qedr_modify_port(struct ib_device *ibdev, u8 port, int mask, ibdev 316 drivers/infiniband/hw/qedr/verbs.c struct ib_device *ibdev = uctx->device; ibdev 320 drivers/infiniband/hw/qedr/verbs.c struct qedr_dev *dev = get_qedr_dev(ibdev); ibdev 436 drivers/infiniband/hw/qedr/verbs.c struct ib_device *ibdev = ibpd->device; ibdev 437 drivers/infiniband/hw/qedr/verbs.c struct qedr_dev *dev = get_qedr_dev(ibdev); ibdev 813 drivers/infiniband/hw/qedr/verbs.c struct ib_device *ibdev = ibcq->device; ibdev 818 drivers/infiniband/hw/qedr/verbs.c struct qedr_dev *dev = get_qedr_dev(ibdev); ibdev 1144 drivers/infiniband/hw/qedr/verbs.c if (rdma_protocol_iwarp(&dev->ibdev, 1)) { ibdev 1163 drivers/infiniband/hw/qedr/verbs.c if (rdma_protocol_iwarp(&dev->ibdev, 1)) ibdev 1197 drivers/infiniband/hw/qedr/verbs.c if (rdma_protocol_iwarp(&dev->ibdev, 1)) { ibdev 1585 drivers/infiniband/hw/qedr/verbs.c if (rdma_protocol_roce(&dev->ibdev, 1)) { ibdev 1604 drivers/infiniband/hw/qedr/verbs.c int alloc_and_init = rdma_protocol_roce(&dev->ibdev, 1); ibdev 1648 drivers/infiniband/hw/qedr/verbs.c if (rdma_protocol_iwarp(&dev->ibdev, 1)) ibdev 1872 drivers/infiniband/hw/qedr/verbs.c if (rdma_protocol_iwarp(&dev->ibdev, 1)) ibdev 1934 drivers/infiniband/hw/qedr/verbs.c if (rdma_protocol_iwarp(&dev->ibdev, 1)) { ibdev 2029 drivers/infiniband/hw/qedr/verbs.c if (rdma_protocol_roce(&dev->ibdev, 1)) { ibdev 2110 drivers/infiniband/hw/qedr/verbs.c struct qedr_dev *dev = get_qedr_dev(&qp->dev->ibdev); ibdev 2126 drivers/infiniband/hw/qedr/verbs.c if (rdma_protocol_roce(&dev->ibdev, 1)) { ibdev 2175 drivers/infiniband/hw/qedr/verbs.c if (rdma_protocol_iwarp(&dev->ibdev, 1)) ibdev 2472 drivers/infiniband/hw/qedr/verbs.c if (rdma_protocol_roce(&dev->ibdev, 1)) { ibdev 2514 drivers/infiniband/hw/qedr/verbs.c if (rdma_protocol_iwarp(&dev->ibdev, 1)) ibdev 2519 drivers/infiniband/hw/qedr/verbs.c if (rdma_protocol_iwarp(&dev->ibdev, 1)) ibdev 3218 drivers/infiniband/hw/qedr/verbs.c if (unlikely(rdma_protocol_iwarp(&dev->ibdev, 1))) { ibdev 3264 drivers/infiniband/hw/qedr/verbs.c if (unlikely(rdma_protocol_iwarp(&dev->ibdev, 1))) { ibdev 3407 drivers/infiniband/hw/qedr/verbs.c if (rdma_protocol_roce(&dev->ibdev, 1)) { ibdev 3640 drivers/infiniband/hw/qedr/verbs.c if (rdma_protocol_iwarp(&dev->ibdev, 1)) { ibdev 4132 drivers/infiniband/hw/qedr/verbs.c int qedr_process_mad(struct ib_device *ibdev, int process_mad_flags, ibdev 4140 drivers/infiniband/hw/qedr/verbs.c struct qedr_dev *dev = get_qedr_dev(ibdev); ibdev 35 drivers/infiniband/hw/qedr/verbs.h int qedr_query_device(struct ib_device *ibdev, ibdev 41 drivers/infiniband/hw/qedr/verbs.h int qedr_iw_query_gid(struct ib_device *ibdev, u8 port, ibdev 94 drivers/infiniband/hw/qedr/verbs.h int qedr_process_mad(struct ib_device *ibdev, int process_mad_flags, ibdev 101 drivers/infiniband/hw/qedr/verbs.h int qedr_port_immutable(struct ib_device *ibdev, u8 port_num, ibdev 1193 drivers/infiniband/hw/qib/qib.h static inline struct qib_devdata *dd_from_ibdev(struct ib_device *ibdev) ibdev 1195 drivers/infiniband/hw/qib/qib.h return dd_from_dev(to_idev(ibdev)); ibdev 1203 drivers/infiniband/hw/qib/qib.h static inline struct qib_ibport *to_iport(struct ib_device *ibdev, u8 port) ibdev 1205 drivers/infiniband/hw/qib/qib.h struct qib_devdata *dd = dd_from_ibdev(ibdev); ibdev 1398 drivers/infiniband/hw/qib/qib.h int qib_create_port_files(struct ib_device *ibdev, u8 port_num, ibdev 84 drivers/infiniband/hw/qib/qib_driver.c struct qib_ibdev *ibdev = container_of(rdi, struct qib_ibdev, rdi); ibdev 85 drivers/infiniband/hw/qib/qib_driver.c struct qib_devdata *dd = container_of(ibdev, ibdev 77 drivers/infiniband/hw/qib/qib_intr.c event.device = &dd->verbs_dev.rdi.ibdev; ibdev 208 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibdev *ibdev = container_of(rdi, struct qib_ibdev, rdi); ibdev 209 drivers/infiniband/hw/qib/qib_mad.c struct qib_devdata *dd = dd_from_dev(ibdev); ibdev 268 drivers/infiniband/hw/qib/qib_mad.c struct ib_device *ibdev) ibdev 273 drivers/infiniband/hw/qib/qib_mad.c memcpy(smp->data, ibdev->node_desc, sizeof(smp->data)); ibdev 278 drivers/infiniband/hw/qib/qib_mad.c static int subn_get_nodeinfo(struct ib_smp *smp, struct ib_device *ibdev, ibdev 282 drivers/infiniband/hw/qib/qib_mad.c struct qib_devdata *dd = dd_from_ibdev(ibdev); ibdev 296 drivers/infiniband/hw/qib/qib_mad.c nip->num_ports = ibdev->phys_port_cnt; ibdev 313 drivers/infiniband/hw/qib/qib_mad.c static int subn_get_guidinfo(struct ib_smp *smp, struct ib_device *ibdev, ibdev 316 drivers/infiniband/hw/qib/qib_mad.c struct qib_devdata *dd = dd_from_ibdev(ibdev); ibdev 454 drivers/infiniband/hw/qib/qib_mad.c static int subn_get_portinfo(struct ib_smp *smp, struct ib_device *ibdev, ibdev 469 drivers/infiniband/hw/qib/qib_mad.c if (port_num > ibdev->phys_port_cnt) { ibdev 475 drivers/infiniband/hw/qib/qib_mad.c ibp = to_iport(ibdev, port_num); ibdev 484 drivers/infiniband/hw/qib/qib_mad.c dd = dd_from_ibdev(ibdev); ibdev 596 drivers/infiniband/hw/qib/qib_mad.c static int subn_get_pkeytable(struct ib_smp *smp, struct ib_device *ibdev, ibdev 607 drivers/infiniband/hw/qib/qib_mad.c struct qib_devdata *dd = dd_from_ibdev(ibdev); ibdev 620 drivers/infiniband/hw/qib/qib_mad.c static int subn_set_guidinfo(struct ib_smp *smp, struct ib_device *ibdev, ibdev 623 drivers/infiniband/hw/qib/qib_mad.c struct qib_devdata *dd = dd_from_ibdev(ibdev); ibdev 642 drivers/infiniband/hw/qib/qib_mad.c return subn_get_guidinfo(smp, ibdev, port); ibdev 653 drivers/infiniband/hw/qib/qib_mad.c static int subn_set_portinfo(struct ib_smp *smp, struct ib_device *ibdev, ibdev 676 drivers/infiniband/hw/qib/qib_mad.c if (port_num > ibdev->phys_port_cnt) ibdev 683 drivers/infiniband/hw/qib/qib_mad.c dd = dd_from_ibdev(ibdev); ibdev 687 drivers/infiniband/hw/qib/qib_mad.c event.device = ibdev; ibdev 882 drivers/infiniband/hw/qib/qib_mad.c ret = subn_get_portinfo(smp, ibdev, port); ibdev 1032 drivers/infiniband/hw/qib/qib_mad.c event.device = &dd->verbs_dev.rdi.ibdev; ibdev 1039 drivers/infiniband/hw/qib/qib_mad.c static int subn_set_pkeytable(struct ib_smp *smp, struct ib_device *ibdev, ibdev 1045 drivers/infiniband/hw/qib/qib_mad.c struct qib_devdata *dd = dd_from_ibdev(ibdev); ibdev 1054 drivers/infiniband/hw/qib/qib_mad.c return subn_get_pkeytable(smp, ibdev, port); ibdev 1057 drivers/infiniband/hw/qib/qib_mad.c static int subn_get_sl_to_vl(struct ib_smp *smp, struct ib_device *ibdev, ibdev 1060 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibport *ibp = to_iport(ibdev, port); ibdev 1075 drivers/infiniband/hw/qib/qib_mad.c static int subn_set_sl_to_vl(struct ib_smp *smp, struct ib_device *ibdev, ibdev 1078 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibport *ibp = to_iport(ibdev, port); ibdev 1091 drivers/infiniband/hw/qib/qib_mad.c qib_set_uevent_bits(ppd_from_ibp(to_iport(ibdev, port)), ibdev 1094 drivers/infiniband/hw/qib/qib_mad.c return subn_get_sl_to_vl(smp, ibdev, port); ibdev 1097 drivers/infiniband/hw/qib/qib_mad.c static int subn_get_vl_arb(struct ib_smp *smp, struct ib_device *ibdev, ibdev 1101 drivers/infiniband/hw/qib/qib_mad.c struct qib_pportdata *ppd = ppd_from_ibp(to_iport(ibdev, port)); ibdev 1119 drivers/infiniband/hw/qib/qib_mad.c static int subn_set_vl_arb(struct ib_smp *smp, struct ib_device *ibdev, ibdev 1123 drivers/infiniband/hw/qib/qib_mad.c struct qib_pportdata *ppd = ppd_from_ibp(to_iport(ibdev, port)); ibdev 1136 drivers/infiniband/hw/qib/qib_mad.c return subn_get_vl_arb(smp, ibdev, port); ibdev 1139 drivers/infiniband/hw/qib/qib_mad.c static int subn_trap_repress(struct ib_smp *smp, struct ib_device *ibdev, ibdev 1152 drivers/infiniband/hw/qib/qib_mad.c struct ib_device *ibdev) ibdev 1156 drivers/infiniband/hw/qib/qib_mad.c struct qib_devdata *dd = dd_from_ibdev(ibdev); ibdev 1183 drivers/infiniband/hw/qib/qib_mad.c struct ib_device *ibdev, u8 port) ibdev 1187 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibdev *dev = to_idev(ibdev); ibdev 1189 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibport *ibp = to_iport(ibdev, port); ibdev 1221 drivers/infiniband/hw/qib/qib_mad.c struct ib_device *ibdev, u8 port) ibdev 1225 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibdev *dev = to_idev(ibdev); ibdev 1227 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibport *ibp = to_iport(ibdev, port); ibdev 1261 drivers/infiniband/hw/qib/qib_mad.c ret = pma_get_portsamplescontrol(pmp, ibdev, port); ibdev 1350 drivers/infiniband/hw/qib/qib_mad.c struct ib_device *ibdev, u8 port) ibdev 1354 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibdev *dev = to_idev(ibdev); ibdev 1356 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibport *ibp = to_iport(ibdev, port); ibdev 1389 drivers/infiniband/hw/qib/qib_mad.c struct ib_device *ibdev, u8 port) ibdev 1393 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibdev *dev = to_idev(ibdev); ibdev 1395 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibport *ibp = to_iport(ibdev, port); ibdev 1431 drivers/infiniband/hw/qib/qib_mad.c struct ib_device *ibdev, u8 port) ibdev 1435 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibport *ibp = to_iport(ibdev, port); ibdev 1529 drivers/infiniband/hw/qib/qib_mad.c struct ib_device *ibdev, u8 port) ibdev 1535 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibport *ibp = to_iport(ibdev, port); ibdev 1659 drivers/infiniband/hw/qib/qib_mad.c struct ib_device *ibdev, u8 port) ibdev 1663 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibport *ibp = to_iport(ibdev, port); ibdev 1706 drivers/infiniband/hw/qib/qib_mad.c struct ib_device *ibdev, u8 port) ibdev 1710 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibport *ibp = to_iport(ibdev, port); ibdev 1765 drivers/infiniband/hw/qib/qib_mad.c return pma_get_portcounters(pmp, ibdev, port); ibdev 1769 drivers/infiniband/hw/qib/qib_mad.c struct ib_device *ibdev, u8 port) ibdev 1771 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibport *ibp = to_iport(ibdev, port); ibdev 1781 drivers/infiniband/hw/qib/qib_mad.c ret = pma_get_portcounters_cong(pmp, ibdev, port); ibdev 1820 drivers/infiniband/hw/qib/qib_mad.c struct ib_device *ibdev, u8 port) ibdev 1824 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibport *ibp = to_iport(ibdev, port); ibdev 1857 drivers/infiniband/hw/qib/qib_mad.c return pma_get_portcounters_ext(pmp, ibdev, port); ibdev 1860 drivers/infiniband/hw/qib/qib_mad.c static int process_subn(struct ib_device *ibdev, int mad_flags, ibdev 1865 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibport *ibp = to_iport(ibdev, port); ibdev 1890 drivers/infiniband/hw/qib/qib_mad.c port_num && port_num <= ibdev->phys_port_cnt && ibdev 1892 drivers/infiniband/hw/qib/qib_mad.c (void) check_mkey(to_iport(ibdev, port_num), smp, 0); ibdev 1901 drivers/infiniband/hw/qib/qib_mad.c ret = subn_get_nodedescription(smp, ibdev); ibdev 1904 drivers/infiniband/hw/qib/qib_mad.c ret = subn_get_nodeinfo(smp, ibdev, port); ibdev 1907 drivers/infiniband/hw/qib/qib_mad.c ret = subn_get_guidinfo(smp, ibdev, port); ibdev 1910 drivers/infiniband/hw/qib/qib_mad.c ret = subn_get_portinfo(smp, ibdev, port); ibdev 1913 drivers/infiniband/hw/qib/qib_mad.c ret = subn_get_pkeytable(smp, ibdev, port); ibdev 1916 drivers/infiniband/hw/qib/qib_mad.c ret = subn_get_sl_to_vl(smp, ibdev, port); ibdev 1919 drivers/infiniband/hw/qib/qib_mad.c ret = subn_get_vl_arb(smp, ibdev, port); ibdev 1941 drivers/infiniband/hw/qib/qib_mad.c ret = subn_set_guidinfo(smp, ibdev, port); ibdev 1944 drivers/infiniband/hw/qib/qib_mad.c ret = subn_set_portinfo(smp, ibdev, port); ibdev 1947 drivers/infiniband/hw/qib/qib_mad.c ret = subn_set_pkeytable(smp, ibdev, port); ibdev 1950 drivers/infiniband/hw/qib/qib_mad.c ret = subn_set_sl_to_vl(smp, ibdev, port); ibdev 1953 drivers/infiniband/hw/qib/qib_mad.c ret = subn_set_vl_arb(smp, ibdev, port); ibdev 1974 drivers/infiniband/hw/qib/qib_mad.c ret = subn_trap_repress(smp, ibdev, port); ibdev 2012 drivers/infiniband/hw/qib/qib_mad.c static int process_perf(struct ib_device *ibdev, u8 port, ibdev 2030 drivers/infiniband/hw/qib/qib_mad.c ret = pma_get_classportinfo(pmp, ibdev); ibdev 2033 drivers/infiniband/hw/qib/qib_mad.c ret = pma_get_portsamplescontrol(pmp, ibdev, port); ibdev 2036 drivers/infiniband/hw/qib/qib_mad.c ret = pma_get_portsamplesresult(pmp, ibdev, port); ibdev 2039 drivers/infiniband/hw/qib/qib_mad.c ret = pma_get_portsamplesresult_ext(pmp, ibdev, port); ibdev 2042 drivers/infiniband/hw/qib/qib_mad.c ret = pma_get_portcounters(pmp, ibdev, port); ibdev 2045 drivers/infiniband/hw/qib/qib_mad.c ret = pma_get_portcounters_ext(pmp, ibdev, port); ibdev 2048 drivers/infiniband/hw/qib/qib_mad.c ret = pma_get_portcounters_cong(pmp, ibdev, port); ibdev 2059 drivers/infiniband/hw/qib/qib_mad.c ret = pma_set_portsamplescontrol(pmp, ibdev, port); ibdev 2062 drivers/infiniband/hw/qib/qib_mad.c ret = pma_set_portcounters(pmp, ibdev, port); ibdev 2065 drivers/infiniband/hw/qib/qib_mad.c ret = pma_set_portcounters_ext(pmp, ibdev, port); ibdev 2068 drivers/infiniband/hw/qib/qib_mad.c ret = pma_set_portcounters_cong(pmp, ibdev, port); ibdev 2096 drivers/infiniband/hw/qib/qib_mad.c struct ib_device *ibdev) ibdev 2116 drivers/infiniband/hw/qib/qib_mad.c struct ib_device *ibdev, u8 port) ibdev 2120 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibport *ibp = to_iport(ibdev, port); ibdev 2132 drivers/infiniband/hw/qib/qib_mad.c struct ib_device *ibdev, u8 port) ibdev 2137 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibport *ibp = to_iport(ibdev, port); ibdev 2163 drivers/infiniband/hw/qib/qib_mad.c struct ib_device *ibdev, u8 port) ibdev 2167 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibport *ibp = to_iport(ibdev, port); ibdev 2216 drivers/infiniband/hw/qib/qib_mad.c struct ib_device *ibdev, u8 port) ibdev 2220 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibport *ibp = to_iport(ibdev, port); ibdev 2244 drivers/infiniband/hw/qib/qib_mad.c struct ib_device *ibdev, u8 port) ibdev 2248 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibport *ibp = to_iport(ibdev, port); ibdev 2305 drivers/infiniband/hw/qib/qib_mad.c static int process_cc(struct ib_device *ibdev, int mad_flags, ibdev 2310 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibport *ibp = to_iport(ibdev, port); ibdev 2329 drivers/infiniband/hw/qib/qib_mad.c ret = cc_get_classportinfo(ccp, ibdev); ibdev 2333 drivers/infiniband/hw/qib/qib_mad.c ret = cc_get_congestion_info(ccp, ibdev, port); ibdev 2337 drivers/infiniband/hw/qib/qib_mad.c ret = cc_get_congestion_setting(ccp, ibdev, port); ibdev 2341 drivers/infiniband/hw/qib/qib_mad.c ret = cc_get_congestion_control_table(ccp, ibdev, port); ibdev 2354 drivers/infiniband/hw/qib/qib_mad.c ret = cc_set_congestion_setting(ccp, ibdev, port); ibdev 2358 drivers/infiniband/hw/qib/qib_mad.c ret = cc_set_congestion_control_table(ccp, ibdev, port); ibdev 2406 drivers/infiniband/hw/qib/qib_mad.c int qib_process_mad(struct ib_device *ibdev, int mad_flags, u8 port, ibdev 2413 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibport *ibp = to_iport(ibdev, port); ibdev 2425 drivers/infiniband/hw/qib/qib_mad.c ret = process_subn(ibdev, mad_flags, port, in_mad, out_mad); ibdev 2429 drivers/infiniband/hw/qib/qib_mad.c ret = process_perf(ibdev, port, in_mad, out_mad); ibdev 2438 drivers/infiniband/hw/qib/qib_mad.c ret = process_cc(ibdev, mad_flags, port, in_mad, out_mad); ibdev 2475 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibdev *ibdev = container_of(rdi, struct qib_ibdev, rdi); ibdev 2476 drivers/infiniband/hw/qib/qib_mad.c struct qib_devdata *dd = container_of(ibdev, ibdev 2489 drivers/infiniband/hw/qib/qib_mad.c struct qib_ibdev *ibdev = container_of(rdi, struct qib_ibdev, rdi); ibdev 2490 drivers/infiniband/hw/qib/qib_mad.c struct qib_devdata *dd = container_of(ibdev, ibdev 566 drivers/infiniband/hw/qib/qib_sysfs.c rdma_device_to_drv_device(device, struct qib_ibdev, rdi.ibdev); ibdev 576 drivers/infiniband/hw/qib/qib_sysfs.c rdma_device_to_drv_device(device, struct qib_ibdev, rdi.ibdev); ibdev 601 drivers/infiniband/hw/qib/qib_sysfs.c rdma_device_to_drv_device(device, struct qib_ibdev, rdi.ibdev); ibdev 613 drivers/infiniband/hw/qib/qib_sysfs.c rdma_device_to_drv_device(device, struct qib_ibdev, rdi.ibdev); ibdev 625 drivers/infiniband/hw/qib/qib_sysfs.c rdma_device_to_drv_device(device, struct qib_ibdev, rdi.ibdev); ibdev 641 drivers/infiniband/hw/qib/qib_sysfs.c rdma_device_to_drv_device(device, struct qib_ibdev, rdi.ibdev); ibdev 653 drivers/infiniband/hw/qib/qib_sysfs.c rdma_device_to_drv_device(device, struct qib_ibdev, rdi.ibdev); ibdev 668 drivers/infiniband/hw/qib/qib_sysfs.c rdma_device_to_drv_device(device, struct qib_ibdev, rdi.ibdev); ibdev 690 drivers/infiniband/hw/qib/qib_sysfs.c rdma_device_to_drv_device(device, struct qib_ibdev, rdi.ibdev); ibdev 741 drivers/infiniband/hw/qib/qib_sysfs.c int qib_create_port_files(struct ib_device *ibdev, u8 port_num, ibdev 745 drivers/infiniband/hw/qib/qib_sysfs.c struct qib_devdata *dd = dd_from_ibdev(ibdev); ibdev 1195 drivers/infiniband/hw/qib/qib_verbs.c struct qib_ibdev *ibdev = container_of(rdi, struct qib_ibdev, rdi); ibdev 1196 drivers/infiniband/hw/qib/qib_verbs.c struct qib_devdata *dd = dd_from_dev(ibdev); ibdev 1279 drivers/infiniband/hw/qib/qib_verbs.c struct qib_ibdev *ibdev = container_of(rdi, struct qib_ibdev, rdi); ibdev 1280 drivers/infiniband/hw/qib/qib_verbs.c struct qib_devdata *dd = dd_from_dev(ibdev); ibdev 1304 drivers/infiniband/hw/qib/qib_verbs.c int qib_check_ah(struct ib_device *ibdev, struct rdma_ah_attr *ah_attr) ibdev 1321 drivers/infiniband/hw/qib/qib_verbs.c static void qib_notify_new_ah(struct ib_device *ibdev, ibdev 1333 drivers/infiniband/hw/qib/qib_verbs.c ibp = to_iport(ibdev, rdma_ah_get_port_num(ah_attr)); ibdev 1349 drivers/infiniband/hw/qib/qib_verbs.c attr.type = rdma_ah_find_type(&dd->verbs_dev.rdi.ibdev, port_num); ibdev 1501 drivers/infiniband/hw/qib/qib_verbs.c struct ib_device *ibdev = &dev->rdi.ibdev; ibdev 1551 drivers/infiniband/hw/qib/qib_verbs.c ibdev->node_guid = ppd->guid; ibdev 1552 drivers/infiniband/hw/qib/qib_verbs.c ibdev->phys_port_cnt = dd->num_pports; ibdev 1553 drivers/infiniband/hw/qib/qib_verbs.c ibdev->dev.parent = &dd->pcidev->dev; ibdev 1555 drivers/infiniband/hw/qib/qib_verbs.c snprintf(ibdev->node_desc, sizeof(ibdev->node_desc), ibdev 1618 drivers/infiniband/hw/qib/qib_verbs.c rdma_set_device_sysfs_group(&dd->verbs_dev.rdi.ibdev, &qib_attr_group); ibdev 1620 drivers/infiniband/hw/qib/qib_verbs.c ib_set_device_ops(ibdev, &qib_dev_ops); ibdev 207 drivers/infiniband/hw/qib/qib_verbs.h static inline struct qib_ibdev *to_idev(struct ib_device *ibdev) ibdev 211 drivers/infiniband/hw/qib/qib_verbs.h rdi = container_of(ibdev, struct rvt_dev_info, ibdev); ibdev 246 drivers/infiniband/hw/qib/qib_verbs.h int qib_process_mad(struct ib_device *ibdev, int mad_flags, u8 port_num, ibdev 301 drivers/infiniband/hw/qib/qib_verbs.h int qib_check_ah(struct ib_device *ibdev, struct rdma_ah_attr *ah_attr); ibdev 101 drivers/infiniband/hw/usnic/usnic_ib.h struct usnic_ib_dev *to_usdev(struct ib_device *ibdev) ibdev 103 drivers/infiniband/hw/usnic/usnic_ib.h return container_of(ibdev, struct usnic_ib_dev, ib_dev); ibdev 225 drivers/infiniband/hw/usnic/usnic_ib_main.c struct ib_device *ibdev; ibdev 229 drivers/infiniband/hw/usnic/usnic_ib_main.c ibdev = ib_device_get_by_netdev(netdev, RDMA_DRIVER_USNIC); ibdev 230 drivers/infiniband/hw/usnic/usnic_ib_main.c if (!ibdev) ibdev 233 drivers/infiniband/hw/usnic/usnic_ib_main.c us_ibdev = container_of(ibdev, struct usnic_ib_dev, ib_dev); ibdev 235 drivers/infiniband/hw/usnic/usnic_ib_main.c ib_device_put(ibdev); ibdev 290 drivers/infiniband/hw/usnic/usnic_ib_main.c struct ib_device *ibdev; ibdev 292 drivers/infiniband/hw/usnic/usnic_ib_main.c ibdev = ib_device_get_by_netdev(netdev, RDMA_DRIVER_USNIC); ibdev 293 drivers/infiniband/hw/usnic/usnic_ib_main.c if (!ibdev) ibdev 296 drivers/infiniband/hw/usnic/usnic_ib_main.c us_ibdev = container_of(ibdev, struct usnic_ib_dev, ib_dev); ibdev 298 drivers/infiniband/hw/usnic/usnic_ib_main.c ib_device_put(ibdev); ibdev 306 drivers/infiniband/hw/usnic/usnic_ib_main.c static int usnic_port_immutable(struct ib_device *ibdev, u8 port_num, ibdev 314 drivers/infiniband/hw/usnic/usnic_ib_main.c err = ib_query_port(ibdev, port_num, &attr); ibdev 275 drivers/infiniband/hw/usnic/usnic_ib_verbs.c int usnic_ib_query_device(struct ib_device *ibdev, ibdev 279 drivers/infiniband/hw/usnic/usnic_ib_verbs.c struct usnic_ib_dev *us_ibdev = to_usdev(ibdev); ibdev 333 drivers/infiniband/hw/usnic/usnic_ib_verbs.c int usnic_ib_query_port(struct ib_device *ibdev, u8 port, ibdev 336 drivers/infiniband/hw/usnic/usnic_ib_verbs.c struct usnic_ib_dev *us_ibdev = to_usdev(ibdev); ibdev 340 drivers/infiniband/hw/usnic/usnic_ib_verbs.c if (ib_get_eth_speed(ibdev, port, &props->active_speed, ibdev 422 drivers/infiniband/hw/usnic/usnic_ib_verbs.c int usnic_ib_query_gid(struct ib_device *ibdev, u8 port, int index, ibdev 426 drivers/infiniband/hw/usnic/usnic_ib_verbs.c struct usnic_ib_dev *us_ibdev = to_usdev(ibdev); ibdev 441 drivers/infiniband/hw/usnic/usnic_ib_verbs.c int usnic_ib_query_pkey(struct ib_device *ibdev, u8 port, u16 index, ibdev 647 drivers/infiniband/hw/usnic/usnic_ib_verbs.c struct ib_device *ibdev = uctx->device; ibdev 649 drivers/infiniband/hw/usnic/usnic_ib_verbs.c struct usnic_ib_dev *us_ibdev = to_usdev(ibdev); ibdev 41 drivers/infiniband/hw/usnic/usnic_ib_verbs.h int usnic_ib_query_device(struct ib_device *ibdev, ibdev 44 drivers/infiniband/hw/usnic/usnic_ib_verbs.h int usnic_ib_query_port(struct ib_device *ibdev, u8 port, ibdev 49 drivers/infiniband/hw/usnic/usnic_ib_verbs.h int usnic_ib_query_gid(struct ib_device *ibdev, u8 port, int index, ibdev 51 drivers/infiniband/hw/usnic/usnic_ib_verbs.h int usnic_ib_query_pkey(struct ib_device *ibdev, u8 port, u16 index, ibdev 258 drivers/infiniband/hw/vmw_pvrdma/pvrdma.h static inline struct pvrdma_dev *to_vdev(struct ib_device *ibdev) ibdev 260 drivers/infiniband/hw/vmw_pvrdma/pvrdma.h return container_of(ibdev, struct pvrdma_dev, ib_dev); ibdev 104 drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c struct ib_device *ibdev = ibcq->device; ibdev 106 drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c struct pvrdma_dev *dev = to_vdev(ibdev); ibdev 124 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c static int pvrdma_port_immutable(struct ib_device *ibdev, u8 port_num, ibdev 127 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c struct pvrdma_dev *dev = to_vdev(ibdev); ibdev 136 drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c err = ib_query_port(ibdev, port_num, &attr); ibdev 65 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c int pvrdma_query_device(struct ib_device *ibdev, ibdev 69 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c struct pvrdma_dev *dev = to_vdev(ibdev); ibdev 128 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c int pvrdma_query_port(struct ib_device *ibdev, u8 port, ibdev 131 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c struct pvrdma_dev *dev = to_vdev(ibdev); ibdev 186 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c int pvrdma_query_gid(struct ib_device *ibdev, u8 port, int index, ibdev 189 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c struct pvrdma_dev *dev = to_vdev(ibdev); ibdev 208 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c int pvrdma_query_pkey(struct ib_device *ibdev, u8 port, u16 index, ibdev 221 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c err = pvrdma_cmd_post(to_vdev(ibdev), &req, &rsp, ibdev 224 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c dev_warn(&to_vdev(ibdev)->pdev->dev, ibdev 234 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c enum rdma_link_layer pvrdma_port_link_layer(struct ib_device *ibdev, ibdev 240 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c int pvrdma_modify_device(struct ib_device *ibdev, int mask, ibdev 247 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c dev_warn(&to_vdev(ibdev)->pdev->dev, ibdev 253 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c spin_lock_irqsave(&to_vdev(ibdev)->desc_lock, flags); ibdev 254 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c memcpy(ibdev->node_desc, props->node_desc, 64); ibdev 255 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c spin_unlock_irqrestore(&to_vdev(ibdev)->desc_lock, flags); ibdev 259 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c mutex_lock(&to_vdev(ibdev)->port_mutex); ibdev 260 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c to_vdev(ibdev)->sys_image_guid = ibdev 262 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c mutex_unlock(&to_vdev(ibdev)->port_mutex); ibdev 277 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c int pvrdma_modify_port(struct ib_device *ibdev, u8 port, int mask, ibdev 281 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c struct pvrdma_dev *vdev = to_vdev(ibdev); ibdev 291 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c ret = ib_query_port(ibdev, port, &attr); ibdev 315 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c struct ib_device *ibdev = uctx->device; ibdev 316 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c struct pvrdma_dev *vdev = to_vdev(ibdev); ibdev 429 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c struct ib_device *ibdev = ibpd->device; ibdev 431 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c struct pvrdma_dev *dev = to_vdev(ibdev); ibdev 383 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h int pvrdma_query_device(struct ib_device *ibdev, ibdev 386 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h int pvrdma_query_port(struct ib_device *ibdev, u8 port, ibdev 388 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h int pvrdma_query_gid(struct ib_device *ibdev, u8 port, ibdev 390 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h int pvrdma_query_pkey(struct ib_device *ibdev, u8 port, ibdev 392 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h enum rdma_link_layer pvrdma_port_link_layer(struct ib_device *ibdev, ibdev 394 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h int pvrdma_modify_device(struct ib_device *ibdev, int mask, ibdev 396 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h int pvrdma_modify_port(struct ib_device *ibdev, u8 port, ibdev 62 drivers/infiniband/sw/rdmavt/ah.c int rvt_check_ah(struct ib_device *ibdev, ibdev 68 drivers/infiniband/sw/rdmavt/ah.c struct rvt_dev_info *rdi = ib_to_rvt(ibdev); ibdev 72 drivers/infiniband/sw/rdmavt/ah.c err = ib_query_port(ibdev, port_num, &port_attr); ibdev 76 drivers/infiniband/sw/rdmavt/ah.c port_num > ibdev->phys_port_cnt) ibdev 85 drivers/infiniband/sw/rdmavt/ah.c return rdi->driver_f.check_ah(ibdev, ah_attr); ibdev 203 drivers/infiniband/sw/rdmavt/cq.c struct ib_device *ibdev = ibcq->device; ibdev 204 drivers/infiniband/sw/rdmavt/cq.c struct rvt_dev_info *rdi = ib_to_rvt(ibdev); ibdev 222 drivers/infiniband/sw/rdmavt/cq.c comp_vector = comp_vector % rdi->ibdev.num_comp_vectors; ibdev 70 drivers/infiniband/sw/rdmavt/mad.c int rvt_process_mad(struct ib_device *ibdev, int mad_flags, u8 port_num, ibdev 82 drivers/infiniband/sw/rdmavt/mad.c if (ibport_num_to_idx(ibdev, port_num) < 0) ibdev 111 drivers/infiniband/sw/rdmavt/mad.c agent = ib_register_mad_agent(&rdi->ibdev, p + 1, ibdev 53 drivers/infiniband/sw/rdmavt/mad.h int rvt_process_mad(struct ib_device *ibdev, int mad_flags, u8 port_num, ibdev 62 drivers/infiniband/sw/rdmavt/pd.c struct ib_device *ibdev = ibpd->device; ibdev 63 drivers/infiniband/sw/rdmavt/pd.c struct rvt_dev_info *dev = ib_to_rvt(ibdev); ibdev 2723 drivers/infiniband/sw/rdmavt/qp.c i->specials = rdi->ibdev.phys_port_cnt * 2; ibdev 2774 drivers/infiniband/sw/rdmavt/qp.c pidx = n % rdi->ibdev.phys_port_cnt; ibdev 2817 drivers/infiniband/sw/rdmavt/qp.c .specials = rdi->ibdev.phys_port_cnt * 2, ibdev 94 drivers/infiniband/sw/rdmavt/vt.c rdi = container_of(_ib_alloc_device(size), struct rvt_dev_info, ibdev); ibdev 102 drivers/infiniband/sw/rdmavt/vt.c ib_dealloc_device(&rdi->ibdev); ibdev 117 drivers/infiniband/sw/rdmavt/vt.c ib_dealloc_device(&rdi->ibdev); ibdev 121 drivers/infiniband/sw/rdmavt/vt.c static int rvt_query_device(struct ib_device *ibdev, ibdev 125 drivers/infiniband/sw/rdmavt/vt.c struct rvt_dev_info *rdi = ib_to_rvt(ibdev); ibdev 156 drivers/infiniband/sw/rdmavt/vt.c static int rvt_query_port(struct ib_device *ibdev, u8 port_num, ibdev 159 drivers/infiniband/sw/rdmavt/vt.c struct rvt_dev_info *rdi = ib_to_rvt(ibdev); ibdev 161 drivers/infiniband/sw/rdmavt/vt.c int port_index = ibport_num_to_idx(ibdev, port_num); ibdev 191 drivers/infiniband/sw/rdmavt/vt.c static int rvt_modify_port(struct ib_device *ibdev, u8 port_num, ibdev 194 drivers/infiniband/sw/rdmavt/vt.c struct rvt_dev_info *rdi = ib_to_rvt(ibdev); ibdev 197 drivers/infiniband/sw/rdmavt/vt.c int port_index = ibport_num_to_idx(ibdev, port_num); ibdev 230 drivers/infiniband/sw/rdmavt/vt.c static int rvt_query_pkey(struct ib_device *ibdev, u8 port_num, u16 index, ibdev 239 drivers/infiniband/sw/rdmavt/vt.c struct rvt_dev_info *rdi = ib_to_rvt(ibdev); ibdev 242 drivers/infiniband/sw/rdmavt/vt.c port_index = ibport_num_to_idx(ibdev, port_num); ibdev 262 drivers/infiniband/sw/rdmavt/vt.c static int rvt_query_gid(struct ib_device *ibdev, u8 port_num, ibdev 274 drivers/infiniband/sw/rdmavt/vt.c port_index = ibport_num_to_idx(ibdev, port_num); ibdev 278 drivers/infiniband/sw/rdmavt/vt.c rdi = ib_to_rvt(ibdev); ibdev 312 drivers/infiniband/sw/rdmavt/vt.c static int rvt_get_port_immutable(struct ib_device *ibdev, u8 port_num, ibdev 315 drivers/infiniband/sw/rdmavt/vt.c struct rvt_dev_info *rdi = ib_to_rvt(ibdev); ibdev 319 drivers/infiniband/sw/rdmavt/vt.c port_index = ibport_num_to_idx(ibdev, port_num); ibdev 325 drivers/infiniband/sw/rdmavt/vt.c err = ib_query_port(ibdev, port_num, &attr); ibdev 446 drivers/infiniband/sw/rdmavt/vt.c if ((!rdi->ibdev.ops.init_port) || ibdev 456 drivers/infiniband/sw/rdmavt/vt.c if (!rdi->ibdev.ops.modify_device) ibdev 461 drivers/infiniband/sw/rdmavt/vt.c if (!rdi->ibdev.ops.query_port) ibdev 467 drivers/infiniband/sw/rdmavt/vt.c if (!rdi->ibdev.ops.modify_port) ibdev 474 drivers/infiniband/sw/rdmavt/vt.c if (!rdi->ibdev.ops.query_gid) ibdev 480 drivers/infiniband/sw/rdmavt/vt.c if (!rdi->ibdev.ops.create_qp) ibdev 491 drivers/infiniband/sw/rdmavt/vt.c if (!rdi->ibdev.ops.modify_qp) ibdev 505 drivers/infiniband/sw/rdmavt/vt.c if (!rdi->ibdev.ops.destroy_qp) ibdev 515 drivers/infiniband/sw/rdmavt/vt.c if (!rdi->ibdev.ops.post_send) ibdev 553 drivers/infiniband/sw/rdmavt/vt.c ib_set_device_ops(&rdi->ibdev, &rvt_dev_ops); ibdev 594 drivers/infiniband/sw/rdmavt/vt.c rdi->ibdev.dev.dma_ops = rdi->ibdev.dev.dma_ops ? : &dma_virt_ops; ibdev 606 drivers/infiniband/sw/rdmavt/vt.c rdi->ibdev.uverbs_cmd_mask = ibdev 637 drivers/infiniband/sw/rdmavt/vt.c rdi->ibdev.node_type = RDMA_NODE_IB_CA; ibdev 638 drivers/infiniband/sw/rdmavt/vt.c if (!rdi->ibdev.num_comp_vectors) ibdev 639 drivers/infiniband/sw/rdmavt/vt.c rdi->ibdev.num_comp_vectors = 1; ibdev 642 drivers/infiniband/sw/rdmavt/vt.c ret = ib_register_device(&rdi->ibdev, dev_name(&rdi->ibdev.dev)); ibdev 677 drivers/infiniband/sw/rdmavt/vt.c ib_unregister_device(&rdi->ibdev); ibdev 99 drivers/infiniband/sw/rdmavt/vt.h static inline int ibport_num_to_idx(struct ib_device *ibdev, u8 port_num) ibdev 101 drivers/infiniband/sw/rdmavt/vt.h struct rvt_dev_info *rdi = ib_to_rvt(ibdev); ibdev 100 drivers/infiniband/sw/rxe/rxe.h struct ib_device *ibdev = ibdev 103 drivers/infiniband/sw/rxe/rxe.h if (!ibdev) ibdev 105 drivers/infiniband/sw/rxe/rxe.h return container_of(ibdev, struct rxe_dev, ib_dev); ibdev 54 drivers/infiniband/sw/rxe/rxe_hw_counters.c int rxe_ib_get_hw_stats(struct ib_device *ibdev, ibdev 58 drivers/infiniband/sw/rxe/rxe_hw_counters.c struct rxe_dev *dev = to_rdev(ibdev); ibdev 70 drivers/infiniband/sw/rxe/rxe_hw_counters.c struct rdma_hw_stats *rxe_ib_alloc_hw_stats(struct ib_device *ibdev, ibdev 59 drivers/infiniband/sw/rxe/rxe_hw_counters.h struct rdma_hw_stats *rxe_ib_alloc_hw_stats(struct ib_device *ibdev, ibdev 61 drivers/infiniband/sw/rxe/rxe_hw_counters.h int rxe_ib_get_hw_stats(struct ib_device *ibdev, ibdev 717 drivers/infiniband/sw/siw/siw.h #define siw_dbg(ibdev, fmt, ...) \ ibdev 718 drivers/infiniband/sw/siw/siw.h ibdev_dbg(ibdev, "%s: " fmt, __func__, ##__VA_ARGS__) ibdev 140 drivers/infiniband/ulp/iser/iser_memory.c struct ib_device *ibdev) ibdev 171 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h struct ib_device *ibdev; ibdev 198 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h struct ib_device *ibdev; ibdev 255 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h dev_err(&cport->ibdev->dev, format, ## arg) ibdev 257 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h dev_info(&cport->ibdev->dev, format, ## arg) ibdev 259 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h dev_dbg(&cport->ibdev->dev, format, ## arg) ibdev 297 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h struct opa_vnic_adapter *opa_vnic_add_netdev(struct ib_device *ibdev, ibdev 323 drivers/infiniband/ulp/opa_vnic/opa_vnic_netdev.c struct opa_vnic_adapter *opa_vnic_add_netdev(struct ib_device *ibdev, ibdev 331 drivers/infiniband/ulp/opa_vnic/opa_vnic_netdev.c netdev = ibdev->ops.alloc_rdma_netdev(ibdev, port_num, ibdev 348 drivers/infiniband/ulp/opa_vnic/opa_vnic_netdev.c rn->hca = ibdev; ibdev 351 drivers/infiniband/ulp/opa_vnic/opa_vnic_netdev.c adapter->ibdev = ibdev; ibdev 363 drivers/infiniband/ulp/opa_vnic/opa_vnic_netdev.c SET_NETDEV_DEV(netdev, ibdev->dev.parent); ibdev 206 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema.c adapter = opa_vnic_add_netdev(cport->ibdev, port->port_num, vport_num); ibdev 744 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema.c ibp = cport->ibdev; ibdev 945 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema.c cport->ibdev, opa_vnic_event); ibdev 950 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema.c port->mad_agent = ib_register_mad_agent(cport->ibdev, i, ibdev 985 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema.c ib_modify_port(cport->ibdev, i, IB_PORT_OPA_MASK_CHG, &pm); ibdev 1008 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema.c cport->ibdev = device; ibdev 1008 drivers/infiniband/ulp/srp/ib_srp.c struct ib_device *ibdev = dev->dev; ibdev 1024 drivers/infiniband/ulp/srp/ib_srp.c ib_dma_unmap_single(ibdev, req->indirect_dma_addr, ibdev 1039 drivers/infiniband/ulp/srp/ib_srp.c struct ib_device *ibdev = srp_dev->dev; ibdev 1070 drivers/infiniband/ulp/srp/ib_srp.c dma_addr = ib_dma_map_single(ibdev, req->indirect_desc, ibdev 1073 drivers/infiniband/ulp/srp/ib_srp.c if (ib_dma_mapping_error(ibdev, dma_addr)) ibdev 1248 drivers/infiniband/ulp/srp/ib_srp.c struct ib_device *ibdev = dev->dev; ibdev 1279 drivers/infiniband/ulp/srp/ib_srp.c ib_dma_unmap_sg(ibdev, scsi_sglist(scmnd), scsi_sg_count(scmnd), ibdev 1808 drivers/infiniband/ulp/srp/ib_srp.c struct ib_device *ibdev; ibdev 1834 drivers/infiniband/ulp/srp/ib_srp.c ibdev = dev->dev; ibdev 1836 drivers/infiniband/ulp/srp/ib_srp.c count = ib_dma_map_sg(ibdev, scat, nents, scmnd->sc_data_direction); ibdev 1890 drivers/infiniband/ulp/srp/ib_srp.c ib_dma_sync_single_for_cpu(ibdev, req->indirect_dma_addr, ibdev 1970 drivers/infiniband/ulp/srp/ib_srp.c ib_dma_sync_single_for_device(ibdev, req->indirect_dma_addr, table_len, ibdev 3777 drivers/infiniband/ulp/srp/ib_srp.c struct ib_device *ibdev = srp_dev->dev; ibdev 3793 drivers/infiniband/ulp/srp/ib_srp.c target_host->max_segment_size = ib_dma_max_seg_size(ibdev); ibdev 3795 drivers/infiniband/ulp/srp/ib_srp.c if (!(ibdev->attrs.device_cap_flags & IB_DEVICE_SG_GAPS_REG)) ibdev 3853 drivers/infiniband/ulp/srp/ib_srp.c bool gaps_reg = (ibdev->attrs.device_cap_flags & ibdev 3896 drivers/infiniband/ulp/srp/ib_srp.c ret = rdma_query_gid(ibdev, host->port, 0, &target->sgid); ibdev 3904 drivers/infiniband/ulp/srp/ib_srp.c ibdev->num_comp_vectors), ibdev 3917 drivers/infiniband/ulp/srp/ib_srp.c const int cv_start = node_idx * ibdev->num_comp_vectors / ibdev 3919 drivers/infiniband/ulp/srp/ib_srp.c const int cv_end = (node_idx + 1) * ibdev->num_comp_vectors / ibdev 4046 drivers/infiniband/ulp/srp/ib_srp.c static DEVICE_ATTR(ibdev, S_IRUGO, show_ibdev, NULL); ibdev 684 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib.c static int mlx5_rdma_setup_rn(struct ib_device *ibdev, u8 port_num, ibdev 720 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib.c rn->hca = ibdev; ibdev 170 drivers/nvme/host/rdma.c static void nvme_rdma_free_qe(struct ib_device *ibdev, struct nvme_rdma_qe *qe, ibdev 173 drivers/nvme/host/rdma.c ib_dma_unmap_single(ibdev, qe->dma, capsule_size, dir); ibdev 177 drivers/nvme/host/rdma.c static int nvme_rdma_alloc_qe(struct ib_device *ibdev, struct nvme_rdma_qe *qe, ibdev 184 drivers/nvme/host/rdma.c qe->dma = ib_dma_map_single(ibdev, qe->data, capsule_size, dir); ibdev 185 drivers/nvme/host/rdma.c if (ib_dma_mapping_error(ibdev, qe->dma)) { ibdev 194 drivers/nvme/host/rdma.c static void nvme_rdma_free_ring(struct ib_device *ibdev, ibdev 201 drivers/nvme/host/rdma.c nvme_rdma_free_qe(ibdev, &ring[i], capsule_size, dir); ibdev 205 drivers/nvme/host/rdma.c static struct nvme_rdma_qe *nvme_rdma_alloc_ring(struct ib_device *ibdev, ibdev 222 drivers/nvme/host/rdma.c if (nvme_rdma_alloc_qe(ibdev, &ring[i], capsule_size, dir)) ibdev 229 drivers/nvme/host/rdma.c nvme_rdma_free_ring(ibdev, ring, i, capsule_size, dir); ibdev 403 drivers/nvme/host/rdma.c struct ib_device *ibdev; ibdev 409 drivers/nvme/host/rdma.c ibdev = dev->dev; ibdev 421 drivers/nvme/host/rdma.c nvme_rdma_free_ring(ibdev, queue->rsp_ring, queue->queue_size, ibdev 427 drivers/nvme/host/rdma.c static int nvme_rdma_get_max_fr_pages(struct ib_device *ibdev) ibdev 430 drivers/nvme/host/rdma.c ibdev->attrs.max_fast_reg_page_list_len - 1); ibdev 435 drivers/nvme/host/rdma.c struct ib_device *ibdev; ibdev 448 drivers/nvme/host/rdma.c ibdev = queue->device->dev; ibdev 463 drivers/nvme/host/rdma.c queue->ib_cq = ib_alloc_cq(ibdev, queue, ibdev 475 drivers/nvme/host/rdma.c queue->rsp_ring = nvme_rdma_alloc_ring(ibdev, queue->queue_size, ibdev 487 drivers/nvme/host/rdma.c pages_per_mr = nvme_rdma_get_max_fr_pages(ibdev) + 1; ibdev 504 drivers/nvme/host/rdma.c nvme_rdma_free_ring(ibdev, queue->rsp_ring, queue->queue_size, ibdev 652 drivers/nvme/host/rdma.c struct ib_device *ibdev = ctrl->device->dev; ibdev 657 drivers/nvme/host/rdma.c nr_read_queues = min_t(unsigned int, ibdev->num_comp_vectors, ibdev 659 drivers/nvme/host/rdma.c nr_default_queues = min_t(unsigned int, ibdev->num_comp_vectors, ibdev 1154 drivers/nvme/host/rdma.c struct ib_device *ibdev = dev->dev; ibdev 1164 drivers/nvme/host/rdma.c ib_dma_unmap_sg(ibdev, req->sg_table.sgl, req->nents, rq_dma_dir(rq)); ibdev 1269 drivers/nvme/host/rdma.c struct ib_device *ibdev = dev->dev; ibdev 1289 drivers/nvme/host/rdma.c count = ib_dma_map_sg(ibdev, req->sg_table.sgl, req->nents, ibdev 1319 drivers/nvme/host/rdma.c ib_dma_unmap_sg(ibdev, req->sg_table.sgl, req->nents, rq_dma_dir(rq)); ibdev 1490 drivers/nvme/host/rdma.c struct ib_device *ibdev = queue->device->dev; ibdev 1499 drivers/nvme/host/rdma.c ib_dma_sync_single_for_cpu(ibdev, qe->dma, len, DMA_FROM_DEVICE); ibdev 1512 drivers/nvme/host/rdma.c ib_dma_sync_single_for_device(ibdev, qe->dma, len, DMA_FROM_DEVICE); ibdev 1813 drivers/nvme/host/rdma.c struct ib_device *ibdev = queue->device->dev; ibdev 1816 drivers/nvme/host/rdma.c ib_dma_unmap_single(ibdev, req->sqe.dma, sizeof(struct nvme_command), ibdev 78 include/linux/dynamic_debug.h const struct ib_device *ibdev, ibdev 1099 include/linux/mlx5/driver.h struct ib_device *ibdev, ibdev 788 include/net/devlink.h struct ib_device *ibdev); ibdev 45 include/rdma/ib_umem.h struct ib_device *ibdev; ibdev 81 include/rdma/ib_verbs.h void ibdev_printk(const char *level, const struct ib_device *ibdev, ibdev 84 include/rdma/ib_verbs.h void ibdev_emerg(const struct ib_device *ibdev, const char *format, ...); ibdev 86 include/rdma/ib_verbs.h void ibdev_alert(const struct ib_device *ibdev, const char *format, ...); ibdev 88 include/rdma/ib_verbs.h void ibdev_crit(const struct ib_device *ibdev, const char *format, ...); ibdev 90 include/rdma/ib_verbs.h void ibdev_err(const struct ib_device *ibdev, const char *format, ...); ibdev 92 include/rdma/ib_verbs.h void ibdev_warn(const struct ib_device *ibdev, const char *format, ...); ibdev 94 include/rdma/ib_verbs.h void ibdev_notice(const struct ib_device *ibdev, const char *format, ...); ibdev 96 include/rdma/ib_verbs.h void ibdev_info(const struct ib_device *ibdev, const char *format, ...); ibdev 104 include/rdma/ib_verbs.h void ibdev_dbg(const struct ib_device *ibdev, const char *format, ...) {} ibdev 107 include/rdma/ib_verbs.h #define ibdev_level_ratelimited(ibdev_level, ibdev, fmt, ...) \ ibdev 113 include/rdma/ib_verbs.h ibdev_level(ibdev, fmt, ##__VA_ARGS__); \ ibdev 116 include/rdma/ib_verbs.h #define ibdev_emerg_ratelimited(ibdev, fmt, ...) \ ibdev 117 include/rdma/ib_verbs.h ibdev_level_ratelimited(ibdev_emerg, ibdev, fmt, ##__VA_ARGS__) ibdev 118 include/rdma/ib_verbs.h #define ibdev_alert_ratelimited(ibdev, fmt, ...) \ ibdev 119 include/rdma/ib_verbs.h ibdev_level_ratelimited(ibdev_alert, ibdev, fmt, ##__VA_ARGS__) ibdev 120 include/rdma/ib_verbs.h #define ibdev_crit_ratelimited(ibdev, fmt, ...) \ ibdev 121 include/rdma/ib_verbs.h ibdev_level_ratelimited(ibdev_crit, ibdev, fmt, ##__VA_ARGS__) ibdev 122 include/rdma/ib_verbs.h #define ibdev_err_ratelimited(ibdev, fmt, ...) \ ibdev 123 include/rdma/ib_verbs.h ibdev_level_ratelimited(ibdev_err, ibdev, fmt, ##__VA_ARGS__) ibdev 124 include/rdma/ib_verbs.h #define ibdev_warn_ratelimited(ibdev, fmt, ...) \ ibdev 125 include/rdma/ib_verbs.h ibdev_level_ratelimited(ibdev_warn, ibdev, fmt, ##__VA_ARGS__) ibdev 126 include/rdma/ib_verbs.h #define ibdev_notice_ratelimited(ibdev, fmt, ...) \ ibdev 127 include/rdma/ib_verbs.h ibdev_level_ratelimited(ibdev_notice, ibdev, fmt, ##__VA_ARGS__) ibdev 128 include/rdma/ib_verbs.h #define ibdev_info_ratelimited(ibdev, fmt, ...) \ ibdev 129 include/rdma/ib_verbs.h ibdev_level_ratelimited(ibdev_info, ibdev, fmt, ##__VA_ARGS__) ibdev 133 include/rdma/ib_verbs.h #define ibdev_dbg_ratelimited(ibdev, fmt, ...) \ ibdev 140 include/rdma/ib_verbs.h __dynamic_ibdev_dbg(&descriptor, ibdev, fmt, \ ibdev 146 include/rdma/ib_verbs.h void ibdev_dbg_ratelimited(const struct ib_device *ibdev, const char *format, ...) {} ibdev 2289 include/rdma/ib_verbs.h const struct cpumask *(*get_vector_affinity)(struct ib_device *ibdev, ibdev 2673 include/rdma/ib_verbs.h int (*get_nl_info)(struct ib_device *ibdev, void *client_data, ibdev 4596 include/rdma/ib_verbs.h void rdma_roce_rescan_device(struct ib_device *ibdev); ibdev 407 include/rdma/rdma_vt.h struct ib_device ibdev; /* Keep this first. Nothing above here */ ibdev 489 include/rdma/rdma_vt.h dev_set_name(&rdi->ibdev.dev, fmt, name, unit); ibdev 490 include/rdma/rdma_vt.h strlcpy(rdi->ibdev.name, dev_name(&rdi->ibdev.dev), IB_DEVICE_NAME_MAX); ibdev 501 include/rdma/rdma_vt.h return dev_name(&rdi->ibdev.dev); ibdev 514 include/rdma/rdma_vt.h static inline struct rvt_dev_info *ib_to_rvt(struct ib_device *ibdev) ibdev 516 include/rdma/rdma_vt.h return container_of(ibdev, struct rvt_dev_info, ibdev); ibdev 560 include/rdma/rdma_vt.h int rvt_check_ah(struct ib_device *ibdev, struct rdma_ah_attr *ah_attr); ibdev 644 lib/dynamic_debug.c const struct ib_device *ibdev, const char *fmt, ...) ibdev 654 lib/dynamic_debug.c if (ibdev && ibdev->dev.parent) { ibdev 657 lib/dynamic_debug.c dev_printk_emit(LOGLEVEL_DEBUG, ibdev->dev.parent, ibdev 660 lib/dynamic_debug.c dev_driver_string(ibdev->dev.parent), ibdev 661 lib/dynamic_debug.c dev_name(ibdev->dev.parent), ibdev 662 lib/dynamic_debug.c dev_name(&ibdev->dev), ibdev 664 lib/dynamic_debug.c } else if (ibdev) { ibdev 665 lib/dynamic_debug.c printk(KERN_DEBUG "%s: %pV", dev_name(&ibdev->dev), &vaf); ibdev 583 net/core/devlink.c struct ib_device *ibdev = devlink_port->type_dev; ibdev 585 net/core/devlink.c if (ibdev && ibdev 587 net/core/devlink.c ibdev->name)) ibdev 6442 net/core/devlink.c struct ib_device *ibdev) ibdev 6444 net/core/devlink.c __devlink_port_type_set(devlink_port, DEVLINK_PORT_TYPE_IB, ibdev); ibdev 267 net/rds/ib.h #define ibdev_to_node(ibdev) dev_to_node((ibdev)->dev.parent) ibdev 239 net/smc/smc_core.c get_device(&ini->ib_dev->ibdev->dev); ibdev 443 net/smc/smc_core.c put_device(&lgr->lnk[SMC_SINGLE_LINK].smcibdev->ibdev->dev); ibdev 159 net/smc/smc_diag.c smc->conn.lgr->lnk[0].smcibdev->ibdev->name, ibdev 160 net/smc/smc_diag.c sizeof(smc->conn.lgr->lnk[0].smcibdev->ibdev->name)); ibdev 151 net/smc/smc_ib.c attr = rdma_get_gid_attr(smcibdev->ibdev, ibport, 0); ibdev 187 net/smc/smc_ib.c attr = rdma_get_gid_attr(smcibdev->ibdev, ibport, i); ibdev 218 net/smc/smc_ib.c rc = ib_query_port(smcibdev->ibdev, ibport, ibdev 291 net/smc/smc_ib.c lnk->roce_pd = ib_alloc_pd(lnk->smcibdev->ibdev, 0); ibdev 412 net/smc/smc_ib.c ib_dma_sync_single_for_cpu(smcibdev->ibdev, ibdev 432 net/smc/smc_ib.c ib_dma_sync_single_for_device(smcibdev->ibdev, ibdev 446 net/smc/smc_ib.c mapped_nents = ib_dma_map_sg(smcibdev->ibdev, ibdev 463 net/smc/smc_ib.c ib_dma_unmap_sg(smcibdev->ibdev, ibdev 482 net/smc/smc_ib.c smcibdev->roce_cq_send = ib_create_cq(smcibdev->ibdev, ibdev 490 net/smc/smc_ib.c smcibdev->roce_cq_recv = ib_create_cq(smcibdev->ibdev, ibdev 520 net/smc/smc_ib.c static void smc_ib_add_dev(struct ib_device *ibdev) ibdev 526 net/smc/smc_ib.c if (ibdev->node_type != RDMA_NODE_IB_CA) ibdev 533 net/smc/smc_ib.c smcibdev->ibdev = ibdev; ibdev 539 net/smc/smc_ib.c ib_set_client_data(ibdev, &smc_ib_client, smcibdev); ibdev 540 net/smc/smc_ib.c INIT_IB_EVENT_HANDLER(&smcibdev->event_handler, smcibdev->ibdev, ibdev 545 net/smc/smc_ib.c port_cnt = smcibdev->ibdev->phys_port_cnt; ibdev 551 net/smc/smc_ib.c smc_pnetid_by_dev_port(ibdev->dev.parent, i, ibdev 558 net/smc/smc_ib.c static void smc_ib_remove_dev(struct ib_device *ibdev, void *client_data) ibdev 562 net/smc/smc_ib.c smcibdev = ib_get_client_data(ibdev, &smc_ib_client); ibdev 563 net/smc/smc_ib.c if (!smcibdev || smcibdev->ibdev != ibdev) ibdev 565 net/smc/smc_ib.c ib_set_client_data(ibdev, &smc_ib_client, NULL); ibdev 34 net/smc/smc_ib.h struct ib_device *ibdev; ibdev 99 net/smc/smc_pnet.c struct smc_ib_device *ibdev; ibdev 129 net/smc/smc_pnet.c list_for_each_entry(ibdev, &smc_ib_devices.list, list) { ibdev 131 net/smc/smc_pnet.c if (ibdev->pnetid_by_user[ibport] && ibdev 134 net/smc/smc_pnet.c ibdev->pnetid[ibport]))) { ibdev 135 net/smc/smc_pnet.c memset(ibdev->pnetid[ibport], 0, ibdev 137 net/smc/smc_pnet.c ibdev->pnetid_by_user[ibport] = false; ibdev 292 net/smc/smc_pnet.c struct smc_ib_device *ibdev; ibdev 295 net/smc/smc_pnet.c list_for_each_entry(ibdev, &smc_ib_devices.list, list) { ibdev 296 net/smc/smc_pnet.c if (!strncmp(ibdev->ibdev->name, ib_name, ibdev 297 net/smc/smc_pnet.c sizeof(ibdev->ibdev->name)) || ibdev 298 net/smc/smc_pnet.c !strncmp(dev_name(ibdev->ibdev->dev.parent), ib_name, ibdev 303 net/smc/smc_pnet.c ibdev = NULL; ibdev 306 net/smc/smc_pnet.c return ibdev; ibdev 398 net/smc/smc_pnet.c dev_name(pnetelem->smcibdev->ibdev->dev.parent)) || ibdev 475 net/smc/smc_pnet.c struct smc_ib_device *ibdev; ibdev 510 net/smc/smc_pnet.c list_for_each_entry(ibdev, &smc_ib_devices.list, list) { ibdev 512 net/smc/smc_pnet.c if (ibdev->pnetid_by_user[ibport]) { ibdev 514 net/smc/smc_pnet.c !smc_pnet_match(ibdev->pnetid[ibport], ibdev 521 net/smc/smc_pnet.c ibdev->pnetid[ibport], ibdev 523 net/smc/smc_pnet.c tmp_entry.smcibdev = ibdev; ibdev 764 net/smc/smc_pnet.c struct smc_ib_device *ibdev; ibdev 767 net/smc/smc_pnet.c list_for_each_entry(ibdev, &smc_ib_devices.list, list) { ibdev 772 net/smc/smc_pnet.c if (!rdma_is_port_valid(ibdev->ibdev, i)) ibdev 774 net/smc/smc_pnet.c if (!ibdev->ibdev->ops.get_netdev) ibdev 776 net/smc/smc_pnet.c ndev = ibdev->ibdev->ops.get_netdev(ibdev->ibdev, i); ibdev 781 net/smc/smc_pnet.c smc_ib_port_active(ibdev, i) && ibdev 782 net/smc/smc_pnet.c !smc_ib_determine_gid(ibdev, i, ini->vlan_id, ibdev 784 net/smc/smc_pnet.c ini->ib_dev = ibdev; ibdev 803 net/smc/smc_pnet.c struct smc_ib_device *ibdev; ibdev 815 net/smc/smc_pnet.c list_for_each_entry(ibdev, &smc_ib_devices.list, list) { ibdev 817 net/smc/smc_pnet.c if (!rdma_is_port_valid(ibdev->ibdev, i)) ibdev 819 net/smc/smc_pnet.c if (smc_pnet_match(ibdev->pnetid[i - 1], ndev_pnetid) && ibdev 820 net/smc/smc_pnet.c smc_ib_port_active(ibdev, i) && ibdev 821 net/smc/smc_pnet.c !smc_ib_determine_gid(ibdev, i, ini->vlan_id, ibdev 823 net/smc/smc_pnet.c ini->ib_dev = ibdev; ibdev 511 net/smc/smc_wr.c struct ib_device *ibdev; ibdev 518 net/smc/smc_wr.c ibdev = lnk->smcibdev->ibdev; ibdev 521 net/smc/smc_wr.c ib_dma_unmap_single(ibdev, lnk->wr_rx_dma_addr, ibdev 527 net/smc/smc_wr.c ib_dma_unmap_single(ibdev, lnk->wr_tx_dma_addr, ibdev 646 net/smc/smc_wr.c struct ib_device *ibdev = lnk->smcibdev->ibdev; ibdev 652 net/smc/smc_wr.c ibdev, lnk->wr_rx_bufs, SMC_WR_BUF_SIZE * lnk->wr_rx_cnt, ibdev 654 net/smc/smc_wr.c if (ib_dma_mapping_error(ibdev, lnk->wr_rx_dma_addr)) { ibdev 660 net/smc/smc_wr.c ibdev, lnk->wr_tx_bufs, SMC_WR_BUF_SIZE * lnk->wr_tx_cnt, ibdev 662 net/smc/smc_wr.c if (ib_dma_mapping_error(ibdev, lnk->wr_tx_dma_addr)) { ibdev 674 net/smc/smc_wr.c ib_dma_unmap_single(ibdev, lnk->wr_rx_dma_addr,