Lines Matching refs:to_mdev
137 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_query_device()
151 err = mlx4_MAD_IFC(to_mdev(ibdev), MLX4_MAD_IFC_IGNORE_KEYS, in mlx4_ib_query_device()
243 struct mlx4_dev *dev = to_mdev(device)->dev; in mlx4_ib_port_link_layer()
267 if (mlx4_is_mfunc(to_mdev(ibdev)->dev) && netw_view) in ib_link_query_port()
270 err = mlx4_MAD_IFC(to_mdev(ibdev), mad_ifc_flags, port, NULL, NULL, in ib_link_query_port()
286 props->gid_tbl_len = to_mdev(ibdev)->dev->caps.gid_table_len[port]; in ib_link_query_port()
287 props->max_msg_sz = to_mdev(ibdev)->dev->caps.max_msg_sz; in ib_link_query_port()
288 props->pkey_tbl_len = to_mdev(ibdev)->dev->caps.pkey_table_len[port]; in ib_link_query_port()
319 err = mlx4_MAD_IFC(to_mdev(ibdev), mad_ifc_flags, port, in ib_link_query_port()
348 struct mlx4_ib_dev *mdev = to_mdev(ibdev); in eth_link_query_port()
429 struct mlx4_ib_dev *dev = to_mdev(ibdev); in __mlx4_ib_query_gid()
482 struct mlx4_ib_dev *dev = to_mdev(ibdev); in iboe_query_gid()
515 if (mlx4_is_mfunc(to_mdev(ibdev)->dev) && netw_view) in __mlx4_ib_query_pkey()
518 err = mlx4_MAD_IFC(to_mdev(ibdev), mad_ifc_flags, port, NULL, NULL, in __mlx4_ib_query_pkey()
548 if (mlx4_is_slave(to_mdev(ibdev)->dev)) in mlx4_ib_modify_device()
551 spin_lock_irqsave(&to_mdev(ibdev)->sm_lock, flags); in mlx4_ib_modify_device()
553 spin_unlock_irqrestore(&to_mdev(ibdev)->sm_lock, flags); in mlx4_ib_modify_device()
559 mailbox = mlx4_alloc_cmd_mailbox(to_mdev(ibdev)->dev); in mlx4_ib_modify_device()
564 mlx4_cmd(to_mdev(ibdev)->dev, mailbox->dma, 1, 0, in mlx4_ib_modify_device()
567 mlx4_free_cmd_mailbox(to_mdev(ibdev)->dev, mailbox); in mlx4_ib_modify_device()
601 struct mlx4_ib_dev *mdev = to_mdev(ibdev); in mlx4_ib_modify_port()
628 mutex_unlock(&to_mdev(ibdev)->cap_mask_mutex); in mlx4_ib_modify_port()
635 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_alloc_ucontext()
660 err = mlx4_uar_alloc(to_mdev(ibdev)->dev, &context->uar); in mlx4_ib_alloc_ucontext()
675 mlx4_uar_free(to_mdev(ibdev)->dev, &context->uar); in mlx4_ib_alloc_ucontext()
687 mlx4_uar_free(to_mdev(ibcontext->device)->dev, &context->uar); in mlx4_ib_dealloc_ucontext()
695 struct mlx4_ib_dev *dev = to_mdev(context->device); in mlx4_ib_mmap()
732 err = mlx4_pd_alloc(to_mdev(ibdev)->dev, &pd->pdn); in mlx4_ib_alloc_pd()
740 mlx4_pd_free(to_mdev(ibdev)->dev, pd->pdn); in mlx4_ib_alloc_pd()
750 mlx4_pd_free(to_mdev(pd->device)->dev, to_mpd(pd)->pdn); in mlx4_ib_dealloc_pd()
763 if (!(to_mdev(ibdev)->dev->caps.flags & MLX4_DEV_CAP_FLAG_XRC)) in mlx4_ib_alloc_xrcd()
770 err = mlx4_xrcd_alloc(to_mdev(ibdev)->dev, &xrcd->xrcdn); in mlx4_ib_alloc_xrcd()
791 mlx4_xrcd_free(to_mdev(ibdev)->dev, xrcd->xrcdn); in mlx4_ib_alloc_xrcd()
801 mlx4_xrcd_free(to_mdev(xrcd->device)->dev, to_mxrcd(xrcd)->xrcdn); in mlx4_ib_dealloc_xrcd()
810 struct mlx4_ib_dev *mdev = to_mdev(ibqp->device); in add_gid_entry()
1030 struct mlx4_ib_dev *mdev = to_mdev(qp->device); in __mlx4_ib_create_flow()
1122 struct mlx4_dev *dev = to_mdev(qp->device)->dev; in mlx4_ib_tunnel_steer_add()
1135 err = mlx4_tunnel_steer_add(to_mdev(qp->device)->dev, ib_spec->eth.val.dst_mac, in mlx4_ib_tunnel_steer_add()
1149 struct mlx4_dev *dev = (to_mdev(qp->device))->dev; in mlx4_ib_create_flow()
1227 (void)__mlx4_ib_destroy_flow(to_mdev(qp->device)->dev, in mlx4_ib_create_flow()
1233 (void)__mlx4_ib_destroy_flow(to_mdev(qp->device)->dev, in mlx4_ib_create_flow()
1246 struct mlx4_ib_dev *mdev = to_mdev(flow_id->qp->device); in mlx4_ib_destroy_flow()
1269 struct mlx4_ib_dev *mdev = to_mdev(ibqp->device); in mlx4_ib_mcg_attach()
1347 struct mlx4_ib_dev *mdev = to_mdev(ibqp->device); in mlx4_ib_mcg_detach()
2727 struct mlx4_ib_dev *ibdev = to_mdev((struct ib_device *) ibdev_ptr); in mlx4_ib_event()