Lines Matching refs:to_mdev
130 struct mlx4_ib_dev *ibdev = to_mdev(device); in mlx4_ib_get_netdev()
197 struct mlx4_ib_dev *ibdev = to_mdev(device); in mlx4_ib_add_gid()
271 struct mlx4_ib_dev *ibdev = to_mdev(device); in mlx4_ib_del_gid()
363 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_query_device()
397 err = mlx4_MAD_IFC(to_mdev(ibdev), MLX4_MAD_IFC_IGNORE_KEYS, in mlx4_ib_query_device()
510 struct mlx4_dev *dev = to_mdev(device)->dev; in mlx4_ib_port_link_layer()
534 if (mlx4_is_mfunc(to_mdev(ibdev)->dev) && netw_view) in ib_link_query_port()
537 err = mlx4_MAD_IFC(to_mdev(ibdev), mad_ifc_flags, port, NULL, NULL, in ib_link_query_port()
553 props->gid_tbl_len = to_mdev(ibdev)->dev->caps.gid_table_len[port]; in ib_link_query_port()
554 props->max_msg_sz = to_mdev(ibdev)->dev->caps.max_msg_sz; in ib_link_query_port()
555 props->pkey_tbl_len = to_mdev(ibdev)->dev->caps.pkey_table_len[port]; in ib_link_query_port()
586 err = mlx4_MAD_IFC(to_mdev(ibdev), mad_ifc_flags, port, in ib_link_query_port()
615 struct mlx4_ib_dev *mdev = to_mdev(ibdev); in eth_link_query_port()
695 struct mlx4_ib_dev *dev = to_mdev(ibdev); in __mlx4_ib_query_gid()
785 if (mlx4_is_mfunc(to_mdev(ibdev)->dev) && netw_view) in __mlx4_ib_query_pkey()
788 err = mlx4_MAD_IFC(to_mdev(ibdev), mad_ifc_flags, port, NULL, NULL, in __mlx4_ib_query_pkey()
818 if (mlx4_is_slave(to_mdev(ibdev)->dev)) in mlx4_ib_modify_device()
821 spin_lock_irqsave(&to_mdev(ibdev)->sm_lock, flags); in mlx4_ib_modify_device()
823 spin_unlock_irqrestore(&to_mdev(ibdev)->sm_lock, flags); in mlx4_ib_modify_device()
829 mailbox = mlx4_alloc_cmd_mailbox(to_mdev(ibdev)->dev); in mlx4_ib_modify_device()
834 mlx4_cmd(to_mdev(ibdev)->dev, mailbox->dma, 1, 0, in mlx4_ib_modify_device()
837 mlx4_free_cmd_mailbox(to_mdev(ibdev)->dev, mailbox); in mlx4_ib_modify_device()
871 struct mlx4_ib_dev *mdev = to_mdev(ibdev); in mlx4_ib_modify_port()
898 mutex_unlock(&to_mdev(ibdev)->cap_mask_mutex); in mlx4_ib_modify_port()
905 struct mlx4_ib_dev *dev = to_mdev(ibdev); in mlx4_ib_alloc_ucontext()
930 err = mlx4_uar_alloc(to_mdev(ibdev)->dev, &context->uar); in mlx4_ib_alloc_ucontext()
945 mlx4_uar_free(to_mdev(ibdev)->dev, &context->uar); in mlx4_ib_alloc_ucontext()
957 mlx4_uar_free(to_mdev(ibcontext->device)->dev, &context->uar); in mlx4_ib_dealloc_ucontext()
1075 struct mlx4_ib_dev *dev = to_mdev(context->device); in mlx4_ib_mmap()
1152 err = mlx4_pd_alloc(to_mdev(ibdev)->dev, &pd->pdn); in mlx4_ib_alloc_pd()
1160 mlx4_pd_free(to_mdev(ibdev)->dev, pd->pdn); in mlx4_ib_alloc_pd()
1170 mlx4_pd_free(to_mdev(pd->device)->dev, to_mpd(pd)->pdn); in mlx4_ib_dealloc_pd()
1184 if (!(to_mdev(ibdev)->dev->caps.flags & MLX4_DEV_CAP_FLAG_XRC)) in mlx4_ib_alloc_xrcd()
1191 err = mlx4_xrcd_alloc(to_mdev(ibdev)->dev, &xrcd->xrcdn); in mlx4_ib_alloc_xrcd()
1213 mlx4_xrcd_free(to_mdev(ibdev)->dev, xrcd->xrcdn); in mlx4_ib_alloc_xrcd()
1223 mlx4_xrcd_free(to_mdev(xrcd->device)->dev, to_mxrcd(xrcd)->xrcdn); in mlx4_ib_dealloc_xrcd()
1232 struct mlx4_ib_dev *mdev = to_mdev(ibqp->device); in add_gid_entry()
1468 struct mlx4_ib_dev *mdev = to_mdev(qp->device); in __mlx4_ib_create_flow()
1560 struct mlx4_dev *dev = to_mdev(qp->device)->dev; in mlx4_ib_tunnel_steer_add()
1573 err = mlx4_tunnel_steer_add(to_mdev(qp->device)->dev, ib_spec->eth.val.dst_mac, in mlx4_ib_tunnel_steer_add()
1587 struct mlx4_dev *dev = (to_mdev(qp->device))->dev; in mlx4_ib_create_flow()
1666 (void)__mlx4_ib_destroy_flow(to_mdev(qp->device)->dev, in mlx4_ib_create_flow()
1672 (void)__mlx4_ib_destroy_flow(to_mdev(qp->device)->dev, in mlx4_ib_create_flow()
1685 struct mlx4_ib_dev *mdev = to_mdev(flow_id->qp->device); in mlx4_ib_destroy_flow()
1708 struct mlx4_ib_dev *mdev = to_mdev(ibqp->device); in mlx4_ib_mcg_attach()
1786 struct mlx4_ib_dev *mdev = to_mdev(ibqp->device); in mlx4_ib_mcg_detach()
2755 struct mlx4_ib_dev *ibdev = to_mdev((struct ib_device *) ibdev_ptr); in mlx4_ib_event()