Lines Matching refs:mdev

615 	struct mlx4_ib_dev *mdev = to_mdev(ibdev);  in eth_link_query_port()  local
616 struct mlx4_ib_iboe *iboe = &mdev->iboe; in eth_link_query_port()
621 int is_bonded = mlx4_is_bonded(mdev->dev); in eth_link_query_port()
623 mailbox = mlx4_alloc_cmd_mailbox(mdev->dev); in eth_link_query_port()
627 err = mlx4_cmd_box(mdev->dev, 0, mailbox->dma, port, 0, in eth_link_query_port()
637 props->gid_tbl_len = mdev->dev->caps.gid_table_len[port]; in eth_link_query_port()
638 props->max_msg_sz = mdev->dev->caps.max_msg_sz; in eth_link_query_port()
664 mlx4_free_cmd_mailbox(mdev->dev, mailbox); in eth_link_query_port()
871 struct mlx4_ib_dev *mdev = to_mdev(ibdev); in mlx4_ib_modify_port() local
872 u8 is_eth = mdev->dev->caps.port_type[port] == MLX4_PORT_TYPE_ETH; in mlx4_ib_modify_port()
884 mutex_lock(&mdev->cap_mask_mutex); in mlx4_ib_modify_port()
893 err = mlx4_ib_SET_PORT(mdev, port, in mlx4_ib_modify_port()
1232 struct mlx4_ib_dev *mdev = to_mdev(ibqp->device); in add_gid_entry() local
1240 if (mlx4_ib_add_mc(mdev, mqp, gid)) { in add_gid_entry()
1268 int mlx4_ib_add_mc(struct mlx4_ib_dev *mdev, struct mlx4_ib_qp *mqp, in mlx4_ib_add_mc() argument
1277 spin_lock_bh(&mdev->iboe.lock); in mlx4_ib_add_mc()
1278 ndev = mdev->iboe.netdevs[mqp->port - 1]; in mlx4_ib_add_mc()
1281 spin_unlock_bh(&mdev->iboe.lock); in mlx4_ib_add_mc()
1423 struct mlx4_ib_dev *mdev, in __mlx4_ib_create_default_rules() argument
1447 ret = parse_flow_attr(mdev->dev, 0, &ib_spec, in __mlx4_ib_create_default_rules()
1468 struct mlx4_ib_dev *mdev = to_mdev(qp->device); in __mlx4_ib_create_flow() local
1490 if (mlx4_map_sw_to_hw_steering_mode(mdev->dev, flow_type) < 0) in __mlx4_ib_create_flow()
1493 mailbox = mlx4_alloc_cmd_mailbox(mdev->dev); in __mlx4_ib_create_flow()
1500 ctrl->type = mlx4_map_sw_to_hw_steering_mode(mdev->dev, flow_type); in __mlx4_ib_create_flow()
1510 mdev, qp, default_table + default_flow, in __mlx4_ib_create_flow()
1513 mlx4_free_cmd_mailbox(mdev->dev, mailbox); in __mlx4_ib_create_flow()
1519 ret = parse_flow_attr(mdev->dev, qp->qp_num, ib_flow, in __mlx4_ib_create_flow()
1522 mlx4_free_cmd_mailbox(mdev->dev, mailbox); in __mlx4_ib_create_flow()
1529 ret = mlx4_cmd_imm(mdev->dev, mailbox->dma, reg_id, size >> 2, 0, in __mlx4_ib_create_flow()
1539 mlx4_free_cmd_mailbox(mdev->dev, mailbox); in __mlx4_ib_create_flow()
1685 struct mlx4_ib_dev *mdev = to_mdev(flow_id->qp->device); in mlx4_ib_destroy_flow() local
1689 err = __mlx4_ib_destroy_flow(mdev->dev, mflow->reg_id[i].id); in mlx4_ib_destroy_flow()
1693 err = __mlx4_ib_destroy_flow(mdev->dev, in mlx4_ib_destroy_flow()
1708 struct mlx4_ib_dev *mdev = to_mdev(ibqp->device); in mlx4_ib_mcg_attach() local
1709 struct mlx4_dev *dev = mdev->dev; in mlx4_ib_mcg_attach()
1715 if (mdev->dev->caps.steering_mode == in mlx4_ib_mcg_attach()
1722 err = mlx4_multicast_attach(mdev->dev, &mqp->mqp, gid->raw, mqp->port, in mlx4_ib_mcg_attach()
1733 err = mlx4_multicast_attach(mdev->dev, &mqp->mqp, gid->raw, in mlx4_ib_mcg_attach()
1756 mlx4_multicast_detach(mdev->dev, &mqp->mqp, gid->raw, in mlx4_ib_mcg_attach()
1759 mlx4_multicast_detach(mdev->dev, &mqp->mqp, gid->raw, in mlx4_ib_mcg_attach()
1786 struct mlx4_ib_dev *mdev = to_mdev(ibqp->device); in mlx4_ib_mcg_detach() local
1787 struct mlx4_dev *dev = mdev->dev; in mlx4_ib_mcg_detach()
1794 if (mdev->dev->caps.steering_mode == in mlx4_ib_mcg_detach()
1814 err = mlx4_multicast_detach(mdev->dev, &mqp->mqp, gid->raw, in mlx4_ib_mcg_detach()
1820 err = mlx4_multicast_detach(mdev->dev, &mqp->mqp, gid->raw, in mlx4_ib_mcg_detach()
1829 spin_lock_bh(&mdev->iboe.lock); in mlx4_ib_mcg_detach()
1830 ndev = ge->added ? mdev->iboe.netdevs[ge->port - 1] : NULL; in mlx4_ib_mcg_detach()
1833 spin_unlock_bh(&mdev->iboe.lock); in mlx4_ib_mcg_detach()
2530 int mlx4_ib_steer_qp_reg(struct mlx4_ib_dev *mdev, struct mlx4_ib_qp *mqp, in mlx4_ib_steer_qp_reg() argument
2558 err = __mlx4_ib_destroy_flow(mdev->dev, mqp->reg_id); in mlx4_ib_steer_qp_reg()