/linux-4.4.14/drivers/infiniband/hw/mthca/ |
H A D | mthca_provider.c | 442 struct mthca_create_srq ucmd; mthca_create_srq() local 457 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { mthca_create_srq() 463 context->db_tab, ucmd.db_index, mthca_create_srq() 464 ucmd.db_page); mthca_create_srq() 469 srq->mr.ibmr.lkey = ucmd.lkey; mthca_create_srq() 470 srq->db_index = ucmd.db_index; mthca_create_srq() 478 context->db_tab, ucmd.db_index); mthca_create_srq() 518 struct mthca_create_qp ucmd; mthca_create_qp() local 539 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { mthca_create_qp() 546 ucmd.sq_db_index, ucmd.sq_db_page); mthca_create_qp() 554 ucmd.rq_db_index, ucmd.rq_db_page); mthca_create_qp() 559 ucmd.sq_db_index); mthca_create_qp() 564 qp->mr.ibmr.lkey = ucmd.lkey; mthca_create_qp() 565 qp->sq.db_index = ucmd.sq_db_index; mthca_create_qp() 566 qp->rq.db_index = ucmd.rq_db_index; mthca_create_qp() 581 ucmd.sq_db_index); mthca_create_qp() 585 ucmd.rq_db_index); mthca_create_qp() 654 struct mthca_create_cq ucmd; mthca_create_cq() local 666 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) mthca_create_cq() 671 ucmd.set_db_index, ucmd.set_db_page); mthca_create_cq() 677 ucmd.arm_db_index, ucmd.arm_db_page); mthca_create_cq() 689 cq->buf.mr.ibmr.lkey = ucmd.lkey; mthca_create_cq() 690 cq->set_ci_db_index = ucmd.set_db_index; mthca_create_cq() 691 cq->arm_db_index = ucmd.arm_db_index; mthca_create_cq() 699 context ? ucmd.pdn : to_mdev(ibdev)->driver_pd.pd_num, mthca_create_cq() 720 to_mucontext(context)->db_tab, ucmd.arm_db_index); mthca_create_cq() 725 to_mucontext(context)->db_tab, ucmd.set_db_index); mthca_create_cq() 779 struct mthca_resize_cq ucmd; mthca_resize_cq() local 800 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { mthca_resize_cq() 804 lkey = ucmd.lkey; mthca_resize_cq() 990 struct mthca_reg_mr ucmd; mthca_reg_user_mr() local 997 if (udata->inlen - sizeof (struct ib_uverbs_cmd_hdr) < sizeof ucmd) { mthca_reg_user_mr() 1004 ucmd.mr_attrs = 0; mthca_reg_user_mr() 1005 } else if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) mthca_reg_user_mr() 1013 ucmd.mr_attrs & MTHCA_MR_DMASYNC); mthca_reg_user_mr()
|
/linux-4.4.14/drivers/infiniband/hw/mlx5/ |
H A D | srq.c | 81 struct mlx5_ib_create_srq ucmd; create_srq_user() local 91 sizeof(ucmd)) ? (sizeof(ucmd) - create_srq_user() 92 sizeof(ucmd.reserved)) : sizeof(ucmd); create_srq_user() 94 if (ib_copy_from_udata(&ucmd, udata, ucmdlen)) { create_srq_user() 99 if (ucmdlen == sizeof(ucmd) && create_srq_user() 100 ucmd.reserved != 0) create_srq_user() 103 srq->wq_sig = !!(ucmd.flags & MLX5_SRQ_FLAG_SIGNATURE); create_srq_user() 105 srq->umem = ib_umem_get(pd->uobject->context, ucmd.buf_addr, buf_size, create_srq_user() 113 mlx5_ib_cont_pages(srq->umem, ucmd.buf_addr, &npages, create_srq_user() 115 err = mlx5_ib_get_buf_offset(ucmd.buf_addr, page_shift, create_srq_user() 132 ucmd.db_addr, &srq->db); create_srq_user()
|
H A D | cq.c | 616 struct mlx5_ib_create_cq ucmd; create_cq_user() local 625 sizeof(ucmd)) ? (sizeof(ucmd) - create_cq_user() 626 sizeof(ucmd.reserved)) : sizeof(ucmd); create_cq_user() 628 if (ib_copy_from_udata(&ucmd, udata, ucmdlen)) create_cq_user() 631 if (ucmdlen == sizeof(ucmd) && create_cq_user() 632 ucmd.reserved != 0) create_cq_user() 635 if (ucmd.cqe_size != 64 && ucmd.cqe_size != 128) create_cq_user() 638 *cqe_size = ucmd.cqe_size; create_cq_user() 640 cq->buf.umem = ib_umem_get(context, ucmd.buf_addr, create_cq_user() 641 entries * ucmd.cqe_size, create_cq_user() 648 err = mlx5_ib_db_map_user(to_mucontext(context), ucmd.db_addr, create_cq_user() 653 mlx5_ib_cont_pages(cq->buf.umem, ucmd.buf_addr, &npages, &page_shift, create_cq_user() 656 ucmd.buf_addr, entries * ucmd.cqe_size, npages, page_shift, ncont); create_cq_user() 962 struct mlx5_ib_resize_cq ucmd; resize_user() local 968 err = ib_copy_from_udata(&ucmd, udata, sizeof(ucmd)); resize_user() 972 if (ucmd.reserved0 || ucmd.reserved1) resize_user() 975 umem = ib_umem_get(context, ucmd.buf_addr, entries * ucmd.cqe_size, resize_user() 982 mlx5_ib_cont_pages(umem, ucmd.buf_addr, &npages, page_shift, resize_user() 986 *cqe_size = ucmd.cqe_size; resize_user()
|
H A D | qp.c | 216 int has_rq, struct mlx5_ib_qp *qp, struct mlx5_ib_create_qp *ucmd) set_rq_size() 230 if (ucmd) { set_rq_size() 231 qp->rq.wqe_cnt = ucmd->rq_wqe_count; set_rq_size() 232 qp->rq.wqe_shift = ucmd->rq_wqe_shift; set_rq_size() 369 struct mlx5_ib_create_qp *ucmd) set_user_buf_size() 379 if (ucmd->sq_wqe_count && ((1 << ilog2(ucmd->sq_wqe_count)) != ucmd->sq_wqe_count)) { set_user_buf_size() 381 ucmd->sq_wqe_count, ucmd->sq_wqe_count); set_user_buf_size() 385 qp->sq.wqe_cnt = ucmd->sq_wqe_count; set_user_buf_size() 599 struct mlx5_ib_create_qp ucmd; create_user_qp() local 608 err = ib_copy_from_udata(&ucmd, udata, sizeof(ucmd)); create_user_qp() 641 err = set_user_buf_size(dev, qp, &ucmd); create_user_qp() 645 if (ucmd.buf_addr && qp->buf_size) { create_user_qp() 646 qp->umem = ib_umem_get(pd->uobject->context, ucmd.buf_addr, create_user_qp() 658 mlx5_ib_cont_pages(qp->umem, ucmd.buf_addr, &npages, &page_shift, create_user_qp() 660 err = mlx5_ib_get_buf_offset(ucmd.buf_addr, page_shift, &offset); create_user_qp() 666 ucmd.buf_addr, qp->buf_size, npages, page_shift, ncont, offset); create_user_qp() 685 err = mlx5_ib_db_map_user(context, ucmd.db_addr, &qp->db); create_user_qp() 864 struct mlx5_ib_create_qp ucmd; create_qp_common() local 887 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) { create_qp_common() 892 qp->wq_sig = !!(ucmd.flags & MLX5_QP_FLAG_SIGNATURE); create_qp_common() 893 qp->scat_cqe = !!(ucmd.flags & MLX5_QP_FLAG_SCATTER_CQE); create_qp_common() 900 qp, (pd && pd->uobject) ? &ucmd : NULL); create_qp_common() 910 mlx5_ib_dbg(dev, "requested sq_wqe_count (%d)\n", ucmd.sq_wqe_count); create_qp_common() 911 if (ucmd.rq_wqe_shift != qp->rq.wqe_shift || create_qp_common() 912 ucmd.rq_wqe_count != qp->rq.wqe_cnt) { create_qp_common() 916 if (ucmd.sq_wqe_count > max_wqes) { create_qp_common() 918 ucmd.sq_wqe_count, max_wqes); create_qp_common() 215 set_rq_size(struct mlx5_ib_dev *dev, struct ib_qp_cap *cap, int has_rq, struct mlx5_ib_qp *qp, struct mlx5_ib_create_qp *ucmd) set_rq_size() argument 367 set_user_buf_size(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, struct mlx5_ib_create_qp *ucmd) set_user_buf_size() argument
|
/linux-4.4.14/drivers/infiniband/hw/mlx4/ |
H A D | srq.c | 110 struct mlx4_ib_create_srq ucmd; mlx4_ib_create_srq() local 112 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { mlx4_ib_create_srq() 117 srq->umem = ib_umem_get(pd->uobject->context, ucmd.buf_addr, mlx4_ib_create_srq() 134 ucmd.db_addr, &srq->db); mlx4_ib_create_srq()
|
H A D | cq.c | 203 struct mlx4_ib_create_cq ucmd; mlx4_ib_create_cq() local 205 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { mlx4_ib_create_cq() 211 ucmd.buf_addr, entries); mlx4_ib_create_cq() 215 err = mlx4_ib_db_map_user(to_mucontext(context), ucmd.db_addr, mlx4_ib_create_cq() 310 struct mlx4_ib_resize_cq ucmd; mlx4_alloc_resize_umem() local 316 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) mlx4_alloc_resize_umem() 324 &cq->resize_umem, ucmd.buf_addr, entries); mlx4_alloc_resize_umem()
|
H A D | qp.c | 533 struct mlx4_ib_create_qp *ucmd) set_user_sq_size() 536 if ((1 << ucmd->log_sq_bb_count) > dev->dev->caps.max_wqes || set_user_sq_size() 537 ucmd->log_sq_stride > set_user_sq_size() 539 ucmd->log_sq_stride < MLX4_IB_MIN_SQ_STRIDE) set_user_sq_size() 542 qp->sq.wqe_cnt = 1 << ucmd->log_sq_bb_count; set_user_sq_size() 543 qp->sq.wqe_shift = ucmd->log_sq_stride; set_user_sq_size() 724 struct mlx4_ib_create_qp ucmd; create_qp_common() local 726 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { create_qp_common() 731 qp->sq_no_prefetch = ucmd.sq_no_prefetch; create_qp_common() 733 err = set_user_sq_size(dev, qp, &ucmd); create_qp_common() 737 qp->umem = ib_umem_get(pd->uobject->context, ucmd.buf_addr, create_qp_common() 755 ucmd.db_addr, &qp->db); create_qp_common() 531 set_user_sq_size(struct mlx4_ib_dev *dev, struct mlx4_ib_qp *qp, struct mlx4_ib_create_qp *ucmd) set_user_sq_size() argument
|
/linux-4.4.14/drivers/block/rsxx/ |
H A D | cregs.c | 678 struct rsxx_reg_access __user *ucmd, rsxx_reg_access() 684 st = copy_from_user(&cmd, ucmd, sizeof(cmd)); rsxx_reg_access() 695 st = put_user(cmd.stat, &ucmd->stat); rsxx_reg_access() 700 st = copy_to_user(ucmd->data, cmd.data, cmd.cnt); rsxx_reg_access() 677 rsxx_reg_access(struct rsxx_cardinfo *card, struct rsxx_reg_access __user *ucmd, int read) rsxx_reg_access() argument
|
H A D | rsxx_priv.h | 427 struct rsxx_reg_access __user *ucmd,
|
/linux-4.4.14/drivers/staging/rdma/ipath/ |
H A D | ipath_file_ops.c | 2238 const struct ipath_cmd __user *ucmd; ipath_write() local 2251 ucmd = (const struct ipath_cmd __user *) data; ipath_write() 2253 if (copy_from_user(&cmd.type, &ucmd->type, sizeof(cmd.type))) { ipath_write() 2266 src = &ucmd->cmd.user_info; ipath_write() 2271 src = &ucmd->cmd.recv_ctrl; ipath_write() 2276 src = &ucmd->cmd.port_info; ipath_write() 2282 src = &ucmd->cmd.tid_info; ipath_write() 2287 src = &ucmd->cmd.part_key; ipath_write() 2292 src = &ucmd->cmd.slave_mask_addr; ipath_write() 2302 src = &ucmd->cmd.poll_type; ipath_write() 2307 src = &ucmd->cmd.armlaunch_ctrl; ipath_write() 2312 src = &ucmd->cmd.sdma_inflight; ipath_write() 2317 src = &ucmd->cmd.sdma_complete; ipath_write()
|
/linux-4.4.14/drivers/infiniband/hw/qib/ |
H A D | qib_file_ops.c | 2064 const struct qib_cmd __user *ucmd; qib_write() local 2080 ucmd = (const struct qib_cmd __user *) data; qib_write() 2082 if (copy_from_user(&cmd.type, &ucmd->type, sizeof(cmd.type))) { qib_write() 2094 src = &ucmd->cmd.user_info; qib_write() 2100 src = &ucmd->cmd.recv_ctrl; qib_write() 2106 src = &ucmd->cmd.ctxt_info; qib_write() 2113 src = &ucmd->cmd.tid_info; qib_write() 2119 src = &ucmd->cmd.part_key; qib_write() 2132 src = &ucmd->cmd.poll_type; qib_write() 2138 src = &ucmd->cmd.armlaunch_ctrl; qib_write() 2144 src = &ucmd->cmd.sdma_inflight; qib_write() 2150 src = &ucmd->cmd.sdma_complete; qib_write() 2156 src = &ucmd->cmd.event_mask; qib_write()
|
/linux-4.4.14/drivers/scsi/sym53c8xx_2/ |
H A D | sym_glue.c | 142 struct sym_ucmd *ucmd = SYM_UCMD_PTR(cmd); sym_xpt_done() local 145 if (ucmd->eh_done) sym_xpt_done() 146 complete(ucmd->eh_done); sym_xpt_done() 593 struct sym_ucmd *ucmd = SYM_UCMD_PTR(cmd); sym_eh_handler() local 673 ucmd->eh_done = &eh_done; sym_eh_handler() 676 ucmd->eh_done = NULL; sym_eh_handler()
|
/linux-4.4.14/drivers/staging/rdma/hfi1/ |
H A D | file_ops.c | 208 const struct hfi1_cmd __user *ucmd; hfi1_file_write() local 228 ucmd = (const struct hfi1_cmd __user *)data; hfi1_file_write() 229 if (copy_from_user(&cmd, ucmd, sizeof(cmd))) { hfi1_file_write()
|
/linux-4.4.14/drivers/nvme/host/ |
H A D | pci.c | 1883 struct nvme_passthru_cmd __user *ucmd) nvme_user_cmd() 1892 if (copy_from_user(&cmd, ucmd, sizeof(cmd))) nvme_user_cmd() 1915 if (put_user(cmd.result, &ucmd->result)) nvme_user_cmd() 1882 nvme_user_cmd(struct nvme_dev *dev, struct nvme_ns *ns, struct nvme_passthru_cmd __user *ucmd) nvme_user_cmd() argument
|