Searched refs:ucmd (Results 1 - 14 of 14) sorted by relevance

/linux-4.4.14/drivers/infiniband/hw/mthca/
H A Dmthca_provider.c442 struct mthca_create_srq ucmd; mthca_create_srq() local
457 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { mthca_create_srq()
463 context->db_tab, ucmd.db_index, mthca_create_srq()
464 ucmd.db_page); mthca_create_srq()
469 srq->mr.ibmr.lkey = ucmd.lkey; mthca_create_srq()
470 srq->db_index = ucmd.db_index; mthca_create_srq()
478 context->db_tab, ucmd.db_index); mthca_create_srq()
518 struct mthca_create_qp ucmd; mthca_create_qp() local
539 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { mthca_create_qp()
546 ucmd.sq_db_index, ucmd.sq_db_page); mthca_create_qp()
554 ucmd.rq_db_index, ucmd.rq_db_page); mthca_create_qp()
559 ucmd.sq_db_index); mthca_create_qp()
564 qp->mr.ibmr.lkey = ucmd.lkey; mthca_create_qp()
565 qp->sq.db_index = ucmd.sq_db_index; mthca_create_qp()
566 qp->rq.db_index = ucmd.rq_db_index; mthca_create_qp()
581 ucmd.sq_db_index); mthca_create_qp()
585 ucmd.rq_db_index); mthca_create_qp()
654 struct mthca_create_cq ucmd; mthca_create_cq() local
666 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) mthca_create_cq()
671 ucmd.set_db_index, ucmd.set_db_page); mthca_create_cq()
677 ucmd.arm_db_index, ucmd.arm_db_page); mthca_create_cq()
689 cq->buf.mr.ibmr.lkey = ucmd.lkey; mthca_create_cq()
690 cq->set_ci_db_index = ucmd.set_db_index; mthca_create_cq()
691 cq->arm_db_index = ucmd.arm_db_index; mthca_create_cq()
699 context ? ucmd.pdn : to_mdev(ibdev)->driver_pd.pd_num, mthca_create_cq()
720 to_mucontext(context)->db_tab, ucmd.arm_db_index); mthca_create_cq()
725 to_mucontext(context)->db_tab, ucmd.set_db_index); mthca_create_cq()
779 struct mthca_resize_cq ucmd; mthca_resize_cq() local
800 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { mthca_resize_cq()
804 lkey = ucmd.lkey; mthca_resize_cq()
990 struct mthca_reg_mr ucmd; mthca_reg_user_mr() local
997 if (udata->inlen - sizeof (struct ib_uverbs_cmd_hdr) < sizeof ucmd) { mthca_reg_user_mr()
1004 ucmd.mr_attrs = 0; mthca_reg_user_mr()
1005 } else if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) mthca_reg_user_mr()
1013 ucmd.mr_attrs & MTHCA_MR_DMASYNC); mthca_reg_user_mr()
/linux-4.4.14/drivers/infiniband/hw/mlx5/
H A Dsrq.c81 struct mlx5_ib_create_srq ucmd; create_srq_user() local
91 sizeof(ucmd)) ? (sizeof(ucmd) - create_srq_user()
92 sizeof(ucmd.reserved)) : sizeof(ucmd); create_srq_user()
94 if (ib_copy_from_udata(&ucmd, udata, ucmdlen)) { create_srq_user()
99 if (ucmdlen == sizeof(ucmd) && create_srq_user()
100 ucmd.reserved != 0) create_srq_user()
103 srq->wq_sig = !!(ucmd.flags & MLX5_SRQ_FLAG_SIGNATURE); create_srq_user()
105 srq->umem = ib_umem_get(pd->uobject->context, ucmd.buf_addr, buf_size, create_srq_user()
113 mlx5_ib_cont_pages(srq->umem, ucmd.buf_addr, &npages, create_srq_user()
115 err = mlx5_ib_get_buf_offset(ucmd.buf_addr, page_shift, create_srq_user()
132 ucmd.db_addr, &srq->db); create_srq_user()
H A Dcq.c616 struct mlx5_ib_create_cq ucmd; create_cq_user() local
625 sizeof(ucmd)) ? (sizeof(ucmd) - create_cq_user()
626 sizeof(ucmd.reserved)) : sizeof(ucmd); create_cq_user()
628 if (ib_copy_from_udata(&ucmd, udata, ucmdlen)) create_cq_user()
631 if (ucmdlen == sizeof(ucmd) && create_cq_user()
632 ucmd.reserved != 0) create_cq_user()
635 if (ucmd.cqe_size != 64 && ucmd.cqe_size != 128) create_cq_user()
638 *cqe_size = ucmd.cqe_size; create_cq_user()
640 cq->buf.umem = ib_umem_get(context, ucmd.buf_addr, create_cq_user()
641 entries * ucmd.cqe_size, create_cq_user()
648 err = mlx5_ib_db_map_user(to_mucontext(context), ucmd.db_addr, create_cq_user()
653 mlx5_ib_cont_pages(cq->buf.umem, ucmd.buf_addr, &npages, &page_shift, create_cq_user()
656 ucmd.buf_addr, entries * ucmd.cqe_size, npages, page_shift, ncont); create_cq_user()
962 struct mlx5_ib_resize_cq ucmd; resize_user() local
968 err = ib_copy_from_udata(&ucmd, udata, sizeof(ucmd)); resize_user()
972 if (ucmd.reserved0 || ucmd.reserved1) resize_user()
975 umem = ib_umem_get(context, ucmd.buf_addr, entries * ucmd.cqe_size, resize_user()
982 mlx5_ib_cont_pages(umem, ucmd.buf_addr, &npages, page_shift, resize_user()
986 *cqe_size = ucmd.cqe_size; resize_user()
H A Dqp.c216 int has_rq, struct mlx5_ib_qp *qp, struct mlx5_ib_create_qp *ucmd) set_rq_size()
230 if (ucmd) { set_rq_size()
231 qp->rq.wqe_cnt = ucmd->rq_wqe_count; set_rq_size()
232 qp->rq.wqe_shift = ucmd->rq_wqe_shift; set_rq_size()
369 struct mlx5_ib_create_qp *ucmd) set_user_buf_size()
379 if (ucmd->sq_wqe_count && ((1 << ilog2(ucmd->sq_wqe_count)) != ucmd->sq_wqe_count)) { set_user_buf_size()
381 ucmd->sq_wqe_count, ucmd->sq_wqe_count); set_user_buf_size()
385 qp->sq.wqe_cnt = ucmd->sq_wqe_count; set_user_buf_size()
599 struct mlx5_ib_create_qp ucmd; create_user_qp() local
608 err = ib_copy_from_udata(&ucmd, udata, sizeof(ucmd)); create_user_qp()
641 err = set_user_buf_size(dev, qp, &ucmd); create_user_qp()
645 if (ucmd.buf_addr && qp->buf_size) { create_user_qp()
646 qp->umem = ib_umem_get(pd->uobject->context, ucmd.buf_addr, create_user_qp()
658 mlx5_ib_cont_pages(qp->umem, ucmd.buf_addr, &npages, &page_shift, create_user_qp()
660 err = mlx5_ib_get_buf_offset(ucmd.buf_addr, page_shift, &offset); create_user_qp()
666 ucmd.buf_addr, qp->buf_size, npages, page_shift, ncont, offset); create_user_qp()
685 err = mlx5_ib_db_map_user(context, ucmd.db_addr, &qp->db); create_user_qp()
864 struct mlx5_ib_create_qp ucmd; create_qp_common() local
887 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) { create_qp_common()
892 qp->wq_sig = !!(ucmd.flags & MLX5_QP_FLAG_SIGNATURE); create_qp_common()
893 qp->scat_cqe = !!(ucmd.flags & MLX5_QP_FLAG_SCATTER_CQE); create_qp_common()
900 qp, (pd && pd->uobject) ? &ucmd : NULL); create_qp_common()
910 mlx5_ib_dbg(dev, "requested sq_wqe_count (%d)\n", ucmd.sq_wqe_count); create_qp_common()
911 if (ucmd.rq_wqe_shift != qp->rq.wqe_shift || create_qp_common()
912 ucmd.rq_wqe_count != qp->rq.wqe_cnt) { create_qp_common()
916 if (ucmd.sq_wqe_count > max_wqes) { create_qp_common()
918 ucmd.sq_wqe_count, max_wqes); create_qp_common()
215 set_rq_size(struct mlx5_ib_dev *dev, struct ib_qp_cap *cap, int has_rq, struct mlx5_ib_qp *qp, struct mlx5_ib_create_qp *ucmd) set_rq_size() argument
367 set_user_buf_size(struct mlx5_ib_dev *dev, struct mlx5_ib_qp *qp, struct mlx5_ib_create_qp *ucmd) set_user_buf_size() argument
/linux-4.4.14/drivers/infiniband/hw/mlx4/
H A Dsrq.c110 struct mlx4_ib_create_srq ucmd; mlx4_ib_create_srq() local
112 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { mlx4_ib_create_srq()
117 srq->umem = ib_umem_get(pd->uobject->context, ucmd.buf_addr, mlx4_ib_create_srq()
134 ucmd.db_addr, &srq->db); mlx4_ib_create_srq()
H A Dcq.c203 struct mlx4_ib_create_cq ucmd; mlx4_ib_create_cq() local
205 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { mlx4_ib_create_cq()
211 ucmd.buf_addr, entries); mlx4_ib_create_cq()
215 err = mlx4_ib_db_map_user(to_mucontext(context), ucmd.db_addr, mlx4_ib_create_cq()
310 struct mlx4_ib_resize_cq ucmd; mlx4_alloc_resize_umem() local
316 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) mlx4_alloc_resize_umem()
324 &cq->resize_umem, ucmd.buf_addr, entries); mlx4_alloc_resize_umem()
H A Dqp.c533 struct mlx4_ib_create_qp *ucmd) set_user_sq_size()
536 if ((1 << ucmd->log_sq_bb_count) > dev->dev->caps.max_wqes || set_user_sq_size()
537 ucmd->log_sq_stride > set_user_sq_size()
539 ucmd->log_sq_stride < MLX4_IB_MIN_SQ_STRIDE) set_user_sq_size()
542 qp->sq.wqe_cnt = 1 << ucmd->log_sq_bb_count; set_user_sq_size()
543 qp->sq.wqe_shift = ucmd->log_sq_stride; set_user_sq_size()
724 struct mlx4_ib_create_qp ucmd; create_qp_common() local
726 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { create_qp_common()
731 qp->sq_no_prefetch = ucmd.sq_no_prefetch; create_qp_common()
733 err = set_user_sq_size(dev, qp, &ucmd); create_qp_common()
737 qp->umem = ib_umem_get(pd->uobject->context, ucmd.buf_addr, create_qp_common()
755 ucmd.db_addr, &qp->db); create_qp_common()
531 set_user_sq_size(struct mlx4_ib_dev *dev, struct mlx4_ib_qp *qp, struct mlx4_ib_create_qp *ucmd) set_user_sq_size() argument
/linux-4.4.14/drivers/block/rsxx/
H A Dcregs.c678 struct rsxx_reg_access __user *ucmd, rsxx_reg_access()
684 st = copy_from_user(&cmd, ucmd, sizeof(cmd)); rsxx_reg_access()
695 st = put_user(cmd.stat, &ucmd->stat); rsxx_reg_access()
700 st = copy_to_user(ucmd->data, cmd.data, cmd.cnt); rsxx_reg_access()
677 rsxx_reg_access(struct rsxx_cardinfo *card, struct rsxx_reg_access __user *ucmd, int read) rsxx_reg_access() argument
H A Drsxx_priv.h427 struct rsxx_reg_access __user *ucmd,
/linux-4.4.14/drivers/staging/rdma/ipath/
H A Dipath_file_ops.c2238 const struct ipath_cmd __user *ucmd; ipath_write() local
2251 ucmd = (const struct ipath_cmd __user *) data; ipath_write()
2253 if (copy_from_user(&cmd.type, &ucmd->type, sizeof(cmd.type))) { ipath_write()
2266 src = &ucmd->cmd.user_info; ipath_write()
2271 src = &ucmd->cmd.recv_ctrl; ipath_write()
2276 src = &ucmd->cmd.port_info; ipath_write()
2282 src = &ucmd->cmd.tid_info; ipath_write()
2287 src = &ucmd->cmd.part_key; ipath_write()
2292 src = &ucmd->cmd.slave_mask_addr; ipath_write()
2302 src = &ucmd->cmd.poll_type; ipath_write()
2307 src = &ucmd->cmd.armlaunch_ctrl; ipath_write()
2312 src = &ucmd->cmd.sdma_inflight; ipath_write()
2317 src = &ucmd->cmd.sdma_complete; ipath_write()
/linux-4.4.14/drivers/infiniband/hw/qib/
H A Dqib_file_ops.c2064 const struct qib_cmd __user *ucmd; qib_write() local
2080 ucmd = (const struct qib_cmd __user *) data; qib_write()
2082 if (copy_from_user(&cmd.type, &ucmd->type, sizeof(cmd.type))) { qib_write()
2094 src = &ucmd->cmd.user_info; qib_write()
2100 src = &ucmd->cmd.recv_ctrl; qib_write()
2106 src = &ucmd->cmd.ctxt_info; qib_write()
2113 src = &ucmd->cmd.tid_info; qib_write()
2119 src = &ucmd->cmd.part_key; qib_write()
2132 src = &ucmd->cmd.poll_type; qib_write()
2138 src = &ucmd->cmd.armlaunch_ctrl; qib_write()
2144 src = &ucmd->cmd.sdma_inflight; qib_write()
2150 src = &ucmd->cmd.sdma_complete; qib_write()
2156 src = &ucmd->cmd.event_mask; qib_write()
/linux-4.4.14/drivers/scsi/sym53c8xx_2/
H A Dsym_glue.c142 struct sym_ucmd *ucmd = SYM_UCMD_PTR(cmd); sym_xpt_done() local
145 if (ucmd->eh_done) sym_xpt_done()
146 complete(ucmd->eh_done); sym_xpt_done()
593 struct sym_ucmd *ucmd = SYM_UCMD_PTR(cmd); sym_eh_handler() local
673 ucmd->eh_done = &eh_done; sym_eh_handler()
676 ucmd->eh_done = NULL; sym_eh_handler()
/linux-4.4.14/drivers/staging/rdma/hfi1/
H A Dfile_ops.c208 const struct hfi1_cmd __user *ucmd; hfi1_file_write() local
228 ucmd = (const struct hfi1_cmd __user *)data; hfi1_file_write()
229 if (copy_from_user(&cmd, ucmd, sizeof(cmd))) { hfi1_file_write()
/linux-4.4.14/drivers/nvme/host/
H A Dpci.c1883 struct nvme_passthru_cmd __user *ucmd) nvme_user_cmd()
1892 if (copy_from_user(&cmd, ucmd, sizeof(cmd))) nvme_user_cmd()
1915 if (put_user(cmd.result, &ucmd->result)) nvme_user_cmd()
1882 nvme_user_cmd(struct nvme_dev *dev, struct nvme_ns *ns, struct nvme_passthru_cmd __user *ucmd) nvme_user_cmd() argument

Completed in 1382 milliseconds