Lines Matching refs:pd

156 struct ib_mr *ehca_get_dma_mr(struct ib_pd *pd, int mr_access_flags)  in ehca_get_dma_mr()  argument
161 struct ehca_pd *e_pd = container_of(pd, struct ehca_pd, ib_pd); in ehca_get_dma_mr()
163 container_of(pd->device, struct ehca_shca, ib_device); in ehca_get_dma_mr()
193 PTR_ERR(ib_mr), pd, mr_access_flags); in ehca_get_dma_mr()
199 struct ib_mr *ehca_reg_phys_mr(struct ib_pd *pd, in ehca_reg_phys_mr() argument
209 container_of(pd->device, struct ehca_shca, ib_device); in ehca_reg_phys_mr()
210 struct ehca_pd *e_pd = container_of(pd, struct ehca_pd, ib_pd); in ehca_reg_phys_mr()
215 ehca_err(pd->device, "bad input values: num_phys_buf=%x " in ehca_reg_phys_mr()
228 ehca_err(pd->device, "bad input values: mr_access_flags=%x", in ehca_reg_phys_mr()
243 ehca_err(pd->device, "bad input values: size=%llx iova_start=%p", in ehca_reg_phys_mr()
251 ehca_err(pd->device, "out of memory"); in ehca_reg_phys_mr()
304 ehca_err(pd->device, "h_ret=%li pd=%p phys_buf_array=%p " in ehca_reg_phys_mr()
306 PTR_ERR(ib_mr), pd, phys_buf_array, in ehca_reg_phys_mr()
313 struct ib_mr *ehca_reg_user_mr(struct ib_pd *pd, u64 start, u64 length, in ehca_reg_user_mr() argument
320 container_of(pd->device, struct ehca_shca, ib_device); in ehca_reg_user_mr()
321 struct ehca_pd *e_pd = container_of(pd, struct ehca_pd, ib_pd); in ehca_reg_user_mr()
328 if (!pd) { in ehca_reg_user_mr()
329 ehca_gen_err("bad pd=%p", pd); in ehca_reg_user_mr()
341 ehca_err(pd->device, "bad input values: mr_access_flags=%x", in ehca_reg_user_mr()
348 ehca_err(pd->device, "bad input values: length=%llx " in ehca_reg_user_mr()
356 ehca_err(pd->device, "out of memory"); in ehca_reg_user_mr()
361 e_mr->umem = ib_umem_get(pd->uobject->context, start, length, in ehca_reg_user_mr()
369 ehca_err(pd->device, "page size not supported, " in ehca_reg_user_mr()
408 ehca_warn(pd->device, "failed to register mr " in ehca_reg_user_mr()
410 ehca_info(pd->device, "try to register mr with " in ehca_reg_user_mr()
433 ehca_err(pd->device, "rc=%li pd=%p mr_access_flags=%x udata=%p", in ehca_reg_user_mr()
434 PTR_ERR(ib_mr), pd, mr_access_flags, udata); in ehca_reg_user_mr()
442 struct ib_pd *pd, in ehca_rereg_phys_mr() argument
472 if (!pd) { in ehca_rereg_phys_mr()
474 "mr_rereg_mask=%x", pd, mr_rereg_mask); in ehca_rereg_phys_mr()
531 new_pd = container_of(mr->pd, struct ehca_pd, ib_pd); in ehca_rereg_phys_mr()
567 new_pd = container_of(pd, struct ehca_pd, ib_pd); in ehca_rereg_phys_mr()
576 mr->pd = pd; in ehca_rereg_phys_mr()
587 ret, mr, mr_rereg_mask, pd, phys_buf_array, in ehca_rereg_phys_mr()
623 mr_attr->pd = mr->pd; in ehca_query_mr()
688 struct ib_mw *ehca_alloc_mw(struct ib_pd *pd, enum ib_mw_type type) in ehca_alloc_mw() argument
693 struct ehca_pd *e_pd = container_of(pd, struct ehca_pd, ib_pd); in ehca_alloc_mw()
695 container_of(pd->device, struct ehca_shca, ib_device); in ehca_alloc_mw()
710 ehca_err(pd->device, "hipz_mw_allocate failed, h_ret=%lli " in ehca_alloc_mw()
725 ehca_err(pd->device, "h_ret=%li pd=%p", PTR_ERR(ib_mw), pd); in ehca_alloc_mw()
765 struct ib_fmr *ehca_alloc_fmr(struct ib_pd *pd, in ehca_alloc_fmr() argument
771 container_of(pd->device, struct ehca_shca, ib_device); in ehca_alloc_fmr()
772 struct ehca_pd *e_pd = container_of(pd, struct ehca_pd, ib_pd); in ehca_alloc_fmr()
788 ehca_err(pd->device, "bad input values: mr_access_flags=%x", in ehca_alloc_fmr()
794 ehca_err(pd->device, "bad input values: mr_access_flags=%x", in ehca_alloc_fmr()
800 ehca_err(pd->device, "bad input values: fmr_attr->max_pages=%x " in ehca_alloc_fmr()
810 ehca_err(pd->device, "unsupported fmr_attr->page_shift=%x", in ehca_alloc_fmr()
864 struct ehca_pd *e_pd = container_of(fmr->pd, struct ehca_pd, ib_pd); in ehca_map_phys_fmr()
1389 container_of(e_fmr->ib.ib_fmr.pd, struct ehca_pd, ib_pd); in ehca_unmap_one_fmr()
1669 e_mr->ib.ib_mr.pd = &e_pd->ib_pd; in ehca_reg_internal_maxmr()
1742 ib_pd = e_maxmr->ib.ib_mr.pd; in ehca_dereg_internal_maxmr()