Lines Matching refs:vsgt
223 struct vmw_sg_table vsgt; member
305 void vmw_piter_start(struct vmw_piter *viter, const struct vmw_sg_table *vsgt, in vmw_piter_start() argument
309 viter->num_pages = vsgt->num_pages; in vmw_piter_start()
310 switch (vsgt->mode) { in vmw_piter_start()
315 viter->pages = vsgt->pages; in vmw_piter_start()
321 viter->addrs = vsgt->addrs; in vmw_piter_start()
322 viter->pages = vsgt->pages; in vmw_piter_start()
329 __sg_page_iter_start(&viter->iter, vsgt->sgt->sgl, in vmw_piter_start()
330 vsgt->sgt->orig_nents, p_offset); in vmw_piter_start()
396 struct vmw_sg_table *vsgt = &vmw_tt->vsgt; in vmw_ttm_map_dma() local
406 vsgt->mode = dev_priv->map_mode; in vmw_ttm_map_dma()
407 vsgt->pages = vmw_tt->dma_ttm.ttm.pages; in vmw_ttm_map_dma()
408 vsgt->num_pages = vmw_tt->dma_ttm.ttm.num_pages; in vmw_ttm_map_dma()
409 vsgt->addrs = vmw_tt->dma_ttm.dma_address; in vmw_ttm_map_dma()
410 vsgt->sgt = &vmw_tt->sgt; in vmw_ttm_map_dma()
419 vmw_tt->sg_alloc_size = sgt_size + sgl_size * vsgt->num_pages; in vmw_ttm_map_dma()
425 ret = sg_alloc_table_from_pages(&vmw_tt->sgt, vsgt->pages, in vmw_ttm_map_dma()
426 vsgt->num_pages, 0, in vmw_ttm_map_dma()
428 vsgt->num_pages << PAGE_SHIFT, in vmw_ttm_map_dma()
433 if (vsgt->num_pages > vmw_tt->sgt.nents) { in vmw_ttm_map_dma()
435 sgl_size * (vsgt->num_pages - in vmw_ttm_map_dma()
452 vmw_tt->vsgt.num_regions = 0; in vmw_ttm_map_dma()
453 for (vmw_piter_start(&iter, vsgt, 0); vmw_piter_next(&iter);) { in vmw_ttm_map_dma()
457 vmw_tt->vsgt.num_regions++; in vmw_ttm_map_dma()
465 sg_free_table(vmw_tt->vsgt.sgt); in vmw_ttm_map_dma()
466 vmw_tt->vsgt.sgt = NULL; in vmw_ttm_map_dma()
485 if (!vmw_tt->vsgt.sgt) in vmw_ttm_unmap_dma()
492 sg_free_table(vmw_tt->vsgt.sgt); in vmw_ttm_unmap_dma()
493 vmw_tt->vsgt.sgt = NULL; in vmw_ttm_unmap_dma()
556 return &vmw_tt->vsgt; in vmw_bo_sg_table()
575 return vmw_gmr_bind(vmw_be->dev_priv, &vmw_be->vsgt, in vmw_ttm_bind()
586 &vmw_be->vsgt, ttm->num_pages, in vmw_ttm_bind()