Lines Matching refs:vsgt

210 	struct vmw_sg_table vsgt;  member
292 void vmw_piter_start(struct vmw_piter *viter, const struct vmw_sg_table *vsgt, in vmw_piter_start() argument
296 viter->num_pages = vsgt->num_pages; in vmw_piter_start()
297 switch (vsgt->mode) { in vmw_piter_start()
302 viter->pages = vsgt->pages; in vmw_piter_start()
308 viter->addrs = vsgt->addrs; in vmw_piter_start()
309 viter->pages = vsgt->pages; in vmw_piter_start()
316 __sg_page_iter_start(&viter->iter, vsgt->sgt->sgl, in vmw_piter_start()
317 vsgt->sgt->orig_nents, p_offset); in vmw_piter_start()
383 struct vmw_sg_table *vsgt = &vmw_tt->vsgt; in vmw_ttm_map_dma() local
393 vsgt->mode = dev_priv->map_mode; in vmw_ttm_map_dma()
394 vsgt->pages = vmw_tt->dma_ttm.ttm.pages; in vmw_ttm_map_dma()
395 vsgt->num_pages = vmw_tt->dma_ttm.ttm.num_pages; in vmw_ttm_map_dma()
396 vsgt->addrs = vmw_tt->dma_ttm.dma_address; in vmw_ttm_map_dma()
397 vsgt->sgt = &vmw_tt->sgt; in vmw_ttm_map_dma()
406 vmw_tt->sg_alloc_size = sgt_size + sgl_size * vsgt->num_pages; in vmw_ttm_map_dma()
412 ret = sg_alloc_table_from_pages(&vmw_tt->sgt, vsgt->pages, in vmw_ttm_map_dma()
413 vsgt->num_pages, 0, in vmw_ttm_map_dma()
415 vsgt->num_pages << PAGE_SHIFT, in vmw_ttm_map_dma()
420 if (vsgt->num_pages > vmw_tt->sgt.nents) { in vmw_ttm_map_dma()
422 sgl_size * (vsgt->num_pages - in vmw_ttm_map_dma()
439 vmw_tt->vsgt.num_regions = 0; in vmw_ttm_map_dma()
440 for (vmw_piter_start(&iter, vsgt, 0); vmw_piter_next(&iter);) { in vmw_ttm_map_dma()
444 vmw_tt->vsgt.num_regions++; in vmw_ttm_map_dma()
452 sg_free_table(vmw_tt->vsgt.sgt); in vmw_ttm_map_dma()
453 vmw_tt->vsgt.sgt = NULL; in vmw_ttm_map_dma()
472 if (!vmw_tt->vsgt.sgt) in vmw_ttm_unmap_dma()
479 sg_free_table(vmw_tt->vsgt.sgt); in vmw_ttm_unmap_dma()
480 vmw_tt->vsgt.sgt = NULL; in vmw_ttm_unmap_dma()
543 return &vmw_tt->vsgt; in vmw_bo_sg_table()
562 return vmw_gmr_bind(vmw_be->dev_priv, &vmw_be->vsgt, in vmw_ttm_bind()
573 &vmw_be->vsgt, ttm->num_pages, in vmw_ttm_bind()