ggtt_mm 1743 drivers/gpu/drm/i915/gvt/cmd_parser.c s->vgpu->gtt.ggtt_mm : s->workload->shadow_mm; ggtt_mm 1826 drivers/gpu/drm/i915/gvt/cmd_parser.c s->vgpu->gtt.ggtt_mm : s->workload->shadow_mm; ggtt_mm 2890 drivers/gpu/drm/i915/gvt/cmd_parser.c ret = copy_gma_to_hva(vgpu, vgpu->gtt.ggtt_mm, ggtt_mm 2901 drivers/gpu/drm/i915/gvt/cmd_parser.c ret = copy_gma_to_hva(vgpu, vgpu->gtt.ggtt_mm, gma_head, gma_tail, ggtt_mm 2964 drivers/gpu/drm/i915/gvt/cmd_parser.c workload->vgpu->gtt.ggtt_mm, ggtt_mm 163 drivers/gpu/drm/i915/gvt/execlist.c hwsp_gpa = intel_vgpu_gma_to_gpa(vgpu->gtt.ggtt_mm, ggtt_mm 251 drivers/gpu/drm/i915/gvt/fb_decoder.c plane->base_gpa = intel_vgpu_gma_to_gpa(vgpu->gtt.ggtt_mm, plane->base); ggtt_mm 374 drivers/gpu/drm/i915/gvt/fb_decoder.c plane->base_gpa = intel_vgpu_gma_to_gpa(vgpu->gtt.ggtt_mm, plane->base); ggtt_mm 478 drivers/gpu/drm/i915/gvt/fb_decoder.c plane->base_gpa = intel_vgpu_gma_to_gpa(vgpu->gtt.ggtt_mm, plane->base); ggtt_mm 607 drivers/gpu/drm/i915/gvt/gtt.c pte_ops->get_entry(mm->ggtt_mm.virtual_ggtt, entry, index, ggtt_mm 618 drivers/gpu/drm/i915/gvt/gtt.c pte_ops->set_entry(mm->ggtt_mm.virtual_ggtt, entry, index, ggtt_mm 1932 drivers/gpu/drm/i915/gvt/gtt.c mm->ggtt_mm.virtual_ggtt = ggtt_mm 1935 drivers/gpu/drm/i915/gvt/gtt.c if (!mm->ggtt_mm.virtual_ggtt) { ggtt_mm 1966 drivers/gpu/drm/i915/gvt/gtt.c vfree(mm->ggtt_mm.virtual_ggtt); ggtt_mm 2145 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_mm *ggtt_mm = vgpu->gtt.ggtt_mm; ggtt_mm 2162 drivers/gpu/drm/i915/gvt/gtt.c ggtt_get_guest_entry(ggtt_mm, &e, index); ggtt_mm 2211 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_mm *ggtt_mm = vgpu->gtt.ggtt_mm; ggtt_mm 2243 drivers/gpu/drm/i915/gvt/gtt.c &ggtt_mm->ggtt_mm.partial_pte_list, list) { ggtt_mm 2263 drivers/gpu/drm/i915/gvt/gtt.c ggtt_set_guest_entry(ggtt_mm, &e, g_gtt_index); ggtt_mm 2276 drivers/gpu/drm/i915/gvt/gtt.c &ggtt_mm->ggtt_mm.partial_pte_list); ggtt_mm 2311 drivers/gpu/drm/i915/gvt/gtt.c ggtt_set_guest_entry(ggtt_mm, &e, g_gtt_index); ggtt_mm 2313 drivers/gpu/drm/i915/gvt/gtt.c ggtt_get_host_entry(ggtt_mm, &e, g_gtt_index); ggtt_mm 2316 drivers/gpu/drm/i915/gvt/gtt.c ggtt_set_host_entry(ggtt_mm, &m, g_gtt_index); ggtt_mm 2467 drivers/gpu/drm/i915/gvt/gtt.c gtt->ggtt_mm = intel_vgpu_create_ggtt_mm(vgpu); ggtt_mm 2468 drivers/gpu/drm/i915/gvt/gtt.c if (IS_ERR(gtt->ggtt_mm)) { ggtt_mm 2470 drivers/gpu/drm/i915/gvt/gtt.c return PTR_ERR(gtt->ggtt_mm); ggtt_mm 2475 drivers/gpu/drm/i915/gvt/gtt.c INIT_LIST_HEAD(>t->ggtt_mm->ggtt_mm.partial_pte_list); ggtt_mm 2504 drivers/gpu/drm/i915/gvt/gtt.c &vgpu->gtt.ggtt_mm->ggtt_mm.partial_pte_list, ggtt_mm 2510 drivers/gpu/drm/i915/gvt/gtt.c intel_vgpu_destroy_mm(vgpu->gtt.ggtt_mm); ggtt_mm 2511 drivers/gpu/drm/i915/gvt/gtt.c vgpu->gtt.ggtt_mm = NULL; ggtt_mm 2796 drivers/gpu/drm/i915/gvt/gtt.c ggtt_get_host_entry(vgpu->gtt.ggtt_mm, &old_entry, index); ggtt_mm 2799 drivers/gpu/drm/i915/gvt/gtt.c ggtt_set_host_entry(vgpu->gtt.ggtt_mm, &entry, index++); ggtt_mm 2806 drivers/gpu/drm/i915/gvt/gtt.c ggtt_get_host_entry(vgpu->gtt.ggtt_mm, &old_entry, index); ggtt_mm 2809 drivers/gpu/drm/i915/gvt/gtt.c ggtt_set_host_entry(vgpu->gtt.ggtt_mm, &entry, index++); ggtt_mm 167 drivers/gpu/drm/i915/gvt/gtt.h } ggtt_mm; ggtt_mm 199 drivers/gpu/drm/i915/gvt/gtt.h struct intel_vgpu_mm *ggtt_mm; ggtt_mm 82 drivers/gpu/drm/i915/gvt/mmio.c pt = vgpu->gtt.ggtt_mm->ggtt_mm.virtual_ggtt + offset; ggtt_mm 189 drivers/gpu/drm/i915/gvt/scheduler.c context_gpa = intel_vgpu_gma_to_gpa(vgpu->gtt.ggtt_mm, ggtt_mm 843 drivers/gpu/drm/i915/gvt/scheduler.c context_gpa = intel_vgpu_gma_to_gpa(vgpu->gtt.ggtt_mm, ggtt_mm 1489 drivers/gpu/drm/i915/gvt/scheduler.c ring_context_gpa = intel_vgpu_gma_to_gpa(vgpu->gtt.ggtt_mm,