Lines Matching refs:ctx_obj

248 u32 intel_execlists_ctx_id(struct drm_i915_gem_object *ctx_obj)  in intel_execlists_ctx_id()  argument
250 u32 lrca = i915_gem_obj_ggtt_offset(ctx_obj); in intel_execlists_ctx_id()
258 struct drm_i915_gem_object *ctx_obj) in execlists_ctx_descriptor() argument
262 uint64_t lrca = i915_gem_obj_ggtt_offset(ctx_obj); in execlists_ctx_descriptor()
271 desc |= (u64)intel_execlists_ctx_id(ctx_obj) << GEN8_CTX_ID_SHIFT; in execlists_ctx_descriptor()
321 static int execlists_update_context(struct drm_i915_gem_object *ctx_obj, in execlists_update_context() argument
328 page = i915_gem_object_get_page(ctx_obj, 1); in execlists_update_context()
439 struct drm_i915_gem_object *ctx_obj = in execlists_check_remove_request() local
441 if (intel_execlists_ctx_id(ctx_obj) == request_id) { in execlists_check_remove_request()
760 struct drm_i915_gem_object *ctx_obj = in intel_execlists_retire_requests() local
763 if (ctx_obj && (ctx != ring->default_context)) in intel_execlists_retire_requests()
837 struct drm_i915_gem_object *ctx_obj = ctx->engine[ring->id].state; in intel_lr_context_pin() local
843 ret = i915_gem_obj_ggtt_pin(ctx_obj, in intel_lr_context_pin()
852 ctx_obj->dirty = true; in intel_lr_context_pin()
858 i915_gem_object_ggtt_unpin(ctx_obj); in intel_lr_context_pin()
868 struct drm_i915_gem_object *ctx_obj = ctx->engine[ring->id].state; in intel_lr_context_unpin() local
871 if (ctx_obj) { in intel_lr_context_unpin()
875 i915_gem_object_ggtt_unpin(ctx_obj); in intel_lr_context_unpin()
1726 populate_lr_context(struct intel_context *ctx, struct drm_i915_gem_object *ctx_obj, in populate_lr_context() argument
1739 ret = i915_gem_object_set_to_cpu_domain(ctx_obj, true); in populate_lr_context()
1745 ret = i915_gem_object_get_pages(ctx_obj); in populate_lr_context()
1751 i915_gem_object_pin_pages(ctx_obj); in populate_lr_context()
1755 page = i915_gem_object_get_page(ctx_obj, 1); in populate_lr_context()
1834 ctx_obj->dirty = 1; in populate_lr_context()
1836 i915_gem_object_unpin_pages(ctx_obj); in populate_lr_context()
1854 struct drm_i915_gem_object *ctx_obj = ctx->engine[i].state; in intel_lr_context_free() local
1856 if (ctx_obj) { in intel_lr_context_free()
1863 i915_gem_object_ggtt_unpin(ctx_obj); in intel_lr_context_free()
1868 drm_gem_object_unreference(&ctx_obj->base); in intel_lr_context_free()
1932 struct drm_i915_gem_object *ctx_obj; in intel_lr_context_deferred_create() local
1942 ctx_obj = i915_gem_alloc_context_obj(dev, context_size); in intel_lr_context_deferred_create()
1943 if (IS_ERR(ctx_obj)) { in intel_lr_context_deferred_create()
1944 ret = PTR_ERR(ctx_obj); in intel_lr_context_deferred_create()
1950 ret = i915_gem_obj_ggtt_pin(ctx_obj, GEN8_LR_CONTEXT_ALIGN, 0); in intel_lr_context_deferred_create()
1954 drm_gem_object_unreference(&ctx_obj->base); in intel_lr_context_deferred_create()
1997 ret = populate_lr_context(ctx, ctx_obj, ring, ringbuf); in intel_lr_context_deferred_create()
2004 ctx->engine[ring->id].state = ctx_obj; in intel_lr_context_deferred_create()
2007 lrc_setup_hardware_status_page(ring, ctx_obj); in intel_lr_context_deferred_create()
2033 i915_gem_object_ggtt_unpin(ctx_obj); in intel_lr_context_deferred_create()
2034 drm_gem_object_unreference(&ctx_obj->base); in intel_lr_context_deferred_create()
2046 struct drm_i915_gem_object *ctx_obj = in intel_lr_context_reset() local
2053 if (!ctx_obj) in intel_lr_context_reset()
2056 if (i915_gem_object_get_pages(ctx_obj)) { in intel_lr_context_reset()
2060 page = i915_gem_object_get_page(ctx_obj, 1); in intel_lr_context_reset()