Lines Matching refs:ctx_obj

273 u32 intel_execlists_ctx_id(struct drm_i915_gem_object *ctx_obj)  in intel_execlists_ctx_id()  argument
275 u32 lrca = i915_gem_obj_ggtt_offset(ctx_obj) + in intel_execlists_ctx_id()
295 struct drm_i915_gem_object *ctx_obj = ctx->engine[ring->id].state; in intel_lr_context_descriptor() local
297 uint64_t lrca = i915_gem_obj_ggtt_offset(ctx_obj) + in intel_lr_context_descriptor()
304 if (IS_GEN8(ctx_obj->base.dev)) in intel_lr_context_descriptor()
308 desc |= (u64)intel_execlists_ctx_id(ctx_obj) << GEN8_CTX_ID_SHIFT; in intel_lr_context_descriptor()
361 struct drm_i915_gem_object *ctx_obj = rq->ctx->engine[ring->id].state; in execlists_update_context() local
366 BUG_ON(!ctx_obj); in execlists_update_context()
367 WARN_ON(!i915_gem_obj_is_pinned(ctx_obj)); in execlists_update_context()
370 page = i915_gem_object_get_page(ctx_obj, LRC_STATE_PN); in execlists_update_context()
476 struct drm_i915_gem_object *ctx_obj = in execlists_check_remove_request() local
478 if (intel_execlists_ctx_id(ctx_obj) == request_id) { in execlists_check_remove_request()
962 struct drm_i915_gem_object *ctx_obj = in intel_execlists_retire_requests() local
965 if (ctx_obj && (ctx != ring->default_context)) in intel_execlists_retire_requests()
1011 struct drm_i915_gem_object *ctx_obj, in intel_lr_context_do_pin() argument
1019 ret = i915_gem_obj_ggtt_pin(ctx_obj, GEN8_LR_CONTEXT_ALIGN, in intel_lr_context_do_pin()
1028 ctx_obj->dirty = true; in intel_lr_context_do_pin()
1037 i915_gem_object_ggtt_unpin(ctx_obj); in intel_lr_context_do_pin()
1046 struct drm_i915_gem_object *ctx_obj = rq->ctx->engine[ring->id].state; in intel_lr_context_pin() local
1050 ret = intel_lr_context_do_pin(ring, ctx_obj, ringbuf); in intel_lr_context_pin()
1064 struct drm_i915_gem_object *ctx_obj = rq->ctx->engine[ring->id].state; in intel_lr_context_unpin() local
1067 if (ctx_obj) { in intel_lr_context_unpin()
1071 i915_gem_object_ggtt_unpin(ctx_obj); in intel_lr_context_unpin()
2230 populate_lr_context(struct intel_context *ctx, struct drm_i915_gem_object *ctx_obj, in populate_lr_context() argument
2243 ret = i915_gem_object_set_to_cpu_domain(ctx_obj, true); in populate_lr_context()
2249 ret = i915_gem_object_get_pages(ctx_obj); in populate_lr_context()
2255 i915_gem_object_pin_pages(ctx_obj); in populate_lr_context()
2259 page = i915_gem_object_get_page(ctx_obj, LRC_STATE_PN); in populate_lr_context()
2362 ctx_obj->dirty = 1; in populate_lr_context()
2364 i915_gem_object_unpin_pages(ctx_obj); in populate_lr_context()
2382 struct drm_i915_gem_object *ctx_obj = ctx->engine[i].state; in intel_lr_context_free() local
2384 if (ctx_obj) { in intel_lr_context_free()
2391 i915_gem_object_ggtt_unpin(ctx_obj); in intel_lr_context_free()
2395 drm_gem_object_unreference(&ctx_obj->base); in intel_lr_context_free()
2460 struct drm_i915_gem_object *ctx_obj; in intel_lr_context_deferred_alloc() local
2473 ctx_obj = i915_gem_alloc_object(dev, context_size); in intel_lr_context_deferred_alloc()
2474 if (!ctx_obj) { in intel_lr_context_deferred_alloc()
2485 ret = populate_lr_context(ctx, ctx_obj, ring, ringbuf); in intel_lr_context_deferred_alloc()
2492 ctx->engine[ring->id].state = ctx_obj; in intel_lr_context_deferred_alloc()
2519 drm_gem_object_unreference(&ctx_obj->base); in intel_lr_context_deferred_alloc()
2533 struct drm_i915_gem_object *ctx_obj = in intel_lr_context_reset() local
2540 if (!ctx_obj) in intel_lr_context_reset()
2543 if (i915_gem_object_get_pages(ctx_obj)) { in intel_lr_context_reset()
2547 page = i915_gem_object_get_page(ctx_obj, LRC_STATE_PN); in intel_lr_context_reset()