blend_tf 1782 drivers/gpu/drm/amd/display/dc/core/dc.c if (srf_update->blend_tf && blend_tf 1783 drivers/gpu/drm/amd/display/dc/core/dc.c (surface->blend_tf != blend_tf 1784 drivers/gpu/drm/amd/display/dc/core/dc.c srf_update->blend_tf)) blend_tf 1785 drivers/gpu/drm/amd/display/dc/core/dc.c memcpy(surface->blend_tf, srf_update->blend_tf, blend_tf 1786 drivers/gpu/drm/amd/display/dc/core/dc.c sizeof(*surface->blend_tf)); blend_tf 64 drivers/gpu/drm/amd/display/dc/core/dc_surface.c plane_state->blend_tf = dc_create_transfer_func(); blend_tf 65 drivers/gpu/drm/amd/display/dc/core/dc_surface.c if (plane_state->blend_tf != NULL) { blend_tf 66 drivers/gpu/drm/amd/display/dc/core/dc_surface.c plane_state->blend_tf->type = TF_TYPE_BYPASS; blend_tf 67 drivers/gpu/drm/amd/display/dc/core/dc_surface.c plane_state->blend_tf->ctx = ctx; blend_tf 94 drivers/gpu/drm/amd/display/dc/core/dc_surface.c if (plane_state->blend_tf != NULL) { blend_tf 96 drivers/gpu/drm/amd/display/dc/core/dc_surface.c plane_state->blend_tf); blend_tf 97 drivers/gpu/drm/amd/display/dc/core/dc_surface.c plane_state->blend_tf = NULL; blend_tf 743 drivers/gpu/drm/amd/display/dc/dc.h struct dc_transfer_func *blend_tf; blend_tf 816 drivers/gpu/drm/amd/display/dc/dc.h const struct dc_transfer_func *blend_tf; blend_tf 697 drivers/gpu/drm/amd/display/dc/dcn20/dcn20_hwseq.c if (plane_state->blend_tf) { blend_tf 698 drivers/gpu/drm/amd/display/dc/dcn20/dcn20_hwseq.c if (plane_state->blend_tf->type == TF_TYPE_HWPWL) blend_tf 699 drivers/gpu/drm/amd/display/dc/dcn20/dcn20_hwseq.c blend_lut = &plane_state->blend_tf->pwl; blend_tf 700 drivers/gpu/drm/amd/display/dc/dcn20/dcn20_hwseq.c else if (plane_state->blend_tf->type == TF_TYPE_DISTRIBUTED_POINTS) { blend_tf 702 drivers/gpu/drm/amd/display/dc/dcn20/dcn20_hwseq.c plane_state->blend_tf,