blend_tf         1782 drivers/gpu/drm/amd/display/dc/core/dc.c 	if (srf_update->blend_tf &&
blend_tf         1783 drivers/gpu/drm/amd/display/dc/core/dc.c 			(surface->blend_tf !=
blend_tf         1784 drivers/gpu/drm/amd/display/dc/core/dc.c 			srf_update->blend_tf))
blend_tf         1785 drivers/gpu/drm/amd/display/dc/core/dc.c 		memcpy(surface->blend_tf, srf_update->blend_tf,
blend_tf         1786 drivers/gpu/drm/amd/display/dc/core/dc.c 		sizeof(*surface->blend_tf));
blend_tf           64 drivers/gpu/drm/amd/display/dc/core/dc_surface.c 	plane_state->blend_tf = dc_create_transfer_func();
blend_tf           65 drivers/gpu/drm/amd/display/dc/core/dc_surface.c 	if (plane_state->blend_tf != NULL) {
blend_tf           66 drivers/gpu/drm/amd/display/dc/core/dc_surface.c 		plane_state->blend_tf->type = TF_TYPE_BYPASS;
blend_tf           67 drivers/gpu/drm/amd/display/dc/core/dc_surface.c 		plane_state->blend_tf->ctx = ctx;
blend_tf           94 drivers/gpu/drm/amd/display/dc/core/dc_surface.c 	if (plane_state->blend_tf != NULL) {
blend_tf           96 drivers/gpu/drm/amd/display/dc/core/dc_surface.c 				plane_state->blend_tf);
blend_tf           97 drivers/gpu/drm/amd/display/dc/core/dc_surface.c 		plane_state->blend_tf = NULL;
blend_tf          743 drivers/gpu/drm/amd/display/dc/dc.h 	struct dc_transfer_func *blend_tf;
blend_tf          816 drivers/gpu/drm/amd/display/dc/dc.h 	const struct dc_transfer_func *blend_tf;
blend_tf          697 drivers/gpu/drm/amd/display/dc/dcn20/dcn20_hwseq.c 	if (plane_state->blend_tf) {
blend_tf          698 drivers/gpu/drm/amd/display/dc/dcn20/dcn20_hwseq.c 		if (plane_state->blend_tf->type == TF_TYPE_HWPWL)
blend_tf          699 drivers/gpu/drm/amd/display/dc/dcn20/dcn20_hwseq.c 			blend_lut = &plane_state->blend_tf->pwl;
blend_tf          700 drivers/gpu/drm/amd/display/dc/dcn20/dcn20_hwseq.c 		else if (plane_state->blend_tf->type == TF_TYPE_DISTRIBUTED_POINTS) {
blend_tf          702 drivers/gpu/drm/amd/display/dc/dcn20/dcn20_hwseq.c 					plane_state->blend_tf,