ufb 32 drivers/gpu/drm/udl/udl_fb.c struct udl_framebuffer ufb; ufb 208 drivers/gpu/drm/udl/udl_fb.c struct drm_device *dev = ufbdev->ufb.base.dev; ufb 283 drivers/gpu/drm/udl/udl_fb.c struct udl_framebuffer *ufb = to_udl_fb(fb); ufb 289 drivers/gpu/drm/udl/udl_fb.c if (!ufb->active_16) ufb 292 drivers/gpu/drm/udl/udl_fb.c if (ufb->obj->base.import_attach) { ufb 293 drivers/gpu/drm/udl/udl_fb.c ret = dma_buf_begin_cpu_access(ufb->obj->base.import_attach->dmabuf, ufb 300 drivers/gpu/drm/udl/udl_fb.c ret = udl_handle_damage(ufb, clips[i].x1, clips[i].y1, ufb 307 drivers/gpu/drm/udl/udl_fb.c if (ufb->obj->base.import_attach) { ufb 308 drivers/gpu/drm/udl/udl_fb.c ret = dma_buf_end_cpu_access(ufb->obj->base.import_attach->dmabuf, ufb 320 drivers/gpu/drm/udl/udl_fb.c struct udl_framebuffer *ufb = to_udl_fb(fb); ufb 322 drivers/gpu/drm/udl/udl_fb.c if (ufb->obj) ufb 323 drivers/gpu/drm/udl/udl_fb.c drm_gem_object_put_unlocked(&ufb->obj->base); ufb 326 drivers/gpu/drm/udl/udl_fb.c kfree(ufb); ufb 337 drivers/gpu/drm/udl/udl_fb.c struct udl_framebuffer *ufb, ufb 343 drivers/gpu/drm/udl/udl_fb.c ufb->obj = obj; ufb 344 drivers/gpu/drm/udl/udl_fb.c drm_helper_mode_fill_fb_struct(dev, &ufb->base, mode_cmd); ufb 345 drivers/gpu/drm/udl/udl_fb.c ret = drm_framebuffer_init(dev, &ufb->base, &udlfb_funcs); ufb 392 drivers/gpu/drm/udl/udl_fb.c ret = udl_framebuffer_init(dev, &ufbdev->ufb, &mode_cmd, obj); ufb 396 drivers/gpu/drm/udl/udl_fb.c fb = &ufbdev->ufb.base; ufb 400 drivers/gpu/drm/udl/udl_fb.c info->screen_base = ufbdev->ufb.obj->vmapping; ufb 402 drivers/gpu/drm/udl/udl_fb.c info->fix.smem_start = (unsigned long)ufbdev->ufb.obj->vmapping; ufb 409 drivers/gpu/drm/udl/udl_fb.c ufbdev->ufb.obj->vmapping); ufb 413 drivers/gpu/drm/udl/udl_fb.c drm_gem_object_put_unlocked(&ufbdev->ufb.obj->base); ufb 427 drivers/gpu/drm/udl/udl_fb.c if (ufbdev->ufb.obj) { ufb 428 drivers/gpu/drm/udl/udl_fb.c drm_framebuffer_unregister_private(&ufbdev->ufb.base); ufb 429 drivers/gpu/drm/udl/udl_fb.c drm_framebuffer_cleanup(&ufbdev->ufb.base); ufb 430 drivers/gpu/drm/udl/udl_fb.c drm_gem_object_put_unlocked(&ufbdev->ufb.obj->base); ufb 501 drivers/gpu/drm/udl/udl_fb.c struct udl_framebuffer *ufb; ufb 517 drivers/gpu/drm/udl/udl_fb.c ufb = kzalloc(sizeof(*ufb), GFP_KERNEL); ufb 518 drivers/gpu/drm/udl/udl_fb.c if (ufb == NULL) ufb 521 drivers/gpu/drm/udl/udl_fb.c ret = udl_framebuffer_init(dev, ufb, mode_cmd, to_udl_bo(obj)); ufb 523 drivers/gpu/drm/udl/udl_fb.c kfree(ufb); ufb 526 drivers/gpu/drm/udl/udl_fb.c return &ufb->base; ufb 304 drivers/gpu/drm/udl/udl_modeset.c struct udl_framebuffer *ufb = to_udl_fb(crtc->primary->fb); ufb 339 drivers/gpu/drm/udl/udl_modeset.c ufb->active_16 = true; ufb 343 drivers/gpu/drm/udl/udl_modeset.c udl_handle_damage(ufb, 0, 0, ufb->base.width, ufb->base.height); ufb 365 drivers/gpu/drm/udl/udl_modeset.c struct udl_framebuffer *ufb = to_udl_fb(fb); ufb 373 drivers/gpu/drm/udl/udl_modeset.c ufb->active_16 = true; ufb 375 drivers/gpu/drm/udl/udl_modeset.c udl_handle_damage(ufb, 0, 0, fb->width, fb->height); ufb 456 drivers/gpu/drm/udl/udl_modeset.c struct udl_framebuffer *ufb; ufb 461 drivers/gpu/drm/udl/udl_modeset.c ufb = to_udl_fb(udl->crtc->primary->fb); ufb 462 drivers/gpu/drm/udl/udl_modeset.c udl_handle_damage(ufb, 0, 0, ufb->base.width, ufb->base.height);