1/* 2 * Copyright 2012 Red Hat 3 * 4 * This file is subject to the terms and conditions of the GNU General 5 * Public License version 2. See the file COPYING in the main 6 * directory of this archive for more details. 7 * 8 * Authors: Matthew Garrett 9 * Dave Airlie 10 */ 11#include <drm/drmP.h> 12#include <drm/drm_crtc_helper.h> 13 14#include "cirrus_drv.h" 15 16 17static void cirrus_user_framebuffer_destroy(struct drm_framebuffer *fb) 18{ 19 struct cirrus_framebuffer *cirrus_fb = to_cirrus_framebuffer(fb); 20 if (cirrus_fb->obj) 21 drm_gem_object_unreference_unlocked(cirrus_fb->obj); 22 drm_framebuffer_cleanup(fb); 23 kfree(fb); 24} 25 26static const struct drm_framebuffer_funcs cirrus_fb_funcs = { 27 .destroy = cirrus_user_framebuffer_destroy, 28}; 29 30int cirrus_framebuffer_init(struct drm_device *dev, 31 struct cirrus_framebuffer *gfb, 32 struct drm_mode_fb_cmd2 *mode_cmd, 33 struct drm_gem_object *obj) 34{ 35 int ret; 36 37 drm_helper_mode_fill_fb_struct(&gfb->base, mode_cmd); 38 gfb->obj = obj; 39 ret = drm_framebuffer_init(dev, &gfb->base, &cirrus_fb_funcs); 40 if (ret) { 41 DRM_ERROR("drm_framebuffer_init failed: %d\n", ret); 42 return ret; 43 } 44 return 0; 45} 46 47static struct drm_framebuffer * 48cirrus_user_framebuffer_create(struct drm_device *dev, 49 struct drm_file *filp, 50 struct drm_mode_fb_cmd2 *mode_cmd) 51{ 52 struct cirrus_device *cdev = dev->dev_private; 53 struct drm_gem_object *obj; 54 struct cirrus_framebuffer *cirrus_fb; 55 int ret; 56 u32 bpp, depth; 57 58 drm_fb_get_bpp_depth(mode_cmd->pixel_format, &depth, &bpp); 59 60 if (!cirrus_check_framebuffer(cdev, mode_cmd->width, mode_cmd->height, 61 bpp, mode_cmd->pitches[0])) 62 return ERR_PTR(-EINVAL); 63 64 obj = drm_gem_object_lookup(dev, filp, mode_cmd->handles[0]); 65 if (obj == NULL) 66 return ERR_PTR(-ENOENT); 67 68 cirrus_fb = kzalloc(sizeof(*cirrus_fb), GFP_KERNEL); 69 if (!cirrus_fb) { 70 drm_gem_object_unreference_unlocked(obj); 71 return ERR_PTR(-ENOMEM); 72 } 73 74 ret = cirrus_framebuffer_init(dev, cirrus_fb, mode_cmd, obj); 75 if (ret) { 76 drm_gem_object_unreference_unlocked(obj); 77 kfree(cirrus_fb); 78 return ERR_PTR(ret); 79 } 80 return &cirrus_fb->base; 81} 82 83static const struct drm_mode_config_funcs cirrus_mode_funcs = { 84 .fb_create = cirrus_user_framebuffer_create, 85}; 86 87/* Unmap the framebuffer from the core and release the memory */ 88static void cirrus_vram_fini(struct cirrus_device *cdev) 89{ 90 iounmap(cdev->rmmio); 91 cdev->rmmio = NULL; 92 if (cdev->mc.vram_base) 93 release_mem_region(cdev->mc.vram_base, cdev->mc.vram_size); 94} 95 96/* Map the framebuffer from the card and configure the core */ 97static int cirrus_vram_init(struct cirrus_device *cdev) 98{ 99 /* BAR 0 is VRAM */ 100 cdev->mc.vram_base = pci_resource_start(cdev->dev->pdev, 0); 101 cdev->mc.vram_size = pci_resource_len(cdev->dev->pdev, 0); 102 103 if (!request_mem_region(cdev->mc.vram_base, cdev->mc.vram_size, 104 "cirrusdrmfb_vram")) { 105 DRM_ERROR("can't reserve VRAM\n"); 106 return -ENXIO; 107 } 108 109 return 0; 110} 111 112/* 113 * Our emulated hardware has two sets of memory. One is video RAM and can 114 * simply be used as a linear framebuffer - the other provides mmio access 115 * to the display registers. The latter can also be accessed via IO port 116 * access, but we map the range and use mmio to program them instead 117 */ 118 119int cirrus_device_init(struct cirrus_device *cdev, 120 struct drm_device *ddev, 121 struct pci_dev *pdev, uint32_t flags) 122{ 123 int ret; 124 125 cdev->dev = ddev; 126 cdev->flags = flags; 127 128 /* Hardcode the number of CRTCs to 1 */ 129 cdev->num_crtc = 1; 130 131 /* BAR 0 is the framebuffer, BAR 1 contains registers */ 132 cdev->rmmio_base = pci_resource_start(cdev->dev->pdev, 1); 133 cdev->rmmio_size = pci_resource_len(cdev->dev->pdev, 1); 134 135 if (!request_mem_region(cdev->rmmio_base, cdev->rmmio_size, 136 "cirrusdrmfb_mmio")) { 137 DRM_ERROR("can't reserve mmio registers\n"); 138 return -ENOMEM; 139 } 140 141 cdev->rmmio = ioremap(cdev->rmmio_base, cdev->rmmio_size); 142 143 if (cdev->rmmio == NULL) 144 return -ENOMEM; 145 146 ret = cirrus_vram_init(cdev); 147 if (ret) { 148 release_mem_region(cdev->rmmio_base, cdev->rmmio_size); 149 return ret; 150 } 151 152 return 0; 153} 154 155void cirrus_device_fini(struct cirrus_device *cdev) 156{ 157 release_mem_region(cdev->rmmio_base, cdev->rmmio_size); 158 cirrus_vram_fini(cdev); 159} 160 161/* 162 * Functions here will be called by the core once it's bound the driver to 163 * a PCI device 164 */ 165 166int cirrus_driver_load(struct drm_device *dev, unsigned long flags) 167{ 168 struct cirrus_device *cdev; 169 int r; 170 171 cdev = kzalloc(sizeof(struct cirrus_device), GFP_KERNEL); 172 if (cdev == NULL) 173 return -ENOMEM; 174 dev->dev_private = (void *)cdev; 175 176 r = cirrus_device_init(cdev, dev, dev->pdev, flags); 177 if (r) { 178 dev_err(&dev->pdev->dev, "Fatal error during GPU init: %d\n", r); 179 goto out; 180 } 181 182 r = cirrus_mm_init(cdev); 183 if (r) { 184 dev_err(&dev->pdev->dev, "fatal err on mm init\n"); 185 goto out; 186 } 187 188 r = cirrus_modeset_init(cdev); 189 if (r) { 190 dev_err(&dev->pdev->dev, "Fatal error during modeset init: %d\n", r); 191 goto out; 192 } 193 194 dev->mode_config.funcs = (void *)&cirrus_mode_funcs; 195 196 return 0; 197out: 198 cirrus_driver_unload(dev); 199 return r; 200} 201 202int cirrus_driver_unload(struct drm_device *dev) 203{ 204 struct cirrus_device *cdev = dev->dev_private; 205 206 if (cdev == NULL) 207 return 0; 208 cirrus_modeset_fini(cdev); 209 cirrus_mm_fini(cdev); 210 cirrus_device_fini(cdev); 211 kfree(cdev); 212 dev->dev_private = NULL; 213 return 0; 214} 215 216int cirrus_gem_create(struct drm_device *dev, 217 u32 size, bool iskernel, 218 struct drm_gem_object **obj) 219{ 220 struct cirrus_bo *cirrusbo; 221 int ret; 222 223 *obj = NULL; 224 225 size = roundup(size, PAGE_SIZE); 226 if (size == 0) 227 return -EINVAL; 228 229 ret = cirrus_bo_create(dev, size, 0, 0, &cirrusbo); 230 if (ret) { 231 if (ret != -ERESTARTSYS) 232 DRM_ERROR("failed to allocate GEM object\n"); 233 return ret; 234 } 235 *obj = &cirrusbo->gem; 236 return 0; 237} 238 239int cirrus_dumb_create(struct drm_file *file, 240 struct drm_device *dev, 241 struct drm_mode_create_dumb *args) 242{ 243 int ret; 244 struct drm_gem_object *gobj; 245 u32 handle; 246 247 args->pitch = args->width * ((args->bpp + 7) / 8); 248 args->size = args->pitch * args->height; 249 250 ret = cirrus_gem_create(dev, args->size, false, 251 &gobj); 252 if (ret) 253 return ret; 254 255 ret = drm_gem_handle_create(file, gobj, &handle); 256 drm_gem_object_unreference_unlocked(gobj); 257 if (ret) 258 return ret; 259 260 args->handle = handle; 261 return 0; 262} 263 264static void cirrus_bo_unref(struct cirrus_bo **bo) 265{ 266 struct ttm_buffer_object *tbo; 267 268 if ((*bo) == NULL) 269 return; 270 271 tbo = &((*bo)->bo); 272 ttm_bo_unref(&tbo); 273 *bo = NULL; 274} 275 276void cirrus_gem_free_object(struct drm_gem_object *obj) 277{ 278 struct cirrus_bo *cirrus_bo = gem_to_cirrus_bo(obj); 279 280 cirrus_bo_unref(&cirrus_bo); 281} 282 283 284static inline u64 cirrus_bo_mmap_offset(struct cirrus_bo *bo) 285{ 286 return drm_vma_node_offset_addr(&bo->bo.vma_node); 287} 288 289int 290cirrus_dumb_mmap_offset(struct drm_file *file, 291 struct drm_device *dev, 292 uint32_t handle, 293 uint64_t *offset) 294{ 295 struct drm_gem_object *obj; 296 int ret; 297 struct cirrus_bo *bo; 298 299 mutex_lock(&dev->struct_mutex); 300 obj = drm_gem_object_lookup(dev, file, handle); 301 if (obj == NULL) { 302 ret = -ENOENT; 303 goto out_unlock; 304 } 305 306 bo = gem_to_cirrus_bo(obj); 307 *offset = cirrus_bo_mmap_offset(bo); 308 309 drm_gem_object_unreference(obj); 310 ret = 0; 311out_unlock: 312 mutex_unlock(&dev->struct_mutex); 313 return ret; 314 315} 316 317bool cirrus_check_framebuffer(struct cirrus_device *cdev, int width, int height, 318 int bpp, int pitch) 319{ 320 const int max_pitch = 0x1FF << 3; /* (4096 - 1) & ~111b bytes */ 321 const int max_size = cdev->mc.vram_size; 322 323 if (bpp > cirrus_bpp) 324 return false; 325 if (bpp > 32) 326 return false; 327 328 if (pitch > max_pitch) 329 return false; 330 331 if (pitch * height > max_size) 332 return false; 333 334 return true; 335} 336