Lines Matching refs:bo

32 	struct qxl_bo *bo;  in qxl_ttm_bo_destroy()  local
35 bo = container_of(tbo, struct qxl_bo, tbo); in qxl_ttm_bo_destroy()
36 qdev = (struct qxl_device *)bo->gem_base.dev->dev_private; in qxl_ttm_bo_destroy()
38 qxl_surface_evict(qdev, bo, false); in qxl_ttm_bo_destroy()
40 list_del_init(&bo->list); in qxl_ttm_bo_destroy()
42 drm_gem_object_release(&bo->gem_base); in qxl_ttm_bo_destroy()
43 kfree(bo); in qxl_ttm_bo_destroy()
46 bool qxl_ttm_bo_is_qxl_bo(struct ttm_buffer_object *bo) in qxl_ttm_bo_is_qxl_bo() argument
48 if (bo->destroy == &qxl_ttm_bo_destroy) in qxl_ttm_bo_is_qxl_bo()
83 struct qxl_bo *bo; in qxl_bo_create() local
92 bo = kzalloc(sizeof(struct qxl_bo), GFP_KERNEL); in qxl_bo_create()
93 if (bo == NULL) in qxl_bo_create()
96 r = drm_gem_object_init(qdev->ddev, &bo->gem_base, size); in qxl_bo_create()
98 kfree(bo); in qxl_bo_create()
101 bo->type = domain; in qxl_bo_create()
102 bo->pin_count = pinned ? 1 : 0; in qxl_bo_create()
103 bo->surface_id = 0; in qxl_bo_create()
104 INIT_LIST_HEAD(&bo->list); in qxl_bo_create()
107 bo->surf = *surf; in qxl_bo_create()
109 qxl_ttm_placement_from_domain(bo, domain, pinned); in qxl_bo_create()
111 r = ttm_bo_init(&qdev->mman.bdev, &bo->tbo, size, type, in qxl_bo_create()
112 &bo->placement, 0, !kernel, NULL, size, in qxl_bo_create()
121 *bo_ptr = bo; in qxl_bo_create()
125 int qxl_bo_kmap(struct qxl_bo *bo, void **ptr) in qxl_bo_kmap() argument
130 if (bo->kptr) { in qxl_bo_kmap()
132 *ptr = bo->kptr; in qxl_bo_kmap()
135 r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, &bo->kmap); in qxl_bo_kmap()
138 bo->kptr = ttm_kmap_obj_virtual(&bo->kmap, &is_iomem); in qxl_bo_kmap()
140 *ptr = bo->kptr; in qxl_bo_kmap()
145 struct qxl_bo *bo, int page_offset) in qxl_bo_kmap_atomic_page() argument
147 struct ttm_mem_type_manager *man = &bo->tbo.bdev->man[bo->tbo.mem.mem_type]; in qxl_bo_kmap_atomic_page()
152 if (bo->tbo.mem.mem_type == TTM_PL_VRAM) in qxl_bo_kmap_atomic_page()
154 else if (bo->tbo.mem.mem_type == TTM_PL_PRIV0) in qxl_bo_kmap_atomic_page()
160 ret = ttm_mem_io_reserve(bo->tbo.bdev, &bo->tbo.mem); in qxl_bo_kmap_atomic_page()
163 return io_mapping_map_atomic_wc(map, bo->tbo.mem.bus.offset + page_offset); in qxl_bo_kmap_atomic_page()
165 if (bo->kptr) { in qxl_bo_kmap_atomic_page()
166 rptr = bo->kptr + (page_offset * PAGE_SIZE); in qxl_bo_kmap_atomic_page()
170 ret = qxl_bo_kmap(bo, &rptr); in qxl_bo_kmap_atomic_page()
178 void qxl_bo_kunmap(struct qxl_bo *bo) in qxl_bo_kunmap() argument
180 if (bo->kptr == NULL) in qxl_bo_kunmap()
182 bo->kptr = NULL; in qxl_bo_kunmap()
183 ttm_bo_kunmap(&bo->kmap); in qxl_bo_kunmap()
187 struct qxl_bo *bo, void *pmap) in qxl_bo_kunmap_atomic_page() argument
189 struct ttm_mem_type_manager *man = &bo->tbo.bdev->man[bo->tbo.mem.mem_type]; in qxl_bo_kunmap_atomic_page()
192 if (bo->tbo.mem.mem_type == TTM_PL_VRAM) in qxl_bo_kunmap_atomic_page()
194 else if (bo->tbo.mem.mem_type == TTM_PL_PRIV0) in qxl_bo_kunmap_atomic_page()
202 ttm_mem_io_free(bo->tbo.bdev, &bo->tbo.mem); in qxl_bo_kunmap_atomic_page()
206 qxl_bo_kunmap(bo); in qxl_bo_kunmap_atomic_page()
209 void qxl_bo_unref(struct qxl_bo **bo) in qxl_bo_unref() argument
213 if ((*bo) == NULL) in qxl_bo_unref()
215 tbo = &((*bo)->tbo); in qxl_bo_unref()
218 *bo = NULL; in qxl_bo_unref()
221 struct qxl_bo *qxl_bo_ref(struct qxl_bo *bo) in qxl_bo_ref() argument
223 ttm_bo_reference(&bo->tbo); in qxl_bo_ref()
224 return bo; in qxl_bo_ref()
227 int qxl_bo_pin(struct qxl_bo *bo, u32 domain, u64 *gpu_addr) in qxl_bo_pin() argument
229 struct qxl_device *qdev = (struct qxl_device *)bo->gem_base.dev->dev_private; in qxl_bo_pin()
232 if (bo->pin_count) { in qxl_bo_pin()
233 bo->pin_count++; in qxl_bo_pin()
235 *gpu_addr = qxl_bo_gpu_offset(bo); in qxl_bo_pin()
238 qxl_ttm_placement_from_domain(bo, domain, true); in qxl_bo_pin()
239 r = ttm_bo_validate(&bo->tbo, &bo->placement, false, false); in qxl_bo_pin()
241 bo->pin_count = 1; in qxl_bo_pin()
243 *gpu_addr = qxl_bo_gpu_offset(bo); in qxl_bo_pin()
246 dev_err(qdev->dev, "%p pin failed\n", bo); in qxl_bo_pin()
250 int qxl_bo_unpin(struct qxl_bo *bo) in qxl_bo_unpin() argument
252 struct qxl_device *qdev = (struct qxl_device *)bo->gem_base.dev->dev_private; in qxl_bo_unpin()
255 if (!bo->pin_count) { in qxl_bo_unpin()
256 dev_warn(qdev->dev, "%p unpin not necessary\n", bo); in qxl_bo_unpin()
259 bo->pin_count--; in qxl_bo_unpin()
260 if (bo->pin_count) in qxl_bo_unpin()
262 for (i = 0; i < bo->placement.num_placement; i++) in qxl_bo_unpin()
263 bo->placements[i].flags &= ~TTM_PL_FLAG_NO_EVICT; in qxl_bo_unpin()
264 r = ttm_bo_validate(&bo->tbo, &bo->placement, false, false); in qxl_bo_unpin()
266 dev_err(qdev->dev, "%p validate failed for unpin\n", bo); in qxl_bo_unpin()
272 struct qxl_bo *bo, *n; in qxl_bo_force_delete() local
277 list_for_each_entry_safe(bo, n, &qdev->gem.objects, list) { in qxl_bo_force_delete()
280 &bo->gem_base, bo, (unsigned long)bo->gem_base.size, in qxl_bo_force_delete()
281 *((unsigned long *)&bo->gem_base.refcount)); in qxl_bo_force_delete()
283 list_del_init(&bo->list); in qxl_bo_force_delete()
286 drm_gem_object_unreference(&bo->gem_base); in qxl_bo_force_delete()
301 int qxl_bo_check_id(struct qxl_device *qdev, struct qxl_bo *bo) in qxl_bo_check_id() argument
304 if (bo->type == QXL_GEM_DOMAIN_SURFACE && bo->surface_id == 0) { in qxl_bo_check_id()
306 ret = qxl_surface_id_alloc(qdev, bo); in qxl_bo_check_id()
310 ret = qxl_hw_surface_alloc(qdev, bo, NULL); in qxl_bo_check_id()