Lines Matching refs:bo

39 		struct ttm_buffer_object *bo = entry->bo;  in ttm_eu_backoff_reservation_reverse()  local
41 __ttm_bo_unreserve(bo); in ttm_eu_backoff_reservation_reverse()
50 struct ttm_buffer_object *bo = entry->bo; in ttm_eu_del_from_lru_locked() local
51 unsigned put_count = ttm_bo_del_from_lru(bo); in ttm_eu_del_from_lru_locked()
53 ttm_bo_list_ref_sub(bo, put_count, true); in ttm_eu_del_from_lru_locked()
67 glob = entry->bo->glob; in ttm_eu_backoff_reservation()
71 struct ttm_buffer_object *bo = entry->bo; in ttm_eu_backoff_reservation() local
73 ttm_bo_add_to_lru(bo); in ttm_eu_backoff_reservation()
74 __ttm_bo_unreserve(bo); in ttm_eu_backoff_reservation()
107 glob = entry->bo->glob; in ttm_eu_reserve_buffers()
113 struct ttm_buffer_object *bo = entry->bo; in ttm_eu_reserve_buffers() local
115 ret = __ttm_bo_reserve(bo, intr, (ticket == NULL), true, in ttm_eu_reserve_buffers()
117 if (!ret && unlikely(atomic_read(&bo->cpu_writers) > 0)) { in ttm_eu_reserve_buffers()
118 __ttm_bo_unreserve(bo); in ttm_eu_reserve_buffers()
134 ret = reservation_object_reserve_shared(bo->resv); in ttm_eu_reserve_buffers()
146 ret = ww_mutex_lock_slow_interruptible(&bo->resv->lock, in ttm_eu_reserve_buffers()
149 ww_mutex_lock_slow(&bo->resv->lock, ticket); in ttm_eu_reserve_buffers()
154 ret = reservation_object_reserve_shared(bo->resv); in ttm_eu_reserve_buffers()
186 struct ttm_buffer_object *bo; in ttm_eu_fence_buffer_objects() local
194 bo = list_first_entry(list, struct ttm_validate_buffer, head)->bo; in ttm_eu_fence_buffer_objects()
195 bdev = bo->bdev; in ttm_eu_fence_buffer_objects()
197 glob = bo->glob; in ttm_eu_fence_buffer_objects()
202 bo = entry->bo; in ttm_eu_fence_buffer_objects()
204 reservation_object_add_shared_fence(bo->resv, fence); in ttm_eu_fence_buffer_objects()
206 reservation_object_add_excl_fence(bo->resv, fence); in ttm_eu_fence_buffer_objects()
207 ttm_bo_add_to_lru(bo); in ttm_eu_fence_buffer_objects()
208 __ttm_bo_unreserve(bo); in ttm_eu_fence_buffer_objects()