Home
last modified time | relevance | path

Searched refs:resv (Results 1 – 92 of 92) sorted by relevance

/linux-4.4.14/fs/ocfs2/
Dreservations.c53 struct ocfs2_alloc_reservation *resv) in ocfs2_resv_window_bits() argument
58 if (!(resv->r_flags & OCFS2_RESV_FLAG_DIR)) { in ocfs2_resv_window_bits()
67 static inline unsigned int ocfs2_resv_end(struct ocfs2_alloc_reservation *resv) in ocfs2_resv_end() argument
69 if (resv->r_len) in ocfs2_resv_end()
70 return resv->r_start + resv->r_len - 1; in ocfs2_resv_end()
71 return resv->r_start; in ocfs2_resv_end()
74 static inline int ocfs2_resv_empty(struct ocfs2_alloc_reservation *resv) in ocfs2_resv_empty() argument
76 return !!(resv->r_len == 0); in ocfs2_resv_empty()
90 struct ocfs2_alloc_reservation *resv; in ocfs2_dump_resv() local
98 resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); in ocfs2_dump_resv()
[all …]
Dreservations.h64 void ocfs2_resv_init_once(struct ocfs2_alloc_reservation *resv);
67 void ocfs2_resv_set_type(struct ocfs2_alloc_reservation *resv,
81 struct ocfs2_alloc_reservation *resv);
137 struct ocfs2_alloc_reservation *resv,
156 struct ocfs2_alloc_reservation *resv,
Dlocalalloc.c56 struct ocfs2_alloc_reservation *resv);
839 struct ocfs2_alloc_reservation *resv) in ocfs2_local_alloc_find_clear_bits() argument
852 if (!resv) { in ocfs2_local_alloc_find_clear_bits()
856 resv = &r; in ocfs2_local_alloc_find_clear_bits()
860 if (ocfs2_resmap_resv_bits(resmap, resv, &bitoff, &numfound) == 0) { in ocfs2_local_alloc_find_clear_bits()
916 ocfs2_resv_discard(resmap, resv); in ocfs2_local_alloc_find_clear_bits()
/linux-4.4.14/drivers/gpu/drm/amd/amdgpu/
Damdgpu_prime.c64 struct reservation_object *resv = attach->dmabuf->resv; in amdgpu_gem_prime_import_sg_table() local
69 ww_mutex_lock(&resv->lock, NULL); in amdgpu_gem_prime_import_sg_table()
71 AMDGPU_GEM_DOMAIN_GTT, 0, sg, resv, &bo); in amdgpu_gem_prime_import_sg_table()
72 ww_mutex_unlock(&resv->lock); in amdgpu_gem_prime_import_sg_table()
115 return bo->tbo.resv; in amdgpu_gem_prime_res_obj()
Damdgpu_sync.c166 struct reservation_object *resv, in amdgpu_sync_resv() argument
175 if (resv == NULL) in amdgpu_sync_resv()
179 f = reservation_object_get_excl(resv); in amdgpu_sync_resv()
182 flist = reservation_object_get_list(resv); in amdgpu_sync_resv()
188 reservation_object_held(resv)); in amdgpu_sync_resv()
Damdgpu_object.c222 struct reservation_object *resv, in amdgpu_bo_create_restricted() argument
276 acc_size, sg, resv, &amdgpu_ttm_bo_destroy); in amdgpu_bo_create_restricted()
291 struct reservation_object *resv, in amdgpu_bo_create() argument
305 resv, bo_ptr); in amdgpu_bo_create()
529 lockdep_assert_held(&bo->tbo.resv->lock.base); in amdgpu_bo_get_tiling_flags()
662 struct reservation_object *resv = bo->tbo.resv; in amdgpu_bo_fence() local
665 reservation_object_add_shared_fence(resv, fence); in amdgpu_bo_fence()
667 reservation_object_add_excl_fence(resv, fence); in amdgpu_bo_fence()
Damdgpu_object.h132 struct reservation_object *resv,
139 struct reservation_object *resv,
Damdgpu_vm.c348 r = reservation_object_reserve_shared(bo->tbo.resv); in amdgpu_vm_clear_bo()
499 amdgpu_sync_resv(adev, &ib->sync, pd->tbo.resv, AMDGPU_FENCE_OWNER_VM); in amdgpu_vm_update_page_directory()
646 amdgpu_sync_resv(adev, &ib->sync, pt->tbo.resv, owner); in amdgpu_vm_update_ptes()
647 r = reservation_object_reserve_shared(pt->tbo.resv); in amdgpu_vm_update_ptes()
1072 struct reservation_object *resv = vm->page_directory->tbo.resv; in amdgpu_vm_bo_map() local
1082 NULL, resv, &pt); in amdgpu_vm_bo_map()
Damdgpu_ttm.c270 bo->resv, &fence); in amdgpu_move_blit()
1014 struct reservation_object *resv, in amdgpu_copy_buffer() argument
1044 if (resv) { in amdgpu_copy_buffer()
1045 r = amdgpu_sync_resv(adev, &ib->sync, resv, in amdgpu_copy_buffer()
Damdgpu_cs.c446 struct reservation_object *resv = e->robj->tbo.resv; in amdgpu_cs_sync_rings() local
447 r = amdgpu_sync_resv(p->adev, &p->ibs[0].sync, resv, p->filp); in amdgpu_cs_sync_rings()
Damdgpu_mn.c155 r = reservation_object_wait_timeout_rcu(bo->tbo.resv, in amdgpu_mn_invalidate_range_start()
Damdgpu_gem.c376 ret = reservation_object_test_signaled_rcu(robj->tbo.resv, true); in amdgpu_gem_wait_idle_ioctl()
378 ret = reservation_object_wait_timeout_rcu(robj->tbo.resv, true, true, timeout); in amdgpu_gem_wait_idle_ioctl()
Damdgpu_display.c222 r = reservation_object_get_fences_rcu(new_rbo->tbo.resv, &work->excl, in amdgpu_crtc_page_flip()
Damdgpu_uvd.c533 r = reservation_object_wait_timeout_rcu(bo->tbo.resv, true, false, in amdgpu_uvd_cs_msg()
Damdgpu.h477 struct reservation_object *resv,
673 struct reservation_object *resv,
/linux-4.4.14/drivers/gpu/drm/radeon/
Dradeon_prime.c64 struct reservation_object *resv = attach->dmabuf->resv; in radeon_gem_prime_import_sg_table() local
69 ww_mutex_lock(&resv->lock, NULL); in radeon_gem_prime_import_sg_table()
71 RADEON_GEM_DOMAIN_GTT, 0, sg, resv, &bo); in radeon_gem_prime_import_sg_table()
72 ww_mutex_unlock(&resv->lock); in radeon_gem_prime_import_sg_table()
116 return bo->tbo.resv; in radeon_gem_prime_res_obj()
Dradeon_benchmark.c38 struct reservation_object *resv) in radeon_benchmark_do_move() argument
51 resv); in radeon_benchmark_do_move()
56 resv); in radeon_benchmark_do_move()
125 dobj->tbo.resv); in radeon_benchmark_move()
136 dobj->tbo.resv); in radeon_benchmark_move()
Dradeon_sync.c91 struct reservation_object *resv, in radeon_sync_resv() argument
101 f = reservation_object_get_excl(resv); in radeon_sync_resv()
108 flist = reservation_object_get_list(resv); in radeon_sync_resv()
114 reservation_object_held(resv)); in radeon_sync_resv()
Drv770_dma.c45 struct reservation_object *resv) in rv770_copy_dma() argument
66 radeon_sync_resv(rdev, &sync, resv, false); in rv770_copy_dma()
Dradeon_object.c181 struct reservation_object *resv, in radeon_bo_create() argument
262 acc_size, sg, resv, &radeon_ttm_bo_destroy); in radeon_bo_create()
599 lockdep_assert_held(&bo->tbo.resv->lock.base); in radeon_bo_get_surface_reg()
725 lockdep_assert_held(&bo->tbo.resv->lock.base); in radeon_bo_get_tiling_flags()
737 lockdep_assert_held(&bo->tbo.resv->lock.base); in radeon_bo_check_tiling()
853 struct reservation_object *resv = bo->tbo.resv; in radeon_bo_fence() local
856 reservation_object_add_shared_fence(resv, &fence->base); in radeon_bo_fence()
858 reservation_object_add_excl_fence(resv, &fence->base); in radeon_bo_fence()
Devergreen_dma.c111 struct reservation_object *resv) in evergreen_copy_dma() argument
132 radeon_sync_resv(rdev, &sync, resv, false); in evergreen_copy_dma()
Dradeon_test.c122 vram_obj->tbo.resv); in radeon_do_test_moves()
126 vram_obj->tbo.resv); in radeon_do_test_moves()
173 vram_obj->tbo.resv); in radeon_do_test_moves()
177 vram_obj->tbo.resv); in radeon_do_test_moves()
Dsi_dma.c234 struct reservation_object *resv) in si_copy_dma() argument
255 radeon_sync_resv(rdev, &sync, resv, false); in si_copy_dma()
Dradeon_asic.h89 struct reservation_object *resv);
160 struct reservation_object *resv);
350 struct reservation_object *resv);
354 struct reservation_object *resv);
475 struct reservation_object *resv);
549 struct reservation_object *resv);
727 struct reservation_object *resv);
798 struct reservation_object *resv);
802 struct reservation_object *resv);
Dr600_dma.c442 struct reservation_object *resv) in r600_copy_dma() argument
463 radeon_sync_resv(rdev, &sync, resv, false); in r600_copy_dma()
Dradeon_cs.c239 struct reservation_object *resv; in radeon_cs_sync_rings() local
241 resv = reloc->robj->tbo.resv; in radeon_cs_sync_rings()
242 r = radeon_sync_resv(p->rdev, &p->ib.sync, resv, in radeon_cs_sync_rings()
Dradeon_object.h129 struct reservation_object *resv,
Dradeon_gem.c114 r = reservation_object_wait_timeout_rcu(robj->tbo.resv, true, true, 30 * HZ); in radeon_gem_set_domain()
444 r = reservation_object_test_signaled_rcu(robj->tbo.resv, true); in radeon_gem_busy_ioctl()
473 ret = reservation_object_wait_timeout_rcu(robj->tbo.resv, true, true, 30 * HZ); in radeon_gem_wait_idle_ioctl()
Dradeon.h619 struct reservation_object *resv,
1921 struct reservation_object *resv);
1927 struct reservation_object *resv);
1934 struct reservation_object *resv);
2747 #define radeon_copy_blit(rdev, s, d, np, resv) (rdev)->asic->copy.blit((rdev), (s), (d), (np), (res… argument
2748 #define radeon_copy_dma(rdev, s, d, np, resv) (rdev)->asic->copy.dma((rdev), (s), (d), (np), (resv)) argument
2749 #define radeon_copy(rdev, s, d, np, resv) (rdev)->asic->copy.copy((rdev), (s), (d), (np), (resv)) argument
Dradeon_vm.c703 radeon_sync_resv(rdev, &ib.sync, pd->tbo.resv, true); in radeon_vm_update_page_directory()
831 radeon_sync_resv(rdev, &ib->sync, pt->tbo.resv, true); in radeon_vm_update_ptes()
832 r = reservation_object_reserve_shared(pt->tbo.resv); in radeon_vm_update_ptes()
Dradeon_mn.c154 r = reservation_object_wait_timeout_rcu(bo->tbo.resv, in radeon_mn_invalidate_range_start()
Dcik_sdma.c582 struct reservation_object *resv) in cik_copy_dma() argument
603 radeon_sync_resv(rdev, &sync, resv, false); in cik_copy_dma()
Dr200.c87 struct reservation_object *resv) in r200_copy_dma() argument
Dradeon_uvd.c436 f = reservation_object_get_excl(bo->tbo.resv); in radeon_uvd_cs_msg()
Dradeon_ttm.c299 fence = radeon_copy(rdev, old_start, new_start, num_pages, bo->resv); in radeon_move_blit()
Dradeon_display.c559 work->fence = fence_get(reservation_object_get_excl(new_rbo->tbo.resv)); in radeon_crtc_page_flip()
Dr600.c2964 struct reservation_object *resv) in r600_copy_cpdma() argument
2985 radeon_sync_resv(rdev, &sync, resv, false); in r600_copy_cpdma()
Dcik.c4070 struct reservation_object *resv) in cik_copy_cpdma() argument
4091 radeon_sync_resv(rdev, &sync, resv, false); in cik_copy_cpdma()
Dr100.c886 struct reservation_object *resv) in r100_copy_blit() argument
/linux-4.4.14/net/sunrpc/
Dsvc.c1069 svc_process_common(struct svc_rqst *rqstp, struct kvec *argv, struct kvec *resv) in svc_process_common() argument
1096 svc_putu32(resv, rqstp->rq_xid); in svc_process_common()
1101 svc_putnl(resv, 1); /* REPLY */ in svc_process_common()
1107 reply_statp = resv->iov_base + resv->iov_len; in svc_process_common()
1109 svc_putnl(resv, 0); /* ACCEPT */ in svc_process_common()
1165 statp = resv->iov_base +resv->iov_len; in svc_process_common()
1166 svc_putnl(resv, RPC_SUCCESS); in svc_process_common()
1198 !xdr(rqstp, resv->iov_base+resv->iov_len, rqstp->rq_resp)) { in svc_process_common()
1215 resv->iov_len = ((void*)statp) - resv->iov_base + 4; in svc_process_common()
1242 svc_putnl(resv, 1); /* REJECT */ in svc_process_common()
[all …]
Dsvcauth_unix.c728 struct kvec *resv = &rqstp->rq_res.head[0]; in svcauth_null_accept() local
757 svc_putnl(resv, RPC_AUTH_NULL); in svcauth_null_accept()
758 svc_putnl(resv, 0); in svcauth_null_accept()
792 struct kvec *resv = &rqstp->rq_res.head[0]; in svcauth_unix_accept() local
836 svc_putnl(resv, RPC_AUTH_NULL); in svcauth_unix_accept()
837 svc_putnl(resv, 0); in svcauth_unix_accept()
Dsvcsock.c1245 struct kvec *resv = &rqstp->rq_res.head[0]; in svc_tcp_prep_reply_hdr() local
1248 svc_putnl(resv, 0); in svc_tcp_prep_reply_hdr()
/linux-4.4.14/drivers/dma-buf/
Ddma-buf.c73 if (dmabuf->resv == (struct reservation_object *)&dmabuf[1]) in dma_buf_release()
74 reservation_object_fini(dmabuf->resv); in dma_buf_release()
138 struct reservation_object *resv; in dma_buf_poll() local
145 if (!dmabuf || !dmabuf->resv) in dma_buf_poll()
148 resv = dmabuf->resv; in dma_buf_poll()
157 seq = read_seqcount_begin(&resv->seq); in dma_buf_poll()
160 fobj = rcu_dereference(resv->fence); in dma_buf_poll()
165 fence_excl = rcu_dereference(resv->fence_excl); in dma_buf_poll()
166 if (read_seqcount_retry(&resv->seq, seq)) { in dma_buf_poll()
287 struct reservation_object *resv = exp_info->resv; in dma_buf_export() local
[all …]
/linux-4.4.14/drivers/gpu/drm/ttm/
Dttm_execbuf_util.c134 ret = reservation_object_reserve_shared(bo->resv); in ttm_eu_reserve_buffers()
146 ret = ww_mutex_lock_slow_interruptible(&bo->resv->lock, in ttm_eu_reserve_buffers()
149 ww_mutex_lock_slow(&bo->resv->lock, ticket); in ttm_eu_reserve_buffers()
154 ret = reservation_object_reserve_shared(bo->resv); in ttm_eu_reserve_buffers()
204 reservation_object_add_shared_fence(bo->resv, fence); in ttm_eu_fence_buffer_objects()
206 reservation_object_add_excl_fence(bo->resv, fence); in ttm_eu_fence_buffer_objects()
Dttm_bo.c153 if (bo->resv == &bo->ttm_resv) in ttm_bo_release_list()
169 lockdep_assert_held(&bo->resv->lock.base); in ttm_bo_add_to_lru()
403 ww_mutex_unlock (&bo->resv->lock); in ttm_bo_cleanup_memtype_use()
412 fobj = reservation_object_get_list(bo->resv); in ttm_bo_flush_all_fences()
413 fence = reservation_object_get_excl(bo->resv); in ttm_bo_flush_all_fences()
419 reservation_object_held(bo->resv)); in ttm_bo_flush_all_fences()
494 ww_mutex_unlock(&bo->resv->lock); in ttm_bo_cleanup_refs_and_unlock()
497 lret = reservation_object_wait_timeout_rcu(bo->resv, in ttm_bo_cleanup_refs_and_unlock()
676 lockdep_assert_held(&bo->resv->lock.base); in ttm_bo_evict()
977 lockdep_assert_held(&bo->resv->lock.base); in ttm_bo_move_buffer()
[all …]
Dttm_bo_util.c472 fbo->resv = &fbo->ttm_resv; in ttm_buffer_object_transfer()
473 reservation_object_init(fbo->resv); in ttm_buffer_object_transfer()
474 ret = ww_mutex_trylock(&fbo->resv->lock); in ttm_buffer_object_transfer()
646 reservation_object_add_excl_fence(bo->resv, fence); in ttm_bo_move_accel_cleanup()
674 reservation_object_add_excl_fence(ghost_obj->resv, fence); in ttm_bo_move_accel_cleanup()
/linux-4.4.14/net/sunrpc/auth_gss/
Dsvcauth_gss.c637 svc_safe_putnetobj(struct kvec *resv, struct xdr_netobj *o) in svc_safe_putnetobj() argument
641 if (resv->iov_len + 4 > PAGE_SIZE) in svc_safe_putnetobj()
643 svc_putnl(resv, o->len); in svc_safe_putnetobj()
644 p = resv->iov_base + resv->iov_len; in svc_safe_putnetobj()
645 resv->iov_len += round_up_to_quad(o->len); in svc_safe_putnetobj()
646 if (resv->iov_len > PAGE_SIZE) in svc_safe_putnetobj()
1074 gss_write_resv(struct kvec *resv, size_t size_limit, in gss_write_resv() argument
1078 if (resv->iov_len + 4 > size_limit) in gss_write_resv()
1080 svc_putnl(resv, RPC_SUCCESS); in gss_write_resv()
1081 if (svc_safe_putnetobj(resv, out_handle)) in gss_write_resv()
[all …]
/linux-4.4.14/mm/
Dhugetlb.c255 static long region_add(struct resv_map *resv, long f, long t) in region_add() argument
257 struct list_head *head = &resv->regions; in region_add()
261 spin_lock(&resv->lock); in region_add()
274 VM_BUG_ON(resv->region_cache_count <= 0); in region_add()
276 resv->region_cache_count--; in region_add()
277 nrg = list_first_entry(&resv->region_cache, struct file_region, in region_add()
323 resv->adds_in_progress--; in region_add()
324 spin_unlock(&resv->lock); in region_add()
351 static long region_chg(struct resv_map *resv, long f, long t) in region_chg() argument
353 struct list_head *head = &resv->regions; in region_chg()
[all …]
/linux-4.4.14/include/drm/ttm/
Dttm_bo_driver.h790 success = ww_mutex_trylock(&bo->resv->lock); in __ttm_bo_reserve()
795 ret = ww_mutex_lock_interruptible(&bo->resv->lock, ticket); in __ttm_bo_reserve()
797 ret = ww_mutex_lock(&bo->resv->lock, ticket); in __ttm_bo_reserve()
883 ret = ww_mutex_lock_slow_interruptible(&bo->resv->lock, in ttm_bo_reserve_slowpath()
886 ww_mutex_lock_slow(&bo->resv->lock, ticket); in ttm_bo_reserve_slowpath()
905 ww_mutex_unlock(&bo->resv->lock); in __ttm_bo_unreserve()
Dttm_bo_api.h257 struct reservation_object *resv; member
491 struct reservation_object *resv,
/linux-4.4.14/arch/sparc/include/asm/
Dvio.h43 u64 resv[6]; member
69 u16 resv; member
77 u64 resv[5]; member
177 u32 resv; member
198 u16 resv; member
240 u32 resv; member
Dhypervisor.h473 unsigned long resv; member
/linux-4.4.14/include/uapi/linux/
Digmp.h70 resv:4; member
72 __u8 resv:4, member
Dvfio.h191 __u32 resv; /* Reserved for alignment */ member
/linux-4.4.14/drivers/gpu/drm/nouveau/
Dnouveau_prime.c64 struct reservation_object *robj = attach->dmabuf->resv; in nouveau_gem_prime_import_sg_table()
114 return nvbo->bo.resv; in nouveau_gem_prime_res_obj()
Dnouveau_fence.c394 struct reservation_object *resv = nvbo->bo.resv; in nouveau_fence_sync() local
400 ret = reservation_object_reserve_shared(resv); in nouveau_fence_sync()
406 fobj = reservation_object_get_list(resv); in nouveau_fence_sync()
407 fence = reservation_object_get_excl(resv); in nouveau_fence_sync()
436 reservation_object_held(resv)); in nouveau_fence_sync()
Dnouveau_gem.c120 struct reservation_object *resv = nvbo->bo.resv; in nouveau_gem_object_unmap() local
124 fobj = reservation_object_get_list(resv); in nouveau_gem_object_unmap()
132 reservation_object_held(resv)); in nouveau_gem_object_unmap()
134 fence = reservation_object_get_excl(nvbo->bo.resv); in nouveau_gem_object_unmap()
873 ret = reservation_object_test_signaled_rcu(nvbo->bo.resv, write) ? 0 : -EBUSY; in nouveau_gem_ioctl_cpu_prep()
877 lret = reservation_object_wait_timeout_rcu(nvbo->bo.resv, write, true, 30 * HZ); in nouveau_gem_ioctl_cpu_prep()
Dnouveau_bo.c1276 struct fence *fence = reservation_object_get_excl(bo->resv); in nouveau_bo_vm_cleanup()
1593 struct reservation_object *resv = nvbo->bo.resv; in nouveau_bo_fence() local
1596 reservation_object_add_excl_fence(resv, &fence->base); in nouveau_bo_fence()
1598 reservation_object_add_shared_fence(resv, &fence->base); in nouveau_bo_fence()
/linux-4.4.14/include/linux/
Ddma-buf.h137 struct reservation_object *resv; member
187 struct reservation_object *resv; member
/linux-4.4.14/include/rdma/
Dib_smi.h55 __be16 resv; member
118 u8 resv; member
Dopa_smi.h61 __be16 resv; member
Dib_mad.h173 __be16 resv; member
/linux-4.4.14/drivers/gpu/drm/msm/
Dmsm_gem_submit.c130 ww_mutex_unlock(&msm_obj->resv->lock); in submit_unlock_unpin_bo()
156 ret = ww_mutex_lock_interruptible(&msm_obj->resv->lock, in submit_validate_objects()
202 ret = ww_mutex_lock_slow_interruptible(&msm_obj->resv->lock, in submit_validate_objects()
Dmsm_gem.h61 struct reservation_object *resv; member
Dmsm_gem.c548 if (msm_obj->resv == &msm_obj->_resv) in msm_gem_free_object()
549 reservation_object_fini(msm_obj->resv); in msm_gem_free_object()
623 msm_obj->resv = &msm_obj->_resv; in msm_gem_new_impl()
624 reservation_object_init(msm_obj->resv); in msm_gem_new_impl()
/linux-4.4.14/fs/nfsd/
Dnfscache.c513 struct kvec *resv = &rqstp->rq_res.head[0], *cachv; in nfsd_cache_update() local
525 len = resv->iov_len - ((char*)statp - (char*)resv->iov_base); in nfsd_cache_update()
/linux-4.4.14/drivers/gpu/drm/qxl/
Dqxl_debugfs.c65 fobj = rcu_dereference(bo->tbo.resv->fence); in qxl_debugfs_buffers_info()
Dqxl_release.c246 ret = reservation_object_reserve_shared(bo->tbo.resv); in qxl_release_validate_bo()
472 reservation_object_add_shared_fence(bo->resv, &release->base); in qxl_release_fence_buffer_objects()
/linux-4.4.14/drivers/nvme/host/
Dlightnvm.c66 __le64 resv; member
133 __le64 resv; member
199 __u8 resv[228]; member
Dscsi.c2364 u32 resv; member
2370 u32 resv; member
/linux-4.4.14/drivers/gpu/drm/vmwgfx/
Dvmwgfx_resource.c578 return reservation_object_test_signaled_rcu(bo->resv, true) ? 0 : -EBUSY; in vmw_user_dmabuf_synccpu_grab()
580 lret = reservation_object_wait_timeout_rcu(bo->resv, true, true, MAX_SCHEDULE_TIMEOUT); in vmw_user_dmabuf_synccpu_grab()
1186 lockdep_assert_held(&res->backup->base.resv->lock.base); in vmw_resource_unreserve()
1193 lockdep_assert_held(&new_backup->base.resv->lock.base); in vmw_resource_unreserve()
1458 reservation_object_add_excl_fence(bo->resv, &fence->base); in vmw_fence_single_bo()
1461 reservation_object_add_excl_fence(bo->resv, &fence->base); in vmw_fence_single_bo()
Dvmwgfx_cotable.c171 lockdep_assert_held(&bo->resv->lock.base); in vmw_cotable_unscrub()
319 lockdep_assert_held(&bo->resv->lock.base); in vmw_cotable_unbind()
Dvmwgfx_dmabuf.c276 lockdep_assert_held(&bo->resv->lock.base); in vmw_bo_pin_reserved()
/linux-4.4.14/drivers/infiniband/hw/qib/
Dqib_mad.h128 __be16 resv; member
/linux-4.4.14/drivers/gpu/drm/virtio/
Dvirtgpu_ioctl.c397 reservation_object_add_excl_fence(qobj->tbo.resv, in virtio_gpu_transfer_from_host_ioctl()
446 reservation_object_add_excl_fence(qobj->tbo.resv, in virtio_gpu_transfer_to_host_ioctl()
Dvirtgpu_display.c95 reservation_object_add_excl_fence(qobj->tbo.resv, in virtio_gpu_crtc_cursor_set()
/linux-4.4.14/drivers/input/mouse/
Dcyapa_gen6.c47 u8 resv; /* Reserved, must be 0 */ member
54 u8 resv; /* Reserved, must be 0 */ member
/linux-4.4.14/drivers/net/ethernet/rocker/
Drocker.h110 u16 resv[5]; member
/linux-4.4.14/arch/sparc/mm/
Dtsb.c259 hp->resv = 0; in setup_tsb_params()
Dinit_64.c1924 ktsb_descr[0].resv = 0; in sun4v_ktsb_init()
1941 ktsb_descr[1].resv = 0; in sun4v_ktsb_init()
/linux-4.4.14/arch/sparc/kernel/
Dpci_fire.c99 u64 resv[6]; member
Dmdesc.c53 u16 resv; member
/linux-4.4.14/drivers/gpu/drm/
Ddrm_prime.c352 exp_info.resv = dev->driver->gem_prime_res_obj(obj); in drm_gem_prime_export()
/linux-4.4.14/drivers/staging/rdma/ehca/
Dehca_reqs.c134 mad_hdr->resv, in trace_ud_wr()
/linux-4.4.14/drivers/scsi/bfa/
Dbfa_defs_svc.h1054 u8 resv[2]; member
/linux-4.4.14/drivers/scsi/lpfc/
Dlpfc_hw.h973 uint8_t resv; member
981 uint8_t resv;
Dlpfc_els.c5322 switch (rscn_did.un.b.resv & RSCN_ADDRESS_FORMAT_MASK) { in lpfc_rscn_payload_check()
/linux-4.4.14/include/linux/mlx4/
Ddevice.h969 __be16 resv; member
/linux-4.4.14/drivers/scsi/
Dipr.h1250 u8 resv; member
/linux-4.4.14/drivers/iommu/
Damd_iommu_init.c126 u64 resv; member
/linux-4.4.14/drivers/net/ethernet/sun/
Dniu.h2821 __le64 resv; member
Dniu.c6676 tp->resv = 0; in niu_start_xmit()