Searched refs:resv (Results 1 - 100 of 100) sorted by relevance

/linux-4.4.14/fs/ocfs2/
H A Dreservations.c53 struct ocfs2_alloc_reservation *resv) ocfs2_resv_window_bits()
58 if (!(resv->r_flags & OCFS2_RESV_FLAG_DIR)) { ocfs2_resv_window_bits()
67 static inline unsigned int ocfs2_resv_end(struct ocfs2_alloc_reservation *resv) ocfs2_resv_end() argument
69 if (resv->r_len) ocfs2_resv_end()
70 return resv->r_start + resv->r_len - 1; ocfs2_resv_end()
71 return resv->r_start; ocfs2_resv_end()
74 static inline int ocfs2_resv_empty(struct ocfs2_alloc_reservation *resv) ocfs2_resv_empty() argument
76 return !!(resv->r_len == 0); ocfs2_resv_empty()
90 struct ocfs2_alloc_reservation *resv; ocfs2_dump_resv() local
98 resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); ocfs2_dump_resv()
101 "\tlast_len: %u\n", resv->r_start, ocfs2_dump_resv()
102 ocfs2_resv_end(resv), resv->r_len, resv->r_last_start, ocfs2_dump_resv()
103 resv->r_last_len); ocfs2_dump_resv()
112 list_for_each_entry(resv, &resmap->m_lru, r_lru) { ocfs2_dump_resv()
114 "last_start: %u\tlast_len: %u\n", i, resv->r_start, ocfs2_dump_resv()
115 ocfs2_resv_end(resv), resv->r_len, resv->r_last_start, ocfs2_dump_resv()
116 resv->r_last_len); ocfs2_dump_resv()
125 struct ocfs2_alloc_reservation *resv) ocfs2_validate_resmap_bits()
128 unsigned int start = resv->r_start; ocfs2_validate_resmap_bits()
129 unsigned int end = ocfs2_resv_end(resv); ocfs2_validate_resmap_bits()
149 struct ocfs2_alloc_reservation *resv; ocfs2_check_resmap() local
153 resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); ocfs2_check_resmap()
155 if (i > 0 && resv->r_start <= off) { ocfs2_check_resmap()
161 if (resv->r_len == 0) { ocfs2_check_resmap()
167 if (resv->r_start > ocfs2_resv_end(resv)) { ocfs2_check_resmap()
173 if (ocfs2_resv_end(resv) >= resmap->m_bitmap_len) { ocfs2_check_resmap()
179 if (ocfs2_validate_resmap_bits(resmap, i, resv)) ocfs2_check_resmap()
182 off = ocfs2_resv_end(resv); ocfs2_check_resmap()
200 void ocfs2_resv_init_once(struct ocfs2_alloc_reservation *resv) ocfs2_resv_init_once() argument
202 memset(resv, 0, sizeof(*resv)); ocfs2_resv_init_once()
203 INIT_LIST_HEAD(&resv->r_lru); ocfs2_resv_init_once()
206 void ocfs2_resv_set_type(struct ocfs2_alloc_reservation *resv, ocfs2_resv_set_type() argument
211 resv->r_flags |= flags; ocfs2_resv_set_type()
228 struct ocfs2_alloc_reservation *resv) ocfs2_resv_mark_lru()
232 if (!list_empty(&resv->r_lru)) ocfs2_resv_mark_lru()
233 list_del_init(&resv->r_lru); ocfs2_resv_mark_lru()
235 list_add_tail(&resv->r_lru, &resmap->m_lru); ocfs2_resv_mark_lru()
238 static void __ocfs2_resv_trunc(struct ocfs2_alloc_reservation *resv) __ocfs2_resv_trunc() argument
240 resv->r_len = 0; __ocfs2_resv_trunc()
241 resv->r_start = 0; __ocfs2_resv_trunc()
245 struct ocfs2_alloc_reservation *resv) ocfs2_resv_remove()
247 if (resv->r_flags & OCFS2_RESV_FLAG_INUSE) { ocfs2_resv_remove()
248 list_del_init(&resv->r_lru); ocfs2_resv_remove()
249 rb_erase(&resv->r_node, &resmap->m_reservations); ocfs2_resv_remove()
250 resv->r_flags &= ~OCFS2_RESV_FLAG_INUSE; ocfs2_resv_remove()
255 struct ocfs2_alloc_reservation *resv) __ocfs2_resv_discard()
259 __ocfs2_resv_trunc(resv); __ocfs2_resv_discard()
264 resv->r_last_len = resv->r_last_start = 0; __ocfs2_resv_discard()
266 ocfs2_resv_remove(resmap, resv); __ocfs2_resv_discard()
269 /* does nothing if 'resv' is null */ ocfs2_resv_discard()
271 struct ocfs2_alloc_reservation *resv) ocfs2_resv_discard()
273 if (resv) { ocfs2_resv_discard()
275 __ocfs2_resv_discard(resmap, resv); ocfs2_resv_discard()
283 struct ocfs2_alloc_reservation *resv; ocfs2_resmap_clear_all_resv() local
288 resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); ocfs2_resmap_clear_all_resv()
290 __ocfs2_resv_discard(resmap, resv); ocfs2_resmap_clear_all_resv()
369 struct ocfs2_alloc_reservation *resv = NULL; ocfs2_find_resv_lhs() local
380 resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); ocfs2_find_resv_lhs()
382 if (resv->r_start <= goal && ocfs2_resv_end(resv) >= goal) ocfs2_find_resv_lhs()
386 if (resv->r_start > goal) { ocfs2_find_resv_lhs()
387 resv = prev_resv; ocfs2_find_resv_lhs()
391 prev_resv = resv; ocfs2_find_resv_lhs()
395 return resv; ocfs2_find_resv_lhs()
471 struct ocfs2_alloc_reservation *resv, __ocfs2_resv_find_window()
489 trace_ocfs2_resv_find_window_begin(resv->r_start, ocfs2_resv_end(resv), __ocfs2_resv_find_window()
512 resv->r_start = cstart; __ocfs2_resv_find_window()
513 resv->r_len = clen; __ocfs2_resv_find_window()
515 ocfs2_resv_insert(resmap, resv); __ocfs2_resv_find_window()
624 resv->r_start = best_start; __ocfs2_resv_find_window()
625 resv->r_len = best_len; __ocfs2_resv_find_window()
626 ocfs2_resv_insert(resmap, resv); __ocfs2_resv_find_window()
631 struct ocfs2_alloc_reservation *resv, ocfs2_cannibalize_resv()
635 int tmpwindow = !!(resv->r_flags & OCFS2_RESV_FLAG_TMP); ocfs2_cannibalize_resv()
639 min_bits = ocfs2_resv_window_bits(resmap, resv) >> 1; ocfs2_cannibalize_resv()
667 resv->r_start = lru_resv->r_start; ocfs2_cannibalize_resv()
668 resv->r_len = lru_resv->r_len; ocfs2_cannibalize_resv()
680 resv->r_start = ocfs2_resv_end(lru_resv) + 1; ocfs2_cannibalize_resv()
681 resv->r_len = shrink; ocfs2_cannibalize_resv()
684 trace_ocfs2_cannibalize_resv_end(resv->r_start, ocfs2_resv_end(resv), ocfs2_cannibalize_resv()
685 resv->r_len, resv->r_last_start, ocfs2_cannibalize_resv()
686 resv->r_last_len); ocfs2_cannibalize_resv()
688 ocfs2_resv_insert(resmap, resv); ocfs2_cannibalize_resv()
692 struct ocfs2_alloc_reservation *resv, ocfs2_resv_find_window()
697 BUG_ON(!ocfs2_resv_empty(resv)); ocfs2_resv_find_window()
704 if (resv->r_last_len) { ocfs2_resv_find_window()
705 goal = resv->r_last_start + resv->r_last_len; ocfs2_resv_find_window()
710 __ocfs2_resv_find_window(resmap, resv, goal, wanted); ocfs2_resv_find_window()
713 if (ocfs2_resv_empty(resv) && goal != 0) ocfs2_resv_find_window()
714 __ocfs2_resv_find_window(resmap, resv, 0, wanted); ocfs2_resv_find_window()
716 if (ocfs2_resv_empty(resv)) { ocfs2_resv_find_window()
721 ocfs2_cannibalize_resv(resmap, resv, wanted); ocfs2_resv_find_window()
724 BUG_ON(ocfs2_resv_empty(resv)); ocfs2_resv_find_window()
728 struct ocfs2_alloc_reservation *resv, ocfs2_resmap_resv_bits()
731 if (resv == NULL || ocfs2_resmap_disabled(resmap)) ocfs2_resmap_resv_bits()
736 if (ocfs2_resv_empty(resv)) { ocfs2_resmap_resv_bits()
742 unsigned int wanted = ocfs2_resv_window_bits(resmap, resv); ocfs2_resmap_resv_bits()
744 if ((resv->r_flags & OCFS2_RESV_FLAG_TMP) || wanted < *clen) ocfs2_resmap_resv_bits()
754 ocfs2_resv_find_window(resmap, resv, wanted); ocfs2_resmap_resv_bits()
755 trace_ocfs2_resmap_resv_bits(resv->r_start, resv->r_len); ocfs2_resmap_resv_bits()
758 BUG_ON(ocfs2_resv_empty(resv)); ocfs2_resmap_resv_bits()
760 *cstart = resv->r_start; ocfs2_resmap_resv_bits()
761 *clen = resv->r_len; ocfs2_resmap_resv_bits()
769 struct ocfs2_alloc_reservation *resv, ocfs2_adjust_resv_from_alloc()
773 unsigned int old_end = ocfs2_resv_end(resv); ocfs2_adjust_resv_from_alloc()
775 BUG_ON(start != resv->r_start || old_end < end); ocfs2_adjust_resv_from_alloc()
781 __ocfs2_resv_discard(resmap, resv); ocfs2_adjust_resv_from_alloc()
792 resv->r_start = end + 1; ocfs2_adjust_resv_from_alloc()
793 resv->r_len = old_end - resv->r_start + 1; ocfs2_adjust_resv_from_alloc()
797 struct ocfs2_alloc_reservation *resv, ocfs2_resmap_claimed_bits()
805 if (resv == NULL) ocfs2_resmap_claimed_bits()
808 BUG_ON(cstart != resv->r_start); ocfs2_resmap_claimed_bits()
812 trace_ocfs2_resmap_claimed_bits_begin(cstart, cend, clen, resv->r_start, ocfs2_resmap_claimed_bits()
813 ocfs2_resv_end(resv), resv->r_len, ocfs2_resmap_claimed_bits()
814 resv->r_last_start, ocfs2_resmap_claimed_bits()
815 resv->r_last_len); ocfs2_resmap_claimed_bits()
817 BUG_ON(cstart < resv->r_start); ocfs2_resmap_claimed_bits()
818 BUG_ON(cstart > ocfs2_resv_end(resv)); ocfs2_resmap_claimed_bits()
819 BUG_ON(cend > ocfs2_resv_end(resv)); ocfs2_resmap_claimed_bits()
821 ocfs2_adjust_resv_from_alloc(resmap, resv, cstart, cend); ocfs2_resmap_claimed_bits()
822 resv->r_last_start = cstart; ocfs2_resmap_claimed_bits()
823 resv->r_last_len = clen; ocfs2_resmap_claimed_bits()
829 if (!ocfs2_resv_empty(resv)) ocfs2_resmap_claimed_bits()
830 ocfs2_resv_mark_lru(resmap, resv); ocfs2_resmap_claimed_bits()
832 trace_ocfs2_resmap_claimed_bits_end(resv->r_start, ocfs2_resv_end(resv), ocfs2_resmap_claimed_bits()
833 resv->r_len, resv->r_last_start, ocfs2_resmap_claimed_bits()
834 resv->r_last_len); ocfs2_resmap_claimed_bits()
52 ocfs2_resv_window_bits(struct ocfs2_reservation_map *resmap, struct ocfs2_alloc_reservation *resv) ocfs2_resv_window_bits() argument
123 ocfs2_validate_resmap_bits(struct ocfs2_reservation_map *resmap, int i, struct ocfs2_alloc_reservation *resv) ocfs2_validate_resmap_bits() argument
227 ocfs2_resv_mark_lru(struct ocfs2_reservation_map *resmap, struct ocfs2_alloc_reservation *resv) ocfs2_resv_mark_lru() argument
244 ocfs2_resv_remove(struct ocfs2_reservation_map *resmap, struct ocfs2_alloc_reservation *resv) ocfs2_resv_remove() argument
254 __ocfs2_resv_discard(struct ocfs2_reservation_map *resmap, struct ocfs2_alloc_reservation *resv) __ocfs2_resv_discard() argument
270 ocfs2_resv_discard(struct ocfs2_reservation_map *resmap, struct ocfs2_alloc_reservation *resv) ocfs2_resv_discard() argument
470 __ocfs2_resv_find_window(struct ocfs2_reservation_map *resmap, struct ocfs2_alloc_reservation *resv, unsigned int goal, unsigned int wanted) __ocfs2_resv_find_window() argument
630 ocfs2_cannibalize_resv(struct ocfs2_reservation_map *resmap, struct ocfs2_alloc_reservation *resv, unsigned int wanted) ocfs2_cannibalize_resv() argument
691 ocfs2_resv_find_window(struct ocfs2_reservation_map *resmap, struct ocfs2_alloc_reservation *resv, unsigned int wanted) ocfs2_resv_find_window() argument
727 ocfs2_resmap_resv_bits(struct ocfs2_reservation_map *resmap, struct ocfs2_alloc_reservation *resv, int *cstart, int *clen) ocfs2_resmap_resv_bits() argument
768 ocfs2_adjust_resv_from_alloc(struct ocfs2_reservation_map *resmap, struct ocfs2_alloc_reservation *resv, unsigned int start, unsigned int end) ocfs2_adjust_resv_from_alloc() argument
796 ocfs2_resmap_claimed_bits(struct ocfs2_reservation_map *resmap, struct ocfs2_alloc_reservation *resv, u32 cstart, u32 clen) ocfs2_resmap_claimed_bits() argument
H A Dreservations.h64 void ocfs2_resv_init_once(struct ocfs2_alloc_reservation *resv);
67 void ocfs2_resv_set_type(struct ocfs2_alloc_reservation *resv,
75 * @resv: the reservation to truncate.
81 struct ocfs2_alloc_reservation *resv);
122 * @resv: reservation to base search from
126 * Using the reservation data from resv, this function will compare
128 * the reservation window is still clear to use. If resv is empty,
137 struct ocfs2_alloc_reservation *resv,
143 * @resv: optional reservation to recalulate based on new bitmap
151 * reservations bitmap. If resv is passed, it's next allocation window will be
156 struct ocfs2_alloc_reservation *resv,
H A Dlocalalloc.c56 struct ocfs2_alloc_reservation *resv);
839 struct ocfs2_alloc_reservation *resv) ocfs2_local_alloc_find_clear_bits()
852 if (!resv) { ocfs2_local_alloc_find_clear_bits()
856 resv = &r; ocfs2_local_alloc_find_clear_bits()
860 if (ocfs2_resmap_resv_bits(resmap, resv, &bitoff, &numfound) == 0) { ocfs2_local_alloc_find_clear_bits()
916 ocfs2_resv_discard(resmap, resv); ocfs2_local_alloc_find_clear_bits()
836 ocfs2_local_alloc_find_clear_bits(struct ocfs2_super *osb, struct ocfs2_dinode *alloc, u32 *numbits, struct ocfs2_alloc_reservation *resv) ocfs2_local_alloc_find_clear_bits() argument
/linux-4.4.14/drivers/gpu/drm/radeon/
H A Dradeon_prime.c64 struct reservation_object *resv = attach->dmabuf->resv; radeon_gem_prime_import_sg_table() local
69 ww_mutex_lock(&resv->lock, NULL); radeon_gem_prime_import_sg_table()
71 RADEON_GEM_DOMAIN_GTT, 0, sg, resv, &bo); radeon_gem_prime_import_sg_table()
72 ww_mutex_unlock(&resv->lock); radeon_gem_prime_import_sg_table()
116 return bo->tbo.resv; radeon_gem_prime_res_obj()
H A Dradeon_benchmark.c38 struct reservation_object *resv) radeon_benchmark_do_move()
51 resv); radeon_benchmark_do_move()
56 resv); radeon_benchmark_do_move()
125 dobj->tbo.resv); radeon_benchmark_move()
136 dobj->tbo.resv); radeon_benchmark_move()
35 radeon_benchmark_do_move(struct radeon_device *rdev, unsigned size, uint64_t saddr, uint64_t daddr, int flag, int n, struct reservation_object *resv) radeon_benchmark_do_move() argument
H A Dradeon_sync.c84 * @resv: reservation object with embedded fence
91 struct reservation_object *resv, radeon_sync_resv()
101 f = reservation_object_get_excl(resv); radeon_sync_resv()
108 flist = reservation_object_get_list(resv); radeon_sync_resv()
114 reservation_object_held(resv)); radeon_sync_resv()
89 radeon_sync_resv(struct radeon_device *rdev, struct radeon_sync *sync, struct reservation_object *resv, bool shared) radeon_sync_resv() argument
H A Drv770_dma.c36 * @resv: reservation object to sync to
45 struct reservation_object *resv) rv770_copy_dma()
66 radeon_sync_resv(rdev, &sync, resv, false); rv770_copy_dma()
42 rv770_copy_dma(struct radeon_device *rdev, uint64_t src_offset, uint64_t dst_offset, unsigned num_gpu_pages, struct reservation_object *resv) rv770_copy_dma() argument
H A Dradeon_object.c181 struct reservation_object *resv, radeon_bo_create()
262 acc_size, sg, resv, &radeon_ttm_bo_destroy); radeon_bo_create()
599 lockdep_assert_held(&bo->tbo.resv->lock.base); radeon_bo_get_surface_reg()
725 lockdep_assert_held(&bo->tbo.resv->lock.base); radeon_bo_get_tiling_flags()
737 lockdep_assert_held(&bo->tbo.resv->lock.base); radeon_bo_check_tiling()
853 struct reservation_object *resv = bo->tbo.resv; radeon_bo_fence() local
856 reservation_object_add_shared_fence(resv, &fence->base); radeon_bo_fence()
858 reservation_object_add_excl_fence(resv, &fence->base); radeon_bo_fence()
178 radeon_bo_create(struct radeon_device *rdev, unsigned long size, int byte_align, bool kernel, u32 domain, u32 flags, struct sg_table *sg, struct reservation_object *resv, struct radeon_bo **bo_ptr) radeon_bo_create() argument
H A Devergreen_dma.c111 struct reservation_object *resv) evergreen_copy_dma()
132 radeon_sync_resv(rdev, &sync, resv, false); evergreen_copy_dma()
107 evergreen_copy_dma(struct radeon_device *rdev, uint64_t src_offset, uint64_t dst_offset, unsigned num_gpu_pages, struct reservation_object *resv) evergreen_copy_dma() argument
H A Dsi_dma.c225 * @resv: reservation object to sync to
234 struct reservation_object *resv) si_copy_dma()
255 radeon_sync_resv(rdev, &sync, resv, false); si_copy_dma()
231 si_copy_dma(struct radeon_device *rdev, uint64_t src_offset, uint64_t dst_offset, unsigned num_gpu_pages, struct reservation_object *resv) si_copy_dma() argument
H A Dradeon_test.c122 vram_obj->tbo.resv); radeon_do_test_moves()
126 vram_obj->tbo.resv); radeon_do_test_moves()
173 vram_obj->tbo.resv); radeon_do_test_moves()
177 vram_obj->tbo.resv); radeon_do_test_moves()
H A Dradeon_asic.h89 struct reservation_object *resv);
160 struct reservation_object *resv);
350 struct reservation_object *resv);
354 struct reservation_object *resv);
475 struct reservation_object *resv);
549 struct reservation_object *resv);
727 struct reservation_object *resv);
798 struct reservation_object *resv);
802 struct reservation_object *resv);
H A Dr600_dma.c433 * @resv: reservation object to sync to
442 struct reservation_object *resv) r600_copy_dma()
463 radeon_sync_resv(rdev, &sync, resv, false); r600_copy_dma()
439 r600_copy_dma(struct radeon_device *rdev, uint64_t src_offset, uint64_t dst_offset, unsigned num_gpu_pages, struct reservation_object *resv) r600_copy_dma() argument
H A Dradeon_cs.c239 struct reservation_object *resv; radeon_cs_sync_rings() local
241 resv = reloc->robj->tbo.resv; radeon_cs_sync_rings()
242 r = radeon_sync_resv(p->rdev, &p->ib.sync, resv, radeon_cs_sync_rings()
H A Dradeon_gem.c114 r = reservation_object_wait_timeout_rcu(robj->tbo.resv, true, true, 30 * HZ); radeon_gem_set_domain()
444 r = reservation_object_test_signaled_rcu(robj->tbo.resv, true); radeon_gem_busy_ioctl()
473 ret = reservation_object_wait_timeout_rcu(robj->tbo.resv, true, true, 30 * HZ); radeon_gem_wait_idle_ioctl()
H A Dradeon_mn.c154 r = reservation_object_wait_timeout_rcu(bo->tbo.resv, radeon_mn_invalidate_range_start()
H A Dradeon_object.h129 struct reservation_object *resv,
H A Dradeon.h619 struct reservation_object *resv,
1921 struct reservation_object *resv);
1927 struct reservation_object *resv);
1934 struct reservation_object *resv);
2747 #define radeon_copy_blit(rdev, s, d, np, resv) (rdev)->asic->copy.blit((rdev), (s), (d), (np), (resv))
2748 #define radeon_copy_dma(rdev, s, d, np, resv) (rdev)->asic->copy.dma((rdev), (s), (d), (np), (resv))
2749 #define radeon_copy(rdev, s, d, np, resv) (rdev)->asic->copy.copy((rdev), (s), (d), (np), (resv))
H A Dcik_sdma.c573 * @resv: reservation object to sync to
582 struct reservation_object *resv) cik_copy_dma()
603 radeon_sync_resv(rdev, &sync, resv, false); cik_copy_dma()
579 cik_copy_dma(struct radeon_device *rdev, uint64_t src_offset, uint64_t dst_offset, unsigned num_gpu_pages, struct reservation_object *resv) cik_copy_dma() argument
H A Dradeon_vm.c703 radeon_sync_resv(rdev, &ib.sync, pd->tbo.resv, true); radeon_vm_update_page_directory()
831 radeon_sync_resv(rdev, &ib->sync, pt->tbo.resv, true); radeon_vm_update_ptes()
832 r = reservation_object_reserve_shared(pt->tbo.resv); radeon_vm_update_ptes()
H A Dr200.c87 struct reservation_object *resv) r200_copy_dma()
83 r200_copy_dma(struct radeon_device *rdev, uint64_t src_offset, uint64_t dst_offset, unsigned num_gpu_pages, struct reservation_object *resv) r200_copy_dma() argument
H A Dradeon_uvd.c436 f = reservation_object_get_excl(bo->tbo.resv); radeon_uvd_cs_msg()
H A Dradeon_ttm.c299 fence = radeon_copy(rdev, old_start, new_start, num_pages, bo->resv); radeon_move_blit()
H A Dradeon_display.c559 work->fence = fence_get(reservation_object_get_excl(new_rbo->tbo.resv)); radeon_crtc_page_flip()
H A Dr600.c2964 struct reservation_object *resv) r600_copy_cpdma()
2985 radeon_sync_resv(rdev, &sync, resv, false); r600_copy_cpdma()
2961 r600_copy_cpdma(struct radeon_device *rdev, uint64_t src_offset, uint64_t dst_offset, unsigned num_gpu_pages, struct reservation_object *resv) r600_copy_cpdma() argument
H A Dcik.c4061 * @resv: reservation object to sync to
4070 struct reservation_object *resv) cik_copy_cpdma()
4091 radeon_sync_resv(rdev, &sync, resv, false); cik_copy_cpdma()
4067 cik_copy_cpdma(struct radeon_device *rdev, uint64_t src_offset, uint64_t dst_offset, unsigned num_gpu_pages, struct reservation_object *resv) cik_copy_cpdma() argument
H A Dr100.c886 struct reservation_object *resv) r100_copy_blit()
882 r100_copy_blit(struct radeon_device *rdev, uint64_t src_offset, uint64_t dst_offset, unsigned num_gpu_pages, struct reservation_object *resv) r100_copy_blit() argument
/linux-4.4.14/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_prime.c64 struct reservation_object *resv = attach->dmabuf->resv; amdgpu_gem_prime_import_sg_table() local
69 ww_mutex_lock(&resv->lock, NULL); amdgpu_gem_prime_import_sg_table()
71 AMDGPU_GEM_DOMAIN_GTT, 0, sg, resv, &bo); amdgpu_gem_prime_import_sg_table()
72 ww_mutex_unlock(&resv->lock); amdgpu_gem_prime_import_sg_table()
115 return bo->tbo.resv; amdgpu_gem_prime_res_obj()
H A Damdgpu_sync.c159 * @resv: reservation object with embedded fence
166 struct reservation_object *resv, amdgpu_sync_resv()
175 if (resv == NULL) amdgpu_sync_resv()
179 f = reservation_object_get_excl(resv); amdgpu_sync_resv()
182 flist = reservation_object_get_list(resv); amdgpu_sync_resv()
188 reservation_object_held(resv)); amdgpu_sync_resv()
164 amdgpu_sync_resv(struct amdgpu_device *adev, struct amdgpu_sync *sync, struct reservation_object *resv, void *owner) amdgpu_sync_resv() argument
H A Damdgpu_object.c222 struct reservation_object *resv, amdgpu_bo_create_restricted()
276 acc_size, sg, resv, &amdgpu_ttm_bo_destroy); amdgpu_bo_create_restricted()
291 struct reservation_object *resv, amdgpu_bo_create()
305 resv, bo_ptr); amdgpu_bo_create()
529 lockdep_assert_held(&bo->tbo.resv->lock.base); amdgpu_bo_get_tiling_flags()
662 struct reservation_object *resv = bo->tbo.resv; amdgpu_bo_fence() local
665 reservation_object_add_shared_fence(resv, fence); amdgpu_bo_fence()
667 reservation_object_add_excl_fence(resv, fence); amdgpu_bo_fence()
217 amdgpu_bo_create_restricted(struct amdgpu_device *adev, unsigned long size, int byte_align, bool kernel, u32 domain, u64 flags, struct sg_table *sg, struct ttm_placement *placement, struct reservation_object *resv, struct amdgpu_bo **bo_ptr) amdgpu_bo_create_restricted() argument
287 amdgpu_bo_create(struct amdgpu_device *adev, unsigned long size, int byte_align, bool kernel, u32 domain, u64 flags, struct sg_table *sg, struct reservation_object *resv, struct amdgpu_bo **bo_ptr) amdgpu_bo_create() argument
H A Damdgpu_object.h132 struct reservation_object *resv,
139 struct reservation_object *resv,
H A Damdgpu_vm.c348 r = reservation_object_reserve_shared(bo->tbo.resv); amdgpu_vm_clear_bo()
499 amdgpu_sync_resv(adev, &ib->sync, pd->tbo.resv, AMDGPU_FENCE_OWNER_VM); amdgpu_vm_update_page_directory()
646 amdgpu_sync_resv(adev, &ib->sync, pt->tbo.resv, owner); amdgpu_vm_update_ptes()
647 r = reservation_object_reserve_shared(pt->tbo.resv); amdgpu_vm_update_ptes()
1072 struct reservation_object *resv = vm->page_directory->tbo.resv; amdgpu_vm_bo_map() local
1082 NULL, resv, &pt); amdgpu_vm_bo_map()
H A Damdgpu_mn.c155 r = reservation_object_wait_timeout_rcu(bo->tbo.resv, amdgpu_mn_invalidate_range_start()
H A Damdgpu_ttm.c270 bo->resv, &fence); amdgpu_move_blit()
1014 struct reservation_object *resv, amdgpu_copy_buffer()
1044 if (resv) { amdgpu_copy_buffer()
1045 r = amdgpu_sync_resv(adev, &ib->sync, resv, amdgpu_copy_buffer()
1010 amdgpu_copy_buffer(struct amdgpu_ring *ring, uint64_t src_offset, uint64_t dst_offset, uint32_t byte_count, struct reservation_object *resv, struct fence **fence) amdgpu_copy_buffer() argument
H A Damdgpu_gem.c376 ret = reservation_object_test_signaled_rcu(robj->tbo.resv, true); amdgpu_gem_wait_idle_ioctl()
378 ret = reservation_object_wait_timeout_rcu(robj->tbo.resv, true, true, timeout); amdgpu_gem_wait_idle_ioctl()
H A Damdgpu_cs.c446 struct reservation_object *resv = e->robj->tbo.resv; amdgpu_cs_sync_rings() local
447 r = amdgpu_sync_resv(p->adev, &p->ibs[0].sync, resv, p->filp); amdgpu_cs_sync_rings()
H A Damdgpu_uvd.c533 r = reservation_object_wait_timeout_rcu(bo->tbo.resv, true, false, amdgpu_uvd_cs_msg()
H A Damdgpu.h477 struct reservation_object *resv,
673 struct reservation_object *resv,
H A Damdgpu_display.c222 r = reservation_object_get_fences_rcu(new_rbo->tbo.resv, &work->excl, amdgpu_crtc_page_flip()
/linux-4.4.14/drivers/gpu/drm/ttm/
H A Dttm_execbuf_util.c134 ret = reservation_object_reserve_shared(bo->resv); list_for_each_entry()
146 ret = ww_mutex_lock_slow_interruptible(&bo->resv->lock, list_for_each_entry()
149 ww_mutex_lock_slow(&bo->resv->lock, ticket); list_for_each_entry()
154 ret = reservation_object_reserve_shared(bo->resv); list_for_each_entry()
204 reservation_object_add_shared_fence(bo->resv, fence); list_for_each_entry()
206 reservation_object_add_excl_fence(bo->resv, fence); list_for_each_entry()
H A Dttm_bo.c153 if (bo->resv == &bo->ttm_resv) ttm_bo_release_list()
169 lockdep_assert_held(&bo->resv->lock.base); ttm_bo_add_to_lru()
403 ww_mutex_unlock (&bo->resv->lock); ttm_bo_cleanup_memtype_use()
412 fobj = reservation_object_get_list(bo->resv); ttm_bo_flush_all_fences()
413 fence = reservation_object_get_excl(bo->resv); ttm_bo_flush_all_fences()
419 reservation_object_held(bo->resv)); ttm_bo_flush_all_fences()
494 ww_mutex_unlock(&bo->resv->lock); ttm_bo_cleanup_refs_and_unlock()
497 lret = reservation_object_wait_timeout_rcu(bo->resv, ttm_bo_cleanup_refs_and_unlock()
676 lockdep_assert_held(&bo->resv->lock.base); ttm_bo_evict()
977 lockdep_assert_held(&bo->resv->lock.base); ttm_bo_move_buffer()
1050 lockdep_assert_held(&bo->resv->lock.base); ttm_bo_validate()
1089 struct reservation_object *resv, ttm_bo_init()
1143 if (resv) { ttm_bo_init()
1144 bo->resv = resv; ttm_bo_init()
1145 lockdep_assert_held(&bo->resv->lock.base); ttm_bo_init()
1147 bo->resv = &bo->ttm_resv; ttm_bo_init()
1165 if (!resv) { ttm_bo_init()
1166 locked = ww_mutex_trylock(&bo->resv->lock); ttm_bo_init()
1173 if (!resv) ttm_bo_init()
1542 struct reservation_object *resv; ttm_bo_wait() local
1547 resv = bo->resv; ttm_bo_wait()
1548 fobj = reservation_object_get_list(resv); ttm_bo_wait()
1549 excl = reservation_object_get_excl(resv); ttm_bo_wait()
1563 reservation_object_held(resv)); ttm_bo_wait()
1580 reservation_object_add_excl_fence(resv, NULL); ttm_bo_wait()
1724 if (!ww_mutex_is_locked(&bo->resv->lock)) ttm_bo_wait_unreserved()
1079 ttm_bo_init(struct ttm_bo_device *bdev, struct ttm_buffer_object *bo, unsigned long size, enum ttm_bo_type type, struct ttm_placement *placement, uint32_t page_alignment, bool interruptible, struct file *persistent_swap_storage, size_t acc_size, struct sg_table *sg, struct reservation_object *resv, void (*destroy) (struct ttm_buffer_object *)) ttm_bo_init() argument
H A Dttm_bo_util.c472 fbo->resv = &fbo->ttm_resv; ttm_buffer_object_transfer()
473 reservation_object_init(fbo->resv); ttm_buffer_object_transfer()
474 ret = ww_mutex_trylock(&fbo->resv->lock); ttm_buffer_object_transfer()
646 reservation_object_add_excl_fence(bo->resv, fence); ttm_bo_move_accel_cleanup()
674 reservation_object_add_excl_fence(ghost_obj->resv, fence); ttm_bo_move_accel_cleanup()
/linux-4.4.14/net/sunrpc/
H A Dsvc.c1069 svc_process_common(struct svc_rqst *rqstp, struct kvec *argv, struct kvec *resv) svc_process_common() argument
1096 svc_putu32(resv, rqstp->rq_xid); svc_process_common()
1101 svc_putnl(resv, 1); /* REPLY */ svc_process_common()
1107 reply_statp = resv->iov_base + resv->iov_len; svc_process_common()
1109 svc_putnl(resv, 0); /* ACCEPT */ svc_process_common()
1165 statp = resv->iov_base +resv->iov_len; svc_process_common()
1166 svc_putnl(resv, RPC_SUCCESS); svc_process_common()
1198 !xdr(rqstp, resv->iov_base+resv->iov_len, rqstp->rq_resp)) { svc_process_common()
1215 resv->iov_len = ((void*)statp) - resv->iov_base + 4; svc_process_common()
1242 svc_putnl(resv, 1); /* REJECT */ svc_process_common()
1243 svc_putnl(resv, 0); /* RPC_MISMATCH */ svc_process_common()
1244 svc_putnl(resv, 2); /* Only RPCv2 supported */ svc_process_common()
1245 svc_putnl(resv, 2); svc_process_common()
1253 svc_putnl(resv, 1); /* REJECT */ svc_process_common()
1254 svc_putnl(resv, 1); /* AUTH_ERROR */ svc_process_common()
1255 svc_putnl(resv, ntohl(auth_stat)); /* status */ svc_process_common()
1261 svc_putnl(resv, RPC_PROG_UNAVAIL); svc_process_common()
1269 svc_putnl(resv, RPC_PROG_MISMATCH); svc_process_common()
1270 svc_putnl(resv, progp->pg_lovers); svc_process_common()
1271 svc_putnl(resv, progp->pg_hivers); svc_process_common()
1278 svc_putnl(resv, RPC_PROC_UNAVAIL); svc_process_common()
1287 svc_putnl(resv, ntohl(rpc_stat)); svc_process_common()
1298 struct kvec *resv = &rqstp->rq_res.head[0]; svc_process() local
1307 resv->iov_base = page_address(rqstp->rq_respages[0]); svc_process()
1308 resv->iov_len = 0; svc_process()
1326 if (likely(svc_process_common(rqstp, argv, resv))) { svc_process()
1349 struct kvec *resv = &rqstp->rq_res.head[0]; bc_svc_process() local
1381 resv->iov_len = 0; bc_svc_process()
1391 proc_error = svc_process_common(rqstp, argv, resv); bc_svc_process()
H A Dsvcauth_unix.c728 struct kvec *resv = &rqstp->rq_res.head[0]; svcauth_null_accept() local
757 svc_putnl(resv, RPC_AUTH_NULL); svcauth_null_accept()
758 svc_putnl(resv, 0); svcauth_null_accept()
792 struct kvec *resv = &rqstp->rq_res.head[0]; svcauth_unix_accept() local
836 svc_putnl(resv, RPC_AUTH_NULL); svcauth_unix_accept()
837 svc_putnl(resv, 0); svcauth_unix_accept()
H A Dsvcsock.c1245 struct kvec *resv = &rqstp->rq_res.head[0]; svc_tcp_prep_reply_hdr() local
1248 svc_putnl(resv, 0); svc_tcp_prep_reply_hdr()
/linux-4.4.14/include/uapi/linux/
H A Digmp.h70 resv:4; member in struct:igmpv3_query
72 __u8 resv:4, member in struct:igmpv3_query
H A Dvfio.h191 __u32 resv; /* Reserved for alignment */ member in struct:vfio_region_info
/linux-4.4.14/net/sunrpc/auth_gss/
H A Dsvcauth_gss.c637 svc_safe_putnetobj(struct kvec *resv, struct xdr_netobj *o) svc_safe_putnetobj() argument
641 if (resv->iov_len + 4 > PAGE_SIZE) svc_safe_putnetobj()
643 svc_putnl(resv, o->len); svc_safe_putnetobj()
644 p = resv->iov_base + resv->iov_len; svc_safe_putnetobj()
645 resv->iov_len += round_up_to_quad(o->len); svc_safe_putnetobj()
646 if (resv->iov_len > PAGE_SIZE) svc_safe_putnetobj()
1074 gss_write_resv(struct kvec *resv, size_t size_limit, gss_write_resv() argument
1078 if (resv->iov_len + 4 > size_limit) gss_write_resv()
1080 svc_putnl(resv, RPC_SUCCESS); gss_write_resv()
1081 if (svc_safe_putnetobj(resv, out_handle)) gss_write_resv()
1083 if (resv->iov_len + 3 * 4 > size_limit) gss_write_resv()
1085 svc_putnl(resv, major_status); gss_write_resv()
1086 svc_putnl(resv, minor_status); gss_write_resv()
1087 svc_putnl(resv, GSS_SEQ_WIN); gss_write_resv()
1088 if (svc_safe_putnetobj(resv, out_token)) gss_write_resv()
1104 struct kvec *resv = &rqstp->rq_res.head[0]; svcauth_gss_legacy_init() local
1129 if (gss_write_resv(resv, PAGE_SIZE, svcauth_gss_legacy_init()
1212 struct kvec *resv = &rqstp->rq_res.head[0]; svcauth_gss_proxy_init() local
1257 if (gss_write_resv(resv, PAGE_SIZE, svcauth_gss_proxy_init()
1400 struct kvec *resv = &rqstp->rq_res.head[0]; svcauth_gss_accept() local
1406 __be32 *reject_stat = resv->iov_base + resv->iov_len; svcauth_gss_accept()
1486 if (resv->iov_len + 4 > PAGE_SIZE) svcauth_gss_accept()
1488 svc_putnl(resv, RPC_SUCCESS); svcauth_gss_accept()
1492 svcdata->verf_start = resv->iov_base + resv->iov_len; svcauth_gss_accept()
1503 svc_putnl(resv, 0); svcauth_gss_accept()
1504 svc_putnl(resv, 0); svcauth_gss_accept()
1512 svc_putnl(resv, 0); svcauth_gss_accept()
1513 svc_putnl(resv, 0); svcauth_gss_accept()
1585 struct kvec *resv; svcauth_gss_wrap_resp_integ() local
1607 resv = &resbuf->tail[0]; svcauth_gss_wrap_resp_integ()
1608 mic.data = (u8 *)resv->iov_base + resv->iov_len + 4; svcauth_gss_wrap_resp_integ()
1611 svc_putnl(resv, mic.len); svcauth_gss_wrap_resp_integ()
1614 resv->iov_len += XDR_QUADLEN(mic.len) << 2; svcauth_gss_wrap_resp_integ()
1617 BUG_ON(resv->iov_len > PAGE_SIZE); svcauth_gss_wrap_resp_integ()
/linux-4.4.14/drivers/dma-buf/
H A Ddma-buf.c73 if (dmabuf->resv == (struct reservation_object *)&dmabuf[1]) dma_buf_release()
74 reservation_object_fini(dmabuf->resv); dma_buf_release()
138 struct reservation_object *resv; dma_buf_poll() local
145 if (!dmabuf || !dmabuf->resv) dma_buf_poll()
148 resv = dmabuf->resv; dma_buf_poll()
157 seq = read_seqcount_begin(&resv->seq); dma_buf_poll()
160 fobj = rcu_dereference(resv->fence); dma_buf_poll()
165 fence_excl = rcu_dereference(resv->fence_excl); dma_buf_poll()
166 if (read_seqcount_retry(&resv->seq, seq)) { dma_buf_poll()
287 struct reservation_object *resv = exp_info->resv; dma_buf_export() local
291 if (!exp_info->resv) dma_buf_export()
294 /* prevent &dma_buf[1] == dma_buf->resv */ dma_buf_export()
326 if (!resv) { dma_buf_export()
327 resv = (struct reservation_object *)&dmabuf[1]; dma_buf_export()
328 reservation_object_init(resv); dma_buf_export()
330 dmabuf->resv = resv; dma_buf_export()
/linux-4.4.14/arch/sparc/include/asm/
H A Dvio.h43 u64 resv[6]; member in struct:vio_rdx
69 u16 resv; member in struct:vio_dring_register
77 u64 resv[5]; member in struct:vio_dring_unregister
177 u32 resv; member in struct:vio_disk_vtoc::__anon2729
198 u16 resv; member in struct:vio_disk_devid
240 u32 resv; member in struct:vio_net_mcast_info
H A Dhypervisor.h473 unsigned long resv; member in struct:hv_tsb_descr
/linux-4.4.14/drivers/gpu/drm/msm/
H A Dmsm_gem.h60 /* normally (resv == &_resv) except for imported bo's */
61 struct reservation_object *resv; member in struct:msm_gem_object
H A Dmsm_gem_submit.c130 ww_mutex_unlock(&msm_obj->resv->lock); submit_unlock_unpin_bo()
156 ret = ww_mutex_lock_interruptible(&msm_obj->resv->lock, submit_validate_objects()
202 ret = ww_mutex_lock_slow_interruptible(&msm_obj->resv->lock, submit_validate_objects()
H A Dmsm_gem.c548 if (msm_obj->resv == &msm_obj->_resv) msm_gem_free_object()
549 reservation_object_fini(msm_obj->resv); msm_gem_free_object()
623 msm_obj->resv = &msm_obj->_resv; msm_gem_new_impl()
624 reservation_object_init(msm_obj->resv); msm_gem_new_impl()
/linux-4.4.14/drivers/gpu/drm/nouveau/
H A Dnouveau_prime.c64 struct reservation_object *robj = attach->dmabuf->resv; nouveau_gem_prime_import_sg_table()
114 return nvbo->bo.resv; nouveau_gem_prime_res_obj()
H A Dnouveau_fence.c394 struct reservation_object *resv = nvbo->bo.resv; nouveau_fence_sync() local
400 ret = reservation_object_reserve_shared(resv); nouveau_fence_sync()
406 fobj = reservation_object_get_list(resv); nouveau_fence_sync()
407 fence = reservation_object_get_excl(resv); nouveau_fence_sync()
436 reservation_object_held(resv)); nouveau_fence_sync()
H A Dnouveau_gem.c120 struct reservation_object *resv = nvbo->bo.resv; nouveau_gem_object_unmap() local
124 fobj = reservation_object_get_list(resv); nouveau_gem_object_unmap()
132 reservation_object_held(resv)); nouveau_gem_object_unmap()
134 fence = reservation_object_get_excl(nvbo->bo.resv); nouveau_gem_object_unmap()
873 ret = reservation_object_test_signaled_rcu(nvbo->bo.resv, write) ? 0 : -EBUSY; nouveau_gem_ioctl_cpu_prep()
877 lret = reservation_object_wait_timeout_rcu(nvbo->bo.resv, write, true, 30 * HZ); nouveau_gem_ioctl_cpu_prep()
H A Dnouveau_bo.c1276 struct fence *fence = reservation_object_get_excl(bo->resv); nouveau_bo_vm_cleanup()
1593 struct reservation_object *resv = nvbo->bo.resv; nouveau_bo_fence() local
1596 reservation_object_add_excl_fence(resv, &fence->base); nouveau_bo_fence()
1598 reservation_object_add_shared_fence(resv, &fence->base); nouveau_bo_fence()
/linux-4.4.14/arch/sparc/include/uapi/asm/
H A Dpsr.h17 * | impl | vers | icc | resv | EC | EF | PIL | S | PS | ET | CWP |
/linux-4.4.14/arch/nios2/mm/
H A Dpgtable.c29 * resv |way |rd | we|pid |dbl|bad|perm|d
/linux-4.4.14/mm/
H A Dhugetlb.c255 static long region_add(struct resv_map *resv, long f, long t) region_add() argument
257 struct list_head *head = &resv->regions; region_add()
261 spin_lock(&resv->lock); region_add()
274 VM_BUG_ON(resv->region_cache_count <= 0); list_for_each_entry()
276 resv->region_cache_count--; list_for_each_entry()
277 nrg = list_first_entry(&resv->region_cache, struct file_region, list_for_each_entry()
323 resv->adds_in_progress--;
324 spin_unlock(&resv->lock);
351 static long region_chg(struct resv_map *resv, long f, long t) region_chg() argument
353 struct list_head *head = &resv->regions; region_chg()
358 spin_lock(&resv->lock); region_chg()
360 resv->adds_in_progress++; region_chg()
366 if (resv->adds_in_progress > resv->region_cache_count) { region_chg()
369 VM_BUG_ON(resv->adds_in_progress - resv->region_cache_count > 1); region_chg()
371 resv->adds_in_progress--; region_chg()
372 spin_unlock(&resv->lock); region_chg()
380 spin_lock(&resv->lock); region_chg()
381 list_add(&trg->link, &resv->region_cache); region_chg()
382 resv->region_cache_count++; region_chg()
396 resv->adds_in_progress--; list_for_each_entry()
397 spin_unlock(&resv->lock); list_for_each_entry()
436 spin_unlock(&resv->lock);
441 spin_unlock(&resv->lock);
456 static void region_abort(struct resv_map *resv, long f, long t) region_abort() argument
458 spin_lock(&resv->lock); region_abort()
459 VM_BUG_ON(!resv->region_cache_count); region_abort()
460 resv->adds_in_progress--; region_abort()
461 spin_unlock(&resv->lock); region_abort()
478 static long region_del(struct resv_map *resv, long f, long t) region_del() argument
480 struct list_head *head = &resv->regions; region_del()
486 spin_lock(&resv->lock); list_for_each_entry_safe()
507 resv->region_cache_count > resv->adds_in_progress) { list_for_each_entry_safe()
508 nrg = list_first_entry(&resv->region_cache, list_for_each_entry_safe()
512 resv->region_cache_count--; list_for_each_entry_safe()
516 spin_unlock(&resv->lock); list_for_each_entry_safe()
554 spin_unlock(&resv->lock);
585 static long region_count(struct resv_map *resv, long f, long t) region_count() argument
587 struct list_head *head = &resv->regions; region_count()
591 spin_lock(&resv->lock); region_count()
607 spin_unlock(&resv->lock);
1788 struct resv_map *resv; __vma_reservation_common() local
1792 resv = vma_resv_map(vma); __vma_reservation_common()
1793 if (!resv) __vma_reservation_common()
1799 ret = region_chg(resv, idx, idx + 1); __vma_reservation_common()
1802 ret = region_add(resv, idx, idx + 1); __vma_reservation_common()
1805 region_abort(resv, idx, idx + 1); __vma_reservation_common()
2963 struct resv_map *resv = vma_resv_map(vma); hugetlb_vm_op_open() local
2973 if (resv && is_vma_resv_set(vma, HPAGE_RESV_OWNER)) hugetlb_vm_op_open()
2974 kref_get(&resv->refs); hugetlb_vm_op_open()
2980 struct resv_map *resv = vma_resv_map(vma); hugetlb_vm_op_close() local
2985 if (!resv || !is_vma_resv_set(vma, HPAGE_RESV_OWNER)) hugetlb_vm_op_close()
2991 reserve = (end - start) - region_count(resv, start, end); hugetlb_vm_op_close()
2993 kref_put(&resv->refs, resv_map_release); hugetlb_vm_op_close()
H A Dcma.c121 * simple by forcing the entire CMA resv range cma_activate_area()
/linux-4.4.14/include/rdma/
H A Dib_smi.h55 __be16 resv; member in struct:ib_smp
118 u8 resv; member in struct:ib_port_info
H A Dib_pma.h64 u8 counter_width; /* resv: 7:3, counter width: 2:0 */
H A Dopa_smi.h61 __be16 resv; member in struct:opa_smp
H A Dib_mad.h173 __be16 resv; member in struct:ib_mad_hdr
/linux-4.4.14/include/linux/
H A Ddma-buf.h122 * @resv: reservation object linked to this dma-buf
137 struct reservation_object *resv; member in struct:dma_buf
175 * @resv: reservation-object, NULL to allocate default one
187 struct reservation_object *resv; member in struct:dma_buf_export_info
/linux-4.4.14/drivers/gpu/drm/qxl/
H A Dqxl_debugfs.c65 fobj = rcu_dereference(bo->tbo.resv->fence); qxl_debugfs_buffers_info()
H A Dqxl_release.c246 ret = reservation_object_reserve_shared(bo->tbo.resv); qxl_release_validate_bo()
472 reservation_object_add_shared_fence(bo->resv, &release->base); qxl_release_fence_buffer_objects()
/linux-4.4.14/drivers/nvme/host/
H A Dlightnvm.c66 __le64 resv; member in struct:nvme_nvm_ph_rw
133 __le64 resv; member in struct:nvme_nvm_erase_blk
199 __u8 resv[228]; member in struct:nvme_nvm_id
H A Dscsi.c2364 u32 resv; member in struct:scsi_unmap_blk_desc
2370 u32 resv; member in struct:scsi_unmap_parm_list
/linux-4.4.14/fs/nfsd/
H A Dnfscache.c513 struct kvec *resv = &rqstp->rq_res.head[0], *cachv; nfsd_cache_update() local
525 len = resv->iov_len - ((char*)statp - (char*)resv->iov_base); nfsd_cache_update()
/linux-4.4.14/include/drm/ttm/
H A Dttm_bo_driver.h790 success = ww_mutex_trylock(&bo->resv->lock); __ttm_bo_reserve()
795 ret = ww_mutex_lock_interruptible(&bo->resv->lock, ticket); __ttm_bo_reserve()
797 ret = ww_mutex_lock(&bo->resv->lock, ticket); __ttm_bo_reserve()
883 ret = ww_mutex_lock_slow_interruptible(&bo->resv->lock, ttm_bo_reserve_slowpath()
886 ww_mutex_lock_slow(&bo->resv->lock, ticket); ttm_bo_reserve_slowpath()
905 ww_mutex_unlock(&bo->resv->lock); __ttm_bo_unreserve()
H A Dttm_bo_api.h215 * Members protected by the bo::resv::reserved lock.
257 struct reservation_object *resv; member in struct:ttm_buffer_object
463 * @resv: Pointer to a reservation_object, or NULL to let ttm allocate one.
491 struct reservation_object *resv,
/linux-4.4.14/drivers/input/mouse/
H A Dcyapa_gen6.c47 u8 resv; /* Reserved, must be 0 */ member in struct:pip_app_cmd_head
48 u8 cmd_code; /* bit7: resv, set to 0; bit6~0: command code.*/
54 u8 resv; /* Reserved, must be 0 */ member in struct:pip_app_resp_head
/linux-4.4.14/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_resource.c578 return reservation_object_test_signaled_rcu(bo->resv, true) ? 0 : -EBUSY; vmw_user_dmabuf_synccpu_grab()
580 lret = reservation_object_wait_timeout_rcu(bo->resv, true, true, MAX_SCHEDULE_TIMEOUT); vmw_user_dmabuf_synccpu_grab()
1186 lockdep_assert_held(&res->backup->base.resv->lock.base); vmw_resource_unreserve()
1193 lockdep_assert_held(&new_backup->base.resv->lock.base); vmw_resource_unreserve()
1458 reservation_object_add_excl_fence(bo->resv, &fence->base); vmw_fence_single_bo()
1461 reservation_object_add_excl_fence(bo->resv, &fence->base); vmw_fence_single_bo()
H A Dvmwgfx_dmabuf.c276 lockdep_assert_held(&bo->resv->lock.base); vmw_bo_pin_reserved()
H A Dvmwgfx_cotable.c171 lockdep_assert_held(&bo->resv->lock.base); vmw_cotable_unscrub()
319 lockdep_assert_held(&bo->resv->lock.base); vmw_cotable_unbind()
/linux-4.4.14/drivers/gpu/drm/virtio/
H A Dvirtgpu_ioctl.c397 reservation_object_add_excl_fence(qobj->tbo.resv, virtio_gpu_transfer_from_host_ioctl()
446 reservation_object_add_excl_fence(qobj->tbo.resv, virtio_gpu_transfer_to_host_ioctl()
H A Dvirtgpu_display.c95 reservation_object_add_excl_fence(qobj->tbo.resv, virtio_gpu_crtc_cursor_set()
/linux-4.4.14/drivers/net/ethernet/sun/
H A Dsunqe.h257 #define MREGS_UTEST_RTRENAB 0x80 /* Enable resv test register */
258 #define MREGS_UTEST_RTRDISAB 0x40 /* Disab resv test register */
H A Dniu.h2821 __le64 resv; member in struct:rx_pkt_hdr0::tx_pkt_hdr
H A Dniu.c6676 tp->resv = 0; niu_start_xmit()
/linux-4.4.14/drivers/infiniband/hw/qib/
H A Dqib_mad.h128 __be16 resv; member in struct:ib_cc_mad
/linux-4.4.14/drivers/net/ethernet/rocker/
H A Drocker.h110 u16 resv[5]; member in struct:rocker_desc
/linux-4.4.14/arch/sparc/kernel/
H A Dpci_fire.c99 u64 resv[6]; member in struct:pci_msiq_entry
H A Dmdesc.c53 u16 resv; member in struct:mdesc_elem
/linux-4.4.14/arch/sparc/mm/
H A Dtsb.c259 hp->resv = 0; setup_tsb_params()
H A Dinit_64.c1924 ktsb_descr[0].resv = 0; sun4v_ktsb_init()
1941 ktsb_descr[1].resv = 0; sun4v_ktsb_init()
/linux-4.4.14/drivers/staging/rdma/ehca/
H A Dehca_reqs.c128 "attr_id=%x resv=%x attr_mod=%x", trace_ud_wr()
134 mad_hdr->resv, trace_ud_wr()
/linux-4.4.14/drivers/gpu/drm/
H A Ddrm_prime.c352 exp_info.resv = dev->driver->gem_prime_res_obj(obj); drm_gem_prime_export()
/linux-4.4.14/fs/xfs/
H A Dxfs_trans_dquot.c694 * Note that q_res_bcount = q_core.d_bcount + resv xfs_trans_dqresv()
H A Dxfs_ioctl.c675 * length of <= 0 for resv/unresv/zero is invalid. length for xfs_ioc_space()
/linux-4.4.14/fs/ceph/
H A Dcaps.c198 dout("reserve caps ctx=%p %d = %d used + %d resv + %d avail\n", ceph_reserve_caps()
213 dout("unreserve caps %d = %d used + %d resv + %d avail\n", ceph_unreserve_caps()
242 dout("get_cap ctx=%p (%d) %d = %d used + %d resv + %d avail\n", ceph_get_cap()
265 dout("put_cap %p %d = %d used + %d resv + %d avail\n", ceph_put_cap()
/linux-4.4.14/include/linux/mlx4/
H A Ddevice.h969 __be16 resv; member in struct:mlx4_mad_ifc
/linux-4.4.14/drivers/scsi/bfa/
H A Dbfa_defs_svc.h1054 u8 resv[2]; member in struct:bfa_fcport_fcf_s
/linux-4.4.14/drivers/scsi/
H A Dipr.h1250 u8 resv; member in struct:ipr_sdt_entry
/linux-4.4.14/drivers/iommu/
H A Damd_iommu_init.c126 u64 resv; member in struct:ivmd_header
/linux-4.4.14/drivers/scsi/lpfc/
H A Dlpfc_hw.h973 uint8_t resv; member in struct:_D_ID::__anon9617::__anon9618
981 uint8_t resv;
H A Dlpfc_els.c5322 switch (rscn_did.un.b.resv & RSCN_ADDRESS_FORMAT_MASK) { lpfc_rscn_payload_check()

Completed in 3071 milliseconds