/linux-4.4.14/fs/ocfs2/ |
H A D | reservations.c | 53 struct ocfs2_alloc_reservation *resv) ocfs2_resv_window_bits() 58 if (!(resv->r_flags & OCFS2_RESV_FLAG_DIR)) { ocfs2_resv_window_bits() 67 static inline unsigned int ocfs2_resv_end(struct ocfs2_alloc_reservation *resv) ocfs2_resv_end() argument 69 if (resv->r_len) ocfs2_resv_end() 70 return resv->r_start + resv->r_len - 1; ocfs2_resv_end() 71 return resv->r_start; ocfs2_resv_end() 74 static inline int ocfs2_resv_empty(struct ocfs2_alloc_reservation *resv) ocfs2_resv_empty() argument 76 return !!(resv->r_len == 0); ocfs2_resv_empty() 90 struct ocfs2_alloc_reservation *resv; ocfs2_dump_resv() local 98 resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); ocfs2_dump_resv() 101 "\tlast_len: %u\n", resv->r_start, ocfs2_dump_resv() 102 ocfs2_resv_end(resv), resv->r_len, resv->r_last_start, ocfs2_dump_resv() 103 resv->r_last_len); ocfs2_dump_resv() 112 list_for_each_entry(resv, &resmap->m_lru, r_lru) { ocfs2_dump_resv() 114 "last_start: %u\tlast_len: %u\n", i, resv->r_start, ocfs2_dump_resv() 115 ocfs2_resv_end(resv), resv->r_len, resv->r_last_start, ocfs2_dump_resv() 116 resv->r_last_len); ocfs2_dump_resv() 125 struct ocfs2_alloc_reservation *resv) ocfs2_validate_resmap_bits() 128 unsigned int start = resv->r_start; ocfs2_validate_resmap_bits() 129 unsigned int end = ocfs2_resv_end(resv); ocfs2_validate_resmap_bits() 149 struct ocfs2_alloc_reservation *resv; ocfs2_check_resmap() local 153 resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); ocfs2_check_resmap() 155 if (i > 0 && resv->r_start <= off) { ocfs2_check_resmap() 161 if (resv->r_len == 0) { ocfs2_check_resmap() 167 if (resv->r_start > ocfs2_resv_end(resv)) { ocfs2_check_resmap() 173 if (ocfs2_resv_end(resv) >= resmap->m_bitmap_len) { ocfs2_check_resmap() 179 if (ocfs2_validate_resmap_bits(resmap, i, resv)) ocfs2_check_resmap() 182 off = ocfs2_resv_end(resv); ocfs2_check_resmap() 200 void ocfs2_resv_init_once(struct ocfs2_alloc_reservation *resv) ocfs2_resv_init_once() argument 202 memset(resv, 0, sizeof(*resv)); ocfs2_resv_init_once() 203 INIT_LIST_HEAD(&resv->r_lru); ocfs2_resv_init_once() 206 void ocfs2_resv_set_type(struct ocfs2_alloc_reservation *resv, ocfs2_resv_set_type() argument 211 resv->r_flags |= flags; ocfs2_resv_set_type() 228 struct ocfs2_alloc_reservation *resv) ocfs2_resv_mark_lru() 232 if (!list_empty(&resv->r_lru)) ocfs2_resv_mark_lru() 233 list_del_init(&resv->r_lru); ocfs2_resv_mark_lru() 235 list_add_tail(&resv->r_lru, &resmap->m_lru); ocfs2_resv_mark_lru() 238 static void __ocfs2_resv_trunc(struct ocfs2_alloc_reservation *resv) __ocfs2_resv_trunc() argument 240 resv->r_len = 0; __ocfs2_resv_trunc() 241 resv->r_start = 0; __ocfs2_resv_trunc() 245 struct ocfs2_alloc_reservation *resv) ocfs2_resv_remove() 247 if (resv->r_flags & OCFS2_RESV_FLAG_INUSE) { ocfs2_resv_remove() 248 list_del_init(&resv->r_lru); ocfs2_resv_remove() 249 rb_erase(&resv->r_node, &resmap->m_reservations); ocfs2_resv_remove() 250 resv->r_flags &= ~OCFS2_RESV_FLAG_INUSE; ocfs2_resv_remove() 255 struct ocfs2_alloc_reservation *resv) __ocfs2_resv_discard() 259 __ocfs2_resv_trunc(resv); __ocfs2_resv_discard() 264 resv->r_last_len = resv->r_last_start = 0; __ocfs2_resv_discard() 266 ocfs2_resv_remove(resmap, resv); __ocfs2_resv_discard() 269 /* does nothing if 'resv' is null */ ocfs2_resv_discard() 271 struct ocfs2_alloc_reservation *resv) ocfs2_resv_discard() 273 if (resv) { ocfs2_resv_discard() 275 __ocfs2_resv_discard(resmap, resv); ocfs2_resv_discard() 283 struct ocfs2_alloc_reservation *resv; ocfs2_resmap_clear_all_resv() local 288 resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); ocfs2_resmap_clear_all_resv() 290 __ocfs2_resv_discard(resmap, resv); ocfs2_resmap_clear_all_resv() 369 struct ocfs2_alloc_reservation *resv = NULL; ocfs2_find_resv_lhs() local 380 resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); ocfs2_find_resv_lhs() 382 if (resv->r_start <= goal && ocfs2_resv_end(resv) >= goal) ocfs2_find_resv_lhs() 386 if (resv->r_start > goal) { ocfs2_find_resv_lhs() 387 resv = prev_resv; ocfs2_find_resv_lhs() 391 prev_resv = resv; ocfs2_find_resv_lhs() 395 return resv; ocfs2_find_resv_lhs() 471 struct ocfs2_alloc_reservation *resv, __ocfs2_resv_find_window() 489 trace_ocfs2_resv_find_window_begin(resv->r_start, ocfs2_resv_end(resv), __ocfs2_resv_find_window() 512 resv->r_start = cstart; __ocfs2_resv_find_window() 513 resv->r_len = clen; __ocfs2_resv_find_window() 515 ocfs2_resv_insert(resmap, resv); __ocfs2_resv_find_window() 624 resv->r_start = best_start; __ocfs2_resv_find_window() 625 resv->r_len = best_len; __ocfs2_resv_find_window() 626 ocfs2_resv_insert(resmap, resv); __ocfs2_resv_find_window() 631 struct ocfs2_alloc_reservation *resv, ocfs2_cannibalize_resv() 635 int tmpwindow = !!(resv->r_flags & OCFS2_RESV_FLAG_TMP); ocfs2_cannibalize_resv() 639 min_bits = ocfs2_resv_window_bits(resmap, resv) >> 1; ocfs2_cannibalize_resv() 667 resv->r_start = lru_resv->r_start; ocfs2_cannibalize_resv() 668 resv->r_len = lru_resv->r_len; ocfs2_cannibalize_resv() 680 resv->r_start = ocfs2_resv_end(lru_resv) + 1; ocfs2_cannibalize_resv() 681 resv->r_len = shrink; ocfs2_cannibalize_resv() 684 trace_ocfs2_cannibalize_resv_end(resv->r_start, ocfs2_resv_end(resv), ocfs2_cannibalize_resv() 685 resv->r_len, resv->r_last_start, ocfs2_cannibalize_resv() 686 resv->r_last_len); ocfs2_cannibalize_resv() 688 ocfs2_resv_insert(resmap, resv); ocfs2_cannibalize_resv() 692 struct ocfs2_alloc_reservation *resv, ocfs2_resv_find_window() 697 BUG_ON(!ocfs2_resv_empty(resv)); ocfs2_resv_find_window() 704 if (resv->r_last_len) { ocfs2_resv_find_window() 705 goal = resv->r_last_start + resv->r_last_len; ocfs2_resv_find_window() 710 __ocfs2_resv_find_window(resmap, resv, goal, wanted); ocfs2_resv_find_window() 713 if (ocfs2_resv_empty(resv) && goal != 0) ocfs2_resv_find_window() 714 __ocfs2_resv_find_window(resmap, resv, 0, wanted); ocfs2_resv_find_window() 716 if (ocfs2_resv_empty(resv)) { ocfs2_resv_find_window() 721 ocfs2_cannibalize_resv(resmap, resv, wanted); ocfs2_resv_find_window() 724 BUG_ON(ocfs2_resv_empty(resv)); ocfs2_resv_find_window() 728 struct ocfs2_alloc_reservation *resv, ocfs2_resmap_resv_bits() 731 if (resv == NULL || ocfs2_resmap_disabled(resmap)) ocfs2_resmap_resv_bits() 736 if (ocfs2_resv_empty(resv)) { ocfs2_resmap_resv_bits() 742 unsigned int wanted = ocfs2_resv_window_bits(resmap, resv); ocfs2_resmap_resv_bits() 744 if ((resv->r_flags & OCFS2_RESV_FLAG_TMP) || wanted < *clen) ocfs2_resmap_resv_bits() 754 ocfs2_resv_find_window(resmap, resv, wanted); ocfs2_resmap_resv_bits() 755 trace_ocfs2_resmap_resv_bits(resv->r_start, resv->r_len); ocfs2_resmap_resv_bits() 758 BUG_ON(ocfs2_resv_empty(resv)); ocfs2_resmap_resv_bits() 760 *cstart = resv->r_start; ocfs2_resmap_resv_bits() 761 *clen = resv->r_len; ocfs2_resmap_resv_bits() 769 struct ocfs2_alloc_reservation *resv, ocfs2_adjust_resv_from_alloc() 773 unsigned int old_end = ocfs2_resv_end(resv); ocfs2_adjust_resv_from_alloc() 775 BUG_ON(start != resv->r_start || old_end < end); ocfs2_adjust_resv_from_alloc() 781 __ocfs2_resv_discard(resmap, resv); ocfs2_adjust_resv_from_alloc() 792 resv->r_start = end + 1; ocfs2_adjust_resv_from_alloc() 793 resv->r_len = old_end - resv->r_start + 1; ocfs2_adjust_resv_from_alloc() 797 struct ocfs2_alloc_reservation *resv, ocfs2_resmap_claimed_bits() 805 if (resv == NULL) ocfs2_resmap_claimed_bits() 808 BUG_ON(cstart != resv->r_start); ocfs2_resmap_claimed_bits() 812 trace_ocfs2_resmap_claimed_bits_begin(cstart, cend, clen, resv->r_start, ocfs2_resmap_claimed_bits() 813 ocfs2_resv_end(resv), resv->r_len, ocfs2_resmap_claimed_bits() 814 resv->r_last_start, ocfs2_resmap_claimed_bits() 815 resv->r_last_len); ocfs2_resmap_claimed_bits() 817 BUG_ON(cstart < resv->r_start); ocfs2_resmap_claimed_bits() 818 BUG_ON(cstart > ocfs2_resv_end(resv)); ocfs2_resmap_claimed_bits() 819 BUG_ON(cend > ocfs2_resv_end(resv)); ocfs2_resmap_claimed_bits() 821 ocfs2_adjust_resv_from_alloc(resmap, resv, cstart, cend); ocfs2_resmap_claimed_bits() 822 resv->r_last_start = cstart; ocfs2_resmap_claimed_bits() 823 resv->r_last_len = clen; ocfs2_resmap_claimed_bits() 829 if (!ocfs2_resv_empty(resv)) ocfs2_resmap_claimed_bits() 830 ocfs2_resv_mark_lru(resmap, resv); ocfs2_resmap_claimed_bits() 832 trace_ocfs2_resmap_claimed_bits_end(resv->r_start, ocfs2_resv_end(resv), ocfs2_resmap_claimed_bits() 833 resv->r_len, resv->r_last_start, ocfs2_resmap_claimed_bits() 834 resv->r_last_len); ocfs2_resmap_claimed_bits() 52 ocfs2_resv_window_bits(struct ocfs2_reservation_map *resmap, struct ocfs2_alloc_reservation *resv) ocfs2_resv_window_bits() argument 123 ocfs2_validate_resmap_bits(struct ocfs2_reservation_map *resmap, int i, struct ocfs2_alloc_reservation *resv) ocfs2_validate_resmap_bits() argument 227 ocfs2_resv_mark_lru(struct ocfs2_reservation_map *resmap, struct ocfs2_alloc_reservation *resv) ocfs2_resv_mark_lru() argument 244 ocfs2_resv_remove(struct ocfs2_reservation_map *resmap, struct ocfs2_alloc_reservation *resv) ocfs2_resv_remove() argument 254 __ocfs2_resv_discard(struct ocfs2_reservation_map *resmap, struct ocfs2_alloc_reservation *resv) __ocfs2_resv_discard() argument 270 ocfs2_resv_discard(struct ocfs2_reservation_map *resmap, struct ocfs2_alloc_reservation *resv) ocfs2_resv_discard() argument 470 __ocfs2_resv_find_window(struct ocfs2_reservation_map *resmap, struct ocfs2_alloc_reservation *resv, unsigned int goal, unsigned int wanted) __ocfs2_resv_find_window() argument 630 ocfs2_cannibalize_resv(struct ocfs2_reservation_map *resmap, struct ocfs2_alloc_reservation *resv, unsigned int wanted) ocfs2_cannibalize_resv() argument 691 ocfs2_resv_find_window(struct ocfs2_reservation_map *resmap, struct ocfs2_alloc_reservation *resv, unsigned int wanted) ocfs2_resv_find_window() argument 727 ocfs2_resmap_resv_bits(struct ocfs2_reservation_map *resmap, struct ocfs2_alloc_reservation *resv, int *cstart, int *clen) ocfs2_resmap_resv_bits() argument 768 ocfs2_adjust_resv_from_alloc(struct ocfs2_reservation_map *resmap, struct ocfs2_alloc_reservation *resv, unsigned int start, unsigned int end) ocfs2_adjust_resv_from_alloc() argument 796 ocfs2_resmap_claimed_bits(struct ocfs2_reservation_map *resmap, struct ocfs2_alloc_reservation *resv, u32 cstart, u32 clen) ocfs2_resmap_claimed_bits() argument
|
H A D | reservations.h | 64 void ocfs2_resv_init_once(struct ocfs2_alloc_reservation *resv); 67 void ocfs2_resv_set_type(struct ocfs2_alloc_reservation *resv, 75 * @resv: the reservation to truncate. 81 struct ocfs2_alloc_reservation *resv); 122 * @resv: reservation to base search from 126 * Using the reservation data from resv, this function will compare 128 * the reservation window is still clear to use. If resv is empty, 137 struct ocfs2_alloc_reservation *resv, 143 * @resv: optional reservation to recalulate based on new bitmap 151 * reservations bitmap. If resv is passed, it's next allocation window will be 156 struct ocfs2_alloc_reservation *resv,
|
H A D | localalloc.c | 56 struct ocfs2_alloc_reservation *resv); 839 struct ocfs2_alloc_reservation *resv) ocfs2_local_alloc_find_clear_bits() 852 if (!resv) { ocfs2_local_alloc_find_clear_bits() 856 resv = &r; ocfs2_local_alloc_find_clear_bits() 860 if (ocfs2_resmap_resv_bits(resmap, resv, &bitoff, &numfound) == 0) { ocfs2_local_alloc_find_clear_bits() 916 ocfs2_resv_discard(resmap, resv); ocfs2_local_alloc_find_clear_bits() 836 ocfs2_local_alloc_find_clear_bits(struct ocfs2_super *osb, struct ocfs2_dinode *alloc, u32 *numbits, struct ocfs2_alloc_reservation *resv) ocfs2_local_alloc_find_clear_bits() argument
|
/linux-4.4.14/drivers/gpu/drm/radeon/ |
H A D | radeon_prime.c | 64 struct reservation_object *resv = attach->dmabuf->resv; radeon_gem_prime_import_sg_table() local 69 ww_mutex_lock(&resv->lock, NULL); radeon_gem_prime_import_sg_table() 71 RADEON_GEM_DOMAIN_GTT, 0, sg, resv, &bo); radeon_gem_prime_import_sg_table() 72 ww_mutex_unlock(&resv->lock); radeon_gem_prime_import_sg_table() 116 return bo->tbo.resv; radeon_gem_prime_res_obj()
|
H A D | radeon_benchmark.c | 38 struct reservation_object *resv) radeon_benchmark_do_move() 51 resv); radeon_benchmark_do_move() 56 resv); radeon_benchmark_do_move() 125 dobj->tbo.resv); radeon_benchmark_move() 136 dobj->tbo.resv); radeon_benchmark_move() 35 radeon_benchmark_do_move(struct radeon_device *rdev, unsigned size, uint64_t saddr, uint64_t daddr, int flag, int n, struct reservation_object *resv) radeon_benchmark_do_move() argument
|
H A D | radeon_sync.c | 84 * @resv: reservation object with embedded fence 91 struct reservation_object *resv, radeon_sync_resv() 101 f = reservation_object_get_excl(resv); radeon_sync_resv() 108 flist = reservation_object_get_list(resv); radeon_sync_resv() 114 reservation_object_held(resv)); radeon_sync_resv() 89 radeon_sync_resv(struct radeon_device *rdev, struct radeon_sync *sync, struct reservation_object *resv, bool shared) radeon_sync_resv() argument
|
H A D | rv770_dma.c | 36 * @resv: reservation object to sync to 45 struct reservation_object *resv) rv770_copy_dma() 66 radeon_sync_resv(rdev, &sync, resv, false); rv770_copy_dma() 42 rv770_copy_dma(struct radeon_device *rdev, uint64_t src_offset, uint64_t dst_offset, unsigned num_gpu_pages, struct reservation_object *resv) rv770_copy_dma() argument
|
H A D | radeon_object.c | 181 struct reservation_object *resv, radeon_bo_create() 262 acc_size, sg, resv, &radeon_ttm_bo_destroy); radeon_bo_create() 599 lockdep_assert_held(&bo->tbo.resv->lock.base); radeon_bo_get_surface_reg() 725 lockdep_assert_held(&bo->tbo.resv->lock.base); radeon_bo_get_tiling_flags() 737 lockdep_assert_held(&bo->tbo.resv->lock.base); radeon_bo_check_tiling() 853 struct reservation_object *resv = bo->tbo.resv; radeon_bo_fence() local 856 reservation_object_add_shared_fence(resv, &fence->base); radeon_bo_fence() 858 reservation_object_add_excl_fence(resv, &fence->base); radeon_bo_fence() 178 radeon_bo_create(struct radeon_device *rdev, unsigned long size, int byte_align, bool kernel, u32 domain, u32 flags, struct sg_table *sg, struct reservation_object *resv, struct radeon_bo **bo_ptr) radeon_bo_create() argument
|
H A D | evergreen_dma.c | 111 struct reservation_object *resv) evergreen_copy_dma() 132 radeon_sync_resv(rdev, &sync, resv, false); evergreen_copy_dma() 107 evergreen_copy_dma(struct radeon_device *rdev, uint64_t src_offset, uint64_t dst_offset, unsigned num_gpu_pages, struct reservation_object *resv) evergreen_copy_dma() argument
|
H A D | si_dma.c | 225 * @resv: reservation object to sync to 234 struct reservation_object *resv) si_copy_dma() 255 radeon_sync_resv(rdev, &sync, resv, false); si_copy_dma() 231 si_copy_dma(struct radeon_device *rdev, uint64_t src_offset, uint64_t dst_offset, unsigned num_gpu_pages, struct reservation_object *resv) si_copy_dma() argument
|
H A D | radeon_test.c | 122 vram_obj->tbo.resv); radeon_do_test_moves() 126 vram_obj->tbo.resv); radeon_do_test_moves() 173 vram_obj->tbo.resv); radeon_do_test_moves() 177 vram_obj->tbo.resv); radeon_do_test_moves()
|
H A D | radeon_asic.h | 89 struct reservation_object *resv); 160 struct reservation_object *resv); 350 struct reservation_object *resv); 354 struct reservation_object *resv); 475 struct reservation_object *resv); 549 struct reservation_object *resv); 727 struct reservation_object *resv); 798 struct reservation_object *resv); 802 struct reservation_object *resv);
|
H A D | r600_dma.c | 433 * @resv: reservation object to sync to 442 struct reservation_object *resv) r600_copy_dma() 463 radeon_sync_resv(rdev, &sync, resv, false); r600_copy_dma() 439 r600_copy_dma(struct radeon_device *rdev, uint64_t src_offset, uint64_t dst_offset, unsigned num_gpu_pages, struct reservation_object *resv) r600_copy_dma() argument
|
H A D | radeon_cs.c | 239 struct reservation_object *resv; radeon_cs_sync_rings() local 241 resv = reloc->robj->tbo.resv; radeon_cs_sync_rings() 242 r = radeon_sync_resv(p->rdev, &p->ib.sync, resv, radeon_cs_sync_rings()
|
H A D | radeon_gem.c | 114 r = reservation_object_wait_timeout_rcu(robj->tbo.resv, true, true, 30 * HZ); radeon_gem_set_domain() 444 r = reservation_object_test_signaled_rcu(robj->tbo.resv, true); radeon_gem_busy_ioctl() 473 ret = reservation_object_wait_timeout_rcu(robj->tbo.resv, true, true, 30 * HZ); radeon_gem_wait_idle_ioctl()
|
H A D | radeon_mn.c | 154 r = reservation_object_wait_timeout_rcu(bo->tbo.resv, radeon_mn_invalidate_range_start()
|
H A D | radeon_object.h | 129 struct reservation_object *resv,
|
H A D | radeon.h | 619 struct reservation_object *resv, 1921 struct reservation_object *resv); 1927 struct reservation_object *resv); 1934 struct reservation_object *resv); 2747 #define radeon_copy_blit(rdev, s, d, np, resv) (rdev)->asic->copy.blit((rdev), (s), (d), (np), (resv)) 2748 #define radeon_copy_dma(rdev, s, d, np, resv) (rdev)->asic->copy.dma((rdev), (s), (d), (np), (resv)) 2749 #define radeon_copy(rdev, s, d, np, resv) (rdev)->asic->copy.copy((rdev), (s), (d), (np), (resv))
|
H A D | cik_sdma.c | 573 * @resv: reservation object to sync to 582 struct reservation_object *resv) cik_copy_dma() 603 radeon_sync_resv(rdev, &sync, resv, false); cik_copy_dma() 579 cik_copy_dma(struct radeon_device *rdev, uint64_t src_offset, uint64_t dst_offset, unsigned num_gpu_pages, struct reservation_object *resv) cik_copy_dma() argument
|
H A D | radeon_vm.c | 703 radeon_sync_resv(rdev, &ib.sync, pd->tbo.resv, true); radeon_vm_update_page_directory() 831 radeon_sync_resv(rdev, &ib->sync, pt->tbo.resv, true); radeon_vm_update_ptes() 832 r = reservation_object_reserve_shared(pt->tbo.resv); radeon_vm_update_ptes()
|
H A D | r200.c | 87 struct reservation_object *resv) r200_copy_dma() 83 r200_copy_dma(struct radeon_device *rdev, uint64_t src_offset, uint64_t dst_offset, unsigned num_gpu_pages, struct reservation_object *resv) r200_copy_dma() argument
|
H A D | radeon_uvd.c | 436 f = reservation_object_get_excl(bo->tbo.resv); radeon_uvd_cs_msg()
|
H A D | radeon_ttm.c | 299 fence = radeon_copy(rdev, old_start, new_start, num_pages, bo->resv); radeon_move_blit()
|
H A D | radeon_display.c | 559 work->fence = fence_get(reservation_object_get_excl(new_rbo->tbo.resv)); radeon_crtc_page_flip()
|
H A D | r600.c | 2964 struct reservation_object *resv) r600_copy_cpdma() 2985 radeon_sync_resv(rdev, &sync, resv, false); r600_copy_cpdma() 2961 r600_copy_cpdma(struct radeon_device *rdev, uint64_t src_offset, uint64_t dst_offset, unsigned num_gpu_pages, struct reservation_object *resv) r600_copy_cpdma() argument
|
H A D | cik.c | 4061 * @resv: reservation object to sync to 4070 struct reservation_object *resv) cik_copy_cpdma() 4091 radeon_sync_resv(rdev, &sync, resv, false); cik_copy_cpdma() 4067 cik_copy_cpdma(struct radeon_device *rdev, uint64_t src_offset, uint64_t dst_offset, unsigned num_gpu_pages, struct reservation_object *resv) cik_copy_cpdma() argument
|
H A D | r100.c | 886 struct reservation_object *resv) r100_copy_blit() 882 r100_copy_blit(struct radeon_device *rdev, uint64_t src_offset, uint64_t dst_offset, unsigned num_gpu_pages, struct reservation_object *resv) r100_copy_blit() argument
|
/linux-4.4.14/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_prime.c | 64 struct reservation_object *resv = attach->dmabuf->resv; amdgpu_gem_prime_import_sg_table() local 69 ww_mutex_lock(&resv->lock, NULL); amdgpu_gem_prime_import_sg_table() 71 AMDGPU_GEM_DOMAIN_GTT, 0, sg, resv, &bo); amdgpu_gem_prime_import_sg_table() 72 ww_mutex_unlock(&resv->lock); amdgpu_gem_prime_import_sg_table() 115 return bo->tbo.resv; amdgpu_gem_prime_res_obj()
|
H A D | amdgpu_sync.c | 159 * @resv: reservation object with embedded fence 166 struct reservation_object *resv, amdgpu_sync_resv() 175 if (resv == NULL) amdgpu_sync_resv() 179 f = reservation_object_get_excl(resv); amdgpu_sync_resv() 182 flist = reservation_object_get_list(resv); amdgpu_sync_resv() 188 reservation_object_held(resv)); amdgpu_sync_resv() 164 amdgpu_sync_resv(struct amdgpu_device *adev, struct amdgpu_sync *sync, struct reservation_object *resv, void *owner) amdgpu_sync_resv() argument
|
H A D | amdgpu_object.c | 222 struct reservation_object *resv, amdgpu_bo_create_restricted() 276 acc_size, sg, resv, &amdgpu_ttm_bo_destroy); amdgpu_bo_create_restricted() 291 struct reservation_object *resv, amdgpu_bo_create() 305 resv, bo_ptr); amdgpu_bo_create() 529 lockdep_assert_held(&bo->tbo.resv->lock.base); amdgpu_bo_get_tiling_flags() 662 struct reservation_object *resv = bo->tbo.resv; amdgpu_bo_fence() local 665 reservation_object_add_shared_fence(resv, fence); amdgpu_bo_fence() 667 reservation_object_add_excl_fence(resv, fence); amdgpu_bo_fence() 217 amdgpu_bo_create_restricted(struct amdgpu_device *adev, unsigned long size, int byte_align, bool kernel, u32 domain, u64 flags, struct sg_table *sg, struct ttm_placement *placement, struct reservation_object *resv, struct amdgpu_bo **bo_ptr) amdgpu_bo_create_restricted() argument 287 amdgpu_bo_create(struct amdgpu_device *adev, unsigned long size, int byte_align, bool kernel, u32 domain, u64 flags, struct sg_table *sg, struct reservation_object *resv, struct amdgpu_bo **bo_ptr) amdgpu_bo_create() argument
|
H A D | amdgpu_object.h | 132 struct reservation_object *resv, 139 struct reservation_object *resv,
|
H A D | amdgpu_vm.c | 348 r = reservation_object_reserve_shared(bo->tbo.resv); amdgpu_vm_clear_bo() 499 amdgpu_sync_resv(adev, &ib->sync, pd->tbo.resv, AMDGPU_FENCE_OWNER_VM); amdgpu_vm_update_page_directory() 646 amdgpu_sync_resv(adev, &ib->sync, pt->tbo.resv, owner); amdgpu_vm_update_ptes() 647 r = reservation_object_reserve_shared(pt->tbo.resv); amdgpu_vm_update_ptes() 1072 struct reservation_object *resv = vm->page_directory->tbo.resv; amdgpu_vm_bo_map() local 1082 NULL, resv, &pt); amdgpu_vm_bo_map()
|
H A D | amdgpu_mn.c | 155 r = reservation_object_wait_timeout_rcu(bo->tbo.resv, amdgpu_mn_invalidate_range_start()
|
H A D | amdgpu_ttm.c | 270 bo->resv, &fence); amdgpu_move_blit() 1014 struct reservation_object *resv, amdgpu_copy_buffer() 1044 if (resv) { amdgpu_copy_buffer() 1045 r = amdgpu_sync_resv(adev, &ib->sync, resv, amdgpu_copy_buffer() 1010 amdgpu_copy_buffer(struct amdgpu_ring *ring, uint64_t src_offset, uint64_t dst_offset, uint32_t byte_count, struct reservation_object *resv, struct fence **fence) amdgpu_copy_buffer() argument
|
H A D | amdgpu_gem.c | 376 ret = reservation_object_test_signaled_rcu(robj->tbo.resv, true); amdgpu_gem_wait_idle_ioctl() 378 ret = reservation_object_wait_timeout_rcu(robj->tbo.resv, true, true, timeout); amdgpu_gem_wait_idle_ioctl()
|
H A D | amdgpu_cs.c | 446 struct reservation_object *resv = e->robj->tbo.resv; amdgpu_cs_sync_rings() local 447 r = amdgpu_sync_resv(p->adev, &p->ibs[0].sync, resv, p->filp); amdgpu_cs_sync_rings()
|
H A D | amdgpu_uvd.c | 533 r = reservation_object_wait_timeout_rcu(bo->tbo.resv, true, false, amdgpu_uvd_cs_msg()
|
H A D | amdgpu.h | 477 struct reservation_object *resv, 673 struct reservation_object *resv,
|
H A D | amdgpu_display.c | 222 r = reservation_object_get_fences_rcu(new_rbo->tbo.resv, &work->excl, amdgpu_crtc_page_flip()
|
/linux-4.4.14/drivers/gpu/drm/ttm/ |
H A D | ttm_execbuf_util.c | 134 ret = reservation_object_reserve_shared(bo->resv); list_for_each_entry() 146 ret = ww_mutex_lock_slow_interruptible(&bo->resv->lock, list_for_each_entry() 149 ww_mutex_lock_slow(&bo->resv->lock, ticket); list_for_each_entry() 154 ret = reservation_object_reserve_shared(bo->resv); list_for_each_entry() 204 reservation_object_add_shared_fence(bo->resv, fence); list_for_each_entry() 206 reservation_object_add_excl_fence(bo->resv, fence); list_for_each_entry()
|
H A D | ttm_bo.c | 153 if (bo->resv == &bo->ttm_resv) ttm_bo_release_list() 169 lockdep_assert_held(&bo->resv->lock.base); ttm_bo_add_to_lru() 403 ww_mutex_unlock (&bo->resv->lock); ttm_bo_cleanup_memtype_use() 412 fobj = reservation_object_get_list(bo->resv); ttm_bo_flush_all_fences() 413 fence = reservation_object_get_excl(bo->resv); ttm_bo_flush_all_fences() 419 reservation_object_held(bo->resv)); ttm_bo_flush_all_fences() 494 ww_mutex_unlock(&bo->resv->lock); ttm_bo_cleanup_refs_and_unlock() 497 lret = reservation_object_wait_timeout_rcu(bo->resv, ttm_bo_cleanup_refs_and_unlock() 676 lockdep_assert_held(&bo->resv->lock.base); ttm_bo_evict() 977 lockdep_assert_held(&bo->resv->lock.base); ttm_bo_move_buffer() 1050 lockdep_assert_held(&bo->resv->lock.base); ttm_bo_validate() 1089 struct reservation_object *resv, ttm_bo_init() 1143 if (resv) { ttm_bo_init() 1144 bo->resv = resv; ttm_bo_init() 1145 lockdep_assert_held(&bo->resv->lock.base); ttm_bo_init() 1147 bo->resv = &bo->ttm_resv; ttm_bo_init() 1165 if (!resv) { ttm_bo_init() 1166 locked = ww_mutex_trylock(&bo->resv->lock); ttm_bo_init() 1173 if (!resv) ttm_bo_init() 1542 struct reservation_object *resv; ttm_bo_wait() local 1547 resv = bo->resv; ttm_bo_wait() 1548 fobj = reservation_object_get_list(resv); ttm_bo_wait() 1549 excl = reservation_object_get_excl(resv); ttm_bo_wait() 1563 reservation_object_held(resv)); ttm_bo_wait() 1580 reservation_object_add_excl_fence(resv, NULL); ttm_bo_wait() 1724 if (!ww_mutex_is_locked(&bo->resv->lock)) ttm_bo_wait_unreserved() 1079 ttm_bo_init(struct ttm_bo_device *bdev, struct ttm_buffer_object *bo, unsigned long size, enum ttm_bo_type type, struct ttm_placement *placement, uint32_t page_alignment, bool interruptible, struct file *persistent_swap_storage, size_t acc_size, struct sg_table *sg, struct reservation_object *resv, void (*destroy) (struct ttm_buffer_object *)) ttm_bo_init() argument
|
H A D | ttm_bo_util.c | 472 fbo->resv = &fbo->ttm_resv; ttm_buffer_object_transfer() 473 reservation_object_init(fbo->resv); ttm_buffer_object_transfer() 474 ret = ww_mutex_trylock(&fbo->resv->lock); ttm_buffer_object_transfer() 646 reservation_object_add_excl_fence(bo->resv, fence); ttm_bo_move_accel_cleanup() 674 reservation_object_add_excl_fence(ghost_obj->resv, fence); ttm_bo_move_accel_cleanup()
|
/linux-4.4.14/net/sunrpc/ |
H A D | svc.c | 1069 svc_process_common(struct svc_rqst *rqstp, struct kvec *argv, struct kvec *resv) svc_process_common() argument 1096 svc_putu32(resv, rqstp->rq_xid); svc_process_common() 1101 svc_putnl(resv, 1); /* REPLY */ svc_process_common() 1107 reply_statp = resv->iov_base + resv->iov_len; svc_process_common() 1109 svc_putnl(resv, 0); /* ACCEPT */ svc_process_common() 1165 statp = resv->iov_base +resv->iov_len; svc_process_common() 1166 svc_putnl(resv, RPC_SUCCESS); svc_process_common() 1198 !xdr(rqstp, resv->iov_base+resv->iov_len, rqstp->rq_resp)) { svc_process_common() 1215 resv->iov_len = ((void*)statp) - resv->iov_base + 4; svc_process_common() 1242 svc_putnl(resv, 1); /* REJECT */ svc_process_common() 1243 svc_putnl(resv, 0); /* RPC_MISMATCH */ svc_process_common() 1244 svc_putnl(resv, 2); /* Only RPCv2 supported */ svc_process_common() 1245 svc_putnl(resv, 2); svc_process_common() 1253 svc_putnl(resv, 1); /* REJECT */ svc_process_common() 1254 svc_putnl(resv, 1); /* AUTH_ERROR */ svc_process_common() 1255 svc_putnl(resv, ntohl(auth_stat)); /* status */ svc_process_common() 1261 svc_putnl(resv, RPC_PROG_UNAVAIL); svc_process_common() 1269 svc_putnl(resv, RPC_PROG_MISMATCH); svc_process_common() 1270 svc_putnl(resv, progp->pg_lovers); svc_process_common() 1271 svc_putnl(resv, progp->pg_hivers); svc_process_common() 1278 svc_putnl(resv, RPC_PROC_UNAVAIL); svc_process_common() 1287 svc_putnl(resv, ntohl(rpc_stat)); svc_process_common() 1298 struct kvec *resv = &rqstp->rq_res.head[0]; svc_process() local 1307 resv->iov_base = page_address(rqstp->rq_respages[0]); svc_process() 1308 resv->iov_len = 0; svc_process() 1326 if (likely(svc_process_common(rqstp, argv, resv))) { svc_process() 1349 struct kvec *resv = &rqstp->rq_res.head[0]; bc_svc_process() local 1381 resv->iov_len = 0; bc_svc_process() 1391 proc_error = svc_process_common(rqstp, argv, resv); bc_svc_process()
|
H A D | svcauth_unix.c | 728 struct kvec *resv = &rqstp->rq_res.head[0]; svcauth_null_accept() local 757 svc_putnl(resv, RPC_AUTH_NULL); svcauth_null_accept() 758 svc_putnl(resv, 0); svcauth_null_accept() 792 struct kvec *resv = &rqstp->rq_res.head[0]; svcauth_unix_accept() local 836 svc_putnl(resv, RPC_AUTH_NULL); svcauth_unix_accept() 837 svc_putnl(resv, 0); svcauth_unix_accept()
|
H A D | svcsock.c | 1245 struct kvec *resv = &rqstp->rq_res.head[0]; svc_tcp_prep_reply_hdr() local 1248 svc_putnl(resv, 0); svc_tcp_prep_reply_hdr()
|
/linux-4.4.14/include/uapi/linux/ |
H A D | igmp.h | 70 resv:4; member in struct:igmpv3_query 72 __u8 resv:4, member in struct:igmpv3_query
|
H A D | vfio.h | 191 __u32 resv; /* Reserved for alignment */ member in struct:vfio_region_info
|
/linux-4.4.14/net/sunrpc/auth_gss/ |
H A D | svcauth_gss.c | 637 svc_safe_putnetobj(struct kvec *resv, struct xdr_netobj *o) svc_safe_putnetobj() argument 641 if (resv->iov_len + 4 > PAGE_SIZE) svc_safe_putnetobj() 643 svc_putnl(resv, o->len); svc_safe_putnetobj() 644 p = resv->iov_base + resv->iov_len; svc_safe_putnetobj() 645 resv->iov_len += round_up_to_quad(o->len); svc_safe_putnetobj() 646 if (resv->iov_len > PAGE_SIZE) svc_safe_putnetobj() 1074 gss_write_resv(struct kvec *resv, size_t size_limit, gss_write_resv() argument 1078 if (resv->iov_len + 4 > size_limit) gss_write_resv() 1080 svc_putnl(resv, RPC_SUCCESS); gss_write_resv() 1081 if (svc_safe_putnetobj(resv, out_handle)) gss_write_resv() 1083 if (resv->iov_len + 3 * 4 > size_limit) gss_write_resv() 1085 svc_putnl(resv, major_status); gss_write_resv() 1086 svc_putnl(resv, minor_status); gss_write_resv() 1087 svc_putnl(resv, GSS_SEQ_WIN); gss_write_resv() 1088 if (svc_safe_putnetobj(resv, out_token)) gss_write_resv() 1104 struct kvec *resv = &rqstp->rq_res.head[0]; svcauth_gss_legacy_init() local 1129 if (gss_write_resv(resv, PAGE_SIZE, svcauth_gss_legacy_init() 1212 struct kvec *resv = &rqstp->rq_res.head[0]; svcauth_gss_proxy_init() local 1257 if (gss_write_resv(resv, PAGE_SIZE, svcauth_gss_proxy_init() 1400 struct kvec *resv = &rqstp->rq_res.head[0]; svcauth_gss_accept() local 1406 __be32 *reject_stat = resv->iov_base + resv->iov_len; svcauth_gss_accept() 1486 if (resv->iov_len + 4 > PAGE_SIZE) svcauth_gss_accept() 1488 svc_putnl(resv, RPC_SUCCESS); svcauth_gss_accept() 1492 svcdata->verf_start = resv->iov_base + resv->iov_len; svcauth_gss_accept() 1503 svc_putnl(resv, 0); svcauth_gss_accept() 1504 svc_putnl(resv, 0); svcauth_gss_accept() 1512 svc_putnl(resv, 0); svcauth_gss_accept() 1513 svc_putnl(resv, 0); svcauth_gss_accept() 1585 struct kvec *resv; svcauth_gss_wrap_resp_integ() local 1607 resv = &resbuf->tail[0]; svcauth_gss_wrap_resp_integ() 1608 mic.data = (u8 *)resv->iov_base + resv->iov_len + 4; svcauth_gss_wrap_resp_integ() 1611 svc_putnl(resv, mic.len); svcauth_gss_wrap_resp_integ() 1614 resv->iov_len += XDR_QUADLEN(mic.len) << 2; svcauth_gss_wrap_resp_integ() 1617 BUG_ON(resv->iov_len > PAGE_SIZE); svcauth_gss_wrap_resp_integ()
|
/linux-4.4.14/drivers/dma-buf/ |
H A D | dma-buf.c | 73 if (dmabuf->resv == (struct reservation_object *)&dmabuf[1]) dma_buf_release() 74 reservation_object_fini(dmabuf->resv); dma_buf_release() 138 struct reservation_object *resv; dma_buf_poll() local 145 if (!dmabuf || !dmabuf->resv) dma_buf_poll() 148 resv = dmabuf->resv; dma_buf_poll() 157 seq = read_seqcount_begin(&resv->seq); dma_buf_poll() 160 fobj = rcu_dereference(resv->fence); dma_buf_poll() 165 fence_excl = rcu_dereference(resv->fence_excl); dma_buf_poll() 166 if (read_seqcount_retry(&resv->seq, seq)) { dma_buf_poll() 287 struct reservation_object *resv = exp_info->resv; dma_buf_export() local 291 if (!exp_info->resv) dma_buf_export() 294 /* prevent &dma_buf[1] == dma_buf->resv */ dma_buf_export() 326 if (!resv) { dma_buf_export() 327 resv = (struct reservation_object *)&dmabuf[1]; dma_buf_export() 328 reservation_object_init(resv); dma_buf_export() 330 dmabuf->resv = resv; dma_buf_export()
|
/linux-4.4.14/arch/sparc/include/asm/ |
H A D | vio.h | 43 u64 resv[6]; member in struct:vio_rdx 69 u16 resv; member in struct:vio_dring_register 77 u64 resv[5]; member in struct:vio_dring_unregister 177 u32 resv; member in struct:vio_disk_vtoc::__anon2729 198 u16 resv; member in struct:vio_disk_devid 240 u32 resv; member in struct:vio_net_mcast_info
|
H A D | hypervisor.h | 473 unsigned long resv; member in struct:hv_tsb_descr
|
/linux-4.4.14/drivers/gpu/drm/msm/ |
H A D | msm_gem.h | 60 /* normally (resv == &_resv) except for imported bo's */ 61 struct reservation_object *resv; member in struct:msm_gem_object
|
H A D | msm_gem_submit.c | 130 ww_mutex_unlock(&msm_obj->resv->lock); submit_unlock_unpin_bo() 156 ret = ww_mutex_lock_interruptible(&msm_obj->resv->lock, submit_validate_objects() 202 ret = ww_mutex_lock_slow_interruptible(&msm_obj->resv->lock, submit_validate_objects()
|
H A D | msm_gem.c | 548 if (msm_obj->resv == &msm_obj->_resv) msm_gem_free_object() 549 reservation_object_fini(msm_obj->resv); msm_gem_free_object() 623 msm_obj->resv = &msm_obj->_resv; msm_gem_new_impl() 624 reservation_object_init(msm_obj->resv); msm_gem_new_impl()
|
/linux-4.4.14/drivers/gpu/drm/nouveau/ |
H A D | nouveau_prime.c | 64 struct reservation_object *robj = attach->dmabuf->resv; nouveau_gem_prime_import_sg_table() 114 return nvbo->bo.resv; nouveau_gem_prime_res_obj()
|
H A D | nouveau_fence.c | 394 struct reservation_object *resv = nvbo->bo.resv; nouveau_fence_sync() local 400 ret = reservation_object_reserve_shared(resv); nouveau_fence_sync() 406 fobj = reservation_object_get_list(resv); nouveau_fence_sync() 407 fence = reservation_object_get_excl(resv); nouveau_fence_sync() 436 reservation_object_held(resv)); nouveau_fence_sync()
|
H A D | nouveau_gem.c | 120 struct reservation_object *resv = nvbo->bo.resv; nouveau_gem_object_unmap() local 124 fobj = reservation_object_get_list(resv); nouveau_gem_object_unmap() 132 reservation_object_held(resv)); nouveau_gem_object_unmap() 134 fence = reservation_object_get_excl(nvbo->bo.resv); nouveau_gem_object_unmap() 873 ret = reservation_object_test_signaled_rcu(nvbo->bo.resv, write) ? 0 : -EBUSY; nouveau_gem_ioctl_cpu_prep() 877 lret = reservation_object_wait_timeout_rcu(nvbo->bo.resv, write, true, 30 * HZ); nouveau_gem_ioctl_cpu_prep()
|
H A D | nouveau_bo.c | 1276 struct fence *fence = reservation_object_get_excl(bo->resv); nouveau_bo_vm_cleanup() 1593 struct reservation_object *resv = nvbo->bo.resv; nouveau_bo_fence() local 1596 reservation_object_add_excl_fence(resv, &fence->base); nouveau_bo_fence() 1598 reservation_object_add_shared_fence(resv, &fence->base); nouveau_bo_fence()
|
/linux-4.4.14/arch/sparc/include/uapi/asm/ |
H A D | psr.h | 17 * | impl | vers | icc | resv | EC | EF | PIL | S | PS | ET | CWP |
|
/linux-4.4.14/arch/nios2/mm/ |
H A D | pgtable.c | 29 * resv |way |rd | we|pid |dbl|bad|perm|d
|
/linux-4.4.14/mm/ |
H A D | hugetlb.c | 255 static long region_add(struct resv_map *resv, long f, long t) region_add() argument 257 struct list_head *head = &resv->regions; region_add() 261 spin_lock(&resv->lock); region_add() 274 VM_BUG_ON(resv->region_cache_count <= 0); list_for_each_entry() 276 resv->region_cache_count--; list_for_each_entry() 277 nrg = list_first_entry(&resv->region_cache, struct file_region, list_for_each_entry() 323 resv->adds_in_progress--; 324 spin_unlock(&resv->lock); 351 static long region_chg(struct resv_map *resv, long f, long t) region_chg() argument 353 struct list_head *head = &resv->regions; region_chg() 358 spin_lock(&resv->lock); region_chg() 360 resv->adds_in_progress++; region_chg() 366 if (resv->adds_in_progress > resv->region_cache_count) { region_chg() 369 VM_BUG_ON(resv->adds_in_progress - resv->region_cache_count > 1); region_chg() 371 resv->adds_in_progress--; region_chg() 372 spin_unlock(&resv->lock); region_chg() 380 spin_lock(&resv->lock); region_chg() 381 list_add(&trg->link, &resv->region_cache); region_chg() 382 resv->region_cache_count++; region_chg() 396 resv->adds_in_progress--; list_for_each_entry() 397 spin_unlock(&resv->lock); list_for_each_entry() 436 spin_unlock(&resv->lock); 441 spin_unlock(&resv->lock); 456 static void region_abort(struct resv_map *resv, long f, long t) region_abort() argument 458 spin_lock(&resv->lock); region_abort() 459 VM_BUG_ON(!resv->region_cache_count); region_abort() 460 resv->adds_in_progress--; region_abort() 461 spin_unlock(&resv->lock); region_abort() 478 static long region_del(struct resv_map *resv, long f, long t) region_del() argument 480 struct list_head *head = &resv->regions; region_del() 486 spin_lock(&resv->lock); list_for_each_entry_safe() 507 resv->region_cache_count > resv->adds_in_progress) { list_for_each_entry_safe() 508 nrg = list_first_entry(&resv->region_cache, list_for_each_entry_safe() 512 resv->region_cache_count--; list_for_each_entry_safe() 516 spin_unlock(&resv->lock); list_for_each_entry_safe() 554 spin_unlock(&resv->lock); 585 static long region_count(struct resv_map *resv, long f, long t) region_count() argument 587 struct list_head *head = &resv->regions; region_count() 591 spin_lock(&resv->lock); region_count() 607 spin_unlock(&resv->lock); 1788 struct resv_map *resv; __vma_reservation_common() local 1792 resv = vma_resv_map(vma); __vma_reservation_common() 1793 if (!resv) __vma_reservation_common() 1799 ret = region_chg(resv, idx, idx + 1); __vma_reservation_common() 1802 ret = region_add(resv, idx, idx + 1); __vma_reservation_common() 1805 region_abort(resv, idx, idx + 1); __vma_reservation_common() 2963 struct resv_map *resv = vma_resv_map(vma); hugetlb_vm_op_open() local 2973 if (resv && is_vma_resv_set(vma, HPAGE_RESV_OWNER)) hugetlb_vm_op_open() 2974 kref_get(&resv->refs); hugetlb_vm_op_open() 2980 struct resv_map *resv = vma_resv_map(vma); hugetlb_vm_op_close() local 2985 if (!resv || !is_vma_resv_set(vma, HPAGE_RESV_OWNER)) hugetlb_vm_op_close() 2991 reserve = (end - start) - region_count(resv, start, end); hugetlb_vm_op_close() 2993 kref_put(&resv->refs, resv_map_release); hugetlb_vm_op_close()
|
H A D | cma.c | 121 * simple by forcing the entire CMA resv range cma_activate_area()
|
/linux-4.4.14/include/rdma/ |
H A D | ib_smi.h | 55 __be16 resv; member in struct:ib_smp 118 u8 resv; member in struct:ib_port_info
|
H A D | ib_pma.h | 64 u8 counter_width; /* resv: 7:3, counter width: 2:0 */
|
H A D | opa_smi.h | 61 __be16 resv; member in struct:opa_smp
|
H A D | ib_mad.h | 173 __be16 resv; member in struct:ib_mad_hdr
|
/linux-4.4.14/include/linux/ |
H A D | dma-buf.h | 122 * @resv: reservation object linked to this dma-buf 137 struct reservation_object *resv; member in struct:dma_buf 175 * @resv: reservation-object, NULL to allocate default one 187 struct reservation_object *resv; member in struct:dma_buf_export_info
|
/linux-4.4.14/drivers/gpu/drm/qxl/ |
H A D | qxl_debugfs.c | 65 fobj = rcu_dereference(bo->tbo.resv->fence); qxl_debugfs_buffers_info()
|
H A D | qxl_release.c | 246 ret = reservation_object_reserve_shared(bo->tbo.resv); qxl_release_validate_bo() 472 reservation_object_add_shared_fence(bo->resv, &release->base); qxl_release_fence_buffer_objects()
|
/linux-4.4.14/drivers/nvme/host/ |
H A D | lightnvm.c | 66 __le64 resv; member in struct:nvme_nvm_ph_rw 133 __le64 resv; member in struct:nvme_nvm_erase_blk 199 __u8 resv[228]; member in struct:nvme_nvm_id
|
H A D | scsi.c | 2364 u32 resv; member in struct:scsi_unmap_blk_desc 2370 u32 resv; member in struct:scsi_unmap_parm_list
|
/linux-4.4.14/fs/nfsd/ |
H A D | nfscache.c | 513 struct kvec *resv = &rqstp->rq_res.head[0], *cachv; nfsd_cache_update() local 525 len = resv->iov_len - ((char*)statp - (char*)resv->iov_base); nfsd_cache_update()
|
/linux-4.4.14/include/drm/ttm/ |
H A D | ttm_bo_driver.h | 790 success = ww_mutex_trylock(&bo->resv->lock); __ttm_bo_reserve() 795 ret = ww_mutex_lock_interruptible(&bo->resv->lock, ticket); __ttm_bo_reserve() 797 ret = ww_mutex_lock(&bo->resv->lock, ticket); __ttm_bo_reserve() 883 ret = ww_mutex_lock_slow_interruptible(&bo->resv->lock, ttm_bo_reserve_slowpath() 886 ww_mutex_lock_slow(&bo->resv->lock, ticket); ttm_bo_reserve_slowpath() 905 ww_mutex_unlock(&bo->resv->lock); __ttm_bo_unreserve()
|
H A D | ttm_bo_api.h | 215 * Members protected by the bo::resv::reserved lock. 257 struct reservation_object *resv; member in struct:ttm_buffer_object 463 * @resv: Pointer to a reservation_object, or NULL to let ttm allocate one. 491 struct reservation_object *resv,
|
/linux-4.4.14/drivers/input/mouse/ |
H A D | cyapa_gen6.c | 47 u8 resv; /* Reserved, must be 0 */ member in struct:pip_app_cmd_head 48 u8 cmd_code; /* bit7: resv, set to 0; bit6~0: command code.*/ 54 u8 resv; /* Reserved, must be 0 */ member in struct:pip_app_resp_head
|
/linux-4.4.14/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_resource.c | 578 return reservation_object_test_signaled_rcu(bo->resv, true) ? 0 : -EBUSY; vmw_user_dmabuf_synccpu_grab() 580 lret = reservation_object_wait_timeout_rcu(bo->resv, true, true, MAX_SCHEDULE_TIMEOUT); vmw_user_dmabuf_synccpu_grab() 1186 lockdep_assert_held(&res->backup->base.resv->lock.base); vmw_resource_unreserve() 1193 lockdep_assert_held(&new_backup->base.resv->lock.base); vmw_resource_unreserve() 1458 reservation_object_add_excl_fence(bo->resv, &fence->base); vmw_fence_single_bo() 1461 reservation_object_add_excl_fence(bo->resv, &fence->base); vmw_fence_single_bo()
|
H A D | vmwgfx_dmabuf.c | 276 lockdep_assert_held(&bo->resv->lock.base); vmw_bo_pin_reserved()
|
H A D | vmwgfx_cotable.c | 171 lockdep_assert_held(&bo->resv->lock.base); vmw_cotable_unscrub() 319 lockdep_assert_held(&bo->resv->lock.base); vmw_cotable_unbind()
|
/linux-4.4.14/drivers/gpu/drm/virtio/ |
H A D | virtgpu_ioctl.c | 397 reservation_object_add_excl_fence(qobj->tbo.resv, virtio_gpu_transfer_from_host_ioctl() 446 reservation_object_add_excl_fence(qobj->tbo.resv, virtio_gpu_transfer_to_host_ioctl()
|
H A D | virtgpu_display.c | 95 reservation_object_add_excl_fence(qobj->tbo.resv, virtio_gpu_crtc_cursor_set()
|
/linux-4.4.14/drivers/net/ethernet/sun/ |
H A D | sunqe.h | 257 #define MREGS_UTEST_RTRENAB 0x80 /* Enable resv test register */ 258 #define MREGS_UTEST_RTRDISAB 0x40 /* Disab resv test register */
|
H A D | niu.h | 2821 __le64 resv; member in struct:rx_pkt_hdr0::tx_pkt_hdr
|
H A D | niu.c | 6676 tp->resv = 0; niu_start_xmit()
|
/linux-4.4.14/drivers/infiniband/hw/qib/ |
H A D | qib_mad.h | 128 __be16 resv; member in struct:ib_cc_mad
|
/linux-4.4.14/drivers/net/ethernet/rocker/ |
H A D | rocker.h | 110 u16 resv[5]; member in struct:rocker_desc
|
/linux-4.4.14/arch/sparc/kernel/ |
H A D | pci_fire.c | 99 u64 resv[6]; member in struct:pci_msiq_entry
|
H A D | mdesc.c | 53 u16 resv; member in struct:mdesc_elem
|
/linux-4.4.14/arch/sparc/mm/ |
H A D | tsb.c | 259 hp->resv = 0; setup_tsb_params()
|
H A D | init_64.c | 1924 ktsb_descr[0].resv = 0; sun4v_ktsb_init() 1941 ktsb_descr[1].resv = 0; sun4v_ktsb_init()
|
/linux-4.4.14/drivers/staging/rdma/ehca/ |
H A D | ehca_reqs.c | 128 "attr_id=%x resv=%x attr_mod=%x", trace_ud_wr() 134 mad_hdr->resv, trace_ud_wr()
|
/linux-4.4.14/drivers/gpu/drm/ |
H A D | drm_prime.c | 352 exp_info.resv = dev->driver->gem_prime_res_obj(obj); drm_gem_prime_export()
|
/linux-4.4.14/fs/xfs/ |
H A D | xfs_trans_dquot.c | 694 * Note that q_res_bcount = q_core.d_bcount + resv xfs_trans_dqresv()
|
H A D | xfs_ioctl.c | 675 * length of <= 0 for resv/unresv/zero is invalid. length for xfs_ioc_space()
|
/linux-4.4.14/fs/ceph/ |
H A D | caps.c | 198 dout("reserve caps ctx=%p %d = %d used + %d resv + %d avail\n", ceph_reserve_caps() 213 dout("unreserve caps %d = %d used + %d resv + %d avail\n", ceph_unreserve_caps() 242 dout("get_cap ctx=%p (%d) %d = %d used + %d resv + %d avail\n", ceph_get_cap() 265 dout("put_cap %p %d = %d used + %d resv + %d avail\n", ceph_put_cap()
|
/linux-4.4.14/include/linux/mlx4/ |
H A D | device.h | 969 __be16 resv; member in struct:mlx4_mad_ifc
|
/linux-4.4.14/drivers/scsi/bfa/ |
H A D | bfa_defs_svc.h | 1054 u8 resv[2]; member in struct:bfa_fcport_fcf_s
|
/linux-4.4.14/drivers/scsi/ |
H A D | ipr.h | 1250 u8 resv; member in struct:ipr_sdt_entry
|
/linux-4.4.14/drivers/iommu/ |
H A D | amd_iommu_init.c | 126 u64 resv; member in struct:ivmd_header
|
/linux-4.4.14/drivers/scsi/lpfc/ |
H A D | lpfc_hw.h | 973 uint8_t resv; member in struct:_D_ID::__anon9617::__anon9618 981 uint8_t resv;
|
H A D | lpfc_els.c | 5322 switch (rscn_did.un.b.resv & RSCN_ADDRESS_FORMAT_MASK) { lpfc_rscn_payload_check()
|