fences             87 drivers/dma-buf/dma-fence-array.c 		if (dma_fence_add_callback(array->fences[i], &cb[i].cb,
fences             89 drivers/dma-buf/dma-fence-array.c 			int error = array->fences[i]->error;
fences            116 drivers/dma-buf/dma-fence-array.c 		dma_fence_put(array->fences[i]);
fences            118 drivers/dma-buf/dma-fence-array.c 	kfree(array->fences);
fences            151 drivers/dma-buf/dma-fence-array.c 					       struct dma_fence **fences,
fences            171 drivers/dma-buf/dma-fence-array.c 	array->fences = fences;
fences            197 drivers/dma-buf/dma-fence-array.c 		if (array->fences[i]->context != context)
fences            525 drivers/dma-buf/dma-fence.c dma_fence_test_signaled_any(struct dma_fence **fences, uint32_t count,
fences            531 drivers/dma-buf/dma-fence.c 		struct dma_fence *fence = fences[i];
fences            562 drivers/dma-buf/dma-fence.c dma_fence_wait_any_timeout(struct dma_fence **fences, uint32_t count,
fences            569 drivers/dma-buf/dma-fence.c 	if (WARN_ON(!fences || !count || timeout < 0))
fences            574 drivers/dma-buf/dma-fence.c 			if (dma_fence_is_signaled(fences[i])) {
fences            590 drivers/dma-buf/dma-fence.c 		struct dma_fence *fence = fences[i];
fences            608 drivers/dma-buf/dma-fence.c 		if (dma_fence_test_signaled_any(fences, count, idx))
fences            621 drivers/dma-buf/dma-fence.c 		dma_fence_remove_callback(fences[i], &cb[i].base);
fences            432 drivers/dma-buf/st-dma-fence.c 	struct dma_fence __rcu **fences;
fences            461 drivers/dma-buf/st-dma-fence.c 		rcu_assign_pointer(t->fences[t->id], f1);
fences            466 drivers/dma-buf/st-dma-fence.c 			f2 = dma_fence_get_rcu_safe(&t->fences[!t->id]);
fences            495 drivers/dma-buf/st-dma-fence.c 		rcu_assign_pointer(t->fences[t->id], NULL);
fences            519 drivers/dma-buf/st-dma-fence.c 			t[i].fences = f;
fences            135 drivers/dma-buf/sync_debug.c 			sync_print_fence(s, array->fences[i], true);
fences            149 drivers/dma-buf/sync_file.c 			       struct dma_fence **fences, int num_fences)
fences            160 drivers/dma-buf/sync_file.c 		sync_file->fence = fences[0];
fences            161 drivers/dma-buf/sync_file.c 		kfree(fences);
fences            163 drivers/dma-buf/sync_file.c 		array = dma_fence_array_create(num_fences, fences,
fences            182 drivers/dma-buf/sync_file.c 		return array->fences;
fences            189 drivers/dma-buf/sync_file.c static void add_fence(struct dma_fence **fences,
fences            192 drivers/dma-buf/sync_file.c 	fences[*i] = fence;
fences            214 drivers/dma-buf/sync_file.c 	struct dma_fence **fences, **nfences, **a_fences, **b_fences;
fences            228 drivers/dma-buf/sync_file.c 	fences = kcalloc(num_fences, sizeof(*fences), GFP_KERNEL);
fences            229 drivers/dma-buf/sync_file.c 	if (!fences)
fences            244 drivers/dma-buf/sync_file.c 			add_fence(fences, &i, pt_a);
fences            248 drivers/dma-buf/sync_file.c 			add_fence(fences, &i, pt_b);
fences            254 drivers/dma-buf/sync_file.c 				add_fence(fences, &i, pt_a);
fences            256 drivers/dma-buf/sync_file.c 				add_fence(fences, &i, pt_b);
fences            264 drivers/dma-buf/sync_file.c 		add_fence(fences, &i, a_fences[i_a]);
fences            267 drivers/dma-buf/sync_file.c 		add_fence(fences, &i, b_fences[i_b]);
fences            270 drivers/dma-buf/sync_file.c 		fences[i++] = dma_fence_get(a_fences[0]);
fences            273 drivers/dma-buf/sync_file.c 		nfences = krealloc(fences, i * sizeof(*fences),
fences            278 drivers/dma-buf/sync_file.c 		fences = nfences;
fences            281 drivers/dma-buf/sync_file.c 	if (sync_file_set_fence(sync_file, fences, i) < 0) {
fences            282 drivers/dma-buf/sync_file.c 		kfree(fences);
fences            403 drivers/dma-buf/sync_file.c 	struct dma_fence **fences;
fences            413 drivers/dma-buf/sync_file.c 	fences = get_fences(sync_file, &num_fences);
fences            437 drivers/dma-buf/sync_file.c 		int status = sync_fill_fence_info(fences[i], &fence_info[i]);
fences           1560 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 				     struct drm_amdgpu_fence *fences)
fences           1570 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		fence = amdgpu_cs_get_fence(adev, filp, &fences[i]);
fences           1605 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 				    struct drm_amdgpu_fence *fences)
fences           1623 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		fence = amdgpu_cs_get_fence(adev, filp, &fences[i]);
fences           1673 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 	struct drm_amdgpu_fence *fences;
fences           1677 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 	fences = kmalloc_array(fence_count, sizeof(struct drm_amdgpu_fence),
fences           1679 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 	if (fences == NULL)
fences           1682 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 	fences_user = u64_to_user_ptr(wait->in.fences);
fences           1683 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 	if (copy_from_user(fences, fences_user,
fences           1690 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		r = amdgpu_cs_wait_all_fences(adev, filp, wait, fences);
fences           1692 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		r = amdgpu_cs_wait_any_fence(adev, filp, wait, fences);
fences           1695 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 	kfree(fences);
fences             90 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c 	ctx->fences = kcalloc(amdgpu_sched_jobs * num_entities,
fences             92 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c 	if (!ctx->fences)
fences            107 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c 		entity->fences = &ctx->fences[amdgpu_sched_jobs * i];
fences            202 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c 	kfree(ctx->fences);
fences            203 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c 	ctx->fences = NULL;
fences            219 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c 			dma_fence_put(ctx->entities[0][i].fences[j]);
fences            220 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c 	kfree(ctx->fences);
fences            476 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c 	other = centity->fences[idx];
fences            483 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c 	centity->fences[idx] = fence;
fences            515 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c 	fence = dma_fence_get(centity->fences[seq & (amdgpu_sched_jobs - 1)]);
fences            550 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c 	other = dma_fence_get(centity->fences[idx]);
fences             34 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.h 	struct dma_fence	**fences;
fences             45 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.h 	struct dma_fence		**fences;
fences            933 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 					  struct dma_fence **fences)
fences            949 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 		ptr = &drv->fences[last_seq];
fences            957 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 		fences[last_seq] = fence;
fences            962 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c static void amdgpu_ib_preempt_signal_fences(struct dma_fence **fences,
fences            969 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 		fence = fences[i];
fences           1007 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 	ptr = &drv->fences[preempt_seq];
fences           1024 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 	struct dma_fence **fences = NULL;
fences           1040 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 	fences = kcalloc(length, sizeof(void *), GFP_KERNEL);
fences           1041 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 	if (!fences)
fences           1068 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 		amdgpu_ib_preempt_fences_swap(ring, fences);
fences           1079 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 		amdgpu_ib_preempt_signal_fences(fences, length);
fences           1090 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 	if (fences)
fences           1091 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c 		kfree(fences);
fences            142 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c 	struct dma_fence **fences;
fences            149 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c 	r = dma_resv_get_fences_rcu(obj, NULL, &count, &fences);
fences            156 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c 		dma_resv_add_excl_fence(obj, fences[0]);
fences            157 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c 		dma_fence_put(fences[0]);
fences            158 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c 		kfree(fences);
fences            162 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c 		array = dma_fence_array_create(count, fences,
fences            176 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c 		dma_fence_put(fences[count]);
fences            177 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c 	kfree(fences);
fences            158 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 	ptr = &ring->fence_drv.fences[seq & ring->fence_drv.num_fences_mask];
fences            261 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 		ptr = &drv->fences[last_seq];
fences            316 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 	ptr = &ring->fence_drv.fences[seq & ring->fence_drv.num_fences_mask];
fences            453 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 	ring->fence_drv.fences = kcalloc(num_hw_submission * 2, sizeof(void *),
fences            455 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 	if (!ring->fence_drv.fences)
fences            547 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 			dma_fence_put(ring->fence_drv.fences[j]);
fences            548 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 		kfree(ring->fence_drv.fences);
fences            549 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c 		ring->fence_drv.fences = NULL;
fences            110 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 	struct dma_fence *fence, **fences;
fences            115 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 	r = dma_resv_get_fences_rcu(resv, NULL, &count, &fences);
fences            125 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 		fence = fences[0];
fences            126 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 		kfree(fences);
fences            131 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 		array = dma_fence_array_create(count, fences, context,
fences            134 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 			kfree(fences);
fences            204 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 	struct dma_fence **fences;
fences            211 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 	fences = kmalloc_array(sizeof(void *), id_mgr->num_ids, GFP_KERNEL);
fences            212 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 	if (!fences)
fences            218 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 		fences[i] = amdgpu_sync_peek_fence(&(*idle)->active, ring);
fences            219 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 		if (!fences[i])
fences            233 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 			dma_fence_get(fences[j]);
fences            235 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 		array = dma_fence_array_create(i, fences, fence_context,
fences            239 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 				dma_fence_put(fences[j]);
fences            240 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 			kfree(fences);
fences            249 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 	kfree(fences);
fences             83 drivers/gpu/drm/amd/amdgpu/amdgpu_ring.h 	struct dma_fence		**fences;
fences            207 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c 				   struct dma_fence **fences,
fences            229 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c 		fences[i] = NULL;
fences            238 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c 			fences[i] = sa_bo->fence;
fences            279 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c 	struct dma_fence *fences[AMDGPU_SA_NUM_FENCE_LISTS];
fences            314 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c 		} while (amdgpu_sa_bo_next_hole(sa_manager, fences, tries));
fences            317 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c 			if (fences[i])
fences            318 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c 				fences[count++] = dma_fence_get(fences[i]);
fences            322 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c 			t = dma_fence_wait_any_timeout(fences, count, false,
fences            326 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c 				dma_fence_put(fences[i]);
fences             52 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c 	hash_init(sync->fences);
fences            136 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c 	hash_for_each_possible(sync->fences, e, node, f->context) {
fences            177 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c 	hash_add(sync->fences, &e->node, f->context);
fences            265 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c 	hash_for_each_safe(sync->fences, i, tmp, e, node) {
fences            307 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c 	hash_for_each_safe(sync->fences, i, tmp, e, node) {
fences            340 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c 	hash_for_each_safe(source->fences, i, tmp, e, node) {
fences            365 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c 	hash_for_each_safe(sync->fences, i, tmp, e, node) {
fences            391 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c 	hash_for_each_safe(sync->fences, i, tmp, e, node) {
fences             38 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.h 	DECLARE_HASHTABLE(fences, 4);
fences            149 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h 			     __field(u32, fences)
fences            156 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h 			   __entry->fences = amdgpu_fence_count_emitted(
fences            161 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h 		      __entry->fences)
fences           1182 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 	unsigned fences = 0, i, j;
fences           1187 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 		fences += amdgpu_fence_count_emitted(&adev->uvd.inst[i].ring);
fences           1189 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 			fences += amdgpu_fence_count_emitted(&adev->uvd.inst[i].ring_enc[j]);
fences           1193 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 	if (fences == 0) {
fences            291 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c 	unsigned int fences = 0, fence[AMDGPU_MAX_VCN_INSTANCES] = {0};
fences            319 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c 		fences += fence[j];
fences            322 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c 	if (fences == 0) {
fences            350 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c 		unsigned int fences = 0;
fences            354 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c 			fences += amdgpu_fence_count_emitted(&adev->vcn.inst[ring->me].ring_enc[i]);
fences            356 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c 		if (fences)
fences           1411 drivers/gpu/drm/drm_gem.c 	struct dma_fence **fences;
fences           1422 drivers/gpu/drm/drm_gem.c 						&fence_count, &fences);
fences           1427 drivers/gpu/drm/drm_gem.c 		ret = drm_gem_fence_array_add(fence_array, fences[i]);
fences           1433 drivers/gpu/drm/drm_gem.c 		dma_fence_put(fences[i]);
fences           1434 drivers/gpu/drm/drm_gem.c 	kfree(fences);
fences           2348 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c __free_fence_array(struct drm_syncobj **fences, unsigned int n)
fences           2351 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		drm_syncobj_put(ptr_mask_bits(fences[n], 2));
fences           2352 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	kvfree(fences);
fences           2361 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	struct drm_syncobj **fences;
fences           2372 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 			    SIZE_MAX / sizeof(*fences)))
fences           2379 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	fences = kvmalloc_array(nfences, sizeof(*fences),
fences           2381 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	if (!fences)
fences           2408 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		fences[n] = ptr_pack_bits(syncobj, fence.flags, 2);
fences           2411 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	return fences;
fences           2414 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	__free_fence_array(fences, n);
fences           2420 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		struct drm_syncobj **fences)
fences           2422 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	if (fences)
fences           2423 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		__free_fence_array(fences, args->num_cliprects);
fences           2428 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		  struct drm_syncobj **fences)
fences           2439 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		syncobj = ptr_unpack_bits(fences[n], &flags, 2);
fences           2458 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		   struct drm_syncobj **fences)
fences           2468 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		syncobj = ptr_unpack_bits(fences[n], &flags, 2);
fences           2481 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		       struct drm_syncobj **fences)
fences           2662 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	if (fences) {
fences           2663 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		err = await_fence_array(&eb, fences);
fences           2693 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	if (fences)
fences           2694 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		signal_fence_array(&eb, fences);
fences           2853 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	struct drm_syncobj **fences = NULL;
fences           2882 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		fences = get_fence_array(args, file);
fences           2883 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		if (IS_ERR(fences)) {
fences           2885 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 			return PTR_ERR(fences);
fences           2889 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	err = i915_gem_do_execbuffer(dev, file, args, exec2_list, fences);
fences           2929 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	put_fence_array(args, fences);
fences            125 drivers/gpu/drm/i915/gem/i915_gem_wait.c 			__fence_set_priority(array->fences[i], attr);
fences            952 drivers/gpu/drm/i915/i915_request.c 		child = array->fences;
fences           1007 drivers/gpu/drm/i915/i915_request.c 		child = array->fences;
fences            453 drivers/gpu/drm/i915/selftests/i915_sw_fence.c 	struct i915_sw_fence **fences;
fences            457 drivers/gpu/drm/i915/selftests/i915_sw_fence.c 	fences = kmalloc_array(nfences, sizeof(*fences), GFP_KERNEL);
fences            458 drivers/gpu/drm/i915/selftests/i915_sw_fence.c 	if (!fences)
fences            462 drivers/gpu/drm/i915/selftests/i915_sw_fence.c 		fences[i] = alloc_fence();
fences            463 drivers/gpu/drm/i915/selftests/i915_sw_fence.c 		if (!fences[i]) {
fences            470 drivers/gpu/drm/i915/selftests/i915_sw_fence.c 			ret = i915_sw_fence_await_sw_fence_gfp(fences[i],
fences            471 drivers/gpu/drm/i915/selftests/i915_sw_fence.c 							       fences[i - 1],
fences            478 drivers/gpu/drm/i915/selftests/i915_sw_fence.c 			i915_sw_fence_commit(fences[i]);
fences            484 drivers/gpu/drm/i915/selftests/i915_sw_fence.c 		if (i915_sw_fence_done(fences[i])) {
fences            490 drivers/gpu/drm/i915/selftests/i915_sw_fence.c 	i915_sw_fence_commit(fences[0]);
fences            492 drivers/gpu/drm/i915/selftests/i915_sw_fence.c 		if (!i915_sw_fence_done(fences[i])) {
fences            500 drivers/gpu/drm/i915/selftests/i915_sw_fence.c 		free_fence(fences[i]);
fences            501 drivers/gpu/drm/i915/selftests/i915_sw_fence.c 	kfree(fences);
fences            395 drivers/gpu/drm/radeon/radeon.h 			  struct radeon_fence **fences,
fences            612 drivers/gpu/drm/radeon/radeon_fence.c 			  struct radeon_fence **fences,
fences            622 drivers/gpu/drm/radeon/radeon_fence.c 		if (!fences[i]) {
fences            626 drivers/gpu/drm/radeon/radeon_fence.c 		seq[i] = fences[i]->seq;
fences            248 drivers/gpu/drm/radeon/radeon_sa.c 				   struct radeon_fence **fences,
fences            278 drivers/gpu/drm/radeon/radeon_sa.c 			fences[i] = sa_bo->fence;
fences            317 drivers/gpu/drm/radeon/radeon_sa.c 	struct radeon_fence *fences[RADEON_NUM_RINGS];
fences            336 drivers/gpu/drm/radeon/radeon_sa.c 			fences[i] = NULL;
fences            350 drivers/gpu/drm/radeon/radeon_sa.c 		} while (radeon_sa_bo_next_hole(sa_manager, fences, tries));
fences            353 drivers/gpu/drm/radeon/radeon_sa.c 			radeon_fence_ref(fences[i]);
fences            356 drivers/gpu/drm/radeon/radeon_sa.c 		r = radeon_fence_wait_any(rdev, fences, false);
fences            358 drivers/gpu/drm/radeon/radeon_sa.c 			radeon_fence_unref(&fences[i]);
fences             36 drivers/gpu/drm/radeon/radeon_trace.h 			     __field(u32, fences)
fences             42 drivers/gpu/drm/radeon/radeon_trace.h 			   __entry->fences = radeon_fence_count_emitted(
fences             47 drivers/gpu/drm/radeon/radeon_trace.h 		      __entry->fences)
fences             97 drivers/gpu/drm/virtio/virtgpu_drv.h 	struct list_head fences;
fences             98 drivers/gpu/drm/virtio/virtgpu_fence.c 	list_add_tail(&fence->node, &drv->fences);
fences            116 drivers/gpu/drm/virtio/virtgpu_fence.c 	list_for_each_entry_safe(fence, tmp, &drv->fences, node) {
fences            145 drivers/gpu/drm/virtio/virtgpu_kms.c 	INIT_LIST_HEAD(&vgdev->fence_drv.fences);
fences            703 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 		struct dma_fence *child = fence_array->fences[i];
fences             43 include/linux/dma-fence-array.h 	struct dma_fence **fences;
fences             78 include/linux/dma-fence-array.h 					       struct dma_fence **fences,
fences            540 include/linux/dma-fence.h signed long dma_fence_wait_any_timeout(struct dma_fence **fences,
fences            445 include/uapi/drm/amdgpu_drm.h 	__u64 fences;