fences 87 drivers/dma-buf/dma-fence-array.c if (dma_fence_add_callback(array->fences[i], &cb[i].cb, fences 89 drivers/dma-buf/dma-fence-array.c int error = array->fences[i]->error; fences 116 drivers/dma-buf/dma-fence-array.c dma_fence_put(array->fences[i]); fences 118 drivers/dma-buf/dma-fence-array.c kfree(array->fences); fences 151 drivers/dma-buf/dma-fence-array.c struct dma_fence **fences, fences 171 drivers/dma-buf/dma-fence-array.c array->fences = fences; fences 197 drivers/dma-buf/dma-fence-array.c if (array->fences[i]->context != context) fences 525 drivers/dma-buf/dma-fence.c dma_fence_test_signaled_any(struct dma_fence **fences, uint32_t count, fences 531 drivers/dma-buf/dma-fence.c struct dma_fence *fence = fences[i]; fences 562 drivers/dma-buf/dma-fence.c dma_fence_wait_any_timeout(struct dma_fence **fences, uint32_t count, fences 569 drivers/dma-buf/dma-fence.c if (WARN_ON(!fences || !count || timeout < 0)) fences 574 drivers/dma-buf/dma-fence.c if (dma_fence_is_signaled(fences[i])) { fences 590 drivers/dma-buf/dma-fence.c struct dma_fence *fence = fences[i]; fences 608 drivers/dma-buf/dma-fence.c if (dma_fence_test_signaled_any(fences, count, idx)) fences 621 drivers/dma-buf/dma-fence.c dma_fence_remove_callback(fences[i], &cb[i].base); fences 432 drivers/dma-buf/st-dma-fence.c struct dma_fence __rcu **fences; fences 461 drivers/dma-buf/st-dma-fence.c rcu_assign_pointer(t->fences[t->id], f1); fences 466 drivers/dma-buf/st-dma-fence.c f2 = dma_fence_get_rcu_safe(&t->fences[!t->id]); fences 495 drivers/dma-buf/st-dma-fence.c rcu_assign_pointer(t->fences[t->id], NULL); fences 519 drivers/dma-buf/st-dma-fence.c t[i].fences = f; fences 135 drivers/dma-buf/sync_debug.c sync_print_fence(s, array->fences[i], true); fences 149 drivers/dma-buf/sync_file.c struct dma_fence **fences, int num_fences) fences 160 drivers/dma-buf/sync_file.c sync_file->fence = fences[0]; fences 161 drivers/dma-buf/sync_file.c kfree(fences); fences 163 drivers/dma-buf/sync_file.c array = dma_fence_array_create(num_fences, fences, fences 182 drivers/dma-buf/sync_file.c return array->fences; fences 189 drivers/dma-buf/sync_file.c static void add_fence(struct dma_fence **fences, fences 192 drivers/dma-buf/sync_file.c fences[*i] = fence; fences 214 drivers/dma-buf/sync_file.c struct dma_fence **fences, **nfences, **a_fences, **b_fences; fences 228 drivers/dma-buf/sync_file.c fences = kcalloc(num_fences, sizeof(*fences), GFP_KERNEL); fences 229 drivers/dma-buf/sync_file.c if (!fences) fences 244 drivers/dma-buf/sync_file.c add_fence(fences, &i, pt_a); fences 248 drivers/dma-buf/sync_file.c add_fence(fences, &i, pt_b); fences 254 drivers/dma-buf/sync_file.c add_fence(fences, &i, pt_a); fences 256 drivers/dma-buf/sync_file.c add_fence(fences, &i, pt_b); fences 264 drivers/dma-buf/sync_file.c add_fence(fences, &i, a_fences[i_a]); fences 267 drivers/dma-buf/sync_file.c add_fence(fences, &i, b_fences[i_b]); fences 270 drivers/dma-buf/sync_file.c fences[i++] = dma_fence_get(a_fences[0]); fences 273 drivers/dma-buf/sync_file.c nfences = krealloc(fences, i * sizeof(*fences), fences 278 drivers/dma-buf/sync_file.c fences = nfences; fences 281 drivers/dma-buf/sync_file.c if (sync_file_set_fence(sync_file, fences, i) < 0) { fences 282 drivers/dma-buf/sync_file.c kfree(fences); fences 403 drivers/dma-buf/sync_file.c struct dma_fence **fences; fences 413 drivers/dma-buf/sync_file.c fences = get_fences(sync_file, &num_fences); fences 437 drivers/dma-buf/sync_file.c int status = sync_fill_fence_info(fences[i], &fence_info[i]); fences 1560 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c struct drm_amdgpu_fence *fences) fences 1570 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c fence = amdgpu_cs_get_fence(adev, filp, &fences[i]); fences 1605 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c struct drm_amdgpu_fence *fences) fences 1623 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c fence = amdgpu_cs_get_fence(adev, filp, &fences[i]); fences 1673 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c struct drm_amdgpu_fence *fences; fences 1677 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c fences = kmalloc_array(fence_count, sizeof(struct drm_amdgpu_fence), fences 1679 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c if (fences == NULL) fences 1682 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c fences_user = u64_to_user_ptr(wait->in.fences); fences 1683 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c if (copy_from_user(fences, fences_user, fences 1690 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c r = amdgpu_cs_wait_all_fences(adev, filp, wait, fences); fences 1692 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c r = amdgpu_cs_wait_any_fence(adev, filp, wait, fences); fences 1695 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c kfree(fences); fences 90 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c ctx->fences = kcalloc(amdgpu_sched_jobs * num_entities, fences 92 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c if (!ctx->fences) fences 107 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c entity->fences = &ctx->fences[amdgpu_sched_jobs * i]; fences 202 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c kfree(ctx->fences); fences 203 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c ctx->fences = NULL; fences 219 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c dma_fence_put(ctx->entities[0][i].fences[j]); fences 220 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c kfree(ctx->fences); fences 476 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c other = centity->fences[idx]; fences 483 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c centity->fences[idx] = fence; fences 515 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c fence = dma_fence_get(centity->fences[seq & (amdgpu_sched_jobs - 1)]); fences 550 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c other = dma_fence_get(centity->fences[idx]); fences 34 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.h struct dma_fence **fences; fences 45 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.h struct dma_fence **fences; fences 933 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c struct dma_fence **fences) fences 949 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c ptr = &drv->fences[last_seq]; fences 957 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c fences[last_seq] = fence; fences 962 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c static void amdgpu_ib_preempt_signal_fences(struct dma_fence **fences, fences 969 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c fence = fences[i]; fences 1007 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c ptr = &drv->fences[preempt_seq]; fences 1024 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c struct dma_fence **fences = NULL; fences 1040 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c fences = kcalloc(length, sizeof(void *), GFP_KERNEL); fences 1041 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c if (!fences) fences 1068 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c amdgpu_ib_preempt_fences_swap(ring, fences); fences 1079 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c amdgpu_ib_preempt_signal_fences(fences, length); fences 1090 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c if (fences) fences 1091 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c kfree(fences); fences 142 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c struct dma_fence **fences; fences 149 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c r = dma_resv_get_fences_rcu(obj, NULL, &count, &fences); fences 156 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c dma_resv_add_excl_fence(obj, fences[0]); fences 157 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c dma_fence_put(fences[0]); fences 158 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c kfree(fences); fences 162 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c array = dma_fence_array_create(count, fences, fences 176 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c dma_fence_put(fences[count]); fences 177 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c kfree(fences); fences 158 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c ptr = &ring->fence_drv.fences[seq & ring->fence_drv.num_fences_mask]; fences 261 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c ptr = &drv->fences[last_seq]; fences 316 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c ptr = &ring->fence_drv.fences[seq & ring->fence_drv.num_fences_mask]; fences 453 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c ring->fence_drv.fences = kcalloc(num_hw_submission * 2, sizeof(void *), fences 455 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c if (!ring->fence_drv.fences) fences 547 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c dma_fence_put(ring->fence_drv.fences[j]); fences 548 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c kfree(ring->fence_drv.fences); fences 549 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c ring->fence_drv.fences = NULL; fences 110 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c struct dma_fence *fence, **fences; fences 115 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c r = dma_resv_get_fences_rcu(resv, NULL, &count, &fences); fences 125 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c fence = fences[0]; fences 126 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c kfree(fences); fences 131 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c array = dma_fence_array_create(count, fences, context, fences 134 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c kfree(fences); fences 204 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c struct dma_fence **fences; fences 211 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c fences = kmalloc_array(sizeof(void *), id_mgr->num_ids, GFP_KERNEL); fences 212 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c if (!fences) fences 218 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c fences[i] = amdgpu_sync_peek_fence(&(*idle)->active, ring); fences 219 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c if (!fences[i]) fences 233 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c dma_fence_get(fences[j]); fences 235 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c array = dma_fence_array_create(i, fences, fence_context, fences 239 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c dma_fence_put(fences[j]); fences 240 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c kfree(fences); fences 249 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c kfree(fences); fences 83 drivers/gpu/drm/amd/amdgpu/amdgpu_ring.h struct dma_fence **fences; fences 207 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c struct dma_fence **fences, fences 229 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c fences[i] = NULL; fences 238 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c fences[i] = sa_bo->fence; fences 279 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c struct dma_fence *fences[AMDGPU_SA_NUM_FENCE_LISTS]; fences 314 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c } while (amdgpu_sa_bo_next_hole(sa_manager, fences, tries)); fences 317 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c if (fences[i]) fences 318 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c fences[count++] = dma_fence_get(fences[i]); fences 322 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c t = dma_fence_wait_any_timeout(fences, count, false, fences 326 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c dma_fence_put(fences[i]); fences 52 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c hash_init(sync->fences); fences 136 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c hash_for_each_possible(sync->fences, e, node, f->context) { fences 177 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c hash_add(sync->fences, &e->node, f->context); fences 265 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c hash_for_each_safe(sync->fences, i, tmp, e, node) { fences 307 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c hash_for_each_safe(sync->fences, i, tmp, e, node) { fences 340 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c hash_for_each_safe(source->fences, i, tmp, e, node) { fences 365 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c hash_for_each_safe(sync->fences, i, tmp, e, node) { fences 391 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c hash_for_each_safe(sync->fences, i, tmp, e, node) { fences 38 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.h DECLARE_HASHTABLE(fences, 4); fences 149 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h __field(u32, fences) fences 156 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h __entry->fences = amdgpu_fence_count_emitted( fences 161 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h __entry->fences) fences 1182 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c unsigned fences = 0, i, j; fences 1187 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c fences += amdgpu_fence_count_emitted(&adev->uvd.inst[i].ring); fences 1189 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c fences += amdgpu_fence_count_emitted(&adev->uvd.inst[i].ring_enc[j]); fences 1193 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c if (fences == 0) { fences 291 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c unsigned int fences = 0, fence[AMDGPU_MAX_VCN_INSTANCES] = {0}; fences 319 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c fences += fence[j]; fences 322 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c if (fences == 0) { fences 350 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c unsigned int fences = 0; fences 354 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c fences += amdgpu_fence_count_emitted(&adev->vcn.inst[ring->me].ring_enc[i]); fences 356 drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c if (fences) fences 1411 drivers/gpu/drm/drm_gem.c struct dma_fence **fences; fences 1422 drivers/gpu/drm/drm_gem.c &fence_count, &fences); fences 1427 drivers/gpu/drm/drm_gem.c ret = drm_gem_fence_array_add(fence_array, fences[i]); fences 1433 drivers/gpu/drm/drm_gem.c dma_fence_put(fences[i]); fences 1434 drivers/gpu/drm/drm_gem.c kfree(fences); fences 2348 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c __free_fence_array(struct drm_syncobj **fences, unsigned int n) fences 2351 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c drm_syncobj_put(ptr_mask_bits(fences[n], 2)); fences 2352 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c kvfree(fences); fences 2361 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c struct drm_syncobj **fences; fences 2372 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c SIZE_MAX / sizeof(*fences))) fences 2379 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c fences = kvmalloc_array(nfences, sizeof(*fences), fences 2381 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (!fences) fences 2408 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c fences[n] = ptr_pack_bits(syncobj, fence.flags, 2); fences 2411 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c return fences; fences 2414 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c __free_fence_array(fences, n); fences 2420 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c struct drm_syncobj **fences) fences 2422 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (fences) fences 2423 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c __free_fence_array(fences, args->num_cliprects); fences 2428 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c struct drm_syncobj **fences) fences 2439 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c syncobj = ptr_unpack_bits(fences[n], &flags, 2); fences 2458 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c struct drm_syncobj **fences) fences 2468 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c syncobj = ptr_unpack_bits(fences[n], &flags, 2); fences 2481 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c struct drm_syncobj **fences) fences 2662 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (fences) { fences 2663 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c err = await_fence_array(&eb, fences); fences 2693 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (fences) fences 2694 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c signal_fence_array(&eb, fences); fences 2853 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c struct drm_syncobj **fences = NULL; fences 2882 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c fences = get_fence_array(args, file); fences 2883 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (IS_ERR(fences)) { fences 2885 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c return PTR_ERR(fences); fences 2889 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c err = i915_gem_do_execbuffer(dev, file, args, exec2_list, fences); fences 2929 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c put_fence_array(args, fences); fences 125 drivers/gpu/drm/i915/gem/i915_gem_wait.c __fence_set_priority(array->fences[i], attr); fences 952 drivers/gpu/drm/i915/i915_request.c child = array->fences; fences 1007 drivers/gpu/drm/i915/i915_request.c child = array->fences; fences 453 drivers/gpu/drm/i915/selftests/i915_sw_fence.c struct i915_sw_fence **fences; fences 457 drivers/gpu/drm/i915/selftests/i915_sw_fence.c fences = kmalloc_array(nfences, sizeof(*fences), GFP_KERNEL); fences 458 drivers/gpu/drm/i915/selftests/i915_sw_fence.c if (!fences) fences 462 drivers/gpu/drm/i915/selftests/i915_sw_fence.c fences[i] = alloc_fence(); fences 463 drivers/gpu/drm/i915/selftests/i915_sw_fence.c if (!fences[i]) { fences 470 drivers/gpu/drm/i915/selftests/i915_sw_fence.c ret = i915_sw_fence_await_sw_fence_gfp(fences[i], fences 471 drivers/gpu/drm/i915/selftests/i915_sw_fence.c fences[i - 1], fences 478 drivers/gpu/drm/i915/selftests/i915_sw_fence.c i915_sw_fence_commit(fences[i]); fences 484 drivers/gpu/drm/i915/selftests/i915_sw_fence.c if (i915_sw_fence_done(fences[i])) { fences 490 drivers/gpu/drm/i915/selftests/i915_sw_fence.c i915_sw_fence_commit(fences[0]); fences 492 drivers/gpu/drm/i915/selftests/i915_sw_fence.c if (!i915_sw_fence_done(fences[i])) { fences 500 drivers/gpu/drm/i915/selftests/i915_sw_fence.c free_fence(fences[i]); fences 501 drivers/gpu/drm/i915/selftests/i915_sw_fence.c kfree(fences); fences 395 drivers/gpu/drm/radeon/radeon.h struct radeon_fence **fences, fences 612 drivers/gpu/drm/radeon/radeon_fence.c struct radeon_fence **fences, fences 622 drivers/gpu/drm/radeon/radeon_fence.c if (!fences[i]) { fences 626 drivers/gpu/drm/radeon/radeon_fence.c seq[i] = fences[i]->seq; fences 248 drivers/gpu/drm/radeon/radeon_sa.c struct radeon_fence **fences, fences 278 drivers/gpu/drm/radeon/radeon_sa.c fences[i] = sa_bo->fence; fences 317 drivers/gpu/drm/radeon/radeon_sa.c struct radeon_fence *fences[RADEON_NUM_RINGS]; fences 336 drivers/gpu/drm/radeon/radeon_sa.c fences[i] = NULL; fences 350 drivers/gpu/drm/radeon/radeon_sa.c } while (radeon_sa_bo_next_hole(sa_manager, fences, tries)); fences 353 drivers/gpu/drm/radeon/radeon_sa.c radeon_fence_ref(fences[i]); fences 356 drivers/gpu/drm/radeon/radeon_sa.c r = radeon_fence_wait_any(rdev, fences, false); fences 358 drivers/gpu/drm/radeon/radeon_sa.c radeon_fence_unref(&fences[i]); fences 36 drivers/gpu/drm/radeon/radeon_trace.h __field(u32, fences) fences 42 drivers/gpu/drm/radeon/radeon_trace.h __entry->fences = radeon_fence_count_emitted( fences 47 drivers/gpu/drm/radeon/radeon_trace.h __entry->fences) fences 97 drivers/gpu/drm/virtio/virtgpu_drv.h struct list_head fences; fences 98 drivers/gpu/drm/virtio/virtgpu_fence.c list_add_tail(&fence->node, &drv->fences); fences 116 drivers/gpu/drm/virtio/virtgpu_fence.c list_for_each_entry_safe(fence, tmp, &drv->fences, node) { fences 145 drivers/gpu/drm/virtio/virtgpu_kms.c INIT_LIST_HEAD(&vgdev->fence_drv.fences); fences 703 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c struct dma_fence *child = fence_array->fences[i]; fences 43 include/linux/dma-fence-array.h struct dma_fence **fences; fences 78 include/linux/dma-fence-array.h struct dma_fence **fences, fences 540 include/linux/dma-fence.h signed long dma_fence_wait_any_timeout(struct dma_fence **fences, fences 445 include/uapi/drm/amdgpu_drm.h __u64 fences;