urelocs 1447 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c struct drm_i915_gem_relocation_entry __user *urelocs; urelocs 1451 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c urelocs = u64_to_user_ptr(entry->relocs_ptr); urelocs 1461 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (unlikely(!access_ok(urelocs, remain*sizeof(*urelocs)))) urelocs 1479 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c copied = __copy_from_user_inatomic(r, urelocs, count * sizeof(r[0])); urelocs 1517 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (unlikely(__put_user(offset, &urelocs[r-stack].presumed_offset))) { urelocs 1523 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c urelocs += ARRAY_SIZE(stack); urelocs 1589 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c struct drm_i915_gem_relocation_entry __user *urelocs; urelocs 1600 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c urelocs = u64_to_user_ptr(eb->exec[i].relocs_ptr); urelocs 1616 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c (char __user *)urelocs + copied, urelocs 1633 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (!user_access_begin(urelocs, size)) urelocs 1638 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c &urelocs[copied].presumed_offset,