urelocs          1447 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	struct drm_i915_gem_relocation_entry __user *urelocs;
urelocs          1451 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	urelocs = u64_to_user_ptr(entry->relocs_ptr);
urelocs          1461 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	if (unlikely(!access_ok(urelocs, remain*sizeof(*urelocs))))
urelocs          1479 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		copied = __copy_from_user_inatomic(r, urelocs, count * sizeof(r[0]));
urelocs          1517 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 				if (unlikely(__put_user(offset, &urelocs[r-stack].presumed_offset))) {
urelocs          1523 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		urelocs += ARRAY_SIZE(stack);
urelocs          1589 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		struct drm_i915_gem_relocation_entry __user *urelocs;
urelocs          1600 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		urelocs = u64_to_user_ptr(eb->exec[i].relocs_ptr);
urelocs          1616 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 					     (char __user *)urelocs + copied,
urelocs          1633 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		if (!user_access_begin(urelocs, size))
urelocs          1638 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 					&urelocs[copied].presumed_offset,