__copy_from_user_inatomic   44 arch/arm/kernel/perf_callchain.c 	err = __copy_from_user_inatomic(&buftail, tail, sizeof(buftail));
__copy_from_user_inatomic   93 arch/arm/oprofile/common.c 	if (__copy_from_user_inatomic(buftail, tail, sizeof(buftail)))
__copy_from_user_inatomic   35 arch/arm64/kernel/perf_callchain.c 	err = __copy_from_user_inatomic(&buftail, tail, sizeof(buftail));
__copy_from_user_inatomic   82 arch/arm64/kernel/perf_callchain.c 	err = __copy_from_user_inatomic(&buftail, tail, sizeof(buftail));
__copy_from_user_inatomic   57 arch/csky/kernel/perf_callchain.c 	if (__copy_from_user_inatomic(&buftail, user_frame_tail,
__copy_from_user_inatomic 1273 arch/mips/kernel/unaligned.c 				res = __copy_from_user_inatomic(fpr, addr,
__copy_from_user_inatomic   24 arch/mips/oprofile/backtrace.c 	if (__copy_from_user_inatomic(result, address, sizeof(unsigned long)))
__copy_from_user_inatomic 1311 arch/nds32/kernel/perf_event_cpu.c 	if (__copy_from_user_inatomic
__copy_from_user_inatomic 1337 arch/nds32/kernel/perf_event_cpu.c 	if (__copy_from_user_inatomic
__copy_from_user_inatomic 1392 arch/nds32/kernel/perf_event_cpu.c 		if (__copy_from_user_inatomic
__copy_from_user_inatomic 1412 arch/nds32/kernel/perf_event_cpu.c 			if (__copy_from_user_inatomic
__copy_from_user_inatomic 1430 arch/nds32/kernel/perf_event_cpu.c 				if (__copy_from_user_inatomic
__copy_from_user_inatomic  940 arch/powerpc/kernel/traps.c 		rc = __copy_from_user_inatomic(vbuf, addr, 16);
__copy_from_user_inatomic   39 arch/powerpc/oprofile/backtrace.c 	if (__copy_from_user_inatomic(stack_frame, p, sizeof(stack_frame)))
__copy_from_user_inatomic   60 arch/powerpc/oprofile/backtrace.c 	if (__copy_from_user_inatomic(stack_frame, (void __user *)sp,
__copy_from_user_inatomic   28 arch/riscv/kernel/perf_callchain.c 	if (__copy_from_user_inatomic(&buftail, user_frame_tail,
__copy_from_user_inatomic   53 arch/sh/oprofile/backtrace.c 	if (__copy_from_user_inatomic(&buf_stack, stackaddr, sizeof(unsigned long)))
__copy_from_user_inatomic 1812 arch/sparc/kernel/perf_event.c 		if (__copy_from_user_inatomic(&sf, usf, sizeof(sf)))
__copy_from_user_inatomic 1836 arch/sparc/kernel/perf_event.c 			if (__copy_from_user_inatomic(&sf, usf, sizeof(sf)))
__copy_from_user_inatomic 1844 arch/sparc/kernel/perf_event.c 			if (__copy_from_user_inatomic(&sf, usf, sizeof(sf)))
__copy_from_user_inatomic  704 arch/x86/include/asm/uaccess.h #define __copy_from_user_nmi __copy_from_user_inatomic
__copy_from_user_inatomic  108 arch/x86/kernel/stacktrace.c 	if (__copy_from_user_inatomic(frame, fp, sizeof(*frame)))
__copy_from_user_inatomic   33 arch/x86/lib/usercopy.c 	ret = __copy_from_user_inatomic(to, from, n);
__copy_from_user_inatomic 1479 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		copied = __copy_from_user_inatomic(r, urelocs, count * sizeof(r[0]));
__copy_from_user_inatomic  394 drivers/gpu/drm/i915/gem/i915_gem_shmem.c 		unwritten = __copy_from_user_inatomic(vaddr + pg,
__copy_from_user_inatomic  229 include/linux/uaccess.h 	return __copy_from_user_inatomic(to, from, n);
__copy_from_user_inatomic  196 kernel/events/internal.h 	ret = __copy_from_user_inatomic(dst, src, n);
__copy_from_user_inatomic 6425 kernel/trace/trace.c 	len = __copy_from_user_inatomic(&entry->buf, ubuf, cnt);
__copy_from_user_inatomic 6504 kernel/trace/trace.c 	len = __copy_from_user_inatomic(&entry->id, ubuf, cnt);
__copy_from_user_inatomic  508 lib/test_kasan.c 	unused = __copy_from_user_inatomic(kmem, usermem, size + 1);
__copy_from_user_inatomic   15 mm/maccess.c   	ret = __copy_from_user_inatomic(dst, src, size);
__copy_from_user_inatomic 2168 mm/memory.c    		if (__copy_from_user_inatomic(kaddr, uaddr, PAGE_SIZE))
__copy_from_user_inatomic 2120 virt/kvm/kvm_main.c 	r = __copy_from_user_inatomic(data, (void __user *)addr + offset, len);