__copy_from_user_inatomic 44 arch/arm/kernel/perf_callchain.c err = __copy_from_user_inatomic(&buftail, tail, sizeof(buftail)); __copy_from_user_inatomic 93 arch/arm/oprofile/common.c if (__copy_from_user_inatomic(buftail, tail, sizeof(buftail))) __copy_from_user_inatomic 35 arch/arm64/kernel/perf_callchain.c err = __copy_from_user_inatomic(&buftail, tail, sizeof(buftail)); __copy_from_user_inatomic 82 arch/arm64/kernel/perf_callchain.c err = __copy_from_user_inatomic(&buftail, tail, sizeof(buftail)); __copy_from_user_inatomic 57 arch/csky/kernel/perf_callchain.c if (__copy_from_user_inatomic(&buftail, user_frame_tail, __copy_from_user_inatomic 1273 arch/mips/kernel/unaligned.c res = __copy_from_user_inatomic(fpr, addr, __copy_from_user_inatomic 24 arch/mips/oprofile/backtrace.c if (__copy_from_user_inatomic(result, address, sizeof(unsigned long))) __copy_from_user_inatomic 1311 arch/nds32/kernel/perf_event_cpu.c if (__copy_from_user_inatomic __copy_from_user_inatomic 1337 arch/nds32/kernel/perf_event_cpu.c if (__copy_from_user_inatomic __copy_from_user_inatomic 1392 arch/nds32/kernel/perf_event_cpu.c if (__copy_from_user_inatomic __copy_from_user_inatomic 1412 arch/nds32/kernel/perf_event_cpu.c if (__copy_from_user_inatomic __copy_from_user_inatomic 1430 arch/nds32/kernel/perf_event_cpu.c if (__copy_from_user_inatomic __copy_from_user_inatomic 940 arch/powerpc/kernel/traps.c rc = __copy_from_user_inatomic(vbuf, addr, 16); __copy_from_user_inatomic 39 arch/powerpc/oprofile/backtrace.c if (__copy_from_user_inatomic(stack_frame, p, sizeof(stack_frame))) __copy_from_user_inatomic 60 arch/powerpc/oprofile/backtrace.c if (__copy_from_user_inatomic(stack_frame, (void __user *)sp, __copy_from_user_inatomic 28 arch/riscv/kernel/perf_callchain.c if (__copy_from_user_inatomic(&buftail, user_frame_tail, __copy_from_user_inatomic 53 arch/sh/oprofile/backtrace.c if (__copy_from_user_inatomic(&buf_stack, stackaddr, sizeof(unsigned long))) __copy_from_user_inatomic 1812 arch/sparc/kernel/perf_event.c if (__copy_from_user_inatomic(&sf, usf, sizeof(sf))) __copy_from_user_inatomic 1836 arch/sparc/kernel/perf_event.c if (__copy_from_user_inatomic(&sf, usf, sizeof(sf))) __copy_from_user_inatomic 1844 arch/sparc/kernel/perf_event.c if (__copy_from_user_inatomic(&sf, usf, sizeof(sf))) __copy_from_user_inatomic 704 arch/x86/include/asm/uaccess.h #define __copy_from_user_nmi __copy_from_user_inatomic __copy_from_user_inatomic 108 arch/x86/kernel/stacktrace.c if (__copy_from_user_inatomic(frame, fp, sizeof(*frame))) __copy_from_user_inatomic 33 arch/x86/lib/usercopy.c ret = __copy_from_user_inatomic(to, from, n); __copy_from_user_inatomic 1479 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c copied = __copy_from_user_inatomic(r, urelocs, count * sizeof(r[0])); __copy_from_user_inatomic 394 drivers/gpu/drm/i915/gem/i915_gem_shmem.c unwritten = __copy_from_user_inatomic(vaddr + pg, __copy_from_user_inatomic 229 include/linux/uaccess.h return __copy_from_user_inatomic(to, from, n); __copy_from_user_inatomic 196 kernel/events/internal.h ret = __copy_from_user_inatomic(dst, src, n); __copy_from_user_inatomic 6425 kernel/trace/trace.c len = __copy_from_user_inatomic(&entry->buf, ubuf, cnt); __copy_from_user_inatomic 6504 kernel/trace/trace.c len = __copy_from_user_inatomic(&entry->id, ubuf, cnt); __copy_from_user_inatomic 508 lib/test_kasan.c unused = __copy_from_user_inatomic(kmem, usermem, size + 1); __copy_from_user_inatomic 15 mm/maccess.c ret = __copy_from_user_inatomic(dst, src, size); __copy_from_user_inatomic 2168 mm/memory.c if (__copy_from_user_inatomic(kaddr, uaddr, PAGE_SIZE)) __copy_from_user_inatomic 2120 virt/kvm/kvm_main.c r = __copy_from_user_inatomic(data, (void __user *)addr + offset, len);