__preempt_count 9 arch/x86/include/asm/preempt.h DECLARE_PER_CPU(int, __preempt_count); __preempt_count 26 arch/x86/include/asm/preempt.h return raw_cpu_read_4(__preempt_count) & ~PREEMPT_NEED_RESCHED; __preempt_count 34 arch/x86/include/asm/preempt.h old = raw_cpu_read_4(__preempt_count); __preempt_count 37 arch/x86/include/asm/preempt.h } while (raw_cpu_cmpxchg_4(__preempt_count, old, new) != old); __preempt_count 46 arch/x86/include/asm/preempt.h per_cpu(__preempt_count, (cpu)) = PREEMPT_ENABLED; \ __preempt_count 60 arch/x86/include/asm/preempt.h raw_cpu_and_4(__preempt_count, ~PREEMPT_NEED_RESCHED); __preempt_count 65 arch/x86/include/asm/preempt.h raw_cpu_or_4(__preempt_count, PREEMPT_NEED_RESCHED); __preempt_count 70 arch/x86/include/asm/preempt.h return !(raw_cpu_read_4(__preempt_count) & PREEMPT_NEED_RESCHED); __preempt_count 79 arch/x86/include/asm/preempt.h raw_cpu_add_4(__preempt_count, val); __preempt_count 84 arch/x86/include/asm/preempt.h raw_cpu_add_4(__preempt_count, -val); __preempt_count 94 arch/x86/include/asm/preempt.h return GEN_UNARY_RMWcc("decl", __preempt_count, e, __percpu_arg([var])); __preempt_count 102 arch/x86/include/asm/preempt.h return unlikely(raw_cpu_read_4(__preempt_count) == preempt_offset); __preempt_count 1705 arch/x86/kernel/cpu/common.c DEFINE_PER_CPU(int, __preempt_count) = INIT_PREEMPT_COUNT; __preempt_count 1706 arch/x86/kernel/cpu/common.c EXPORT_PER_CPU_SYMBOL(__preempt_count); __preempt_count 1762 arch/x86/kernel/cpu/common.c DEFINE_PER_CPU(int, __preempt_count) = INIT_PREEMPT_COUNT; __preempt_count 1763 arch/x86/kernel/cpu/common.c EXPORT_PER_CPU_SYMBOL(__preempt_count);