__preempt_count     9 arch/x86/include/asm/preempt.h DECLARE_PER_CPU(int, __preempt_count);
__preempt_count    26 arch/x86/include/asm/preempt.h 	return raw_cpu_read_4(__preempt_count) & ~PREEMPT_NEED_RESCHED;
__preempt_count    34 arch/x86/include/asm/preempt.h 		old = raw_cpu_read_4(__preempt_count);
__preempt_count    37 arch/x86/include/asm/preempt.h 	} while (raw_cpu_cmpxchg_4(__preempt_count, old, new) != old);
__preempt_count    46 arch/x86/include/asm/preempt.h 	per_cpu(__preempt_count, (cpu)) = PREEMPT_ENABLED; \
__preempt_count    60 arch/x86/include/asm/preempt.h 	raw_cpu_and_4(__preempt_count, ~PREEMPT_NEED_RESCHED);
__preempt_count    65 arch/x86/include/asm/preempt.h 	raw_cpu_or_4(__preempt_count, PREEMPT_NEED_RESCHED);
__preempt_count    70 arch/x86/include/asm/preempt.h 	return !(raw_cpu_read_4(__preempt_count) & PREEMPT_NEED_RESCHED);
__preempt_count    79 arch/x86/include/asm/preempt.h 	raw_cpu_add_4(__preempt_count, val);
__preempt_count    84 arch/x86/include/asm/preempt.h 	raw_cpu_add_4(__preempt_count, -val);
__preempt_count    94 arch/x86/include/asm/preempt.h 	return GEN_UNARY_RMWcc("decl", __preempt_count, e, __percpu_arg([var]));
__preempt_count   102 arch/x86/include/asm/preempt.h 	return unlikely(raw_cpu_read_4(__preempt_count) == preempt_offset);
__preempt_count  1705 arch/x86/kernel/cpu/common.c DEFINE_PER_CPU(int, __preempt_count) = INIT_PREEMPT_COUNT;
__preempt_count  1706 arch/x86/kernel/cpu/common.c EXPORT_PER_CPU_SYMBOL(__preempt_count);
__preempt_count  1762 arch/x86/kernel/cpu/common.c DEFINE_PER_CPU(int, __preempt_count) = INIT_PREEMPT_COUNT;
__preempt_count  1763 arch/x86/kernel/cpu/common.c EXPORT_PER_CPU_SYMBOL(__preempt_count);