preempt_count      24 arch/alpha/include/asm/thread_info.h 	int			preempt_count; /* 0 => preemptable, <0 => BUG */
preempt_count      39 arch/alpha/include/asm/thread_info.h 	.preempt_count	= INIT_PREEMPT_COUNT,	\
preempt_count      41 arch/arc/include/asm/thread_info.h 	int preempt_count;		/* 0 => preemptable, <0 => BUG */
preempt_count      58 arch/arc/include/asm/thread_info.h 	.preempt_count  = INIT_PREEMPT_COUNT,	\
preempt_count      31 arch/arc/kernel/asm-offsets.c 	       offsetof(struct thread_info, preempt_count));
preempt_count      48 arch/arm/include/asm/thread_info.h 	int			preempt_count;	/* 0 => preemptable, <0 => bug */
preempt_count      74 arch/arm/include/asm/thread_info.h 	.preempt_count	= INIT_PREEMPT_COUNT,				\
preempt_count      55 arch/arm/kernel/asm-offsets.c   DEFINE(TI_PREEMPT,		offsetof(struct thread_info, preempt_count));
preempt_count      22 arch/arm64/include/asm/preempt.h 	task_thread_info(p)->preempt_count = FORK_PREEMPT_COUNT; \
preempt_count      26 arch/arm64/include/asm/preempt.h 	task_thread_info(p)->preempt_count = PREEMPT_ENABLED; \
preempt_count      61 arch/arm64/include/asm/preempt.h 	u64 pc = READ_ONCE(ti->preempt_count);
preempt_count      73 arch/arm64/include/asm/preempt.h 	return !pc || !READ_ONCE(ti->preempt_count);
preempt_count      78 arch/arm64/include/asm/preempt.h 	u64 pc = READ_ONCE(current_thread_info()->preempt_count);
preempt_count      33 arch/arm64/include/asm/thread_info.h 		u64		preempt_count;	/* 0 => preemptible, <0 => bug */
preempt_count     108 arch/arm64/include/asm/thread_info.h 	.preempt_count	= INIT_PREEMPT_COUNT,				\
preempt_count      32 arch/arm64/kernel/asm-offsets.c   DEFINE(TSK_TI_PREEMPT,	offsetof(struct task_struct, thread_info.preempt_count));
preempt_count      42 arch/c6x/include/asm/thread_info.h 	int			preempt_count;	/* 0 = preemptable, <0 = BUG */
preempt_count      56 arch/c6x/include/asm/thread_info.h 	.preempt_count	= INIT_PREEMPT_COUNT,	\
preempt_count     104 arch/c6x/kernel/asm-offsets.c 	OFFSET(THREAD_INFO_PREEMPT_COUNT, thread_info, preempt_count);
preempt_count      19 arch/csky/include/asm/thread_info.h 	int			preempt_count;
preempt_count      30 arch/csky/include/asm/thread_info.h 	.preempt_count  = INIT_PREEMPT_COUNT,	\
preempt_count      29 arch/csky/kernel/asm-offsets.c 	DEFINE(TINFO_PREEMPT,     offsetof(struct thread_info, preempt_count));
preempt_count      33 arch/h8300/include/asm/thread_info.h 	int		   preempt_count;	/* 0 => preemptable, <0 => BUG */
preempt_count      45 arch/h8300/include/asm/thread_info.h 	.preempt_count = INIT_PREEMPT_COUNT,	\
preempt_count      65 arch/h8300/kernel/asm-offsets.c 	OFFSET(TI_PRE, thread_info, preempt_count);
preempt_count      39 arch/hexagon/include/asm/thread_info.h 	int                     preempt_count;  /* 0=>preemptible,<0=>BUG */
preempt_count      68 arch/hexagon/include/asm/thread_info.h 	.preempt_count  = 1,                    \
preempt_count      31 arch/ia64/include/asm/thread_info.h 	int preempt_count;		/* 0=premptable, <0=BUG; will also serve as bh-counter */
preempt_count      52 arch/ia64/include/asm/thread_info.h 	.preempt_count	= INIT_PREEMPT_COUNT,	\
preempt_count      41 arch/ia64/kernel/asm-offsets.c 	DEFINE(TI_PRE_COUNT, offsetof(struct thread_info, preempt_count));
preempt_count    1814 arch/ia64/kernel/mca.c 	ti->preempt_count = 1;
preempt_count      31 arch/m68k/include/asm/thread_info.h 	int			preempt_count;	/* 0 => preemptable, <0 => BUG */
preempt_count      41 arch/m68k/include/asm/thread_info.h 	.preempt_count	= INIT_PREEMPT_COUNT,	\
preempt_count      42 arch/m68k/kernel/asm-offsets.c 	DEFINE(TINFO_PREEMPT, offsetof(struct thread_info, preempt_count));
preempt_count      71 arch/microblaze/include/asm/thread_info.h 	__s32			preempt_count; /* 0 => preemptable,< 0 => BUG*/
preempt_count      85 arch/microblaze/include/asm/thread_info.h 	.preempt_count	= INIT_PREEMPT_COUNT,	\
preempt_count      94 arch/microblaze/kernel/asm-offsets.c 	DEFINE(TI_PREEMPT_COUNT, offsetof(struct thread_info, preempt_count));
preempt_count      30 arch/mips/include/asm/thread_info.h 	int			preempt_count;	/* 0 => preemptable, <0 => BUG */
preempt_count      48 arch/mips/include/asm/thread_info.h 	.preempt_count	= INIT_PREEMPT_COUNT,	\
preempt_count     100 arch/mips/kernel/asm-offsets.c 	OFFSET(TI_PRE_COUNT, thread_info, preempt_count);
preempt_count      27 arch/nds32/include/asm/thread_info.h 	__s32 preempt_count;	/* 0 => preemptable, <0 => bug */
preempt_count      32 arch/nds32/include/asm/thread_info.h 	.preempt_count	= INIT_PREEMPT_COUNT,				\
preempt_count      14 arch/nds32/kernel/asm-offsets.c 	       offsetof(struct task_struct, thread_info.preempt_count));
preempt_count      44 arch/nios2/include/asm/thread_info.h 	int			preempt_count;	/* 0 => preemptable,<0 => BUG */
preempt_count      62 arch/nios2/include/asm/thread_info.h 	.preempt_count	= INIT_PREEMPT_COUNT,	\
preempt_count      70 arch/nios2/kernel/asm-offsets.c 	OFFSET(TI_PREEMPT_COUNT, thread_info, preempt_count);
preempt_count      49 arch/openrisc/include/asm/thread_info.h 	__s32			preempt_count; /* 0 => preemptable, <0 => BUG */
preempt_count      73 arch/openrisc/include/asm/thread_info.h 	.preempt_count	= INIT_PREEMPT_COUNT,		\
preempt_count      51 arch/openrisc/kernel/asm-offsets.c 	DEFINE(TI_PREEMPT, offsetof(struct thread_info, preempt_count));
preempt_count      16 arch/parisc/include/asm/thread_info.h 	int preempt_count;		/* 0=premptable, <0=BUG; will also serve as bh-counter */
preempt_count      25 arch/parisc/include/asm/thread_info.h 	.preempt_count	= INIT_PREEMPT_COUNT,	\
preempt_count     235 arch/parisc/kernel/asm-offsets.c 	DEFINE(TI_PRE_COUNT, offsetof(struct thread_info, preempt_count));
preempt_count      31 arch/powerpc/include/asm/thread_info.h 	int		preempt_count;		/* 0 => preemptable,
preempt_count      53 arch/powerpc/include/asm/thread_info.h 	.preempt_count = INIT_PREEMPT_COUNT,	\
preempt_count     169 arch/powerpc/kernel/asm-offsets.c 	OFFSET(TI_PREEMPT, thread_info, preempt_count);
preempt_count     321 arch/powerpc/kernel/machine_kexec_64.c 	current_thread_info()->preempt_count = HARDIRQ_OFFSET;
preempt_count      82 arch/powerpc/platforms/cell/smp.c 	task_thread_info(paca_ptrs[lcpu]->__current)->preempt_count	= 0;
preempt_count     110 arch/powerpc/platforms/pseries/smp.c 	task_thread_info(paca_ptrs[lcpu]->__current)->preempt_count	= 0;
preempt_count      37 arch/riscv/include/asm/thread_info.h 	int                     preempt_count;  /* 0=>preemptible, <0=>BUG */
preempt_count      57 arch/riscv/include/asm/thread_info.h 	.preempt_count	= INIT_PREEMPT_COUNT,	\
preempt_count      34 arch/riscv/kernel/asm-offsets.c 	OFFSET(TASK_TI_PREEMPT_COUNT, task_struct, thread_info.preempt_count);
preempt_count     132 arch/s390/include/asm/lowcore.h 	__s32	preempt_count;			/* 0x03a8 */
preempt_count      17 arch/s390/include/asm/preempt.h 	return READ_ONCE(S390_lowcore.preempt_count) & ~PREEMPT_NEED_RESCHED;
preempt_count      25 arch/s390/include/asm/preempt.h 		old = READ_ONCE(S390_lowcore.preempt_count);
preempt_count      28 arch/s390/include/asm/preempt.h 	} while (__atomic_cmpxchg(&S390_lowcore.preempt_count,
preempt_count      35 arch/s390/include/asm/preempt.h 	S390_lowcore.preempt_count = PREEMPT_ENABLED; \
preempt_count      40 arch/s390/include/asm/preempt.h 	__atomic_and(~PREEMPT_NEED_RESCHED, &S390_lowcore.preempt_count);
preempt_count      45 arch/s390/include/asm/preempt.h 	__atomic_or(PREEMPT_NEED_RESCHED, &S390_lowcore.preempt_count);
preempt_count      50 arch/s390/include/asm/preempt.h 	return !(READ_ONCE(S390_lowcore.preempt_count) & PREEMPT_NEED_RESCHED);
preempt_count      56 arch/s390/include/asm/preempt.h 		__atomic_add_const(val, &S390_lowcore.preempt_count);
preempt_count      58 arch/s390/include/asm/preempt.h 		__atomic_add(val, &S390_lowcore.preempt_count);
preempt_count      68 arch/s390/include/asm/preempt.h 	return __atomic_add(-1, &S390_lowcore.preempt_count) == 1;
preempt_count      73 arch/s390/include/asm/preempt.h 	return unlikely(READ_ONCE(S390_lowcore.preempt_count) ==
preempt_count      83 arch/s390/include/asm/preempt.h 	return READ_ONCE(S390_lowcore.preempt_count);
preempt_count      88 arch/s390/include/asm/preempt.h 	S390_lowcore.preempt_count = pc;
preempt_count      94 arch/s390/include/asm/preempt.h 	S390_lowcore.preempt_count = PREEMPT_ENABLED; \
preempt_count     112 arch/s390/include/asm/preempt.h 	S390_lowcore.preempt_count += val;
preempt_count     117 arch/s390/include/asm/preempt.h 	S390_lowcore.preempt_count -= val;
preempt_count     122 arch/s390/include/asm/preempt.h 	return !--S390_lowcore.preempt_count && tif_need_resched();
preempt_count     127 arch/s390/include/asm/preempt.h 	return unlikely(preempt_count() == preempt_offset &&
preempt_count     177 arch/s390/kernel/asm-offsets.c 	OFFSET(__LC_PREEMPT_COUNT, lowcore, preempt_count);
preempt_count     177 arch/s390/kernel/early.c 	S390_lowcore.preempt_count = INIT_PREEMPT_COUNT;
preempt_count     413 arch/s390/kernel/setup.c 	lc->preempt_count = S390_lowcore.preempt_count;
preempt_count      34 arch/sh/include/asm/thread_info.h 	int			preempt_count; /* 0 => preemptable, <0 => BUG */
preempt_count      62 arch/sh/include/asm/thread_info.h 	.preempt_count	= INIT_PREEMPT_COUNT,	\
preempt_count      27 arch/sh/kernel/asm-offsets.c 	DEFINE(TI_PRE_COUNT,	offsetof(struct thread_info, preempt_count));
preempt_count      95 arch/sh/kernel/irq.c 		irqctx->tinfo.preempt_count =
preempt_count      96 arch/sh/kernel/irq.c 			(irqctx->tinfo.preempt_count & ~SOFTIRQ_MASK) |
preempt_count      97 arch/sh/kernel/irq.c 			(curctx->tinfo.preempt_count & SOFTIRQ_MASK);
preempt_count     129 arch/sh/kernel/irq.c 	irqctx->tinfo.preempt_count	= HARDIRQ_OFFSET;
preempt_count     137 arch/sh/kernel/irq.c 	irqctx->tinfo.preempt_count	= 0;
preempt_count      33 arch/sparc/include/asm/thread_info_32.h 	int			preempt_count;	/* 0 => preemptable,
preempt_count      63 arch/sparc/include/asm/thread_info_32.h 	.preempt_count	=	INIT_PREEMPT_COUNT,	\
preempt_count      47 arch/sparc/include/asm/thread_info_64.h 	int			preempt_count;	/* 0 => preemptable, <0 => BUG */
preempt_count     120 arch/sparc/include/asm/thread_info_64.h 	.preempt_count	=	INIT_PREEMPT_COUNT,	\
preempt_count     377 arch/sparc/kernel/traps_32.c 	    TI_PREEMPT     != offsetof(struct thread_info, preempt_count) ||
preempt_count    2867 arch/sparc/kernel/traps_64.c 					      preempt_count) ||
preempt_count      23 arch/um/include/asm/thread_info.h 	int			preempt_count;  /* 0 => preemptable,
preempt_count      38 arch/um/include/asm/thread_info.h 	.preempt_count = INIT_PREEMPT_COUNT,	\
preempt_count      66 arch/unicore32/include/asm/thread_info.h 	int			preempt_count;	/* 0 => preemptable */
preempt_count      83 arch/unicore32/include/asm/thread_info.h 	.preempt_count	= INIT_PREEMPT_COUNT,				\
preempt_count      39 arch/unicore32/kernel/asm-offsets.c 	DEFINE(TI_PREEMPT,	offsetof(struct thread_info, preempt_count));
preempt_count     127 arch/x86/kernel/kvm.c 		    ? preempt_count() > 1 || rcu_preempt_depth()
preempt_count      53 arch/xtensa/include/asm/thread_info.h 	__s32			preempt_count;	/* 0 => preemptable,< 0 => BUG*/
preempt_count      79 arch/xtensa/include/asm/thread_info.h 	.preempt_count	= INIT_PREEMPT_COUNT,	\
preempt_count      89 arch/xtensa/kernel/asm-offsets.c 	OFFSET(TI_PRE_COUNT, thread_info, preempt_count);
preempt_count      11 include/asm-generic/preempt.h 	return READ_ONCE(current_thread_info()->preempt_count);
preempt_count      16 include/asm-generic/preempt.h 	return &current_thread_info()->preempt_count;
preempt_count      28 include/asm-generic/preempt.h 	task_thread_info(p)->preempt_count = FORK_PREEMPT_COUNT; \
preempt_count      32 include/asm-generic/preempt.h 	task_thread_info(p)->preempt_count = PREEMPT_ENABLED; \
preempt_count      77 include/asm-generic/preempt.h 	return unlikely(preempt_count() == preempt_offset &&
preempt_count      94 include/linux/bit_spinlock.h 	return preempt_count();
preempt_count     268 include/linux/ftrace.h 		WARN_ON_ONCE(!preempt_count() || !irqs_disabled());
preempt_count     281 include/linux/ftrace.h 		WARN_ON_ONCE(!preempt_count() || !irqs_disabled());
preempt_count      80 include/linux/preempt.h #define hardirq_count()	(preempt_count() & HARDIRQ_MASK)
preempt_count      81 include/linux/preempt.h #define softirq_count()	(preempt_count() & SOFTIRQ_MASK)
preempt_count      82 include/linux/preempt.h #define irq_count()	(preempt_count() & (HARDIRQ_MASK | SOFTIRQ_MASK \
preempt_count     102 include/linux/preempt.h #define in_nmi()		(preempt_count() & NMI_MASK)
preempt_count     103 include/linux/preempt.h #define in_task()		(!(preempt_count() & \
preempt_count     142 include/linux/preempt.h #define in_atomic()	(preempt_count() != 0)
preempt_count     148 include/linux/preempt.h #define in_atomic_preempt_off() (preempt_count() != PREEMPT_DISABLE_OFFSET)
preempt_count     183 include/linux/preempt.h #define preemptible()	(preempt_count() == 0 && !irqs_disabled())
preempt_count      63 include/linux/trace_events.h 	unsigned char		preempt_count;
preempt_count     932 init/main.c    	int count = preempt_count();
preempt_count     945 init/main.c    	if (preempt_count() != count) {
preempt_count     755 kernel/exit.c  			preempt_count());
preempt_count     345 kernel/rcu/rcutorture.c 		if (preempt_count() & (SOFTIRQ_MASK | HARDIRQ_MASK))
preempt_count     357 kernel/rcu/rcutorture.c 	if (!preempt_count() &&
preempt_count    2354 kernel/rcu/tree.c 	if (!(preempt_count() & PREEMPT_MASK)) {
preempt_count     612 kernel/rcu/tree_exp.h 		if (!(preempt_count() & (PREEMPT_MASK | SOFTIRQ_MASK)) ||
preempt_count     662 kernel/rcu/tree_exp.h 	if (!(preempt_count() & (PREEMPT_MASK | SOFTIRQ_MASK)) ||
preempt_count     599 kernel/rcu/tree_plugin.h 			!!(preempt_count() & (PREEMPT_MASK | SOFTIRQ_MASK));
preempt_count     690 kernel/rcu/tree_plugin.h 	    (preempt_count() & (PREEMPT_MASK | SOFTIRQ_MASK))) {
preempt_count    3197 kernel/sched/core.c 	if (WARN_ONCE(preempt_count() != 2*PREEMPT_DISABLE_OFFSET,
preempt_count    3199 kernel/sched/core.c 		      current->comm, current->pid, preempt_count()))
preempt_count    3762 kernel/sched/core.c 	if (preempt_count() == val) {
preempt_count    3777 kernel/sched/core.c 	if (DEBUG_LOCKS_WARN_ON((preempt_count() < 0)))
preempt_count    3785 kernel/sched/core.c 	DEBUG_LOCKS_WARN_ON((preempt_count() & PREEMPT_MASK) >=
preempt_count    3799 kernel/sched/core.c 	if (preempt_count() == val)
preempt_count    3809 kernel/sched/core.c 	if (DEBUG_LOCKS_WARN_ON(val > preempt_count()))
preempt_count    3815 kernel/sched/core.c 			!(preempt_count() & PREEMPT_MASK)))
preempt_count    3851 kernel/sched/core.c 		prev->comm, prev->pid, preempt_count());
preempt_count    4320 kernel/sched/core.c 	BUG_ON(preempt_count() || !irqs_disabled());
preempt_count    6723 kernel/sched/core.c 	int nested = preempt_count() + rcu_preempt_depth();
preempt_count    6804 kernel/sched/core.c 	if (preempt_count() > preempt_offset)
preempt_count     132 kernel/softirq.c 	if (preempt_count() == cnt) {
preempt_count     146 kernel/softirq.c 	if (preempt_count() == cnt)
preempt_count     287 kernel/softirq.c 		prev_count = preempt_count();
preempt_count     294 kernel/softirq.c 		if (unlikely(prev_count != preempt_count())) {
preempt_count     297 kernel/softirq.c 			       prev_count, preempt_count());
preempt_count     527 kernel/stop_machine.c 		WARN_ONCE(preempt_count(),
preempt_count    1382 kernel/time/timer.c 	int count = preempt_count();
preempt_count    1409 kernel/time/timer.c 	if (count != preempt_count()) {
preempt_count    1411 kernel/time/timer.c 			  fn, count, preempt_count());
preempt_count      80 kernel/trace/blktrace.c 		pc = preempt_count();
preempt_count     253 kernel/trace/blktrace.c 		pc = preempt_count();
preempt_count    2695 kernel/trace/ring_buffer.c 	unsigned long pc = preempt_count();
preempt_count     845 kernel/trace/trace.c 	pc = preempt_count();
preempt_count     895 kernel/trace/trace.c 	pc = preempt_count();
preempt_count    2345 kernel/trace/trace.c 	entry->preempt_count		= pc & 0xff;
preempt_count    2931 kernel/trace/trace.c 			     flags, skip, preempt_count(), NULL);
preempt_count    3122 kernel/trace/trace.c 	pc = preempt_count();
preempt_count    3182 kernel/trace/trace.c 	pc = preempt_count();
preempt_count    6417 kernel/trace/trace.c 					    irq_flags, preempt_count());
preempt_count    6497 kernel/trace/trace.c 					    irq_flags, preempt_count());
preempt_count      62 kernel/trace/trace_branch.c 	pc = preempt_count();
preempt_count     419 kernel/trace/trace_event_perf.c 	int pc = preempt_count();
preempt_count     182 kernel/trace/trace_events.c 	__common_field(unsigned char, preempt_count);
preempt_count     257 kernel/trace/trace_events.c 	fbuffer->pc = preempt_count();
preempt_count    3369 kernel/trace/trace_events.c 	pc = preempt_count();
preempt_count     138 kernel/trace/trace_functions.c 	pc = preempt_count();
preempt_count     199 kernel/trace/trace_functions.c 		pc = preempt_count();
preempt_count     408 kernel/trace/trace_functions.c 	pc = preempt_count();
preempt_count     177 kernel/trace/trace_functions_graph.c 		pc = preempt_count();
preempt_count     258 kernel/trace/trace_functions_graph.c 		pc = preempt_count();
preempt_count     113 kernel/trace/trace_hwlat.c 	pc = preempt_count();
preempt_count     122 kernel/trace/trace_irqsoff.c 	if (!irqs_disabled_flags(*flags) && !preempt_count())
preempt_count     150 kernel/trace/trace_irqsoff.c 	trace_function(tr, ip, parent_ip, flags, preempt_count());
preempt_count     198 kernel/trace/trace_irqsoff.c 	pc = preempt_count();
preempt_count     217 kernel/trace/trace_irqsoff.c 	pc = preempt_count();
preempt_count     333 kernel/trace/trace_irqsoff.c 	pc = preempt_count();
preempt_count     441 kernel/trace/trace_irqsoff.c 	int pc = preempt_count();
preempt_count     451 kernel/trace/trace_irqsoff.c 	int pc = preempt_count();
preempt_count     612 kernel/trace/trace_irqsoff.c 	unsigned int pc = preempt_count();
preempt_count     621 kernel/trace/trace_irqsoff.c 	unsigned int pc = preempt_count();
preempt_count     664 kernel/trace/trace_irqsoff.c 	int pc = preempt_count();
preempt_count     672 kernel/trace/trace_irqsoff.c 	int pc = preempt_count();
preempt_count    1192 kernel/trace/trace_kprobe.c 	pc = preempt_count();
preempt_count    1240 kernel/trace/trace_kprobe.c 	pc = preempt_count();
preempt_count     303 kernel/trace/trace_mmiotrace.c 	int pc = preempt_count();
preempt_count     333 kernel/trace/trace_mmiotrace.c 	int pc = preempt_count();
preempt_count     470 kernel/trace/trace_output.c 	if (entry->preempt_count)
preempt_count     471 kernel/trace/trace_output.c 		trace_seq_printf(s, "%x", entry->preempt_count);
preempt_count     629 kernel/trace/trace_output.c 			entry->preempt_count, iter->idx);
preempt_count      78 kernel/trace/trace_sched_wakeup.c 	*pc = preempt_count();
preempt_count     458 kernel/trace/trace_sched_wakeup.c 	pc = preempt_count();
preempt_count     553 kernel/trace/trace_sched_wakeup.c 	pc = preempt_count();
preempt_count     884 kernel/trace/trace_selftest.c 	if (preempt_count()) {
preempt_count     946 kernel/trace/trace_selftest.c 	if (preempt_count()) {
preempt_count     341 kernel/trace/trace_syscalls.c 	pc = preempt_count();
preempt_count     387 kernel/trace/trace_syscalls.c 	pc = preempt_count();
preempt_count    2284 kernel/workqueue.c 		       current->comm, preempt_count(), task_pid_nr(current),
preempt_count    1139 lib/locking-selftest.c 	unsigned long saved_preempt_count = preempt_count();
preempt_count      16 lib/smp_processor_id.c 	if (likely(preempt_count()))
preempt_count      44 lib/smp_processor_id.c 		what1, what2, preempt_count() - 1, current->comm, current->pid);
preempt_count      51 security/apparmor/include/path.h #define AA_BUG_PREEMPT_ENABLED(X) AA_BUG(preempt_count() <= 0, X)
preempt_count      16 tools/testing/radix-tree/linux.c int preempt_count;
preempt_count       5 tools/testing/radix-tree/linux/preempt.h extern int preempt_count;
preempt_count       7 tools/testing/radix-tree/linux/preempt.h #define preempt_disable()	uatomic_inc(&preempt_count)
preempt_count       8 tools/testing/radix-tree/linux/preempt.h #define preempt_enable()	uatomic_dec(&preempt_count)
preempt_count     244 tools/testing/radix-tree/main.c 		nr_allocated, preempt_count);
preempt_count     248 tools/testing/radix-tree/main.c 		nr_allocated, preempt_count);
preempt_count     252 tools/testing/radix-tree/main.c 		nr_allocated, preempt_count);
preempt_count     256 tools/testing/radix-tree/main.c 		nr_allocated, preempt_count);
preempt_count     260 tools/testing/radix-tree/main.c 		nr_allocated, preempt_count);
preempt_count     264 tools/testing/radix-tree/main.c 		nr_allocated, preempt_count);
preempt_count     269 tools/testing/radix-tree/main.c 		nr_allocated, preempt_count);
preempt_count     273 tools/testing/radix-tree/main.c 		nr_allocated, preempt_count);
preempt_count     281 tools/testing/radix-tree/main.c 		nr_allocated, preempt_count);
preempt_count     324 tools/testing/radix-tree/main.c 		nr_allocated, preempt_count);
preempt_count      40 tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/preempt.h 	return !preempt_count();