preempt_count 24 arch/alpha/include/asm/thread_info.h int preempt_count; /* 0 => preemptable, <0 => BUG */ preempt_count 39 arch/alpha/include/asm/thread_info.h .preempt_count = INIT_PREEMPT_COUNT, \ preempt_count 41 arch/arc/include/asm/thread_info.h int preempt_count; /* 0 => preemptable, <0 => BUG */ preempt_count 58 arch/arc/include/asm/thread_info.h .preempt_count = INIT_PREEMPT_COUNT, \ preempt_count 31 arch/arc/kernel/asm-offsets.c offsetof(struct thread_info, preempt_count)); preempt_count 48 arch/arm/include/asm/thread_info.h int preempt_count; /* 0 => preemptable, <0 => bug */ preempt_count 74 arch/arm/include/asm/thread_info.h .preempt_count = INIT_PREEMPT_COUNT, \ preempt_count 55 arch/arm/kernel/asm-offsets.c DEFINE(TI_PREEMPT, offsetof(struct thread_info, preempt_count)); preempt_count 22 arch/arm64/include/asm/preempt.h task_thread_info(p)->preempt_count = FORK_PREEMPT_COUNT; \ preempt_count 26 arch/arm64/include/asm/preempt.h task_thread_info(p)->preempt_count = PREEMPT_ENABLED; \ preempt_count 61 arch/arm64/include/asm/preempt.h u64 pc = READ_ONCE(ti->preempt_count); preempt_count 73 arch/arm64/include/asm/preempt.h return !pc || !READ_ONCE(ti->preempt_count); preempt_count 78 arch/arm64/include/asm/preempt.h u64 pc = READ_ONCE(current_thread_info()->preempt_count); preempt_count 33 arch/arm64/include/asm/thread_info.h u64 preempt_count; /* 0 => preemptible, <0 => bug */ preempt_count 108 arch/arm64/include/asm/thread_info.h .preempt_count = INIT_PREEMPT_COUNT, \ preempt_count 32 arch/arm64/kernel/asm-offsets.c DEFINE(TSK_TI_PREEMPT, offsetof(struct task_struct, thread_info.preempt_count)); preempt_count 42 arch/c6x/include/asm/thread_info.h int preempt_count; /* 0 = preemptable, <0 = BUG */ preempt_count 56 arch/c6x/include/asm/thread_info.h .preempt_count = INIT_PREEMPT_COUNT, \ preempt_count 104 arch/c6x/kernel/asm-offsets.c OFFSET(THREAD_INFO_PREEMPT_COUNT, thread_info, preempt_count); preempt_count 19 arch/csky/include/asm/thread_info.h int preempt_count; preempt_count 30 arch/csky/include/asm/thread_info.h .preempt_count = INIT_PREEMPT_COUNT, \ preempt_count 29 arch/csky/kernel/asm-offsets.c DEFINE(TINFO_PREEMPT, offsetof(struct thread_info, preempt_count)); preempt_count 33 arch/h8300/include/asm/thread_info.h int preempt_count; /* 0 => preemptable, <0 => BUG */ preempt_count 45 arch/h8300/include/asm/thread_info.h .preempt_count = INIT_PREEMPT_COUNT, \ preempt_count 65 arch/h8300/kernel/asm-offsets.c OFFSET(TI_PRE, thread_info, preempt_count); preempt_count 39 arch/hexagon/include/asm/thread_info.h int preempt_count; /* 0=>preemptible,<0=>BUG */ preempt_count 68 arch/hexagon/include/asm/thread_info.h .preempt_count = 1, \ preempt_count 31 arch/ia64/include/asm/thread_info.h int preempt_count; /* 0=premptable, <0=BUG; will also serve as bh-counter */ preempt_count 52 arch/ia64/include/asm/thread_info.h .preempt_count = INIT_PREEMPT_COUNT, \ preempt_count 41 arch/ia64/kernel/asm-offsets.c DEFINE(TI_PRE_COUNT, offsetof(struct thread_info, preempt_count)); preempt_count 1814 arch/ia64/kernel/mca.c ti->preempt_count = 1; preempt_count 31 arch/m68k/include/asm/thread_info.h int preempt_count; /* 0 => preemptable, <0 => BUG */ preempt_count 41 arch/m68k/include/asm/thread_info.h .preempt_count = INIT_PREEMPT_COUNT, \ preempt_count 42 arch/m68k/kernel/asm-offsets.c DEFINE(TINFO_PREEMPT, offsetof(struct thread_info, preempt_count)); preempt_count 71 arch/microblaze/include/asm/thread_info.h __s32 preempt_count; /* 0 => preemptable,< 0 => BUG*/ preempt_count 85 arch/microblaze/include/asm/thread_info.h .preempt_count = INIT_PREEMPT_COUNT, \ preempt_count 94 arch/microblaze/kernel/asm-offsets.c DEFINE(TI_PREEMPT_COUNT, offsetof(struct thread_info, preempt_count)); preempt_count 30 arch/mips/include/asm/thread_info.h int preempt_count; /* 0 => preemptable, <0 => BUG */ preempt_count 48 arch/mips/include/asm/thread_info.h .preempt_count = INIT_PREEMPT_COUNT, \ preempt_count 100 arch/mips/kernel/asm-offsets.c OFFSET(TI_PRE_COUNT, thread_info, preempt_count); preempt_count 27 arch/nds32/include/asm/thread_info.h __s32 preempt_count; /* 0 => preemptable, <0 => bug */ preempt_count 32 arch/nds32/include/asm/thread_info.h .preempt_count = INIT_PREEMPT_COUNT, \ preempt_count 14 arch/nds32/kernel/asm-offsets.c offsetof(struct task_struct, thread_info.preempt_count)); preempt_count 44 arch/nios2/include/asm/thread_info.h int preempt_count; /* 0 => preemptable,<0 => BUG */ preempt_count 62 arch/nios2/include/asm/thread_info.h .preempt_count = INIT_PREEMPT_COUNT, \ preempt_count 70 arch/nios2/kernel/asm-offsets.c OFFSET(TI_PREEMPT_COUNT, thread_info, preempt_count); preempt_count 49 arch/openrisc/include/asm/thread_info.h __s32 preempt_count; /* 0 => preemptable, <0 => BUG */ preempt_count 73 arch/openrisc/include/asm/thread_info.h .preempt_count = INIT_PREEMPT_COUNT, \ preempt_count 51 arch/openrisc/kernel/asm-offsets.c DEFINE(TI_PREEMPT, offsetof(struct thread_info, preempt_count)); preempt_count 16 arch/parisc/include/asm/thread_info.h int preempt_count; /* 0=premptable, <0=BUG; will also serve as bh-counter */ preempt_count 25 arch/parisc/include/asm/thread_info.h .preempt_count = INIT_PREEMPT_COUNT, \ preempt_count 235 arch/parisc/kernel/asm-offsets.c DEFINE(TI_PRE_COUNT, offsetof(struct thread_info, preempt_count)); preempt_count 31 arch/powerpc/include/asm/thread_info.h int preempt_count; /* 0 => preemptable, preempt_count 53 arch/powerpc/include/asm/thread_info.h .preempt_count = INIT_PREEMPT_COUNT, \ preempt_count 169 arch/powerpc/kernel/asm-offsets.c OFFSET(TI_PREEMPT, thread_info, preempt_count); preempt_count 321 arch/powerpc/kernel/machine_kexec_64.c current_thread_info()->preempt_count = HARDIRQ_OFFSET; preempt_count 82 arch/powerpc/platforms/cell/smp.c task_thread_info(paca_ptrs[lcpu]->__current)->preempt_count = 0; preempt_count 110 arch/powerpc/platforms/pseries/smp.c task_thread_info(paca_ptrs[lcpu]->__current)->preempt_count = 0; preempt_count 37 arch/riscv/include/asm/thread_info.h int preempt_count; /* 0=>preemptible, <0=>BUG */ preempt_count 57 arch/riscv/include/asm/thread_info.h .preempt_count = INIT_PREEMPT_COUNT, \ preempt_count 34 arch/riscv/kernel/asm-offsets.c OFFSET(TASK_TI_PREEMPT_COUNT, task_struct, thread_info.preempt_count); preempt_count 132 arch/s390/include/asm/lowcore.h __s32 preempt_count; /* 0x03a8 */ preempt_count 17 arch/s390/include/asm/preempt.h return READ_ONCE(S390_lowcore.preempt_count) & ~PREEMPT_NEED_RESCHED; preempt_count 25 arch/s390/include/asm/preempt.h old = READ_ONCE(S390_lowcore.preempt_count); preempt_count 28 arch/s390/include/asm/preempt.h } while (__atomic_cmpxchg(&S390_lowcore.preempt_count, preempt_count 35 arch/s390/include/asm/preempt.h S390_lowcore.preempt_count = PREEMPT_ENABLED; \ preempt_count 40 arch/s390/include/asm/preempt.h __atomic_and(~PREEMPT_NEED_RESCHED, &S390_lowcore.preempt_count); preempt_count 45 arch/s390/include/asm/preempt.h __atomic_or(PREEMPT_NEED_RESCHED, &S390_lowcore.preempt_count); preempt_count 50 arch/s390/include/asm/preempt.h return !(READ_ONCE(S390_lowcore.preempt_count) & PREEMPT_NEED_RESCHED); preempt_count 56 arch/s390/include/asm/preempt.h __atomic_add_const(val, &S390_lowcore.preempt_count); preempt_count 58 arch/s390/include/asm/preempt.h __atomic_add(val, &S390_lowcore.preempt_count); preempt_count 68 arch/s390/include/asm/preempt.h return __atomic_add(-1, &S390_lowcore.preempt_count) == 1; preempt_count 73 arch/s390/include/asm/preempt.h return unlikely(READ_ONCE(S390_lowcore.preempt_count) == preempt_count 83 arch/s390/include/asm/preempt.h return READ_ONCE(S390_lowcore.preempt_count); preempt_count 88 arch/s390/include/asm/preempt.h S390_lowcore.preempt_count = pc; preempt_count 94 arch/s390/include/asm/preempt.h S390_lowcore.preempt_count = PREEMPT_ENABLED; \ preempt_count 112 arch/s390/include/asm/preempt.h S390_lowcore.preempt_count += val; preempt_count 117 arch/s390/include/asm/preempt.h S390_lowcore.preempt_count -= val; preempt_count 122 arch/s390/include/asm/preempt.h return !--S390_lowcore.preempt_count && tif_need_resched(); preempt_count 127 arch/s390/include/asm/preempt.h return unlikely(preempt_count() == preempt_offset && preempt_count 177 arch/s390/kernel/asm-offsets.c OFFSET(__LC_PREEMPT_COUNT, lowcore, preempt_count); preempt_count 177 arch/s390/kernel/early.c S390_lowcore.preempt_count = INIT_PREEMPT_COUNT; preempt_count 413 arch/s390/kernel/setup.c lc->preempt_count = S390_lowcore.preempt_count; preempt_count 34 arch/sh/include/asm/thread_info.h int preempt_count; /* 0 => preemptable, <0 => BUG */ preempt_count 62 arch/sh/include/asm/thread_info.h .preempt_count = INIT_PREEMPT_COUNT, \ preempt_count 27 arch/sh/kernel/asm-offsets.c DEFINE(TI_PRE_COUNT, offsetof(struct thread_info, preempt_count)); preempt_count 95 arch/sh/kernel/irq.c irqctx->tinfo.preempt_count = preempt_count 96 arch/sh/kernel/irq.c (irqctx->tinfo.preempt_count & ~SOFTIRQ_MASK) | preempt_count 97 arch/sh/kernel/irq.c (curctx->tinfo.preempt_count & SOFTIRQ_MASK); preempt_count 129 arch/sh/kernel/irq.c irqctx->tinfo.preempt_count = HARDIRQ_OFFSET; preempt_count 137 arch/sh/kernel/irq.c irqctx->tinfo.preempt_count = 0; preempt_count 33 arch/sparc/include/asm/thread_info_32.h int preempt_count; /* 0 => preemptable, preempt_count 63 arch/sparc/include/asm/thread_info_32.h .preempt_count = INIT_PREEMPT_COUNT, \ preempt_count 47 arch/sparc/include/asm/thread_info_64.h int preempt_count; /* 0 => preemptable, <0 => BUG */ preempt_count 120 arch/sparc/include/asm/thread_info_64.h .preempt_count = INIT_PREEMPT_COUNT, \ preempt_count 377 arch/sparc/kernel/traps_32.c TI_PREEMPT != offsetof(struct thread_info, preempt_count) || preempt_count 2867 arch/sparc/kernel/traps_64.c preempt_count) || preempt_count 23 arch/um/include/asm/thread_info.h int preempt_count; /* 0 => preemptable, preempt_count 38 arch/um/include/asm/thread_info.h .preempt_count = INIT_PREEMPT_COUNT, \ preempt_count 66 arch/unicore32/include/asm/thread_info.h int preempt_count; /* 0 => preemptable */ preempt_count 83 arch/unicore32/include/asm/thread_info.h .preempt_count = INIT_PREEMPT_COUNT, \ preempt_count 39 arch/unicore32/kernel/asm-offsets.c DEFINE(TI_PREEMPT, offsetof(struct thread_info, preempt_count)); preempt_count 127 arch/x86/kernel/kvm.c ? preempt_count() > 1 || rcu_preempt_depth() preempt_count 53 arch/xtensa/include/asm/thread_info.h __s32 preempt_count; /* 0 => preemptable,< 0 => BUG*/ preempt_count 79 arch/xtensa/include/asm/thread_info.h .preempt_count = INIT_PREEMPT_COUNT, \ preempt_count 89 arch/xtensa/kernel/asm-offsets.c OFFSET(TI_PRE_COUNT, thread_info, preempt_count); preempt_count 11 include/asm-generic/preempt.h return READ_ONCE(current_thread_info()->preempt_count); preempt_count 16 include/asm-generic/preempt.h return ¤t_thread_info()->preempt_count; preempt_count 28 include/asm-generic/preempt.h task_thread_info(p)->preempt_count = FORK_PREEMPT_COUNT; \ preempt_count 32 include/asm-generic/preempt.h task_thread_info(p)->preempt_count = PREEMPT_ENABLED; \ preempt_count 77 include/asm-generic/preempt.h return unlikely(preempt_count() == preempt_offset && preempt_count 94 include/linux/bit_spinlock.h return preempt_count(); preempt_count 268 include/linux/ftrace.h WARN_ON_ONCE(!preempt_count() || !irqs_disabled()); preempt_count 281 include/linux/ftrace.h WARN_ON_ONCE(!preempt_count() || !irqs_disabled()); preempt_count 80 include/linux/preempt.h #define hardirq_count() (preempt_count() & HARDIRQ_MASK) preempt_count 81 include/linux/preempt.h #define softirq_count() (preempt_count() & SOFTIRQ_MASK) preempt_count 82 include/linux/preempt.h #define irq_count() (preempt_count() & (HARDIRQ_MASK | SOFTIRQ_MASK \ preempt_count 102 include/linux/preempt.h #define in_nmi() (preempt_count() & NMI_MASK) preempt_count 103 include/linux/preempt.h #define in_task() (!(preempt_count() & \ preempt_count 142 include/linux/preempt.h #define in_atomic() (preempt_count() != 0) preempt_count 148 include/linux/preempt.h #define in_atomic_preempt_off() (preempt_count() != PREEMPT_DISABLE_OFFSET) preempt_count 183 include/linux/preempt.h #define preemptible() (preempt_count() == 0 && !irqs_disabled()) preempt_count 63 include/linux/trace_events.h unsigned char preempt_count; preempt_count 932 init/main.c int count = preempt_count(); preempt_count 945 init/main.c if (preempt_count() != count) { preempt_count 755 kernel/exit.c preempt_count()); preempt_count 345 kernel/rcu/rcutorture.c if (preempt_count() & (SOFTIRQ_MASK | HARDIRQ_MASK)) preempt_count 357 kernel/rcu/rcutorture.c if (!preempt_count() && preempt_count 2354 kernel/rcu/tree.c if (!(preempt_count() & PREEMPT_MASK)) { preempt_count 612 kernel/rcu/tree_exp.h if (!(preempt_count() & (PREEMPT_MASK | SOFTIRQ_MASK)) || preempt_count 662 kernel/rcu/tree_exp.h if (!(preempt_count() & (PREEMPT_MASK | SOFTIRQ_MASK)) || preempt_count 599 kernel/rcu/tree_plugin.h !!(preempt_count() & (PREEMPT_MASK | SOFTIRQ_MASK)); preempt_count 690 kernel/rcu/tree_plugin.h (preempt_count() & (PREEMPT_MASK | SOFTIRQ_MASK))) { preempt_count 3197 kernel/sched/core.c if (WARN_ONCE(preempt_count() != 2*PREEMPT_DISABLE_OFFSET, preempt_count 3199 kernel/sched/core.c current->comm, current->pid, preempt_count())) preempt_count 3762 kernel/sched/core.c if (preempt_count() == val) { preempt_count 3777 kernel/sched/core.c if (DEBUG_LOCKS_WARN_ON((preempt_count() < 0))) preempt_count 3785 kernel/sched/core.c DEBUG_LOCKS_WARN_ON((preempt_count() & PREEMPT_MASK) >= preempt_count 3799 kernel/sched/core.c if (preempt_count() == val) preempt_count 3809 kernel/sched/core.c if (DEBUG_LOCKS_WARN_ON(val > preempt_count())) preempt_count 3815 kernel/sched/core.c !(preempt_count() & PREEMPT_MASK))) preempt_count 3851 kernel/sched/core.c prev->comm, prev->pid, preempt_count()); preempt_count 4320 kernel/sched/core.c BUG_ON(preempt_count() || !irqs_disabled()); preempt_count 6723 kernel/sched/core.c int nested = preempt_count() + rcu_preempt_depth(); preempt_count 6804 kernel/sched/core.c if (preempt_count() > preempt_offset) preempt_count 132 kernel/softirq.c if (preempt_count() == cnt) { preempt_count 146 kernel/softirq.c if (preempt_count() == cnt) preempt_count 287 kernel/softirq.c prev_count = preempt_count(); preempt_count 294 kernel/softirq.c if (unlikely(prev_count != preempt_count())) { preempt_count 297 kernel/softirq.c prev_count, preempt_count()); preempt_count 527 kernel/stop_machine.c WARN_ONCE(preempt_count(), preempt_count 1382 kernel/time/timer.c int count = preempt_count(); preempt_count 1409 kernel/time/timer.c if (count != preempt_count()) { preempt_count 1411 kernel/time/timer.c fn, count, preempt_count()); preempt_count 80 kernel/trace/blktrace.c pc = preempt_count(); preempt_count 253 kernel/trace/blktrace.c pc = preempt_count(); preempt_count 2695 kernel/trace/ring_buffer.c unsigned long pc = preempt_count(); preempt_count 845 kernel/trace/trace.c pc = preempt_count(); preempt_count 895 kernel/trace/trace.c pc = preempt_count(); preempt_count 2345 kernel/trace/trace.c entry->preempt_count = pc & 0xff; preempt_count 2931 kernel/trace/trace.c flags, skip, preempt_count(), NULL); preempt_count 3122 kernel/trace/trace.c pc = preempt_count(); preempt_count 3182 kernel/trace/trace.c pc = preempt_count(); preempt_count 6417 kernel/trace/trace.c irq_flags, preempt_count()); preempt_count 6497 kernel/trace/trace.c irq_flags, preempt_count()); preempt_count 62 kernel/trace/trace_branch.c pc = preempt_count(); preempt_count 419 kernel/trace/trace_event_perf.c int pc = preempt_count(); preempt_count 182 kernel/trace/trace_events.c __common_field(unsigned char, preempt_count); preempt_count 257 kernel/trace/trace_events.c fbuffer->pc = preempt_count(); preempt_count 3369 kernel/trace/trace_events.c pc = preempt_count(); preempt_count 138 kernel/trace/trace_functions.c pc = preempt_count(); preempt_count 199 kernel/trace/trace_functions.c pc = preempt_count(); preempt_count 408 kernel/trace/trace_functions.c pc = preempt_count(); preempt_count 177 kernel/trace/trace_functions_graph.c pc = preempt_count(); preempt_count 258 kernel/trace/trace_functions_graph.c pc = preempt_count(); preempt_count 113 kernel/trace/trace_hwlat.c pc = preempt_count(); preempt_count 122 kernel/trace/trace_irqsoff.c if (!irqs_disabled_flags(*flags) && !preempt_count()) preempt_count 150 kernel/trace/trace_irqsoff.c trace_function(tr, ip, parent_ip, flags, preempt_count()); preempt_count 198 kernel/trace/trace_irqsoff.c pc = preempt_count(); preempt_count 217 kernel/trace/trace_irqsoff.c pc = preempt_count(); preempt_count 333 kernel/trace/trace_irqsoff.c pc = preempt_count(); preempt_count 441 kernel/trace/trace_irqsoff.c int pc = preempt_count(); preempt_count 451 kernel/trace/trace_irqsoff.c int pc = preempt_count(); preempt_count 612 kernel/trace/trace_irqsoff.c unsigned int pc = preempt_count(); preempt_count 621 kernel/trace/trace_irqsoff.c unsigned int pc = preempt_count(); preempt_count 664 kernel/trace/trace_irqsoff.c int pc = preempt_count(); preempt_count 672 kernel/trace/trace_irqsoff.c int pc = preempt_count(); preempt_count 1192 kernel/trace/trace_kprobe.c pc = preempt_count(); preempt_count 1240 kernel/trace/trace_kprobe.c pc = preempt_count(); preempt_count 303 kernel/trace/trace_mmiotrace.c int pc = preempt_count(); preempt_count 333 kernel/trace/trace_mmiotrace.c int pc = preempt_count(); preempt_count 470 kernel/trace/trace_output.c if (entry->preempt_count) preempt_count 471 kernel/trace/trace_output.c trace_seq_printf(s, "%x", entry->preempt_count); preempt_count 629 kernel/trace/trace_output.c entry->preempt_count, iter->idx); preempt_count 78 kernel/trace/trace_sched_wakeup.c *pc = preempt_count(); preempt_count 458 kernel/trace/trace_sched_wakeup.c pc = preempt_count(); preempt_count 553 kernel/trace/trace_sched_wakeup.c pc = preempt_count(); preempt_count 884 kernel/trace/trace_selftest.c if (preempt_count()) { preempt_count 946 kernel/trace/trace_selftest.c if (preempt_count()) { preempt_count 341 kernel/trace/trace_syscalls.c pc = preempt_count(); preempt_count 387 kernel/trace/trace_syscalls.c pc = preempt_count(); preempt_count 2284 kernel/workqueue.c current->comm, preempt_count(), task_pid_nr(current), preempt_count 1139 lib/locking-selftest.c unsigned long saved_preempt_count = preempt_count(); preempt_count 16 lib/smp_processor_id.c if (likely(preempt_count())) preempt_count 44 lib/smp_processor_id.c what1, what2, preempt_count() - 1, current->comm, current->pid); preempt_count 51 security/apparmor/include/path.h #define AA_BUG_PREEMPT_ENABLED(X) AA_BUG(preempt_count() <= 0, X) preempt_count 16 tools/testing/radix-tree/linux.c int preempt_count; preempt_count 5 tools/testing/radix-tree/linux/preempt.h extern int preempt_count; preempt_count 7 tools/testing/radix-tree/linux/preempt.h #define preempt_disable() uatomic_inc(&preempt_count) preempt_count 8 tools/testing/radix-tree/linux/preempt.h #define preempt_enable() uatomic_dec(&preempt_count) preempt_count 244 tools/testing/radix-tree/main.c nr_allocated, preempt_count); preempt_count 248 tools/testing/radix-tree/main.c nr_allocated, preempt_count); preempt_count 252 tools/testing/radix-tree/main.c nr_allocated, preempt_count); preempt_count 256 tools/testing/radix-tree/main.c nr_allocated, preempt_count); preempt_count 260 tools/testing/radix-tree/main.c nr_allocated, preempt_count); preempt_count 264 tools/testing/radix-tree/main.c nr_allocated, preempt_count); preempt_count 269 tools/testing/radix-tree/main.c nr_allocated, preempt_count); preempt_count 273 tools/testing/radix-tree/main.c nr_allocated, preempt_count); preempt_count 281 tools/testing/radix-tree/main.c nr_allocated, preempt_count); preempt_count 324 tools/testing/radix-tree/main.c nr_allocated, preempt_count); preempt_count 40 tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/preempt.h return !preempt_count();