Searched refs:preempt_count (Results 1 - 117 of 117) sorted by relevance

/linux-4.1.27/include/asm-generic/
H A Dpreempt.h8 static __always_inline int preempt_count(void) preempt_count() function
10 return current_thread_info()->preempt_count; preempt_count()
15 return &current_thread_info()->preempt_count; preempt_count_ptr()
27 task_thread_info(p)->preempt_count = PREEMPT_DISABLED; \
31 task_thread_info(p)->preempt_count = PREEMPT_ENABLED; \
48 * The various preempt_count add/sub methods
76 return unlikely(preempt_count() == preempt_offset && should_resched()
/linux-4.1.27/arch/m32r/kernel/
H A Dasm-offsets.c10 OFFSET(TI_PRE_COUNT, thread_info, preempt_count); foo()
H A Dentry.S177 ld r9, @(TI_PRE_COUNT, r8) ; non-zero preempt_count ?
/linux-4.1.27/include/linux/
H A Dpreempt_mask.h53 #define hardirq_count() (preempt_count() & HARDIRQ_MASK)
54 #define softirq_count() (preempt_count() & SOFTIRQ_MASK)
55 #define irq_count() (preempt_count() & (HARDIRQ_MASK | SOFTIRQ_MASK \
72 #define in_nmi() (preempt_count() & NMI_MASK)
75 * The preempt_count offset after preempt_disable();
84 * The preempt_count offset after spin_lock()
89 * The preempt_count offset needed for things like:
110 #define in_atomic() ((preempt_count() & ~PREEMPT_ACTIVE) != 0)
117 ((preempt_count() & ~PREEMPT_ACTIVE) != PREEMPT_DISABLE_OFFSET)
120 # define preemptible() (preempt_count() == 0 && !irqs_disabled())
H A Dbit_spinlock.h93 return preempt_count(); bit_spin_is_locked()
H A Dpreempt.h6 * preempt_count (used for kernel preemption, interrupt count, etc.)
H A Dftrace_event.h67 unsigned char preempt_count; member in struct:trace_entry
H A Drcupdate.h512 return lockdep_opinion || preempt_count() != 0 || irqs_disabled(); rcu_read_lock_sched_held()
539 return preempt_count() != 0 || irqs_disabled(); rcu_read_lock_sched_held()
/linux-4.1.27/lib/
H A Dsmp_processor_id.c15 if (likely(preempt_count())) check_preemption_disabled()
43 what1, what2, preempt_count() - 1, current->comm, current->pid); check_preemption_disabled()
H A Dlocking-selftest.c972 unsigned long saved_preempt_count = preempt_count(); dotest()
/linux-4.1.27/arch/avr32/kernel/
H A Dasm-offsets.c17 OFFSET(TI_preempt_count, thread_info, preempt_count); foo()
/linux-4.1.27/arch/sh/kernel/
H A Dirq.c91 * Copy the softirq bits in preempt_count so that the handle_one_irq()
94 irqctx->tinfo.preempt_count = handle_one_irq()
95 (irqctx->tinfo.preempt_count & ~SOFTIRQ_MASK) | handle_one_irq()
96 (curctx->tinfo.preempt_count & SOFTIRQ_MASK); handle_one_irq()
128 irqctx->tinfo.preempt_count = HARDIRQ_OFFSET; irq_ctx_init()
136 irqctx->tinfo.preempt_count = 0; irq_ctx_init()
H A Dasm-offsets.c26 DEFINE(TI_PRE_COUNT, offsetof(struct thread_info, preempt_count)); main()
H A Dentry-common.S97 mov.l @(TI_PRE_COUNT,r8), r0 ! current_thread_info->preempt_count
/linux-4.1.27/arch/x86/include/asm/
H A Dpreempt.h20 static __always_inline int preempt_count(void) preempt_count() function
67 * The various preempt_count add/sub methods
82 * a decrement which hits zero means we have no preempt_count and should
H A Dthread_info.h176 * preempt_count needs to be 1 initially, until the scheduler is functional.
/linux-4.1.27/arch/score/include/asm/
H A Dthread_info.h36 int preempt_count; member in struct:thread_info
50 * preempt_count needs to be 1 initially, until the scheduler is functional.
56 .preempt_count = 1, \
/linux-4.1.27/arch/c6x/include/asm/
H A Dthread_info.h45 int preempt_count; /* 0 = preemptable, <0 = BUG */ member in struct:thread_info
52 * preempt_count needs to be 1 initially, until the scheduler is functional.
59 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/um/include/asm/
H A Dthread_info.h19 int preempt_count; /* 0 => preemptable, member in struct:thread_info
32 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/parisc/include/asm/
H A Dthread_info.h15 int preempt_count; /* 0=premptable, <0=BUG; will also serve as bh-counter */ member in struct:thread_info
24 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/m68k/include/asm/
H A Dthread_info.h30 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info
40 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/sparc/include/asm/
H A Dthread_info_32.h32 int preempt_count; /* 0 => preemptable, member in struct:thread_info
62 .preempt_count = INIT_PREEMPT_COUNT, \
90 #define TI_PREEMPT 0x10 /* preempt_count */
H A Dthread_info_64.h46 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info
119 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/nios2/include/asm/
H A Dthread_info.h44 int preempt_count; /* 0 => preemptable,<0 => BUG */ member in struct:thread_info
55 * preempt_count needs to be 1 initially, until the scheduler is functional.
62 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/openrisc/include/asm/
H A Dthread_info.h53 __s32 preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info
69 * preempt_count needs to be 1 initially, until the scheduler is functional.
77 .preempt_count = 1, \
/linux-4.1.27/arch/arc/include/asm/
H A Dthread_info.h43 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info
53 * preempt_count needs to be 1 initially, until the scheduler is functional.
60 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/cris/include/arch-v10/arch/
H A Doffset.h22 #define TI_preempt_count 16 /* offsetof(struct thread_info, preempt_count) */
/linux-4.1.27/arch/cris/include/arch-v32/arch/
H A Doffset.h24 #define TI_preempt_count 16 /* offsetof(struct thread_info, preempt_count) */
/linux-4.1.27/arch/cris/kernel/
H A Dasm-offsets.c40 ENTRY(preempt_count); main()
/linux-4.1.27/arch/s390/include/asm/
H A Dthread_info.h35 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info
50 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/avr32/include/asm/
H A Dthread_info.h25 __s32 preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info
39 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/blackfin/include/asm/
H A Dthread_info.h42 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info
57 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/cris/include/asm/
H A Dthread_info.h33 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info
54 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/metag/kernel/
H A Dirq.c84 * Copy the softirq bits in preempt_count so that the do_IRQ()
87 irqctx->tinfo.preempt_count = do_IRQ()
88 (irqctx->tinfo.preempt_count & ~SOFTIRQ_MASK) | do_IRQ()
89 (curctx->tinfo.preempt_count & SOFTIRQ_MASK); do_IRQ()
136 irqctx->tinfo.preempt_count = HARDIRQ_OFFSET; irq_ctx_init()
144 irqctx->tinfo.preempt_count = 0; irq_ctx_init()
/linux-4.1.27/arch/xtensa/include/asm/
H A Dthread_info.h50 __s32 preempt_count; /* 0 => preemptable,< 0 => BUG*/ member in struct:thread_info
76 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/unicore32/include/asm/
H A Dthread_info.h69 int preempt_count; /* 0 => preemptable */ member in struct:thread_info
86 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/metag/include/asm/
H A Dthread_info.h34 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info
72 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/arm64/include/asm/
H A Dthread_info.h49 int preempt_count; /* 0 => preemptable, <0 => bug */ member in struct:thread_info
57 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/frv/include/asm/
H A Dthread_info.h37 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info
63 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/hexagon/include/asm/
H A Dthread_info.h52 int preempt_count; /* 0=>preemptible,<0=>BUG */ member in struct:thread_info
81 .preempt_count = 1, \
/linux-4.1.27/arch/m32r/include/asm/
H A Dthread_info.h30 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info
54 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/alpha/include/asm/
H A Dthread_info.h23 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info
38 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/arc/kernel/
H A Dasm-offsets.c34 offsetof(struct thread_info, preempt_count)); main()
H A Dentry.S207 ; bump thread_info->preempt_count (Disable preemption)
690 ; decrement thread_info->preempt_count (re-enable preemption)
/linux-4.1.27/kernel/trace/
H A Dtrace_irqsoff.c43 return ((trace_type & TRACER_PREEMPT_OFF) && preempt_count()); preempt_trace()
148 trace_function(tr, ip, parent_ip, flags, preempt_count()); irqsoff_tracer_call()
188 pc = preempt_count(); irqsoff_graph_entry()
205 pc = preempt_count(); irqsoff_graph_return()
326 pc = preempt_count(); check_critical_timing()
391 __trace_function(tr, ip, parent_ip, flags, preempt_count()); start_critical_timing()
425 __trace_function(tr, ip, parent_ip, flags, preempt_count()); stop_critical_timing()
H A Dtrace_mmiotrace.c305 int pc = preempt_count(); __trace_mmiotrace_rw()
335 int pc = preempt_count(); __trace_mmiotrace_map()
H A Dtrace_output.c470 if (entry->preempt_count) trace_print_lat_fmt()
471 trace_seq_printf(s, "%x", entry->preempt_count); trace_print_lat_fmt()
614 entry->preempt_count, iter->idx); trace_print_lat_context()
H A Dtrace_sched_wakeup.c87 *pc = preempt_count(); func_prolog_preempt_disable()
450 pc = preempt_count(); probe_wakeup_sched_switch()
544 pc = preempt_count(); probe_wakeup()
H A Dtrace_functions.c137 pc = preempt_count(); function_trace_call()
180 pc = preempt_count(); function_stack_trace_call()
H A Dtrace_syscalls.c325 pc = preempt_count(); ftrace_syscall_enter()
370 pc = preempt_count(); ftrace_syscall_exit()
H A Dtrace_branch.c61 pc = preempt_count(); probe_likely_condition()
H A Dtrace_event_perf.c277 pc = preempt_count(); perf_trace_buf_prepare()
H A Dtrace_functions_graph.c343 pc = preempt_count(); trace_graph_entry()
428 pc = preempt_count(); trace_graph_return()
H A Dtrace_kprobe.c935 pc = preempt_count(); __kprobe_trace_func()
983 pc = preempt_count(); __kretprobe_trace_func()
H A Dtrace_selftest.c875 if (preempt_count()) { trace_selftest_startup_preemptoff()
937 if (preempt_count()) { trace_selftest_startup_preemptirqsoff()
H A Dtrace_events.c163 __common_field(unsigned char, preempt_count); trace_define_common_fields()
200 fbuffer->pc = preempt_count(); ftrace_event_buffer_reserve()
2887 pc = preempt_count(); function_test_events_call()
H A Dblktrace.c80 pc = preempt_count(); trace_note()
234 pc = preempt_count(); __blk_add_trace()
H A Dtrace.c524 pc = preempt_count(); __trace_puts()
574 pc = preempt_count(); __trace_bputs()
1632 entry->preempt_count = pc & 0xff; tracing_generic_entry_update()
1918 flags, skip, preempt_count(), NULL); trace_dump_stack()
1978 ftrace_trace_userstack(tr, flags, preempt_count()); __trace_userstack()
2147 pc = preempt_count(); trace_vbprintk()
2203 pc = preempt_count(); __trace_array_vprintk()
5129 irq_flags, preempt_count()); tracing_mark_write()
/linux-4.1.27/kernel/
H A Dsoftirq.c81 * preempt_count and SOFTIRQ_OFFSET usage:
82 * - preempt_count is changed by SOFTIRQ_OFFSET on entering or leaving
84 * - preempt_count is changed by SOFTIRQ_DISABLE_OFFSET (= 2 * SOFTIRQ_OFFSET)
106 * We must manually increment preempt_count here and manually __local_bh_disable_ip()
117 if (preempt_count() == cnt) { __local_bh_disable_ip()
268 prev_count = preempt_count(); __do_softirq()
275 if (unlikely(prev_count != preempt_count())) { __do_softirq()
276 pr_err("huh, entered softirq %u %s %p with preempt_count %08x, exited with %08x?\n", __do_softirq()
278 prev_count, preempt_count()); __do_softirq()
H A Dstop_machine.c479 WARN_ONCE(preempt_count(), cpu_stopper_thread()
H A Dexit.c710 pr_info("note: %s[%d] exited with preempt_count %d\n", do_exit()
712 preempt_count()); do_exit()
H A Dworkqueue.c2088 current->comm, preempt_count(), task_pid_nr(current),
/linux-4.1.27/arch/tile/include/asm/
H A Dthread_info.h33 int preempt_count; /* 0 => preemptable, member in struct:thread_info
55 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/microblaze/include/asm/
H A Dthread_info.h71 __s32 preempt_count; /* 0 => preemptable,< 0 => BUG*/ member in struct:thread_info
85 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/sh/include/asm/
H A Dthread_info.h33 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info
61 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/mips/include/asm/
H A Dthread_info.h29 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info
48 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/mn10300/include/asm/
H A Dthread_info.h46 __s32 preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info
78 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/powerpc/include/asm/
H A Dthread_info.h43 int preempt_count; /* 0 => preemptable, member in struct:thread_info
58 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/ia64/include/asm/
H A Dthread_info.h28 int preempt_count; /* 0=premptable, <0=BUG; will also serve as bh-counter */ member in struct:thread_info
45 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/arm/include/asm/
H A Dthread_info.h52 int preempt_count; /* 0 => preemptable, <0 => bug */ member in struct:thread_info
75 .preempt_count = INIT_PREEMPT_COUNT, \
/linux-4.1.27/arch/tile/kernel/
H A Dasm-offsets.c62 offsetof(struct thread_info, preempt_count)); foo()
/linux-4.1.27/arch/nios2/kernel/
H A Dasm-offsets.c83 OFFSET(TI_PREEMPT_COUNT, thread_info, preempt_count); main()
/linux-4.1.27/arch/openrisc/kernel/
H A Dasm-offsets.c55 DEFINE(TI_PREEMPT, offsetof(struct thread_info, preempt_count)); main()
/linux-4.1.27/arch/powerpc/kernel/
H A Dirq.c531 /* Copy the preempt_count so that the [soft]irq checks work. */ do_IRQ()
532 irqtp->preempt_count = curtp->preempt_count; do_IRQ()
581 tp->preempt_count = 0; for_each_possible_cpu()
587 tp->preempt_count = 0; for_each_possible_cpu()
592 tp->preempt_count = HARDIRQ_OFFSET; for_each_possible_cpu()
H A Dmachine_kexec_64.c337 * We setup preempt_count to avoid using VMX in memcpy. default_machine_kexec()
342 kexec_stack.thread_info.preempt_count = HARDIRQ_OFFSET; default_machine_kexec()
H A Dasm-offsets.c162 DEFINE(TI_PREEMPT, offsetof(struct thread_info, preempt_count)); main()
H A Dentry_32.S811 /* check current_thread_info->preempt_count */
H A Dentry_64.S703 /* Check that preempt_count() == 0 and interrupts are enabled */
/linux-4.1.27/arch/powerpc/platforms/cell/
H A Dsmp.c86 task_thread_info(paca[lcpu].__current)->preempt_count = 0; smp_startup_cpu()
/linux-4.1.27/arch/mn10300/kernel/
H A Dasm-offsets.c28 OFFSET(TI_preempt_count, thread_info, preempt_count); foo()
H A Dentry.S179 mov (TI_preempt_count,a2),d0 # non-zero preempt_count ?
/linux-4.1.27/arch/c6x/kernel/
H A Dasm-offsets.c103 OFFSET(THREAD_INFO_PREEMPT_COUNT, thread_info, preempt_count); foo()
/linux-4.1.27/arch/frv/kernel/
H A Dasm-offsets.c40 OFFSET(TI_PREEMPT_COUNT, thread_info, preempt_count); foo()
/linux-4.1.27/arch/m68k/kernel/
H A Dasm-offsets.c41 DEFINE(TINFO_PREEMPT, offsetof(struct thread_info, preempt_count)); main()
/linux-4.1.27/arch/xtensa/kernel/
H A Dasm-offsets.c85 OFFSET(TI_PRE_COUNT, thread_info, preempt_count); main()
H A Dentry.S465 /* Check current_thread_info->preempt_count */
/linux-4.1.27/arch/unicore32/kernel/
H A Dasm-offsets.c42 DEFINE(TI_PREEMPT, offsetof(struct thread_info, preempt_count)); main()
/linux-4.1.27/arch/microblaze/kernel/
H A Dasm-offsets.c94 DEFINE(TI_PREEMPT_COUNT, offsetof(struct thread_info, preempt_count)); main()
H A Dentry.S723 /* MS: get preempt_count from thread info */
/linux-4.1.27/arch/powerpc/platforms/pseries/
H A Dsmp.c118 task_thread_info(paca[lcpu].__current)->preempt_count = 0; smp_startup_cpu()
/linux-4.1.27/arch/s390/kernel/
H A Dasm-offsets.c40 DEFINE(__TI_precount, offsetof(struct thread_info, preempt_count)); main()
/linux-4.1.27/arch/score/kernel/
H A Dasm-offsets.c106 OFFSET(TI_PRE_COUNT, thread_info, preempt_count); output_thread_info_defines()
/linux-4.1.27/arch/arm64/kernel/
H A Dasm-offsets.c37 DEFINE(TI_PREEMPT, offsetof(struct thread_info, preempt_count)); main()
/linux-4.1.27/arch/blackfin/kernel/
H A Dasm-offsets.c49 OFFSET(TI_PREEMPT, thread_info, preempt_count); main()
/linux-4.1.27/arch/arm/kernel/
H A Dasm-offsets.c67 DEFINE(TI_PREEMPT, offsetof(struct thread_info, preempt_count)); main()
/linux-4.1.27/arch/ia64/kernel/
H A Dentry.S705 * With CONFIG_PREEMPT, we also check for extra work when the preempt_count
716 (pKStk) ld4 r21=[r20] // r21 <- preempt_count
719 cmp.eq p6,p0=r21,r0 // p6 <- pUStk || (preempt_count == 0)
849 * With CONFIG_PREEMPT, we also check for extra work when the preempt_count
860 (pKStk) ld4 r21=[r20] // r21 <- preempt_count
863 cmp.eq p6,p0=r21,r0 // p6 <- pUStk || (preempt_count == 0)
H A Dasm-offsets.c40 DEFINE(TI_PRE_COUNT, offsetof(struct thread_info, preempt_count)); foo()
H A Dmca.c1820 ti->preempt_count = 1; format_mca_init_stack()
/linux-4.1.27/init/
H A Dmain.c779 int count = preempt_count(); do_one_initcall()
793 if (preempt_count() != count) { do_one_initcall()
/linux-4.1.27/kernel/sched/
H A Dcputime.c44 * Called before incrementing preempt_count on {soft,}irq_enter
45 * and before decrementing preempt_count on {soft,}irq_exit.
H A Dcore.c1546 * Fold TIF_NEED_RESCHED into the preempt_count; anybody setting scheduler_ipi()
2565 if (DEBUG_LOCKS_WARN_ON((preempt_count() < 0))) preempt_count_add()
2573 DEBUG_LOCKS_WARN_ON((preempt_count() & PREEMPT_MASK) >= preempt_count_add()
2576 if (preempt_count() == val) { preempt_count_add()
2593 if (DEBUG_LOCKS_WARN_ON(val > preempt_count())) preempt_count_sub()
2599 !(preempt_count() & PREEMPT_MASK))) preempt_count_sub()
2603 if (preempt_count() == val) preempt_count_sub()
2621 prev->comm, prev->pid, preempt_count()); __schedule_bug()
2769 if (prev->state && !(preempt_count() & PREEMPT_ACTIVE)) { __schedule()
2860 * Returns with preemption disabled. Note: preempt_count must be 1
2893 * If there is a non-zero preempt_count or interrupts are disabled, preempt_schedule()
2957 BUG_ON(preempt_count() || !irqs_disabled()); preempt_schedule_irq()
7284 int nested = (preempt_count() & ~PREEMPT_ACTIVE) + rcu_preempt_depth(); preempt_count_equals()
/linux-4.1.27/arch/x86/kernel/
H A Dprocess_32.c282 * preempt_count of all tasks was equal here and this would not be __switch_to()
H A Dkvm.c143 n.halted = is_idle_task(current) || preempt_count() > 1; kvm_async_pf_task_wait()
H A Dprocess_64.c404 * preempt_count of all tasks was equal here and this would not be __switch_to()
H A Dentry_64.S696 * or not. While this is essentially redundant with preempt_count it is
/linux-4.1.27/include/trace/events/
H A Dsched.h106 if (preempt_count() & PREEMPT_ACTIVE) __trace_sched_switch_state()
/linux-4.1.27/arch/sparc/kernel/
H A Dtraps_32.c438 TI_PREEMPT != offsetof(struct thread_info, preempt_count) || trap_init()
H A Dtraps_64.c2701 preempt_count) || trap_init()
/linux-4.1.27/arch/mips/kernel/
H A Dasm-offsets.c98 OFFSET(TI_PRE_COUNT, thread_info, preempt_count); output_thread_info_defines()
/linux-4.1.27/arch/parisc/kernel/
H A Dasm-offsets.c248 DEFINE(TI_PRE_COUNT, offsetof(struct thread_info, preempt_count)); main()
H A Dentry.S970 * and preempt_count is 0. otherwise, we continue on
978 /* current_thread_info()->preempt_count */
981 cmpib,COND(<>) 0, %r19, intr_restore /* if preempt_count > 0 */
/linux-4.1.27/kernel/time/
H A Dtimer.c1131 int count = preempt_count(); call_timer_fn()
1158 if (count != preempt_count()) { call_timer_fn()
1160 fn, count, preempt_count()); call_timer_fn()
H A Dclocksource.c30 #include <linux/sched.h> /* for spin_unlock_irq() using preempt_count() m68k */
/linux-4.1.27/include/trace/
H A Dftrace.h540 * pc = preempt_count();
/linux-4.1.27/arch/blackfin/mach-common/
H A Dentry.S915 /* check preempt_count */
/linux-4.1.27/kernel/rcu/
H A Drcutorture.c289 if (!preempt_count() && rcu_read_delay()
/linux-4.1.27/mm/
H A Dmemory.c3732 if (preempt_count()) print_vma_addr()

Completed in 1142 milliseconds