/linux-4.1.27/include/asm-generic/ |
H A D | preempt.h | 8 static __always_inline int preempt_count(void) preempt_count() function 10 return current_thread_info()->preempt_count; preempt_count() 15 return ¤t_thread_info()->preempt_count; preempt_count_ptr() 27 task_thread_info(p)->preempt_count = PREEMPT_DISABLED; \ 31 task_thread_info(p)->preempt_count = PREEMPT_ENABLED; \ 48 * The various preempt_count add/sub methods 76 return unlikely(preempt_count() == preempt_offset && should_resched()
|
/linux-4.1.27/arch/m32r/kernel/ |
H A D | asm-offsets.c | 10 OFFSET(TI_PRE_COUNT, thread_info, preempt_count); foo()
|
H A D | entry.S | 177 ld r9, @(TI_PRE_COUNT, r8) ; non-zero preempt_count ?
|
/linux-4.1.27/include/linux/ |
H A D | preempt_mask.h | 53 #define hardirq_count() (preempt_count() & HARDIRQ_MASK) 54 #define softirq_count() (preempt_count() & SOFTIRQ_MASK) 55 #define irq_count() (preempt_count() & (HARDIRQ_MASK | SOFTIRQ_MASK \ 72 #define in_nmi() (preempt_count() & NMI_MASK) 75 * The preempt_count offset after preempt_disable(); 84 * The preempt_count offset after spin_lock() 89 * The preempt_count offset needed for things like: 110 #define in_atomic() ((preempt_count() & ~PREEMPT_ACTIVE) != 0) 117 ((preempt_count() & ~PREEMPT_ACTIVE) != PREEMPT_DISABLE_OFFSET) 120 # define preemptible() (preempt_count() == 0 && !irqs_disabled())
|
H A D | bit_spinlock.h | 93 return preempt_count(); bit_spin_is_locked()
|
H A D | preempt.h | 6 * preempt_count (used for kernel preemption, interrupt count, etc.)
|
H A D | ftrace_event.h | 67 unsigned char preempt_count; member in struct:trace_entry
|
H A D | rcupdate.h | 512 return lockdep_opinion || preempt_count() != 0 || irqs_disabled(); rcu_read_lock_sched_held() 539 return preempt_count() != 0 || irqs_disabled(); rcu_read_lock_sched_held()
|
/linux-4.1.27/lib/ |
H A D | smp_processor_id.c | 15 if (likely(preempt_count())) check_preemption_disabled() 43 what1, what2, preempt_count() - 1, current->comm, current->pid); check_preemption_disabled()
|
H A D | locking-selftest.c | 972 unsigned long saved_preempt_count = preempt_count(); dotest()
|
/linux-4.1.27/arch/avr32/kernel/ |
H A D | asm-offsets.c | 17 OFFSET(TI_preempt_count, thread_info, preempt_count); foo()
|
/linux-4.1.27/arch/sh/kernel/ |
H A D | irq.c | 91 * Copy the softirq bits in preempt_count so that the handle_one_irq() 94 irqctx->tinfo.preempt_count = handle_one_irq() 95 (irqctx->tinfo.preempt_count & ~SOFTIRQ_MASK) | handle_one_irq() 96 (curctx->tinfo.preempt_count & SOFTIRQ_MASK); handle_one_irq() 128 irqctx->tinfo.preempt_count = HARDIRQ_OFFSET; irq_ctx_init() 136 irqctx->tinfo.preempt_count = 0; irq_ctx_init()
|
H A D | asm-offsets.c | 26 DEFINE(TI_PRE_COUNT, offsetof(struct thread_info, preempt_count)); main()
|
H A D | entry-common.S | 97 mov.l @(TI_PRE_COUNT,r8), r0 ! current_thread_info->preempt_count
|
/linux-4.1.27/arch/x86/include/asm/ |
H A D | preempt.h | 20 static __always_inline int preempt_count(void) preempt_count() function 67 * The various preempt_count add/sub methods 82 * a decrement which hits zero means we have no preempt_count and should
|
H A D | thread_info.h | 176 * preempt_count needs to be 1 initially, until the scheduler is functional.
|
/linux-4.1.27/arch/score/include/asm/ |
H A D | thread_info.h | 36 int preempt_count; member in struct:thread_info 50 * preempt_count needs to be 1 initially, until the scheduler is functional. 56 .preempt_count = 1, \
|
/linux-4.1.27/arch/c6x/include/asm/ |
H A D | thread_info.h | 45 int preempt_count; /* 0 = preemptable, <0 = BUG */ member in struct:thread_info 52 * preempt_count needs to be 1 initially, until the scheduler is functional. 59 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/um/include/asm/ |
H A D | thread_info.h | 19 int preempt_count; /* 0 => preemptable, member in struct:thread_info 32 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/parisc/include/asm/ |
H A D | thread_info.h | 15 int preempt_count; /* 0=premptable, <0=BUG; will also serve as bh-counter */ member in struct:thread_info 24 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/m68k/include/asm/ |
H A D | thread_info.h | 30 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 40 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/sparc/include/asm/ |
H A D | thread_info_32.h | 32 int preempt_count; /* 0 => preemptable, member in struct:thread_info 62 .preempt_count = INIT_PREEMPT_COUNT, \ 90 #define TI_PREEMPT 0x10 /* preempt_count */
|
H A D | thread_info_64.h | 46 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 119 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/nios2/include/asm/ |
H A D | thread_info.h | 44 int preempt_count; /* 0 => preemptable,<0 => BUG */ member in struct:thread_info 55 * preempt_count needs to be 1 initially, until the scheduler is functional. 62 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/openrisc/include/asm/ |
H A D | thread_info.h | 53 __s32 preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 69 * preempt_count needs to be 1 initially, until the scheduler is functional. 77 .preempt_count = 1, \
|
/linux-4.1.27/arch/arc/include/asm/ |
H A D | thread_info.h | 43 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 53 * preempt_count needs to be 1 initially, until the scheduler is functional. 60 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/cris/include/arch-v10/arch/ |
H A D | offset.h | 22 #define TI_preempt_count 16 /* offsetof(struct thread_info, preempt_count) */
|
/linux-4.1.27/arch/cris/include/arch-v32/arch/ |
H A D | offset.h | 24 #define TI_preempt_count 16 /* offsetof(struct thread_info, preempt_count) */
|
/linux-4.1.27/arch/cris/kernel/ |
H A D | asm-offsets.c | 40 ENTRY(preempt_count); main()
|
/linux-4.1.27/arch/s390/include/asm/ |
H A D | thread_info.h | 35 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 50 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/avr32/include/asm/ |
H A D | thread_info.h | 25 __s32 preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 39 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/blackfin/include/asm/ |
H A D | thread_info.h | 42 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 57 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/cris/include/asm/ |
H A D | thread_info.h | 33 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 54 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/metag/kernel/ |
H A D | irq.c | 84 * Copy the softirq bits in preempt_count so that the do_IRQ() 87 irqctx->tinfo.preempt_count = do_IRQ() 88 (irqctx->tinfo.preempt_count & ~SOFTIRQ_MASK) | do_IRQ() 89 (curctx->tinfo.preempt_count & SOFTIRQ_MASK); do_IRQ() 136 irqctx->tinfo.preempt_count = HARDIRQ_OFFSET; irq_ctx_init() 144 irqctx->tinfo.preempt_count = 0; irq_ctx_init()
|
/linux-4.1.27/arch/xtensa/include/asm/ |
H A D | thread_info.h | 50 __s32 preempt_count; /* 0 => preemptable,< 0 => BUG*/ member in struct:thread_info 76 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/unicore32/include/asm/ |
H A D | thread_info.h | 69 int preempt_count; /* 0 => preemptable */ member in struct:thread_info 86 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/metag/include/asm/ |
H A D | thread_info.h | 34 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 72 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/arm64/include/asm/ |
H A D | thread_info.h | 49 int preempt_count; /* 0 => preemptable, <0 => bug */ member in struct:thread_info 57 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/frv/include/asm/ |
H A D | thread_info.h | 37 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 63 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/hexagon/include/asm/ |
H A D | thread_info.h | 52 int preempt_count; /* 0=>preemptible,<0=>BUG */ member in struct:thread_info 81 .preempt_count = 1, \
|
/linux-4.1.27/arch/m32r/include/asm/ |
H A D | thread_info.h | 30 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 54 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/alpha/include/asm/ |
H A D | thread_info.h | 23 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 38 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/arc/kernel/ |
H A D | asm-offsets.c | 34 offsetof(struct thread_info, preempt_count)); main()
|
H A D | entry.S | 207 ; bump thread_info->preempt_count (Disable preemption) 690 ; decrement thread_info->preempt_count (re-enable preemption)
|
/linux-4.1.27/kernel/trace/ |
H A D | trace_irqsoff.c | 43 return ((trace_type & TRACER_PREEMPT_OFF) && preempt_count()); preempt_trace() 148 trace_function(tr, ip, parent_ip, flags, preempt_count()); irqsoff_tracer_call() 188 pc = preempt_count(); irqsoff_graph_entry() 205 pc = preempt_count(); irqsoff_graph_return() 326 pc = preempt_count(); check_critical_timing() 391 __trace_function(tr, ip, parent_ip, flags, preempt_count()); start_critical_timing() 425 __trace_function(tr, ip, parent_ip, flags, preempt_count()); stop_critical_timing()
|
H A D | trace_mmiotrace.c | 305 int pc = preempt_count(); __trace_mmiotrace_rw() 335 int pc = preempt_count(); __trace_mmiotrace_map()
|
H A D | trace_output.c | 470 if (entry->preempt_count) trace_print_lat_fmt() 471 trace_seq_printf(s, "%x", entry->preempt_count); trace_print_lat_fmt() 614 entry->preempt_count, iter->idx); trace_print_lat_context()
|
H A D | trace_sched_wakeup.c | 87 *pc = preempt_count(); func_prolog_preempt_disable() 450 pc = preempt_count(); probe_wakeup_sched_switch() 544 pc = preempt_count(); probe_wakeup()
|
H A D | trace_functions.c | 137 pc = preempt_count(); function_trace_call() 180 pc = preempt_count(); function_stack_trace_call()
|
H A D | trace_syscalls.c | 325 pc = preempt_count(); ftrace_syscall_enter() 370 pc = preempt_count(); ftrace_syscall_exit()
|
H A D | trace_branch.c | 61 pc = preempt_count(); probe_likely_condition()
|
H A D | trace_event_perf.c | 277 pc = preempt_count(); perf_trace_buf_prepare()
|
H A D | trace_functions_graph.c | 343 pc = preempt_count(); trace_graph_entry() 428 pc = preempt_count(); trace_graph_return()
|
H A D | trace_kprobe.c | 935 pc = preempt_count(); __kprobe_trace_func() 983 pc = preempt_count(); __kretprobe_trace_func()
|
H A D | trace_selftest.c | 875 if (preempt_count()) { trace_selftest_startup_preemptoff() 937 if (preempt_count()) { trace_selftest_startup_preemptirqsoff()
|
H A D | trace_events.c | 163 __common_field(unsigned char, preempt_count); trace_define_common_fields() 200 fbuffer->pc = preempt_count(); ftrace_event_buffer_reserve() 2887 pc = preempt_count(); function_test_events_call()
|
H A D | blktrace.c | 80 pc = preempt_count(); trace_note() 234 pc = preempt_count(); __blk_add_trace()
|
H A D | trace.c | 524 pc = preempt_count(); __trace_puts() 574 pc = preempt_count(); __trace_bputs() 1632 entry->preempt_count = pc & 0xff; tracing_generic_entry_update() 1918 flags, skip, preempt_count(), NULL); trace_dump_stack() 1978 ftrace_trace_userstack(tr, flags, preempt_count()); __trace_userstack() 2147 pc = preempt_count(); trace_vbprintk() 2203 pc = preempt_count(); __trace_array_vprintk() 5129 irq_flags, preempt_count()); tracing_mark_write()
|
/linux-4.1.27/kernel/ |
H A D | softirq.c | 81 * preempt_count and SOFTIRQ_OFFSET usage: 82 * - preempt_count is changed by SOFTIRQ_OFFSET on entering or leaving 84 * - preempt_count is changed by SOFTIRQ_DISABLE_OFFSET (= 2 * SOFTIRQ_OFFSET) 106 * We must manually increment preempt_count here and manually __local_bh_disable_ip() 117 if (preempt_count() == cnt) { __local_bh_disable_ip() 268 prev_count = preempt_count(); __do_softirq() 275 if (unlikely(prev_count != preempt_count())) { __do_softirq() 276 pr_err("huh, entered softirq %u %s %p with preempt_count %08x, exited with %08x?\n", __do_softirq() 278 prev_count, preempt_count()); __do_softirq()
|
H A D | stop_machine.c | 479 WARN_ONCE(preempt_count(), cpu_stopper_thread()
|
H A D | exit.c | 710 pr_info("note: %s[%d] exited with preempt_count %d\n", do_exit() 712 preempt_count()); do_exit()
|
H A D | workqueue.c | 2088 current->comm, preempt_count(), task_pid_nr(current),
|
/linux-4.1.27/arch/tile/include/asm/ |
H A D | thread_info.h | 33 int preempt_count; /* 0 => preemptable, member in struct:thread_info 55 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/microblaze/include/asm/ |
H A D | thread_info.h | 71 __s32 preempt_count; /* 0 => preemptable,< 0 => BUG*/ member in struct:thread_info 85 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/sh/include/asm/ |
H A D | thread_info.h | 33 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 61 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/mips/include/asm/ |
H A D | thread_info.h | 29 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 48 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/mn10300/include/asm/ |
H A D | thread_info.h | 46 __s32 preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 78 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/powerpc/include/asm/ |
H A D | thread_info.h | 43 int preempt_count; /* 0 => preemptable, member in struct:thread_info 58 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/ia64/include/asm/ |
H A D | thread_info.h | 28 int preempt_count; /* 0=premptable, <0=BUG; will also serve as bh-counter */ member in struct:thread_info 45 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/arm/include/asm/ |
H A D | thread_info.h | 52 int preempt_count; /* 0 => preemptable, <0 => bug */ member in struct:thread_info 75 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.1.27/arch/tile/kernel/ |
H A D | asm-offsets.c | 62 offsetof(struct thread_info, preempt_count)); foo()
|
/linux-4.1.27/arch/nios2/kernel/ |
H A D | asm-offsets.c | 83 OFFSET(TI_PREEMPT_COUNT, thread_info, preempt_count); main()
|
/linux-4.1.27/arch/openrisc/kernel/ |
H A D | asm-offsets.c | 55 DEFINE(TI_PREEMPT, offsetof(struct thread_info, preempt_count)); main()
|
/linux-4.1.27/arch/powerpc/kernel/ |
H A D | irq.c | 531 /* Copy the preempt_count so that the [soft]irq checks work. */ do_IRQ() 532 irqtp->preempt_count = curtp->preempt_count; do_IRQ() 581 tp->preempt_count = 0; for_each_possible_cpu() 587 tp->preempt_count = 0; for_each_possible_cpu() 592 tp->preempt_count = HARDIRQ_OFFSET; for_each_possible_cpu()
|
H A D | machine_kexec_64.c | 337 * We setup preempt_count to avoid using VMX in memcpy. default_machine_kexec() 342 kexec_stack.thread_info.preempt_count = HARDIRQ_OFFSET; default_machine_kexec()
|
H A D | asm-offsets.c | 162 DEFINE(TI_PREEMPT, offsetof(struct thread_info, preempt_count)); main()
|
H A D | entry_32.S | 811 /* check current_thread_info->preempt_count */
|
H A D | entry_64.S | 703 /* Check that preempt_count() == 0 and interrupts are enabled */
|
/linux-4.1.27/arch/powerpc/platforms/cell/ |
H A D | smp.c | 86 task_thread_info(paca[lcpu].__current)->preempt_count = 0; smp_startup_cpu()
|
/linux-4.1.27/arch/mn10300/kernel/ |
H A D | asm-offsets.c | 28 OFFSET(TI_preempt_count, thread_info, preempt_count); foo()
|
H A D | entry.S | 179 mov (TI_preempt_count,a2),d0 # non-zero preempt_count ?
|
/linux-4.1.27/arch/c6x/kernel/ |
H A D | asm-offsets.c | 103 OFFSET(THREAD_INFO_PREEMPT_COUNT, thread_info, preempt_count); foo()
|
/linux-4.1.27/arch/frv/kernel/ |
H A D | asm-offsets.c | 40 OFFSET(TI_PREEMPT_COUNT, thread_info, preempt_count); foo()
|
/linux-4.1.27/arch/m68k/kernel/ |
H A D | asm-offsets.c | 41 DEFINE(TINFO_PREEMPT, offsetof(struct thread_info, preempt_count)); main()
|
/linux-4.1.27/arch/xtensa/kernel/ |
H A D | asm-offsets.c | 85 OFFSET(TI_PRE_COUNT, thread_info, preempt_count); main()
|
H A D | entry.S | 465 /* Check current_thread_info->preempt_count */
|
/linux-4.1.27/arch/unicore32/kernel/ |
H A D | asm-offsets.c | 42 DEFINE(TI_PREEMPT, offsetof(struct thread_info, preempt_count)); main()
|
/linux-4.1.27/arch/microblaze/kernel/ |
H A D | asm-offsets.c | 94 DEFINE(TI_PREEMPT_COUNT, offsetof(struct thread_info, preempt_count)); main()
|
H A D | entry.S | 723 /* MS: get preempt_count from thread info */
|
/linux-4.1.27/arch/powerpc/platforms/pseries/ |
H A D | smp.c | 118 task_thread_info(paca[lcpu].__current)->preempt_count = 0; smp_startup_cpu()
|
/linux-4.1.27/arch/s390/kernel/ |
H A D | asm-offsets.c | 40 DEFINE(__TI_precount, offsetof(struct thread_info, preempt_count)); main()
|
/linux-4.1.27/arch/score/kernel/ |
H A D | asm-offsets.c | 106 OFFSET(TI_PRE_COUNT, thread_info, preempt_count); output_thread_info_defines()
|
/linux-4.1.27/arch/arm64/kernel/ |
H A D | asm-offsets.c | 37 DEFINE(TI_PREEMPT, offsetof(struct thread_info, preempt_count)); main()
|
/linux-4.1.27/arch/blackfin/kernel/ |
H A D | asm-offsets.c | 49 OFFSET(TI_PREEMPT, thread_info, preempt_count); main()
|
/linux-4.1.27/arch/arm/kernel/ |
H A D | asm-offsets.c | 67 DEFINE(TI_PREEMPT, offsetof(struct thread_info, preempt_count)); main()
|
/linux-4.1.27/arch/ia64/kernel/ |
H A D | entry.S | 705 * With CONFIG_PREEMPT, we also check for extra work when the preempt_count 716 (pKStk) ld4 r21=[r20] // r21 <- preempt_count 719 cmp.eq p6,p0=r21,r0 // p6 <- pUStk || (preempt_count == 0) 849 * With CONFIG_PREEMPT, we also check for extra work when the preempt_count 860 (pKStk) ld4 r21=[r20] // r21 <- preempt_count 863 cmp.eq p6,p0=r21,r0 // p6 <- pUStk || (preempt_count == 0)
|
H A D | asm-offsets.c | 40 DEFINE(TI_PRE_COUNT, offsetof(struct thread_info, preempt_count)); foo()
|
H A D | mca.c | 1820 ti->preempt_count = 1; format_mca_init_stack()
|
/linux-4.1.27/init/ |
H A D | main.c | 779 int count = preempt_count(); do_one_initcall() 793 if (preempt_count() != count) { do_one_initcall()
|
/linux-4.1.27/kernel/sched/ |
H A D | cputime.c | 44 * Called before incrementing preempt_count on {soft,}irq_enter 45 * and before decrementing preempt_count on {soft,}irq_exit.
|
H A D | core.c | 1546 * Fold TIF_NEED_RESCHED into the preempt_count; anybody setting scheduler_ipi() 2565 if (DEBUG_LOCKS_WARN_ON((preempt_count() < 0))) preempt_count_add() 2573 DEBUG_LOCKS_WARN_ON((preempt_count() & PREEMPT_MASK) >= preempt_count_add() 2576 if (preempt_count() == val) { preempt_count_add() 2593 if (DEBUG_LOCKS_WARN_ON(val > preempt_count())) preempt_count_sub() 2599 !(preempt_count() & PREEMPT_MASK))) preempt_count_sub() 2603 if (preempt_count() == val) preempt_count_sub() 2621 prev->comm, prev->pid, preempt_count()); __schedule_bug() 2769 if (prev->state && !(preempt_count() & PREEMPT_ACTIVE)) { __schedule() 2860 * Returns with preemption disabled. Note: preempt_count must be 1 2893 * If there is a non-zero preempt_count or interrupts are disabled, preempt_schedule() 2957 BUG_ON(preempt_count() || !irqs_disabled()); preempt_schedule_irq() 7284 int nested = (preempt_count() & ~PREEMPT_ACTIVE) + rcu_preempt_depth(); preempt_count_equals()
|
/linux-4.1.27/arch/x86/kernel/ |
H A D | process_32.c | 282 * preempt_count of all tasks was equal here and this would not be __switch_to()
|
H A D | kvm.c | 143 n.halted = is_idle_task(current) || preempt_count() > 1; kvm_async_pf_task_wait()
|
H A D | process_64.c | 404 * preempt_count of all tasks was equal here and this would not be __switch_to()
|
H A D | entry_64.S | 696 * or not. While this is essentially redundant with preempt_count it is
|
/linux-4.1.27/include/trace/events/ |
H A D | sched.h | 106 if (preempt_count() & PREEMPT_ACTIVE) __trace_sched_switch_state()
|
/linux-4.1.27/arch/sparc/kernel/ |
H A D | traps_32.c | 438 TI_PREEMPT != offsetof(struct thread_info, preempt_count) || trap_init()
|
H A D | traps_64.c | 2701 preempt_count) || trap_init()
|
/linux-4.1.27/arch/mips/kernel/ |
H A D | asm-offsets.c | 98 OFFSET(TI_PRE_COUNT, thread_info, preempt_count); output_thread_info_defines()
|
/linux-4.1.27/arch/parisc/kernel/ |
H A D | asm-offsets.c | 248 DEFINE(TI_PRE_COUNT, offsetof(struct thread_info, preempt_count)); main()
|
H A D | entry.S | 970 * and preempt_count is 0. otherwise, we continue on 978 /* current_thread_info()->preempt_count */ 981 cmpib,COND(<>) 0, %r19, intr_restore /* if preempt_count > 0 */
|
/linux-4.1.27/kernel/time/ |
H A D | timer.c | 1131 int count = preempt_count(); call_timer_fn() 1158 if (count != preempt_count()) { call_timer_fn() 1160 fn, count, preempt_count()); call_timer_fn()
|
H A D | clocksource.c | 30 #include <linux/sched.h> /* for spin_unlock_irq() using preempt_count() m68k */
|
/linux-4.1.27/include/trace/ |
H A D | ftrace.h | 540 * pc = preempt_count();
|
/linux-4.1.27/arch/blackfin/mach-common/ |
H A D | entry.S | 915 /* check preempt_count */
|
/linux-4.1.27/kernel/rcu/ |
H A D | rcutorture.c | 289 if (!preempt_count() && rcu_read_delay()
|
/linux-4.1.27/mm/ |
H A D | memory.c | 3732 if (preempt_count()) print_vma_addr()
|