/linux-4.4.14/include/asm-generic/ |
H A D | preempt.h | 8 static __always_inline int preempt_count(void) preempt_count() function 10 return current_thread_info()->preempt_count; preempt_count() 15 return ¤t_thread_info()->preempt_count; preempt_count_ptr() 27 task_thread_info(p)->preempt_count = FORK_PREEMPT_COUNT; \ 31 task_thread_info(p)->preempt_count = PREEMPT_ENABLED; \ 48 * The various preempt_count add/sub methods 76 return unlikely(preempt_count() == preempt_offset && should_resched()
|
/linux-4.4.14/arch/m32r/kernel/ |
H A D | asm-offsets.c | 10 OFFSET(TI_PRE_COUNT, thread_info, preempt_count); foo()
|
H A D | entry.S | 177 ld r9, @(TI_PRE_COUNT, r8) ; non-zero preempt_count ?
|
/linux-4.4.14/lib/ |
H A D | smp_processor_id.c | 15 if (likely(preempt_count())) check_preemption_disabled() 43 what1, what2, preempt_count() - 1, current->comm, current->pid); check_preemption_disabled()
|
H A D | locking-selftest.c | 972 unsigned long saved_preempt_count = preempt_count(); dotest()
|
/linux-4.4.14/arch/avr32/kernel/ |
H A D | asm-offsets.c | 17 OFFSET(TI_preempt_count, thread_info, preempt_count); foo()
|
/linux-4.4.14/include/linux/ |
H A D | preempt.h | 6 * preempt_count (used for kernel preemption, interrupt count, etc.) 58 /* preempt_count() and related functions, depends on PREEMPT_NEED_RESCHED */ 61 #define hardirq_count() (preempt_count() & HARDIRQ_MASK) 62 #define softirq_count() (preempt_count() & SOFTIRQ_MASK) 63 #define irq_count() (preempt_count() & (HARDIRQ_MASK | SOFTIRQ_MASK \ 80 #define in_nmi() (preempt_count() & NMI_MASK) 83 * The preempt_count offset after preempt_disable(); 92 * The preempt_count offset after spin_lock() 97 * The preempt_count offset needed for things like: 118 #define in_atomic() (preempt_count() != 0) 124 #define in_atomic_preempt_off() (preempt_count() != PREEMPT_DISABLE_OFFSET) 159 #define preemptible() (preempt_count() == 0 && !irqs_disabled())
|
H A D | bit_spinlock.h | 93 return preempt_count(); bit_spin_is_locked()
|
H A D | trace_events.h | 67 unsigned char preempt_count; member in struct:trace_entry
|
H A D | sched.h | 616 * Initial preempt_count value; reflects the preempt_count schedule invariant 619 * preempt_count() == 2*PREEMPT_DISABLE_OFFSET
|
H A D | rcupdate.h | 529 return preempt_count() != 0 || irqs_disabled(); rcu_read_lock_sched_held()
|
/linux-4.4.14/arch/sh/kernel/ |
H A D | irq.c | 91 * Copy the softirq bits in preempt_count so that the handle_one_irq() 94 irqctx->tinfo.preempt_count = handle_one_irq() 95 (irqctx->tinfo.preempt_count & ~SOFTIRQ_MASK) | handle_one_irq() 96 (curctx->tinfo.preempt_count & SOFTIRQ_MASK); handle_one_irq() 128 irqctx->tinfo.preempt_count = HARDIRQ_OFFSET; irq_ctx_init() 136 irqctx->tinfo.preempt_count = 0; irq_ctx_init()
|
H A D | asm-offsets.c | 26 DEFINE(TI_PRE_COUNT, offsetof(struct thread_info, preempt_count)); main()
|
H A D | entry-common.S | 97 mov.l @(TI_PRE_COUNT,r8), r0 ! current_thread_info->preempt_count
|
/linux-4.4.14/arch/x86/include/asm/ |
H A D | preempt.h | 20 static __always_inline int preempt_count(void) preempt_count() function 64 * The various preempt_count add/sub methods 79 * a decrement which hits zero means we have no preempt_count and should
|
H A D | thread_info.h | 160 * preempt_count needs to be 1 initially, until the scheduler is functional.
|
/linux-4.4.14/arch/c6x/include/asm/ |
H A D | thread_info.h | 45 int preempt_count; /* 0 = preemptable, <0 = BUG */ member in struct:thread_info 52 * preempt_count needs to be 1 initially, until the scheduler is functional. 59 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/score/include/asm/ |
H A D | thread_info.h | 36 int preempt_count; member in struct:thread_info 50 * preempt_count needs to be 1 initially, until the scheduler is functional. 56 .preempt_count = 1, \
|
/linux-4.4.14/arch/um/include/asm/ |
H A D | thread_info.h | 19 int preempt_count; /* 0 => preemptable, member in struct:thread_info 32 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/m68k/include/asm/ |
H A D | thread_info.h | 30 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 40 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/parisc/include/asm/ |
H A D | thread_info.h | 15 int preempt_count; /* 0=premptable, <0=BUG; will also serve as bh-counter */ member in struct:thread_info 24 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/arc/include/asm/ |
H A D | thread_info.h | 44 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 54 * preempt_count needs to be 1 initially, until the scheduler is functional. 61 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/sparc/include/asm/ |
H A D | thread_info_32.h | 32 int preempt_count; /* 0 => preemptable, member in struct:thread_info 62 .preempt_count = INIT_PREEMPT_COUNT, \ 90 #define TI_PREEMPT 0x10 /* preempt_count */
|
H A D | thread_info_64.h | 46 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 119 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/nios2/include/asm/ |
H A D | thread_info.h | 44 int preempt_count; /* 0 => preemptable,<0 => BUG */ member in struct:thread_info 55 * preempt_count needs to be 1 initially, until the scheduler is functional. 62 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/openrisc/include/asm/ |
H A D | thread_info.h | 53 __s32 preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 69 * preempt_count needs to be 1 initially, until the scheduler is functional. 77 .preempt_count = 1, \
|
/linux-4.4.14/arch/cris/include/arch-v10/arch/ |
H A D | offset.h | 22 #define TI_preempt_count 16 /* offsetof(struct thread_info, preempt_count) */
|
/linux-4.4.14/arch/cris/include/arch-v32/arch/ |
H A D | offset.h | 24 #define TI_preempt_count 16 /* offsetof(struct thread_info, preempt_count) */
|
/linux-4.4.14/arch/cris/kernel/ |
H A D | asm-offsets.c | 40 ENTRY(preempt_count); main()
|
/linux-4.4.14/arch/s390/include/asm/ |
H A D | thread_info.h | 37 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 52 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/cris/include/asm/ |
H A D | thread_info.h | 33 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 54 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/h8300/include/asm/ |
H A D | thread_info.h | 32 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 45 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/avr32/include/asm/ |
H A D | thread_info.h | 25 __s32 preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 39 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/blackfin/include/asm/ |
H A D | thread_info.h | 42 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 57 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/metag/kernel/ |
H A D | irq.c | 84 * Copy the softirq bits in preempt_count so that the do_IRQ() 87 irqctx->tinfo.preempt_count = do_IRQ() 88 (irqctx->tinfo.preempt_count & ~SOFTIRQ_MASK) | do_IRQ() 89 (curctx->tinfo.preempt_count & SOFTIRQ_MASK); do_IRQ() 134 irqctx->tinfo.preempt_count = HARDIRQ_OFFSET; irq_ctx_init() 142 irqctx->tinfo.preempt_count = 0; irq_ctx_init()
|
/linux-4.4.14/arch/unicore32/include/asm/ |
H A D | thread_info.h | 69 int preempt_count; /* 0 => preemptable */ member in struct:thread_info 86 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/xtensa/include/asm/ |
H A D | thread_info.h | 50 __s32 preempt_count; /* 0 => preemptable,< 0 => BUG*/ member in struct:thread_info 76 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/alpha/include/asm/ |
H A D | thread_info.h | 23 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 38 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/kernel/trace/ |
H A D | trace_irqsoff.c | 42 return ((trace_type & TRACER_PREEMPT_OFF) && preempt_count()); preempt_trace() 117 if (!irqs_disabled_flags(*flags) && !preempt_count()) func_prolog_dec() 145 trace_function(tr, ip, parent_ip, flags, preempt_count()); irqsoff_tracer_call() 181 pc = preempt_count(); irqsoff_graph_entry() 198 pc = preempt_count(); irqsoff_graph_return() 317 pc = preempt_count(); check_critical_timing() 382 __trace_function(tr, ip, parent_ip, flags, preempt_count()); start_critical_timing() 416 __trace_function(tr, ip, parent_ip, flags, preempt_count()); stop_critical_timing()
|
H A D | trace_mmiotrace.c | 305 int pc = preempt_count(); __trace_mmiotrace_rw() 335 int pc = preempt_count(); __trace_mmiotrace_map()
|
H A D | trace_output.c | 426 if (entry->preempt_count) trace_print_lat_fmt() 427 trace_seq_printf(s, "%x", entry->preempt_count); trace_print_lat_fmt() 575 entry->preempt_count, iter->idx); trace_print_lat_context()
|
H A D | trace_sched_wakeup.c | 84 *pc = preempt_count(); func_prolog_preempt_disable() 454 pc = preempt_count(); probe_wakeup_sched_switch() 548 pc = preempt_count(); probe_wakeup()
|
H A D | trace_functions.c | 137 pc = preempt_count(); function_trace_call() 180 pc = preempt_count(); function_stack_trace_call()
|
H A D | trace_syscalls.c | 326 pc = preempt_count(); ftrace_syscall_enter() 371 pc = preempt_count(); ftrace_syscall_exit()
|
H A D | trace_branch.c | 61 pc = preempt_count(); probe_likely_condition()
|
H A D | trace_event_perf.c | 277 pc = preempt_count(); perf_trace_buf_prepare()
|
H A D | trace_functions_graph.c | 345 pc = preempt_count(); trace_graph_entry() 427 pc = preempt_count(); trace_graph_return()
|
H A D | trace_kprobe.c | 923 pc = preempt_count(); __kprobe_trace_func() 971 pc = preempt_count(); __kretprobe_trace_func()
|
H A D | trace_selftest.c | 875 if (preempt_count()) { trace_selftest_startup_preemptoff() 937 if (preempt_count()) { trace_selftest_startup_preemptirqsoff()
|
H A D | blktrace.c | 80 pc = preempt_count(); trace_note() 234 pc = preempt_count(); __blk_add_trace()
|
H A D | trace.c | 553 pc = preempt_count(); __trace_puts() 603 pc = preempt_count(); __trace_bputs() 1654 entry->preempt_count = pc & 0xff; tracing_generic_entry_update() 1915 flags, skip, preempt_count(), NULL); trace_dump_stack() 1975 ftrace_trace_userstack(tr, flags, preempt_count()); __trace_userstack() 2144 pc = preempt_count(); trace_vbprintk() 2200 pc = preempt_count(); __trace_array_vprintk() 5162 irq_flags, preempt_count()); tracing_mark_write()
|
H A D | trace_events.c | 189 __common_field(unsigned char, preempt_count); trace_define_common_fields() 246 fbuffer->pc = preempt_count(); trace_event_buffer_reserve() 3379 pc = preempt_count(); function_test_events_call()
|
/linux-4.4.14/arch/metag/include/asm/ |
H A D | thread_info.h | 34 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 72 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/frv/include/asm/ |
H A D | thread_info.h | 37 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 63 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/hexagon/include/asm/ |
H A D | thread_info.h | 52 int preempt_count; /* 0=>preemptible,<0=>BUG */ member in struct:thread_info 81 .preempt_count = 1, \
|
/linux-4.4.14/arch/m32r/include/asm/ |
H A D | thread_info.h | 30 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 54 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/arm64/include/asm/ |
H A D | thread_info.h | 51 int preempt_count; /* 0 => preemptable, <0 => bug */ member in struct:thread_info 59 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/arc/kernel/ |
H A D | asm-offsets.c | 34 offsetof(struct thread_info, preempt_count)); main()
|
H A D | entry-compact.S | 189 ; bump thread_info->preempt_count (Disable preemption) 381 ; decrement thread_info->preempt_count (re-enable preemption)
|
/linux-4.4.14/kernel/ |
H A D | softirq.c | 81 * preempt_count and SOFTIRQ_OFFSET usage: 82 * - preempt_count is changed by SOFTIRQ_OFFSET on entering or leaving 84 * - preempt_count is changed by SOFTIRQ_DISABLE_OFFSET (= 2 * SOFTIRQ_OFFSET) 106 * We must manually increment preempt_count here and manually __local_bh_disable_ip() 117 if (preempt_count() == cnt) { __local_bh_disable_ip() 268 prev_count = preempt_count(); __do_softirq() 275 if (unlikely(prev_count != preempt_count())) { __do_softirq() 276 pr_err("huh, entered softirq %u %s %p with preempt_count %08x, exited with %08x?\n", __do_softirq() 278 prev_count, preempt_count()); __do_softirq()
|
H A D | stop_machine.c | 462 WARN_ONCE(preempt_count(), cpu_stopper_thread()
|
H A D | exit.c | 710 pr_info("note: %s[%d] exited with preempt_count %d\n", do_exit() 712 preempt_count()); do_exit()
|
H A D | workqueue.c | 2081 current->comm, preempt_count(), task_pid_nr(current),
|
/linux-4.4.14/arch/tile/include/asm/ |
H A D | thread_info.h | 33 int preempt_count; /* 0 => preemptable, member in struct:thread_info 56 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/powerpc/include/asm/ |
H A D | thread_info.h | 43 int preempt_count; /* 0 => preemptable, member in struct:thread_info 58 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/mips/include/asm/ |
H A D | thread_info.h | 29 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 48 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/sh/include/asm/ |
H A D | thread_info.h | 33 int preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 61 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/microblaze/include/asm/ |
H A D | thread_info.h | 71 __s32 preempt_count; /* 0 => preemptable,< 0 => BUG*/ member in struct:thread_info 85 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/mn10300/include/asm/ |
H A D | thread_info.h | 46 __s32 preempt_count; /* 0 => preemptable, <0 => BUG */ member in struct:thread_info 78 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/ia64/include/asm/ |
H A D | thread_info.h | 28 int preempt_count; /* 0=premptable, <0=BUG; will also serve as bh-counter */ member in struct:thread_info 45 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/arm/include/asm/ |
H A D | thread_info.h | 51 int preempt_count; /* 0 => preemptable, <0 => bug */ member in struct:thread_info 74 .preempt_count = INIT_PREEMPT_COUNT, \
|
/linux-4.4.14/arch/tile/kernel/ |
H A D | asm-offsets.c | 62 offsetof(struct thread_info, preempt_count)); foo()
|
/linux-4.4.14/arch/nios2/kernel/ |
H A D | asm-offsets.c | 83 OFFSET(TI_PREEMPT_COUNT, thread_info, preempt_count); main()
|
/linux-4.4.14/arch/openrisc/kernel/ |
H A D | asm-offsets.c | 55 DEFINE(TI_PREEMPT, offsetof(struct thread_info, preempt_count)); main()
|
/linux-4.4.14/arch/h8300/kernel/ |
H A D | asm-offsets.c | 64 OFFSET(TI_PRE, thread_info, preempt_count); main()
|
/linux-4.4.14/arch/powerpc/kernel/ |
H A D | irq.c | 531 /* Copy the preempt_count so that the [soft]irq checks work. */ do_IRQ() 532 irqtp->preempt_count = curtp->preempt_count; do_IRQ() 581 tp->preempt_count = 0; for_each_possible_cpu() 587 tp->preempt_count = 0; for_each_possible_cpu() 592 tp->preempt_count = HARDIRQ_OFFSET; for_each_possible_cpu()
|
H A D | machine_kexec_64.c | 353 * We setup preempt_count to avoid using VMX in memcpy. default_machine_kexec() 358 kexec_stack.thread_info.preempt_count = HARDIRQ_OFFSET; default_machine_kexec()
|
H A D | asm-offsets.c | 162 DEFINE(TI_PREEMPT, offsetof(struct thread_info, preempt_count)); main()
|
H A D | entry_32.S | 816 /* check current_thread_info->preempt_count */
|
H A D | entry_64.S | 750 /* Check that preempt_count() == 0 and interrupts are enabled */
|
/linux-4.4.14/arch/c6x/kernel/ |
H A D | asm-offsets.c | 103 OFFSET(THREAD_INFO_PREEMPT_COUNT, thread_info, preempt_count); foo()
|
/linux-4.4.14/arch/powerpc/platforms/cell/ |
H A D | smp.c | 86 task_thread_info(paca[lcpu].__current)->preempt_count = 0; smp_startup_cpu()
|
/linux-4.4.14/arch/m68k/kernel/ |
H A D | asm-offsets.c | 41 DEFINE(TINFO_PREEMPT, offsetof(struct thread_info, preempt_count)); main()
|
/linux-4.4.14/arch/mn10300/kernel/ |
H A D | asm-offsets.c | 28 OFFSET(TI_preempt_count, thread_info, preempt_count); foo()
|
H A D | entry.S | 179 mov (TI_preempt_count,a2),d0 # non-zero preempt_count ?
|
/linux-4.4.14/arch/frv/kernel/ |
H A D | asm-offsets.c | 40 OFFSET(TI_PREEMPT_COUNT, thread_info, preempt_count); foo()
|
/linux-4.4.14/arch/unicore32/kernel/ |
H A D | asm-offsets.c | 42 DEFINE(TI_PREEMPT, offsetof(struct thread_info, preempt_count)); main()
|
/linux-4.4.14/arch/xtensa/kernel/ |
H A D | asm-offsets.c | 85 OFFSET(TI_PRE_COUNT, thread_info, preempt_count); main()
|
H A D | entry.S | 530 /* Check current_thread_info->preempt_count */
|
/linux-4.4.14/arch/microblaze/kernel/ |
H A D | asm-offsets.c | 94 DEFINE(TI_PREEMPT_COUNT, offsetof(struct thread_info, preempt_count)); main()
|
H A D | entry.S | 723 /* MS: get preempt_count from thread info */
|
/linux-4.4.14/arch/powerpc/platforms/pseries/ |
H A D | smp.c | 118 task_thread_info(paca[lcpu].__current)->preempt_count = 0; smp_startup_cpu()
|
/linux-4.4.14/arch/s390/kernel/ |
H A D | asm-offsets.c | 45 OFFSET(__TI_precount, thread_info, preempt_count); main()
|
/linux-4.4.14/arch/score/kernel/ |
H A D | asm-offsets.c | 106 OFFSET(TI_PRE_COUNT, thread_info, preempt_count); output_thread_info_defines()
|
/linux-4.4.14/arch/arm64/kernel/ |
H A D | asm-offsets.c | 37 DEFINE(TI_PREEMPT, offsetof(struct thread_info, preempt_count)); main()
|
/linux-4.4.14/arch/blackfin/kernel/ |
H A D | asm-offsets.c | 49 OFFSET(TI_PREEMPT, thread_info, preempt_count); main()
|
/linux-4.4.14/arch/arm/kernel/ |
H A D | asm-offsets.c | 67 DEFINE(TI_PREEMPT, offsetof(struct thread_info, preempt_count)); main()
|
/linux-4.4.14/arch/ia64/kernel/ |
H A D | entry.S | 694 * With CONFIG_PREEMPT, we also check for extra work when the preempt_count 705 (pKStk) ld4 r21=[r20] // r21 <- preempt_count 708 cmp.eq p6,p0=r21,r0 // p6 <- pUStk || (preempt_count == 0) 838 * With CONFIG_PREEMPT, we also check for extra work when the preempt_count 849 (pKStk) ld4 r21=[r20] // r21 <- preempt_count 852 cmp.eq p6,p0=r21,r0 // p6 <- pUStk || (preempt_count == 0)
|
H A D | asm-offsets.c | 40 DEFINE(TI_PRE_COUNT, offsetof(struct thread_info, preempt_count)); foo()
|
H A D | mca.c | 1820 ti->preempt_count = 1; format_mca_init_stack()
|
/linux-4.4.14/init/ |
H A D | main.c | 784 int count = preempt_count(); do_one_initcall() 798 if (preempt_count() != count) { do_one_initcall()
|
/linux-4.4.14/kernel/sched/ |
H A D | cputime.c | 44 * Called before incrementing preempt_count on {soft,}irq_enter 45 * and before decrementing preempt_count on {soft,}irq_exit.
|
H A D | core.c | 1812 * Fold TIF_NEED_RESCHED into the preempt_count; anybody setting scheduler_ipi() 2546 * The previous task will have left us with a preempt_count of 2 2556 if (WARN_ONCE(preempt_count() != 2*PREEMPT_DISABLE_OFFSET, 2557 "corrupted preempt_count: %s/%d/0x%x\n", 2558 current->comm, current->pid, preempt_count())) 2649 * finish_task_switch() will drop rq->lock() and lower preempt_count 2932 if (DEBUG_LOCKS_WARN_ON((preempt_count() < 0))) preempt_count_add() 2940 DEBUG_LOCKS_WARN_ON((preempt_count() & PREEMPT_MASK) >= preempt_count_add() 2943 if (preempt_count() == val) { preempt_count_add() 2960 if (DEBUG_LOCKS_WARN_ON(val > preempt_count())) preempt_count_sub() 2966 !(preempt_count() & PREEMPT_MASK))) preempt_count_sub() 2970 if (preempt_count() == val) preempt_count_sub() 2988 prev->comm, prev->pid, preempt_count()); __schedule_bug() 3238 * Returns with preemption disabled. Note: preempt_count must be 1 3270 * If there is a non-zero preempt_count or interrupts are disabled, preempt_schedule() 3331 BUG_ON(preempt_count() || !irqs_disabled()); preempt_schedule_irq() 7542 int nested = preempt_count() + rcu_preempt_depth(); preempt_count_equals()
|
/linux-4.4.14/arch/sparc/kernel/ |
H A D | traps_32.c | 438 TI_PREEMPT != offsetof(struct thread_info, preempt_count) || trap_init()
|
H A D | traps_64.c | 2701 preempt_count) || trap_init()
|
/linux-4.4.14/arch/mips/kernel/ |
H A D | asm-offsets.c | 98 OFFSET(TI_PRE_COUNT, thread_info, preempt_count); output_thread_info_defines()
|
/linux-4.4.14/kernel/time/ |
H A D | timer.c | 1156 int count = preempt_count(); call_timer_fn() 1183 if (count != preempt_count()) { call_timer_fn() 1185 fn, count, preempt_count()); call_timer_fn()
|
H A D | clocksource.c | 32 #include <linux/sched.h> /* for spin_unlock_irq() using preempt_count() m68k */
|
/linux-4.4.14/arch/parisc/kernel/ |
H A D | entry.S | 982 * and preempt_count is 0. otherwise, we continue on 990 /* current_thread_info()->preempt_count */ 993 cmpib,COND(<>) 0, %r19, intr_restore /* if preempt_count > 0 */
|
H A D | asm-offsets.c | 248 DEFINE(TI_PRE_COUNT, offsetof(struct thread_info, preempt_count)); main()
|
/linux-4.4.14/arch/x86/kernel/ |
H A D | kvm.c | 143 n.halted = is_idle_task(current) || preempt_count() > 1; kvm_async_pf_task_wait()
|
/linux-4.4.14/include/trace/ |
H A D | trace_events.h | 540 * pc = preempt_count();
|
/linux-4.4.14/kernel/rcu/ |
H A D | update.c | 109 return lockdep_opinion || preempt_count() != 0 || irqs_disabled(); rcu_read_lock_sched_held()
|
H A D | rcutorture.c | 290 if (!preempt_count() && rcu_read_delay()
|
/linux-4.4.14/arch/x86/entry/ |
H A D | entry_64.S | 531 * or not. While this is essentially redundant with preempt_count it is
|
/linux-4.4.14/arch/blackfin/mach-common/ |
H A D | entry.S | 915 /* check preempt_count */
|
/linux-4.4.14/mm/ |
H A D | memory.c | 3808 if (preempt_count()) print_vma_addr()
|