preempt 217 arch/arm/include/asm/assembler.h ldr \tmp, [\ti, #TI_PREEMPT] @ get preempt count preempt 223 arch/arm/include/asm/assembler.h ldr \tmp, [\ti, #TI_PREEMPT] @ get preempt count preempt 12 arch/arm64/include/asm/preempt.h return READ_ONCE(current_thread_info()->preempt.count); preempt 18 arch/arm64/include/asm/preempt.h WRITE_ONCE(current_thread_info()->preempt.count, pc); preempt 31 arch/arm64/include/asm/preempt.h current_thread_info()->preempt.need_resched = 0; preempt 36 arch/arm64/include/asm/preempt.h current_thread_info()->preempt.need_resched = 1; preempt 41 arch/arm64/include/asm/preempt.h return !current_thread_info()->preempt.need_resched; preempt 46 arch/arm64/include/asm/preempt.h u32 pc = READ_ONCE(current_thread_info()->preempt.count); preempt 48 arch/arm64/include/asm/preempt.h WRITE_ONCE(current_thread_info()->preempt.count, pc); preempt 53 arch/arm64/include/asm/preempt.h u32 pc = READ_ONCE(current_thread_info()->preempt.count); preempt 55 arch/arm64/include/asm/preempt.h WRITE_ONCE(current_thread_info()->preempt.count, pc); preempt 64 arch/arm64/include/asm/preempt.h WRITE_ONCE(ti->preempt.count, --pc); preempt 42 arch/arm64/include/asm/thread_info.h } preempt; preempt 923 drivers/char/lp.c ppdev_cb.preempt = lp_preempt; preempt 357 drivers/gpu/drm/i915/gt/intel_lrc.c bool preempt = false; preempt 365 drivers/gpu/drm/i915/gt/intel_lrc.c preempt = rq_prio(next) > last_prio; preempt 369 drivers/gpu/drm/i915/gt/intel_lrc.c if (preempt) preempt 370 drivers/gpu/drm/i915/gt/intel_lrc.c return preempt; preempt 34 drivers/gpu/drm/msm/adreno/a5xx_gpu.h struct a5xx_preempt_record *preempt[MSM_GPU_MAX_RINGS]; preempt 136 drivers/gpu/drm/msm/adreno/a5xx_preempt.c a5xx_gpu->preempt[ring->id]->wptr = get_wptr(ring); preempt 210 drivers/gpu/drm/msm/adreno/a5xx_preempt.c a5xx_gpu->preempt[i]->wptr = 0; preempt 211 drivers/gpu/drm/msm/adreno/a5xx_preempt.c a5xx_gpu->preempt[i]->rptr = 0; preempt 212 drivers/gpu/drm/msm/adreno/a5xx_preempt.c a5xx_gpu->preempt[i]->rbase = gpu->rb[i]->iova; preempt 243 drivers/gpu/drm/msm/adreno/a5xx_preempt.c a5xx_gpu->preempt[ring->id] = ptr; preempt 1288 drivers/net/plip/plip.c plip_cb.preempt = plip_preempt; preempt 772 drivers/parport/share.c tmp->preempt = pf; preempt 861 drivers/parport/share.c if (!par_dev_cb->preempt || !par_dev_cb->wakeup) { preempt 902 drivers/parport/share.c par_dev->preempt = par_dev_cb->preempt; preempt 1150 drivers/parport/share.c if (oldcad->preempt) { preempt 1151 drivers/parport/share.c if (oldcad->preempt(oldcad->private)) preempt 142 include/linux/parport.h int (*preempt)(void *); preempt 321 include/linux/parport.h int (*preempt)(void *); preempt 49 include/linux/rcutiny.h #define rcu_note_context_switch(preempt) \ preempt 21 include/linux/rcutree.h void rcu_note_context_switch(bool preempt); preempt 1588 include/linux/sched.h extern int yield_to(struct task_struct *p, bool preempt); preempt 108 include/trace/events/sched.h static inline long __trace_sched_switch_state(bool preempt, struct task_struct *p) preempt 120 include/trace/events/sched.h if (preempt) preempt 140 include/trace/events/sched.h TP_PROTO(bool preempt, preempt 144 include/trace/events/sched.h TP_ARGS(preempt, prev, next), preempt 160 include/trace/events/sched.h __entry->prev_state = __trace_sched_switch_state(preempt, prev); preempt 285 kernel/rcu/tree_plugin.h void rcu_note_context_switch(bool preempt) preempt 293 kernel/rcu/tree_plugin.h WARN_ON_ONCE(!preempt && t->rcu_read_lock_nesting > 0); preempt 844 kernel/rcu/tree_plugin.h void rcu_note_context_switch(bool preempt) preempt 854 kernel/rcu/tree_plugin.h if (!preempt) preempt 3873 kernel/sched/core.c static inline void schedule_debug(struct task_struct *prev, bool preempt) preempt 3881 kernel/sched/core.c if (!preempt && prev->state && prev->non_block_count) { preempt 3997 kernel/sched/core.c static void __sched notrace __schedule(bool preempt) preempt 4009 kernel/sched/core.c schedule_debug(prev, preempt); preempt 4015 kernel/sched/core.c rcu_note_context_switch(preempt); preempt 4033 kernel/sched/core.c if (!preempt && prev->state) { preempt 4074 kernel/sched/core.c trace_sched_switch(preempt, prev, next); preempt 5692 kernel/sched/core.c int __sched yield_to(struct task_struct *p, bool preempt) preempt 5728 kernel/sched/core.c yielded = curr->sched_class->yield_to_task(rq, p, preempt); preempt 5735 kernel/sched/core.c if (preempt && rq != p_rq) preempt 6733 kernel/sched/fair.c goto preempt; preempt 6752 kernel/sched/fair.c goto preempt; preempt 6757 kernel/sched/fair.c preempt: preempt 6969 kernel/sched/fair.c static bool yield_to_task_fair(struct rq *rq, struct task_struct *p, bool preempt) preempt 1718 kernel/sched/sched.h bool (*yield_to_task)(struct rq *rq, struct task_struct *p, bool preempt); preempt 389 kernel/trace/fgraph.c ftrace_graph_probe_sched_switch(void *ignore, bool preempt, preempt 6423 kernel/trace/ftrace.c ftrace_filter_pid_sched_switch_probe(void *data, bool preempt, preempt 541 kernel/trace/trace_events.c event_filter_pid_sched_switch_probe_pre(void *data, bool preempt, preempt 555 kernel/trace/trace_events.c event_filter_pid_sched_switch_probe_post(void *data, bool preempt, preempt 24 kernel/trace/trace_sched_switch.c probe_sched_switch(void *ignore, bool preempt, preempt 431 kernel/trace/trace_sched_wakeup.c probe_wakeup_sched_switch(void *ignore, bool preempt, preempt 911 sound/drivers/mts64.c .preempt = NULL, preempt 710 sound/drivers/portman2x4.c .preempt = NULL,