/linux-4.4.14/include/linux/ |
D | rcutiny.h | 62 synchronize_sched(); /* Only one CPU, so pretty fast anyway!!! */ in synchronize_rcu_expedited() 72 synchronize_sched(); in synchronize_rcu_bh() 77 synchronize_sched(); in synchronize_rcu_bh_expedited() 82 synchronize_sched(); in synchronize_sched_expedited()
|
D | rcupdate.h | 218 void synchronize_sched(void); 312 synchronize_sched(); in synchronize_rcu()
|
D | tracepoint.h | 94 synchronize_sched(); in tracepoint_synchronize_unregister()
|
/linux-4.4.14/Documentation/RCU/ |
D | NMI-RCU.txt | 84 One way to accomplish this is via synchronize_sched(), perhaps as 88 synchronize_sched(); 91 This works because synchronize_sched() blocks until all CPUs complete 93 Since NMI handlers disable preemption, synchronize_sched() is guaranteed 95 to free up the handler's data as soon as synchronize_sched() returns.
|
D | checklist.txt | 176 synchronize_rcu_bh(), synchronize_sched(), synchronize_srcu(), 206 updater uses call_rcu_sched() or synchronize_sched(), then 328 need to use synchronize_irq() or synchronize_sched().
|
D | UP.txt | 89 It -is- safe for synchronize_sched() and synchronize_rcu_bh() to return
|
D | whatisRCU.txt | 334 c. synchronize_sched() rcu_read_lock_sched() / rcu_read_unlock_sched() 858 rcu_read_lock_sched synchronize_sched rcu_barrier_sched
|
D | RTFP.txt | 1295 in favor of synchronize_rcu() and synchronize_sched().
|
/linux-4.4.14/kernel/ |
D | membarrier.c | 61 synchronize_sched(); in SYSCALL_DEFINE2()
|
D | kprobes.c | 230 synchronize_sched(); in collect_garbage_slots() 549 synchronize_sched(); in kprobe_optimizer() 1701 synchronize_sched(); in unregister_kprobes() 1773 synchronize_sched(); in unregister_jprobes() 1926 synchronize_sched(); in unregister_kretprobes()
|
D | module.c | 2036 synchronize_sched(); in free_module() 3323 synchronize_sched(); in do_init_module() 3574 synchronize_sched(); in load_module() 3589 synchronize_sched(); in load_module()
|
/linux-4.4.14/kernel/rcu/ |
D | tiny.c | 192 void synchronize_sched(void) in synchronize_sched() function 200 EXPORT_SYMBOL_GPL(synchronize_sched);
|
D | sync.c | 47 .sync = synchronize_sched,
|
D | update.c | 697 synchronize_sched(); in rcu_tasks_kthread() 775 synchronize_sched(); in rcu_tasks_kthread()
|
D | tree.c | 3201 void synchronize_sched(void) in synchronize_sched() function 3214 EXPORT_SYMBOL_GPL(synchronize_sched); 3343 synchronize_sched(); in cond_synchronize_sched()
|
D | rcutorture.c | 636 .sync = synchronize_sched,
|
/linux-4.4.14/arch/sparc/oprofile/ |
D | init.c | 56 synchronize_sched(); /* Allow already-started NMIs to complete. */ in timer_stop()
|
/linux-4.4.14/tools/lib/lockdep/uinclude/linux/ |
D | kernel.h | 33 #define synchronize_sched() macro
|
/linux-4.4.14/kernel/trace/ |
D | trace_events_filter.c | 1882 synchronize_sched(); in replace_system_preds() 1900 synchronize_sched(); in replace_system_preds() 2046 synchronize_sched(); in apply_event_filter() 2073 synchronize_sched(); in apply_event_filter() 2103 synchronize_sched(); in apply_subsystem_event_filter()
|
D | trace_events_trigger.c | 37 synchronize_sched(); /* make sure current triggers exit before free */ in trigger_data_free() 734 synchronize_sched(); in set_trigger_filter()
|
D | trace_events.c | 603 synchronize_sched(); in __ftrace_clear_event_pids() 1721 synchronize_sched(); in ftrace_event_pid_write() 3036 synchronize_sched(); in event_trace_del_tracer()
|
D | ring_buffer.c | 1798 synchronize_sched(); in ring_buffer_resize() 4061 synchronize_sched(); in ring_buffer_read_prepare_sync() 4233 synchronize_sched(); in ring_buffer_reset_cpu()
|
D | trace_uprobe.c | 970 synchronize_sched(); in probe_event_disable()
|
D | trace_kprobe.c | 424 synchronize_sched(); in disable_trace_kprobe()
|
D | trace.c | 1313 synchronize_sched(); in tracing_reset() 1330 synchronize_sched(); in tracing_reset_online_cpus() 4428 synchronize_sched(); in tracing_set_tracer()
|
D | ftrace.c | 3895 synchronize_sched(); in __unregister_ftrace_function_probe()
|
/linux-4.4.14/fs/ |
D | file.c | 178 synchronize_sched(); in expand_fdtable()
|
/linux-4.4.14/Documentation/ |
D | kprobes.txt | 236 rather, it calls synchronize_sched() for safety first, because it's 238 optimized region(*). As you know, synchronize_sched() can ensure 239 that all interruptions that were active when synchronize_sched()
|
/linux-4.4.14/drivers/net/ethernet/sis/ |
D | sis190.c | 1147 synchronize_sched(); in sis190_down()
|
/linux-4.4.14/arch/x86/kernel/cpu/mcheck/ |
D | mce.c | 1871 synchronize_sched(); in mce_chrdev_read()
|
/linux-4.4.14/kernel/locking/ |
D | lockdep.c | 3953 synchronize_sched(); in lockdep_free_key_range()
|
/linux-4.4.14/drivers/char/ipmi/ |
D | ipmi_si_intf.c | 3737 synchronize_sched(); in try_smi_init()
|
/linux-4.4.14/drivers/net/ethernet/realtek/ |
D | r8169.c | 6853 synchronize_sched(); in rtl_reset_work() 7602 synchronize_sched(); in rtl8169_down()
|