/linux-4.1.27/include/linux/ |
H A D | vermagic.h | 10 #define MODULE_VERMAGIC_PREEMPT "preempt "
|
H A D | bottom_half.h | 4 #include <linux/preempt.h>
|
H A D | preempt.h | 5 * include/linux/preempt.h - macros for accessing and manipulating 18 #include <asm/preempt.h> 111 * Even if we don't have any preemption, we need preempt disable/enable 113 * that can cause faults and scheduling migrate into our preempt-protected
|
H A D | hardirq.h | 31 * because NMI handlers may not preempt and the ops are
|
H A D | bit_spinlock.h | 5 #include <linux/preempt.h>
|
H A D | preempt_mask.h | 4 #include <linux/preempt.h>
|
H A D | spinlock_api_up.h | 23 * only thing we have to do is to keep the preempt counts and irq
|
H A D | uaccess.h | 4 #include <linux/preempt.h>
|
H A D | percpu.h | 5 #include <linux/preempt.h>
|
H A D | smp.h | 55 #include <linux/preempt.h>
|
H A D | spinlock_api_smp.h | 102 * not re-enabled during lock-acquire (which the preempt-spin-ops do):
|
H A D | stop_machine.h | 110 * holding a spinlock or inside any other preempt-disabled region when
|
H A D | spinlock.h | 38 * builds. (which are NOPs on non-debug, non-preempt 50 #include <linux/preempt.h>
|
H A D | rwlock_api_smp.h | 142 * not re-enabled during lock-acquire (which the preempt-spin-ops do):
|
H A D | ftrace_event.h | 502 * @pc: The state of the preempt count at the start of the event. 530 * @pc: The state of the preempt count at the start of the event.
|
H A D | seqlock.h | 36 #include <linux/preempt.h>
|
H A D | pagemap.h | 157 * disabling preempt, and hence no need for the "speculative get" that page_cache_get_speculative()
|
H A D | parport.h | 141 int (*preempt)(void *); member in struct:pardevice
|
H A D | percpu-defs.h | 401 * If there is no other protection through preempt disable and/or disabling
|
H A D | radix-tree.h | 24 #include <linux/preempt.h>
|
H A D | bio.h | 489 * might not be a highmem page, but the preempt/irq count bvec_kmap_irq()
|
H A D | interrupt.h | 8 #include <linux/preempt.h>
|
H A D | kvm_host.h | 19 #include <linux/preempt.h>
|
H A D | fs.h | 709 * with respect to the local cpu (unlike with preempt disabled), 714 * and 64bit archs it makes no difference if preempt is enabled or not.
|
H A D | sched.h | 2268 extern int yield_to(struct task_struct *p, bool preempt); sched_autogroup_exit()
|
/linux-4.1.27/arch/x86/include/asm/ |
H A D | preempt.h | 18 * that think a non-zero value indicates we cannot preempt. 43 * We fold the NEED_RESCHED bit into the preempt count such that 48 * need to resched (the bit is cleared) and can resched (no preempt count).
|
H A D | i387.h | 29 * Careful: __kernel_fpu_begin/end() must be called with preempt disabled 30 * and they don't touch the preempt state on their own. 31 * If you enable preemption after __kernel_fpu_begin(), preempt notifier 54 /* Must be called with preempt disabled */
|
H A D | mc146818rtc.h | 37 * All of these below must be called with interrupts off, preempt
|
H A D | kvm_emulate.h | 203 void (*get_fpu)(struct x86_emulate_ctxt *ctxt); /* disables preempt */ 204 void (*put_fpu)(struct x86_emulate_ctxt *ctxt); /* reenables preempt */
|
H A D | fpu-internal.h | 267 * These must be called with preempt disabled. Returns
|
/linux-4.1.27/arch/sh/kernel/ |
H A D | idle.c | 15 #include <linux/preempt.h>
|
H A D | kprobes.c | 14 #include <linux/preempt.h>
|
/linux-4.1.27/drivers/xen/ |
H A D | Makefile | 5 obj-y += grant-table.o features.o balloon.o manage.o preempt.o
|
/linux-4.1.27/arch/x86/kernel/kprobes/ |
H A D | ftrace.c | 23 #include <linux/preempt.h> 65 /* Disable irq for emulating a breakpoint and avoiding preempt */ kprobe_ftrace_handler()
|
H A D | opt.c | 26 #include <linux/preempt.h>
|
H A D | core.c | 47 #include <linux/preempt.h>
|
/linux-4.1.27/kernel/locking/ |
H A D | rtmutex-debug.c | 9 * This code is based on the rt.c implementation in the preempt-rt tree. 17 * See rt.c in preempt-rt for proper credits and further information
|
H A D | spinlock.c | 18 #include <linux/preempt.h> 27 * not re-enabled during lock-acquire (which the preempt-spin-ops do):
|
H A D | percpu-rwsem.c | 36 * serialize with the preempt-disabled section below.
|
/linux-4.1.27/arch/x86/lib/ |
H A D | msr.c | 2 #include <linux/preempt.h>
|
H A D | delay.c | 17 #include <linux/preempt.h>
|
H A D | msr-smp.c | 2 #include <linux/preempt.h>
|
/linux-4.1.27/arch/s390/include/asm/ |
H A D | facility.h | 11 #include <linux/preempt.h>
|
H A D | percpu.h | 4 #include <linux/preempt.h>
|
/linux-4.1.27/arch/parisc/lib/ |
H A D | delay.c | 15 #include <linux/preempt.h>
|
/linux-4.1.27/arch/m68k/include/asm/ |
H A D | irqflags.h | 8 #include <linux/preempt.h>
|
/linux-4.1.27/arch/metag/include/asm/ |
H A D | irqflags.h | 82 /* Avoid circular include dependencies through <linux/preempt.h> */
|
/linux-4.1.27/arch/arm/mach-orion5x/ |
H A D | common.h | 80 * These are not preempt-safe. Locks, if needed, must be taken
|
/linux-4.1.27/drivers/staging/speakup/ |
H A D | fakekey.c | 22 #include <linux/preempt.h>
|
/linux-4.1.27/arch/mips/lib/ |
H A D | mips-atomic.c | 14 #include <linux/preempt.h>
|
/linux-4.1.27/tools/perf/scripts/python/Perf-Trace-Util/ |
H A D | Context.c | 77 "Get the common preempt count event field value."},
|
/linux-4.1.27/arch/arm/include/asm/ |
H A D | assembler.h | 191 * Increment/decrement the preempt count. 195 ldr \tmp, [\ti, #TI_PREEMPT] @ get preempt count 201 ldr \tmp, [\ti, #TI_PREEMPT] @ get preempt count
|
H A D | futex.h | 72 #include <linux/preempt.h>
|
/linux-4.1.27/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_mqd_manager.h | 39 * @destroy_mqd: Destroys the HQD slot and by that preempt the relevant queue.
|
H A D | kfd_mqd_manager_cik.c | 254 * preempt type here is ignored because there is only one way 255 * to preempt sdma queue
|
/linux-4.1.27/arch/x86/xen/ |
H A D | irq.c | 68 /* There's a one instruction preempt window here. We need to xen_irq_disable()
|
H A D | enlighten.c | 18 #include <linux/preempt.h>
|
H A D | mmu.c | 768 * traverse, because we may wrap the preempt count (8 xen_pin_page()
|
/linux-4.1.27/arch/powerpc/lib/ |
H A D | xor_vmx.c | 22 #include <linux/preempt.h>
|
/linux-4.1.27/arch/powerpc/mm/ |
H A D | mmu_context_hash32.c | 90 * Free a context ID. Make sure to call this with preempt disabled!
|
H A D | tlb_nohash.c | 36 #include <linux/preempt.h> 247 * We snapshot the PID with preempt disabled. At this point, it can still 407 * while preempt & CPU mask mucking around, or even the IPI but tlb_flush_pgtable()
|
H A D | tlb_hash64.c | 139 * Must be called from within some kind of spinlock/non-preempt region...
|
H A D | tlb_nohash_low.S | 45 * the PID and I don't want to preempt when that happens.
|
/linux-4.1.27/arch/arm/kernel/ |
H A D | smp_tlb.c | 10 #include <linux/preempt.h>
|
H A D | entry-armv.S | 210 ldr r8, [tsk, #TI_PREEMPT] @ get preempt count 212 teq r8, #0 @ if preempt count != 0
|
/linux-4.1.27/arch/arc/include/asm/ |
H A D | futex.h | 15 #include <linux/preempt.h>
|
H A D | smp.h | 89 * asm/bitops.h -> linux/spinlock.h -> linux/preempt.h
|
/linux-4.1.27/kernel/trace/ |
H A D | trace_irqsoff.c | 89 * Prologue for the preempt and irqs off function tracers. 109 * Does not matter if we preempt. We test the flags func_prolog_dec() 111 * If we preempt and get a false positive, the flags func_prolog_dec() 338 /* Skip 5 functions to get to the irq/preempt enable function */ check_critical_timing()
|
H A D | ring_buffer_benchmark.c | 255 * If we are a non preempt kernel, the 10 second run will ring_buffer_producer()
|
H A D | trace_selftest.c | 898 * in case of parallels max preempt off latencies. trace_selftest_startup_preemptoff() 957 /* reverse the order of preempt vs irqs */ trace_selftest_startup_preemptirqsoff() 964 * in case of parallels max irqs/preempt off latencies. trace_selftest_startup_preemptirqsoff() 993 /* reverse the order of preempt vs irqs */ trace_selftest_startup_preemptirqsoff()
|
H A D | trace_output.c | 422 * trace_print_lat_fmt - print the irq, preempt and lockdep fields 426 * Prints the generic fields of irqs off, in hard or softirq, preempt
|
H A D | trace_functions_graph.c | 1244 seq_printf(s, "#%.*s|| / _--=> preempt-depth \n", size, spaces); print_lat_header()
|
H A D | ring_buffer.c | 2643 * The lock and unlock are done within a preempt disable section. 2658 * preempt other contexts. A SoftIRQ never preempts an IRQ 4852 /* No sleep, but for non preempt, let others run */ rb_hammer_test()
|
H A D | trace.c | 2565 "# ||| / _--=> preempt-depth \n" print_lat_help_header() 2595 "# || / _--=> preempt-depth\n" print_func_help_header_irq() 2631 "preempt", print_trace_header()
|
/linux-4.1.27/arch/blackfin/mach-common/ |
H A D | interrupt.S | 126 r7 = [p5 + TI_PREEMPT]; /* get preempt count */ 134 [p5 + TI_PREEMPT] = r7; /* restore preempt count */
|
H A D | entry.S | 627 * preempt the current domain while the pipeline log is being played
|
/linux-4.1.27/arch/x86/kernel/cpu/mcheck/ |
H A D | mce-inject.c | 20 #include <linux/preempt.h>
|
H A D | mce_amd.c | 208 /* cpu init entry point, called from mce.c with preempt off */ mce_amd_feature_init()
|
H A D | mce.c | 501 /* Always runs in MCE context with preempt off */ mce_ring_add() 1687 * Must be called with preempt off:
|
/linux-4.1.27/arch/sparc/mm/ |
H A D | tlb.c | 10 #include <linux/preempt.h>
|
H A D | tsb.c | 7 #include <linux/preempt.h>
|
/linux-4.1.27/crypto/ |
H A D | xor.c | 24 #include <linux/preempt.h>
|
/linux-4.1.27/fs/xfs/ |
H A D | xfs_stats.h | 221 * We don't disable preempt, not too worried about poking the
|
/linux-4.1.27/arch/powerpc/platforms/cell/spufs/ |
H A D | fault.c | 162 * time slicing will not preempt the context while the page fault spufs_handle_class1()
|
H A D | run.c | 19 * It should be impossible to preempt a context while an exception spufs_stop_callback()
|
H A D | sched.c | 624 * find_victim - find a lower priority context to preempt
|
/linux-4.1.27/arch/mips/kernel/ |
H A D | entry.S | 31 local_irq_disable # preempt stop
|
H A D | kprobes.c | 27 #include <linux/preempt.h>
|
/linux-4.1.27/arch/powerpc/include/asm/ |
H A D | lppaca.h | 85 * hypervisor preempt). An even value implies that the processor is
|
/linux-4.1.27/arch/ia64/sn/kernel/ |
H A D | mca.c | 42 * code. SAL calls are run with preempt disabled so this routine must not
|
/linux-4.1.27/kernel/ |
H A D | panic.c | 90 * not have preempt disabled. Some functions called from here want panic() 91 * preempt to be disabled. No point enabling it later though... panic()
|
H A D | cpu.c | 378 * By now we've cleared cpu_active_mask, wait for all preempt-disabled _cpu_down() 395 * So now all preempt/rcu users must observe !cpu_active(). _cpu_down()
|
H A D | stop_machine.c | 224 * our local migrate thread gets the chance to preempt us. 480 "cpu_stop: %s(%p) leaked preempt count\n", cpu_stopper_thread()
|
H A D | reboot.c | 126 * preempt all other restart handlers
|
H A D | softirq.c | 103 * The preempt tracer hooks into preempt_count_add and will break __local_bh_disable_ip()
|
H A D | module.c | 431 * (optional) module which owns it. Needs preempt disabled or module_mutex. */ find_symbol() 3152 * walking this with preempt disabled. In all the failure paths, do_init_module() 3861 * Must be called with preempt disabled or module mutex held so that 3904 * Must be called with preempt disabled or module mutex held so that 3927 /* Most callers should already have preempt disabled, but make sure */ print_modules()
|
/linux-4.1.27/block/ |
H A D | blk-lib.c | 126 * us to schedule out to avoid softlocking if preempt blkdev_issue_discard()
|
H A D | cfq-iosched.c | 3235 * preempt async queue, limiting the sync queue doesn't make cfq_may_dispatch() 3783 * Check if new_cfqq should preempt the currently active queue. Return 0 for 3784 * no or if we aren't sure, a 1 will cause a preempt. 3803 * Don't allow a non-RT request to preempt an ongoing RT cfqq timeslice. cfq_should_preempt() 3850 * current cfqq, let it preempt cfq_should_preempt() 3859 * cfqq preempts the active queue. if we allowed preempt with no slice left, 3866 cfq_log_cfqq(cfqd, cfqq, "preempt"); cfq_preempt_queue() 3870 * workload type is changed, don't save slice, otherwise preempt cfq_preempt_queue()
|
H A D | blk-core.c | 3048 * preempt will imply a full memory barrier blk_start_plug()
|
/linux-4.1.27/arch/powerpc/kvm/ |
H A D | book3s_hv_builtin.c | 11 #include <linux/preempt.h>
|
H A D | e500.c | 117 /* Invalidate all id mappings on local core -- call with preempt disabled */ local_sid_destroy_all()
|
H A D | book3s_pr.c | 193 * our preempt notifiers. Don't bother touching this svcpu then. kvmppc_copy_from_svcpu()
|
H A D | book3s_hv.c | 24 #include <linux/preempt.h>
|
/linux-4.1.27/arch/arm64/kernel/ |
H A D | cpuinfo.c | 27 #include <linux/preempt.h>
|
H A D | efi.c | 23 #include <linux/preempt.h>
|
H A D | entry.S | 369 ldr w24, [tsk, #TI_PREEMPT] // get preempt count 370 cbnz w24, 1f // preempt count != 0
|
/linux-4.1.27/arch/m68k/coldfire/ |
H A D | entry.S | 124 jmp preempt_schedule_irq /* preempt the kernel */
|
/linux-4.1.27/arch/arm/kvm/ |
H A D | psci.c | 18 #include <linux/preempt.h>
|
/linux-4.1.27/include/asm-generic/ |
H A D | tlb.h | 42 * We use the sched RCU variant, as that guarantees that IRQ/preempt disabling
|
/linux-4.1.27/arch/arm/vfp/ |
H A D | vfpmodule.c | 156 * - we could be preempted if tree preempt rcu is enabled, so 163 * - we could be preempted if tree preempt rcu is enabled, so
|
/linux-4.1.27/arch/arc/kernel/ |
H A D | entry.S | 624 ; Can't preempt if preemption disabled 694 ; paranoid check, given A1 was active when A2 happened, preempt count
|
/linux-4.1.27/fs/ocfs2/cluster/ |
H A D | quorum.c | 72 /* panic spins with interrupts enabled. with preempt o2quo_fence_self()
|
/linux-4.1.27/arch/score/kernel/ |
H A D | entry.S | 238 disable_irq # preempt stop
|
/linux-4.1.27/drivers/parport/ |
H A D | share.c | 571 tmp->preempt = pf; parport_register_device() 787 if (oldcad->preempt) { parport_claim() 788 if (oldcad->preempt(oldcad->private)) parport_claim()
|
/linux-4.1.27/kernel/sched/ |
H A D | sched.h | 112 * Tells if entity @a should preempt entity @b. 774 * preempt-disabled sections. 1169 bool (*yield_to_task) (struct rq *rq, struct task_struct *p, bool preempt);
|
H A D | deadline.c | 234 * If we cannot preempt any rq, fall back to pick any dl_task_offline_migration() 1389 * can be sent to some other CPU where they can preempt 1522 * - it will preempt the last one we pulled (if any). pull_dl_task()
|
H A D | core.c | 2894 * we do not want to preempt the current task. Just return.. preempt_schedule() 2909 * recursion and tracing preempt enabling caused by the tracing 2911 * from userspace or just about to enter userspace, a preempt enable 2929 * Needs preempt disabled in case user_exit() is traced preempt_schedule_context() 3042 * --> -dl task blocks on mutex A and could preempt the rt_mutex_setprio() 4221 * no need to preempt or enable interrupts: SYSCALL_DEFINE0() 4319 * @preempt: whether task preemption is allowed or not 4329 int __sched yield_to(struct task_struct *p, bool preempt) yield_to() argument 4365 yielded = curr->sched_class->yield_to_task(rq, p, preempt); yield_to() 4372 if (preempt && rq != p_rq) yield_to() 4636 /* Set the preempt count _outside_ the spinlocks! */ init_idle()
|
H A D | fair.c | 4826 * preempt must be disabled. 4958 * Should 'se' preempt 'curr'. 5054 goto preempt; check_preempt_wakeup() 5057 * Batch and idle tasks do not preempt non-idle tasks (their preemption check_preempt_wakeup() 5073 goto preempt; check_preempt_wakeup() 5078 preempt: check_preempt_wakeup() 5269 static bool yield_to_task_fair(struct rq *rq, struct task_struct *p, bool preempt) yield_to_task_fair() argument 7233 * Drop the rq->lock, but keep IRQ/preempt disabled. idle_balance() 7881 * Priority of the task has changed. Check to see if we preempt 7957 * if we can still preempt the current task. switched_to_fair()
|
H A D | rt.c | 1589 * "this_cpu" is cheaper to preempt than a for_each_domain() 2166 * we may need to preempt the current running task. switched_to_rt()
|
/linux-4.1.27/drivers/rtc/ |
H A D | rtc-bfin.c | 125 * shouldn't be an issue on an SMP or preempt system as this function should
|
/linux-4.1.27/arch/x86/mm/ |
H A D | kmmio.c | 18 #include <linux/preempt.h>
|
/linux-4.1.27/arch/mn10300/kernel/ |
H A D | kprobes.c | 23 #include <linux/preempt.h>
|
/linux-4.1.27/arch/powerpc/kernel/ |
H A D | kprobes.c | 31 #include <linux/preempt.h>
|
H A D | vector.S | 297 * with preempt disabled.
|
H A D | irq.c | 212 * From this point onward, we can take interrupts, preempt, arch_local_irq_restore()
|
H A D | setup-common.c | 220 /* We only show online cpus: disable preempt (overzealous, I show_cpuinfo()
|
H A D | entry_64.S | 700 /* Check if we need to preempt */
|
/linux-4.1.27/arch/microblaze/kernel/ |
H A D | entry.S | 731 preempt: label 738 bnei r5, preempt /* if non zero jump to resched */
|
/linux-4.1.27/drivers/md/ |
H A D | dm-stats.c | 6 #include <linux/preempt.h>
|
/linux-4.1.27/arch/x86/kernel/ |
H A D | nmi.c | 433 * If an NMI executes an iret, another NMI can preempt it. We do not
|
/linux-4.1.27/arch/tile/kernel/ |
H A D | process.c | 16 #include <linux/preempt.h>
|
/linux-4.1.27/drivers/char/hw_random/ |
H A D | n2-drv.c | 12 #include <linux/preempt.h>
|
/linux-4.1.27/drivers/acpi/acpica/ |
H A D | dswexec.c | 551 * Tell the walk loop to preempt this running method and acpi_ds_exec_end_op()
|
/linux-4.1.27/drivers/virtio/ |
H A D | virtio_balloon.c | 376 * and always have work to do. Be nice if preempt disabled. balloon()
|
/linux-4.1.27/init/ |
H A D | main.c | 408 /* Call into cpu_idle with preempt disabled */ rest_init()
|
/linux-4.1.27/arch/s390/kernel/ |
H A D | kprobes.c | 25 #include <linux/preempt.h>
|
/linux-4.1.27/mm/ |
H A D | zbud.c | 50 #include <linux/preempt.h>
|
H A D | huge_memory.c | 2238 * ptl mostly unnecessary, but preempt has to __collapse_huge_page_copy()
|
H A D | slab.c | 61 * and local interrupts are disabled so slab code is preempt-safe.
|
H A D | memcontrol.c | 2111 * This must be called under preempt disabled or must be called by
|
/linux-4.1.27/sound/drivers/ |
H A D | mts64.c | 965 NULL, /* preempt */ snd_mts64_probe()
|
H A D | portman2x4.c | 764 NULL, /* preempt */ snd_portman_probe()
|
/linux-4.1.27/drivers/scsi/ |
H A D | constants.c | 171 {0x4, "Persistent reserve out, preempt"}, 172 {0x5, "Persistent reserve out, preempt and abort"},
|
/linux-4.1.27/kernel/time/ |
H A D | timer.c | 1159 WARN_ONCE(1, "timer: %pF preempt leak: %08x -> %08x\n", call_timer_fn() 1162 * Restore the preempt count. That gives us a decent call_timer_fn()
|
/linux-4.1.27/drivers/ide/ |
H A D | ide-io.c | 535 * We let requests forced at head of queue with ide-preempt do_ide_request()
|
/linux-4.1.27/drivers/target/ |
H A D | target_core_pr.c | 2836 * server shall perform a preempt by doing the following in an core_scsi3_pro_preempt() 2926 * with a zero SA rservation key, preempt the existing core_scsi3_pro_preempt() 2956 * reservation holder to preempt itself (i.e., a core_scsi3_pro_preempt() 2971 * server shall perform a preempt by doing the following as core_scsi3_pro_preempt()
|
/linux-4.1.27/drivers/input/misc/ |
H A D | wistron_btns.c | 30 #include <linux/preempt.h>
|
/linux-4.1.27/arch/sparc/kernel/ |
H A D | smp_64.c | 132 /* idle thread is expected to have preempt disabled */ smp_callin()
|
/linux-4.1.27/arch/metag/kernel/ |
H A D | traps.c | 19 #include <linux/preempt.h>
|
/linux-4.1.27/drivers/usb/host/ |
H A D | ohci-hcd.c | 577 /* 2msec timelimit here means no irqs/preempt */ ohci_run()
|
/linux-4.1.27/arch/ia64/kernel/ |
H A D | kprobes.c | 30 #include <linux/preempt.h>
|
/linux-4.1.27/net/mac80211/ |
H A D | mesh_hwmp.c | 1115 * this function is considered "using" the associated mpath, so preempt a path
|
/linux-4.1.27/kernel/rcu/ |
H A D | srcu.c | 32 #include <linux/preempt.h>
|
H A D | tree_plugin.h | 702 * Wait for an RCU-preempt grace period, but expedite it. The basic 906 * Wait for an rcu-preempt grace period, but make it happen quickly.
|
/linux-4.1.27/drivers/net/ethernet/sis/ |
H A D | sis190.c | 690 * again too early (hint: think preempt and unclocked smp systems). sis190_tx_interrupt()
|
/linux-4.1.27/arch/mips/mm/ |
H A D | c-r4k.c | 16 #include <linux/preempt.h>
|
/linux-4.1.27/drivers/infiniband/hw/qib/ |
H A D | qib_init.c | 1044 * This is only roughly accurate, since even with preempt we qib_verify_pioperf()
|
/linux-4.1.27/ipc/ |
H A D | sem.c | 695 * Hold preempt off so that we don't get preempted and have the wake_up_sem_queue_prepare()
|
/linux-4.1.27/arch/parisc/kernel/ |
H A D | entry.S | 968 /* preempt the current task on returning to kernel
|
/linux-4.1.27/kernel/irq/ |
H A D | manage.c | 816 * interrupts rely on the implicit bh/preempt disable of the hard irq irq_thread_check_affinity()
|
/linux-4.1.27/drivers/net/ethernet/3com/ |
H A D | typhoon.c | 637 * preempt or do anything other than take interrupts. So, don't typhoon_issue_command()
|
/linux-4.1.27/drivers/infiniband/hw/ipath/ |
H A D | ipath_driver.c | 359 * this is only roughly accurate, since even with preempt we ipath_verify_pioperf()
|
/linux-4.1.27/drivers/usb/gadget/function/ |
H A D | f_mass_storage.c | 415 * If a lower-or-equal priority exception is in progress, preempt it raise_exception()
|
/linux-4.1.27/net/core/ |
H A D | sock.c | 1948 * disabled, use cond_resched_softirq() to preempt.
|
/linux-4.1.27/drivers/net/bonding/ |
H A D | bond_main.c | 73 #include <linux/preempt.h>
|
/linux-4.1.27/drivers/net/ethernet/neterion/vxge/ |
H A D | vxge-main.c | 1838 * away which can preempt this NAPI thread */ vxge_poll_msix()
|
/linux-4.1.27/drivers/media/pci/bt8xx/ |
H A D | bttv-driver.c | 763 capturing is off, it shall no longer "preempt" VBI capturing, disclaim_video_lines()
|
/linux-4.1.27/fs/ |
H A D | buffer.c | 1418 * or with preempt disabled.
|
/linux-4.1.27/tools/lib/traceevent/ |
H A D | event-parse.c | 4944 * need rescheduling, in hard/soft interrupt, preempt count
|
/linux-4.1.27/fs/dlm/ |
H A D | lock.c | 1544 in the cancel reply should be 0). We preempt the cancel reply _remove_from_waiters()
|