Searched refs:stopper (Results 1 - 7 of 7) sorted by relevance
/linux-4.4.14/kernel/ |
H A D | stop_machine.c | 36 /* the actual stopper, one per every possible cpu, enabled on online cpus */ 41 bool enabled; /* is this stopper enabled? */ 76 static void __cpu_stop_queue_work(struct cpu_stopper *stopper, __cpu_stop_queue_work() argument 79 list_add_tail(&work->list, &stopper->works); __cpu_stop_queue_work() 80 wake_up_process(stopper->thread); __cpu_stop_queue_work() 83 /* queue @work to @stopper. if offline, @work is completed immediately */ cpu_stop_queue_work() 86 struct cpu_stopper *stopper = &per_cpu(cpu_stopper, cpu); cpu_stop_queue_work() local 89 spin_lock_irqsave(&stopper->lock, flags); cpu_stop_queue_work() 90 if (stopper->enabled) cpu_stop_queue_work() 91 __cpu_stop_queue_work(stopper, work); cpu_stop_queue_work() 94 spin_unlock_irqrestore(&stopper->lock, flags); cpu_stop_queue_work() 301 * and will remain untouched until stopper starts executing @fn. 325 * preempted by a stopper which might wait for other stoppers queue_stop_cpus_work() 421 struct cpu_stopper *stopper = &per_cpu(cpu_stopper, cpu); cpu_stop_should_run() local 425 spin_lock_irqsave(&stopper->lock, flags); cpu_stop_should_run() 426 run = !list_empty(&stopper->works); cpu_stop_should_run() 427 spin_unlock_irqrestore(&stopper->lock, flags); cpu_stop_should_run() 433 struct cpu_stopper *stopper = &per_cpu(cpu_stopper, cpu); cpu_stopper_thread() local 439 spin_lock_irq(&stopper->lock); cpu_stopper_thread() 440 if (!list_empty(&stopper->works)) { cpu_stopper_thread() 441 work = list_first_entry(&stopper->works, cpu_stopper_thread() 445 spin_unlock_irq(&stopper->lock); cpu_stopper_thread() 474 struct cpu_stopper *stopper = &per_cpu(cpu_stopper, cpu); stop_machine_park() local 476 * Lockless. cpu_stopper_thread() will take stopper->lock and flush stop_machine_park() 480 stopper->enabled = false; stop_machine_park() 481 kthread_park(stopper->thread); stop_machine_park() 493 struct cpu_stopper *stopper = &per_cpu(cpu_stopper, cpu); cpu_stop_park() local 495 WARN_ON(!list_empty(&stopper->works)); cpu_stop_park() 500 struct cpu_stopper *stopper = &per_cpu(cpu_stopper, cpu); stop_machine_unpark() local 502 stopper->enabled = true; stop_machine_unpark() 503 kthread_unpark(stopper->thread); stop_machine_unpark() 521 struct cpu_stopper *stopper = &per_cpu(cpu_stopper, cpu); for_each_possible_cpu() local 523 spin_lock_init(&stopper->lock); for_each_possible_cpu() 524 INIT_LIST_HEAD(&stopper->works); for_each_possible_cpu()
|
H A D | cpu.c | 333 /* Park the stopper thread */ take_cpu_down()
|
/linux-4.4.14/drivers/acpi/acpica/ |
H A D | exdebug.c | 354 * not be cleared by the trace stopper during the first match acpi_ex_interpreter_trace_enabled()
|
/linux-4.4.14/drivers/scsi/ |
H A D | advansys.c | 1849 ADV_CARR_T *icq_sp; /* Initiator command queue stopper pointer. */ 1850 ADV_CARR_T *irq_sp; /* Initiator response queue stopper pointer. */ 4262 * insert stopper carrier to terminate list adv_get_next_carrier() 6164 * Check if the IRQ stopper carrier contains a completed request. AdvISR() 6195 * Advance the stopper pointer to the next carrier AdvISR() 6197 * stopper carrier. AdvISR() 8424 * Use the current stopper to send the ADV_SCSI_REQ_Q command to AdvExeScsiQueue() 8425 * the microcode. The newly allocated stopper will become the new AdvExeScsiQueue() 8426 * stopper. AdvExeScsiQueue() 8431 * Set the 'next_vpa' pointer for the old stopper to be the AdvExeScsiQueue() 8432 * physical address of the new stopper. The RISC can only AdvExeScsiQueue() 8438 * Set the host adapter stopper pointer to point to the new carrier. AdvExeScsiQueue() 8460 * address of the new carrier stopper to the COMMA register. AdvExeScsiQueue()
|
/linux-4.4.14/drivers/block/paride/ |
H A D | pf.c | 265 static int pf_mask; /* stopper for pseudo-int */
|
/linux-4.4.14/kernel/sched/ |
H A D | core.c | 1056 * 2) stopper starts to run (implicitly forcing the migrated thread 1061 * 5) stopper completes and stop_one_cpu() returns and the migration 1119 * migration_cpu_stop - this will be executed by a highprio stopper thread
|
H A D | fair.c | 7342 * active_load_balance_cpu_stop is run by cpu stopper. It pushes
|
Completed in 444 milliseconds