Home
last modified time | relevance | path

Searched refs:this_cpu (Results 1 – 56 of 56) sorted by relevance

/linux-4.4.14/arch/x86/kernel/apic/
Dx2apic_cluster.c31 unsigned int cpu, this_cpu; in __x2apic_send_IPI_mask() local
39 this_cpu = smp_processor_id(); in __x2apic_send_IPI_mask()
59 if (apic_dest == APIC_DEST_ALLINC || i != this_cpu) in __x2apic_send_IPI_mask()
133 unsigned int this_cpu = smp_processor_id(); in init_x2apic_ldr() local
136 per_cpu(x86_cpu_to_logical_apicid, this_cpu) = apic_read(APIC_LDR); in init_x2apic_ldr()
138 cpumask_set_cpu(this_cpu, per_cpu(cpus_in_cluster, this_cpu)); in init_x2apic_ldr()
140 if (x2apic_cluster(this_cpu) != x2apic_cluster(cpu)) in init_x2apic_ldr()
142 cpumask_set_cpu(this_cpu, per_cpu(cpus_in_cluster, cpu)); in init_x2apic_ldr()
143 cpumask_set_cpu(cpu, per_cpu(cpus_in_cluster, this_cpu)); in init_x2apic_ldr()
153 unsigned int this_cpu = (unsigned long)hcpu; in update_clusterinfo() local
[all …]
Dipi.c42 unsigned int this_cpu = smp_processor_id(); in default_send_IPI_mask_allbutself_phys() local
50 if (query_cpu == this_cpu) in default_send_IPI_mask_allbutself_phys()
85 unsigned int this_cpu = smp_processor_id(); in default_send_IPI_mask_allbutself_logical() local
91 if (query_cpu == this_cpu) in default_send_IPI_mask_allbutself_logical()
Dx2apic_phys.c43 unsigned long this_cpu; in __x2apic_send_IPI_mask() local
50 this_cpu = smp_processor_id(); in __x2apic_send_IPI_mask()
52 if (apic_dest == APIC_DEST_ALLBUT && this_cpu == query_cpu) in __x2apic_send_IPI_mask()
Dapic_numachip.c133 unsigned int this_cpu = smp_processor_id(); in numachip_send_IPI_mask_allbutself() local
137 if (cpu != this_cpu) in numachip_send_IPI_mask_allbutself()
144 unsigned int this_cpu = smp_processor_id(); in numachip_send_IPI_allbutself() local
148 if (cpu != this_cpu) in numachip_send_IPI_allbutself()
Dx2apic_uv_x.c274 unsigned int this_cpu = smp_processor_id(); in uv_send_IPI_mask_allbutself() local
278 if (cpu != this_cpu) in uv_send_IPI_mask_allbutself()
285 unsigned int this_cpu = smp_processor_id(); in uv_send_IPI_allbutself() local
289 if (cpu != this_cpu) in uv_send_IPI_allbutself()
/linux-4.4.14/net/netfilter/
Dnft_counter.c39 struct nft_counter_percpu *this_cpu; in nft_counter_eval() local
42 this_cpu = this_cpu_ptr(priv->counter); in nft_counter_eval()
43 u64_stats_update_begin(&this_cpu->syncp); in nft_counter_eval()
44 this_cpu->counter.bytes += pkt->skb->len; in nft_counter_eval()
45 this_cpu->counter.packets++; in nft_counter_eval()
46 u64_stats_update_end(&this_cpu->syncp); in nft_counter_eval()
99 struct nft_counter_percpu *this_cpu; in nft_counter_init() local
106 this_cpu = this_cpu_ptr(cpu_stats); in nft_counter_init()
108 this_cpu->counter.packets = in nft_counter_init()
112 this_cpu->counter.bytes = in nft_counter_init()
[all …]
/linux-4.4.14/arch/parisc/kernel/
Dsmp.c123 int this_cpu = smp_processor_id(); in ipi_interrupt() local
124 struct cpuinfo_parisc *p = &per_cpu(cpu_data, this_cpu); in ipi_interrupt()
129 spinlock_t *lock = &per_cpu(ipi_lock, this_cpu); in ipi_interrupt()
147 smp_debug(100, KERN_DEBUG "CPU%d IPI_NOP\n", this_cpu); in ipi_interrupt()
151 smp_debug(100, KERN_DEBUG "CPU%d IPI_RESCHEDULE\n", this_cpu); in ipi_interrupt()
157 smp_debug(100, KERN_DEBUG "CPU%d IPI_CALL_FUNC\n", this_cpu); in ipi_interrupt()
162 smp_debug(100, KERN_DEBUG "CPU%d IPI_CPU_START\n", this_cpu); in ipi_interrupt()
166 smp_debug(100, KERN_DEBUG "CPU%d IPI_CPU_STOP\n", this_cpu); in ipi_interrupt()
171 smp_debug(100, KERN_DEBUG "CPU%d is alive!\n", this_cpu); in ipi_interrupt()
176 this_cpu, which); in ipi_interrupt()
/linux-4.4.14/lib/
Dsmp_processor_id.c13 int this_cpu = raw_smp_processor_id(); in check_preemption_disabled() local
25 if (cpumask_equal(tsk_cpus_allowed(current), cpumask_of(this_cpu))) in check_preemption_disabled()
51 return this_cpu; in check_preemption_disabled()
Dnmi_backtrace.c56 int i, cpu, this_cpu = get_cpu(); in nmi_trigger_all_cpu_backtrace() local
69 cpumask_clear_cpu(this_cpu, to_cpumask(backtrace_mask)); in nmi_trigger_all_cpu_backtrace()
/linux-4.4.14/arch/x86/kernel/cpu/
Dcommon.c93 static const struct cpu_dev *this_cpu = &default_cpu; variable
354 if (!this_cpu) in table_lookup_model()
357 info = this_cpu->legacy_models; in table_lookup_model()
456 if (this_cpu->legacy_cache_size) in cpu_detect_cache_sizes()
457 l2size = this_cpu->legacy_cache_size(c, l2size); in cpu_detect_cache_sizes()
480 if (this_cpu->c_detect_tlb) in cpu_detect_tlb()
481 this_cpu->c_detect_tlb(c); in cpu_detect_tlb()
556 this_cpu = cpu_devs[i]; in get_cpu_vendor()
557 c->x86_vendor = this_cpu->c_x86_vendor; in get_cpu_vendor()
567 this_cpu = &default_cpu; in get_cpu_vendor()
[all …]
/linux-4.4.14/arch/alpha/kernel/
Dsmp.c525 int this_cpu = smp_processor_id(); in handle_ipi() local
526 unsigned long *pending_ipis = &ipi_data[this_cpu].bits; in handle_ipi()
531 this_cpu, *pending_ipis, regs->pc)); in handle_ipi()
558 this_cpu, which); in handle_ipi()
566 cpu_data[this_cpu].ipi_count++; in handle_ipi()
657 int cpu, this_cpu = smp_processor_id(); in flush_tlb_mm() local
659 if (!cpu_online(cpu) || cpu == this_cpu) in flush_tlb_mm()
706 int cpu, this_cpu = smp_processor_id(); in flush_tlb_page() local
708 if (!cpu_online(cpu) || cpu == this_cpu) in flush_tlb_page()
762 int cpu, this_cpu = smp_processor_id(); in flush_icache_user_range() local
[all …]
/linux-4.4.14/arch/x86/lib/
Dmsr-smp.c10 int this_cpu = raw_smp_processor_id(); in __rdmsr_on_cpu() local
13 reg = per_cpu_ptr(rv->msrs, this_cpu); in __rdmsr_on_cpu()
24 int this_cpu = raw_smp_processor_id(); in __wrmsr_on_cpu() local
27 reg = per_cpu_ptr(rv->msrs, this_cpu); in __wrmsr_on_cpu()
102 int this_cpu; in __rwmsr_on_cpus() local
109 this_cpu = get_cpu(); in __rwmsr_on_cpus()
111 if (cpumask_test_cpu(this_cpu, mask)) in __rwmsr_on_cpus()
/linux-4.4.14/kernel/trace/
Dtrace_clock.c95 int this_cpu; in trace_clock_global() local
100 this_cpu = raw_smp_processor_id(); in trace_clock_global()
101 now = sched_clock_cpu(this_cpu); in trace_clock_global()
/linux-4.4.14/arch/sparc/kernel/
Dprocess_64.c195 int this_cpu) in __global_reg_self() argument
201 rp = &global_cpu_snapshot[this_cpu].reg; in __global_reg_self()
247 int this_cpu, cpu; in arch_trigger_all_cpu_backtrace() local
254 this_cpu = raw_smp_processor_id(); in arch_trigger_all_cpu_backtrace()
259 __global_reg_self(tp, regs, this_cpu); in arch_trigger_all_cpu_backtrace()
266 if (!include_self && cpu == this_cpu) in arch_trigger_all_cpu_backtrace()
275 (cpu == this_cpu ? '*' : ' '), cpu, in arch_trigger_all_cpu_backtrace()
312 static void __global_pmu_self(int this_cpu) in __global_pmu_self() argument
320 pp = &global_cpu_snapshot[this_cpu].pmu; in __global_pmu_self()
346 int this_cpu, cpu; in pmu_snapshot_all_cpus() local
[all …]
Dsmp_64.c623 int retries, this_cpu, prev_sent, i, saw_cpu_error; in hypervisor_xcall_deliver() local
627 this_cpu = smp_processor_id(); in hypervisor_xcall_deliver()
710 this_cpu, saw_cpu_error - 1); in hypervisor_xcall_deliver()
716 this_cpu, retries); in hypervisor_xcall_deliver()
721 this_cpu, status); in hypervisor_xcall_deliver()
724 this_cpu, cnt, tb->cpu_list_pa, tb->cpu_mondo_block_pa); in hypervisor_xcall_deliver()
727 printk(KERN_CRIT "CPU[%d]: CPU list [ ", this_cpu); in hypervisor_xcall_deliver()
738 int this_cpu, i, cnt; in xcall_deliver() local
755 this_cpu = smp_processor_id(); in xcall_deliver()
756 tb = &trap_block[this_cpu]; in xcall_deliver()
[all …]
Dnmi.c71 int this_cpu = smp_processor_id(); in die_nmi() local
78 panic("Watchdog detected hard LOCKUP on cpu %d", this_cpu); in die_nmi()
80 WARN(1, "Watchdog detected hard LOCKUP on cpu %d", this_cpu); in die_nmi()
Dchmc.c591 unsigned long ret, this_cpu; in chmc_read_mcreg() local
595 this_cpu = real_hard_smp_processor_id(); in chmc_read_mcreg()
597 if (p->portid == this_cpu) { in chmc_read_mcreg()
Dentry.h246 void sun4v_register_mondo_queues(int this_cpu);
Dirq_64.c1000 void notrace sun4v_register_mondo_queues(int this_cpu) in sun4v_register_mondo_queues() argument
1002 struct trap_per_cpu *tb = &trap_block[this_cpu]; in sun4v_register_mondo_queues()
/linux-4.4.14/Documentation/arm/
Dvlocks.txt35 bool vlock_trylock(int this_cpu)
38 currently_voting[this_cpu] = 1;
41 currently_voting[this_cpu] = 0;
46 last_vote = this_cpu;
47 currently_voting[this_cpu] = 0;
56 if (last_vote == this_cpu)
99 my_town = towns[(this_cpu >> 4) & 0xf];
100 I_won = vlock_trylock(my_town, this_cpu & 0xf);
103 my_state = states[(this_cpu >> 8) & 0xf];
104 I_won = vlock_lock(my_state, this_cpu & 0xf));
[all …]
/linux-4.4.14/arch/s390/kernel/
Dmachine_kexec.c58 int cpu, this_cpu; in setup_regs() local
62 this_cpu = smp_find_processor_id(stap()); in setup_regs()
63 add_elf_notes(this_cpu); in setup_regs()
65 if (cpu == this_cpu) in setup_regs()
/linux-4.4.14/arch/blackfin/kernel/
Dnmi.c146 unsigned int this_cpu = smp_processor_id(); in check_nmi_wdt_touched() local
151 if (!atomic_read(&nmi_touched[this_cpu])) in check_nmi_wdt_touched()
154 atomic_set(&nmi_touched[this_cpu], 0); in check_nmi_wdt_touched()
156 cpumask_clear_cpu(this_cpu, &mask); in check_nmi_wdt_touched()
/linux-4.4.14/kernel/
Dsmp.c276 int this_cpu; in smp_call_function_single() local
283 this_cpu = get_cpu(); in smp_call_function_single()
291 WARN_ON_ONCE(cpu_online(this_cpu) && irqs_disabled() in smp_call_function_single()
408 int cpu, next_cpu, this_cpu = smp_processor_id(); in smp_call_function_many() local
416 WARN_ON_ONCE(cpu_online(this_cpu) && irqs_disabled() in smp_call_function_many()
421 if (cpu == this_cpu) in smp_call_function_many()
430 if (next_cpu == this_cpu) in smp_call_function_many()
442 cpumask_clear_cpu(this_cpu, cfd->cpumask); in smp_call_function_many()
Dwatchdog.c330 int this_cpu = smp_processor_id(); in watchdog_overflow_callback() local
337 pr_emerg("Watchdog detected hard LOCKUP on cpu %d", this_cpu); in watchdog_overflow_callback()
/linux-4.4.14/arch/arm/common/
DbL_switcher.c151 unsigned int mpidr, this_cpu, that_cpu; in bL_switch_to() local
157 this_cpu = smp_processor_id(); in bL_switch_to()
161 BUG_ON(cpu_logical_map(this_cpu) != ob_mpidr); in bL_switch_to()
166 that_cpu = bL_switcher_cpu_pairing[this_cpu]; in bL_switch_to()
172 this_cpu, ob_mpidr, ib_mpidr); in bL_switch_to()
174 this_cpu = smp_processor_id(); in bL_switch_to()
182 ipi_nr = register_ipi_completion(&inbound_alive, this_cpu); in bL_switch_to()
229 cpu_logical_map(this_cpu) = ib_mpidr; in bL_switch_to()
239 pr_debug("after switch: CPU %d MPIDR %#x\n", this_cpu, mpidr); in bL_switch_to()
/linux-4.4.14/Documentation/
Dthis_cpu_ops.txt1 this_cpu operations
4 this_cpu operations are a way of optimizing access to per cpu
10 this_cpu operations add a per cpu variable offset to the processor
37 The main use of the this_cpu operations has been to optimize counter
40 The following this_cpu() operations with implied preemption protection
61 Inner working of this_cpu operations
87 Consider the following this_cpu operation:
111 after the this_cpu instruction is executed. In general this means that
159 cpu variable. Most this_cpu operations take a cpu variable.
197 if we do not make use of this_cpu ops later to manipulate fields:
[all …]
Dlocal_ops.txt12 Please use the this_cpu operations instead unless there is really a special purpose.
13 Most uses of local_t in the kernel have been replaced by this_cpu operations.
14 this_cpu operations combine the relocation with the local_t like semantics in
D00-INDEX439 - List rationale behind and the way to use this_cpu operations.
/linux-4.4.14/init/
Dcalibrate.c278 int this_cpu = smp_processor_id(); in calibrate_delay() local
280 if (per_cpu(cpu_loops_per_jiffy, this_cpu)) { in calibrate_delay()
281 lpj = per_cpu(cpu_loops_per_jiffy, this_cpu); in calibrate_delay()
305 per_cpu(cpu_loops_per_jiffy, this_cpu) = lpj; in calibrate_delay()
/linux-4.4.14/arch/x86/kernel/cpu/mcheck/
Dtherm_throt.c154 unsigned int this_cpu = smp_processor_id(); in therm_throt_process() local
157 struct thermal_state *pstate = &per_cpu(thermal_state, this_cpu); in therm_throt_process()
194 this_cpu, in therm_throt_process()
202 this_cpu, in therm_throt_process()
213 unsigned int this_cpu = smp_processor_id(); in thresh_event_valid() local
214 struct thermal_state *pstate = &per_cpu(thermal_state, this_cpu); in thresh_event_valid()
/linux-4.4.14/arch/arm/kernel/
Dsmp_tlb.c126 int this_cpu; in broadcast_tlb_mm_a15_erratum() local
132 this_cpu = get_cpu(); in broadcast_tlb_mm_a15_erratum()
133 a15_erratum_get_cpumask(this_cpu, mm, &mask); in broadcast_tlb_mm_a15_erratum()
/linux-4.4.14/drivers/cpuidle/
Dcpuidle-ux500.c29 int this_cpu = smp_processor_id(); in ux500_enter_idle() local
52 if (!prcmu_is_cpu_in_wfi(this_cpu ? 0 : 1)) in ux500_enter_idle()
Dcoupled.c356 static void cpuidle_coupled_poke_others(int this_cpu, in cpuidle_coupled_poke_others() argument
362 if (cpu != this_cpu && cpu_online(cpu)) in cpuidle_coupled_poke_others()
/linux-4.4.14/arch/x86/kernel/
Dirq.c350 unsigned int this_cpu, vector, this_count, count; in check_irq_vectors_for_cpu_disable() local
355 this_cpu = smp_processor_id(); in check_irq_vectors_for_cpu_disable()
357 cpumask_clear_cpu(this_cpu, &online_new); in check_irq_vectors_for_cpu_disable()
372 cpumask_clear_cpu(this_cpu, &affinity_new); in check_irq_vectors_for_cpu_disable()
402 if (cpu == this_cpu) in check_irq_vectors_for_cpu_disable()
423 this_cpu, this_count, count); in check_irq_vectors_for_cpu_disable()
/linux-4.4.14/arch/arm/include/asm/
Dmmu_context.h32 void a15_erratum_get_cpumask(int this_cpu, struct mm_struct *mm,
35 static inline void a15_erratum_get_cpumask(int this_cpu, struct mm_struct *mm, in a15_erratum_get_cpumask() argument
/linux-4.4.14/arch/blackfin/mach-common/
Ddpmc.c87 unsigned int this_cpu = smp_processor_id(); in bfin_wakeup_cpu() local
91 cpumask_clear_cpu(this_cpu, &mask); in bfin_wakeup_cpu()
/linux-4.4.14/kernel/sched/
Ddeadline.c1331 int this_cpu = smp_processor_id(); in find_later_rq() local
1369 if (!cpumask_test_cpu(this_cpu, later_mask)) in find_later_rq()
1370 this_cpu = -1; in find_later_rq()
1380 if (this_cpu != -1 && in find_later_rq()
1381 cpumask_test_cpu(this_cpu, sched_domain_span(sd))) { in find_later_rq()
1383 return this_cpu; in find_later_rq()
1403 if (this_cpu != -1) in find_later_rq()
1404 return this_cpu; in find_later_rq()
1582 int this_cpu = this_rq->cpu, cpu; in pull_dl_task() local
1598 if (this_cpu == cpu) in pull_dl_task()
[all …]
Drt.c1563 int this_cpu = smp_processor_id(); in find_lowest_rq() local
1591 if (!cpumask_test_cpu(this_cpu, lowest_mask)) in find_lowest_rq()
1592 this_cpu = -1; /* Skip this_cpu opt if not among lowest */ in find_lowest_rq()
1603 if (this_cpu != -1 && in find_lowest_rq()
1604 cpumask_test_cpu(this_cpu, sched_domain_span(sd))) { in find_lowest_rq()
1606 return this_cpu; in find_lowest_rq()
1624 if (this_cpu != -1) in find_lowest_rq()
1625 return this_cpu; in find_lowest_rq()
1904 int this_cpu; in try_to_push_tasks() local
1907 this_cpu = rt_rq->push_cpu; in try_to_push_tasks()
[all …]
Dfair.c4665 int idx, this_cpu, prev_cpu; in wake_affine() local
4671 this_cpu = smp_processor_id(); in wake_affine()
4674 this_load = target_load(this_cpu, idx); in wake_affine()
4685 this_load += effective_load(tg, this_cpu, -weight, -weight); in wake_affine()
4705 prev_eff_load *= capacity_of(this_cpu); in wake_affine()
4709 effective_load(tg, this_cpu, weight, weight); in wake_affine()
4733 int this_cpu, int sd_flag) in find_idlest_group() argument
4753 local_group = cpumask_test_cpu(this_cpu, in find_idlest_group()
4789 find_idlest_cpu(struct sched_group *group, struct task_struct *p, int this_cpu) in find_idlest_cpu() argument
4794 int least_loaded_cpu = this_cpu; in find_idlest_cpu()
[all …]
Ddebug.c644 unsigned int this_cpu = raw_smp_processor_id(); in proc_sched_show_task() local
647 t0 = cpu_clock(this_cpu); in proc_sched_show_task()
648 t1 = cpu_clock(this_cpu); in proc_sched_show_task()
Dcore.c1670 int this_cpu = smp_processor_id(); in ttwu_stat() local
1672 if (cpu == this_cpu) { in ttwu_stat()
1680 for_each_domain(this_cpu, sd) { in ttwu_stat()
1883 bool cpus_share_cache(int this_cpu, int that_cpu) in cpus_share_cache() argument
1885 return per_cpu(sd_llc_id, this_cpu) == per_cpu(sd_llc_id, that_cpu); in cpus_share_cache()
/linux-4.4.14/arch/ia64/kernel/
Dsmp.c99 int this_cpu = get_cpu(); in handle_IPI() local
129 this_cpu, which); in handle_IPI()
Dperfmon.c5263 int this_cpu = smp_processor_id(); in pfm_overflow_handler() local
5298 pfm_stats[this_cpu].pfm_smpl_handler_calls++; in pfm_overflow_handler()
5321 pfm_stats[this_cpu].pfm_smpl_handler_cycles += end_cycles - start_cycles; in pfm_overflow_handler()
5463 int this_cpu = smp_processor_id(); in pfm_do_interrupt_handler() local
5466 pfm_stats[this_cpu].pfm_ovfl_intr_count++; in pfm_do_interrupt_handler()
5498 pfm_stats[this_cpu].pfm_spurious_ovfl_intr_count++; in pfm_do_interrupt_handler()
5510 this_cpu, task_pid_nr(task)); in pfm_do_interrupt_handler()
5515 this_cpu, in pfm_do_interrupt_handler()
5526 int this_cpu; in pfm_interrupt_handler() local
5530 this_cpu = get_cpu(); in pfm_interrupt_handler()
[all …]
Dprocess.c215 unsigned int this_cpu = smp_processor_id(); in play_dead() local
223 ia64_jump_to_sal(&sal_boot_rendez_state[this_cpu]); in play_dead()
/linux-4.4.14/arch/arm/mm/
Dcontext.c54 void a15_erratum_get_cpumask(int this_cpu, struct mm_struct *mm, in a15_erratum_get_cpumask() argument
64 if (cpu == this_cpu) in a15_erratum_get_cpumask()
/linux-4.4.14/tools/perf/
Dbuiltin-sched.c1350 int cpu, this_cpu = sample->cpu; in map_switch_event() local
1352 BUG_ON(this_cpu >= MAX_CPUS || this_cpu < 0); in map_switch_event()
1354 if (this_cpu > sched->max_cpu) in map_switch_event()
1355 sched->max_cpu = this_cpu; in map_switch_event()
1357 timestamp0 = sched->cpu_last_switched[this_cpu]; in map_switch_event()
1358 sched->cpu_last_switched[this_cpu] = timestamp; in map_switch_event()
1373 sched->curr_thread[this_cpu] = thread__get(sched_in); in map_switch_event()
1404 if (cpu != this_cpu) in map_switch_event()
1434 int this_cpu = sample->cpu, err = 0; in process_sched_switch_event() local
1438 if (sched->curr_pid[this_cpu] != (u32)-1) { in process_sched_switch_event()
[all …]
/linux-4.4.14/arch/sparc/mm/
Dinit_64.c224 static inline void set_dcache_dirty(struct page *page, int this_cpu) in set_dcache_dirty() argument
226 unsigned long mask = this_cpu; in set_dcache_dirty()
293 int this_cpu = get_cpu(); in flush_dcache() local
298 if (cpu == this_cpu) in flush_dcache()
363 int this_cpu; in flush_dcache_page() local
375 this_cpu = get_cpu(); in flush_dcache_page()
383 if (dirty_cpu == this_cpu) in flush_dcache_page()
387 set_dcache_dirty(page, this_cpu); in flush_dcache_page()
/linux-4.4.14/drivers/cpufreq/
Dacpi-cpufreq.c327 int this_cpu; in drv_write() local
329 this_cpu = get_cpu(); in drv_write()
330 if (cpumask_test_cpu(this_cpu, cmd->mask)) in drv_write()
/linux-4.4.14/arch/x86/xen/
Dsmp.c682 unsigned int this_cpu = smp_processor_id(); in xen_send_IPI_mask_allbutself() local
689 if (this_cpu == cpu) in xen_send_IPI_mask_allbutself()
/linux-4.4.14/kernel/printk/
Dprintk.c1669 int this_cpu; in vprintk_emit() local
1685 this_cpu = smp_processor_id(); in vprintk_emit()
1690 if (unlikely(logbuf_cpu == this_cpu)) { in vprintk_emit()
1708 logbuf_cpu = this_cpu; in vprintk_emit()
/linux-4.4.14/tools/perf/bench/
Dnuma.c1154 int this_cpu; in worker_thread() local
1163 this_cpu = g->threads[task_nr].curr_cpu; in worker_thread()
1164 if (this_cpu < g->p.nr_cpus/2) in worker_thread()
/linux-4.4.14/fs/
Deventpoll.c508 int this_cpu = get_cpu(); in ep_poll_safewake() local
511 ep_poll_wakeup_proc, NULL, wq, (void *) (long) this_cpu); in ep_poll_safewake()
/linux-4.4.14/tools/power/x86/turbostat/
Dturbostat.c1519 int this_cpu; in get_cpu_position_in_core() local
1533 fscanf(filep, "%d", &this_cpu); in get_cpu_position_in_core()
1534 if (this_cpu == cpu) { in get_cpu_position_in_core()
/linux-4.4.14/include/linux/
Dsched.h1120 bool cpus_share_cache(int this_cpu, int that_cpu);
1163 static inline bool cpus_share_cache(int this_cpu, int that_cpu) in cpus_share_cache() argument
/linux-4.4.14/drivers/gpu/drm/i915/
Di915_gem.c1173 unsigned this_cpu; in busywait_stop() local
1175 if (time_after(local_clock_us(&this_cpu), timeout)) in busywait_stop()
1178 return this_cpu != cpu; in busywait_stop()
/linux-4.4.14/drivers/net/ethernet/cavium/liquidio/
Dlio_main.c1858 int this_cpu = smp_processor_id(); in liquidio_napi_drv_callback() local
1860 if (droq->cpu_id == this_cpu) { in liquidio_napi_drv_callback()