/linux-4.1.27/kernel/ |
D | irq_work.c | 98 if (llist_add(&work->llnode, this_cpu_ptr(&lazy_list)) && in irq_work_queue() 102 if (llist_add(&work->llnode, this_cpu_ptr(&raised_list))) in irq_work_queue() 116 raised = this_cpu_ptr(&raised_list); in irq_work_needs_cpu() 117 lazy = this_cpu_ptr(&lazy_list); in irq_work_needs_cpu() 171 irq_work_run_list(this_cpu_ptr(&raised_list)); in irq_work_run() 172 irq_work_run_list(this_cpu_ptr(&lazy_list)); in irq_work_run() 178 struct llist_head *raised = this_cpu_ptr(&raised_list); in irq_work_tick() 182 irq_work_run_list(this_cpu_ptr(&lazy_list)); in irq_work_tick()
|
D | user-return-notifier.c | 17 hlist_add_head(&urn->link, this_cpu_ptr(&return_notifier_list)); in user_return_notifier_register() 28 if (hlist_empty(this_cpu_ptr(&return_notifier_list))) in user_return_notifier_unregister()
|
D | smp.c | 221 head = this_cpu_ptr(&call_single_queue); in flush_smp_call_function_queue() 296 csd = this_cpu_ptr(&csd_data); in smp_call_function_single() 439 cfd = this_cpu_ptr(&cfd_data); in smp_call_function_many()
|
D | softirq.c | 492 __this_cpu_write(tasklet_vec.tail, this_cpu_ptr(&tasklet_vec.head)); in tasklet_action() 528 __this_cpu_write(tasklet_hi_vec.tail, this_cpu_ptr(&tasklet_hi_vec.head)); in tasklet_hi_action()
|
/linux-4.1.27/arch/x86/kernel/cpu/mcheck/ |
D | mce_intel.c | 103 if (machine_check_poll(MCP_TIMESTAMP, this_cpu_ptr(&mce_banks_owned))) in mce_intel_cmci_poll() 126 owned = this_cpu_ptr(mce_banks_owned); in cmci_toggle_interrupt_mode() 223 machine_check_poll(MCP_TIMESTAMP, this_cpu_ptr(&mce_banks_owned)); in intel_threshold_interrupt() 234 unsigned long *owned = (void *)this_cpu_ptr(&mce_banks_owned); in cmci_discover() 256 __clear_bit(i, this_cpu_ptr(mce_poll_banks)); in cmci_discover() 280 __clear_bit(i, this_cpu_ptr(mce_poll_banks)); in cmci_discover() 291 WARN_ON(!test_bit(i, this_cpu_ptr(mce_poll_banks))); in cmci_discover() 316 machine_check_poll(MCP_TIMESTAMP, this_cpu_ptr(&mce_banks_owned)); in cmci_recheck() 325 if (!test_bit(bank, this_cpu_ptr(mce_banks_owned))) in __cmci_disable_bank() 330 __clear_bit(bank, this_cpu_ptr(mce_banks_owned)); in __cmci_disable_bank()
|
D | mce-inject.c | 86 struct mce *m = this_cpu_ptr(&injectm); in mce_raise_notify() 100 struct mce *m = this_cpu_ptr(&injectm); in mce_irq_ipi() 112 struct mce *m = this_cpu_ptr(&injectm); in raise_local()
|
D | mce.c | 400 return *(u64 *)((char *)this_cpu_ptr(&injectm) + offset); in mce_rdmsrl() 422 *(u64 *)((char *)this_cpu_ptr(&injectm) + offset) = v; in mce_wrmsrl() 478 struct mce_ring *r = this_cpu_ptr(&mce_ring); in mce_ring_empty() 490 r = this_cpu_ptr(&mce_ring); in mce_ring_get() 504 struct mce_ring *r = this_cpu_ptr(&mce_ring); in mce_ring_add() 526 schedule_work(this_cpu_ptr(&mce_work)); in mce_schedule_work() 551 irq_work_queue(this_cpu_ptr(&mce_irq_work)); in mce_report_event() 1063 final = this_cpu_ptr(&mces_seen); in do_machine_check() 1299 struct timer_list *t = this_cpu_ptr(&mce_timer); in mce_timer_fn() 1307 if (mce_available(this_cpu_ptr(&cpu_info))) { in mce_timer_fn() [all …]
|
/linux-4.1.27/arch/powerpc/include/asm/ |
D | xics.h | 101 struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr); in xics_push_cppr() 114 struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr); in xics_pop_cppr() 124 struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr); in xics_set_base_cppr() 136 struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr); in xics_cppr_top()
|
D | tlbflush.h | 110 struct ppc64_tlb_batch *batch = this_cpu_ptr(&ppc64_tlb_batch); in arch_enter_lazy_mmu_mode() 117 struct ppc64_tlb_batch *batch = this_cpu_ptr(&ppc64_tlb_batch); in arch_leave_lazy_mmu_mode()
|
/linux-4.1.27/arch/s390/kernel/ |
D | perf_cpum_cf.c | 176 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in cpumf_pmu_enable() 199 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in cpumf_pmu_disable() 233 cpuhw = this_cpu_ptr(&cpu_hw_events); in cpumf_measurement_alert() 253 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in setup_pmc_cpu() 478 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in cpumf_pmu_start() 509 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in cpumf_pmu_stop() 530 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in cpumf_pmu_add() 554 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in cpumf_pmu_del() 578 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in cpumf_pmu_start_txn() 592 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in cpumf_pmu_cancel_txn() [all …]
|
D | nmi.c | 58 mcck = *this_cpu_ptr(&cpu_mcck); in s390_handle_mcck() 59 memset(this_cpu_ptr(&cpu_mcck), 0, sizeof(mcck)); in s390_handle_mcck() 253 mcck = this_cpu_ptr(&cpu_mcck); in s390_do_machine_check()
|
D | perf_cpum_sf.c | 565 struct cpu_hw_sf *cpusf = this_cpu_ptr(&cpu_hw_sf); in setup_pmc_cpu() 852 struct cpu_hw_sf *cpuhw = this_cpu_ptr(&cpu_hw_sf); in cpumsf_pmu_enable() 901 struct cpu_hw_sf *cpuhw = this_cpu_ptr(&cpu_hw_sf); in cpumsf_pmu_disable() 1309 struct cpu_hw_sf *cpuhw = this_cpu_ptr(&cpu_hw_sf); in cpumsf_pmu_start() 1330 struct cpu_hw_sf *cpuhw = this_cpu_ptr(&cpu_hw_sf); in cpumsf_pmu_stop() 1349 struct cpu_hw_sf *cpuhw = this_cpu_ptr(&cpu_hw_sf); in cpumsf_pmu_add() 1399 struct cpu_hw_sf *cpuhw = this_cpu_ptr(&cpu_hw_sf); in cpumsf_pmu_del() 1466 cpuhw = this_cpu_ptr(&cpu_hw_sf); in cpumf_measurement_alert()
|
D | processor.c | 34 struct cpuid *id = this_cpu_ptr(&cpu_id); in cpu_init()
|
D | idle.c | 24 struct s390_idle_data *idle = this_cpu_ptr(&s390_idle); in enabled_wait()
|
D | vtime.c | 92 cycles_old = this_cpu_ptr(mt_cycles); in do_account_vtime()
|
D | time.c | 96 cd = this_cpu_ptr(&comparators); in clock_comparator_work() 377 atomic_t *sw_ptr = this_cpu_ptr(&clock_sync_word); in disable_sync_clock() 394 atomic_t *sw_ptr = this_cpu_ptr(&clock_sync_word); in enable_sync_clock()
|
/linux-4.1.27/block/ |
D | blk-softirq.c | 26 cpu_list = this_cpu_ptr(&blk_cpu_done); in blk_done_softirq() 47 list = this_cpu_ptr(&blk_cpu_done); in trigger_softirq() 93 this_cpu_ptr(&blk_cpu_done)); in blk_cpu_notify() 138 list = this_cpu_ptr(&blk_cpu_done); in __blk_complete_request()
|
D | blk-iopoll.c | 35 list_add_tail(&iop->list, this_cpu_ptr(&blk_cpu_iopoll)); in blk_iopoll_sched() 79 struct list_head *list = this_cpu_ptr(&blk_cpu_iopoll); in blk_iopoll_softirq() 201 this_cpu_ptr(&blk_cpu_iopoll)); in blk_iopoll_cpu_notify()
|
/linux-4.1.27/kernel/time/ |
D | tick-sched.c | 208 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in __tick_nohz_full_check() 238 irq_work_queue(this_cpu_ptr(&nohz_full_kick_work)); in tick_nohz_full_kick() 842 ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_idle_enter() 859 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_irq_exit() 874 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_get_sleep_length() 952 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_idle_exit() 986 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_handler() 1010 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_switch_to_nohz() 1072 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_irq_enter() 1150 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_setup_sched_timer() [all …]
|
D | tick-common.c | 227 struct tick_device *td = this_cpu_ptr(&tick_cpu_device); in tick_install_replacement() 388 struct tick_device *td = this_cpu_ptr(&tick_cpu_device); in tick_suspend_local() 402 struct tick_device *td = this_cpu_ptr(&tick_cpu_device); in tick_resume_local()
|
D | hrtimer.c | 569 struct hrtimer_cpu_base *cpu_base = this_cpu_ptr(&hrtimer_bases); in hrtimer_reprogram() 649 struct hrtimer_cpu_base *base = this_cpu_ptr(&hrtimer_bases); in retrigger_next_event() 923 reprogram = base->cpu_base == this_cpu_ptr(&hrtimer_bases); in remove_hrtimer() 983 } else if (new_base->cpu_base == this_cpu_ptr(&hrtimer_bases) && in __hrtimer_start_range_ns() 1123 struct hrtimer_cpu_base *cpu_base = this_cpu_ptr(&hrtimer_bases); in hrtimer_get_next_event() 1247 struct hrtimer_cpu_base *cpu_base = this_cpu_ptr(&hrtimer_bases); in hrtimer_interrupt() 1373 td = this_cpu_ptr(&tick_cpu_device); in __hrtimer_peek_ahead_timers() 1437 struct hrtimer_cpu_base *cpu_base = this_cpu_ptr(&hrtimer_bases); in hrtimer_run_queues() 1676 new_base = this_cpu_ptr(&hrtimer_bases); in migrate_hrtimers()
|
D | tick-broadcast.c | 241 struct tick_device *td = this_cpu_ptr(&tick_cpu_device); in tick_receive_broadcast() 346 td = this_cpu_ptr(&tick_cpu_device); in tick_broadcast_control() 561 struct tick_device *td = this_cpu_ptr(&tick_cpu_device); in tick_check_oneshot_broadcast_this_cpu() 705 td = this_cpu_ptr(&tick_cpu_device); in tick_broadcast_oneshot_control()
|
D | tick-oneshot.c | 62 struct tick_device *td = this_cpu_ptr(&tick_cpu_device); in tick_switch_to_oneshot()
|
/linux-4.1.27/arch/x86/kernel/cpu/ |
D | perf_event_intel_bts.c | 221 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in __bts_event_start() 250 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in bts_event_start() 274 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in bts_event_stop() 287 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in intel_bts_enable_local() 295 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in intel_bts_disable_local() 381 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in intel_bts_interrupt() 422 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bts_event_del() 423 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in bts_event_del() 446 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in bts_event_add() 447 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bts_event_add()
|
D | perf_event_intel_ds.c | 478 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_disable_bts() 495 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_drain_bts_buffer() 689 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_pebs_enable() 704 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_pebs_disable() 722 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_pebs_enable_all() 730 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_pebs_disable_all() 738 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_pebs_fixup_ip() 860 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in __intel_pmu_pebs_event() 970 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_drain_pebs_core() 1011 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_drain_pebs_nhm()
|
D | perf_event_intel_lbr.c | 137 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in __intel_pmu_lbr_enable() 262 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_sched_task() 309 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_enable() 337 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_disable() 362 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_enable_all() 370 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_disable_all() 462 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_lbr_read()
|
D | perf_event_intel_pt.c | 866 struct pt *pt = this_cpu_ptr(&pt_ctx); in intel_pt_interrupt() 925 struct pt *pt = this_cpu_ptr(&pt_ctx); in pt_event_start() 944 struct pt *pt = this_cpu_ptr(&pt_ctx); in pt_event_stop() 959 struct pt *pt = this_cpu_ptr(&pt_ctx); in pt_event_stop() 978 struct pt *pt = this_cpu_ptr(&pt_ctx); in pt_event_del() 998 struct pt *pt = this_cpu_ptr(&pt_ctx); in pt_event_add()
|
D | perf_event_amd_ibs.c | 369 struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu); in perf_ibs_start() 389 struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu); in perf_ibs_stop() 423 struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu); in perf_ibs_add() 441 struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu); in perf_ibs_del() 518 struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu); in perf_ibs_handle_irq()
|
D | perf_event.c | 575 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in x86_pmu_disable_all() 593 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in x86_pmu_disable() 610 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in x86_pmu_enable_all() 998 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in x86_pmu_enable() 1152 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in x86_pmu_add() 1201 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in x86_pmu_start() 1286 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in x86_pmu_stop() 1308 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in x86_pmu_del() 1365 cpuc = this_cpu_ptr(&cpu_hw_events); in x86_pmu_handle_irq() 1776 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in x86_pmu_commit_txn()
|
D | perf_event_amd.c | 705 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_pmu_enable_virt() 717 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in amd_pmu_disable_virt()
|
D | perf_event_intel.c | 1251 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in __intel_pmu_disable_all() 1271 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in __intel_pmu_enable_all() 1311 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_nhm_workaround() 1410 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_disable_event() 1474 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_enable_event() 1580 cpuc = this_cpu_ptr(&cpu_hw_events); in intel_pmu_handle_irq() 2348 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in intel_guest_get_msrs() 2369 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in core_guest_get_msrs() 2403 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in core_pmu_enable_all()
|
D | perf_event_knc.c | 220 cpuc = this_cpu_ptr(&cpu_hw_events); in knc_pmu_handle_irq()
|
D | perf_event_p4.c | 918 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in p4_pmu_disable_all() 987 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in p4_pmu_enable_all() 1007 cpuc = this_cpu_ptr(&cpu_hw_events); in p4_pmu_handle_irq()
|
D | perf_event_intel_cqm.c | 972 struct intel_cqm_state *state = this_cpu_ptr(&cqm_state); in intel_cqm_event_start() 996 struct intel_cqm_state *state = this_cpu_ptr(&cqm_state); in intel_cqm_event_stop()
|
/linux-4.1.27/arch/arm64/kernel/ |
D | hw_breakpoint.c | 241 slots = this_cpu_ptr(bp_on_reg); in hw_breakpoint_control() 248 slots = this_cpu_ptr(wp_on_reg); in hw_breakpoint_control() 549 slots = this_cpu_ptr(bp_on_reg); in toggle_bp_registers() 553 slots = this_cpu_ptr(wp_on_reg); in toggle_bp_registers() 590 slots = this_cpu_ptr(bp_on_reg); in breakpoint_handler() 640 kernel_step = this_cpu_ptr(&stepping_kernel_bp); in breakpoint_handler() 667 slots = this_cpu_ptr(wp_on_reg); in watchpoint_handler() 742 kernel_step = this_cpu_ptr(&stepping_kernel_bp); in watchpoint_handler() 766 kernel_step = this_cpu_ptr(&stepping_kernel_bp); in reinstall_suspended_bps() 864 for (slots = this_cpu_ptr(bp_on_reg), i = 0; i < core_num_brps; ++i) { in hw_breakpoint_reset() [all …]
|
D | fpsimd.c | 231 struct fpsimd_partial_state *s = this_cpu_ptr( in kernel_neon_begin_partial() 255 struct fpsimd_partial_state *s = this_cpu_ptr( in kernel_neon_end()
|
D | cpuinfo.c | 245 struct cpuinfo_arm64 *info = this_cpu_ptr(&cpu_data); in cpuinfo_store_cpu()
|
D | perf_event.c | 1122 cpuc = this_cpu_ptr(&cpu_hw_events); in armv8pmu_handle_irq() 1376 return this_cpu_ptr(&cpu_hw_events); in armpmu_get_cpu_events()
|
/linux-4.1.27/arch/arm/kernel/ |
D | perf_event_xscale.c | 141 struct pmu_hw_events *cpuc = this_cpu_ptr(cpu_pmu->hw_events); in xscale1pmu_handle_irq() 201 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in xscale1pmu_enable_event() 237 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in xscale1pmu_disable_event() 290 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in xscale1pmu_start() 302 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in xscale1pmu_stop() 488 struct pmu_hw_events *cpuc = this_cpu_ptr(cpu_pmu->hw_events); in xscale2pmu_handle_irq() 542 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in xscale2pmu_enable_event() 588 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in xscale2pmu_disable_event() 654 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in xscale2pmu_start() 666 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in xscale2pmu_stop()
|
D | perf_event_v6.c | 265 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in armv6pmu_enable_event() 303 struct pmu_hw_events *cpuc = this_cpu_ptr(cpu_pmu->hw_events); in armv6pmu_handle_irq() 359 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in armv6pmu_start() 371 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in armv6pmu_stop() 412 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in armv6pmu_disable_event() 447 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in armv6mpcore_pmu_disable_event()
|
D | hw_breakpoint.c | 348 slots = this_cpu_ptr(bp_on_reg); in arch_install_hw_breakpoint() 354 slots = this_cpu_ptr(wp_on_reg); in arch_install_hw_breakpoint() 400 slots = this_cpu_ptr(bp_on_reg); in arch_uninstall_hw_breakpoint() 405 slots = this_cpu_ptr(wp_on_reg); in arch_uninstall_hw_breakpoint() 701 slots = this_cpu_ptr(wp_on_reg); in watchpoint_handler() 772 slots = this_cpu_ptr(wp_on_reg); in watchpoint_single_step_handler() 806 slots = this_cpu_ptr(bp_on_reg); in breakpoint_handler()
|
D | perf_event.c | 210 struct pmu_hw_events *hw_events = this_cpu_ptr(armpmu->hw_events); in armpmu_del() 227 struct pmu_hw_events *hw_events = this_cpu_ptr(armpmu->hw_events); in armpmu_add() 489 struct pmu_hw_events *hw_events = this_cpu_ptr(armpmu->hw_events); in armpmu_enable()
|
D | perf_event_v7.c | 764 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in armv7pmu_enable_event() 810 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in armv7pmu_disable_event() 842 struct pmu_hw_events *cpuc = this_cpu_ptr(cpu_pmu->hw_events); in armv7pmu_handle_irq() 902 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in armv7pmu_start() 913 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in armv7pmu_stop() 1347 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in krait_pmu_disable_event() 1373 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in krait_pmu_enable_event() 1680 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in scorpion_pmu_disable_event() 1706 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in scorpion_pmu_enable_event()
|
/linux-4.1.27/arch/ia64/include/asm/sn/ |
D | arch.h | 60 #define sn_hub_info this_cpu_ptr(&__sn_hub_info) 75 #define sn_cnodeid_to_nasid this_cpu_ptr(&__sn_cnodeid_to_nasid[0])
|
/linux-4.1.27/arch/powerpc/platforms/powernv/ |
D | opal-tracepoints.c | 51 depth = this_cpu_ptr(&opal_trace_depth); in __trace_opal_entry() 72 depth = this_cpu_ptr(&opal_trace_depth); in __trace_opal_exit()
|
D | subcore.c | 291 if (this_cpu_ptr(&split_state)->master) { in cpu_update_split_mode() 305 if (this_cpu_ptr(&split_state)->master) { in cpu_update_split_mode() 354 this_cpu_ptr(&split_state)->master = 1; in set_subcores_per_core()
|
/linux-4.1.27/drivers/clocksource/ |
D | mips-gic-timer.c | 87 gic_clockevent_cpu_init(this_cpu_ptr(&gic_clockevent_device)); in gic_cpu_notifier() 90 gic_clockevent_cpu_exit(this_cpu_ptr(&gic_clockevent_device)); in gic_cpu_notifier() 110 gic_clockevent_cpu_init(this_cpu_ptr(&gic_clockevent_device)); in gic_clockevent_init()
|
D | arm_global_timer.c | 230 gt_clockevents_init(this_cpu_ptr(gt_evt)); in gt_cpu_notify() 233 gt_clockevents_stop(this_cpu_ptr(gt_evt)); in gt_cpu_notify() 306 gt_clockevents_init(this_cpu_ptr(gt_evt)); in global_timer_of_register()
|
D | exynos_mct.c | 386 struct mct_clock_event_device *mevt = this_cpu_ptr(&percpu_mct_tick); in exynos4_tick_set_next_event() 396 struct mct_clock_event_device *mevt = this_cpu_ptr(&percpu_mct_tick); in exynos4_tick_set_mode() 506 mevt = this_cpu_ptr(&percpu_mct_tick); in exynos4_mct_cpu_notify() 510 mevt = this_cpu_ptr(&percpu_mct_tick); in exynos4_mct_cpu_notify() 525 struct mct_clock_event_device *mevt = this_cpu_ptr(&percpu_mct_tick); in exynos4_timer_resources()
|
D | timer-atlas7.c | 232 sirfsoc_local_timer_setup(this_cpu_ptr(sirfsoc_clockevent)); in sirfsoc_cpu_notify() 235 sirfsoc_local_timer_stop(this_cpu_ptr(sirfsoc_clockevent)); in sirfsoc_cpu_notify() 254 sirfsoc_local_timer_setup(this_cpu_ptr(sirfsoc_clockevent)); in sirfsoc_clockevent_init()
|
D | time-armada-370-xp.c | 213 armada_370_xp_timer_setup(this_cpu_ptr(armada_370_xp_evt)); in armada_370_xp_timer_cpu_notify() 216 armada_370_xp_timer_stop(this_cpu_ptr(armada_370_xp_evt)); in armada_370_xp_timer_cpu_notify() 310 armada_370_xp_timer_setup(this_cpu_ptr(armada_370_xp_evt)); in armada_370_xp_timer_common_init()
|
D | qcom-timer.c | 165 msm_local_timer_setup(this_cpu_ptr(msm_evt)); in msm_timer_cpu_notify() 168 msm_local_timer_stop(this_cpu_ptr(msm_evt)); in msm_timer_cpu_notify()
|
D | metag_generic.c | 93 struct clock_event_device *evt = this_cpu_ptr(&local_clockevent); in metag_timer_interrupt()
|
D | arm_arch_timer.c | 521 arch_timer_setup(this_cpu_ptr(arch_timer_evt)); in arch_timer_cpu_notify() 524 arch_timer_stop(this_cpu_ptr(arch_timer_evt)); in arch_timer_cpu_notify() 606 arch_timer_setup(this_cpu_ptr(arch_timer_evt)); in arch_timer_register()
|
/linux-4.1.27/arch/x86/xen/ |
D | multicalls.c | 57 struct mc_buffer *b = this_cpu_ptr(&mc_buffer); in xen_mc_flush() 134 struct mc_buffer *b = this_cpu_ptr(&mc_buffer); in __xen_mc_entry() 165 struct mc_buffer *b = this_cpu_ptr(&mc_buffer); in xen_mc_extend_args() 195 struct mc_buffer *b = this_cpu_ptr(&mc_buffer); in xen_mc_callback()
|
D | time.c | 83 state = this_cpu_ptr(&xen_runstate); in get_runstate_snapshot() 126 snap = this_cpu_ptr(&xen_runstate_snapshot); in do_stolen_accounting() 400 struct clock_event_device *evt = this_cpu_ptr(&xen_clock_events.evt); in xen_timer_interrupt() 455 clockevents_register_device(this_cpu_ptr(&xen_clock_events.evt)); in xen_setup_cpu_clockevents()
|
D | spinlock.c | 112 struct xen_lock_waiting *w = this_cpu_ptr(&lock_waiting); in xen_lock_spinning()
|
D | enlighten.c | 870 const struct desc_ptr *desc = this_cpu_ptr(&idt_desc); in xen_copy_trap_info() 887 memcpy(this_cpu_ptr(&idt_desc), desc, sizeof(idt_desc)); in xen_load_idt()
|
/linux-4.1.27/arch/x86/mm/kmemcheck/ |
D | kmemcheck.c | 143 struct kmemcheck_context *data = this_cpu_ptr(&kmemcheck_context); in kmemcheck_active() 151 struct kmemcheck_context *data = this_cpu_ptr(&kmemcheck_context); in kmemcheck_save_addr() 159 struct kmemcheck_context *data = this_cpu_ptr(&kmemcheck_context); in kmemcheck_show_all() 172 struct kmemcheck_context *data = this_cpu_ptr(&kmemcheck_context); in kmemcheck_hide_all() 188 struct kmemcheck_context *data = this_cpu_ptr(&kmemcheck_context); in kmemcheck_show() 229 struct kmemcheck_context *data = this_cpu_ptr(&kmemcheck_context); in kmemcheck_hide() 531 struct kmemcheck_context *data = this_cpu_ptr(&kmemcheck_context); in kmemcheck_access()
|
/linux-4.1.27/drivers/oprofile/ |
D | cpu_buffer.c | 48 struct oprofile_cpu_buffer *cpu_buf = this_cpu_ptr(&op_cpu_buffer); in oprofile_cpu_buffer_inc_smpl_lost() 300 struct oprofile_cpu_buffer *cpu_buf = this_cpu_ptr(&op_cpu_buffer); in __oprofile_add_ext_sample() 360 struct oprofile_cpu_buffer *cpu_buf = this_cpu_ptr(&op_cpu_buffer); in oprofile_write_reserve() 415 struct oprofile_cpu_buffer *cpu_buf = this_cpu_ptr(&op_cpu_buffer); in oprofile_add_pc() 421 struct oprofile_cpu_buffer *cpu_buf = this_cpu_ptr(&op_cpu_buffer); in oprofile_add_trace()
|
D | timer_int.c | 35 struct hrtimer *hrtimer = this_cpu_ptr(&oprofile_hrtimer); in __oprofile_hrtimer_start()
|
/linux-4.1.27/include/linux/ |
D | kernel_stat.h | 47 #define kstat_this_cpu this_cpu_ptr(&kstat) 48 #define kcpustat_this_cpu this_cpu_ptr(&kernel_cpustat)
|
D | percpu-defs.h | 233 #define this_cpu_ptr(ptr) \ macro 239 #define this_cpu_ptr(ptr) raw_cpu_ptr(ptr) macro 252 #define this_cpu_ptr(ptr) raw_cpu_ptr(ptr) macro 265 this_cpu_ptr(&var); \ 281 this_cpu_ptr(var); \
|
D | ftrace.h | 212 (*this_cpu_ptr(ops->disabled))--; in ftrace_function_local_enable() 229 (*this_cpu_ptr(ops->disabled))++; in ftrace_function_local_disable() 244 return *this_cpu_ptr(ops->disabled); in ftrace_function_local_disabled()
|
D | if_macvlan.h | 73 pcpu_stats = this_cpu_ptr(vlan->pcpu_stats); in macvlan_count_rx()
|
D | kprobes.h | 361 return this_cpu_ptr(&kprobe_ctlblk); in get_kprobe_ctlblk()
|
D | cpumask.h | 667 #define this_cpu_cpumask_var_ptr(x) this_cpu_ptr(x)
|
D | perf_event.h | 788 struct pt_regs *regs = this_cpu_ptr(&__perf_regs[0]); in perf_sw_event_sched()
|
/linux-4.1.27/arch/tile/kernel/ |
D | messaging.c | 31 HV_MsgState *state = this_cpu_ptr(&msg_state); in init_messaging() 70 HV_MsgState *state = this_cpu_ptr(&msg_state); in hv_message_intr()
|
D | time.c | 165 struct clock_event_device *evt = this_cpu_ptr(&tile_timer); in setup_tile_timer() 185 struct clock_event_device *evt = this_cpu_ptr(&tile_timer); in do_timer_interrupt()
|
D | perf_event.c | 593 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in tile_pmu_stop() 619 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in tile_pmu_start() 653 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in tile_pmu_add() 709 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in tile_pmu_del() 883 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in tile_pmu_handle_irq()
|
D | single_step.c | 743 unsigned long *ss_pc = this_cpu_ptr(&ss_saved_pc); in gx_singlestep_handle() 770 unsigned long *ss_pc = this_cpu_ptr(&ss_saved_pc); in single_step_once()
|
/linux-4.1.27/arch/x86/kernel/ |
D | irq_64.c | 55 irq_stack_top = (u64)this_cpu_ptr(irq_stack_union.irq_stack) + in stack_overflow_check() 61 oist = this_cpu_ptr(&orig_ist); in stack_overflow_check()
|
D | hw_breakpoint.c | 111 struct perf_event **slot = this_cpu_ptr(&bp_per_reg[i]); in arch_install_hw_breakpoint() 125 dr7 = this_cpu_ptr(&cpu_dr7); in arch_install_hw_breakpoint() 151 struct perf_event **slot = this_cpu_ptr(&bp_per_reg[i]); in arch_uninstall_hw_breakpoint() 162 dr7 = this_cpu_ptr(&cpu_dr7); in arch_uninstall_hw_breakpoint()
|
D | kvm.c | 329 if (__test_and_clear_bit(KVM_PV_EOI_BIT, this_cpu_ptr(&kvm_apic_eoi))) in kvm_guest_apic_eoi_write() 340 u64 pa = slow_virt_to_phys(this_cpu_ptr(&apf_reason)); in kvm_guest_cpu_init() 356 pa = slow_virt_to_phys(this_cpu_ptr(&kvm_apic_eoi)) in kvm_guest_cpu_init() 735 w = this_cpu_ptr(&klock_waiting); in kvm_lock_spinning()
|
D | apb_timer.c | 141 struct apbt_dev *adev = this_cpu_ptr(&cpu_apbt_dev); in apbt_clockevent_register() 193 adev = this_cpu_ptr(&cpu_apbt_dev); in apbt_setup_secondary_clock()
|
/linux-4.1.27/arch/sh/kernel/ |
D | perf_event.c | 222 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_stop() 240 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_start() 257 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_del() 267 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sh_pmu_add()
|
D | kprobes.c | 105 struct kprobe *saved = this_cpu_ptr(&saved_next_opcode); in arch_remove_kprobe() 114 saved = this_cpu_ptr(&saved_next_opcode2); in arch_remove_kprobe() 156 op1 = this_cpu_ptr(&saved_next_opcode); in prepare_singlestep() 157 op2 = this_cpu_ptr(&saved_next_opcode2); in prepare_singlestep() 386 p = this_cpu_ptr(&saved_next_opcode); in post_kprobe_handler() 398 p = this_cpu_ptr(&saved_next_opcode2); in post_kprobe_handler()
|
D | localtimer.c | 35 struct clock_event_device *clk = this_cpu_ptr(&local_clockevent); in local_timer_interrupt()
|
D | hw_breakpoint.c | 55 struct perf_event **slot = this_cpu_ptr(&bp_per_reg[i]); in arch_install_hw_breakpoint() 87 struct perf_event **slot = this_cpu_ptr(&bp_per_reg[i]); in arch_uninstall_hw_breakpoint()
|
/linux-4.1.27/arch/powerpc/kernel/ |
D | mce.c | 77 struct machine_check_event *mce = this_cpu_ptr(&mce_event[index]); in save_mce_event() 156 mc_evt = this_cpu_ptr(&mce_event[index]); in get_mce_event() 193 memcpy(this_cpu_ptr(&mce_event_queue[index]), &evt, sizeof(evt)); in machine_check_queue_event() 214 this_cpu_ptr(&mce_event_queue[index])); in machine_check_process_queued_event()
|
D | hw_breakpoint.c | 66 struct perf_event **slot = this_cpu_ptr(&bp_per_reg); in arch_install_hw_breakpoint() 91 struct perf_event **slot = this_cpu_ptr(&bp_per_reg); in arch_uninstall_hw_breakpoint()
|
D | time.c | 486 u64 *next_tb = this_cpu_ptr(&decrementers_next_tb); in __timer_interrupt() 487 struct clock_event_device *evt = this_cpu_ptr(&decrementers); in __timer_interrupt() 516 struct cpu_usage *cu = this_cpu_ptr(&cpu_usage_array); in __timer_interrupt() 531 u64 *next_tb = this_cpu_ptr(&decrementers_next_tb); in timer_interrupt() 875 u64 *next_tb = this_cpu_ptr(&decrementers_next_tb); in tick_broadcast_ipi_handler()
|
D | process.c | 499 memcpy(this_cpu_ptr(¤t_brk), brk, sizeof(*brk)); in __set_breakpoint() 860 if (unlikely(!hw_brk_match(this_cpu_ptr(¤t_brk), &new->thread.hw_brk))) in __switch_to() 874 struct cpu_usage *cu = this_cpu_ptr(&cpu_usage_array); in __switch_to() 884 batch = this_cpu_ptr(&ppc64_tlb_batch); in __switch_to() 907 batch = this_cpu_ptr(&ppc64_tlb_batch); in __switch_to()
|
D | kgdb.c | 158 this_cpu_ptr(&kgdb_thread_info); in kgdb_singlestep()
|
D | irq.c | 116 u64 *next_tb = this_cpu_ptr(&decrementers_next_tb); in decrementer_check_overflow()
|
D | smp.c | 246 struct cpu_messages *info = this_cpu_ptr(&ipi_message); in smp_ipi_demux()
|
/linux-4.1.27/arch/blackfin/kernel/ |
D | perf_event.c | 303 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bfin_pmu_stop() 321 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bfin_pmu_start() 338 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bfin_pmu_del() 348 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bfin_pmu_add() 424 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bfin_pmu_enable()
|
/linux-4.1.27/kernel/locking/ |
D | lglock.c | 25 lock = this_cpu_ptr(lg->lock); in lg_local_lock() 35 lock = this_cpu_ptr(lg->lock); in lg_local_unlock()
|
D | osq_lock.c | 86 struct optimistic_spin_node *node = this_cpu_ptr(&osq_node); in osq_lock() 193 node = this_cpu_ptr(&osq_node); in osq_unlock()
|
/linux-4.1.27/kernel/trace/ |
D | trace_event_perf.c | 250 list = this_cpu_ptr(pcpu_list); in perf_trace_add() 284 *regs = this_cpu_ptr(&__perf_regs[*rctxp]); in perf_trace_buf_prepare() 285 raw_data = this_cpu_ptr(perf_trace_buf[*rctxp]); in perf_trace_buf_prepare() 310 head = this_cpu_ptr(event_function.perf_events); in perf_ftrace_function_call()
|
D | trace_syscalls.c | 567 head = this_cpu_ptr(sys_data->enter_event->perf_events); in perf_syscall_enter() 641 head = this_cpu_ptr(sys_data->exit_event->perf_events); in perf_syscall_exit()
|
D | trace_branch.c | 57 data = this_cpu_ptr(tr->trace_buffer.data); in probe_likely_condition()
|
D | trace_kprobe.c | 1147 head = this_cpu_ptr(call->perf_events); in kprobe_perf_func() 1182 head = this_cpu_ptr(call->perf_events); in kretprobe_perf_func()
|
/linux-4.1.27/arch/alpha/kernel/ |
D | perf_event.c | 434 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_add() 486 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_del() 534 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_stop() 554 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_start() 727 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_enable() 753 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_pmu_disable() 818 cpuc = this_cpu_ptr(&cpu_hw_events); in alpha_perf_event_irq_handler()
|
/linux-4.1.27/arch/metag/kernel/perf/ |
D | perf_event.c | 261 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in metag_pmu_start() 309 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in metag_pmu_add() 351 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in metag_pmu_del() 600 struct cpu_hw_events *events = this_cpu_ptr(&cpu_hw_events); in metag_pmu_enable_counter() 673 struct cpu_hw_events *events = this_cpu_ptr(&cpu_hw_events); in metag_pmu_disable_counter() 721 struct cpu_hw_events *events = this_cpu_ptr(&cpu_hw_events); in metag_pmu_write_counter() 754 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in metag_pmu_counter_overflow()
|
/linux-4.1.27/arch/arc/kernel/ |
D | time.c | 202 struct clock_event_device *evt = this_cpu_ptr(&arc_clockevent_device); in timer_irq_handler() 222 struct clock_event_device *evt = this_cpu_ptr(&arc_clockevent_device); in arc_local_timer_setup()
|
D | smp.c | 310 *this_cpu_ptr(&ipi_data), smp_processor_id()); in do_IPI() 319 pending = xchg(this_cpu_ptr(&ipi_data), 0); in do_IPI()
|
/linux-4.1.27/arch/powerpc/platforms/pseries/ |
D | hvCall_inst.c | 113 h = this_cpu_ptr(&hcall_stats[opcode / 4]); in probe_hcall_entry() 126 h = this_cpu_ptr(&hcall_stats[opcode / 4]); in probe_hcall_exit()
|
D | ras.c | 308 memcpy(this_cpu_ptr(&mce_data_buf), h, sizeof(__u64)); in fwnmi_get_errinfo() 309 errhdr = (struct rtas_error_log *)this_cpu_ptr(&mce_data_buf); in fwnmi_get_errinfo()
|
D | lpar.c | 524 struct ppc64_tlb_batch *batch = this_cpu_ptr(&ppc64_tlb_batch); in pSeries_lpar_flush_hash_range() 714 depth = this_cpu_ptr(&hcall_trace_depth); in __trace_hcall_entry() 739 depth = this_cpu_ptr(&hcall_trace_depth); in __trace_hcall_exit()
|
D | dtl.c | 78 struct dtl_ring *dtlr = this_cpu_ptr(&dtl_rings); in consume_dtle()
|
/linux-4.1.27/drivers/cpuidle/governors/ |
D | menu.c | 286 struct menu_device *data = this_cpu_ptr(&menu_devices); in menu_select() 369 struct menu_device *data = this_cpu_ptr(&menu_devices); in menu_reflect() 382 struct menu_device *data = this_cpu_ptr(&menu_devices); in menu_update()
|
D | ladder.c | 69 struct ladder_device *ldev = this_cpu_ptr(&ladder_devices); in ladder_select_state() 168 struct ladder_device *ldev = this_cpu_ptr(&ladder_devices); in ladder_reflect()
|
/linux-4.1.27/arch/xtensa/kernel/ |
D | time.c | 150 setup_irq(this_cpu_ptr(&ccount_timer)->evt.irq, &timer_irqaction); in time_init() 161 struct clock_event_device *evt = &this_cpu_ptr(&ccount_timer)->evt; in timer_interrupt()
|
D | traps.c | 334 unsigned long excsave1 = (unsigned long)this_cpu_ptr(exc_table); in trap_init_excsave()
|
/linux-4.1.27/Documentation/ |
D | this_cpu_ops.txt | 127 y = this_cpu_ptr(&x) 131 processor. this_cpu_ptr avoids multiple steps that the common 163 &x and hence p is the *offset* of a per cpu variable. this_cpu_ptr() 196 The calculation of the pointer may require the use of this_cpu_ptr() 201 pp = this_cpu_ptr(&p); 250 &this_cpu_ptr(pp)->n vs this_cpu_ptr(&pp->n) 294 struct data *p = this_cpu_ptr(&datap); 296 but, passing of pointers calculated via this_cpu_ptr to other cpus is
|
D | local_ops.txt | 96 this_cpu_ptr() instead. 98 local_inc(this_cpu_ptr(&counters)); 142 local_inc(this_cpu_ptr(&counters));
|
/linux-4.1.27/lib/ |
D | percpu_ida.c | 152 tags = this_cpu_ptr(pool->tag_cpu); in percpu_ida_alloc() 200 tags = this_cpu_ptr(pool->tag_cpu); in percpu_ida_alloc() 225 tags = this_cpu_ptr(pool->tag_cpu); in percpu_ida_free()
|
D | radix-tree.c | 198 rtp = this_cpu_ptr(&radix_tree_preloads); in radix_tree_node_alloc() 259 rtp = this_cpu_ptr(&radix_tree_preloads); in __radix_tree_preload() 266 rtp = this_cpu_ptr(&radix_tree_preloads); in __radix_tree_preload()
|
/linux-4.1.27/arch/sparc/kernel/ |
D | perf_event.c | 1018 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sparc_pmu_enable() 1036 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sparc_pmu_disable() 1070 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sparc_pmu_start() 1085 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sparc_pmu_stop() 1101 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sparc_pmu_del() 1136 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sparc_pmu_read() 1148 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in perf_stop_nmi_watchdog() 1359 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in sparc_pmu_add() 1499 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in sparc_pmu_start_txn() 1512 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in sparc_pmu_cancel_txn() [all …]
|
D | pci_sun4v.c | 52 struct iommu_batch *p = this_cpu_ptr(&iommu_batch); in iommu_batch_start() 98 struct iommu_batch *p = this_cpu_ptr(&iommu_batch); in iommu_batch_new_entry() 110 struct iommu_batch *p = this_cpu_ptr(&iommu_batch); in iommu_batch_add() 124 struct iommu_batch *p = this_cpu_ptr(&iommu_batch); in iommu_batch_end()
|
D | sun4d_smp.c | 207 struct sun4d_ipi_work *work = this_cpu_ptr(&sun4d_ipi_work); in sun4d_ipi_interrupt()
|
D | leon_smp.c | 346 struct leon_ipi_work *work = this_cpu_ptr(&leon_ipi_work); in leonsmp_ipi_interrupt()
|
D | time_64.c | 765 sevt = this_cpu_ptr(&sparc64_events); in setup_sparc64_timer()
|
/linux-4.1.27/kernel/events/ |
D | callchain.c | 140 *rctx = get_recursion_context(this_cpu_ptr(callchain_recursion)); in get_callchain_entry() 156 put_recursion_context(this_cpu_ptr(callchain_recursion), rctx); in put_callchain_entry()
|
D | core.c | 340 return this_cpu_ptr(ctx->pmu->pmu_cpu_context); in __get_cpu_context() 411 info = this_cpu_ptr(cgrp->info); in __update_cgrp_time() 459 info = this_cpu_ptr(cgrp->info); in perf_cgroup_set_timestamp() 492 cpuctx = this_cpu_ptr(pmu->pmu_cpu_context); in perf_cgroup_switch() 788 cpuctx = this_cpu_ptr(pmu->pmu_cpu_context); in perf_cpu_hrtimer_cancel() 844 int *count = this_cpu_ptr(pmu->pmu_disable_count); in perf_pmu_disable() 851 int *count = this_cpu_ptr(pmu->pmu_disable_count); in perf_pmu_enable() 866 struct list_head *head = this_cpu_ptr(&active_ctx_list); in perf_event_ctx_activate() 2621 cpuctx = this_cpu_ptr(pmu->pmu_cpu_context); in perf_pmu_sched_task() 2670 if (atomic_read(this_cpu_ptr(&perf_cgroup_events))) in __perf_event_task_sched_out() [all …]
|
/linux-4.1.27/arch/sparc/include/asm/ |
D | cpudata_32.h | 29 #define local_cpu_data() (*this_cpu_ptr(&__cpu_data))
|
D | cpudata_64.h | 34 #define local_cpu_data() (*this_cpu_ptr(&__cpu_data))
|
/linux-4.1.27/net/can/ |
D | raw.c | 135 if (this_cpu_ptr(ro->uniq)->skb == oskb && in raw_rcv() 136 this_cpu_ptr(ro->uniq)->skbcnt == can_skb_prv(oskb)->skbcnt) { in raw_rcv() 140 if (this_cpu_ptr(ro->uniq)->join_rx_count < ro->count) in raw_rcv() 146 this_cpu_ptr(ro->uniq)->skb = oskb; in raw_rcv() 147 this_cpu_ptr(ro->uniq)->skbcnt = can_skb_prv(oskb)->skbcnt; in raw_rcv() 148 this_cpu_ptr(ro->uniq)->join_rx_count = 1; in raw_rcv()
|
/linux-4.1.27/kernel/rcu/ |
D | tree.c | 291 rdtp = this_cpu_ptr(&rcu_dynticks); in rcu_momentary_dyntick_idle() 585 struct rcu_dynticks *rdtp = this_cpu_ptr(&rcu_dynticks); in rcu_eqs_enter_common() 599 rdp = this_cpu_ptr(rsp->rda); in rcu_eqs_enter_common() 631 rdtp = this_cpu_ptr(&rcu_dynticks); in rcu_eqs_enter() 703 rdtp = this_cpu_ptr(&rcu_dynticks); in rcu_irq_exit() 724 struct rcu_dynticks *rdtp = this_cpu_ptr(&rcu_dynticks); in rcu_eqs_exit_common() 756 rdtp = this_cpu_ptr(&rcu_dynticks); in rcu_eqs_exit() 828 rdtp = this_cpu_ptr(&rcu_dynticks); in rcu_irq_enter() 851 struct rcu_dynticks *rdtp = this_cpu_ptr(&rcu_dynticks); in rcu_nmi_enter() 887 struct rcu_dynticks *rdtp = this_cpu_ptr(&rcu_dynticks); in rcu_nmi_exit() [all …]
|
D | tree_plugin.h | 153 rdp = this_cpu_ptr(rcu_preempt_state.rda); in rcu_preempt_note_context_switch() 503 rcu_do_batch(&rcu_preempt_state, this_cpu_ptr(&rcu_preempt_data)); in rcu_preempt_do_callbacks() 1201 rcu_do_batch(&rcu_sched_state, this_cpu_ptr(&rcu_sched_data)); in rcu_kthread_do_work() 1202 rcu_do_batch(&rcu_bh_state, this_cpu_ptr(&rcu_bh_data)); in rcu_kthread_do_work() 1231 unsigned int *statusp = this_cpu_ptr(&rcu_cpu_kthread_status); in rcu_cpu_kthread() 1232 char work, *workp = this_cpu_ptr(&rcu_cpu_has_work); in rcu_cpu_kthread() 1446 struct rcu_dynticks *rdtp = this_cpu_ptr(&rcu_dynticks); in rcu_try_advance_all_cbs() 1456 rdp = this_cpu_ptr(rsp->rda); in rcu_try_advance_all_cbs() 1486 struct rcu_dynticks *rdtp = this_cpu_ptr(&rcu_dynticks); in rcu_needs_cpu() 1531 struct rcu_dynticks *rdtp = this_cpu_ptr(&rcu_dynticks); in rcu_prepare_for_idle() [all …]
|
/linux-4.1.27/arch/powerpc/platforms/cell/ |
D | interrupt.c | 85 struct iic *iic = this_cpu_ptr(&cpu_iic); in iic_eoi() 151 iic = this_cpu_ptr(&cpu_iic); in iic_get_irq() 166 out_be64(&this_cpu_ptr(&cpu_iic)->regs->prio, 0xff); in iic_setup_cpu()
|
/linux-4.1.27/arch/powerpc/perf/ |
D | core-book3s.c | 351 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in power_pmu_bhrb_enable() 367 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in power_pmu_bhrb_disable() 1161 cpuhw = this_cpu_ptr(&cpu_hw_events); in power_pmu_disable() 1228 cpuhw = this_cpu_ptr(&cpu_hw_events); in power_pmu_enable() 1420 cpuhw = this_cpu_ptr(&cpu_hw_events); in power_pmu_add() 1486 cpuhw = this_cpu_ptr(&cpu_hw_events); in power_pmu_del() 1592 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in power_pmu_start_txn() 1606 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in power_pmu_cancel_txn() 1624 cpuhw = this_cpu_ptr(&cpu_hw_events); in power_pmu_commit_txn() 1983 cpuhw = this_cpu_ptr(&cpu_hw_events); in record_and_restart() [all …]
|
D | core-fsl-emb.c | 213 cpuhw = this_cpu_ptr(&cpu_hw_events); in fsl_emb_pmu_disable() 252 cpuhw = this_cpu_ptr(&cpu_hw_events); in fsl_emb_pmu_enable() 662 struct cpu_hw_events *cpuhw = this_cpu_ptr(&cpu_hw_events); in perf_event_interrupt()
|
/linux-4.1.27/arch/sparc/mm/ |
D | tlb.c | 55 struct tlb_batch *tb = this_cpu_ptr(&tlb_batch); in arch_enter_lazy_mmu_mode() 62 struct tlb_batch *tb = this_cpu_ptr(&tlb_batch); in arch_leave_lazy_mmu_mode()
|
/linux-4.1.27/arch/x86/oprofile/ |
D | nmi_int.c | 67 model->check_ctrs(regs, this_cpu_ptr(&cpu_msrs)); in profile_exceptions_notify() 71 model->stop(this_cpu_ptr(&cpu_msrs)); in profile_exceptions_notify() 94 struct op_msrs const *msrs = this_cpu_ptr(&cpu_msrs); in nmi_cpu_start() 114 struct op_msrs const *msrs = this_cpu_ptr(&cpu_msrs); in nmi_cpu_stop()
|
/linux-4.1.27/drivers/net/ethernet/tile/ |
D | tilegx.c | 425 struct tile_net_info *info = this_cpu_ptr(&per_cpu_info); in tile_net_provide_needed_buffers() 553 struct tile_net_info *info = this_cpu_ptr(&per_cpu_info); in tile_net_receive_skb() 587 struct tile_net_info *info = this_cpu_ptr(&per_cpu_info); in tile_net_handle_packet() 653 struct tile_net_info *info = this_cpu_ptr(&per_cpu_info); in tile_net_poll() 702 struct tile_net_info *info = this_cpu_ptr(&per_cpu_info); in tile_net_handle_ingress_irq() 765 struct tile_net_info *info = this_cpu_ptr(&per_cpu_info); in tile_net_schedule_egress_timer() 782 struct tile_net_info *info = this_cpu_ptr(&per_cpu_info); in tile_net_handle_egress_timer() 1930 struct tile_net_info *info = this_cpu_ptr(&per_cpu_info); in tile_net_tx_tso() 1999 struct tile_net_info *info = this_cpu_ptr(&per_cpu_info); in tile_net_tx() 2141 struct tile_net_info *info = this_cpu_ptr(&per_cpu_info); in tile_net_netpoll() [all …]
|
D | tilepro.c | 999 info = this_cpu_ptr(&hv_xgbe0); in tile_net_register() 1001 info = this_cpu_ptr(&hv_xgbe1); in tile_net_register() 1003 info = this_cpu_ptr(&hv_gbe0); in tile_net_register() 1005 info = this_cpu_ptr(&hv_gbe1); in tile_net_register()
|
/linux-4.1.27/arch/mips/cavium-octeon/ |
D | octeon-irq.c | 293 raw_spinlock_t *lock = this_cpu_ptr(&octeon_irq_ciu_spinlock); in octeon_irq_ciu_enable_local() 299 pen = this_cpu_ptr(&octeon_irq_ciu0_en_mirror); in octeon_irq_ciu_enable_local() 308 pen = this_cpu_ptr(&octeon_irq_ciu1_en_mirror); in octeon_irq_ciu_enable_local() 325 raw_spinlock_t *lock = this_cpu_ptr(&octeon_irq_ciu_spinlock); in octeon_irq_ciu_disable_local() 331 pen = this_cpu_ptr(&octeon_irq_ciu0_en_mirror); in octeon_irq_ciu_disable_local() 340 pen = this_cpu_ptr(&octeon_irq_ciu1_en_mirror); in octeon_irq_ciu_disable_local() 521 set_bit(cd->bit, this_cpu_ptr(&octeon_irq_ciu0_en_mirror)); in octeon_irq_ciu_enable_local_v2() 525 set_bit(cd->bit, this_cpu_ptr(&octeon_irq_ciu1_en_mirror)); in octeon_irq_ciu_enable_local_v2() 540 clear_bit(cd->bit, this_cpu_ptr(&octeon_irq_ciu0_en_mirror)); in octeon_irq_ciu_disable_local_v2() 544 clear_bit(cd->bit, this_cpu_ptr(&octeon_irq_ciu1_en_mirror)); in octeon_irq_ciu_disable_local_v2() [all …]
|
/linux-4.1.27/drivers/leds/trigger/ |
D | ledtrig-cpu.c | 50 struct led_trigger_cpu *trig = this_cpu_ptr(&cpu_trig); in ledtrig_cpu()
|
/linux-4.1.27/arch/powerpc/kvm/ |
D | e500.c | 83 entry->pentry = this_cpu_ptr(&pcpu_sids.entry[sid]); in local_sid_setup_one() 112 entry->pentry == this_cpu_ptr(&pcpu_sids.entry[entry->val])) in local_sid_lookup() 121 memset(this_cpu_ptr(&pcpu_sids), 0, sizeof(pcpu_sids)); in local_sid_destroy_all()
|
/linux-4.1.27/include/net/ |
D | gro_cells.h | 27 cell = this_cpu_ptr(gcells->cells); in gro_cells_receive()
|
D | sch_generic.h | 529 this_cpu_ptr(sch->cpu_bstats); in qdisc_bstats_update_cpu() 566 struct gnet_stats_queue *qstats = this_cpu_ptr(sch->cpu_qstats); in qdisc_qstats_drop_cpu()
|
/linux-4.1.27/net/core/ |
D | flow.c | 203 fcp = this_cpu_ptr(fc->percpu); in flow_cache_lookup() 293 fcp = this_cpu_ptr(fc->percpu); in flow_cache_flush_tasklet() 335 tasklet = &this_cpu_ptr(info->cache->percpu)->flush_tasklet; in flow_cache_flush_per_cpu()
|
D | drop_monitor.c | 149 data = this_cpu_ptr(&dm_cpu_data); in trace_drop_common()
|
D | dev.c | 2211 sd = this_cpu_ptr(&softnet_data); in __netif_reschedule() 3276 struct softnet_data *mysd = this_cpu_ptr(&softnet_data); in rps_ipi_queued() 3303 sd = this_cpu_ptr(&softnet_data); in skb_flow_limit() 3453 struct softnet_data *sd = this_cpu_ptr(&softnet_data); in net_tx_action() 3882 struct softnet_data *sd = this_cpu_ptr(&softnet_data); in flush_backlog() 4444 ____napi_schedule(this_cpu_ptr(&softnet_data), n); in __napi_schedule() 4457 ____napi_schedule(this_cpu_ptr(&softnet_data), n); in __napi_schedule_irqoff() 4681 struct softnet_data *sd = this_cpu_ptr(&softnet_data); in net_rx_action()
|
/linux-4.1.27/arch/mips/kernel/ |
D | perf_event_mipsxx.c | 343 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in mipsxx_pmu_enable_event() 363 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in mipsxx_pmu_disable_event() 463 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in mipspmu_add() 499 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in mipspmu_del() 1342 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in pause_local_counters() 1358 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in resume_local_counters() 1369 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in mipsxx_pmu_handle_shared_irq()
|
D | pm-cps.c | 124 cpumask_t *coupled_mask = this_cpu_ptr(&online_coupled); in cps_pm_enter_state()
|
/linux-4.1.27/arch/ia64/include/asm/uv/ |
D | uv_hub.h | 111 #define uv_hub_info this_cpu_ptr(&__uv_hub_info)
|
/linux-4.1.27/net/netfilter/ipvs/ |
D | ip_vs_core.c | 121 s = this_cpu_ptr(dest->stats.cpustats); in ip_vs_in_stats() 129 s = this_cpu_ptr(svc->stats.cpustats); in ip_vs_in_stats() 136 s = this_cpu_ptr(ipvs->tot_stats.cpustats); in ip_vs_in_stats() 155 s = this_cpu_ptr(dest->stats.cpustats); in ip_vs_out_stats() 163 s = this_cpu_ptr(svc->stats.cpustats); in ip_vs_out_stats() 170 s = this_cpu_ptr(ipvs->tot_stats.cpustats); in ip_vs_out_stats() 185 s = this_cpu_ptr(cp->dest->stats.cpustats); in ip_vs_conn_stats() 190 s = this_cpu_ptr(svc->stats.cpustats); in ip_vs_conn_stats() 195 s = this_cpu_ptr(ipvs->tot_stats.cpustats); in ip_vs_conn_stats()
|
/linux-4.1.27/net/xfrm/ |
D | xfrm_ipcomp.c | 149 scratch = *this_cpu_ptr(ipcomp_scratches); in ipcomp_compress() 150 tfm = *this_cpu_ptr(ipcd->tfms); in ipcomp_compress()
|
/linux-4.1.27/arch/x86/kernel/apic/ |
D | hw_nmi.c | 145 struct nmi_seq_buf *s = this_cpu_ptr(&nmi_print_seq); in nmi_vprintk()
|
D | apic.c | 541 struct clock_event_device *levt = this_cpu_ptr(&lapic_events); in setup_APIC_timer() 676 struct clock_event_device *levt = this_cpu_ptr(&lapic_events); in calibrate_APIC_clock()
|
/linux-4.1.27/drivers/net/ |
D | loopback.c | 87 lb_stats = this_cpu_ptr(dev->lstats); in loopback_xmit()
|
D | nlmon.c | 18 struct pcpu_lstats *stats = this_cpu_ptr(dev->lstats); in nlmon_xmit()
|
D | dummy.c | 81 struct pcpu_dstats *dstats = this_cpu_ptr(dev->dstats); in dummy_xmit()
|
D | veth.c | 122 struct pcpu_vstats *stats = this_cpu_ptr(dev->vstats); in veth_xmit()
|
/linux-4.1.27/net/openvswitch/ |
D | vport.c | 478 stats = this_cpu_ptr(vport->percpu_stats); in ovs_vport_receive() 513 stats = this_cpu_ptr(vport->percpu_stats); in ovs_vport_send()
|
D | actions.c | 101 fifo = this_cpu_ptr(action_fifos); in add_deferred_actions() 939 struct action_fifo *fifo = this_cpu_ptr(action_fifos); in process_deferred_actions()
|
/linux-4.1.27/arch/x86/crypto/sha-mb/ |
D | sha1_mb.c | 497 this_cpu_ptr(sha1_mb_alg_state.alg_cstate); in sha1_mb_update() 556 this_cpu_ptr(sha1_mb_alg_state.alg_cstate); in sha1_mb_finup() 616 this_cpu_ptr(sha1_mb_alg_state.alg_cstate); in sha1_mb_final()
|
/linux-4.1.27/drivers/s390/cio/ |
D | cio.c | 564 irb = this_cpu_ptr(&cio_irb); in do_cio_interrupt() 614 irb = this_cpu_ptr(&cio_irb); in cio_tsch() 752 tsch(ti.schid, this_cpu_ptr(&cio_irb)); in __clear_io_subchannel_easy()
|
D | ccwreq.c | 255 struct irb *irb = this_cpu_ptr(&cio_irb); in ccw_request_handler()
|
D | eadm_sch.c | 137 struct irb *irb = this_cpu_ptr(&cio_irb); in eadm_subchannel_irq()
|
D | device_fsm.c | 742 irb = this_cpu_ptr(&cio_irb); in ccw_device_irq() 808 irb = this_cpu_ptr(&cio_irb); in ccw_device_w4sense()
|
/linux-4.1.27/kernel/sched/ |
D | clock.c | 137 return this_cpu_ptr(&sched_clock_data); in this_scd()
|
D | cpuacct.c | 271 kcpustat = this_cpu_ptr(ca->cpustat); in cpuacct_account_field()
|
/linux-4.1.27/arch/parisc/mm/ |
D | fault.c | 152 d = this_cpu_ptr(&exception_data); in fixup_exception()
|
/linux-4.1.27/arch/tile/mm/ |
D | highmem.c | 106 amp = this_cpu_ptr(&s.per_type[type]); in kmap_atomic_register()
|
/linux-4.1.27/drivers/scsi/ |
D | scsi_logging.c | 39 buf = this_cpu_ptr(&scsi_format_log); in scsi_log_reserve_buffer() 62 buf = this_cpu_ptr(&scsi_format_log); in scsi_log_release_buffer()
|
/linux-4.1.27/arch/powerpc/sysdev/xics/ |
D | xics-common.c | 156 struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr); in xics_teardown_cpu()
|
/linux-4.1.27/arch/blackfin/mach-common/ |
D | smp.c | 149 bfin_ipi_data = this_cpu_ptr(&bfin_ipi); in ipi_handler_int1()
|
/linux-4.1.27/net/netfilter/ |
D | nf_tables_core.c | 209 stats = this_cpu_ptr(rcu_dereference(nft_base_chain(basechain)->stats)); in nft_do_chain()
|
/linux-4.1.27/drivers/net/team/ |
D | team_mode_loadbalance.c | 197 pcpu_stats = this_cpu_ptr(lb_priv->pcpu_stats); in lb_update_tx_stats() 198 port_stats = this_cpu_ptr(lb_port_priv->pcpu_stats); in lb_update_tx_stats()
|
/linux-4.1.27/net/ipv4/ |
D | ip_input.c | 347 struct ip_rt_acct *st = this_cpu_ptr(ip_rt_acct); in ip_rcv_finish()
|
D | syncookies.c | 63 tmp = this_cpu_ptr(ipv4_cookie_scratch); in cookie_hash()
|
D | ip_vti.c | 121 tstats = this_cpu_ptr(dev->tstats); in vti_rcv_cb()
|
D | icmp.c | 208 return *this_cpu_ptr(net->ipv4.icmp_sk); in icmp_sk()
|
/linux-4.1.27/net/bridge/ |
D | br_input.c | 33 struct pcpu_sw_netstats *brstats = this_cpu_ptr(br->stats); in br_pass_frame_up()
|
D | br_device.c | 40 struct pcpu_sw_netstats *brstats = this_cpu_ptr(br->stats); in br_dev_xmit()
|
D | br_netfilter.c | 824 data = this_cpu_ptr(&brnf_frag_data_storage); in br_nf_push_frag_xmit() 862 data = this_cpu_ptr(&brnf_frag_data_storage); in br_nf_dev_queue_xmit()
|
/linux-4.1.27/arch/x86/platform/uv/ |
D | uv_time.c | 368 struct clock_event_device *ced = this_cpu_ptr(&cpu_ced); in uv_rtc_register_clockevents()
|
/linux-4.1.27/arch/x86/include/asm/uv/ |
D | uv_hub.h | 167 #define uv_hub_info this_cpu_ptr(&__uv_hub_info)
|
/linux-4.1.27/net/8021q/ |
D | vlan_core.c | 53 rx_stats = this_cpu_ptr(vlan_dev_priv(vlan_dev)->vlan_pcpu_stats); in vlan_do_receive()
|
D | vlan_dev.c | 134 stats = this_cpu_ptr(vlan->vlan_pcpu_stats); in vlan_dev_hard_start_xmit()
|
/linux-4.1.27/net/ipv6/ |
D | syncookies.c | 71 tmp = this_cpu_ptr(ipv6_cookie_scratch); in cookie_hash()
|
D | ip6_vti.c | 372 tstats = this_cpu_ptr(dev->tstats); in vti6_rcv_cb() 487 struct pcpu_sw_netstats *tstats = this_cpu_ptr(dev->tstats); in vti6_xmit()
|
/linux-4.1.27/include/asm-generic/ |
D | percpu.h | 112 __ret = *this_cpu_ptr(&(pcp)); \
|
/linux-4.1.27/fs/gfs2/ |
D | lock_dlm.c | 83 lks = this_cpu_ptr(gl->gl_sbd->sd_lkstats); in gfs2_update_reply_times() 111 lks = this_cpu_ptr(gl->gl_sbd->sd_lkstats); in gfs2_update_request_times()
|
D | incore.h | 838 this_cpu_ptr(sdp->sd_lkstats)->lkstats[gl->gl_name.ln_type].stats[which]++; in gfs2_sbstats_inc()
|
/linux-4.1.27/mm/ |
D | vmstat.c | 512 drain_zone_pages(zone, this_cpu_ptr(&p->pcp)); in refresh_cpu_vm_stats() 1372 schedule_delayed_work(this_cpu_ptr(&vmstat_work), in vmstat_update()
|
D | page-writeback.c | 1598 p = this_cpu_ptr(&bdp_ratelimits); in balance_dirty_pages_ratelimited() 1610 p = this_cpu_ptr(&dirty_throttle_leaks); in balance_dirty_pages_ratelimited()
|
D | swap.c | 485 pvec = this_cpu_ptr(&lru_rotate_pvecs); in rotate_reclaimable_page()
|
/linux-4.1.27/arch/ia64/kernel/ |
D | traps.c | 302 struct fpu_swa_msg *cp = this_cpu_ptr(&cpulast); in handle_fpu_swa()
|
/linux-4.1.27/arch/parisc/lib/ |
D | memcpy.c | 473 d = this_cpu_ptr(&exception_data); in pa_memcpy()
|
/linux-4.1.27/arch/powerpc/platforms/ps3/ |
D | interrupt.c | 714 struct ps3_private *pd = this_cpu_ptr(&ps3_private); in ps3_get_irq()
|
/linux-4.1.27/arch/powerpc/mm/ |
D | hash_native_64.c | 645 struct ppc64_tlb_batch *batch = this_cpu_ptr(&ppc64_tlb_batch); in native_flush_hash_range()
|
D | hugetlbpage.c | 475 batchp = this_cpu_ptr(&hugepd_freelist_cur); in hugepd_free()
|
/linux-4.1.27/arch/mips/dec/ |
D | setup.c | 768 fpu_kstat_irq = this_cpu_ptr(desc_fpu->kstat_irqs); in arch_init_irq()
|
/linux-4.1.27/net/rds/ |
D | ib_rdma.c | 270 flag = this_cpu_ptr(&clean_list_grace); in rds_ib_reuse_fmr()
|
/linux-4.1.27/drivers/irqchip/ |
D | irq-gic-v3.c | 54 #define gic_data_rdist() (this_cpu_ptr(gic_data.rdists.rdist))
|
/linux-4.1.27/drivers/net/ipvlan/ |
D | ipvlan_core.c | 28 pcptr = this_cpu_ptr(ipvlan->pcpu_stats); in ipvlan_count_rx()
|
D | ipvlan_main.c | 187 pcptr = this_cpu_ptr(ipvlan->pcpu_stats); in ipvlan_start_xmit()
|
/linux-4.1.27/arch/arm/kvm/ |
D | arm.c | 288 vcpu->arch.host_cpu_context = this_cpu_ptr(kvm_host_cpu_state); in kvm_arch_vcpu_load()
|
/linux-4.1.27/fs/ |
D | aio.c | 881 kcpu = this_cpu_ptr(ctx->cpu); in put_reqs_available() 899 kcpu = this_cpu_ptr(ctx->cpu); in get_reqs_available()
|
/linux-4.1.27/crypto/ |
D | mcryptd.c | 107 cpu_queue = this_cpu_ptr(queue->cpu_queue); in mcryptd_enqueue_request()
|
/linux-4.1.27/arch/s390/oprofile/ |
D | hwsampler.c | 181 struct hws_cpu_buffer *cb = this_cpu_ptr(&sampler_cpu_buffer); in hws_ext_handler()
|
/linux-4.1.27/fs/fscache/ |
D | object.c | 805 wait_queue_head_t *cong_wq = this_cpu_ptr(&fscache_object_cong_wait); in fscache_object_sleep_till_congested()
|
/linux-4.1.27/include/trace/ |
D | ftrace.h | 816 head = this_cpu_ptr(event_call->perf_events); \
|
/linux-4.1.27/drivers/infiniband/hw/ehca/ |
D | ehca_irq.c | 793 struct ehca_cpu_comp_task *cct = this_cpu_ptr(pool->cpu_comp_tasks); in comp_task()
|