/linux-4.4.14/arch/x86/platform/uv/ |
D | uv_time.c | 62 int next_cpu; member 182 head->next_cpu = -1; in uv_rtc_allocate_timers() 199 head->next_cpu = -1; in uv_rtc_find_next_timer() 208 head->next_cpu = bcpu; in uv_rtc_find_next_timer() 232 int next_cpu; in uv_rtc_set_timer() local 236 next_cpu = head->next_cpu; in uv_rtc_set_timer() 240 if (next_cpu < 0 || bcpu == next_cpu || in uv_rtc_set_timer() 241 expires < head->cpu[next_cpu].expires) { in uv_rtc_set_timer() 242 head->next_cpu = bcpu; in uv_rtc_set_timer() 272 if ((head->next_cpu == bcpu && uv_read_rtc(NULL) >= *t) || force) in uv_rtc_unset_timer() [all …]
|
/linux-4.4.14/arch/parisc/kernel/ |
D | irq.c | 351 static int next_cpu = -1; in txn_alloc_addr() local 353 next_cpu++; /* assign to "next" CPU we want this bugger on */ in txn_alloc_addr() 356 while ((next_cpu < nr_cpu_ids) && in txn_alloc_addr() 357 (!per_cpu(cpu_data, next_cpu).txn_addr || in txn_alloc_addr() 358 !cpu_online(next_cpu))) in txn_alloc_addr() 359 next_cpu++; in txn_alloc_addr() 361 if (next_cpu >= nr_cpu_ids) in txn_alloc_addr() 362 next_cpu = 0; /* nothing else, assign monarch */ in txn_alloc_addr() 364 return txn_affinity_addr(virt_irq, next_cpu); in txn_alloc_addr()
|
/linux-4.4.14/kernel/ |
D | smp.c | 408 int cpu, next_cpu, this_cpu = smp_processor_id(); in smp_call_function_many() local 429 next_cpu = cpumask_next_and(cpu, mask, cpu_online_mask); in smp_call_function_many() 430 if (next_cpu == this_cpu) in smp_call_function_many() 431 next_cpu = cpumask_next_and(next_cpu, mask, cpu_online_mask); in smp_call_function_many() 434 if (next_cpu >= nr_cpu_ids) { in smp_call_function_many()
|
/linux-4.4.14/kernel/trace/ |
D | trace_entries.h | 121 __field( unsigned int, next_cpu ) \ 138 __entry->next_cpu), 158 __entry->next_cpu),
|
D | trace_output.c | 867 field->next_cpu, in trace_ctxwake_print() 901 field->next_cpu, in trace_ctxwake_raw() 937 SEQ_PUT_HEX_FIELD(s, field->next_cpu); in trace_ctxwake_hex() 968 SEQ_PUT_FIELD(s, field->next_cpu); in trace_ctxwake_bin()
|
D | trace_sched_wakeup.c | 392 entry->next_cpu = task_cpu(next); in tracing_sched_switch_trace() 420 entry->next_cpu = task_cpu(wakee); in tracing_sched_wakeup_trace()
|
D | trace.c | 2314 int next_cpu = -1; in __find_next_entry() local 2344 next_cpu = cpu; in __find_next_entry() 2354 *ent_cpu = next_cpu; in __find_next_entry()
|
/linux-4.4.14/kernel/time/ |
D | clocksource.c | 174 int next_cpu, reset_pending; in clocksource_watchdog() local 272 next_cpu = cpumask_next(raw_smp_processor_id(), cpu_online_mask); in clocksource_watchdog() 273 if (next_cpu >= nr_cpu_ids) in clocksource_watchdog() 274 next_cpu = cpumask_first(cpu_online_mask); in clocksource_watchdog() 276 add_timer_on(&watchdog_timer, next_cpu); in clocksource_watchdog()
|
D | tick-broadcast.c | 603 int cpu, next_cpu = 0; in tick_handle_oneshot_broadcast() local 624 next_cpu = cpu; in tick_handle_oneshot_broadcast() 661 tick_broadcast_set_event(dev, next_cpu, next_event); in tick_handle_oneshot_broadcast()
|
/linux-4.4.14/block/ |
D | blk-mq.c | 843 int cpu = hctx->next_cpu, next_cpu; in blk_mq_hctx_next_cpu() local 845 next_cpu = cpumask_next(hctx->next_cpu, hctx->cpumask); in blk_mq_hctx_next_cpu() 846 if (next_cpu >= nr_cpu_ids) in blk_mq_hctx_next_cpu() 847 next_cpu = cpumask_first(hctx->cpumask); in blk_mq_hctx_next_cpu() 849 hctx->next_cpu = next_cpu; in blk_mq_hctx_next_cpu() 855 return hctx->next_cpu; in blk_mq_hctx_next_cpu() 1867 hctx->next_cpu = cpumask_first(hctx->cpumask); in blk_mq_map_swqueue()
|
/linux-4.4.14/arch/x86/kernel/apic/ |
D | vector.c | 148 goto next_cpu; in __assign_irq_vector() 173 goto next_cpu; in __assign_irq_vector() 192 next_cpu: in __assign_irq_vector()
|
/linux-4.4.14/include/linux/ |
D | blk-mq.h | 31 int next_cpu; member
|
/linux-4.4.14/net/core/ |
D | dev.c | 3228 struct rps_dev_flow *rflow, u16 next_cpu) in set_rps_cpu() argument 3230 if (next_cpu < nr_cpu_ids) { in set_rps_cpu() 3243 rxq_index = cpu_rmap_lookup_index(dev->rx_cpu_rmap, next_cpu); in set_rps_cpu() 3264 per_cpu(softnet_data, next_cpu).input_queue_head; in set_rps_cpu() 3267 rflow->cpu = next_cpu; in set_rps_cpu() 3315 u32 next_cpu; in get_rps_cpu() local 3323 next_cpu = ident & rps_cpu_mask; in get_rps_cpu() 3342 if (unlikely(tcpu != next_cpu) && in get_rps_cpu() 3346 tcpu = next_cpu; in get_rps_cpu() 3347 rflow = set_rps_cpu(dev, skb, rflow, next_cpu); in get_rps_cpu()
|