old_cpu 81 arch/mips/sibyte/bcm1480/irq.c int i = 0, old_cpu, cpu, int_on, k; old_cpu 94 arch/mips/sibyte/bcm1480/irq.c old_cpu = bcm1480_irq_owner[irq]; old_cpu 101 arch/mips/sibyte/bcm1480/irq.c cur_ints = ____raw_readq(IOADDR(A_BCM1480_IMR_MAPPER(old_cpu) + R_BCM1480_IMR_INTERRUPT_MASK_H + (k*BCM1480_IMR_HL_SPACING))); old_cpu 106 arch/mips/sibyte/bcm1480/irq.c ____raw_writeq(cur_ints, IOADDR(A_BCM1480_IMR_MAPPER(old_cpu) + R_BCM1480_IMR_INTERRUPT_MASK_H + (k*BCM1480_IMR_HL_SPACING))); old_cpu 73 arch/mips/sibyte/sb1250/irq.c int i = 0, old_cpu, cpu, int_on; old_cpu 87 arch/mips/sibyte/sb1250/irq.c old_cpu = sb1250_irq_owner[irq]; old_cpu 88 arch/mips/sibyte/sb1250/irq.c cur_ints = ____raw_readq(IOADDR(A_IMR_MAPPER(old_cpu) + old_cpu 94 arch/mips/sibyte/sb1250/irq.c ____raw_writeq(cur_ints, IOADDR(A_IMR_MAPPER(old_cpu) + old_cpu 598 arch/powerpc/kernel/fadump.c int old_cpu, this_cpu; old_cpu 611 arch/powerpc/kernel/fadump.c old_cpu = cmpxchg(&crashing_cpu, -1, this_cpu); old_cpu 613 arch/powerpc/kernel/fadump.c if (old_cpu != -1) { old_cpu 320 arch/powerpc/perf/imc-pmu.c static void nest_change_cpu_context(int old_cpu, int new_cpu) old_cpu 324 arch/powerpc/perf/imc-pmu.c if (old_cpu < 0 || new_cpu < 0) old_cpu 328 arch/powerpc/perf/imc-pmu.c perf_pmu_migrate_context(&(*pn)->pmu, old_cpu, new_cpu); old_cpu 1157 arch/x86/events/intel/uncore.c static void uncore_change_type_ctx(struct intel_uncore_type *type, int old_cpu, old_cpu 1164 arch/x86/events/intel/uncore.c die = topology_logical_die_id(old_cpu < 0 ? new_cpu : old_cpu); old_cpu 1170 arch/x86/events/intel/uncore.c if (old_cpu < 0) { old_cpu 1176 arch/x86/events/intel/uncore.c WARN_ON_ONCE(box->cpu != old_cpu); old_cpu 1182 arch/x86/events/intel/uncore.c perf_pmu_migrate_context(&pmu->pmu, old_cpu, new_cpu); old_cpu 1188 arch/x86/events/intel/uncore.c int old_cpu, int new_cpu) old_cpu 1191 arch/x86/events/intel/uncore.c uncore_change_type_ctx(*uncores, old_cpu, new_cpu); old_cpu 797 drivers/infiniband/hw/hfi1/affinity.c int i, old_cpu; old_cpu 807 drivers/infiniband/hw/hfi1/affinity.c old_cpu = sde->cpu; old_cpu 830 drivers/infiniband/hw/hfi1/affinity.c if (cpumask_test_cpu(old_cpu, &other_msix->mask)) old_cpu 833 drivers/infiniband/hw/hfi1/affinity.c cpumask_clear_cpu(old_cpu, &set->mask); old_cpu 834 drivers/infiniband/hw/hfi1/affinity.c cpumask_clear_cpu(old_cpu, &set->used); old_cpu 203 drivers/irqchip/irq-bcm6345-l1.c unsigned int old_cpu = cpu_for_irq(intc, d); old_cpu 219 drivers/irqchip/irq-bcm6345-l1.c if (old_cpu != new_cpu) { old_cpu 220 drivers/irqchip/irq-bcm6345-l1.c enabled = intc->cpus[old_cpu]->enable_cache[word] & mask; old_cpu 563 kernel/debug/kdb/kdb_io.c int this_cpu, old_cpu; old_cpu 576 kernel/debug/kdb/kdb_io.c old_cpu = cmpxchg(&kdb_printf_cpu, -1, this_cpu); old_cpu 577 kernel/debug/kdb/kdb_io.c if (old_cpu == -1 || old_cpu == this_cpu) old_cpu 850 kernel/debug/kdb/kdb_io.c smp_store_release(&kdb_printf_cpu, old_cpu); old_cpu 964 kernel/kexec_core.c int old_cpu, this_cpu; old_cpu 972 kernel/kexec_core.c old_cpu = atomic_cmpxchg(&panic_cpu, PANIC_CPU_INVALID, this_cpu); old_cpu 973 kernel/kexec_core.c if (old_cpu == PANIC_CPU_INVALID) { old_cpu 126 kernel/panic.c int old_cpu, cpu; old_cpu 129 kernel/panic.c old_cpu = atomic_cmpxchg(&panic_cpu, PANIC_CPU_INVALID, cpu); old_cpu 131 kernel/panic.c if (old_cpu == PANIC_CPU_INVALID) old_cpu 133 kernel/panic.c else if (old_cpu != cpu) old_cpu 173 kernel/panic.c int old_cpu, this_cpu; old_cpu 201 kernel/panic.c old_cpu = atomic_cmpxchg(&panic_cpu, PANIC_CPU_INVALID, this_cpu); old_cpu 203 kernel/panic.c if (old_cpu != PANIC_CPU_INVALID && old_cpu != this_cpu)