cpu_llc_id 410 arch/x86/events/amd/uncore.c uncore->id = per_cpu(cpu_llc_id, cpu); cpu_llc_id 29 arch/x86/include/asm/smp.h DECLARE_PER_CPU_READ_MOSTLY(u16, cpu_llc_id); cpu_llc_id 175 arch/x86/kernel/apic/apic_numachip.c this_cpu_write(cpu_llc_id, node); cpu_llc_id 376 arch/x86/kernel/cpu/amd.c per_cpu(cpu_llc_id, cpu) = node_id; cpu_llc_id 401 arch/x86/kernel/cpu/amd.c per_cpu(cpu_llc_id, cpu) = c->phys_proc_id; cpu_llc_id 406 arch/x86/kernel/cpu/amd.c return per_cpu(cpu_llc_id, cpu); cpu_llc_id 425 arch/x86/kernel/cpu/amd.c node = per_cpu(cpu_llc_id, cpu); cpu_llc_id 660 arch/x86/kernel/cpu/cacheinfo.c per_cpu(cpu_llc_id, cpu) = node_id; cpu_llc_id 666 arch/x86/kernel/cpu/cacheinfo.c per_cpu(cpu_llc_id, cpu) = c->apicid >> 3; cpu_llc_id 682 arch/x86/kernel/cpu/cacheinfo.c per_cpu(cpu_llc_id, cpu) = c->apicid >> bits; cpu_llc_id 700 arch/x86/kernel/cpu/cacheinfo.c per_cpu(cpu_llc_id, cpu) = c->apicid >> 3; cpu_llc_id 848 arch/x86/kernel/cpu/cacheinfo.c per_cpu(cpu_llc_id, cpu) = l2_id; cpu_llc_id 855 arch/x86/kernel/cpu/cacheinfo.c per_cpu(cpu_llc_id, cpu) = l3_id; cpu_llc_id 867 arch/x86/kernel/cpu/cacheinfo.c if (per_cpu(cpu_llc_id, cpu) == BAD_APICID) cpu_llc_id 868 arch/x86/kernel/cpu/cacheinfo.c per_cpu(cpu_llc_id, cpu) = c->phys_proc_id; cpu_llc_id 78 arch/x86/kernel/cpu/common.c DEFINE_PER_CPU_READ_MOSTLY(u16, cpu_llc_id) = BAD_APICID; cpu_llc_id 102 arch/x86/kernel/cpu/hygon.c per_cpu(cpu_llc_id, cpu) = node_id; cpu_llc_id 125 arch/x86/kernel/cpu/hygon.c per_cpu(cpu_llc_id, cpu) = c->phys_proc_id; cpu_llc_id 137 arch/x86/kernel/cpu/hygon.c node = per_cpu(cpu_llc_id, cpu); cpu_llc_id 442 arch/x86/kernel/smpboot.c per_cpu(cpu_llc_id, cpu1) == per_cpu(cpu_llc_id, cpu2)) { cpu_llc_id 486 arch/x86/kernel/smpboot.c if (per_cpu(cpu_llc_id, cpu1) == BAD_APICID) cpu_llc_id 490 arch/x86/kernel/smpboot.c if (per_cpu(cpu_llc_id, cpu1) != per_cpu(cpu_llc_id, cpu2))