__ia64_per_cpu_var 69 arch/ia64/include/asm/mmu_context.h if (unlikely(__ia64_per_cpu_var(ia64_need_tlb_flush))) { __ia64_per_cpu_var 71 arch/ia64/include/asm/mmu_context.h if (__ia64_per_cpu_var(ia64_need_tlb_flush)) { __ia64_per_cpu_var 73 arch/ia64/include/asm/mmu_context.h __ia64_per_cpu_var(ia64_need_tlb_flush) = 0; __ia64_per_cpu_var 23 arch/ia64/include/asm/percpu.h #define __my_cpu_offset __ia64_per_cpu_var(local_per_cpu_offset) __ia64_per_cpu_var 241 arch/ia64/include/asm/processor.h #define local_cpu_data (&__ia64_per_cpu_var(ia64_cpu_info)) __ia64_per_cpu_var 579 arch/ia64/kernel/perfmon.c #define pfm_get_cpu_var(v) __ia64_per_cpu_var(v) __ia64_per_cpu_var 980 arch/ia64/kernel/setup.c cpu_info = cpu_data + ((char *) &__ia64_per_cpu_var(ia64_cpu_info) - __per_cpu_start); __ia64_per_cpu_var 100 arch/ia64/kernel/smp.c unsigned long *pending_ipis = &__ia64_per_cpu_var(ipi_operation); __ia64_per_cpu_var 258 arch/ia64/kernel/smp.c unsigned short *counts = __ia64_per_cpu_var(shadow_flush_counts);