__ia64_per_cpu_var   69 arch/ia64/include/asm/mmu_context.h 	if (unlikely(__ia64_per_cpu_var(ia64_need_tlb_flush))) {
__ia64_per_cpu_var   71 arch/ia64/include/asm/mmu_context.h 		if (__ia64_per_cpu_var(ia64_need_tlb_flush)) {
__ia64_per_cpu_var   73 arch/ia64/include/asm/mmu_context.h 			__ia64_per_cpu_var(ia64_need_tlb_flush) = 0;
__ia64_per_cpu_var   23 arch/ia64/include/asm/percpu.h #define __my_cpu_offset	__ia64_per_cpu_var(local_per_cpu_offset)
__ia64_per_cpu_var  241 arch/ia64/include/asm/processor.h #define local_cpu_data		(&__ia64_per_cpu_var(ia64_cpu_info))
__ia64_per_cpu_var  579 arch/ia64/kernel/perfmon.c #define pfm_get_cpu_var(v)		__ia64_per_cpu_var(v)
__ia64_per_cpu_var  980 arch/ia64/kernel/setup.c 	cpu_info = cpu_data + ((char *) &__ia64_per_cpu_var(ia64_cpu_info) - __per_cpu_start);
__ia64_per_cpu_var  100 arch/ia64/kernel/smp.c 	unsigned long *pending_ipis = &__ia64_per_cpu_var(ipi_operation);
__ia64_per_cpu_var  258 arch/ia64/kernel/smp.c 	unsigned short *counts = __ia64_per_cpu_var(shadow_flush_counts);