CPU_ARCH_ARMv6 89 arch/arm/kernel/elf.c if (cpu_architecture() < CPU_ARCH_ARMv6) CPU_ARCH_ARMv6 265 arch/arm/kernel/setup.c cpu_arch = CPU_ARCH_ARMv6; CPU_ARCH_ARMv6 301 arch/arm/kernel/setup.c case CPU_ARCH_ARMv6: CPU_ARCH_ARMv6 316 arch/arm/kernel/setup.c if (arch >= CPU_ARCH_ARMv6) { CPU_ARCH_ARMv6 334 arch/arm/kernel/setup.c arch = CPU_ARCH_ARMv6; CPU_ARCH_ARMv6 96 arch/arm/mm/alignment.c return cpu_architecture() >= CPU_ARCH_ARMv6 && get_cr() & CR_U; CPU_ARCH_ARMv6 1041 arch/arm/mm/alignment.c if (cpu_architecture() <= CPU_ARCH_ARMv6) { CPU_ARCH_ARMv6 598 arch/arm/mm/fault.c if (cpu_architecture() >= CPU_ARCH_ARMv6) { CPU_ARCH_ARMv6 590 arch/arm/mm/init.c if (cpu_architecture() < CPU_ARCH_ARMv6) CPU_ARCH_ARMv6 315 arch/arm/mm/ioremap.c (((cpu_architecture() >= CPU_ARCH_ARMv6) && (get_cr() & CR_XP)) || CPU_ARCH_ARMv6 179 arch/arm/mm/mmu.c if (cpu_architecture() >= CPU_ARCH_ARMv6 && selected != cachepolicy) { CPU_ARCH_ARMv6 441 arch/arm/mm/mmu.c if (cpu_arch < CPU_ARCH_ARMv6) { CPU_ARCH_ARMv6 475 arch/arm/mm/mmu.c if ((cpu_arch < CPU_ARCH_ARMv6 || !(cr & CR_XP)) && !cpu_is_xsc3()) CPU_ARCH_ARMv6 489 arch/arm/mm/mmu.c } else if (cpu_arch < CPU_ARCH_ARMv6) { CPU_ARCH_ARMv6 501 arch/arm/mm/mmu.c if (cpu_is_xsc3() || (cpu_arch >= CPU_ARCH_ARMv6 && (cr & CR_XP))) { CPU_ARCH_ARMv6 571 arch/arm/mm/mmu.c if (cpu_arch == CPU_ARCH_ARMv6) CPU_ARCH_ARMv6 587 arch/arm/mm/mmu.c if (cpu_arch >= CPU_ARCH_ARMv6 && (cr & CR_XP)) { CPU_ARCH_ARMv6 626 arch/arm/mm/mmu.c if (cpu_arch >= CPU_ARCH_ARMv6) { CPU_ARCH_ARMv6 859 arch/arm/mm/mmu.c if (!(cpu_architecture() >= CPU_ARCH_ARMv6 || cpu_is_xsc3())) { CPU_ARCH_ARMv6 723 arch/arm/vfp/vfpmodule.c if (cpu_arch >= CPU_ARCH_ARMv6)