CPU_ARCH_ARMv6     89 arch/arm/kernel/elf.c 	if (cpu_architecture() < CPU_ARCH_ARMv6)
CPU_ARCH_ARMv6    265 arch/arm/kernel/setup.c 			cpu_arch = CPU_ARCH_ARMv6;
CPU_ARCH_ARMv6    301 arch/arm/kernel/setup.c 	case CPU_ARCH_ARMv6:
CPU_ARCH_ARMv6    316 arch/arm/kernel/setup.c 	if (arch >= CPU_ARCH_ARMv6) {
CPU_ARCH_ARMv6    334 arch/arm/kernel/setup.c 			arch = CPU_ARCH_ARMv6;
CPU_ARCH_ARMv6     96 arch/arm/mm/alignment.c 	return cpu_architecture() >= CPU_ARCH_ARMv6 && get_cr() & CR_U;
CPU_ARCH_ARMv6   1041 arch/arm/mm/alignment.c 	if (cpu_architecture() <= CPU_ARCH_ARMv6) {
CPU_ARCH_ARMv6    598 arch/arm/mm/fault.c 	if (cpu_architecture() >= CPU_ARCH_ARMv6) {
CPU_ARCH_ARMv6    590 arch/arm/mm/init.c 	if (cpu_architecture() < CPU_ARCH_ARMv6)
CPU_ARCH_ARMv6    315 arch/arm/mm/ioremap.c 	    (((cpu_architecture() >= CPU_ARCH_ARMv6) && (get_cr() & CR_XP)) ||
CPU_ARCH_ARMv6    179 arch/arm/mm/mmu.c 	if (cpu_architecture() >= CPU_ARCH_ARMv6 && selected != cachepolicy) {
CPU_ARCH_ARMv6    441 arch/arm/mm/mmu.c 	if (cpu_arch < CPU_ARCH_ARMv6) {
CPU_ARCH_ARMv6    475 arch/arm/mm/mmu.c 	if ((cpu_arch < CPU_ARCH_ARMv6 || !(cr & CR_XP)) && !cpu_is_xsc3())
CPU_ARCH_ARMv6    489 arch/arm/mm/mmu.c 	} else if (cpu_arch < CPU_ARCH_ARMv6) {
CPU_ARCH_ARMv6    501 arch/arm/mm/mmu.c 	if (cpu_is_xsc3() || (cpu_arch >= CPU_ARCH_ARMv6 && (cr & CR_XP))) {
CPU_ARCH_ARMv6    571 arch/arm/mm/mmu.c 	if (cpu_arch == CPU_ARCH_ARMv6)
CPU_ARCH_ARMv6    587 arch/arm/mm/mmu.c 	if (cpu_arch >= CPU_ARCH_ARMv6 && (cr & CR_XP)) {
CPU_ARCH_ARMv6    626 arch/arm/mm/mmu.c 	if (cpu_arch >= CPU_ARCH_ARMv6) {
CPU_ARCH_ARMv6    859 arch/arm/mm/mmu.c 	if (!(cpu_architecture() >= CPU_ARCH_ARMv6 || cpu_is_xsc3())) {
CPU_ARCH_ARMv6    723 arch/arm/vfp/vfpmodule.c 	if (cpu_arch >= CPU_ARCH_ARMv6)