KERNEL_START       24 arch/alpha/include/uapi/asm/setup.h #define SWAPPER_PGD	KERNEL_START
KERNEL_START      267 arch/arm/mm/init.c 	memblock_reserve(__pa(KERNEL_START), KERNEL_END - KERNEL_START);
KERNEL_START     1453 arch/arm/mm/mmu.c 	phys_addr_t kernel_x_start = round_down(__pa(KERNEL_START), SECTION_SIZE);
KERNEL_START       53 arch/arm/mm/physaddr.c 	VIRTUAL_BUG_ON(x < (unsigned long)KERNEL_START ||
KERNEL_START      254 arch/arm/mm/pmsa-v8.c 	subtract_range(mem, ARRAY_SIZE(mem), __pa(KERNEL_START), __pa(KERNEL_END));
KERNEL_START      255 arch/arm/mm/pmsa-v8.c 	subtract_range(io, ARRAY_SIZE(io),  __pa(KERNEL_START), __pa(KERNEL_END));
KERNEL_START      279 arch/arm/mm/pmsa-v8.c 	err |= pmsav8_setup_fixed(region++, __pa(KERNEL_START), __pa(KERNEL_END));
KERNEL_START       27 arch/arm64/mm/physaddr.c 	VIRTUAL_BUG_ON(x < (unsigned long) KERNEL_START ||
KERNEL_START      234 arch/ia64/include/asm/page.h #define LOAD_OFFSET		(KERNEL_START - KERNEL_TR_PAGE_SIZE)
KERNEL_START      394 arch/ia64/kernel/efi.c 		if ((vaddr & mask) == (KERNEL_START & mask)) {
KERNEL_START      356 arch/ia64/kernel/setup.c 	rsvd_region[n].start = (unsigned long) ia64_imva((void *)KERNEL_START);
KERNEL_START     2279 arch/ia64/kernel/unwind.c 	init_unwind_table(&unw.kernel_table, "kernel", KERNEL_START, (unsigned long) __gp,
KERNEL_START      189 arch/parisc/kernel/unwind.c 	unwind_table_init(&kernel_unwind_table, "kernel", KERNEL_START,