KERNEL_START 24 arch/alpha/include/uapi/asm/setup.h #define SWAPPER_PGD KERNEL_START KERNEL_START 267 arch/arm/mm/init.c memblock_reserve(__pa(KERNEL_START), KERNEL_END - KERNEL_START); KERNEL_START 1453 arch/arm/mm/mmu.c phys_addr_t kernel_x_start = round_down(__pa(KERNEL_START), SECTION_SIZE); KERNEL_START 53 arch/arm/mm/physaddr.c VIRTUAL_BUG_ON(x < (unsigned long)KERNEL_START || KERNEL_START 254 arch/arm/mm/pmsa-v8.c subtract_range(mem, ARRAY_SIZE(mem), __pa(KERNEL_START), __pa(KERNEL_END)); KERNEL_START 255 arch/arm/mm/pmsa-v8.c subtract_range(io, ARRAY_SIZE(io), __pa(KERNEL_START), __pa(KERNEL_END)); KERNEL_START 279 arch/arm/mm/pmsa-v8.c err |= pmsav8_setup_fixed(region++, __pa(KERNEL_START), __pa(KERNEL_END)); KERNEL_START 27 arch/arm64/mm/physaddr.c VIRTUAL_BUG_ON(x < (unsigned long) KERNEL_START || KERNEL_START 234 arch/ia64/include/asm/page.h #define LOAD_OFFSET (KERNEL_START - KERNEL_TR_PAGE_SIZE) KERNEL_START 394 arch/ia64/kernel/efi.c if ((vaddr & mask) == (KERNEL_START & mask)) { KERNEL_START 356 arch/ia64/kernel/setup.c rsvd_region[n].start = (unsigned long) ia64_imva((void *)KERNEL_START); KERNEL_START 2279 arch/ia64/kernel/unwind.c init_unwind_table(&unw.kernel_table, "kernel", KERNEL_START, (unsigned long) __gp, KERNEL_START 189 arch/parisc/kernel/unwind.c unwind_table_init(&kernel_unwind_table, "kernel", KERNEL_START,