k_start 33 arch/powerpc/mm/kasan/kasan_init_32.c static int __ref kasan_init_shadow_page_tables(unsigned long k_start, unsigned long k_end) k_start 39 arch/powerpc/mm/kasan/kasan_init_32.c pmd = pmd_offset(pud_offset(pgd_offset_k(k_start), k_start), k_start); k_start 41 arch/powerpc/mm/kasan/kasan_init_32.c for (k_cur = k_start; k_cur != k_end; k_cur = k_next, pmd++) { k_start 83 arch/powerpc/mm/kasan/kasan_init_32.c unsigned long k_start = (unsigned long)kasan_mem_to_shadow(start); k_start 89 arch/powerpc/mm/kasan/kasan_init_32.c ret = kasan_init_shadow_page_tables(k_start, k_end); k_start 94 arch/powerpc/mm/kasan/kasan_init_32.c block = memblock_alloc(k_end - k_start, PAGE_SIZE); k_start 96 arch/powerpc/mm/kasan/kasan_init_32.c for (k_cur = k_start & PAGE_MASK; k_cur < k_end; k_cur += PAGE_SIZE) { k_start 98 arch/powerpc/mm/kasan/kasan_init_32.c void *va = block ? block + k_cur - k_start : kasan_get_one_page(); k_start 106 arch/powerpc/mm/kasan/kasan_init_32.c flush_tlb_kernel_range(k_start, k_end); k_start 113 arch/powerpc/mm/kasan/kasan_init_32.c unsigned long k_start = KASAN_SHADOW_START; k_start 120 arch/powerpc/mm/kasan/kasan_init_32.c for (k_cur = k_start & PAGE_MASK; k_cur != k_end; k_cur += PAGE_SIZE) {