k_start            33 arch/powerpc/mm/kasan/kasan_init_32.c static int __ref kasan_init_shadow_page_tables(unsigned long k_start, unsigned long k_end)
k_start            39 arch/powerpc/mm/kasan/kasan_init_32.c 	pmd = pmd_offset(pud_offset(pgd_offset_k(k_start), k_start), k_start);
k_start            41 arch/powerpc/mm/kasan/kasan_init_32.c 	for (k_cur = k_start; k_cur != k_end; k_cur = k_next, pmd++) {
k_start            83 arch/powerpc/mm/kasan/kasan_init_32.c 	unsigned long k_start = (unsigned long)kasan_mem_to_shadow(start);
k_start            89 arch/powerpc/mm/kasan/kasan_init_32.c 	ret = kasan_init_shadow_page_tables(k_start, k_end);
k_start            94 arch/powerpc/mm/kasan/kasan_init_32.c 		block = memblock_alloc(k_end - k_start, PAGE_SIZE);
k_start            96 arch/powerpc/mm/kasan/kasan_init_32.c 	for (k_cur = k_start & PAGE_MASK; k_cur < k_end; k_cur += PAGE_SIZE) {
k_start            98 arch/powerpc/mm/kasan/kasan_init_32.c 		void *va = block ? block + k_cur - k_start : kasan_get_one_page();
k_start           106 arch/powerpc/mm/kasan/kasan_init_32.c 	flush_tlb_kernel_range(k_start, k_end);
k_start           113 arch/powerpc/mm/kasan/kasan_init_32.c 	unsigned long k_start = KASAN_SHADOW_START;
k_start           120 arch/powerpc/mm/kasan/kasan_init_32.c 	for (k_cur = k_start & PAGE_MASK; k_cur != k_end; k_cur += PAGE_SIZE) {