new_pgd            32 arch/arm/mm/pgd.c 	pgd_t *new_pgd, *init_pgd;
new_pgd            37 arch/arm/mm/pgd.c 	new_pgd = __pgd_alloc();
new_pgd            38 arch/arm/mm/pgd.c 	if (!new_pgd)
new_pgd            41 arch/arm/mm/pgd.c 	memset(new_pgd, 0, USER_PTRS_PER_PGD * sizeof(pgd_t));
new_pgd            47 arch/arm/mm/pgd.c 	memcpy(new_pgd + USER_PTRS_PER_PGD, init_pgd + USER_PTRS_PER_PGD,
new_pgd            50 arch/arm/mm/pgd.c 	clean_dcache_area(new_pgd, PTRS_PER_PGD * sizeof(pgd_t));
new_pgd            56 arch/arm/mm/pgd.c 	new_pud = pud_alloc(mm, new_pgd + pgd_index(MODULES_VADDR),
new_pgd            72 arch/arm/mm/pgd.c 		new_pud = pud_alloc(mm, new_pgd, 0);
new_pgd           103 arch/arm/mm/pgd.c 	return new_pgd;
new_pgd           111 arch/arm/mm/pgd.c 	__pgd_free(new_pgd);
new_pgd            93 arch/m68k/include/asm/mcf_pgalloc.h 	pgd_t *new_pgd;
new_pgd            95 arch/m68k/include/asm/mcf_pgalloc.h 	new_pgd = (pgd_t *)__get_free_page(GFP_DMA | __GFP_NOWARN);
new_pgd            96 arch/m68k/include/asm/mcf_pgalloc.h 	if (!new_pgd)
new_pgd            98 arch/m68k/include/asm/mcf_pgalloc.h 	memcpy(new_pgd, swapper_pg_dir, PAGE_SIZE);
new_pgd            99 arch/m68k/include/asm/mcf_pgalloc.h 	memset(new_pgd, 0, PAGE_OFFSET >> PGDIR_SHIFT);
new_pgd           100 arch/m68k/include/asm/mcf_pgalloc.h 	return new_pgd;
new_pgd            53 arch/m68k/include/asm/sun3_pgalloc.h      pgd_t *new_pgd;
new_pgd            55 arch/m68k/include/asm/sun3_pgalloc.h      new_pgd = (pgd_t *)get_zeroed_page(GFP_KERNEL);
new_pgd            56 arch/m68k/include/asm/sun3_pgalloc.h      memcpy(new_pgd, swapper_pg_dir, PAGE_SIZE);
new_pgd            57 arch/m68k/include/asm/sun3_pgalloc.h      memset(new_pgd, 0, (PAGE_OFFSET >> PGDIR_SHIFT));
new_pgd            58 arch/m68k/include/asm/sun3_pgalloc.h      return new_pgd;
new_pgd            15 arch/nds32/mm/mm-nds32.c 	pgd_t *new_pgd, *init_pgd;
new_pgd            18 arch/nds32/mm/mm-nds32.c 	new_pgd = (pgd_t *) __get_free_pages(GFP_KERNEL, 0);
new_pgd            19 arch/nds32/mm/mm-nds32.c 	if (!new_pgd)
new_pgd            22 arch/nds32/mm/mm-nds32.c 		(*new_pgd) = 1;
new_pgd            23 arch/nds32/mm/mm-nds32.c 		new_pgd++;
new_pgd            25 arch/nds32/mm/mm-nds32.c 	new_pgd -= PTRS_PER_PGD;
new_pgd            29 arch/nds32/mm/mm-nds32.c 	memcpy(new_pgd + FIRST_KERNEL_PGD_NR, init_pgd + FIRST_KERNEL_PGD_NR,
new_pgd            32 arch/nds32/mm/mm-nds32.c 	cpu_dcache_wb_range((unsigned long)new_pgd,
new_pgd            33 arch/nds32/mm/mm-nds32.c 			    (unsigned long)new_pgd +
new_pgd            35 arch/nds32/mm/mm-nds32.c 	inc_zone_page_state(virt_to_page((unsigned long *)new_pgd),
new_pgd            38 arch/nds32/mm/mm-nds32.c 	return new_pgd;
new_pgd            26 arch/unicore32/mm/pgd.c 	pgd_t *new_pgd, *init_pgd;
new_pgd            30 arch/unicore32/mm/pgd.c 	new_pgd = (pgd_t *)__get_free_pages(GFP_KERNEL, 0);
new_pgd            31 arch/unicore32/mm/pgd.c 	if (!new_pgd)
new_pgd            34 arch/unicore32/mm/pgd.c 	memset(new_pgd, 0, FIRST_KERNEL_PGD_NR * sizeof(pgd_t));
new_pgd            40 arch/unicore32/mm/pgd.c 	memcpy(new_pgd + FIRST_KERNEL_PGD_NR, init_pgd + FIRST_KERNEL_PGD_NR,
new_pgd            43 arch/unicore32/mm/pgd.c 	clean_dcache_area(new_pgd, PTRS_PER_PGD * sizeof(pgd_t));
new_pgd            50 arch/unicore32/mm/pgd.c 		new_pmd = pmd_alloc(mm, (pud_t *)new_pgd, 0);
new_pgd            65 arch/unicore32/mm/pgd.c 	return new_pgd;
new_pgd            71 arch/unicore32/mm/pgd.c 	free_pages((unsigned long)new_pgd, 0);
new_pgd            74 arch/x86/power/hibernate_64.c 		pgd_t new_pgd = __pgd(__pa(p4d) | pgprot_val(pgtable_prot));
new_pgd            77 arch/x86/power/hibernate_64.c 		set_pgd(pgd + pgd_index(restore_jump_address), new_pgd);
new_pgd            80 arch/x86/power/hibernate_64.c 		pgd_t new_pgd = __pgd(__pa(pud) | pgprot_val(pgtable_prot));
new_pgd            81 arch/x86/power/hibernate_64.c 		set_pgd(pgd + pgd_index(restore_jump_address), new_pgd);