new_pgd 32 arch/arm/mm/pgd.c pgd_t *new_pgd, *init_pgd; new_pgd 37 arch/arm/mm/pgd.c new_pgd = __pgd_alloc(); new_pgd 38 arch/arm/mm/pgd.c if (!new_pgd) new_pgd 41 arch/arm/mm/pgd.c memset(new_pgd, 0, USER_PTRS_PER_PGD * sizeof(pgd_t)); new_pgd 47 arch/arm/mm/pgd.c memcpy(new_pgd + USER_PTRS_PER_PGD, init_pgd + USER_PTRS_PER_PGD, new_pgd 50 arch/arm/mm/pgd.c clean_dcache_area(new_pgd, PTRS_PER_PGD * sizeof(pgd_t)); new_pgd 56 arch/arm/mm/pgd.c new_pud = pud_alloc(mm, new_pgd + pgd_index(MODULES_VADDR), new_pgd 72 arch/arm/mm/pgd.c new_pud = pud_alloc(mm, new_pgd, 0); new_pgd 103 arch/arm/mm/pgd.c return new_pgd; new_pgd 111 arch/arm/mm/pgd.c __pgd_free(new_pgd); new_pgd 93 arch/m68k/include/asm/mcf_pgalloc.h pgd_t *new_pgd; new_pgd 95 arch/m68k/include/asm/mcf_pgalloc.h new_pgd = (pgd_t *)__get_free_page(GFP_DMA | __GFP_NOWARN); new_pgd 96 arch/m68k/include/asm/mcf_pgalloc.h if (!new_pgd) new_pgd 98 arch/m68k/include/asm/mcf_pgalloc.h memcpy(new_pgd, swapper_pg_dir, PAGE_SIZE); new_pgd 99 arch/m68k/include/asm/mcf_pgalloc.h memset(new_pgd, 0, PAGE_OFFSET >> PGDIR_SHIFT); new_pgd 100 arch/m68k/include/asm/mcf_pgalloc.h return new_pgd; new_pgd 53 arch/m68k/include/asm/sun3_pgalloc.h pgd_t *new_pgd; new_pgd 55 arch/m68k/include/asm/sun3_pgalloc.h new_pgd = (pgd_t *)get_zeroed_page(GFP_KERNEL); new_pgd 56 arch/m68k/include/asm/sun3_pgalloc.h memcpy(new_pgd, swapper_pg_dir, PAGE_SIZE); new_pgd 57 arch/m68k/include/asm/sun3_pgalloc.h memset(new_pgd, 0, (PAGE_OFFSET >> PGDIR_SHIFT)); new_pgd 58 arch/m68k/include/asm/sun3_pgalloc.h return new_pgd; new_pgd 15 arch/nds32/mm/mm-nds32.c pgd_t *new_pgd, *init_pgd; new_pgd 18 arch/nds32/mm/mm-nds32.c new_pgd = (pgd_t *) __get_free_pages(GFP_KERNEL, 0); new_pgd 19 arch/nds32/mm/mm-nds32.c if (!new_pgd) new_pgd 22 arch/nds32/mm/mm-nds32.c (*new_pgd) = 1; new_pgd 23 arch/nds32/mm/mm-nds32.c new_pgd++; new_pgd 25 arch/nds32/mm/mm-nds32.c new_pgd -= PTRS_PER_PGD; new_pgd 29 arch/nds32/mm/mm-nds32.c memcpy(new_pgd + FIRST_KERNEL_PGD_NR, init_pgd + FIRST_KERNEL_PGD_NR, new_pgd 32 arch/nds32/mm/mm-nds32.c cpu_dcache_wb_range((unsigned long)new_pgd, new_pgd 33 arch/nds32/mm/mm-nds32.c (unsigned long)new_pgd + new_pgd 35 arch/nds32/mm/mm-nds32.c inc_zone_page_state(virt_to_page((unsigned long *)new_pgd), new_pgd 38 arch/nds32/mm/mm-nds32.c return new_pgd; new_pgd 26 arch/unicore32/mm/pgd.c pgd_t *new_pgd, *init_pgd; new_pgd 30 arch/unicore32/mm/pgd.c new_pgd = (pgd_t *)__get_free_pages(GFP_KERNEL, 0); new_pgd 31 arch/unicore32/mm/pgd.c if (!new_pgd) new_pgd 34 arch/unicore32/mm/pgd.c memset(new_pgd, 0, FIRST_KERNEL_PGD_NR * sizeof(pgd_t)); new_pgd 40 arch/unicore32/mm/pgd.c memcpy(new_pgd + FIRST_KERNEL_PGD_NR, init_pgd + FIRST_KERNEL_PGD_NR, new_pgd 43 arch/unicore32/mm/pgd.c clean_dcache_area(new_pgd, PTRS_PER_PGD * sizeof(pgd_t)); new_pgd 50 arch/unicore32/mm/pgd.c new_pmd = pmd_alloc(mm, (pud_t *)new_pgd, 0); new_pgd 65 arch/unicore32/mm/pgd.c return new_pgd; new_pgd 71 arch/unicore32/mm/pgd.c free_pages((unsigned long)new_pgd, 0); new_pgd 74 arch/x86/power/hibernate_64.c pgd_t new_pgd = __pgd(__pa(p4d) | pgprot_val(pgtable_prot)); new_pgd 77 arch/x86/power/hibernate_64.c set_pgd(pgd + pgd_index(restore_jump_address), new_pgd); new_pgd 80 arch/x86/power/hibernate_64.c pgd_t new_pgd = __pgd(__pa(pud) | pgprot_val(pgtable_prot)); new_pgd 81 arch/x86/power/hibernate_64.c set_pgd(pgd + pgd_index(restore_jump_address), new_pgd);