/linux-4.1.27/arch/powerpc/mm/ |
D | hugepage-hash64.c | 28 unsigned long old_pmd, new_pmd; in __hash_page_thp() local 52 new_pmd = old_pmd | _PAGE_BUSY | _PAGE_ACCESSED; in __hash_page_thp() 54 new_pmd |= _PAGE_DIRTY; in __hash_page_thp() 56 old_pmd, new_pmd)); in __hash_page_thp() 61 rflags = new_pmd & _PAGE_USER; in __hash_page_thp() 62 if ((new_pmd & _PAGE_USER) && !((new_pmd & _PAGE_RW) && in __hash_page_thp() 63 (new_pmd & _PAGE_DIRTY))) in __hash_page_thp() 68 rflags |= ((new_pmd & _PAGE_EXEC) ? 0 : HPTE_R_N); in __hash_page_thp() 132 new_pmd |= _PAGE_HASHPTE; in __hash_page_thp() 135 rflags |= (new_pmd & (_PAGE_WRITETHRU | _PAGE_NO_CACHE | in __hash_page_thp() [all …]
|
/linux-4.1.27/arch/arm/mm/ |
D | pgd.c | 37 pmd_t *new_pmd, *init_pmd; in pgd_alloc() local 64 new_pmd = pmd_alloc(mm, new_pud, 0); in pgd_alloc() 65 if (!new_pmd) in pgd_alloc() 79 new_pmd = pmd_alloc(mm, new_pud, 0); in pgd_alloc() 80 if (!new_pmd) in pgd_alloc() 83 new_pte = pte_alloc_map(mm, NULL, new_pmd, 0); in pgd_alloc() 99 pmd_free(mm, new_pmd); in pgd_alloc()
|
/linux-4.1.27/arch/unicore32/mm/ |
D | pgd.c | 30 pmd_t *new_pmd, *init_pmd; in get_pgd_slow() local 53 new_pmd = pmd_alloc(mm, (pud_t *)new_pgd, 0); in get_pgd_slow() 54 if (!new_pmd) in get_pgd_slow() 57 new_pte = pte_alloc_map(mm, NULL, new_pmd, 0); in get_pgd_slow() 71 pmd_free(mm, new_pmd); in get_pgd_slow()
|
/linux-4.1.27/mm/ |
D | mremap.c | 90 struct vm_area_struct *new_vma, pmd_t *new_pmd, in move_ptes() argument 133 new_pte = pte_offset_map(new_pmd, new_addr); in move_ptes() 134 new_ptl = pte_lockptr(mm, new_pmd); in move_ptes() 168 pmd_t *old_pmd, *new_pmd; in move_page_tables() local 190 new_pmd = alloc_new_pmd(vma->vm_mm, vma, new_addr); in move_page_tables() 191 if (!new_pmd) in move_page_tables() 203 old_pmd, new_pmd); in move_page_tables() 215 if (pmd_none(*new_pmd) && __pte_alloc(new_vma->vm_mm, new_vma, in move_page_tables() 216 new_pmd, new_addr)) in move_page_tables() 224 new_vma, new_pmd, new_addr, need_rmap_locks); in move_page_tables()
|
D | huge_memory.c | 1430 pmd_t *old_pmd, pmd_t *new_pmd) in move_huge_pmd() argument 1448 if (WARN_ON(!pmd_none(*new_pmd))) { in move_huge_pmd() 1449 VM_BUG_ON(pmd_trans_huge(*new_pmd)); in move_huge_pmd() 1459 new_ptl = pmd_lockptr(mm, new_pmd); in move_huge_pmd() 1463 VM_BUG_ON(!pmd_none(*new_pmd)); in move_huge_pmd() 1468 pgtable_trans_huge_deposit(mm, new_pmd, pgtable); in move_huge_pmd() 1470 set_pmd_at(mm, new_addr, new_pmd, pmd_mksoft_dirty(pmd)); in move_huge_pmd()
|
/linux-4.1.27/arch/arm/include/asm/ |
D | kvm_mmu.h | 72 static inline void kvm_set_pmd(pmd_t *pmd, pmd_t new_pmd) in kvm_set_pmd() argument 74 *pmd = new_pmd; in kvm_set_pmd()
|
/linux-4.1.27/arch/arm/kvm/ |
D | mmu.c | 870 *cache, phys_addr_t addr, const pmd_t *new_pmd) in stage2_set_pmd_huge() argument 886 VM_BUG_ON(pmd_present(*pmd) && pmd_pfn(*pmd) != pmd_pfn(*new_pmd)); in stage2_set_pmd_huge() 896 kvm_set_pmd(pmd, *new_pmd); in stage2_set_pmd_huge() 1317 pmd_t new_pmd = pfn_pmd(pfn, mem_type); in user_mem_abort() local 1318 new_pmd = pmd_mkhuge(new_pmd); in user_mem_abort() 1320 kvm_set_s2pmd_writable(&new_pmd); in user_mem_abort() 1324 ret = stage2_set_pmd_huge(kvm, memcache, fault_ipa, &new_pmd); in user_mem_abort()
|
/linux-4.1.27/include/linux/ |
D | huge_mm.h | 32 pmd_t *old_pmd, pmd_t *new_pmd);
|