pmd_entry 75 arch/ia64/include/asm/pgalloc.h pmd_populate(struct mm_struct *mm, pmd_t * pmd_entry, pgtable_t pte) pmd_entry 77 arch/ia64/include/asm/pgalloc.h pmd_val(*pmd_entry) = page_to_phys(pte); pmd_entry 82 arch/ia64/include/asm/pgalloc.h pmd_populate_kernel(struct mm_struct *mm, pmd_t * pmd_entry, pte_t * pte) pmd_entry 84 arch/ia64/include/asm/pgalloc.h pmd_val(*pmd_entry) = __pa(pte); pmd_entry 143 arch/powerpc/mm/book3s64/subpage_prot.c .pmd_entry = subpage_walk_pmd_entry, pmd_entry 2530 arch/s390/mm/gmap.c .pmd_entry = __zap_zero_pages, pmd_entry 63 arch/sh/include/asm/pgtable_64.h #define pmd_present(pmd_entry) (pmd_val(pmd_entry) & _PAGE_PRESENT) pmd_entry 65 arch/sh/include/asm/pgtable_64.h #define pmd_none(pmd_entry) (pmd_val((pmd_entry)) == _PMD_EMPTY) pmd_entry 66 arch/sh/include/asm/pgtable_64.h #define pmd_bad(pmd_entry) ((pmd_val(pmd_entry) & (~PAGE_MASK & ~_PAGE_USER)) != _KERNPG_TABLE) pmd_entry 68 arch/sh/include/asm/pgtable_64.h #define pmd_page_vaddr(pmd_entry) \ pmd_entry 69 arch/sh/include/asm/pgtable_64.h ((unsigned long) __va(pmd_val(pmd_entry) & PAGE_MASK)) pmd_entry 122 arch/x86/kernel/head64.c pmdval_t *pmd, pmd_entry; pmd_entry 207 arch/x86/kernel/head64.c pmd_entry = __PAGE_KERNEL_LARGE_EXEC & ~_PAGE_GLOBAL; pmd_entry 210 arch/x86/kernel/head64.c pmd_entry &= *mask_ptr; pmd_entry 211 arch/x86/kernel/head64.c pmd_entry += sme_get_me_mask(); pmd_entry 212 arch/x86/kernel/head64.c pmd_entry += physaddr; pmd_entry 217 arch/x86/kernel/head64.c pmd[idx % PTRS_PER_PMD] = pmd_entry + i * PMD_SIZE; pmd_entry 1424 fs/dax.c pmd_t pmd_entry; pmd_entry 1452 fs/dax.c pmd_entry = mk_pmd(zero_page, vmf->vma->vm_page_prot); pmd_entry 1453 fs/dax.c pmd_entry = pmd_mkhuge(pmd_entry); pmd_entry 1454 fs/dax.c set_pmd_at(vmf->vma->vm_mm, pmd_addr, vmf->pmd, pmd_entry); pmd_entry 740 fs/proc/task_mmu.c .pmd_entry = smaps_pte_range, pmd_entry 745 fs/proc/task_mmu.c .pmd_entry = smaps_pte_range, pmd_entry 1130 fs/proc/task_mmu.c .pmd_entry = clear_refs_pte_range, pmd_entry 1505 fs/proc/task_mmu.c .pmd_entry = pagemap_pmd_range, pmd_entry 1810 fs/proc/task_mmu.c .pmd_entry = gather_pte_stats, pmd_entry 31 include/linux/pagewalk.h int (*pmd_entry)(pmd_t *pmd, unsigned long addr, pmd_entry 857 mm/hmm.c .pmd_entry = hmm_vma_walk_pmd, pmd_entry 5020 mm/hugetlb.c pmd_t *pmd, pmd_entry; pmd_entry 5038 mm/hugetlb.c pmd_entry = READ_ONCE(*pmd); pmd_entry 5039 mm/hugetlb.c if (sz != PMD_SIZE && pmd_none(pmd_entry)) pmd_entry 5042 mm/hugetlb.c if (pmd_huge(pmd_entry) || !pmd_present(pmd_entry)) pmd_entry 219 mm/madvise.c .pmd_entry = swapin_walk_pmd_entry, pmd_entry 474 mm/madvise.c .pmd_entry = madvise_cold_or_pageout_pte_range, pmd_entry 696 mm/madvise.c .pmd_entry = madvise_free_pte_range, pmd_entry 5711 mm/memcontrol.c .pmd_entry = mem_cgroup_count_precharge_pte_range, pmd_entry 5989 mm/memcontrol.c .pmd_entry = mem_cgroup_move_charge_pte_range, pmd_entry 660 mm/mempolicy.c .pmd_entry = queue_pages_pte_range, pmd_entry 2353 mm/migrate.c .pmd_entry = migrate_vma_collect_pmd, pmd_entry 197 mm/mincore.c .pmd_entry = mincore_pte_range, pmd_entry 52 mm/pagewalk.c if (ops->pmd_entry) pmd_entry 53 mm/pagewalk.c err = ops->pmd_entry(pmd, addr, next, walk); pmd_entry 111 mm/pagewalk.c if (ops->pmd_entry || ops->pte_entry) pmd_entry 138 mm/pagewalk.c if (ops->pmd_entry || ops->pte_entry) pmd_entry 165 mm/pagewalk.c if (ops->pmd_entry || ops->pte_entry)