pmd_entry          75 arch/ia64/include/asm/pgalloc.h pmd_populate(struct mm_struct *mm, pmd_t * pmd_entry, pgtable_t pte)
pmd_entry          77 arch/ia64/include/asm/pgalloc.h 	pmd_val(*pmd_entry) = page_to_phys(pte);
pmd_entry          82 arch/ia64/include/asm/pgalloc.h pmd_populate_kernel(struct mm_struct *mm, pmd_t * pmd_entry, pte_t * pte)
pmd_entry          84 arch/ia64/include/asm/pgalloc.h 	pmd_val(*pmd_entry) = __pa(pte);
pmd_entry         143 arch/powerpc/mm/book3s64/subpage_prot.c 	.pmd_entry	= subpage_walk_pmd_entry,
pmd_entry        2530 arch/s390/mm/gmap.c 	.pmd_entry	= __zap_zero_pages,
pmd_entry          63 arch/sh/include/asm/pgtable_64.h #define pmd_present(pmd_entry)	(pmd_val(pmd_entry) & _PAGE_PRESENT)
pmd_entry          65 arch/sh/include/asm/pgtable_64.h #define pmd_none(pmd_entry)	(pmd_val((pmd_entry)) == _PMD_EMPTY)
pmd_entry          66 arch/sh/include/asm/pgtable_64.h #define pmd_bad(pmd_entry)	((pmd_val(pmd_entry) & (~PAGE_MASK & ~_PAGE_USER)) != _KERNPG_TABLE)
pmd_entry          68 arch/sh/include/asm/pgtable_64.h #define pmd_page_vaddr(pmd_entry) \
pmd_entry          69 arch/sh/include/asm/pgtable_64.h 	((unsigned long) __va(pmd_val(pmd_entry) & PAGE_MASK))
pmd_entry         122 arch/x86/kernel/head64.c 	pmdval_t *pmd, pmd_entry;
pmd_entry         207 arch/x86/kernel/head64.c 	pmd_entry = __PAGE_KERNEL_LARGE_EXEC & ~_PAGE_GLOBAL;
pmd_entry         210 arch/x86/kernel/head64.c 	pmd_entry &= *mask_ptr;
pmd_entry         211 arch/x86/kernel/head64.c 	pmd_entry += sme_get_me_mask();
pmd_entry         212 arch/x86/kernel/head64.c 	pmd_entry +=  physaddr;
pmd_entry         217 arch/x86/kernel/head64.c 		pmd[idx % PTRS_PER_PMD] = pmd_entry + i * PMD_SIZE;
pmd_entry        1424 fs/dax.c       	pmd_t pmd_entry;
pmd_entry        1452 fs/dax.c       	pmd_entry = mk_pmd(zero_page, vmf->vma->vm_page_prot);
pmd_entry        1453 fs/dax.c       	pmd_entry = pmd_mkhuge(pmd_entry);
pmd_entry        1454 fs/dax.c       	set_pmd_at(vmf->vma->vm_mm, pmd_addr, vmf->pmd, pmd_entry);
pmd_entry         740 fs/proc/task_mmu.c 	.pmd_entry		= smaps_pte_range,
pmd_entry         745 fs/proc/task_mmu.c 	.pmd_entry		= smaps_pte_range,
pmd_entry        1130 fs/proc/task_mmu.c 	.pmd_entry		= clear_refs_pte_range,
pmd_entry        1505 fs/proc/task_mmu.c 	.pmd_entry	= pagemap_pmd_range,
pmd_entry        1810 fs/proc/task_mmu.c 	.pmd_entry = gather_pte_stats,
pmd_entry          31 include/linux/pagewalk.h 	int (*pmd_entry)(pmd_t *pmd, unsigned long addr,
pmd_entry         857 mm/hmm.c       	.pmd_entry	= hmm_vma_walk_pmd,
pmd_entry        5020 mm/hugetlb.c   	pmd_t *pmd, pmd_entry;
pmd_entry        5038 mm/hugetlb.c   	pmd_entry = READ_ONCE(*pmd);
pmd_entry        5039 mm/hugetlb.c   	if (sz != PMD_SIZE && pmd_none(pmd_entry))
pmd_entry        5042 mm/hugetlb.c   	if (pmd_huge(pmd_entry) || !pmd_present(pmd_entry))
pmd_entry         219 mm/madvise.c   	.pmd_entry		= swapin_walk_pmd_entry,
pmd_entry         474 mm/madvise.c   	.pmd_entry = madvise_cold_or_pageout_pte_range,
pmd_entry         696 mm/madvise.c   	.pmd_entry		= madvise_free_pte_range,
pmd_entry        5711 mm/memcontrol.c 	.pmd_entry	= mem_cgroup_count_precharge_pte_range,
pmd_entry        5989 mm/memcontrol.c 	.pmd_entry	= mem_cgroup_move_charge_pte_range,
pmd_entry         660 mm/mempolicy.c 	.pmd_entry		= queue_pages_pte_range,
pmd_entry        2353 mm/migrate.c   	.pmd_entry		= migrate_vma_collect_pmd,
pmd_entry         197 mm/mincore.c   	.pmd_entry		= mincore_pte_range,
pmd_entry          52 mm/pagewalk.c  		if (ops->pmd_entry)
pmd_entry          53 mm/pagewalk.c  			err = ops->pmd_entry(pmd, addr, next, walk);
pmd_entry         111 mm/pagewalk.c  		if (ops->pmd_entry || ops->pte_entry)
pmd_entry         138 mm/pagewalk.c  		if (ops->pmd_entry || ops->pte_entry)
pmd_entry         165 mm/pagewalk.c  		if (ops->pmd_entry || ops->pte_entry)