pmd_pfn 313 arch/arm/include/asm/kvm_mmu.h kvm_pfn_t pfn = pmd_pfn(pmd); pmd_pfn 719 arch/arm64/mm/mmu.c return pfn_valid(pmd_pfn(pmd)); pmd_pfn 606 arch/mips/include/asm/pgtable.h return pfn_to_page(pmd_pfn(pmd)); pmd_pfn 133 arch/powerpc/mm/book3s64/hash_hugepage.c pa = pmd_pfn(__pmd(old_pmd)) << PAGE_SHIFT; pmd_pfn 1291 arch/s390/include/asm/pgtable.h #define pmd_page(pmd) pfn_to_page(pmd_pfn(pmd)) pmd_pfn 120 arch/s390/mm/pageattr.c pte_addr = pmd_pfn(*pmdp) << PAGE_SHIFT; pmd_pfn 104 arch/sparc/mm/fault_64.c pa = pmd_pfn(*pmdp) << PAGE_SHIFT; pmd_pfn 1689 arch/sparc/mm/init_64.c return pfn_valid(pmd_pfn(*pmd)); pmd_pfn 592 arch/x86/include/asm/pgtable.h return pfn_pmd(pmd_pfn(pmd), pmd_pfn 805 arch/x86/include/asm/pgtable.h #define pmd_page(pmd) pfn_to_page(pmd_pfn(pmd)) pmd_pfn 187 arch/x86/mm/fault.c BUG_ON(pmd_pfn(*pmd) != pmd_pfn(*pmd_k)); pmd_pfn 320 arch/x86/mm/fault.c if (!low_pfn(pmd_pfn(*pmd)) || !pmd_present(*pmd) || pmd_large(*pmd)) pmd_pfn 1373 arch/x86/mm/init_64.c return pfn_valid(pmd_pfn(*pmd)); pmd_pfn 211 arch/x86/mm/mem_encrypt.c pfn = pmd_pfn(*(pmd_t *)kpte); pmd_pfn 690 arch/x86/mm/pageattr.c phys_addr = (phys_addr_t)pmd_pfn(*(pmd_t *)pte) << PAGE_SHIFT; pmd_pfn 765 arch/x86/mm/pageattr.c old_pfn = pmd_pfn(*(pmd_t *)kpte); pmd_pfn 971 arch/x86/mm/pageattr.c ref_pfn = pmd_pfn(*(pmd_t *)kpte); pmd_pfn 816 fs/dax.c if (pfn != pmd_pfn(*pmdp)) pmd_pfn 1389 fs/proc/task_mmu.c frame = pmd_pfn(pmd) + pmd_pfn 1939 mm/gup.c fault_pfn = pmd_pfn(orig) + ((addr & ~PMD_MASK) >> PAGE_SHIFT); pmd_pfn 422 mm/hmm.c pfn = pmd_pfn(pmd) + ((addr & ~PMD_MASK) >> PAGE_SHIFT); pmd_pfn 780 mm/huge_memory.c if (pmd_pfn(*pmd) != pfn_t_to_pfn(pfn)) { pmd_pfn 938 mm/huge_memory.c unsigned long pfn = pmd_pfn(*pmd); pmd_pfn 633 mm/memory.c unsigned long pfn = pmd_pfn(pmd); pmd_pfn 1108 virt/kvm/arm/mmu.c WARN_ON_ONCE(pmd_pfn(old_pmd) != pmd_pfn(*new_pmd)); pmd_pfn 1882 virt/kvm/arm/mmu.c pfn = pmd_pfn(*pmd);