pmd_p 963 arch/s390/mm/gmap.c int pmd_p = pmd_val(*pmdp) & _SEGMENT_ENTRY_PROTECT; pmd_p 967 arch/s390/mm/gmap.c if ((pmd_i && (prot != PROT_NONE)) || (pmd_p && (prot == PROT_WRITE))) pmd_p 975 arch/s390/mm/gmap.c if (prot == PROT_READ && !pmd_p) { pmd_p 136 arch/x86/kernel/espfix_64.c pmd_t pmd, *pmd_p; pmd_p 169 arch/x86/kernel/espfix_64.c pmd_p = (pmd_t *)page_address(page); pmd_p 170 arch/x86/kernel/espfix_64.c pud = __pud(__pa(pmd_p) | (PGTABLE_PROT & ptemask)); pmd_p 171 arch/x86/kernel/espfix_64.c paravirt_alloc_pmd(&init_mm, __pa(pmd_p) >> PAGE_SHIFT); pmd_p 176 arch/x86/kernel/espfix_64.c pmd_p = pmd_offset(&pud, addr); pmd_p 177 arch/x86/kernel/espfix_64.c pmd = *pmd_p; pmd_p 185 arch/x86/kernel/espfix_64.c set_pmd(&pmd_p[n], pmd); pmd_p 306 arch/x86/kernel/head64.c pmdval_t *pmd_p; pmd_p 354 arch/x86/kernel/head64.c pmd_p = (pmdval_t *)((pud & PTE_PFN_MASK) + __START_KERNEL_map - phys_base); pmd_p 361 arch/x86/kernel/head64.c pmd_p = (pmdval_t *)early_dynamic_pgts[next_early_pgt++]; pmd_p 362 arch/x86/kernel/head64.c memset(pmd_p, 0, sizeof(*pmd_p) * PTRS_PER_PMD); pmd_p 363 arch/x86/kernel/head64.c *pud_p = (pudval_t)pmd_p - __START_KERNEL_map + phys_base + _KERNPG_TABLE; pmd_p 365 arch/x86/kernel/head64.c pmd_p[pmd_index(address)] = pmd;