pmd_p             963 arch/s390/mm/gmap.c 	int pmd_p = pmd_val(*pmdp) & _SEGMENT_ENTRY_PROTECT;
pmd_p             967 arch/s390/mm/gmap.c 	if ((pmd_i && (prot != PROT_NONE)) || (pmd_p && (prot == PROT_WRITE)))
pmd_p             975 arch/s390/mm/gmap.c 	if (prot == PROT_READ && !pmd_p) {
pmd_p             136 arch/x86/kernel/espfix_64.c 	pmd_t pmd, *pmd_p;
pmd_p             169 arch/x86/kernel/espfix_64.c 		pmd_p = (pmd_t *)page_address(page);
pmd_p             170 arch/x86/kernel/espfix_64.c 		pud = __pud(__pa(pmd_p) | (PGTABLE_PROT & ptemask));
pmd_p             171 arch/x86/kernel/espfix_64.c 		paravirt_alloc_pmd(&init_mm, __pa(pmd_p) >> PAGE_SHIFT);
pmd_p             176 arch/x86/kernel/espfix_64.c 	pmd_p = pmd_offset(&pud, addr);
pmd_p             177 arch/x86/kernel/espfix_64.c 	pmd = *pmd_p;
pmd_p             185 arch/x86/kernel/espfix_64.c 			set_pmd(&pmd_p[n], pmd);
pmd_p             306 arch/x86/kernel/head64.c 	pmdval_t *pmd_p;
pmd_p             354 arch/x86/kernel/head64.c 		pmd_p = (pmdval_t *)((pud & PTE_PFN_MASK) + __START_KERNEL_map - phys_base);
pmd_p             361 arch/x86/kernel/head64.c 		pmd_p = (pmdval_t *)early_dynamic_pgts[next_early_pgt++];
pmd_p             362 arch/x86/kernel/head64.c 		memset(pmd_p, 0, sizeof(*pmd_p) * PTRS_PER_PMD);
pmd_p             363 arch/x86/kernel/head64.c 		*pud_p = (pudval_t)pmd_p - __START_KERNEL_map + phys_base + _KERNPG_TABLE;
pmd_p             365 arch/x86/kernel/head64.c 	pmd_p[pmd_index(address)] = pmd;