pmd_k 34 arch/arc/mm/fault.c pmd_t *pmd, *pmd_k; pmd_k 48 arch/arc/mm/fault.c pmd_k = pmd_offset(pud_k, address); pmd_k 49 arch/arc/mm/fault.c if (!pmd_present(*pmd_k)) pmd_k 52 arch/arc/mm/fault.c set_pmd(pmd, *pmd_k); pmd_k 115 arch/arc/mm/highmem.c pmd_t *pmd_k; pmd_k 120 arch/arc/mm/highmem.c pmd_k = pmd_offset(pud_k, kvaddr); pmd_k 127 arch/arc/mm/highmem.c pmd_populate_kernel(&init_mm, pmd_k, pte_k); pmd_k 412 arch/arm/mm/fault.c pmd_t *pmd, *pmd_k; pmd_k 439 arch/arm/mm/fault.c pmd_k = pmd_offset(pud_k, addr); pmd_k 457 arch/arm/mm/fault.c if (pmd_none(pmd_k[index])) pmd_k 460 arch/arm/mm/fault.c copy_pmd(pmd, pmd_k); pmd_k 80 arch/csky/mm/fault.c pmd_t *pmd, *pmd_k; pmd_k 99 arch/csky/mm/fault.c pmd_k = pmd_offset(pud_k, address); pmd_k 100 arch/csky/mm/fault.c if (!pmd_present(*pmd_k)) pmd_k 102 arch/csky/mm/fault.c set_pmd(pmd, *pmd_k); pmd_k 104 arch/csky/mm/fault.c pte_k = pte_offset_kernel(pmd_k, address); pmd_k 298 arch/mips/mm/fault.c pmd_t *pmd, *pmd_k; pmd_k 314 arch/mips/mm/fault.c pmd_k = pmd_offset(pud_k, address); pmd_k 315 arch/mips/mm/fault.c if (!pmd_present(*pmd_k)) pmd_k 317 arch/mips/mm/fault.c set_pmd(pmd, *pmd_k); pmd_k 319 arch/mips/mm/fault.c pte_k = pte_offset_kernel(pmd_k, address); pmd_k 363 arch/nds32/mm/fault.c pmd_t *pmd, *pmd_k; pmd_k 378 arch/nds32/mm/fault.c pmd_k = pmd_offset(pud_k, addr); pmd_k 379 arch/nds32/mm/fault.c if (!pmd_present(*pmd_k)) pmd_k 383 arch/nds32/mm/fault.c set_pmd(pmd, *pmd_k); pmd_k 385 arch/nds32/mm/fault.c BUG_ON(pmd_page(*pmd) != pmd_page(*pmd_k)); pmd_k 401 arch/nds32/mm/fault.c pte_k = pte_offset_kernel(pmd_k, addr); pmd_k 249 arch/nios2/mm/fault.c pmd_t *pmd, *pmd_k; pmd_k 264 arch/nios2/mm/fault.c pmd_k = pmd_offset(pud_k, address); pmd_k 265 arch/nios2/mm/fault.c if (!pmd_present(*pmd_k)) pmd_k 267 arch/nios2/mm/fault.c set_pmd(pmd, *pmd_k); pmd_k 269 arch/nios2/mm/fault.c pte_k = pte_offset_kernel(pmd_k, address); pmd_k 300 arch/openrisc/mm/fault.c pmd_t *pmd, *pmd_k; pmd_k 331 arch/openrisc/mm/fault.c pmd_k = pmd_offset(pud_k, address); pmd_k 333 arch/openrisc/mm/fault.c if (!pmd_present(*pmd_k)) pmd_k 336 arch/openrisc/mm/fault.c set_pmd(pmd, *pmd_k); pmd_k 344 arch/openrisc/mm/fault.c pte_k = pte_offset_kernel(pmd_k, address); pmd_k 218 arch/riscv/mm/fault.c pmd_t *pmd, *pmd_k; pmd_k 257 arch/riscv/mm/fault.c pmd_k = pmd_offset(pud_k, addr); pmd_k 258 arch/riscv/mm/fault.c if (!pmd_present(*pmd_k)) pmd_k 260 arch/riscv/mm/fault.c set_pmd(pmd, *pmd_k); pmd_k 268 arch/riscv/mm/fault.c pte_k = pte_offset_kernel(pmd_k, addr); pmd_k 140 arch/sh/kernel/io_trapped.c pmd_t *pmd_k; pmd_k 152 arch/sh/kernel/io_trapped.c pmd_k = pmd_offset(pud_k, address); pmd_k 153 arch/sh/kernel/io_trapped.c if (!pmd_present(*pmd_k)) pmd_k 156 arch/sh/kernel/io_trapped.c pte_k = pte_offset_kernel(pmd_k, address); pmd_k 111 arch/sh/mm/fault.c pmd_t *pmd, *pmd_k; pmd_k 128 arch/sh/mm/fault.c pmd_k = pmd_offset(pud_k, address); pmd_k 129 arch/sh/mm/fault.c if (!pmd_present(*pmd_k)) pmd_k 133 arch/sh/mm/fault.c set_pmd(pmd, *pmd_k); pmd_k 140 arch/sh/mm/fault.c BUG_ON(pmd_page(*pmd) != pmd_page(*pmd_k)); pmd_k 144 arch/sh/mm/fault.c return pmd_k; pmd_k 159 arch/sh/mm/fault.c pmd_t *pmd_k; pmd_k 174 arch/sh/mm/fault.c pmd_k = vmalloc_sync_one(pgd_k, address); pmd_k 175 arch/sh/mm/fault.c if (!pmd_k) pmd_k 178 arch/sh/mm/fault.c pte_k = pte_offset_kernel(pmd_k, address); pmd_k 354 arch/sparc/mm/fault_32.c pmd_t *pmd, *pmd_k; pmd_k 367 arch/sparc/mm/fault_32.c pmd_k = pmd_offset(pgd_k, address); pmd_k 369 arch/sparc/mm/fault_32.c if (pmd_present(*pmd) || !pmd_present(*pmd_k)) pmd_k 372 arch/sparc/mm/fault_32.c *pmd = *pmd_k; pmd_k 340 arch/unicore32/mm/fault.c pmd_t *pmd, *pmd_k; pmd_k 356 arch/unicore32/mm/fault.c pmd_k = pmd_offset((pud_t *) pgd_k, addr); pmd_k 359 arch/unicore32/mm/fault.c if (pmd_none(*pmd_k)) pmd_k 362 arch/unicore32/mm/fault.c set_pmd(pmd, *pmd_k); pmd_k 155 arch/x86/mm/fault.c pmd_t *pmd, *pmd_k; pmd_k 179 arch/x86/mm/fault.c pmd_k = pmd_offset(pud_k, address); pmd_k 181 arch/x86/mm/fault.c if (pmd_present(*pmd) != pmd_present(*pmd_k)) pmd_k 182 arch/x86/mm/fault.c set_pmd(pmd, *pmd_k); pmd_k 184 arch/x86/mm/fault.c if (!pmd_present(*pmd_k)) pmd_k 187 arch/x86/mm/fault.c BUG_ON(pmd_pfn(*pmd) != pmd_pfn(*pmd_k)); pmd_k 189 arch/x86/mm/fault.c return pmd_k; pmd_k 237 arch/x86/mm/fault.c pmd_t *pmd_k; pmd_k 252 arch/x86/mm/fault.c pmd_k = vmalloc_sync_one(__va(pgd_paddr), address); pmd_k 253 arch/x86/mm/fault.c if (!pmd_k) pmd_k 256 arch/x86/mm/fault.c if (pmd_large(*pmd_k)) pmd_k 259 arch/x86/mm/fault.c pte_k = pte_offset_kernel(pmd_k, address); pmd_k 200 arch/xtensa/mm/fault.c pmd_t *pmd, *pmd_k; pmd_k 215 arch/xtensa/mm/fault.c pmd_k = pmd_offset(pgd_k, address); pmd_k 216 arch/xtensa/mm/fault.c if (!pmd_present(*pmd) || !pmd_present(*pmd_k)) pmd_k 219 arch/xtensa/mm/fault.c pmd_val(*pmd) = pmd_val(*pmd_k); pmd_k 220 arch/xtensa/mm/fault.c pte_k = pte_offset_kernel(pmd_k, address);