PMD_SIZE 34 arch/alpha/include/asm/pgtable.h #define PMD_MASK (~(PMD_SIZE-1)) PMD_SIZE 829 arch/arc/mm/tlb.c BUILD_BUG_ON(!IS_ALIGNED((CONFIG_ARC_KVADDR_SIZE << 20), PMD_SIZE)); PMD_SIZE 835 arch/arc/mm/tlb.c BUILD_BUG_ON(!IS_ALIGNED(STACK_TOP, PMD_SIZE)); PMD_SIZE 7 arch/arm/include/asm/highmem.h #define PKMAP_BASE (PAGE_OFFSET - PMD_SIZE) PMD_SIZE 312 arch/arm/include/asm/kvm_mmu.h unsigned long size = PMD_SIZE; PMD_SIZE 58 arch/arm/include/asm/memory.h #define MODULES_END (PAGE_OFFSET - PMD_SIZE) PMD_SIZE 86 arch/arm/include/asm/pgtable-2level.h #define PMD_MASK (~(PMD_SIZE-1)) PMD_SIZE 49 arch/arm/include/asm/stage2_pgtable.h phys_addr_t boundary = (addr + PMD_SIZE) & PMD_MASK; PMD_SIZE 66 arch/arm/include/asm/stage2_pgtable.h #define S2_PMD_SIZE PMD_SIZE PMD_SIZE 454 arch/arm/mm/dma-mapping.c addr += PMD_SIZE) PMD_SIZE 342 arch/arm/mm/dump.c addr = start + i * PMD_SIZE; PMD_SIZE 349 arch/arm/mm/dump.c if (SECTION_SIZE < PMD_SIZE && pmd_large(pmd[1])) { PMD_SIZE 173 arch/arm/mm/ioremap.c addr += PMD_SIZE; PMD_SIZE 212 arch/arm/mm/ioremap.c addr += PMD_SIZE; PMD_SIZE 249 arch/arm/mm/ioremap.c addr += PMD_SIZE; PMD_SIZE 1094 arch/arm/mm/mmu.c next = (addr + PMD_SIZE - 1) & PMD_MASK; PMD_SIZE 1186 arch/arm/mm/mmu.c if (!IS_ALIGNED(reg->base, PMD_SIZE)) { PMD_SIZE 1189 arch/arm/mm/mmu.c len = round_up(reg->base, PMD_SIZE) - reg->base; PMD_SIZE 1229 arch/arm/mm/mmu.c if (!IS_ALIGNED(block_start, PMD_SIZE)) PMD_SIZE 1231 arch/arm/mm/mmu.c else if (!IS_ALIGNED(block_end, PMD_SIZE)) PMD_SIZE 1250 arch/arm/mm/mmu.c memblock_limit = round_down(memblock_limit, PMD_SIZE); PMD_SIZE 1275 arch/arm/mm/mmu.c for (addr = 0; addr < MODULES_VADDR; addr += PMD_SIZE) PMD_SIZE 1280 arch/arm/mm/mmu.c addr = ((unsigned long)_exiprom + PMD_SIZE - 1) & PMD_MASK; PMD_SIZE 1282 arch/arm/mm/mmu.c for ( ; addr < PAGE_OFFSET; addr += PMD_SIZE) PMD_SIZE 1297 arch/arm/mm/mmu.c addr < VMALLOC_START; addr += PMD_SIZE) PMD_SIZE 1352 arch/arm/mm/mmu.c for (addr = VMALLOC_START; addr < (FIXADDR_TOP & PMD_MASK); addr += PMD_SIZE) PMD_SIZE 353 arch/arm64/include/asm/kvm_mmu.h kvm_flush_dcache_to_poc(page_address(page), PMD_SIZE); PMD_SIZE 51 arch/arm64/include/asm/pgtable-hwdef.h #define PMD_MASK (~(PMD_SIZE-1)) PMD_SIZE 99 arch/arm64/include/asm/pgtable-hwdef.h #define CONT_PMD_SIZE (CONT_PMDS * PMD_SIZE) PMD_SIZE 31 arch/arm64/mm/hugetlbpage.c case PMD_SIZE: PMD_SIZE 77 arch/arm64/mm/hugetlbpage.c *pgsize = PMD_SIZE; PMD_SIZE 93 arch/arm64/mm/hugetlbpage.c case PMD_SIZE: PMD_SIZE 97 arch/arm64/mm/hugetlbpage.c *pgsize = PMD_SIZE; PMD_SIZE 245 arch/arm64/mm/hugetlbpage.c } else if (sz == PMD_SIZE) { PMD_SIZE 285 arch/arm64/mm/hugetlbpage.c if (!(sz == PMD_SIZE || sz == CONT_PMD_SIZE) && PMD_SIZE 306 arch/arm64/mm/hugetlbpage.c } else if (pagesize != PUD_SIZE && pagesize != PMD_SIZE) { PMD_SIZE 460 arch/arm64/mm/hugetlbpage.c add_huge_page_size(PMD_SIZE); PMD_SIZE 476 arch/arm64/mm/hugetlbpage.c case PMD_SIZE: PMD_SIZE 760 arch/arm64/mm/mmu.c p = vmemmap_alloc_block_buf(PMD_SIZE, node); PMD_SIZE 1039 arch/arm64/mm/mmu.c } while (pmdp++, next += PMD_SIZE, next != end); PMD_SIZE 152 arch/csky/mm/highmem.c vaddr += PMD_SIZE; PMD_SIZE 100 arch/ia64/include/asm/pgtable.h #define PMD_MASK (~(PMD_SIZE-1)) PMD_SIZE 403 arch/ia64/mm/init.c end_address += PMD_SIZE; PMD_SIZE 39 arch/m68k/include/asm/pgtable_mm.h #define PMD_MASK (~(PMD_SIZE-1)) PMD_SIZE 118 arch/m68k/sun3x/dvma.c end3 = (vaddr + (PMD_SIZE-1)) & PMD_MASK; PMD_SIZE 144 arch/microblaze/include/asm/pgtable.h #define PMD_MASK (~(PMD_SIZE-1)) PMD_SIZE 52 arch/mips/include/asm/pgtable-64.h #define PMD_MASK (~(PMD_SIZE-1)) PMD_SIZE 263 arch/mips/mm/init.c vaddr += PMD_SIZE; PMD_SIZE 31 arch/mips/mm/ioremap.c if (end > PMD_SIZE) PMD_SIZE 32 arch/mips/mm/ioremap.c end = PMD_SIZE; PMD_SIZE 63 arch/mips/mm/ioremap.c address = (address + PMD_SIZE) & PMD_MASK; PMD_SIZE 46 arch/nds32/include/asm/pgtable.h #define PMD_MASK (~(PMD_SIZE-1)) PMD_SIZE 100 arch/nds32/kernel/vdso.c end = (start + vdso_mapping_len + PMD_SIZE - 1) & PMD_MASK; PMD_SIZE 33 arch/nios2/mm/ioremap.c if (end > PMD_SIZE) PMD_SIZE 34 arch/nios2/mm/ioremap.c end = PMD_SIZE; PMD_SIZE 70 arch/nios2/mm/ioremap.c address = (address + PMD_SIZE) & PMD_MASK; PMD_SIZE 137 arch/parisc/include/asm/pgtable.h #define PMD_MASK (~(PMD_SIZE-1)) PMD_SIZE 86 arch/parisc/kernel/pci-dma.c if (end > PMD_SIZE) PMD_SIZE 87 arch/parisc/kernel/pci-dma.c end = PMD_SIZE; PMD_SIZE 121 arch/parisc/kernel/pci-dma.c vaddr = (vaddr + PMD_SIZE) & PMD_MASK; PMD_SIZE 122 arch/parisc/kernel/pci-dma.c orig_vaddr += PMD_SIZE; PMD_SIZE 166 arch/parisc/kernel/pci-dma.c if (end > PMD_SIZE) PMD_SIZE 167 arch/parisc/kernel/pci-dma.c end = PMD_SIZE; PMD_SIZE 206 arch/parisc/kernel/pci-dma.c vaddr = (vaddr + PMD_SIZE) & PMD_MASK; PMD_SIZE 207 arch/parisc/kernel/pci-dma.c orig_vaddr += PMD_SIZE; PMD_SIZE 237 arch/powerpc/include/asm/book3s/64/pgtable.h #define PMD_MASK (~(PMD_SIZE-1)) PMD_SIZE 32 arch/powerpc/include/asm/nohash/64/pgtable-4k.h #define PMD_MASK (~(PMD_SIZE-1)) PMD_SIZE 840 arch/powerpc/kvm/book3s_64_mmu_radix.c (gpa & (PMD_SIZE - PAGE_SIZE)) == PMD_SIZE 841 arch/powerpc/kvm/book3s_64_mmu_radix.c (hva & (PMD_SIZE - PAGE_SIZE))) { PMD_SIZE 1293 arch/powerpc/kvm/book3s_64_mmu_radix.c gpa = (gpa & PMD_MASK) + PMD_SIZE; PMD_SIZE 241 arch/powerpc/mm/book3s64/hash_tlb.c addr = _ALIGN_DOWN(addr, PMD_SIZE); PMD_SIZE 87 arch/powerpc/mm/book3s64/radix_pgtable.c if (map_page_size == PMD_SIZE) { PMD_SIZE 148 arch/powerpc/mm/book3s64/radix_pgtable.c if (map_page_size == PMD_SIZE) { PMD_SIZE 275 arch/powerpc/mm/book3s64/radix_pgtable.c } else if (IS_ALIGNED(addr, PMD_SIZE) && gap >= PMD_SIZE && PMD_SIZE 277 arch/powerpc/mm/book3s64/radix_pgtable.c mapping_size = PMD_SIZE; PMD_SIZE 805 arch/powerpc/mm/book3s64/radix_pgtable.c split_kernel_mapping(addr, end, PMD_SIZE, (pte_t *)pmd); PMD_SIZE 1177 arch/powerpc/mm/book3s64/radix_pgtable.c flush_tlb_kernel_range(addr, addr + PMD_SIZE); PMD_SIZE 891 arch/powerpc/mm/book3s64/radix_tlb.c hstart = (start + PMD_SIZE - 1) & PMD_MASK; PMD_SIZE 909 arch/powerpc/mm/book3s64/radix_tlb.c PMD_SIZE, MMU_PAGE_2M); PMD_SIZE 919 arch/powerpc/mm/book3s64/radix_tlb.c PMD_SIZE, MMU_PAGE_2M); PMD_SIZE 930 arch/powerpc/mm/book3s64/radix_tlb.c hstart, hend, pid, PMD_SIZE, MMU_PAGE_2M, false); PMD_SIZE 415 arch/powerpc/mm/nohash/tlb.c unsigned long end = address + PMD_SIZE; PMD_SIZE 413 arch/powerpc/mm/ptdump/hashpagetable.c addr = start + i * PMD_SIZE; PMD_SIZE 272 arch/powerpc/mm/ptdump/ptdump.c addr = start + i * PMD_SIZE; PMD_SIZE 277 arch/powerpc/mm/ptdump/ptdump.c note_page(st, addr, 3, pmd_val(*pmd), PMD_SIZE); PMD_SIZE 19 arch/riscv/include/asm/pgtable-64.h #define PMD_MASK (~(PMD_SIZE - 1)) PMD_SIZE 110 arch/riscv/include/asm/pgtable.h #define FIXADDR_SIZE PMD_SIZE PMD_SIZE 253 arch/riscv/mm/init.c if (sz == PMD_SIZE) { PMD_SIZE 277 arch/riscv/mm/init.c #define PTE_PARENT_SIZE PMD_SIZE PMD_SIZE 374 arch/riscv/mm/init.c (uintptr_t)fixmap_pte, PMD_SIZE, PAGE_TABLE); PMD_SIZE 379 arch/riscv/mm/init.c load_pa, PMD_SIZE, PAGE_KERNEL_EXEC); PMD_SIZE 177 arch/s390/mm/dump_pagetables.c addr += PMD_SIZE; PMD_SIZE 343 arch/s390/mm/gmap.c offset = (offset & (PTRS_PER_PMD - 1)) * PMD_SIZE; PMD_SIZE 402 arch/s390/mm/gmap.c if ((to | len) & (PMD_SIZE - 1)) PMD_SIZE 409 arch/s390/mm/gmap.c for (off = 0; off < len; off += PMD_SIZE) PMD_SIZE 434 arch/s390/mm/gmap.c if ((from | to | len) & (PMD_SIZE - 1)) PMD_SIZE 442 arch/s390/mm/gmap.c for (off = 0; off < len; off += PMD_SIZE) { PMD_SIZE 701 arch/s390/mm/gmap.c gaddr = (gaddr + PMD_SIZE) & PMD_MASK) { PMD_SIZE 719 arch/s390/mm/gmap.c size = min(to - gaddr, PMD_SIZE - (gaddr & ~PMD_MASK)); PMD_SIZE 144 arch/s390/mm/hugetlbpage.c size = PMD_SIZE; PMD_SIZE 207 arch/s390/mm/hugetlbpage.c else if (sz == PMD_SIZE) PMD_SIZE 263 arch/s390/mm/hugetlbpage.c if (MACHINE_HAS_EDAT1 && size == PMD_SIZE) { PMD_SIZE 150 arch/s390/mm/kasan_init.c IS_ALIGNED(address, PMD_SIZE) && PMD_SIZE 151 arch/s390/mm/kasan_init.c end - address >= PMD_SIZE) { PMD_SIZE 154 arch/s390/mm/kasan_init.c address = (address + PMD_SIZE) & PMD_MASK; PMD_SIZE 158 arch/s390/mm/kasan_init.c if (has_edat && address && end - address >= PMD_SIZE && PMD_SIZE 169 arch/s390/mm/kasan_init.c address = (address + PMD_SIZE) & PMD_MASK; PMD_SIZE 176 arch/s390/mm/kasan_init.c address = (address + PMD_SIZE) & PMD_MASK; PMD_SIZE 168 arch/s390/mm/pageattr.c if (addr & ~PMD_MASK || addr + PMD_SIZE > next) { PMD_SIZE 206 arch/s390/mm/pageattr.c pmd_addr += PMD_SIZE; PMD_SIZE 123 arch/s390/mm/vmem.c !(address & ~PMD_MASK) && (address + PMD_SIZE <= end) && PMD_SIZE 126 arch/s390/mm/vmem.c address += PMD_SIZE; PMD_SIZE 190 arch/s390/mm/vmem.c address += PMD_SIZE; PMD_SIZE 195 arch/s390/mm/vmem.c address += PMD_SIZE; PMD_SIZE 267 arch/s390/mm/vmem.c new_page = vmemmap_alloc_block(PMD_SIZE, node); PMD_SIZE 271 arch/s390/mm/vmem.c address = (address + PMD_SIZE) & PMD_MASK; PMD_SIZE 279 arch/s390/mm/vmem.c address = (address + PMD_SIZE) & PMD_MASK; PMD_SIZE 25 arch/sh/include/asm/pgtable-3level.h #define PMD_MASK (~(PMD_SIZE-1)) PMD_SIZE 27 arch/sh/include/asm/pgtable-3level.h #define PTRS_PER_PMD ((1 << PGDIR_SHIFT) / PMD_SIZE) PMD_SIZE 190 arch/sh/mm/init.c vaddr += PMD_SIZE; PMD_SIZE 330 arch/sh/mm/init.c end = (FIXADDR_TOP + PMD_SIZE - 1) & PMD_MASK; PMD_SIZE 39 arch/sparc/include/asm/pgtable_32.h #define PMD_MASK (~(PMD_SIZE-1)) PMD_SIZE 53 arch/sparc/include/asm/pgtable_64.h #define PMD_MASK (~(PMD_SIZE-1)) PMD_SIZE 292 arch/sparc/mm/hugetlbpage.c if (sz >= PMD_SIZE) PMD_SIZE 332 arch/sparc/mm/hugetlbpage.c else if (size >= PMD_SIZE) PMD_SIZE 369 arch/sparc/mm/hugetlbpage.c else if (size >= PMD_SIZE) PMD_SIZE 499 arch/sparc/mm/hugetlbpage.c addr += PMD_SIZE; PMD_SIZE 509 arch/sparc/mm/hugetlbpage.c end -= PMD_SIZE; PMD_SIZE 454 arch/sparc/mm/init_64.c } else if (hugepage_size >= PMD_SIZE) { PMD_SIZE 461 arch/sparc/mm/init_64.c if (hugepage_size >= PMD_SIZE) { PMD_SIZE 1752 arch/sparc/mm/init_64.c return vstart + PMD_SIZE; PMD_SIZE 1769 arch/sparc/mm/init_64.c pte_val += PMD_SIZE; PMD_SIZE 1770 arch/sparc/mm/init_64.c vstart += PMD_SIZE; PMD_SIZE 1780 arch/sparc/mm/init_64.c if (guard && !(vstart & ~PMD_MASK) && (vend - vstart) >= PMD_SIZE) PMD_SIZE 1850 arch/sparc/mm/init_64.c this_end = (vstart + PMD_SIZE) & PMD_MASK; PMD_SIZE 2611 arch/sparc/mm/init_64.c vend = ALIGN(vend, PMD_SIZE); PMD_SIZE 2612 arch/sparc/mm/init_64.c for (; vstart < vend; vstart += PMD_SIZE) { PMD_SIZE 2628 arch/sparc/mm/init_64.c void *block = vmemmap_alloc_block(PMD_SIZE, node); PMD_SIZE 717 arch/sparc/mm/srmmu.c if (start > (0xffffffffUL - PMD_SIZE)) PMD_SIZE 719 arch/sparc/mm/srmmu.c start = (start + PMD_SIZE) & PMD_MASK; PMD_SIZE 748 arch/sparc/mm/srmmu.c if (start > (0xffffffffUL - PMD_SIZE)) PMD_SIZE 750 arch/sparc/mm/srmmu.c start = (start + PMD_SIZE) & PMD_MASK; PMD_SIZE 29 arch/um/include/asm/pgtable-3level.h #define PMD_MASK (~(PMD_SIZE-1)) PMD_SIZE 116 arch/um/kernel/mem.c vaddr += PMD_SIZE; PMD_SIZE 385 arch/um/kernel/tlb.c last = ADD_ROUND(addr, PMD_SIZE); PMD_SIZE 423 arch/x86/boot/compressed/kaslr.c add_identity_map(0, PMD_SIZE); PMD_SIZE 135 arch/x86/boot/compressed/kaslr_64.c start = round_down(start, PMD_SIZE); PMD_SIZE 136 arch/x86/boot/compressed/kaslr_64.c end = round_up(end, PMD_SIZE); PMD_SIZE 224 arch/x86/entry/vdso/vma.c end = (start + len + PMD_SIZE - 1) & PMD_MASK; PMD_SIZE 13 arch/x86/include/asm/pgtable_32_types.h # define PMD_MASK (~(PMD_SIZE - 1)) PMD_SIZE 56 arch/x86/include/asm/pgtable_32_types.h #define LDT_END_ADDR (LDT_BASE_ADDR + PMD_SIZE) PMD_SIZE 99 arch/x86/include/asm/pgtable_64_types.h #define PMD_MASK (~(PMD_SIZE - 1)) PMD_SIZE 214 arch/x86/kernel/head64.c for (i = 0; i < DIV_ROUND_UP(_end - _text, PMD_SIZE); i++) { PMD_SIZE 217 arch/x86/kernel/head64.c pmd[idx % PTRS_PER_PMD] = pmd_entry + i * PMD_SIZE; PMD_SIZE 269 arch/x86/kernel/head64.c for (; vaddr < vaddr_end; vaddr += PMD_SIZE) { PMD_SIZE 201 arch/x86/kernel/setup_percpu.c atom_size = PMD_SIZE; PMD_SIZE 190 arch/x86/mm/cpu_entry_area.c for (; start < end && start >= CPU_ENTRY_AREA_BASE; start += PMD_SIZE) PMD_SIZE 201 arch/x86/mm/fault.c address += PMD_SIZE) { PMD_SIZE 192 arch/x86/mm/hugetlbpage.c if (ps == PMD_SIZE) { PMD_SIZE 11 arch/x86/mm/ident_map.c for (; addr < end; addr += PMD_SIZE) { PMD_SIZE 289 arch/x86/mm/init.c unsigned long start = round_down(mr[i].start, PMD_SIZE); PMD_SIZE 290 arch/x86/mm/init.c unsigned long end = round_up(mr[i].end, PMD_SIZE); PMD_SIZE 356 arch/x86/mm/init.c end_pfn = PFN_DOWN(PMD_SIZE); PMD_SIZE 358 arch/x86/mm/init.c end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); PMD_SIZE 360 arch/x86/mm/init.c end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); PMD_SIZE 370 arch/x86/mm/init.c start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); PMD_SIZE 372 arch/x86/mm/init.c end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); PMD_SIZE 375 arch/x86/mm/init.c if (end_pfn > round_down(limit_pfn, PFN_DOWN(PMD_SIZE))) PMD_SIZE 376 arch/x86/mm/init.c end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); PMD_SIZE 397 arch/x86/mm/init.c start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); PMD_SIZE 398 arch/x86/mm/init.c end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); PMD_SIZE 570 arch/x86/mm/init.c addr = memblock_find_in_range(map_start, map_end, PMD_SIZE, PMD_SIZE); PMD_SIZE 571 arch/x86/mm/init.c real_end = addr + PMD_SIZE; PMD_SIZE 574 arch/x86/mm/init.c step_size = PMD_SIZE; PMD_SIZE 621 arch/x86/mm/init.c unsigned long step_size = PMD_SIZE; PMD_SIZE 149 arch/x86/mm/init_32.c vaddr += PMD_SIZE; PMD_SIZE 234 arch/x86/mm/init_32.c vaddr += PMD_SIZE; PMD_SIZE 549 arch/x86/mm/init_32.c end = (FIXADDR_TOP + PMD_SIZE - 1) & PMD_MASK; PMD_SIZE 372 arch/x86/mm/init_64.c for (; size; phys += PMD_SIZE, size -= PMD_SIZE) { PMD_SIZE 424 arch/x86/mm/init_64.c unsigned long end = roundup((unsigned long)_brk_end, PMD_SIZE) - 1; PMD_SIZE 435 arch/x86/mm/init_64.c for (; vaddr + PMD_SIZE - 1 < vaddr_end; pmd++, vaddr += PMD_SIZE) { PMD_SIZE 515 arch/x86/mm/init_64.c paddr_next = (paddr & PMD_MASK) + PMD_SIZE; PMD_SIZE 898 arch/x86/mm/init_64.c vmem_altmap_free(altmap, PMD_SIZE / PAGE_SIZE); PMD_SIZE 900 arch/x86/mm/init_64.c free_pagetable(page, get_order(PMD_SIZE)); PMD_SIZE 1045 arch/x86/mm/init_64.c if (IS_ALIGNED(addr, PMD_SIZE) && PMD_SIZE 1046 arch/x86/mm/init_64.c IS_ALIGNED(next, PMD_SIZE)) { PMD_SIZE 1061 arch/x86/mm/init_64.c PMD_SIZE)) { PMD_SIZE 1324 arch/x86/mm/init_64.c all_end = roundup((unsigned long)_brk_end, PMD_SIZE); PMD_SIZE 1479 arch/x86/mm/init_64.c p = altmap_alloc_block_buf(PMD_SIZE, altmap); PMD_SIZE 1481 arch/x86/mm/init_64.c p = vmemmap_alloc_block_buf(PMD_SIZE, node); PMD_SIZE 1499 arch/x86/mm/init_64.c addr_end = addr + PMD_SIZE; PMD_SIZE 1500 arch/x86/mm/init_64.c p_end = p + PMD_SIZE; PMD_SIZE 1592 arch/x86/mm/init_64.c nr_pmd_pages = 1 << get_order(PMD_SIZE); PMD_SIZE 48 arch/x86/mm/kasan_init_64.c ((end - addr) == PMD_SIZE) && PMD_SIZE 49 arch/x86/mm/kasan_init_64.c IS_ALIGNED(addr, PMD_SIZE)) { PMD_SIZE 50 arch/x86/mm/kasan_init_64.c p = early_alloc(PMD_SIZE, nid, false); PMD_SIZE 54 arch/x86/mm/kasan_init_64.c memblock_free(__pa(p), PMD_SIZE); PMD_SIZE 132 arch/x86/mm/mem_encrypt.c vaddr += PMD_SIZE; PMD_SIZE 133 arch/x86/mm/mem_encrypt.c paddr += PMD_SIZE; PMD_SIZE 134 arch/x86/mm/mem_encrypt.c size = (size <= PMD_SIZE) ? 0 : size - PMD_SIZE; PMD_SIZE 213 arch/x86/mm/pageattr.c return __pa_symbol(roundup(_brk_end, PMD_SIZE) - 1) >> PAGE_SHIFT; PMD_SIZE 981 arch/x86/mm/pageattr.c lpinc = PMD_SIZE; PMD_SIZE 1121 arch/x86/mm/pageattr.c if (start & (PMD_SIZE - 1)) { PMD_SIZE 1122 arch/x86/mm/pageattr.c unsigned long next_page = (start + PMD_SIZE) & PMD_MASK; PMD_SIZE 1134 arch/x86/mm/pageattr.c while (end - start >= PMD_SIZE) { PMD_SIZE 1138 arch/x86/mm/pageattr.c __unmap_pmd_range(pud, pmd, start, start + PMD_SIZE); PMD_SIZE 1140 arch/x86/mm/pageattr.c start += PMD_SIZE; PMD_SIZE 1251 arch/x86/mm/pageattr.c if (start & (PMD_SIZE - 1)) { PMD_SIZE 1253 arch/x86/mm/pageattr.c unsigned long next_page = (start + PMD_SIZE) & PMD_MASK; PMD_SIZE 1280 arch/x86/mm/pageattr.c while (end - start >= PMD_SIZE) { PMD_SIZE 1294 arch/x86/mm/pageattr.c start += PMD_SIZE; PMD_SIZE 1295 arch/x86/mm/pageattr.c cpa->pfn += PMD_SIZE >> PAGE_SHIFT; PMD_SIZE 1296 arch/x86/mm/pageattr.c cur_pages += PMD_SIZE >> PAGE_SHIFT; PMD_SIZE 729 arch/x86/mm/pgtable.c mtrr = mtrr_type_lookup(addr, addr + PMD_SIZE, &uniform); PMD_SIZE 733 arch/x86/mm/pgtable.c __func__, addr, addr + PMD_SIZE); PMD_SIZE 341 arch/x86/mm/pti.c addr = round_up(addr + 1, PMD_SIZE); PMD_SIZE 378 arch/x86/mm/pti.c addr += PMD_SIZE; PMD_SIZE 458 arch/x86/platform/efi/efi_64.c if (!(pa & (PMD_SIZE - 1))) { PMD_SIZE 461 arch/x86/platform/efi/efi_64.c u64 pa_offset = pa & (PMD_SIZE - 1); PMD_SIZE 468 arch/x86/platform/efi/efi_64.c efi_va -= PMD_SIZE; PMD_SIZE 1096 arch/x86/xen/mmu_pv.c unsigned long kernel_end = roundup((unsigned long)_brk_end, PMD_SIZE) - 1; PMD_SIZE 1102 arch/x86/xen/mmu_pv.c pmd++, vaddr += PMD_SIZE) { PMD_SIZE 1145 arch/x86/xen/mmu_pv.c xen_free_ro_pages(pa, PMD_SIZE); PMD_SIZE 1244 arch/x86/xen/mmu_pv.c size = roundup(size, PMD_SIZE); PMD_SIZE 1273 arch/x86/xen/mmu_pv.c xen_cleanhighmap(addr, roundup(addr + size, PMD_SIZE * 2)); PMD_SIZE 2069 arch/x86/xen/mmu_pv.c n_pt = roundup(size, PMD_SIZE) >> PMD_SHIFT; PMD_SIZE 409 arch/x86/xen/p2m.c (unsigned long)(p2m + pfn) + i * PMD_SIZE); PMD_SIZE 425 arch/x86/xen/p2m.c PMD_SIZE * PMDS_PER_MID_PAGE); PMD_SIZE 426 arch/x86/xen/p2m.c vm_area_register_early(&vm, PMD_SIZE * PMDS_PER_MID_PAGE); PMD_SIZE 493 arch/x86/xen/p2m.c vaddr = addr & ~(PMD_SIZE * PMDS_PER_MID_PAGE - 1); PMD_SIZE 522 arch/x86/xen/p2m.c vaddr += PMD_SIZE; PMD_SIZE 31 arch/xtensa/mm/kasan_init.c for (vaddr = 0; vaddr < KASAN_SHADOW_SIZE; vaddr += PMD_SIZE, ++pmd) { PMD_SIZE 118 drivers/dax/device.c unsigned int fault_size = PMD_SIZE; PMD_SIZE 124 drivers/dax/device.c if (dax_region->align > PMD_SIZE) { PMD_SIZE 143 drivers/dax/device.c (pmd_addr + PMD_SIZE) > vmf->vma->vm_end) PMD_SIZE 147 drivers/dax/device.c phys = dax_pgoff_to_phys(dev_dax, pgoff, PMD_SIZE); PMD_SIZE 236 drivers/dax/device.c fault_size = PMD_SIZE; PMD_SIZE 49 fs/dax.c #define PG_PMD_COLOUR ((PMD_SIZE >> PAGE_SHIFT) - 1) PMD_SIZE 50 fs/dax.c #define PG_PMD_NR (PMD_SIZE >> PAGE_SHIFT) PMD_SIZE 308 fs/dax.c return PMD_SIZE; PMD_SIZE 1511 fs/dax.c if ((pmd_addr + PMD_SIZE) > vma->vm_end) PMD_SIZE 1553 fs/dax.c error = ops->iomap_begin(inode, pos, PMD_SIZE, iomap_flags, &iomap); PMD_SIZE 1557 fs/dax.c if (iomap.offset + iomap.length < pos + PMD_SIZE) PMD_SIZE 1564 fs/dax.c error = dax_iomap_pfn(&iomap, pos, PMD_SIZE, &pfn); PMD_SIZE 1585 fs/dax.c trace_dax_pmd_insert_mapping(inode, vmf, PMD_SIZE, pfn, entry); PMD_SIZE 1601 fs/dax.c int copied = PMD_SIZE; PMD_SIZE 1611 fs/dax.c ops->iomap_end(inode, pos, PMD_SIZE, copied, iomap_flags, PMD_SIZE 1256 fs/proc/task_mmu.c #define PAGEMAP_WALK_SIZE (PMD_SIZE) PMD_SIZE 23 include/asm-generic/pgtable-nopmd.h #define PMD_MASK (~(PMD_SIZE-1)) PMD_SIZE 550 include/asm-generic/pgtable.h ({ unsigned long __boundary = ((addr) + PMD_SIZE) & PMD_MASK; \ PMD_SIZE 520 include/asm-generic/tlb.h if (_sz == PMD_SIZE) \ PMD_SIZE 539 include/linux/hugetlb.h if (huge_page_size(h) == PMD_SIZE) PMD_SIZE 444 include/linux/mmu_notifier.h PMD_SIZE); \ PMD_SIZE 466 include/linux/mmu_notifier.h ___address + PMD_SIZE); \ PMD_SIZE 88 lib/ioremap.c if ((end - addr) != PMD_SIZE) PMD_SIZE 91 lib/ioremap.c if (!IS_ALIGNED(addr, PMD_SIZE)) PMD_SIZE 94 lib/ioremap.c if (!IS_ALIGNED(phys_addr, PMD_SIZE)) PMD_SIZE 567 mm/huge_memory.c ret = __thp_get_unmapped_area(filp, addr, len, off, flags, PMD_SIZE); PMD_SIZE 1909 mm/huge_memory.c flush_tlb_range(vma, old_addr, old_addr + PMD_SIZE); PMD_SIZE 4993 mm/hugetlb.c BUG_ON(sz != PMD_SIZE); PMD_SIZE 5039 mm/hugetlb.c if (sz != PMD_SIZE && pmd_none(pmd_entry)) PMD_SIZE 122 mm/kasan/init.c if (IS_ALIGNED(addr, PMD_SIZE) && end - addr >= PMD_SIZE) { PMD_SIZE 390 mm/kasan/init.c if (IS_ALIGNED(addr, PMD_SIZE) && PMD_SIZE 391 mm/kasan/init.c IS_ALIGNED(next, PMD_SIZE)) PMD_SIZE 1138 mm/memory-failure.c if (huge_page_size(page_hstate(head)) > PMD_SIZE) { PMD_SIZE 343 mm/memory.c addr += PMD_SIZE; PMD_SIZE 353 mm/memory.c end -= PMD_SIZE; PMD_SIZE 391 mm/memory.c while (next && next->vm_start <= vma->vm_end + PMD_SIZE PMD_SIZE 4164 mm/memory.c (address & PMD_MASK) + PMD_SIZE); PMD_SIZE 204 mm/mremap.c || old_end - old_addr < PMD_SIZE) PMD_SIZE 231 mm/mremap.c flush_tlb_range(vma, old_addr, old_addr + PMD_SIZE); PMD_SIZE 258 mm/mremap.c next = (old_addr + PMD_SIZE) & PMD_MASK; PMD_SIZE 285 mm/mremap.c } else if (extent == PMD_SIZE) { PMD_SIZE 306 mm/mremap.c next = (new_addr + PMD_SIZE) & PMD_MASK; PMD_SIZE 229 mm/page_vma_mapped.c if (pvmw->address % PMD_SIZE == 0) { PMD_SIZE 442 mm/sparse.c return ALIGN(sizeof(struct page) * PAGES_PER_SECTION, PMD_SIZE); PMD_SIZE 637 mm/swap_state.c PFN_DOWN((faddr & PMD_MASK) + PMD_SIZE)); PMD_SIZE 1719 virt/kvm/arm/mmu.c if (vma_pagesize == PMD_SIZE || PMD_SIZE 1785 virt/kvm/arm/mmu.c if (fault_supports_stage2_huge_mapping(memslot, hva, PMD_SIZE) && PMD_SIZE 1787 virt/kvm/arm/mmu.c vma_pagesize = PMD_SIZE; PMD_SIZE 1821 virt/kvm/arm/mmu.c } else if (vma_pagesize == PMD_SIZE) { PMD_SIZE 2107 virt/kvm/arm/mmu.c WARN_ON(size != PAGE_SIZE && size != PMD_SIZE && size != PUD_SIZE); PMD_SIZE 2125 virt/kvm/arm/mmu.c WARN_ON(size != PAGE_SIZE && size != PMD_SIZE && size != PUD_SIZE);