PUD_SIZE 68 arch/arm/include/asm/stage2_pgtable.h #define S2_PUD_SIZE PUD_SIZE PUD_SIZE 365 arch/arm/mm/dump.c addr = start + i * PUD_SIZE; PUD_SIZE 361 arch/arm64/include/asm/kvm_mmu.h kvm_flush_dcache_to_poc(page_address(page), PUD_SIZE); PUD_SIZE 61 arch/arm64/include/asm/pgtable-hwdef.h #define PUD_MASK (~(PUD_SIZE-1)) PUD_SIZE 24 arch/arm64/include/asm/pgtable.h #define VMALLOC_END (- PUD_SIZE - VMEMMAP_SIZE - SZ_64K) PUD_SIZE 29 arch/arm64/mm/hugetlbpage.c case PUD_SIZE: PUD_SIZE 91 arch/arm64/mm/hugetlbpage.c case PUD_SIZE: PUD_SIZE 229 arch/arm64/mm/hugetlbpage.c if (sz == PUD_SIZE) { PUD_SIZE 273 arch/arm64/mm/hugetlbpage.c if (sz != PUD_SIZE && pud_none(pud)) PUD_SIZE 306 arch/arm64/mm/hugetlbpage.c } else if (pagesize != PUD_SIZE && pagesize != PMD_SIZE) { PUD_SIZE 457 arch/arm64/mm/hugetlbpage.c add_huge_page_size(PUD_SIZE); PUD_SIZE 473 arch/arm64/mm/hugetlbpage.c case PUD_SIZE: PUD_SIZE 1036 arch/arm64/mm/mmu.c end = addr + PUD_SIZE; PUD_SIZE 112 arch/ia64/include/asm/pgtable.h #define PUD_MASK (~(PUD_SIZE-1)) PUD_SIZE 397 arch/ia64/mm/init.c end_address += PUD_SIZE; PUD_SIZE 62 arch/mips/include/asm/pgtable-64.h #define PUD_MASK (~(PUD_SIZE-1)) PUD_SIZE 242 arch/powerpc/include/asm/book3s/64/pgtable.h #define PUD_MASK (~(PUD_SIZE-1)) PUD_SIZE 37 arch/powerpc/include/asm/nohash/64/pgtable-4k.h #define PUD_MASK (~(PUD_SIZE-1)) PUD_SIZE 836 arch/powerpc/kvm/book3s_64_mmu_radix.c (gpa & (PUD_SIZE - PAGE_SIZE)) == PUD_SIZE 837 arch/powerpc/kvm/book3s_64_mmu_radix.c (hva & (PUD_SIZE - PAGE_SIZE))) { PUD_SIZE 1281 arch/powerpc/kvm/book3s_64_mmu_radix.c gpa = (gpa & PUD_MASK) + PUD_SIZE; PUD_SIZE 77 arch/powerpc/mm/book3s64/radix_pgtable.c if (map_page_size == PUD_SIZE) { PUD_SIZE 141 arch/powerpc/mm/book3s64/radix_pgtable.c if (map_page_size == PUD_SIZE) { PUD_SIZE 271 arch/powerpc/mm/book3s64/radix_pgtable.c if (IS_ALIGNED(addr, PUD_SIZE) && gap >= PUD_SIZE && PUD_SIZE 273 arch/powerpc/mm/book3s64/radix_pgtable.c mapping_size = PUD_SIZE; PUD_SIZE 830 arch/powerpc/mm/book3s64/radix_pgtable.c split_kernel_mapping(addr, end, PUD_SIZE, (pte_t *)pud); PUD_SIZE 1131 arch/powerpc/mm/book3s64/radix_pgtable.c flush_tlb_kernel_range(addr, addr + PUD_SIZE); PUD_SIZE 898 arch/powerpc/mm/book3s64/radix_tlb.c gstart = (start + PUD_SIZE - 1) & PUD_MASK; PUD_SIZE 912 arch/powerpc/mm/book3s64/radix_tlb.c PUD_SIZE, MMU_PAGE_1G); PUD_SIZE 922 arch/powerpc/mm/book3s64/radix_tlb.c PUD_SIZE, MMU_PAGE_1G); PUD_SIZE 933 arch/powerpc/mm/book3s64/radix_tlb.c gstart, gend, pid, PUD_SIZE, MMU_PAGE_1G, false); PUD_SIZE 427 arch/powerpc/mm/ptdump/hashpagetable.c addr = start + i * PUD_SIZE; PUD_SIZE 288 arch/powerpc/mm/ptdump/ptdump.c addr = start + i * PUD_SIZE; PUD_SIZE 293 arch/powerpc/mm/ptdump/ptdump.c note_page(st, addr, 2, pud_val(*pud), PUD_SIZE); PUD_SIZE 23 arch/riscv/mm/hugetlbpage.c } else if (IS_ENABLED(CONFIG_64BIT) && ps == PUD_SIZE) { PUD_SIZE 208 arch/s390/mm/dump_pagetables.c addr += PUD_SIZE; PUD_SIZE 140 arch/s390/mm/hugetlbpage.c size = PUD_SIZE; PUD_SIZE 205 arch/s390/mm/hugetlbpage.c if (sz == PUD_SIZE) PUD_SIZE 265 arch/s390/mm/hugetlbpage.c } else if (MACHINE_HAS_EDAT2 && size == PUD_SIZE) { PUD_SIZE 136 arch/s390/mm/kasan_init.c IS_ALIGNED(address, PUD_SIZE) && PUD_SIZE 137 arch/s390/mm/kasan_init.c end - address >= PUD_SIZE) { PUD_SIZE 140 arch/s390/mm/kasan_init.c address = (address + PUD_SIZE) & PUD_MASK; PUD_SIZE 278 arch/s390/mm/kasan_init.c BUILD_BUG_ON(!IS_ALIGNED(KASAN_SHADOW_START, PUD_SIZE)); PUD_SIZE 279 arch/s390/mm/kasan_init.c BUILD_BUG_ON(!IS_ALIGNED(KASAN_SHADOW_END, PUD_SIZE)); PUD_SIZE 245 arch/s390/mm/pageattr.c if (addr & ~PUD_MASK || addr + PUD_SIZE > next) { PUD_SIZE 108 arch/s390/mm/vmem.c !(address & ~PUD_MASK) && (address + PUD_SIZE <= end) && PUD_SIZE 111 arch/s390/mm/vmem.c address += PUD_SIZE; PUD_SIZE 179 arch/s390/mm/vmem.c address += PUD_SIZE; PUD_SIZE 184 arch/s390/mm/vmem.c address += PUD_SIZE; PUD_SIZE 61 arch/sparc/include/asm/pgtable_64.h #define PUD_MASK (~(PUD_SIZE-1)) PUD_SIZE 287 arch/sparc/mm/hugetlbpage.c if (sz >= PUD_SIZE) PUD_SIZE 330 arch/sparc/mm/hugetlbpage.c if (size >= PUD_SIZE) PUD_SIZE 367 arch/sparc/mm/hugetlbpage.c if (size >= PUD_SIZE) PUD_SIZE 446 arch/sparc/mm/init_64.c if (hugepage_size >= PUD_SIZE) { PUD_SIZE 1712 arch/sparc/mm/init_64.c return vstart + PUD_SIZE; PUD_SIZE 1722 arch/sparc/mm/init_64.c pte_val += PUD_SIZE; PUD_SIZE 1723 arch/sparc/mm/init_64.c vstart += PUD_SIZE; PUD_SIZE 1732 arch/sparc/mm/init_64.c if (guard && !(vstart & ~PUD_MASK) && (vend - vstart) >= PUD_SIZE) PUD_SIZE 369 arch/um/kernel/tlb.c last = ADD_ROUND(addr, PUD_SIZE); PUD_SIZE 223 arch/x86/boot/compressed/kaslr.c if (memparse(p, &p) != PUD_SIZE) { PUD_SIZE 514 arch/x86/boot/compressed/kaslr.c addr = ALIGN(region->start, PUD_SIZE); PUD_SIZE 520 arch/x86/boot/compressed/kaslr.c while (size > PUD_SIZE && max_gb_huge_pages) { PUD_SIZE 521 arch/x86/boot/compressed/kaslr.c size -= PUD_SIZE; PUD_SIZE 544 arch/x86/boot/compressed/kaslr.c size = region->size - (addr - region->start) - i * PUD_SIZE; PUD_SIZE 546 arch/x86/boot/compressed/kaslr.c tmp.start = addr + i * PUD_SIZE; PUD_SIZE 101 arch/x86/include/asm/pgtable_64_types.h #define PUD_MASK (~(PUD_SIZE - 1)) PUD_SIZE 433 arch/x86/kernel/head64.c BUILD_BUG_ON(MODULES_LEN + KERNEL_IMAGE_SIZE > 2*PUD_SIZE); PUD_SIZE 194 arch/x86/mm/hugetlbpage.c } else if (ps == PUD_SIZE && boot_cpu_has(X86_FEATURE_GBPAGES)) { PUD_SIZE 30 arch/x86/mm/ident_map.c next = (addr & PUD_MASK) + PUD_SIZE; PUD_SIZE 302 arch/x86/mm/init.c unsigned long start = round_down(mr[i].start, PUD_SIZE); PUD_SIZE 303 arch/x86/mm/init.c unsigned long end = round_up(mr[i].end, PUD_SIZE); PUD_SIZE 374 arch/x86/mm/init.c end_pfn = round_up(pfn, PFN_DOWN(PUD_SIZE)); PUD_SIZE 387 arch/x86/mm/init.c start_pfn = round_up(pfn, PFN_DOWN(PUD_SIZE)); PUD_SIZE 388 arch/x86/mm/init.c end_pfn = round_down(limit_pfn, PFN_DOWN(PUD_SIZE)); PUD_SIZE 602 arch/x86/mm/init_64.c paddr_next = (paddr & PUD_MASK) + PUD_SIZE; PUD_SIZE 1101 arch/x86/mm/init_64.c if (IS_ALIGNED(addr, PUD_SIZE) && PUD_SIZE 1102 arch/x86/mm/init_64.c IS_ALIGNED(next, PUD_SIZE)) { PUD_SIZE 1105 arch/x86/mm/init_64.c get_order(PUD_SIZE)); PUD_SIZE 1117 arch/x86/mm/init_64.c PUD_SIZE)) { PUD_SIZE 1119 arch/x86/mm/init_64.c get_order(PUD_SIZE)); PUD_SIZE 85 arch/x86/mm/kasan_init_64.c ((end - addr) == PUD_SIZE) && PUD_SIZE 86 arch/x86/mm/kasan_init_64.c IS_ALIGNED(addr, PUD_SIZE)) { PUD_SIZE 87 arch/x86/mm/kasan_init_64.c p = early_alloc(PUD_SIZE, nid, false); PUD_SIZE 91 arch/x86/mm/kasan_init_64.c memblock_free(__pa(p), PUD_SIZE); PUD_SIZE 146 arch/x86/mm/kaslr.c vaddr = round_up(vaddr + 1, PUD_SIZE); PUD_SIZE 264 arch/x86/mm/mem_encrypt_identity.c entries += (DIV_ROUND_UP(len, PUD_SIZE) + 1) * sizeof(pmd_t) * PTRS_PER_PMD; PUD_SIZE 275 arch/x86/mm/mem_encrypt_identity.c tables += DIV_ROUND_UP(entries, PUD_SIZE) * sizeof(pmd_t) * PTRS_PER_PMD; PUD_SIZE 1165 arch/x86/mm/pageattr.c if (start & (PUD_SIZE - 1)) { PUD_SIZE 1166 arch/x86/mm/pageattr.c unsigned long next_page = (start + PUD_SIZE) & PUD_MASK; PUD_SIZE 1178 arch/x86/mm/pageattr.c while (end - start >= PUD_SIZE) { PUD_SIZE 1183 arch/x86/mm/pageattr.c unmap_pmd_range(pud, start, start + PUD_SIZE); PUD_SIZE 1185 arch/x86/mm/pageattr.c start += PUD_SIZE; PUD_SIZE 1328 arch/x86/mm/pageattr.c if (start & (PUD_SIZE - 1)) { PUD_SIZE 1330 arch/x86/mm/pageattr.c unsigned long next_page = (start + PUD_SIZE) & PUD_MASK; PUD_SIZE 1363 arch/x86/mm/pageattr.c while (boot_cpu_has(X86_FEATURE_GBPAGES) && end - start >= PUD_SIZE) { PUD_SIZE 1367 arch/x86/mm/pageattr.c start += PUD_SIZE; PUD_SIZE 1368 arch/x86/mm/pageattr.c cpa->pfn += PUD_SIZE >> PAGE_SHIFT; PUD_SIZE 1369 arch/x86/mm/pageattr.c cur_pages += PUD_SIZE >> PAGE_SHIFT; PUD_SIZE 700 arch/x86/mm/pgtable.c mtrr = mtrr_type_lookup(addr, addr + PUD_SIZE, &uniform); PUD_SIZE 334 arch/x86/mm/pti.c addr = round_up(addr + 1, PUD_SIZE); PUD_SIZE 128 arch/x86/platform/efi/efi_64.c addr_pud = addr_p4d + j * PUD_SIZE; PUD_SIZE 1168 arch/x86/xen/mmu_pv.c xen_free_ro_pages(pa, PUD_SIZE); PUD_SIZE 2070 arch/x86/xen/mmu_pv.c n_pmd = roundup(size, PUD_SIZE) >> PUD_SHIFT; PUD_SIZE 167 drivers/dax/device.c unsigned int fault_size = PUD_SIZE; PUD_SIZE 174 drivers/dax/device.c if (dax_region->align > PUD_SIZE) { PUD_SIZE 193 drivers/dax/device.c (pud_addr + PUD_SIZE) > vmf->vma->vm_end) PUD_SIZE 197 drivers/dax/device.c phys = dax_pgoff_to_phys(dev_dax, pgoff, PUD_SIZE); PUD_SIZE 240 drivers/dax/device.c fault_size = PUD_SIZE; PUD_SIZE 20 include/asm-generic/pgtable-nop4d-hack.h #define PUD_MASK (~(PUD_SIZE-1)) PUD_SIZE 24 include/asm-generic/pgtable-nopud.h #define PUD_MASK (~(PUD_SIZE-1)) PUD_SIZE 543 include/asm-generic/pgtable.h ({ unsigned long __boundary = ((addr) + PUD_SIZE) & PUD_MASK; \ PUD_SIZE 522 include/asm-generic/tlb.h else if (_sz == PUD_SIZE) \ PUD_SIZE 131 lib/ioremap.c if ((end - addr) != PUD_SIZE) PUD_SIZE 134 lib/ioremap.c if (!IS_ALIGNED(addr, PUD_SIZE)) PUD_SIZE 137 lib/ioremap.c if (!IS_ALIGNED(phys_addr, PUD_SIZE)) PUD_SIZE 4810 mm/hugetlb.c unsigned long s_end = sbase + PUD_SIZE; PUD_SIZE 4831 mm/hugetlb.c unsigned long end = base + PUD_SIZE; PUD_SIZE 4854 mm/hugetlb.c for (check_addr = *start; check_addr < *end; check_addr += PUD_SIZE) { PUD_SIZE 4856 mm/hugetlb.c unsigned long a_end = a_start + PUD_SIZE; PUD_SIZE 4990 mm/hugetlb.c if (sz == PUD_SIZE) { PUD_SIZE 5031 mm/hugetlb.c if (sz != PUD_SIZE && pud_none(pud_entry)) PUD_SIZE 154 mm/kasan/init.c if (IS_ALIGNED(addr, PUD_SIZE) && end - addr >= PUD_SIZE) { PUD_SIZE 415 mm/kasan/init.c if (IS_ALIGNED(addr, PUD_SIZE) && PUD_SIZE 416 mm/kasan/init.c IS_ALIGNED(next, PUD_SIZE)) PUD_SIZE 1720 virt/kvm/arm/mmu.c (vma_pagesize == PUD_SIZE && kvm_stage2_has_pmd(kvm))) PUD_SIZE 1810 virt/kvm/arm/mmu.c if (vma_pagesize == PUD_SIZE) { PUD_SIZE 2107 virt/kvm/arm/mmu.c WARN_ON(size != PAGE_SIZE && size != PMD_SIZE && size != PUD_SIZE); PUD_SIZE 2125 virt/kvm/arm/mmu.c WARN_ON(size != PAGE_SIZE && size != PMD_SIZE && size != PUD_SIZE);