PUD_SIZE           68 arch/arm/include/asm/stage2_pgtable.h #define S2_PUD_SIZE				PUD_SIZE
PUD_SIZE          365 arch/arm/mm/dump.c 		addr = start + i * PUD_SIZE;
PUD_SIZE          361 arch/arm64/include/asm/kvm_mmu.h 		kvm_flush_dcache_to_poc(page_address(page), PUD_SIZE);
PUD_SIZE           61 arch/arm64/include/asm/pgtable-hwdef.h #define PUD_MASK		(~(PUD_SIZE-1))
PUD_SIZE           24 arch/arm64/include/asm/pgtable.h #define VMALLOC_END		(- PUD_SIZE - VMEMMAP_SIZE - SZ_64K)
PUD_SIZE           29 arch/arm64/mm/hugetlbpage.c 	case PUD_SIZE:
PUD_SIZE           91 arch/arm64/mm/hugetlbpage.c 	case PUD_SIZE:
PUD_SIZE          229 arch/arm64/mm/hugetlbpage.c 	if (sz == PUD_SIZE) {
PUD_SIZE          273 arch/arm64/mm/hugetlbpage.c 	if (sz != PUD_SIZE && pud_none(pud))
PUD_SIZE          306 arch/arm64/mm/hugetlbpage.c 	} else if (pagesize != PUD_SIZE && pagesize != PMD_SIZE) {
PUD_SIZE          457 arch/arm64/mm/hugetlbpage.c 	add_huge_page_size(PUD_SIZE);
PUD_SIZE          473 arch/arm64/mm/hugetlbpage.c 	case PUD_SIZE:
PUD_SIZE         1036 arch/arm64/mm/mmu.c 	end = addr + PUD_SIZE;
PUD_SIZE          112 arch/ia64/include/asm/pgtable.h #define PUD_MASK	(~(PUD_SIZE-1))
PUD_SIZE          397 arch/ia64/mm/init.c 			end_address += PUD_SIZE;
PUD_SIZE           62 arch/mips/include/asm/pgtable-64.h #define PUD_MASK	(~(PUD_SIZE-1))
PUD_SIZE          242 arch/powerpc/include/asm/book3s/64/pgtable.h #define PUD_MASK	(~(PUD_SIZE-1))
PUD_SIZE           37 arch/powerpc/include/asm/nohash/64/pgtable-4k.h #define PUD_MASK	(~(PUD_SIZE-1))
PUD_SIZE          836 arch/powerpc/kvm/book3s_64_mmu_radix.c 	    (gpa & (PUD_SIZE - PAGE_SIZE)) ==
PUD_SIZE          837 arch/powerpc/kvm/book3s_64_mmu_radix.c 	    (hva & (PUD_SIZE - PAGE_SIZE))) {
PUD_SIZE         1281 arch/powerpc/kvm/book3s_64_mmu_radix.c 			gpa = (gpa & PUD_MASK) + PUD_SIZE;
PUD_SIZE           77 arch/powerpc/mm/book3s64/radix_pgtable.c 	if (map_page_size == PUD_SIZE) {
PUD_SIZE          141 arch/powerpc/mm/book3s64/radix_pgtable.c 	if (map_page_size == PUD_SIZE) {
PUD_SIZE          271 arch/powerpc/mm/book3s64/radix_pgtable.c 		if (IS_ALIGNED(addr, PUD_SIZE) && gap >= PUD_SIZE &&
PUD_SIZE          273 arch/powerpc/mm/book3s64/radix_pgtable.c 			mapping_size = PUD_SIZE;
PUD_SIZE          830 arch/powerpc/mm/book3s64/radix_pgtable.c 			split_kernel_mapping(addr, end, PUD_SIZE, (pte_t *)pud);
PUD_SIZE         1131 arch/powerpc/mm/book3s64/radix_pgtable.c 	flush_tlb_kernel_range(addr, addr + PUD_SIZE);
PUD_SIZE          898 arch/powerpc/mm/book3s64/radix_tlb.c 			gstart = (start + PUD_SIZE - 1) & PUD_MASK;
PUD_SIZE          912 arch/powerpc/mm/book3s64/radix_tlb.c 						PUD_SIZE, MMU_PAGE_1G);
PUD_SIZE          922 arch/powerpc/mm/book3s64/radix_tlb.c 						PUD_SIZE, MMU_PAGE_1G);
PUD_SIZE          933 arch/powerpc/mm/book3s64/radix_tlb.c 					gstart, gend, pid, PUD_SIZE, MMU_PAGE_1G, false);
PUD_SIZE          427 arch/powerpc/mm/ptdump/hashpagetable.c 		addr = start + i * PUD_SIZE;
PUD_SIZE          288 arch/powerpc/mm/ptdump/ptdump.c 		addr = start + i * PUD_SIZE;
PUD_SIZE          293 arch/powerpc/mm/ptdump/ptdump.c 			note_page(st, addr, 2, pud_val(*pud), PUD_SIZE);
PUD_SIZE           23 arch/riscv/mm/hugetlbpage.c 	} else if (IS_ENABLED(CONFIG_64BIT) && ps == PUD_SIZE) {
PUD_SIZE          208 arch/s390/mm/dump_pagetables.c 		addr += PUD_SIZE;
PUD_SIZE          140 arch/s390/mm/hugetlbpage.c 		size = PUD_SIZE;
PUD_SIZE          205 arch/s390/mm/hugetlbpage.c 			if (sz == PUD_SIZE)
PUD_SIZE          265 arch/s390/mm/hugetlbpage.c 	} else if (MACHINE_HAS_EDAT2 && size == PUD_SIZE) {
PUD_SIZE          136 arch/s390/mm/kasan_init.c 			    IS_ALIGNED(address, PUD_SIZE) &&
PUD_SIZE          137 arch/s390/mm/kasan_init.c 			    end - address >= PUD_SIZE) {
PUD_SIZE          140 arch/s390/mm/kasan_init.c 				address = (address + PUD_SIZE) & PUD_MASK;
PUD_SIZE          278 arch/s390/mm/kasan_init.c 		BUILD_BUG_ON(!IS_ALIGNED(KASAN_SHADOW_START, PUD_SIZE));
PUD_SIZE          279 arch/s390/mm/kasan_init.c 		BUILD_BUG_ON(!IS_ALIGNED(KASAN_SHADOW_END, PUD_SIZE));
PUD_SIZE          245 arch/s390/mm/pageattr.c 			if (addr & ~PUD_MASK || addr + PUD_SIZE > next) {
PUD_SIZE          108 arch/s390/mm/vmem.c 		    !(address & ~PUD_MASK) && (address + PUD_SIZE <= end) &&
PUD_SIZE          111 arch/s390/mm/vmem.c 			address += PUD_SIZE;
PUD_SIZE          179 arch/s390/mm/vmem.c 			address += PUD_SIZE;
PUD_SIZE          184 arch/s390/mm/vmem.c 			address += PUD_SIZE;
PUD_SIZE           61 arch/sparc/include/asm/pgtable_64.h #define PUD_MASK	(~(PUD_SIZE-1))
PUD_SIZE          287 arch/sparc/mm/hugetlbpage.c 	if (sz >= PUD_SIZE)
PUD_SIZE          330 arch/sparc/mm/hugetlbpage.c 	if (size >= PUD_SIZE)
PUD_SIZE          367 arch/sparc/mm/hugetlbpage.c 	if (size >= PUD_SIZE)
PUD_SIZE          446 arch/sparc/mm/init_64.c 		if (hugepage_size >= PUD_SIZE) {
PUD_SIZE         1712 arch/sparc/mm/init_64.c 		return vstart + PUD_SIZE;
PUD_SIZE         1722 arch/sparc/mm/init_64.c 		pte_val += PUD_SIZE;
PUD_SIZE         1723 arch/sparc/mm/init_64.c 		vstart += PUD_SIZE;
PUD_SIZE         1732 arch/sparc/mm/init_64.c 	if (guard && !(vstart & ~PUD_MASK) && (vend - vstart) >= PUD_SIZE)
PUD_SIZE          369 arch/um/kernel/tlb.c 			last = ADD_ROUND(addr, PUD_SIZE);
PUD_SIZE          223 arch/x86/boot/compressed/kaslr.c 		if (memparse(p, &p) != PUD_SIZE) {
PUD_SIZE          514 arch/x86/boot/compressed/kaslr.c 	addr = ALIGN(region->start, PUD_SIZE);
PUD_SIZE          520 arch/x86/boot/compressed/kaslr.c 	while (size > PUD_SIZE && max_gb_huge_pages) {
PUD_SIZE          521 arch/x86/boot/compressed/kaslr.c 		size -= PUD_SIZE;
PUD_SIZE          544 arch/x86/boot/compressed/kaslr.c 	size  = region->size - (addr - region->start) - i * PUD_SIZE;
PUD_SIZE          546 arch/x86/boot/compressed/kaslr.c 		tmp.start = addr + i * PUD_SIZE;
PUD_SIZE          101 arch/x86/include/asm/pgtable_64_types.h #define PUD_MASK	(~(PUD_SIZE - 1))
PUD_SIZE          433 arch/x86/kernel/head64.c 	BUILD_BUG_ON(MODULES_LEN + KERNEL_IMAGE_SIZE > 2*PUD_SIZE);
PUD_SIZE          194 arch/x86/mm/hugetlbpage.c 	} else if (ps == PUD_SIZE && boot_cpu_has(X86_FEATURE_GBPAGES)) {
PUD_SIZE           30 arch/x86/mm/ident_map.c 		next = (addr & PUD_MASK) + PUD_SIZE;
PUD_SIZE          302 arch/x86/mm/init.c 			unsigned long start = round_down(mr[i].start, PUD_SIZE);
PUD_SIZE          303 arch/x86/mm/init.c 			unsigned long end = round_up(mr[i].end, PUD_SIZE);
PUD_SIZE          374 arch/x86/mm/init.c 	end_pfn = round_up(pfn, PFN_DOWN(PUD_SIZE));
PUD_SIZE          387 arch/x86/mm/init.c 	start_pfn = round_up(pfn, PFN_DOWN(PUD_SIZE));
PUD_SIZE          388 arch/x86/mm/init.c 	end_pfn = round_down(limit_pfn, PFN_DOWN(PUD_SIZE));
PUD_SIZE          602 arch/x86/mm/init_64.c 		paddr_next = (paddr & PUD_MASK) + PUD_SIZE;
PUD_SIZE         1101 arch/x86/mm/init_64.c 			if (IS_ALIGNED(addr, PUD_SIZE) &&
PUD_SIZE         1102 arch/x86/mm/init_64.c 			    IS_ALIGNED(next, PUD_SIZE)) {
PUD_SIZE         1105 arch/x86/mm/init_64.c 						       get_order(PUD_SIZE));
PUD_SIZE         1117 arch/x86/mm/init_64.c 						PUD_SIZE)) {
PUD_SIZE         1119 arch/x86/mm/init_64.c 						       get_order(PUD_SIZE));
PUD_SIZE           85 arch/x86/mm/kasan_init_64.c 		    ((end - addr) == PUD_SIZE) &&
PUD_SIZE           86 arch/x86/mm/kasan_init_64.c 		    IS_ALIGNED(addr, PUD_SIZE)) {
PUD_SIZE           87 arch/x86/mm/kasan_init_64.c 			p = early_alloc(PUD_SIZE, nid, false);
PUD_SIZE           91 arch/x86/mm/kasan_init_64.c 				memblock_free(__pa(p), PUD_SIZE);
PUD_SIZE          146 arch/x86/mm/kaslr.c 		vaddr = round_up(vaddr + 1, PUD_SIZE);
PUD_SIZE          264 arch/x86/mm/mem_encrypt_identity.c 	entries += (DIV_ROUND_UP(len, PUD_SIZE) + 1) * sizeof(pmd_t) * PTRS_PER_PMD;
PUD_SIZE          275 arch/x86/mm/mem_encrypt_identity.c 	tables += DIV_ROUND_UP(entries, PUD_SIZE) * sizeof(pmd_t) * PTRS_PER_PMD;
PUD_SIZE         1165 arch/x86/mm/pageattr.c 	if (start & (PUD_SIZE - 1)) {
PUD_SIZE         1166 arch/x86/mm/pageattr.c 		unsigned long next_page = (start + PUD_SIZE) & PUD_MASK;
PUD_SIZE         1178 arch/x86/mm/pageattr.c 	while (end - start >= PUD_SIZE) {
PUD_SIZE         1183 arch/x86/mm/pageattr.c 			unmap_pmd_range(pud, start, start + PUD_SIZE);
PUD_SIZE         1185 arch/x86/mm/pageattr.c 		start += PUD_SIZE;
PUD_SIZE         1328 arch/x86/mm/pageattr.c 	if (start & (PUD_SIZE - 1)) {
PUD_SIZE         1330 arch/x86/mm/pageattr.c 		unsigned long next_page = (start + PUD_SIZE) & PUD_MASK;
PUD_SIZE         1363 arch/x86/mm/pageattr.c 	while (boot_cpu_has(X86_FEATURE_GBPAGES) && end - start >= PUD_SIZE) {
PUD_SIZE         1367 arch/x86/mm/pageattr.c 		start	  += PUD_SIZE;
PUD_SIZE         1368 arch/x86/mm/pageattr.c 		cpa->pfn  += PUD_SIZE >> PAGE_SHIFT;
PUD_SIZE         1369 arch/x86/mm/pageattr.c 		cur_pages += PUD_SIZE >> PAGE_SHIFT;
PUD_SIZE          700 arch/x86/mm/pgtable.c 	mtrr = mtrr_type_lookup(addr, addr + PUD_SIZE, &uniform);
PUD_SIZE          334 arch/x86/mm/pti.c 			addr = round_up(addr + 1, PUD_SIZE);
PUD_SIZE          128 arch/x86/platform/efi/efi_64.c 				addr_pud = addr_p4d + j * PUD_SIZE;
PUD_SIZE         1168 arch/x86/xen/mmu_pv.c 		xen_free_ro_pages(pa, PUD_SIZE);
PUD_SIZE         2070 arch/x86/xen/mmu_pv.c 	n_pmd = roundup(size, PUD_SIZE) >> PUD_SHIFT;
PUD_SIZE          167 drivers/dax/device.c 	unsigned int fault_size = PUD_SIZE;
PUD_SIZE          174 drivers/dax/device.c 	if (dax_region->align > PUD_SIZE) {
PUD_SIZE          193 drivers/dax/device.c 			(pud_addr + PUD_SIZE) > vmf->vma->vm_end)
PUD_SIZE          197 drivers/dax/device.c 	phys = dax_pgoff_to_phys(dev_dax, pgoff, PUD_SIZE);
PUD_SIZE          240 drivers/dax/device.c 		fault_size = PUD_SIZE;
PUD_SIZE           20 include/asm-generic/pgtable-nop4d-hack.h #define PUD_MASK	(~(PUD_SIZE-1))
PUD_SIZE           24 include/asm-generic/pgtable-nopud.h #define PUD_MASK  	(~(PUD_SIZE-1))
PUD_SIZE          543 include/asm-generic/pgtable.h ({	unsigned long __boundary = ((addr) + PUD_SIZE) & PUD_MASK;	\
PUD_SIZE          522 include/asm-generic/tlb.h 		else if (_sz == PUD_SIZE)			\
PUD_SIZE          131 lib/ioremap.c  	if ((end - addr) != PUD_SIZE)
PUD_SIZE          134 lib/ioremap.c  	if (!IS_ALIGNED(addr, PUD_SIZE))
PUD_SIZE          137 lib/ioremap.c  	if (!IS_ALIGNED(phys_addr, PUD_SIZE))
PUD_SIZE         4810 mm/hugetlb.c   	unsigned long s_end = sbase + PUD_SIZE;
PUD_SIZE         4831 mm/hugetlb.c   	unsigned long end = base + PUD_SIZE;
PUD_SIZE         4854 mm/hugetlb.c   	for (check_addr = *start; check_addr < *end; check_addr += PUD_SIZE) {
PUD_SIZE         4856 mm/hugetlb.c   		unsigned long a_end = a_start + PUD_SIZE;
PUD_SIZE         4990 mm/hugetlb.c   		if (sz == PUD_SIZE) {
PUD_SIZE         5031 mm/hugetlb.c   	if (sz != PUD_SIZE && pud_none(pud_entry))
PUD_SIZE          154 mm/kasan/init.c 		if (IS_ALIGNED(addr, PUD_SIZE) && end - addr >= PUD_SIZE) {
PUD_SIZE          415 mm/kasan/init.c 			if (IS_ALIGNED(addr, PUD_SIZE) &&
PUD_SIZE          416 mm/kasan/init.c 			    IS_ALIGNED(next, PUD_SIZE))
PUD_SIZE         1720 virt/kvm/arm/mmu.c 	    (vma_pagesize == PUD_SIZE && kvm_stage2_has_pmd(kvm)))
PUD_SIZE         1810 virt/kvm/arm/mmu.c 	if (vma_pagesize == PUD_SIZE) {
PUD_SIZE         2107 virt/kvm/arm/mmu.c 	WARN_ON(size != PAGE_SIZE && size != PMD_SIZE && size != PUD_SIZE);
PUD_SIZE         2125 virt/kvm/arm/mmu.c 	WARN_ON(size != PAGE_SIZE && size != PMD_SIZE && size != PUD_SIZE);