PMD_SIZE           34 arch/alpha/include/asm/pgtable.h #define PMD_MASK	(~(PMD_SIZE-1))
PMD_SIZE          829 arch/arc/mm/tlb.c 	BUILD_BUG_ON(!IS_ALIGNED((CONFIG_ARC_KVADDR_SIZE << 20), PMD_SIZE));
PMD_SIZE          835 arch/arc/mm/tlb.c 	BUILD_BUG_ON(!IS_ALIGNED(STACK_TOP, PMD_SIZE));
PMD_SIZE            7 arch/arm/include/asm/highmem.h #define PKMAP_BASE		(PAGE_OFFSET - PMD_SIZE)
PMD_SIZE          312 arch/arm/include/asm/kvm_mmu.h 	unsigned long size = PMD_SIZE;
PMD_SIZE           58 arch/arm/include/asm/memory.h #define MODULES_END		(PAGE_OFFSET - PMD_SIZE)
PMD_SIZE           86 arch/arm/include/asm/pgtable-2level.h #define PMD_MASK		(~(PMD_SIZE-1))
PMD_SIZE           49 arch/arm/include/asm/stage2_pgtable.h 	phys_addr_t boundary = (addr + PMD_SIZE) & PMD_MASK;
PMD_SIZE           66 arch/arm/include/asm/stage2_pgtable.h #define S2_PMD_SIZE				PMD_SIZE
PMD_SIZE          454 arch/arm/mm/dma-mapping.c 		     addr += PMD_SIZE)
PMD_SIZE          342 arch/arm/mm/dump.c 		addr = start + i * PMD_SIZE;
PMD_SIZE          349 arch/arm/mm/dump.c 		if (SECTION_SIZE < PMD_SIZE && pmd_large(pmd[1])) {
PMD_SIZE          173 arch/arm/mm/ioremap.c 		addr += PMD_SIZE;
PMD_SIZE          212 arch/arm/mm/ioremap.c 		addr += PMD_SIZE;
PMD_SIZE          249 arch/arm/mm/ioremap.c 			addr += PMD_SIZE;
PMD_SIZE         1094 arch/arm/mm/mmu.c 		next = (addr + PMD_SIZE - 1) & PMD_MASK;
PMD_SIZE         1186 arch/arm/mm/mmu.c 			if (!IS_ALIGNED(reg->base, PMD_SIZE)) {
PMD_SIZE         1189 arch/arm/mm/mmu.c 				len = round_up(reg->base, PMD_SIZE) - reg->base;
PMD_SIZE         1229 arch/arm/mm/mmu.c 				if (!IS_ALIGNED(block_start, PMD_SIZE))
PMD_SIZE         1231 arch/arm/mm/mmu.c 				else if (!IS_ALIGNED(block_end, PMD_SIZE))
PMD_SIZE         1250 arch/arm/mm/mmu.c 	memblock_limit = round_down(memblock_limit, PMD_SIZE);
PMD_SIZE         1275 arch/arm/mm/mmu.c 	for (addr = 0; addr < MODULES_VADDR; addr += PMD_SIZE)
PMD_SIZE         1280 arch/arm/mm/mmu.c 	addr = ((unsigned long)_exiprom + PMD_SIZE - 1) & PMD_MASK;
PMD_SIZE         1282 arch/arm/mm/mmu.c 	for ( ; addr < PAGE_OFFSET; addr += PMD_SIZE)
PMD_SIZE         1297 arch/arm/mm/mmu.c 	     addr < VMALLOC_START; addr += PMD_SIZE)
PMD_SIZE         1352 arch/arm/mm/mmu.c 	for (addr = VMALLOC_START; addr < (FIXADDR_TOP & PMD_MASK); addr += PMD_SIZE)
PMD_SIZE          353 arch/arm64/include/asm/kvm_mmu.h 		kvm_flush_dcache_to_poc(page_address(page), PMD_SIZE);
PMD_SIZE           51 arch/arm64/include/asm/pgtable-hwdef.h #define PMD_MASK		(~(PMD_SIZE-1))
PMD_SIZE           99 arch/arm64/include/asm/pgtable-hwdef.h #define CONT_PMD_SIZE		(CONT_PMDS * PMD_SIZE)
PMD_SIZE           31 arch/arm64/mm/hugetlbpage.c 	case PMD_SIZE:
PMD_SIZE           77 arch/arm64/mm/hugetlbpage.c 		*pgsize = PMD_SIZE;
PMD_SIZE           93 arch/arm64/mm/hugetlbpage.c 	case PMD_SIZE:
PMD_SIZE           97 arch/arm64/mm/hugetlbpage.c 		*pgsize = PMD_SIZE;
PMD_SIZE          245 arch/arm64/mm/hugetlbpage.c 	} else if (sz == PMD_SIZE) {
PMD_SIZE          285 arch/arm64/mm/hugetlbpage.c 	if (!(sz == PMD_SIZE || sz == CONT_PMD_SIZE) &&
PMD_SIZE          306 arch/arm64/mm/hugetlbpage.c 	} else if (pagesize != PUD_SIZE && pagesize != PMD_SIZE) {
PMD_SIZE          460 arch/arm64/mm/hugetlbpage.c 	add_huge_page_size(PMD_SIZE);
PMD_SIZE          476 arch/arm64/mm/hugetlbpage.c 	case PMD_SIZE:
PMD_SIZE          760 arch/arm64/mm/mmu.c 			p = vmemmap_alloc_block_buf(PMD_SIZE, node);
PMD_SIZE         1039 arch/arm64/mm/mmu.c 	} while (pmdp++, next += PMD_SIZE, next != end);
PMD_SIZE          152 arch/csky/mm/highmem.c 				vaddr += PMD_SIZE;
PMD_SIZE          100 arch/ia64/include/asm/pgtable.h #define PMD_MASK	(~(PMD_SIZE-1))
PMD_SIZE          403 arch/ia64/mm/init.c 			end_address += PMD_SIZE;
PMD_SIZE           39 arch/m68k/include/asm/pgtable_mm.h #define PMD_MASK	(~(PMD_SIZE-1))
PMD_SIZE          118 arch/m68k/sun3x/dvma.c 				end3 = (vaddr + (PMD_SIZE-1)) & PMD_MASK;
PMD_SIZE          144 arch/microblaze/include/asm/pgtable.h #define PMD_MASK	(~(PMD_SIZE-1))
PMD_SIZE           52 arch/mips/include/asm/pgtable-64.h #define PMD_MASK	(~(PMD_SIZE-1))
PMD_SIZE          263 arch/mips/mm/init.c 				vaddr += PMD_SIZE;
PMD_SIZE           31 arch/mips/mm/ioremap.c 	if (end > PMD_SIZE)
PMD_SIZE           32 arch/mips/mm/ioremap.c 		end = PMD_SIZE;
PMD_SIZE           63 arch/mips/mm/ioremap.c 		address = (address + PMD_SIZE) & PMD_MASK;
PMD_SIZE           46 arch/nds32/include/asm/pgtable.h #define PMD_MASK		(~(PMD_SIZE-1))
PMD_SIZE          100 arch/nds32/kernel/vdso.c 	end = (start + vdso_mapping_len + PMD_SIZE - 1) & PMD_MASK;
PMD_SIZE           33 arch/nios2/mm/ioremap.c 	if (end > PMD_SIZE)
PMD_SIZE           34 arch/nios2/mm/ioremap.c 		end = PMD_SIZE;
PMD_SIZE           70 arch/nios2/mm/ioremap.c 		address = (address + PMD_SIZE) & PMD_MASK;
PMD_SIZE          137 arch/parisc/include/asm/pgtable.h #define PMD_MASK	(~(PMD_SIZE-1))
PMD_SIZE           86 arch/parisc/kernel/pci-dma.c 	if (end > PMD_SIZE)
PMD_SIZE           87 arch/parisc/kernel/pci-dma.c 		end = PMD_SIZE;
PMD_SIZE          121 arch/parisc/kernel/pci-dma.c 		vaddr = (vaddr + PMD_SIZE) & PMD_MASK;
PMD_SIZE          122 arch/parisc/kernel/pci-dma.c 		orig_vaddr += PMD_SIZE;
PMD_SIZE          166 arch/parisc/kernel/pci-dma.c 	if (end > PMD_SIZE)
PMD_SIZE          167 arch/parisc/kernel/pci-dma.c 		end = PMD_SIZE;
PMD_SIZE          206 arch/parisc/kernel/pci-dma.c 		vaddr = (vaddr + PMD_SIZE) & PMD_MASK;
PMD_SIZE          207 arch/parisc/kernel/pci-dma.c 		orig_vaddr += PMD_SIZE;
PMD_SIZE          237 arch/powerpc/include/asm/book3s/64/pgtable.h #define PMD_MASK	(~(PMD_SIZE-1))
PMD_SIZE           32 arch/powerpc/include/asm/nohash/64/pgtable-4k.h #define PMD_MASK	(~(PMD_SIZE-1))
PMD_SIZE          840 arch/powerpc/kvm/book3s_64_mmu_radix.c 		   (gpa & (PMD_SIZE - PAGE_SIZE)) ==
PMD_SIZE          841 arch/powerpc/kvm/book3s_64_mmu_radix.c 		   (hva & (PMD_SIZE - PAGE_SIZE))) {
PMD_SIZE         1293 arch/powerpc/kvm/book3s_64_mmu_radix.c 			gpa = (gpa & PMD_MASK) + PMD_SIZE;
PMD_SIZE          241 arch/powerpc/mm/book3s64/hash_tlb.c 	addr = _ALIGN_DOWN(addr, PMD_SIZE);
PMD_SIZE           87 arch/powerpc/mm/book3s64/radix_pgtable.c 	if (map_page_size == PMD_SIZE) {
PMD_SIZE          148 arch/powerpc/mm/book3s64/radix_pgtable.c 	if (map_page_size == PMD_SIZE) {
PMD_SIZE          275 arch/powerpc/mm/book3s64/radix_pgtable.c 		} else if (IS_ALIGNED(addr, PMD_SIZE) && gap >= PMD_SIZE &&
PMD_SIZE          277 arch/powerpc/mm/book3s64/radix_pgtable.c 			mapping_size = PMD_SIZE;
PMD_SIZE          805 arch/powerpc/mm/book3s64/radix_pgtable.c 			split_kernel_mapping(addr, end, PMD_SIZE, (pte_t *)pmd);
PMD_SIZE         1177 arch/powerpc/mm/book3s64/radix_pgtable.c 	flush_tlb_kernel_range(addr, addr + PMD_SIZE);
PMD_SIZE          891 arch/powerpc/mm/book3s64/radix_tlb.c 			hstart = (start + PMD_SIZE - 1) & PMD_MASK;
PMD_SIZE          909 arch/powerpc/mm/book3s64/radix_tlb.c 						PMD_SIZE, MMU_PAGE_2M);
PMD_SIZE          919 arch/powerpc/mm/book3s64/radix_tlb.c 						PMD_SIZE, MMU_PAGE_2M);
PMD_SIZE          930 arch/powerpc/mm/book3s64/radix_tlb.c 					hstart, hend, pid, PMD_SIZE, MMU_PAGE_2M, false);
PMD_SIZE          415 arch/powerpc/mm/nohash/tlb.c 		unsigned long end = address + PMD_SIZE;
PMD_SIZE          413 arch/powerpc/mm/ptdump/hashpagetable.c 		addr = start + i * PMD_SIZE;
PMD_SIZE          272 arch/powerpc/mm/ptdump/ptdump.c 		addr = start + i * PMD_SIZE;
PMD_SIZE          277 arch/powerpc/mm/ptdump/ptdump.c 			note_page(st, addr, 3, pmd_val(*pmd), PMD_SIZE);
PMD_SIZE           19 arch/riscv/include/asm/pgtable-64.h #define PMD_MASK        (~(PMD_SIZE - 1))
PMD_SIZE          110 arch/riscv/include/asm/pgtable.h #define FIXADDR_SIZE     PMD_SIZE
PMD_SIZE          253 arch/riscv/mm/init.c 	if (sz == PMD_SIZE) {
PMD_SIZE          277 arch/riscv/mm/init.c #define PTE_PARENT_SIZE		PMD_SIZE
PMD_SIZE          374 arch/riscv/mm/init.c 			   (uintptr_t)fixmap_pte, PMD_SIZE, PAGE_TABLE);
PMD_SIZE          379 arch/riscv/mm/init.c 			   load_pa, PMD_SIZE, PAGE_KERNEL_EXEC);
PMD_SIZE          177 arch/s390/mm/dump_pagetables.c 		addr += PMD_SIZE;
PMD_SIZE          343 arch/s390/mm/gmap.c 	offset = (offset & (PTRS_PER_PMD - 1)) * PMD_SIZE;
PMD_SIZE          402 arch/s390/mm/gmap.c 	if ((to | len) & (PMD_SIZE - 1))
PMD_SIZE          409 arch/s390/mm/gmap.c 	for (off = 0; off < len; off += PMD_SIZE)
PMD_SIZE          434 arch/s390/mm/gmap.c 	if ((from | to | len) & (PMD_SIZE - 1))
PMD_SIZE          442 arch/s390/mm/gmap.c 	for (off = 0; off < len; off += PMD_SIZE) {
PMD_SIZE          701 arch/s390/mm/gmap.c 	     gaddr = (gaddr + PMD_SIZE) & PMD_MASK) {
PMD_SIZE          719 arch/s390/mm/gmap.c 		size = min(to - gaddr, PMD_SIZE - (gaddr & ~PMD_MASK));
PMD_SIZE          144 arch/s390/mm/hugetlbpage.c 		size = PMD_SIZE;
PMD_SIZE          207 arch/s390/mm/hugetlbpage.c 			else if (sz == PMD_SIZE)
PMD_SIZE          263 arch/s390/mm/hugetlbpage.c 	if (MACHINE_HAS_EDAT1 && size == PMD_SIZE) {
PMD_SIZE          150 arch/s390/mm/kasan_init.c 			    IS_ALIGNED(address, PMD_SIZE) &&
PMD_SIZE          151 arch/s390/mm/kasan_init.c 			    end - address >= PMD_SIZE) {
PMD_SIZE          154 arch/s390/mm/kasan_init.c 				address = (address + PMD_SIZE) & PMD_MASK;
PMD_SIZE          158 arch/s390/mm/kasan_init.c 			if (has_edat && address && end - address >= PMD_SIZE &&
PMD_SIZE          169 arch/s390/mm/kasan_init.c 				address = (address + PMD_SIZE) & PMD_MASK;
PMD_SIZE          176 arch/s390/mm/kasan_init.c 			address = (address + PMD_SIZE) & PMD_MASK;
PMD_SIZE          168 arch/s390/mm/pageattr.c 			if (addr & ~PMD_MASK || addr + PMD_SIZE > next) {
PMD_SIZE          206 arch/s390/mm/pageattr.c 		pmd_addr += PMD_SIZE;
PMD_SIZE          123 arch/s390/mm/vmem.c 		    !(address & ~PMD_MASK) && (address + PMD_SIZE <= end) &&
PMD_SIZE          126 arch/s390/mm/vmem.c 			address += PMD_SIZE;
PMD_SIZE          190 arch/s390/mm/vmem.c 			address += PMD_SIZE;
PMD_SIZE          195 arch/s390/mm/vmem.c 			address += PMD_SIZE;
PMD_SIZE          267 arch/s390/mm/vmem.c 				new_page = vmemmap_alloc_block(PMD_SIZE, node);
PMD_SIZE          271 arch/s390/mm/vmem.c 				address = (address + PMD_SIZE) & PMD_MASK;
PMD_SIZE          279 arch/s390/mm/vmem.c 			address = (address + PMD_SIZE) & PMD_MASK;
PMD_SIZE           25 arch/sh/include/asm/pgtable-3level.h #define PMD_MASK	(~(PMD_SIZE-1))
PMD_SIZE           27 arch/sh/include/asm/pgtable-3level.h #define PTRS_PER_PMD	((1 << PGDIR_SHIFT) / PMD_SIZE)
PMD_SIZE          190 arch/sh/mm/init.c 				vaddr += PMD_SIZE;
PMD_SIZE          330 arch/sh/mm/init.c 	end = (FIXADDR_TOP + PMD_SIZE - 1) & PMD_MASK;
PMD_SIZE           39 arch/sparc/include/asm/pgtable_32.h #define PMD_MASK        	(~(PMD_SIZE-1))
PMD_SIZE           53 arch/sparc/include/asm/pgtable_64.h #define PMD_MASK	(~(PMD_SIZE-1))
PMD_SIZE          292 arch/sparc/mm/hugetlbpage.c 	if (sz >= PMD_SIZE)
PMD_SIZE          332 arch/sparc/mm/hugetlbpage.c 	else if (size >= PMD_SIZE)
PMD_SIZE          369 arch/sparc/mm/hugetlbpage.c 	else if (size >= PMD_SIZE)
PMD_SIZE          499 arch/sparc/mm/hugetlbpage.c 		addr += PMD_SIZE;
PMD_SIZE          509 arch/sparc/mm/hugetlbpage.c 		end -= PMD_SIZE;
PMD_SIZE          454 arch/sparc/mm/init_64.c 		} else if (hugepage_size >= PMD_SIZE) {
PMD_SIZE          461 arch/sparc/mm/init_64.c 		if (hugepage_size >= PMD_SIZE) {
PMD_SIZE         1752 arch/sparc/mm/init_64.c 		return vstart + PMD_SIZE;
PMD_SIZE         1769 arch/sparc/mm/init_64.c 		pte_val += PMD_SIZE;
PMD_SIZE         1770 arch/sparc/mm/init_64.c 		vstart += PMD_SIZE;
PMD_SIZE         1780 arch/sparc/mm/init_64.c 	if (guard && !(vstart & ~PMD_MASK) && (vend - vstart) >= PMD_SIZE)
PMD_SIZE         1850 arch/sparc/mm/init_64.c 		this_end = (vstart + PMD_SIZE) & PMD_MASK;
PMD_SIZE         2611 arch/sparc/mm/init_64.c 	vend = ALIGN(vend, PMD_SIZE);
PMD_SIZE         2612 arch/sparc/mm/init_64.c 	for (; vstart < vend; vstart += PMD_SIZE) {
PMD_SIZE         2628 arch/sparc/mm/init_64.c 			void *block = vmemmap_alloc_block(PMD_SIZE, node);
PMD_SIZE          717 arch/sparc/mm/srmmu.c 		if (start > (0xffffffffUL - PMD_SIZE))
PMD_SIZE          719 arch/sparc/mm/srmmu.c 		start = (start + PMD_SIZE) & PMD_MASK;
PMD_SIZE          748 arch/sparc/mm/srmmu.c 		if (start > (0xffffffffUL - PMD_SIZE))
PMD_SIZE          750 arch/sparc/mm/srmmu.c 		start = (start + PMD_SIZE) & PMD_MASK;
PMD_SIZE           29 arch/um/include/asm/pgtable-3level.h #define PMD_MASK	(~(PMD_SIZE-1))
PMD_SIZE          116 arch/um/kernel/mem.c 			vaddr += PMD_SIZE;
PMD_SIZE          385 arch/um/kernel/tlb.c 			last = ADD_ROUND(addr, PMD_SIZE);
PMD_SIZE          423 arch/x86/boot/compressed/kaslr.c 	add_identity_map(0, PMD_SIZE);
PMD_SIZE          135 arch/x86/boot/compressed/kaslr_64.c 	start = round_down(start, PMD_SIZE);
PMD_SIZE          136 arch/x86/boot/compressed/kaslr_64.c 	end = round_up(end, PMD_SIZE);
PMD_SIZE          224 arch/x86/entry/vdso/vma.c 	end = (start + len + PMD_SIZE - 1) & PMD_MASK;
PMD_SIZE           13 arch/x86/include/asm/pgtable_32_types.h # define PMD_MASK	(~(PMD_SIZE - 1))
PMD_SIZE           56 arch/x86/include/asm/pgtable_32_types.h #define LDT_END_ADDR		(LDT_BASE_ADDR + PMD_SIZE)
PMD_SIZE           99 arch/x86/include/asm/pgtable_64_types.h #define PMD_MASK	(~(PMD_SIZE - 1))
PMD_SIZE          214 arch/x86/kernel/head64.c 	for (i = 0; i < DIV_ROUND_UP(_end - _text, PMD_SIZE); i++) {
PMD_SIZE          217 arch/x86/kernel/head64.c 		pmd[idx % PTRS_PER_PMD] = pmd_entry + i * PMD_SIZE;
PMD_SIZE          269 arch/x86/kernel/head64.c 		for (; vaddr < vaddr_end; vaddr += PMD_SIZE) {
PMD_SIZE          201 arch/x86/kernel/setup_percpu.c 		atom_size = PMD_SIZE;
PMD_SIZE          190 arch/x86/mm/cpu_entry_area.c 	for (; start < end && start >= CPU_ENTRY_AREA_BASE; start += PMD_SIZE)
PMD_SIZE          201 arch/x86/mm/fault.c 	     address += PMD_SIZE) {
PMD_SIZE          192 arch/x86/mm/hugetlbpage.c 	if (ps == PMD_SIZE) {
PMD_SIZE           11 arch/x86/mm/ident_map.c 	for (; addr < end; addr += PMD_SIZE) {
PMD_SIZE          289 arch/x86/mm/init.c 			unsigned long start = round_down(mr[i].start, PMD_SIZE);
PMD_SIZE          290 arch/x86/mm/init.c 			unsigned long end = round_up(mr[i].end, PMD_SIZE);
PMD_SIZE          356 arch/x86/mm/init.c 		end_pfn = PFN_DOWN(PMD_SIZE);
PMD_SIZE          358 arch/x86/mm/init.c 		end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE));
PMD_SIZE          360 arch/x86/mm/init.c 	end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE));
PMD_SIZE          370 arch/x86/mm/init.c 	start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE));
PMD_SIZE          372 arch/x86/mm/init.c 	end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE));
PMD_SIZE          375 arch/x86/mm/init.c 	if (end_pfn > round_down(limit_pfn, PFN_DOWN(PMD_SIZE)))
PMD_SIZE          376 arch/x86/mm/init.c 		end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE));
PMD_SIZE          397 arch/x86/mm/init.c 	start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE));
PMD_SIZE          398 arch/x86/mm/init.c 	end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE));
PMD_SIZE          570 arch/x86/mm/init.c 	addr = memblock_find_in_range(map_start, map_end, PMD_SIZE, PMD_SIZE);
PMD_SIZE          571 arch/x86/mm/init.c 	real_end = addr + PMD_SIZE;
PMD_SIZE          574 arch/x86/mm/init.c 	step_size = PMD_SIZE;
PMD_SIZE          621 arch/x86/mm/init.c 	unsigned long step_size = PMD_SIZE;
PMD_SIZE          149 arch/x86/mm/init_32.c 			vaddr += PMD_SIZE;
PMD_SIZE          234 arch/x86/mm/init_32.c 			vaddr += PMD_SIZE;
PMD_SIZE          549 arch/x86/mm/init_32.c 	end = (FIXADDR_TOP + PMD_SIZE - 1) & PMD_MASK;
PMD_SIZE          372 arch/x86/mm/init_64.c 	for (; size; phys += PMD_SIZE, size -= PMD_SIZE) {
PMD_SIZE          424 arch/x86/mm/init_64.c 	unsigned long end = roundup((unsigned long)_brk_end, PMD_SIZE) - 1;
PMD_SIZE          435 arch/x86/mm/init_64.c 	for (; vaddr + PMD_SIZE - 1 < vaddr_end; pmd++, vaddr += PMD_SIZE) {
PMD_SIZE          515 arch/x86/mm/init_64.c 		paddr_next = (paddr & PMD_MASK) + PMD_SIZE;
PMD_SIZE          898 arch/x86/mm/init_64.c 		vmem_altmap_free(altmap, PMD_SIZE / PAGE_SIZE);
PMD_SIZE          900 arch/x86/mm/init_64.c 		free_pagetable(page, get_order(PMD_SIZE));
PMD_SIZE         1045 arch/x86/mm/init_64.c 			if (IS_ALIGNED(addr, PMD_SIZE) &&
PMD_SIZE         1046 arch/x86/mm/init_64.c 			    IS_ALIGNED(next, PMD_SIZE)) {
PMD_SIZE         1061 arch/x86/mm/init_64.c 						PMD_SIZE)) {
PMD_SIZE         1324 arch/x86/mm/init_64.c 	all_end = roundup((unsigned long)_brk_end, PMD_SIZE);
PMD_SIZE         1479 arch/x86/mm/init_64.c 				p = altmap_alloc_block_buf(PMD_SIZE, altmap);
PMD_SIZE         1481 arch/x86/mm/init_64.c 				p = vmemmap_alloc_block_buf(PMD_SIZE, node);
PMD_SIZE         1499 arch/x86/mm/init_64.c 				addr_end = addr + PMD_SIZE;
PMD_SIZE         1500 arch/x86/mm/init_64.c 				p_end = p + PMD_SIZE;
PMD_SIZE         1592 arch/x86/mm/init_64.c 			nr_pmd_pages = 1 << get_order(PMD_SIZE);
PMD_SIZE           48 arch/x86/mm/kasan_init_64.c 		    ((end - addr) == PMD_SIZE) &&
PMD_SIZE           49 arch/x86/mm/kasan_init_64.c 		    IS_ALIGNED(addr, PMD_SIZE)) {
PMD_SIZE           50 arch/x86/mm/kasan_init_64.c 			p = early_alloc(PMD_SIZE, nid, false);
PMD_SIZE           54 arch/x86/mm/kasan_init_64.c 				memblock_free(__pa(p), PMD_SIZE);
PMD_SIZE          132 arch/x86/mm/mem_encrypt.c 		vaddr += PMD_SIZE;
PMD_SIZE          133 arch/x86/mm/mem_encrypt.c 		paddr += PMD_SIZE;
PMD_SIZE          134 arch/x86/mm/mem_encrypt.c 		size = (size <= PMD_SIZE) ? 0 : size - PMD_SIZE;
PMD_SIZE          213 arch/x86/mm/pageattr.c 	return __pa_symbol(roundup(_brk_end, PMD_SIZE) - 1) >> PAGE_SHIFT;
PMD_SIZE          981 arch/x86/mm/pageattr.c 		lpinc = PMD_SIZE;
PMD_SIZE         1121 arch/x86/mm/pageattr.c 	if (start & (PMD_SIZE - 1)) {
PMD_SIZE         1122 arch/x86/mm/pageattr.c 		unsigned long next_page = (start + PMD_SIZE) & PMD_MASK;
PMD_SIZE         1134 arch/x86/mm/pageattr.c 	while (end - start >= PMD_SIZE) {
PMD_SIZE         1138 arch/x86/mm/pageattr.c 			__unmap_pmd_range(pud, pmd, start, start + PMD_SIZE);
PMD_SIZE         1140 arch/x86/mm/pageattr.c 		start += PMD_SIZE;
PMD_SIZE         1251 arch/x86/mm/pageattr.c 	if (start & (PMD_SIZE - 1)) {
PMD_SIZE         1253 arch/x86/mm/pageattr.c 		unsigned long next_page = (start + PMD_SIZE) & PMD_MASK;
PMD_SIZE         1280 arch/x86/mm/pageattr.c 	while (end - start >= PMD_SIZE) {
PMD_SIZE         1294 arch/x86/mm/pageattr.c 		start	  += PMD_SIZE;
PMD_SIZE         1295 arch/x86/mm/pageattr.c 		cpa->pfn  += PMD_SIZE >> PAGE_SHIFT;
PMD_SIZE         1296 arch/x86/mm/pageattr.c 		cur_pages += PMD_SIZE >> PAGE_SHIFT;
PMD_SIZE          729 arch/x86/mm/pgtable.c 	mtrr = mtrr_type_lookup(addr, addr + PMD_SIZE, &uniform);
PMD_SIZE          733 arch/x86/mm/pgtable.c 			     __func__, addr, addr + PMD_SIZE);
PMD_SIZE          341 arch/x86/mm/pti.c 			addr = round_up(addr + 1, PMD_SIZE);
PMD_SIZE          378 arch/x86/mm/pti.c 			addr += PMD_SIZE;
PMD_SIZE          458 arch/x86/platform/efi/efi_64.c 	if (!(pa & (PMD_SIZE - 1))) {
PMD_SIZE          461 arch/x86/platform/efi/efi_64.c 		u64 pa_offset = pa & (PMD_SIZE - 1);
PMD_SIZE          468 arch/x86/platform/efi/efi_64.c 			efi_va -= PMD_SIZE;
PMD_SIZE         1096 arch/x86/xen/mmu_pv.c 	unsigned long kernel_end = roundup((unsigned long)_brk_end, PMD_SIZE) - 1;
PMD_SIZE         1102 arch/x86/xen/mmu_pv.c 			pmd++, vaddr += PMD_SIZE) {
PMD_SIZE         1145 arch/x86/xen/mmu_pv.c 		xen_free_ro_pages(pa, PMD_SIZE);
PMD_SIZE         1244 arch/x86/xen/mmu_pv.c 	size = roundup(size, PMD_SIZE);
PMD_SIZE         1273 arch/x86/xen/mmu_pv.c 	xen_cleanhighmap(addr, roundup(addr + size, PMD_SIZE * 2));
PMD_SIZE         2069 arch/x86/xen/mmu_pv.c 	n_pt = roundup(size, PMD_SIZE) >> PMD_SHIFT;
PMD_SIZE          409 arch/x86/xen/p2m.c 				(unsigned long)(p2m + pfn) + i * PMD_SIZE);
PMD_SIZE          425 arch/x86/xen/p2m.c 			PMD_SIZE * PMDS_PER_MID_PAGE);
PMD_SIZE          426 arch/x86/xen/p2m.c 	vm_area_register_early(&vm, PMD_SIZE * PMDS_PER_MID_PAGE);
PMD_SIZE          493 arch/x86/xen/p2m.c 	vaddr = addr & ~(PMD_SIZE * PMDS_PER_MID_PAGE - 1);
PMD_SIZE          522 arch/x86/xen/p2m.c 		vaddr += PMD_SIZE;
PMD_SIZE           31 arch/xtensa/mm/kasan_init.c 	for (vaddr = 0; vaddr < KASAN_SHADOW_SIZE; vaddr += PMD_SIZE, ++pmd) {
PMD_SIZE          118 drivers/dax/device.c 	unsigned int fault_size = PMD_SIZE;
PMD_SIZE          124 drivers/dax/device.c 	if (dax_region->align > PMD_SIZE) {
PMD_SIZE          143 drivers/dax/device.c 			(pmd_addr + PMD_SIZE) > vmf->vma->vm_end)
PMD_SIZE          147 drivers/dax/device.c 	phys = dax_pgoff_to_phys(dev_dax, pgoff, PMD_SIZE);
PMD_SIZE          236 drivers/dax/device.c 		fault_size = PMD_SIZE;
PMD_SIZE           49 fs/dax.c       #define PG_PMD_COLOUR	((PMD_SIZE >> PAGE_SHIFT) - 1)
PMD_SIZE           50 fs/dax.c       #define PG_PMD_NR	(PMD_SIZE >> PAGE_SHIFT)
PMD_SIZE          308 fs/dax.c       		return PMD_SIZE;
PMD_SIZE         1511 fs/dax.c       	if ((pmd_addr + PMD_SIZE) > vma->vm_end)
PMD_SIZE         1553 fs/dax.c       	error = ops->iomap_begin(inode, pos, PMD_SIZE, iomap_flags, &iomap);
PMD_SIZE         1557 fs/dax.c       	if (iomap.offset + iomap.length < pos + PMD_SIZE)
PMD_SIZE         1564 fs/dax.c       		error = dax_iomap_pfn(&iomap, pos, PMD_SIZE, &pfn);
PMD_SIZE         1585 fs/dax.c       		trace_dax_pmd_insert_mapping(inode, vmf, PMD_SIZE, pfn, entry);
PMD_SIZE         1601 fs/dax.c       		int copied = PMD_SIZE;
PMD_SIZE         1611 fs/dax.c       		ops->iomap_end(inode, pos, PMD_SIZE, copied, iomap_flags,
PMD_SIZE         1256 fs/proc/task_mmu.c #define PAGEMAP_WALK_SIZE	(PMD_SIZE)
PMD_SIZE           23 include/asm-generic/pgtable-nopmd.h #define PMD_MASK  	(~(PMD_SIZE-1))
PMD_SIZE          550 include/asm-generic/pgtable.h ({	unsigned long __boundary = ((addr) + PMD_SIZE) & PMD_MASK;	\
PMD_SIZE          520 include/asm-generic/tlb.h 		if (_sz == PMD_SIZE)				\
PMD_SIZE          539 include/linux/hugetlb.h 	if (huge_page_size(h) == PMD_SIZE)
PMD_SIZE          444 include/linux/mmu_notifier.h 							PMD_SIZE);	\
PMD_SIZE          466 include/linux/mmu_notifier.h 					    ___address + PMD_SIZE);	\
PMD_SIZE           88 lib/ioremap.c  	if ((end - addr) != PMD_SIZE)
PMD_SIZE           91 lib/ioremap.c  	if (!IS_ALIGNED(addr, PMD_SIZE))
PMD_SIZE           94 lib/ioremap.c  	if (!IS_ALIGNED(phys_addr, PMD_SIZE))
PMD_SIZE          567 mm/huge_memory.c 	ret = __thp_get_unmapped_area(filp, addr, len, off, flags, PMD_SIZE);
PMD_SIZE         1909 mm/huge_memory.c 			flush_tlb_range(vma, old_addr, old_addr + PMD_SIZE);
PMD_SIZE         4993 mm/hugetlb.c   			BUG_ON(sz != PMD_SIZE);
PMD_SIZE         5039 mm/hugetlb.c   	if (sz != PMD_SIZE && pmd_none(pmd_entry))
PMD_SIZE          122 mm/kasan/init.c 		if (IS_ALIGNED(addr, PMD_SIZE) && end - addr >= PMD_SIZE) {
PMD_SIZE          390 mm/kasan/init.c 			if (IS_ALIGNED(addr, PMD_SIZE) &&
PMD_SIZE          391 mm/kasan/init.c 			    IS_ALIGNED(next, PMD_SIZE))
PMD_SIZE         1138 mm/memory-failure.c 	if (huge_page_size(page_hstate(head)) > PMD_SIZE) {
PMD_SIZE          343 mm/memory.c    		addr += PMD_SIZE;
PMD_SIZE          353 mm/memory.c    		end -= PMD_SIZE;
PMD_SIZE          391 mm/memory.c    			while (next && next->vm_start <= vma->vm_end + PMD_SIZE
PMD_SIZE         4164 mm/memory.c    						(address & PMD_MASK) + PMD_SIZE);
PMD_SIZE          204 mm/mremap.c    	    || old_end - old_addr < PMD_SIZE)
PMD_SIZE          231 mm/mremap.c    	flush_tlb_range(vma, old_addr, old_addr + PMD_SIZE);
PMD_SIZE          258 mm/mremap.c    		next = (old_addr + PMD_SIZE) & PMD_MASK;
PMD_SIZE          285 mm/mremap.c    		} else if (extent == PMD_SIZE) {
PMD_SIZE          306 mm/mremap.c    		next = (new_addr + PMD_SIZE) & PMD_MASK;
PMD_SIZE          229 mm/page_vma_mapped.c 			if (pvmw->address % PMD_SIZE == 0) {
PMD_SIZE          442 mm/sparse.c    	return ALIGN(sizeof(struct page) * PAGES_PER_SECTION, PMD_SIZE);
PMD_SIZE          637 mm/swap_state.c 		    PFN_DOWN((faddr & PMD_MASK) + PMD_SIZE));
PMD_SIZE         1719 virt/kvm/arm/mmu.c 	if (vma_pagesize == PMD_SIZE ||
PMD_SIZE         1785 virt/kvm/arm/mmu.c 		if (fault_supports_stage2_huge_mapping(memslot, hva, PMD_SIZE) &&
PMD_SIZE         1787 virt/kvm/arm/mmu.c 			vma_pagesize = PMD_SIZE;
PMD_SIZE         1821 virt/kvm/arm/mmu.c 	} else if (vma_pagesize == PMD_SIZE) {
PMD_SIZE         2107 virt/kvm/arm/mmu.c 	WARN_ON(size != PAGE_SIZE && size != PMD_SIZE && size != PUD_SIZE);
PMD_SIZE         2125 virt/kvm/arm/mmu.c 	WARN_ON(size != PAGE_SIZE && size != PMD_SIZE && size != PUD_SIZE);