/linux-4.4.14/arch/x86/include/asm/ |
D | pgtable_64_types.h | 50 #define PUD_SIZE (_AC(1, UL) << PUD_SHIFT) macro 51 #define PUD_MASK (~(PUD_SIZE - 1))
|
/linux-4.4.14/include/asm-generic/ |
D | pgtable-nopud.h | 17 #define PUD_SIZE (1UL << PUD_SHIFT) macro 18 #define PUD_MASK (~(PUD_SIZE-1))
|
D | 4level-fixup.h | 8 #define PUD_SIZE PGDIR_SIZE macro
|
D | pgtable.h | 339 ({ unsigned long __boundary = ((addr) + PUD_SIZE) & PUD_MASK; \
|
/linux-4.4.14/arch/powerpc/include/asm/ |
D | pgtable-ppc64-4k.h | 35 #define PUD_SIZE (1UL << PUD_SHIFT) macro 36 #define PUD_MASK (~(PUD_SIZE-1))
|
/linux-4.4.14/lib/ |
D | ioremap.c | 111 ((next - addr) == PUD_SIZE) && in ioremap_pud_range() 112 IS_ALIGNED(phys_addr + addr, PUD_SIZE)) { in ioremap_pud_range()
|
/linux-4.4.14/arch/s390/mm/ |
D | vmem.c | 99 !(address & ~PUD_MASK) && (address + PUD_SIZE <= end)) { in vmem_add_mem() 103 address += PUD_SIZE; in vmem_add_mem() 165 address += PUD_SIZE; in vmem_remove_range() 170 address += PUD_SIZE; in vmem_remove_range()
|
D | dump_pagetables.c | 166 addr += PUD_SIZE; in walk_pud_level()
|
/linux-4.4.14/mm/kasan/ |
D | kasan_init.c | 92 if (IS_ALIGNED(addr, PUD_SIZE) && end - addr >= PUD_SIZE) { in zero_pud_populate()
|
/linux-4.4.14/arch/x86/mm/ |
D | init.c | 230 unsigned long start = round_down(mr[i].start, PUD_SIZE); in adjust_range_page_size_mask() 231 unsigned long end = round_up(mr[i].end, PUD_SIZE); in adjust_range_page_size_mask() 302 end_pfn = round_up(pfn, PFN_DOWN(PUD_SIZE)); in split_mem_range() 315 start_pfn = round_up(pfn, PFN_DOWN(PUD_SIZE)); in split_mem_range() 316 end_pfn = round_down(limit_pfn, PFN_DOWN(PUD_SIZE)); in split_mem_range()
|
D | pageattr.c | 839 if (start & (PUD_SIZE - 1)) { in unmap_pud_range() 840 unsigned long next_page = (start + PUD_SIZE) & PUD_MASK; in unmap_pud_range() 852 while (end - start >= PUD_SIZE) { in unmap_pud_range() 857 unmap_pmd_range(pud, start, start + PUD_SIZE); in unmap_pud_range() 859 start += PUD_SIZE; in unmap_pud_range() 1015 if (start & (PUD_SIZE - 1)) { in populate_pud() 1017 unsigned long next_page = (start + PUD_SIZE) & PUD_MASK; in populate_pud() 1050 while (end - start >= PUD_SIZE) { in populate_pud() 1054 start += PUD_SIZE; in populate_pud() 1055 cpa->pfn += PUD_SIZE; in populate_pud() [all …]
|
D | init_64.c | 79 next = (addr & PUD_MASK) + PUD_SIZE; in ident_pud_init() 534 next = (addr & PUD_MASK) + PUD_SIZE; in phys_pud_init() 935 if (IS_ALIGNED(addr, PUD_SIZE) && in remove_pud_table() 936 IS_ALIGNED(next, PUD_SIZE)) { in remove_pud_table() 939 get_order(PUD_SIZE)); in remove_pud_table() 951 PUD_SIZE)) { in remove_pud_table() 953 get_order(PUD_SIZE)); in remove_pud_table()
|
D | hugetlbpage.c | 165 } else if (ps == PUD_SIZE && cpu_has_gbpages) { in setup_hugepagesz()
|
D | pgtable.c | 588 mtrr = mtrr_type_lookup(addr, addr + PUD_SIZE, &uniform); in pud_set_huge()
|
/linux-4.4.14/arch/tile/mm/ |
D | hugetlbpage.c | 263 } else if (ps >= PUD_SIZE) { in __setup_hugepagesz() 267 if (hv_jpage_size != PUD_SIZE) { in __setup_hugepagesz() 269 PUD_SIZE >> 20, hv_jpage_size); in __setup_hugepagesz()
|
/linux-4.4.14/arch/arm64/include/asm/ |
D | pgtable-hwdef.h | 69 #define PUD_SIZE (_AC(1, UL) << PUD_SHIFT) macro 70 #define PUD_MASK (~(PUD_SIZE-1))
|
D | kvm_mmu.h | 266 kvm_flush_dcache_to_poc(page_address(page), PUD_SIZE);
|
D | pgtable.h | 43 #define VMEMMAP_SIZE ALIGN((1UL << (VA_BITS - PAGE_SHIFT)) * sizeof(struct page), PUD_SIZE) 52 #define VMALLOC_END (PAGE_OFFSET - PUD_SIZE - VMEMMAP_SIZE - SZ_64K)
|
/linux-4.4.14/arch/arm64/mm/ |
D | hugetlbpage.c | 49 } else if (ps == PUD_SIZE) { in setup_hugepagesz()
|
D | dump.c | 285 addr = start + i * PUD_SIZE; in walk_pud()
|
/linux-4.4.14/arch/tile/include/asm/ |
D | hugetlb.h | 106 if (pagesize != PUD_SIZE && pagesize != PMD_SIZE) in arch_make_huge_pte()
|
/linux-4.4.14/arch/x86/kernel/ |
D | head64.c | 151 BUILD_BUG_ON(MODULES_LEN + KERNEL_IMAGE_SIZE > 2*PUD_SIZE); in x86_64_start_kernel()
|
/linux-4.4.14/arch/frv/include/asm/ |
D | pgtable.h | 129 #define PUD_SIZE (1UL << PUD_SHIFT) macro 130 #define PUD_MASK (~(PUD_SIZE - 1))
|
/linux-4.4.14/arch/ia64/include/asm/ |
D | pgtable.h | 110 #define PUD_SIZE (1UL << PUD_SHIFT) macro 111 #define PUD_MASK (~(PUD_SIZE-1))
|
/linux-4.4.14/arch/arm/mm/ |
D | dump.c | 294 addr = start + i * PUD_SIZE; in walk_pud()
|
/linux-4.4.14/arch/x86/xen/ |
D | mmu.c | 1173 va += PUD_SIZE; in xen_cleanmfnmap() 1176 xen_free_ro_pages(pa, PUD_SIZE); in xen_cleanmfnmap() 1177 va += PUD_SIZE; in xen_cleanmfnmap() 1262 xen_cleanhighmap(MODULES_VADDR, roundup(MODULES_VADDR, PUD_SIZE) - 1); in xen_pagetable_cleanhighmap() 2101 n_pmd = roundup(size, PUD_SIZE) >> PUD_SHIFT; in xen_relocate_p2m()
|
/linux-4.4.14/arch/sparc/include/asm/ |
D | pgtable_64.h | 57 #define PUD_SIZE (_AC(1,UL) << PUD_SHIFT) macro 58 #define PUD_MASK (~(PUD_SIZE-1))
|
/linux-4.4.14/arch/sparc/mm/ |
D | init_64.c | 1487 return vstart + PUD_SIZE; in kernel_map_hugepud() 1497 pte_val += PUD_SIZE; in kernel_map_hugepud() 1498 vstart += PUD_SIZE; in kernel_map_hugepud() 1507 if (guard && !(vstart & ~PUD_MASK) && (vend - vstart) >= PUD_SIZE) in kernel_can_map_hugepud()
|
/linux-4.4.14/arch/um/kernel/ |
D | tlb.c | 339 last = ADD_ROUND(addr, PUD_SIZE); in flush_tlb_kernel_range_common()
|
/linux-4.4.14/arch/ia64/mm/ |
D | init.c | 400 end_address += PUD_SIZE; in vmemmap_find_next_valid_pfn()
|
/linux-4.4.14/arch/s390/include/asm/ |
D | pgtable.h | 72 #define PUD_SIZE (1UL << PUD_SHIFT) macro 73 #define PUD_MASK (~(PUD_SIZE-1))
|
/linux-4.4.14/mm/ |
D | hugetlb.c | 4147 unsigned long s_end = sbase + PUD_SIZE; in page_table_shareable() 4168 unsigned long end = base + PUD_SIZE; in vma_shareable() 4290 if (sz == PUD_SIZE) { in huge_pte_alloc()
|