Home
last modified time | relevance | path

Searched refs:PUD_SIZE (Results 1 – 33 of 33) sorted by relevance

/linux-4.1.27/arch/x86/include/asm/
Dpgtable_64_types.h50 #define PUD_SIZE (_AC(1, UL) << PUD_SHIFT) macro
51 #define PUD_MASK (~(PUD_SIZE - 1))
/linux-4.1.27/include/asm-generic/
Dpgtable-nopud.h17 #define PUD_SIZE (1UL << PUD_SHIFT) macro
18 #define PUD_MASK (~(PUD_SIZE-1))
D4level-fixup.h8 #define PUD_SIZE PGDIR_SIZE macro
Dpgtable.h296 ({ unsigned long __boundary = ((addr) + PUD_SIZE) & PUD_MASK; \
/linux-4.1.27/arch/powerpc/include/asm/
Dpgtable-ppc64-4k.h35 #define PUD_SIZE (1UL << PUD_SHIFT) macro
36 #define PUD_MASK (~(PUD_SIZE-1))
/linux-4.1.27/arch/x86/mm/
Dkasan_init_64.c111 while (IS_ALIGNED(addr, PUD_SIZE) && addr + PUD_SIZE <= end) { in zero_pud_populate()
115 addr += PUD_SIZE; in zero_pud_populate()
Dinit.c227 unsigned long start = round_down(mr[i].start, PUD_SIZE); in adjust_range_page_size_mask()
228 unsigned long end = round_up(mr[i].end, PUD_SIZE); in adjust_range_page_size_mask()
299 end_pfn = round_up(pfn, PFN_DOWN(PUD_SIZE)); in split_mem_range()
312 start_pfn = round_up(pfn, PFN_DOWN(PUD_SIZE)); in split_mem_range()
313 end_pfn = round_down(limit_pfn, PFN_DOWN(PUD_SIZE)); in split_mem_range()
Dpageattr.c812 if (start & (PUD_SIZE - 1)) { in unmap_pud_range()
813 unsigned long next_page = (start + PUD_SIZE) & PUD_MASK; in unmap_pud_range()
825 while (end - start >= PUD_SIZE) { in unmap_pud_range()
830 unmap_pmd_range(pud, start, start + PUD_SIZE); in unmap_pud_range()
832 start += PUD_SIZE; in unmap_pud_range()
988 if (start & (PUD_SIZE - 1)) { in populate_pud()
990 unsigned long next_page = (start + PUD_SIZE) & PUD_MASK; in populate_pud()
1023 while (end - start >= PUD_SIZE) { in populate_pud()
1027 start += PUD_SIZE; in populate_pud()
1028 cpa->pfn += PUD_SIZE; in populate_pud()
[all …]
Dinit_64.c79 next = (addr & PUD_MASK) + PUD_SIZE; in ident_pud_init()
534 next = (addr & PUD_MASK) + PUD_SIZE; in phys_pud_init()
935 if (IS_ALIGNED(addr, PUD_SIZE) && in remove_pud_table()
936 IS_ALIGNED(next, PUD_SIZE)) { in remove_pud_table()
939 get_order(PUD_SIZE)); in remove_pud_table()
951 PUD_SIZE)) { in remove_pud_table()
953 get_order(PUD_SIZE)); in remove_pud_table()
Dhugetlbpage.c165 } else if (ps == PUD_SIZE && cpu_has_gbpages) { in setup_hugepagesz()
Dpgtable.c574 mtrr = mtrr_type_lookup(addr, addr + PUD_SIZE); in pud_set_huge()
/linux-4.1.27/lib/
Dioremap.c111 ((next - addr) == PUD_SIZE) && in ioremap_pud_range()
112 IS_ALIGNED(phys_addr + addr, PUD_SIZE)) { in ioremap_pud_range()
/linux-4.1.27/arch/s390/mm/
Dvmem.c99 !(address & ~PUD_MASK) && (address + PUD_SIZE <= end)) { in vmem_add_mem()
103 address += PUD_SIZE; in vmem_add_mem()
165 address += PUD_SIZE; in vmem_remove_range()
170 address += PUD_SIZE; in vmem_remove_range()
Ddump_pagetables.c166 addr += PUD_SIZE; in walk_pud_level()
/linux-4.1.27/arch/arm64/include/asm/
Dpgtable-hwdef.h36 #define PUD_SIZE (_AC(1, UL) << PUD_SHIFT) macro
37 #define PUD_MASK (~(PUD_SIZE-1))
Dkvm_mmu.h266 kvm_flush_dcache_to_poc(page_address(page), PUD_SIZE);
Dpgtable.h42 #define VMEMMAP_SIZE ALIGN((1UL << (VA_BITS - PAGE_SHIFT)) * sizeof(struct page), PUD_SIZE)
44 #define VMALLOC_END (PAGE_OFFSET - PUD_SIZE - VMEMMAP_SIZE - SZ_64K)
/linux-4.1.27/arch/tile/mm/
Dhugetlbpage.c268 } else if (ps >= PUD_SIZE) { in __setup_hugepagesz()
272 if (hv_jpage_size != PUD_SIZE) { in __setup_hugepagesz()
274 PUD_SIZE >> 20, hv_jpage_size); in __setup_hugepagesz()
/linux-4.1.27/arch/arm64/mm/
Dhugetlbpage.c60 } else if (ps == PUD_SIZE) { in setup_hugepagesz()
Ddump.c269 addr = start + i * PUD_SIZE; in walk_pud()
Dmmu.c362 limit = PHYS_OFFSET + PUD_SIZE; in map_mem()
/linux-4.1.27/arch/tile/include/asm/
Dhugetlb.h119 if (pagesize != PUD_SIZE && pagesize != PMD_SIZE) in arch_make_huge_pte()
/linux-4.1.27/arch/x86/kernel/
Dhead64.c151 BUILD_BUG_ON(MODULES_LEN + KERNEL_IMAGE_SIZE > 2*PUD_SIZE); in x86_64_start_kernel()
/linux-4.1.27/arch/frv/include/asm/
Dpgtable.h129 #define PUD_SIZE (1UL << PUD_SHIFT) macro
130 #define PUD_MASK (~(PUD_SIZE - 1))
/linux-4.1.27/arch/ia64/include/asm/
Dpgtable.h110 #define PUD_SIZE (1UL << PUD_SHIFT) macro
111 #define PUD_MASK (~(PUD_SIZE-1))
/linux-4.1.27/arch/arm/mm/
Ddump.c294 addr = start + i * PUD_SIZE; in walk_pud()
/linux-4.1.27/arch/sparc/include/asm/
Dpgtable_64.h57 #define PUD_SIZE (_AC(1,UL) << PUD_SHIFT) macro
58 #define PUD_MASK (~(PUD_SIZE-1))
/linux-4.1.27/arch/sparc/mm/
Dinit_64.c1430 return vstart + PUD_SIZE; in kernel_map_hugepud()
1440 pte_val += PUD_SIZE; in kernel_map_hugepud()
1441 vstart += PUD_SIZE; in kernel_map_hugepud()
1450 if (guard && !(vstart & ~PUD_MASK) && (vend - vstart) >= PUD_SIZE) in kernel_can_map_hugepud()
/linux-4.1.27/arch/s390/include/asm/
Dpgtable.h72 #define PUD_SIZE (1UL << PUD_SHIFT) macro
73 #define PUD_MASK (~(PUD_SIZE-1))
/linux-4.1.27/arch/ia64/mm/
Dinit.c407 end_address += PUD_SIZE; in vmemmap_find_next_valid_pfn()
/linux-4.1.27/arch/um/kernel/
Dtlb.c329 last = ADD_ROUND(addr, PUD_SIZE); in flush_tlb_kernel_range_common()
/linux-4.1.27/mm/
Dhugetlb.c3680 unsigned long s_end = sbase + PUD_SIZE; in page_table_shareable()
3701 unsigned long end = base + PUD_SIZE; in vma_shareable()
3818 if (sz == PUD_SIZE) { in huge_pte_alloc()
/linux-4.1.27/arch/x86/xen/
Dmmu.c1161 xen_cleanhighmap(MODULES_VADDR, roundup(MODULES_VADDR, PUD_SIZE) - 1); in xen_pagetable_p2m_free()