Home
last modified time | relevance | path

Searched refs:PMD_SIZE (Results 1 – 85 of 85) sorted by relevance

/linux-4.4.14/arch/x86/mm/
Dinit.c217 unsigned long start = round_down(mr[i].start, PMD_SIZE); in adjust_range_page_size_mask()
218 unsigned long end = round_up(mr[i].end, PMD_SIZE); in adjust_range_page_size_mask()
284 end_pfn = PFN_DOWN(PMD_SIZE); in split_mem_range()
286 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
288 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
298 start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
300 end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
303 if (end_pfn > round_down(limit_pfn, PFN_DOWN(PMD_SIZE))) in split_mem_range()
304 end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
325 start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
[all …]
Dinit_64.c63 for (; addr < end; addr += PMD_SIZE) { in ident_pmd_init()
336 for (; size; phys += PMD_SIZE, size -= PMD_SIZE) { in __init_extra_mapping()
382 unsigned long end = roundup((unsigned long)_brk_end, PMD_SIZE) - 1; in cleanup_highmap()
393 for (; vaddr + PMD_SIZE - 1 < vaddr_end; pmd++, vaddr += PMD_SIZE) { in cleanup_highmap()
460 next = (address & PMD_MASK) + PMD_SIZE; in phys_pmd_init()
879 if (IS_ALIGNED(addr, PMD_SIZE) && in remove_pmd_table()
880 IS_ALIGNED(next, PMD_SIZE)) { in remove_pmd_table()
883 get_order(PMD_SIZE)); in remove_pmd_table()
895 PMD_SIZE)) { in remove_pmd_table()
897 get_order(PMD_SIZE)); in remove_pmd_table()
[all …]
Dpageattr.c101 return __pa_symbol(roundup(_brk_end, PMD_SIZE)) >> PAGE_SHIFT; in highmap_end_pfn()
795 if (start & (PMD_SIZE - 1)) { in unmap_pmd_range()
796 unsigned long next_page = (start + PMD_SIZE) & PMD_MASK; in unmap_pmd_range()
808 while (end - start >= PMD_SIZE) { in unmap_pmd_range()
812 __unmap_pmd_range(pud, pmd, start, start + PMD_SIZE); in unmap_pmd_range()
814 start += PMD_SIZE; in unmap_pmd_range()
938 if (start & (PMD_SIZE - 1)) { in populate_pmd()
940 unsigned long next_page = (start + PMD_SIZE) & PMD_MASK; in populate_pmd()
967 while (end - start >= PMD_SIZE) { in populate_pmd()
981 start += PMD_SIZE; in populate_pmd()
[all …]
Dinit_32.c148 vaddr += PMD_SIZE; in page_table_range_init_count()
233 vaddr += PMD_SIZE; in page_table_range_init()
529 end = (FIXADDR_TOP + PMD_SIZE - 1) & PMD_MASK; in early_ioremap_page_table_range_init()
Dpgtable.c613 mtrr = mtrr_type_lookup(addr, addr + PMD_SIZE, &uniform); in pmd_set_huge()
617 __func__, addr, addr + PMD_SIZE); in pmd_set_huge()
Dhugetlbpage.c163 if (ps == PMD_SIZE) { in setup_hugepagesz()
Dfault.c239 address += PMD_SIZE) { in vmalloc_sync_all()
/linux-4.4.14/arch/sh/include/asm/
Dpgtable-3level.h22 #define PMD_SIZE (1UL << PMD_SHIFT) macro
23 #define PMD_MASK (~(PMD_SIZE-1))
25 #define PTRS_PER_PMD ((1 << PGDIR_SHIFT) / PMD_SIZE)
/linux-4.4.14/arch/tile/mm/
Dhugetlbpage.c72 if (sz >= PMD_SIZE) { in huge_pte_alloc()
73 BUG_ON(sz != PMD_SIZE && in huge_pte_alloc()
74 sz != (PMD_SIZE << huge_shift[HUGE_SHIFT_PMD])); in huge_pte_alloc()
84 BUG_ON(sz != PMD_SIZE); in huge_pte_alloc()
274 } else if (ps >= PMD_SIZE) { in __setup_hugepagesz()
329 BUILD_BUG_ON(ADDITIONAL_HUGE_SIZE >= PMD_SIZE || in add_default_hugepagesz()
Dinit.c181 start = round_down(start, PMD_SIZE); in page_table_range_init()
182 end = round_up(end, PMD_SIZE); in page_table_range_init()
183 for (vaddr = start; vaddr < end; vaddr += PMD_SIZE) { in page_table_range_init()
/linux-4.4.14/arch/s390/mm/
Dvmem.c116 !(address & ~PMD_MASK) && (address + PMD_SIZE <= end)) { in vmem_add_mem()
121 address += PMD_SIZE; in vmem_add_mem()
175 address += PMD_SIZE; in vmem_remove_range()
180 address += PMD_SIZE; in vmem_remove_range()
230 new_page = vmemmap_alloc_block(PMD_SIZE, node); in vmemmap_populate()
235 address = (address + PMD_SIZE) & PMD_MASK; in vmemmap_populate()
243 address = (address + PMD_SIZE) & PMD_MASK; in vmemmap_populate()
Dpgtable.c285 offset = (offset & (PTRS_PER_PMD - 1)) * PMD_SIZE; in __gmap_segment_gaddr()
342 if ((to | len) & (PMD_SIZE - 1)) in gmap_unmap_segment()
349 for (off = 0; off < len; off += PMD_SIZE) in gmap_unmap_segment()
373 if ((from | to | len) & (PMD_SIZE - 1)) in gmap_map_segment()
381 for (off = 0; off < len; off += PMD_SIZE) { in gmap_map_segment()
630 gaddr = (gaddr + PMD_SIZE) & PMD_MASK) { in gmap_discard()
640 size = min(to - gaddr, PMD_SIZE - (gaddr & ~PMD_MASK)); in gmap_discard()
Ddump_pagetables.c144 addr += PMD_SIZE; in walk_pmd_level()
/linux-4.4.14/arch/x86/include/asm/
Dpgtable_32_types.h11 # define PMD_SIZE (1UL << PMD_SHIFT) macro
12 # define PMD_MASK (~(PMD_SIZE - 1))
Dpgtable_64_types.h48 #define PMD_SIZE (_AC(1, UL) << PMD_SHIFT) macro
49 #define PMD_MASK (~(PMD_SIZE - 1))
/linux-4.4.14/arch/powerpc/include/asm/
Dpgtable-ppc64-64k.h27 #define PMD_SIZE (1UL << PMD_SHIFT) macro
28 #define PMD_MASK (~(PMD_SIZE-1))
Dpgtable-ppc64-4k.h27 #define PMD_SIZE (1UL << PMD_SHIFT) macro
28 #define PMD_MASK (~(PMD_SIZE-1))
/linux-4.4.14/arch/parisc/kernel/
Dpci-dma.c91 if (end > PMD_SIZE) in map_pte_uncached()
92 end = PMD_SIZE; in map_pte_uncached()
126 vaddr = (vaddr + PMD_SIZE) & PMD_MASK; in map_pmd_uncached()
127 orig_vaddr += PMD_SIZE; in map_pmd_uncached()
171 if (end > PMD_SIZE) in unmap_uncached_pte()
172 end = PMD_SIZE; in unmap_uncached_pte()
211 vaddr = (vaddr + PMD_SIZE) & PMD_MASK; in unmap_uncached_pmd()
212 orig_vaddr += PMD_SIZE; in unmap_uncached_pmd()
/linux-4.4.14/include/asm-generic/
Dpgtable-nopmd.h21 #define PMD_SIZE (1UL << PMD_SHIFT) macro
22 #define PMD_MASK (~(PMD_SIZE-1))
Dpgtable.h346 ({ unsigned long __boundary = ((addr) + PMD_SIZE) & PMD_MASK; \
/linux-4.4.14/arch/nios2/mm/
Dioremap.c33 if (end > PMD_SIZE) in remap_area_pte()
34 end = PMD_SIZE; in remap_area_pte()
70 address = (address + PMD_SIZE) & PMD_MASK; in remap_area_pmd()
/linux-4.4.14/arch/mips/mm/
Dioremap.c29 if (end > PMD_SIZE) in remap_area_pte()
30 end = PMD_SIZE; in remap_area_pte()
61 address = (address + PMD_SIZE) & PMD_MASK; in remap_area_pmd()
Dinit.c247 vaddr += PMD_SIZE; in fixrange_init()
/linux-4.4.14/arch/m68k/include/asm/
Dpgtable_mm.h37 #define PMD_SIZE (1UL << PMD_SHIFT) macro
38 #define PMD_MASK (~(PMD_SIZE-1))
/linux-4.4.14/arch/arm/include/asm/
Dpgtable-2level.h88 #define PMD_SIZE (1UL << PMD_SHIFT) macro
89 #define PMD_MASK (~(PMD_SIZE-1))
Dkvm_mmu.h148 ({ u64 __boundary = ((addr) + PMD_SIZE) & PMD_MASK; \
248 unsigned long size = PMD_SIZE; in __kvm_flush_dcache_pmd()
Dhighmem.h6 #define PKMAP_BASE (PAGE_OFFSET - PMD_SIZE)
Dmemory.h67 #define MODULES_END (PAGE_OFFSET - PMD_SIZE)
Dpgtable-3level.h50 #define PMD_SIZE (1UL << PMD_SHIFT) macro
/linux-4.4.14/arch/um/include/asm/
Dpgtable-3level.h27 #define PMD_SIZE (1UL << PMD_SHIFT) macro
28 #define PMD_MASK (~(PMD_SIZE-1))
/linux-4.4.14/arch/tile/include/asm/
Dpgtable_64.h32 #define PMD_SIZE HPAGE_SIZE macro
33 #define PMD_MASK (~(PMD_SIZE-1))
Dhugetlb.h106 if (pagesize != PUD_SIZE && pagesize != PMD_SIZE) in arch_make_huge_pte()
/linux-4.4.14/lib/
Dioremap.c85 ((next - addr) == PMD_SIZE) && in ioremap_pmd_range()
86 IS_ALIGNED(phys_addr + addr, PMD_SIZE)) { in ioremap_pmd_range()
/linux-4.4.14/mm/kasan/
Dkasan_init.c71 if (IS_ALIGNED(addr, PMD_SIZE) && end - addr >= PMD_SIZE) { in zero_pmd_populate()
/linux-4.4.14/fs/
Ddax.c524 #define PG_PMD_COLOUR ((PMD_SIZE >> PAGE_SHIFT) - 1)
554 if ((pmd_addr + PMD_SIZE) > vma->vm_end) in __dax_pmd_fault()
568 bh.b_size = PMD_SIZE; in __dax_pmd_fault()
579 if (!buffer_size_valid(&bh) || bh.b_size < PMD_SIZE) in __dax_pmd_fault()
588 unmap_mapping_range(mapping, pgoff << PAGE_SHIFT, PMD_SIZE, 0); in __dax_pmd_fault()
633 if ((length < PMD_SIZE) || (pfn & PG_PMD_COLOUR)) in __dax_pmd_fault()
/linux-4.4.14/arch/arm64/include/asm/
Dpgtable-hwdef.h59 #define PMD_SIZE (_AC(1, UL) << PMD_SHIFT) macro
60 #define PMD_MASK (~(PMD_SIZE-1))
Dkvm_mmu.h260 kvm_flush_dcache_to_poc(page_address(page), PMD_SIZE);
/linux-4.4.14/arch/arm64/mm/
Dhugetlbpage.c47 if (ps == PMD_SIZE) { in setup_hugepagesz()
Dmmu.c126 addr += PMD_SIZE; in split_pud()
538 p = vmemmap_alloc_block_buf(PMD_SIZE, node); in vmemmap_populate()
Ddump.c268 addr = start + i * PMD_SIZE; in walk_pmd()
/linux-4.4.14/arch/metag/include/asm/
Dhighmem.h26 #define PKMAP_BASE (FIXADDR_START - PMD_SIZE)
/linux-4.4.14/arch/x86/xen/
Dp2m.c407 (unsigned long)(p2m + pfn) + i * PMD_SIZE); in xen_rebuild_p2m_list()
423 PMD_SIZE * PMDS_PER_MID_PAGE); in xen_vmalloc_p2m_tree()
424 vm_area_register_early(&vm, PMD_SIZE * PMDS_PER_MID_PAGE); in xen_vmalloc_p2m_tree()
491 vaddr = addr & ~(PMD_SIZE * PMDS_PER_MID_PAGE - 1); in alloc_p2m_pmd()
520 vaddr += PMD_SIZE; in alloc_p2m_pmd()
Dmmu.c1111 unsigned long kernel_end = roundup((unsigned long)_brk_end, PMD_SIZE) - 1; in xen_cleanhighmap()
1117 pmd++, vaddr += PMD_SIZE) { in xen_cleanhighmap()
1182 xen_free_ro_pages(pa, PMD_SIZE); in xen_cleanmfnmap()
1194 va += PMD_SIZE; in xen_cleanmfnmap()
1228 size = roundup(size, PMD_SIZE); in xen_pagetable_p2m_free()
1254 size = roundup(xen_start_info->nr_pt_frames * PAGE_SIZE, PMD_SIZE); in xen_pagetable_cleanhighmap()
2100 n_pt = roundup(size, PMD_SIZE) >> PMD_SHIFT; in xen_relocate_p2m()
/linux-4.4.14/arch/arm/mm/
Dmmu.c1041 next = (addr + PMD_SIZE - 1) & PMD_MASK; in fill_pmd_gaps()
1173 if (!IS_ALIGNED(block_start, PMD_SIZE)) in sanity_check_meminfo()
1175 else if (!IS_ALIGNED(block_end, PMD_SIZE)) in sanity_check_meminfo()
1193 memblock_limit = round_down(memblock_limit, PMD_SIZE); in sanity_check_meminfo()
1208 for (addr = 0; addr < MODULES_VADDR; addr += PMD_SIZE) in prepare_page_table()
1213 addr = ((unsigned long)_etext + PMD_SIZE - 1) & PMD_MASK; in prepare_page_table()
1215 for ( ; addr < PAGE_OFFSET; addr += PMD_SIZE) in prepare_page_table()
1230 addr < VMALLOC_START; addr += PMD_SIZE) in prepare_page_table()
1285 for (addr = VMALLOC_START; addr < (FIXADDR_TOP & PMD_MASK); addr += PMD_SIZE) in devicemaps_init()
Dioremap.c171 addr += PMD_SIZE; in unmap_area_sections()
210 addr += PMD_SIZE; in remap_area_sections()
247 addr += PMD_SIZE; in remap_area_supersections()
Ddump.c276 addr = start + i * PMD_SIZE; in walk_pmd()
282 if (SECTION_SIZE < PMD_SIZE && pmd_large(pmd[1])) in walk_pmd()
Ddma-mapping.c443 addr += PMD_SIZE) in dma_contiguous_remap()
/linux-4.4.14/mm/
Dsparse-vmemmap.c204 size = ALIGN(size, PMD_SIZE); in sparse_mem_maps_populate_node()
206 PMD_SIZE, __pa(MAX_DMA_ADDRESS)); in sparse_mem_maps_populate_node()
Dmremap.c183 next = (old_addr + PMD_SIZE) & PMD_MASK; in move_page_tables()
219 next = (new_addr + PMD_SIZE) & PMD_MASK; in move_page_tables()
Dmemory.c506 addr += PMD_SIZE; in free_pgd_range()
516 end -= PMD_SIZE; in free_pgd_range()
550 while (next && next->vm_start <= vma->vm_end + PMD_SIZE in free_pgtables()
Dhugetlb.c4293 BUG_ON(sz != PMD_SIZE); in huge_pte_alloc()
/linux-4.4.14/arch/mips/include/asm/
Dpgtable-64.h53 #define PMD_SIZE (1UL << PMD_SHIFT) macro
54 #define PMD_MASK (~(PMD_SIZE-1))
/linux-4.4.14/arch/m32r/include/asm/
Dpgtable.h50 #define PMD_SIZE (1UL << PMD_SHIFT) macro
51 #define PMD_MASK (~(PMD_SIZE - 1))
/linux-4.4.14/arch/x86/platform/efi/
Defi_64.c245 if (!(pa & (PMD_SIZE - 1))) { in efi_map_region()
248 u64 pa_offset = pa & (PMD_SIZE - 1); in efi_map_region()
255 efi_va -= PMD_SIZE; in efi_map_region()
/linux-4.4.14/arch/metag/mm/
Dinit.c280 vaddr += PMD_SIZE; in allocate_pgtables()
304 end = (FIXADDR_TOP + PMD_SIZE - 1) & PMD_MASK; in fixedrange_init()
/linux-4.4.14/include/linux/
Dmmu_notifier.h332 PMD_SIZE); \
354 ___address + PMD_SIZE); \
Dhugetlb.h471 if (huge_page_size(h) == PMD_SIZE) in huge_pte_lockptr()
/linux-4.4.14/arch/avr32/include/asm/
Dpgtable.h27 #define PMD_SIZE (1UL << PMD_SHIFT) macro
28 #define PMD_MASK (~(PMD_SIZE-1))
/linux-4.4.14/arch/sparc/include/asm/
Dpgtable_32.h37 #define PMD_SIZE (1UL << PMD_SHIFT) macro
38 #define PMD_MASK (~(PMD_SIZE-1))
Dpgtable_64.h49 #define PMD_SIZE (_AC(1,UL) << PMD_SHIFT) macro
50 #define PMD_MASK (~(PMD_SIZE-1))
/linux-4.4.14/arch/alpha/include/asm/
Dpgtable.h32 #define PMD_SIZE (1UL << PMD_SHIFT) macro
33 #define PMD_MASK (~(PMD_SIZE-1))
/linux-4.4.14/arch/mn10300/mm/
Dpgtable.c41 if (vaddr & (PMD_SIZE-1)) { /* vaddr is misaligned */ in set_pmd_pfn()
/linux-4.4.14/arch/sh/mm/
Dinit.c184 vaddr += PMD_SIZE; in page_table_range_init()
374 end = (FIXADDR_TOP + PMD_SIZE - 1) & PMD_MASK; in paging_init()
/linux-4.4.14/arch/m68k/sun3x/
Ddvma.c124 end3 = (vaddr + (PMD_SIZE-1)) & PMD_MASK; in dvma_map_cpu()
/linux-4.4.14/arch/microblaze/include/asm/
Dpgtable.h143 #define PMD_SIZE (1UL << PMD_SHIFT) macro
144 #define PMD_MASK (~(PMD_SIZE-1))
/linux-4.4.14/arch/x86/kernel/
Dhead_64.S139 addq $PMD_SIZE, %rax
502 KERNEL_IMAGE_SIZE/PMD_SIZE)
Dsetup_percpu.c200 atom_size = PMD_SIZE; in setup_per_cpu_areas()
/linux-4.4.14/arch/frv/include/asm/
Dpgtable.h135 #define PMD_SIZE (1UL << PMD_SHIFT) macro
136 #define PMD_MASK (~(PMD_SIZE - 1))
/linux-4.4.14/arch/sparc/mm/
Dinit_64.c1527 return vstart + PMD_SIZE; in kernel_map_hugepmd()
1544 pte_val += PMD_SIZE; in kernel_map_hugepmd()
1545 vstart += PMD_SIZE; in kernel_map_hugepmd()
1555 if (guard && !(vstart & ~PMD_MASK) && (vend - vstart) >= PMD_SIZE) in kernel_can_map_hugepmd()
1616 this_end = (vstart + PMD_SIZE) & PMD_MASK; in kernel_map_range()
2412 vend = ALIGN(vend, PMD_SIZE); in vmemmap_populate()
2413 for (; vstart < vend; vstart += PMD_SIZE) { in vmemmap_populate()
2440 void *block = vmemmap_alloc_block(PMD_SIZE, node); in vmemmap_populate()
Dsrmmu.c684 if (start > (0xffffffffUL - PMD_SIZE)) in srmmu_early_allocate_ptable_skeleton()
686 start = (start + PMD_SIZE) & PMD_MASK; in srmmu_early_allocate_ptable_skeleton()
715 if (start > (0xffffffffUL - PMD_SIZE)) in srmmu_allocate_ptable_skeleton()
717 start = (start + PMD_SIZE) & PMD_MASK; in srmmu_allocate_ptable_skeleton()
/linux-4.4.14/arch/powerpc/mm/
Dtlb_hash64.c238 addr = _ALIGN_DOWN(addr, PMD_SIZE); in flush_tlb_pmd_range()
Dtlb_nohash.c404 unsigned long end = address + PMD_SIZE; in tlb_flush_pgtable()
/linux-4.4.14/arch/um/kernel/
Dmem.c108 vaddr += PMD_SIZE; in fixrange_init()
Dtlb.c356 last = ADD_ROUND(addr, PMD_SIZE); in flush_tlb_kernel_range_common()
/linux-4.4.14/arch/x86/entry/vdso/
Dvma.c70 end = (start + len + PMD_SIZE - 1) & PMD_MASK; in vdso_addr()
/linux-4.4.14/arch/ia64/include/asm/
Dpgtable.h98 #define PMD_SIZE (1UL << PMD_SHIFT) macro
99 #define PMD_MASK (~(PMD_SIZE-1))
/linux-4.4.14/arch/parisc/include/asm/
Dpgtable.h114 #define PMD_SIZE (1UL << PMD_SHIFT) macro
115 #define PMD_MASK (~(PMD_SIZE-1))
/linux-4.4.14/Documentation/vm/
Dsplit_page_table_lock44 takes pmd split lock for PMD_SIZE page, mm->page_table_lock
/linux-4.4.14/arch/mn10300/include/asm/
Dpgtable.h65 #define PMD_SIZE (1UL << PMD_SHIFT) macro
/linux-4.4.14/drivers/nvdimm/
Dpmem.c266 offset = ALIGN(SZ_8K + 64 * npfns, PMD_SIZE); in nd_pfn_init()
/linux-4.4.14/arch/ia64/mm/
Dinit.c406 end_address += PMD_SIZE; in vmemmap_find_next_valid_pfn()
/linux-4.4.14/arch/s390/include/asm/
Dpgtable.h70 #define PMD_SIZE (1UL << PMD_SHIFT) macro
71 #define PMD_MASK (~(PMD_SIZE-1))
/linux-4.4.14/fs/ext4/
Dfile.c257 PMD_SIZE / PAGE_SIZE)); in ext4_dax_pmd_fault()
/linux-4.4.14/fs/proc/
Dtask_mmu.c1002 #define PAGEMAP_WALK_SIZE (PMD_SIZE)
/linux-4.4.14/arch/arm/kvm/
Dmmu.c1324 coherent_cache_guest_page(vcpu, pfn, PMD_SIZE, fault_ipa_uncached); in user_mem_abort()