/linux-4.4.14/arch/unicore32/include/asm/ |
D | tlbflush.h | 94 : : "r" (uaddr & PAGE_MASK) : "cc"); in local_flush_tlb_page() 97 : : "r" (uaddr & PAGE_MASK) : "cc"); in local_flush_tlb_page() 101 : : "r" (uaddr & PAGE_MASK) : "cc"); in local_flush_tlb_page() 111 : : "r" (kaddr & PAGE_MASK) : "cc"); in local_flush_tlb_kernel_page() 114 : : "r" (kaddr & PAGE_MASK) : "cc"); in local_flush_tlb_kernel_page() 118 : : "r" (kaddr & PAGE_MASK) : "cc"); in local_flush_tlb_kernel_page()
|
D | page.h | 18 #define PAGE_MASK (~(PAGE_SIZE-1)) macro
|
D | cacheflush.h | 153 __cpuc_coherent_user_range((start) & PAGE_MASK, PAGE_ALIGN(end))
|
/linux-4.4.14/arch/sh/mm/ |
D | tlbflush_64.c | 72 page &= PAGE_MASK; in local_flush_tlb_page() 94 start &= PAGE_MASK; in local_flush_tlb_range() 95 end &= PAGE_MASK; in local_flush_tlb_range() 105 pteh_epn = pteh & PAGE_MASK; in local_flush_tlb_range() 106 pteh_low = pteh & ~PAGE_MASK; in local_flush_tlb_range() 118 pteh_epn = pteh & PAGE_MASK; in local_flush_tlb_range() 119 pteh_low = pteh & ~PAGE_MASK; in local_flush_tlb_range()
|
D | tlbflush_32.c | 25 page &= PAGE_MASK; in local_flush_tlb_page() 60 start &= PAGE_MASK; in local_flush_tlb_range() 62 end &= PAGE_MASK; in local_flush_tlb_range() 93 start &= PAGE_MASK; in local_flush_tlb_kernel_range() 95 end &= PAGE_MASK; in local_flush_tlb_kernel_range()
|
D | ioremap.c | 67 offset = phys_addr & ~PAGE_MASK; in __ioremap_caller() 68 phys_addr &= PAGE_MASK; in __ioremap_caller() 129 p = remove_vm_area((void *)(vaddr & PAGE_MASK)); in __iounmap()
|
D | mmap.c | 74 info.align_mask = do_colour_align ? (PAGE_MASK & shm_align_mask) : 0; in arch_get_unmapped_area() 124 info.align_mask = do_colour_align ? (PAGE_MASK & shm_align_mask) : 0; in arch_get_unmapped_area_topdown() 134 if (addr & ~PAGE_MASK) { in arch_get_unmapped_area_topdown()
|
D | tlb-sh5.c | 126 pteh &= PAGE_MASK; in sh64_setup_tlb_slot() 129 ptel &= PAGE_MASK; in sh64_setup_tlb_slot() 163 paddr &= ~PAGE_MASK; in tlb_wire_entry()
|
D | ioremap_fixed.c | 58 offset = phys_addr & ~PAGE_MASK; in ioremap_fixed() 59 phys_addr &= PAGE_MASK; in ioremap_fixed()
|
D | cache.c | 64 void *vto = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); in copy_to_user_page() 83 void *vfrom = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); in copy_from_user_page() 111 if (pages_do_alias((unsigned long)vto, vaddr & PAGE_MASK) || in copy_user_highpage() 127 if (pages_do_alias((unsigned long)kaddr, vaddr & PAGE_MASK)) in clear_user_highpage()
|
D | cache-sh5.c | 173 aligned_start = start & PAGE_MASK; in sh64_icache_inv_user_page_range() 174 after_last_page_start = PAGE_SIZE + ((end - 1) & PAGE_MASK); in sh64_icache_inv_user_page_range() 413 paddr = pte_val(entry) & PAGE_MASK; in sh64_dcache_purge_user_pages() 476 start &= PAGE_MASK; /* should already be so */ in sh64_dcache_purge_user_range()
|
D | gup.c | 173 start &= PAGE_MASK; in __get_user_pages_fast() 226 start &= PAGE_MASK; in get_user_pages_fast()
|
D | kmap.c | 57 unsigned long vaddr = (unsigned long)kvaddr & PAGE_MASK; in kunmap_coherent()
|
D | tlbex_32.c | 71 local_flush_tlb_one(get_asid(), address & PAGE_MASK); in handle_tlbmiss()
|
/linux-4.4.14/arch/microblaze/mm/ |
D | pgtable.c | 63 p = addr & PAGE_MASK; in __ioremap() 120 return (void __iomem *) (v + ((unsigned long)addr & ~PAGE_MASK)); in __ioremap() 133 vfree((void *) (PAGE_MASK & (unsigned long) addr)); in iounmap() 198 pgd = pgd_offset(mm, addr & PAGE_MASK); in get_pteptr() 200 pmd = pmd_offset(pgd, addr & PAGE_MASK); in get_pteptr() 202 pte = pte_offset_kernel(pmd, addr & PAGE_MASK); in get_pteptr() 232 pa = (pte_val(*pte) & PAGE_MASK) | (addr & ~PAGE_MASK); in iopa()
|
D | highmem.c | 61 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic()
|
/linux-4.4.14/arch/m68k/mm/ |
D | cache.c | 49 return (mmusr & PAGE_MASK) | (vaddr & ~PAGE_MASK); in virt_to_phys_slow() 67 return (*descaddr & PAGE_MASK) | (vaddr & ~PAGE_MASK); in virt_to_phys_slow() 87 address &= PAGE_MASK; in flush_icache_range()
|
D | memory.c | 40 unsigned long page = ptable & PAGE_MASK; in init_pointer_table() 104 unsigned long page = (unsigned long)ptable & PAGE_MASK; in free_pointer_table() 216 pushcl040(paddr & PAGE_MASK); in cache_clear() 222 paddr &= PAGE_MASK; in cache_clear() 270 paddr &= PAGE_MASK; in cache_push()
|
D | mcfmmu.c | 55 bootmem_end = (next_pgtable + size + PAGE_SIZE) & PAGE_MASK; in paging_init() 136 mmutr = (mmuar & PAGE_MASK) | (asid << MMUTR_IDN) | MMUTR_V; in cf_tlb_miss() 141 mmu_write(MMUDR, (pte_val(*pte) & PAGE_MASK) | in cf_tlb_miss()
|
D | motorola.c | 95 if (((unsigned long)last_pgtable & ~PAGE_MASK) == 0) { in kernel_ptr_table() 127 printk ("\npa=%#lx va=%#lx ", physaddr & PAGE_MASK, in map_node()
|
D | sun3mmu.c | 61 bootmem_end = (next_pgtable + size + PAGE_SIZE) & PAGE_MASK; in paging_init()
|
/linux-4.4.14/arch/x86/kernel/ |
D | livepatch.c | 80 numpages = ((loc & PAGE_MASK) == ((loc + size) & PAGE_MASK)) ? 1 : 2; in klp_write_module_reloc() 83 set_memory_rw(loc & PAGE_MASK, numpages); in klp_write_module_reloc() 88 set_memory_ro(loc & PAGE_MASK, numpages); in klp_write_module_reloc()
|
D | sys_x86_64.c | 92 if (off & ~PAGE_MASK) in SYSCALL_DEFINE6() 204 if (!(addr & ~PAGE_MASK)) in arch_get_unmapped_area_topdown()
|
/linux-4.4.14/arch/score/mm/ |
D | tlb-score.c | 95 start &= PAGE_MASK; in local_flush_tlb_range() 97 end &= PAGE_MASK; in local_flush_tlb_range() 133 start &= PAGE_MASK; in local_flush_tlb_kernel_range() 135 end &= PAGE_MASK; in local_flush_tlb_kernel_range() 167 page &= PAGE_MASK; in local_flush_tlb_page() 197 page &= (PAGE_MASK << 1); in local_flush_tlb_one() 227 address &= PAGE_MASK; in __update_tlb()
|
/linux-4.4.14/arch/m68k/kernel/ |
D | sys_m68k.c | 62 _paddr = (_mmusr & MMU_R_040) ? (_mmusr & PAGE_MASK) : 0; \ 103 paddr += addr & ~(PAGE_MASK | 15); in cache_flush_040() 106 unsigned long tmp = PAGE_SIZE - (addr & ~PAGE_MASK); in cache_flush_040() 124 i = (PAGE_SIZE - (paddr & ~PAGE_MASK)) >> 4; in cache_flush_040() 179 len += (addr & ~PAGE_MASK) + (PAGE_SIZE - 1); in cache_flush_040() 267 unsigned long tmp = PAGE_SIZE - (addr & ~PAGE_MASK); in cache_flush_060() 285 i = (PAGE_SIZE - (paddr & ~PAGE_MASK)) >> 4; in cache_flush_060() 318 addr &= PAGE_MASK; in cache_flush_060() 340 len += (addr & ~PAGE_MASK) + (PAGE_SIZE - 1); in cache_flush_060() 341 addr &= PAGE_MASK; /* Workaround for bug in some in cache_flush_060()
|
D | machine_kexec.c | 55 ((relocate_kernel_t) reboot_code_buffer)(image->head & PAGE_MASK, in machine_kexec()
|
/linux-4.4.14/arch/sparc/mm/ |
D | iommu.c | 158 start &= PAGE_MASK; in iommu_flush_iotlb() 214 off = (unsigned long)vaddr & ~PAGE_MASK; in iommu_get_scsi_one() 216 page = virt_to_page((unsigned long)vaddr & PAGE_MASK); in iommu_get_scsi_one() 229 unsigned long page = ((unsigned long) vaddr) & PAGE_MASK; in iommu_get_scsi_one_pflush() 304 off = vaddr & ~PAGE_MASK; in iommu_release_scsi_one() 306 iommu_release_one(dev, vaddr & PAGE_MASK, npages); in iommu_release_scsi_one() 317 iommu_release_one(dev, sg->dma_address & PAGE_MASK, n); in iommu_release_scsi_sgl() 333 BUG_ON((va & ~PAGE_MASK) != 0); in iommu_map_dma_area() 334 BUG_ON((addr & ~PAGE_MASK) != 0); in iommu_map_dma_area() 335 BUG_ON((len & ~PAGE_MASK) != 0); in iommu_map_dma_area() [all …]
|
D | fault_32.c | 442 if (((sp + 0x38) & PAGE_MASK) != (sp & PAGE_MASK)) in window_overflow_fault() 451 if (((sp + 0x38) & PAGE_MASK) != (sp & PAGE_MASK)) in window_underflow_fault() 463 if (((sp + 0x38) & PAGE_MASK) != (sp & PAGE_MASK)) in window_ret_fault()
|
D | io-unit.c | 100 npages = ((vaddr & ~PAGE_MASK) + size + (PAGE_SIZE-1)) >> PAGE_SHIFT; in iounit_get_area() 132 iopte = MKIOPTE(__pa(vaddr & PAGE_MASK)); in iounit_get_area() 133 vaddr = IOUNIT_DMA_BASE + (scan << PAGE_SHIFT) + (vaddr & ~PAGE_MASK); in iounit_get_area() 175 len = ((vaddr & ~PAGE_MASK) + len + (PAGE_SIZE-1)) >> PAGE_SHIFT; in iounit_release_scsi_one() 192 len = ((sg->dma_address & ~PAGE_MASK) + sg->length + (PAGE_SIZE-1)) >> PAGE_SHIFT; in iounit_release_scsi_sgl()
|
D | hugetlbpage.c | 40 info.align_mask = PAGE_MASK & ~HPAGE_MASK; in hugetlb_get_unmapped_area_bottomup() 44 if ((addr & ~PAGE_MASK) && task_size > VA_EXCLUDE_END) { in hugetlb_get_unmapped_area_bottomup() 71 info.align_mask = PAGE_MASK & ~HPAGE_MASK; in hugetlb_get_unmapped_area_topdown() 81 if (addr & ~PAGE_MASK) { in hugetlb_get_unmapped_area_topdown()
|
D | gup.c | 172 start &= PAGE_MASK; in __get_user_pages_fast() 202 start &= PAGE_MASK; in get_user_pages_fast()
|
D | viking.S | 101 andncc %g3, PAGE_MASK, %g0 184 and %o1, PAGE_MASK, %o1 271 and %o1, PAGE_MASK, %o1
|
D | highmem.c | 89 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic()
|
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/subdev/devinit/ |
D | fbmem.h | 63 u8 __iomem *p = io_mapping_map_atomic_wc(fb, off & PAGE_MASK); in fbmem_peek() 64 u32 val = ioread32(p + (off & ~PAGE_MASK)); in fbmem_peek() 72 u8 __iomem *p = io_mapping_map_atomic_wc(fb, off & PAGE_MASK); in fbmem_poke() 73 iowrite32(val, p + (off & ~PAGE_MASK)); in fbmem_poke()
|
/linux-4.4.14/drivers/infiniband/hw/mlx5/ |
D | doorbell.c | 55 if (page->user_virt == (virt & PAGE_MASK)) in mlx5_ib_db_map_user() 64 page->user_virt = (virt & PAGE_MASK); in mlx5_ib_db_map_user() 66 page->umem = ib_umem_get(&context->ibucontext, virt & PAGE_MASK, in mlx5_ib_db_map_user() 77 db->dma = sg_dma_address(page->umem->sg_head.sgl) + (virt & ~PAGE_MASK); in mlx5_ib_db_map_user()
|
/linux-4.4.14/drivers/infiniband/hw/mlx4/ |
D | doorbell.c | 53 if (page->user_virt == (virt & PAGE_MASK)) in mlx4_ib_db_map_user() 62 page->user_virt = (virt & PAGE_MASK); in mlx4_ib_db_map_user() 64 page->umem = ib_umem_get(&context->ibucontext, virt & PAGE_MASK, in mlx4_ib_db_map_user() 75 db->dma = sg_dma_address(page->umem->sg_head.sgl) + (virt & ~PAGE_MASK); in mlx4_ib_db_map_user()
|
/linux-4.4.14/arch/mips/mm/ |
D | tlb-r8k.c | 87 start &= PAGE_MASK; in local_flush_tlb_range() 89 end &= PAGE_MASK; in local_flush_tlb_range() 127 start &= PAGE_MASK; in local_flush_tlb_kernel_range() 129 end &= PAGE_MASK; in local_flush_tlb_kernel_range() 159 page &= PAGE_MASK; in local_flush_tlb_page() 200 address &= PAGE_MASK; in __update_tlb()
|
D | tlb-r3k.c | 100 start &= PAGE_MASK; in local_flush_tlb_range() 102 end &= PAGE_MASK; in local_flush_tlb_range() 136 start &= PAGE_MASK; in local_flush_tlb_kernel_range() 138 end &= PAGE_MASK; in local_flush_tlb_kernel_range() 172 page &= PAGE_MASK; in local_flush_tlb_page() 212 address &= PAGE_MASK; in __update_tlb()
|
D | ioremap.c | 156 offset = phys_addr & ~PAGE_MASK; in __ioremap() 157 phys_addr &= PAGE_MASK; in __ioremap() 184 p = remove_vm_area((void *) (PAGE_MASK & (unsigned long __force) addr)); in __iounmap()
|
D | tlb-r4k.c | 178 start &= (PAGE_MASK << 1); in local_flush_tlb_kernel_range() 180 end &= (PAGE_MASK << 1); in local_flush_tlb_kernel_range() 220 page &= (PAGE_MASK << 1); in local_flush_tlb_page() 259 page &= (PAGE_MASK << 1); in local_flush_tlb_one() 304 address &= (PAGE_MASK << 1); in __update_tlb()
|
D | init.c | 82 zero_page_mask = ((PAGE_SIZE << order) - 1) & PAGE_MASK; in setup_zero_pages() 109 write_c0_entryhi(vaddr & (PAGE_MASK << 1)); in __kmap_pgprot() 178 pages_do_alias((unsigned long)vto, vaddr & PAGE_MASK)) in copy_user_highpage() 191 void *vto = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); in copy_to_user_page() 209 void *vfrom = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); in copy_from_user_page()
|
D | dma-default.c | 275 offset &= ~PAGE_MASK; in __dma_sync() 297 dma_addr & ~PAGE_MASK, size, direction); in mips_dma_unmap_page() 353 dma_handle & ~PAGE_MASK, size, direction); in mips_dma_sync_single_for_cpu() 362 dma_handle & ~PAGE_MASK, size, direction); in mips_dma_sync_single_for_device()
|
D | mmap.c | 100 info.align_mask = do_color_align ? (PAGE_MASK & shm_align_mask) : 0; in arch_get_unmapped_area_common() 109 if (!(addr & ~PAGE_MASK)) in arch_get_unmapped_area_common()
|
/linux-4.4.14/arch/ia64/mm/ |
D | ioremap.c | 71 page_base = phys_addr & PAGE_MASK; in ioremap() 79 offset = phys_addr & ~PAGE_MASK; in ioremap() 80 phys_addr &= PAGE_MASK; in ioremap() 123 vunmap((void *) ((unsigned long) addr & PAGE_MASK)); in iounmap()
|
D | init.c | 119 vma->vm_start = current->thread.rbs_bot & PAGE_MASK; in ia64_init_addr_space() 195 end = end & PAGE_MASK; in free_initrd_mem() 443 start_page = (unsigned long) map_start & PAGE_MASK; in create_mem_map_page_table() 533 && ((((u64)pg & PAGE_MASK) == (((u64)(pg + 1) - 1) & PAGE_MASK)) in ia64_pfn_valid()
|
/linux-4.4.14/arch/microblaze/pci/ |
D | indirect_pci.c | 153 resource_size_t base = cfg_addr & PAGE_MASK; in setup_indirect_pci() 157 hose->cfg_addr = mbase + (cfg_addr & ~PAGE_MASK); in setup_indirect_pci() 158 if ((cfg_data & PAGE_MASK) != base) in setup_indirect_pci() 159 mbase = ioremap(cfg_data & PAGE_MASK, PAGE_SIZE); in setup_indirect_pci() 160 hose->cfg_data = mbase + (cfg_data & ~PAGE_MASK); in setup_indirect_pci()
|
/linux-4.4.14/arch/mips/lib/ |
D | r3k_dump_tlb.c | 47 if ((entryhi & PAGE_MASK) != KSEG0 && in dump_tlb() 57 entryhi & PAGE_MASK, in dump_tlb() 59 entrylo0 & PAGE_MASK, in dump_tlb()
|
D | dump_tlb.c | 136 pa = (pa << 6) & PAGE_MASK; in dump_tlb() 151 pa = (pa << 6) & PAGE_MASK; in dump_tlb()
|
/linux-4.4.14/arch/s390/pci/ |
D | pci_mmio.c | 46 if (length <= 0 || PAGE_SIZE - (mmio_addr & ~PAGE_MASK) < length) in SYSCALL_DEFINE3() 58 io_addr = (void __iomem *)((pfn << PAGE_SHIFT) | (mmio_addr & ~PAGE_MASK)); in SYSCALL_DEFINE3() 86 if (length <= 0 || PAGE_SIZE - (mmio_addr & ~PAGE_MASK) < length) in SYSCALL_DEFINE3() 98 io_addr = (void __iomem *)((pfn << PAGE_SHIFT) | (mmio_addr & ~PAGE_MASK)); in SYSCALL_DEFINE3()
|
/linux-4.4.14/arch/powerpc/sysdev/ |
D | indirect_pci.c | 167 resource_size_t base = cfg_addr & PAGE_MASK; in setup_indirect_pci() 171 hose->cfg_addr = mbase + (cfg_addr & ~PAGE_MASK); in setup_indirect_pci() 172 if ((cfg_data & PAGE_MASK) != base) in setup_indirect_pci() 173 mbase = ioremap(cfg_data & PAGE_MASK, PAGE_SIZE); in setup_indirect_pci() 174 hose->cfg_data = mbase + (cfg_data & ~PAGE_MASK); in setup_indirect_pci()
|
/linux-4.4.14/arch/cris/mm/ |
D | ioremap.c | 44 offset = phys_addr & ~PAGE_MASK; in __ioremap_prot() 45 phys_addr &= PAGE_MASK; in __ioremap_prot() 88 return vfree((void *) (PAGE_MASK & (unsigned long) addr)); in iounmap()
|
/linux-4.4.14/arch/arc/mm/ |
D | ioremap.c | 62 off = paddr & ~PAGE_MASK; in ioremap_prot() 63 paddr &= PAGE_MASK; in ioremap_prot() 89 vfree((void *)(PAGE_MASK & (unsigned long __force)addr)); in iounmap()
|
D | tlb.c | 346 start &= PAGE_MASK; in local_flush_tlb_range() 379 start &= PAGE_MASK; in local_flush_tlb_kernel_range() 408 tlb_entry_erase((page & PAGE_MASK) | hw_pid(vma->vm_mm, cpu)); in local_flush_tlb_page() 553 vaddr &= PAGE_MASK; in create_tlb() 598 unsigned long vaddr = vaddr_unaligned & PAGE_MASK; in update_mmu_cache() 599 phys_addr_t paddr = pte_val(*ptep) & PAGE_MASK; in update_mmu_cache() 903 pd0[way] &= PAGE_MASK; in do_tlb_overlap_fault()
|
D | mmap.c | 75 info.align_mask = do_align ? (PAGE_MASK & (SHMLBA - 1)) : 0; in arch_get_unmapped_area()
|
/linux-4.4.14/arch/metag/mm/ |
D | ioremap.c | 49 offset = phys_addr & ~PAGE_MASK; in __ioremap() 50 phys_addr &= PAGE_MASK; in __ioremap() 81 p = remove_vm_area((void *)(PAGE_MASK & (unsigned long __force)addr)); in __iounmap()
|
/linux-4.4.14/arch/avr32/mm/ |
D | ioremap.c | 52 offset = phys_addr & ~PAGE_MASK; in __ioremap() 53 phys_addr &= PAGE_MASK; in __ioremap() 85 p = remove_vm_area((void *)(PAGE_MASK & (unsigned long __force)addr)); in __iounmap()
|
D | tlb.c | 157 page &= PAGE_MASK; in flush_tlb_page() 196 start &= PAGE_MASK; in flush_tlb_range() 198 end &= PAGE_MASK; in flush_tlb_range() 235 start &= PAGE_MASK; in flush_tlb_kernel_range() 237 end &= PAGE_MASK; in flush_tlb_kernel_range()
|
/linux-4.4.14/arch/m32r/mm/ |
D | ioremap.c | 83 offset = phys_addr & ~PAGE_MASK; in __ioremap() 84 phys_addr &= PAGE_MASK; in __ioremap() 109 vfree((void *) (PAGE_MASK & (unsigned long) addr)); in iounmap()
|
D | fault.c | 197 addr = (address & PAGE_MASK); in do_page_fault() 261 page &= PAGE_MASK; in do_page_fault() 334 addr = (address & PAGE_MASK); in do_page_fault() 363 vaddr = (vaddr & PAGE_MASK) | get_asid(); in update_mmu_cache() 464 page &= PAGE_MASK; in local_flush_tlb_page() 494 start &= PAGE_MASK; in local_flush_tlb_range() 496 end &= PAGE_MASK; in local_flush_tlb_range()
|
/linux-4.4.14/arch/parisc/mm/ |
D | ioremap.c | 72 offset = phys_addr & ~PAGE_MASK; in __ioremap() 73 phys_addr &= PAGE_MASK; in __ioremap() 97 return vfree((void *) (PAGE_MASK & (unsigned long __force) addr)); in iounmap()
|
/linux-4.4.14/arch/arm64/mm/ |
D | ioremap.c | 36 unsigned long offset = phys_addr & ~PAGE_MASK; in __ioremap_caller() 45 phys_addr &= PAGE_MASK; in __ioremap_caller() 85 unsigned long addr = (unsigned long)io_addr & PAGE_MASK; in __iounmap()
|
D | init.c | 153 memory_limit = memparse(p, &p) & PAGE_MASK; in early_mem() 226 pgend = (unsigned long)__pa(end_pg) & PAGE_MASK; in free_memmap()
|
D | mmu.c | 254 addr = virt & PAGE_MASK; in __create_mapping() 255 length = PAGE_ALIGN(size + (virt & ~PAGE_MASK)); in __create_mapping() 283 __create_mapping(&init_mm, pgd_offset_k(virt & PAGE_MASK), phys, virt, in create_mapping() 304 return __create_mapping(&init_mm, pgd_offset_k(virt & PAGE_MASK), in create_mapping_late()
|
/linux-4.4.14/arch/openrisc/mm/ |
D | ioremap.c | 57 offset = addr & ~PAGE_MASK; in __ioremap() 58 p = addr & PAGE_MASK; in __ioremap() 107 return vfree((void *)(PAGE_MASK & (unsigned long)addr)); in iounmap()
|
/linux-4.4.14/arch/powerpc/mm/ |
D | pgtable_32.c | 203 p = addr & PAGE_MASK; in __ioremap_caller() 271 return (void __iomem *) (v + ((unsigned long)addr & ~PAGE_MASK)); in __ioremap_caller() 284 vunmap((void *) (PAGE_MASK & (unsigned long)addr)); in iounmap() 375 pgd = pgd_offset(mm, addr & PAGE_MASK); in get_pteptr() 377 pud = pud_offset(pgd, addr & PAGE_MASK); in get_pteptr() 379 pmd = pmd_offset(pud, addr & PAGE_MASK); in get_pteptr() 381 pte = pte_offset_map(pmd, addr & PAGE_MASK); in get_pteptr()
|
D | tlb_hash32.c | 45 ptephys = __pa(ptep) & PAGE_MASK; in flush_hash_entry() 102 start &= PAGE_MASK; in flush_range() 105 end = (end - 1) | ~PAGE_MASK; in flush_range()
|
D | pgtable_64.c | 178 WARN_ON(pa & ~PAGE_MASK); in __ioremap_at() 179 WARN_ON(((unsigned long)ea) & ~PAGE_MASK); in __ioremap_at() 180 WARN_ON(size & ~PAGE_MASK); in __ioremap_at() 197 WARN_ON(((unsigned long)ea) & ~PAGE_MASK); in __iounmap_at() 198 WARN_ON(size & ~PAGE_MASK); in __iounmap_at() 218 paligned = addr & PAGE_MASK; in __ioremap_caller() 244 ret += addr & ~PAGE_MASK; in __ioremap_caller() 312 PCI_FIX_ADDR(token) & PAGE_MASK); in __iounmap() 378 if (((unsigned long)pte_frag & ~PAGE_MASK) == 0) in get_from_cache()
|
D | highmem.c | 57 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic()
|
D | slice.c | 276 info.align_mask = PAGE_MASK & ((1ul << pshift) - 1); in slice_find_area_bottomup() 301 if (!(found & ~PAGE_MASK)) in slice_find_area_bottomup() 319 info.align_mask = PAGE_MASK & ((1ul << pshift) - 1); in slice_find_area_topdown() 344 if (!(found & ~PAGE_MASK)) in slice_find_area_topdown()
|
/linux-4.4.14/arch/mn10300/mm/ |
D | cache-inv-icache.c | 38 off = start & ~PAGE_MASK; in flush_icache_page_range() 110 start_page = start & PAGE_MASK; in flush_icache_range() 111 end_page = (end - 1) & PAGE_MASK; in flush_icache_range()
|
D | cache-flush-icache.c | 60 off = start & ~PAGE_MASK; in flush_icache_page_range() 135 start_page = start & PAGE_MASK; in flush_icache_range() 136 end_page = (end - 1) & PAGE_MASK; in flush_icache_range()
|
D | mmu-context.c | 34 addr &= PAGE_MASK; in update_mmu_cache()
|
/linux-4.4.14/arch/arc/include/asm/ |
D | pgtable.h | 113 #define _PAGE_CHG_MASK (PAGE_MASK | _PAGE_ACCESSED | _PAGE_DIRTY) 139 #define PTE_BITS_NON_RWX_IN_PD1 (0xff00000000 | PAGE_MASK | _PAGE_CACHEABLE) 141 #define PTE_BITS_NON_RWX_IN_PD1 (PAGE_MASK | _PAGE_CACHEABLE) 254 #define pmd_page(pmd) virt_to_page(pmd_val(pmd) & PAGE_MASK) 257 #define pmd_page_vaddr(pmd) (pmd_val(pmd) & PAGE_MASK) 270 #define pmd_bad(x) ((pmd_val(x) & ~PAGE_MASK))
|
/linux-4.4.14/arch/powerpc/include/asm/ |
D | pte-44x.h | 89 #define _PMD_PRESENT_MASK (PAGE_MASK) 90 #define _PMD_BAD (~PAGE_MASK)
|
D | pte-fsl-booke.h | 34 #define _PMD_PRESENT_MASK (PAGE_MASK) 35 #define _PMD_BAD (~PAGE_MASK)
|
D | pte-hash32.h | 39 #define _PMD_PRESENT_MASK (PAGE_MASK) 40 #define _PMD_BAD (~PAGE_MASK)
|
D | pte-book3e.h | 82 #define _PMD_PRESENT_MASK (PAGE_MASK) 83 #define _PMD_BAD (~PAGE_MASK)
|
D | pgtable-ppc32.h | 73 #define IOREMAP_TOP ((KVIRT_TOP - CONFIG_CONSISTENT_SIZE) & PAGE_MASK) 238 unsigned long ptephys = __pa(ptep) & PAGE_MASK; in __ptep_test_and_clear_young() 288 ((unsigned long) __va(pmd_val(pmd) & PAGE_MASK)) 293 ((unsigned long) (pmd_val(pmd) & PAGE_MASK))
|
/linux-4.4.14/include/linux/ |
D | pagemap.h | 99 #define PAGE_CACHE_MASK PAGE_MASK 568 if (((unsigned long)uaddr & PAGE_MASK) != in fault_in_pages_writeable() 569 ((unsigned long)end & PAGE_MASK)) in fault_in_pages_writeable() 587 if (((unsigned long)uaddr & PAGE_MASK) != in fault_in_pages_readable() 588 ((unsigned long)end & PAGE_MASK)) { in fault_in_pages_readable() 622 if (((unsigned long)uaddr & PAGE_MASK) == in fault_in_multipages_writeable() 623 ((unsigned long)end & PAGE_MASK)) in fault_in_multipages_writeable() 647 if (((unsigned long)uaddr & PAGE_MASK) == in fault_in_multipages_readable() 648 ((unsigned long)end & PAGE_MASK)) { in fault_in_multipages_readable()
|
D | pfn.h | 8 #define PFN_ALIGN(x) (((unsigned long)(x) + (PAGE_SIZE - 1)) & PAGE_MASK)
|
/linux-4.4.14/arch/x86/mm/kmemcheck/ |
D | kmemcheck.c | 355 unsigned long page = addr & PAGE_MASK; in kmemcheck_read() 357 unsigned long next_page = next_addr & PAGE_MASK; in kmemcheck_read() 391 unsigned long page = addr & PAGE_MASK; in kmemcheck_write() 393 unsigned long next_page = next_addr & PAGE_MASK; in kmemcheck_write() 425 page = src_addr & PAGE_MASK; in kmemcheck_copy() 427 next_page = next_addr & PAGE_MASK; in kmemcheck_copy() 469 page = dst_addr & PAGE_MASK; in kmemcheck_copy() 471 next_page = next_addr & PAGE_MASK; in kmemcheck_copy()
|
D | shadow.c | 41 unsigned long page = addr & PAGE_MASK; in mark_shadow() 42 unsigned long last_page = last_addr & PAGE_MASK; in mark_shadow()
|
/linux-4.4.14/arch/mn10300/include/asm/ |
D | page.h | 19 #define PAGE_MASK (~(PAGE_SIZE - 1)) macro 23 #define PAGE_MASK +(~(PAGE_SIZE - 1)) macro 43 #define PTE_MASK PAGE_MASK
|
D | tlbflush.h | 64 addr &= PAGE_MASK; in local_flush_tlb_page()
|
/linux-4.4.14/arch/microblaze/kernel/ |
D | sys_microblaze.c | 40 if (pgoff & ~PAGE_MASK) in SYSCALL_DEFINE6() 50 if (pgoff & (~PAGE_MASK >> 12)) in SYSCALL_DEFINE6()
|
D | hw_exception_handler.S | 637 andi r5, r4, PAGE_MASK /* Extract L2 (pte) base address */ 662 andi r4, r4, PAGE_MASK | TLB_EX | TLB_WR | \ 738 andi r5, r4, PAGE_MASK /* Extract L2 (pte) base address */ 762 andi r4, r4, PAGE_MASK | TLB_EX | TLB_WR | \ 809 andi r5, r4, PAGE_MASK /* Extract L2 (pte) base address */ 833 andi r4, r4, PAGE_MASK | TLB_EX | TLB_WR | \ 887 andi r3, r3, PAGE_MASK
|
/linux-4.4.14/arch/arm/xen/ |
D | mm.c | 90 dma_cache_maint(handle & PAGE_MASK, handle & ~PAGE_MASK, size, dir, DMA_UNMAP); in __xen_dma_page_dev_to_cpu() 96 dma_cache_maint(handle & PAGE_MASK, handle & ~PAGE_MASK, size, dir, DMA_MAP); in __xen_dma_page_cpu_to_dev()
|
/linux-4.4.14/drivers/media/pci/ivtv/ |
D | ivtv-udma.c | 28 dma_page->uaddr = first & PAGE_MASK; in ivtv_udma_get_page_info() 29 dma_page->offset = first & ~PAGE_MASK; in ivtv_udma_get_page_info() 30 dma_page->tail = 1 + ((first+size-1) & ~PAGE_MASK); in ivtv_udma_get_page_info() 31 dma_page->first = (first & PAGE_MASK) >> PAGE_SHIFT; in ivtv_udma_get_page_info() 32 dma_page->last = ((first+size-1) & PAGE_MASK) >> PAGE_SHIFT; in ivtv_udma_get_page_info()
|
/linux-4.4.14/arch/parisc/kernel/ |
D | sys_parisc.c | 128 info.align_mask = last_mmap ? (PAGE_MASK & (SHM_COLOUR - 1)) : 0; in arch_get_unmapped_area() 133 if (do_color_align && !last_mmap && !(addr & ~PAGE_MASK)) in arch_get_unmapped_area() 188 info.align_mask = last_mmap ? (PAGE_MASK & (SHM_COLOUR - 1)) : 0; in arch_get_unmapped_area_topdown() 191 if (!(addr & ~PAGE_MASK)) in arch_get_unmapped_area_topdown() 204 if (do_color_align && !last_mmap && !(addr & ~PAGE_MASK)) in arch_get_unmapped_area_topdown() 278 if (!(offset & ~PAGE_MASK)) { in sys_mmap()
|
/linux-4.4.14/arch/hexagon/mm/ |
D | ioremap.c | 28 unsigned long offset = phys_addr & ~PAGE_MASK; in ioremap_nocache() 56 vunmap((void *) ((unsigned long) addr & PAGE_MASK)); in __iounmap()
|
/linux-4.4.14/arch/um/include/asm/ |
D | pgtable-3level.h | 57 #define pud_bad(x) ((pud_val(x) & (~PAGE_MASK & ~_PAGE_USER)) != _KERNPG_TABLE) 89 #define pud_page(pud) phys_to_page(pud_val(pud) & PAGE_MASK) 90 #define pud_page_vaddr(pud) ((unsigned long) __va(pud_val(pud) & PAGE_MASK))
|
D | pgtable.h | 57 #define _PAGE_CHG_MASK (PAGE_MASK | _PAGE_ACCESSED | _PAGE_DIRTY) 100 #define pmd_bad(x) ((pmd_val(x) & (~PAGE_MASK & ~_PAGE_USER)) != _KERNPG_TABLE) 111 #define pmd_page(pmd) phys_to_page(pmd_val(pmd) & PAGE_MASK) 317 #define pmd_page_vaddr(pmd) ((unsigned long) __va(pmd_val(pmd) & PAGE_MASK)) 321 ((unsigned long) __va(pmd_val(pmd) & PAGE_MASK))
|
/linux-4.4.14/arch/s390/mm/ |
D | mmap.c | 78 gap &= PAGE_MASK; in mmap_base() 159 if (addr & ~PAGE_MASK) { in arch_get_unmapped_area_topdown() 190 if (!(area & ~PAGE_MASK)) in s390_get_unmapped_area() 212 if (!(area & ~PAGE_MASK)) in s390_get_unmapped_area_topdown()
|
D | init.c | 73 zero_page_mask = ((PAGE_SIZE << order) - 1) & PAGE_MASK; in setup_zero_pages() 92 init_mm.context.asce = (__pa(init_mm.pgd) & PAGE_MASK) | asce_bits; in paging_init()
|
D | gup.c | 185 start &= PAGE_MASK; in __get_user_pages_fast() 236 start &= PAGE_MASK; in get_user_pages_fast()
|
/linux-4.4.14/drivers/infiniband/hw/cxgb3/ |
D | iwch_mem.c | 148 if (i != 0 && buffer_list[i].addr & ~PAGE_MASK) in build_phys_page_list() 151 (buffer_list[i].size & ~PAGE_MASK)) in build_phys_page_list() 157 mask |= buffer_list[i].addr & PAGE_MASK; in build_phys_page_list() 162 PAGE_SIZE - 1) & PAGE_MASK; in build_phys_page_list()
|
/linux-4.4.14/arch/h8300/mm/ |
D | init.c | 71 unsigned long end_mem = memory_end & PAGE_MASK; in paging_init() 107 high_memory = (void *) (memory_end & PAGE_MASK); in mem_init()
|
/linux-4.4.14/arch/sparc/kernel/ |
D | sys_sparc_64.c | 129 info.align_mask = do_color_align ? (PAGE_MASK & (SHMLBA - 1)) : 0; in arch_get_unmapped_area() 133 if ((addr & ~PAGE_MASK) && task_size > VA_EXCLUDE_END) { in arch_get_unmapped_area() 192 info.align_mask = do_color_align ? (PAGE_MASK & (SHMLBA - 1)) : 0; in arch_get_unmapped_area_topdown() 202 if (addr & ~PAGE_MASK) { in arch_get_unmapped_area_topdown() 238 if (!(addr & ~PAGE_MASK)) { in get_fb_unmapped_area() 249 } while ((addr & ~PAGE_MASK) && align_goal > PAGE_SIZE); in get_fb_unmapped_area() 254 if (addr & ~PAGE_MASK) in get_fb_unmapped_area() 456 if (off & ~PAGE_MASK) in SYSCALL_DEFINE6()
|
D | ioport.c | 138 unsigned long vaddr = (unsigned long) virtual & PAGE_MASK; in iounmap() 218 unsigned long offset = ((unsigned long) pa) & (~PAGE_MASK); in _sparc_ioremap() 221 (offset + sz + PAGE_SIZE-1) & PAGE_MASK, in _sparc_ioremap() 229 pa &= PAGE_MASK; in _sparc_ioremap()
|
/linux-4.4.14/arch/xtensa/mm/ |
D | tlb.c | 108 start &= PAGE_MASK; in local_flush_tlb_range() 156 start &= PAGE_MASK; in local_flush_tlb_kernel_range() 221 unsigned vpn = (r0 & PAGE_MASK) | (e << PAGE_SHIFT); in check_tlb_entry() 238 if ((pte ^ r1) & PAGE_MASK) { in check_tlb_entry()
|
D | init.c | 96 (start < end) != (PAGE_ALIGN(start) < (end & PAGE_MASK))) { in add_sysmem_bank() 103 end &= PAGE_MASK; in add_sysmem_bank() 177 start = start & PAGE_MASK; in mem_reserve()
|
/linux-4.4.14/arch/m68k/sun3/ |
D | config.c | 117 memory_start = ((memory_start + (PAGE_SIZE-1)) & PAGE_MASK); in sun3_bootmem_alloc() 118 memory_end = memory_end & PAGE_MASK; in sun3_bootmem_alloc() 128 availmem = (availmem + (PAGE_SIZE-1)) & PAGE_MASK; in sun3_bootmem_alloc()
|
D | dvma.c | 41 return (vaddr + (kaddr & ~PAGE_MASK)); in dvma_page()
|
/linux-4.4.14/arch/xtensa/include/asm/ |
D | page.h | 34 #define PAGE_MASK (~(PAGE_SIZE-1)) macro 76 # define DCACHE_ALIAS_MASK (PAGE_MASK & (DCACHE_WAY_SIZE - 1)) 87 # define ICACHE_ALIAS_MASK (PAGE_MASK & (ICACHE_WAY_SIZE - 1))
|
D | pgtable.h | 159 #define _PAGE_CHG_MASK (PAGE_MASK | _PAGE_ACCESSED | _PAGE_DIRTY) 181 # define _PAGE_CHG_MASK (PAGE_MASK | _PAGE_ACCESSED | _PAGE_DIRTY) 238 #define pmd_page_vaddr(pmd) ((unsigned long)(pmd_val(pmd) & PAGE_MASK)) 256 #define pmd_present(pmd) (pmd_val(pmd) & PAGE_MASK) 257 #define pmd_bad(pmd) (pmd_val(pmd) & ~PAGE_MASK)
|
/linux-4.4.14/arch/ia64/include/asm/ |
D | tlbflush.h | 78 flush_tlb_range(vma, (addr & PAGE_MASK), (addr & PAGE_MASK) + PAGE_SIZE); in flush_tlb_page()
|
D | cacheflush.h | 43 unsigned long _addr = (unsigned long) page_address(page) + ((user_addr) & ~PAGE_MASK); \
|
/linux-4.4.14/arch/x86/include/asm/ |
D | page_types.h | 10 #define PAGE_MASK (~(PAGE_SIZE-1)) macro 24 #define PHYSICAL_PAGE_MASK (((signed long)PAGE_MASK) & __PHYSICAL_MASK)
|
D | pmem.h | 136 if (size == PAGE_SIZE && ((unsigned long)vaddr & ~PAGE_MASK) == 0) in arch_clear_pmem()
|
/linux-4.4.14/arch/x86/mm/ |
D | ioremap.c | 129 offset = phys_addr & ~PAGE_MASK; in __ioremap_caller() 352 (PAGE_MASK & (unsigned long __force)addr); in iounmap() 398 unsigned long start = phys & PAGE_MASK; in xlate_dev_mem_ptr() 399 unsigned long offset = phys & ~PAGE_MASK; in xlate_dev_mem_ptr() 419 iounmap((void __iomem *)((unsigned long)addr & PAGE_MASK)); in unxlate_dev_mem_ptr()
|
D | hugetlbpage.c | 86 info.align_mask = PAGE_MASK & ~huge_page_mask(h); in hugetlb_get_unmapped_area_bottomup() 103 info.align_mask = PAGE_MASK & ~huge_page_mask(h); in hugetlb_get_unmapped_area_topdown() 113 if (addr & ~PAGE_MASK) { in hugetlb_get_unmapped_area_topdown()
|
D | init_64.c | 223 if (!ptr || ((unsigned long)ptr & ~PAGE_MASK)) { in spp_getpage() 412 next = (addr & PAGE_MASK) + PAGE_SIZE; in phys_pte_init() 415 !e820_any_mapped(addr & PAGE_MASK, next, E820_RAM) && in phys_pte_init() 416 !e820_any_mapped(addr & PAGE_MASK, next, E820_RESERVED_KERN)) in phys_pte_init() 438 last_map_addr = (addr & PAGE_MASK) + PAGE_SIZE; in phys_pte_init() 801 next = (addr + PAGE_SIZE) & PAGE_MASK; in remove_pte_table() 813 phys_addr = pte_val(*pte) + (addr & PAGE_MASK); in remove_pte_table() 1326 next = (addr + PAGE_SIZE) & PAGE_MASK; in register_page_bootmem_memmap() 1333 next = (addr + PAGE_SIZE) & PAGE_MASK; in register_page_bootmem_memmap() 1339 next = (addr + PAGE_SIZE) & PAGE_MASK; in register_page_bootmem_memmap()
|
/linux-4.4.14/arch/openrisc/include/asm/ |
D | pgtable.h | 156 #define _PAGE_CHG_MASK (PAGE_MASK | _PAGE_ACCESSED | _PAGE_DIRTY) 218 ((unsigned long)(address)>>(PAGE_SHIFT-SIZEOF_PTR_LOG2)&PTR_MASK&~PAGE_MASK) 228 #define pmd_bad(x) ((pmd_val(x) & (~PAGE_MASK)) != _KERNPG_TABLE) 351 return (unsigned long)__va(pte_val(pte) & PAGE_MASK); in __pte_page() 372 #define pmd_page_kernel(pmd) ((unsigned long) __va(pmd_val(pmd) & PAGE_MASK))
|
D | fixmap.h | 57 #define __virt_to_fix(x) ((FIXADDR_TOP - ((x)&PAGE_MASK)) >> PAGE_SHIFT)
|
/linux-4.4.14/arch/nios2/mm/ |
D | ioremap.c | 152 offset = phys_addr & ~PAGE_MASK; in __ioremap() 153 phys_addr &= PAGE_MASK; in __ioremap() 182 p = remove_vm_area((void *) (PAGE_MASK & (unsigned long __force) addr)); in __iounmap()
|
/linux-4.4.14/drivers/infiniband/hw/usnic/ |
D | usnic_uiom.c | 129 npages = PAGE_ALIGN(size + (addr & ~PAGE_MASK)) >> PAGE_SHIFT; in usnic_uiom_get_pages() 143 cur_base = addr & PAGE_MASK; in usnic_uiom_get_pages() 227 vpn_start = (uiomr->va & PAGE_MASK) >> PAGE_SHIFT; in __usnic_uiom_reg_release() 258 long int va = uiomr->va & PAGE_MASK; in usnic_uiom_map_sorted_intervals() 355 va_base = addr & PAGE_MASK; in usnic_uiom_reg_get() 356 offset = addr & ~PAGE_MASK; in usnic_uiom_reg_get() 358 vpn_start = (addr & PAGE_MASK) >> PAGE_SHIFT; in usnic_uiom_reg_get()
|
/linux-4.4.14/arch/mips/kernel/ |
D | machine_kexec.c | 76 (unsigned long) phys_to_virt(image->head & PAGE_MASK); in machine_kexec() 92 phys_to_virt(entry & PAGE_MASK) : ptr + 1) { in machine_kexec()
|
D | syscall.c | 66 if (offset & ~PAGE_MASK) in SYSCALL_DEFINE6() 79 if (pgoff & (~PAGE_MASK >> 12)) in SYSCALL_DEFINE6()
|
/linux-4.4.14/arch/blackfin/mm/ |
D | init.c | 47 unsigned long end_mem = memory_end & PAGE_MASK; in paging_init() 95 high_memory = (void *)(memory_end & PAGE_MASK); in mem_init()
|
/linux-4.4.14/arch/cris/include/asm/ |
D | pgtable.h | 90 ((unsigned long)(address)>>(PAGE_SHIFT-SIZEOF_PTR_LOG2)&PTR_MASK&~PAGE_MASK) 103 #define pmd_bad(x) ((pmd_val(x) & (~PAGE_MASK & ~_PAGE_KERNEL)) != _PAGE_TABLE) 210 return (unsigned long)__va(pte_val(pte) & PAGE_MASK); in __pte_page() 229 #define pmd_page_vaddr(pmd) ((unsigned long) __va(pmd_val(pmd) & PAGE_MASK))
|
/linux-4.4.14/sound/core/ |
D | sgbuf.c | 48 if (!(sgbuf->table[i].addr & ~PAGE_MASK)) in snd_free_sgbuf_pages() 51 tmpb.addr = sgbuf->table[i].addr & PAGE_MASK; in snd_free_sgbuf_pages() 52 tmpb.bytes = (sgbuf->table[i].addr & ~PAGE_MASK) << PAGE_SHIFT; in snd_free_sgbuf_pages()
|
/linux-4.4.14/arch/arm/include/asm/ |
D | tlbflush.h | 424 uaddr = (uaddr & PAGE_MASK) | ASID(vma->vm_mm); in __local_flush_tlb_page() 445 uaddr = (uaddr & PAGE_MASK) | ASID(vma->vm_mm); in local_flush_tlb_page() 462 uaddr = (uaddr & PAGE_MASK) | ASID(vma->vm_mm); in __flush_tlb_page() 469 tlb_op(TLB_V7_UIS_PAGE, "c8, c3, 3", uaddr & PAGE_MASK); in __flush_tlb_page() 498 kaddr &= PAGE_MASK; in local_flush_tlb_kernel_page() 516 kaddr &= PAGE_MASK; in __flush_tlb_kernel_page()
|
D | dma-mapping.h | 141 unsigned int offset = paddr & ~PAGE_MASK; in phys_to_dma() 147 unsigned int offset = dev_addr & ~PAGE_MASK; in dma_to_phys()
|
D | cacheflush.h | 233 __cpuc_flush_user_range(start & PAGE_MASK, PAGE_ALIGN(end), in vivt_flush_cache_range() 243 unsigned long addr = user_addr & PAGE_MASK; in vivt_flush_cache_page()
|
/linux-4.4.14/kernel/ |
D | kexec_core.c | 170 if ((mstart & ~PAGE_MASK) || (mend & ~PAGE_MASK)) in sanity_check_segment_list() 499 destination &= PAGE_MASK; in kimage_set_destination() 510 page &= PAGE_MASK; in kimage_add_page() 537 phys_to_virt((entry & PAGE_MASK)) : ptr + 1) 596 destination = entry & PAGE_MASK; in kimage_dst_used() 679 old_addr = *old & PAGE_MASK; in kimage_alloc_page() 682 *old = addr | (*old & ~PAGE_MASK); in kimage_alloc_page() 744 ptr += maddr & ~PAGE_MASK; in kimage_load_normal_segment() 746 PAGE_SIZE - (maddr & ~PAGE_MASK)); in kimage_load_normal_segment() 803 ptr += maddr & ~PAGE_MASK; in kimage_load_crash_segment() [all …]
|
/linux-4.4.14/arch/arm/mm/ |
D | mmap.c | 100 info.align_mask = do_align ? (PAGE_MASK & (SHMLBA - 1)) : 0; in arch_get_unmapped_area() 151 info.align_mask = do_align ? (PAGE_MASK & (SHMLBA - 1)) : 0; in arch_get_unmapped_area_topdown() 161 if (addr & ~PAGE_MASK) { in arch_get_unmapped_area_topdown()
|
D | dma-mapping.c | 112 handle & ~PAGE_MASK, size, dir); in arm_dma_unmap_page() 1293 size = PAGE_ALIGN((iova & ~PAGE_MASK) + size); in __iommu_remove_mapping() 1294 iova &= PAGE_MASK; in __iommu_remove_mapping() 1564 if (s->offset || (size & ~PAGE_MASK) || size + s->length > max) { in __iommu_map_sg() 1784 dma_addr_t iova = handle & PAGE_MASK; in arm_coherent_iommu_unmap_page() 1785 int offset = handle & ~PAGE_MASK; in arm_coherent_iommu_unmap_page() 1809 dma_addr_t iova = handle & PAGE_MASK; in arm_iommu_unmap_page() 1811 int offset = handle & ~PAGE_MASK; in arm_iommu_unmap_page() 1828 dma_addr_t iova = handle & PAGE_MASK; in arm_iommu_sync_single_for_cpu() 1830 unsigned int offset = handle & ~PAGE_MASK; in arm_iommu_sync_single_for_cpu() [all …]
|
/linux-4.4.14/arch/frv/mm/ |
D | elf-fdpic.c | 89 if (!(addr & ~PAGE_MASK)) in arch_get_unmapped_area() 97 if (!(addr & ~PAGE_MASK)) in arch_get_unmapped_area()
|
D | cache-page.c | 59 start = (start & ~PAGE_MASK) | (unsigned long) vaddr; in flush_icache_user_range()
|
/linux-4.4.14/arch/m32r/include/asm/ |
D | page.h | 9 #define PAGE_MASK (~(PAGE_SIZE-1)) macro 30 #define PTE_MASK PAGE_MASK
|
D | pgtable.h | 183 #define pmd_bad(x) ((pmd_val(x) & ~PAGE_MASK) != _KERNPG_TABLE) 296 pmd_val(*pmdp) = (((unsigned long) ptep) & PAGE_MASK); in pmd_set() 300 ((unsigned long) __va(pmd_val(pmd) & PAGE_MASK))
|
/linux-4.4.14/arch/avr32/include/asm/ |
D | page.h | 16 #define PAGE_MASK (~(PAGE_SIZE-1)) macro 17 #define PTE_MASK PAGE_MASK
|
/linux-4.4.14/arch/x86/kvm/ |
D | x86.h | 90 vcpu->arch.mmio_gva = gva & PAGE_MASK; in vcpu_cache_mmio_info() 109 if (gva != MMIO_GVA_ANY && vcpu->arch.mmio_gva != (gva & PAGE_MASK)) in vcpu_clear_mmio_info() 118 vcpu->arch.mmio_gva == (gva & PAGE_MASK)) in vcpu_match_mmio_gva()
|
/linux-4.4.14/mm/ |
D | msync.c | 46 len = (len + ~PAGE_MASK) & PAGE_MASK; in SYSCALL_DEFINE3()
|
D | early_ioremap.c | 130 phys_addr &= PAGE_MASK; in __early_ioremap() 241 p = early_memremap(src & PAGE_MASK, clen + slop); in copy_from_early_mem()
|
D | slob.c | 147 slob_t *base = (slob_t *)((unsigned long)s & PAGE_MASK); in set_slob() 172 slob_t *base = (slob_t *)((unsigned long)s & PAGE_MASK); in slob_next() 187 return !((unsigned long)slob_next(s) & ~PAGE_MASK); in slob_last() 374 SLOB_UNITS(PAGE_SIZE)) & PAGE_MASK)); in slob_free()
|
D | madvise.c | 477 if (start & ~PAGE_MASK) in SYSCALL_DEFINE3() 479 len = (len_in + ~PAGE_MASK) & PAGE_MASK; in SYSCALL_DEFINE3()
|
/linux-4.4.14/fs/ |
D | binfmt_aout.c | 277 error = vm_brk(text_addr & PAGE_MASK, map_size); in load_aout_binary() 278 if (error != (text_addr & PAGE_MASK)) in load_aout_binary() 292 if ((fd_offset & ~PAGE_MASK) != 0 && printk_ratelimit()) in load_aout_binary() 299 if (!bprm->file->f_op->mmap||((fd_offset & ~PAGE_MASK) != 0)) { in load_aout_binary() 374 if ((N_TXTOFF(ex) & ~PAGE_MASK) != 0) { in load_aout_library()
|
/linux-4.4.14/arch/x86/platform/ce4100/ |
D | ce4100.c | 106 up->mapbase & PAGE_MASK); in ce4100_serial_fixup() 109 up->membase += up->mapbase & ~PAGE_MASK; in ce4100_serial_fixup()
|
/linux-4.4.14/drivers/target/tcm_fc/ |
D | tfc_io.c | 157 from += mem_off & ~PAGE_MASK; in ft_queue_data_in() 159 (mem_off & ~PAGE_MASK))); in ft_queue_data_in() 317 to += mem_off & ~PAGE_MASK; in ft_recv_write_data() 319 (mem_off & ~PAGE_MASK))); in ft_recv_write_data()
|
/linux-4.4.14/tools/virtio/linux/ |
D | kernel.h | 21 #define PAGE_MASK (~(PAGE_SIZE-1)) macro 36 #define virt_to_page(p) ((struct page *)((unsigned long)p & PAGE_MASK))
|
/linux-4.4.14/arch/arm/kernel/ |
D | patch.c | 41 return (void *) (__fix_to_virt(fixmap) + (uintaddr & ~PAGE_MASK)); in patch_map() 78 twopage = (uintaddr & ~PAGE_MASK) == PAGE_SIZE - 2; in __patch_text_real()
|
/linux-4.4.14/arch/cris/arch-v10/mm/ |
D | tlb.c | 111 addr &= PAGE_MASK; /* perhaps not necessary */ in flush_tlb_page() 123 (tlb_hi & PAGE_MASK) == addr) { in flush_tlb_page()
|
/linux-4.4.14/arch/sh/kernel/ |
D | machine_kexec.c | 89 phys_to_virt(entry & PAGE_MASK) : ptr + 1) { in machine_kexec() 134 phys_to_virt(*ptr & PAGE_MASK) : ptr + 1) { in machine_kexec()
|
D | sys_sh.c | 35 if (off & ~PAGE_MASK) in old_mmap()
|
/linux-4.4.14/arch/cris/arch-v32/mm/ |
D | tlb.c | 135 addr &= PAGE_MASK; in __flush_tlb_page() 151 ((tlb_hi & PAGE_MASK) == addr)) { in __flush_tlb_page()
|
D | mmu.S | 128 and.w PAGE_MASK, $acr ; Remove PMD flags 188 move.w PAGE_MASK, $acr 192 and.w PAGE_MASK, $r10 ; Get VPN (virtual address)
|
/linux-4.4.14/arch/powerpc/platforms/powermac/ |
D | udbg_scc.c | 116 sccc = ioremap(addr & PAGE_MASK, PAGE_SIZE) ; in udbg_scc_init() 117 sccc += addr & ~PAGE_MASK; in udbg_scc_init()
|
/linux-4.4.14/arch/m68k/sun3x/ |
D | dvma.c | 89 kaddr &= PAGE_MASK; in dvma_map_cpu() 90 vaddr &= PAGE_MASK; in dvma_map_cpu()
|
/linux-4.4.14/drivers/media/v4l2-core/ |
D | videobuf-dma-sg.c | 172 first = (data & PAGE_MASK) >> PAGE_SHIFT; in videobuf_dma_init_user_locked() 173 last = ((data+size-1) & PAGE_MASK) >> PAGE_SHIFT; in videobuf_dma_init_user_locked() 174 dma->offset = data & ~PAGE_MASK; in videobuf_dma_init_user_locked() 185 data & PAGE_MASK, dma->nr_pages, in videobuf_dma_init_user_locked() 303 & PAGE_MASK; in videobuf_dma_map() 304 dma->sglist[0].offset = dma->bus_addr & ~PAGE_MASK; in videobuf_dma_map()
|
/linux-4.4.14/arch/x86/ia32/ |
D | ia32_aout.c | 322 error = vm_brk(text_addr & PAGE_MASK, map_size); in load_aout_binary() 324 if (error != (text_addr & PAGE_MASK)) in load_aout_binary() 341 if ((fd_offset & ~PAGE_MASK) != 0 && in load_aout_binary() 351 if (!bprm->file->f_op->mmap || (fd_offset & ~PAGE_MASK) != 0) { in load_aout_binary() 426 if ((N_TXTOFF(ex) & ~PAGE_MASK) != 0) { in load_aout_library()
|
/linux-4.4.14/arch/alpha/kernel/ |
D | core_irongate.c | 354 if (addr & ~PAGE_MASK) { in irongate_ioremap() 401 vaddr = (unsigned long)area->addr + (addr & ~PAGE_MASK); in irongate_ioremap() 417 return vfree((void *)(PAGE_MASK & addr)); in irongate_iounmap()
|
D | pci_iommu.c | 312 paddr &= PAGE_MASK; in pci_map_single_1() 317 ret += (unsigned long)cpu_addr & ~PAGE_MASK; in pci_map_single_1() 530 } else if (((next_paddr | addr) & ~PAGE_MASK) == 0 && virt_ok) { in sg_classify() 594 paddr &= ~PAGE_MASK; in sg_fill() 633 paddr &= PAGE_MASK; in sg_fill()
|
/linux-4.4.14/arch/tile/mm/ |
D | hugetlbpage.c | 175 info.align_mask = PAGE_MASK & ~huge_page_mask(h); in hugetlb_get_unmapped_area_bottomup() 192 info.align_mask = PAGE_MASK & ~huge_page_mask(h); in hugetlb_get_unmapped_area_topdown() 202 if (addr & ~PAGE_MASK) { in hugetlb_get_unmapped_area_topdown()
|
/linux-4.4.14/arch/arc/include/uapi/asm/ |
D | page.h | 33 #define PAGE_MASK (~(PAGE_SIZE-1)) macro
|
/linux-4.4.14/drivers/video/fbdev/ |
D | igafb.c | 543 par->mmap_map[0].poff = par->frame_buffer_phys & PAGE_MASK; in igafb_init() 544 par->mmap_map[0].size = info->fix.smem_len & PAGE_MASK; in igafb_init() 549 par->mmap_map[1].voff = par->frame_buffer_phys & PAGE_MASK; in igafb_init() 550 par->mmap_map[1].poff = info->fix.smem_start & PAGE_MASK; in igafb_init()
|
/linux-4.4.14/drivers/gpu/drm/qxl/ |
D | qxl_ioctl.c | 88 reloc_page = qxl_bo_kmap_atomic_page(qdev, info->dst_bo, info->dst_offset & PAGE_MASK); in apply_reloc() 89 *(uint64_t *)(reloc_page + (info->dst_offset & ~PAGE_MASK)) = qxl_bo_physical_address(qdev, in apply_reloc() 104 reloc_page = qxl_bo_kmap_atomic_page(qdev, info->dst_bo, info->dst_offset & PAGE_MASK); in apply_surf_reloc() 105 *(uint32_t *)(reloc_page + (info->dst_offset & ~PAGE_MASK)) = id; in apply_surf_reloc()
|
/linux-4.4.14/arch/mn10300/kernel/ |
D | sys_mn10300.c | 30 if (offset & ~PAGE_MASK) in old_mmap()
|
/linux-4.4.14/arch/um/kernel/ |
D | syscall.c | 22 if (offset & ~PAGE_MASK) in old_mmap()
|
D | um_arch.c | 295 uml_physmem = (unsigned long) __binary_start & PAGE_MASK; in linux_main() 303 iomem_size = (iomem_size + PAGE_SIZE - 1) & PAGE_MASK; in linux_main()
|
D | tlb.c | 218 ret = add_mmap(addr, pte_val(*pte) & PAGE_MASK, in update_pte_range() 381 pte_val(*pte) & PAGE_MASK, in flush_tlb_kernel_range_common() 404 address &= PAGE_MASK; in flush_tlb_page() 437 fd = phys_mapping(pte_val(*pte) & PAGE_MASK, &offset); in flush_tlb_page()
|
/linux-4.4.14/arch/mips/include/asm/ |
D | page.h | 36 #define PAGE_MASK (~((1 << PAGE_SHIFT) - 1)) macro 104 if (pages_do_alias((unsigned long) addr, vaddr & PAGE_MASK)) in clear_user_page()
|
D | pgtable-64.h | 184 if (unlikely(pmd_val(pmd) & ~PAGE_MASK)) in pmd_bad() 211 return pud_val(pud) & ~PAGE_MASK; in pud_bad()
|
/linux-4.4.14/arch/avr32/mach-at32ap/ |
D | pm.c | 46 page_addr = pm_sram_start & PAGE_MASK; in avr32_pm_map_sram() 59 tlbehi |= vaddr & PAGE_MASK; in avr32_pm_map_sram()
|
/linux-4.4.14/arch/sh/include/asm/ |
D | pgtable_64.h | 69 #define pmd_bad(pmd_entry) ((pmd_val(pmd_entry) & (~PAGE_MASK & ~_PAGE_USER)) != _KERNPG_TABLE) 72 ((unsigned long) __va(pmd_val(pmd_entry) & PAGE_MASK)) 84 ((pte_t *) ((pmd_val(*(dir))) & PAGE_MASK) + pte_index((addr)))
|
D | page.h | 24 #define PAGE_MASK (~(PAGE_SIZE-1)) macro 25 #define PTE_MASK PAGE_MASK
|
D | pgtable-3level.h | 48 #define pud_bad(x) (pud_val(x) & ~PAGE_MASK)
|
/linux-4.4.14/arch/score/include/asm/ |
D | pgtable-bits.h | 22 (PAGE_MASK | _PAGE_ACCESSED | _PAGE_MODIFIED | _PAGE_CACHE)
|
D | fixmap.h | 63 ((FIXADDR_TOP - ((x) & PAGE_MASK)) >> PAGE_SHIFT)
|
/linux-4.4.14/drivers/tty/serial/ |
D | earlycon.c | 51 set_fixmap_io(FIX_EARLYCON_MEM_BASE, paddr & PAGE_MASK); in earlycon_map() 53 base += paddr & ~PAGE_MASK; in earlycon_map()
|
/linux-4.4.14/arch/unicore32/mm/ |
D | ioremap.c | 188 unsigned long offset = phys_addr & ~PAGE_MASK; in __uc32_ioremap_caller() 237 void *addr = (void *)(PAGE_MASK & (unsigned long)io_addr); in __uc32_iounmap()
|
D | cache-ucv2.S | 107 sub r9, r9, #1 @ PAGE_MASK 161 sub r9, r9, #1 @ PAGE_MASK
|
/linux-4.4.14/arch/tile/kernel/ |
D | machine_kexec.c | 159 if (((unsigned long)desc & PAGE_MASK) != (unsigned long)pg) { in kexec_bn2cl() 179 phys_to_virt((entry & PAGE_MASK)) : ptr + 1) { in kexec_find_and_set_command_line()
|
/linux-4.4.14/arch/microblaze/include/asm/ |
D | pgtable.h | 217 #define _PMD_PRESENT PAGE_MASK 235 #define _PAGE_CHG_MASK (PAGE_MASK | _PAGE_ACCESSED | _PAGE_DIRTY) 326 ((unsigned long) __va(pgd_val(pgd) & PAGE_MASK)) 470 #define pmd_page_kernel(pmd) ((unsigned long) (pmd_val(pmd) & PAGE_MASK))
|
/linux-4.4.14/arch/score/kernel/ |
D | sys_score.c | 47 if (unlikely(offset & ~PAGE_MASK)) in sys_mmap()
|
/linux-4.4.14/arch/powerpc/boot/ |
D | page.h | 22 #define PAGE_MASK (~(PAGE_SIZE-1)) macro
|
/linux-4.4.14/drivers/mtd/onenand/ |
D | omap2.c | 320 if (((size_t)buf & PAGE_MASK) != in omap3_onenand_read_bufferram() 321 ((size_t)(buf + count - 1) & PAGE_MASK)) in omap3_onenand_read_bufferram() 326 buf = page_address(p1) + ((size_t)buf & ~PAGE_MASK); in omap3_onenand_read_bufferram() 397 if (((size_t)buf & PAGE_MASK) != in omap3_onenand_write_bufferram() 398 ((size_t)(buf + count - 1) & PAGE_MASK)) in omap3_onenand_write_bufferram() 403 buf = page_address(p1) + ((size_t)buf & ~PAGE_MASK); in omap3_onenand_write_bufferram()
|
/linux-4.4.14/arch/c6x/mm/ |
D | init.c | 61 high_memory = (void *)(memory_end & PAGE_MASK); in mem_init()
|
/linux-4.4.14/arch/alpha/mm/ |
D | init.c | 181 (((unsigned long)kernel_end + ~PAGE_MASK) & PAGE_MASK); in callback_init()
|
/linux-4.4.14/arch/arm/mach-orion5x/ |
D | common.c | 372 (!t->u.mem.size || t->u.mem.size & ~PAGE_MASK || in tag_fixup_mem32() 373 t->u.mem.start & ~PAGE_MASK)) { in tag_fixup_mem32()
|
/linux-4.4.14/arch/frv/include/asm/ |
D | mem-layout.h | 32 #define PAGE_MASK (~(PAGE_SIZE-1)) macro
|
/linux-4.4.14/arch/arm64/include/asm/ |
D | page.h | 35 #define PAGE_MASK (~(PAGE_SIZE-1)) macro
|
/linux-4.4.14/arch/um/include/shared/ |
D | common-offsets.h | 6 DEFINE(UM_KERN_PAGE_MASK, PAGE_MASK);
|
D | kern_util.h | 20 ((((unsigned long) addr) + PAGE_SIZE - 1) & PAGE_MASK)
|
/linux-4.4.14/arch/m68k/include/asm/ |
D | page.h | 15 #define PAGE_MASK (~(PAGE_SIZE-1)) macro
|
D | mmu_context.h | 143 mmu_write(MMUTR, (mmuar & PAGE_MASK) | (asid << MMUTR_IDN) | in load_ksp_mmu() 147 mmu_write(MMUDR, (pte_val(*pte) & PAGE_MASK) | in load_ksp_mmu()
|
/linux-4.4.14/arch/x86/include/asm/xen/ |
D | page.h | 181 unsigned offset = phys.paddr & ~PAGE_MASK; in phys_to_machine() 187 unsigned offset = machine.maddr & ~PAGE_MASK; in machine_to_phys()
|
/linux-4.4.14/arch/unicore32/kernel/ |
D | setup.c | 148 size -= start & ~PAGE_MASK; in uc32_add_memory() 151 bank->size = size & PAGE_MASK; in uc32_add_memory()
|
/linux-4.4.14/drivers/lguest/ |
D | lg.h | 148 #define pgd_flags(x) (pgd_val(x) & ~PAGE_MASK) 150 #define pmd_flags(x) (pmd_val(x) & ~PAGE_MASK)
|
/linux-4.4.14/net/rds/ |
D | iw_rdma.c | 286 if (dma_addr & PAGE_MASK) { 289 dma_addr &= ~PAGE_MASK; 291 if (end_addr & PAGE_MASK) { 294 end_addr = (end_addr + PAGE_MASK) & ~PAGE_MASK;
|
/linux-4.4.14/arch/sparc/include/asm/ |
D | ross.h | 178 page &= PAGE_MASK; in hyper_flush_cache_page()
|
D | iommu_32.h | 118 sbus_writel(ba & PAGE_MASK, ®s->pageflush); in iommu_invalidate_page()
|
/linux-4.4.14/arch/c6x/include/asm/ |
D | pgtable.h | 36 #define pmd_bad(x) (pmd_val(x) & ~PAGE_MASK)
|
/linux-4.4.14/arch/um/kernel/skas/ |
D | uaccess.c | 73 (addr & ~PAGE_MASK); in do_op_one_page() 109 while (addr < ((addr + remain) & PAGE_MASK)) { in buffer_op()
|
/linux-4.4.14/arch/arm/lib/ |
D | uaccess_with_memcpy.c | 118 tocopy = (~(unsigned long)to & ~PAGE_MASK) + 1; in __copy_to_user_memcpy() 184 tocopy = (~(unsigned long)addr & ~PAGE_MASK) + 1; in __clear_user_memset()
|
/linux-4.4.14/drivers/net/ethernet/ibm/ehea/ |
D | ehea_phyp.h | 67 epas->kernel.addr = ioremap((paddr_kernel & PAGE_MASK), PAGE_SIZE) + in hcp_epas_ctor() 68 (paddr_kernel & ~PAGE_MASK); in hcp_epas_ctor() 75 iounmap((void __iomem *)((u64)epas->kernel.addr & PAGE_MASK)); in hcp_epas_dtor()
|
/linux-4.4.14/drivers/vfio/ |
D | vfio_iommu_spapr_tce.c | 110 if ((vaddr & ~PAGE_MASK) || (size & ~PAGE_MASK)) in tce_iommu_unregister_pages() 127 if ((vaddr & ~PAGE_MASK) || (size & ~PAGE_MASK) || in tce_iommu_register_pages() 449 if (get_user_pages_fast(tce & PAGE_MASK, 1, in tce_iommu_use_page() 469 unsigned long offset = tce & IOMMU_PAGE_MASK(tbl) & ~PAGE_MASK; in tce_iommu_build() 531 hpa |= tce & IOMMU_PAGE_MASK(tbl) & ~PAGE_MASK; in tce_iommu_build_v2()
|
/linux-4.4.14/drivers/vfio/platform/ |
D | vfio_platform_common.c | 104 if (!(vdev->regions[i].addr & ~PAGE_MASK) && in vfio_platform_regions_init() 105 !(vdev->regions[i].size & ~PAGE_MASK)) in vfio_platform_regions_init() 510 if (vma->vm_start & ~PAGE_MASK) in vfio_platform_mmap() 512 if (vma->vm_end & ~PAGE_MASK) in vfio_platform_mmap()
|
/linux-4.4.14/arch/cris/include/arch-v10/arch/ |
D | mmu.h | 72 #define _PAGE_CHG_MASK (PAGE_MASK | _PAGE_ACCESSED | _PAGE_MODIFIED)
|
/linux-4.4.14/include/asm-generic/ |
D | fixmap.h | 21 #define __virt_to_fix(x) ((FIXADDR_TOP - ((x)&PAGE_MASK)) >> PAGE_SHIFT)
|