Home
last modified time | relevance | path

Searched refs:PAGE_MASK (Results 1 – 200 of 662) sorted by relevance

1234

/linux-4.4.14/arch/unicore32/include/asm/
Dtlbflush.h94 : : "r" (uaddr & PAGE_MASK) : "cc"); in local_flush_tlb_page()
97 : : "r" (uaddr & PAGE_MASK) : "cc"); in local_flush_tlb_page()
101 : : "r" (uaddr & PAGE_MASK) : "cc"); in local_flush_tlb_page()
111 : : "r" (kaddr & PAGE_MASK) : "cc"); in local_flush_tlb_kernel_page()
114 : : "r" (kaddr & PAGE_MASK) : "cc"); in local_flush_tlb_kernel_page()
118 : : "r" (kaddr & PAGE_MASK) : "cc"); in local_flush_tlb_kernel_page()
Dpage.h18 #define PAGE_MASK (~(PAGE_SIZE-1)) macro
Dcacheflush.h153 __cpuc_coherent_user_range((start) & PAGE_MASK, PAGE_ALIGN(end))
/linux-4.4.14/arch/sh/mm/
Dtlbflush_64.c72 page &= PAGE_MASK; in local_flush_tlb_page()
94 start &= PAGE_MASK; in local_flush_tlb_range()
95 end &= PAGE_MASK; in local_flush_tlb_range()
105 pteh_epn = pteh & PAGE_MASK; in local_flush_tlb_range()
106 pteh_low = pteh & ~PAGE_MASK; in local_flush_tlb_range()
118 pteh_epn = pteh & PAGE_MASK; in local_flush_tlb_range()
119 pteh_low = pteh & ~PAGE_MASK; in local_flush_tlb_range()
Dtlbflush_32.c25 page &= PAGE_MASK; in local_flush_tlb_page()
60 start &= PAGE_MASK; in local_flush_tlb_range()
62 end &= PAGE_MASK; in local_flush_tlb_range()
93 start &= PAGE_MASK; in local_flush_tlb_kernel_range()
95 end &= PAGE_MASK; in local_flush_tlb_kernel_range()
Dioremap.c67 offset = phys_addr & ~PAGE_MASK; in __ioremap_caller()
68 phys_addr &= PAGE_MASK; in __ioremap_caller()
129 p = remove_vm_area((void *)(vaddr & PAGE_MASK)); in __iounmap()
Dmmap.c74 info.align_mask = do_colour_align ? (PAGE_MASK & shm_align_mask) : 0; in arch_get_unmapped_area()
124 info.align_mask = do_colour_align ? (PAGE_MASK & shm_align_mask) : 0; in arch_get_unmapped_area_topdown()
134 if (addr & ~PAGE_MASK) { in arch_get_unmapped_area_topdown()
Dtlb-sh5.c126 pteh &= PAGE_MASK; in sh64_setup_tlb_slot()
129 ptel &= PAGE_MASK; in sh64_setup_tlb_slot()
163 paddr &= ~PAGE_MASK; in tlb_wire_entry()
Dioremap_fixed.c58 offset = phys_addr & ~PAGE_MASK; in ioremap_fixed()
59 phys_addr &= PAGE_MASK; in ioremap_fixed()
Dcache.c64 void *vto = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); in copy_to_user_page()
83 void *vfrom = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); in copy_from_user_page()
111 if (pages_do_alias((unsigned long)vto, vaddr & PAGE_MASK) || in copy_user_highpage()
127 if (pages_do_alias((unsigned long)kaddr, vaddr & PAGE_MASK)) in clear_user_highpage()
Dcache-sh5.c173 aligned_start = start & PAGE_MASK; in sh64_icache_inv_user_page_range()
174 after_last_page_start = PAGE_SIZE + ((end - 1) & PAGE_MASK); in sh64_icache_inv_user_page_range()
413 paddr = pte_val(entry) & PAGE_MASK; in sh64_dcache_purge_user_pages()
476 start &= PAGE_MASK; /* should already be so */ in sh64_dcache_purge_user_range()
Dgup.c173 start &= PAGE_MASK; in __get_user_pages_fast()
226 start &= PAGE_MASK; in get_user_pages_fast()
Dkmap.c57 unsigned long vaddr = (unsigned long)kvaddr & PAGE_MASK; in kunmap_coherent()
Dtlbex_32.c71 local_flush_tlb_one(get_asid(), address & PAGE_MASK); in handle_tlbmiss()
/linux-4.4.14/arch/microblaze/mm/
Dpgtable.c63 p = addr & PAGE_MASK; in __ioremap()
120 return (void __iomem *) (v + ((unsigned long)addr & ~PAGE_MASK)); in __ioremap()
133 vfree((void *) (PAGE_MASK & (unsigned long) addr)); in iounmap()
198 pgd = pgd_offset(mm, addr & PAGE_MASK); in get_pteptr()
200 pmd = pmd_offset(pgd, addr & PAGE_MASK); in get_pteptr()
202 pte = pte_offset_kernel(pmd, addr & PAGE_MASK); in get_pteptr()
232 pa = (pte_val(*pte) & PAGE_MASK) | (addr & ~PAGE_MASK); in iopa()
Dhighmem.c61 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic()
/linux-4.4.14/arch/m68k/mm/
Dcache.c49 return (mmusr & PAGE_MASK) | (vaddr & ~PAGE_MASK); in virt_to_phys_slow()
67 return (*descaddr & PAGE_MASK) | (vaddr & ~PAGE_MASK); in virt_to_phys_slow()
87 address &= PAGE_MASK; in flush_icache_range()
Dmemory.c40 unsigned long page = ptable & PAGE_MASK; in init_pointer_table()
104 unsigned long page = (unsigned long)ptable & PAGE_MASK; in free_pointer_table()
216 pushcl040(paddr & PAGE_MASK); in cache_clear()
222 paddr &= PAGE_MASK; in cache_clear()
270 paddr &= PAGE_MASK; in cache_push()
Dmcfmmu.c55 bootmem_end = (next_pgtable + size + PAGE_SIZE) & PAGE_MASK; in paging_init()
136 mmutr = (mmuar & PAGE_MASK) | (asid << MMUTR_IDN) | MMUTR_V; in cf_tlb_miss()
141 mmu_write(MMUDR, (pte_val(*pte) & PAGE_MASK) | in cf_tlb_miss()
Dmotorola.c95 if (((unsigned long)last_pgtable & ~PAGE_MASK) == 0) { in kernel_ptr_table()
127 printk ("\npa=%#lx va=%#lx ", physaddr & PAGE_MASK, in map_node()
Dsun3mmu.c61 bootmem_end = (next_pgtable + size + PAGE_SIZE) & PAGE_MASK; in paging_init()
/linux-4.4.14/arch/x86/kernel/
Dlivepatch.c80 numpages = ((loc & PAGE_MASK) == ((loc + size) & PAGE_MASK)) ? 1 : 2; in klp_write_module_reloc()
83 set_memory_rw(loc & PAGE_MASK, numpages); in klp_write_module_reloc()
88 set_memory_ro(loc & PAGE_MASK, numpages); in klp_write_module_reloc()
Dsys_x86_64.c92 if (off & ~PAGE_MASK) in SYSCALL_DEFINE6()
204 if (!(addr & ~PAGE_MASK)) in arch_get_unmapped_area_topdown()
/linux-4.4.14/arch/score/mm/
Dtlb-score.c95 start &= PAGE_MASK; in local_flush_tlb_range()
97 end &= PAGE_MASK; in local_flush_tlb_range()
133 start &= PAGE_MASK; in local_flush_tlb_kernel_range()
135 end &= PAGE_MASK; in local_flush_tlb_kernel_range()
167 page &= PAGE_MASK; in local_flush_tlb_page()
197 page &= (PAGE_MASK << 1); in local_flush_tlb_one()
227 address &= PAGE_MASK; in __update_tlb()
/linux-4.4.14/arch/m68k/kernel/
Dsys_m68k.c62 _paddr = (_mmusr & MMU_R_040) ? (_mmusr & PAGE_MASK) : 0; \
103 paddr += addr & ~(PAGE_MASK | 15); in cache_flush_040()
106 unsigned long tmp = PAGE_SIZE - (addr & ~PAGE_MASK); in cache_flush_040()
124 i = (PAGE_SIZE - (paddr & ~PAGE_MASK)) >> 4; in cache_flush_040()
179 len += (addr & ~PAGE_MASK) + (PAGE_SIZE - 1); in cache_flush_040()
267 unsigned long tmp = PAGE_SIZE - (addr & ~PAGE_MASK); in cache_flush_060()
285 i = (PAGE_SIZE - (paddr & ~PAGE_MASK)) >> 4; in cache_flush_060()
318 addr &= PAGE_MASK; in cache_flush_060()
340 len += (addr & ~PAGE_MASK) + (PAGE_SIZE - 1); in cache_flush_060()
341 addr &= PAGE_MASK; /* Workaround for bug in some in cache_flush_060()
Dmachine_kexec.c55 ((relocate_kernel_t) reboot_code_buffer)(image->head & PAGE_MASK, in machine_kexec()
/linux-4.4.14/arch/sparc/mm/
Diommu.c158 start &= PAGE_MASK; in iommu_flush_iotlb()
214 off = (unsigned long)vaddr & ~PAGE_MASK; in iommu_get_scsi_one()
216 page = virt_to_page((unsigned long)vaddr & PAGE_MASK); in iommu_get_scsi_one()
229 unsigned long page = ((unsigned long) vaddr) & PAGE_MASK; in iommu_get_scsi_one_pflush()
304 off = vaddr & ~PAGE_MASK; in iommu_release_scsi_one()
306 iommu_release_one(dev, vaddr & PAGE_MASK, npages); in iommu_release_scsi_one()
317 iommu_release_one(dev, sg->dma_address & PAGE_MASK, n); in iommu_release_scsi_sgl()
333 BUG_ON((va & ~PAGE_MASK) != 0); in iommu_map_dma_area()
334 BUG_ON((addr & ~PAGE_MASK) != 0); in iommu_map_dma_area()
335 BUG_ON((len & ~PAGE_MASK) != 0); in iommu_map_dma_area()
[all …]
Dfault_32.c442 if (((sp + 0x38) & PAGE_MASK) != (sp & PAGE_MASK)) in window_overflow_fault()
451 if (((sp + 0x38) & PAGE_MASK) != (sp & PAGE_MASK)) in window_underflow_fault()
463 if (((sp + 0x38) & PAGE_MASK) != (sp & PAGE_MASK)) in window_ret_fault()
Dio-unit.c100 npages = ((vaddr & ~PAGE_MASK) + size + (PAGE_SIZE-1)) >> PAGE_SHIFT; in iounit_get_area()
132 iopte = MKIOPTE(__pa(vaddr & PAGE_MASK)); in iounit_get_area()
133 vaddr = IOUNIT_DMA_BASE + (scan << PAGE_SHIFT) + (vaddr & ~PAGE_MASK); in iounit_get_area()
175 len = ((vaddr & ~PAGE_MASK) + len + (PAGE_SIZE-1)) >> PAGE_SHIFT; in iounit_release_scsi_one()
192 len = ((sg->dma_address & ~PAGE_MASK) + sg->length + (PAGE_SIZE-1)) >> PAGE_SHIFT; in iounit_release_scsi_sgl()
Dhugetlbpage.c40 info.align_mask = PAGE_MASK & ~HPAGE_MASK; in hugetlb_get_unmapped_area_bottomup()
44 if ((addr & ~PAGE_MASK) && task_size > VA_EXCLUDE_END) { in hugetlb_get_unmapped_area_bottomup()
71 info.align_mask = PAGE_MASK & ~HPAGE_MASK; in hugetlb_get_unmapped_area_topdown()
81 if (addr & ~PAGE_MASK) { in hugetlb_get_unmapped_area_topdown()
Dgup.c172 start &= PAGE_MASK; in __get_user_pages_fast()
202 start &= PAGE_MASK; in get_user_pages_fast()
Dviking.S101 andncc %g3, PAGE_MASK, %g0
184 and %o1, PAGE_MASK, %o1
271 and %o1, PAGE_MASK, %o1
Dhighmem.c89 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic()
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/subdev/devinit/
Dfbmem.h63 u8 __iomem *p = io_mapping_map_atomic_wc(fb, off & PAGE_MASK); in fbmem_peek()
64 u32 val = ioread32(p + (off & ~PAGE_MASK)); in fbmem_peek()
72 u8 __iomem *p = io_mapping_map_atomic_wc(fb, off & PAGE_MASK); in fbmem_poke()
73 iowrite32(val, p + (off & ~PAGE_MASK)); in fbmem_poke()
/linux-4.4.14/drivers/infiniband/hw/mlx5/
Ddoorbell.c55 if (page->user_virt == (virt & PAGE_MASK)) in mlx5_ib_db_map_user()
64 page->user_virt = (virt & PAGE_MASK); in mlx5_ib_db_map_user()
66 page->umem = ib_umem_get(&context->ibucontext, virt & PAGE_MASK, in mlx5_ib_db_map_user()
77 db->dma = sg_dma_address(page->umem->sg_head.sgl) + (virt & ~PAGE_MASK); in mlx5_ib_db_map_user()
/linux-4.4.14/drivers/infiniband/hw/mlx4/
Ddoorbell.c53 if (page->user_virt == (virt & PAGE_MASK)) in mlx4_ib_db_map_user()
62 page->user_virt = (virt & PAGE_MASK); in mlx4_ib_db_map_user()
64 page->umem = ib_umem_get(&context->ibucontext, virt & PAGE_MASK, in mlx4_ib_db_map_user()
75 db->dma = sg_dma_address(page->umem->sg_head.sgl) + (virt & ~PAGE_MASK); in mlx4_ib_db_map_user()
/linux-4.4.14/arch/mips/mm/
Dtlb-r8k.c87 start &= PAGE_MASK; in local_flush_tlb_range()
89 end &= PAGE_MASK; in local_flush_tlb_range()
127 start &= PAGE_MASK; in local_flush_tlb_kernel_range()
129 end &= PAGE_MASK; in local_flush_tlb_kernel_range()
159 page &= PAGE_MASK; in local_flush_tlb_page()
200 address &= PAGE_MASK; in __update_tlb()
Dtlb-r3k.c100 start &= PAGE_MASK; in local_flush_tlb_range()
102 end &= PAGE_MASK; in local_flush_tlb_range()
136 start &= PAGE_MASK; in local_flush_tlb_kernel_range()
138 end &= PAGE_MASK; in local_flush_tlb_kernel_range()
172 page &= PAGE_MASK; in local_flush_tlb_page()
212 address &= PAGE_MASK; in __update_tlb()
Dioremap.c156 offset = phys_addr & ~PAGE_MASK; in __ioremap()
157 phys_addr &= PAGE_MASK; in __ioremap()
184 p = remove_vm_area((void *) (PAGE_MASK & (unsigned long __force) addr)); in __iounmap()
Dtlb-r4k.c178 start &= (PAGE_MASK << 1); in local_flush_tlb_kernel_range()
180 end &= (PAGE_MASK << 1); in local_flush_tlb_kernel_range()
220 page &= (PAGE_MASK << 1); in local_flush_tlb_page()
259 page &= (PAGE_MASK << 1); in local_flush_tlb_one()
304 address &= (PAGE_MASK << 1); in __update_tlb()
Dinit.c82 zero_page_mask = ((PAGE_SIZE << order) - 1) & PAGE_MASK; in setup_zero_pages()
109 write_c0_entryhi(vaddr & (PAGE_MASK << 1)); in __kmap_pgprot()
178 pages_do_alias((unsigned long)vto, vaddr & PAGE_MASK)) in copy_user_highpage()
191 void *vto = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); in copy_to_user_page()
209 void *vfrom = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); in copy_from_user_page()
Ddma-default.c275 offset &= ~PAGE_MASK; in __dma_sync()
297 dma_addr & ~PAGE_MASK, size, direction); in mips_dma_unmap_page()
353 dma_handle & ~PAGE_MASK, size, direction); in mips_dma_sync_single_for_cpu()
362 dma_handle & ~PAGE_MASK, size, direction); in mips_dma_sync_single_for_device()
Dmmap.c100 info.align_mask = do_color_align ? (PAGE_MASK & shm_align_mask) : 0; in arch_get_unmapped_area_common()
109 if (!(addr & ~PAGE_MASK)) in arch_get_unmapped_area_common()
/linux-4.4.14/arch/ia64/mm/
Dioremap.c71 page_base = phys_addr & PAGE_MASK; in ioremap()
79 offset = phys_addr & ~PAGE_MASK; in ioremap()
80 phys_addr &= PAGE_MASK; in ioremap()
123 vunmap((void *) ((unsigned long) addr & PAGE_MASK)); in iounmap()
Dinit.c119 vma->vm_start = current->thread.rbs_bot & PAGE_MASK; in ia64_init_addr_space()
195 end = end & PAGE_MASK; in free_initrd_mem()
443 start_page = (unsigned long) map_start & PAGE_MASK; in create_mem_map_page_table()
533 && ((((u64)pg & PAGE_MASK) == (((u64)(pg + 1) - 1) & PAGE_MASK)) in ia64_pfn_valid()
/linux-4.4.14/arch/microblaze/pci/
Dindirect_pci.c153 resource_size_t base = cfg_addr & PAGE_MASK; in setup_indirect_pci()
157 hose->cfg_addr = mbase + (cfg_addr & ~PAGE_MASK); in setup_indirect_pci()
158 if ((cfg_data & PAGE_MASK) != base) in setup_indirect_pci()
159 mbase = ioremap(cfg_data & PAGE_MASK, PAGE_SIZE); in setup_indirect_pci()
160 hose->cfg_data = mbase + (cfg_data & ~PAGE_MASK); in setup_indirect_pci()
/linux-4.4.14/arch/mips/lib/
Dr3k_dump_tlb.c47 if ((entryhi & PAGE_MASK) != KSEG0 && in dump_tlb()
57 entryhi & PAGE_MASK, in dump_tlb()
59 entrylo0 & PAGE_MASK, in dump_tlb()
Ddump_tlb.c136 pa = (pa << 6) & PAGE_MASK; in dump_tlb()
151 pa = (pa << 6) & PAGE_MASK; in dump_tlb()
/linux-4.4.14/arch/s390/pci/
Dpci_mmio.c46 if (length <= 0 || PAGE_SIZE - (mmio_addr & ~PAGE_MASK) < length) in SYSCALL_DEFINE3()
58 io_addr = (void __iomem *)((pfn << PAGE_SHIFT) | (mmio_addr & ~PAGE_MASK)); in SYSCALL_DEFINE3()
86 if (length <= 0 || PAGE_SIZE - (mmio_addr & ~PAGE_MASK) < length) in SYSCALL_DEFINE3()
98 io_addr = (void __iomem *)((pfn << PAGE_SHIFT) | (mmio_addr & ~PAGE_MASK)); in SYSCALL_DEFINE3()
/linux-4.4.14/arch/powerpc/sysdev/
Dindirect_pci.c167 resource_size_t base = cfg_addr & PAGE_MASK; in setup_indirect_pci()
171 hose->cfg_addr = mbase + (cfg_addr & ~PAGE_MASK); in setup_indirect_pci()
172 if ((cfg_data & PAGE_MASK) != base) in setup_indirect_pci()
173 mbase = ioremap(cfg_data & PAGE_MASK, PAGE_SIZE); in setup_indirect_pci()
174 hose->cfg_data = mbase + (cfg_data & ~PAGE_MASK); in setup_indirect_pci()
/linux-4.4.14/arch/cris/mm/
Dioremap.c44 offset = phys_addr & ~PAGE_MASK; in __ioremap_prot()
45 phys_addr &= PAGE_MASK; in __ioremap_prot()
88 return vfree((void *) (PAGE_MASK & (unsigned long) addr)); in iounmap()
/linux-4.4.14/arch/arc/mm/
Dioremap.c62 off = paddr & ~PAGE_MASK; in ioremap_prot()
63 paddr &= PAGE_MASK; in ioremap_prot()
89 vfree((void *)(PAGE_MASK & (unsigned long __force)addr)); in iounmap()
Dtlb.c346 start &= PAGE_MASK; in local_flush_tlb_range()
379 start &= PAGE_MASK; in local_flush_tlb_kernel_range()
408 tlb_entry_erase((page & PAGE_MASK) | hw_pid(vma->vm_mm, cpu)); in local_flush_tlb_page()
553 vaddr &= PAGE_MASK; in create_tlb()
598 unsigned long vaddr = vaddr_unaligned & PAGE_MASK; in update_mmu_cache()
599 phys_addr_t paddr = pte_val(*ptep) & PAGE_MASK; in update_mmu_cache()
903 pd0[way] &= PAGE_MASK; in do_tlb_overlap_fault()
Dmmap.c75 info.align_mask = do_align ? (PAGE_MASK & (SHMLBA - 1)) : 0; in arch_get_unmapped_area()
/linux-4.4.14/arch/metag/mm/
Dioremap.c49 offset = phys_addr & ~PAGE_MASK; in __ioremap()
50 phys_addr &= PAGE_MASK; in __ioremap()
81 p = remove_vm_area((void *)(PAGE_MASK & (unsigned long __force)addr)); in __iounmap()
/linux-4.4.14/arch/avr32/mm/
Dioremap.c52 offset = phys_addr & ~PAGE_MASK; in __ioremap()
53 phys_addr &= PAGE_MASK; in __ioremap()
85 p = remove_vm_area((void *)(PAGE_MASK & (unsigned long __force)addr)); in __iounmap()
Dtlb.c157 page &= PAGE_MASK; in flush_tlb_page()
196 start &= PAGE_MASK; in flush_tlb_range()
198 end &= PAGE_MASK; in flush_tlb_range()
235 start &= PAGE_MASK; in flush_tlb_kernel_range()
237 end &= PAGE_MASK; in flush_tlb_kernel_range()
/linux-4.4.14/arch/m32r/mm/
Dioremap.c83 offset = phys_addr & ~PAGE_MASK; in __ioremap()
84 phys_addr &= PAGE_MASK; in __ioremap()
109 vfree((void *) (PAGE_MASK & (unsigned long) addr)); in iounmap()
Dfault.c197 addr = (address & PAGE_MASK); in do_page_fault()
261 page &= PAGE_MASK; in do_page_fault()
334 addr = (address & PAGE_MASK); in do_page_fault()
363 vaddr = (vaddr & PAGE_MASK) | get_asid(); in update_mmu_cache()
464 page &= PAGE_MASK; in local_flush_tlb_page()
494 start &= PAGE_MASK; in local_flush_tlb_range()
496 end &= PAGE_MASK; in local_flush_tlb_range()
/linux-4.4.14/arch/parisc/mm/
Dioremap.c72 offset = phys_addr & ~PAGE_MASK; in __ioremap()
73 phys_addr &= PAGE_MASK; in __ioremap()
97 return vfree((void *) (PAGE_MASK & (unsigned long __force) addr)); in iounmap()
/linux-4.4.14/arch/arm64/mm/
Dioremap.c36 unsigned long offset = phys_addr & ~PAGE_MASK; in __ioremap_caller()
45 phys_addr &= PAGE_MASK; in __ioremap_caller()
85 unsigned long addr = (unsigned long)io_addr & PAGE_MASK; in __iounmap()
Dinit.c153 memory_limit = memparse(p, &p) & PAGE_MASK; in early_mem()
226 pgend = (unsigned long)__pa(end_pg) & PAGE_MASK; in free_memmap()
Dmmu.c254 addr = virt & PAGE_MASK; in __create_mapping()
255 length = PAGE_ALIGN(size + (virt & ~PAGE_MASK)); in __create_mapping()
283 __create_mapping(&init_mm, pgd_offset_k(virt & PAGE_MASK), phys, virt, in create_mapping()
304 return __create_mapping(&init_mm, pgd_offset_k(virt & PAGE_MASK), in create_mapping_late()
/linux-4.4.14/arch/openrisc/mm/
Dioremap.c57 offset = addr & ~PAGE_MASK; in __ioremap()
58 p = addr & PAGE_MASK; in __ioremap()
107 return vfree((void *)(PAGE_MASK & (unsigned long)addr)); in iounmap()
/linux-4.4.14/arch/powerpc/mm/
Dpgtable_32.c203 p = addr & PAGE_MASK; in __ioremap_caller()
271 return (void __iomem *) (v + ((unsigned long)addr & ~PAGE_MASK)); in __ioremap_caller()
284 vunmap((void *) (PAGE_MASK & (unsigned long)addr)); in iounmap()
375 pgd = pgd_offset(mm, addr & PAGE_MASK); in get_pteptr()
377 pud = pud_offset(pgd, addr & PAGE_MASK); in get_pteptr()
379 pmd = pmd_offset(pud, addr & PAGE_MASK); in get_pteptr()
381 pte = pte_offset_map(pmd, addr & PAGE_MASK); in get_pteptr()
Dtlb_hash32.c45 ptephys = __pa(ptep) & PAGE_MASK; in flush_hash_entry()
102 start &= PAGE_MASK; in flush_range()
105 end = (end - 1) | ~PAGE_MASK; in flush_range()
Dpgtable_64.c178 WARN_ON(pa & ~PAGE_MASK); in __ioremap_at()
179 WARN_ON(((unsigned long)ea) & ~PAGE_MASK); in __ioremap_at()
180 WARN_ON(size & ~PAGE_MASK); in __ioremap_at()
197 WARN_ON(((unsigned long)ea) & ~PAGE_MASK); in __iounmap_at()
198 WARN_ON(size & ~PAGE_MASK); in __iounmap_at()
218 paligned = addr & PAGE_MASK; in __ioremap_caller()
244 ret += addr & ~PAGE_MASK; in __ioremap_caller()
312 PCI_FIX_ADDR(token) & PAGE_MASK); in __iounmap()
378 if (((unsigned long)pte_frag & ~PAGE_MASK) == 0) in get_from_cache()
Dhighmem.c57 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic()
Dslice.c276 info.align_mask = PAGE_MASK & ((1ul << pshift) - 1); in slice_find_area_bottomup()
301 if (!(found & ~PAGE_MASK)) in slice_find_area_bottomup()
319 info.align_mask = PAGE_MASK & ((1ul << pshift) - 1); in slice_find_area_topdown()
344 if (!(found & ~PAGE_MASK)) in slice_find_area_topdown()
/linux-4.4.14/arch/mn10300/mm/
Dcache-inv-icache.c38 off = start & ~PAGE_MASK; in flush_icache_page_range()
110 start_page = start & PAGE_MASK; in flush_icache_range()
111 end_page = (end - 1) & PAGE_MASK; in flush_icache_range()
Dcache-flush-icache.c60 off = start & ~PAGE_MASK; in flush_icache_page_range()
135 start_page = start & PAGE_MASK; in flush_icache_range()
136 end_page = (end - 1) & PAGE_MASK; in flush_icache_range()
Dmmu-context.c34 addr &= PAGE_MASK; in update_mmu_cache()
/linux-4.4.14/arch/arc/include/asm/
Dpgtable.h113 #define _PAGE_CHG_MASK (PAGE_MASK | _PAGE_ACCESSED | _PAGE_DIRTY)
139 #define PTE_BITS_NON_RWX_IN_PD1 (0xff00000000 | PAGE_MASK | _PAGE_CACHEABLE)
141 #define PTE_BITS_NON_RWX_IN_PD1 (PAGE_MASK | _PAGE_CACHEABLE)
254 #define pmd_page(pmd) virt_to_page(pmd_val(pmd) & PAGE_MASK)
257 #define pmd_page_vaddr(pmd) (pmd_val(pmd) & PAGE_MASK)
270 #define pmd_bad(x) ((pmd_val(x) & ~PAGE_MASK))
/linux-4.4.14/arch/powerpc/include/asm/
Dpte-44x.h89 #define _PMD_PRESENT_MASK (PAGE_MASK)
90 #define _PMD_BAD (~PAGE_MASK)
Dpte-fsl-booke.h34 #define _PMD_PRESENT_MASK (PAGE_MASK)
35 #define _PMD_BAD (~PAGE_MASK)
Dpte-hash32.h39 #define _PMD_PRESENT_MASK (PAGE_MASK)
40 #define _PMD_BAD (~PAGE_MASK)
Dpte-book3e.h82 #define _PMD_PRESENT_MASK (PAGE_MASK)
83 #define _PMD_BAD (~PAGE_MASK)
Dpgtable-ppc32.h73 #define IOREMAP_TOP ((KVIRT_TOP - CONFIG_CONSISTENT_SIZE) & PAGE_MASK)
238 unsigned long ptephys = __pa(ptep) & PAGE_MASK; in __ptep_test_and_clear_young()
288 ((unsigned long) __va(pmd_val(pmd) & PAGE_MASK))
293 ((unsigned long) (pmd_val(pmd) & PAGE_MASK))
/linux-4.4.14/include/linux/
Dpagemap.h99 #define PAGE_CACHE_MASK PAGE_MASK
568 if (((unsigned long)uaddr & PAGE_MASK) != in fault_in_pages_writeable()
569 ((unsigned long)end & PAGE_MASK)) in fault_in_pages_writeable()
587 if (((unsigned long)uaddr & PAGE_MASK) != in fault_in_pages_readable()
588 ((unsigned long)end & PAGE_MASK)) { in fault_in_pages_readable()
622 if (((unsigned long)uaddr & PAGE_MASK) == in fault_in_multipages_writeable()
623 ((unsigned long)end & PAGE_MASK)) in fault_in_multipages_writeable()
647 if (((unsigned long)uaddr & PAGE_MASK) == in fault_in_multipages_readable()
648 ((unsigned long)end & PAGE_MASK)) { in fault_in_multipages_readable()
Dpfn.h8 #define PFN_ALIGN(x) (((unsigned long)(x) + (PAGE_SIZE - 1)) & PAGE_MASK)
/linux-4.4.14/arch/x86/mm/kmemcheck/
Dkmemcheck.c355 unsigned long page = addr & PAGE_MASK; in kmemcheck_read()
357 unsigned long next_page = next_addr & PAGE_MASK; in kmemcheck_read()
391 unsigned long page = addr & PAGE_MASK; in kmemcheck_write()
393 unsigned long next_page = next_addr & PAGE_MASK; in kmemcheck_write()
425 page = src_addr & PAGE_MASK; in kmemcheck_copy()
427 next_page = next_addr & PAGE_MASK; in kmemcheck_copy()
469 page = dst_addr & PAGE_MASK; in kmemcheck_copy()
471 next_page = next_addr & PAGE_MASK; in kmemcheck_copy()
Dshadow.c41 unsigned long page = addr & PAGE_MASK; in mark_shadow()
42 unsigned long last_page = last_addr & PAGE_MASK; in mark_shadow()
/linux-4.4.14/arch/mn10300/include/asm/
Dpage.h19 #define PAGE_MASK (~(PAGE_SIZE - 1)) macro
23 #define PAGE_MASK +(~(PAGE_SIZE - 1)) macro
43 #define PTE_MASK PAGE_MASK
Dtlbflush.h64 addr &= PAGE_MASK; in local_flush_tlb_page()
/linux-4.4.14/arch/microblaze/kernel/
Dsys_microblaze.c40 if (pgoff & ~PAGE_MASK) in SYSCALL_DEFINE6()
50 if (pgoff & (~PAGE_MASK >> 12)) in SYSCALL_DEFINE6()
Dhw_exception_handler.S637 andi r5, r4, PAGE_MASK /* Extract L2 (pte) base address */
662 andi r4, r4, PAGE_MASK | TLB_EX | TLB_WR | \
738 andi r5, r4, PAGE_MASK /* Extract L2 (pte) base address */
762 andi r4, r4, PAGE_MASK | TLB_EX | TLB_WR | \
809 andi r5, r4, PAGE_MASK /* Extract L2 (pte) base address */
833 andi r4, r4, PAGE_MASK | TLB_EX | TLB_WR | \
887 andi r3, r3, PAGE_MASK
/linux-4.4.14/arch/arm/xen/
Dmm.c90 dma_cache_maint(handle & PAGE_MASK, handle & ~PAGE_MASK, size, dir, DMA_UNMAP); in __xen_dma_page_dev_to_cpu()
96 dma_cache_maint(handle & PAGE_MASK, handle & ~PAGE_MASK, size, dir, DMA_MAP); in __xen_dma_page_cpu_to_dev()
/linux-4.4.14/drivers/media/pci/ivtv/
Divtv-udma.c28 dma_page->uaddr = first & PAGE_MASK; in ivtv_udma_get_page_info()
29 dma_page->offset = first & ~PAGE_MASK; in ivtv_udma_get_page_info()
30 dma_page->tail = 1 + ((first+size-1) & ~PAGE_MASK); in ivtv_udma_get_page_info()
31 dma_page->first = (first & PAGE_MASK) >> PAGE_SHIFT; in ivtv_udma_get_page_info()
32 dma_page->last = ((first+size-1) & PAGE_MASK) >> PAGE_SHIFT; in ivtv_udma_get_page_info()
/linux-4.4.14/arch/parisc/kernel/
Dsys_parisc.c128 info.align_mask = last_mmap ? (PAGE_MASK & (SHM_COLOUR - 1)) : 0; in arch_get_unmapped_area()
133 if (do_color_align && !last_mmap && !(addr & ~PAGE_MASK)) in arch_get_unmapped_area()
188 info.align_mask = last_mmap ? (PAGE_MASK & (SHM_COLOUR - 1)) : 0; in arch_get_unmapped_area_topdown()
191 if (!(addr & ~PAGE_MASK)) in arch_get_unmapped_area_topdown()
204 if (do_color_align && !last_mmap && !(addr & ~PAGE_MASK)) in arch_get_unmapped_area_topdown()
278 if (!(offset & ~PAGE_MASK)) { in sys_mmap()
/linux-4.4.14/arch/hexagon/mm/
Dioremap.c28 unsigned long offset = phys_addr & ~PAGE_MASK; in ioremap_nocache()
56 vunmap((void *) ((unsigned long) addr & PAGE_MASK)); in __iounmap()
/linux-4.4.14/arch/um/include/asm/
Dpgtable-3level.h57 #define pud_bad(x) ((pud_val(x) & (~PAGE_MASK & ~_PAGE_USER)) != _KERNPG_TABLE)
89 #define pud_page(pud) phys_to_page(pud_val(pud) & PAGE_MASK)
90 #define pud_page_vaddr(pud) ((unsigned long) __va(pud_val(pud) & PAGE_MASK))
Dpgtable.h57 #define _PAGE_CHG_MASK (PAGE_MASK | _PAGE_ACCESSED | _PAGE_DIRTY)
100 #define pmd_bad(x) ((pmd_val(x) & (~PAGE_MASK & ~_PAGE_USER)) != _KERNPG_TABLE)
111 #define pmd_page(pmd) phys_to_page(pmd_val(pmd) & PAGE_MASK)
317 #define pmd_page_vaddr(pmd) ((unsigned long) __va(pmd_val(pmd) & PAGE_MASK))
321 ((unsigned long) __va(pmd_val(pmd) & PAGE_MASK))
/linux-4.4.14/arch/s390/mm/
Dmmap.c78 gap &= PAGE_MASK; in mmap_base()
159 if (addr & ~PAGE_MASK) { in arch_get_unmapped_area_topdown()
190 if (!(area & ~PAGE_MASK)) in s390_get_unmapped_area()
212 if (!(area & ~PAGE_MASK)) in s390_get_unmapped_area_topdown()
Dinit.c73 zero_page_mask = ((PAGE_SIZE << order) - 1) & PAGE_MASK; in setup_zero_pages()
92 init_mm.context.asce = (__pa(init_mm.pgd) & PAGE_MASK) | asce_bits; in paging_init()
Dgup.c185 start &= PAGE_MASK; in __get_user_pages_fast()
236 start &= PAGE_MASK; in get_user_pages_fast()
/linux-4.4.14/drivers/infiniband/hw/cxgb3/
Diwch_mem.c148 if (i != 0 && buffer_list[i].addr & ~PAGE_MASK) in build_phys_page_list()
151 (buffer_list[i].size & ~PAGE_MASK)) in build_phys_page_list()
157 mask |= buffer_list[i].addr & PAGE_MASK; in build_phys_page_list()
162 PAGE_SIZE - 1) & PAGE_MASK; in build_phys_page_list()
/linux-4.4.14/arch/h8300/mm/
Dinit.c71 unsigned long end_mem = memory_end & PAGE_MASK; in paging_init()
107 high_memory = (void *) (memory_end & PAGE_MASK); in mem_init()
/linux-4.4.14/arch/sparc/kernel/
Dsys_sparc_64.c129 info.align_mask = do_color_align ? (PAGE_MASK & (SHMLBA - 1)) : 0; in arch_get_unmapped_area()
133 if ((addr & ~PAGE_MASK) && task_size > VA_EXCLUDE_END) { in arch_get_unmapped_area()
192 info.align_mask = do_color_align ? (PAGE_MASK & (SHMLBA - 1)) : 0; in arch_get_unmapped_area_topdown()
202 if (addr & ~PAGE_MASK) { in arch_get_unmapped_area_topdown()
238 if (!(addr & ~PAGE_MASK)) { in get_fb_unmapped_area()
249 } while ((addr & ~PAGE_MASK) && align_goal > PAGE_SIZE); in get_fb_unmapped_area()
254 if (addr & ~PAGE_MASK) in get_fb_unmapped_area()
456 if (off & ~PAGE_MASK) in SYSCALL_DEFINE6()
Dioport.c138 unsigned long vaddr = (unsigned long) virtual & PAGE_MASK; in iounmap()
218 unsigned long offset = ((unsigned long) pa) & (~PAGE_MASK); in _sparc_ioremap()
221 (offset + sz + PAGE_SIZE-1) & PAGE_MASK, in _sparc_ioremap()
229 pa &= PAGE_MASK; in _sparc_ioremap()
/linux-4.4.14/arch/xtensa/mm/
Dtlb.c108 start &= PAGE_MASK; in local_flush_tlb_range()
156 start &= PAGE_MASK; in local_flush_tlb_kernel_range()
221 unsigned vpn = (r0 & PAGE_MASK) | (e << PAGE_SHIFT); in check_tlb_entry()
238 if ((pte ^ r1) & PAGE_MASK) { in check_tlb_entry()
Dinit.c96 (start < end) != (PAGE_ALIGN(start) < (end & PAGE_MASK))) { in add_sysmem_bank()
103 end &= PAGE_MASK; in add_sysmem_bank()
177 start = start & PAGE_MASK; in mem_reserve()
/linux-4.4.14/arch/m68k/sun3/
Dconfig.c117 memory_start = ((memory_start + (PAGE_SIZE-1)) & PAGE_MASK); in sun3_bootmem_alloc()
118 memory_end = memory_end & PAGE_MASK; in sun3_bootmem_alloc()
128 availmem = (availmem + (PAGE_SIZE-1)) & PAGE_MASK; in sun3_bootmem_alloc()
Ddvma.c41 return (vaddr + (kaddr & ~PAGE_MASK)); in dvma_page()
/linux-4.4.14/arch/xtensa/include/asm/
Dpage.h34 #define PAGE_MASK (~(PAGE_SIZE-1)) macro
76 # define DCACHE_ALIAS_MASK (PAGE_MASK & (DCACHE_WAY_SIZE - 1))
87 # define ICACHE_ALIAS_MASK (PAGE_MASK & (ICACHE_WAY_SIZE - 1))
Dpgtable.h159 #define _PAGE_CHG_MASK (PAGE_MASK | _PAGE_ACCESSED | _PAGE_DIRTY)
181 # define _PAGE_CHG_MASK (PAGE_MASK | _PAGE_ACCESSED | _PAGE_DIRTY)
238 #define pmd_page_vaddr(pmd) ((unsigned long)(pmd_val(pmd) & PAGE_MASK))
256 #define pmd_present(pmd) (pmd_val(pmd) & PAGE_MASK)
257 #define pmd_bad(pmd) (pmd_val(pmd) & ~PAGE_MASK)
/linux-4.4.14/arch/ia64/include/asm/
Dtlbflush.h78 flush_tlb_range(vma, (addr & PAGE_MASK), (addr & PAGE_MASK) + PAGE_SIZE); in flush_tlb_page()
Dcacheflush.h43 unsigned long _addr = (unsigned long) page_address(page) + ((user_addr) & ~PAGE_MASK); \
/linux-4.4.14/arch/x86/include/asm/
Dpage_types.h10 #define PAGE_MASK (~(PAGE_SIZE-1)) macro
24 #define PHYSICAL_PAGE_MASK (((signed long)PAGE_MASK) & __PHYSICAL_MASK)
Dpmem.h136 if (size == PAGE_SIZE && ((unsigned long)vaddr & ~PAGE_MASK) == 0) in arch_clear_pmem()
/linux-4.4.14/arch/x86/mm/
Dioremap.c129 offset = phys_addr & ~PAGE_MASK; in __ioremap_caller()
352 (PAGE_MASK & (unsigned long __force)addr); in iounmap()
398 unsigned long start = phys & PAGE_MASK; in xlate_dev_mem_ptr()
399 unsigned long offset = phys & ~PAGE_MASK; in xlate_dev_mem_ptr()
419 iounmap((void __iomem *)((unsigned long)addr & PAGE_MASK)); in unxlate_dev_mem_ptr()
Dhugetlbpage.c86 info.align_mask = PAGE_MASK & ~huge_page_mask(h); in hugetlb_get_unmapped_area_bottomup()
103 info.align_mask = PAGE_MASK & ~huge_page_mask(h); in hugetlb_get_unmapped_area_topdown()
113 if (addr & ~PAGE_MASK) { in hugetlb_get_unmapped_area_topdown()
Dinit_64.c223 if (!ptr || ((unsigned long)ptr & ~PAGE_MASK)) { in spp_getpage()
412 next = (addr & PAGE_MASK) + PAGE_SIZE; in phys_pte_init()
415 !e820_any_mapped(addr & PAGE_MASK, next, E820_RAM) && in phys_pte_init()
416 !e820_any_mapped(addr & PAGE_MASK, next, E820_RESERVED_KERN)) in phys_pte_init()
438 last_map_addr = (addr & PAGE_MASK) + PAGE_SIZE; in phys_pte_init()
801 next = (addr + PAGE_SIZE) & PAGE_MASK; in remove_pte_table()
813 phys_addr = pte_val(*pte) + (addr & PAGE_MASK); in remove_pte_table()
1326 next = (addr + PAGE_SIZE) & PAGE_MASK; in register_page_bootmem_memmap()
1333 next = (addr + PAGE_SIZE) & PAGE_MASK; in register_page_bootmem_memmap()
1339 next = (addr + PAGE_SIZE) & PAGE_MASK; in register_page_bootmem_memmap()
/linux-4.4.14/arch/openrisc/include/asm/
Dpgtable.h156 #define _PAGE_CHG_MASK (PAGE_MASK | _PAGE_ACCESSED | _PAGE_DIRTY)
218 ((unsigned long)(address)>>(PAGE_SHIFT-SIZEOF_PTR_LOG2)&PTR_MASK&~PAGE_MASK)
228 #define pmd_bad(x) ((pmd_val(x) & (~PAGE_MASK)) != _KERNPG_TABLE)
351 return (unsigned long)__va(pte_val(pte) & PAGE_MASK); in __pte_page()
372 #define pmd_page_kernel(pmd) ((unsigned long) __va(pmd_val(pmd) & PAGE_MASK))
Dfixmap.h57 #define __virt_to_fix(x) ((FIXADDR_TOP - ((x)&PAGE_MASK)) >> PAGE_SHIFT)
/linux-4.4.14/arch/nios2/mm/
Dioremap.c152 offset = phys_addr & ~PAGE_MASK; in __ioremap()
153 phys_addr &= PAGE_MASK; in __ioremap()
182 p = remove_vm_area((void *) (PAGE_MASK & (unsigned long __force) addr)); in __iounmap()
/linux-4.4.14/drivers/infiniband/hw/usnic/
Dusnic_uiom.c129 npages = PAGE_ALIGN(size + (addr & ~PAGE_MASK)) >> PAGE_SHIFT; in usnic_uiom_get_pages()
143 cur_base = addr & PAGE_MASK; in usnic_uiom_get_pages()
227 vpn_start = (uiomr->va & PAGE_MASK) >> PAGE_SHIFT; in __usnic_uiom_reg_release()
258 long int va = uiomr->va & PAGE_MASK; in usnic_uiom_map_sorted_intervals()
355 va_base = addr & PAGE_MASK; in usnic_uiom_reg_get()
356 offset = addr & ~PAGE_MASK; in usnic_uiom_reg_get()
358 vpn_start = (addr & PAGE_MASK) >> PAGE_SHIFT; in usnic_uiom_reg_get()
/linux-4.4.14/arch/mips/kernel/
Dmachine_kexec.c76 (unsigned long) phys_to_virt(image->head & PAGE_MASK); in machine_kexec()
92 phys_to_virt(entry & PAGE_MASK) : ptr + 1) { in machine_kexec()
Dsyscall.c66 if (offset & ~PAGE_MASK) in SYSCALL_DEFINE6()
79 if (pgoff & (~PAGE_MASK >> 12)) in SYSCALL_DEFINE6()
/linux-4.4.14/arch/blackfin/mm/
Dinit.c47 unsigned long end_mem = memory_end & PAGE_MASK; in paging_init()
95 high_memory = (void *)(memory_end & PAGE_MASK); in mem_init()
/linux-4.4.14/arch/cris/include/asm/
Dpgtable.h90 ((unsigned long)(address)>>(PAGE_SHIFT-SIZEOF_PTR_LOG2)&PTR_MASK&~PAGE_MASK)
103 #define pmd_bad(x) ((pmd_val(x) & (~PAGE_MASK & ~_PAGE_KERNEL)) != _PAGE_TABLE)
210 return (unsigned long)__va(pte_val(pte) & PAGE_MASK); in __pte_page()
229 #define pmd_page_vaddr(pmd) ((unsigned long) __va(pmd_val(pmd) & PAGE_MASK))
/linux-4.4.14/sound/core/
Dsgbuf.c48 if (!(sgbuf->table[i].addr & ~PAGE_MASK)) in snd_free_sgbuf_pages()
51 tmpb.addr = sgbuf->table[i].addr & PAGE_MASK; in snd_free_sgbuf_pages()
52 tmpb.bytes = (sgbuf->table[i].addr & ~PAGE_MASK) << PAGE_SHIFT; in snd_free_sgbuf_pages()
/linux-4.4.14/arch/arm/include/asm/
Dtlbflush.h424 uaddr = (uaddr & PAGE_MASK) | ASID(vma->vm_mm); in __local_flush_tlb_page()
445 uaddr = (uaddr & PAGE_MASK) | ASID(vma->vm_mm); in local_flush_tlb_page()
462 uaddr = (uaddr & PAGE_MASK) | ASID(vma->vm_mm); in __flush_tlb_page()
469 tlb_op(TLB_V7_UIS_PAGE, "c8, c3, 3", uaddr & PAGE_MASK); in __flush_tlb_page()
498 kaddr &= PAGE_MASK; in local_flush_tlb_kernel_page()
516 kaddr &= PAGE_MASK; in __flush_tlb_kernel_page()
Ddma-mapping.h141 unsigned int offset = paddr & ~PAGE_MASK; in phys_to_dma()
147 unsigned int offset = dev_addr & ~PAGE_MASK; in dma_to_phys()
Dcacheflush.h233 __cpuc_flush_user_range(start & PAGE_MASK, PAGE_ALIGN(end), in vivt_flush_cache_range()
243 unsigned long addr = user_addr & PAGE_MASK; in vivt_flush_cache_page()
/linux-4.4.14/kernel/
Dkexec_core.c170 if ((mstart & ~PAGE_MASK) || (mend & ~PAGE_MASK)) in sanity_check_segment_list()
499 destination &= PAGE_MASK; in kimage_set_destination()
510 page &= PAGE_MASK; in kimage_add_page()
537 phys_to_virt((entry & PAGE_MASK)) : ptr + 1)
596 destination = entry & PAGE_MASK; in kimage_dst_used()
679 old_addr = *old & PAGE_MASK; in kimage_alloc_page()
682 *old = addr | (*old & ~PAGE_MASK); in kimage_alloc_page()
744 ptr += maddr & ~PAGE_MASK; in kimage_load_normal_segment()
746 PAGE_SIZE - (maddr & ~PAGE_MASK)); in kimage_load_normal_segment()
803 ptr += maddr & ~PAGE_MASK; in kimage_load_crash_segment()
[all …]
/linux-4.4.14/arch/arm/mm/
Dmmap.c100 info.align_mask = do_align ? (PAGE_MASK & (SHMLBA - 1)) : 0; in arch_get_unmapped_area()
151 info.align_mask = do_align ? (PAGE_MASK & (SHMLBA - 1)) : 0; in arch_get_unmapped_area_topdown()
161 if (addr & ~PAGE_MASK) { in arch_get_unmapped_area_topdown()
Ddma-mapping.c112 handle & ~PAGE_MASK, size, dir); in arm_dma_unmap_page()
1293 size = PAGE_ALIGN((iova & ~PAGE_MASK) + size); in __iommu_remove_mapping()
1294 iova &= PAGE_MASK; in __iommu_remove_mapping()
1564 if (s->offset || (size & ~PAGE_MASK) || size + s->length > max) { in __iommu_map_sg()
1784 dma_addr_t iova = handle & PAGE_MASK; in arm_coherent_iommu_unmap_page()
1785 int offset = handle & ~PAGE_MASK; in arm_coherent_iommu_unmap_page()
1809 dma_addr_t iova = handle & PAGE_MASK; in arm_iommu_unmap_page()
1811 int offset = handle & ~PAGE_MASK; in arm_iommu_unmap_page()
1828 dma_addr_t iova = handle & PAGE_MASK; in arm_iommu_sync_single_for_cpu()
1830 unsigned int offset = handle & ~PAGE_MASK; in arm_iommu_sync_single_for_cpu()
[all …]
/linux-4.4.14/arch/frv/mm/
Delf-fdpic.c89 if (!(addr & ~PAGE_MASK)) in arch_get_unmapped_area()
97 if (!(addr & ~PAGE_MASK)) in arch_get_unmapped_area()
Dcache-page.c59 start = (start & ~PAGE_MASK) | (unsigned long) vaddr; in flush_icache_user_range()
/linux-4.4.14/arch/m32r/include/asm/
Dpage.h9 #define PAGE_MASK (~(PAGE_SIZE-1)) macro
30 #define PTE_MASK PAGE_MASK
Dpgtable.h183 #define pmd_bad(x) ((pmd_val(x) & ~PAGE_MASK) != _KERNPG_TABLE)
296 pmd_val(*pmdp) = (((unsigned long) ptep) & PAGE_MASK); in pmd_set()
300 ((unsigned long) __va(pmd_val(pmd) & PAGE_MASK))
/linux-4.4.14/arch/avr32/include/asm/
Dpage.h16 #define PAGE_MASK (~(PAGE_SIZE-1)) macro
17 #define PTE_MASK PAGE_MASK
/linux-4.4.14/arch/x86/kvm/
Dx86.h90 vcpu->arch.mmio_gva = gva & PAGE_MASK; in vcpu_cache_mmio_info()
109 if (gva != MMIO_GVA_ANY && vcpu->arch.mmio_gva != (gva & PAGE_MASK)) in vcpu_clear_mmio_info()
118 vcpu->arch.mmio_gva == (gva & PAGE_MASK)) in vcpu_match_mmio_gva()
/linux-4.4.14/mm/
Dmsync.c46 len = (len + ~PAGE_MASK) & PAGE_MASK; in SYSCALL_DEFINE3()
Dearly_ioremap.c130 phys_addr &= PAGE_MASK; in __early_ioremap()
241 p = early_memremap(src & PAGE_MASK, clen + slop); in copy_from_early_mem()
Dslob.c147 slob_t *base = (slob_t *)((unsigned long)s & PAGE_MASK); in set_slob()
172 slob_t *base = (slob_t *)((unsigned long)s & PAGE_MASK); in slob_next()
187 return !((unsigned long)slob_next(s) & ~PAGE_MASK); in slob_last()
374 SLOB_UNITS(PAGE_SIZE)) & PAGE_MASK)); in slob_free()
Dmadvise.c477 if (start & ~PAGE_MASK) in SYSCALL_DEFINE3()
479 len = (len_in + ~PAGE_MASK) & PAGE_MASK; in SYSCALL_DEFINE3()
/linux-4.4.14/fs/
Dbinfmt_aout.c277 error = vm_brk(text_addr & PAGE_MASK, map_size); in load_aout_binary()
278 if (error != (text_addr & PAGE_MASK)) in load_aout_binary()
292 if ((fd_offset & ~PAGE_MASK) != 0 && printk_ratelimit()) in load_aout_binary()
299 if (!bprm->file->f_op->mmap||((fd_offset & ~PAGE_MASK) != 0)) { in load_aout_binary()
374 if ((N_TXTOFF(ex) & ~PAGE_MASK) != 0) { in load_aout_library()
/linux-4.4.14/arch/x86/platform/ce4100/
Dce4100.c106 up->mapbase & PAGE_MASK); in ce4100_serial_fixup()
109 up->membase += up->mapbase & ~PAGE_MASK; in ce4100_serial_fixup()
/linux-4.4.14/drivers/target/tcm_fc/
Dtfc_io.c157 from += mem_off & ~PAGE_MASK; in ft_queue_data_in()
159 (mem_off & ~PAGE_MASK))); in ft_queue_data_in()
317 to += mem_off & ~PAGE_MASK; in ft_recv_write_data()
319 (mem_off & ~PAGE_MASK))); in ft_recv_write_data()
/linux-4.4.14/tools/virtio/linux/
Dkernel.h21 #define PAGE_MASK (~(PAGE_SIZE-1)) macro
36 #define virt_to_page(p) ((struct page *)((unsigned long)p & PAGE_MASK))
/linux-4.4.14/arch/arm/kernel/
Dpatch.c41 return (void *) (__fix_to_virt(fixmap) + (uintaddr & ~PAGE_MASK)); in patch_map()
78 twopage = (uintaddr & ~PAGE_MASK) == PAGE_SIZE - 2; in __patch_text_real()
/linux-4.4.14/arch/cris/arch-v10/mm/
Dtlb.c111 addr &= PAGE_MASK; /* perhaps not necessary */ in flush_tlb_page()
123 (tlb_hi & PAGE_MASK) == addr) { in flush_tlb_page()
/linux-4.4.14/arch/sh/kernel/
Dmachine_kexec.c89 phys_to_virt(entry & PAGE_MASK) : ptr + 1) { in machine_kexec()
134 phys_to_virt(*ptr & PAGE_MASK) : ptr + 1) { in machine_kexec()
Dsys_sh.c35 if (off & ~PAGE_MASK) in old_mmap()
/linux-4.4.14/arch/cris/arch-v32/mm/
Dtlb.c135 addr &= PAGE_MASK; in __flush_tlb_page()
151 ((tlb_hi & PAGE_MASK) == addr)) { in __flush_tlb_page()
Dmmu.S128 and.w PAGE_MASK, $acr ; Remove PMD flags
188 move.w PAGE_MASK, $acr
192 and.w PAGE_MASK, $r10 ; Get VPN (virtual address)
/linux-4.4.14/arch/powerpc/platforms/powermac/
Dudbg_scc.c116 sccc = ioremap(addr & PAGE_MASK, PAGE_SIZE) ; in udbg_scc_init()
117 sccc += addr & ~PAGE_MASK; in udbg_scc_init()
/linux-4.4.14/arch/m68k/sun3x/
Ddvma.c89 kaddr &= PAGE_MASK; in dvma_map_cpu()
90 vaddr &= PAGE_MASK; in dvma_map_cpu()
/linux-4.4.14/drivers/media/v4l2-core/
Dvideobuf-dma-sg.c172 first = (data & PAGE_MASK) >> PAGE_SHIFT; in videobuf_dma_init_user_locked()
173 last = ((data+size-1) & PAGE_MASK) >> PAGE_SHIFT; in videobuf_dma_init_user_locked()
174 dma->offset = data & ~PAGE_MASK; in videobuf_dma_init_user_locked()
185 data & PAGE_MASK, dma->nr_pages, in videobuf_dma_init_user_locked()
303 & PAGE_MASK; in videobuf_dma_map()
304 dma->sglist[0].offset = dma->bus_addr & ~PAGE_MASK; in videobuf_dma_map()
/linux-4.4.14/arch/x86/ia32/
Dia32_aout.c322 error = vm_brk(text_addr & PAGE_MASK, map_size); in load_aout_binary()
324 if (error != (text_addr & PAGE_MASK)) in load_aout_binary()
341 if ((fd_offset & ~PAGE_MASK) != 0 && in load_aout_binary()
351 if (!bprm->file->f_op->mmap || (fd_offset & ~PAGE_MASK) != 0) { in load_aout_binary()
426 if ((N_TXTOFF(ex) & ~PAGE_MASK) != 0) { in load_aout_library()
/linux-4.4.14/arch/alpha/kernel/
Dcore_irongate.c354 if (addr & ~PAGE_MASK) { in irongate_ioremap()
401 vaddr = (unsigned long)area->addr + (addr & ~PAGE_MASK); in irongate_ioremap()
417 return vfree((void *)(PAGE_MASK & addr)); in irongate_iounmap()
Dpci_iommu.c312 paddr &= PAGE_MASK; in pci_map_single_1()
317 ret += (unsigned long)cpu_addr & ~PAGE_MASK; in pci_map_single_1()
530 } else if (((next_paddr | addr) & ~PAGE_MASK) == 0 && virt_ok) { in sg_classify()
594 paddr &= ~PAGE_MASK; in sg_fill()
633 paddr &= PAGE_MASK; in sg_fill()
/linux-4.4.14/arch/tile/mm/
Dhugetlbpage.c175 info.align_mask = PAGE_MASK & ~huge_page_mask(h); in hugetlb_get_unmapped_area_bottomup()
192 info.align_mask = PAGE_MASK & ~huge_page_mask(h); in hugetlb_get_unmapped_area_topdown()
202 if (addr & ~PAGE_MASK) { in hugetlb_get_unmapped_area_topdown()
/linux-4.4.14/arch/arc/include/uapi/asm/
Dpage.h33 #define PAGE_MASK (~(PAGE_SIZE-1)) macro
/linux-4.4.14/drivers/video/fbdev/
Digafb.c543 par->mmap_map[0].poff = par->frame_buffer_phys & PAGE_MASK; in igafb_init()
544 par->mmap_map[0].size = info->fix.smem_len & PAGE_MASK; in igafb_init()
549 par->mmap_map[1].voff = par->frame_buffer_phys & PAGE_MASK; in igafb_init()
550 par->mmap_map[1].poff = info->fix.smem_start & PAGE_MASK; in igafb_init()
/linux-4.4.14/drivers/gpu/drm/qxl/
Dqxl_ioctl.c88 reloc_page = qxl_bo_kmap_atomic_page(qdev, info->dst_bo, info->dst_offset & PAGE_MASK); in apply_reloc()
89 *(uint64_t *)(reloc_page + (info->dst_offset & ~PAGE_MASK)) = qxl_bo_physical_address(qdev, in apply_reloc()
104 reloc_page = qxl_bo_kmap_atomic_page(qdev, info->dst_bo, info->dst_offset & PAGE_MASK); in apply_surf_reloc()
105 *(uint32_t *)(reloc_page + (info->dst_offset & ~PAGE_MASK)) = id; in apply_surf_reloc()
/linux-4.4.14/arch/mn10300/kernel/
Dsys_mn10300.c30 if (offset & ~PAGE_MASK) in old_mmap()
/linux-4.4.14/arch/um/kernel/
Dsyscall.c22 if (offset & ~PAGE_MASK) in old_mmap()
Dum_arch.c295 uml_physmem = (unsigned long) __binary_start & PAGE_MASK; in linux_main()
303 iomem_size = (iomem_size + PAGE_SIZE - 1) & PAGE_MASK; in linux_main()
Dtlb.c218 ret = add_mmap(addr, pte_val(*pte) & PAGE_MASK, in update_pte_range()
381 pte_val(*pte) & PAGE_MASK, in flush_tlb_kernel_range_common()
404 address &= PAGE_MASK; in flush_tlb_page()
437 fd = phys_mapping(pte_val(*pte) & PAGE_MASK, &offset); in flush_tlb_page()
/linux-4.4.14/arch/mips/include/asm/
Dpage.h36 #define PAGE_MASK (~((1 << PAGE_SHIFT) - 1)) macro
104 if (pages_do_alias((unsigned long) addr, vaddr & PAGE_MASK)) in clear_user_page()
Dpgtable-64.h184 if (unlikely(pmd_val(pmd) & ~PAGE_MASK)) in pmd_bad()
211 return pud_val(pud) & ~PAGE_MASK; in pud_bad()
/linux-4.4.14/arch/avr32/mach-at32ap/
Dpm.c46 page_addr = pm_sram_start & PAGE_MASK; in avr32_pm_map_sram()
59 tlbehi |= vaddr & PAGE_MASK; in avr32_pm_map_sram()
/linux-4.4.14/arch/sh/include/asm/
Dpgtable_64.h69 #define pmd_bad(pmd_entry) ((pmd_val(pmd_entry) & (~PAGE_MASK & ~_PAGE_USER)) != _KERNPG_TABLE)
72 ((unsigned long) __va(pmd_val(pmd_entry) & PAGE_MASK))
84 ((pte_t *) ((pmd_val(*(dir))) & PAGE_MASK) + pte_index((addr)))
Dpage.h24 #define PAGE_MASK (~(PAGE_SIZE-1)) macro
25 #define PTE_MASK PAGE_MASK
Dpgtable-3level.h48 #define pud_bad(x) (pud_val(x) & ~PAGE_MASK)
/linux-4.4.14/arch/score/include/asm/
Dpgtable-bits.h22 (PAGE_MASK | _PAGE_ACCESSED | _PAGE_MODIFIED | _PAGE_CACHE)
Dfixmap.h63 ((FIXADDR_TOP - ((x) & PAGE_MASK)) >> PAGE_SHIFT)
/linux-4.4.14/drivers/tty/serial/
Dearlycon.c51 set_fixmap_io(FIX_EARLYCON_MEM_BASE, paddr & PAGE_MASK); in earlycon_map()
53 base += paddr & ~PAGE_MASK; in earlycon_map()
/linux-4.4.14/arch/unicore32/mm/
Dioremap.c188 unsigned long offset = phys_addr & ~PAGE_MASK; in __uc32_ioremap_caller()
237 void *addr = (void *)(PAGE_MASK & (unsigned long)io_addr); in __uc32_iounmap()
Dcache-ucv2.S107 sub r9, r9, #1 @ PAGE_MASK
161 sub r9, r9, #1 @ PAGE_MASK
/linux-4.4.14/arch/tile/kernel/
Dmachine_kexec.c159 if (((unsigned long)desc & PAGE_MASK) != (unsigned long)pg) { in kexec_bn2cl()
179 phys_to_virt((entry & PAGE_MASK)) : ptr + 1) { in kexec_find_and_set_command_line()
/linux-4.4.14/arch/microblaze/include/asm/
Dpgtable.h217 #define _PMD_PRESENT PAGE_MASK
235 #define _PAGE_CHG_MASK (PAGE_MASK | _PAGE_ACCESSED | _PAGE_DIRTY)
326 ((unsigned long) __va(pgd_val(pgd) & PAGE_MASK))
470 #define pmd_page_kernel(pmd) ((unsigned long) (pmd_val(pmd) & PAGE_MASK))
/linux-4.4.14/arch/score/kernel/
Dsys_score.c47 if (unlikely(offset & ~PAGE_MASK)) in sys_mmap()
/linux-4.4.14/arch/powerpc/boot/
Dpage.h22 #define PAGE_MASK (~(PAGE_SIZE-1)) macro
/linux-4.4.14/drivers/mtd/onenand/
Domap2.c320 if (((size_t)buf & PAGE_MASK) != in omap3_onenand_read_bufferram()
321 ((size_t)(buf + count - 1) & PAGE_MASK)) in omap3_onenand_read_bufferram()
326 buf = page_address(p1) + ((size_t)buf & ~PAGE_MASK); in omap3_onenand_read_bufferram()
397 if (((size_t)buf & PAGE_MASK) != in omap3_onenand_write_bufferram()
398 ((size_t)(buf + count - 1) & PAGE_MASK)) in omap3_onenand_write_bufferram()
403 buf = page_address(p1) + ((size_t)buf & ~PAGE_MASK); in omap3_onenand_write_bufferram()
/linux-4.4.14/arch/c6x/mm/
Dinit.c61 high_memory = (void *)(memory_end & PAGE_MASK); in mem_init()
/linux-4.4.14/arch/alpha/mm/
Dinit.c181 (((unsigned long)kernel_end + ~PAGE_MASK) & PAGE_MASK); in callback_init()
/linux-4.4.14/arch/arm/mach-orion5x/
Dcommon.c372 (!t->u.mem.size || t->u.mem.size & ~PAGE_MASK || in tag_fixup_mem32()
373 t->u.mem.start & ~PAGE_MASK)) { in tag_fixup_mem32()
/linux-4.4.14/arch/frv/include/asm/
Dmem-layout.h32 #define PAGE_MASK (~(PAGE_SIZE-1)) macro
/linux-4.4.14/arch/arm64/include/asm/
Dpage.h35 #define PAGE_MASK (~(PAGE_SIZE-1)) macro
/linux-4.4.14/arch/um/include/shared/
Dcommon-offsets.h6 DEFINE(UM_KERN_PAGE_MASK, PAGE_MASK);
Dkern_util.h20 ((((unsigned long) addr) + PAGE_SIZE - 1) & PAGE_MASK)
/linux-4.4.14/arch/m68k/include/asm/
Dpage.h15 #define PAGE_MASK (~(PAGE_SIZE-1)) macro
Dmmu_context.h143 mmu_write(MMUTR, (mmuar & PAGE_MASK) | (asid << MMUTR_IDN) | in load_ksp_mmu()
147 mmu_write(MMUDR, (pte_val(*pte) & PAGE_MASK) | in load_ksp_mmu()
/linux-4.4.14/arch/x86/include/asm/xen/
Dpage.h181 unsigned offset = phys.paddr & ~PAGE_MASK; in phys_to_machine()
187 unsigned offset = machine.maddr & ~PAGE_MASK; in machine_to_phys()
/linux-4.4.14/arch/unicore32/kernel/
Dsetup.c148 size -= start & ~PAGE_MASK; in uc32_add_memory()
151 bank->size = size & PAGE_MASK; in uc32_add_memory()
/linux-4.4.14/drivers/lguest/
Dlg.h148 #define pgd_flags(x) (pgd_val(x) & ~PAGE_MASK)
150 #define pmd_flags(x) (pmd_val(x) & ~PAGE_MASK)
/linux-4.4.14/net/rds/
Diw_rdma.c286 if (dma_addr & PAGE_MASK) {
289 dma_addr &= ~PAGE_MASK;
291 if (end_addr & PAGE_MASK) {
294 end_addr = (end_addr + PAGE_MASK) & ~PAGE_MASK;
/linux-4.4.14/arch/sparc/include/asm/
Dross.h178 page &= PAGE_MASK; in hyper_flush_cache_page()
Diommu_32.h118 sbus_writel(ba & PAGE_MASK, &regs->pageflush); in iommu_invalidate_page()
/linux-4.4.14/arch/c6x/include/asm/
Dpgtable.h36 #define pmd_bad(x) (pmd_val(x) & ~PAGE_MASK)
/linux-4.4.14/arch/um/kernel/skas/
Duaccess.c73 (addr & ~PAGE_MASK); in do_op_one_page()
109 while (addr < ((addr + remain) & PAGE_MASK)) { in buffer_op()
/linux-4.4.14/arch/arm/lib/
Duaccess_with_memcpy.c118 tocopy = (~(unsigned long)to & ~PAGE_MASK) + 1; in __copy_to_user_memcpy()
184 tocopy = (~(unsigned long)addr & ~PAGE_MASK) + 1; in __clear_user_memset()
/linux-4.4.14/drivers/net/ethernet/ibm/ehea/
Dehea_phyp.h67 epas->kernel.addr = ioremap((paddr_kernel & PAGE_MASK), PAGE_SIZE) + in hcp_epas_ctor()
68 (paddr_kernel & ~PAGE_MASK); in hcp_epas_ctor()
75 iounmap((void __iomem *)((u64)epas->kernel.addr & PAGE_MASK)); in hcp_epas_dtor()
/linux-4.4.14/drivers/vfio/
Dvfio_iommu_spapr_tce.c110 if ((vaddr & ~PAGE_MASK) || (size & ~PAGE_MASK)) in tce_iommu_unregister_pages()
127 if ((vaddr & ~PAGE_MASK) || (size & ~PAGE_MASK) || in tce_iommu_register_pages()
449 if (get_user_pages_fast(tce & PAGE_MASK, 1, in tce_iommu_use_page()
469 unsigned long offset = tce & IOMMU_PAGE_MASK(tbl) & ~PAGE_MASK; in tce_iommu_build()
531 hpa |= tce & IOMMU_PAGE_MASK(tbl) & ~PAGE_MASK; in tce_iommu_build_v2()
/linux-4.4.14/drivers/vfio/platform/
Dvfio_platform_common.c104 if (!(vdev->regions[i].addr & ~PAGE_MASK) && in vfio_platform_regions_init()
105 !(vdev->regions[i].size & ~PAGE_MASK)) in vfio_platform_regions_init()
510 if (vma->vm_start & ~PAGE_MASK) in vfio_platform_mmap()
512 if (vma->vm_end & ~PAGE_MASK) in vfio_platform_mmap()
/linux-4.4.14/arch/cris/include/arch-v10/arch/
Dmmu.h72 #define _PAGE_CHG_MASK (PAGE_MASK | _PAGE_ACCESSED | _PAGE_MODIFIED)
/linux-4.4.14/include/asm-generic/
Dfixmap.h21 #define __virt_to_fix(x) ((FIXADDR_TOP - ((x)&PAGE_MASK)) >> PAGE_SHIFT)

1234