/linux-4.1.27/arch/x86/um/ |
H A D | mem_64.c | 7 if (vma->vm_mm && vma->vm_start == um_vdso_addr) arch_vma_name()
|
H A D | mem_32.c | 21 gate_vma.vm_start = FIXADDR_USER_START; gate_vma_init() 53 return (addr >= vma->vm_start) && (addr < vma->vm_end); in_gate_area()
|
/linux-4.1.27/arch/sparc/include/asm/ |
H A D | tlb_32.h | 6 flush_cache_range(vma, vma->vm_start, vma->vm_end); \ 11 flush_tlb_range(vma, vma->vm_start, vma->vm_end); \
|
/linux-4.1.27/arch/ia64/include/asm/ |
H A D | fb.h | 12 if (efi_range_is_wc(vma->vm_start, vma->vm_end - vma->vm_start)) fb_pgprotect()
|
/linux-4.1.27/arch/metag/include/asm/ |
H A D | tlb.h | 14 flush_cache_range(vma, vma->vm_start, vma->vm_end); \ 20 flush_tlb_range(vma, vma->vm_start, vma->vm_end); \
|
/linux-4.1.27/arch/parisc/include/asm/ |
H A D | tlb.h | 11 flush_cache_range(vma, vma->vm_start, vma->vm_end); \ 16 flush_tlb_range(vma, vma->vm_start, vma->vm_end); \
|
/linux-4.1.27/arch/avr32/include/asm/ |
H A D | tlb.h | 12 flush_cache_range(vma, vma->vm_start, vma->vm_end) 15 flush_tlb_range(vma, vma->vm_start, vma->vm_end)
|
/linux-4.1.27/arch/powerpc/include/asm/ |
H A D | fb.h | 12 vma->vm_end - vma->vm_start, fb_pgprotect()
|
/linux-4.1.27/arch/metag/kernel/ |
H A D | dma.c | 90 * .vm_start = VMALLOC_START, 94 * However, vmalloc_head.vm_start is variable (typically, it is dependent on 101 unsigned long vm_start; member in struct:metag_vm_region 109 .vm_start = CONSISTENT_START, 117 unsigned long addr = head->vm_start, end = head->vm_end - size; metag_vm_region_alloc() 130 if ((addr + size) <= c->vm_start) metag_vm_region_alloc() 142 new->vm_start = addr; metag_vm_region_alloc() 162 if (c->vm_active && c->vm_start == addr) metag_vm_region_find() 225 unsigned long vaddr = c->vm_start; dma_alloc_coherent() 258 return (void *)c->vm_start; dma_alloc_coherent() 287 if ((c->vm_end - c->vm_start) != size) { dma_free_coherent() 289 __func__, c->vm_end - c->vm_start, size); dma_free_coherent() 291 size = c->vm_end - c->vm_start; dma_free_coherent() 294 ptep = consistent_pte + CONSISTENT_OFFSET(c->vm_start); dma_free_coherent() 295 addr = c->vm_start; dma_free_coherent() 317 flush_tlb_kernel_range(c->vm_start, c->vm_end); dma_free_coherent() 343 user_size = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; dma_mmap() 352 kern_size = (c->vm_end - c->vm_start) >> PAGE_SHIFT; dma_mmap() 356 ret = remap_pfn_range(vma, vma->vm_start, dma_mmap()
|
/linux-4.1.27/drivers/infiniband/hw/ehca/ |
H A D | ehca_uverbs.c | 75 ehca_gen_err("Invalid vma struct vm_start=%lx vm_end=%lx", ehca_mm_open() 76 vma->vm_start, vma->vm_end); ehca_mm_open() 81 ehca_gen_err("Use count overflow vm_start=%lx vm_end=%lx", ehca_mm_open() 82 vma->vm_start, vma->vm_end); ehca_mm_open() 83 ehca_gen_dbg("vm_start=%lx vm_end=%lx count=%x", ehca_mm_open() 84 vma->vm_start, vma->vm_end, *count); ehca_mm_open() 91 ehca_gen_err("Invalid vma struct vm_start=%lx vm_end=%lx", ehca_mm_close() 92 vma->vm_start, vma->vm_end); ehca_mm_close() 96 ehca_gen_dbg("vm_start=%lx vm_end=%lx count=%x", ehca_mm_close() 97 vma->vm_start, vma->vm_end, *count); ehca_mm_close() 111 vsize = vma->vm_end - vma->vm_start; ehca_mmap_fw() 113 ehca_gen_err("invalid vsize=%lx", vma->vm_end - vma->vm_start); ehca_mmap_fw() 121 ret = remap_4k_pfn(vma, vma->vm_start, physical >> EHCA_PAGESHIFT, ehca_mmap_fw() 143 start = vma->vm_start; ehca_mmap_queue()
|
/linux-4.1.27/arch/xtensa/include/asm/ |
H A D | tlb.h | 29 flush_cache_range(vma, vma->vm_start, vma->vm_end); \ 35 flush_tlb_range(vma, vma->vm_start, vma->vm_end); \
|
/linux-4.1.27/arch/arc/include/asm/ |
H A D | tlb.h | 32 flush_cache_range(vma, vma->vm_start, vma->vm_end); \ 39 flush_tlb_range(vma, vma->vm_start, vma->vm_end); \
|
/linux-4.1.27/drivers/xen/xenfs/ |
H A D | xenstored.c | 35 size_t size = vma->vm_end - vma->vm_start; xsd_kva_mmap() 40 if (remap_pfn_range(vma, vma->vm_start, xsd_kva_mmap()
|
/linux-4.1.27/mm/ |
H A D | msync.c | 68 if (start < vma->vm_start) { SYSCALL_DEFINE3() 69 start = vma->vm_start; SYSCALL_DEFINE3() 74 /* Here vma->vm_start <= start < vma->vm_end. */ SYSCALL_DEFINE3() 81 fstart = (start - vma->vm_start) + SYSCALL_DEFINE3()
|
H A D | nommu.c | 141 return vma->vm_end - vma->vm_start; kobjsize() 597 BUG_ON(unlikely(last->vm_end <= last->vm_start)); validate_nommu_regions() 604 BUG_ON(unlikely(region->vm_end <= region->vm_start)); validate_nommu_regions() 606 BUG_ON(unlikely(region->vm_start < last->vm_top)); validate_nommu_regions() 632 if (region->vm_start < pregion->vm_start) add_nommu_region() 634 else if (region->vm_start > pregion->vm_start) add_nommu_region() 681 * will equal vm_start 691 if (region->vm_top > region->vm_start) __releases() 702 free_page_series(region->vm_start, region->vm_top); __releases() 726 long start = vma->vm_start & PAGE_MASK; protect_vma() 776 if (vma->vm_start < pvma->vm_start) add_vma_to_mm() 778 else if (vma->vm_start > pvma->vm_start) { add_vma_to_mm() 882 if (vma->vm_start > addr) find_vma() 931 if (vma->vm_start < addr) find_vma_exact() 933 if (vma->vm_start > addr) find_vma_exact() 1226 region->vm_start = (unsigned long) base; do_mmap_private() 1227 region->vm_end = region->vm_start + len; do_mmap_private() 1228 region->vm_top = region->vm_start + (total << PAGE_SHIFT); do_mmap_private() 1230 vma->vm_start = region->vm_start; do_mmap_private() 1231 vma->vm_end = region->vm_start + len; do_mmap_private() 1258 free_page_series(region->vm_start, region->vm_top); do_mmap_private() 1259 region->vm_start = vma->vm_start = 0; do_mmap_private() 1362 rpglen = pregion->vm_end - pregion->vm_start; do_mmap_pgoff() 1381 start = pregion->vm_start; do_mmap_pgoff() 1383 vma->vm_start = start; do_mmap_pgoff() 1394 vma->vm_start = 0; do_mmap_pgoff() 1429 vma->vm_start = region->vm_start = addr; do_mmap_pgoff() 1450 memset((void *)region->vm_start, 0, do_mmap_pgoff() 1451 region->vm_end - region->vm_start); do_mmap_pgoff() 1454 result = vma->vm_start; do_mmap_pgoff() 1464 flush_icache_range(region->vm_start, region->vm_end); do_mmap_pgoff() 1591 npages = (addr - vma->vm_start) >> PAGE_SHIFT; split_vma() 1596 region->vm_start = new->vm_start = addr; split_vma() 1607 vma->vm_region->vm_start = vma->vm_start = addr; split_vma() 1636 if (from > vma->vm_start) shrink_vma() 1639 vma->vm_start = to; shrink_vma() 1648 if (from > region->vm_start) { shrink_vma() 1652 region->vm_start = to; shrink_vma() 1698 if (start > vma->vm_start) { do_munmap() 1710 if (start == vma->vm_start && end == vma->vm_end) do_munmap() 1712 if (start < vma->vm_start || end > vma->vm_end) { do_munmap() 1724 if (start != vma->vm_start && end != vma->vm_end) { do_munmap() 1820 if (vma->vm_end != vma->vm_start + old_len) do_mremap() 1826 if (new_len > vma->vm_region->vm_end - vma->vm_region->vm_start) do_mremap() 1830 vma->vm_end = vma->vm_start + new_len; do_mremap() 1831 return vma->vm_start; do_mremap() 1868 unsigned long vm_len = vma->vm_end - vma->vm_start; vm_iomap_memory() 1871 return io_remap_pfn_range(vma, vma->vm_start, pfn, vm_len, vma->vm_page_prot); vm_iomap_memory() 1878 unsigned int size = vma->vm_end - vma->vm_start; remap_vmalloc_range() 1883 vma->vm_start = (unsigned long)(addr + (pgoff << PAGE_SHIFT)); remap_vmalloc_range() 1884 vma->vm_end = vma->vm_start + size; remap_vmalloc_range() 2126 r_size = region->vm_top - region->vm_start; nommu_shrink_inode_mappings()
|
H A D | madvise.c | 103 pgoff = vma->vm_pgoff + ((start - vma->vm_start) >> PAGE_SHIFT); madvise_behavior() 113 if (start != vma->vm_start) { madvise_behavior() 196 index = ((start - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff; force_shm_swapin_readahead() 248 start = ((start - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff; madvise_willneed() 251 end = ((end - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff; madvise_willneed() 314 offset = (loff_t)(start - vma->vm_start) madvise_remove() 503 if (vma && start > vma->vm_start) SYSCALL_DEFINE3() 514 if (start < vma->vm_start) { SYSCALL_DEFINE3() 516 start = vma->vm_start; SYSCALL_DEFINE3() 521 /* Here vma->vm_start <= start < (end|vma->vm_end) */ SYSCALL_DEFINE3() 526 /* Here vma->vm_start <= start < tmp <= (end|vma->vm_end). */ SYSCALL_DEFINE3()
|
H A D | mremap.c | 269 new_pgoff = vma->vm_pgoff + ((old_addr - vma->vm_start) >> PAGE_SHIFT); move_vma() 301 excess = vma->vm_end - vma->vm_start - old_len; move_vma() 302 if (old_addr > vma->vm_start && move_vma() 347 if (!vma || vma->vm_start > addr) vma_to_resize() 363 pgoff = (addr - vma->vm_start) >> PAGE_SHIFT; vma_to_resize() 437 ((addr - vma->vm_start) >> PAGE_SHIFT), mremap_to() 457 if (vma->vm_next && vma->vm_next->vm_start < end) /* intersection */ vma_expandable() 459 if (get_unmapped_area(NULL, vma->vm_start, end - vma->vm_start, vma_expandable() 539 if (vma_adjust(vma, vma->vm_start, addr + new_len, SYSCALL_DEFINE5() 568 ((addr - vma->vm_start) >> PAGE_SHIFT), SYSCALL_DEFINE5()
|
H A D | mprotect.c | 291 pgoff = vma->vm_pgoff + ((start - vma->vm_start) >> PAGE_SHIFT); mprotect_fixup() 301 if (start != vma->vm_start) { mprotect_fixup() 374 if (vma->vm_start >= end) SYSCALL_DEFINE3() 376 start = vma->vm_start; SYSCALL_DEFINE3() 381 if (vma->vm_start > start) SYSCALL_DEFINE3() 390 if (start > vma->vm_start) SYSCALL_DEFINE3() 396 /* Here we know that vma->vm_start <= nstart < vma->vm_end. */ SYSCALL_DEFINE3() 425 if (!vma || vma->vm_start != nstart) { SYSCALL_DEFINE3()
|
H A D | mmap.c | 358 max = vma->vm_start; vma_compute_subtree_gap() 386 if (vma->vm_start < prev) { browse_rb() 387 pr_emerg("vm_start %lx < prev %lx\n", browse_rb() 388 vma->vm_start, prev); browse_rb() 391 if (vma->vm_start < pend) { browse_rb() 392 pr_emerg("vm_start %lx < pend %lx\n", browse_rb() 393 vma->vm_start, pend); browse_rb() 396 if (vma->vm_start > vma->vm_end) { browse_rb() 397 pr_emerg("vm_start %lx > vm_end %lx\n", browse_rb() 398 vma->vm_start, vma->vm_end); browse_rb() 409 prev = vma->vm_start; browse_rb() 483 * Update augmented rbtree rb_subtree_gap values after vma->vm_start or 525 * Before updating the vma's vm_start / vm_end / vm_pgoff fields, the 570 if (vma_tmp->vm_start < end) find_vma_links() 599 max(addr, vma->vm_start)) >> PAGE_SHIFT; count_vma_pages_range() 605 if (vma->vm_start > end) count_vma_pages_range() 608 overlap_len = min(end, vma->vm_end) - vma->vm_start; count_vma_pages_range() 697 if (find_vma_links(mm, vma->vm_start, vma->vm_end, __insert_vm_struct() 720 * We cannot adjust vm_start, vm_end, vm_pgoff fields of a vma that 752 } else if (end > next->vm_start) { vma_adjust() 757 adjust_next = (end - next->vm_start) >> PAGE_SHIFT; vma_adjust() 789 uprobe_munmap(vma, vma->vm_start, vma->vm_end); vma_adjust() 792 uprobe_munmap(next, next->vm_start, next->vm_end); vma_adjust() 827 if (start != vma->vm_start) { vma_adjust() 828 vma->vm_start = start; vma_adjust() 837 next->vm_start += adjust_next << PAGE_SHIFT; vma_adjust() 892 uprobe_munmap(next, next->vm_start, next->vm_end); vma_adjust() 1067 if (next && end == next->vm_start && vma_merge() 1074 err = vma_adjust(prev, prev->vm_start, vma_merge() 1077 err = vma_adjust(prev, prev->vm_start, vma_merge() 1088 if (next && end == next->vm_start && vma_merge() 1093 err = vma_adjust(prev, prev->vm_start, vma_merge() 1122 return a->vm_end == b->vm_start && anon_vma_compatible() 1126 b->vm_pgoff == a->vm_pgoff + ((b->vm_start - a->vm_start) >> PAGE_SHIFT); anon_vma_compatible() 1594 vma->vm_start = addr; mmap_region() 1630 WARN_ON_ONCE(addr != vma->vm_start); mmap_region() 1632 addr = vma->vm_start; mmap_region() 1682 unmap_region(mm, vma, prev, vma->vm_start, vma->vm_end); mmap_region() 1703 * - gap_end = vma->vm_start >= info->low_limit + length; unmapped_area() 1734 gap_end = vma->vm_start; unmapped_area() 1773 gap_end = vma->vm_start; unmapped_area() 1850 gap_end = vma->vm_start; unmapped_area_topdown() 1927 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area() 1967 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area_topdown() 2051 if (tmp->vm_start <= addr) find_vma() 2122 new_start = (vma->vm_flags & VM_GROWSUP) ? vma->vm_start : acct_stack_growth() 2164 * vma->vm_start/vm_end cannot change under us because the caller expand_upwards() 2174 size = address - vma->vm_start; expand_upwards() 2214 * vma is the first one with address < vma->vm_start. Have to extend vma. 2231 * vma->vm_start/vm_end cannot change under us because the caller expand_downwards() 2238 if (address < vma->vm_start) { expand_downwards() 2242 grow = (vma->vm_start - address) >> PAGE_SHIFT; expand_downwards() 2261 vma->vm_start = address; expand_downwards() 2295 if (next && next->vm_start == address + PAGE_SIZE) { expand_stack() 2309 if (vma && (vma->vm_start <= addr)) find_extend_vma() 2341 if (vma->vm_start <= addr) find_extend_vma() 2345 start = vma->vm_start; find_extend_vma() 2397 next ? next->vm_start : USER_PGTABLES_CEILING); unmap_region() 2419 } while (vma && vma->vm_start < end); detach_vmas_to_be_unmapped() 2458 new->vm_start = addr; __split_vma() 2459 new->vm_pgoff += ((addr - vma->vm_start) >> PAGE_SHIFT); __split_vma() 2478 ((addr - new->vm_start) >> PAGE_SHIFT), new); __split_vma() 2480 err = vma_adjust(vma, vma->vm_start, addr, vma->vm_pgoff, new); __split_vma() 2539 if (vma->vm_start >= end) do_munmap() 2549 if (start > vma->vm_start) { do_munmap() 2568 if (last && end > last->vm_start) { do_munmap() 2580 while (tmp && tmp->vm_start < end) { do_munmap() 2657 if (start < vma->vm_start) SYSCALL_DEFINE5() 2665 if (next->vm_start != next->vm_prev->vm_end) SYSCALL_DEFINE5() 2693 for (tmp = vma; tmp->vm_start >= start + size; SYSCALL_DEFINE5() 2696 max(tmp->vm_start, start), SYSCALL_DEFINE5() 2794 vma->vm_start = addr; do_brk() 2895 vma->vm_pgoff = vma->vm_start >> PAGE_SHIFT; insert_vm_struct() 2897 if (find_vma_links(mm, vma->vm_start, vma->vm_end, insert_vm_struct() 2917 unsigned long vma_start = vma->vm_start; copy_vma() 2940 if (unlikely(vma_start >= new_vma->vm_start && copy_vma() 2962 new_vma->vm_start = addr; copy_vma() 3077 vma->vm_start = addr; __install_special_mapping()
|
H A D | vmacache.c | 102 if (vma->vm_start <= addr && vma->vm_end > addr) { vmacache_find() 126 if (vma && vma->vm_start == start && vma->vm_end == end) { vmacache_find_exact()
|
H A D | pagewalk.c | 259 } else if (start < vma->vm_start) { /* outside vma */ walk_page_range() 261 next = min(end, vma->vm_start); walk_page_range() 298 err = walk_page_test(vma->vm_start, vma->vm_end, walk); walk_page_vma() 303 return __walk_page_range(vma->vm_start, vma->vm_end, walk); walk_page_vma()
|
H A D | interval_tree.c | 21 return v->vm_pgoff + ((v->vm_end - v->vm_start) >> PAGE_SHIFT) - 1; vma_last_pgoff()
|
H A D | mlock.c | 511 pgoff = vma->vm_pgoff + ((start - vma->vm_start) >> PAGE_SHIFT); mlock_fixup() 519 if (start != vma->vm_start) { mlock_fixup() 570 if (!vma || vma->vm_start > start) do_mlock() 574 if (start > vma->vm_start) do_mlock() 580 /* Here we know that vma->vm_start <= nstart < vma->vm_end. */ do_mlock() 599 if (!vma || vma->vm_start != nstart) { do_mlock() 676 mlock_fixup(vma, &prev, vma->vm_start, vma->vm_end, newflags); do_mlockall()
|
H A D | memory.c | 533 unsigned long addr = vma->vm_start; free_pgtables() 544 floor, next? next->vm_start: ceiling); free_pgtables() 549 while (next && next->vm_start <= vma->vm_end + PMD_SIZE free_pgtables() 557 floor, next? next->vm_start: ceiling); free_pgtables() 723 * pfn_of_page == vma->vm_pgoff + ((addr - vma->vm_start) >> PAGE_SHIFT) 775 off = (addr - vma->vm_start) >> PAGE_SHIFT; vm_normal_page() 1010 unsigned long addr = vma->vm_start; copy_page_range() 1199 pr_err("%s: mmap_sem is unlocked! addr=0x%lx end=0x%lx vma->vm_start=0x%lx vma->vm_end=0x%lx\n", zap_pmd_range() 1201 vma->vm_start, zap_pmd_range() 1276 unsigned long start = max(vma->vm_start, start_addr); unmap_single_vma() 1282 if (end <= vma->vm_start) unmap_single_vma() 1339 for ( ; vma && vma->vm_start < end_addr; vma = vma->vm_next) unmap_vmas() 1364 for ( ; vma && vma->vm_start < end; vma = vma->vm_next) zap_page_range() 1410 if (address < vma->vm_start || address + size > vma->vm_end || zap_vma_ptes() 1505 if (addr < vma->vm_start || addr >= vma->vm_end) vm_insert_page() 1580 if (addr < vma->vm_start || addr >= vma->vm_end) vm_insert_pfn() 1596 if (addr < vma->vm_start || addr >= vma->vm_end) vm_insert_mixed() 1721 if (addr != vma->vm_start || end != vma->vm_end) remap_pfn_range() 1789 vm_len = vma->vm_end - vma->vm_start; vm_iomap_memory() 1794 return io_remap_pfn_range(vma, vma->vm_start, pfn, vm_len, vma->vm_page_prot); vm_iomap_memory() 2383 ((zba - vba) << PAGE_SHIFT) + vma->vm_start, unmap_mapping_range_tree() 2384 ((zea - vba + 1) << PAGE_SHIFT) + vma->vm_start, unmap_mapping_range_tree() 2630 if ((vma->vm_flags & VM_GROWSDOWN) && address == vma->vm_start) { check_stack_guard_page() 2648 if (next && next->vm_start == address + PAGE_SIZE) check_stack_guard_page() 2889 start_addr = max(address & mask, vma->vm_start); do_fault_around() 3101 - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff; do_fault() 3647 if (!vma || vma->vm_start > addr) __access_remote_vm() 3747 vma->vm_start, print_vma_addr() 3748 vma->vm_end - vma->vm_start); print_vma_addr()
|
H A D | debug.c | 160 vma, (void *)vma->vm_start, (void *)vma->vm_end, vma->vm_next, dump_vma()
|
H A D | mempolicy.c | 599 if (vma->vm_start > start) queue_pages_test_walk() 600 start = vma->vm_start; queue_pages_test_walk() 605 if (qp->prev && qp->prev->vm_end < vma->vm_start) queue_pages_test_walk() 670 vma->vm_start, vma->vm_end, vma->vm_pgoff, vma_replace_policy() 707 if (!vma || vma->vm_start > start) mbind_range() 711 if (start > vma->vm_start) mbind_range() 714 for (; vma && vma->vm_start < end; prev = vma, vma = next) { mbind_range() 716 vmstart = max(start, vma->vm_start); mbind_range() 723 ((vmstart - vma->vm_start) >> PAGE_SHIFT); mbind_range() 735 if (vma->vm_start != vmstart) { mbind_range() 972 queue_pages_range(mm, mm->mmap->vm_start, mm->task_size, &nmask, migrate_to_node() 1599 pol = vma->vm_ops->get_policy(vma, vma->vm_start); vma_policy_mof() 1772 off += (addr - vma->vm_start) >> shift; interleave_nid() 2270 BUG_ON(addr < vma->vm_start); mpol_misplaced() 2273 pgoff += (addr - vma->vm_start) >> PAGE_SHIFT; mpol_misplaced()
|
H A D | rmap.c | 570 return vma->vm_start + ((pgoff - vma->vm_pgoff) << PAGE_SHIFT); __vma_address() 579 VM_BUG_ON_VMA(address < vma->vm_start || address >= vma->vm_end, vma); vma_address() 606 if (unlikely(address < vma->vm_start || address >= vma->vm_end)) page_address_in_vma() 704 if (unlikely(address < vma->vm_start || address >= vma->vm_end)) page_mapped_in_vma() 1076 VM_BUG_ON_VMA(address < vma->vm_start || address >= vma->vm_end, vma); page_add_new_anon_rmap() 1595 BUG_ON(address < vma->vm_start || address >= vma->vm_end); hugepage_add_new_anon_rmap()
|
H A D | huge_memory.c | 797 if (haddr < vma->vm_start || haddr + HPAGE_PMD_SIZE > vma->vm_end) do_huge_pmd_anonymous_page() 2067 hstart = (vma->vm_start + ~HPAGE_PMD_MASK) & HPAGE_PMD_MASK; khugepaged_enter_vma_merge() 2471 hstart = (vma->vm_start + ~HPAGE_PMD_MASK) & HPAGE_PMD_MASK; collapse_huge_page() 2705 hstart = (vma->vm_start + ~HPAGE_PMD_MASK) & HPAGE_PMD_MASK; 2892 BUG_ON(vma->vm_start > haddr || vma->vm_end < haddr + HPAGE_PMD_SIZE); __split_huge_page_pmd() 2977 (start & HPAGE_PMD_MASK) >= vma->vm_start && __vma_adjust_trans_huge() 2987 (end & HPAGE_PMD_MASK) >= vma->vm_start && __vma_adjust_trans_huge() 2992 * If we're also updating the vma->vm_next->vm_start, if the new __vma_adjust_trans_huge() 2993 * vm_next->vm_start isn't page aligned and it could previously __vma_adjust_trans_huge() 2998 unsigned long nstart = next->vm_start; __vma_adjust_trans_huge() 3001 (nstart & HPAGE_PMD_MASK) >= next->vm_start && __vma_adjust_trans_huge()
|
H A D | mincore.c | 186 if (!vma || addr < vma->vm_start) do_mincore()
|
H A D | gup.c | 552 if (!vma || address < vma->vm_start) fixup_user_fault() 849 VM_BUG_ON_VMA(start < vma->vm_start, vma); populate_vma_page_range() 907 if (!vma || vma->vm_start >= end) __mm_populate() 916 if (nstart < vma->vm_start) __mm_populate() 917 nstart = vma->vm_start; __mm_populate()
|
/linux-4.1.27/arch/powerpc/mm/ |
H A D | dma-noncoherent.c | 76 * .vm_start = VMALLOC_START, 80 * However, vmalloc_head.vm_start is variable (typically, it is dependent on 86 unsigned long vm_start; member in struct:ppc_vm_region 92 .vm_start = CONSISTENT_BASE, 99 unsigned long addr = head->vm_start, end = head->vm_end - size; ppc_vm_region_alloc() 112 if ((addr + size) <= c->vm_start) ppc_vm_region_alloc() 124 new->vm_start = addr; ppc_vm_region_alloc() 142 if (c->vm_start == addr) ppc_vm_region_find() 218 unsigned long vaddr = c->vm_start; __dma_alloc_coherent() 244 return (void *)c->vm_start; __dma_alloc_coherent() 270 if ((c->vm_end - c->vm_start) != size) { __dma_free_coherent() 272 __func__, c->vm_end - c->vm_start, size); __dma_free_coherent() 274 size = c->vm_end - c->vm_start; __dma_free_coherent() 277 addr = c->vm_start; __dma_free_coherent() 297 flush_tlb_kernel_range(c->vm_start, c->vm_end); __dma_free_coherent()
|
H A D | copro_fault.c | 55 if (ea < vma->vm_start) { copro_handle_mm_fault()
|
H A D | subpage-prot.c | 159 if (vma && ((addr + len) <= vma->vm_start)) subpage_mark_vma_nohuge() 163 if (vma->vm_start >= (addr + len)) subpage_mark_vma_nohuge()
|
H A D | tlb_hash32.c | 150 flush_range(mp->vm_mm, mp->vm_start, mp->vm_end); flush_tlb_mm()
|
/linux-4.1.27/arch/mips/include/asm/ |
H A D | tlb.h | 11 flush_cache_range(vma, vma->vm_start, vma->vm_end); \
|
/linux-4.1.27/drivers/gpu/drm/ |
H A D | drm_vm.c | 67 if (efi_range_is_wc(vma->vm_start, vma->vm_end - drm_io_prot() 68 vma->vm_start)) drm_io_prot() 128 vma->vm_start; drm_do_vm_fault() 198 offset = (unsigned long)vmf->virtual_address - vma->vm_start; drm_do_vm_shm_fault() 228 vma->vm_start, vma->vm_end - vma->vm_start); drm_vm_shm_close() 304 offset = (unsigned long)vmf->virtual_address - vma->vm_start; /* vm_[pg]off[set] should be 0 */ drm_do_vm_dma_fault() 340 offset = (unsigned long)vmf->virtual_address - vma->vm_start; drm_do_vm_sg_fault() 412 vma->vm_start, vma->vm_end - vma->vm_start); drm_vm_open_locked() 438 vma->vm_start, vma->vm_end - vma->vm_start); drm_vm_close_locked() 482 unsigned long length = vma->vm_end - vma->vm_start; drm_mmap_dma() 487 vma->vm_start, vma->vm_end, vma->vm_pgoff); drm_mmap_dma() 549 vma->vm_start, vma->vm_end, vma->vm_pgoff); drm_mmap_locked() 576 if (map->size < vma->vm_end - vma->vm_start) drm_mmap_locked() 615 if (io_remap_pfn_range(vma, vma->vm_start, drm_mmap_locked() 617 vma->vm_end - vma->vm_start, drm_mmap_locked() 623 vma->vm_start, vma->vm_end, (unsigned long long)(map->offset + offset)); drm_mmap_locked() 630 if (remap_pfn_range(vma, vma->vm_start, drm_mmap_locked() 632 vma->vm_end - vma->vm_start, vma->vm_page_prot)) drm_mmap_locked() 708 (void *)vma->vm_start, (void *)vma->vm_end, drm_vma_info()
|
/linux-4.1.27/arch/arc/kernel/ |
H A D | arc_hostlink.c | 25 if (io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, arc_hl_mmap() 26 vma->vm_end - vma->vm_start, arc_hl_mmap()
|
H A D | troubleshoot.c | 99 if (vma && (vma->vm_start <= address)) { show_faulting_vma() 110 vma->vm_start < TASK_UNMAPPED_BASE ? show_faulting_vma() 111 address : address - vma->vm_start, show_faulting_vma() 112 nm, vma->vm_start, vma->vm_end); show_faulting_vma()
|
/linux-4.1.27/arch/nios2/include/asm/ |
H A D | tlb.h | 25 flush_cache_range(vma, vma->vm_start, vma->vm_end); \
|
/linux-4.1.27/arch/nios2/kernel/ |
H A D | sys_nios2.c | 42 if (vma == NULL || addr < vma->vm_start || addr + len > vma->vm_end) sys_cacheflush()
|
/linux-4.1.27/arch/um/drivers/ |
H A D | mmapper_kern.c | 56 size = vma->vm_end - vma->vm_start; mmapper_mmap() 64 if (remap_pfn_range(vma, vma->vm_start, p_buf >> PAGE_SHIFT, size, mmapper_mmap()
|
/linux-4.1.27/arch/powerpc/kernel/ |
H A D | proc_powerpc.c | 46 if ((vma->vm_end - vma->vm_start) > PAGE_SIZE) page_map_mmap() 49 remap_pfn_range(vma, vma->vm_start, page_map_mmap()
|
H A D | dma.c | 122 return remap_pfn_range(vma, vma->vm_start, dma_direct_mmap_coherent() 124 vma->vm_end - vma->vm_start, dma_direct_mmap_coherent()
|
H A D | vdso.c | 168 follow_page(vma, vma->vm_start + i*PAGE_SIZE, 0) dump_vdso_pages() 179 follow_page(vma, vma->vm_start + i*PAGE_SIZE, 0) dump_vdso_pages() 288 if (vma->vm_mm && vma->vm_start == vma->vm_mm->context.vdso_base) arch_vma_name()
|
H A D | pci-common.c | 439 * is described by vm_start and vm_end members of VMA, the base physical 463 ret = remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_page_range() 464 vma->vm_end - vma->vm_start, vma->vm_page_prot); pci_mmap_page_range() 562 resource_size_t size = vma->vm_end - vma->vm_start; pci_mmap_legacy_page_range() 602 return remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_legacy_page_range() 603 vma->vm_end - vma->vm_start, pci_mmap_legacy_page_range()
|
/linux-4.1.27/arch/cris/arch-v32/drivers/pci/ |
H A D | bios.c | 31 if (remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_page_range() 32 vma->vm_end - vma->vm_start, pci_mmap_page_range()
|
/linux-4.1.27/fs/proc/ |
H A D | task_nommu.c | 34 size += region->vm_end - region->vm_start; task_mem() 36 size = vma->vm_end - vma->vm_start; task_mem() 89 vsize += vma->vm_end - vma->vm_start; task_vsize() 111 size += region->vm_end - region->vm_start; task_statm() 172 vma->vm_start, nommu_vma_show() 193 if (!is_pid || (vma->vm_start <= mm->start_stack && nommu_vma_show()
|
H A D | nommu.c | 56 region->vm_start, nommu_region_show()
|
H A D | task_mmu.c | 137 m->version = m_next_vma(m->private, vma) ? vma->vm_start : -1UL; m_cache_vma() 173 m->version = vma->vm_start; m_start() 289 start = vma->vm_start; show_map_vma() 332 if (vma->vm_start <= mm->brk && show_map_vma() 344 if (!is_pid || (vma->vm_start <= mm->start_stack && show_map_vma() 644 (vma->vm_end - vma->vm_start) >> 10, show_smap() 1019 hole_end = min(end, vma->vm_start); pagemap_pte_hole() 1500 pol = __get_vma_policy(vma, vma->vm_start); show_numa_map() 1508 seq_printf(m, "%08lx %s", vma->vm_start, buffer); show_numa_map() 1513 } else if (vma->vm_start <= mm->brk && vma->vm_end >= mm->start_brk) { show_numa_map() 1522 if (!is_pid || (vma->vm_start <= mm->start_stack && show_numa_map()
|
H A D | vmcore.c | 409 size_t size = vma->vm_end - vma->vm_start; mmap_vmcore() 433 if (remap_pfn_range(vma, vma->vm_start, pfn, tsz, mmap_vmcore() 449 if (remap_vmalloc_range_partial(vma, vma->vm_start + len, mmap_vmcore() 466 if (vmcore_remap_oldmem_pfn(vma, vma->vm_start + len, mmap_vmcore() 481 do_munmap(vma->vm_mm, vma->vm_start, len); mmap_vmcore()
|
H A D | base.c | 1659 unsigned long vm_start, vm_end; map_files_d_revalidate() local 1684 if (!dname_to_vma_addr(dentry, &vm_start, &vm_end)) { map_files_d_revalidate() 1686 exact_vma_exists = !!find_exact_vma(mm, vm_start, vm_end); map_files_d_revalidate() 1721 unsigned long vm_start, vm_end; proc_map_files_get_link() local 1737 rc = dname_to_vma_addr(dentry, &vm_start, &vm_end); proc_map_files_get_link() 1743 vma = find_exact_vma(mm, vm_start, vm_end); proc_map_files_get_link() 1796 unsigned long vm_start, vm_end; proc_map_files_lookup() local 1816 if (dname_to_vma_addr(dentry, &vm_start, &vm_end)) proc_map_files_lookup() 1824 vma = find_exact_vma(mm, vm_start, vm_end); proc_map_files_lookup() 1920 vma->vm_start, vma->vm_end); proc_map_files_readdir()
|
/linux-4.1.27/drivers/xen/ |
H A D | gntdev.c | 241 unsigned int pgnr = (addr - map->vma->vm_start) >> PAGE_SHIFT; find_grant_ptes() 452 if (map->vma->vm_start >= end) unmap_if_in_range() 456 mstart = max(start, map->vma->vm_start); unmap_if_in_range() 460 map->vma->vm_start, map->vma->vm_end, unmap_if_in_range() 463 (mstart - map->vma->vm_start) >> PAGE_SHIFT, unmap_if_in_range() 505 map->vma->vm_start, map->vma->vm_end); mn_release() 514 map->vma->vm_start, map->vma->vm_end); mn_release() 782 int count = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; gntdev_mmap() 790 index, count, vma->vm_start, vma->vm_pgoff); gntdev_mmap() 830 err = apply_to_page_range(vma->vm_mm, vma->vm_start, gntdev_mmap() 831 vma->vm_end - vma->vm_start, gntdev_mmap() 845 err = vm_insert_page(vma, vma->vm_start + i*PAGE_SIZE, gntdev_mmap() 862 apply_to_page_range(vma->vm_mm, vma->vm_start, gntdev_mmap() 863 vma->vm_end - vma->vm_start, gntdev_mmap() 867 map->pages_vm_start = vma->vm_start; gntdev_mmap()
|
H A D | privcmd.c | 265 if (!vma || (msg->va != vma->vm_start) || vma->vm_private_data) privcmd_ioctl_mmap() 270 state.va = vma->vm_start; privcmd_ioctl_mmap() 491 if (m.addr != vma->vm_start || privcmd_ioctl_mmap_batch() 503 if (m.addr < vma->vm_start || privcmd_ioctl_mmap_batch() 584 int numpgs = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; privcmd_close() 602 vma, vma->vm_start, vma->vm_end, privcmd_fault()
|
H A D | gntalloc.c | 506 int count = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; gntalloc_mmap() 544 rv = vm_insert_page(vma, vma->vm_start + i * PAGE_SIZE, gntalloc_mmap()
|
/linux-4.1.27/drivers/media/v4l2-core/ |
H A D | videobuf-vmalloc.c | 59 map->count, vma->vm_start, vma->vm_end); videobuf_vm_open() 71 map->count, vma->vm_start, vma->vm_end); videobuf_vm_close() 250 buf->baddr = vma->vm_start; __videobuf_mmap_mapper() 256 pages = PAGE_ALIGN(vma->vm_end - vma->vm_start); __videobuf_mmap_mapper() 277 map, q, vma->vm_start, vma->vm_end, __videobuf_mmap_mapper()
|
H A D | videobuf-dma-contig.c | 71 map, map->count, vma->vm_start, vma->vm_end); videobuf_vm_open() 83 map, map->count, vma->vm_start, vma->vm_end); videobuf_vm_close() 295 buf->baddr = vma->vm_start; __videobuf_mmap_mapper() 306 size = vma->vm_end - vma->vm_start; __videobuf_mmap_mapper() 331 map, q, vma->vm_start, vma->vm_end, __videobuf_mmap_mapper()
|
H A D | videobuf2-memops.c | 151 __func__, h, atomic_read(h->refcount), vma->vm_start, vb2_common_vm_open() 169 __func__, h, atomic_read(h->refcount), vma->vm_start, vb2_common_vm_close()
|
H A D | videobuf-dma-sg.c | 389 map->count, vma->vm_start, vma->vm_end); videobuf_vm_open() 402 map->count, vma->vm_start, vma->vm_end); videobuf_vm_close() 441 vma->vm_start, vma->vm_end); videobuf_vm_fault() 629 q->bufs[i]->baddr = vma->vm_start + size; __videobuf_mmap_mapper() 640 map, q, vma->vm_start, vma->vm_end, vma->vm_pgoff, first, last); __videobuf_mmap_mapper()
|
H A D | videobuf2-dma-sg.c | 397 unsigned long uaddr = vma->vm_start; vb2_dma_sg_mmap() 398 unsigned long usize = vma->vm_end - vma->vm_start; vb2_dma_sg_mmap()
|
/linux-4.1.27/drivers/misc/mic/host/ |
H A D | mic_fops.c | 196 unsigned long pa, size = vma->vm_end - vma->vm_start, size_rem = size; mic_mmap() 210 err = remap_pfn_range(vma, vma->vm_start + offset, mic_mmap() 217 pa, vma->vm_start + offset); mic_mmap()
|
/linux-4.1.27/drivers/sbus/char/ |
H A D | flash.c | 68 if (vma->vm_end - (vma->vm_start + (vma->vm_pgoff << PAGE_SHIFT)) > size) flash_mmap() 69 size = vma->vm_end - (vma->vm_start + (vma->vm_pgoff << PAGE_SHIFT)); flash_mmap() 73 if (io_remap_pfn_range(vma, vma->vm_start, addr, size, vma->vm_page_prot)) flash_mmap()
|
/linux-4.1.27/arch/microblaze/kernel/ |
H A D | dma.c | 162 unsigned long user_count = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; dma_direct_mmap_coherent() 176 return remap_pfn_range(vma, vma->vm_start, pfn + off, dma_direct_mmap_coherent() 177 vma->vm_end - vma->vm_start, vma->vm_page_prot); dma_direct_mmap_coherent()
|
/linux-4.1.27/arch/parisc/mm/ |
H A D | fault.c | 126 if (tree->vm_start > addr) { 132 if (prev->vm_next->vm_start > addr) 195 pr_warn(" vm_start = 0x%08lx, vm_end = 0x%08lx\n", show_signal_msg() 196 vma->vm_start, vma->vm_end); show_signal_msg() 229 if (!vma || address < vma->vm_start) do_page_fault() 306 address < vma->vm_start || address > vma->vm_end) { do_page_fault()
|
/linux-4.1.27/drivers/staging/unisys/visorchipset/ |
H A D | file.c | 85 if (remap_pfn_range(vma, vma->vm_start, visorchipset_mmap() 87 vma->vm_end - vma->vm_start, visorchipset_mmap()
|
/linux-4.1.27/arch/sh/mm/ |
H A D | mmap.c | 66 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area() 116 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area_topdown()
|
H A D | fault.c | 456 if (likely(vma->vm_start <= address)) do_page_fault()
|
/linux-4.1.27/arch/tile/mm/ |
H A D | elf.c | 78 if (vma->vm_start == (ELF_ET_DYN_BASE & PAGE_MASK)) { notify_exec() 82 snprintf(buf, sizeof(buf), "0x%lx:@", vma->vm_start); notify_exec()
|
H A D | hugetlbpage.c | 240 (!vma || addr + len <= vma->vm_start)) hugetlb_get_unmapped_area()
|
/linux-4.1.27/drivers/xen/xenbus/ |
H A D | xenbus_dev_backend.c | 95 size_t size = vma->vm_end - vma->vm_start; xenbus_backend_mmap() 103 if (remap_pfn_range(vma, vma->vm_start, xenbus_backend_mmap()
|
/linux-4.1.27/arch/sh/kernel/vsyscall/ |
H A D | vsyscall.c | 90 if (vma->vm_mm && vma->vm_start == (long)vma->vm_mm->context.vdso) arch_vma_name()
|
/linux-4.1.27/arch/alpha/kernel/ |
H A D | pci-sysfs.c | 30 return io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, hose_mmap_page_range() 31 vma->vm_end - vma->vm_start, hose_mmap_page_range() 41 nr = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; __pci_mmap_fits() 258 nr = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; __legacy_mmap_fits()
|
/linux-4.1.27/arch/arc/mm/ |
H A D | mmap.c | 67 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area()
|
H A D | fault.c | 99 if (vma->vm_start <= address) do_page_fault()
|
/linux-4.1.27/arch/tile/kernel/ |
H A D | vdso.c | 115 if (vma->vm_mm && vma->vm_start == VDSO_BASE) arch_vma_name() 118 if (vma->vm_start == MEM_USER_INTRPT) arch_vma_name()
|
H A D | stack.c | 331 if (vma == NULL || address < vma->vm_start) { describe_addr() 350 vma->vm_start, vma->vm_end - vma->vm_start); describe_addr()
|
/linux-4.1.27/drivers/char/ |
H A D | uv_mmtimer.c | 151 if (vma->vm_end - vma->vm_start != PAGE_SIZE) uv_mmtimer_mmap() 166 if (remap_pfn_range(vma, vma->vm_start, uv_mmtimer_addr >> PAGE_SHIFT, uv_mmtimer_mmap()
|
H A D | mspec.c | 97 unsigned long vm_start; /* Original (unsplit) base. */ member in struct:vma_data 171 last_index = (vdata->vm_end - vdata->vm_start) >> PAGE_SHIFT; mspec_close() 281 vdata->vm_start = vma->vm_start; mspec_mmap()
|
H A D | bsr.c | 127 unsigned long size = vma->vm_end - vma->vm_start; bsr_mmap() 135 ret = remap_4k_pfn(vma, vma->vm_start, dev->bsr_addr >> 12, bsr_mmap() 138 ret = io_remap_pfn_range(vma, vma->vm_start, bsr_mmap()
|
H A D | mem.c | 323 size_t size = vma->vm_end - vma->vm_start; mmap_mem() 346 vma->vm_start, mmap_mem()
|
H A D | mmtimer.c | 441 if (vma->vm_end - vma->vm_start != PAGE_SIZE) mmtimer_mmap() 456 if (remap_pfn_range(vma, vma->vm_start, mmtimer_addr >> PAGE_SHIFT, mmtimer_mmap()
|
/linux-4.1.27/arch/arm/mm/ |
H A D | mmap.c | 92 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area() 143 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area_topdown()
|
H A D | fault-armv.c | 141 pgoff = vma->vm_pgoff + ((addr - vma->vm_start) >> PAGE_SHIFT); make_coherent() 160 aliases += adjust_pte(mpnt, mpnt->vm_start + offset, pfn); make_coherent()
|
H A D | flush.c | 252 flush_cache_page(mpnt, mpnt->vm_start + offset, page_to_pfn(page)); __flush_dcache_aliases()
|
/linux-4.1.27/sound/soc/pxa/ |
H A D | mmp-pcm.c | 129 return remap_pfn_range(vma, vma->vm_start, mmp_pcm_mmap() 131 vma->vm_end - vma->vm_start, vma->vm_page_prot); mmp_pcm_mmap()
|
/linux-4.1.27/arch/xtensa/kernel/ |
H A D | syscall.c | 90 if (!vmm || addr + len <= vmm->vm_start) arch_get_unmapped_area()
|
H A D | pci.c | 358 * is described by vm_start and vm_end members of VMA, the base physical 377 ret = io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_page_range() 378 vma->vm_end - vma->vm_start,vma->vm_page_prot); pci_mmap_page_range()
|
/linux-4.1.27/arch/sh/kernel/ |
H A D | sys_sh.c | 73 if (vma == NULL || addr < vma->vm_start || addr + len > vma->vm_end) { sys_cacheflush()
|
/linux-4.1.27/drivers/vfio/platform/ |
H A D | vfio_platform_common.c | 422 req_len = vma->vm_end - vma->vm_start; vfio_platform_mmap_mmio() 433 return remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, vfio_platform_mmap_mmio() 444 if (vma->vm_end < vma->vm_start) vfio_platform_mmap() 450 if (vma->vm_start & ~PAGE_MASK) vfio_platform_mmap()
|
/linux-4.1.27/arch/mips/kernel/ |
H A D | vdso.c | 119 if (vma->vm_mm && vma->vm_start == (long)vma->vm_mm->context.vdso) arch_vma_name()
|
/linux-4.1.27/arch/hexagon/kernel/ |
H A D | vdso.c | 97 if (vma->vm_mm && vma->vm_start == (long)vma->vm_mm->context.vdso) arch_vma_name()
|
/linux-4.1.27/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_doorbell.c | 129 if (vma->vm_end - vma->vm_start != doorbell_process_allocation()) kfd_doorbell_mmap() 150 (unsigned long long) vma->vm_start, address, vma->vm_flags, kfd_doorbell_mmap() 155 vma->vm_start, kfd_doorbell_mmap()
|
/linux-4.1.27/arch/metag/mm/ |
H A D | hugetlbpage.c | 50 if (addr + len > vma->vm_start) prepare_hugepage_range() 53 (ALIGN_HUGEPT(addr + len) > vma->vm_start)) prepare_hugepage_range() 162 <= vma->vm_start)) { hugetlb_get_unmapped_area_existing()
|
H A D | fault.c | 118 if (!vma || address < vma->vm_start) do_page_fault()
|
/linux-4.1.27/arch/ia64/mm/ |
H A D | fault.c | 138 if (( !vma && prev_vma ) || (address < vma->vm_start) ) ia64_do_page_fault() 211 if (REGION_NUMBER(address) != REGION_NUMBER(vma->vm_start) ia64_do_page_fault() 218 if (REGION_NUMBER(address) != REGION_NUMBER(vma->vm_start) ia64_do_page_fault()
|
H A D | init.c | 120 vma->vm_start = current->thread.rbs_bot & PAGE_MASK; ia64_init_addr_space() 121 vma->vm_end = vma->vm_start + PAGE_SIZE; ia64_init_addr_space() 286 gate_vma.vm_start = FIXADDR_USER_START; gate_vma_init()
|
/linux-4.1.27/sound/soc/blackfin/ |
H A D | bf5xx-i2s-pcm.c | 220 size_t size = vma->vm_end - vma->vm_start; bf5xx_pcm_mmap() 221 vma->vm_start = (unsigned long)runtime->dma_area; bf5xx_pcm_mmap() 222 vma->vm_end = vma->vm_start + size; bf5xx_pcm_mmap()
|
H A D | bf5xx-ac97-pcm.c | 275 size_t size = vma->vm_end - vma->vm_start; bf5xx_pcm_mmap() 276 vma->vm_start = (unsigned long)runtime->dma_area; bf5xx_pcm_mmap() 277 vma->vm_end = vma->vm_start + size; bf5xx_pcm_mmap()
|
/linux-4.1.27/arch/parisc/kernel/ |
H A D | cache.c | 314 addr = mpnt->vm_start + offset; flush_dcache_page() 495 usize += vma->vm_end - vma->vm_start; mm_total_size() 528 flush_user_dcache_range_asm(vma->vm_start, vma->vm_end); flush_cache_mm() 531 flush_user_icache_range_asm(vma->vm_start, vma->vm_end); flush_cache_mm() 540 for (addr = vma->vm_start; addr < vma->vm_end; flush_cache_mm()
|
H A D | pci.c | 246 return remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_page_range() 247 vma->vm_end - vma->vm_start, vma->vm_page_prot); pci_mmap_page_range()
|
H A D | sys_parisc.c | 120 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area() 180 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area_topdown()
|
/linux-4.1.27/arch/ia64/pci/ |
H A D | pci.c | 593 unsigned long size = vma->vm_end - vma->vm_start; pci_mmap_page_range() 623 efi_range_is_wc(vma->vm_start, vma->vm_end - vma->vm_start)) pci_mmap_page_range() 628 if (remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_page_range() 629 vma->vm_end - vma->vm_start, vma->vm_page_prot)) pci_mmap_page_range() 664 unsigned long size = vma->vm_end - vma->vm_start; pci_mmap_legacy_page_range() 688 if (remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_legacy_page_range()
|
/linux-4.1.27/arch/arm64/mm/ |
H A D | dma-mapping.c | 311 unsigned long nr_vma_pages = (vma->vm_end - vma->vm_start) >> __dma_common_mmap() 321 ret = remap_pfn_range(vma, vma->vm_start, __dma_common_mmap() 323 vma->vm_end - vma->vm_start, __dma_common_mmap()
|
/linux-4.1.27/fs/ |
H A D | binfmt_elf_fdpic.c | 1197 kdcore("%08lx: %08lx: no (IO)", vma->vm_start, vma->vm_flags); maydump() 1205 kdcore("%08lx: %08lx: no (!read)", vma->vm_start, vma->vm_flags); maydump() 1213 kdcore("%08lx: %08lx: %s (share)", vma->vm_start, maydump() 1219 kdcore("%08lx: %08lx: %s (share)", vma->vm_start, maydump() 1228 kdcore("%08lx: %08lx: %s (!anon)", vma->vm_start, maydump() 1235 kdcore("%08lx: %08lx: %s", vma->vm_start, vma->vm_flags, maydump() 1488 for (addr = vma->vm_start; addr < vma->vm_end; elf_fdpic_dump_segments() 1504 if (!dump_emit(cprm, (void *) vma->vm_start, elf_fdpic_dump_segments() 1505 vma->vm_end - vma->vm_start)) elf_fdpic_dump_segments() 1519 size += vma->vm_end - vma->vm_start; elf_core_vma_data_size() 1712 sz = vma->vm_end - vma->vm_start; elf_fdpic_core_dump() 1716 phdr.p_vaddr = vma->vm_start; elf_fdpic_core_dump()
|
H A D | exec.c | 201 unsigned long size = bprm->vma->vm_end - bprm->vma->vm_start; get_arg_page() 270 vma->vm_start = vma->vm_end - PAGE_SIZE; __bprm_mm_init() 579 unsigned long old_start = vma->vm_start; shift_arg_pages() 616 vma->vm_next ? vma->vm_next->vm_start : USER_PGTABLES_CEILING); shift_arg_pages() 625 vma->vm_next ? vma->vm_next->vm_start : USER_PGTABLES_CEILING); shift_arg_pages() 666 if (vma->vm_end - vma->vm_start > stack_base) setup_arg_pages() 671 stack_shift = vma->vm_start - stack_base; setup_arg_pages() 679 unlikely(vma->vm_end - vma->vm_start >= stack_top - mmap_min_addr)) setup_arg_pages() 707 ret = mprotect_fixup(vma, &prev, vma->vm_start, vma->vm_end, setup_arg_pages() 724 stack_size = vma->vm_end - vma->vm_start; setup_arg_pages() 732 stack_base = vma->vm_start + rlim_stack; setup_arg_pages() 739 stack_base = vma->vm_start - stack_expand; setup_arg_pages()
|
H A D | binfmt_elf.c | 1276 u32 __user *header = (u32 __user *) vma->vm_start; vma_dump_size() 1308 return vma->vm_end - vma->vm_start; vma_dump_size() 1550 *start_end_ofs++ = vma->vm_start; fill_files_note() 2235 phdr.p_vaddr = vma->vm_start; elf_core_dump() 2238 phdr.p_memsz = vma->vm_end - vma->vm_start; elf_core_dump() 2270 end = vma->vm_start + vma_filesz[i++]; elf_core_dump() 2272 for (addr = vma->vm_start; addr < end; addr += PAGE_SIZE) { elf_core_dump()
|
/linux-4.1.27/drivers/auxdisplay/ |
H A D | cfag12864bfb.c | 69 return vm_insert_page(vma, vma->vm_start, cfag12864bfb_mmap()
|
/linux-4.1.27/ipc/ |
H A D | shm.c | 1309 (vma->vm_start - addr)/PAGE_SIZE == vma->vm_pgoff) { SYSCALL_DEFINE1() 1319 do_munmap(mm, vma->vm_start, vma->vm_end - vma->vm_start); SYSCALL_DEFINE1() 1344 ((vma->vm_start - addr)/PAGE_SIZE == vma->vm_pgoff) && SYSCALL_DEFINE1() 1346 do_munmap(mm, vma->vm_start, vma->vm_end - vma->vm_start); SYSCALL_DEFINE1() 1353 if (vma && vma->vm_start == addr && vma->vm_ops == &shm_vm_ops) { SYSCALL_DEFINE1() 1354 do_munmap(mm, vma->vm_start, vma->vm_end - vma->vm_start); SYSCALL_DEFINE1()
|
/linux-4.1.27/arch/sh/include/asm/ |
H A D | tlb.h | 77 flush_cache_range(vma, vma->vm_start, vma->vm_end); tlb_start_vma()
|
/linux-4.1.27/arch/nios2/mm/ |
H A D | init.c | 141 return (vma->vm_start == KUSER_BASE) ? "[kuser]" : NULL; arch_vma_name()
|
H A D | cacheflush.c | 107 flush_cache_page(mpnt, mpnt->vm_start + offset, flush_aliases()
|
H A D | fault.c | 96 if (vma->vm_start <= address) do_page_fault()
|
/linux-4.1.27/arch/frv/mm/ |
H A D | elf-fdpic.c | 77 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area()
|
H A D | fault.c | 92 if (vma->vm_start <= ear0) do_page_fault()
|
/linux-4.1.27/arch/x86/kernel/ |
H A D | sys_x86_64.c | 146 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area() 189 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area_topdown()
|
H A D | vsyscall_64.c | 284 .vm_start = VSYSCALL_ADDR, 309 return (addr >= vma->vm_start) && (addr < vma->vm_end); in_gate_area()
|
/linux-4.1.27/drivers/base/ |
H A D | dma-coherent.c | 255 int user_count = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; dma_mmap_from_coherent() 261 *ret = remap_pfn_range(vma, vma->vm_start, pfn, dma_mmap_from_coherent()
|
H A D | dma-mapping.c | 251 unsigned long user_count = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; dma_common_mmap() 262 ret = remap_pfn_range(vma, vma->vm_start, dma_common_mmap()
|
/linux-4.1.27/drivers/video/fbdev/ |
H A D | sbuslib.c | 54 size = vma->vm_end - vma->vm_start; sbusfb_mmap_helper() 85 vma->vm_start + page, sbusfb_mmap_helper()
|
H A D | amba-clcd.c | 412 if (off <= len && vma->vm_end - vma->vm_start <= len - off && clcdfb_mmap() 735 user_size = vma->vm_end - vma->vm_start; clcdfb_of_vram_mmap() 741 return remap_pfn_range(vma, vma->vm_start, clcdfb_of_vram_mmap()
|
H A D | igafb.c | 232 size = vma->vm_end - vma->vm_start; igafb_mmap() 261 if (remap_pfn_range(vma, vma->vm_start + page, igafb_mmap()
|
H A D | pxa3xx-gcu.c | 484 unsigned int size = vma->vm_end - vma->vm_start; pxa3xx_gcu_mmap() 504 return io_remap_pfn_range(vma, vma->vm_start, pxa3xx_gcu_mmap()
|
H A D | vfb.c | 417 unsigned long start = vma->vm_start; vfb_mmap() 418 unsigned long size = vma->vm_end - vma->vm_start; vfb_mmap()
|
/linux-4.1.27/arch/s390/mm/ |
H A D | mmap.c | 107 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area() 147 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area_topdown()
|
/linux-4.1.27/arch/sh/drivers/pci/ |
H A D | pci.c | 288 return remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_page_range() 289 vma->vm_end - vma->vm_start, pci_mmap_page_range()
|
/linux-4.1.27/arch/mips/pci/ |
H A D | pci.c | 345 return remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_page_range() 346 vma->vm_end - vma->vm_start, vma->vm_page_prot); pci_mmap_page_range()
|
/linux-4.1.27/arch/mn10300/unit-asb2305/ |
H A D | pci-asb2305.c | 231 if (io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_page_range() 232 vma->vm_end - vma->vm_start, pci_mmap_page_range()
|
/linux-4.1.27/sound/usb/usx2y/ |
H A D | usX2Yhwdep.c | 42 area->vm_start, snd_us428ctls_vm_fault() 63 unsigned long size = (unsigned long)(area->vm_end - area->vm_start); snd_us428ctls_mmap()
|
/linux-4.1.27/sound/soc/samsung/ |
H A D | idma.c | 252 size = vma->vm_end - vma->vm_start; idma_mmap() 254 ret = io_remap_pfn_range(vma, vma->vm_start, idma_mmap()
|
/linux-4.1.27/arch/arm/kvm/ |
H A D | mmu.c | 766 hva_t vm_start, vm_end; stage2_unmap_memslot() local 768 if (!vma || vma->vm_start >= reg_end) stage2_unmap_memslot() 774 vm_start = max(hva, vma->vm_start); stage2_unmap_memslot() 778 gpa_t gpa = addr + (vm_start - memslot->userspace_addr); stage2_unmap_memslot() 779 unmap_stage2_range(kvm, gpa, vm_end - vm_start); stage2_unmap_memslot() 1774 hva_t vm_start, vm_end; kvm_arch_prepare_memory_region() local 1776 if (!vma || vma->vm_start >= reg_end) kvm_arch_prepare_memory_region() 1791 vm_start = max(hva, vma->vm_start); kvm_arch_prepare_memory_region() 1796 (vm_start - mem->userspace_addr); kvm_arch_prepare_memory_region() 1800 pa += vm_start - vma->vm_start; kvm_arch_prepare_memory_region() 1807 vm_end - vm_start, kvm_arch_prepare_memory_region()
|
/linux-4.1.27/drivers/infiniband/hw/cxgb4/ |
H A D | provider.c | 164 int len = vma->vm_end - vma->vm_start; c4iw_mmap() 175 if (vma->vm_start & (PAGE_SIZE-1)) c4iw_mmap() 195 ret = io_remap_pfn_range(vma, vma->vm_start, c4iw_mmap() 215 ret = io_remap_pfn_range(vma, vma->vm_start, c4iw_mmap() 223 ret = remap_pfn_range(vma, vma->vm_start, c4iw_mmap()
|
/linux-4.1.27/drivers/infiniband/hw/ipath/ |
H A D | ipath_file_ops.c | 993 if ((vma->vm_end - vma->vm_start) > len) { ipath_mmap_mem() 996 vma->vm_end - vma->vm_start, len); ipath_mmap_mem() 1014 ret = remap_pfn_range(vma, vma->vm_start, pfn, ipath_mmap_mem() 1039 if ((vma->vm_end - vma->vm_start) > PAGE_SIZE) { mmap_ureg() 1041 "%lx > PAGE\n", vma->vm_end - vma->vm_start); mmap_ureg() 1048 ret = io_remap_pfn_range(vma, vma->vm_start, mmap_ureg() 1050 vma->vm_end - vma->vm_start, mmap_ureg() 1070 if ((vma->vm_end - vma->vm_start) > (piocnt * dd->ipath_palign)) { mmap_piobufs() 1073 vma->vm_end - vma->vm_start); mmap_piobufs() 1094 ret = io_remap_pfn_range(vma, vma->vm_start, phys >> PAGE_SHIFT, mmap_piobufs() 1095 vma->vm_end - vma->vm_start, mmap_piobufs() 1112 if ((vma->vm_end - vma->vm_start) > total_size) { mmap_rcvegrbufs() 1115 vma->vm_end - vma->vm_start, mmap_rcvegrbufs() 1130 start = vma->vm_start; mmap_rcvegrbufs() 1224 len = vma->vm_end - vma->vm_start; mmap_kvaddr() 1284 ipath_cdbg(MM, "pgaddr %llx vm_start=%lx len %lx port %u:%u:%u\n", ipath_mmap() 1285 (unsigned long long) pgaddr, vma->vm_start, ipath_mmap() 1286 vma->vm_end - vma->vm_start, dd->ipath_unit, ipath_mmap() 1354 vma->vm_end - vma->vm_start); ipath_mmap()
|
H A D | ipath_mmap.c | 93 unsigned long size = vma->vm_end - vma->vm_start; ipath_mmap()
|
/linux-4.1.27/sound/oss/ |
H A D | soundcard.c | 468 size = vma->vm_end - vma->vm_start; sound_mmap() 473 if (remap_pfn_range(vma, vma->vm_start, sound_mmap() 475 vma->vm_end - vma->vm_start, vma->vm_page_prot)) { sound_mmap()
|
/linux-4.1.27/drivers/android/ |
H A D | binder.c | 2813 proc->pid, vma->vm_start, vma->vm_end, binder_vma_open() 2814 (vma->vm_end - vma->vm_start) / SZ_1K, vma->vm_flags, binder_vma_open() 2824 proc->pid, vma->vm_start, vma->vm_end, binder_vma_close() 2825 (vma->vm_end - vma->vm_start) / SZ_1K, vma->vm_flags, binder_vma_close() 2854 if ((vma->vm_end - vma->vm_start) > SZ_4M) binder_mmap() 2855 vma->vm_end = vma->vm_start + SZ_4M; binder_mmap() 2859 proc->pid, vma->vm_start, vma->vm_end, binder_mmap() 2860 (vma->vm_end - vma->vm_start) / SZ_1K, vma->vm_flags, binder_mmap() 2877 area = get_vm_area(vma->vm_end - vma->vm_start, VM_IOREMAP); binder_mmap() 2884 proc->user_buffer_offset = vma->vm_start - (uintptr_t)proc->buffer; binder_mmap() 2889 while (CACHE_COLOUR((vma->vm_start ^ (uint32_t)proc->buffer))) { binder_mmap() 2890 pr_info("binder_mmap: %d %lx-%lx maps %p bad alignment\n", proc->pid, vma->vm_start, vma->vm_end, proc->buffer); binder_mmap() 2891 vma->vm_start += PAGE_SIZE; binder_mmap() 2895 proc->pages = kzalloc(sizeof(proc->pages[0]) * ((vma->vm_end - vma->vm_start) / PAGE_SIZE), GFP_KERNEL); binder_mmap() 2901 proc->buffer_size = vma->vm_end - vma->vm_start; binder_mmap() 2923 proc->pid, vma->vm_start, vma->vm_end, proc->buffer);*/ binder_mmap() 2938 proc->pid, vma->vm_start, vma->vm_end, failure_string, ret); binder_mmap()
|
/linux-4.1.27/drivers/misc/sgi-gru/ |
H A D | grukdump.c | 155 hdr.vaddr = gts->ts_vma->vm_start; gru_dump_context()
|
H A D | grufile.c | 112 if (vma->vm_start & (GRU_GSEG_PAGESIZE - 1) || gru_file_mmap() 126 file, vma->vm_start, vma, vma->vm_private_data); gru_file_mmap()
|
H A D | grutables.h | 398 #define TSID(a, v) (((a) - (v)->vm_start) / GRU_GSEG_PAGESIZE) 399 #define UGRUADDR(gts) ((gts)->ts_vma->vm_start + \
|
/linux-4.1.27/drivers/infiniband/hw/qib/ |
H A D | qib_file_ops.c | 736 if ((vma->vm_end - vma->vm_start) > len) { qib_mmap_mem() 739 vma->vm_end - vma->vm_start, len); qib_mmap_mem() 761 ret = remap_pfn_range(vma, vma->vm_start, pfn, qib_mmap_mem() 784 if ((vma->vm_end - vma->vm_start) > sz) { mmap_ureg() 787 vma->vm_end - vma->vm_start); mmap_ureg() 794 ret = io_remap_pfn_range(vma, vma->vm_start, mmap_ureg() 796 vma->vm_end - vma->vm_start, mmap_ureg() 816 if ((vma->vm_end - vma->vm_start) > (piocnt * dd->palign)) { mmap_piobufs() 819 vma->vm_end - vma->vm_start); mmap_piobufs() 844 ret = io_remap_pfn_range(vma, vma->vm_start, phys >> PAGE_SHIFT, mmap_piobufs() 845 vma->vm_end - vma->vm_start, mmap_piobufs() 862 if ((vma->vm_end - vma->vm_start) > total_size) { mmap_rcvegrbufs() 865 vma->vm_end - vma->vm_start, mmap_rcvegrbufs() 881 start = vma->vm_start; mmap_rcvegrbufs() 974 len = vma->vm_end - vma->vm_start; mmap_kvaddr() 1100 vma->vm_end - vma->vm_start); qib_mmapf()
|
H A D | qib_mmap.c | 93 unsigned long size = vma->vm_end - vma->vm_start; qib_mmap()
|
/linux-4.1.27/arch/x86/mm/ |
H A D | hugetlbpage.c | 147 (!vma || addr + len <= vma->vm_start)) hugetlb_get_unmapped_area()
|
H A D | mpx.c | 624 * Find the first overlapping vma. If vma->vm_start > start, there zap_bt_entries() 629 if (!vma || vma->vm_start > start) zap_bt_entries() 639 while (vma && vma->vm_start < end) { zap_bt_entries() 653 addr = vma->vm_start; zap_bt_entries() 783 if (next && (mm->bd_addr + MPX_GET_BD_ENTRY_OFFSET(next->vm_start)) unmap_edge_bts() 911 } while (vma && vma->vm_start < end); mpx_notify_unmap()
|
H A D | pat.c | 807 unsigned long vma_size = vma->vm_end - vma->vm_start; track_pfn_copy() 815 if (follow_phys(vma, vma->vm_start, 0, &prot, &paddr)) { track_pfn_copy() 838 if (addr == vma->vm_start && size == (vma->vm_end - vma->vm_start)) { track_pfn_remap() 903 if (follow_phys(vma, vma->vm_start, 0, &prot, &paddr)) { untrack_pfn() 908 size = vma->vm_end - vma->vm_start; untrack_pfn()
|
/linux-4.1.27/arch/sparc/mm/ |
H A D | hugetlbpage.c | 118 (!vma || addr + len <= vma->vm_start)) hugetlb_get_unmapped_area()
|
H A D | fault_32.c | 213 if (vma->vm_start <= address) do_sparc_fault() 398 if (vma->vm_start <= address) force_user_fault()
|
/linux-4.1.27/drivers/char/agp/ |
H A D | alpha-agp.c | 22 dma_addr = (unsigned long)vmf->virtual_address - vma->vm_start alpha_core_agp_vm_fault()
|
H A D | frontend.c | 582 size = vma->vm_end - vma->vm_start; agp_mmap() 603 } else if (io_remap_pfn_range(vma, vma->vm_start, agp_mmap() 620 } else if (io_remap_pfn_range(vma, vma->vm_start, agp_mmap()
|
/linux-4.1.27/drivers/gpu/drm/udl/ |
H A D | udl_gem.c | 110 page_offset = ((unsigned long)vmf->virtual_address - vma->vm_start) >> udl_gem_fault()
|
H A D | udl_fb.c | 257 unsigned long start = vma->vm_start; udl_fb_mmap() 258 unsigned long size = vma->vm_end - vma->vm_start; udl_fb_mmap()
|
/linux-4.1.27/drivers/uio/ |
H A D | uio.c | 646 if (vma->vm_end - vma->vm_start > mem->size) uio_mmap_physical() 662 vma->vm_start, uio_mmap_physical() 664 vma->vm_end - vma->vm_start, uio_mmap_physical() 676 if (vma->vm_end < vma->vm_start) uio_mmap()
|
/linux-4.1.27/arch/mips/mm/ |
H A D | mmap.c | 95 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area_common()
|
H A D | fault.c | 107 if (vma->vm_start <= address) __do_page_fault()
|
/linux-4.1.27/arch/powerpc/kvm/ |
H A D | e500_mmu_host.c | 363 if (vma && hva >= vma->vm_start && kvmppc_e500_shadow_map() 379 ((vma->vm_end - vma->vm_start) >> PAGE_SHIFT); kvmppc_e500_shadow_map() 381 pfn = start + ((hva - vma->vm_start) >> PAGE_SHIFT); kvmppc_e500_shadow_map() 426 } else if (vma && hva >= vma->vm_start && kvmppc_e500_shadow_map()
|
/linux-4.1.27/arch/hexagon/mm/ |
H A D | vm_fault.c | 75 if (vma->vm_start <= address) do_page_fault()
|
/linux-4.1.27/arch/m68k/mm/ |
H A D | fault.c | 97 if (vma->vm_start <= address) do_page_fault()
|
/linux-4.1.27/arch/x86/pci/ |
H A D | i386.c | 447 if (io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_page_range() 448 vma->vm_end - vma->vm_start, pci_mmap_page_range()
|
/linux-4.1.27/arch/unicore32/kernel/ |
H A D | pci.c | 380 if (remap_pfn_range(vma, vma->vm_start, phys, pci_mmap_page_range() 381 vma->vm_end - vma->vm_start, pci_mmap_page_range()
|
H A D | process.c | 327 return (vma->vm_start == 0xffff0000) ? "[vectors]" : NULL; arch_vma_name()
|
/linux-4.1.27/arch/mn10300/mm/ |
H A D | fault.c | 182 if (vma->vm_start <= address) do_page_fault() 197 vma->vm_start, vma->vm_end); do_page_fault()
|
/linux-4.1.27/arch/arm/kernel/ |
H A D | process.c | 300 .vm_start = 0xffff0000, 319 return (addr >= gate_vma.vm_start) && (addr < gate_vma.vm_end); in_gate_area()
|
H A D | bios32.c | 629 if (remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_page_range() 630 vma->vm_end - vma->vm_start, pci_mmap_page_range()
|
/linux-4.1.27/arch/arm/mach-integrator/ |
H A D | impd1.c | 244 size = vma->vm_end - vma->vm_start; impd1fb_clcd_mmap() 246 return remap_pfn_range(vma, vma->vm_start, start, size, impd1fb_clcd_mmap()
|
/linux-4.1.27/sound/soc/atmel/ |
H A D | atmel-pcm-pdc.c | 75 return remap_pfn_range(vma, vma->vm_start, atmel_pcm_mmap() 77 vma->vm_end - vma->vm_start, vma->vm_page_prot); atmel_pcm_mmap()
|
/linux-4.1.27/drivers/misc/genwqe/ |
H A D | card_dev.c | 397 unsigned long vsize = vma->vm_end - vma->vm_start; genwqe_vma_close() 406 dma_map = __genwqe_search_mapping(cfile, vma->vm_start, vsize, genwqe_vma_close() 411 __func__, vma->vm_start, vma->vm_pgoff << PAGE_SHIFT, genwqe_vma_close() 441 unsigned long pfn, vsize = vma->vm_end - vma->vm_start; genwqe_mmap() 457 dma_map->u_vaddr = (void *)vma->vm_start; genwqe_mmap() 472 vma->vm_start, genwqe_mmap()
|
/linux-4.1.27/drivers/firewire/ |
H A D | core-cdev.c | 74 unsigned long vm_start; member in struct:client 1085 payload = (unsigned long)a->data - client->vm_start; ioctl_queue_iso() 1157 a->data = client->vm_start + payload; ioctl_queue_iso() 1687 if (vma->vm_start & ~PAGE_MASK) fw_device_op_mmap() 1690 client->vm_start = vma->vm_start; fw_device_op_mmap() 1691 size = vma->vm_end - vma->vm_start; fw_device_op_mmap()
|
H A D | core-iso.c | 113 uaddr = vma->vm_start; fw_iso_buffer_map_vma()
|
/linux-4.1.27/arch/sparc/kernel/ |
H A D | pci.c | 747 user_size = vma->vm_end - vma->vm_start; __pci_mmap_make_offset_bus() 799 user_size = vma->vm_end - vma->vm_start; __pci_mmap_make_offset() 849 * for this architecture. The region in the process to map is described by vm_start 869 ret = io_remap_pfn_range(vma, vma->vm_start, pci_mmap_page_range() 871 vma->vm_end - vma->vm_start, pci_mmap_page_range()
|
H A D | sys_sparc_64.c | 121 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area() 184 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area_topdown()
|
/linux-4.1.27/drivers/infiniband/hw/usnic/ |
H A D | usnic_ib_verbs.c | 693 if ((vma->vm_end - vma->vm_start) != bar->len) { usnic_ib_mmap() 696 vma->vm_end - vma->vm_start); usnic_ib_mmap() 707 vma->vm_start, usnic_ib_mmap()
|
/linux-4.1.27/drivers/dma-buf/ |
H A D | dma-buf.c | 89 if (vma->vm_pgoff + ((vma->vm_end - vma->vm_start) >> PAGE_SHIFT) > dma_buf_mmap_internal() 671 if (pgoff + ((vma->vm_end - vma->vm_start) >> PAGE_SHIFT) < pgoff) dma_buf_mmap() 675 if (pgoff + ((vma->vm_end - vma->vm_start) >> PAGE_SHIFT) > dma_buf_mmap()
|
/linux-4.1.27/tools/vm/ |
H A D | page-types.c | 784 unsigned long vm_start; parse_pid() local 793 &vm_start, parse_pid() 803 pg_start[nr_vmas] = vm_start / page_size; parse_pid()
|
/linux-4.1.27/include/linux/ |
H A D | mm_types.h | 230 unsigned long vm_start; /* start address of region */ member in struct:vm_region 233 unsigned long vm_pgoff; /* the offset in vm_file corresponding to vm_start */ 250 unsigned long vm_start; /* Our start address within vm_mm. */ member in struct:vm_area_struct
|
H A D | mm.h | 1260 (vma->vm_start == addr) && stack_guard_page_start() 1267 return vma && (vma->vm_start == addr) && (vma->vm_flags & VM_GROWSUP); vma_growsup() 1991 if (vma && end_addr <= vma->vm_start) find_vma_intersection() 1998 return (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; vma_pages() 2001 /* Look up the first VMA which exactly match the interval vm_start ... vm_end */ find_exact_vma() 2003 unsigned long vm_start, unsigned long vm_end) find_exact_vma() 2005 struct vm_area_struct *vma = find_vma(mm, vm_start); find_exact_vma() 2007 if (vma && (vma->vm_start != vm_start || vma->vm_end != vm_end)) find_exact_vma() 2002 find_exact_vma(struct mm_struct *mm, unsigned long vm_start, unsigned long vm_end) find_exact_vma() argument
|
/linux-4.1.27/arch/microblaze/pci/ |
H A D | pci-common.c | 309 * is described by vm_start and vm_end members of VMA, the base physical 333 ret = remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_page_range() 334 vma->vm_end - vma->vm_start, vma->vm_page_prot); pci_mmap_page_range() 432 resource_size_t size = vma->vm_end - vma->vm_start; pci_mmap_legacy_page_range() 476 return remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_legacy_page_range() 477 vma->vm_end - vma->vm_start, pci_mmap_legacy_page_range()
|
/linux-4.1.27/drivers/staging/lustre/lustre/llite/ |
H A D | llite_mmap.c | 60 policy->l_extent.start = ((addr - vma->vm_start) & CFS_PAGE_MASK) + policy_from_vma() 75 vma != NULL && vma->vm_start < (addr + count); vma = vma->vm_next) { our_vma()
|
/linux-4.1.27/drivers/oprofile/ |
H A D | buffer_sync.c | 260 if (addr < vma->vm_start || addr >= vma->vm_end) lookup_dcookie() 266 vma->vm_start; lookup_dcookie()
|
/linux-4.1.27/drivers/gpu/drm/exynos/ |
H A D | exynos_drm_gem.c | 332 vm_size = vma->vm_end - vma->vm_start; exynos_drm_gem_mmap_buffer() 604 vma->vm_start) >> PAGE_SHIFT; exynos_drm_gem_fault()
|
H A D | exynos_drm_fbdev.c | 50 vm_size = vma->vm_end - vma->vm_start; exynos_drm_fb_mmap()
|
/linux-4.1.27/arch/um/kernel/ |
H A D | tlb.c | 530 fix_range(mm, vma->vm_start, vma->vm_end, 0); flush_tlb_mm() 541 fix_range(mm, vma->vm_start, vma->vm_end, 1); force_flush_all()
|
H A D | trap.c | 51 else if (vma->vm_start <= address) handle_page_fault()
|
/linux-4.1.27/kernel/events/ |
H A D | uprobes.c | 114 * We keep the vma's vm_start rather than a pointer to the vma 141 return vma->vm_start + offset - ((loff_t)vma->vm_pgoff << PAGE_SHIFT); offset_to_vaddr() 146 return ((loff_t)vma->vm_pgoff << PAGE_SHIFT) + (vaddr - vma->vm_start); vaddr_to_offset() 814 if (vma->vm_start > info->vaddr || register_for_each_vma() 982 uprobe->offset >= offset + vma->vm_end - vma->vm_start) unapply_uprobe() 1075 build_probe_list(inode, vma, vma->vm_start, vma->vm_end, &tmp_list); uprobe_mmap() 1662 if (vma_has_uprobes(vma, vma->vm_start, vma->vm_end)) mmf_recalc_uprobes() 1702 if (vma && vma->vm_start <= bp_vaddr) { find_active_uprobe()
|
/linux-4.1.27/drivers/misc/cxl/ |
H A D | context.c | 148 u64 len = vma->vm_end - vma->vm_start; cxl_context_iomap()
|
H A D | fault.c | 278 for (ea = vma->vm_start; ea < vma->vm_end; cxl_prefault_vma()
|
/linux-4.1.27/drivers/staging/android/ion/ |
H A D | ion_heap.c | 75 unsigned long addr = vma->vm_start; ion_heap_map_user()
|
/linux-4.1.27/drivers/gpu/drm/gma500/ |
H A D | gem.c | 207 page_offset = ((unsigned long) vmf->virtual_address - vma->vm_start) psb_gem_fault()
|
/linux-4.1.27/arch/xtensa/mm/ |
H A D | fault.c | 83 if (vma->vm_start <= address) do_page_fault()
|
/linux-4.1.27/drivers/gpu/drm/vgem/ |
H A D | vgem_drv.c | 98 page_offset = ((unsigned long)vmf->virtual_address - vma->vm_start) >> vgem_gem_fault()
|
/linux-4.1.27/arch/s390/kernel/ |
H A D | vdso.c | 249 if (vma->vm_mm && vma->vm_start == vma->vm_mm->context.vdso_base) arch_vma_name()
|
/linux-4.1.27/arch/score/mm/ |
H A D | fault.c | 86 if (vma->vm_start <= address) do_page_fault()
|
/linux-4.1.27/arch/avr32/mm/ |
H A D | fault.c | 97 if (vma->vm_start <= address) do_page_fault()
|
/linux-4.1.27/arch/arm/include/asm/ |
H A D | tlb.h | 198 flush_cache_range(vma, vma->vm_start, vma->vm_end); tlb_start_vma()
|
/linux-4.1.27/arch/alpha/mm/ |
H A D | fault.c | 124 if (vma->vm_start <= address) do_page_fault()
|
/linux-4.1.27/drivers/media/platform/omap/ |
H A D | omap_vout.c | 216 physp = (vma->vm_pgoff << PAGE_SHIFT) + (virtp - vma->vm_start); omap_vout_uservirt_to_phys() 866 "vm_open [vma=%08lx-%08lx]\n", vma->vm_start, vma->vm_end); omap_vout_vm_open() 875 "vm_close [vma=%08lx-%08lx]\n", vma->vm_start, vma->vm_end); omap_vout_vm_close() 888 unsigned long start = vma->vm_start; omap_vout_mmap() 889 unsigned long size = (vma->vm_end - vma->vm_start); omap_vout_mmap() 895 vma->vm_pgoff, vma->vm_start, vma->vm_end); omap_vout_mmap() 921 q->bufs[i]->baddr = vma->vm_start; omap_vout_mmap()
|
/linux-4.1.27/arch/powerpc/oprofile/cell/ |
H A D | spu_task_sync.c | 341 if (vma->vm_start > spu_ref || vma->vm_end <= spu_ref) get_exec_dcookie_and_offset() 343 my_offset = spu_ref - vma->vm_start; get_exec_dcookie_and_offset()
|
/linux-4.1.27/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_provider.c | 348 int len = vma->vm_end - vma->vm_start; iwch_mmap() 359 if (vma->vm_start & (PAGE_SIZE-1)) { iwch_mmap() 386 ret = io_remap_pfn_range(vma, vma->vm_start, iwch_mmap() 394 ret = remap_pfn_range(vma, vma->vm_start, iwch_mmap()
|
/linux-4.1.27/arch/microblaze/mm/ |
H A D | fault.c | 153 if (vma->vm_start <= address) do_page_fault()
|
/linux-4.1.27/drivers/gpu/drm/ttm/ |
H A D | ttm_bo_vm.c | 177 page_offset = ((address - vma->vm_start) >> PAGE_SHIFT) + ttm_bo_vm_fault()
|
/linux-4.1.27/arch/openrisc/mm/ |
H A D | fault.c | 117 if (vma->vm_start <= address) do_page_fault()
|
/linux-4.1.27/arch/blackfin/kernel/ |
H A D | ptrace.c | 130 valid = vma && start >= vma->vm_start && start + len <= vma->vm_end; is_user_addr_valid()
|
/linux-4.1.27/arch/cris/mm/ |
H A D | fault.c | 126 if (vma->vm_start <= address) do_page_fault()
|
/linux-4.1.27/arch/m68k/kernel/ |
H A D | sys_m68k.c | 404 if (!vma || addr < vma->vm_start || addr + len > vma->vm_end) sys_cacheflush()
|