Searched refs:vm_start (Results 1 - 200 of 289) sorted by relevance

12

/linux-4.1.27/arch/x86/um/
H A Dmem_64.c7 if (vma->vm_mm && vma->vm_start == um_vdso_addr) arch_vma_name()
H A Dmem_32.c21 gate_vma.vm_start = FIXADDR_USER_START; gate_vma_init()
53 return (addr >= vma->vm_start) && (addr < vma->vm_end); in_gate_area()
/linux-4.1.27/arch/sparc/include/asm/
H A Dtlb_32.h6 flush_cache_range(vma, vma->vm_start, vma->vm_end); \
11 flush_tlb_range(vma, vma->vm_start, vma->vm_end); \
/linux-4.1.27/arch/ia64/include/asm/
H A Dfb.h12 if (efi_range_is_wc(vma->vm_start, vma->vm_end - vma->vm_start)) fb_pgprotect()
/linux-4.1.27/arch/metag/include/asm/
H A Dtlb.h14 flush_cache_range(vma, vma->vm_start, vma->vm_end); \
20 flush_tlb_range(vma, vma->vm_start, vma->vm_end); \
/linux-4.1.27/arch/parisc/include/asm/
H A Dtlb.h11 flush_cache_range(vma, vma->vm_start, vma->vm_end); \
16 flush_tlb_range(vma, vma->vm_start, vma->vm_end); \
/linux-4.1.27/arch/avr32/include/asm/
H A Dtlb.h12 flush_cache_range(vma, vma->vm_start, vma->vm_end)
15 flush_tlb_range(vma, vma->vm_start, vma->vm_end)
/linux-4.1.27/arch/powerpc/include/asm/
H A Dfb.h12 vma->vm_end - vma->vm_start, fb_pgprotect()
/linux-4.1.27/arch/metag/kernel/
H A Ddma.c90 * .vm_start = VMALLOC_START,
94 * However, vmalloc_head.vm_start is variable (typically, it is dependent on
101 unsigned long vm_start; member in struct:metag_vm_region
109 .vm_start = CONSISTENT_START,
117 unsigned long addr = head->vm_start, end = head->vm_end - size; metag_vm_region_alloc()
130 if ((addr + size) <= c->vm_start) metag_vm_region_alloc()
142 new->vm_start = addr; metag_vm_region_alloc()
162 if (c->vm_active && c->vm_start == addr) metag_vm_region_find()
225 unsigned long vaddr = c->vm_start; dma_alloc_coherent()
258 return (void *)c->vm_start; dma_alloc_coherent()
287 if ((c->vm_end - c->vm_start) != size) { dma_free_coherent()
289 __func__, c->vm_end - c->vm_start, size); dma_free_coherent()
291 size = c->vm_end - c->vm_start; dma_free_coherent()
294 ptep = consistent_pte + CONSISTENT_OFFSET(c->vm_start); dma_free_coherent()
295 addr = c->vm_start; dma_free_coherent()
317 flush_tlb_kernel_range(c->vm_start, c->vm_end); dma_free_coherent()
343 user_size = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; dma_mmap()
352 kern_size = (c->vm_end - c->vm_start) >> PAGE_SHIFT; dma_mmap()
356 ret = remap_pfn_range(vma, vma->vm_start, dma_mmap()
/linux-4.1.27/drivers/infiniband/hw/ehca/
H A Dehca_uverbs.c75 ehca_gen_err("Invalid vma struct vm_start=%lx vm_end=%lx", ehca_mm_open()
76 vma->vm_start, vma->vm_end); ehca_mm_open()
81 ehca_gen_err("Use count overflow vm_start=%lx vm_end=%lx", ehca_mm_open()
82 vma->vm_start, vma->vm_end); ehca_mm_open()
83 ehca_gen_dbg("vm_start=%lx vm_end=%lx count=%x", ehca_mm_open()
84 vma->vm_start, vma->vm_end, *count); ehca_mm_open()
91 ehca_gen_err("Invalid vma struct vm_start=%lx vm_end=%lx", ehca_mm_close()
92 vma->vm_start, vma->vm_end); ehca_mm_close()
96 ehca_gen_dbg("vm_start=%lx vm_end=%lx count=%x", ehca_mm_close()
97 vma->vm_start, vma->vm_end, *count); ehca_mm_close()
111 vsize = vma->vm_end - vma->vm_start; ehca_mmap_fw()
113 ehca_gen_err("invalid vsize=%lx", vma->vm_end - vma->vm_start); ehca_mmap_fw()
121 ret = remap_4k_pfn(vma, vma->vm_start, physical >> EHCA_PAGESHIFT, ehca_mmap_fw()
143 start = vma->vm_start; ehca_mmap_queue()
/linux-4.1.27/arch/xtensa/include/asm/
H A Dtlb.h29 flush_cache_range(vma, vma->vm_start, vma->vm_end); \
35 flush_tlb_range(vma, vma->vm_start, vma->vm_end); \
/linux-4.1.27/arch/arc/include/asm/
H A Dtlb.h32 flush_cache_range(vma, vma->vm_start, vma->vm_end); \
39 flush_tlb_range(vma, vma->vm_start, vma->vm_end); \
/linux-4.1.27/drivers/xen/xenfs/
H A Dxenstored.c35 size_t size = vma->vm_end - vma->vm_start; xsd_kva_mmap()
40 if (remap_pfn_range(vma, vma->vm_start, xsd_kva_mmap()
/linux-4.1.27/mm/
H A Dmsync.c68 if (start < vma->vm_start) { SYSCALL_DEFINE3()
69 start = vma->vm_start; SYSCALL_DEFINE3()
74 /* Here vma->vm_start <= start < vma->vm_end. */ SYSCALL_DEFINE3()
81 fstart = (start - vma->vm_start) + SYSCALL_DEFINE3()
H A Dnommu.c141 return vma->vm_end - vma->vm_start; kobjsize()
597 BUG_ON(unlikely(last->vm_end <= last->vm_start)); validate_nommu_regions()
604 BUG_ON(unlikely(region->vm_end <= region->vm_start)); validate_nommu_regions()
606 BUG_ON(unlikely(region->vm_start < last->vm_top)); validate_nommu_regions()
632 if (region->vm_start < pregion->vm_start) add_nommu_region()
634 else if (region->vm_start > pregion->vm_start) add_nommu_region()
681 * will equal vm_start
691 if (region->vm_top > region->vm_start) __releases()
702 free_page_series(region->vm_start, region->vm_top); __releases()
726 long start = vma->vm_start & PAGE_MASK; protect_vma()
776 if (vma->vm_start < pvma->vm_start) add_vma_to_mm()
778 else if (vma->vm_start > pvma->vm_start) { add_vma_to_mm()
882 if (vma->vm_start > addr) find_vma()
931 if (vma->vm_start < addr) find_vma_exact()
933 if (vma->vm_start > addr) find_vma_exact()
1226 region->vm_start = (unsigned long) base; do_mmap_private()
1227 region->vm_end = region->vm_start + len; do_mmap_private()
1228 region->vm_top = region->vm_start + (total << PAGE_SHIFT); do_mmap_private()
1230 vma->vm_start = region->vm_start; do_mmap_private()
1231 vma->vm_end = region->vm_start + len; do_mmap_private()
1258 free_page_series(region->vm_start, region->vm_top); do_mmap_private()
1259 region->vm_start = vma->vm_start = 0; do_mmap_private()
1362 rpglen = pregion->vm_end - pregion->vm_start; do_mmap_pgoff()
1381 start = pregion->vm_start; do_mmap_pgoff()
1383 vma->vm_start = start; do_mmap_pgoff()
1394 vma->vm_start = 0; do_mmap_pgoff()
1429 vma->vm_start = region->vm_start = addr; do_mmap_pgoff()
1450 memset((void *)region->vm_start, 0, do_mmap_pgoff()
1451 region->vm_end - region->vm_start); do_mmap_pgoff()
1454 result = vma->vm_start; do_mmap_pgoff()
1464 flush_icache_range(region->vm_start, region->vm_end); do_mmap_pgoff()
1591 npages = (addr - vma->vm_start) >> PAGE_SHIFT; split_vma()
1596 region->vm_start = new->vm_start = addr; split_vma()
1607 vma->vm_region->vm_start = vma->vm_start = addr; split_vma()
1636 if (from > vma->vm_start) shrink_vma()
1639 vma->vm_start = to; shrink_vma()
1648 if (from > region->vm_start) { shrink_vma()
1652 region->vm_start = to; shrink_vma()
1698 if (start > vma->vm_start) { do_munmap()
1710 if (start == vma->vm_start && end == vma->vm_end) do_munmap()
1712 if (start < vma->vm_start || end > vma->vm_end) { do_munmap()
1724 if (start != vma->vm_start && end != vma->vm_end) { do_munmap()
1820 if (vma->vm_end != vma->vm_start + old_len) do_mremap()
1826 if (new_len > vma->vm_region->vm_end - vma->vm_region->vm_start) do_mremap()
1830 vma->vm_end = vma->vm_start + new_len; do_mremap()
1831 return vma->vm_start; do_mremap()
1868 unsigned long vm_len = vma->vm_end - vma->vm_start; vm_iomap_memory()
1871 return io_remap_pfn_range(vma, vma->vm_start, pfn, vm_len, vma->vm_page_prot); vm_iomap_memory()
1878 unsigned int size = vma->vm_end - vma->vm_start; remap_vmalloc_range()
1883 vma->vm_start = (unsigned long)(addr + (pgoff << PAGE_SHIFT)); remap_vmalloc_range()
1884 vma->vm_end = vma->vm_start + size; remap_vmalloc_range()
2126 r_size = region->vm_top - region->vm_start; nommu_shrink_inode_mappings()
H A Dmadvise.c103 pgoff = vma->vm_pgoff + ((start - vma->vm_start) >> PAGE_SHIFT); madvise_behavior()
113 if (start != vma->vm_start) { madvise_behavior()
196 index = ((start - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff; force_shm_swapin_readahead()
248 start = ((start - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff; madvise_willneed()
251 end = ((end - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff; madvise_willneed()
314 offset = (loff_t)(start - vma->vm_start) madvise_remove()
503 if (vma && start > vma->vm_start) SYSCALL_DEFINE3()
514 if (start < vma->vm_start) { SYSCALL_DEFINE3()
516 start = vma->vm_start; SYSCALL_DEFINE3()
521 /* Here vma->vm_start <= start < (end|vma->vm_end) */ SYSCALL_DEFINE3()
526 /* Here vma->vm_start <= start < tmp <= (end|vma->vm_end). */ SYSCALL_DEFINE3()
H A Dmremap.c269 new_pgoff = vma->vm_pgoff + ((old_addr - vma->vm_start) >> PAGE_SHIFT); move_vma()
301 excess = vma->vm_end - vma->vm_start - old_len; move_vma()
302 if (old_addr > vma->vm_start && move_vma()
347 if (!vma || vma->vm_start > addr) vma_to_resize()
363 pgoff = (addr - vma->vm_start) >> PAGE_SHIFT; vma_to_resize()
437 ((addr - vma->vm_start) >> PAGE_SHIFT), mremap_to()
457 if (vma->vm_next && vma->vm_next->vm_start < end) /* intersection */ vma_expandable()
459 if (get_unmapped_area(NULL, vma->vm_start, end - vma->vm_start, vma_expandable()
539 if (vma_adjust(vma, vma->vm_start, addr + new_len, SYSCALL_DEFINE5()
568 ((addr - vma->vm_start) >> PAGE_SHIFT), SYSCALL_DEFINE5()
H A Dmprotect.c291 pgoff = vma->vm_pgoff + ((start - vma->vm_start) >> PAGE_SHIFT); mprotect_fixup()
301 if (start != vma->vm_start) { mprotect_fixup()
374 if (vma->vm_start >= end) SYSCALL_DEFINE3()
376 start = vma->vm_start; SYSCALL_DEFINE3()
381 if (vma->vm_start > start) SYSCALL_DEFINE3()
390 if (start > vma->vm_start) SYSCALL_DEFINE3()
396 /* Here we know that vma->vm_start <= nstart < vma->vm_end. */ SYSCALL_DEFINE3()
425 if (!vma || vma->vm_start != nstart) { SYSCALL_DEFINE3()
H A Dmmap.c358 max = vma->vm_start; vma_compute_subtree_gap()
386 if (vma->vm_start < prev) { browse_rb()
387 pr_emerg("vm_start %lx < prev %lx\n", browse_rb()
388 vma->vm_start, prev); browse_rb()
391 if (vma->vm_start < pend) { browse_rb()
392 pr_emerg("vm_start %lx < pend %lx\n", browse_rb()
393 vma->vm_start, pend); browse_rb()
396 if (vma->vm_start > vma->vm_end) { browse_rb()
397 pr_emerg("vm_start %lx > vm_end %lx\n", browse_rb()
398 vma->vm_start, vma->vm_end); browse_rb()
409 prev = vma->vm_start; browse_rb()
483 * Update augmented rbtree rb_subtree_gap values after vma->vm_start or
525 * Before updating the vma's vm_start / vm_end / vm_pgoff fields, the
570 if (vma_tmp->vm_start < end) find_vma_links()
599 max(addr, vma->vm_start)) >> PAGE_SHIFT; count_vma_pages_range()
605 if (vma->vm_start > end) count_vma_pages_range()
608 overlap_len = min(end, vma->vm_end) - vma->vm_start; count_vma_pages_range()
697 if (find_vma_links(mm, vma->vm_start, vma->vm_end, __insert_vm_struct()
720 * We cannot adjust vm_start, vm_end, vm_pgoff fields of a vma that
752 } else if (end > next->vm_start) { vma_adjust()
757 adjust_next = (end - next->vm_start) >> PAGE_SHIFT; vma_adjust()
789 uprobe_munmap(vma, vma->vm_start, vma->vm_end); vma_adjust()
792 uprobe_munmap(next, next->vm_start, next->vm_end); vma_adjust()
827 if (start != vma->vm_start) { vma_adjust()
828 vma->vm_start = start; vma_adjust()
837 next->vm_start += adjust_next << PAGE_SHIFT; vma_adjust()
892 uprobe_munmap(next, next->vm_start, next->vm_end); vma_adjust()
1067 if (next && end == next->vm_start && vma_merge()
1074 err = vma_adjust(prev, prev->vm_start, vma_merge()
1077 err = vma_adjust(prev, prev->vm_start, vma_merge()
1088 if (next && end == next->vm_start && vma_merge()
1093 err = vma_adjust(prev, prev->vm_start, vma_merge()
1122 return a->vm_end == b->vm_start && anon_vma_compatible()
1126 b->vm_pgoff == a->vm_pgoff + ((b->vm_start - a->vm_start) >> PAGE_SHIFT); anon_vma_compatible()
1594 vma->vm_start = addr; mmap_region()
1630 WARN_ON_ONCE(addr != vma->vm_start); mmap_region()
1632 addr = vma->vm_start; mmap_region()
1682 unmap_region(mm, vma, prev, vma->vm_start, vma->vm_end); mmap_region()
1703 * - gap_end = vma->vm_start >= info->low_limit + length; unmapped_area()
1734 gap_end = vma->vm_start; unmapped_area()
1773 gap_end = vma->vm_start; unmapped_area()
1850 gap_end = vma->vm_start; unmapped_area_topdown()
1927 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area()
1967 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area_topdown()
2051 if (tmp->vm_start <= addr) find_vma()
2122 new_start = (vma->vm_flags & VM_GROWSUP) ? vma->vm_start : acct_stack_growth()
2164 * vma->vm_start/vm_end cannot change under us because the caller expand_upwards()
2174 size = address - vma->vm_start; expand_upwards()
2214 * vma is the first one with address < vma->vm_start. Have to extend vma.
2231 * vma->vm_start/vm_end cannot change under us because the caller expand_downwards()
2238 if (address < vma->vm_start) { expand_downwards()
2242 grow = (vma->vm_start - address) >> PAGE_SHIFT; expand_downwards()
2261 vma->vm_start = address; expand_downwards()
2295 if (next && next->vm_start == address + PAGE_SIZE) { expand_stack()
2309 if (vma && (vma->vm_start <= addr)) find_extend_vma()
2341 if (vma->vm_start <= addr) find_extend_vma()
2345 start = vma->vm_start; find_extend_vma()
2397 next ? next->vm_start : USER_PGTABLES_CEILING); unmap_region()
2419 } while (vma && vma->vm_start < end); detach_vmas_to_be_unmapped()
2458 new->vm_start = addr; __split_vma()
2459 new->vm_pgoff += ((addr - vma->vm_start) >> PAGE_SHIFT); __split_vma()
2478 ((addr - new->vm_start) >> PAGE_SHIFT), new); __split_vma()
2480 err = vma_adjust(vma, vma->vm_start, addr, vma->vm_pgoff, new); __split_vma()
2539 if (vma->vm_start >= end) do_munmap()
2549 if (start > vma->vm_start) { do_munmap()
2568 if (last && end > last->vm_start) { do_munmap()
2580 while (tmp && tmp->vm_start < end) { do_munmap()
2657 if (start < vma->vm_start) SYSCALL_DEFINE5()
2665 if (next->vm_start != next->vm_prev->vm_end) SYSCALL_DEFINE5()
2693 for (tmp = vma; tmp->vm_start >= start + size; SYSCALL_DEFINE5()
2696 max(tmp->vm_start, start), SYSCALL_DEFINE5()
2794 vma->vm_start = addr; do_brk()
2895 vma->vm_pgoff = vma->vm_start >> PAGE_SHIFT; insert_vm_struct()
2897 if (find_vma_links(mm, vma->vm_start, vma->vm_end, insert_vm_struct()
2917 unsigned long vma_start = vma->vm_start; copy_vma()
2940 if (unlikely(vma_start >= new_vma->vm_start && copy_vma()
2962 new_vma->vm_start = addr; copy_vma()
3077 vma->vm_start = addr; __install_special_mapping()
H A Dvmacache.c102 if (vma->vm_start <= addr && vma->vm_end > addr) { vmacache_find()
126 if (vma && vma->vm_start == start && vma->vm_end == end) { vmacache_find_exact()
H A Dpagewalk.c259 } else if (start < vma->vm_start) { /* outside vma */ walk_page_range()
261 next = min(end, vma->vm_start); walk_page_range()
298 err = walk_page_test(vma->vm_start, vma->vm_end, walk); walk_page_vma()
303 return __walk_page_range(vma->vm_start, vma->vm_end, walk); walk_page_vma()
H A Dinterval_tree.c21 return v->vm_pgoff + ((v->vm_end - v->vm_start) >> PAGE_SHIFT) - 1; vma_last_pgoff()
H A Dmlock.c511 pgoff = vma->vm_pgoff + ((start - vma->vm_start) >> PAGE_SHIFT); mlock_fixup()
519 if (start != vma->vm_start) { mlock_fixup()
570 if (!vma || vma->vm_start > start) do_mlock()
574 if (start > vma->vm_start) do_mlock()
580 /* Here we know that vma->vm_start <= nstart < vma->vm_end. */ do_mlock()
599 if (!vma || vma->vm_start != nstart) { do_mlock()
676 mlock_fixup(vma, &prev, vma->vm_start, vma->vm_end, newflags); do_mlockall()
H A Dmemory.c533 unsigned long addr = vma->vm_start; free_pgtables()
544 floor, next? next->vm_start: ceiling); free_pgtables()
549 while (next && next->vm_start <= vma->vm_end + PMD_SIZE free_pgtables()
557 floor, next? next->vm_start: ceiling); free_pgtables()
723 * pfn_of_page == vma->vm_pgoff + ((addr - vma->vm_start) >> PAGE_SHIFT)
775 off = (addr - vma->vm_start) >> PAGE_SHIFT; vm_normal_page()
1010 unsigned long addr = vma->vm_start; copy_page_range()
1199 pr_err("%s: mmap_sem is unlocked! addr=0x%lx end=0x%lx vma->vm_start=0x%lx vma->vm_end=0x%lx\n", zap_pmd_range()
1201 vma->vm_start, zap_pmd_range()
1276 unsigned long start = max(vma->vm_start, start_addr); unmap_single_vma()
1282 if (end <= vma->vm_start) unmap_single_vma()
1339 for ( ; vma && vma->vm_start < end_addr; vma = vma->vm_next) unmap_vmas()
1364 for ( ; vma && vma->vm_start < end; vma = vma->vm_next) zap_page_range()
1410 if (address < vma->vm_start || address + size > vma->vm_end || zap_vma_ptes()
1505 if (addr < vma->vm_start || addr >= vma->vm_end) vm_insert_page()
1580 if (addr < vma->vm_start || addr >= vma->vm_end) vm_insert_pfn()
1596 if (addr < vma->vm_start || addr >= vma->vm_end) vm_insert_mixed()
1721 if (addr != vma->vm_start || end != vma->vm_end) remap_pfn_range()
1789 vm_len = vma->vm_end - vma->vm_start; vm_iomap_memory()
1794 return io_remap_pfn_range(vma, vma->vm_start, pfn, vm_len, vma->vm_page_prot); vm_iomap_memory()
2383 ((zba - vba) << PAGE_SHIFT) + vma->vm_start, unmap_mapping_range_tree()
2384 ((zea - vba + 1) << PAGE_SHIFT) + vma->vm_start, unmap_mapping_range_tree()
2630 if ((vma->vm_flags & VM_GROWSDOWN) && address == vma->vm_start) { check_stack_guard_page()
2648 if (next && next->vm_start == address + PAGE_SIZE) check_stack_guard_page()
2889 start_addr = max(address & mask, vma->vm_start); do_fault_around()
3101 - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff; do_fault()
3647 if (!vma || vma->vm_start > addr) __access_remote_vm()
3747 vma->vm_start, print_vma_addr()
3748 vma->vm_end - vma->vm_start); print_vma_addr()
H A Ddebug.c160 vma, (void *)vma->vm_start, (void *)vma->vm_end, vma->vm_next, dump_vma()
H A Dmempolicy.c599 if (vma->vm_start > start) queue_pages_test_walk()
600 start = vma->vm_start; queue_pages_test_walk()
605 if (qp->prev && qp->prev->vm_end < vma->vm_start) queue_pages_test_walk()
670 vma->vm_start, vma->vm_end, vma->vm_pgoff, vma_replace_policy()
707 if (!vma || vma->vm_start > start) mbind_range()
711 if (start > vma->vm_start) mbind_range()
714 for (; vma && vma->vm_start < end; prev = vma, vma = next) { mbind_range()
716 vmstart = max(start, vma->vm_start); mbind_range()
723 ((vmstart - vma->vm_start) >> PAGE_SHIFT); mbind_range()
735 if (vma->vm_start != vmstart) { mbind_range()
972 queue_pages_range(mm, mm->mmap->vm_start, mm->task_size, &nmask, migrate_to_node()
1599 pol = vma->vm_ops->get_policy(vma, vma->vm_start); vma_policy_mof()
1772 off += (addr - vma->vm_start) >> shift; interleave_nid()
2270 BUG_ON(addr < vma->vm_start); mpol_misplaced()
2273 pgoff += (addr - vma->vm_start) >> PAGE_SHIFT; mpol_misplaced()
H A Drmap.c570 return vma->vm_start + ((pgoff - vma->vm_pgoff) << PAGE_SHIFT); __vma_address()
579 VM_BUG_ON_VMA(address < vma->vm_start || address >= vma->vm_end, vma); vma_address()
606 if (unlikely(address < vma->vm_start || address >= vma->vm_end)) page_address_in_vma()
704 if (unlikely(address < vma->vm_start || address >= vma->vm_end)) page_mapped_in_vma()
1076 VM_BUG_ON_VMA(address < vma->vm_start || address >= vma->vm_end, vma); page_add_new_anon_rmap()
1595 BUG_ON(address < vma->vm_start || address >= vma->vm_end); hugepage_add_new_anon_rmap()
H A Dhuge_memory.c797 if (haddr < vma->vm_start || haddr + HPAGE_PMD_SIZE > vma->vm_end) do_huge_pmd_anonymous_page()
2067 hstart = (vma->vm_start + ~HPAGE_PMD_MASK) & HPAGE_PMD_MASK; khugepaged_enter_vma_merge()
2471 hstart = (vma->vm_start + ~HPAGE_PMD_MASK) & HPAGE_PMD_MASK; collapse_huge_page()
2705 hstart = (vma->vm_start + ~HPAGE_PMD_MASK) & HPAGE_PMD_MASK;
2892 BUG_ON(vma->vm_start > haddr || vma->vm_end < haddr + HPAGE_PMD_SIZE); __split_huge_page_pmd()
2977 (start & HPAGE_PMD_MASK) >= vma->vm_start && __vma_adjust_trans_huge()
2987 (end & HPAGE_PMD_MASK) >= vma->vm_start && __vma_adjust_trans_huge()
2992 * If we're also updating the vma->vm_next->vm_start, if the new __vma_adjust_trans_huge()
2993 * vm_next->vm_start isn't page aligned and it could previously __vma_adjust_trans_huge()
2998 unsigned long nstart = next->vm_start; __vma_adjust_trans_huge()
3001 (nstart & HPAGE_PMD_MASK) >= next->vm_start && __vma_adjust_trans_huge()
H A Dmincore.c186 if (!vma || addr < vma->vm_start) do_mincore()
H A Dgup.c552 if (!vma || address < vma->vm_start) fixup_user_fault()
849 VM_BUG_ON_VMA(start < vma->vm_start, vma); populate_vma_page_range()
907 if (!vma || vma->vm_start >= end) __mm_populate()
916 if (nstart < vma->vm_start) __mm_populate()
917 nstart = vma->vm_start; __mm_populate()
/linux-4.1.27/arch/powerpc/mm/
H A Ddma-noncoherent.c76 * .vm_start = VMALLOC_START,
80 * However, vmalloc_head.vm_start is variable (typically, it is dependent on
86 unsigned long vm_start; member in struct:ppc_vm_region
92 .vm_start = CONSISTENT_BASE,
99 unsigned long addr = head->vm_start, end = head->vm_end - size; ppc_vm_region_alloc()
112 if ((addr + size) <= c->vm_start) ppc_vm_region_alloc()
124 new->vm_start = addr; ppc_vm_region_alloc()
142 if (c->vm_start == addr) ppc_vm_region_find()
218 unsigned long vaddr = c->vm_start; __dma_alloc_coherent()
244 return (void *)c->vm_start; __dma_alloc_coherent()
270 if ((c->vm_end - c->vm_start) != size) { __dma_free_coherent()
272 __func__, c->vm_end - c->vm_start, size); __dma_free_coherent()
274 size = c->vm_end - c->vm_start; __dma_free_coherent()
277 addr = c->vm_start; __dma_free_coherent()
297 flush_tlb_kernel_range(c->vm_start, c->vm_end); __dma_free_coherent()
H A Dcopro_fault.c55 if (ea < vma->vm_start) { copro_handle_mm_fault()
H A Dsubpage-prot.c159 if (vma && ((addr + len) <= vma->vm_start)) subpage_mark_vma_nohuge()
163 if (vma->vm_start >= (addr + len)) subpage_mark_vma_nohuge()
H A Dtlb_hash32.c150 flush_range(mp->vm_mm, mp->vm_start, mp->vm_end); flush_tlb_mm()
/linux-4.1.27/arch/mips/include/asm/
H A Dtlb.h11 flush_cache_range(vma, vma->vm_start, vma->vm_end); \
/linux-4.1.27/drivers/gpu/drm/
H A Ddrm_vm.c67 if (efi_range_is_wc(vma->vm_start, vma->vm_end - drm_io_prot()
68 vma->vm_start)) drm_io_prot()
128 vma->vm_start; drm_do_vm_fault()
198 offset = (unsigned long)vmf->virtual_address - vma->vm_start; drm_do_vm_shm_fault()
228 vma->vm_start, vma->vm_end - vma->vm_start); drm_vm_shm_close()
304 offset = (unsigned long)vmf->virtual_address - vma->vm_start; /* vm_[pg]off[set] should be 0 */ drm_do_vm_dma_fault()
340 offset = (unsigned long)vmf->virtual_address - vma->vm_start; drm_do_vm_sg_fault()
412 vma->vm_start, vma->vm_end - vma->vm_start); drm_vm_open_locked()
438 vma->vm_start, vma->vm_end - vma->vm_start); drm_vm_close_locked()
482 unsigned long length = vma->vm_end - vma->vm_start; drm_mmap_dma()
487 vma->vm_start, vma->vm_end, vma->vm_pgoff); drm_mmap_dma()
549 vma->vm_start, vma->vm_end, vma->vm_pgoff); drm_mmap_locked()
576 if (map->size < vma->vm_end - vma->vm_start) drm_mmap_locked()
615 if (io_remap_pfn_range(vma, vma->vm_start, drm_mmap_locked()
617 vma->vm_end - vma->vm_start, drm_mmap_locked()
623 vma->vm_start, vma->vm_end, (unsigned long long)(map->offset + offset)); drm_mmap_locked()
630 if (remap_pfn_range(vma, vma->vm_start, drm_mmap_locked()
632 vma->vm_end - vma->vm_start, vma->vm_page_prot)) drm_mmap_locked()
708 (void *)vma->vm_start, (void *)vma->vm_end, drm_vma_info()
/linux-4.1.27/arch/arc/kernel/
H A Darc_hostlink.c25 if (io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, arc_hl_mmap()
26 vma->vm_end - vma->vm_start, arc_hl_mmap()
H A Dtroubleshoot.c99 if (vma && (vma->vm_start <= address)) { show_faulting_vma()
110 vma->vm_start < TASK_UNMAPPED_BASE ? show_faulting_vma()
111 address : address - vma->vm_start, show_faulting_vma()
112 nm, vma->vm_start, vma->vm_end); show_faulting_vma()
/linux-4.1.27/arch/nios2/include/asm/
H A Dtlb.h25 flush_cache_range(vma, vma->vm_start, vma->vm_end); \
/linux-4.1.27/arch/nios2/kernel/
H A Dsys_nios2.c42 if (vma == NULL || addr < vma->vm_start || addr + len > vma->vm_end) sys_cacheflush()
/linux-4.1.27/arch/um/drivers/
H A Dmmapper_kern.c56 size = vma->vm_end - vma->vm_start; mmapper_mmap()
64 if (remap_pfn_range(vma, vma->vm_start, p_buf >> PAGE_SHIFT, size, mmapper_mmap()
/linux-4.1.27/arch/powerpc/kernel/
H A Dproc_powerpc.c46 if ((vma->vm_end - vma->vm_start) > PAGE_SIZE) page_map_mmap()
49 remap_pfn_range(vma, vma->vm_start, page_map_mmap()
H A Ddma.c122 return remap_pfn_range(vma, vma->vm_start, dma_direct_mmap_coherent()
124 vma->vm_end - vma->vm_start, dma_direct_mmap_coherent()
H A Dvdso.c168 follow_page(vma, vma->vm_start + i*PAGE_SIZE, 0) dump_vdso_pages()
179 follow_page(vma, vma->vm_start + i*PAGE_SIZE, 0) dump_vdso_pages()
288 if (vma->vm_mm && vma->vm_start == vma->vm_mm->context.vdso_base) arch_vma_name()
H A Dpci-common.c439 * is described by vm_start and vm_end members of VMA, the base physical
463 ret = remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_page_range()
464 vma->vm_end - vma->vm_start, vma->vm_page_prot); pci_mmap_page_range()
562 resource_size_t size = vma->vm_end - vma->vm_start; pci_mmap_legacy_page_range()
602 return remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_legacy_page_range()
603 vma->vm_end - vma->vm_start, pci_mmap_legacy_page_range()
/linux-4.1.27/arch/cris/arch-v32/drivers/pci/
H A Dbios.c31 if (remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_page_range()
32 vma->vm_end - vma->vm_start, pci_mmap_page_range()
/linux-4.1.27/fs/proc/
H A Dtask_nommu.c34 size += region->vm_end - region->vm_start; task_mem()
36 size = vma->vm_end - vma->vm_start; task_mem()
89 vsize += vma->vm_end - vma->vm_start; task_vsize()
111 size += region->vm_end - region->vm_start; task_statm()
172 vma->vm_start, nommu_vma_show()
193 if (!is_pid || (vma->vm_start <= mm->start_stack && nommu_vma_show()
H A Dnommu.c56 region->vm_start, nommu_region_show()
H A Dtask_mmu.c137 m->version = m_next_vma(m->private, vma) ? vma->vm_start : -1UL; m_cache_vma()
173 m->version = vma->vm_start; m_start()
289 start = vma->vm_start; show_map_vma()
332 if (vma->vm_start <= mm->brk && show_map_vma()
344 if (!is_pid || (vma->vm_start <= mm->start_stack && show_map_vma()
644 (vma->vm_end - vma->vm_start) >> 10, show_smap()
1019 hole_end = min(end, vma->vm_start); pagemap_pte_hole()
1500 pol = __get_vma_policy(vma, vma->vm_start); show_numa_map()
1508 seq_printf(m, "%08lx %s", vma->vm_start, buffer); show_numa_map()
1513 } else if (vma->vm_start <= mm->brk && vma->vm_end >= mm->start_brk) { show_numa_map()
1522 if (!is_pid || (vma->vm_start <= mm->start_stack && show_numa_map()
H A Dvmcore.c409 size_t size = vma->vm_end - vma->vm_start; mmap_vmcore()
433 if (remap_pfn_range(vma, vma->vm_start, pfn, tsz, mmap_vmcore()
449 if (remap_vmalloc_range_partial(vma, vma->vm_start + len, mmap_vmcore()
466 if (vmcore_remap_oldmem_pfn(vma, vma->vm_start + len, mmap_vmcore()
481 do_munmap(vma->vm_mm, vma->vm_start, len); mmap_vmcore()
H A Dbase.c1659 unsigned long vm_start, vm_end; map_files_d_revalidate() local
1684 if (!dname_to_vma_addr(dentry, &vm_start, &vm_end)) { map_files_d_revalidate()
1686 exact_vma_exists = !!find_exact_vma(mm, vm_start, vm_end); map_files_d_revalidate()
1721 unsigned long vm_start, vm_end; proc_map_files_get_link() local
1737 rc = dname_to_vma_addr(dentry, &vm_start, &vm_end); proc_map_files_get_link()
1743 vma = find_exact_vma(mm, vm_start, vm_end); proc_map_files_get_link()
1796 unsigned long vm_start, vm_end; proc_map_files_lookup() local
1816 if (dname_to_vma_addr(dentry, &vm_start, &vm_end)) proc_map_files_lookup()
1824 vma = find_exact_vma(mm, vm_start, vm_end); proc_map_files_lookup()
1920 vma->vm_start, vma->vm_end); proc_map_files_readdir()
/linux-4.1.27/drivers/xen/
H A Dgntdev.c241 unsigned int pgnr = (addr - map->vma->vm_start) >> PAGE_SHIFT; find_grant_ptes()
452 if (map->vma->vm_start >= end) unmap_if_in_range()
456 mstart = max(start, map->vma->vm_start); unmap_if_in_range()
460 map->vma->vm_start, map->vma->vm_end, unmap_if_in_range()
463 (mstart - map->vma->vm_start) >> PAGE_SHIFT, unmap_if_in_range()
505 map->vma->vm_start, map->vma->vm_end); mn_release()
514 map->vma->vm_start, map->vma->vm_end); mn_release()
782 int count = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; gntdev_mmap()
790 index, count, vma->vm_start, vma->vm_pgoff); gntdev_mmap()
830 err = apply_to_page_range(vma->vm_mm, vma->vm_start, gntdev_mmap()
831 vma->vm_end - vma->vm_start, gntdev_mmap()
845 err = vm_insert_page(vma, vma->vm_start + i*PAGE_SIZE, gntdev_mmap()
862 apply_to_page_range(vma->vm_mm, vma->vm_start, gntdev_mmap()
863 vma->vm_end - vma->vm_start, gntdev_mmap()
867 map->pages_vm_start = vma->vm_start; gntdev_mmap()
H A Dprivcmd.c265 if (!vma || (msg->va != vma->vm_start) || vma->vm_private_data) privcmd_ioctl_mmap()
270 state.va = vma->vm_start; privcmd_ioctl_mmap()
491 if (m.addr != vma->vm_start || privcmd_ioctl_mmap_batch()
503 if (m.addr < vma->vm_start || privcmd_ioctl_mmap_batch()
584 int numpgs = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; privcmd_close()
602 vma, vma->vm_start, vma->vm_end, privcmd_fault()
H A Dgntalloc.c506 int count = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; gntalloc_mmap()
544 rv = vm_insert_page(vma, vma->vm_start + i * PAGE_SIZE, gntalloc_mmap()
/linux-4.1.27/drivers/media/v4l2-core/
H A Dvideobuf-vmalloc.c59 map->count, vma->vm_start, vma->vm_end); videobuf_vm_open()
71 map->count, vma->vm_start, vma->vm_end); videobuf_vm_close()
250 buf->baddr = vma->vm_start; __videobuf_mmap_mapper()
256 pages = PAGE_ALIGN(vma->vm_end - vma->vm_start); __videobuf_mmap_mapper()
277 map, q, vma->vm_start, vma->vm_end, __videobuf_mmap_mapper()
H A Dvideobuf-dma-contig.c71 map, map->count, vma->vm_start, vma->vm_end); videobuf_vm_open()
83 map, map->count, vma->vm_start, vma->vm_end); videobuf_vm_close()
295 buf->baddr = vma->vm_start; __videobuf_mmap_mapper()
306 size = vma->vm_end - vma->vm_start; __videobuf_mmap_mapper()
331 map, q, vma->vm_start, vma->vm_end, __videobuf_mmap_mapper()
H A Dvideobuf2-memops.c151 __func__, h, atomic_read(h->refcount), vma->vm_start, vb2_common_vm_open()
169 __func__, h, atomic_read(h->refcount), vma->vm_start, vb2_common_vm_close()
H A Dvideobuf-dma-sg.c389 map->count, vma->vm_start, vma->vm_end); videobuf_vm_open()
402 map->count, vma->vm_start, vma->vm_end); videobuf_vm_close()
441 vma->vm_start, vma->vm_end); videobuf_vm_fault()
629 q->bufs[i]->baddr = vma->vm_start + size; __videobuf_mmap_mapper()
640 map, q, vma->vm_start, vma->vm_end, vma->vm_pgoff, first, last); __videobuf_mmap_mapper()
H A Dvideobuf2-dma-sg.c397 unsigned long uaddr = vma->vm_start; vb2_dma_sg_mmap()
398 unsigned long usize = vma->vm_end - vma->vm_start; vb2_dma_sg_mmap()
/linux-4.1.27/drivers/misc/mic/host/
H A Dmic_fops.c196 unsigned long pa, size = vma->vm_end - vma->vm_start, size_rem = size; mic_mmap()
210 err = remap_pfn_range(vma, vma->vm_start + offset, mic_mmap()
217 pa, vma->vm_start + offset); mic_mmap()
/linux-4.1.27/drivers/sbus/char/
H A Dflash.c68 if (vma->vm_end - (vma->vm_start + (vma->vm_pgoff << PAGE_SHIFT)) > size) flash_mmap()
69 size = vma->vm_end - (vma->vm_start + (vma->vm_pgoff << PAGE_SHIFT)); flash_mmap()
73 if (io_remap_pfn_range(vma, vma->vm_start, addr, size, vma->vm_page_prot)) flash_mmap()
/linux-4.1.27/arch/microblaze/kernel/
H A Ddma.c162 unsigned long user_count = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; dma_direct_mmap_coherent()
176 return remap_pfn_range(vma, vma->vm_start, pfn + off, dma_direct_mmap_coherent()
177 vma->vm_end - vma->vm_start, vma->vm_page_prot); dma_direct_mmap_coherent()
/linux-4.1.27/arch/parisc/mm/
H A Dfault.c126 if (tree->vm_start > addr) {
132 if (prev->vm_next->vm_start > addr)
195 pr_warn(" vm_start = 0x%08lx, vm_end = 0x%08lx\n", show_signal_msg()
196 vma->vm_start, vma->vm_end); show_signal_msg()
229 if (!vma || address < vma->vm_start) do_page_fault()
306 address < vma->vm_start || address > vma->vm_end) { do_page_fault()
/linux-4.1.27/drivers/staging/unisys/visorchipset/
H A Dfile.c85 if (remap_pfn_range(vma, vma->vm_start, visorchipset_mmap()
87 vma->vm_end - vma->vm_start, visorchipset_mmap()
/linux-4.1.27/arch/sh/mm/
H A Dmmap.c66 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area()
116 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area_topdown()
H A Dfault.c456 if (likely(vma->vm_start <= address)) do_page_fault()
/linux-4.1.27/arch/tile/mm/
H A Delf.c78 if (vma->vm_start == (ELF_ET_DYN_BASE & PAGE_MASK)) { notify_exec()
82 snprintf(buf, sizeof(buf), "0x%lx:@", vma->vm_start); notify_exec()
H A Dhugetlbpage.c240 (!vma || addr + len <= vma->vm_start)) hugetlb_get_unmapped_area()
/linux-4.1.27/drivers/xen/xenbus/
H A Dxenbus_dev_backend.c95 size_t size = vma->vm_end - vma->vm_start; xenbus_backend_mmap()
103 if (remap_pfn_range(vma, vma->vm_start, xenbus_backend_mmap()
/linux-4.1.27/arch/sh/kernel/vsyscall/
H A Dvsyscall.c90 if (vma->vm_mm && vma->vm_start == (long)vma->vm_mm->context.vdso) arch_vma_name()
/linux-4.1.27/arch/alpha/kernel/
H A Dpci-sysfs.c30 return io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, hose_mmap_page_range()
31 vma->vm_end - vma->vm_start, hose_mmap_page_range()
41 nr = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; __pci_mmap_fits()
258 nr = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; __legacy_mmap_fits()
/linux-4.1.27/arch/arc/mm/
H A Dmmap.c67 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area()
H A Dfault.c99 if (vma->vm_start <= address) do_page_fault()
/linux-4.1.27/arch/tile/kernel/
H A Dvdso.c115 if (vma->vm_mm && vma->vm_start == VDSO_BASE) arch_vma_name()
118 if (vma->vm_start == MEM_USER_INTRPT) arch_vma_name()
H A Dstack.c331 if (vma == NULL || address < vma->vm_start) { describe_addr()
350 vma->vm_start, vma->vm_end - vma->vm_start); describe_addr()
/linux-4.1.27/drivers/char/
H A Duv_mmtimer.c151 if (vma->vm_end - vma->vm_start != PAGE_SIZE) uv_mmtimer_mmap()
166 if (remap_pfn_range(vma, vma->vm_start, uv_mmtimer_addr >> PAGE_SHIFT, uv_mmtimer_mmap()
H A Dmspec.c97 unsigned long vm_start; /* Original (unsplit) base. */ member in struct:vma_data
171 last_index = (vdata->vm_end - vdata->vm_start) >> PAGE_SHIFT; mspec_close()
281 vdata->vm_start = vma->vm_start; mspec_mmap()
H A Dbsr.c127 unsigned long size = vma->vm_end - vma->vm_start; bsr_mmap()
135 ret = remap_4k_pfn(vma, vma->vm_start, dev->bsr_addr >> 12, bsr_mmap()
138 ret = io_remap_pfn_range(vma, vma->vm_start, bsr_mmap()
H A Dmem.c323 size_t size = vma->vm_end - vma->vm_start; mmap_mem()
346 vma->vm_start, mmap_mem()
H A Dmmtimer.c441 if (vma->vm_end - vma->vm_start != PAGE_SIZE) mmtimer_mmap()
456 if (remap_pfn_range(vma, vma->vm_start, mmtimer_addr >> PAGE_SHIFT, mmtimer_mmap()
/linux-4.1.27/arch/arm/mm/
H A Dmmap.c92 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area()
143 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area_topdown()
H A Dfault-armv.c141 pgoff = vma->vm_pgoff + ((addr - vma->vm_start) >> PAGE_SHIFT); make_coherent()
160 aliases += adjust_pte(mpnt, mpnt->vm_start + offset, pfn); make_coherent()
H A Dflush.c252 flush_cache_page(mpnt, mpnt->vm_start + offset, page_to_pfn(page)); __flush_dcache_aliases()
/linux-4.1.27/sound/soc/pxa/
H A Dmmp-pcm.c129 return remap_pfn_range(vma, vma->vm_start, mmp_pcm_mmap()
131 vma->vm_end - vma->vm_start, vma->vm_page_prot); mmp_pcm_mmap()
/linux-4.1.27/arch/xtensa/kernel/
H A Dsyscall.c90 if (!vmm || addr + len <= vmm->vm_start) arch_get_unmapped_area()
H A Dpci.c358 * is described by vm_start and vm_end members of VMA, the base physical
377 ret = io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_page_range()
378 vma->vm_end - vma->vm_start,vma->vm_page_prot); pci_mmap_page_range()
/linux-4.1.27/arch/sh/kernel/
H A Dsys_sh.c73 if (vma == NULL || addr < vma->vm_start || addr + len > vma->vm_end) { sys_cacheflush()
/linux-4.1.27/drivers/vfio/platform/
H A Dvfio_platform_common.c422 req_len = vma->vm_end - vma->vm_start; vfio_platform_mmap_mmio()
433 return remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, vfio_platform_mmap_mmio()
444 if (vma->vm_end < vma->vm_start) vfio_platform_mmap()
450 if (vma->vm_start & ~PAGE_MASK) vfio_platform_mmap()
/linux-4.1.27/arch/mips/kernel/
H A Dvdso.c119 if (vma->vm_mm && vma->vm_start == (long)vma->vm_mm->context.vdso) arch_vma_name()
/linux-4.1.27/arch/hexagon/kernel/
H A Dvdso.c97 if (vma->vm_mm && vma->vm_start == (long)vma->vm_mm->context.vdso) arch_vma_name()
/linux-4.1.27/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_doorbell.c129 if (vma->vm_end - vma->vm_start != doorbell_process_allocation()) kfd_doorbell_mmap()
150 (unsigned long long) vma->vm_start, address, vma->vm_flags, kfd_doorbell_mmap()
155 vma->vm_start, kfd_doorbell_mmap()
/linux-4.1.27/arch/metag/mm/
H A Dhugetlbpage.c50 if (addr + len > vma->vm_start) prepare_hugepage_range()
53 (ALIGN_HUGEPT(addr + len) > vma->vm_start)) prepare_hugepage_range()
162 <= vma->vm_start)) { hugetlb_get_unmapped_area_existing()
H A Dfault.c118 if (!vma || address < vma->vm_start) do_page_fault()
/linux-4.1.27/arch/ia64/mm/
H A Dfault.c138 if (( !vma && prev_vma ) || (address < vma->vm_start) ) ia64_do_page_fault()
211 if (REGION_NUMBER(address) != REGION_NUMBER(vma->vm_start) ia64_do_page_fault()
218 if (REGION_NUMBER(address) != REGION_NUMBER(vma->vm_start) ia64_do_page_fault()
H A Dinit.c120 vma->vm_start = current->thread.rbs_bot & PAGE_MASK; ia64_init_addr_space()
121 vma->vm_end = vma->vm_start + PAGE_SIZE; ia64_init_addr_space()
286 gate_vma.vm_start = FIXADDR_USER_START; gate_vma_init()
/linux-4.1.27/sound/soc/blackfin/
H A Dbf5xx-i2s-pcm.c220 size_t size = vma->vm_end - vma->vm_start; bf5xx_pcm_mmap()
221 vma->vm_start = (unsigned long)runtime->dma_area; bf5xx_pcm_mmap()
222 vma->vm_end = vma->vm_start + size; bf5xx_pcm_mmap()
H A Dbf5xx-ac97-pcm.c275 size_t size = vma->vm_end - vma->vm_start; bf5xx_pcm_mmap()
276 vma->vm_start = (unsigned long)runtime->dma_area; bf5xx_pcm_mmap()
277 vma->vm_end = vma->vm_start + size; bf5xx_pcm_mmap()
/linux-4.1.27/arch/parisc/kernel/
H A Dcache.c314 addr = mpnt->vm_start + offset; flush_dcache_page()
495 usize += vma->vm_end - vma->vm_start; mm_total_size()
528 flush_user_dcache_range_asm(vma->vm_start, vma->vm_end); flush_cache_mm()
531 flush_user_icache_range_asm(vma->vm_start, vma->vm_end); flush_cache_mm()
540 for (addr = vma->vm_start; addr < vma->vm_end; flush_cache_mm()
H A Dpci.c246 return remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_page_range()
247 vma->vm_end - vma->vm_start, vma->vm_page_prot); pci_mmap_page_range()
H A Dsys_parisc.c120 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area()
180 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area_topdown()
/linux-4.1.27/arch/ia64/pci/
H A Dpci.c593 unsigned long size = vma->vm_end - vma->vm_start; pci_mmap_page_range()
623 efi_range_is_wc(vma->vm_start, vma->vm_end - vma->vm_start)) pci_mmap_page_range()
628 if (remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_page_range()
629 vma->vm_end - vma->vm_start, vma->vm_page_prot)) pci_mmap_page_range()
664 unsigned long size = vma->vm_end - vma->vm_start; pci_mmap_legacy_page_range()
688 if (remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_legacy_page_range()
/linux-4.1.27/arch/arm64/mm/
H A Ddma-mapping.c311 unsigned long nr_vma_pages = (vma->vm_end - vma->vm_start) >> __dma_common_mmap()
321 ret = remap_pfn_range(vma, vma->vm_start, __dma_common_mmap()
323 vma->vm_end - vma->vm_start, __dma_common_mmap()
/linux-4.1.27/fs/
H A Dbinfmt_elf_fdpic.c1197 kdcore("%08lx: %08lx: no (IO)", vma->vm_start, vma->vm_flags); maydump()
1205 kdcore("%08lx: %08lx: no (!read)", vma->vm_start, vma->vm_flags); maydump()
1213 kdcore("%08lx: %08lx: %s (share)", vma->vm_start, maydump()
1219 kdcore("%08lx: %08lx: %s (share)", vma->vm_start, maydump()
1228 kdcore("%08lx: %08lx: %s (!anon)", vma->vm_start, maydump()
1235 kdcore("%08lx: %08lx: %s", vma->vm_start, vma->vm_flags, maydump()
1488 for (addr = vma->vm_start; addr < vma->vm_end; elf_fdpic_dump_segments()
1504 if (!dump_emit(cprm, (void *) vma->vm_start, elf_fdpic_dump_segments()
1505 vma->vm_end - vma->vm_start)) elf_fdpic_dump_segments()
1519 size += vma->vm_end - vma->vm_start; elf_core_vma_data_size()
1712 sz = vma->vm_end - vma->vm_start; elf_fdpic_core_dump()
1716 phdr.p_vaddr = vma->vm_start; elf_fdpic_core_dump()
H A Dexec.c201 unsigned long size = bprm->vma->vm_end - bprm->vma->vm_start; get_arg_page()
270 vma->vm_start = vma->vm_end - PAGE_SIZE; __bprm_mm_init()
579 unsigned long old_start = vma->vm_start; shift_arg_pages()
616 vma->vm_next ? vma->vm_next->vm_start : USER_PGTABLES_CEILING); shift_arg_pages()
625 vma->vm_next ? vma->vm_next->vm_start : USER_PGTABLES_CEILING); shift_arg_pages()
666 if (vma->vm_end - vma->vm_start > stack_base) setup_arg_pages()
671 stack_shift = vma->vm_start - stack_base; setup_arg_pages()
679 unlikely(vma->vm_end - vma->vm_start >= stack_top - mmap_min_addr)) setup_arg_pages()
707 ret = mprotect_fixup(vma, &prev, vma->vm_start, vma->vm_end, setup_arg_pages()
724 stack_size = vma->vm_end - vma->vm_start; setup_arg_pages()
732 stack_base = vma->vm_start + rlim_stack; setup_arg_pages()
739 stack_base = vma->vm_start - stack_expand; setup_arg_pages()
H A Dbinfmt_elf.c1276 u32 __user *header = (u32 __user *) vma->vm_start; vma_dump_size()
1308 return vma->vm_end - vma->vm_start; vma_dump_size()
1550 *start_end_ofs++ = vma->vm_start; fill_files_note()
2235 phdr.p_vaddr = vma->vm_start; elf_core_dump()
2238 phdr.p_memsz = vma->vm_end - vma->vm_start; elf_core_dump()
2270 end = vma->vm_start + vma_filesz[i++]; elf_core_dump()
2272 for (addr = vma->vm_start; addr < end; addr += PAGE_SIZE) { elf_core_dump()
/linux-4.1.27/drivers/auxdisplay/
H A Dcfag12864bfb.c69 return vm_insert_page(vma, vma->vm_start, cfag12864bfb_mmap()
/linux-4.1.27/ipc/
H A Dshm.c1309 (vma->vm_start - addr)/PAGE_SIZE == vma->vm_pgoff) { SYSCALL_DEFINE1()
1319 do_munmap(mm, vma->vm_start, vma->vm_end - vma->vm_start); SYSCALL_DEFINE1()
1344 ((vma->vm_start - addr)/PAGE_SIZE == vma->vm_pgoff) && SYSCALL_DEFINE1()
1346 do_munmap(mm, vma->vm_start, vma->vm_end - vma->vm_start); SYSCALL_DEFINE1()
1353 if (vma && vma->vm_start == addr && vma->vm_ops == &shm_vm_ops) { SYSCALL_DEFINE1()
1354 do_munmap(mm, vma->vm_start, vma->vm_end - vma->vm_start); SYSCALL_DEFINE1()
/linux-4.1.27/arch/sh/include/asm/
H A Dtlb.h77 flush_cache_range(vma, vma->vm_start, vma->vm_end); tlb_start_vma()
/linux-4.1.27/arch/nios2/mm/
H A Dinit.c141 return (vma->vm_start == KUSER_BASE) ? "[kuser]" : NULL; arch_vma_name()
H A Dcacheflush.c107 flush_cache_page(mpnt, mpnt->vm_start + offset, flush_aliases()
H A Dfault.c96 if (vma->vm_start <= address) do_page_fault()
/linux-4.1.27/arch/frv/mm/
H A Delf-fdpic.c77 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area()
H A Dfault.c92 if (vma->vm_start <= ear0) do_page_fault()
/linux-4.1.27/arch/x86/kernel/
H A Dsys_x86_64.c146 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area()
189 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area_topdown()
H A Dvsyscall_64.c284 .vm_start = VSYSCALL_ADDR,
309 return (addr >= vma->vm_start) && (addr < vma->vm_end); in_gate_area()
/linux-4.1.27/drivers/base/
H A Ddma-coherent.c255 int user_count = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; dma_mmap_from_coherent()
261 *ret = remap_pfn_range(vma, vma->vm_start, pfn, dma_mmap_from_coherent()
H A Ddma-mapping.c251 unsigned long user_count = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; dma_common_mmap()
262 ret = remap_pfn_range(vma, vma->vm_start, dma_common_mmap()
/linux-4.1.27/drivers/video/fbdev/
H A Dsbuslib.c54 size = vma->vm_end - vma->vm_start; sbusfb_mmap_helper()
85 vma->vm_start + page, sbusfb_mmap_helper()
H A Damba-clcd.c412 if (off <= len && vma->vm_end - vma->vm_start <= len - off && clcdfb_mmap()
735 user_size = vma->vm_end - vma->vm_start; clcdfb_of_vram_mmap()
741 return remap_pfn_range(vma, vma->vm_start, clcdfb_of_vram_mmap()
H A Digafb.c232 size = vma->vm_end - vma->vm_start; igafb_mmap()
261 if (remap_pfn_range(vma, vma->vm_start + page, igafb_mmap()
H A Dpxa3xx-gcu.c484 unsigned int size = vma->vm_end - vma->vm_start; pxa3xx_gcu_mmap()
504 return io_remap_pfn_range(vma, vma->vm_start, pxa3xx_gcu_mmap()
H A Dvfb.c417 unsigned long start = vma->vm_start; vfb_mmap()
418 unsigned long size = vma->vm_end - vma->vm_start; vfb_mmap()
/linux-4.1.27/arch/s390/mm/
H A Dmmap.c107 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area()
147 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area_topdown()
/linux-4.1.27/arch/sh/drivers/pci/
H A Dpci.c288 return remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_page_range()
289 vma->vm_end - vma->vm_start, pci_mmap_page_range()
/linux-4.1.27/arch/mips/pci/
H A Dpci.c345 return remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_page_range()
346 vma->vm_end - vma->vm_start, vma->vm_page_prot); pci_mmap_page_range()
/linux-4.1.27/arch/mn10300/unit-asb2305/
H A Dpci-asb2305.c231 if (io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_page_range()
232 vma->vm_end - vma->vm_start, pci_mmap_page_range()
/linux-4.1.27/sound/usb/usx2y/
H A DusX2Yhwdep.c42 area->vm_start, snd_us428ctls_vm_fault()
63 unsigned long size = (unsigned long)(area->vm_end - area->vm_start); snd_us428ctls_mmap()
/linux-4.1.27/sound/soc/samsung/
H A Didma.c252 size = vma->vm_end - vma->vm_start; idma_mmap()
254 ret = io_remap_pfn_range(vma, vma->vm_start, idma_mmap()
/linux-4.1.27/arch/arm/kvm/
H A Dmmu.c766 hva_t vm_start, vm_end; stage2_unmap_memslot() local
768 if (!vma || vma->vm_start >= reg_end) stage2_unmap_memslot()
774 vm_start = max(hva, vma->vm_start); stage2_unmap_memslot()
778 gpa_t gpa = addr + (vm_start - memslot->userspace_addr); stage2_unmap_memslot()
779 unmap_stage2_range(kvm, gpa, vm_end - vm_start); stage2_unmap_memslot()
1774 hva_t vm_start, vm_end; kvm_arch_prepare_memory_region() local
1776 if (!vma || vma->vm_start >= reg_end) kvm_arch_prepare_memory_region()
1791 vm_start = max(hva, vma->vm_start); kvm_arch_prepare_memory_region()
1796 (vm_start - mem->userspace_addr); kvm_arch_prepare_memory_region()
1800 pa += vm_start - vma->vm_start; kvm_arch_prepare_memory_region()
1807 vm_end - vm_start, kvm_arch_prepare_memory_region()
/linux-4.1.27/drivers/infiniband/hw/cxgb4/
H A Dprovider.c164 int len = vma->vm_end - vma->vm_start; c4iw_mmap()
175 if (vma->vm_start & (PAGE_SIZE-1)) c4iw_mmap()
195 ret = io_remap_pfn_range(vma, vma->vm_start, c4iw_mmap()
215 ret = io_remap_pfn_range(vma, vma->vm_start, c4iw_mmap()
223 ret = remap_pfn_range(vma, vma->vm_start, c4iw_mmap()
/linux-4.1.27/drivers/infiniband/hw/ipath/
H A Dipath_file_ops.c993 if ((vma->vm_end - vma->vm_start) > len) { ipath_mmap_mem()
996 vma->vm_end - vma->vm_start, len); ipath_mmap_mem()
1014 ret = remap_pfn_range(vma, vma->vm_start, pfn, ipath_mmap_mem()
1039 if ((vma->vm_end - vma->vm_start) > PAGE_SIZE) { mmap_ureg()
1041 "%lx > PAGE\n", vma->vm_end - vma->vm_start); mmap_ureg()
1048 ret = io_remap_pfn_range(vma, vma->vm_start, mmap_ureg()
1050 vma->vm_end - vma->vm_start, mmap_ureg()
1070 if ((vma->vm_end - vma->vm_start) > (piocnt * dd->ipath_palign)) { mmap_piobufs()
1073 vma->vm_end - vma->vm_start); mmap_piobufs()
1094 ret = io_remap_pfn_range(vma, vma->vm_start, phys >> PAGE_SHIFT, mmap_piobufs()
1095 vma->vm_end - vma->vm_start, mmap_piobufs()
1112 if ((vma->vm_end - vma->vm_start) > total_size) { mmap_rcvegrbufs()
1115 vma->vm_end - vma->vm_start, mmap_rcvegrbufs()
1130 start = vma->vm_start; mmap_rcvegrbufs()
1224 len = vma->vm_end - vma->vm_start; mmap_kvaddr()
1284 ipath_cdbg(MM, "pgaddr %llx vm_start=%lx len %lx port %u:%u:%u\n", ipath_mmap()
1285 (unsigned long long) pgaddr, vma->vm_start, ipath_mmap()
1286 vma->vm_end - vma->vm_start, dd->ipath_unit, ipath_mmap()
1354 vma->vm_end - vma->vm_start); ipath_mmap()
H A Dipath_mmap.c93 unsigned long size = vma->vm_end - vma->vm_start; ipath_mmap()
/linux-4.1.27/sound/oss/
H A Dsoundcard.c468 size = vma->vm_end - vma->vm_start; sound_mmap()
473 if (remap_pfn_range(vma, vma->vm_start, sound_mmap()
475 vma->vm_end - vma->vm_start, vma->vm_page_prot)) { sound_mmap()
/linux-4.1.27/drivers/android/
H A Dbinder.c2813 proc->pid, vma->vm_start, vma->vm_end, binder_vma_open()
2814 (vma->vm_end - vma->vm_start) / SZ_1K, vma->vm_flags, binder_vma_open()
2824 proc->pid, vma->vm_start, vma->vm_end, binder_vma_close()
2825 (vma->vm_end - vma->vm_start) / SZ_1K, vma->vm_flags, binder_vma_close()
2854 if ((vma->vm_end - vma->vm_start) > SZ_4M) binder_mmap()
2855 vma->vm_end = vma->vm_start + SZ_4M; binder_mmap()
2859 proc->pid, vma->vm_start, vma->vm_end, binder_mmap()
2860 (vma->vm_end - vma->vm_start) / SZ_1K, vma->vm_flags, binder_mmap()
2877 area = get_vm_area(vma->vm_end - vma->vm_start, VM_IOREMAP); binder_mmap()
2884 proc->user_buffer_offset = vma->vm_start - (uintptr_t)proc->buffer; binder_mmap()
2889 while (CACHE_COLOUR((vma->vm_start ^ (uint32_t)proc->buffer))) { binder_mmap()
2890 pr_info("binder_mmap: %d %lx-%lx maps %p bad alignment\n", proc->pid, vma->vm_start, vma->vm_end, proc->buffer); binder_mmap()
2891 vma->vm_start += PAGE_SIZE; binder_mmap()
2895 proc->pages = kzalloc(sizeof(proc->pages[0]) * ((vma->vm_end - vma->vm_start) / PAGE_SIZE), GFP_KERNEL); binder_mmap()
2901 proc->buffer_size = vma->vm_end - vma->vm_start; binder_mmap()
2923 proc->pid, vma->vm_start, vma->vm_end, proc->buffer);*/ binder_mmap()
2938 proc->pid, vma->vm_start, vma->vm_end, failure_string, ret); binder_mmap()
/linux-4.1.27/drivers/misc/sgi-gru/
H A Dgrukdump.c155 hdr.vaddr = gts->ts_vma->vm_start; gru_dump_context()
H A Dgrufile.c112 if (vma->vm_start & (GRU_GSEG_PAGESIZE - 1) || gru_file_mmap()
126 file, vma->vm_start, vma, vma->vm_private_data); gru_file_mmap()
H A Dgrutables.h398 #define TSID(a, v) (((a) - (v)->vm_start) / GRU_GSEG_PAGESIZE)
399 #define UGRUADDR(gts) ((gts)->ts_vma->vm_start + \
/linux-4.1.27/drivers/infiniband/hw/qib/
H A Dqib_file_ops.c736 if ((vma->vm_end - vma->vm_start) > len) { qib_mmap_mem()
739 vma->vm_end - vma->vm_start, len); qib_mmap_mem()
761 ret = remap_pfn_range(vma, vma->vm_start, pfn, qib_mmap_mem()
784 if ((vma->vm_end - vma->vm_start) > sz) { mmap_ureg()
787 vma->vm_end - vma->vm_start); mmap_ureg()
794 ret = io_remap_pfn_range(vma, vma->vm_start, mmap_ureg()
796 vma->vm_end - vma->vm_start, mmap_ureg()
816 if ((vma->vm_end - vma->vm_start) > (piocnt * dd->palign)) { mmap_piobufs()
819 vma->vm_end - vma->vm_start); mmap_piobufs()
844 ret = io_remap_pfn_range(vma, vma->vm_start, phys >> PAGE_SHIFT, mmap_piobufs()
845 vma->vm_end - vma->vm_start, mmap_piobufs()
862 if ((vma->vm_end - vma->vm_start) > total_size) { mmap_rcvegrbufs()
865 vma->vm_end - vma->vm_start, mmap_rcvegrbufs()
881 start = vma->vm_start; mmap_rcvegrbufs()
974 len = vma->vm_end - vma->vm_start; mmap_kvaddr()
1100 vma->vm_end - vma->vm_start); qib_mmapf()
H A Dqib_mmap.c93 unsigned long size = vma->vm_end - vma->vm_start; qib_mmap()
/linux-4.1.27/arch/x86/mm/
H A Dhugetlbpage.c147 (!vma || addr + len <= vma->vm_start)) hugetlb_get_unmapped_area()
H A Dmpx.c624 * Find the first overlapping vma. If vma->vm_start > start, there zap_bt_entries()
629 if (!vma || vma->vm_start > start) zap_bt_entries()
639 while (vma && vma->vm_start < end) { zap_bt_entries()
653 addr = vma->vm_start; zap_bt_entries()
783 if (next && (mm->bd_addr + MPX_GET_BD_ENTRY_OFFSET(next->vm_start)) unmap_edge_bts()
911 } while (vma && vma->vm_start < end); mpx_notify_unmap()
H A Dpat.c807 unsigned long vma_size = vma->vm_end - vma->vm_start; track_pfn_copy()
815 if (follow_phys(vma, vma->vm_start, 0, &prot, &paddr)) { track_pfn_copy()
838 if (addr == vma->vm_start && size == (vma->vm_end - vma->vm_start)) { track_pfn_remap()
903 if (follow_phys(vma, vma->vm_start, 0, &prot, &paddr)) { untrack_pfn()
908 size = vma->vm_end - vma->vm_start; untrack_pfn()
/linux-4.1.27/arch/sparc/mm/
H A Dhugetlbpage.c118 (!vma || addr + len <= vma->vm_start)) hugetlb_get_unmapped_area()
H A Dfault_32.c213 if (vma->vm_start <= address) do_sparc_fault()
398 if (vma->vm_start <= address) force_user_fault()
/linux-4.1.27/drivers/char/agp/
H A Dalpha-agp.c22 dma_addr = (unsigned long)vmf->virtual_address - vma->vm_start alpha_core_agp_vm_fault()
H A Dfrontend.c582 size = vma->vm_end - vma->vm_start; agp_mmap()
603 } else if (io_remap_pfn_range(vma, vma->vm_start, agp_mmap()
620 } else if (io_remap_pfn_range(vma, vma->vm_start, agp_mmap()
/linux-4.1.27/drivers/gpu/drm/udl/
H A Dudl_gem.c110 page_offset = ((unsigned long)vmf->virtual_address - vma->vm_start) >> udl_gem_fault()
H A Dudl_fb.c257 unsigned long start = vma->vm_start; udl_fb_mmap()
258 unsigned long size = vma->vm_end - vma->vm_start; udl_fb_mmap()
/linux-4.1.27/drivers/uio/
H A Duio.c646 if (vma->vm_end - vma->vm_start > mem->size) uio_mmap_physical()
662 vma->vm_start, uio_mmap_physical()
664 vma->vm_end - vma->vm_start, uio_mmap_physical()
676 if (vma->vm_end < vma->vm_start) uio_mmap()
/linux-4.1.27/arch/mips/mm/
H A Dmmap.c95 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area_common()
H A Dfault.c107 if (vma->vm_start <= address) __do_page_fault()
/linux-4.1.27/arch/powerpc/kvm/
H A De500_mmu_host.c363 if (vma && hva >= vma->vm_start && kvmppc_e500_shadow_map()
379 ((vma->vm_end - vma->vm_start) >> PAGE_SHIFT); kvmppc_e500_shadow_map()
381 pfn = start + ((hva - vma->vm_start) >> PAGE_SHIFT); kvmppc_e500_shadow_map()
426 } else if (vma && hva >= vma->vm_start && kvmppc_e500_shadow_map()
/linux-4.1.27/arch/hexagon/mm/
H A Dvm_fault.c75 if (vma->vm_start <= address) do_page_fault()
/linux-4.1.27/arch/m68k/mm/
H A Dfault.c97 if (vma->vm_start <= address) do_page_fault()
/linux-4.1.27/arch/x86/pci/
H A Di386.c447 if (io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_page_range()
448 vma->vm_end - vma->vm_start, pci_mmap_page_range()
/linux-4.1.27/arch/unicore32/kernel/
H A Dpci.c380 if (remap_pfn_range(vma, vma->vm_start, phys, pci_mmap_page_range()
381 vma->vm_end - vma->vm_start, pci_mmap_page_range()
H A Dprocess.c327 return (vma->vm_start == 0xffff0000) ? "[vectors]" : NULL; arch_vma_name()
/linux-4.1.27/arch/mn10300/mm/
H A Dfault.c182 if (vma->vm_start <= address) do_page_fault()
197 vma->vm_start, vma->vm_end); do_page_fault()
/linux-4.1.27/arch/arm/kernel/
H A Dprocess.c300 .vm_start = 0xffff0000,
319 return (addr >= gate_vma.vm_start) && (addr < gate_vma.vm_end); in_gate_area()
H A Dbios32.c629 if (remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_page_range()
630 vma->vm_end - vma->vm_start, pci_mmap_page_range()
/linux-4.1.27/arch/arm/mach-integrator/
H A Dimpd1.c244 size = vma->vm_end - vma->vm_start; impd1fb_clcd_mmap()
246 return remap_pfn_range(vma, vma->vm_start, start, size, impd1fb_clcd_mmap()
/linux-4.1.27/sound/soc/atmel/
H A Datmel-pcm-pdc.c75 return remap_pfn_range(vma, vma->vm_start, atmel_pcm_mmap()
77 vma->vm_end - vma->vm_start, vma->vm_page_prot); atmel_pcm_mmap()
/linux-4.1.27/drivers/misc/genwqe/
H A Dcard_dev.c397 unsigned long vsize = vma->vm_end - vma->vm_start; genwqe_vma_close()
406 dma_map = __genwqe_search_mapping(cfile, vma->vm_start, vsize, genwqe_vma_close()
411 __func__, vma->vm_start, vma->vm_pgoff << PAGE_SHIFT, genwqe_vma_close()
441 unsigned long pfn, vsize = vma->vm_end - vma->vm_start; genwqe_mmap()
457 dma_map->u_vaddr = (void *)vma->vm_start; genwqe_mmap()
472 vma->vm_start, genwqe_mmap()
/linux-4.1.27/drivers/firewire/
H A Dcore-cdev.c74 unsigned long vm_start; member in struct:client
1085 payload = (unsigned long)a->data - client->vm_start; ioctl_queue_iso()
1157 a->data = client->vm_start + payload; ioctl_queue_iso()
1687 if (vma->vm_start & ~PAGE_MASK) fw_device_op_mmap()
1690 client->vm_start = vma->vm_start; fw_device_op_mmap()
1691 size = vma->vm_end - vma->vm_start; fw_device_op_mmap()
H A Dcore-iso.c113 uaddr = vma->vm_start; fw_iso_buffer_map_vma()
/linux-4.1.27/arch/sparc/kernel/
H A Dpci.c747 user_size = vma->vm_end - vma->vm_start; __pci_mmap_make_offset_bus()
799 user_size = vma->vm_end - vma->vm_start; __pci_mmap_make_offset()
849 * for this architecture. The region in the process to map is described by vm_start
869 ret = io_remap_pfn_range(vma, vma->vm_start, pci_mmap_page_range()
871 vma->vm_end - vma->vm_start, pci_mmap_page_range()
H A Dsys_sparc_64.c121 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area()
184 (!vma || addr + len <= vma->vm_start)) arch_get_unmapped_area_topdown()
/linux-4.1.27/drivers/infiniband/hw/usnic/
H A Dusnic_ib_verbs.c693 if ((vma->vm_end - vma->vm_start) != bar->len) { usnic_ib_mmap()
696 vma->vm_end - vma->vm_start); usnic_ib_mmap()
707 vma->vm_start, usnic_ib_mmap()
/linux-4.1.27/drivers/dma-buf/
H A Ddma-buf.c89 if (vma->vm_pgoff + ((vma->vm_end - vma->vm_start) >> PAGE_SHIFT) > dma_buf_mmap_internal()
671 if (pgoff + ((vma->vm_end - vma->vm_start) >> PAGE_SHIFT) < pgoff) dma_buf_mmap()
675 if (pgoff + ((vma->vm_end - vma->vm_start) >> PAGE_SHIFT) > dma_buf_mmap()
/linux-4.1.27/tools/vm/
H A Dpage-types.c784 unsigned long vm_start; parse_pid() local
793 &vm_start, parse_pid()
803 pg_start[nr_vmas] = vm_start / page_size; parse_pid()
/linux-4.1.27/include/linux/
H A Dmm_types.h230 unsigned long vm_start; /* start address of region */ member in struct:vm_region
233 unsigned long vm_pgoff; /* the offset in vm_file corresponding to vm_start */
250 unsigned long vm_start; /* Our start address within vm_mm. */ member in struct:vm_area_struct
H A Dmm.h1260 (vma->vm_start == addr) && stack_guard_page_start()
1267 return vma && (vma->vm_start == addr) && (vma->vm_flags & VM_GROWSUP); vma_growsup()
1991 if (vma && end_addr <= vma->vm_start) find_vma_intersection()
1998 return (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; vma_pages()
2001 /* Look up the first VMA which exactly match the interval vm_start ... vm_end */ find_exact_vma()
2003 unsigned long vm_start, unsigned long vm_end) find_exact_vma()
2005 struct vm_area_struct *vma = find_vma(mm, vm_start); find_exact_vma()
2007 if (vma && (vma->vm_start != vm_start || vma->vm_end != vm_end)) find_exact_vma()
2002 find_exact_vma(struct mm_struct *mm, unsigned long vm_start, unsigned long vm_end) find_exact_vma() argument
/linux-4.1.27/arch/microblaze/pci/
H A Dpci-common.c309 * is described by vm_start and vm_end members of VMA, the base physical
333 ret = remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_page_range()
334 vma->vm_end - vma->vm_start, vma->vm_page_prot); pci_mmap_page_range()
432 resource_size_t size = vma->vm_end - vma->vm_start; pci_mmap_legacy_page_range()
476 return remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, pci_mmap_legacy_page_range()
477 vma->vm_end - vma->vm_start, pci_mmap_legacy_page_range()
/linux-4.1.27/drivers/staging/lustre/lustre/llite/
H A Dllite_mmap.c60 policy->l_extent.start = ((addr - vma->vm_start) & CFS_PAGE_MASK) + policy_from_vma()
75 vma != NULL && vma->vm_start < (addr + count); vma = vma->vm_next) { our_vma()
/linux-4.1.27/drivers/oprofile/
H A Dbuffer_sync.c260 if (addr < vma->vm_start || addr >= vma->vm_end) lookup_dcookie()
266 vma->vm_start; lookup_dcookie()
/linux-4.1.27/drivers/gpu/drm/exynos/
H A Dexynos_drm_gem.c332 vm_size = vma->vm_end - vma->vm_start; exynos_drm_gem_mmap_buffer()
604 vma->vm_start) >> PAGE_SHIFT; exynos_drm_gem_fault()
H A Dexynos_drm_fbdev.c50 vm_size = vma->vm_end - vma->vm_start; exynos_drm_fb_mmap()
/linux-4.1.27/arch/um/kernel/
H A Dtlb.c530 fix_range(mm, vma->vm_start, vma->vm_end, 0); flush_tlb_mm()
541 fix_range(mm, vma->vm_start, vma->vm_end, 1); force_flush_all()
H A Dtrap.c51 else if (vma->vm_start <= address) handle_page_fault()
/linux-4.1.27/kernel/events/
H A Duprobes.c114 * We keep the vma's vm_start rather than a pointer to the vma
141 return vma->vm_start + offset - ((loff_t)vma->vm_pgoff << PAGE_SHIFT); offset_to_vaddr()
146 return ((loff_t)vma->vm_pgoff << PAGE_SHIFT) + (vaddr - vma->vm_start); vaddr_to_offset()
814 if (vma->vm_start > info->vaddr || register_for_each_vma()
982 uprobe->offset >= offset + vma->vm_end - vma->vm_start) unapply_uprobe()
1075 build_probe_list(inode, vma, vma->vm_start, vma->vm_end, &tmp_list); uprobe_mmap()
1662 if (vma_has_uprobes(vma, vma->vm_start, vma->vm_end)) mmf_recalc_uprobes()
1702 if (vma && vma->vm_start <= bp_vaddr) { find_active_uprobe()
/linux-4.1.27/drivers/misc/cxl/
H A Dcontext.c148 u64 len = vma->vm_end - vma->vm_start; cxl_context_iomap()
H A Dfault.c278 for (ea = vma->vm_start; ea < vma->vm_end; cxl_prefault_vma()
/linux-4.1.27/drivers/staging/android/ion/
H A Dion_heap.c75 unsigned long addr = vma->vm_start; ion_heap_map_user()
/linux-4.1.27/drivers/gpu/drm/gma500/
H A Dgem.c207 page_offset = ((unsigned long) vmf->virtual_address - vma->vm_start) psb_gem_fault()
/linux-4.1.27/arch/xtensa/mm/
H A Dfault.c83 if (vma->vm_start <= address) do_page_fault()
/linux-4.1.27/drivers/gpu/drm/vgem/
H A Dvgem_drv.c98 page_offset = ((unsigned long)vmf->virtual_address - vma->vm_start) >> vgem_gem_fault()
/linux-4.1.27/arch/s390/kernel/
H A Dvdso.c249 if (vma->vm_mm && vma->vm_start == vma->vm_mm->context.vdso_base) arch_vma_name()
/linux-4.1.27/arch/score/mm/
H A Dfault.c86 if (vma->vm_start <= address) do_page_fault()
/linux-4.1.27/arch/avr32/mm/
H A Dfault.c97 if (vma->vm_start <= address) do_page_fault()
/linux-4.1.27/arch/arm/include/asm/
H A Dtlb.h198 flush_cache_range(vma, vma->vm_start, vma->vm_end); tlb_start_vma()
/linux-4.1.27/arch/alpha/mm/
H A Dfault.c124 if (vma->vm_start <= address) do_page_fault()
/linux-4.1.27/drivers/media/platform/omap/
H A Domap_vout.c216 physp = (vma->vm_pgoff << PAGE_SHIFT) + (virtp - vma->vm_start); omap_vout_uservirt_to_phys()
866 "vm_open [vma=%08lx-%08lx]\n", vma->vm_start, vma->vm_end); omap_vout_vm_open()
875 "vm_close [vma=%08lx-%08lx]\n", vma->vm_start, vma->vm_end); omap_vout_vm_close()
888 unsigned long start = vma->vm_start; omap_vout_mmap()
889 unsigned long size = (vma->vm_end - vma->vm_start); omap_vout_mmap()
895 vma->vm_pgoff, vma->vm_start, vma->vm_end); omap_vout_mmap()
921 q->bufs[i]->baddr = vma->vm_start; omap_vout_mmap()
/linux-4.1.27/arch/powerpc/oprofile/cell/
H A Dspu_task_sync.c341 if (vma->vm_start > spu_ref || vma->vm_end <= spu_ref) get_exec_dcookie_and_offset()
343 my_offset = spu_ref - vma->vm_start; get_exec_dcookie_and_offset()
/linux-4.1.27/drivers/infiniband/hw/cxgb3/
H A Diwch_provider.c348 int len = vma->vm_end - vma->vm_start; iwch_mmap()
359 if (vma->vm_start & (PAGE_SIZE-1)) { iwch_mmap()
386 ret = io_remap_pfn_range(vma, vma->vm_start, iwch_mmap()
394 ret = remap_pfn_range(vma, vma->vm_start, iwch_mmap()
/linux-4.1.27/arch/microblaze/mm/
H A Dfault.c153 if (vma->vm_start <= address) do_page_fault()
/linux-4.1.27/drivers/gpu/drm/ttm/
H A Dttm_bo_vm.c177 page_offset = ((address - vma->vm_start) >> PAGE_SHIFT) + ttm_bo_vm_fault()
/linux-4.1.27/arch/openrisc/mm/
H A Dfault.c117 if (vma->vm_start <= address) do_page_fault()
/linux-4.1.27/arch/blackfin/kernel/
H A Dptrace.c130 valid = vma && start >= vma->vm_start && start + len <= vma->vm_end; is_user_addr_valid()
/linux-4.1.27/arch/cris/mm/
H A Dfault.c126 if (vma->vm_start <= address) do_page_fault()
/linux-4.1.27/arch/m68k/kernel/
H A Dsys_m68k.c404 if (!vma || addr < vma->vm_start || addr + len > vma->vm_end) sys_cacheflush()

Completed in 6520 milliseconds

12