Lines Matching refs:vm_start
141 return vma->vm_end - vma->vm_start; in kobjsize()
597 BUG_ON(unlikely(last->vm_end <= last->vm_start)); in validate_nommu_regions()
604 BUG_ON(unlikely(region->vm_end <= region->vm_start)); in validate_nommu_regions()
606 BUG_ON(unlikely(region->vm_start < last->vm_top)); in validate_nommu_regions()
632 if (region->vm_start < pregion->vm_start) in add_nommu_region()
634 else if (region->vm_start > pregion->vm_start) in add_nommu_region()
691 if (region->vm_top > region->vm_start) in __put_nommu_region()
702 free_page_series(region->vm_start, region->vm_top); in __put_nommu_region()
726 long start = vma->vm_start & PAGE_MASK; in protect_vma()
776 if (vma->vm_start < pvma->vm_start) in add_vma_to_mm()
778 else if (vma->vm_start > pvma->vm_start) { in add_vma_to_mm()
882 if (vma->vm_start > addr) in find_vma()
931 if (vma->vm_start < addr) in find_vma_exact()
933 if (vma->vm_start > addr) in find_vma_exact()
1226 region->vm_start = (unsigned long) base; in do_mmap_private()
1227 region->vm_end = region->vm_start + len; in do_mmap_private()
1228 region->vm_top = region->vm_start + (total << PAGE_SHIFT); in do_mmap_private()
1230 vma->vm_start = region->vm_start; in do_mmap_private()
1231 vma->vm_end = region->vm_start + len; in do_mmap_private()
1258 free_page_series(region->vm_start, region->vm_top); in do_mmap_private()
1259 region->vm_start = vma->vm_start = 0; in do_mmap_private()
1362 rpglen = pregion->vm_end - pregion->vm_start; in do_mmap_pgoff()
1381 start = pregion->vm_start; in do_mmap_pgoff()
1383 vma->vm_start = start; in do_mmap_pgoff()
1394 vma->vm_start = 0; in do_mmap_pgoff()
1429 vma->vm_start = region->vm_start = addr; in do_mmap_pgoff()
1450 memset((void *)region->vm_start, 0, in do_mmap_pgoff()
1451 region->vm_end - region->vm_start); in do_mmap_pgoff()
1454 result = vma->vm_start; in do_mmap_pgoff()
1464 flush_icache_range(region->vm_start, region->vm_end); in do_mmap_pgoff()
1591 npages = (addr - vma->vm_start) >> PAGE_SHIFT; in split_vma()
1596 region->vm_start = new->vm_start = addr; in split_vma()
1607 vma->vm_region->vm_start = vma->vm_start = addr; in split_vma()
1636 if (from > vma->vm_start) in shrink_vma()
1639 vma->vm_start = to; in shrink_vma()
1648 if (from > region->vm_start) { in shrink_vma()
1652 region->vm_start = to; in shrink_vma()
1698 if (start > vma->vm_start) { in do_munmap()
1710 if (start == vma->vm_start && end == vma->vm_end) in do_munmap()
1712 if (start < vma->vm_start || end > vma->vm_end) { in do_munmap()
1724 if (start != vma->vm_start && end != vma->vm_end) { in do_munmap()
1820 if (vma->vm_end != vma->vm_start + old_len) in do_mremap()
1826 if (new_len > vma->vm_region->vm_end - vma->vm_region->vm_start) in do_mremap()
1830 vma->vm_end = vma->vm_start + new_len; in do_mremap()
1831 return vma->vm_start; in do_mremap()
1868 unsigned long vm_len = vma->vm_end - vma->vm_start; in vm_iomap_memory()
1871 return io_remap_pfn_range(vma, vma->vm_start, pfn, vm_len, vma->vm_page_prot); in vm_iomap_memory()
1878 unsigned int size = vma->vm_end - vma->vm_start; in remap_vmalloc_range()
1883 vma->vm_start = (unsigned long)(addr + (pgoff << PAGE_SHIFT)); in remap_vmalloc_range()
1884 vma->vm_end = vma->vm_start + size; in remap_vmalloc_range()
2126 r_size = region->vm_top - region->vm_start; in nommu_shrink_inode_mappings()