Lines Matching refs:next
271 struct vm_area_struct *next = vma->vm_next; in remove_vma() local
280 return next; in remove_vma()
708 struct vm_area_struct *next; in __vma_unlink() local
711 prev->vm_next = next = vma->vm_next; in __vma_unlink()
712 if (next) in __vma_unlink()
713 next->vm_prev = prev; in __vma_unlink()
730 struct vm_area_struct *next = vma->vm_next; in vma_adjust() local
740 if (next && !insert) { in vma_adjust()
743 if (end >= next->vm_end) { in vma_adjust()
748 again: remove_next = 1 + (end > next->vm_end); in vma_adjust()
749 end = next->vm_end; in vma_adjust()
750 exporter = next; in vma_adjust()
752 } else if (end > next->vm_start) { in vma_adjust()
757 adjust_next = (end - next->vm_start) >> PAGE_SHIFT; in vma_adjust()
758 exporter = next; in vma_adjust()
768 importer = next; in vma_adjust()
792 uprobe_munmap(next, next->vm_start, next->vm_end); in vma_adjust()
810 anon_vma = next->anon_vma; in vma_adjust()
812 VM_BUG_ON_VMA(adjust_next && next->anon_vma && in vma_adjust()
813 anon_vma != next->anon_vma, next); in vma_adjust()
817 anon_vma_interval_tree_pre_update_vma(next); in vma_adjust()
824 vma_interval_tree_remove(next, root); in vma_adjust()
837 next->vm_start += adjust_next << PAGE_SHIFT; in vma_adjust()
838 next->vm_pgoff += adjust_next; in vma_adjust()
843 vma_interval_tree_insert(next, root); in vma_adjust()
853 __vma_unlink(mm, next, vma); in vma_adjust()
855 __remove_shared_vm_struct(next, file, mapping); in vma_adjust()
867 if (!next) in vma_adjust()
870 vma_gap_update(next); in vma_adjust()
877 anon_vma_interval_tree_post_update_vma(next); in vma_adjust()
887 uprobe_mmap(next); in vma_adjust()
892 uprobe_munmap(next, next->vm_start, next->vm_end); in vma_adjust()
895 if (next->anon_vma) in vma_adjust()
896 anon_vma_merge(vma, next); in vma_adjust()
898 mpol_put(vma_policy(next)); in vma_adjust()
899 kmem_cache_free(vm_area_cachep, next); in vma_adjust()
905 next = vma->vm_next; in vma_adjust()
908 else if (next) in vma_adjust()
909 vma_gap_update(next); in vma_adjust()
1039 struct vm_area_struct *area, *next; in vma_merge() local
1050 next = prev->vm_next; in vma_merge()
1052 next = mm->mmap; in vma_merge()
1053 area = next; in vma_merge()
1054 if (next && next->vm_end == end) /* cases 6, 7, 8 */ in vma_merge()
1055 next = next->vm_next; in vma_merge()
1067 if (next && end == next->vm_start && in vma_merge()
1068 mpol_equal(policy, vma_policy(next)) && in vma_merge()
1069 can_vma_merge_before(next, vm_flags, in vma_merge()
1072 next->anon_vma, NULL)) { in vma_merge()
1075 next->vm_end, prev->vm_pgoff, NULL); in vma_merge()
1088 if (next && end == next->vm_start && in vma_merge()
1089 mpol_equal(policy, vma_policy(next)) && in vma_merge()
1090 can_vma_merge_before(next, vm_flags, in vma_merge()
1096 err = vma_adjust(area, addr, next->vm_end, in vma_merge()
1097 next->vm_pgoff - pglen, NULL); in vma_merge()
2291 struct vm_area_struct *next; in expand_stack() local
2294 next = vma->vm_next; in expand_stack()
2295 if (next && next->vm_start == address + PAGE_SIZE) { in expand_stack()
2296 if (!(next->vm_flags & VM_GROWSUP)) in expand_stack()
2389 struct vm_area_struct *next = prev ? prev->vm_next : mm->mmap; in unmap_region() local
2397 next ? next->vm_start : USER_PGTABLES_CEILING); in unmap_region()
2661 struct vm_area_struct *next; in SYSCALL_DEFINE5() local
2663 for (next = vma->vm_next; next; next = next->vm_next) { in SYSCALL_DEFINE5()
2665 if (next->vm_start != next->vm_prev->vm_end) in SYSCALL_DEFINE5()
2668 if (next->vm_file != vma->vm_file) in SYSCALL_DEFINE5()
2671 if (next->vm_flags != vma->vm_flags) in SYSCALL_DEFINE5()
2674 if (start + size <= next->vm_end) in SYSCALL_DEFINE5()
2678 if (!next) in SYSCALL_DEFINE5()