Lines Matching refs:prev
62 struct vm_area_struct *vma, struct vm_area_struct *prev,
382 unsigned long prev = 0, pend = 0; in browse_rb() local
387 if (vma->vm_start < prev) { in browse_rb()
389 vma->vm_start, prev); in browse_rb()
410 prev = vma->vm_start; in browse_rb()
661 struct vm_area_struct *prev, struct rb_node **rb_link, in __vma_link() argument
664 __vma_link_list(mm, vma, prev, rb_parent); in __vma_link()
669 struct vm_area_struct *prev, struct rb_node **rb_link, in vma_link() argument
679 __vma_link(mm, vma, prev, rb_link, rb_parent); in vma_link()
695 struct vm_area_struct *prev; in __insert_vm_struct() local
699 &prev, &rb_link, &rb_parent)) in __insert_vm_struct()
701 __vma_link(mm, vma, prev, rb_link, rb_parent); in __insert_vm_struct()
707 struct vm_area_struct *prev) in __vma_unlink() argument
712 prev->vm_next = next = vma->vm_next; in __vma_unlink()
714 next->vm_prev = prev; in __vma_unlink()
1041 struct vm_area_struct *prev, unsigned long addr, in vma_merge() argument
1058 if (prev) in vma_merge()
1059 next = prev->vm_next; in vma_merge()
1069 if (prev && prev->vm_end == addr && in vma_merge()
1070 mpol_equal(vma_policy(prev), policy) && in vma_merge()
1071 can_vma_merge_after(prev, vm_flags, in vma_merge()
1083 is_mergeable_anon_vma(prev->anon_vma, in vma_merge()
1086 err = vma_adjust(prev, prev->vm_start, in vma_merge()
1087 next->vm_end, prev->vm_pgoff, NULL); in vma_merge()
1089 err = vma_adjust(prev, prev->vm_start, in vma_merge()
1090 end, prev->vm_pgoff, NULL); in vma_merge()
1093 khugepaged_enter_vma_merge(prev, vm_flags); in vma_merge()
1094 return prev; in vma_merge()
1105 if (prev && addr < prev->vm_end) /* case 4 */ in vma_merge()
1106 err = vma_adjust(prev, prev->vm_start, in vma_merge()
1107 addr, prev->vm_pgoff, NULL); in vma_merge()
1545 struct vm_area_struct *vma, *prev; in mmap_region() local
1568 while (find_vma_links(mm, addr, addr + len, &prev, &rb_link, in mmap_region()
1587 vma = vma_merge(mm, prev, addr, addr + len, vm_flags, in mmap_region()
1650 vma_link(mm, vma, prev, rb_link, rb_parent); in mmap_region()
1692 unmap_region(mm, vma, prev, vma->vm_start, vma->vm_end); in mmap_region()
1776 struct rb_node *prev = &vma->vm_rb; in unmapped_area() local
1777 if (!rb_parent(prev)) in unmapped_area()
1779 vma = rb_entry(rb_parent(prev), in unmapped_area()
1781 if (prev == vma->vm_rb.rb_left) { in unmapped_area()
1879 struct rb_node *prev = &vma->vm_rb; in unmapped_area_topdown() local
1880 if (!rb_parent(prev)) in unmapped_area_topdown()
1882 vma = rb_entry(rb_parent(prev), in unmapped_area_topdown()
1884 if (prev == vma->vm_rb.rb_right) { in unmapped_area_topdown()
2320 struct vm_area_struct *vma, *prev; in find_extend_vma() local
2323 vma = find_vma_prev(mm, addr, &prev); in find_extend_vma()
2326 if (!prev || expand_stack(prev, addr)) in find_extend_vma()
2328 if (prev->vm_flags & VM_LOCKED) in find_extend_vma()
2329 populate_vma_page_range(prev, addr, prev->vm_end, NULL); in find_extend_vma()
2330 return prev; in find_extend_vma()
2335 struct vm_area_struct *prev; in expand_stack() local
2338 prev = vma->vm_prev; in expand_stack()
2339 if (prev && prev->vm_end == address) { in expand_stack()
2340 if (!(prev->vm_flags & VM_GROWSDOWN)) in expand_stack()
2401 struct vm_area_struct *vma, struct vm_area_struct *prev, in unmap_region() argument
2404 struct vm_area_struct *next = prev ? prev->vm_next : mm->mmap; in unmap_region()
2411 free_pgtables(&tlb, vma, prev ? prev->vm_end : FIRST_USER_ADDRESS, in unmap_region()
2422 struct vm_area_struct *prev, unsigned long end) in detach_vmas_to_be_unmapped() argument
2427 insertion_point = (prev ? &prev->vm_next : &mm->mmap); in detach_vmas_to_be_unmapped()
2437 vma->vm_prev = prev; in detach_vmas_to_be_unmapped()
2440 mm->highest_vm_end = prev ? prev->vm_end : 0; in detach_vmas_to_be_unmapped()
2535 struct vm_area_struct *vma, *prev, *last; in do_munmap() local
2548 prev = vma->vm_prev; in do_munmap()
2577 prev = vma; in do_munmap()
2587 vma = prev ? prev->vm_next : mm->mmap; in do_munmap()
2606 detach_vmas_to_be_unmapped(mm, vma, prev, end); in do_munmap()
2607 unmap_region(mm, vma, prev, start, end); in do_munmap()
2746 struct vm_area_struct *vma, *prev; in do_brk() local
2775 while (find_vma_links(mm, addr, addr + len, &prev, &rb_link, in do_brk()
2792 vma = vma_merge(mm, prev, addr, addr + len, flags, in do_brk()
2813 vma_link(mm, vma, prev, rb_link, rb_parent); in do_brk()
2892 struct vm_area_struct *prev; in insert_vm_struct() local
2896 &prev, &rb_link, &rb_parent)) in insert_vm_struct()
2919 vma_link(mm, vma, prev, rb_link, rb_parent); in insert_vm_struct()
2934 struct vm_area_struct *new_vma, *prev; in copy_vma() local
2947 if (find_vma_links(mm, addr, addr + len, &prev, &rb_link, &rb_parent)) in copy_vma()
2949 new_vma = vma_merge(mm, prev, addr, addr + len, vma->vm_flags, in copy_vma()
2991 vma_link(mm, new_vma, prev, rb_link, rb_parent); in copy_vma()