Lines Matching refs:vm_rb
362 if (vma->vm_rb.rb_left) { in vma_compute_subtree_gap()
363 subtree_gap = rb_entry(vma->vm_rb.rb_left, in vma_compute_subtree_gap()
364 struct vm_area_struct, vm_rb)->rb_subtree_gap; in vma_compute_subtree_gap()
368 if (vma->vm_rb.rb_right) { in vma_compute_subtree_gap()
369 subtree_gap = rb_entry(vma->vm_rb.rb_right, in vma_compute_subtree_gap()
370 struct vm_area_struct, vm_rb)->rb_subtree_gap; in vma_compute_subtree_gap()
386 vma = rb_entry(nd, struct vm_area_struct, vm_rb); in browse_rb()
429 vma = rb_entry(nd, struct vm_area_struct, vm_rb); in validate_mm_rb()
480 RB_DECLARE_CALLBACKS(static, vma_gap_callbacks, struct vm_area_struct, vm_rb, in RB_DECLARE_CALLBACKS() argument
494 vma_gap_callbacks_propagate(&vma->vm_rb, NULL); in RB_DECLARE_CALLBACKS()
503 rb_insert_augmented(&vma->vm_rb, root, &vma_gap_callbacks); in vma_rb_insert()
519 rb_erase_augmented(&vma->vm_rb, root, &vma_gap_callbacks); in vma_rb_erase()
567 vma_tmp = rb_entry(__rb_parent, struct vm_area_struct, vm_rb); in find_vma_links()
582 *pprev = rb_entry(rb_prev, struct vm_area_struct, vm_rb); in find_vma_links()
634 rb_link_node(&vma->vm_rb, rb_parent, rb_link); in __vma_link_rb()
1738 vma = rb_entry(mm->mm_rb.rb_node, struct vm_area_struct, vm_rb); in unmapped_area()
1745 if (gap_end >= low_limit && vma->vm_rb.rb_left) { in unmapped_area()
1747 rb_entry(vma->vm_rb.rb_left, in unmapped_area()
1748 struct vm_area_struct, vm_rb); in unmapped_area()
1764 if (vma->vm_rb.rb_right) { in unmapped_area()
1766 rb_entry(vma->vm_rb.rb_right, in unmapped_area()
1767 struct vm_area_struct, vm_rb); in unmapped_area()
1776 struct rb_node *prev = &vma->vm_rb; in unmapped_area()
1780 struct vm_area_struct, vm_rb); in unmapped_area()
1781 if (prev == vma->vm_rb.rb_left) { in unmapped_area()
1841 vma = rb_entry(mm->mm_rb.rb_node, struct vm_area_struct, vm_rb); in unmapped_area_topdown()
1848 if (gap_start <= high_limit && vma->vm_rb.rb_right) { in unmapped_area_topdown()
1850 rb_entry(vma->vm_rb.rb_right, in unmapped_area_topdown()
1851 struct vm_area_struct, vm_rb); in unmapped_area_topdown()
1867 if (vma->vm_rb.rb_left) { in unmapped_area_topdown()
1869 rb_entry(vma->vm_rb.rb_left, in unmapped_area_topdown()
1870 struct vm_area_struct, vm_rb); in unmapped_area_topdown()
1879 struct rb_node *prev = &vma->vm_rb; in unmapped_area_topdown()
1883 struct vm_area_struct, vm_rb); in unmapped_area_topdown()
1884 if (prev == vma->vm_rb.rb_right) { in unmapped_area_topdown()
2056 tmp = rb_entry(rb_node, struct vm_area_struct, vm_rb); in find_vma()
2090 *pprev = rb_entry(rb_node, struct vm_area_struct, vm_rb); in find_vma_prev()