mm_rb 74 arch/unicore32/include/asm/mmu_context.h rb_erase(&high_vma->vm_rb, &mm->mm_rb); \ mm_rb 90 arch/x86/kernel/tboot.c .mm_rb = RB_ROOT, mm_rb 60 drivers/firmware/efi/efi.c .mm_rb = RB_ROOT, mm_rb 29 fs/proc/task_nommu.c for (p = rb_first(&mm->mm_rb); p; p = rb_next(p)) { mm_rb 90 fs/proc/task_nommu.c for (p = rb_first(&mm->mm_rb); p; p = rb_next(p)) { mm_rb 108 fs/proc/task_nommu.c for (p = rb_first(&mm->mm_rb); p; p = rb_next(p)) { mm_rb 220 fs/proc/task_nommu.c for (p = rb_first(&mm->mm_rb); p; p = rb_next(p)) mm_rb 373 include/linux/mm_types.h struct rb_root mm_rb; mm_rb 507 kernel/fork.c rb_link = &mm->mm_rb.rb_node; mm_rb 1007 kernel/fork.c mm->mm_rb = RB_ROOT; mm_rb 30 mm/init-mm.c .mm_rb = RB_ROOT, mm_rb 328 mm/mmap.c struct rb_root *root = &mm->mm_rb; mm_rb 532 mm/mmap.c __rb_link = &mm->mm_rb.rb_node; mm_rb 609 mm/mmap.c vma_rb_insert(vma, &mm->mm_rb); mm_rb 685 mm/mmap.c vma_rb_erase_ignore(vma, &mm->mm_rb, ignore); mm_rb 1906 mm/mmap.c if (RB_EMPTY_ROOT(&mm->mm_rb)) mm_rb 1908 mm/mmap.c vma = rb_entry(mm->mm_rb.rb_node, struct vm_area_struct, vm_rb); mm_rb 2010 mm/mmap.c if (RB_EMPTY_ROOT(&mm->mm_rb)) mm_rb 2012 mm/mmap.c vma = rb_entry(mm->mm_rb.rb_node, struct vm_area_struct, vm_rb); mm_rb 2245 mm/mmap.c rb_node = mm->mm_rb.rb_node; mm_rb 2281 mm/mmap.c struct rb_node *rb_node = rb_last(&mm->mm_rb); mm_rb 2635 mm/mmap.c vma_rb_erase(vma, &mm->mm_rb); mm_rb 610 mm/nommu.c p = &mm->mm_rb.rb_node; mm_rb 637 mm/nommu.c rb_insert_color(&vma->vm_rb, &mm->mm_rb); mm_rb 678 mm/nommu.c rb_erase(&vma->vm_rb, &mm->mm_rb);