mm_rb              74 arch/unicore32/include/asm/mmu_context.h 		rb_erase(&high_vma->vm_rb, &mm->mm_rb); \
mm_rb              90 arch/x86/kernel/tboot.c 	.mm_rb          = RB_ROOT,
mm_rb              60 drivers/firmware/efi/efi.c 	.mm_rb			= RB_ROOT,
mm_rb              29 fs/proc/task_nommu.c 	for (p = rb_first(&mm->mm_rb); p; p = rb_next(p)) {
mm_rb              90 fs/proc/task_nommu.c 	for (p = rb_first(&mm->mm_rb); p; p = rb_next(p)) {
mm_rb             108 fs/proc/task_nommu.c 	for (p = rb_first(&mm->mm_rb); p; p = rb_next(p)) {
mm_rb             220 fs/proc/task_nommu.c 	for (p = rb_first(&mm->mm_rb); p; p = rb_next(p))
mm_rb             373 include/linux/mm_types.h 		struct rb_root mm_rb;
mm_rb             507 kernel/fork.c  	rb_link = &mm->mm_rb.rb_node;
mm_rb            1007 kernel/fork.c  	mm->mm_rb = RB_ROOT;
mm_rb              30 mm/init-mm.c   	.mm_rb		= RB_ROOT,
mm_rb             328 mm/mmap.c      	struct rb_root *root = &mm->mm_rb;
mm_rb             532 mm/mmap.c      	__rb_link = &mm->mm_rb.rb_node;
mm_rb             609 mm/mmap.c      	vma_rb_insert(vma, &mm->mm_rb);
mm_rb             685 mm/mmap.c      	vma_rb_erase_ignore(vma, &mm->mm_rb, ignore);
mm_rb            1906 mm/mmap.c      	if (RB_EMPTY_ROOT(&mm->mm_rb))
mm_rb            1908 mm/mmap.c      	vma = rb_entry(mm->mm_rb.rb_node, struct vm_area_struct, vm_rb);
mm_rb            2010 mm/mmap.c      	if (RB_EMPTY_ROOT(&mm->mm_rb))
mm_rb            2012 mm/mmap.c      	vma = rb_entry(mm->mm_rb.rb_node, struct vm_area_struct, vm_rb);
mm_rb            2245 mm/mmap.c      	rb_node = mm->mm_rb.rb_node;
mm_rb            2281 mm/mmap.c      		struct rb_node *rb_node = rb_last(&mm->mm_rb);
mm_rb            2635 mm/mmap.c      		vma_rb_erase(vma, &mm->mm_rb);
mm_rb             610 mm/nommu.c     	p = &mm->mm_rb.rb_node;
mm_rb             637 mm/nommu.c     	rb_insert_color(&vma->vm_rb, &mm->mm_rb);
mm_rb             678 mm/nommu.c     	rb_erase(&vma->vm_rb, &mm->mm_rb);