old_mm 26 arch/sparc/include/asm/mmu_context_32.h void switch_mm(struct mm_struct *old_mm, struct mm_struct *mm, old_mm 81 arch/sparc/include/asm/mmu_context_64.h static inline void switch_mm(struct mm_struct *old_mm, struct mm_struct *mm, struct task_struct *tsk) old_mm 435 arch/sparc/mm/srmmu.c static inline void alloc_context(struct mm_struct *old_mm, struct mm_struct *mm) old_mm 448 arch/sparc/mm/srmmu.c if (ctxp->ctx_mm == old_mm) old_mm 493 arch/sparc/mm/srmmu.c void switch_mm(struct mm_struct *old_mm, struct mm_struct *mm, old_mm 500 arch/sparc/mm/srmmu.c alloc_context(old_mm, mm); old_mm 359 arch/x86/kernel/ldt.c int ldt_dup_context(struct mm_struct *old_mm, struct mm_struct *mm) old_mm 364 arch/x86/kernel/ldt.c if (!old_mm) old_mm 367 arch/x86/kernel/ldt.c mutex_lock(&old_mm->context.lock); old_mm 368 arch/x86/kernel/ldt.c if (!old_mm->context.ldt) old_mm 371 arch/x86/kernel/ldt.c new_ldt = alloc_ldt_struct(old_mm->context.ldt->nr_entries); old_mm 377 arch/x86/kernel/ldt.c memcpy(new_ldt->entries, old_mm->context.ldt->entries, old_mm 390 arch/x86/kernel/ldt.c mutex_unlock(&old_mm->context.lock); old_mm 1013 fs/exec.c struct mm_struct *old_mm, *active_mm; old_mm 1017 fs/exec.c old_mm = current->mm; old_mm 1018 fs/exec.c exec_mm_release(tsk, old_mm); old_mm 1020 fs/exec.c if (old_mm) { old_mm 1021 fs/exec.c sync_mm_rss(old_mm); old_mm 1028 fs/exec.c down_read(&old_mm->mmap_sem); old_mm 1029 fs/exec.c if (unlikely(old_mm->core_state)) { old_mm 1030 fs/exec.c up_read(&old_mm->mmap_sem); old_mm 1043 fs/exec.c if (old_mm) { old_mm 1044 fs/exec.c up_read(&old_mm->mmap_sem); old_mm 1045 fs/exec.c BUG_ON(active_mm != old_mm); old_mm 1046 fs/exec.c setmax_mm_hiwater_rss(&tsk->signal->maxrss, old_mm); old_mm 1047 fs/exec.c mm_update_next_owner(old_mm); old_mm 1048 fs/exec.c mmput(old_mm);