next_mm 136 arch/alpha/include/asm/mmu_context.h ev5_switch_mm(struct mm_struct *prev_mm, struct mm_struct *next_mm, next_mm 149 arch/alpha/include/asm/mmu_context.h mmc = next_mm->context[cpu]; next_mm 151 arch/alpha/include/asm/mmu_context.h mmc = __get_new_mm_context(next_mm, cpu); next_mm 152 arch/alpha/include/asm/mmu_context.h next_mm->context[cpu] = mmc; next_mm 166 arch/alpha/include/asm/mmu_context.h ev4_switch_mm(struct mm_struct *prev_mm, struct mm_struct *next_mm, next_mm 177 arch/alpha/include/asm/mmu_context.h if (prev_mm != next_mm) next_mm 182 arch/alpha/include/asm/mmu_context.h ev5_switch_mm(prev_mm, next_mm, next); next_mm 205 arch/alpha/include/asm/mmu_context.h ev5_activate_mm(struct mm_struct *prev_mm, struct mm_struct *next_mm) next_mm 207 arch/alpha/include/asm/mmu_context.h __load_new_mm_context(next_mm); next_mm 211 arch/alpha/include/asm/mmu_context.h ev4_activate_mm(struct mm_struct *prev_mm, struct mm_struct *next_mm) next_mm 213 arch/alpha/include/asm/mmu_context.h __load_new_mm_context(next_mm); next_mm 41 arch/alpha/mm/fault.c __load_new_mm_context(struct mm_struct *next_mm) next_mm 46 arch/alpha/mm/fault.c mmc = __get_new_mm_context(next_mm, smp_processor_id()); next_mm 47 arch/alpha/mm/fault.c next_mm->context[smp_processor_id()] = mmc; next_mm 51 arch/alpha/mm/fault.c pcb->ptbr = ((unsigned long) next_mm->pgd - IDENT_ADDR) >> PAGE_SHIFT; next_mm 197 arch/ia64/include/asm/mmu_context.h #define switch_mm(prev_mm,next_mm,next_task) activate_mm(prev_mm, next_mm) next_mm 206 arch/m68k/include/asm/mmu_context.h struct mm_struct *next_mm) next_mm 208 arch/m68k/include/asm/mmu_context.h activate_context(next_mm); next_mm 296 arch/m68k/include/asm/mmu_context.h struct mm_struct *next_mm) next_mm 298 arch/m68k/include/asm/mmu_context.h next_mm->context = virt_to_phys(next_mm->pgd); next_mm 301 arch/m68k/include/asm/mmu_context.h switch_mm_0230(next_mm); next_mm 303 arch/m68k/include/asm/mmu_context.h switch_mm_0460(next_mm); next_mm 323 arch/m68k/include/asm/mmu_context.h static inline void activate_mm(struct mm_struct *prev_mm, struct mm_struct *next_mm) next_mm 215 arch/x86/mm/tlb.c unsigned long prev_mm, next_mm; next_mm 248 arch/x86/mm/tlb.c next_mm = mm_mangle_tif_spec_ib(next); next_mm 255 arch/x86/mm/tlb.c if (next_mm != prev_mm && next_mm 256 arch/x86/mm/tlb.c (next_mm | prev_mm) & LAST_USER_MM_IBPB) next_mm 259 arch/x86/mm/tlb.c this_cpu_write(cpu_tlbstate.last_user_mm_ibpb, next_mm); next_mm 42 include/asm-generic/mmu_context.h struct mm_struct *next_mm) next_mm 2485 kernel/sched/sched.h struct mm_struct *next_mm) next_mm 2489 kernel/sched/sched.h if (prev_mm == next_mm) next_mm 2492 kernel/sched/sched.h membarrier_state = atomic_read(&next_mm->membarrier_state); next_mm 2501 kernel/sched/sched.h struct mm_struct *next_mm) next_mm 2282 mm/ksm.c next_mm: next_mm 2377 mm/ksm.c goto next_mm;