Lines Matching refs:mm
24 void __check_vmalloc_seq(struct mm_struct *mm);
28 void check_and_switch_context(struct mm_struct *mm, struct task_struct *tsk);
29 #define init_new_context(tsk,mm) ({ atomic64_set(&mm->context.id, 0); 0; }) argument
32 void a15_erratum_get_cpumask(int this_cpu, struct mm_struct *mm,
35 static inline void a15_erratum_get_cpumask(int this_cpu, struct mm_struct *mm, in a15_erratum_get_cpumask() argument
45 static inline void check_and_switch_context(struct mm_struct *mm, in check_and_switch_context() argument
48 if (unlikely(mm->context.vmalloc_seq != init_mm.context.vmalloc_seq)) in check_and_switch_context()
49 __check_vmalloc_seq(mm); in check_and_switch_context()
59 mm->context.switch_pending = 1; in check_and_switch_context()
61 cpu_switch_mm(mm->pgd, mm); in check_and_switch_context()
68 struct mm_struct *mm = current->mm; in finish_arch_post_lock_switch() local
70 if (mm && mm->context.switch_pending) { in finish_arch_post_lock_switch()
78 if (mm->context.switch_pending) { in finish_arch_post_lock_switch()
79 mm->context.switch_pending = 0; in finish_arch_post_lock_switch()
80 cpu_switch_mm(mm->pgd, mm); in finish_arch_post_lock_switch()
88 #define init_new_context(tsk,mm) 0 argument
92 #define destroy_context(mm) do { } while(0) argument
105 enter_lazy_tlb(struct mm_struct *mm, struct task_struct *tsk) in enter_lazy_tlb() argument
140 #define deactivate_mm(tsk,mm) do { } while (0) argument