tlb_gen            28 arch/x86/include/asm/mmu.h 	atomic64_t tlb_gen;
tlb_gen           192 arch/x86/include/asm/mmu_context.h 	atomic64_set(&mm->context.tlb_gen, 0);
tlb_gen           153 arch/x86/include/asm/tlbflush.h 	u64 tlb_gen;
tlb_gen           599 arch/x86/include/asm/tlbflush.h 	return atomic64_inc_return(&mm->context.tlb_gen);
tlb_gen            96 arch/x86/mm/tlb.c 		*need_flush = (this_cpu_read(cpu_tlbstate.ctxs[asid].tlb_gen) <
tlb_gen           361 arch/x86/mm/tlb.c 		next_tlb_gen = atomic64_read(&next->context.tlb_gen);
tlb_gen           362 arch/x86/mm/tlb.c 		if (this_cpu_read(cpu_tlbstate.ctxs[prev_asid].tlb_gen) ==
tlb_gen           405 arch/x86/mm/tlb.c 		next_tlb_gen = atomic64_read(&next->context.tlb_gen);
tlb_gen           416 arch/x86/mm/tlb.c 		this_cpu_write(cpu_tlbstate.ctxs[new_asid].tlb_gen, next_tlb_gen);
tlb_gen           486 arch/x86/mm/tlb.c 	u64 tlb_gen = atomic64_read(&init_mm.context.tlb_gen);
tlb_gen           508 arch/x86/mm/tlb.c 	this_cpu_write(cpu_tlbstate.ctxs[0].tlb_gen, tlb_gen);
tlb_gen           535 arch/x86/mm/tlb.c 	u64 mm_tlb_gen = atomic64_read(&loaded_mm->context.tlb_gen);
tlb_gen           536 arch/x86/mm/tlb.c 	u64 local_tlb_gen = this_cpu_read(cpu_tlbstate.ctxs[loaded_mm_asid].tlb_gen);
tlb_gen           635 arch/x86/mm/tlb.c 	this_cpu_write(cpu_tlbstate.ctxs[loaded_mm_asid].tlb_gen, mm_tlb_gen);