tlb_gen 28 arch/x86/include/asm/mmu.h atomic64_t tlb_gen; tlb_gen 192 arch/x86/include/asm/mmu_context.h atomic64_set(&mm->context.tlb_gen, 0); tlb_gen 153 arch/x86/include/asm/tlbflush.h u64 tlb_gen; tlb_gen 599 arch/x86/include/asm/tlbflush.h return atomic64_inc_return(&mm->context.tlb_gen); tlb_gen 96 arch/x86/mm/tlb.c *need_flush = (this_cpu_read(cpu_tlbstate.ctxs[asid].tlb_gen) < tlb_gen 361 arch/x86/mm/tlb.c next_tlb_gen = atomic64_read(&next->context.tlb_gen); tlb_gen 362 arch/x86/mm/tlb.c if (this_cpu_read(cpu_tlbstate.ctxs[prev_asid].tlb_gen) == tlb_gen 405 arch/x86/mm/tlb.c next_tlb_gen = atomic64_read(&next->context.tlb_gen); tlb_gen 416 arch/x86/mm/tlb.c this_cpu_write(cpu_tlbstate.ctxs[new_asid].tlb_gen, next_tlb_gen); tlb_gen 486 arch/x86/mm/tlb.c u64 tlb_gen = atomic64_read(&init_mm.context.tlb_gen); tlb_gen 508 arch/x86/mm/tlb.c this_cpu_write(cpu_tlbstate.ctxs[0].tlb_gen, tlb_gen); tlb_gen 535 arch/x86/mm/tlb.c u64 mm_tlb_gen = atomic64_read(&loaded_mm->context.tlb_gen); tlb_gen 536 arch/x86/mm/tlb.c u64 local_tlb_gen = this_cpu_read(cpu_tlbstate.ctxs[loaded_mm_asid].tlb_gen); tlb_gen 635 arch/x86/mm/tlb.c this_cpu_write(cpu_tlbstate.ctxs[loaded_mm_asid].tlb_gen, mm_tlb_gen);