/linux-4.4.14/arch/microblaze/kernel/ |
H A D | asm-offsets.c | 93 DEFINE(TI_CPU_CONTEXT, offsetof(struct thread_info, cpu_context)); main() 97 /* struct cpu_context */ main() 98 DEFINE(CC_R1, offsetof(struct cpu_context, r1)); /* r1 */ main() 99 DEFINE(CC_R2, offsetof(struct cpu_context, r2)); main() 101 DEFINE(CC_R13, offsetof(struct cpu_context, r13)); main() 102 DEFINE(CC_R14, offsetof(struct cpu_context, r14)); main() 103 DEFINE(CC_R15, offsetof(struct cpu_context, r15)); main() 104 DEFINE(CC_R16, offsetof(struct cpu_context, r16)); main() 105 DEFINE(CC_R17, offsetof(struct cpu_context, r17)); main() 106 DEFINE(CC_R18, offsetof(struct cpu_context, r18)); main() 108 DEFINE(CC_R19, offsetof(struct cpu_context, r19)); main() 109 DEFINE(CC_R20, offsetof(struct cpu_context, r20)); main() 110 DEFINE(CC_R21, offsetof(struct cpu_context, r21)); main() 111 DEFINE(CC_R22, offsetof(struct cpu_context, r22)); main() 112 DEFINE(CC_R23, offsetof(struct cpu_context, r23)); main() 113 DEFINE(CC_R24, offsetof(struct cpu_context, r24)); main() 114 DEFINE(CC_R25, offsetof(struct cpu_context, r25)); main() 115 DEFINE(CC_R26, offsetof(struct cpu_context, r26)); main() 116 DEFINE(CC_R27, offsetof(struct cpu_context, r27)); main() 117 DEFINE(CC_R28, offsetof(struct cpu_context, r28)); main() 118 DEFINE(CC_R29, offsetof(struct cpu_context, r29)); main() 119 DEFINE(CC_R30, offsetof(struct cpu_context, r30)); main() 121 DEFINE(CC_MSR, offsetof(struct cpu_context, msr)); main() 122 DEFINE(CC_EAR, offsetof(struct cpu_context, ear)); main() 123 DEFINE(CC_ESR, offsetof(struct cpu_context, esr)); main() 124 DEFINE(CC_FSR, offsetof(struct cpu_context, fsr)); main()
|
H A D | process.c | 64 memset(&ti->cpu_context, 0, sizeof(struct cpu_context)); copy_thread() 65 ti->cpu_context.r1 = (unsigned long)childregs; copy_thread() 66 ti->cpu_context.r20 = (unsigned long)usp; /* fn */ copy_thread() 67 ti->cpu_context.r19 = (unsigned long)arg; copy_thread() 71 ti->cpu_context.msr = childregs->msr & ~MSR_IE; copy_thread() 73 ti->cpu_context.r15 = (unsigned long)ret_from_kernel_thread - 8; copy_thread() 80 memset(&ti->cpu_context, 0, sizeof(struct cpu_context)); copy_thread() 81 ti->cpu_context.r1 = (unsigned long)childregs; copy_thread() 83 ti->cpu_context.msr = (unsigned long)childregs->msr; copy_thread() 103 ti->cpu_context.msr = (childregs->msr|MSR_VM); copy_thread() 104 ti->cpu_context.msr &= ~MSR_UMS; /* switch_to to kernel mode */ copy_thread() 105 ti->cpu_context.msr &= ~MSR_IE; copy_thread() 107 ti->cpu_context.r15 = (unsigned long)ret_from_fork - 8; copy_thread() 125 struct cpu_context *ctx = thread_saved_pc() 126 &(((struct thread_info *)(tsk->stack))->cpu_context); thread_saved_pc()
|
H A D | traps.c | 41 (task->stack))->cpu_context.r1; show_stack()
|
H A D | unwind.c | 295 const struct cpu_context *cpu_context = microblaze_unwind() local 296 &thread_info->cpu_context; microblaze_unwind() 300 cpu_context->r1, microblaze_unwind() 301 cpu_context->r15, trace); microblaze_unwind()
|
H A D | entry.S | 875 /* save registers in cpu_context */ 877 /* give start of cpu_context for previous process */
|
H A D | entry-nommu.S | 391 /* save registers in cpu_context */
|
/linux-4.4.14/arch/avr32/include/asm/ |
H A D | switch_to.h | 36 struct cpu_context; 39 struct cpu_context *, 40 struct cpu_context *); 44 last = __switch_to(prev, &prev->thread.cpu_context + 1, \ 45 &next->thread.cpu_context); \
|
H A D | processor.h | 98 struct cpu_context { struct 114 struct cpu_context cpu_context; member in struct:thread_struct 120 .cpu_context = { \ 142 #define thread_saved_pc(tsk) ((tsk)->thread.cpu_context.pc) 153 #define KSTK_EIP(tsk) ((tsk)->thread.cpu_context.pc) 154 #define KSTK_ESP(tsk) ((tsk)->thread.cpu_context.ksp)
|
/linux-4.4.14/arch/mips/mm/ |
H A D | tlb-r3k.c | 73 if (cpu_context(cpu, mm) != 0) { local_flush_tlb_mm() 75 printk("[tlbmm<%lu>]", (unsigned long)cpu_context(cpu, mm)); local_flush_tlb_mm() 87 if (cpu_context(cpu, mm) != 0) { local_flush_tlb_range() 92 cpu_context(cpu, mm) & ASID_MASK, start, end); local_flush_tlb_range() 98 int newpid = cpu_context(cpu, mm) & ASID_MASK; local_flush_tlb_range() 164 if (cpu_context(cpu, vma->vm_mm) != 0) { local_flush_tlb_page() 169 printk("[tlbpage<%lu,0x%08lx>]", cpu_context(cpu, vma->vm_mm), page); local_flush_tlb_page() 171 newpid = cpu_context(cpu, vma->vm_mm) & ASID_MASK; local_flush_tlb_page() 205 if ((pid != (cpu_context(cpu, vma->vm_mm) & ASID_MASK)) || (cpu_context(cpu, vma->vm_mm) == 0)) { __update_tlb() 207 (cpu_context(cpu, vma->vm_mm)), pid); __update_tlb()
|
H A D | tlb-r8k.c | 57 if (cpu_context(cpu, mm) != 0) local_flush_tlb_mm() 69 if (!cpu_context(cpu, mm)) local_flush_tlb_range() 155 if (!cpu_context(cpu, vma->vm_mm)) local_flush_tlb_page()
|
H A D | c-tx39.c | 152 if (cpu_context(smp_processor_id(), mm) != 0) tx39_flush_cache_mm() 161 if (!(cpu_context(smp_processor_id(), vma->vm_mm))) tx39_flush_cache_range() 180 if (cpu_context(smp_processor_id(), mm) == 0) tx39_flush_cache_page()
|
H A D | tlb-r4k.c | 110 if (cpu_context(cpu, mm) != 0) { local_flush_tlb_mm() 123 if (cpu_context(cpu, mm) != 0) { local_flush_tlb_range() 215 if (cpu_context(cpu, vma->vm_mm) != 0) { local_flush_tlb_page()
|
H A D | c-r3k.c | 248 cpu_context(smp_processor_id(), mm), addr); r3k_flush_cache_page() 251 if (cpu_context(smp_processor_id(), mm) == 0) r3k_flush_cache_page()
|
H A D | c-r4k.c | 468 if (cpu_context(i, mm)) has_valid_asid() 473 return cpu_context(smp_processor_id(), mm); has_valid_asid() 608 if (cpu_context(cpu, mm) != 0) local_r4k_flush_cache_page()
|
/linux-4.4.14/arch/unicore32/include/asm/ |
H A D | suspend.h | 21 struct cpu_context_save cpu_context; /* cpu context */ member in struct:swsusp_arch_regs
|
H A D | thread_info.h | 65 * __switch_to() assumes cpu_context follows immediately after cpu_domain. 74 struct cpu_context_save cpu_context; /* cpu context */ member in struct:thread_info 105 ((unsigned long)(task_thread_info(tsk)->cpu_context.pc)) 107 ((unsigned long)(task_thread_info(tsk)->cpu_context.sp)) 109 ((unsigned long)(task_thread_info(tsk)->cpu_context.fp))
|
/linux-4.4.14/arch/sh/mm/ |
H A D | tlbflush_32.c | 19 if (vma->vm_mm && cpu_context(cpu, vma->vm_mm) != NO_CONTEXT) { local_flush_tlb_page() 45 if (cpu_context(cpu, mm) != NO_CONTEXT) { local_flush_tlb_range() 52 cpu_context(cpu, mm) = NO_CONTEXT; local_flush_tlb_range() 112 if (cpu_context(cpu, mm) != NO_CONTEXT) { local_flush_tlb_mm() 116 cpu_context(cpu, mm) = NO_CONTEXT; local_flush_tlb_mm()
|
H A D | tlbflush_64.c | 89 if (cpu_context(cpu, mm) == NO_CONTEXT) local_flush_tlb_range() 133 if (cpu_context(cpu, mm) == NO_CONTEXT) local_flush_tlb_mm() 138 cpu_context(cpu, mm) = NO_CONTEXT; local_flush_tlb_mm()
|
H A D | cache-sh4.c | 192 if (cpu_context(smp_processor_id(), mm) == NO_CONTEXT) sh4_flush_cache_mm() 223 if (cpu_context(smp_processor_id(), vma->vm_mm) == NO_CONTEXT) sh4_flush_cache_page() 286 if (cpu_context(smp_processor_id(), vma->vm_mm) == NO_CONTEXT) sh4_flush_cache_range()
|
/linux-4.4.14/arch/avr32/kernel/ |
H A D | process.c | 129 fp = tsk->thread.cpu_context.r7; show_trace_log_lvl() 186 sp = tsk->thread.cpu_context.ksp; show_stack_log_lvl() 290 p->thread.cpu_context.r0 = arg; copy_thread() 291 p->thread.cpu_context.r1 = usp; /* fn */ copy_thread() 292 p->thread.cpu_context.r2 = (unsigned long)syscall_return; copy_thread() 293 p->thread.cpu_context.pc = (unsigned long)ret_from_kernel_thread; copy_thread() 300 p->thread.cpu_context.pc = (unsigned long)ret_from_fork; copy_thread() 303 p->thread.cpu_context.sr = MODE_SUPERVISOR | SR_GM; copy_thread() 304 p->thread.cpu_context.ksp = (unsigned long)childregs; copy_thread() 335 unsigned long fp = p->thread.cpu_context.r7; get_wchan() 348 unsigned long sp = p->thread.cpu_context.ksp + 16; get_wchan()
|
/linux-4.4.14/arch/arm/kernel/ |
H A D | kgdb.c | 90 gdb_regs[_R4] = ti->cpu_context.r4; sleeping_thread_to_gdb_regs() 91 gdb_regs[_R5] = ti->cpu_context.r5; sleeping_thread_to_gdb_regs() 92 gdb_regs[_R6] = ti->cpu_context.r6; sleeping_thread_to_gdb_regs() 93 gdb_regs[_R7] = ti->cpu_context.r7; sleeping_thread_to_gdb_regs() 94 gdb_regs[_R8] = ti->cpu_context.r8; sleeping_thread_to_gdb_regs() 95 gdb_regs[_R9] = ti->cpu_context.r9; sleeping_thread_to_gdb_regs() 96 gdb_regs[_R10] = ti->cpu_context.sl; sleeping_thread_to_gdb_regs() 97 gdb_regs[_FP] = ti->cpu_context.fp; sleeping_thread_to_gdb_regs() 98 gdb_regs[_SPT] = ti->cpu_context.sp; sleeping_thread_to_gdb_regs() 99 gdb_regs[_PC] = ti->cpu_context.pc; sleeping_thread_to_gdb_regs()
|
H A D | xscale-cp0.c | 39 thread->cpu_context.extra[0] = 0; dsp_do() 40 thread->cpu_context.extra[1] = 0; dsp_do() 44 dsp_save_state(current_thread_info()->cpu_context.extra); dsp_do() 45 dsp_load_state(thread->cpu_context.extra); dsp_do()
|
H A D | process.c | 230 memset(&thread->cpu_context, 0, sizeof(struct cpu_context_save)); copy_thread() 249 thread->cpu_context.r4 = stk_sz; copy_thread() 250 thread->cpu_context.r5 = stack_start; copy_thread() 253 thread->cpu_context.pc = (unsigned long)ret_from_fork; copy_thread() 254 thread->cpu_context.sp = (unsigned long)childregs; copy_thread()
|
H A D | asm-offsets.c | 72 DEFINE(TI_CPU_SAVE, offsetof(struct thread_info, cpu_context)); main()
|
/linux-4.4.14/arch/mips/include/asm/ |
H A D | mmu_context.h | 85 #define cpu_context(cpu, mm) ((mm)->context.asid[cpu]) macro 86 #define cpu_asid(cpu, mm) (cpu_context((cpu), (mm)) & ASID_MASK) 119 cpu_context(cpu, mm) = asid_cache(cpu) = asid; get_new_mmu_context() 132 cpu_context(i, mm) = 0; init_new_context() 148 if ((cpu_context(cpu, next) ^ asid_cache(cpu)) & ASID_VERSION_MASK) switch_mm() 218 cpu_context(cpu, mm) = 0; drop_mmu_context()
|
/linux-4.4.14/arch/arm/include/asm/ |
H A D | thread_info.h | 47 * __switch_to() assumes cpu_context follows immediately after cpu_domain. 56 struct cpu_context_save cpu_context; /* cpu context */ member in struct:thread_info 98 ((unsigned long)(task_thread_info(tsk)->cpu_context.pc)) 100 ((unsigned long)(task_thread_info(tsk)->cpu_context.sp)) 104 ((unsigned long)(task_thread_info(tsk)->cpu_context.fp)) 107 ((unsigned long)(task_thread_info(tsk)->cpu_context.r7))
|
/linux-4.4.14/arch/arm64/include/asm/ |
H A D | thread_info.h | 45 * __switch_to() assumes cpu_context follows immediately after cpu_domain. 83 ((unsigned long)(tsk->thread.cpu_context.pc)) 85 ((unsigned long)(tsk->thread.cpu_context.sp)) 87 ((unsigned long)(tsk->thread.cpu_context.fp))
|
H A D | processor.h | 63 struct cpu_context { struct 80 struct cpu_context cpu_context; /* cpu context */ member in struct:thread_struct
|
/linux-4.4.14/arch/sh/include/asm/ |
H A D | mmu_context.h | 38 #define cpu_context(cpu, mm) ((mm)->context.id[cpu]) macro 41 (cpu_context((cpu), (mm)) & MMU_CONTEXT_ASID_MASK) 62 if (((cpu_context(cpu, mm) ^ asid) & MMU_CONTEXT_VERSION_MASK) == 0) get_mmu_context() 90 cpu_context(cpu, mm) = asid_cache(cpu) = asid; get_mmu_context() 103 cpu_context(i, mm) = NO_CONTEXT; init_new_context()
|
/linux-4.4.14/arch/unicore32/kernel/ |
H A D | process.c | 237 memset(&thread->cpu_context, 0, sizeof(struct cpu_context_save)); copy_thread() 238 thread->cpu_context.sp = (unsigned long)childregs; copy_thread() 240 thread->cpu_context.pc = (unsigned long)ret_from_kernel_thread; copy_thread() 241 thread->cpu_context.r4 = stack_start; copy_thread() 242 thread->cpu_context.r5 = stk_sz; copy_thread() 245 thread->cpu_context.pc = (unsigned long)ret_from_fork; copy_thread()
|
H A D | asm-offsets.c | 46 DEFINE(TI_CPU_SAVE, offsetof(struct thread_info, cpu_context)); main() 101 cpu_context)); main()
|
/linux-4.4.14/arch/arm64/kernel/ |
H A D | process.c | 252 memset(&p->thread.cpu_context, 0, sizeof(struct cpu_context)); copy_thread() 283 p->thread.cpu_context.x19 = stack_start; copy_thread() 284 p->thread.cpu_context.x20 = stk_sz; copy_thread() 286 p->thread.cpu_context.pc = (unsigned long)ret_from_fork; copy_thread() 287 p->thread.cpu_context.sp = (unsigned long)childregs; copy_thread()
|
H A D | asm-offsets.c | 42 DEFINE(THREAD_CPU_CONTEXT, offsetof(struct task_struct, thread.cpu_context)); main()
|
/linux-4.4.14/arch/microblaze/include/asm/ |
H A D | thread_info.h | 31 struct cpu_context { struct 74 struct cpu_context cpu_context; member in struct:thread_info
|
/linux-4.4.14/arch/mips/kernel/ |
H A D | smp.c | 332 if (cpu != smp_processor_id() && cpu_context(cpu, mm)) for_each_online_cpu() 333 cpu_context(cpu, mm) = 0; for_each_online_cpu() 371 if (cpu != smp_processor_id() && cpu_context(cpu, mm)) for_each_online_cpu() 372 cpu_context(cpu, mm) = 0; for_each_online_cpu() 417 if (cpu != smp_processor_id() && cpu_context(cpu, vma->vm_mm)) for_each_online_cpu() 418 cpu_context(cpu, vma->vm_mm) = 0; for_each_online_cpu()
|
/linux-4.4.14/arch/sh/kernel/ |
H A D | smp.c | 368 cpu_context(i, mm) = 0; flush_tlb_mm() 405 cpu_context(i, mm) = 0; flush_tlb_range() 448 cpu_context(i, vma->vm_mm) = 0; flush_tlb_page()
|
/linux-4.4.14/arch/mips/kvm/ |
H A D | tlb.c | 615 cpu_context(cpu, mm) = asid_cache(cpu) = asid; kvm_get_new_mmu_context() 686 kvm_debug("[%d]: cpu_context: %#lx\n", cpu, kvm_arch_vcpu_load() 687 cpu_context(cpu, current->mm)); kvm_arch_vcpu_load() 760 if (((cpu_context(cpu, current->mm) ^ asid_cache(cpu)) & kvm_arch_vcpu_put() 763 cpu_context(cpu, current->mm)); kvm_arch_vcpu_put()
|