/linux-4.4.14/tools/perf/util/ |
D | thread.c | 13 int thread__init_map_groups(struct thread *thread, struct machine *machine) in thread__init_map_groups() argument 15 struct thread *leader; in thread__init_map_groups() 16 pid_t pid = thread->pid_; in thread__init_map_groups() 18 if (pid == thread->tid || pid == -1) { in thread__init_map_groups() 19 thread->mg = map_groups__new(machine); in thread__init_map_groups() 23 thread->mg = map_groups__get(leader->mg); in thread__init_map_groups() 26 return thread->mg ? 0 : -1; in thread__init_map_groups() 29 struct thread *thread__new(pid_t pid, pid_t tid) in thread__new() 33 struct thread *thread = zalloc(sizeof(*thread)); in thread__new() local 35 if (thread != NULL) { in thread__new() [all …]
|
D | thread.h | 15 struct thread { struct 40 struct thread *thread__new(pid_t pid, pid_t tid); argument 41 int thread__init_map_groups(struct thread *thread, struct machine *machine); 42 void thread__delete(struct thread *thread); 44 struct thread *thread__get(struct thread *thread); 45 void thread__put(struct thread *thread); 47 static inline void __thread__zput(struct thread **thread) in __thread__zput() argument 49 thread__put(*thread); in __thread__zput() 50 *thread = NULL; in __thread__zput() 53 #define thread__zput(thread) __thread__zput(&thread) argument [all …]
|
D | unwind.h | 19 struct thread *thread, 24 int unwind__prepare_access(struct thread *thread); 25 void unwind__flush_access(struct thread *thread); 26 void unwind__finish_access(struct thread *thread); 28 static inline int unwind__prepare_access(struct thread *thread __maybe_unused) in unwind__prepare_access() 33 static inline void unwind__flush_access(struct thread *thread __maybe_unused) {} in unwind__flush_access() 34 static inline void unwind__finish_access(struct thread *thread __maybe_unused) {} in unwind__finish_access() 40 struct thread *thread __maybe_unused, in unwind__get_entries() 47 static inline int unwind__prepare_access(struct thread *thread __maybe_unused) in unwind__prepare_access() 52 static inline void unwind__flush_access(struct thread *thread __maybe_unused) {} in unwind__flush_access() [all …]
|
D | thread-stack.c | 126 static struct thread_stack *thread_stack__new(struct thread *thread, in thread_stack__new() argument 140 if (thread->mg && thread->mg->machine) in thread_stack__new() 141 ts->kernel_start = machine__kernel_start(thread->mg->machine); in thread_stack__new() 195 static int thread_stack__call_return(struct thread *thread, in thread_stack__call_return() argument 202 .thread = thread, in thread_stack__call_return() 222 static int __thread_stack__flush(struct thread *thread, struct thread_stack *ts) in __thread_stack__flush() argument 233 err = thread_stack__call_return(thread, ts, --ts->cnt, in __thread_stack__flush() 245 int thread_stack__flush(struct thread *thread) in thread_stack__flush() argument 247 if (thread->ts) in thread_stack__flush() 248 return __thread_stack__flush(thread, thread->ts); in thread_stack__flush() [all …]
|
D | thread-stack.h | 24 struct thread; 59 struct thread *thread; member 94 int thread_stack__event(struct thread *thread, u32 flags, u64 from_ip, 96 void thread_stack__set_trace_nr(struct thread *thread, u64 trace_nr); 97 void thread_stack__sample(struct thread *thread, struct ip_callchain *chain, 99 int thread_stack__flush(struct thread *thread); 100 void thread_stack__free(struct thread *thread); 106 int thread_stack__process(struct thread *thread, struct comm *comm,
|
D | db-export.c | 122 int db_export__thread(struct db_export *dbe, struct thread *thread, in db_export__thread() argument 125 struct thread *main_thread; in db_export__thread() 129 if (thread->db_id) in db_export__thread() 132 thread->db_id = ++dbe->thread_last_db_id; in db_export__thread() 134 if (thread->pid_ != -1) { in db_export__thread() 135 if (thread->pid_ == thread->tid) { in db_export__thread() 136 main_thread = thread; in db_export__thread() 139 thread->pid_, in db_export__thread() 140 thread->pid_); in db_export__thread() 148 err = db_export__comm_thread(dbe, comm, thread); in db_export__thread() [all …]
|
D | machine.c | 17 static void __machine__remove_thread(struct machine *machine, struct thread *th, bool lock); 52 struct thread *thread = machine__findnew_thread(machine, -1, in machine__init() local 56 if (thread == NULL) in machine__init() 60 thread__set_comm(thread, comm, 0); in machine__init() 61 thread__put(thread); in machine__init() 115 struct thread *t = rb_entry(nd, struct thread, rb_node); in machine__delete_threads() 313 struct thread *th, pid_t pid) in machine__update_thread_pid() 315 struct thread *leader; in machine__update_thread_pid() 358 static struct thread *____machine__findnew_thread(struct machine *machine, in ____machine__findnew_thread() 364 struct thread *th; in ____machine__findnew_thread() [all …]
|
D | db-export.h | 24 struct thread; 51 int (*export_thread)(struct db_export *dbe, struct thread *thread, 55 struct comm *comm, struct thread *thread); 85 int db_export__thread(struct db_export *dbe, struct thread *thread, 88 struct thread *main_thread); 90 struct thread *thread);
|
D | unwind-libunwind.c | 90 struct thread *thread; member 320 thread__find_addr_map(ui->thread, PERF_RECORD_MISC_USER, in find_map() 422 thread__find_addr_map(ui->thread, PERF_RECORD_MISC_USER, in access_dso_mem() 524 static int entry(u64 ip, struct thread *thread, in entry() argument 530 thread__find_addr_location(thread, PERF_RECORD_MISC_USER, in entry() 573 int unwind__prepare_access(struct thread *thread) in unwind__prepare_access() argument 587 thread__set_priv(thread, addr_space); in unwind__prepare_access() 592 void unwind__flush_access(struct thread *thread) in unwind__flush_access() argument 599 addr_space = thread__priv(thread); in unwind__flush_access() 603 void unwind__finish_access(struct thread *thread) in unwind__finish_access() argument [all …]
|
D | thread_map.h | 38 static inline pid_t thread_map__pid(struct thread_map *map, int thread) in thread_map__pid() argument 40 return map->map[thread].pid; in thread_map__pid() 44 thread_map__set_pid(struct thread_map *map, int thread, pid_t pid) in thread_map__set_pid() argument 46 map->map[thread].pid = pid; in thread_map__set_pid() 49 static inline char *thread_map__comm(struct thread_map *map, int thread) in thread_map__comm() argument 51 return map->map[thread].comm; in thread_map__comm()
|
D | evlist.c | 339 int cpu, thread; in perf_evlist__disable() local 349 for (thread = 0; thread < nr_threads; thread++) in perf_evlist__disable() 350 ioctl(FD(pos, cpu, thread), in perf_evlist__disable() 360 int cpu, thread; in perf_evlist__enable() local 370 for (thread = 0; thread < nr_threads; thread++) in perf_evlist__enable() 371 ioctl(FD(pos, cpu, thread), in perf_evlist__enable() 387 int cpu, thread, err; in perf_evlist__disable_event() local 395 for (thread = 0; thread < nr_threads; thread++) { in perf_evlist__disable_event() 396 err = ioctl(FD(evsel, cpu, thread), in perf_evlist__disable_event() 408 int cpu, thread, err; in perf_evlist__enable_event() local [all …]
|
D | machine.h | 15 struct thread; 35 struct thread *last_match; 78 struct thread *machine__find_thread(struct machine *machine, pid_t pid, 81 struct thread *thread); 138 void machine__remove_thread(struct machine *machine, struct thread *th); 144 int thread__resolve_callchain(struct thread *thread, 165 struct thread *__machine__findnew_thread(struct machine *machine, pid_t pid, pid_t tid); 166 struct thread *machine__findnew_thread(struct machine *machine, pid_t pid, pid_t tid); 226 int (*fn)(struct thread *thread, void *p), 229 int (*fn)(struct thread *thread, void *p),
|
D | unwind-libdw.c | 29 thread__find_addr_location(ui->thread, in __report_module() 92 thread__find_addr_map(ui->thread, PERF_RECORD_MISC_USER, in access_dso_mem() 167 struct thread *thread, in unwind__get_entries() argument 173 .thread = thread, in unwind__get_entries() 174 .machine = thread->mg->machine, in unwind__get_entries() 197 if (!dwfl_attach_state(ui.dwfl, EM_NONE, thread->tid, &callbacks, &ui)) in unwind__get_entries() 200 err = dwfl_getthread_frames(ui.dwfl, thread->tid, frame_callback, &ui); in unwind__get_entries()
|
D | vdso.c | 140 struct thread *thread) in machine__thread_dso_type() argument 146 map = map_groups__first(thread->mg, MAP__FUNCTION); in machine__thread_dso_type() 253 struct thread *thread, in __machine__findnew_vdso_compat() argument 259 dso_type = machine__thread_dso_type(machine, thread); in __machine__findnew_vdso_compat() 287 struct thread *thread __maybe_unused) in machine__findnew_vdso() 301 if (__machine__findnew_vdso_compat(machine, thread, vdso_info, &dso)) in machine__findnew_vdso()
|
D | unwind-libdw.h | 9 bool libdw__arch_set_initial_registers(Dwfl_Thread *thread, void *arg); 15 struct thread *thread; member
|
D | event.c | 492 int err = -1, thread, j; in perf_event__synthesize_thread_map() local 507 for (thread = 0; thread < threads->nr; ++thread) { in perf_event__synthesize_thread_map() 510 thread_map__pid(threads, thread), 0, in perf_event__synthesize_thread_map() 521 if ((int) comm_event->comm.pid != thread_map__pid(threads, thread)) { in perf_event__synthesize_thread_map() 898 void thread__find_addr_map(struct thread *thread, u8 cpumode, in thread__find_addr_map() argument 902 struct map_groups *mg = thread->mg; in thread__find_addr_map() 907 al->thread = thread; in thread__find_addr_map() 974 void thread__find_addr_location(struct thread *thread, in thread__find_addr_location() argument 978 thread__find_addr_map(thread, cpumode, type, addr, al); in thread__find_addr_location() 981 thread->mg->machine->symbol_filter); in thread__find_addr_location() [all …]
|
D | vdso.h | 24 struct thread; 26 struct dso *machine__findnew_vdso(struct machine *machine, struct thread *thread);
|
D | session.h | 16 struct thread; 67 struct thread *thread, 91 struct thread *perf_session__findnew(struct perf_session *session, pid_t pid); 92 struct thread *perf_session__register_idle_thread(struct perf_session *session);
|
D | stat.c | 243 int cpu, int thread, in process_counter_values() argument 264 perf_evsel__compute_deltas(evsel, cpu, thread, count); in process_counter_values() 288 int cpu, thread; in process_counter_maps() local 293 for (thread = 0; thread < nthreads; thread++) { in process_counter_maps() 295 if (process_counter_values(config, counter, cpu, thread, in process_counter_maps() 296 perf_counts(counter->counts, cpu, thread))) in process_counter_maps()
|
D | evsel.c | 907 int cpu, thread; in perf_evsel__alloc_fd() local 916 for (thread = 0; thread < nthreads; thread++) { in perf_evsel__alloc_fd() 917 FD(evsel, cpu, thread) = -1; in perf_evsel__alloc_fd() 928 int cpu, thread; in perf_evsel__run_ioctl() local 934 for (thread = 0; thread < nthreads; thread++) { in perf_evsel__run_ioctl() 935 int fd = FD(evsel, cpu, thread), in perf_evsel__run_ioctl() 1038 int cpu, thread; in perf_evsel__close_fd() local 1044 for (thread = 0; thread < nthreads; ++thread) { in perf_evsel__close_fd() 1045 close(FD(evsel, cpu, thread)); in perf_evsel__close_fd() 1046 FD(evsel, cpu, thread) = -1; in perf_evsel__close_fd() [all …]
|
D | evsel.h | 152 void perf_evsel__compute_deltas(struct perf_evsel *evsel, int cpu, int thread, 275 int perf_evsel__read(struct perf_evsel *evsel, int cpu, int thread, 279 int cpu, int thread, bool scale); 289 int cpu, int thread) in perf_evsel__read_on_cpu() argument 291 return __perf_evsel__read_on_cpu(evsel, cpu, thread, false); in perf_evsel__read_on_cpu() 302 int cpu, int thread) in perf_evsel__read_on_cpu_scaled() argument 304 return __perf_evsel__read_on_cpu(evsel, cpu, thread, true); in perf_evsel__read_on_cpu_scaled()
|
D | build-id.c | 32 struct thread *thread = machine__findnew_thread(machine, sample->pid, in build_id__mark_dso_hit() local 35 if (thread == NULL) { in build_id__mark_dso_hit() 41 thread__find_addr_map(thread, cpumode, MAP__FUNCTION, sample->ip, &al); in build_id__mark_dso_hit() 46 thread__put(thread); in build_id__mark_dso_hit() 56 struct thread *thread = machine__findnew_thread(machine, in perf_event__exit_del_thread() local 63 if (thread) { in perf_event__exit_del_thread() 64 machine__remove_thread(machine, thread); in perf_event__exit_del_thread() 65 thread__put(thread); in perf_event__exit_del_thread()
|
/linux-4.4.14/arch/mips/include/asm/ |
D | asmmacro-32.h | 15 .macro fpu_save_single thread tmp=t0 19 s.d $f0, THREAD_FPR0(\thread) 20 s.d $f2, THREAD_FPR2(\thread) 21 s.d $f4, THREAD_FPR4(\thread) 22 s.d $f6, THREAD_FPR6(\thread) 23 s.d $f8, THREAD_FPR8(\thread) 24 s.d $f10, THREAD_FPR10(\thread) 25 s.d $f12, THREAD_FPR12(\thread) 26 s.d $f14, THREAD_FPR14(\thread) 27 s.d $f16, THREAD_FPR16(\thread) [all …]
|
D | asmmacro.h | 59 .macro fpu_save_16even thread tmp=t0 63 sdc1 $f0, THREAD_FPR0(\thread) 64 sdc1 $f2, THREAD_FPR2(\thread) 65 sdc1 $f4, THREAD_FPR4(\thread) 66 sdc1 $f6, THREAD_FPR6(\thread) 67 sdc1 $f8, THREAD_FPR8(\thread) 68 sdc1 $f10, THREAD_FPR10(\thread) 69 sdc1 $f12, THREAD_FPR12(\thread) 70 sdc1 $f14, THREAD_FPR14(\thread) 71 sdc1 $f16, THREAD_FPR16(\thread) [all …]
|
D | asmmacro-64.h | 16 .macro cpu_save_nonscratch thread 17 LONG_S s0, THREAD_REG16(\thread) 18 LONG_S s1, THREAD_REG17(\thread) 19 LONG_S s2, THREAD_REG18(\thread) 20 LONG_S s3, THREAD_REG19(\thread) 21 LONG_S s4, THREAD_REG20(\thread) 22 LONG_S s5, THREAD_REG21(\thread) 23 LONG_S s6, THREAD_REG22(\thread) 24 LONG_S s7, THREAD_REG23(\thread) 25 LONG_S sp, THREAD_REG29(\thread) [all …]
|
D | dsp.h | 45 tsk->thread.dsp.dspr[0] = mfhi1(); \ 46 tsk->thread.dsp.dspr[1] = mflo1(); \ 47 tsk->thread.dsp.dspr[2] = mfhi2(); \ 48 tsk->thread.dsp.dspr[3] = mflo2(); \ 49 tsk->thread.dsp.dspr[4] = mfhi3(); \ 50 tsk->thread.dsp.dspr[5] = mflo3(); \ 51 tsk->thread.dsp.dspcontrol = rddsp(DSP_MASK); \ 62 mthi1(tsk->thread.dsp.dspr[0]); \ 63 mtlo1(tsk->thread.dsp.dspr[1]); \ 64 mthi2(tsk->thread.dsp.dspr[2]); \ [all …]
|
D | cop2.h | 19 #define cop2_save(r) octeon_cop2_save(&(r)->thread.cp2) 20 #define cop2_restore(r) octeon_cop2_restore(&(r)->thread.cp2) 30 #define cop2_save(r) nlm_cop2_save(&(r)->thread.cp2) 31 #define cop2_restore(r) nlm_cop2_restore(&(r)->thread.cp2)
|
/linux-4.4.14/tools/perf/tests/ |
D | dwarf-unwind.c | 71 static int unwind_thread(struct thread *thread) in unwind_thread() argument 79 if (test__arch_unwind_sample(&sample, thread)) { in unwind_thread() 84 err = unwind__get_entries(unwind_entry, &cnt, thread, in unwind_thread() 106 struct thread *thread = *(struct thread **)p1; in compare() local 109 global_unwind_retval = unwind_thread(thread); in compare() 115 static int krava_3(struct thread *thread) in krava_3() argument 117 struct thread *array[2] = {thread, thread}; in krava_3() 129 _bsearch(array, &thread, 2, sizeof(struct thread **), compare); in krava_3() 134 static int krava_2(struct thread *thread) in krava_2() argument 136 return krava_3(thread); in krava_2() [all …]
|
D | hists_common.c | 91 struct thread *thread; in setup_fake_machine() local 93 thread = machine__findnew_thread(machine, fake_threads[i].pid, in setup_fake_machine() 95 if (thread == NULL) in setup_fake_machine() 98 thread__set_comm(thread, fake_threads[i].comm, 0); in setup_fake_machine() 99 thread__put(thread); in setup_fake_machine() 177 i, thread__comm_str(he->thread), in print_hists_in() 204 i, thread__comm_str(he->thread), he->thread->tid, in print_hists_out()
|
D | hists_link.c | 16 struct thread *thread; member 99 fake_common_samples[k].thread = al.thread; in add_hist_entries() 125 fake_samples[i][k].thread = al.thread; in add_hist_entries() 140 struct thread *t, struct map *m, struct symbol *s) in find_sample() 143 if (samples->thread == t && samples->map == m && in find_sample() 174 he->thread, he->ms.map, he->ms.sym)) { in __validate_match() 226 he->thread, he->ms.map, he->ms.sym) && in __validate_link() 229 he->thread, he->ms.map, he->ms.sym)) { in __validate_link()
|
D | code-reading.c | 178 struct thread *thread, struct state *state) in read_object_code() argument 189 thread__find_addr_map(thread, cpumode, MAP__FUNCTION, addr, &al); in read_object_code() 213 ret_len = dso__data_read_offset(al.map->dso, thread->mg->machine, in read_object_code() 295 struct thread *thread; in process_sample_event() local 304 thread = machine__findnew_thread(machine, sample.pid, sample.tid); in process_sample_event() 305 if (!thread) { in process_sample_event() 312 ret = read_object_code(sample.ip, READLEN, cpumode, thread, state); in process_sample_event() 313 thread__put(thread); in process_sample_event() 438 struct thread *thread; in do_test_code_reading() local 506 thread = machine__findnew_thread(machine, pid, pid); in do_test_code_reading() [all …]
|
D | mmap-thread-lookup.c | 184 struct thread *thread; in mmap_events() local 186 thread = machine__findnew_thread(machine, getpid(), td->tid); in mmap_events() 190 thread__find_addr_map(thread, in mmap_events() 194 thread__put(thread); in mmap_events()
|
D | tests.h | 74 struct thread; 77 struct thread *thread);
|
D | thread-mg-share.c | 13 struct thread *leader; in test__thread_mg_share() 14 struct thread *t1, *t2, *t3; in test__thread_mg_share() 18 struct thread *other, *other_leader; in test__thread_mg_share()
|
D | hists_filter.c | 16 struct thread *thread; member 94 fake_samples[i].thread = al.thread; in add_hist_entries() 170 hists->thread_filter = fake_samples[9].thread; in test__hists_filter() 292 hists->thread_filter = fake_samples[1].thread; in test__hists_filter()
|
/linux-4.4.14/arch/score/kernel/ |
D | ptrace.c | 180 child->thread.single_step = 1; in user_enable_single_step() 181 child->thread.ss_nextcnt = 1; in user_enable_single_step() 191 child->thread.ss_nextcnt = 2; in user_enable_single_step() 198 child->thread.ss_nextcnt = 2; in user_enable_single_step() 211 child->thread.ss_nextcnt = 2; in user_enable_single_step() 220 child->thread.ss_nextcnt = 2; in user_enable_single_step() 228 if (child->thread.ss_nextcnt == 1) { in user_enable_single_step() 240 child->thread.insn1_type = 0; in user_enable_single_step() 241 child->thread.addr1 = epc; in user_enable_single_step() 243 child->thread.insn1 = (short)epc_insn; in user_enable_single_step() [all …]
|
D | asm-offsets.c | 117 OFFSET(THREAD_REG0, task_struct, thread.reg0); in output_thread_defines() 118 OFFSET(THREAD_REG2, task_struct, thread.reg2); in output_thread_defines() 119 OFFSET(THREAD_REG3, task_struct, thread.reg3); in output_thread_defines() 120 OFFSET(THREAD_REG12, task_struct, thread.reg12); in output_thread_defines() 121 OFFSET(THREAD_REG13, task_struct, thread.reg13); in output_thread_defines() 122 OFFSET(THREAD_REG14, task_struct, thread.reg14); in output_thread_defines() 123 OFFSET(THREAD_REG15, task_struct, thread.reg15); in output_thread_defines() 124 OFFSET(THREAD_REG16, task_struct, thread.reg16); in output_thread_defines() 125 OFFSET(THREAD_REG17, task_struct, thread.reg17); in output_thread_defines() 126 OFFSET(THREAD_REG18, task_struct, thread.reg18); in output_thread_defines() [all …]
|
D | process.c | 78 p->thread.reg0 = (unsigned long) childregs; in copy_thread() 81 p->thread.reg12 = usp; in copy_thread() 82 p->thread.reg13 = arg; in copy_thread() 83 p->thread.reg3 = (unsigned long) ret_from_kernel_thread; in copy_thread() 90 p->thread.reg3 = (unsigned long) ret_from_fork; in copy_thread() 93 p->thread.cp0_psr = 0; in copy_thread()
|
/linux-4.4.14/arch/powerpc/kernel/ |
D | ptrace.c | 173 return task->thread.regs->msr | task->thread.fpexc_mode; in get_user_msr() 178 task->thread.regs->msr &= ~MSR_DEBUGCHANGE; in set_user_msr() 179 task->thread.regs->msr |= msr & MSR_DEBUGCHANGE; in set_user_msr() 186 *data = task->thread.dscr; in get_user_dscr() 192 task->thread.dscr = dscr; in set_user_dscr() 193 task->thread.dscr_inherit = 1; in set_user_dscr() 214 task->thread.regs->trap = trap & 0xfff0; in set_user_trap() 223 if ((task->thread.regs == NULL) || !data) in ptrace_get_reg() 235 *data = ((unsigned long *)task->thread.regs)[regno]; in ptrace_get_reg() 247 if (task->thread.regs == NULL) in ptrace_put_reg() [all …]
|
D | process.c | 86 if (tsk == current && tsk->thread.regs && in giveup_fpu_maybe_transactional() 87 MSR_TM_ACTIVE(tsk->thread.regs->msr) && in giveup_fpu_maybe_transactional() 89 tsk->thread.ckpt_regs.msr = tsk->thread.regs->msr; in giveup_fpu_maybe_transactional() 104 if (tsk == current && tsk->thread.regs && in giveup_altivec_maybe_transactional() 105 MSR_TM_ACTIVE(tsk->thread.regs->msr) && in giveup_altivec_maybe_transactional() 107 tsk->thread.ckpt_regs.msr = tsk->thread.regs->msr; in giveup_altivec_maybe_transactional() 126 if (tsk->thread.regs) { in flush_fp_to_thread() 136 if (tsk->thread.regs->msr & MSR_FP) { in flush_fp_to_thread() 160 if (current->thread.regs && (current->thread.regs->msr & MSR_FP)) in enable_kernel_fp() 176 if (current->thread.regs && (current->thread.regs->msr & MSR_VEC)) in enable_kernel_altivec() [all …]
|
D | signal_32.c | 272 buf[i] = task->thread.TS_FPR(i); in copy_fpr_to_user() 273 buf[i] = task->thread.fp_state.fpscr; in copy_fpr_to_user() 286 task->thread.TS_FPR(i) = buf[i]; in copy_fpr_from_user() 287 task->thread.fp_state.fpscr = buf[i]; in copy_fpr_from_user() 300 buf[i] = task->thread.fp_state.fpr[i][TS_VSRLOWOFFSET]; in copy_vsx_to_user() 313 task->thread.fp_state.fpr[i][TS_VSRLOWOFFSET] = buf[i]; in copy_vsx_from_user() 326 buf[i] = task->thread.TS_TRANS_FPR(i); in copy_transact_fpr_to_user() 327 buf[i] = task->thread.transact_fp.fpscr; in copy_transact_fpr_to_user() 340 task->thread.TS_TRANS_FPR(i) = buf[i]; in copy_transact_fpr_from_user() 341 task->thread.transact_fp.fpscr = buf[i]; in copy_transact_fpr_from_user() [all …]
|
D | signal_64.c | 115 if (current->thread.used_vr) { in setup_sigcontext() 118 err |= __copy_to_user(v_regs, ¤t->thread.vr_state, in setup_sigcontext() 129 current->thread.vrsave = mfspr(SPRN_VRSAVE); in setup_sigcontext() 130 err |= __put_user(current->thread.vrsave, (u32 __user *)&v_regs[33]); in setup_sigcontext() 149 if (current->thread.used_vsr && ctx_has_vsx_region) { in setup_sigcontext() 219 if (current->thread.used_vr) { in setup_tm_sigcontexts() 222 err |= __copy_to_user(v_regs, ¤t->thread.vr_state, in setup_tm_sigcontexts() 229 ¤t->thread.transact_vr, in setup_tm_sigcontexts() 233 ¤t->thread.vr_state, in setup_tm_sigcontexts() 245 current->thread.vrsave = mfspr(SPRN_VRSAVE); in setup_tm_sigcontexts() [all …]
|
D | ptrace32.c | 95 CHECK_FULL_REGS(child->thread.regs); in compat_arch_ptrace() 107 tmp = ((unsigned int *)child->thread.fp_state.fpr) in compat_arch_ptrace() 145 CHECK_FULL_REGS(child->thread.regs); in compat_arch_ptrace() 149 tmp = child->thread.fp_state.fpr[numReg - PT_FPR0][0]; in compat_arch_ptrace() 198 CHECK_FULL_REGS(child->thread.regs); in compat_arch_ptrace() 208 ((unsigned int *)child->thread.fp_state.fpr) in compat_arch_ptrace() 237 CHECK_FULL_REGS(child->thread.regs); in compat_arch_ptrace() 252 tmp = &child->thread.fp_state.fpr[numReg - PT_FPR0][0]; in compat_arch_ptrace() 269 ret = put_user(child->thread.debug.dac1, (u32 __user *)data); in compat_arch_ptrace() 272 (child->thread.hw_brk.address & (~HW_BRK_TYPE_DABR)) | in compat_arch_ptrace() [all …]
|
D | uprobes.c | 68 autask->saved_trap_nr = current->thread.trap_nr; in arch_uprobe_pre_xol() 69 current->thread.trap_nr = UPROBE_TRAP_NR; in arch_uprobe_pre_xol() 99 if (t->thread.trap_nr != UPROBE_TRAP_NR) in arch_uprobe_xol_was_trapped() 116 WARN_ON_ONCE(current->thread.trap_nr != UPROBE_TRAP_NR); in arch_uprobe_post_xol() 118 current->thread.trap_nr = utask->autask.saved_trap_nr; in arch_uprobe_post_xol() 171 current->thread.trap_nr = utask->autask.saved_trap_nr; in arch_uprobe_abort_xol()
|
D | hw_breakpoint.c | 74 if (current->thread.last_hit_ubp != bp) in arch_install_hw_breakpoint() 114 bp->ctx->task->thread.last_hit_ubp = NULL; in arch_unregister_hw_breakpoint() 196 if (likely(!tsk->thread.last_hit_ubp)) in thread_change_pc() 199 info = counter_arch_bp(tsk->thread.last_hit_ubp); in thread_change_pc() 202 tsk->thread.last_hit_ubp = NULL; in thread_change_pc() 259 current->thread.last_hit_ubp = bp; in hw_breakpoint_handler() 302 bp = current->thread.last_hit_ubp; in single_step_dabr_instruction() 320 current->thread.last_hit_ubp = NULL; in single_step_dabr_instruction() 357 struct thread_struct *t = &tsk->thread; in flush_ptrace_hw_breakpoint()
|
D | traps.c | 263 current->thread.trap_nr = code; in _exception() 383 #define single_stepping(regs) (current->thread.debug.dbcr0 & DBCR0_IC) 384 #define clear_single_step(regs) (current->thread.debug.dbcr0 &= ~DBCR0_IC) 862 code = __parse_fpscr(current->thread.fp_state.fpscr); in parse_fpe() 1091 current->thread.dscr = regs->gpr[rd]; in emulate_instruction() 1092 current->thread.dscr_inherit = 1; in emulate_instruction() 1093 mtspr(SPRN_DSCR, current->thread.dscr); in emulate_instruction() 1122 code = __parse_fpscr(current->thread.fp_state.fpscr); in emulate_math() 1273 if (!(current->thread.align_ctl & PR_UNALIGN_SIGBUS)) in alignment_exception() 1419 current->thread.dscr = regs->gpr[rd]; in facility_unavailable_exception() [all …]
|
D | signal.c | 127 if (current->thread.hw_brk.address && in do_signal() 128 current->thread.hw_brk.type) in do_signal() 129 __set_breakpoint(¤t->thread.hw_brk); in do_signal() 193 return current->thread.ckpt_regs.gpr[1]; in get_tm_stackpointer()
|
/linux-4.4.14/arch/parisc/kernel/ |
D | asm-offsets.c | 66 DEFINE(TASK_REGS, offsetof(struct task_struct, thread.regs)); in main() 67 DEFINE(TASK_PT_PSW, offsetof(struct task_struct, thread.regs.gr[ 0])); in main() 68 DEFINE(TASK_PT_GR1, offsetof(struct task_struct, thread.regs.gr[ 1])); in main() 69 DEFINE(TASK_PT_GR2, offsetof(struct task_struct, thread.regs.gr[ 2])); in main() 70 DEFINE(TASK_PT_GR3, offsetof(struct task_struct, thread.regs.gr[ 3])); in main() 71 DEFINE(TASK_PT_GR4, offsetof(struct task_struct, thread.regs.gr[ 4])); in main() 72 DEFINE(TASK_PT_GR5, offsetof(struct task_struct, thread.regs.gr[ 5])); in main() 73 DEFINE(TASK_PT_GR6, offsetof(struct task_struct, thread.regs.gr[ 6])); in main() 74 DEFINE(TASK_PT_GR7, offsetof(struct task_struct, thread.regs.gr[ 7])); in main() 75 DEFINE(TASK_PT_GR8, offsetof(struct task_struct, thread.regs.gr[ 8])); in main() [all …]
|
/linux-4.4.14/arch/metag/kernel/ |
D | process.c | 191 memset(&tsk->thread.kernel_context, 0, in copy_thread() 192 sizeof(tsk->thread.kernel_context)); in copy_thread() 194 tsk->thread.kernel_context = __TBISwitchInit(kernel_context, in copy_thread() 211 tsk->thread.int_depth = 2; in copy_thread() 225 tsk->thread.int_depth = 1; in copy_thread() 232 tsk->thread.tls_ptr = in copy_thread() 236 if (tsk->thread.fpu_context) { in copy_thread() 239 ctx = kmemdup(tsk->thread.fpu_context, in copy_thread() 241 tsk->thread.fpu_context = ctx; in copy_thread() 246 if (tsk->thread.dsp_context) { in copy_thread() [all …]
|
D | smp.c | 71 static int boot_secondary(unsigned int thread, struct task_struct *idle) in boot_secondary() argument 81 core_reg_write(TXUPC_ID, 0, thread, (unsigned int)secondary_startup); in boot_secondary() 82 core_reg_write(TXUPC_ID, 1, thread, 0); in boot_secondary() 88 core_reg_write(TXUCT_ID, TXSTATUS_REGNUM, thread, TXSTATUS_PSTAT_BIT); in boot_secondary() 91 val = core_reg_read(TXUCT_ID, TXPRIVEXT_REGNUM, thread); in boot_secondary() 92 core_reg_write(TXUCT_ID, TXPRIVEXT_REGNUM, thread, val & ~0x80); in boot_secondary() 98 val = core_reg_read(TXUCT_ID, TXENABLE_REGNUM, thread); in boot_secondary() 99 core_reg_write(TXUCT_ID, TXENABLE_REGNUM, thread, val | 0x1); in boot_secondary() 121 static void describe_cachepart_change(unsigned int thread, const char *label, in describe_cachepart_change() argument 132 pr_info("Thread %d: %s partition changed:", thread, label); in describe_cachepart_change() [all …]
|
D | core_reg.c | 41 void core_reg_write(int unit, int reg, int thread, unsigned int val) in core_reg_write() argument 47 void __iomem *cu_reg = __CU_addr(thread, reg); in core_reg_write() 62 val = UNIT_VAL(unit) | REG_VAL(reg) | THREAD_VAL(thread); in core_reg_write() 83 unsigned int core_reg_read(int unit, int reg, int thread) in core_reg_read() argument 90 void __iomem *cu_reg = __CU_addr(thread, reg); in core_reg_read() 102 val = (UNIT_VAL(unit) | REG_VAL(reg) | THREAD_VAL(thread) | in core_reg_read()
|
D | traps.c | 215 dsp_ctx = current->thread.dsp_context; in nest_interrupts() 221 current->thread.dsp_context = dsp_ctx; in nest_interrupts() 224 current->thread.user_flags |= (D0_8 & 0xffff0000); in nest_interrupts() 241 current->thread.user_flags = savemask; in head_end() 282 BUG_ON(current->thread.int_depth > HARDIRQ_BITS); in head_end() 284 depth = current->thread.int_depth++; in head_end() 290 current->thread.txdefr_failure &= ~(1 << depth); in head_end() 292 current->thread.txdefr_failure |= (1 << depth); in head_end() 328 if (current->thread.fpu_context && in tail_end_sys() 329 current->thread.fpu_context->needs_restore) { in tail_end_sys() [all …]
|
/linux-4.4.14/arch/um/kernel/ |
D | process.c | 85 to->thread.prev_sched = from; in __switch_to() 88 switch_threads(&from->thread.switch_buf, &to->thread.switch_buf); in __switch_to() 91 return current->thread.prev_sched; in __switch_to() 96 struct pt_regs *regs = ¤t->thread.regs; in interrupt_end() 124 if (current->thread.prev_sched != NULL) in new_thread_handler() 125 schedule_tail(current->thread.prev_sched); in new_thread_handler() 126 current->thread.prev_sched = NULL; in new_thread_handler() 128 fn = current->thread.request.u.thread.proc; in new_thread_handler() 129 arg = current->thread.request.u.thread.arg; in new_thread_handler() 135 userspace(¤t->thread.regs.regs); in new_thread_handler() [all …]
|
D | trap.c | 169 current->thread.arch.faultinfo = fi; in bad_segv() 176 do_signal(¤t->thread.regs); in fatal_sigsegv() 213 current->thread.segv_regs = container_of(regs, struct pt_regs, regs); in segv() 242 catcher = current->thread.fault_catcher; in segv() 246 current->thread.fault_addr = (void *) address; in segv() 249 else if (current->thread.fault_addr != NULL) in segv() 267 current->thread.arch.faultinfo = fi; in segv() 273 current->thread.arch.faultinfo = fi; in segv() 279 current->thread.segv_regs = NULL; in segv() 310 current->thread.arch.faultinfo = *fi; in relay_signal() [all …]
|
/linux-4.4.14/arch/sh/kernel/cpu/sh4/ |
D | fpu.c | 88 :"0"((char *)(&tsk->thread.xstate->hardfpu.status)), in save_fpu() 138 :"0" (tsk->thread.xstate), "r" (FPSCR_RCHG) in restore_fpu() 234 if ((tsk->thread.xstate->hardfpu.fpscr & FPSCR_CAUSE_ERROR)) in ieee_fpe_handler() 236 denormal_to_double(&tsk->thread.xstate->hardfpu, in ieee_fpe_handler() 252 hx = tsk->thread.xstate->hardfpu.fp_regs[n]; in ieee_fpe_handler() 253 hy = tsk->thread.xstate->hardfpu.fp_regs[m]; in ieee_fpe_handler() 254 fpscr = tsk->thread.xstate->hardfpu.fpscr; in ieee_fpe_handler() 264 | tsk->thread.xstate->hardfpu.fp_regs[n + 1]; in ieee_fpe_handler() 266 | tsk->thread.xstate->hardfpu.fp_regs[m + 1]; in ieee_fpe_handler() 268 tsk->thread.xstate->hardfpu.fp_regs[n] = llx >> 32; in ieee_fpe_handler() [all …]
|
/linux-4.4.14/arch/s390/kernel/ |
D | process.c | 59 sf = (struct stack_frame *) (tsk->thread.ksp & PSW_ADDR_INSN); in thread_saved_pc() 89 kfree(tsk->thread.fpu.regs); in arch_release_task_struct() 104 dst->thread.fpu.regs = kzalloc(fpu_regs_size, GFP_KERNEL|__GFP_REPEAT); in arch_dup_task_struct() 105 if (!dst->thread.fpu.regs) in arch_dup_task_struct() 114 dst->thread.fpu.fpc = current->thread.fpu.fpc; in arch_dup_task_struct() 115 memcpy(dst->thread.fpu.regs, current->thread.fpu.regs, fpu_regs_size); in arch_dup_task_struct() 131 p->thread.ksp = (unsigned long) frame; in copy_thread() 133 save_access_regs(&p->thread.acrs[0]); in copy_thread() 135 p->thread.mm_segment = get_fs(); in copy_thread() 137 memset(&p->thread.per_user, 0, sizeof(p->thread.per_user)); in copy_thread() [all …]
|
D | ptrace.c | 44 struct thread_struct *thread = &task->thread; in update_cr_regs() local 54 if (task->thread.per_flags & PER_FLAG_NO_TE) in update_cr_regs() 61 if (task->thread.per_flags & PER_FLAG_TE_ABORT_RAND) { in update_cr_regs() 62 if (task->thread.per_flags & PER_FLAG_TE_ABORT_RAND_TEND) in update_cr_regs() 71 new.control = thread->per_user.control; in update_cr_regs() 72 new.start = thread->per_user.start; in update_cr_regs() 73 new.end = thread->per_user.end; in update_cr_regs() 126 memset(&task->thread.per_user, 0, sizeof(task->thread.per_user)); in ptrace_disable() 127 memset(&task->thread.per_event, 0, sizeof(task->thread.per_event)); in ptrace_disable() 130 task->thread.per_flags = 0; in ptrace_disable() [all …]
|
D | runtime_instr.c | 50 if (!task->thread.ri_cb) in exit_thread_runtime_instr() 53 kfree(task->thread.ri_cb); in exit_thread_runtime_instr() 54 task->thread.ri_cb = NULL; in exit_thread_runtime_instr() 74 if (!current->thread.ri_cb) { in SYSCALL_DEFINE1() 79 cb = current->thread.ri_cb; in SYSCALL_DEFINE1() 87 current->thread.ri_cb = cb; in SYSCALL_DEFINE1()
|
D | uprobes.c | 68 regs->psw.addr >= current->thread.per_user.start && in check_per_event() 69 regs->psw.addr <= current->thread.per_user.end) in check_per_event() 98 if (check_per_event(current->thread.per_event.cause, in arch_uprobe_post_xol() 99 current->thread.per_user.control, regs)) { in arch_uprobe_post_xol() 101 current->thread.per_event.address = utask->vaddr; in arch_uprobe_post_xol() 137 current->thread.per_event.address = current->utask->vaddr; in arch_uprobe_abort_xol() 242 if (!(current->thread.per_user.control & PER_EVENT_STORE)) in sim_stor_event() 244 if ((void *)current->thread.per_user.start > (addr + len)) in sim_stor_event() 246 if ((void *)current->thread.per_user.end < addr) in sim_stor_event() 248 current->thread.per_event.address = regs->psw.addr; in sim_stor_event() [all …]
|
D | signal.c | 107 save_access_regs(current->thread.acrs); in store_sigregs() 114 restore_access_regs(current->thread.acrs); in load_sigregs() 128 memcpy(&user_sregs.regs.acrs, current->thread.acrs, in save_sigregs() 130 fpregs_store(&user_sregs.fpregs, ¤t->thread.fpu); in save_sigregs() 165 memcpy(¤t->thread.acrs, &user_sregs.regs.acrs, in restore_sigregs() 166 sizeof(current->thread.acrs)); in restore_sigregs() 168 fpregs_load(&user_sregs.fpregs, ¤t->thread.fpu); in restore_sigregs() 184 vxrs[i] = *((__u64 *)(current->thread.fpu.vxrs + i) + 1); in save_sigregs_ext() 188 current->thread.fpu.vxrs + __NUM_VXRS_LOW, in save_sigregs_ext() 205 __copy_from_user(current->thread.fpu.vxrs + __NUM_VXRS_LOW, in restore_sigregs_ext() [all …]
|
/linux-4.4.14/arch/sparc/kernel/ |
D | sigutil_32.c | 20 fpsave(¤t->thread.float_regs[0], ¤t->thread.fsr, in save_fpu_state() 21 ¤t->thread.fpqueue[0], ¤t->thread.fpqdepth); in save_fpu_state() 28 fpsave(¤t->thread.float_regs[0], ¤t->thread.fsr, in save_fpu_state() 29 ¤t->thread.fpqueue[0], ¤t->thread.fpqdepth); in save_fpu_state() 35 ¤t->thread.float_regs[0], in save_fpu_state() 37 err |= __put_user(current->thread.fsr, &fpu->si_fsr); in save_fpu_state() 38 err |= __put_user(current->thread.fpqdepth, &fpu->si_fpqdepth); in save_fpu_state() 39 if (current->thread.fpqdepth != 0) in save_fpu_state() 41 ¤t->thread.fpqueue[0], in save_fpu_state() 70 err = __copy_from_user(¤t->thread.float_regs[0], &fpu->si_float_regs[0], in restore_fpu_state() [all …]
|
D | process_32.c | 196 fpsave(¤t->thread.float_regs[0], ¤t->thread.fsr, 197 ¤t->thread.fpqueue[0], ¤t->thread.fpqdepth); 217 fpsave(¤t->thread.float_regs[0], ¤t->thread.fsr, 218 ¤t->thread.fpqueue[0], ¤t->thread.fpqdepth); 227 if (current->thread.flags & SPARC_FLAG_KTHREAD) { 228 current->thread.flags &= ~SPARC_FLAG_KTHREAD; 232 current->thread.kregs = (struct pt_regs *) 321 fpsave(&p->thread.float_regs[0], &p->thread.fsr, 322 &p->thread.fpqueue[0], &p->thread.fpqdepth); 343 p->thread.kregs = childregs; [all …]
|
D | traps_32.c | 195 fpsave(&fptask->thread.float_regs[0], &fptask->thread.fsr, in do_fpd_trap() 196 &fptask->thread.fpqueue[0], &fptask->thread.fpqdepth); in do_fpd_trap() 200 fpload(¤t->thread.float_regs[0], ¤t->thread.fsr); in do_fpd_trap() 211 fpload(¤t->thread.float_regs[0], ¤t->thread.fsr); in do_fpd_trap() 248 fpsave(&fpt->thread.float_regs[0], &fpt->thread.fsr, 249 &fpt->thread.fpqueue[0], &fpt->thread.fpqdepth); 251 printk("Hmm, FP exception, fsr was %016lx\n", fpt->thread.fsr); 254 switch ((fpt->thread.fsr & 0x1c000)) { 279 fpload(¤t->thread.float_regs[0], ¤t->thread.fsr); 302 fsr = fpt->thread.fsr;
|
/linux-4.4.14/arch/m68k/kernel/ |
D | process.c | 46 struct switch_stack *sw = (struct switch_stack *)tsk->thread.ksp; in thread_saved_pc() 105 current->thread.fs = __USER_DS; in flush_thread() 142 p->thread.ksp = (unsigned long)frame; in copy_thread() 143 p->thread.esp0 = (unsigned long)&frame->regs; in copy_thread() 149 p->thread.fs = get_fs().seg; in copy_thread() 158 p->thread.usp = 0; in copy_thread() 165 p->thread.usp = usp ?: rdusp(); in copy_thread() 173 asm volatile ("fsave %0" : : "m" (p->thread.fpstate[0]) : "memory"); in copy_thread() 175 if (!CPU_IS_060 ? p->thread.fpstate[0] : p->thread.fpstate[2]) { in copy_thread() 182 : "m" (p->thread.fp[0]), in copy_thread() [all …]
|
D | ptrace.c | 79 addr = &task->thread.usp; in get_reg() 81 addr = (unsigned long *)(task->thread.esp0 + regoff[regno]); in get_reg() 86 long stkadj = *(long *)(task->thread.esp0 + PT_REG(stkadj)); in get_reg() 104 addr = &task->thread.usp; in put_reg() 106 addr = (unsigned long *)(task->thread.esp0 + regoff[regno]); in put_reg() 111 long stkadj = *(long *)(task->thread.esp0 + PT_REG(stkadj)); in put_reg() 178 tmp = child->thread.fp[regno - 21]; in arch_ptrace() 219 child->thread.fp[regno - 21] = data; in arch_ptrace() 249 if (copy_to_user(datap, &child->thread.fp, in arch_ptrace() 255 if (copy_from_user(&child->thread.fp, datap, in arch_ptrace()
|
/linux-4.4.14/arch/x86/kernel/ |
D | process_64.c | 146 struct desc_struct *desc = t->thread.tls_array; in set_32bit_tls() 153 return get_desc_base(&t->thread.tls_array[tls]); in read_32bit_tls() 163 p->thread.sp0 = (unsigned long)task_stack_page(p) + THREAD_SIZE; in copy_thread_tls() 165 p->thread.sp = (unsigned long) childregs; in copy_thread_tls() 167 p->thread.io_bitmap_ptr = NULL; in copy_thread_tls() 169 savesegment(gs, p->thread.gsindex); in copy_thread_tls() 170 p->thread.gs = p->thread.gsindex ? 0 : me->thread.gs; in copy_thread_tls() 171 savesegment(fs, p->thread.fsindex); in copy_thread_tls() 172 p->thread.fs = p->thread.fsindex ? 0 : me->thread.fs; in copy_thread_tls() 173 savesegment(es, p->thread.es); in copy_thread_tls() [all …]
|
D | process_32.c | 66 return ((unsigned long *)tsk->thread.sp)[3]; in thread_saved_pc() 139 p->thread.sp = (unsigned long) childregs; in copy_thread_tls() 140 p->thread.sp0 = (unsigned long) (childregs+1); in copy_thread_tls() 141 memset(p->thread.ptrace_bps, 0, sizeof(p->thread.ptrace_bps)); in copy_thread_tls() 146 p->thread.ip = (unsigned long) ret_from_kernel_thread; in copy_thread_tls() 156 p->thread.io_bitmap_ptr = NULL; in copy_thread_tls() 164 p->thread.ip = (unsigned long) ret_from_fork; in copy_thread_tls() 167 p->thread.io_bitmap_ptr = NULL; in copy_thread_tls() 172 p->thread.io_bitmap_ptr = kmemdup(tsk->thread.io_bitmap_ptr, in copy_thread_tls() 174 if (!p->thread.io_bitmap_ptr) { in copy_thread_tls() [all …]
|
D | ptrace.c | 283 return task->thread.fsindex; in get_segment_reg() 289 return task->thread.gsindex; in get_segment_reg() 295 return task->thread.ds; in get_segment_reg() 301 return task->thread.es; in get_segment_reg() 325 if ((value == FS_TLS_SEL && task->thread.fsindex == 0 && in set_segment_reg() 326 task->thread.fs != 0) || in set_segment_reg() 327 (value == 0 && task->thread.fsindex == FS_TLS_SEL && in set_segment_reg() 328 task->thread.fs == 0)) in set_segment_reg() 330 task->thread.fsindex = value; in set_segment_reg() 332 loadsegment(fs, task->thread.fsindex); in set_segment_reg() [all …]
|
D | traps.c | 203 tsk->thread.error_code = error_code; in do_trap_no_signal() 204 tsk->thread.trap_nr = trapnr; in do_trap_no_signal() 262 tsk->thread.error_code = error_code; in do_trap() 263 tsk->thread.trap_nr = trapnr; in do_trap() 347 tsk->thread.error_code = error_code; in do_double_fault() 348 tsk->thread.trap_nr = X86_TRAP_DF; in do_double_fault() 459 tsk->thread.error_code = error_code; in do_general_protection() 460 tsk->thread.trap_nr = X86_TRAP_GP; in do_general_protection() 467 tsk->thread.error_code = error_code; in do_general_protection() 468 tsk->thread.trap_nr = X86_TRAP_GP; in do_general_protection() [all …]
|
D | process.c | 88 dst->thread.vm86 = NULL; in arch_dup_task_struct() 91 return fpu__copy(&dst->thread.fpu, &src->thread.fpu); in arch_dup_task_struct() 100 struct thread_struct *t = &me->thread; in exit_thread() 128 memset(tsk->thread.tls_array, 0, sizeof(tsk->thread.tls_array)); in flush_thread() 130 fpu__clear(&tsk->thread.fpu); in flush_thread() 196 prev = &prev_p->thread; in __switch_to_xtra() 197 next = &next_p->thread; in __switch_to_xtra() 552 sp = READ_ONCE(p->thread.sp); in get_wchan()
|
D | vm86_32.c | 84 #define VFLAGS (*(unsigned short *)&(current->thread.vm86->veflags)) 85 #define VEFLAGS (current->thread.vm86->veflags) 98 struct vm86 *vm86 = current->thread.vm86; in save_v86_state() 148 tsk->thread.sp0 = vm86->saved_sp0; in save_v86_state() 149 tsk->thread.sysenter_cs = __KERNEL_CS; in save_v86_state() 150 load_sp0(tss, &tsk->thread); in save_v86_state() 231 struct vm86 *vm86 = tsk->thread.vm86; in do_sys_vm86() 265 tsk->thread.vm86 = vm86; in do_sys_vm86() 354 vm86->saved_sp0 = tsk->thread.sp0; in do_sys_vm86() 359 tsk->thread.sp0 += 16; in do_sys_vm86() [all …]
|
D | tls.c | 21 struct thread_struct *t = ¤t->thread; in get_free_idx() 85 struct thread_struct *t = &p->thread; in set_tls_desc() 103 if (t == ¤t->thread) in set_tls_desc() 188 &p->thread.tls_array[idx - GDT_ENTRY_TLS_MIN]); in do_get_thread_area() 203 struct thread_struct *t = &target->thread; in regset_tls_active() 224 tls = &target->thread.tls_array[pos]; in regset_tls_get()
|
D | hw_breakpoint.c | 378 struct thread_struct *thread = ¤t->thread; in aout_dump_debugregs() local 381 bp = thread->ptrace_bps[i]; in aout_dump_debugregs() 394 dump->u_debugreg[6] = current->thread.debugreg6; in aout_dump_debugregs() 406 struct thread_struct *t = &tsk->thread; in flush_ptrace_hw_breakpoint() 423 set_debugreg(current->thread.debugreg6, 6); in hw_breakpoint_restore() 471 current->thread.debugreg6 &= ~DR_TRAP_BITS; in hw_breakpoint_handler() 518 if ((current->thread.debugreg6 & DR_TRAP_BITS) || in hw_breakpoint_handler()
|
/linux-4.4.14/arch/frv/kernel/ |
D | process.c | 140 p->thread.frame = childregs; in copy_thread() 141 p->thread.curr = p; in copy_thread() 142 p->thread.sp = (unsigned long) childregs; in copy_thread() 143 p->thread.fp = 0; in copy_thread() 144 p->thread.lr = 0; in copy_thread() 145 p->thread.frame0 = childregs; in copy_thread() 150 p->thread.pc = (unsigned long) ret_from_kernel_thread; in copy_thread() 151 save_user_regs(p->thread.user); in copy_thread() 158 p->thread.pc = (unsigned long) ret_from_fork; in copy_thread() 164 save_user_regs(p->thread.user); in copy_thread() [all …]
|
D | ptrace.c | 45 const struct user_int_regs *iregs = &target->thread.user->i; in genregs_get() 65 struct user_int_regs *iregs = &target->thread.user->i; in genregs_set() 110 const struct user_fpmedia_regs *fpregs = &target->thread.user->f; in fpmregs_get() 130 struct user_fpmedia_regs *fpregs = &target->thread.user->f; in fpmregs_set() 205 struct user_context *user = task->thread.user; in get_reg() 219 struct user_context *user = task->thread.user; in put_reg() 243 child->thread.frame0->__status |= REG__STATUS_STEP; in user_enable_single_step() 248 child->thread.frame0->__status &= ~REG__STATUS_STEP; in user_disable_single_step() 323 0, sizeof(child->thread.user->i), in arch_ptrace() 329 0, sizeof(child->thread.user->i), in arch_ptrace() [all …]
|
/linux-4.4.14/arch/mn10300/kernel/ |
D | fpu.c | 52 info.si_addr = (void *) tsk->thread.uregs->pc; in fpu_exception() 57 fpcr = tsk->thread.fpu_state.fpcr; in fpu_exception() 88 if (tsk->thread.fpu_flags & THREAD_HAS_FPU) { in fpu_setup_sigcontext() 89 fpu_save(&tsk->thread.fpu_state); in fpu_setup_sigcontext() 90 tsk->thread.uregs->epsw &= ~EPSW_FE; in fpu_setup_sigcontext() 91 tsk->thread.fpu_flags &= ~THREAD_HAS_FPU; in fpu_setup_sigcontext() 95 fpu_save(&tsk->thread.fpu_state); in fpu_setup_sigcontext() 96 fpu_state_owner->thread.uregs->epsw &= ~EPSW_FE; in fpu_setup_sigcontext() 108 &tsk->thread.fpu_state, in fpu_setup_sigcontext() 125 if (tsk->thread.fpu_flags & THREAD_HAS_FPU) { in fpu_kill_state() [all …]
|
D | process.c | 44 return ((unsigned long *) tsk->thread.sp)[3]; in thread_saved_pc() 159 p->thread.uregs = c_regs; in copy_thread() 161 p->thread.a3 = (unsigned long) c_regs; in copy_thread() 162 p->thread.sp = c_ksp; in copy_thread() 163 p->thread.wchan = p->thread.pc; in copy_thread() 164 p->thread.usp = c_usp; in copy_thread() 172 p->thread.pc = (unsigned long) ret_from_kernel_thread; in copy_thread() 184 p->thread.pc = (unsigned long) ret_from_fork; in copy_thread() 191 return p->thread.wchan; in get_wchan()
|
/linux-4.4.14/samples/pktgen/ |
D | pktgen_bench_xmit_mode_netif_receive.sh | 49 for ((thread = 0; thread < $THREADS; thread++)); do 52 dev=${DEV}@${thread} 55 pg_thread $thread "rem_device_all" 56 pg_thread $thread "add_device" $dev 82 for ((thread = 0; thread < $THREADS; thread++)); do 83 dev=${DEV}@${thread}
|
D | pktgen_sample03_burst_single_flow.sh | 41 for ((thread = 0; thread < $THREADS; thread++)); do 42 dev=${DEV}@${thread} 45 pg_thread $thread "rem_device_all" 46 pg_thread $thread "add_device" $dev 72 for ((thread = 0; thread < $THREADS; thread++)); do 73 dev=${DEV}@${thread}
|
D | pktgen_sample02_multiqueue.sh | 33 for ((thread = 0; thread < $THREADS; thread++)); do 36 dev=${DEV}@${thread} 39 pg_thread $thread "rem_device_all" 40 pg_thread $thread "add_device" $dev 71 for ((thread = 0; thread < $THREADS; thread++)); do 72 dev=${DEV}@${thread}
|
D | pktgen_sample01_simple.sh | 35 thread=0 36 pg_thread $thread "rem_device_all" 37 pg_thread $thread "add_device" $DEV
|
/linux-4.4.14/drivers/acpi/acpica/ |
D | exmutex.c | 55 struct acpi_thread_state *thread); 71 struct acpi_thread_state *thread = obj_desc->mutex.owner_thread; in acpi_ex_unlink_mutex() local 73 if (!thread) { in acpi_ex_unlink_mutex() 95 thread->acquired_mutex_list = obj_desc->mutex.next; in acpi_ex_unlink_mutex() 114 struct acpi_thread_state *thread) in acpi_ex_link_mutex() argument 118 list_head = thread->acquired_mutex_list; in acpi_ex_link_mutex() 133 thread->acquired_mutex_list = obj_desc; in acpi_ex_link_mutex() 238 if (!walk_state->thread) { in acpi_ex_acquire_mutex() 249 if (walk_state->thread->current_sync_level > obj_desc->mutex.sync_level) { in acpi_ex_acquire_mutex() 253 walk_state->thread->current_sync_level)); in acpi_ex_acquire_mutex() [all …]
|
D | dswstate.c | 447 *thread) in acpi_ds_get_current_walk_state() 451 if (!thread) { in acpi_ds_get_current_walk_state() 456 thread->walk_state_list)); in acpi_ds_get_current_walk_state() 458 return (thread->walk_state_list); in acpi_ds_get_current_walk_state() 476 struct acpi_thread_state *thread) in acpi_ds_push_walk_state() argument 480 walk_state->next = thread->walk_state_list; in acpi_ds_push_walk_state() 481 thread->walk_state_list = walk_state; in acpi_ds_push_walk_state() 500 struct acpi_walk_state *acpi_ds_pop_walk_state(struct acpi_thread_state *thread) in acpi_ds_pop_walk_state() argument 506 walk_state = thread->walk_state_list; in acpi_ds_pop_walk_state() 512 thread->walk_state_list = walk_state->next; in acpi_ds_pop_walk_state() [all …]
|
D | psparse.c | 444 struct acpi_thread_state *thread; in acpi_ps_parse_aml() local 461 thread = acpi_ut_create_thread_state(); in acpi_ps_parse_aml() 462 if (!thread) { in acpi_ps_parse_aml() 476 walk_state->thread = thread; in acpi_ps_parse_aml() 483 walk_state->thread->current_sync_level = in acpi_ps_parse_aml() 487 acpi_ds_push_walk_state(walk_state, thread); in acpi_ps_parse_aml() 493 acpi_gbl_current_walk_list = thread; in acpi_ps_parse_aml() 521 acpi_ds_call_control_method(thread, walk_state, in acpi_ps_parse_aml() 532 walk_state = acpi_ds_get_current_walk_state(thread); in acpi_ps_parse_aml() 562 walk_state = acpi_ds_pop_walk_state(thread); in acpi_ps_parse_aml() [all …]
|
D | dsdebug.c | 128 struct acpi_thread_state *thread; in acpi_ds_dump_method_stack() local 154 thread = walk_state->thread; in acpi_ds_dump_method_stack() 155 if (!thread) { in acpi_ds_dump_method_stack() 170 next_walk_state = thread->walk_state_list; in acpi_ds_dump_method_stack()
|
D | dsmethod.c | 375 && (walk_state->thread->current_sync_level > in acpi_ds_begin_method_execution() 380 walk_state->thread->current_sync_level)); in acpi_ds_begin_method_execution() 391 (walk_state->thread->thread_id != in acpi_ds_begin_method_execution() 410 walk_state->thread->current_sync_level; in acpi_ds_begin_method_execution() 413 walk_state->thread->thread_id; in acpi_ds_begin_method_execution() 414 walk_state->thread->current_sync_level = in acpi_ds_begin_method_execution() 475 acpi_ds_call_control_method(struct acpi_thread_state *thread, in acpi_ds_call_control_method() argument 516 NULL, obj_desc, thread); in acpi_ds_call_control_method() 724 walk_state->thread->current_sync_level = in acpi_ds_terminate_control_method()
|
/linux-4.4.14/arch/arm/vfp/ |
D | vfpmodule.c | 65 static bool vfp_state_in_hw(unsigned int cpu, struct thread_info *thread) in vfp_state_in_hw() argument 68 if (thread->vfpstate.hard.cpu != cpu) in vfp_state_in_hw() 71 return vfp_current_hw_state[cpu] == &thread->vfpstate; in vfp_state_in_hw() 79 static void vfp_force_reload(unsigned int cpu, struct thread_info *thread) in vfp_force_reload() argument 81 if (vfp_state_in_hw(cpu, thread)) { in vfp_force_reload() 86 thread->vfpstate.hard.cpu = NR_CPUS; in vfp_force_reload() 93 static void vfp_thread_flush(struct thread_info *thread) in vfp_thread_flush() argument 95 union vfp_state *vfp = &thread->vfpstate; in vfp_thread_flush() 121 static void vfp_thread_exit(struct thread_info *thread) in vfp_thread_exit() argument 124 union vfp_state *vfp = &thread->vfpstate; in vfp_thread_exit() [all …]
|
/linux-4.4.14/arch/mips/kernel/ |
D | asm-offsets.c | 110 OFFSET(THREAD_REG16, task_struct, thread.reg16); in output_thread_defines() 111 OFFSET(THREAD_REG17, task_struct, thread.reg17); in output_thread_defines() 112 OFFSET(THREAD_REG18, task_struct, thread.reg18); in output_thread_defines() 113 OFFSET(THREAD_REG19, task_struct, thread.reg19); in output_thread_defines() 114 OFFSET(THREAD_REG20, task_struct, thread.reg20); in output_thread_defines() 115 OFFSET(THREAD_REG21, task_struct, thread.reg21); in output_thread_defines() 116 OFFSET(THREAD_REG22, task_struct, thread.reg22); in output_thread_defines() 117 OFFSET(THREAD_REG23, task_struct, thread.reg23); in output_thread_defines() 118 OFFSET(THREAD_REG29, task_struct, thread.reg29); in output_thread_defines() 119 OFFSET(THREAD_REG30, task_struct, thread.reg30); in output_thread_defines() [all …]
|
D | ptrace32.c | 128 tmp = child->thread.fpu.fcr31; in compat_arch_ptrace() 152 tmp = child->thread.dsp.dspcontrol; in compat_arch_ptrace() 202 memset(&child->thread.fpu, ~0, in compat_arch_ptrace() 203 sizeof(child->thread.fpu)); in compat_arch_ptrace() 204 child->thread.fpu.fcr31 = 0; in compat_arch_ptrace() 229 child->thread.fpu.fcr31 = data; in compat_arch_ptrace() 248 child->thread.dsp.dspcontrol = data; in compat_arch_ptrace()
|
D | ptrace.c | 58 memset(&target->thread.fpu.fpr, ~0, sizeof(target->thread.fpu.fpr)); in init_fp_ctx() 92 fcr31 = child->thread.fpu.fcr31; in ptrace_setfcr31() 94 child->thread.fpu.fcr31 = (value & ~mask) | (fcr31 & mask); in ptrace_setfcr31() 167 __put_user(child->thread.fpu.fcr31, data + 64); in ptrace_getfpregs() 222 __put_user(child->thread.watch.mips3264.watchlo[i], in ptrace_get_watch_regs() 224 __put_user(child->thread.watch.mips3264.watchhi[i] & 0xfff, in ptrace_get_watch_regs() 273 child->thread.watch.mips3264.watchlo[i] = lt[i]; in ptrace_set_watch_regs() 275 child->thread.watch.mips3264.watchhi[i] = ht[i]; in ptrace_set_watch_regs() 453 if (sizeof(target->thread.fpu.fpr[i]) == sizeof(elf_fpreg_t)) in fpr_get() 455 &target->thread.fpu, in fpr_get() [all …]
|
D | traps.c | 203 regs.regs[29] = task->thread.reg29; in show_stack() 205 regs.cp0_epc = task->thread.reg31; in show_stack() 383 if (notify_die(DIE_OOPS, str, regs, 0, current->thread.trap_nr, in die() 469 if (notify_die(DIE_OOPS, "bus error", regs, 0, current->thread.trap_nr, in do_be() 796 sig = fpu_emulator_cop1Handler(regs, ¤t->thread.fpu, 1, in simulate_fp() 798 fcr31 = current->thread.fpu.fcr31; in simulate_fp() 804 current->thread.fpu.fcr31 &= ~FPU_CSR_ALL_X; in simulate_fp() 825 if (notify_die(DIE_FP, "FP exception", regs, 0, current->thread.trap_nr, in do_fpe() 850 sig = fpu_emulator_cop1Handler(regs, ¤t->thread.fpu, 1, in do_fpe() 852 fcr31 = current->thread.fpu.fcr31; in do_fpe() [all …]
|
D | signal.c | 71 struct mips_abi *abi = current->thread.abi; in copy_fp_to_sigcontext() 80 __put_user(get_fpr64(¤t->thread.fpu.fpr[i], 0), in copy_fp_to_sigcontext() 83 err |= __put_user(current->thread.fpu.fcr31, csr); in copy_fp_to_sigcontext() 90 struct mips_abi *abi = current->thread.abi; in copy_fp_from_sigcontext() 100 set_fpr64(¤t->thread.fpu.fpr[i], 0, fpr_val); in copy_fp_from_sigcontext() 102 err |= __get_user(current->thread.fpu.fcr31, csr); in copy_fp_from_sigcontext() 112 struct mips_abi *abi = current->thread.abi; in save_hw_fp_context() 121 struct mips_abi *abi = current->thread.abi; in restore_hw_fp_context() 177 err = __put_user(current->thread.fpu.msacsr, &msa->csr); in save_msa_extcontext() 180 val = get_fpr64(¤t->thread.fpu.fpr[i], 1); in save_msa_extcontext() [all …]
|
D | process.c | 127 p->thread.cp0_status = read_c0_status() & ~(ST0_CU2|ST0_CU1); in copy_thread() 130 unsigned long status = p->thread.cp0_status; in copy_thread() 133 p->thread.reg16 = usp; /* fn */ in copy_thread() 134 p->thread.reg17 = kthread_arg; in copy_thread() 135 p->thread.reg29 = childksp; in copy_thread() 136 p->thread.reg31 = (unsigned long) ret_from_kernel_thread; in copy_thread() 155 p->thread.reg29 = (unsigned long) childregs; in copy_thread() 156 p->thread.reg31 = (unsigned long) ret_from_fork; in copy_thread() 422 struct thread_struct *t = &tsk->thread; in thread_saved_pc() 534 sp = task->thread.reg29 + schedule_mfi.frame_size; in get_wchan()
|
/linux-4.4.14/drivers/dma/ |
D | dmatest.c | 176 struct dmatest_thread *thread; in is_threaded_test_run() local 178 list_for_each_entry(thread, &dtc->threads, node) { in is_threaded_test_run() 179 if (!thread->done) in is_threaded_test_run() 402 struct dmatest_thread *thread = data; in dmatest_func() local 428 info = thread->info; in dmatest_func() 430 chan = thread->chan; in dmatest_func() 432 if (thread->type == DMA_MEMCPY) in dmatest_func() 434 else if (thread->type == DMA_XOR) { in dmatest_func() 438 } else if (thread->type == DMA_PQ) { in dmatest_func() 452 thread->srcs = kcalloc(src_cnt+1, sizeof(u8 *), GFP_KERNEL); in dmatest_func() [all …]
|
/linux-4.4.14/arch/mn10300/include/asm/ |
D | fpu.h | 41 : "i"(THREAD_USING_FPU), "a"(&tsk->thread.fpu_flags) in set_using_fpu() 50 : "i"(THREAD_USING_FPU), "a"(&tsk->thread.fpu_flags) in clear_using_fpu() 54 #define is_using_fpu(tsk) ((tsk)->thread.fpu_flags & THREAD_USING_FPU) 67 if (tsk->thread.fpu_flags & THREAD_HAS_FPU) { in unlazy_fpu() 68 fpu_save(&tsk->thread.fpu_state); in unlazy_fpu() 69 tsk->thread.fpu_flags &= ~THREAD_HAS_FPU; in unlazy_fpu() 70 tsk->thread.uregs->epsw &= ~EPSW_FE; in unlazy_fpu() 74 fpu_save(&tsk->thread.fpu_state); in unlazy_fpu() 97 if (tsk->thread.fpu_flags & THREAD_HAS_FPU) { in flush_fpu() 98 tsk->thread.fpu_flags &= ~THREAD_HAS_FPU; in flush_fpu() [all …]
|
D | switch_to.h | 24 if ((prev)->thread.fpu_flags & THREAD_HAS_FPU) { \ 25 (prev)->thread.fpu_flags &= ~THREAD_HAS_FPU; \ 26 (prev)->thread.uregs->epsw &= ~EPSW_FE; \ 27 fpu_save(&(prev)->thread.fpu_state); \ 43 current->thread.wchan = (u_long) __builtin_return_address(0); \ 44 (last) = __switch_to(&(prev)->thread, &(next)->thread, (prev)); \ 46 current->thread.wchan = 0; \
|
/linux-4.4.14/arch/m68k/mm/ |
D | fault.c | 25 siginfo.si_signo = current->thread.signo; in send_fault_sig() 26 siginfo.si_code = current->thread.code; in send_fault_sig() 27 siginfo.si_addr = (void *)current->thread.faddr; in send_fault_sig() 196 current->thread.signo = SIGBUS; in do_page_fault() 197 current->thread.faddr = address; in do_page_fault() 201 current->thread.signo = SIGBUS; in do_page_fault() 202 current->thread.code = BUS_ADRERR; in do_page_fault() 203 current->thread.faddr = address; in do_page_fault() 207 current->thread.signo = SIGSEGV; in do_page_fault() 208 current->thread.code = SEGV_MAPERR; in do_page_fault() [all …]
|
/linux-4.4.14/drivers/android/ |
D | binder.c | 1318 struct binder_thread *thread, in binder_transaction() argument 1336 e->from_thread = thread->pid; in binder_transaction() 1342 in_reply_to = thread->transaction_stack; in binder_transaction() 1345 proc->pid, thread->pid); in binder_transaction() 1350 if (in_reply_to->to_thread != thread) { in binder_transaction() 1352 proc->pid, thread->pid, in_reply_to->debug_id, in binder_transaction() 1361 thread->transaction_stack = in_reply_to->to_parent; in binder_transaction() 1369 proc->pid, thread->pid, in binder_transaction() 1386 proc->pid, thread->pid); in binder_transaction() 1409 if (!(tr->flags & TF_ONE_WAY) && thread->transaction_stack) { in binder_transaction() [all …]
|
/linux-4.4.14/arch/arm/kernel/ |
D | process.c | 203 struct thread_info *thread = current_thread_info(); in flush_thread() local 208 memset(thread->used_cp, 0, sizeof(thread->used_cp)); in flush_thread() 209 memset(&tsk->thread.debug, 0, sizeof(struct debug_info)); in flush_thread() 210 memset(&thread->fpstate, 0, sizeof(union fp_state)); in flush_thread() 214 thread_notify(THREAD_NOTIFY_FLUSH, thread); in flush_thread() 227 struct thread_info *thread = task_thread_info(p); in copy_thread() local 230 memset(&thread->cpu_context, 0, sizeof(struct cpu_context_save)); in copy_thread() 239 thread->cpu_domain = get_domain(); in copy_thread() 249 thread->cpu_context.r4 = stk_sz; in copy_thread() 250 thread->cpu_context.r5 = stack_start; in copy_thread() [all …]
|
D | ptrace.c | 305 struct thread_info *thread = task_thread_info(tsk); in ptrace_getwmmxregs() local 307 if (!test_ti_thread_flag(thread, TIF_USING_IWMMXT)) in ptrace_getwmmxregs() 309 iwmmxt_task_disable(thread); /* force it to ram */ in ptrace_getwmmxregs() 310 return copy_to_user(ufp, &thread->fpstate.iwmmxt, IWMMXT_SIZE) in ptrace_getwmmxregs() 319 struct thread_info *thread = task_thread_info(tsk); in ptrace_setwmmxregs() local 321 if (!test_ti_thread_flag(thread, TIF_USING_IWMMXT)) in ptrace_setwmmxregs() 323 iwmmxt_task_release(thread); /* force a reload */ in ptrace_setwmmxregs() 324 return copy_from_user(&thread->fpstate.iwmmxt, ufp, IWMMXT_SIZE) in ptrace_setwmmxregs() 336 struct thread_info *thread = task_thread_info(tsk); in ptrace_getcrunchregs() local 338 crunch_task_disable(thread); /* force it to ram */ in ptrace_getcrunchregs() [all …]
|
D | xscale-cp0.c | 35 struct thread_info *thread = t; in dsp_do() local 39 thread->cpu_context.extra[0] = 0; in dsp_do() 40 thread->cpu_context.extra[1] = 0; in dsp_do() 45 dsp_load_state(thread->cpu_context.extra); in dsp_do() 60 struct thread_info *thread = t; in iwmmxt_do() local 73 iwmmxt_task_release(thread); in iwmmxt_do() 77 iwmmxt_task_switch(thread); in iwmmxt_do()
|
D | pj4-cp0.c | 24 struct thread_info *thread = t; in iwmmxt_do() local 37 iwmmxt_task_release(thread); in iwmmxt_do() 41 iwmmxt_task_switch(thread); in iwmmxt_do()
|
/linux-4.4.14/arch/sh/kernel/ |
D | process.c | 28 if (src->thread.xstate) { in arch_dup_task_struct() 29 dst->thread.xstate = kmem_cache_alloc(task_xstate_cachep, in arch_dup_task_struct() 31 if (!dst->thread.xstate) in arch_dup_task_struct() 33 memcpy(dst->thread.xstate, src->thread.xstate, xstate_size); in arch_dup_task_struct() 41 if (tsk->thread.xstate) { in free_thread_xstate() 42 kmem_cache_free(task_xstate_cachep, tsk->thread.xstate); in free_thread_xstate() 43 tsk->thread.xstate = NULL; in free_thread_xstate()
|
D | process_32.c | 140 p->thread.dsp_status = tsk->thread.dsp_status; in copy_thread() 144 memset(p->thread.ptrace_bps, 0, sizeof(p->thread.ptrace_bps)); in copy_thread() 147 p->thread.sp = (unsigned long) childregs; in copy_thread() 150 p->thread.pc = (unsigned long) ret_from_kernel_thread; in copy_thread() 159 p->thread.fpu_counter = 0; in copy_thread() 172 p->thread.pc = (unsigned long) ret_from_fork; in copy_thread() 183 struct thread_struct *next_t = &next->thread; in __switch_to() 192 if (next->thread.fpu_counter > 5) in __switch_to() 210 if (next->thread.fpu_counter > 5) in __switch_to() 230 unsigned long schedule_frame = (unsigned long)p->thread.sp; in get_wchan()
|
D | traps_64.c | 355 current->thread.xstate->hardfpu.fp_regs[destreg] = buflo; in misaligned_fpu_load() 359 current->thread.xstate->hardfpu.fp_regs[destreg] = buflo; in misaligned_fpu_load() 360 current->thread.xstate->hardfpu.fp_regs[destreg+1] = bufhi; in misaligned_fpu_load() 363 current->thread.xstate->hardfpu.fp_regs[destreg] = bufhi; in misaligned_fpu_load() 364 current->thread.xstate->hardfpu.fp_regs[destreg+1] = buflo; in misaligned_fpu_load() 366 current->thread.xstate->hardfpu.fp_regs[destreg] = buflo; in misaligned_fpu_load() 367 current->thread.xstate->hardfpu.fp_regs[destreg+1] = bufhi; in misaligned_fpu_load() 422 buflo = current->thread.xstate->hardfpu.fp_regs[srcreg]; in misaligned_fpu_store() 426 buflo = current->thread.xstate->hardfpu.fp_regs[srcreg]; in misaligned_fpu_store() 427 bufhi = current->thread.xstate->hardfpu.fp_regs[srcreg+1]; in misaligned_fpu_store() [all …]
|
D | ptrace_32.c | 81 struct thread_struct *thread = &tsk->thread; in set_single_step() local 85 bp = thread->ptrace_bps[0]; in set_single_step() 98 thread->ptrace_bps[0] = bp; in set_single_step() 199 &target->thread.xstate->hardfpu, 0, -1); in fpregs_get() 202 &target->thread.xstate->softfpu, 0, -1); in fpregs_get() 220 &target->thread.xstate->hardfpu, 0, -1); in fpregs_set() 223 &target->thread.xstate->softfpu, 0, -1); in fpregs_set() 240 (struct pt_dspregs *)&target->thread.dsp_status.dsp_regs; in dspregs_get() 258 (struct pt_dspregs *)&target->thread.dsp_status.dsp_regs; in dspregs_set() 401 tmp = ((unsigned long *)child->thread.xstate) in arch_ptrace() [all …]
|
D | process_64.c | 282 tsk->thread.kregs = regs; in show_regs() 332 if(current->thread.kregs==&fake_swapper_regs) { in flush_thread() 333 current->thread.kregs = in flush_thread() 335 current->thread.uregs = current->thread.kregs; in flush_thread() 361 memcpy(fpu, &tsk->thread.xstate->hardfpu, sizeof(*fpu)); in dump_fpu() 391 p->thread.sp = (unsigned long) childregs; in copy_thread() 399 p->thread.pc = (unsigned long) ret_from_kernel_thread; in copy_thread() 412 p->thread.uregs = childregs; in copy_thread() 417 p->thread.pc = (unsigned long) ret_from_fork; in copy_thread() 451 sh64_switch_to_fp = (long) p->thread.sp; in get_wchan()
|
/linux-4.4.14/arch/unicore32/kernel/ |
D | process.c | 213 struct thread_info *thread = current_thread_info(); in flush_thread() local 216 memset(thread->used_cp, 0, sizeof(thread->used_cp)); in flush_thread() 217 memset(&tsk->thread.debug, 0, sizeof(struct debug_info)); in flush_thread() 219 memset(&thread->fpstate, 0, sizeof(struct fp_state)); in flush_thread() 234 struct thread_info *thread = task_thread_info(p); in copy_thread() local 237 memset(&thread->cpu_context, 0, sizeof(struct cpu_context_save)); in copy_thread() 238 thread->cpu_context.sp = (unsigned long)childregs; in copy_thread() 240 thread->cpu_context.pc = (unsigned long)ret_from_kernel_thread; in copy_thread() 241 thread->cpu_context.r4 = stack_start; in copy_thread() 242 thread->cpu_context.r5 = stk_sz; in copy_thread() [all …]
|
D | traps.c | 179 static int __die(const char *str, int err, struct thread_info *thread, in __die() argument 182 struct task_struct *tsk = thread->task; in __die() 190 ret = notify_die(DIE_OOPS, str, regs, err, tsk->thread.trap_no, \ in __die() 198 TASK_COMM_LEN, tsk->comm, task_pid_nr(tsk), thread + 1); in __die() 217 struct thread_info *thread = current_thread_info(); in die() local 225 ret = __die(str, err, thread, regs); in die() 244 current->thread.error_code = err; in uc32_notify_die() 245 current->thread.trap_no = trap; in uc32_notify_die()
|
/linux-4.4.14/arch/powerpc/include/asm/ |
D | switch_to.h | 90 t->thread.ebbrr = 0; in clear_task_ebb() 91 t->thread.ebbhr = 0; in clear_task_ebb() 92 t->thread.bescr = 0; in clear_task_ebb() 93 t->thread.mmcr2 = 0; in clear_task_ebb() 94 t->thread.mmcr0 = 0; in clear_task_ebb() 95 t->thread.siar = 0; in clear_task_ebb() 96 t->thread.sdar = 0; in clear_task_ebb() 97 t->thread.sier = 0; in clear_task_ebb() 98 t->thread.used_ebb = 0; in clear_task_ebb()
|
D | tm.h | 13 extern void do_load_up_transact_fpu(struct thread_struct *thread); 14 extern void do_load_up_transact_altivec(struct thread_struct *thread); 18 extern void tm_reclaim(struct thread_struct *thread, 21 extern void tm_recheckpoint(struct thread_struct *thread, 24 extern void tm_save_sprs(struct thread_struct *thread); 25 extern void tm_restore_sprs(struct thread_struct *thread);
|
D | processor.h | 360 ((tsk)->thread.regs? (tsk)->thread.regs->nip: 0) 362 #define task_pt_regs(tsk) ((struct pt_regs *)(tsk)->thread.regs) 366 #define KSTK_EIP(tsk) ((tsk)->thread.regs? (tsk)->thread.regs->nip: 0) 367 #define KSTK_ESP(tsk) ((tsk)->thread.regs? (tsk)->thread.regs->gpr[1]: 0)
|
/linux-4.4.14/tools/testing/selftests/powerpc/dscr/ |
D | dscr_default_test.c | 23 unsigned long thread = (unsigned long)in; in do_test() local 47 "but is %ld\n", thread, d, cur_dscr); in do_test() 48 result[thread] = 1; in do_test() 49 pthread_exit(&result[thread]); in do_test() 54 "but is %ld\n", thread, d, cur_dscr_usr); in do_test() 55 result[thread] = 1; in do_test() 56 pthread_exit(&result[thread]); in do_test() 59 result[thread] = 0; in do_test() 60 pthread_exit(&result[thread]); in do_test()
|
/linux-4.4.14/drivers/gpu/drm/amd/amdkfd/ |
D | kfd_process.c | 60 static struct kfd_process *find_process(const struct task_struct *thread); 61 static struct kfd_process *create_process(const struct task_struct *thread); 78 struct kfd_process *kfd_create_process(const struct task_struct *thread) in kfd_create_process() argument 84 if (thread->mm == NULL) in kfd_create_process() 88 if (thread->group_leader->mm != thread->mm) in kfd_create_process() 92 down_write(&thread->mm->mmap_sem); in kfd_create_process() 102 process = find_process(thread); in kfd_create_process() 107 process = create_process(thread); in kfd_create_process() 111 up_write(&thread->mm->mmap_sem); in kfd_create_process() 116 struct kfd_process *kfd_get_process(const struct task_struct *thread) in kfd_get_process() argument [all …]
|
/linux-4.4.14/arch/sh/kernel/cpu/sh2a/ |
D | fpu.c | 55 : "0" ((char *)(&tsk->thread.xstate->hardfpu.status)), in save_fpu() 87 : "0" (tsk->thread.xstate), "r" (FPSCR_RCHG) in restore_fpu() 459 if ((tsk->thread.xstate->hardfpu.fpscr & FPSCR_FPU_ERROR)) { in ieee_fpe_handler() 461 denormal_to_double (&tsk->thread.xstate->hardfpu, in ieee_fpe_handler() 476 hx = tsk->thread.xstate->hardfpu.fp_regs[n]; in ieee_fpe_handler() 477 hy = tsk->thread.xstate->hardfpu.fp_regs[m]; in ieee_fpe_handler() 478 fpscr = tsk->thread.xstate->hardfpu.fpscr; in ieee_fpe_handler() 488 | tsk->thread.xstate->hardfpu.fp_regs[n+1]; in ieee_fpe_handler() 490 | tsk->thread.xstate->hardfpu.fp_regs[m+1]; in ieee_fpe_handler() 495 tsk->thread.xstate->hardfpu.fp_regs[n] = llx >> 32; in ieee_fpe_handler() [all …]
|
/linux-4.4.14/arch/sh/include/asm/ |
D | switch_to_32.h | 7 (!!(tsk->thread.dsp_status.status & SR_DSP)) 12 (u32 *)&tsk->thread.dsp_status; \ 35 (u32 *)&tsk->thread.dsp_status + 14; \ 84 __ts1 = (u32 *)&prev->thread.sp; \ 85 __ts2 = (u32 *)&prev->thread.pc; \ 88 __ts6 = (u32 *)&next->thread.sp; \ 89 __ts7 = next->thread.pc; \
|
D | switch_to_64.h | 27 struct pt_regs *regs = next->thread.uregs; \ 30 last = sh64_switch_to(prev, &prev->thread, next, \ 31 &next->thread); \
|
/linux-4.4.14/arch/powerpc/math-emu/ |
D | math.c | 331 op0 = (void *)¤t->thread.TS_FPR((insn >> 21) & 0x1f); in do_mathemu() 332 op1 = (void *)¤t->thread.TS_FPR((insn >> 16) & 0x1f); in do_mathemu() 333 op2 = (void *)¤t->thread.TS_FPR((insn >> 11) & 0x1f); in do_mathemu() 337 op0 = (void *)¤t->thread.TS_FPR((insn >> 21) & 0x1f); in do_mathemu() 338 op1 = (void *)¤t->thread.TS_FPR((insn >> 16) & 0x1f); in do_mathemu() 339 op2 = (void *)¤t->thread.TS_FPR((insn >> 6) & 0x1f); in do_mathemu() 343 op0 = (void *)¤t->thread.TS_FPR((insn >> 21) & 0x1f); in do_mathemu() 344 op1 = (void *)¤t->thread.TS_FPR((insn >> 16) & 0x1f); in do_mathemu() 345 op2 = (void *)¤t->thread.TS_FPR((insn >> 11) & 0x1f); in do_mathemu() 346 op3 = (void *)¤t->thread.TS_FPR((insn >> 6) & 0x1f); in do_mathemu() [all …]
|
D | math_efp.c | 203 vc.wp[0] = current->thread.evr[fc]; in do_spe_mathemu() 205 va.wp[0] = current->thread.evr[fa]; in do_spe_mathemu() 207 vb.wp[0] = current->thread.evr[fb]; in do_spe_mathemu() 680 &= ~(FP_EX_INVALID | FP_EX_UNDERFLOW) | current->thread.spefscr_last; in do_spe_mathemu() 683 current->thread.spefscr_last = __FPU_FPSCR; in do_spe_mathemu() 685 current->thread.evr[fc] = vc.wp[0]; in do_spe_mathemu() 695 if (current->thread.fpexc_mode & PR_FP_EXC_SW_ENABLE) { in do_spe_mathemu() 697 && (current->thread.fpexc_mode & PR_FP_EXC_DIV)) in do_spe_mathemu() 700 && (current->thread.fpexc_mode & PR_FP_EXC_OVF)) in do_spe_mathemu() 703 && (current->thread.fpexc_mode & PR_FP_EXC_UND)) in do_spe_mathemu() [all …]
|
/linux-4.4.14/drivers/media/pci/saa7134/ |
D | saa7134-tvaudio.c | 316 if (dev->thread.scan1 == dev->thread.scan2 && in tvaudio_sleep() 326 return dev->thread.scan1 != dev->thread.scan2; in tvaudio_sleep() 490 dev->thread.scan1 = dev->thread.scan2; in tvaudio_thread() 492 dev->thread.scan1); in tvaudio_thread() 531 if (dev->thread.scan1 != dev->thread.scan2) in tvaudio_thread() 605 if (UNSET == dev->thread.mode) { in tvaudio_thread() 609 mode = dev->thread.mode; in tvaudio_thread() 619 dev->thread.stopped = 1; in tvaudio_thread() 785 dev->thread.scan1 = dev->thread.scan2; in tvaudio_thread_ddep() 787 dev->thread.scan1); in tvaudio_thread_ddep() [all …]
|
/linux-4.4.14/arch/c6x/kernel/ |
D | process.c | 107 current->thread.usp = usp; in start_thread() 125 p->thread.pc = (unsigned long) ret_from_kernel_thread; in copy_thread() 133 p->thread.pc = (unsigned long) ret_from_fork; in copy_thread() 137 p->thread.usp = childregs->sp; in copy_thread() 139 p->thread.wchan = p->thread.pc; in copy_thread() 156 return p->thread.wchan; in get_wchan()
|
/linux-4.4.14/drivers/staging/lustre/lustre/ptlrpc/ |
D | sec_gc.c | 149 struct ptlrpc_thread *thread = arg; in sec_gc_main() local 155 thread_set_flags(thread, SVC_RUNNING); in sec_gc_main() 156 wake_up(&thread->t_ctl_waitq); in sec_gc_main() 161 thread_clear_flags(thread, SVC_SIGNAL); in sec_gc_main() 188 l_wait_event(thread->t_ctl_waitq, in sec_gc_main() 189 thread_is_stopping(thread) || in sec_gc_main() 190 thread_is_signal(thread), in sec_gc_main() 193 if (thread_test_and_clear_flags(thread, SVC_STOPPING)) in sec_gc_main() 197 thread_set_flags(thread, SVC_STOPPED); in sec_gc_main() 198 wake_up(&thread->t_ctl_waitq); in sec_gc_main()
|
D | service.c | 1446 struct ptlrpc_thread *thread) in ptlrpc_server_handle_req_in() argument 1563 req->rq_svc_thread = thread; in ptlrpc_server_handle_req_in() 1587 struct ptlrpc_thread *thread) in ptlrpc_server_handle_request() argument 1639 request->rq_session.lc_thread = thread; in ptlrpc_server_handle_request() 1645 request->rq_svc_thread = thread; in ptlrpc_server_handle_request() 1646 if (thread) in ptlrpc_server_handle_request() 1913 ptlrpc_thread_stopping(struct ptlrpc_thread *thread) in ptlrpc_thread_stopping() argument 1915 return thread_is_stopping(thread) || in ptlrpc_thread_stopping() 1916 thread->t_svcpt->scp_service->srv_is_stopping; in ptlrpc_thread_stopping() 1945 struct ptlrpc_thread *thread) in ptlrpc_wait_event() argument [all …]
|
D | pinger.c | 224 struct ptlrpc_thread *thread = arg; in ptlrpc_pinger_main() local 227 thread_set_flags(thread, SVC_RUNNING); in ptlrpc_pinger_main() 228 wake_up(&thread->t_ctl_waitq); in ptlrpc_pinger_main() 273 l_wait_event(thread->t_ctl_waitq, in ptlrpc_pinger_main() 274 thread_is_stopping(thread) || in ptlrpc_pinger_main() 275 thread_is_event(thread), in ptlrpc_pinger_main() 277 if (thread_test_and_clear_flags(thread, SVC_STOPPING)) in ptlrpc_pinger_main() 280 thread_test_and_clear_flags(thread, SVC_EVENT); in ptlrpc_pinger_main() 284 thread_set_flags(thread, SVC_STOPPED); in ptlrpc_pinger_main() 285 wake_up(&thread->t_ctl_waitq); in ptlrpc_pinger_main()
|
/linux-4.4.14/arch/sparc/include/asm/ |
D | switch_to_32.h | 20 fpsave(&(prv)->thread.float_regs[0], &(prv)->thread.fsr, \ 21 &(prv)->thread.fpqueue[0], &(prv)->thread.fpqdepth); \ 23 (prv)->thread.kregs->psr &= ~PSR_EF; \ 33 (nxt)->thread.kregs->psr&=~PSR_EF; \
|
D | sfp-machine_32.h | 185 #define FP_ROUNDMODE ((current->thread.fsr >> 30) & 0x3) 187 #define FP_ROUNDMODE ((last_task_used_math->thread.fsr >> 30) & 0x3) 201 #define FP_INHIBIT_RESULTS ((current->thread.fsr >> 23) & _fex) 203 #define FP_INHIBIT_RESULTS ((last_task_used_math->thread.fsr >> 23) & _fex) 207 #define FP_TRAPPING_EXCEPTIONS ((current->thread.fsr >> 23) & 0x1f) 209 #define FP_TRAPPING_EXCEPTIONS ((last_task_used_math->thread.fsr >> 23) & 0x1f)
|
D | processor_32.h | 112 #define task_pt_regs(tsk) ((tsk)->thread.kregs) 113 #define KSTK_EIP(tsk) ((tsk)->thread.kregs->pc) 114 #define KSTK_ESP(tsk) ((tsk)->thread.kregs->u_regs[UREG_FP])
|
/linux-4.4.14/tools/perf/arch/x86/tests/ |
D | dwarf-unwind.c | 13 struct thread *thread, u64 *regs) in sample_ustack() argument 28 map = map_groups__find(thread->mg, MAP__VARIABLE, (u64) sp); in sample_ustack() 45 struct thread *thread) in test__arch_unwind_sample() argument 61 return sample_ustack(sample, thread, buf); in test__arch_unwind_sample()
|
/linux-4.4.14/tools/perf/arch/arm64/tests/ |
D | dwarf-unwind.c | 12 struct thread *thread, u64 *regs) in sample_ustack() argument 27 map = map_groups__find(thread->mg, MAP__VARIABLE, (u64) sp); in sample_ustack() 44 struct thread *thread) in test__arch_unwind_sample() argument 60 return sample_ustack(sample, thread, buf); in test__arch_unwind_sample()
|
/linux-4.4.14/tools/perf/arch/arm/tests/ |
D | dwarf-unwind.c | 12 struct thread *thread, u64 *regs) in sample_ustack() argument 27 map = map_groups__find(thread->mg, MAP__VARIABLE, (u64) sp); in sample_ustack() 44 struct thread *thread) in test__arch_unwind_sample() argument 60 return sample_ustack(sample, thread, buf); in test__arch_unwind_sample()
|
/linux-4.4.14/arch/sh/kernel/cpu/ |
D | fpu.c | 18 if (!tsk->thread.xstate) { in init_fpu() 19 tsk->thread.xstate = kmem_cache_alloc(task_xstate_cachep, in init_fpu() 21 if (!tsk->thread.xstate) in init_fpu() 26 struct sh_fpu_hard_struct *fp = &tsk->thread.xstate->hardfpu; in init_fpu() 30 struct sh_fpu_soft_struct *fp = &tsk->thread.xstate->softfpu; in init_fpu() 47 tsk->thread.fpu_counter++; in __fpu_state_restore()
|
/linux-4.4.14/arch/x86/um/ |
D | syscalls_64.c | 35 ret = restore_registers(pid, ¤t->thread.regs.regs); in arch_prctl() 58 current->thread.arch.fs = (unsigned long) ptr; in arch_prctl() 59 ret = save_registers(pid, ¤t->thread.regs.regs); in arch_prctl() 62 ret = save_registers(pid, ¤t->thread.regs.regs); in arch_prctl() 82 if ((to->thread.arch.fs == 0) || (to->mm == NULL)) in arch_switch_to() 85 arch_prctl(to, ARCH_SET_FS, (void __user *) to->thread.arch.fs); in arch_switch_to()
|
D | ptrace_32.c | 109 child->thread.regs.regs.gp[HOST_EFLAGS] |= value; in putreg() 112 child->thread.regs.regs.syscall = value; in putreg() 117 child->thread.regs.regs.gp[reg_offsets[regno]] = value; in putreg() 134 child->thread.arch.debugregs[addr] = data; in poke_user() 147 return child->thread.regs.regs.syscall; in getreg() 170 return mask & child->thread.regs.regs.gp[reg_offsets[regno]]; in getreg() 189 tmp = child->thread.arch.debugregs[addr]; in peek_user()
|
D | tls_32.c | 65 struct thread_struct *t = &task->thread; in get_free_idx() 99 &to->thread.arch.tls_array[idx - GDT_ENTRY_TLS_MIN]; in load_TLS() 139 &task->thread.arch.tls_array[i - GDT_ENTRY_TLS_MIN]; in needs_TLS_update() 163 &task->thread.arch.tls_array[i - GDT_ENTRY_TLS_MIN]; in clear_flushed_tls() 206 struct thread_struct *t = &task->thread; in set_tls_entry() 224 (void __user *) UPT_SI(&new->thread.regs.regs), in arch_copy_tls() 243 struct thread_struct *t = &task->thread; in get_tls_entry()
|
D | tls_64.c | 15 t->thread.arch.fs = t->thread.regs.regs.gp[R8 / sizeof(long)]; in arch_copy_tls()
|
/linux-4.4.14/arch/um/kernel/skas/ |
D | process.c | 40 init_task.thread.request.u.thread.proc = start_kernel_proc; in start_uml() 41 init_task.thread.request.u.thread.arg = NULL; in start_uml() 43 &init_task.thread.switch_buf); in start_uml()
|
/linux-4.4.14/arch/s390/include/asm/ |
D | switch_to.h | 34 save_access_regs(&prev->thread.acrs[0]); \ 35 save_ri_cb(prev->thread.ri_cb); \ 40 restore_access_regs(&next->thread.acrs[0]); \ 41 restore_ri_cb(next->thread.ri_cb, prev->thread.ri_cb); \
|
/linux-4.4.14/arch/arm64/kernel/ |
D | fpsimd.c | 136 fpsimd_save_state(¤t->thread.fpsimd_state); in fpsimd_thread_switch() 146 struct fpsimd_state *st = &next->thread.fpsimd_state; in fpsimd_thread_switch() 160 memset(¤t->thread.fpsimd_state, 0, sizeof(struct fpsimd_state)); in fpsimd_flush_thread() 173 fpsimd_save_state(¤t->thread.fpsimd_state); in fpsimd_preserve_current_state() 186 struct fpsimd_state *st = ¤t->thread.fpsimd_state; in fpsimd_restore_current_state() 205 struct fpsimd_state *st = ¤t->thread.fpsimd_state; in fpsimd_update_current_state() 218 t->thread.fpsimd_state.cpu = NR_CPUS; in fpsimd_flush_task_state() 247 fpsimd_save_state(¤t->thread.fpsimd_state); in kernel_neon_begin_partial() 274 fpsimd_save_state(¤t->thread.fpsimd_state); in fpsimd_cpu_pm_notifier()
|
D | process.c | 214 current->thread.tp_value = 0; in tls_thread_flush() 252 memset(&p->thread.cpu_context, 0, sizeof(struct cpu_context)); in copy_thread() 279 p->thread.tp_value = childregs->regs[3]; in copy_thread() 283 p->thread.cpu_context.x19 = stack_start; in copy_thread() 284 p->thread.cpu_context.x20 = stk_sz; in copy_thread() 286 p->thread.cpu_context.pc = (unsigned long)ret_from_fork; in copy_thread() 287 p->thread.cpu_context.sp = (unsigned long)childregs; in copy_thread() 303 next->thread.tp_value : 0; in tls_thread_switch()
|
/linux-4.4.14/tools/perf/ |
D | builtin-script.c | 372 struct thread *thread __maybe_unused, in print_sample_iregs() 389 struct thread *thread, in print_sample_start() argument 399 printf("%8.8s ", thread__comm_str(thread)); in print_sample_start() 401 printf("%s ", thread__comm_str(thread)); in print_sample_start() 403 printf("%16s ", thread__comm_str(thread)); in print_sample_start() 443 struct thread *thread __maybe_unused, in print_sample_brstack() 465 struct thread *thread __maybe_unused, in print_sample_brstacksym() 483 thread__find_addr_map(thread, cpumode, MAP__FUNCTION, from, &alf); in print_sample_brstacksym() 487 thread__find_addr_map(thread, cpumode, MAP__FUNCTION, to, &alt); in print_sample_brstacksym() 505 struct thread *thread, in print_sample_addr() argument [all …]
|
D | builtin-trace.c | 275 struct thread *thread; member 1372 static struct thread_trace *thread__trace(struct thread *thread, FILE *fp) in thread__trace() argument 1376 if (thread == NULL) in thread__trace() 1379 if (thread__priv(thread) == NULL) in thread__trace() 1380 thread__set_priv(thread, thread_trace__new()); in thread__trace() 1382 if (thread__priv(thread) == NULL) in thread__trace() 1385 ttrace = thread__priv(thread); in thread__trace() 1417 struct thread *current; 1453 static int trace__set_fd_pathname(struct thread *thread, int fd, const char *pathname) in trace__set_fd_pathname() argument 1455 struct thread_trace *ttrace = thread__priv(thread); in trace__set_fd_pathname() [all …]
|
D | builtin-sched.c | 45 pthread_t thread; member 91 struct thread *thread; member 141 struct thread *curr_thread[MAX_CPUS]; 572 err = pthread_create(&task->thread, &attr, thread_func, parms); in create_tasks() 765 struct thread *child, *parent; in replay_fork_event() 816 thread_atoms_search(struct rb_root *root, struct thread *thread, in thread_atoms_search() argument 820 struct work_atoms key = { .thread = thread }; in thread_atoms_search() 834 BUG_ON(thread != atoms->thread); in thread_atoms_search() 866 static int thread_atoms_insert(struct perf_sched *sched, struct thread *thread) in thread_atoms_insert() argument 874 atoms->thread = thread__get(thread); in thread_atoms_insert() [all …]
|
/linux-4.4.14/arch/arm/mach-ep93xx/ |
D | crunch.c | 26 void crunch_task_release(struct thread_info *thread) in crunch_task_release() argument 29 if (crunch_owner == &thread->crunchstate) in crunch_task_release() 41 struct thread_info *thread = (struct thread_info *)t; in crunch_do() local 45 crunch_state = &thread->crunchstate; in crunch_do() 57 crunch_task_release(thread); in crunch_do()
|
/linux-4.4.14/drivers/staging/lustre/lustre/llite/ |
D | statahead.c | 956 struct ptlrpc_thread *thread = &sai->sai_agl_thread; in ll_agl_thread() local 959 thread->t_pid = current_pid(); in ll_agl_thread() 966 if (thread_is_init(thread)) in ll_agl_thread() 970 thread_set_flags(thread, SVC_RUNNING); in ll_agl_thread() 972 wake_up(&thread->t_ctl_waitq); in ll_agl_thread() 975 l_wait_event(thread->t_ctl_waitq, in ll_agl_thread() 977 !thread_is_running(thread), in ll_agl_thread() 980 if (!thread_is_running(thread)) in ll_agl_thread() 1006 thread_set_flags(thread, SVC_STOPPED); in ll_agl_thread() 1008 wake_up(&thread->t_ctl_waitq); in ll_agl_thread() [all …]
|
/linux-4.4.14/arch/h8300/kernel/ |
D | ptrace.c | 43 return task->thread.usp + sizeof(long)*2; in h8300_get_reg() 46 return *(unsigned short *)(task->thread.esp0 + in h8300_get_reg() 49 return *(unsigned long *)(task->thread.esp0 + in h8300_get_reg() 61 task->thread.usp = data - sizeof(long)*2; in h8300_put_reg() 63 oldccr = *(unsigned short *)(task->thread.esp0 + in h8300_put_reg() 68 *(unsigned short *)(task->thread.esp0 + in h8300_put_reg() 72 oldexr = *(unsigned short *)(task->thread.esp0 + in h8300_put_reg() 77 *(unsigned short *)(task->thread.esp0 + in h8300_put_reg() 81 *(unsigned long *)(task->thread.esp0 + in h8300_put_reg()
|
D | ptrace_h.c | 20 if ((long)child->thread.breakinfo.addr != -1L) { in user_disable_single_step() 21 *(child->thread.breakinfo.addr) = child->thread.breakinfo.inst; in user_disable_single_step() 22 child->thread.breakinfo.addr = (unsigned short *)-1L; in user_disable_single_step() 244 child->thread.breakinfo.addr = next; in user_enable_single_step() 245 child->thread.breakinfo.inst = *next; in user_enable_single_step() 251 if ((unsigned long)current->thread.breakinfo.addr == bp) { in trace_trap()
|
D | process.c | 122 p->thread.usp = usp ?: rdusp(); in copy_thread() 124 p->thread.ksp = (unsigned long)childregs; in copy_thread() 131 return ((struct pt_regs *)tsk->thread.esp0)->pc; in thread_saved_pc() 144 fp = ((struct pt_regs *)p->thread.ksp)->er6; in get_wchan()
|
/linux-4.4.14/arch/s390/mm/ |
D | fault.c | 116 return current->thread.mm_segment.ar4; in user_space_fault() 419 current->thread.gmap_addr = address; in do_exception() 487 current->thread.gmap_pfault = 1; in do_exception() 503 address = __gmap_link(gmap, current->thread.gmap_addr, in do_exception() 663 if (tsk->thread.pfault_wait == 1) { in pfault_interrupt() 669 tsk->thread.pfault_wait = 0; in pfault_interrupt() 670 list_del(&tsk->thread.list); in pfault_interrupt() 682 tsk->thread.pfault_wait = -1; in pfault_interrupt() 688 if (tsk->thread.pfault_wait == 1) { in pfault_interrupt() 692 } else if (tsk->thread.pfault_wait == -1) { in pfault_interrupt() [all …]
|
/linux-4.4.14/tools/usb/ |
D | ffs-test.c | 286 struct thread; 288 static ssize_t read_wrap(struct thread *t, void *buf, size_t nbytes); 289 static ssize_t write_wrap(struct thread *t, const void *buf, size_t nbytes); 290 static ssize_t ep0_consume(struct thread *t, const void *buf, size_t nbytes); 291 static ssize_t fill_in_buf(struct thread *t, void *buf, size_t nbytes); 292 static ssize_t empty_out_buf(struct thread *t, const void *buf, size_t nbytes); 295 static struct thread { struct 299 ssize_t (*in)(struct thread *, void *, size_t); argument 302 ssize_t (*out)(struct thread *, const void *, size_t); argument 331 static void init_thread(struct thread *t) in init_thread() argument [all …]
|
/linux-4.4.14/arch/sparc/math-emu/ |
D | math_32.c | 171 printk("fpqdepth is %ld\n", fpt->thread.fpqdepth); in do_mathemu() 172 for (i = 0; i < fpt->thread.fpqdepth; i++) in do_mathemu() 173 printk("%d: %08lx at %08lx\n", i, fpt->thread.fpqueue[i].insn, in do_mathemu() 174 (unsigned long)fpt->thread.fpqueue[i].insn_addr); in do_mathemu() 177 if (fpt->thread.fpqdepth == 0) { /* no queue, guilty insn is at regs->pc */ in do_mathemu() 182 retcode = do_one_mathemu(insn, &fpt->thread.fsr, fpt->thread.float_regs); in do_mathemu() 193 for (i = 0; i < fpt->thread.fpqdepth; i++) { in do_mathemu() 194 retcode = do_one_mathemu(fpt->thread.fpqueue[i].insn, &(fpt->thread.fsr), fpt->thread.float_regs); in do_mathemu() 200 fpt->thread.fsr &= ~(0x3000 | FSR_CEXC_MASK); in do_mathemu() 202 fpt->thread.fsr &= ~0x3000; in do_mathemu() [all …]
|
/linux-4.4.14/drivers/base/ |
D | devtmpfs.c | 29 static struct task_struct *thread; variable 87 if (!thread) in devtmpfs_create_node() 113 wake_up_process(thread); in devtmpfs_create_node() 126 if (!thread) in devtmpfs_delete_node() 143 wake_up_process(thread); in devtmpfs_delete_node() 163 d_inode(dentry)->i_private = &thread; in dev_mkdir() 223 d_inode(dentry)->i_private = &thread; in handle_create() 239 if (d_inode(dentry)->i_private == &thread) in dev_rmdir() 280 if (inode->i_private != &thread) in dev_mynode() 354 if (!thread) in devtmpfs_mount() [all …]
|
/linux-4.4.14/arch/m32r/include/asm/ |
D | switch_to.h | 45 "r" (&(prev->thread.sp)), "r" (&(next->thread.sp)), \ 46 "r" (&(prev->thread.lr)), "r" (&(next->thread.lr)) \
|
/linux-4.4.14/arch/c6x/include/asm/ |
D | switch_to.h | 26 current->thread.wchan = (u_long) __builtin_return_address(0); \ 27 (last) = __switch_to(&(prev)->thread, \ 28 &(next)->thread, (prev)); \ 30 current->thread.wchan = 0; \
|
D | processor.h | 108 (*(unsigned long *)&(tsk)->thread.b15_14) 110 (*(((unsigned long *)&(tsk)->thread.b15_14) + 1)) 113 (*(((unsigned long *)&(tsk)->thread.b15_14) + 1)) 115 (*(unsigned long *)&(tsk)->thread.b15_14)
|
/linux-4.4.14/arch/avr32/kernel/ |
D | process.c | 129 fp = tsk->thread.cpu_context.r7; in show_trace_log_lvl() 186 sp = tsk->thread.cpu_context.ksp; in show_stack_log_lvl() 290 p->thread.cpu_context.r0 = arg; in copy_thread() 291 p->thread.cpu_context.r1 = usp; /* fn */ in copy_thread() 292 p->thread.cpu_context.r2 = (unsigned long)syscall_return; in copy_thread() 293 p->thread.cpu_context.pc = (unsigned long)ret_from_kernel_thread; in copy_thread() 300 p->thread.cpu_context.pc = (unsigned long)ret_from_fork; in copy_thread() 303 p->thread.cpu_context.sr = MODE_SUPERVISOR | SR_GM; in copy_thread() 304 p->thread.cpu_context.ksp = (unsigned long)childregs; in copy_thread() 335 unsigned long fp = p->thread.cpu_context.r7; in get_wchan() [all …]
|
/linux-4.4.14/arch/parisc/include/asm/ |
D | processor.h | 35 #define TASK_SIZE_OF(tsk) ((tsk)->thread.task_size) 37 #define TASK_UNMAPPED_BASE (current->thread.map_base) 133 #define task_pt_regs(tsk) ((struct pt_regs *)&((tsk)->thread.regs)) 145 (task)->thread.flags = (((task)->thread.flags & ~PARISC_UAC_MASK) \ 153 put_user(((task)->thread.flags & PARISC_UAC_MASK) \ 308 #define KSTK_EIP(tsk) ((tsk)->thread.regs.iaoq[0]) 309 #define KSTK_ESP(tsk) ((tsk)->thread.regs.gr[30])
|
/linux-4.4.14/drivers/net/wireless/rsi/ |
D | rsi_common.h | 61 struct rsi_thread *thread, in rsi_create_kthread() argument 65 init_completion(&thread->completion); in rsi_create_kthread() 66 thread->task = kthread_run(func_ptr, common, "%s", name); in rsi_create_kthread() 67 if (IS_ERR(thread->task)) in rsi_create_kthread() 68 return (int)PTR_ERR(thread->task); in rsi_create_kthread()
|
/linux-4.4.14/arch/x86/kernel/fpu/ |
D | regset.c | 15 struct fpu *target_fpu = &target->thread.fpu; in regset_fpregs_active() 22 struct fpu *target_fpu = &target->thread.fpu; in regset_xregset_fpregs_active() 31 struct fpu *fpu = &target->thread.fpu; in xfpregs_get() 47 struct fpu *fpu = &target->thread.fpu; in xfpregs_set() 78 struct fpu *fpu = &target->thread.fpu; in xstateregs_get() 107 struct fpu *fpu = &target->thread.fpu; in xstateregs_set() 207 struct fxregs_state *fxsave = &tsk->thread.fpu.state.fxsave; in convert_from_fxsr() 227 env->fos = tsk->thread.ds; in convert_from_fxsr() 245 struct fxregs_state *fxsave = &tsk->thread.fpu.state.fxsave; in convert_to_fxsr() 273 struct fpu *fpu = &target->thread.fpu; in fpregs_get() [all …]
|
D | core.c | 76 return !current->thread.fpu.fpregs_active && (read_cr0() & X86_CR0_TS); in interrupted_kernel_fpu_idle() 110 struct fpu *fpu = ¤t->thread.fpu; in __kernel_fpu_begin() 127 struct fpu *fpu = ¤t->thread.fpu; in __kernel_fpu_end() 188 WARN_ON_FPU(fpu != ¤t->thread.fpu); in fpu__save() 234 WARN_ON_FPU(src_fpu != ¤t->thread.fpu); in fpu_copy() 285 WARN_ON_FPU(fpu != ¤t->thread.fpu); in fpu__activate_curr() 341 WARN_ON_FPU(fpu == ¤t->thread.fpu); in fpu__activate_fpstate_write() 424 WARN_ON_FPU(fpu != ¤t->thread.fpu); /* Almost certainly an anomaly */ in fpu__clear()
|
D | signal.c | 57 struct xregs_state *xsave = &tsk->thread.fpu.state.xsave; in save_fsave_header() 155 struct xregs_state *xsave = ¤t->thread.fpu.state.xsave; in copy_fpstate_to_sigframe() 176 copy_fxregs_to_kernel(&tsk->thread.fpu); in copy_fpstate_to_sigframe() 178 fpstate_sanitize_xstate(&tsk->thread.fpu); in copy_fpstate_to_sigframe() 198 struct xregs_state *xsave = &tsk->thread.fpu.state.xsave; in sanitize_restored_xstate() 252 struct fpu *fpu = &tsk->thread.fpu; in __fpu__restore_sig() 297 struct fpu *fpu = &tsk->thread.fpu; in __fpu__restore_sig()
|
/linux-4.4.14/arch/tile/kernel/ |
D | process.c | 124 p->thread.ksp = ksp; in copy_thread() 127 p->thread.creator_pid = current->pid; in copy_thread() 136 p->thread.pc = (unsigned long) ret_from_kernel_thread; in copy_thread() 144 p->thread.pc = (unsigned long) ret_from_fork; in copy_thread() 172 p->thread.usp0 = childregs->sp; in copy_thread() 187 memset(&p->thread.tile_dma_state, 0, sizeof(struct tile_dma_state)); in copy_thread() 188 memset(&p->thread.dma_async_tlb, 0, sizeof(struct async_tlb)); in copy_thread() 192 p->thread.proc_status = 0; in copy_thread() 196 memset(&p->thread.hardwall[0], 0, in copy_thread() 205 save_arch_state(&p->thread); in copy_thread() [all …]
|
/linux-4.4.14/Documentation/ |
D | robust-futex-ABI.txt | 9 The interesting data as to what futexes a thread is holding is kept on a 15 1) a one time call, per thread, to tell the kernel where its list of 18 by the exiting thread. 34 A thread that anticipates possibly using robust_futexes should first 42 bits on 64 bit arch's, and local byte order. Each thread should have 43 its own thread private 'head'. 45 If a thread is running in 32 bit compatibility mode on a 64 native arch 62 word' holds 3 flag bits in the upper 3 bits, and the thread id (TID) 63 of the thread holding the lock in the bottom 29 bits. See further 68 and is needed to correctly resolve races should a thread exit while [all …]
|
/linux-4.4.14/arch/x86/include/asm/ |
D | switch_to.h | 55 : [prev_sp] "=m" (prev->thread.sp), \ 56 [prev_ip] "=m" (prev->thread.ip), \ 66 : [next_sp] "m" (next->thread.sp), \ 67 [next_ip] "m" (next->thread.ip), \ 123 [threadrsp] "i" (offsetof(struct task_struct, thread.sp)), \
|
D | processor.h | 441 native_load_sp0(struct tss_struct *tss, struct thread_struct *thread) in native_load_sp0() argument 443 tss->x86_tss.sp0 = thread->sp0; in native_load_sp0() 446 if (unlikely(tss->x86_tss.ss1 != thread->sysenter_cs)) { in native_load_sp0() 447 tss->x86_tss.ss1 = thread->sysenter_cs; in native_load_sp0() 448 wrmsr(MSR_IA32_SYSENTER_CS, thread->sysenter_cs, 0); in native_load_sp0() 478 struct thread_struct *thread) in load_sp0() argument 480 native_load_sp0(tss, thread); in load_sp0() 769 #define thread_saved_pc(t) (*(unsigned long *)((t)->thread.sp - 8)) 771 #define task_pt_regs(tsk) ((struct pt_regs *)(tsk)->thread.sp0 - 1)
|
/linux-4.4.14/arch/blackfin/include/asm/ |
D | processor.h | 88 #define thread_saved_pc(tsk) (tsk->thread.pc) 95 if ((tsk)->thread.esp0 > PAGE_SIZE && \ 96 MAP_NR((tsk)->thread.esp0) < max_mapnr) \ 97 eip = ((struct pt_regs *) (tsk)->thread.esp0)->pc; \ 99 #define KSTK_ESP(tsk) ((tsk) == current ? rdusp() : (tsk)->thread.usp)
|
/linux-4.4.14/arch/ia64/kernel/ |
D | process.c | 176 if (current->thread.pfm_needs_checking) in do_notify_resume_user() 269 if ((task->thread.flags & IA64_THREAD_DBG_VALID) != 0) in ia64_save_extra() 270 ia64_save_debug_regs(&task->thread.dbr[0]); in ia64_save_extra() 273 if ((task->thread.flags & IA64_THREAD_PM_VALID) != 0) in ia64_save_extra() 289 if ((task->thread.flags & IA64_THREAD_DBG_VALID) != 0) in ia64_load_extra() 290 ia64_load_debug_regs(&task->thread.dbr[0]); in ia64_load_extra() 293 if ((task->thread.flags & IA64_THREAD_PM_VALID) != 0) in ia64_load_extra() 352 p->thread.ksp = (unsigned long) child_stack - 16; in copy_thread() 373 p->thread.flags = ((current->thread.flags & ~THREAD_FLAGS_TO_CLEAR) in copy_thread() 436 if (current->thread.pfm_context) in copy_thread() [all …]
|
/linux-4.4.14/drivers/vfio/ |
D | virqfd.c | 55 virqfd->thread) in virqfd_wakeup() 103 if (virqfd->thread) in virqfd_inject() 104 virqfd->thread(virqfd->opaque, virqfd->data); in virqfd_inject() 109 void (*thread)(void *, void *), in vfio_virqfd_enable() 125 virqfd->thread = thread; in vfio_virqfd_enable() 176 if ((!handler || handler(opaque, data)) && thread) in vfio_virqfd_enable()
|
/linux-4.4.14/drivers/irqchip/ |
D | irq-metag.c | 163 unsigned int thread = hard_processor_id(); in metag_internal_irq_unmask() local 168 metag_out32(TBI_TRIG_VEC(TBID_SIGNUM_TR1(thread)), vec_addr); in metag_internal_irq_unmask() 195 unsigned int cpu, thread; in metag_internal_irq_set_affinity() local 205 thread = cpu_2_hwthread_id[cpu]; in metag_internal_irq_set_affinity() 207 metag_out32(TBI_TRIG_VEC(TBID_SIGNUM_TR1(thread)), in metag_internal_irq_set_affinity() 284 unsigned int thread = cpu_2_hwthread_id[cpu]; in metag_internal_irq_init_cpu() local 285 unsigned int signum = TBID_SIGNUM_TR1(thread); in metag_internal_irq_init_cpu()
|
D | irq-metag-ext.c | 134 int thread = hard_processor_id(); in meta_intc_startup_irq() local 141 metag_out32(TBI_TRIG_VEC(TBID_SIGNUM_TR2(thread)), vec_addr); in meta_intc_startup_irq() 338 unsigned int thread = hard_processor_id(); in meta_intc_unmask_edge_irq_nomask() local 343 metag_out32(TBI_TRIG_VEC(TBID_SIGNUM_TR2(thread)), vec_addr); in meta_intc_unmask_edge_irq_nomask() 379 unsigned int thread = hard_processor_id(); in meta_intc_unmask_level_irq_nomask() local 384 metag_out32(TBI_TRIG_VEC(TBID_SIGNUM_TR2(thread)), vec_addr); in meta_intc_unmask_level_irq_nomask() 508 unsigned int cpu, thread; in meta_intc_set_affinity() local 518 thread = cpu_2_hwthread_id[cpu]; in meta_intc_set_affinity() 520 metag_out32(TBI_TRIG_VEC(TBID_SIGNUM_TR2(thread)), vec_addr); in meta_intc_set_affinity() 781 unsigned int thread = cpu_2_hwthread_id[cpu]; in meta_intc_init_cpu() local [all …]
|
/linux-4.4.14/drivers/thermal/ |
D | intel_powerclamp.c | 511 struct task_struct *thread; in start_power_clamp() local 536 thread = kthread_create_on_node(clamp_thread, in start_power_clamp() 541 if (likely(!IS_ERR(thread))) { in start_power_clamp() 542 kthread_bind(thread, cpu); in start_power_clamp() 543 wake_up_process(thread); in start_power_clamp() 544 *p = thread; in start_power_clamp() 556 struct task_struct *thread; in end_power_clamp() local 568 thread = *per_cpu_ptr(powerclamp_thread, i); in end_power_clamp() 569 kthread_stop(thread); in end_power_clamp() 578 struct task_struct *thread; in powerclamp_cpu_callback() local [all …]
|
/linux-4.4.14/tools/perf/arch/x86/include/ |
D | arch-tests.h | 11 struct thread; 14 struct thread *thread);
|
/linux-4.4.14/arch/m68k/include/asm/ |
D | processor.h | 170 if ((tsk)->thread.esp0 > PAGE_SIZE && \ 171 (virt_addr_valid((tsk)->thread.esp0))) \ 172 eip = ((struct pt_regs *) (tsk)->thread.esp0)->pc; \ 174 #define KSTK_ESP(tsk) ((tsk) == current ? rdusp() : (tsk)->thread.usp) 176 #define task_pt_regs(tsk) ((struct pt_regs *) ((tsk)->thread.esp0))
|
/linux-4.4.14/arch/ia64/include/asm/ |
D | switch_to.h | 41 ((t)->thread.flags & (IA64_THREAD_DBG_VALID|IA64_THREAD_PM_VALID) \ 63 (prev)->thread.flags |= IA64_THREAD_FPH_VALID; \ 64 __ia64_save_fpu((prev)->thread.fph); \ 68 if (unlikely((current->thread.flags & IA64_THREAD_MIGRATION) && \
|
D | processor.h | 42 #define TASK_UNMAPPED_BASE (current->thread.map_base) 253 (task)->thread.flags = (((task)->thread.flags & ~IA64_THREAD_UAC_MASK) \ 259 put_user(((task)->thread.flags & IA64_THREAD_UAC_MASK) >> IA64_THREAD_UAC_SHIFT, \ 265 (task)->thread.flags = (((task)->thread.flags & ~IA64_THREAD_FPEMU_MASK) \ 271 put_user(((task)->thread.flags & IA64_THREAD_FPEMU_MASK) >> IA64_THREAD_FPEMU_SHIFT, \ 317 regs->ar_bspstore = current->thread.rbs_bot; \ 354 #define KSTK_ESP(tsk) ((tsk)->thread.ksp) 404 (__ia64_islfo_task->thread.last_fph_cpu == smp_processor_id() \ 414 __ia64_slfo_task->thread.last_fph_cpu = smp_processor_id(); \ 419 #define ia64_drop_fpu(t) ((t)->thread.last_fph_cpu = -1)
|
/linux-4.4.14/arch/frv/include/asm/ |
D | switch_to.h | 29 (prev)->thread.sched_lr = \ 31 (last) = __switch_to(&(prev)->thread, &(next)->thread, (prev)); \
|
/linux-4.4.14/arch/metag/include/asm/ |
D | hwthread.h | 26 void __iomem *__CU_addr(unsigned int thread, unsigned int regnum) in __CU_addr() argument 30 WARN_ON(thread == BAD_HWTHREAD_ID); in __CU_addr() 34 thread_offset = TnUCTRX_STRIDE * thread; in __CU_addr()
|
D | processor.h | 114 current->thread.int_depth = 1; \ 143 ((unsigned long)(tsk)->thread.kernel_context->CurrPC) 145 ((unsigned long)(tsk)->thread.kernel_context->AX[0].U0) 147 ((unsigned long)(tsk)->thread.kernel_context->AX[1].U0)
|
/linux-4.4.14/arch/cris/arch-v10/kernel/ |
D | process.c | 114 p->thread.ksp = (unsigned long) swstack; in copy_thread() 115 p->thread.usp = 0; in copy_thread() 132 p->thread.usp = usp ?: rdusp(); in copy_thread() 136 p->thread.ksp = (unsigned long) swstack; in copy_thread() 157 esp = p->thread.esp; in get_wchan()
|
/linux-4.4.14/arch/m32r/kernel/ |
D | process.c | 44 return tsk->thread.lr; in thread_saved_pc() 116 memset(¤t->thread.debug_trap, 0, sizeof(struct debug_trap)); in flush_thread() 143 tsk->thread.lr = (unsigned long)ret_from_kernel_thread; in copy_thread() 150 tsk->thread.lr = (unsigned long)ret_from_fork; in copy_thread() 152 tsk->thread.sp = (unsigned long)childregs; in copy_thread()
|
/linux-4.4.14/kernel/debug/ |
D | gdbstub.c | 495 struct task_struct *thread; in gdb_get_regs_helper() local 499 thread = kgdb_usethread; in gdb_get_regs_helper() 500 if (!thread) { in gdb_get_regs_helper() 501 thread = kgdb_info[ks->cpu].task; in gdb_get_regs_helper() 512 if (thread == kgdb_info[i].task) in gdb_get_regs_helper() 532 sleeping_thread_to_gdb_regs(gdb_regs, thread); in gdb_get_regs_helper() 800 struct task_struct *thread; in gdb_cmd_task() local 807 thread = getthread(ks->linux_regs, ks->threadid); in gdb_cmd_task() 808 if (!thread && ks->threadid > 0) { in gdb_cmd_task() 812 kgdb_usethread = thread; in gdb_cmd_task() [all …]
|
/linux-4.4.14/drivers/block/paride/ |
D | Transition-notes | 49 1) at any time at most one thread of execution can be in that area or 51 2) When there is such a thread, pd_busy is set or pd_lock is held by 52 that thread. 53 3) When there is such a thread, ps_tq_active is 0 or ps_spinlock is 54 held by that thread. 55 4) When there is such a thread, all PIA belonging to pd.c have NULL 56 ->claim_cont or pi_spinlock is held by thread in question. 60 (1) can become not true if some thread enters that area while another is there. 62 In the first case the thread was already in the area. In the second, 63 the thread was holding pd_lock and found pd_busy not set, which would [all …]
|
/linux-4.4.14/kernel/ |
D | seccomp.c | 256 struct task_struct *thread, *caller; in seccomp_can_sync_threads() local 263 for_each_thread(caller, thread) { in seccomp_can_sync_threads() 267 if (thread == caller) in seccomp_can_sync_threads() 270 if (thread->seccomp.mode == SECCOMP_MODE_DISABLED || in seccomp_can_sync_threads() 271 (thread->seccomp.mode == SECCOMP_MODE_FILTER && in seccomp_can_sync_threads() 272 is_ancestor(thread->seccomp.filter, in seccomp_can_sync_threads() 277 failed = task_pid_vnr(thread); in seccomp_can_sync_threads() 297 struct task_struct *thread, *caller; in seccomp_sync_threads() local 304 for_each_thread(caller, thread) { in seccomp_sync_threads() 306 if (thread == caller) in seccomp_sync_threads() [all …]
|
/linux-4.4.14/arch/h8300/include/asm/ |
D | processor.h | 129 if ((tsk)->thread.esp0 > PAGE_SIZE && \ 130 MAP_NR((tsk)->thread.esp0) < max_mapnr) \ 131 eip = ((struct pt_regs *) (tsk)->thread.esp0)->pc; \ 134 #define KSTK_ESP(tsk) ((tsk) == current ? rdusp() : (tsk)->thread.usp)
|
/linux-4.4.14/arch/sh/mm/ |
D | alignment.c | 72 if (current->thread.flags & SH_THREAD_UAC_SIGBUS) { in unaligned_user_action() 77 if (current->thread.flags & SH_THREAD_UAC_NOPRINT) in unaligned_user_action() 85 return put_user(tsk->thread.flags & SH_THREAD_UAC_MASK, in get_unalign_ctl() 91 tsk->thread.flags = (tsk->thread.flags & ~SH_THREAD_UAC_MASK) | in set_unalign_ctl()
|
/linux-4.4.14/arch/arm/probes/uprobes/ |
D | core.c | 143 utask->autask.saved_trap_no = current->thread.trap_no; in arch_uprobe_pre_xol() 144 current->thread.trap_no = UPROBE_TRAP_NR; in arch_uprobe_pre_xol() 154 WARN_ON_ONCE(current->thread.trap_no != UPROBE_TRAP_NR); in arch_uprobe_post_xol() 156 current->thread.trap_no = utask->autask.saved_trap_no; in arch_uprobe_post_xol() 167 if (t->thread.trap_no != UPROBE_TRAP_NR) in arch_uprobe_xol_was_trapped() 177 current->thread.trap_no = utask->autask.saved_trap_no; in arch_uprobe_abort_xol()
|
/linux-4.4.14/Documentation/filesystems/nfs/ |
D | knfsd-stats.txt | 29 for each NFS thread pool. 36 The id number of the NFS thread pool to which this line applies. 40 at zero. The maximum value depends on the thread pool mode, but 42 Note that in the default case there will be a single thread pool 62 an nfsd thread to service it, i.e. no nfsd thread was considered 71 This can happen because there are too few nfsd threads in the thread 72 pool for the NFS workload (the workload is thread-limited), in which 77 Counts how many times an idle nfsd thread is woken to try to 86 Counts how many times an nfsd thread triggered an idle timeout, 107 one of three ways. An nfsd thread can be woken (threads-woken counts [all …]
|
/linux-4.4.14/Documentation/kdump/ |
D | gdbmacros.txt | 25 set var $stackp = $next_t.thread.esp 39 set var $stackp = $next_t.thread.esp 54 dump all thread stack traces on a kernel compiled with !CONFIG_FRAME_POINTER 66 set var $stackp = $next_t.thread.esp 82 set var $stackp = $next_t.thread.esp 98 dump all thread stack traces on a kernel compiled with CONFIG_FRAME_POINTER 129 set var $stackp = $pid_task.thread.esp 171 printf "Trapno %ld, cr2 0x%lx, error_code %ld\n", $pid_task.thread.trap_no, \ 172 $pid_task.thread.cr2, $pid_task.thread.error_code 176 Run info threads and lookup pid of thread #1
|
/linux-4.4.14/arch/xtensa/kernel/ |
D | process.c | 204 p->thread.sp = (unsigned long)childregs; in copy_thread() 211 p->thread.ra = MAKE_RA_FOR_CALL( in copy_thread() 257 p->thread.ra = MAKE_RA_FOR_CALL( in copy_thread() 293 sp = p->thread.sp; in get_wchan() 294 pc = MAKE_PC_FROM_RA(p->thread.ra, p->thread.sp); in get_wchan()
|
/linux-4.4.14/arch/x86/um/asm/ |
D | processor_32.h | 38 static inline void arch_flush_thread(struct arch_thread *thread) in arch_flush_thread() argument 41 memset(&thread->tls_array, 0, sizeof(thread->tls_array)); in arch_flush_thread()
|
/linux-4.4.14/drivers/scsi/bnx2i/ |
D | bnx2i_init.c | 423 struct task_struct *thread; in bnx2i_percpu_thread_create() local 427 thread = kthread_create_on_node(bnx2i_percpu_io_thread, (void *)p, in bnx2i_percpu_thread_create() 431 if (likely(!IS_ERR(thread))) { in bnx2i_percpu_thread_create() 432 kthread_bind(thread, cpu); in bnx2i_percpu_thread_create() 433 p->iothread = thread; in bnx2i_percpu_thread_create() 434 wake_up_process(thread); in bnx2i_percpu_thread_create() 442 struct task_struct *thread; in bnx2i_percpu_thread_destroy() local 448 thread = p->iothread; in bnx2i_percpu_thread_destroy() 460 if (thread) in bnx2i_percpu_thread_destroy() 461 kthread_stop(thread); in bnx2i_percpu_thread_destroy()
|
/linux-4.4.14/arch/mips/mm/ |
D | fault.c | 63 current->thread.trap_nr, SIGSEGV) == NOTIFY_STOP) in __do_page_fault() 208 tsk->thread.cp0_badvaddr = address; in __do_page_fault() 209 tsk->thread.error_code = write; in __do_page_fault() 225 current->thread.trap_nr = (regs->cp0_cause >> 2) & 0x1f; in __do_page_fault() 237 current->thread.cp0_baduaddr = address; in __do_page_fault() 284 current->thread.trap_nr = (regs->cp0_cause >> 2) & 0x1f; in __do_page_fault() 285 tsk->thread.cp0_badvaddr = address; in __do_page_fault()
|
/linux-4.4.14/Documentation/devicetree/bindings/arm/ |
D | topology.txt | 14 - thread 19 The bottom hierarchy level sits at core or thread level depending on whether 23 threads existing in the system and map to the hierarchy level "thread" above. 74 - thread node 78 The nodes describing the CPU topology (cluster/core/thread) can only 79 be defined within the cpu-map node and every core/thread in the system 89 cluster/core/thread) (where N = {0, 1, ...} is the node number; nodes which 97 3 - cluster/core/thread node bindings 100 Bindings for cluster/cpu/thread nodes are defined as follows: 124 thread nodes. [all …]
|
/linux-4.4.14/arch/cris/arch-v32/kernel/ |
D | process.c | 123 p->thread.ksp = (unsigned long) swstack; in copy_thread() 124 p->thread.usp = 0; in copy_thread() 149 p->thread.usp = usp ?: rdusp(); in copy_thread() 150 p->thread.ksp = (unsigned long) swstack; in copy_thread()
|
/linux-4.4.14/arch/nios2/include/asm/ |
D | processor.h | 84 #define thread_saved_pc(tsk) ((tsk)->thread.kregs->ea) 92 #define KSTK_EIP(tsk) ((tsk)->thread.kregs->ea) 93 #define KSTK_ESP(tsk) ((tsk)->thread.kregs->sp)
|
/linux-4.4.14/tools/perf/ui/browsers/ |
D | hists.c | 1256 static struct thread *hist_browser__selected_thread(struct hist_browser *browser) in hist_browser__selected_thread() 1258 return browser->he_selection->thread; in hist_browser__selected_thread() 1274 const struct thread *thread = hists->thread_filter; in hists__browser_title() local 1321 if (thread) in hists__browser_title() 1324 (thread->comm_set ? thread__comm_str(thread) : ""), in hists__browser_title() 1325 thread->tid); in hists__browser_title() 1440 struct thread *thread; member 1498 struct thread *thread = act->thread; in do_zoom_thread() local 1507 thread->comm_set ? thread__comm_str(thread) : "", in do_zoom_thread() 1508 thread->tid); in do_zoom_thread() [all …]
|
/linux-4.4.14/arch/x86/mm/ |
D | fault.c | 311 if (!v8086_mode(regs) || !tsk->thread.vm86) in check_v8086_mode() 316 tsk->thread.vm86->screen_bitmap |= 1 << bit; in check_v8086_mode() 647 tsk->thread.cr2 = address; in pgtable_bad() 648 tsk->thread.trap_nr = X86_TRAP_PF; in pgtable_bad() 649 tsk->thread.error_code = error_code; in pgtable_bad() 682 tsk->thread.trap_nr = X86_TRAP_PF; in no_context() 683 tsk->thread.error_code = error_code | PF_USER; in no_context() 684 tsk->thread.cr2 = address; in no_context() 724 tsk->thread.cr2 = address; in no_context() 725 tsk->thread.trap_nr = X86_TRAP_PF; in no_context() [all …]
|
/linux-4.4.14/drivers/staging/lustre/lustre/include/ |
D | lustre_net.h | 1817 static inline int thread_is_init(struct ptlrpc_thread *thread) in thread_is_init() argument 1819 return thread->t_flags == 0; in thread_is_init() 1822 static inline int thread_is_stopped(struct ptlrpc_thread *thread) in thread_is_stopped() argument 1824 return !!(thread->t_flags & SVC_STOPPED); in thread_is_stopped() 1827 static inline int thread_is_stopping(struct ptlrpc_thread *thread) in thread_is_stopping() argument 1829 return !!(thread->t_flags & SVC_STOPPING); in thread_is_stopping() 1832 static inline int thread_is_starting(struct ptlrpc_thread *thread) in thread_is_starting() argument 1834 return !!(thread->t_flags & SVC_STARTING); in thread_is_starting() 1837 static inline int thread_is_running(struct ptlrpc_thread *thread) in thread_is_running() argument 1839 return !!(thread->t_flags & SVC_RUNNING); in thread_is_running() [all …]
|
/linux-4.4.14/arch/arm64/include/asm/ |
D | thread_info.h | 83 ((unsigned long)(tsk->thread.cpu_context.pc)) 85 ((unsigned long)(tsk->thread.cpu_context.sp)) 87 ((unsigned long)(tsk->thread.cpu_context.fp))
|
/linux-4.4.14/arch/powerpc/platforms/cell/ |
D | cbe_regs.c | 266 struct cbe_thread_map *thread = &cbe_thread_map[i]; in cbe_regs_init() local 268 if (thread->cpu_node == cpu) { in cbe_regs_init() 269 thread->regs = map; in cbe_regs_init() 270 thread->cbe_id = cbe_id; in cbe_regs_init() 271 map->be_node = thread->be_node; in cbe_regs_init() 273 if(thread->thread_id == 0) in cbe_regs_init()
|
/linux-4.4.14/crypto/ |
D | algboss.c | 97 struct task_struct *thread; in cryptomgr_schedule_probe() local 193 thread = kthread_run(cryptomgr_probe, param, "cryptomgr_probe"); in cryptomgr_schedule_probe() 194 if (IS_ERR(thread)) in cryptomgr_schedule_probe() 235 struct task_struct *thread; in cryptomgr_schedule_test() local 260 thread = kthread_run(cryptomgr_test, param, "cryptomgr_test"); in cryptomgr_schedule_test() 261 if (IS_ERR(thread)) in cryptomgr_schedule_test()
|