Searched refs:ret_stack (Results 1 – 12 of 12) sorted by relevance
29 if (!current->ret_stack) in push_return_trace()40 current->ret_stack[index].ret = ret; in push_return_trace()41 current->ret_stack[index].func = func; in push_return_trace()42 current->ret_stack[index].calltime = time; in push_return_trace()64 *ret = current->ret_stack[index].ret; in pop_return_trace()65 trace->func = current->ret_stack[index].func; in pop_return_trace()66 trace->calltime = current->ret_stack[index].calltime; in pop_return_trace()
124 if (!current->ret_stack) in ftrace_push_return_trace()164 current->ret_stack[index].ret = ret; in ftrace_push_return_trace()165 current->ret_stack[index].func = func; in ftrace_push_return_trace()166 current->ret_stack[index].calltime = calltime; in ftrace_push_return_trace()167 current->ret_stack[index].subtime = 0; in ftrace_push_return_trace()168 current->ret_stack[index].fp = frame_pointer; in ftrace_push_return_trace()216 if (unlikely(current->ret_stack[index].fp != frame_pointer)) { in ftrace_pop_return_trace()220 current->ret_stack[index].fp, in ftrace_pop_return_trace()222 (void *)current->ret_stack[index].func, in ftrace_pop_return_trace()223 current->ret_stack[index].ret); in ftrace_pop_return_trace()[all …]
922 current->ret_stack[index - 1].subtime += calltime; in profile_graph_return()924 if (current->ret_stack[index].subtime < calltime) in profile_graph_return()925 calltime -= current->ret_stack[index].subtime; in profile_graph_return()5676 if (t->ret_stack == NULL) { in alloc_retstack_tasklist()5682 t->ret_stack = ret_stack_list[start++]; in alloc_retstack_tasklist()5723 next->ret_stack[index].calltime += timestamp; in ftrace_graph_probe_sched_switch()5741 if (!idle_task(cpu)->ret_stack) in start_graph_tracing()5898 graph_init_task(struct task_struct *t, struct ftrace_ret_stack *ret_stack) in graph_init_task() argument5905 t->ret_stack = ret_stack; in graph_init_task()5919 if (t->ret_stack) in ftrace_graph_init_idle_task()[all …]
60 if (t->ret_stack && index >= graph) { in __save_stack_trace()61 pc = t->ret_stack[index - graph].ret; in __save_stack_trace()
1734 if (current->ret_stack && index >= graph) { in perf_callchain_kernel()1735 pc = current->ret_stack[index - graph].ret; in perf_callchain_kernel()
2369 if (tsk->ret_stack && index >= graph) { in show_stack()2370 pc = tsk->ret_stack[index - graph].ret; in show_stack()
66 if (!task->ret_stack || index < *graph) in print_ftrace_graph_addr()70 ret_addr = task->ret_stack[index].ret; in print_ftrace_graph_addr()
621 pc = current->ret_stack[index].ret; in dwarf_unwind_stack()
57 if (!task->ret_stack || index < *graph) in print_ftrace_graph_addr()61 ret_addr = task->ret_stack[index].ret; in print_ftrace_graph_addr()
738 #define INIT_FTRACE_GRAPH .ret_stack = NULL,
1693 struct ftrace_ret_stack *ret_stack; member
1576 (void *)current->ret_stack[curr_frame].ret); in show_stack()