max_stack 79 arch/arm/kernel/perf_callchain.c while ((entry->nr < entry->max_stack) && max_stack 118 arch/arm64/kernel/perf_callchain.c while (entry->nr < entry->max_stack && max_stack 128 arch/arm64/kernel/perf_callchain.c while ((entry->nr < entry->max_stack) && max_stack 106 arch/csky/kernel/perf_callchain.c while (fp && !(fp & 0x3) && entry->nr < entry->max_stack) max_stack 36 arch/mips/kernel/perf_event.c if (entry->nr >= entry->max_stack) max_stack 60 arch/mips/kernel/perf_event.c if (entry->nr >= entry->max_stack) max_stack 247 arch/powerpc/perf/callchain.c while (entry->nr < entry->max_stack) { max_stack 453 arch/powerpc/perf/callchain.c while (entry->nr < entry->max_stack) { max_stack 73 arch/riscv/kernel/perf_callchain.c while (fp && !(fp & 0x3) && entry->nr < entry->max_stack) max_stack 1784 arch/sparc/kernel/perf_event.c } while (entry->nr < entry->max_stack); max_stack 1818 arch/sparc/kernel/perf_event.c } while (entry->nr < entry->max_stack); max_stack 1850 arch/sparc/kernel/perf_event.c } while (entry->nr < entry->max_stack); max_stack 2445 arch/x86/events/core.c while (entry->nr < entry->max_stack) { max_stack 2502 arch/x86/events/core.c while (entry->nr < entry->max_stack) { max_stack 332 arch/xtensa/kernel/perf_event.c xtensa_backtrace_kernel(regs, entry->max_stack, max_stack 339 arch/xtensa/kernel/perf_event.c xtensa_backtrace_user(regs, entry->max_stack, max_stack 485 drivers/net/ethernet/netronome/nfp/bpf/offload.c unsigned int fw_mtu, pkt_off, max_stack, max_prog_len; max_stack 497 drivers/net/ethernet/netronome/nfp/bpf/offload.c max_stack = nn_readb(nn, NFP_NET_CFG_BPF_STACK_SZ) * 64; max_stack 498 drivers/net/ethernet/netronome/nfp/bpf/offload.c if (nfp_prog->stack_size > max_stack) { max_stack 762 drivers/net/ethernet/netronome/nfp/bpf/verifier.c unsigned int max_stack; max_stack 790 drivers/net/ethernet/netronome/nfp/bpf/verifier.c max_stack = nn_readb(nn, NFP_NET_CFG_BPF_STACK_SZ) * 64; max_stack 792 drivers/net/ethernet/netronome/nfp/bpf/verifier.c if (nfp_prog->stack_size > max_stack) { max_stack 794 drivers/net/ethernet/netronome/nfp/bpf/verifier.c nfp_prog->stack_size, max_stack); max_stack 68 include/linux/perf_event.h u32 max_stack; max_stack 1194 include/linux/perf_event.h u32 max_stack, bool crosstask, bool add_mark); max_stack 1196 include/linux/perf_event.h extern int get_callchain_buffers(int max_stack); max_stack 1217 include/linux/perf_event.h if (ctx->nr < ctx->max_stack && !ctx->contexts_maxed) { max_stack 179 kernel/events/callchain.c u32 max_stack, bool crosstask, bool add_mark) max_stack 193 kernel/events/callchain.c ctx.max_stack = max_stack; max_stack 6568 kernel/events/core.c const u32 max_stack = event->attr.sample_max_stack; max_stack 6575 kernel/events/core.c max_stack, crosstask, true); max_stack 86 tools/perf/builtin-report.c int max_stack; max_stack 297 tools/perf/builtin-report.c ret = hist_entry_iter__add(&iter, &al, rep->max_stack, rep); max_stack 1066 tools/perf/builtin-report.c .max_stack = PERF_MAX_STACK_DEPTH, max_stack 1122 tools/perf/builtin-report.c OPT_INTEGER(0, "max-stack", &report.max_stack, max_stack 1264 tools/perf/builtin-report.c (int)itrace_synth_opts.callchain_sz > report.max_stack) max_stack 1265 tools/perf/builtin-report.c report.max_stack = itrace_synth_opts.callchain_sz; max_stack 234 tools/perf/builtin-sched.c unsigned int max_stack; max_stack 2158 tools/perf/builtin-sched.c NULL, NULL, sched->max_stack + 2) != 0) { max_stack 3368 tools/perf/builtin-sched.c .max_stack = 5, max_stack 3412 tools/perf/builtin-sched.c OPT_UINTEGER(0, "max-stack", &sched.max_stack, max_stack 829 tools/perf/builtin-top.c err = hist_entry_iter__add(&iter, &al, top->max_stack, top); max_stack 1417 tools/perf/builtin-top.c .max_stack = sysctl__max_stack(), max_stack 1488 tools/perf/builtin-top.c OPT_INTEGER(0, "max-stack", &top.max_stack, max_stack 137 tools/perf/builtin-trace.c unsigned int max_stack; max_stack 2061 tools/perf/builtin-trace.c int max_stack = evsel->core.attr.sample_max_stack ? max_stack 2063 tools/perf/builtin-trace.c trace->max_stack; max_stack 2069 tools/perf/builtin-trace.c err = thread__resolve_callchain(al.thread, cursor, evsel, sample, NULL, NULL, max_stack); max_stack 3443 tools/perf/builtin-trace.c evsel->core.attr.sample_max_stack = trace->max_stack; max_stack 4097 tools/perf/builtin-trace.c .max_stack = UINT_MAX, max_stack 4160 tools/perf/builtin-trace.c OPT_UINTEGER(0, "max-stack", &trace.max_stack, max_stack 4271 tools/perf/builtin-trace.c if (trace.max_stack == UINT_MAX) { max_stack 4272 tools/perf/builtin-trace.c trace.max_stack = input_name ? PERF_MAX_STACK_DEPTH : sysctl__max_stack(); max_stack 1086 tools/perf/util/callchain.c int max_stack) max_stack 1094 tools/perf/util/callchain.c parent, al, max_stack); max_stack 103 tools/perf/util/callchain.h u16 max_stack; max_stack 244 tools/perf/util/callchain.h int max_stack); max_stack 698 tools/perf/util/evsel.c attr->sample_max_stack = param->max_stack; max_stack 785 tools/perf/util/evsel.c int max_stack = 0; max_stack 825 tools/perf/util/evsel.c max_stack = term->val.max_stack; max_stack 855 tools/perf/util/evsel.c if ((callgraph_buf != NULL) || (dump_size > 0) || max_stack) { max_stack 858 tools/perf/util/evsel.c if (max_stack) { max_stack 859 tools/perf/util/evsel.c param.max_stack = max_stack; max_stack 40 tools/perf/util/evsel_config.h int max_stack; max_stack 2173 tools/perf/util/machine.c int max_stack) max_stack 2176 tools/perf/util/machine.c int chain_nr = min(max_stack, (int)chain->nr), i; max_stack 2281 tools/perf/util/machine.c int max_stack) max_stack 2296 tools/perf/util/machine.c root_al, max_stack); max_stack 2320 tools/perf/util/machine.c int nr = min(max_stack, (int)branch->nr); max_stack 2388 tools/perf/util/machine.c i < chain_nr && nr_entries < max_stack; i++) { max_stack 2485 tools/perf/util/machine.c int max_stack) max_stack 2498 tools/perf/util/machine.c thread, sample, max_stack); max_stack 2507 tools/perf/util/machine.c int max_stack) max_stack 2517 tools/perf/util/machine.c max_stack); max_stack 2522 tools/perf/util/machine.c max_stack); max_stack 2526 tools/perf/util/machine.c max_stack); max_stack 2532 tools/perf/util/machine.c max_stack); max_stack 183 tools/perf/util/machine.h int max_stack); max_stack 1216 tools/perf/util/parse-events.c ADD_CONFIG_TERM(MAX_STACK, max_stack, term->val.num); max_stack 34 tools/perf/util/top.h int max_stack; max_stack 190 tools/perf/util/unwind-libdw.c return entry(pc, ui) || !(--ui->max_stack) ? max_stack 197 tools/perf/util/unwind-libdw.c int max_stack) max_stack 205 tools/perf/util/unwind-libdw.c .max_stack = max_stack, max_stack 213 tools/perf/util/unwind-libdw.c ui = zalloc(sizeof(ui_buf) + sizeof(ui_buf.entries[0]) * max_stack); max_stack 237 tools/perf/util/unwind-libdw.c if (err && ui->max_stack != max_stack) max_stack 21 tools/perf/util/unwind-libdw.h int max_stack; max_stack 641 tools/perf/util/unwind-libunwind-local.c void *arg, int max_stack) max_stack 644 tools/perf/util/unwind-libunwind-local.c unw_word_t ips[max_stack]; max_stack 660 tools/perf/util/unwind-libunwind-local.c if (max_stack - 1 > 0) { max_stack 671 tools/perf/util/unwind-libunwind-local.c while (!ret && (unw_step(&c) > 0) && i < max_stack) { max_stack 687 tools/perf/util/unwind-libunwind-local.c max_stack = i; max_stack 693 tools/perf/util/unwind-libunwind-local.c for (i = 0; i < max_stack && !ret; i++) { max_stack 697 tools/perf/util/unwind-libunwind-local.c j = max_stack - i - 1; max_stack 706 tools/perf/util/unwind-libunwind-local.c struct perf_sample *data, int max_stack) max_stack 717 tools/perf/util/unwind-libunwind-local.c if (max_stack <= 0) max_stack 720 tools/perf/util/unwind-libunwind-local.c return get_entries(&ui, cb, arg, max_stack); max_stack 85 tools/perf/util/unwind-libunwind.c struct perf_sample *data, int max_stack) max_stack 88 tools/perf/util/unwind-libunwind.c return thread->mg->unwind_libunwind_ops->get_entries(cb, arg, thread, data, max_stack); max_stack 28 tools/perf/util/unwind.h struct perf_sample *data, int max_stack); max_stack 34 tools/perf/util/unwind.h struct perf_sample *data, int max_stack); max_stack 71 tools/perf/util/unwind.h int max_stack __maybe_unused)