btm 37 kernel/trace/bpf_trace.c struct bpf_trace_module *btm; btm 41 kernel/trace/bpf_trace.c list_for_each_entry(btm, &bpf_trace_modules, list) { btm 42 kernel/trace/bpf_trace.c for (i = 0; i < btm->module->num_bpf_raw_events; ++i) { btm 43 kernel/trace/bpf_trace.c btp = &btm->module->bpf_raw_events[i]; btm 45 kernel/trace/bpf_trace.c if (try_module_get(btm->module)) btm 1476 kernel/trace/bpf_trace.c struct bpf_trace_module *btm, *tmp; btm 1487 kernel/trace/bpf_trace.c btm = kzalloc(sizeof(*btm), GFP_KERNEL); btm 1488 kernel/trace/bpf_trace.c if (btm) { btm 1489 kernel/trace/bpf_trace.c btm->module = module; btm 1490 kernel/trace/bpf_trace.c list_add(&btm->list, &bpf_trace_modules); btm 1494 kernel/trace/bpf_trace.c list_for_each_entry_safe(btm, tmp, &bpf_trace_modules, list) { btm 1495 kernel/trace/bpf_trace.c if (btm->module == module) { btm 1496 kernel/trace/bpf_trace.c list_del(&btm->list); btm 1497 kernel/trace/bpf_trace.c kfree(btm);