btm                37 kernel/trace/bpf_trace.c 	struct bpf_trace_module *btm;
btm                41 kernel/trace/bpf_trace.c 	list_for_each_entry(btm, &bpf_trace_modules, list) {
btm                42 kernel/trace/bpf_trace.c 		for (i = 0; i < btm->module->num_bpf_raw_events; ++i) {
btm                43 kernel/trace/bpf_trace.c 			btp = &btm->module->bpf_raw_events[i];
btm                45 kernel/trace/bpf_trace.c 				if (try_module_get(btm->module))
btm              1476 kernel/trace/bpf_trace.c 	struct bpf_trace_module *btm, *tmp;
btm              1487 kernel/trace/bpf_trace.c 		btm = kzalloc(sizeof(*btm), GFP_KERNEL);
btm              1488 kernel/trace/bpf_trace.c 		if (btm) {
btm              1489 kernel/trace/bpf_trace.c 			btm->module = module;
btm              1490 kernel/trace/bpf_trace.c 			list_add(&btm->list, &bpf_trace_modules);
btm              1494 kernel/trace/bpf_trace.c 		list_for_each_entry_safe(btm, tmp, &bpf_trace_modules, list) {
btm              1495 kernel/trace/bpf_trace.c 			if (btm->module == module) {
btm              1496 kernel/trace/bpf_trace.c 				list_del(&btm->list);
btm              1497 kernel/trace/bpf_trace.c 				kfree(btm);