child_ctx 11515 kernel/events/core.c struct perf_event_context *child_ctx, child_ctx 11532 kernel/events/core.c raw_spin_lock_irq(&child_ctx->lock); child_ctx 11533 kernel/events/core.c WARN_ON_ONCE(child_ctx->is_active); child_ctx 11537 kernel/events/core.c list_del_event(child_event, child_ctx); child_ctx 11539 kernel/events/core.c raw_spin_unlock_irq(&child_ctx->lock); child_ctx 11572 kernel/events/core.c struct perf_event_context *child_ctx, *clone_ctx = NULL; child_ctx 11577 kernel/events/core.c child_ctx = perf_pin_task_context(child, ctxn); child_ctx 11578 kernel/events/core.c if (!child_ctx) child_ctx 11591 kernel/events/core.c mutex_lock(&child_ctx->mutex); child_ctx 11598 kernel/events/core.c raw_spin_lock_irq(&child_ctx->lock); child_ctx 11599 kernel/events/core.c task_ctx_sched_out(__get_cpu_context(child_ctx), child_ctx, EVENT_ALL); child_ctx 11606 kernel/events/core.c put_ctx(child_ctx); /* cannot be last */ child_ctx 11607 kernel/events/core.c WRITE_ONCE(child_ctx->task, TASK_TOMBSTONE); child_ctx 11610 kernel/events/core.c clone_ctx = unclone_ctx(child_ctx); child_ctx 11611 kernel/events/core.c raw_spin_unlock_irq(&child_ctx->lock); child_ctx 11621 kernel/events/core.c perf_event_task(child, child_ctx, 0); child_ctx 11623 kernel/events/core.c list_for_each_entry_safe(child_event, next, &child_ctx->event_list, event_entry) child_ctx 11624 kernel/events/core.c perf_event_exit_event(child_event, child_ctx, child); child_ctx 11626 kernel/events/core.c mutex_unlock(&child_ctx->mutex); child_ctx 11628 kernel/events/core.c put_ctx(child_ctx); child_ctx 11796 kernel/events/core.c struct perf_event_context *child_ctx) child_ctx 11821 kernel/events/core.c !child_ctx->task_ctx_data) { child_ctx 11824 kernel/events/core.c child_ctx->task_ctx_data = kzalloc(pmu->task_ctx_size, child_ctx 11826 kernel/events/core.c if (!child_ctx->task_ctx_data) { child_ctx 11847 kernel/events/core.c get_ctx(child_ctx); child_ctx 11869 kernel/events/core.c child_event->ctx = child_ctx; child_ctx 11883 kernel/events/core.c raw_spin_lock_irqsave(&child_ctx->lock, flags); child_ctx 11884 kernel/events/core.c add_event_to_ctx(child_event, child_ctx); child_ctx 11885 kernel/events/core.c raw_spin_unlock_irqrestore(&child_ctx->lock, flags); child_ctx 11910 kernel/events/core.c struct perf_event_context *child_ctx) child_ctx 11917 kernel/events/core.c child, NULL, child_ctx); child_ctx 11927 kernel/events/core.c child, leader, child_ctx); child_ctx 11956 kernel/events/core.c struct perf_event_context *child_ctx; child_ctx 11963 kernel/events/core.c child_ctx = child->perf_event_ctxp[ctxn]; child_ctx 11964 kernel/events/core.c if (!child_ctx) { child_ctx 11971 kernel/events/core.c child_ctx = alloc_perf_context(parent_ctx->pmu, child); child_ctx 11972 kernel/events/core.c if (!child_ctx) child_ctx 11975 kernel/events/core.c child->perf_event_ctxp[ctxn] = child_ctx; child_ctx 11979 kernel/events/core.c child, child_ctx); child_ctx 11992 kernel/events/core.c struct perf_event_context *child_ctx, *parent_ctx; child_ctx 12054 kernel/events/core.c child_ctx = child->perf_event_ctxp[ctxn]; child_ctx 12056 kernel/events/core.c if (child_ctx && inherited_all) { child_ctx 12066 kernel/events/core.c child_ctx->parent_ctx = cloned_ctx; child_ctx 12067 kernel/events/core.c child_ctx->parent_gen = parent_ctx->parent_gen; child_ctx 12069 kernel/events/core.c child_ctx->parent_ctx = parent_ctx; child_ctx 12070 kernel/events/core.c child_ctx->parent_gen = parent_ctx->generation; child_ctx 12072 kernel/events/core.c get_ctx(child_ctx->parent_ctx);