cpus_ptr 180 arch/mips/kernel/mips-mt-fpaff.c cpumask_or(&allowed, &p->thread.user_cpus_allowed, p->cpus_ptr); cpus_ptr 131 arch/powerpc/platforms/cell/spufs/sched.c cpumask_copy(&ctx->cpus_allowed, current->cpus_ptr); cpus_ptr 1502 arch/x86/kernel/cpu/resctrl/pseudo_lock.c if (!cpumask_subset(current->cpus_ptr, &plr->d->cpu_mask)) { cpus_ptr 1043 drivers/infiniband/hw/hfi1/affinity.c *proc_mask = current->cpus_ptr; cpus_ptr 1626 drivers/infiniband/hw/qib/qib_file_ops.c const unsigned int cpu = cpumask_first(current->cpus_ptr); cpus_ptr 384 fs/proc/array.c cpumask_pr_args(task->cpus_ptr)); cpus_ptr 386 fs/proc/array.c cpumask_pr_args(task->cpus_ptr)); cpus_ptr 701 include/linux/sched.h const cpumask_t *cpus_ptr; cpus_ptr 73 init/init_task.c .cpus_ptr = &init_task.cpus_mask, cpus_ptr 2865 kernel/cgroup/cpuset.c set_cpus_allowed_ptr(task, current->cpus_ptr); cpus_ptr 914 kernel/fork.c if (orig->cpus_ptr == &orig->cpus_mask) cpus_ptr 915 kernel/fork.c tsk->cpus_ptr = &tsk->cpus_mask; cpus_ptr 1456 kernel/sched/core.c if (!cpumask_test_cpu(cpu, p->cpus_ptr)) cpus_ptr 1652 kernel/sched/core.c if (cpumask_equal(p->cpus_ptr, new_mask)) cpus_ptr 1812 kernel/sched/core.c if (!cpumask_test_cpu(arg->dst_cpu, arg->src_task->cpus_ptr)) cpus_ptr 1815 kernel/sched/core.c if (!cpumask_test_cpu(arg->src_cpu, arg->dst_task->cpus_ptr)) cpus_ptr 1857 kernel/sched/core.c if (!cpumask_test_cpu(arg.dst_cpu, arg.src_task->cpus_ptr)) cpus_ptr 1860 kernel/sched/core.c if (!cpumask_test_cpu(arg.src_cpu, arg.dst_task->cpus_ptr)) cpus_ptr 2045 kernel/sched/core.c if (cpumask_test_cpu(dest_cpu, p->cpus_ptr)) cpus_ptr 2052 kernel/sched/core.c for_each_cpu(dest_cpu, p->cpus_ptr) { cpus_ptr 2106 kernel/sched/core.c cpu = cpumask_any(p->cpus_ptr); cpus_ptr 4923 kernel/sched/core.c if (!cpumask_subset(span, p->cpus_ptr) || cpus_ptr 6132 kernel/sched/core.c if (!cpumask_test_cpu(target_cpu, p->cpus_ptr)) cpus_ptr 123 kernel/sched/cpudeadline.c cpumask_and(later_mask, cp->free_cpus, p->cpus_ptr)) { cpus_ptr 130 kernel/sched/cpudeadline.c if (cpumask_test_cpu(best_cpu, p->cpus_ptr) && cpus_ptr 97 kernel/sched/cpupri.c if (cpumask_any_and(p->cpus_ptr, vec->mask) >= nr_cpu_ids) cpus_ptr 101 kernel/sched/cpupri.c cpumask_and(lowest_mask, p->cpus_ptr, vec->mask); cpus_ptr 542 kernel/sched/deadline.c cpu = cpumask_any_and(cpu_active_mask, p->cpus_ptr); cpus_ptr 1843 kernel/sched/deadline.c cpumask_test_cpu(cpu, p->cpus_ptr)) cpus_ptr 1993 kernel/sched/deadline.c !cpumask_test_cpu(later_rq->cpu, task->cpus_ptr) || cpus_ptr 1631 kernel/sched/fair.c if (!cpumask_test_cpu(env->src_cpu, cur->cpus_ptr)) cpus_ptr 1729 kernel/sched/fair.c if (!cpumask_test_cpu(cpu, env->p->cpus_ptr)) cpus_ptr 5601 kernel/sched/fair.c p->cpus_ptr)) cpus_ptr 5728 kernel/sched/fair.c for_each_cpu_and(i, sched_group_span(group), p->cpus_ptr) { cpus_ptr 5777 kernel/sched/fair.c if (!cpumask_intersects(sched_domain_span(sd), p->cpus_ptr)) cpus_ptr 5894 kernel/sched/fair.c cpumask_and(cpus, sched_domain_span(sd), p->cpus_ptr); cpus_ptr 5928 kernel/sched/fair.c if (!cpumask_test_cpu(cpu, p->cpus_ptr)) cpus_ptr 5992 kernel/sched/fair.c cpumask_and(cpus, sched_domain_span(sd), p->cpus_ptr); cpus_ptr 6035 kernel/sched/fair.c cpumask_test_cpu(p->recent_used_cpu, p->cpus_ptr)) { cpus_ptr 6398 kernel/sched/fair.c if (!cpumask_test_cpu(cpu, p->cpus_ptr)) cpus_ptr 6488 kernel/sched/fair.c cpumask_test_cpu(cpu, p->cpus_ptr); cpus_ptr 7262 kernel/sched/fair.c if (!cpumask_test_cpu(env->dst_cpu, p->cpus_ptr)) { cpus_ptr 7282 kernel/sched/fair.c if (cpumask_test_cpu(cpu, p->cpus_ptr)) { cpus_ptr 9012 kernel/sched/fair.c if (!cpumask_test_cpu(this_cpu, busiest->curr->cpus_ptr)) { cpus_ptr 1607 kernel/sched/rt.c cpumask_test_cpu(cpu, p->cpus_ptr)) cpus_ptr 1744 kernel/sched/rt.c !cpumask_test_cpu(lowest_rq->cpu, task->cpus_ptr) || cpus_ptr 282 kernel/trace/trace_hwlat.c if (!cpumask_equal(current_mask, current->cpus_ptr)) cpus_ptr 26 lib/smp_processor_id.c if (cpumask_equal(current->cpus_ptr, cpumask_of(this_cpu))) cpus_ptr 37 samples/trace_events/trace-events-sample.c current->cpus_ptr);