sd_flag 1602 kernel/sched/deadline.c select_task_rq_dl(struct task_struct *p, int cpu, int sd_flag, int flags) sd_flag 1607 kernel/sched/deadline.c if (sd_flag != SD_BALANCE_WAKE) sd_flag 5581 kernel/sched/fair.c int this_cpu, int sd_flag) sd_flag 5674 kernel/sched/fair.c if (sd_flag & SD_BALANCE_FORK) sd_flag 5773 kernel/sched/fair.c int cpu, int prev_cpu, int sd_flag) sd_flag 5784 kernel/sched/fair.c if (!(sd_flag & SD_BALANCE_FORK)) sd_flag 5792 kernel/sched/fair.c if (!(sd->flags & sd_flag)) { sd_flag 5797 kernel/sched/fair.c group = find_idlest_group(sd, p, cpu, sd_flag); sd_flag 5817 kernel/sched/fair.c if (tmp->flags & sd_flag) sd_flag 6469 kernel/sched/fair.c select_task_rq_fair(struct task_struct *p, int prev_cpu, int sd_flag, int wake_flags) sd_flag 6477 kernel/sched/fair.c if (sd_flag & SD_BALANCE_WAKE) { sd_flag 6509 kernel/sched/fair.c if (tmp->flags & sd_flag) sd_flag 6517 kernel/sched/fair.c new_cpu = find_idlest_cpu(sd, p, cpu, prev_cpu, sd_flag); sd_flag 6518 kernel/sched/fair.c } else if (sd_flag & SD_BALANCE_WAKE) { /* XXX always ? */ sd_flag 364 kernel/sched/idle.c select_task_rq_idle(struct task_struct *p, int cpu, int sd_flag, int flags) sd_flag 1390 kernel/sched/rt.c select_task_rq_rt(struct task_struct *p, int cpu, int sd_flag, int flags) sd_flag 1396 kernel/sched/rt.c if (sd_flag != SD_BALANCE_WAKE && sd_flag != SD_BALANCE_FORK) sd_flag 1741 kernel/sched/sched.h int (*select_task_rq)(struct task_struct *p, int task_cpu, int sd_flag, int flags); sd_flag 14 kernel/sched/stop_task.c select_task_rq_stop(struct task_struct *p, int cpu, int sd_flag, int flags)