sd_flag          1602 kernel/sched/deadline.c select_task_rq_dl(struct task_struct *p, int cpu, int sd_flag, int flags)
sd_flag          1607 kernel/sched/deadline.c 	if (sd_flag != SD_BALANCE_WAKE)
sd_flag          5581 kernel/sched/fair.c 		  int this_cpu, int sd_flag)
sd_flag          5674 kernel/sched/fair.c 	if (sd_flag & SD_BALANCE_FORK)
sd_flag          5773 kernel/sched/fair.c 				  int cpu, int prev_cpu, int sd_flag)
sd_flag          5784 kernel/sched/fair.c 	if (!(sd_flag & SD_BALANCE_FORK))
sd_flag          5792 kernel/sched/fair.c 		if (!(sd->flags & sd_flag)) {
sd_flag          5797 kernel/sched/fair.c 		group = find_idlest_group(sd, p, cpu, sd_flag);
sd_flag          5817 kernel/sched/fair.c 			if (tmp->flags & sd_flag)
sd_flag          6469 kernel/sched/fair.c select_task_rq_fair(struct task_struct *p, int prev_cpu, int sd_flag, int wake_flags)
sd_flag          6477 kernel/sched/fair.c 	if (sd_flag & SD_BALANCE_WAKE) {
sd_flag          6509 kernel/sched/fair.c 		if (tmp->flags & sd_flag)
sd_flag          6517 kernel/sched/fair.c 		new_cpu = find_idlest_cpu(sd, p, cpu, prev_cpu, sd_flag);
sd_flag          6518 kernel/sched/fair.c 	} else if (sd_flag & SD_BALANCE_WAKE) { /* XXX always ? */
sd_flag           364 kernel/sched/idle.c select_task_rq_idle(struct task_struct *p, int cpu, int sd_flag, int flags)
sd_flag          1390 kernel/sched/rt.c select_task_rq_rt(struct task_struct *p, int cpu, int sd_flag, int flags)
sd_flag          1396 kernel/sched/rt.c 	if (sd_flag != SD_BALANCE_WAKE && sd_flag != SD_BALANCE_FORK)
sd_flag          1741 kernel/sched/sched.h 	int  (*select_task_rq)(struct task_struct *p, int task_cpu, int sd_flag, int flags);
sd_flag            14 kernel/sched/stop_task.c select_task_rq_stop(struct task_struct *p, int cpu, int sd_flag, int flags)