Lines Matching refs:dst_cpu

1371 				int src_nid, int dst_cpu)  in should_numa_migrate_memory()  argument
1374 int dst_nid = cpu_to_node(dst_cpu); in should_numa_migrate_memory()
1377 this_cpupid = cpu_pid_to_cpupid(dst_cpu, current->pid); in should_numa_migrate_memory()
1491 int dst_cpu, dst_nid; member
1600 struct rq *rq = cpu_rq(env->dst_cpu); in task_numa_assign()
1603 if (env->best_cpu != env->dst_cpu && xchg(&rq->numa_migrate_on, 1)) { in task_numa_assign()
1605 int start = env->dst_cpu; in task_numa_assign()
1614 env->dst_cpu = cpu; in task_numa_assign()
1615 rq = cpu_rq(env->dst_cpu); in task_numa_assign()
1629 if (env->best_cpu != -1 && env->best_cpu != env->dst_cpu) { in task_numa_assign()
1641 env->best_cpu = env->dst_cpu; in task_numa_assign()
1689 struct rq *dst_rq = cpu_rq(env->dst_cpu); in task_numa_compare()
1826 cpu = env->dst_cpu; in task_numa_compare()
1837 env->dst_cpu = cpu; in task_numa_compare()
1894 env->dst_cpu = env->dst_stats.idle_cpu; in task_numa_find_cpu()
1916 env->dst_cpu = cpu; in task_numa_find_cpu()
6604 static unsigned long cpu_util_next(int cpu, struct task_struct *p, int dst_cpu) in cpu_util_next() argument
6615 if (task_cpu(p) == cpu && dst_cpu != cpu) in cpu_util_next()
6617 else if (task_cpu(p) != cpu && dst_cpu == cpu) in cpu_util_next()
6629 if (dst_cpu == cpu) in cpu_util_next()
6646 compute_energy(struct task_struct *p, int dst_cpu, struct perf_domain *pd) in compute_energy() argument
6666 unsigned long util_freq = cpu_util_next(cpu, p, dst_cpu); in compute_energy()
6679 if (cpu == dst_cpu) { in compute_energy()
7620 int dst_cpu; member
7675 if (!sched_core_cookie_match(cpu_rq(env->dst_cpu), p)) in task_hot()
7705 dst_nid = cpu_to_node(env->dst_cpu); in migrate_degrades_locality()
7763 if (throttled_lb_pair(task_group(p), env->src_cpu, env->dst_cpu)) in can_migrate_task()
7770 if (!cpumask_test_cpu(env->dst_cpu, p->cpus_ptr)) { in can_migrate_task()
7846 set_task_cpu(p, env->dst_cpu); in detach_task()
8590 static bool asym_smt_can_pull_tasks(int dst_cpu, struct sd_lb_stats *sds, in asym_smt_can_pull_tasks() argument
8618 return sched_asym_prefer(dst_cpu, sg->asym_prefer_cpu); in asym_smt_can_pull_tasks()
8629 return sched_asym_prefer(dst_cpu, sg->asym_prefer_cpu); in asym_smt_can_pull_tasks()
8640 return sched_asym_prefer(dst_cpu, sg->asym_prefer_cpu); in asym_smt_can_pull_tasks()
8656 return asym_smt_can_pull_tasks(env->dst_cpu, sds, sgs, group); in sched_asym()
8658 return sched_asym_prefer(env->dst_cpu, group->asym_prefer_cpu); in sched_asym()
8771 (!capacity_greater(capacity_of(env->dst_cpu), sg->sgc->max_capacity) || in update_sd_pick_busiest()
8855 (capacity_greater(sg->sgc->min_capacity, capacity_of(env->dst_cpu)))) in update_sd_pick_busiest()
9230 local_group = cpumask_test_cpu(env->dst_cpu, sched_group_span(sg)); in update_sd_lb_stats()
9237 update_group_capacity(env->sd, env->dst_cpu); in update_sd_lb_stats()
9659 !capacity_greater(capacity_of(env->dst_cpu), capacity) && in find_busiest_queue()
9665 sched_asym_prefer(i, env->dst_cpu) && in find_busiest_queue()
9758 sched_asym_prefer(env->dst_cpu, env->src_cpu); in asym_active_balance()
9797 (capacity_of(env->src_cpu)*sd->imbalance_pct < capacity_of(env->dst_cpu)*100)) in need_active_balance()
9818 if (!cpumask_test_cpu(env->dst_cpu, env->cpus)) in should_we_balance()
9834 return cpu == env->dst_cpu; in should_we_balance()
9838 return group_balance_cpu(sg) == env->dst_cpu; in should_we_balance()
9858 .dst_cpu = this_cpu, in load_balance()
9963 __cpumask_clear_cpu(env.dst_cpu, env.cpus); in load_balance()
9966 env.dst_cpu = env.new_dst_cpu; in load_balance()
10198 .dst_cpu = target_cpu, in active_load_balance_cpu_stop()