Skip to content

Commit

Permalink
mm: numa: Revert temporarily disabling of NUMA migration
Browse files Browse the repository at this point in the history
With the scan rate code working (at least for multi-instance specjbb),
the large hammer that is "sched: Do not migrate memory immediately after
switching node" can be replaced with something smarter. Revert temporarily
migration disabling and all traces of numa_migrate_seq.

Signed-off-by: Rik van Riel <[email protected]>
Signed-off-by: Mel Gorman <[email protected]>
Cc: Andrea Arcangeli <[email protected]>
Cc: Johannes Weiner <[email protected]>
Cc: Srikar Dronamraju <[email protected]>
Signed-off-by: Peter Zijlstra <[email protected]>
Link: http://lkml.kernel.org/r/[email protected]
Signed-off-by: Ingo Molnar <[email protected]>
  • Loading branch information
Rik van Riel authored and Ingo Molnar committed Oct 9, 2013
1 parent 930aa17 commit 1e3646f
Show file tree
Hide file tree
Showing 4 changed files with 1 addition and 39 deletions.
1 change: 0 additions & 1 deletion include/linux/sched.h
Original file line number Diff line number Diff line change
Expand Up @@ -1340,7 +1340,6 @@ struct task_struct {
#endif
#ifdef CONFIG_NUMA_BALANCING
int numa_scan_seq;
int numa_migrate_seq;
unsigned int numa_scan_period;
unsigned int numa_scan_period_max;
unsigned long numa_migrate_retry;
Expand Down
2 changes: 0 additions & 2 deletions kernel/sched/core.c
Original file line number Diff line number Diff line change
Expand Up @@ -1731,7 +1731,6 @@ static void __sched_fork(unsigned long clone_flags, struct task_struct *p)

p->node_stamp = 0ULL;
p->numa_scan_seq = p->mm ? p->mm->numa_scan_seq : 0;
p->numa_migrate_seq = 1;
p->numa_scan_period = sysctl_numa_balancing_scan_delay;
p->numa_work.next = &p->numa_work;
p->numa_faults = NULL;
Expand Down Expand Up @@ -4488,7 +4487,6 @@ void sched_setnuma(struct task_struct *p, int nid)
p->sched_class->put_prev_task(rq, p);

p->numa_preferred_nid = nid;
p->numa_migrate_seq = 1;

if (running)
p->sched_class->set_curr_task(rq);
Expand Down
25 changes: 1 addition & 24 deletions kernel/sched/fair.c
Original file line number Diff line number Diff line change
Expand Up @@ -1261,16 +1261,8 @@ static void numa_migrate_preferred(struct task_struct *p)
{
/* Success if task is already running on preferred CPU */
p->numa_migrate_retry = 0;
if (cpu_to_node(task_cpu(p)) == p->numa_preferred_nid) {
/*
* If migration is temporarily disabled due to a task migration
* then re-enable it now as the task is running on its
* preferred node and memory should migrate locally
*/
if (!p->numa_migrate_seq)
p->numa_migrate_seq++;
if (cpu_to_node(task_cpu(p)) == p->numa_preferred_nid)
return;
}

/* This task has no NUMA fault statistics yet */
if (unlikely(p->numa_preferred_nid == -1))
Expand Down Expand Up @@ -1367,7 +1359,6 @@ static void task_numa_placement(struct task_struct *p)
if (p->numa_scan_seq == seq)
return;
p->numa_scan_seq = seq;
p->numa_migrate_seq++;
p->numa_scan_period_max = task_scan_max(p);

/* If the task is part of a group prevent parallel updates to group stats */
Expand Down Expand Up @@ -4730,20 +4721,6 @@ static void move_task(struct task_struct *p, struct lb_env *env)
set_task_cpu(p, env->dst_cpu);
activate_task(env->dst_rq, p, 0);
check_preempt_curr(env->dst_rq, p, 0);
#ifdef CONFIG_NUMA_BALANCING
if (p->numa_preferred_nid != -1) {
int src_nid = cpu_to_node(env->src_cpu);
int dst_nid = cpu_to_node(env->dst_cpu);

/*
* If the load balancer has moved the task then limit
* migrations from taking place in the short term in
* case this is a short-lived migration.
*/
if (src_nid != dst_nid && dst_nid != p->numa_preferred_nid)
p->numa_migrate_seq = 0;
}
#endif
}

/*
Expand Down
12 changes: 0 additions & 12 deletions mm/mempolicy.c
Original file line number Diff line number Diff line change
Expand Up @@ -2404,18 +2404,6 @@ int mpol_misplaced(struct page *page, struct vm_area_struct *vma, unsigned long
last_cpupid = page_cpupid_xchg_last(page, this_cpupid);
if (!cpupid_pid_unset(last_cpupid) && cpupid_to_nid(last_cpupid) != thisnid)
goto out;

#ifdef CONFIG_NUMA_BALANCING
/*
* If the scheduler has just moved us away from our
* preferred node, do not bother migrating pages yet.
* This way a short and temporary process migration will
* not cause excessive memory migration.
*/
if (thisnid != current->numa_preferred_nid &&
!current->numa_migrate_seq)
goto out;
#endif
}

if (curnid != polnid)
Expand Down

0 comments on commit 1e3646f

Please sign in to comment.