summaryrefslogtreecommitdiffstats
path: root/debian/patches-rt/0003-sched-hotplug-Ensure-only-per-cpu-kthreads-run-durin.patch
diff options
context:
space:
mode:
Diffstat (limited to 'debian/patches-rt/0003-sched-hotplug-Ensure-only-per-cpu-kthreads-run-durin.patch')
-rw-r--r--debian/patches-rt/0003-sched-hotplug-Ensure-only-per-cpu-kthreads-run-durin.patch22
1 files changed, 11 insertions, 11 deletions
diff --git a/debian/patches-rt/0003-sched-hotplug-Ensure-only-per-cpu-kthreads-run-durin.patch b/debian/patches-rt/0003-sched-hotplug-Ensure-only-per-cpu-kthreads-run-durin.patch
index 8e9877f29..8d26bbaa9 100644
--- a/debian/patches-rt/0003-sched-hotplug-Ensure-only-per-cpu-kthreads-run-durin.patch
+++ b/debian/patches-rt/0003-sched-hotplug-Ensure-only-per-cpu-kthreads-run-durin.patch
@@ -2,7 +2,7 @@ From: Peter Zijlstra <peterz@infradead.org>
Date: Fri, 23 Oct 2020 12:12:01 +0200
Subject: [PATCH 03/19] sched/hotplug: Ensure only per-cpu kthreads run during
hotplug
-Origin: https://www.kernel.org/pub/linux/kernel/projects/rt/5.10/older/patches-5.10.8-rt24.tar.xz
+Origin: https://www.kernel.org/pub/linux/kernel/projects/rt/5.10/older/patches-5.10.21-rt34.tar.xz
In preparation for migrate_disable(), make sure only per-cpu kthreads
are allowed to run on !active CPUs.
@@ -28,7 +28,7 @@ Signed-off-by: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
--- a/kernel/sched/core.c
+++ b/kernel/sched/core.c
-@@ -3514,8 +3514,10 @@ static inline struct callback_head *spli
+@@ -3511,8 +3511,10 @@ static inline struct callback_head *spli
struct callback_head *head = rq->balance_callback;
lockdep_assert_held(&rq->lock);
@@ -40,7 +40,7 @@ Signed-off-by: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
return head;
}
-@@ -3536,6 +3538,21 @@ static inline void balance_callbacks(str
+@@ -3533,6 +3535,21 @@ static inline void balance_callbacks(str
}
}
@@ -62,7 +62,7 @@ Signed-off-by: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
#else
static inline void __balance_callbacks(struct rq *rq)
-@@ -3551,6 +3568,10 @@ static inline void balance_callbacks(str
+@@ -3548,6 +3565,10 @@ static inline void balance_callbacks(str
{
}
@@ -73,7 +73,7 @@ Signed-off-by: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
#endif
static inline void
-@@ -3578,7 +3599,7 @@ static inline void finish_lock_switch(st
+@@ -3575,7 +3596,7 @@ static inline void finish_lock_switch(st
* prev into current:
*/
spin_acquire(&rq->lock.dep_map, 0, 0, _THIS_IP_);
@@ -82,7 +82,7 @@ Signed-off-by: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
raw_spin_unlock_irq(&rq->lock);
}
-@@ -6834,6 +6855,90 @@ static void migrate_tasks(struct rq *dea
+@@ -6831,6 +6852,90 @@ static void migrate_tasks(struct rq *dea
rq->stop = stop;
}
@@ -173,7 +173,7 @@ Signed-off-by: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
#endif /* CONFIG_HOTPLUG_CPU */
void set_rq_online(struct rq *rq)
-@@ -6919,6 +7024,8 @@ int sched_cpu_activate(unsigned int cpu)
+@@ -6916,6 +7021,8 @@ int sched_cpu_activate(unsigned int cpu)
struct rq *rq = cpu_rq(cpu);
struct rq_flags rf;
@@ -182,7 +182,7 @@ Signed-off-by: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
#ifdef CONFIG_SCHED_SMT
/*
* When going up, increment the number of cores with SMT present.
-@@ -6966,6 +7073,8 @@ int sched_cpu_deactivate(unsigned int cp
+@@ -6963,6 +7070,8 @@ int sched_cpu_deactivate(unsigned int cp
*/
synchronize_rcu();
@@ -191,7 +191,7 @@ Signed-off-by: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
#ifdef CONFIG_SCHED_SMT
/*
* When going down, decrement the number of cores with SMT present.
-@@ -6979,6 +7088,7 @@ int sched_cpu_deactivate(unsigned int cp
+@@ -6976,6 +7085,7 @@ int sched_cpu_deactivate(unsigned int cp
ret = cpuset_cpu_inactive(cpu);
if (ret) {
@@ -209,7 +209,7 @@ Signed-off-by: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
unsigned char nohz_idle_balance;
unsigned char idle_balance;
-@@ -1379,6 +1380,9 @@ init_numa_balancing(unsigned long clone_
+@@ -1380,6 +1381,9 @@ init_numa_balancing(unsigned long clone_
#ifdef CONFIG_SMP
@@ -219,7 +219,7 @@ Signed-off-by: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
static inline void
queue_balance_callback(struct rq *rq,
struct callback_head *head,
-@@ -1386,12 +1390,13 @@ queue_balance_callback(struct rq *rq,
+@@ -1387,12 +1391,13 @@ queue_balance_callback(struct rq *rq,
{
lockdep_assert_held(&rq->lock);