hc
2024-12-19 9370bb92b2d16684ee45cf24e879c93c509162da
kernel/arch/arm/include/asm/switch_to.h
....@@ -4,20 +4,13 @@
44
55 #include <linux/thread_info.h>
66
7
-#if defined CONFIG_PREEMPT_RT_FULL && defined CONFIG_HIGHMEM
8
-void switch_kmaps(struct task_struct *prev_p, struct task_struct *next_p);
9
-#else
10
-static inline void
11
-switch_kmaps(struct task_struct *prev_p, struct task_struct *next_p) { }
12
-#endif
13
-
147 /*
158 * For v7 SMP cores running a preemptible kernel we may be pre-empted
169 * during a TLB maintenance operation, so execute an inner-shareable dsb
1710 * to ensure that the maintenance completes in case we migrate to another
1811 * CPU.
1912 */
20
-#if defined(CONFIG_PREEMPT) && defined(CONFIG_SMP) && defined(CONFIG_CPU_V7)
13
+#if defined(CONFIG_PREEMPTION) && defined(CONFIG_SMP) && defined(CONFIG_CPU_V7)
2114 #define __complete_pending_tlbi() dsb(ish)
2215 #else
2316 #define __complete_pending_tlbi()
....@@ -33,7 +26,6 @@
3326 #define switch_to(prev,next,last) \
3427 do { \
3528 __complete_pending_tlbi(); \
36
- switch_kmaps(prev, next); \
3729 last = __switch_to(prev,task_thread_info(prev), task_thread_info(next)); \
3830 } while (0)
3931