| .. | .. |
|---|
| 4 | 4 | |
|---|
| 5 | 5 | #include <linux/thread_info.h> |
|---|
| 6 | 6 | |
|---|
| 7 | | -#if defined CONFIG_PREEMPT_RT_FULL && defined CONFIG_HIGHMEM |
|---|
| 8 | | -void switch_kmaps(struct task_struct *prev_p, struct task_struct *next_p); |
|---|
| 9 | | -#else |
|---|
| 10 | | -static inline void |
|---|
| 11 | | -switch_kmaps(struct task_struct *prev_p, struct task_struct *next_p) { } |
|---|
| 12 | | -#endif |
|---|
| 13 | | - |
|---|
| 14 | 7 | /* |
|---|
| 15 | 8 | * For v7 SMP cores running a preemptible kernel we may be pre-empted |
|---|
| 16 | 9 | * during a TLB maintenance operation, so execute an inner-shareable dsb |
|---|
| 17 | 10 | * to ensure that the maintenance completes in case we migrate to another |
|---|
| 18 | 11 | * CPU. |
|---|
| 19 | 12 | */ |
|---|
| 20 | | -#if defined(CONFIG_PREEMPT) && defined(CONFIG_SMP) && defined(CONFIG_CPU_V7) |
|---|
| 13 | +#if defined(CONFIG_PREEMPTION) && defined(CONFIG_SMP) && defined(CONFIG_CPU_V7) |
|---|
| 21 | 14 | #define __complete_pending_tlbi() dsb(ish) |
|---|
| 22 | 15 | #else |
|---|
| 23 | 16 | #define __complete_pending_tlbi() |
|---|
| .. | .. |
|---|
| 33 | 26 | #define switch_to(prev,next,last) \ |
|---|
| 34 | 27 | do { \ |
|---|
| 35 | 28 | __complete_pending_tlbi(); \ |
|---|
| 36 | | - switch_kmaps(prev, next); \ |
|---|
| 37 | 29 | last = __switch_to(prev,task_thread_info(prev), task_thread_info(next)); \ |
|---|
| 38 | 30 | } while (0) |
|---|
| 39 | 31 | |
|---|