hc
2023-12-11 d2ccde1c8e90d38cee87a1b0309ad2827f3fd30d
kernel/kernel/locking/spinlock.c
....@@ -22,6 +22,13 @@
2222 #include <linux/debug_locks.h>
2323 #include <linux/export.h>
2424
25
+#ifdef CONFIG_MMIOWB
26
+#ifndef arch_mmiowb_state
27
+DEFINE_PER_CPU(struct mmiowb_state, __mmiowb_state);
28
+EXPORT_PER_CPU_SYMBOL(__mmiowb_state);
29
+#endif
30
+#endif
31
+
2532 /*
2633 * If lockdep is enabled then we use the non-preemption spin-ops
2734 * even on CONFIG_PREEMPT, because lockdep assumes that interrupts are
....@@ -118,7 +125,7 @@
118125 */
119126 BUILD_LOCK_OPS(spin, raw_spinlock);
120127
121
-#ifndef CONFIG_PREEMPT_RT_FULL
128
+#ifndef CONFIG_PREEMPT_RT
122129 BUILD_LOCK_OPS(read, rwlock);
123130 BUILD_LOCK_OPS(write, rwlock);
124131 #endif
....@@ -205,7 +212,7 @@
205212 EXPORT_SYMBOL(_raw_spin_unlock_bh);
206213 #endif
207214
208
-#ifndef CONFIG_PREEMPT_RT_FULL
215
+#ifndef CONFIG_PREEMPT_RT
209216
210217 #ifndef CONFIG_INLINE_READ_TRYLOCK
211218 int __lockfunc _raw_read_trylock(rwlock_t *lock)
....@@ -351,7 +358,7 @@
351358 EXPORT_SYMBOL(_raw_write_unlock_bh);
352359 #endif
353360
354
-#endif /* !PREEMPT_RT_FULL */
361
+#endif /* !PREEMPT_RT */
355362
356363 #ifdef CONFIG_DEBUG_LOCK_ALLOC
357364