.. | .. |
---|
74 | 74 | { |
---|
75 | 75 | int owner; |
---|
76 | 76 | |
---|
77 | | - asm volatile( |
---|
| 77 | + asm_inline volatile( |
---|
78 | 78 | ALTERNATIVE("", ".long 0xb2fa0040", 49) /* NIAI 4 */ |
---|
79 | 79 | " l %0,%1\n" |
---|
80 | 80 | : "=d" (owner) : "Q" (*lock) : "memory"); |
---|
.. | .. |
---|
85 | 85 | { |
---|
86 | 86 | int expected = old; |
---|
87 | 87 | |
---|
88 | | - asm volatile( |
---|
| 88 | + asm_inline volatile( |
---|
89 | 89 | ALTERNATIVE("", ".long 0xb2fa0080", 49) /* NIAI 8 */ |
---|
90 | 90 | " cs %0,%3,%1\n" |
---|
91 | 91 | : "=d" (old), "=Q" (*lock) |
---|
.. | .. |
---|
242 | 242 | |
---|
243 | 243 | void arch_spin_lock_wait(arch_spinlock_t *lp) |
---|
244 | 244 | { |
---|
245 | | - /* Use classic spinlocks + niai if the steal time is >= 10% */ |
---|
246 | 245 | if (test_cpu_flag(CIF_DEDICATED_CPU)) |
---|
247 | 246 | arch_spin_lock_queued(lp); |
---|
248 | 247 | else |
---|