| .. | .. |
|---|
| 74 | 74 | { |
|---|
| 75 | 75 | int owner; |
|---|
| 76 | 76 | |
|---|
| 77 | | - asm volatile( |
|---|
| 77 | + asm_inline volatile( |
|---|
| 78 | 78 | ALTERNATIVE("", ".long 0xb2fa0040", 49) /* NIAI 4 */ |
|---|
| 79 | 79 | " l %0,%1\n" |
|---|
| 80 | 80 | : "=d" (owner) : "Q" (*lock) : "memory"); |
|---|
| .. | .. |
|---|
| 85 | 85 | { |
|---|
| 86 | 86 | int expected = old; |
|---|
| 87 | 87 | |
|---|
| 88 | | - asm volatile( |
|---|
| 88 | + asm_inline volatile( |
|---|
| 89 | 89 | ALTERNATIVE("", ".long 0xb2fa0080", 49) /* NIAI 8 */ |
|---|
| 90 | 90 | " cs %0,%3,%1\n" |
|---|
| 91 | 91 | : "=d" (old), "=Q" (*lock) |
|---|
| .. | .. |
|---|
| 242 | 242 | |
|---|
| 243 | 243 | void arch_spin_lock_wait(arch_spinlock_t *lp) |
|---|
| 244 | 244 | { |
|---|
| 245 | | - /* Use classic spinlocks + niai if the steal time is >= 10% */ |
|---|
| 246 | 245 | if (test_cpu_flag(CIF_DEDICATED_CPU)) |
|---|
| 247 | 246 | arch_spin_lock_queued(lp); |
|---|
| 248 | 247 | else |
|---|