.. | .. |
---|
| 1 | +// SPDX-License-Identifier: GPL-2.0-or-later |
---|
1 | 2 | /* |
---|
2 | 3 | * Generic implementation of 64-bit atomics using spinlocks, |
---|
3 | 4 | * useful on processors that don't have 64-bit atomic instructions. |
---|
4 | 5 | * |
---|
5 | 6 | * Copyright © 2009 Paul Mackerras, IBM Corp. <paulus@au1.ibm.com> |
---|
6 | | - * |
---|
7 | | - * This program is free software; you can redistribute it and/or |
---|
8 | | - * modify it under the terms of the GNU General Public License |
---|
9 | | - * as published by the Free Software Foundation; either version |
---|
10 | | - * 2 of the License, or (at your option) any later version. |
---|
11 | 7 | */ |
---|
12 | 8 | #include <linux/types.h> |
---|
13 | 9 | #include <linux/cache.h> |
---|
.. | .. |
---|
46 | 42 | return &atomic64_lock[addr & (NR_LOCKS - 1)].lock; |
---|
47 | 43 | } |
---|
48 | 44 | |
---|
49 | | -long long atomic64_read(const atomic64_t *v) |
---|
| 45 | +s64 atomic64_read(const atomic64_t *v) |
---|
50 | 46 | { |
---|
51 | 47 | unsigned long flags; |
---|
52 | 48 | raw_spinlock_t *lock = lock_addr(v); |
---|
53 | | - long long val; |
---|
| 49 | + s64 val; |
---|
54 | 50 | |
---|
55 | 51 | raw_spin_lock_irqsave(lock, flags); |
---|
56 | 52 | val = v->counter; |
---|
.. | .. |
---|
59 | 55 | } |
---|
60 | 56 | EXPORT_SYMBOL(atomic64_read); |
---|
61 | 57 | |
---|
62 | | -void atomic64_set(atomic64_t *v, long long i) |
---|
| 58 | +void atomic64_set(atomic64_t *v, s64 i) |
---|
63 | 59 | { |
---|
64 | 60 | unsigned long flags; |
---|
65 | 61 | raw_spinlock_t *lock = lock_addr(v); |
---|
.. | .. |
---|
71 | 67 | EXPORT_SYMBOL(atomic64_set); |
---|
72 | 68 | |
---|
73 | 69 | #define ATOMIC64_OP(op, c_op) \ |
---|
74 | | -void atomic64_##op(long long a, atomic64_t *v) \ |
---|
| 70 | +void atomic64_##op(s64 a, atomic64_t *v) \ |
---|
75 | 71 | { \ |
---|
76 | 72 | unsigned long flags; \ |
---|
77 | 73 | raw_spinlock_t *lock = lock_addr(v); \ |
---|
.. | .. |
---|
83 | 79 | EXPORT_SYMBOL(atomic64_##op); |
---|
84 | 80 | |
---|
85 | 81 | #define ATOMIC64_OP_RETURN(op, c_op) \ |
---|
86 | | -long long atomic64_##op##_return(long long a, atomic64_t *v) \ |
---|
| 82 | +s64 atomic64_##op##_return(s64 a, atomic64_t *v) \ |
---|
87 | 83 | { \ |
---|
88 | 84 | unsigned long flags; \ |
---|
89 | 85 | raw_spinlock_t *lock = lock_addr(v); \ |
---|
90 | | - long long val; \ |
---|
| 86 | + s64 val; \ |
---|
91 | 87 | \ |
---|
92 | 88 | raw_spin_lock_irqsave(lock, flags); \ |
---|
93 | 89 | val = (v->counter c_op a); \ |
---|
.. | .. |
---|
97 | 93 | EXPORT_SYMBOL(atomic64_##op##_return); |
---|
98 | 94 | |
---|
99 | 95 | #define ATOMIC64_FETCH_OP(op, c_op) \ |
---|
100 | | -long long atomic64_fetch_##op(long long a, atomic64_t *v) \ |
---|
| 96 | +s64 atomic64_fetch_##op(s64 a, atomic64_t *v) \ |
---|
101 | 97 | { \ |
---|
102 | 98 | unsigned long flags; \ |
---|
103 | 99 | raw_spinlock_t *lock = lock_addr(v); \ |
---|
104 | | - long long val; \ |
---|
| 100 | + s64 val; \ |
---|
105 | 101 | \ |
---|
106 | 102 | raw_spin_lock_irqsave(lock, flags); \ |
---|
107 | 103 | val = v->counter; \ |
---|
.. | .. |
---|
134 | 130 | #undef ATOMIC64_OP_RETURN |
---|
135 | 131 | #undef ATOMIC64_OP |
---|
136 | 132 | |
---|
137 | | -long long atomic64_dec_if_positive(atomic64_t *v) |
---|
| 133 | +s64 atomic64_dec_if_positive(atomic64_t *v) |
---|
138 | 134 | { |
---|
139 | 135 | unsigned long flags; |
---|
140 | 136 | raw_spinlock_t *lock = lock_addr(v); |
---|
141 | | - long long val; |
---|
| 137 | + s64 val; |
---|
142 | 138 | |
---|
143 | 139 | raw_spin_lock_irqsave(lock, flags); |
---|
144 | 140 | val = v->counter - 1; |
---|
.. | .. |
---|
149 | 145 | } |
---|
150 | 146 | EXPORT_SYMBOL(atomic64_dec_if_positive); |
---|
151 | 147 | |
---|
152 | | -long long atomic64_cmpxchg(atomic64_t *v, long long o, long long n) |
---|
| 148 | +s64 atomic64_cmpxchg(atomic64_t *v, s64 o, s64 n) |
---|
153 | 149 | { |
---|
154 | 150 | unsigned long flags; |
---|
155 | 151 | raw_spinlock_t *lock = lock_addr(v); |
---|
156 | | - long long val; |
---|
| 152 | + s64 val; |
---|
157 | 153 | |
---|
158 | 154 | raw_spin_lock_irqsave(lock, flags); |
---|
159 | 155 | val = v->counter; |
---|
.. | .. |
---|
164 | 160 | } |
---|
165 | 161 | EXPORT_SYMBOL(atomic64_cmpxchg); |
---|
166 | 162 | |
---|
167 | | -long long atomic64_xchg(atomic64_t *v, long long new) |
---|
| 163 | +s64 atomic64_xchg(atomic64_t *v, s64 new) |
---|
168 | 164 | { |
---|
169 | 165 | unsigned long flags; |
---|
170 | 166 | raw_spinlock_t *lock = lock_addr(v); |
---|
171 | | - long long val; |
---|
| 167 | + s64 val; |
---|
172 | 168 | |
---|
173 | 169 | raw_spin_lock_irqsave(lock, flags); |
---|
174 | 170 | val = v->counter; |
---|
.. | .. |
---|
178 | 174 | } |
---|
179 | 175 | EXPORT_SYMBOL(atomic64_xchg); |
---|
180 | 176 | |
---|
181 | | -long long atomic64_fetch_add_unless(atomic64_t *v, long long a, long long u) |
---|
| 177 | +s64 atomic64_fetch_add_unless(atomic64_t *v, s64 a, s64 u) |
---|
182 | 178 | { |
---|
183 | 179 | unsigned long flags; |
---|
184 | 180 | raw_spinlock_t *lock = lock_addr(v); |
---|
185 | | - long long val; |
---|
| 181 | + s64 val; |
---|
186 | 182 | |
---|
187 | 183 | raw_spin_lock_irqsave(lock, flags); |
---|
188 | 184 | val = v->counter; |
---|