hc
2024-02-20 102a0743326a03cd1a1202ceda21e175b7d3575c
kernel/lib/atomic64.c
....@@ -1,13 +1,9 @@
1
+// SPDX-License-Identifier: GPL-2.0-or-later
12 /*
23 * Generic implementation of 64-bit atomics using spinlocks,
34 * useful on processors that don't have 64-bit atomic instructions.
45 *
56 * Copyright © 2009 Paul Mackerras, IBM Corp. <paulus@au1.ibm.com>
6
- *
7
- * This program is free software; you can redistribute it and/or
8
- * modify it under the terms of the GNU General Public License
9
- * as published by the Free Software Foundation; either version
10
- * 2 of the License, or (at your option) any later version.
117 */
128 #include <linux/types.h>
139 #include <linux/cache.h>
....@@ -46,11 +42,11 @@
4642 return &atomic64_lock[addr & (NR_LOCKS - 1)].lock;
4743 }
4844
49
-long long atomic64_read(const atomic64_t *v)
45
+s64 atomic64_read(const atomic64_t *v)
5046 {
5147 unsigned long flags;
5248 raw_spinlock_t *lock = lock_addr(v);
53
- long long val;
49
+ s64 val;
5450
5551 raw_spin_lock_irqsave(lock, flags);
5652 val = v->counter;
....@@ -59,7 +55,7 @@
5955 }
6056 EXPORT_SYMBOL(atomic64_read);
6157
62
-void atomic64_set(atomic64_t *v, long long i)
58
+void atomic64_set(atomic64_t *v, s64 i)
6359 {
6460 unsigned long flags;
6561 raw_spinlock_t *lock = lock_addr(v);
....@@ -71,7 +67,7 @@
7167 EXPORT_SYMBOL(atomic64_set);
7268
7369 #define ATOMIC64_OP(op, c_op) \
74
-void atomic64_##op(long long a, atomic64_t *v) \
70
+void atomic64_##op(s64 a, atomic64_t *v) \
7571 { \
7672 unsigned long flags; \
7773 raw_spinlock_t *lock = lock_addr(v); \
....@@ -83,11 +79,11 @@
8379 EXPORT_SYMBOL(atomic64_##op);
8480
8581 #define ATOMIC64_OP_RETURN(op, c_op) \
86
-long long atomic64_##op##_return(long long a, atomic64_t *v) \
82
+s64 atomic64_##op##_return(s64 a, atomic64_t *v) \
8783 { \
8884 unsigned long flags; \
8985 raw_spinlock_t *lock = lock_addr(v); \
90
- long long val; \
86
+ s64 val; \
9187 \
9288 raw_spin_lock_irqsave(lock, flags); \
9389 val = (v->counter c_op a); \
....@@ -97,11 +93,11 @@
9793 EXPORT_SYMBOL(atomic64_##op##_return);
9894
9995 #define ATOMIC64_FETCH_OP(op, c_op) \
100
-long long atomic64_fetch_##op(long long a, atomic64_t *v) \
96
+s64 atomic64_fetch_##op(s64 a, atomic64_t *v) \
10197 { \
10298 unsigned long flags; \
10399 raw_spinlock_t *lock = lock_addr(v); \
104
- long long val; \
100
+ s64 val; \
105101 \
106102 raw_spin_lock_irqsave(lock, flags); \
107103 val = v->counter; \
....@@ -134,11 +130,11 @@
134130 #undef ATOMIC64_OP_RETURN
135131 #undef ATOMIC64_OP
136132
137
-long long atomic64_dec_if_positive(atomic64_t *v)
133
+s64 atomic64_dec_if_positive(atomic64_t *v)
138134 {
139135 unsigned long flags;
140136 raw_spinlock_t *lock = lock_addr(v);
141
- long long val;
137
+ s64 val;
142138
143139 raw_spin_lock_irqsave(lock, flags);
144140 val = v->counter - 1;
....@@ -149,11 +145,11 @@
149145 }
150146 EXPORT_SYMBOL(atomic64_dec_if_positive);
151147
152
-long long atomic64_cmpxchg(atomic64_t *v, long long o, long long n)
148
+s64 atomic64_cmpxchg(atomic64_t *v, s64 o, s64 n)
153149 {
154150 unsigned long flags;
155151 raw_spinlock_t *lock = lock_addr(v);
156
- long long val;
152
+ s64 val;
157153
158154 raw_spin_lock_irqsave(lock, flags);
159155 val = v->counter;
....@@ -164,11 +160,11 @@
164160 }
165161 EXPORT_SYMBOL(atomic64_cmpxchg);
166162
167
-long long atomic64_xchg(atomic64_t *v, long long new)
163
+s64 atomic64_xchg(atomic64_t *v, s64 new)
168164 {
169165 unsigned long flags;
170166 raw_spinlock_t *lock = lock_addr(v);
171
- long long val;
167
+ s64 val;
172168
173169 raw_spin_lock_irqsave(lock, flags);
174170 val = v->counter;
....@@ -178,11 +174,11 @@
178174 }
179175 EXPORT_SYMBOL(atomic64_xchg);
180176
181
-long long atomic64_fetch_add_unless(atomic64_t *v, long long a, long long u)
177
+s64 atomic64_fetch_add_unless(atomic64_t *v, s64 a, s64 u)
182178 {
183179 unsigned long flags;
184180 raw_spinlock_t *lock = lock_addr(v);
185
- long long val;
181
+ s64 val;
186182
187183 raw_spin_lock_irqsave(lock, flags);
188184 val = v->counter;