Thomas Gleixner | caab277 | 2019-06-03 07:44:50 +0200 | [diff] [blame] | 1 | /* SPDX-License-Identifier: GPL-2.0-only */ |
Catalin Marinas | 6170a97 | 2012-03-05 11:49:29 +0000 | [diff] [blame] | 2 | /* |
| 3 | * Based on arch/arm/include/asm/atomic.h |
| 4 | * |
| 5 | * Copyright (C) 1996 Russell King. |
| 6 | * Copyright (C) 2002 Deep Blue Solutions Ltd. |
| 7 | * Copyright (C) 2012 ARM Ltd. |
Catalin Marinas | 6170a97 | 2012-03-05 11:49:29 +0000 | [diff] [blame] | 8 | */ |
| 9 | #ifndef __ASM_ATOMIC_H |
| 10 | #define __ASM_ATOMIC_H |
| 11 | |
| 12 | #include <linux/compiler.h> |
| 13 | #include <linux/types.h> |
| 14 | |
| 15 | #include <asm/barrier.h> |
Will Deacon | c342f78 | 2015-04-23 20:08:49 +0100 | [diff] [blame] | 16 | #include <asm/cmpxchg.h> |
Catalin Marinas | 6170a97 | 2012-03-05 11:49:29 +0000 | [diff] [blame] | 17 | #include <asm/lse.h> |
| 18 | |
Will Deacon | 0533f97 | 2019-08-29 11:49:10 +0100 | [diff] [blame] | 19 | #define ATOMIC_OP(op) \ |
| 20 | static inline void arch_##op(int i, atomic_t *v) \ |
| 21 | { \ |
| 22 | __lse_ll_sc_body(op, i, v); \ |
| 23 | } |
Will Deacon | c0385b2 | 2015-02-03 12:39:03 +0000 | [diff] [blame] | 24 | |
Will Deacon | 0533f97 | 2019-08-29 11:49:10 +0100 | [diff] [blame] | 25 | ATOMIC_OP(atomic_andnot) |
| 26 | ATOMIC_OP(atomic_or) |
| 27 | ATOMIC_OP(atomic_xor) |
| 28 | ATOMIC_OP(atomic_add) |
| 29 | ATOMIC_OP(atomic_and) |
| 30 | ATOMIC_OP(atomic_sub) |
Will Deacon | c0385b2 | 2015-02-03 12:39:03 +0000 | [diff] [blame] | 31 | |
Will Deacon | 5aad6cd | 2019-08-29 14:33:23 +0100 | [diff] [blame] | 32 | #undef ATOMIC_OP |
Will Deacon | c275f76 | 2015-02-03 11:26:53 +0000 | [diff] [blame] | 33 | |
Will Deacon | 0533f97 | 2019-08-29 11:49:10 +0100 | [diff] [blame] | 34 | #define ATOMIC_FETCH_OP(name, op) \ |
| 35 | static inline int arch_##op##name(int i, atomic_t *v) \ |
| 36 | { \ |
| 37 | return __lse_ll_sc_body(op##name, i, v); \ |
| 38 | } |
| 39 | |
| 40 | #define ATOMIC_FETCH_OPS(op) \ |
| 41 | ATOMIC_FETCH_OP(_relaxed, op) \ |
| 42 | ATOMIC_FETCH_OP(_acquire, op) \ |
| 43 | ATOMIC_FETCH_OP(_release, op) \ |
| 44 | ATOMIC_FETCH_OP( , op) |
| 45 | |
| 46 | ATOMIC_FETCH_OPS(atomic_fetch_andnot) |
| 47 | ATOMIC_FETCH_OPS(atomic_fetch_or) |
| 48 | ATOMIC_FETCH_OPS(atomic_fetch_xor) |
| 49 | ATOMIC_FETCH_OPS(atomic_fetch_add) |
| 50 | ATOMIC_FETCH_OPS(atomic_fetch_and) |
| 51 | ATOMIC_FETCH_OPS(atomic_fetch_sub) |
| 52 | ATOMIC_FETCH_OPS(atomic_add_return) |
| 53 | ATOMIC_FETCH_OPS(atomic_sub_return) |
| 54 | |
Will Deacon | 5aad6cd | 2019-08-29 14:33:23 +0100 | [diff] [blame] | 55 | #undef ATOMIC_FETCH_OP |
| 56 | #undef ATOMIC_FETCH_OPS |
Will Deacon | 0533f97 | 2019-08-29 11:49:10 +0100 | [diff] [blame] | 57 | |
| 58 | #define ATOMIC64_OP(op) \ |
| 59 | static inline void arch_##op(long i, atomic64_t *v) \ |
| 60 | { \ |
| 61 | __lse_ll_sc_body(op, i, v); \ |
| 62 | } |
| 63 | |
| 64 | ATOMIC64_OP(atomic64_andnot) |
| 65 | ATOMIC64_OP(atomic64_or) |
| 66 | ATOMIC64_OP(atomic64_xor) |
| 67 | ATOMIC64_OP(atomic64_add) |
| 68 | ATOMIC64_OP(atomic64_and) |
| 69 | ATOMIC64_OP(atomic64_sub) |
| 70 | |
Will Deacon | 5aad6cd | 2019-08-29 14:33:23 +0100 | [diff] [blame] | 71 | #undef ATOMIC64_OP |
Will Deacon | 0533f97 | 2019-08-29 11:49:10 +0100 | [diff] [blame] | 72 | |
| 73 | #define ATOMIC64_FETCH_OP(name, op) \ |
| 74 | static inline long arch_##op##name(long i, atomic64_t *v) \ |
| 75 | { \ |
| 76 | return __lse_ll_sc_body(op##name, i, v); \ |
| 77 | } |
| 78 | |
| 79 | #define ATOMIC64_FETCH_OPS(op) \ |
| 80 | ATOMIC64_FETCH_OP(_relaxed, op) \ |
| 81 | ATOMIC64_FETCH_OP(_acquire, op) \ |
| 82 | ATOMIC64_FETCH_OP(_release, op) \ |
| 83 | ATOMIC64_FETCH_OP( , op) |
| 84 | |
| 85 | ATOMIC64_FETCH_OPS(atomic64_fetch_andnot) |
| 86 | ATOMIC64_FETCH_OPS(atomic64_fetch_or) |
| 87 | ATOMIC64_FETCH_OPS(atomic64_fetch_xor) |
| 88 | ATOMIC64_FETCH_OPS(atomic64_fetch_add) |
| 89 | ATOMIC64_FETCH_OPS(atomic64_fetch_and) |
| 90 | ATOMIC64_FETCH_OPS(atomic64_fetch_sub) |
| 91 | ATOMIC64_FETCH_OPS(atomic64_add_return) |
| 92 | ATOMIC64_FETCH_OPS(atomic64_sub_return) |
| 93 | |
Will Deacon | 5aad6cd | 2019-08-29 14:33:23 +0100 | [diff] [blame] | 94 | #undef ATOMIC64_FETCH_OP |
| 95 | #undef ATOMIC64_FETCH_OPS |
| 96 | |
Will Deacon | 0533f97 | 2019-08-29 11:49:10 +0100 | [diff] [blame] | 97 | static inline long arch_atomic64_dec_if_positive(atomic64_t *v) |
| 98 | { |
| 99 | return __lse_ll_sc_body(atomic64_dec_if_positive, v); |
| 100 | } |
Will Deacon | c342f78 | 2015-04-23 20:08:49 +0100 | [diff] [blame] | 101 | |
Peter Zijlstra | 37f8173 | 2020-01-24 22:13:03 +0100 | [diff] [blame] | 102 | #define arch_atomic_read(v) __READ_ONCE((v)->counter) |
| 103 | #define arch_atomic_set(v, i) __WRITE_ONCE(((v)->counter), (i)) |
Will Deacon | 305d454 | 2015-10-08 20:15:18 +0100 | [diff] [blame] | 104 | |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 105 | #define arch_atomic_add_return_relaxed arch_atomic_add_return_relaxed |
| 106 | #define arch_atomic_add_return_acquire arch_atomic_add_return_acquire |
| 107 | #define arch_atomic_add_return_release arch_atomic_add_return_release |
| 108 | #define arch_atomic_add_return arch_atomic_add_return |
Will Deacon | 305d454 | 2015-10-08 20:15:18 +0100 | [diff] [blame] | 109 | |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 110 | #define arch_atomic_sub_return_relaxed arch_atomic_sub_return_relaxed |
| 111 | #define arch_atomic_sub_return_acquire arch_atomic_sub_return_acquire |
| 112 | #define arch_atomic_sub_return_release arch_atomic_sub_return_release |
| 113 | #define arch_atomic_sub_return arch_atomic_sub_return |
Will Deacon | 305d454 | 2015-10-08 20:15:18 +0100 | [diff] [blame] | 114 | |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 115 | #define arch_atomic_fetch_add_relaxed arch_atomic_fetch_add_relaxed |
| 116 | #define arch_atomic_fetch_add_acquire arch_atomic_fetch_add_acquire |
| 117 | #define arch_atomic_fetch_add_release arch_atomic_fetch_add_release |
| 118 | #define arch_atomic_fetch_add arch_atomic_fetch_add |
Peter Zijlstra | e490f9b | 2016-04-18 01:16:09 +0200 | [diff] [blame] | 119 | |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 120 | #define arch_atomic_fetch_sub_relaxed arch_atomic_fetch_sub_relaxed |
| 121 | #define arch_atomic_fetch_sub_acquire arch_atomic_fetch_sub_acquire |
| 122 | #define arch_atomic_fetch_sub_release arch_atomic_fetch_sub_release |
| 123 | #define arch_atomic_fetch_sub arch_atomic_fetch_sub |
Peter Zijlstra | e490f9b | 2016-04-18 01:16:09 +0200 | [diff] [blame] | 124 | |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 125 | #define arch_atomic_fetch_and_relaxed arch_atomic_fetch_and_relaxed |
| 126 | #define arch_atomic_fetch_and_acquire arch_atomic_fetch_and_acquire |
| 127 | #define arch_atomic_fetch_and_release arch_atomic_fetch_and_release |
| 128 | #define arch_atomic_fetch_and arch_atomic_fetch_and |
Peter Zijlstra | e490f9b | 2016-04-18 01:16:09 +0200 | [diff] [blame] | 129 | |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 130 | #define arch_atomic_fetch_andnot_relaxed arch_atomic_fetch_andnot_relaxed |
| 131 | #define arch_atomic_fetch_andnot_acquire arch_atomic_fetch_andnot_acquire |
| 132 | #define arch_atomic_fetch_andnot_release arch_atomic_fetch_andnot_release |
| 133 | #define arch_atomic_fetch_andnot arch_atomic_fetch_andnot |
Peter Zijlstra | e490f9b | 2016-04-18 01:16:09 +0200 | [diff] [blame] | 134 | |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 135 | #define arch_atomic_fetch_or_relaxed arch_atomic_fetch_or_relaxed |
| 136 | #define arch_atomic_fetch_or_acquire arch_atomic_fetch_or_acquire |
| 137 | #define arch_atomic_fetch_or_release arch_atomic_fetch_or_release |
| 138 | #define arch_atomic_fetch_or arch_atomic_fetch_or |
Peter Zijlstra | e490f9b | 2016-04-18 01:16:09 +0200 | [diff] [blame] | 139 | |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 140 | #define arch_atomic_fetch_xor_relaxed arch_atomic_fetch_xor_relaxed |
| 141 | #define arch_atomic_fetch_xor_acquire arch_atomic_fetch_xor_acquire |
| 142 | #define arch_atomic_fetch_xor_release arch_atomic_fetch_xor_release |
| 143 | #define arch_atomic_fetch_xor arch_atomic_fetch_xor |
Peter Zijlstra | e490f9b | 2016-04-18 01:16:09 +0200 | [diff] [blame] | 144 | |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 145 | #define arch_atomic_xchg_relaxed(v, new) \ |
| 146 | arch_xchg_relaxed(&((v)->counter), (new)) |
| 147 | #define arch_atomic_xchg_acquire(v, new) \ |
| 148 | arch_xchg_acquire(&((v)->counter), (new)) |
| 149 | #define arch_atomic_xchg_release(v, new) \ |
| 150 | arch_xchg_release(&((v)->counter), (new)) |
| 151 | #define arch_atomic_xchg(v, new) \ |
| 152 | arch_xchg(&((v)->counter), (new)) |
Will Deacon | 305d454 | 2015-10-08 20:15:18 +0100 | [diff] [blame] | 153 | |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 154 | #define arch_atomic_cmpxchg_relaxed(v, old, new) \ |
| 155 | arch_cmpxchg_relaxed(&((v)->counter), (old), (new)) |
| 156 | #define arch_atomic_cmpxchg_acquire(v, old, new) \ |
| 157 | arch_cmpxchg_acquire(&((v)->counter), (old), (new)) |
| 158 | #define arch_atomic_cmpxchg_release(v, old, new) \ |
| 159 | arch_cmpxchg_release(&((v)->counter), (old), (new)) |
| 160 | #define arch_atomic_cmpxchg(v, old, new) \ |
| 161 | arch_cmpxchg(&((v)->counter), (old), (new)) |
Catalin Marinas | 6170a97 | 2012-03-05 11:49:29 +0000 | [diff] [blame] | 162 | |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 163 | #define arch_atomic_andnot arch_atomic_andnot |
Will Deacon | c275f76 | 2015-02-03 11:26:53 +0000 | [diff] [blame] | 164 | |
Catalin Marinas | 6170a97 | 2012-03-05 11:49:29 +0000 | [diff] [blame] | 165 | /* |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 166 | * 64-bit arch_atomic operations. |
Catalin Marinas | 6170a97 | 2012-03-05 11:49:29 +0000 | [diff] [blame] | 167 | */ |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 168 | #define ATOMIC64_INIT ATOMIC_INIT |
| 169 | #define arch_atomic64_read arch_atomic_read |
| 170 | #define arch_atomic64_set arch_atomic_set |
Will Deacon | 305d454 | 2015-10-08 20:15:18 +0100 | [diff] [blame] | 171 | |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 172 | #define arch_atomic64_add_return_relaxed arch_atomic64_add_return_relaxed |
| 173 | #define arch_atomic64_add_return_acquire arch_atomic64_add_return_acquire |
| 174 | #define arch_atomic64_add_return_release arch_atomic64_add_return_release |
| 175 | #define arch_atomic64_add_return arch_atomic64_add_return |
Will Deacon | 305d454 | 2015-10-08 20:15:18 +0100 | [diff] [blame] | 176 | |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 177 | #define arch_atomic64_sub_return_relaxed arch_atomic64_sub_return_relaxed |
| 178 | #define arch_atomic64_sub_return_acquire arch_atomic64_sub_return_acquire |
| 179 | #define arch_atomic64_sub_return_release arch_atomic64_sub_return_release |
| 180 | #define arch_atomic64_sub_return arch_atomic64_sub_return |
Will Deacon | 305d454 | 2015-10-08 20:15:18 +0100 | [diff] [blame] | 181 | |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 182 | #define arch_atomic64_fetch_add_relaxed arch_atomic64_fetch_add_relaxed |
| 183 | #define arch_atomic64_fetch_add_acquire arch_atomic64_fetch_add_acquire |
| 184 | #define arch_atomic64_fetch_add_release arch_atomic64_fetch_add_release |
| 185 | #define arch_atomic64_fetch_add arch_atomic64_fetch_add |
Peter Zijlstra | e490f9b | 2016-04-18 01:16:09 +0200 | [diff] [blame] | 186 | |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 187 | #define arch_atomic64_fetch_sub_relaxed arch_atomic64_fetch_sub_relaxed |
| 188 | #define arch_atomic64_fetch_sub_acquire arch_atomic64_fetch_sub_acquire |
| 189 | #define arch_atomic64_fetch_sub_release arch_atomic64_fetch_sub_release |
| 190 | #define arch_atomic64_fetch_sub arch_atomic64_fetch_sub |
Peter Zijlstra | e490f9b | 2016-04-18 01:16:09 +0200 | [diff] [blame] | 191 | |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 192 | #define arch_atomic64_fetch_and_relaxed arch_atomic64_fetch_and_relaxed |
| 193 | #define arch_atomic64_fetch_and_acquire arch_atomic64_fetch_and_acquire |
| 194 | #define arch_atomic64_fetch_and_release arch_atomic64_fetch_and_release |
| 195 | #define arch_atomic64_fetch_and arch_atomic64_fetch_and |
Peter Zijlstra | e490f9b | 2016-04-18 01:16:09 +0200 | [diff] [blame] | 196 | |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 197 | #define arch_atomic64_fetch_andnot_relaxed arch_atomic64_fetch_andnot_relaxed |
| 198 | #define arch_atomic64_fetch_andnot_acquire arch_atomic64_fetch_andnot_acquire |
| 199 | #define arch_atomic64_fetch_andnot_release arch_atomic64_fetch_andnot_release |
| 200 | #define arch_atomic64_fetch_andnot arch_atomic64_fetch_andnot |
Peter Zijlstra | e490f9b | 2016-04-18 01:16:09 +0200 | [diff] [blame] | 201 | |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 202 | #define arch_atomic64_fetch_or_relaxed arch_atomic64_fetch_or_relaxed |
| 203 | #define arch_atomic64_fetch_or_acquire arch_atomic64_fetch_or_acquire |
| 204 | #define arch_atomic64_fetch_or_release arch_atomic64_fetch_or_release |
| 205 | #define arch_atomic64_fetch_or arch_atomic64_fetch_or |
Peter Zijlstra | e490f9b | 2016-04-18 01:16:09 +0200 | [diff] [blame] | 206 | |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 207 | #define arch_atomic64_fetch_xor_relaxed arch_atomic64_fetch_xor_relaxed |
| 208 | #define arch_atomic64_fetch_xor_acquire arch_atomic64_fetch_xor_acquire |
| 209 | #define arch_atomic64_fetch_xor_release arch_atomic64_fetch_xor_release |
| 210 | #define arch_atomic64_fetch_xor arch_atomic64_fetch_xor |
Peter Zijlstra | e490f9b | 2016-04-18 01:16:09 +0200 | [diff] [blame] | 211 | |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 212 | #define arch_atomic64_xchg_relaxed arch_atomic_xchg_relaxed |
| 213 | #define arch_atomic64_xchg_acquire arch_atomic_xchg_acquire |
| 214 | #define arch_atomic64_xchg_release arch_atomic_xchg_release |
| 215 | #define arch_atomic64_xchg arch_atomic_xchg |
Will Deacon | 305d454 | 2015-10-08 20:15:18 +0100 | [diff] [blame] | 216 | |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 217 | #define arch_atomic64_cmpxchg_relaxed arch_atomic_cmpxchg_relaxed |
| 218 | #define arch_atomic64_cmpxchg_acquire arch_atomic_cmpxchg_acquire |
| 219 | #define arch_atomic64_cmpxchg_release arch_atomic_cmpxchg_release |
| 220 | #define arch_atomic64_cmpxchg arch_atomic_cmpxchg |
Catalin Marinas | 6170a97 | 2012-03-05 11:49:29 +0000 | [diff] [blame] | 221 | |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 222 | #define arch_atomic64_andnot arch_atomic64_andnot |
Catalin Marinas | 6170a97 | 2012-03-05 11:49:29 +0000 | [diff] [blame] | 223 | |
Mark Rutland | c0df108 | 2018-09-04 11:48:30 +0100 | [diff] [blame] | 224 | #define arch_atomic64_dec_if_positive arch_atomic64_dec_if_positive |
| 225 | |
Peter Zijlstra | 37f8173 | 2020-01-24 22:13:03 +0100 | [diff] [blame] | 226 | #define ARCH_ATOMIC |
Mark Rutland | b3a2a05 | 2018-06-21 13:13:20 +0100 | [diff] [blame] | 227 | |
Mark Rutland | b907b80 | 2019-07-08 17:36:40 +0100 | [diff] [blame] | 228 | #endif /* __ASM_ATOMIC_H */ |