Commit | Line | Data |
---|---|---|
c275f76b WD |
1 | /* |
2 | * Based on arch/arm/include/asm/atomic.h | |
3 | * | |
4 | * Copyright (C) 1996 Russell King. | |
5 | * Copyright (C) 2002 Deep Blue Solutions Ltd. | |
6 | * Copyright (C) 2012 ARM Ltd. | |
7 | * | |
8 | * This program is free software; you can redistribute it and/or modify | |
9 | * it under the terms of the GNU General Public License version 2 as | |
10 | * published by the Free Software Foundation. | |
11 | * | |
12 | * This program is distributed in the hope that it will be useful, | |
13 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | |
14 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | |
15 | * GNU General Public License for more details. | |
16 | * | |
17 | * You should have received a copy of the GNU General Public License | |
18 | * along with this program. If not, see <http://www.gnu.org/licenses/>. | |
19 | */ | |
20 | ||
21 | #ifndef __ASM_ATOMIC_LL_SC_H | |
22 | #define __ASM_ATOMIC_LL_SC_H | |
23 | ||
c0385b24 WD |
24 | #ifndef __ARM64_IN_ATOMIC_IMPL |
25 | #error "please don't include this file directly" | |
26 | #endif | |
27 | ||
c275f76b WD |
28 | /* |
29 | * AArch64 UP and SMP safe atomic ops. We use load exclusive and | |
30 | * store exclusive to ensure that these are atomic. We may loop | |
31 | * to ensure that the update happens. | |
32 | * | |
33 | * NOTE: these functions do *not* follow the PCS and must explicitly | |
34 | * save any clobbered registers other than x0 (regardless of return | |
35 | * value). This is achieved through -fcall-saved-* compiler flags for | |
36 | * this file, which unfortunately don't work on a per-function basis | |
37 | * (the optimize attribute silently ignores these options). | |
38 | */ | |
39 | ||
c275f76b WD |
40 | #define ATOMIC_OP(op, asm_op) \ |
41 | __LL_SC_INLINE void \ | |
42 | __LL_SC_PREFIX(atomic_##op(int i, atomic_t *v)) \ | |
43 | { \ | |
44 | unsigned long tmp; \ | |
45 | int result; \ | |
46 | \ | |
47 | asm volatile("// atomic_" #op "\n" \ | |
0ea366f5 | 48 | " prfm pstl1strm, %2\n" \ |
c275f76b WD |
49 | "1: ldxr %w0, %2\n" \ |
50 | " " #asm_op " %w0, %w0, %w3\n" \ | |
51 | " stxr %w1, %w0, %2\n" \ | |
52 | " cbnz %w1, 1b" \ | |
53 | : "=&r" (result), "=&r" (tmp), "+Q" (v->counter) \ | |
54 | : "Ir" (i)); \ | |
55 | } \ | |
c0385b24 | 56 | __LL_SC_EXPORT(atomic_##op); |
c275f76b WD |
57 | |
58 | #define ATOMIC_OP_RETURN(op, asm_op) \ | |
59 | __LL_SC_INLINE int \ | |
60 | __LL_SC_PREFIX(atomic_##op##_return(int i, atomic_t *v)) \ | |
61 | { \ | |
62 | unsigned long tmp; \ | |
63 | int result; \ | |
64 | \ | |
65 | asm volatile("// atomic_" #op "_return\n" \ | |
0ea366f5 | 66 | " prfm pstl1strm, %2\n" \ |
c275f76b WD |
67 | "1: ldxr %w0, %2\n" \ |
68 | " " #asm_op " %w0, %w0, %w3\n" \ | |
69 | " stlxr %w1, %w0, %2\n" \ | |
70 | " cbnz %w1, 1b" \ | |
71 | : "=&r" (result), "=&r" (tmp), "+Q" (v->counter) \ | |
72 | : "Ir" (i) \ | |
73 | : "memory"); \ | |
74 | \ | |
75 | smp_mb(); \ | |
76 | return result; \ | |
c0385b24 WD |
77 | } \ |
78 | __LL_SC_EXPORT(atomic_##op##_return); | |
c275f76b WD |
79 | |
80 | #define ATOMIC_OPS(op, asm_op) \ | |
81 | ATOMIC_OP(op, asm_op) \ | |
82 | ATOMIC_OP_RETURN(op, asm_op) | |
83 | ||
84 | ATOMIC_OPS(add, add) | |
85 | ATOMIC_OPS(sub, sub) | |
86 | ||
87 | ATOMIC_OP(and, and) | |
88 | ATOMIC_OP(andnot, bic) | |
89 | ATOMIC_OP(or, orr) | |
90 | ATOMIC_OP(xor, eor) | |
91 | ||
92 | #undef ATOMIC_OPS | |
93 | #undef ATOMIC_OP_RETURN | |
94 | #undef ATOMIC_OP | |
95 | ||
c275f76b WD |
96 | #define ATOMIC64_OP(op, asm_op) \ |
97 | __LL_SC_INLINE void \ | |
98 | __LL_SC_PREFIX(atomic64_##op(long i, atomic64_t *v)) \ | |
99 | { \ | |
100 | long result; \ | |
101 | unsigned long tmp; \ | |
102 | \ | |
103 | asm volatile("// atomic64_" #op "\n" \ | |
0ea366f5 | 104 | " prfm pstl1strm, %2\n" \ |
c275f76b WD |
105 | "1: ldxr %0, %2\n" \ |
106 | " " #asm_op " %0, %0, %3\n" \ | |
107 | " stxr %w1, %0, %2\n" \ | |
108 | " cbnz %w1, 1b" \ | |
109 | : "=&r" (result), "=&r" (tmp), "+Q" (v->counter) \ | |
110 | : "Ir" (i)); \ | |
111 | } \ | |
c0385b24 | 112 | __LL_SC_EXPORT(atomic64_##op); |
c275f76b WD |
113 | |
114 | #define ATOMIC64_OP_RETURN(op, asm_op) \ | |
115 | __LL_SC_INLINE long \ | |
116 | __LL_SC_PREFIX(atomic64_##op##_return(long i, atomic64_t *v)) \ | |
117 | { \ | |
118 | long result; \ | |
119 | unsigned long tmp; \ | |
120 | \ | |
121 | asm volatile("// atomic64_" #op "_return\n" \ | |
0ea366f5 | 122 | " prfm pstl1strm, %2\n" \ |
c275f76b WD |
123 | "1: ldxr %0, %2\n" \ |
124 | " " #asm_op " %0, %0, %3\n" \ | |
125 | " stlxr %w1, %0, %2\n" \ | |
126 | " cbnz %w1, 1b" \ | |
127 | : "=&r" (result), "=&r" (tmp), "+Q" (v->counter) \ | |
128 | : "Ir" (i) \ | |
129 | : "memory"); \ | |
130 | \ | |
131 | smp_mb(); \ | |
132 | return result; \ | |
c0385b24 WD |
133 | } \ |
134 | __LL_SC_EXPORT(atomic64_##op##_return); | |
c275f76b WD |
135 | |
136 | #define ATOMIC64_OPS(op, asm_op) \ | |
137 | ATOMIC64_OP(op, asm_op) \ | |
138 | ATOMIC64_OP_RETURN(op, asm_op) | |
139 | ||
140 | ATOMIC64_OPS(add, add) | |
141 | ATOMIC64_OPS(sub, sub) | |
142 | ||
143 | ATOMIC64_OP(and, and) | |
144 | ATOMIC64_OP(andnot, bic) | |
145 | ATOMIC64_OP(or, orr) | |
146 | ATOMIC64_OP(xor, eor) | |
147 | ||
148 | #undef ATOMIC64_OPS | |
149 | #undef ATOMIC64_OP_RETURN | |
150 | #undef ATOMIC64_OP | |
151 | ||
c275f76b WD |
152 | __LL_SC_INLINE long |
153 | __LL_SC_PREFIX(atomic64_dec_if_positive(atomic64_t *v)) | |
154 | { | |
155 | long result; | |
156 | unsigned long tmp; | |
157 | ||
158 | asm volatile("// atomic64_dec_if_positive\n" | |
0ea366f5 | 159 | " prfm pstl1strm, %2\n" |
c275f76b WD |
160 | "1: ldxr %0, %2\n" |
161 | " subs %0, %0, #1\n" | |
db26217e | 162 | " b.lt 2f\n" |
c275f76b WD |
163 | " stlxr %w1, %0, %2\n" |
164 | " cbnz %w1, 1b\n" | |
165 | " dmb ish\n" | |
166 | "2:" | |
167 | : "=&r" (result), "=&r" (tmp), "+Q" (v->counter) | |
168 | : | |
169 | : "cc", "memory"); | |
170 | ||
171 | return result; | |
172 | } | |
c0385b24 | 173 | __LL_SC_EXPORT(atomic64_dec_if_positive); |
c275f76b | 174 | |
4e39715f | 175 | #define __CMPXCHG_CASE(w, sz, name, mb, rel, cl) \ |
c342f782 WD |
176 | __LL_SC_INLINE unsigned long \ |
177 | __LL_SC_PREFIX(__cmpxchg_case_##name(volatile void *ptr, \ | |
178 | unsigned long old, \ | |
179 | unsigned long new)) \ | |
180 | { \ | |
181 | unsigned long tmp, oldval; \ | |
182 | \ | |
183 | asm volatile( \ | |
7f08a414 | 184 | " prfm pstl1strm, %[v]\n" \ |
c342f782 WD |
185 | "1: ldxr" #sz "\t%" #w "[oldval], %[v]\n" \ |
186 | " eor %" #w "[tmp], %" #w "[oldval], %" #w "[old]\n" \ | |
187 | " cbnz %" #w "[tmp], 2f\n" \ | |
4e39715f | 188 | " st" #rel "xr" #sz "\t%w[tmp], %" #w "[new], %[v]\n" \ |
c342f782 WD |
189 | " cbnz %w[tmp], 1b\n" \ |
190 | " " #mb "\n" \ | |
191 | " mov %" #w "[oldval], %" #w "[old]\n" \ | |
192 | "2:" \ | |
193 | : [tmp] "=&r" (tmp), [oldval] "=&r" (oldval), \ | |
194 | [v] "+Q" (*(unsigned long *)ptr) \ | |
195 | : [old] "Lr" (old), [new] "r" (new) \ | |
196 | : cl); \ | |
197 | \ | |
198 | return oldval; \ | |
199 | } \ | |
200 | __LL_SC_EXPORT(__cmpxchg_case_##name); | |
201 | ||
4e39715f WD |
202 | __CMPXCHG_CASE(w, b, 1, , , ) |
203 | __CMPXCHG_CASE(w, h, 2, , , ) | |
204 | __CMPXCHG_CASE(w, , 4, , , ) | |
205 | __CMPXCHG_CASE( , , 8, , , ) | |
206 | __CMPXCHG_CASE(w, b, mb_1, dmb ish, l, "memory") | |
207 | __CMPXCHG_CASE(w, h, mb_2, dmb ish, l, "memory") | |
208 | __CMPXCHG_CASE(w, , mb_4, dmb ish, l, "memory") | |
209 | __CMPXCHG_CASE( , , mb_8, dmb ish, l, "memory") | |
c342f782 WD |
210 | |
211 | #undef __CMPXCHG_CASE | |
212 | ||
4e39715f | 213 | #define __CMPXCHG_DBL(name, mb, rel, cl) \ |
e9a4b795 WD |
214 | __LL_SC_INLINE int \ |
215 | __LL_SC_PREFIX(__cmpxchg_double##name(unsigned long old1, \ | |
216 | unsigned long old2, \ | |
217 | unsigned long new1, \ | |
218 | unsigned long new2, \ | |
219 | volatile void *ptr)) \ | |
220 | { \ | |
221 | unsigned long tmp, ret; \ | |
222 | \ | |
223 | asm volatile("// __cmpxchg_double" #name "\n" \ | |
0ea366f5 | 224 | " prfm pstl1strm, %2\n" \ |
e9a4b795 WD |
225 | "1: ldxp %0, %1, %2\n" \ |
226 | " eor %0, %0, %3\n" \ | |
227 | " eor %1, %1, %4\n" \ | |
228 | " orr %1, %0, %1\n" \ | |
229 | " cbnz %1, 2f\n" \ | |
4e39715f | 230 | " st" #rel "xp %w0, %5, %6, %2\n" \ |
e9a4b795 WD |
231 | " cbnz %w0, 1b\n" \ |
232 | " " #mb "\n" \ | |
233 | "2:" \ | |
234 | : "=&r" (tmp), "=&r" (ret), "+Q" (*(unsigned long *)ptr) \ | |
235 | : "r" (old1), "r" (old2), "r" (new1), "r" (new2) \ | |
236 | : cl); \ | |
237 | \ | |
238 | return ret; \ | |
239 | } \ | |
240 | __LL_SC_EXPORT(__cmpxchg_double##name); | |
241 | ||
4e39715f WD |
242 | __CMPXCHG_DBL( , , , ) |
243 | __CMPXCHG_DBL(_mb, dmb ish, l, "memory") | |
e9a4b795 WD |
244 | |
245 | #undef __CMPXCHG_DBL | |
246 | ||
c275f76b | 247 | #endif /* __ASM_ATOMIC_LL_SC_H */ |