]>
Commit | Line | Data |
---|---|---|
2874c5fd | 1 | /* SPDX-License-Identifier: GPL-2.0-or-later */ |
09d4e0ed PM |
2 | /* |
3 | * Generic implementation of 64-bit atomics using spinlocks, | |
4 | * useful on processors that don't have 64-bit atomic instructions. | |
5 | * | |
6 | * Copyright © 2009 Paul Mackerras, IBM Corp. <paulus@au1.ibm.com> | |
09d4e0ed PM |
7 | */ |
8 | #ifndef _ASM_GENERIC_ATOMIC64_H | |
9 | #define _ASM_GENERIC_ATOMIC64_H | |
ade5ef92 | 10 | #include <linux/types.h> |
09d4e0ed PM |
11 | |
12 | typedef struct { | |
9255813d | 13 | s64 counter; |
09d4e0ed PM |
14 | } atomic64_t; |
15 | ||
16 | #define ATOMIC64_INIT(i) { (i) } | |
17 | ||
1bdadf46 MR |
18 | extern s64 generic_atomic64_read(const atomic64_t *v); |
19 | extern void generic_atomic64_set(atomic64_t *v, s64 i); | |
9d664c0a | 20 | |
560cb12a | 21 | #define ATOMIC64_OP(op) \ |
1bdadf46 | 22 | extern void generic_atomic64_##op(s64 a, atomic64_t *v); |
560cb12a PZ |
23 | |
24 | #define ATOMIC64_OP_RETURN(op) \ | |
1bdadf46 | 25 | extern s64 generic_atomic64_##op##_return(s64 a, atomic64_t *v); |
560cb12a | 26 | |
28aa2bda | 27 | #define ATOMIC64_FETCH_OP(op) \ |
1bdadf46 | 28 | extern s64 generic_atomic64_fetch_##op(s64 a, atomic64_t *v); |
28aa2bda PZ |
29 | |
30 | #define ATOMIC64_OPS(op) ATOMIC64_OP(op) ATOMIC64_OP_RETURN(op) ATOMIC64_FETCH_OP(op) | |
560cb12a PZ |
31 | |
32 | ATOMIC64_OPS(add) | |
33 | ATOMIC64_OPS(sub) | |
34 | ||
28aa2bda PZ |
35 | #undef ATOMIC64_OPS |
36 | #define ATOMIC64_OPS(op) ATOMIC64_OP(op) ATOMIC64_FETCH_OP(op) | |
37 | ||
38 | ATOMIC64_OPS(and) | |
39 | ATOMIC64_OPS(or) | |
40 | ATOMIC64_OPS(xor) | |
e6942b7d | 41 | |
560cb12a | 42 | #undef ATOMIC64_OPS |
28aa2bda | 43 | #undef ATOMIC64_FETCH_OP |
560cb12a PZ |
44 | #undef ATOMIC64_OP_RETURN |
45 | #undef ATOMIC64_OP | |
46 | ||
1bdadf46 MR |
47 | extern s64 generic_atomic64_dec_if_positive(atomic64_t *v); |
48 | extern s64 generic_atomic64_cmpxchg(atomic64_t *v, s64 o, s64 n); | |
49 | extern s64 generic_atomic64_xchg(atomic64_t *v, s64 new); | |
50 | extern s64 generic_atomic64_fetch_add_unless(atomic64_t *v, s64 a, s64 u); | |
51 | ||
1bdadf46 MR |
52 | #define arch_atomic64_read generic_atomic64_read |
53 | #define arch_atomic64_set generic_atomic64_set | |
54 | #define arch_atomic64_set_release generic_atomic64_set | |
55 | ||
56 | #define arch_atomic64_add generic_atomic64_add | |
57 | #define arch_atomic64_add_return generic_atomic64_add_return | |
58 | #define arch_atomic64_fetch_add generic_atomic64_fetch_add | |
59 | #define arch_atomic64_sub generic_atomic64_sub | |
60 | #define arch_atomic64_sub_return generic_atomic64_sub_return | |
61 | #define arch_atomic64_fetch_sub generic_atomic64_fetch_sub | |
62 | ||
63 | #define arch_atomic64_and generic_atomic64_and | |
64 | #define arch_atomic64_fetch_and generic_atomic64_fetch_and | |
65 | #define arch_atomic64_or generic_atomic64_or | |
66 | #define arch_atomic64_fetch_or generic_atomic64_fetch_or | |
67 | #define arch_atomic64_xor generic_atomic64_xor | |
68 | #define arch_atomic64_fetch_xor generic_atomic64_fetch_xor | |
69 | ||
70 | #define arch_atomic64_dec_if_positive generic_atomic64_dec_if_positive | |
71 | #define arch_atomic64_cmpxchg generic_atomic64_cmpxchg | |
72 | #define arch_atomic64_xchg generic_atomic64_xchg | |
73 | #define arch_atomic64_fetch_add_unless generic_atomic64_fetch_add_unless | |
74 | ||
09d4e0ed | 75 | #endif /* _ASM_GENERIC_ATOMIC64_H */ |