]>
git.proxmox.com Git - mirror_ubuntu-bionic-kernel.git/blob - include/asm-sparc64/spinlock.h
1 /* spinlock.h: 64-bit Sparc spinlock support.
3 * Copyright (C) 1997 David S. Miller (davem@caip.rutgers.edu)
6 #ifndef __SPARC64_SPINLOCK_H
7 #define __SPARC64_SPINLOCK_H
9 #include <linux/config.h>
10 #include <linux/threads.h> /* For NR_CPUS */
14 /* To get debugging spinlocks which detect and catch
15 * deadlock situations, set CONFIG_DEBUG_SPINLOCK
16 * and rebuild your kernel.
19 /* All of these locking primitives are expected to work properly
20 * even in an RMO memory model, which currently is what the kernel
23 * There is another issue. Because we play games to save cycles
24 * in the non-contention case, we need to be extra careful about
25 * branch targets into the "spinning" code. They live in their
26 * own section, but the newer V9 branches have a shorter range
27 * than the traditional 32-bit sparc branch variants. The rule
28 * is that the branches that go into and out of the spinner sections
29 * must be pre-V9 branches.
32 #ifndef CONFIG_DEBUG_SPINLOCK
35 volatile unsigned char lock
;
37 unsigned int break_lock
;
40 #define SPIN_LOCK_UNLOCKED (spinlock_t) {0,}
42 #define spin_lock_init(lp) do { *(lp)= SPIN_LOCK_UNLOCKED; } while(0)
43 #define spin_is_locked(lp) ((lp)->lock != 0)
45 #define spin_unlock_wait(lp) \
49 static inline void _raw_spin_lock(spinlock_t
*lock
)
54 "1: ldstub [%1], %0\n"
55 " membar #StoreLoad | #StoreStore\n"
63 " ba,a,pt %%xcc, 1b\n"
70 static inline int _raw_spin_trylock(spinlock_t
*lock
)
76 " membar #StoreLoad | #StoreStore"
81 return (result
== 0UL);
84 static inline void _raw_spin_unlock(spinlock_t
*lock
)
87 " membar #StoreStore | #LoadStore\n"
94 static inline void _raw_spin_lock_flags(spinlock_t
*lock
, unsigned long flags
)
96 unsigned long tmp1
, tmp2
;
99 "1: ldstub [%2], %0\n"
100 " membar #StoreLoad | #StoreStore\n"
104 "2: rdpr %%pil, %1\n"
107 " membar #LoadLoad\n"
113 : "=&r" (tmp1
), "=&r" (tmp2
)
114 : "r"(lock
), "r"(flags
)
118 #else /* !(CONFIG_DEBUG_SPINLOCK) */
121 volatile unsigned char lock
;
122 unsigned int owner_pc
, owner_cpu
;
123 #ifdef CONFIG_PREEMPT
124 unsigned int break_lock
;
127 #define SPIN_LOCK_UNLOCKED (spinlock_t) { 0, 0, 0xff }
128 #define spin_lock_init(lp) do { *(lp)= SPIN_LOCK_UNLOCKED; } while(0)
129 #define spin_is_locked(__lock) ((__lock)->lock != 0)
130 #define spin_unlock_wait(__lock) \
133 } while((__lock)->lock)
135 extern void _do_spin_lock(spinlock_t
*lock
, char *str
, unsigned long caller
);
136 extern void _do_spin_unlock(spinlock_t
*lock
);
137 extern int _do_spin_trylock(spinlock_t
*lock
, unsigned long caller
);
139 #define _raw_spin_trylock(lp) \
140 _do_spin_trylock(lp, (unsigned long) __builtin_return_address(0))
141 #define _raw_spin_lock(lock) \
142 _do_spin_lock(lock, "spin_lock", \
143 (unsigned long) __builtin_return_address(0))
144 #define _raw_spin_unlock(lock) _do_spin_unlock(lock)
145 #define _raw_spin_lock_flags(lock, flags) _raw_spin_lock(lock)
147 #endif /* CONFIG_DEBUG_SPINLOCK */
149 /* Multi-reader locks, these are much saner than the 32-bit Sparc ones... */
151 #ifndef CONFIG_DEBUG_SPINLOCK
154 volatile unsigned int lock
;
155 #ifdef CONFIG_PREEMPT
156 unsigned int break_lock
;
159 #define RW_LOCK_UNLOCKED (rwlock_t) {0,}
160 #define rwlock_init(lp) do { *(lp) = RW_LOCK_UNLOCKED; } while(0)
162 static void inline __read_lock(rwlock_t
*lock
)
164 unsigned long tmp1
, tmp2
;
166 __asm__
__volatile__ (
170 " cas [%2], %0, %1\n"
172 " membar #StoreLoad | #StoreStore\n"
173 " bne,pn %%icc, 1b\n"
177 " membar #LoadLoad\n"
180 " ba,a,pt %%xcc, 4b\n"
182 : "=&r" (tmp1
), "=&r" (tmp2
)
187 static void inline __read_unlock(rwlock_t
*lock
)
189 unsigned long tmp1
, tmp2
;
191 __asm__
__volatile__(
192 " membar #StoreLoad | #LoadLoad\n"
195 " cas [%2], %0, %1\n"
197 " bne,pn %%xcc, 1b\n"
199 : "=&r" (tmp1
), "=&r" (tmp2
)
204 static void inline __write_lock(rwlock_t
*lock
)
206 unsigned long mask
, tmp1
, tmp2
;
210 __asm__
__volatile__(
214 " cas [%2], %0, %1\n"
216 " membar #StoreLoad | #StoreStore\n"
217 " bne,pn %%icc, 1b\n"
221 " membar #LoadLoad\n"
224 " ba,a,pt %%xcc, 4b\n"
226 : "=&r" (tmp1
), "=&r" (tmp2
)
227 : "r" (lock
), "r" (mask
)
231 static void inline __write_unlock(rwlock_t
*lock
)
233 __asm__
__volatile__(
234 " membar #LoadStore | #StoreStore\n"
241 static int inline __write_trylock(rwlock_t
*lock
)
243 unsigned long mask
, tmp1
, tmp2
, result
;
247 __asm__
__volatile__(
252 " cas [%3], %0, %1\n"
254 " membar #StoreLoad | #StoreStore\n"
255 " bne,pn %%icc, 1b\n"
259 : "=&r" (tmp1
), "=&r" (tmp2
), "=&r" (result
)
260 : "r" (lock
), "r" (mask
)
266 #define _raw_read_lock(p) __read_lock(p)
267 #define _raw_read_unlock(p) __read_unlock(p)
268 #define _raw_write_lock(p) __write_lock(p)
269 #define _raw_write_unlock(p) __write_unlock(p)
270 #define _raw_write_trylock(p) __write_trylock(p)
272 #else /* !(CONFIG_DEBUG_SPINLOCK) */
275 volatile unsigned long lock
;
276 unsigned int writer_pc
, writer_cpu
;
277 unsigned int reader_pc
[NR_CPUS
];
278 #ifdef CONFIG_PREEMPT
279 unsigned int break_lock
;
282 #define RW_LOCK_UNLOCKED (rwlock_t) { 0, 0, 0xff, { } }
283 #define rwlock_init(lp) do { *(lp) = RW_LOCK_UNLOCKED; } while(0)
285 extern void _do_read_lock(rwlock_t
*rw
, char *str
, unsigned long caller
);
286 extern void _do_read_unlock(rwlock_t
*rw
, char *str
, unsigned long caller
);
287 extern void _do_write_lock(rwlock_t
*rw
, char *str
, unsigned long caller
);
288 extern void _do_write_unlock(rwlock_t
*rw
, unsigned long caller
);
289 extern int _do_write_trylock(rwlock_t
*rw
, char *str
, unsigned long caller
);
291 #define _raw_read_lock(lock) \
292 do { unsigned long flags; \
293 local_irq_save(flags); \
294 _do_read_lock(lock, "read_lock", \
295 (unsigned long) __builtin_return_address(0)); \
296 local_irq_restore(flags); \
299 #define _raw_read_unlock(lock) \
300 do { unsigned long flags; \
301 local_irq_save(flags); \
302 _do_read_unlock(lock, "read_unlock", \
303 (unsigned long) __builtin_return_address(0)); \
304 local_irq_restore(flags); \
307 #define _raw_write_lock(lock) \
308 do { unsigned long flags; \
309 local_irq_save(flags); \
310 _do_write_lock(lock, "write_lock", \
311 (unsigned long) __builtin_return_address(0)); \
312 local_irq_restore(flags); \
315 #define _raw_write_unlock(lock) \
316 do { unsigned long flags; \
317 local_irq_save(flags); \
318 _do_write_unlock(lock, \
319 (unsigned long) __builtin_return_address(0)); \
320 local_irq_restore(flags); \
323 #define _raw_write_trylock(lock) \
324 ({ unsigned long flags; \
326 local_irq_save(flags); \
327 val = _do_write_trylock(lock, "write_trylock", \
328 (unsigned long) __builtin_return_address(0)); \
329 local_irq_restore(flags); \
333 #endif /* CONFIG_DEBUG_SPINLOCK */
335 #define _raw_read_trylock(lock) generic_raw_read_trylock(lock)
336 #define read_can_lock(rw) (!((rw)->lock & 0x80000000UL))
337 #define write_can_lock(rw) (!(rw)->lock)
339 #endif /* !(__ASSEMBLY__) */
341 #endif /* !(__SPARC64_SPINLOCK_H) */