Robin Getz | 96f1050 | 2009-09-24 14:11:24 +0000 | [diff] [blame] | 1 | /* |
| 2 | * Copyright 2004-2009 Analog Devices Inc. |
| 3 | * |
| 4 | * Licensed under the GPL-2 or later. |
| 5 | */ |
| 6 | |
Bryan Wu | 1394f03 | 2007-05-06 14:50:22 -0700 | [diff] [blame] | 7 | #ifndef __BFIN_SPINLOCK_H |
| 8 | #define __BFIN_SPINLOCK_H |
| 9 | |
Mike Frysinger | 3d15063 | 2009-06-13 11:21:51 -0400 | [diff] [blame] | 10 | #ifndef CONFIG_SMP |
| 11 | # include <asm-generic/spinlock.h> |
| 12 | #else |
| 13 | |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 14 | #include <asm/atomic.h> |
Bryan Wu | 1394f03 | 2007-05-06 14:50:22 -0700 | [diff] [blame] | 15 | |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 16 | asmlinkage int __raw_spin_is_locked_asm(volatile int *ptr); |
| 17 | asmlinkage void __raw_spin_lock_asm(volatile int *ptr); |
| 18 | asmlinkage int __raw_spin_trylock_asm(volatile int *ptr); |
| 19 | asmlinkage void __raw_spin_unlock_asm(volatile int *ptr); |
Graf Yang | 71a6628 | 2010-03-12 04:24:21 +0000 | [diff] [blame] | 20 | asmlinkage void __raw_read_lock_asm(volatile int *ptr); |
| 21 | asmlinkage int __raw_read_trylock_asm(volatile int *ptr); |
| 22 | asmlinkage void __raw_read_unlock_asm(volatile int *ptr); |
| 23 | asmlinkage void __raw_write_lock_asm(volatile int *ptr); |
| 24 | asmlinkage int __raw_write_trylock_asm(volatile int *ptr); |
| 25 | asmlinkage void __raw_write_unlock_asm(volatile int *ptr); |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 26 | |
Thomas Gleixner | 0199c4e | 2009-12-02 20:01:25 +0100 | [diff] [blame] | 27 | static inline int arch_spin_is_locked(arch_spinlock_t *lock) |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 28 | { |
| 29 | return __raw_spin_is_locked_asm(&lock->lock); |
| 30 | } |
| 31 | |
Thomas Gleixner | 0199c4e | 2009-12-02 20:01:25 +0100 | [diff] [blame] | 32 | static inline void arch_spin_lock(arch_spinlock_t *lock) |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 33 | { |
| 34 | __raw_spin_lock_asm(&lock->lock); |
| 35 | } |
| 36 | |
Thomas Gleixner | 0199c4e | 2009-12-02 20:01:25 +0100 | [diff] [blame] | 37 | #define arch_spin_lock_flags(lock, flags) arch_spin_lock(lock) |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 38 | |
Thomas Gleixner | 0199c4e | 2009-12-02 20:01:25 +0100 | [diff] [blame] | 39 | static inline int arch_spin_trylock(arch_spinlock_t *lock) |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 40 | { |
| 41 | return __raw_spin_trylock_asm(&lock->lock); |
| 42 | } |
| 43 | |
Thomas Gleixner | 0199c4e | 2009-12-02 20:01:25 +0100 | [diff] [blame] | 44 | static inline void arch_spin_unlock(arch_spinlock_t *lock) |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 45 | { |
| 46 | __raw_spin_unlock_asm(&lock->lock); |
| 47 | } |
| 48 | |
Thomas Gleixner | 0199c4e | 2009-12-02 20:01:25 +0100 | [diff] [blame] | 49 | static inline void arch_spin_unlock_wait(arch_spinlock_t *lock) |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 50 | { |
Thomas Gleixner | 0199c4e | 2009-12-02 20:01:25 +0100 | [diff] [blame] | 51 | while (arch_spin_is_locked(lock)) |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 52 | cpu_relax(); |
| 53 | } |
| 54 | |
Thomas Gleixner | e593194 | 2009-12-03 20:08:46 +0100 | [diff] [blame] | 55 | static inline int arch_read_can_lock(arch_rwlock_t *rw) |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 56 | { |
| 57 | return __raw_uncached_fetch_asm(&rw->lock) > 0; |
| 58 | } |
| 59 | |
Thomas Gleixner | e593194 | 2009-12-03 20:08:46 +0100 | [diff] [blame] | 60 | static inline int arch_write_can_lock(arch_rwlock_t *rw) |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 61 | { |
| 62 | return __raw_uncached_fetch_asm(&rw->lock) == RW_LOCK_BIAS; |
| 63 | } |
| 64 | |
Thomas Gleixner | e593194 | 2009-12-03 20:08:46 +0100 | [diff] [blame] | 65 | static inline void arch_read_lock(arch_rwlock_t *rw) |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 66 | { |
Graf Yang | 71a6628 | 2010-03-12 04:24:21 +0000 | [diff] [blame] | 67 | __raw_read_lock_asm(&rw->lock); |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 68 | } |
| 69 | |
Graf Yang | 54d756e | 2009-09-21 11:51:31 +0000 | [diff] [blame] | 70 | #define arch_read_lock_flags(lock, flags) arch_read_lock(lock) |
| 71 | |
Thomas Gleixner | e593194 | 2009-12-03 20:08:46 +0100 | [diff] [blame] | 72 | static inline int arch_read_trylock(arch_rwlock_t *rw) |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 73 | { |
Graf Yang | 71a6628 | 2010-03-12 04:24:21 +0000 | [diff] [blame] | 74 | return __raw_read_trylock_asm(&rw->lock); |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 75 | } |
| 76 | |
Thomas Gleixner | e593194 | 2009-12-03 20:08:46 +0100 | [diff] [blame] | 77 | static inline void arch_read_unlock(arch_rwlock_t *rw) |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 78 | { |
Graf Yang | 71a6628 | 2010-03-12 04:24:21 +0000 | [diff] [blame] | 79 | __raw_read_unlock_asm(&rw->lock); |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 80 | } |
| 81 | |
Thomas Gleixner | e593194 | 2009-12-03 20:08:46 +0100 | [diff] [blame] | 82 | static inline void arch_write_lock(arch_rwlock_t *rw) |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 83 | { |
Graf Yang | 71a6628 | 2010-03-12 04:24:21 +0000 | [diff] [blame] | 84 | __raw_write_lock_asm(&rw->lock); |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 85 | } |
| 86 | |
Graf Yang | 54d756e | 2009-09-21 11:51:31 +0000 | [diff] [blame] | 87 | #define arch_write_lock_flags(lock, flags) arch_write_lock(lock) |
| 88 | |
Thomas Gleixner | e593194 | 2009-12-03 20:08:46 +0100 | [diff] [blame] | 89 | static inline int arch_write_trylock(arch_rwlock_t *rw) |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 90 | { |
Graf Yang | 71a6628 | 2010-03-12 04:24:21 +0000 | [diff] [blame] | 91 | return __raw_write_trylock_asm(&rw->lock); |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 92 | } |
| 93 | |
Thomas Gleixner | e593194 | 2009-12-03 20:08:46 +0100 | [diff] [blame] | 94 | static inline void arch_write_unlock(arch_rwlock_t *rw) |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 95 | { |
Graf Yang | 71a6628 | 2010-03-12 04:24:21 +0000 | [diff] [blame] | 96 | __raw_write_unlock_asm(&rw->lock); |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 97 | } |
| 98 | |
Thomas Gleixner | 0199c4e | 2009-12-02 20:01:25 +0100 | [diff] [blame] | 99 | #define arch_spin_relax(lock) cpu_relax() |
| 100 | #define arch_read_relax(lock) cpu_relax() |
| 101 | #define arch_write_relax(lock) cpu_relax() |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 102 | |
Mike Frysinger | 3d15063 | 2009-06-13 11:21:51 -0400 | [diff] [blame] | 103 | #endif |
| 104 | |
Graf Yang | 6b3087c | 2009-01-07 23:14:39 +0800 | [diff] [blame] | 105 | #endif /* !__BFIN_SPINLOCK_H */ |