Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux
1
fork

Configure Feed

Select the types of activity you want to include in your feed.

at 4d8e74ad4585672489da6145b3328d415f50db82 144 lines 4.6 kB view raw
1#ifndef __LINUX_SPINLOCK_API_UP_H 2#define __LINUX_SPINLOCK_API_UP_H 3 4#ifndef __LINUX_INSIDE_SPINLOCK_H 5# error "please don't include this file directly" 6#endif 7 8/* 9 * include/linux/spinlock_api_up.h 10 * 11 * spinlock API implementation on UP-nondebug (inlined implementation) 12 * 13 * portions Copyright 2005, Red Hat, Inc., Ingo Molnar 14 * Released under the General Public License (GPL). 15 */ 16 17#define in_lock_functions(ADDR) 0 18 19#define assert_raw_spin_locked(lock) do { (void)(lock); } while (0) 20 21/* 22 * In the UP-nondebug case there's no real locking going on, so the 23 * only thing we have to do is to keep the preempt counts and irq 24 * flags straight, to suppress compiler warnings of unused lock 25 * variables, and to add the proper checker annotations: 26 */ 27#define ___LOCK_(lock) \ 28 do { __acquire(lock); (void)(lock); } while (0) 29 30#define ___LOCK_shared(lock) \ 31 do { __acquire_shared(lock); (void)(lock); } while (0) 32 33#define __LOCK(lock, ...) \ 34 do { preempt_disable(); ___LOCK_##__VA_ARGS__(lock); } while (0) 35 36#define __LOCK_BH(lock, ...) \ 37 do { __local_bh_disable_ip(_THIS_IP_, SOFTIRQ_LOCK_OFFSET); ___LOCK_##__VA_ARGS__(lock); } while (0) 38 39#define __LOCK_IRQ(lock, ...) \ 40 do { local_irq_disable(); __LOCK(lock, ##__VA_ARGS__); } while (0) 41 42#define __LOCK_IRQSAVE(lock, flags, ...) \ 43 do { local_irq_save(flags); __LOCK(lock, ##__VA_ARGS__); } while (0) 44 45#define ___UNLOCK_(lock) \ 46 do { __release(lock); (void)(lock); } while (0) 47 48#define ___UNLOCK_shared(lock) \ 49 do { __release_shared(lock); (void)(lock); } while (0) 50 51#define __UNLOCK(lock, ...) \ 52 do { preempt_enable(); ___UNLOCK_##__VA_ARGS__(lock); } while (0) 53 54#define __UNLOCK_BH(lock, ...) \ 55 do { __local_bh_enable_ip(_THIS_IP_, SOFTIRQ_LOCK_OFFSET); \ 56 ___UNLOCK_##__VA_ARGS__(lock); } while (0) 57 58#define __UNLOCK_IRQ(lock, ...) \ 59 do { local_irq_enable(); __UNLOCK(lock, ##__VA_ARGS__); } while (0) 60 61#define __UNLOCK_IRQRESTORE(lock, flags, ...) \ 62 do { local_irq_restore(flags); __UNLOCK(lock, ##__VA_ARGS__); } while (0) 63 64#define _raw_spin_lock(lock) __LOCK(lock) 65#define _raw_spin_lock_nested(lock, subclass) __LOCK(lock) 66#define _raw_read_lock(lock) __LOCK(lock, shared) 67#define _raw_write_lock(lock) __LOCK(lock) 68#define _raw_write_lock_nested(lock, subclass) __LOCK(lock) 69#define _raw_spin_lock_bh(lock) __LOCK_BH(lock) 70#define _raw_read_lock_bh(lock) __LOCK_BH(lock, shared) 71#define _raw_write_lock_bh(lock) __LOCK_BH(lock) 72#define _raw_spin_lock_irq(lock) __LOCK_IRQ(lock) 73#define _raw_read_lock_irq(lock) __LOCK_IRQ(lock, shared) 74#define _raw_write_lock_irq(lock) __LOCK_IRQ(lock) 75#define _raw_spin_lock_irqsave(lock, flags) __LOCK_IRQSAVE(lock, flags) 76#define _raw_read_lock_irqsave(lock, flags) __LOCK_IRQSAVE(lock, flags, shared) 77#define _raw_write_lock_irqsave(lock, flags) __LOCK_IRQSAVE(lock, flags) 78 79static __always_inline int _raw_spin_trylock(raw_spinlock_t *lock) 80 __cond_acquires(true, lock) 81{ 82 __LOCK(lock); 83 return 1; 84} 85 86static __always_inline int _raw_spin_trylock_bh(raw_spinlock_t *lock) 87 __cond_acquires(true, lock) 88{ 89 __LOCK_BH(lock); 90 return 1; 91} 92 93static __always_inline int _raw_spin_trylock_irq(raw_spinlock_t *lock) 94 __cond_acquires(true, lock) 95{ 96 __LOCK_IRQ(lock); 97 return 1; 98} 99 100static __always_inline int _raw_spin_trylock_irqsave(raw_spinlock_t *lock, unsigned long *flags) 101 __cond_acquires(true, lock) 102{ 103 __LOCK_IRQSAVE(lock, *(flags)); 104 return 1; 105} 106 107static __always_inline int _raw_read_trylock(rwlock_t *lock) 108 __cond_acquires_shared(true, lock) 109{ 110 __LOCK(lock, shared); 111 return 1; 112} 113 114static __always_inline int _raw_write_trylock(rwlock_t *lock) 115 __cond_acquires(true, lock) 116{ 117 __LOCK(lock); 118 return 1; 119} 120 121static __always_inline int _raw_write_trylock_irqsave(rwlock_t *lock, unsigned long *flags) 122 __cond_acquires(true, lock) 123{ 124 __LOCK_IRQSAVE(lock, *(flags)); 125 return 1; 126} 127 128#define _raw_spin_unlock(lock) __UNLOCK(lock) 129#define _raw_read_unlock(lock) __UNLOCK(lock, shared) 130#define _raw_write_unlock(lock) __UNLOCK(lock) 131#define _raw_spin_unlock_bh(lock) __UNLOCK_BH(lock) 132#define _raw_write_unlock_bh(lock) __UNLOCK_BH(lock) 133#define _raw_read_unlock_bh(lock) __UNLOCK_BH(lock, shared) 134#define _raw_spin_unlock_irq(lock) __UNLOCK_IRQ(lock) 135#define _raw_read_unlock_irq(lock) __UNLOCK_IRQ(lock, shared) 136#define _raw_write_unlock_irq(lock) __UNLOCK_IRQ(lock) 137#define _raw_spin_unlock_irqrestore(lock, flags) \ 138 __UNLOCK_IRQRESTORE(lock, flags) 139#define _raw_read_unlock_irqrestore(lock, flags) \ 140 __UNLOCK_IRQRESTORE(lock, flags, shared) 141#define _raw_write_unlock_irqrestore(lock, flags) \ 142 __UNLOCK_IRQRESTORE(lock, flags) 143 144#endif /* __LINUX_SPINLOCK_API_UP_H */