| 
									
										
										
										
											2011-05-04 14:38:26 -04:00
										 |  |  | /*
 | 
					
						
							|  |  |  |  * Copyright 2011 Tilera Corporation. All Rights Reserved. | 
					
						
							|  |  |  |  * | 
					
						
							|  |  |  |  *   This program is free software; you can redistribute it and/or | 
					
						
							|  |  |  |  *   modify it under the terms of the GNU General Public License | 
					
						
							|  |  |  |  *   as published by the Free Software Foundation, version 2. | 
					
						
							|  |  |  |  * | 
					
						
							|  |  |  |  *   This program is distributed in the hope that it will be useful, but | 
					
						
							|  |  |  |  *   WITHOUT ANY WARRANTY; without even the implied warranty of | 
					
						
							|  |  |  |  *   MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or | 
					
						
							|  |  |  |  *   NON INFRINGEMENT.  See the GNU General Public License for | 
					
						
							|  |  |  |  *   more details. | 
					
						
							|  |  |  |  * | 
					
						
							|  |  |  |  * 64-bit SMP ticket spinlocks, allowing only a single CPU anywhere | 
					
						
							|  |  |  |  * (the type definitions are in asm/spinlock_types.h) | 
					
						
							|  |  |  |  */ | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | #ifndef _ASM_TILE_SPINLOCK_64_H
 | 
					
						
							|  |  |  | #define _ASM_TILE_SPINLOCK_64_H
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | /* Shifts and masks for the various fields in "lock". */ | 
					
						
							|  |  |  | #define __ARCH_SPIN_CURRENT_SHIFT	17
 | 
					
						
							|  |  |  | #define __ARCH_SPIN_NEXT_MASK		0x7fff
 | 
					
						
							|  |  |  | #define __ARCH_SPIN_NEXT_OVERFLOW	0x8000
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | /*
 | 
					
						
							|  |  |  |  * Return the "current" portion of a ticket lock value, | 
					
						
							|  |  |  |  * i.e. the number that currently owns the lock. | 
					
						
							|  |  |  |  */ | 
					
						
							| 
									
										
										
										
											2013-08-09 16:53:50 -04:00
										 |  |  | static inline u32 arch_spin_current(u32 val) | 
					
						
							| 
									
										
										
										
											2011-05-04 14:38:26 -04:00
										 |  |  | { | 
					
						
							|  |  |  | 	return val >> __ARCH_SPIN_CURRENT_SHIFT; | 
					
						
							|  |  |  | } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | /*
 | 
					
						
							|  |  |  |  * Return the "next" portion of a ticket lock value, | 
					
						
							|  |  |  |  * i.e. the number that the next task to try to acquire the lock will get. | 
					
						
							|  |  |  |  */ | 
					
						
							| 
									
										
										
										
											2013-08-09 16:53:50 -04:00
										 |  |  | static inline u32 arch_spin_next(u32 val) | 
					
						
							| 
									
										
										
										
											2011-05-04 14:38:26 -04:00
										 |  |  | { | 
					
						
							|  |  |  | 	return val & __ARCH_SPIN_NEXT_MASK; | 
					
						
							|  |  |  | } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | /* The lock is locked if a task would have to wait to get it. */ | 
					
						
							|  |  |  | static inline int arch_spin_is_locked(arch_spinlock_t *lock) | 
					
						
							|  |  |  | { | 
					
						
							|  |  |  | 	u32 val = lock->lock; | 
					
						
							|  |  |  | 	return arch_spin_current(val) != arch_spin_next(val); | 
					
						
							|  |  |  | } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | /* Bump the current ticket so the next task owns the lock. */ | 
					
						
							|  |  |  | static inline void arch_spin_unlock(arch_spinlock_t *lock) | 
					
						
							|  |  |  | { | 
					
						
							|  |  |  | 	wmb();  /* guarantee anything modified under the lock is visible */ | 
					
						
							|  |  |  | 	__insn_fetchadd4(&lock->lock, 1U << __ARCH_SPIN_CURRENT_SHIFT); | 
					
						
							|  |  |  | } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | void arch_spin_unlock_wait(arch_spinlock_t *lock); | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | void arch_spin_lock_slow(arch_spinlock_t *lock, u32 val); | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | /* Grab the "next" ticket number and bump it atomically.
 | 
					
						
							|  |  |  |  * If the current ticket is not ours, go to the slow path. | 
					
						
							|  |  |  |  * We also take the slow path if the "next" value overflows. | 
					
						
							|  |  |  |  */ | 
					
						
							|  |  |  | static inline void arch_spin_lock(arch_spinlock_t *lock) | 
					
						
							|  |  |  | { | 
					
						
							|  |  |  | 	u32 val = __insn_fetchadd4(&lock->lock, 1); | 
					
						
							|  |  |  | 	u32 ticket = val & (__ARCH_SPIN_NEXT_MASK | __ARCH_SPIN_NEXT_OVERFLOW); | 
					
						
							|  |  |  | 	if (unlikely(arch_spin_current(val) != ticket)) | 
					
						
							|  |  |  | 		arch_spin_lock_slow(lock, ticket); | 
					
						
							|  |  |  | } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | /* Try to get the lock, and return whether we succeeded. */ | 
					
						
							|  |  |  | int arch_spin_trylock(arch_spinlock_t *lock); | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | /* We cannot take an interrupt after getting a ticket, so don't enable them. */ | 
					
						
							|  |  |  | #define arch_spin_lock_flags(lock, flags) arch_spin_lock(lock)
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | /*
 | 
					
						
							|  |  |  |  * Read-write spinlocks, allowing multiple readers | 
					
						
							|  |  |  |  * but only one writer. | 
					
						
							|  |  |  |  * | 
					
						
							|  |  |  |  * We use fetchadd() for readers, and fetchor() with the sign bit | 
					
						
							|  |  |  |  * for writers. | 
					
						
							|  |  |  |  */ | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | #define __WRITE_LOCK_BIT (1 << 31)
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | static inline int arch_write_val_locked(int val) | 
					
						
							|  |  |  | { | 
					
						
							|  |  |  | 	return val < 0;  /* Optimize "val & __WRITE_LOCK_BIT". */ | 
					
						
							|  |  |  | } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | /**
 | 
					
						
							|  |  |  |  * read_can_lock - would read_trylock() succeed? | 
					
						
							|  |  |  |  * @lock: the rwlock in question. | 
					
						
							|  |  |  |  */ | 
					
						
							|  |  |  | static inline int arch_read_can_lock(arch_rwlock_t *rw) | 
					
						
							|  |  |  | { | 
					
						
							|  |  |  | 	return !arch_write_val_locked(rw->lock); | 
					
						
							|  |  |  | } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | /**
 | 
					
						
							|  |  |  |  * write_can_lock - would write_trylock() succeed? | 
					
						
							|  |  |  |  * @lock: the rwlock in question. | 
					
						
							|  |  |  |  */ | 
					
						
							|  |  |  | static inline int arch_write_can_lock(arch_rwlock_t *rw) | 
					
						
							|  |  |  | { | 
					
						
							|  |  |  | 	return rw->lock == 0; | 
					
						
							|  |  |  | } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | extern void __read_lock_failed(arch_rwlock_t *rw); | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | static inline void arch_read_lock(arch_rwlock_t *rw) | 
					
						
							|  |  |  | { | 
					
						
							|  |  |  | 	u32 val = __insn_fetchaddgez4(&rw->lock, 1); | 
					
						
							|  |  |  | 	if (unlikely(arch_write_val_locked(val))) | 
					
						
							|  |  |  | 		__read_lock_failed(rw); | 
					
						
							|  |  |  | } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | extern void __write_lock_failed(arch_rwlock_t *rw, u32 val); | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | static inline void arch_write_lock(arch_rwlock_t *rw) | 
					
						
							|  |  |  | { | 
					
						
							|  |  |  | 	u32 val = __insn_fetchor4(&rw->lock, __WRITE_LOCK_BIT); | 
					
						
							|  |  |  | 	if (unlikely(val != 0)) | 
					
						
							|  |  |  | 		__write_lock_failed(rw, val); | 
					
						
							|  |  |  | } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | static inline void arch_read_unlock(arch_rwlock_t *rw) | 
					
						
							|  |  |  | { | 
					
						
							|  |  |  | 	__insn_mf(); | 
					
						
							|  |  |  | 	__insn_fetchadd4(&rw->lock, -1); | 
					
						
							|  |  |  | } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | static inline void arch_write_unlock(arch_rwlock_t *rw) | 
					
						
							|  |  |  | { | 
					
						
							|  |  |  | 	__insn_mf(); | 
					
						
							| 
									
										
										
										
											2012-03-30 15:46:29 -04:00
										 |  |  | 	__insn_exch4(&rw->lock, 0);  /* Avoid waiting in the write buffer. */ | 
					
						
							| 
									
										
										
										
											2011-05-04 14:38:26 -04:00
										 |  |  | } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | static inline int arch_read_trylock(arch_rwlock_t *rw) | 
					
						
							|  |  |  | { | 
					
						
							|  |  |  | 	return !arch_write_val_locked(__insn_fetchaddgez4(&rw->lock, 1)); | 
					
						
							|  |  |  | } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | static inline int arch_write_trylock(arch_rwlock_t *rw) | 
					
						
							|  |  |  | { | 
					
						
							|  |  |  | 	u32 val = __insn_fetchor4(&rw->lock, __WRITE_LOCK_BIT); | 
					
						
							|  |  |  | 	if (likely(val == 0)) | 
					
						
							|  |  |  | 		return 1; | 
					
						
							|  |  |  | 	if (!arch_write_val_locked(val)) | 
					
						
							|  |  |  | 		__insn_fetchand4(&rw->lock, ~__WRITE_LOCK_BIT); | 
					
						
							|  |  |  | 	return 0; | 
					
						
							|  |  |  | } | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | #define arch_read_lock_flags(lock, flags) arch_read_lock(lock)
 | 
					
						
							|  |  |  | #define arch_write_lock_flags(lock, flags) arch_write_lock(lock)
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | #endif /* _ASM_TILE_SPINLOCK_64_H */
 |