 ce3609f934
			
		
	
	
	ce3609f934
	
	
	
		
			
			Implement the new smp_mb__* ops as per the old ones. Signed-off-by: Peter Zijlstra <peterz@infradead.org> Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com> Acked-by: Chris Metcalf <cmetcalf@tilera.com> Link: http://lkml.kernel.org/n/tip-euuabnf5a3u23fy4fq8m3jcg@git.kernel.org Cc: Akinobu Mita <akinobu.mita@gmail.com> Cc: Chen Gang <gang.chen@asianux.com> Cc: Geert Uytterhoeven <geert@linux-m68k.org> Cc: Linus Torvalds <torvalds@linux-foundation.org> Cc: Mathieu Desnoyers <mathieu.desnoyers@polymtl.ca> Cc: linux-kernel@vger.kernel.org Signed-off-by: Ingo Molnar <mingo@kernel.org>
		
			
				
	
	
		
			113 lines
		
	
	
	
		
			3.2 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			113 lines
		
	
	
	
		
			3.2 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * Copyright 2011 Tilera Corporation. All Rights Reserved.
 | |
|  *
 | |
|  *   This program is free software; you can redistribute it and/or
 | |
|  *   modify it under the terms of the GNU General Public License
 | |
|  *   as published by the Free Software Foundation, version 2.
 | |
|  *
 | |
|  *   This program is distributed in the hope that it will be useful, but
 | |
|  *   WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
|  *   MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or
 | |
|  *   NON INFRINGEMENT.  See the GNU General Public License for
 | |
|  *   more details.
 | |
|  *
 | |
|  * Do not include directly; use <linux/atomic.h>.
 | |
|  */
 | |
| 
 | |
| #ifndef _ASM_TILE_ATOMIC_64_H
 | |
| #define _ASM_TILE_ATOMIC_64_H
 | |
| 
 | |
| #ifndef __ASSEMBLY__
 | |
| 
 | |
| #include <asm/barrier.h>
 | |
| #include <arch/spr_def.h>
 | |
| 
 | |
| /* First, the 32-bit atomic ops that are "real" on our 64-bit platform. */
 | |
| 
 | |
| #define atomic_set(v, i) ((v)->counter = (i))
 | |
| 
 | |
| /*
 | |
|  * The smp_mb() operations throughout are to support the fact that
 | |
|  * Linux requires memory barriers before and after the operation,
 | |
|  * on any routine which updates memory and returns a value.
 | |
|  */
 | |
| 
 | |
| static inline void atomic_add(int i, atomic_t *v)
 | |
| {
 | |
| 	__insn_fetchadd4((void *)&v->counter, i);
 | |
| }
 | |
| 
 | |
| static inline int atomic_add_return(int i, atomic_t *v)
 | |
| {
 | |
| 	int val;
 | |
| 	smp_mb();  /* barrier for proper semantics */
 | |
| 	val = __insn_fetchadd4((void *)&v->counter, i) + i;
 | |
| 	barrier();  /* the "+ i" above will wait on memory */
 | |
| 	return val;
 | |
| }
 | |
| 
 | |
| static inline int __atomic_add_unless(atomic_t *v, int a, int u)
 | |
| {
 | |
| 	int guess, oldval = v->counter;
 | |
| 	do {
 | |
| 		if (oldval == u)
 | |
| 			break;
 | |
| 		guess = oldval;
 | |
| 		oldval = cmpxchg(&v->counter, guess, guess + a);
 | |
| 	} while (guess != oldval);
 | |
| 	return oldval;
 | |
| }
 | |
| 
 | |
| /* Now the true 64-bit operations. */
 | |
| 
 | |
| #define ATOMIC64_INIT(i)	{ (i) }
 | |
| 
 | |
| #define atomic64_read(v)		((v)->counter)
 | |
| #define atomic64_set(v, i) ((v)->counter = (i))
 | |
| 
 | |
| static inline void atomic64_add(long i, atomic64_t *v)
 | |
| {
 | |
| 	__insn_fetchadd((void *)&v->counter, i);
 | |
| }
 | |
| 
 | |
| static inline long atomic64_add_return(long i, atomic64_t *v)
 | |
| {
 | |
| 	int val;
 | |
| 	smp_mb();  /* barrier for proper semantics */
 | |
| 	val = __insn_fetchadd((void *)&v->counter, i) + i;
 | |
| 	barrier();  /* the "+ i" above will wait on memory */
 | |
| 	return val;
 | |
| }
 | |
| 
 | |
| static inline long atomic64_add_unless(atomic64_t *v, long a, long u)
 | |
| {
 | |
| 	long guess, oldval = v->counter;
 | |
| 	do {
 | |
| 		if (oldval == u)
 | |
| 			break;
 | |
| 		guess = oldval;
 | |
| 		oldval = cmpxchg(&v->counter, guess, guess + a);
 | |
| 	} while (guess != oldval);
 | |
| 	return oldval != u;
 | |
| }
 | |
| 
 | |
| #define atomic64_sub_return(i, v)	atomic64_add_return(-(i), (v))
 | |
| #define atomic64_sub(i, v)		atomic64_add(-(i), (v))
 | |
| #define atomic64_inc_return(v)		atomic64_add_return(1, (v))
 | |
| #define atomic64_dec_return(v)		atomic64_sub_return(1, (v))
 | |
| #define atomic64_inc(v)			atomic64_add(1, (v))
 | |
| #define atomic64_dec(v)			atomic64_sub(1, (v))
 | |
| 
 | |
| #define atomic64_inc_and_test(v)	(atomic64_inc_return(v) == 0)
 | |
| #define atomic64_dec_and_test(v)	(atomic64_dec_return(v) == 0)
 | |
| #define atomic64_sub_and_test(i, v)	(atomic64_sub_return((i), (v)) == 0)
 | |
| #define atomic64_add_negative(i, v)	(atomic64_add_return((i), (v)) < 0)
 | |
| 
 | |
| #define atomic64_inc_not_zero(v)	atomic64_add_unless((v), 1, 0)
 | |
| 
 | |
| /* Define this to indicate that cmpxchg is an efficient operation. */
 | |
| #define __HAVE_ARCH_CMPXCHG
 | |
| 
 | |
| #endif /* !__ASSEMBLY__ */
 | |
| 
 | |
| #endif /* _ASM_TILE_ATOMIC_64_H */
 |