Sam Ravnborg did the build-test that the direct header file move works, I'm just committing it. This is a pure move: mkdir arch/alpha/include git mv include/asm-alpha arch/alpha/include/asm with no other changes. Requested-and-tested-by: Sam Ravnborg <sam@ravnborg.org> Cc: Richard Henderson <rth@twiddle.net> Cc: Ivan Kokshaysky <ink@jurassic.park.msu.ru> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
		
			
				
	
	
		
			259 lines
		
	
	
	
		
			5.6 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			259 lines
		
	
	
	
		
			5.6 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
#ifndef _ALPHA_RWSEM_H
 | 
						|
#define _ALPHA_RWSEM_H
 | 
						|
 | 
						|
/*
 | 
						|
 * Written by Ivan Kokshaysky <ink@jurassic.park.msu.ru>, 2001.
 | 
						|
 * Based on asm-alpha/semaphore.h and asm-i386/rwsem.h
 | 
						|
 */
 | 
						|
 | 
						|
#ifndef _LINUX_RWSEM_H
 | 
						|
#error "please don't include asm/rwsem.h directly, use linux/rwsem.h instead"
 | 
						|
#endif
 | 
						|
 | 
						|
#ifdef __KERNEL__
 | 
						|
 | 
						|
#include <linux/compiler.h>
 | 
						|
#include <linux/list.h>
 | 
						|
#include <linux/spinlock.h>
 | 
						|
 | 
						|
struct rwsem_waiter;
 | 
						|
 | 
						|
extern struct rw_semaphore *rwsem_down_read_failed(struct rw_semaphore *sem);
 | 
						|
extern struct rw_semaphore *rwsem_down_write_failed(struct rw_semaphore *sem);
 | 
						|
extern struct rw_semaphore *rwsem_wake(struct rw_semaphore *);
 | 
						|
extern struct rw_semaphore *rwsem_downgrade_wake(struct rw_semaphore *sem);
 | 
						|
 | 
						|
/*
 | 
						|
 * the semaphore definition
 | 
						|
 */
 | 
						|
struct rw_semaphore {
 | 
						|
	long			count;
 | 
						|
#define RWSEM_UNLOCKED_VALUE		0x0000000000000000L
 | 
						|
#define RWSEM_ACTIVE_BIAS		0x0000000000000001L
 | 
						|
#define RWSEM_ACTIVE_MASK		0x00000000ffffffffL
 | 
						|
#define RWSEM_WAITING_BIAS		(-0x0000000100000000L)
 | 
						|
#define RWSEM_ACTIVE_READ_BIAS		RWSEM_ACTIVE_BIAS
 | 
						|
#define RWSEM_ACTIVE_WRITE_BIAS		(RWSEM_WAITING_BIAS + RWSEM_ACTIVE_BIAS)
 | 
						|
	spinlock_t		wait_lock;
 | 
						|
	struct list_head	wait_list;
 | 
						|
};
 | 
						|
 | 
						|
#define __RWSEM_INITIALIZER(name) \
 | 
						|
	{ RWSEM_UNLOCKED_VALUE, SPIN_LOCK_UNLOCKED, \
 | 
						|
	LIST_HEAD_INIT((name).wait_list) }
 | 
						|
 | 
						|
#define DECLARE_RWSEM(name) \
 | 
						|
	struct rw_semaphore name = __RWSEM_INITIALIZER(name)
 | 
						|
 | 
						|
static inline void init_rwsem(struct rw_semaphore *sem)
 | 
						|
{
 | 
						|
	sem->count = RWSEM_UNLOCKED_VALUE;
 | 
						|
	spin_lock_init(&sem->wait_lock);
 | 
						|
	INIT_LIST_HEAD(&sem->wait_list);
 | 
						|
}
 | 
						|
 | 
						|
static inline void __down_read(struct rw_semaphore *sem)
 | 
						|
{
 | 
						|
	long oldcount;
 | 
						|
#ifndef	CONFIG_SMP
 | 
						|
	oldcount = sem->count;
 | 
						|
	sem->count += RWSEM_ACTIVE_READ_BIAS;
 | 
						|
#else
 | 
						|
	long temp;
 | 
						|
	__asm__ __volatile__(
 | 
						|
	"1:	ldq_l	%0,%1\n"
 | 
						|
	"	addq	%0,%3,%2\n"
 | 
						|
	"	stq_c	%2,%1\n"
 | 
						|
	"	beq	%2,2f\n"
 | 
						|
	"	mb\n"
 | 
						|
	".subsection 2\n"
 | 
						|
	"2:	br	1b\n"
 | 
						|
	".previous"
 | 
						|
	:"=&r" (oldcount), "=m" (sem->count), "=&r" (temp)
 | 
						|
	:"Ir" (RWSEM_ACTIVE_READ_BIAS), "m" (sem->count) : "memory");
 | 
						|
#endif
 | 
						|
	if (unlikely(oldcount < 0))
 | 
						|
		rwsem_down_read_failed(sem);
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * trylock for reading -- returns 1 if successful, 0 if contention
 | 
						|
 */
 | 
						|
static inline int __down_read_trylock(struct rw_semaphore *sem)
 | 
						|
{
 | 
						|
	long old, new, res;
 | 
						|
 | 
						|
	res = sem->count;
 | 
						|
	do {
 | 
						|
		new = res + RWSEM_ACTIVE_READ_BIAS;
 | 
						|
		if (new <= 0)
 | 
						|
			break;
 | 
						|
		old = res;
 | 
						|
		res = cmpxchg(&sem->count, old, new);
 | 
						|
	} while (res != old);
 | 
						|
	return res >= 0 ? 1 : 0;
 | 
						|
}
 | 
						|
 | 
						|
static inline void __down_write(struct rw_semaphore *sem)
 | 
						|
{
 | 
						|
	long oldcount;
 | 
						|
#ifndef	CONFIG_SMP
 | 
						|
	oldcount = sem->count;
 | 
						|
	sem->count += RWSEM_ACTIVE_WRITE_BIAS;
 | 
						|
#else
 | 
						|
	long temp;
 | 
						|
	__asm__ __volatile__(
 | 
						|
	"1:	ldq_l	%0,%1\n"
 | 
						|
	"	addq	%0,%3,%2\n"
 | 
						|
	"	stq_c	%2,%1\n"
 | 
						|
	"	beq	%2,2f\n"
 | 
						|
	"	mb\n"
 | 
						|
	".subsection 2\n"
 | 
						|
	"2:	br	1b\n"
 | 
						|
	".previous"
 | 
						|
	:"=&r" (oldcount), "=m" (sem->count), "=&r" (temp)
 | 
						|
	:"Ir" (RWSEM_ACTIVE_WRITE_BIAS), "m" (sem->count) : "memory");
 | 
						|
#endif
 | 
						|
	if (unlikely(oldcount))
 | 
						|
		rwsem_down_write_failed(sem);
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * trylock for writing -- returns 1 if successful, 0 if contention
 | 
						|
 */
 | 
						|
static inline int __down_write_trylock(struct rw_semaphore *sem)
 | 
						|
{
 | 
						|
	long ret = cmpxchg(&sem->count, RWSEM_UNLOCKED_VALUE,
 | 
						|
			   RWSEM_ACTIVE_WRITE_BIAS);
 | 
						|
	if (ret == RWSEM_UNLOCKED_VALUE)
 | 
						|
		return 1;
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static inline void __up_read(struct rw_semaphore *sem)
 | 
						|
{
 | 
						|
	long oldcount;
 | 
						|
#ifndef	CONFIG_SMP
 | 
						|
	oldcount = sem->count;
 | 
						|
	sem->count -= RWSEM_ACTIVE_READ_BIAS;
 | 
						|
#else
 | 
						|
	long temp;
 | 
						|
	__asm__ __volatile__(
 | 
						|
	"	mb\n"
 | 
						|
	"1:	ldq_l	%0,%1\n"
 | 
						|
	"	subq	%0,%3,%2\n"
 | 
						|
	"	stq_c	%2,%1\n"
 | 
						|
	"	beq	%2,2f\n"
 | 
						|
	".subsection 2\n"
 | 
						|
	"2:	br	1b\n"
 | 
						|
	".previous"
 | 
						|
	:"=&r" (oldcount), "=m" (sem->count), "=&r" (temp)
 | 
						|
	:"Ir" (RWSEM_ACTIVE_READ_BIAS), "m" (sem->count) : "memory");
 | 
						|
#endif
 | 
						|
	if (unlikely(oldcount < 0))
 | 
						|
		if ((int)oldcount - RWSEM_ACTIVE_READ_BIAS == 0)
 | 
						|
			rwsem_wake(sem);
 | 
						|
}
 | 
						|
 | 
						|
static inline void __up_write(struct rw_semaphore *sem)
 | 
						|
{
 | 
						|
	long count;
 | 
						|
#ifndef	CONFIG_SMP
 | 
						|
	sem->count -= RWSEM_ACTIVE_WRITE_BIAS;
 | 
						|
	count = sem->count;
 | 
						|
#else
 | 
						|
	long temp;
 | 
						|
	__asm__ __volatile__(
 | 
						|
	"	mb\n"
 | 
						|
	"1:	ldq_l	%0,%1\n"
 | 
						|
	"	subq	%0,%3,%2\n"
 | 
						|
	"	stq_c	%2,%1\n"
 | 
						|
	"	beq	%2,2f\n"
 | 
						|
	"	subq	%0,%3,%0\n"
 | 
						|
	".subsection 2\n"
 | 
						|
	"2:	br	1b\n"
 | 
						|
	".previous"
 | 
						|
	:"=&r" (count), "=m" (sem->count), "=&r" (temp)
 | 
						|
	:"Ir" (RWSEM_ACTIVE_WRITE_BIAS), "m" (sem->count) : "memory");
 | 
						|
#endif
 | 
						|
	if (unlikely(count))
 | 
						|
		if ((int)count == 0)
 | 
						|
			rwsem_wake(sem);
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * downgrade write lock to read lock
 | 
						|
 */
 | 
						|
static inline void __downgrade_write(struct rw_semaphore *sem)
 | 
						|
{
 | 
						|
	long oldcount;
 | 
						|
#ifndef	CONFIG_SMP
 | 
						|
	oldcount = sem->count;
 | 
						|
	sem->count -= RWSEM_WAITING_BIAS;
 | 
						|
#else
 | 
						|
	long temp;
 | 
						|
	__asm__ __volatile__(
 | 
						|
	"1:	ldq_l	%0,%1\n"
 | 
						|
	"	addq	%0,%3,%2\n"
 | 
						|
	"	stq_c	%2,%1\n"
 | 
						|
	"	beq	%2,2f\n"
 | 
						|
	"	mb\n"
 | 
						|
	".subsection 2\n"
 | 
						|
	"2:	br	1b\n"
 | 
						|
	".previous"
 | 
						|
	:"=&r" (oldcount), "=m" (sem->count), "=&r" (temp)
 | 
						|
	:"Ir" (-RWSEM_WAITING_BIAS), "m" (sem->count) : "memory");
 | 
						|
#endif
 | 
						|
	if (unlikely(oldcount < 0))
 | 
						|
		rwsem_downgrade_wake(sem);
 | 
						|
}
 | 
						|
 | 
						|
static inline void rwsem_atomic_add(long val, struct rw_semaphore *sem)
 | 
						|
{
 | 
						|
#ifndef	CONFIG_SMP
 | 
						|
	sem->count += val;
 | 
						|
#else
 | 
						|
	long temp;
 | 
						|
	__asm__ __volatile__(
 | 
						|
	"1:	ldq_l	%0,%1\n"
 | 
						|
	"	addq	%0,%2,%0\n"
 | 
						|
	"	stq_c	%0,%1\n"
 | 
						|
	"	beq	%0,2f\n"
 | 
						|
	".subsection 2\n"
 | 
						|
	"2:	br	1b\n"
 | 
						|
	".previous"
 | 
						|
	:"=&r" (temp), "=m" (sem->count)
 | 
						|
	:"Ir" (val), "m" (sem->count));
 | 
						|
#endif
 | 
						|
}
 | 
						|
 | 
						|
static inline long rwsem_atomic_update(long val, struct rw_semaphore *sem)
 | 
						|
{
 | 
						|
#ifndef	CONFIG_SMP
 | 
						|
	sem->count += val;
 | 
						|
	return sem->count;
 | 
						|
#else
 | 
						|
	long ret, temp;
 | 
						|
	__asm__ __volatile__(
 | 
						|
	"1:	ldq_l	%0,%1\n"
 | 
						|
	"	addq 	%0,%3,%2\n"
 | 
						|
	"	addq	%0,%3,%0\n"
 | 
						|
	"	stq_c	%2,%1\n"
 | 
						|
	"	beq	%2,2f\n"
 | 
						|
	".subsection 2\n"
 | 
						|
	"2:	br	1b\n"
 | 
						|
	".previous"
 | 
						|
	:"=&r" (ret), "=m" (sem->count), "=&r" (temp)
 | 
						|
	:"Ir" (val), "m" (sem->count));
 | 
						|
 | 
						|
	return ret;
 | 
						|
#endif
 | 
						|
}
 | 
						|
 | 
						|
static inline int rwsem_is_locked(struct rw_semaphore *sem)
 | 
						|
{
 | 
						|
	return (sem->count != 0);
 | 
						|
}
 | 
						|
 | 
						|
#endif /* __KERNEL__ */
 | 
						|
#endif /* _ALPHA_RWSEM_H */
 |