 8b23281605
			
		
	
	
	8b23281605
	
	
	
		
			
			In debugging the failure of the glibc tst-cond18 test on parisc, I realized that futexes need to use the same locks the lws calls. This fixes all the pthread 'cond' tests. Sadly, there are still problems with thread cancellation. [jejb: checkpatch fixes] Signed-off-by: John David Anglin <dave.anglin@bell.net> Signed-off-by: James Bottomley <JBottomley@Parallels.com>
		
			
				
	
	
		
			154 lines
		
	
	
	
		
			3.5 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			154 lines
		
	
	
	
		
			3.5 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
| #ifndef _ASM_PARISC_FUTEX_H
 | |
| #define _ASM_PARISC_FUTEX_H
 | |
| 
 | |
| #ifdef __KERNEL__
 | |
| 
 | |
| #include <linux/futex.h>
 | |
| #include <linux/uaccess.h>
 | |
| #include <asm/atomic.h>
 | |
| #include <asm/errno.h>
 | |
| 
 | |
| /* The following has to match the LWS code in syscall.S.  We have
 | |
|    sixteen four-word locks. */
 | |
| 
 | |
| static inline void
 | |
| _futex_spin_lock_irqsave(u32 __user *uaddr, unsigned long int *flags)
 | |
| {
 | |
| 	extern u32 lws_lock_start[];
 | |
| 	long index = ((long)uaddr & 0xf0) >> 2;
 | |
| 	arch_spinlock_t *s = (arch_spinlock_t *)&lws_lock_start[index];
 | |
| 	local_irq_save(*flags);
 | |
| 	arch_spin_lock(s);
 | |
| }
 | |
| 
 | |
| static inline void
 | |
| _futex_spin_unlock_irqrestore(u32 __user *uaddr, unsigned long int *flags)
 | |
| {
 | |
| 	extern u32 lws_lock_start[];
 | |
| 	long index = ((long)uaddr & 0xf0) >> 2;
 | |
| 	arch_spinlock_t *s = (arch_spinlock_t *)&lws_lock_start[index];
 | |
| 	arch_spin_unlock(s);
 | |
| 	local_irq_restore(*flags);
 | |
| }
 | |
| 
 | |
| static inline int
 | |
| futex_atomic_op_inuser (int encoded_op, u32 __user *uaddr)
 | |
| {
 | |
| 	unsigned long int flags;
 | |
| 	u32 val;
 | |
| 	int op = (encoded_op >> 28) & 7;
 | |
| 	int cmp = (encoded_op >> 24) & 15;
 | |
| 	int oparg = (encoded_op << 8) >> 20;
 | |
| 	int cmparg = (encoded_op << 20) >> 20;
 | |
| 	int oldval = 0, ret;
 | |
| 	if (encoded_op & (FUTEX_OP_OPARG_SHIFT << 28))
 | |
| 		oparg = 1 << oparg;
 | |
| 
 | |
| 	if (!access_ok(VERIFY_WRITE, uaddr, sizeof(*uaddr)))
 | |
| 		return -EFAULT;
 | |
| 
 | |
| 	pagefault_disable();
 | |
| 
 | |
| 	_futex_spin_lock_irqsave(uaddr, &flags);
 | |
| 
 | |
| 	switch (op) {
 | |
| 	case FUTEX_OP_SET:
 | |
| 		/* *(int *)UADDR2 = OPARG; */
 | |
| 		ret = get_user(oldval, uaddr);
 | |
| 		if (!ret)
 | |
| 			ret = put_user(oparg, uaddr);
 | |
| 		break;
 | |
| 	case FUTEX_OP_ADD:
 | |
| 		/* *(int *)UADDR2 += OPARG; */
 | |
| 		ret = get_user(oldval, uaddr);
 | |
| 		if (!ret) {
 | |
| 			val = oldval + oparg;
 | |
| 			ret = put_user(val, uaddr);
 | |
| 		}
 | |
| 		break;
 | |
| 	case FUTEX_OP_OR:
 | |
| 		/* *(int *)UADDR2 |= OPARG; */
 | |
| 		ret = get_user(oldval, uaddr);
 | |
| 		if (!ret) {
 | |
| 			val = oldval | oparg;
 | |
| 			ret = put_user(val, uaddr);
 | |
| 		}
 | |
| 		break;
 | |
| 	case FUTEX_OP_ANDN:
 | |
| 		/* *(int *)UADDR2 &= ~OPARG; */
 | |
| 		ret = get_user(oldval, uaddr);
 | |
| 		if (!ret) {
 | |
| 			val = oldval & ~oparg;
 | |
| 			ret = put_user(val, uaddr);
 | |
| 		}
 | |
| 		break;
 | |
| 	case FUTEX_OP_XOR:
 | |
| 		/* *(int *)UADDR2 ^= OPARG; */
 | |
| 		ret = get_user(oldval, uaddr);
 | |
| 		if (!ret) {
 | |
| 			val = oldval ^ oparg;
 | |
| 			ret = put_user(val, uaddr);
 | |
| 		}
 | |
| 		break;
 | |
| 	default:
 | |
| 		ret = -ENOSYS;
 | |
| 	}
 | |
| 
 | |
| 	_futex_spin_unlock_irqrestore(uaddr, &flags);
 | |
| 
 | |
| 	pagefault_enable();
 | |
| 
 | |
| 	if (!ret) {
 | |
| 		switch (cmp) {
 | |
| 		case FUTEX_OP_CMP_EQ: ret = (oldval == cmparg); break;
 | |
| 		case FUTEX_OP_CMP_NE: ret = (oldval != cmparg); break;
 | |
| 		case FUTEX_OP_CMP_LT: ret = (oldval < cmparg); break;
 | |
| 		case FUTEX_OP_CMP_GE: ret = (oldval >= cmparg); break;
 | |
| 		case FUTEX_OP_CMP_LE: ret = (oldval <= cmparg); break;
 | |
| 		case FUTEX_OP_CMP_GT: ret = (oldval > cmparg); break;
 | |
| 		default: ret = -ENOSYS;
 | |
| 		}
 | |
| 	}
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| /* Non-atomic version */
 | |
| static inline int
 | |
| futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr,
 | |
| 			      u32 oldval, u32 newval)
 | |
| {
 | |
| 	int ret;
 | |
| 	u32 val;
 | |
| 	unsigned long flags;
 | |
| 
 | |
| 	/* futex.c wants to do a cmpxchg_inatomic on kernel NULL, which is
 | |
| 	 * our gateway page, and causes no end of trouble...
 | |
| 	 */
 | |
| 	if (segment_eq(KERNEL_DS, get_fs()) && !uaddr)
 | |
| 		return -EFAULT;
 | |
| 
 | |
| 	if (!access_ok(VERIFY_WRITE, uaddr, sizeof(u32)))
 | |
| 		return -EFAULT;
 | |
| 
 | |
| 	/* HPPA has no cmpxchg in hardware and therefore the
 | |
| 	 * best we can do here is use an array of locks. The
 | |
| 	 * lock selected is based on a hash of the userspace
 | |
| 	 * address. This should scale to a couple of CPUs.
 | |
| 	 */
 | |
| 
 | |
| 	_futex_spin_lock_irqsave(uaddr, &flags);
 | |
| 
 | |
| 	ret = get_user(val, uaddr);
 | |
| 
 | |
| 	if (!ret && val == oldval)
 | |
| 		ret = put_user(newval, uaddr);
 | |
| 
 | |
| 	*uval = val;
 | |
| 
 | |
| 	_futex_spin_unlock_irqrestore(uaddr, &flags);
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| #endif /*__KERNEL__*/
 | |
| #endif /*_ASM_PARISC_FUTEX_H*/
 |