This rwlock uses the arch_spin_lock_t as a waitqueue, and assuming the arch_spin_lock_t is a fair lock (ticket,mcs etc..) the resulting rwlock is a fair lock. It fits in the same 8 bytes as the regular rwlock_t by folding the reader and writer count into a single integer, using the remaining 4 bytes for the arch_spinlock_t. Architectures that can single-copy adress bytes can optimize queue_write_unlock() with a 0 write to the LSB (the write count). Performance as measured by Davidlohr Bueso (rwlock_t -> qrwlock_t): +--------------+-------------+---------------+ | Workload | #users | delta | +--------------+-------------+---------------+ | alltests | > 1400 | -4.83% | | custom | 0-100,> 100 | +1.43%,-1.57% | | high_systime | > 1000 | -2.61 | | shared | all | +0.32 | +--------------+-------------+---------------+ http://www.stgolabs.net/qrwlock-stuff/aim7-results-vs-rwsem_optsin/ Signed-off-by: Waiman Long <Waiman.Long@hp.com> [peterz: near complete rewrite] Signed-off-by: Peter Zijlstra <peterz@infradead.org> Cc: Arnd Bergmann <arnd@arndb.de> Cc: Linus Torvalds <torvalds@linux-foundation.org> Cc: "Paul E.McKenney" <paulmck@linux.vnet.ibm.com> Cc: linux-arch@vger.kernel.org Cc: linux-kernel@vger.kernel.org Link: http://lkml.kernel.org/n/tip-gac1nnl3wvs2ij87zv2xkdzq@git.kernel.org Signed-off-by: Ingo Molnar <mingo@kernel.org>
		
			
				
	
	
		
			133 lines
		
	
	
	
		
			3.6 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			133 lines
		
	
	
	
		
			3.6 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
 * Queue read/write lock
 | 
						|
 *
 | 
						|
 * This program is free software; you can redistribute it and/or modify
 | 
						|
 * it under the terms of the GNU General Public License as published by
 | 
						|
 * the Free Software Foundation; either version 2 of the License, or
 | 
						|
 * (at your option) any later version.
 | 
						|
 *
 | 
						|
 * This program is distributed in the hope that it will be useful,
 | 
						|
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
						|
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | 
						|
 * GNU General Public License for more details.
 | 
						|
 *
 | 
						|
 * (C) Copyright 2013-2014 Hewlett-Packard Development Company, L.P.
 | 
						|
 *
 | 
						|
 * Authors: Waiman Long <waiman.long@hp.com>
 | 
						|
 */
 | 
						|
#include <linux/smp.h>
 | 
						|
#include <linux/bug.h>
 | 
						|
#include <linux/cpumask.h>
 | 
						|
#include <linux/percpu.h>
 | 
						|
#include <linux/hardirq.h>
 | 
						|
#include <linux/mutex.h>
 | 
						|
#include <asm/qrwlock.h>
 | 
						|
 | 
						|
/**
 | 
						|
 * rspin_until_writer_unlock - inc reader count & spin until writer is gone
 | 
						|
 * @lock  : Pointer to queue rwlock structure
 | 
						|
 * @writer: Current queue rwlock writer status byte
 | 
						|
 *
 | 
						|
 * In interrupt context or at the head of the queue, the reader will just
 | 
						|
 * increment the reader count & wait until the writer releases the lock.
 | 
						|
 */
 | 
						|
static __always_inline void
 | 
						|
rspin_until_writer_unlock(struct qrwlock *lock, u32 cnts)
 | 
						|
{
 | 
						|
	while ((cnts & _QW_WMASK) == _QW_LOCKED) {
 | 
						|
		arch_mutex_cpu_relax();
 | 
						|
		cnts = smp_load_acquire((u32 *)&lock->cnts);
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
 * queue_read_lock_slowpath - acquire read lock of a queue rwlock
 | 
						|
 * @lock: Pointer to queue rwlock structure
 | 
						|
 */
 | 
						|
void queue_read_lock_slowpath(struct qrwlock *lock)
 | 
						|
{
 | 
						|
	u32 cnts;
 | 
						|
 | 
						|
	/*
 | 
						|
	 * Readers come here when they cannot get the lock without waiting
 | 
						|
	 */
 | 
						|
	if (unlikely(in_interrupt())) {
 | 
						|
		/*
 | 
						|
		 * Readers in interrupt context will spin until the lock is
 | 
						|
		 * available without waiting in the queue.
 | 
						|
		 */
 | 
						|
		cnts = smp_load_acquire((u32 *)&lock->cnts);
 | 
						|
		rspin_until_writer_unlock(lock, cnts);
 | 
						|
		return;
 | 
						|
	}
 | 
						|
	atomic_sub(_QR_BIAS, &lock->cnts);
 | 
						|
 | 
						|
	/*
 | 
						|
	 * Put the reader into the wait queue
 | 
						|
	 */
 | 
						|
	arch_spin_lock(&lock->lock);
 | 
						|
 | 
						|
	/*
 | 
						|
	 * At the head of the wait queue now, wait until the writer state
 | 
						|
	 * goes to 0 and then try to increment the reader count and get
 | 
						|
	 * the lock. It is possible that an incoming writer may steal the
 | 
						|
	 * lock in the interim, so it is necessary to check the writer byte
 | 
						|
	 * to make sure that the write lock isn't taken.
 | 
						|
	 */
 | 
						|
	while (atomic_read(&lock->cnts) & _QW_WMASK)
 | 
						|
		arch_mutex_cpu_relax();
 | 
						|
 | 
						|
	cnts = atomic_add_return(_QR_BIAS, &lock->cnts) - _QR_BIAS;
 | 
						|
	rspin_until_writer_unlock(lock, cnts);
 | 
						|
 | 
						|
	/*
 | 
						|
	 * Signal the next one in queue to become queue head
 | 
						|
	 */
 | 
						|
	arch_spin_unlock(&lock->lock);
 | 
						|
}
 | 
						|
EXPORT_SYMBOL(queue_read_lock_slowpath);
 | 
						|
 | 
						|
/**
 | 
						|
 * queue_write_lock_slowpath - acquire write lock of a queue rwlock
 | 
						|
 * @lock : Pointer to queue rwlock structure
 | 
						|
 */
 | 
						|
void queue_write_lock_slowpath(struct qrwlock *lock)
 | 
						|
{
 | 
						|
	u32 cnts;
 | 
						|
 | 
						|
	/* Put the writer into the wait queue */
 | 
						|
	arch_spin_lock(&lock->lock);
 | 
						|
 | 
						|
	/* Try to acquire the lock directly if no reader is present */
 | 
						|
	if (!atomic_read(&lock->cnts) &&
 | 
						|
	    (atomic_cmpxchg(&lock->cnts, 0, _QW_LOCKED) == 0))
 | 
						|
		goto unlock;
 | 
						|
 | 
						|
	/*
 | 
						|
	 * Set the waiting flag to notify readers that a writer is pending,
 | 
						|
	 * or wait for a previous writer to go away.
 | 
						|
	 */
 | 
						|
	for (;;) {
 | 
						|
		cnts = atomic_read(&lock->cnts);
 | 
						|
		if (!(cnts & _QW_WMASK) &&
 | 
						|
		    (atomic_cmpxchg(&lock->cnts, cnts,
 | 
						|
				    cnts | _QW_WAITING) == cnts))
 | 
						|
			break;
 | 
						|
 | 
						|
		arch_mutex_cpu_relax();
 | 
						|
	}
 | 
						|
 | 
						|
	/* When no more readers, set the locked flag */
 | 
						|
	for (;;) {
 | 
						|
		cnts = atomic_read(&lock->cnts);
 | 
						|
		if ((cnts == _QW_WAITING) &&
 | 
						|
		    (atomic_cmpxchg(&lock->cnts, _QW_WAITING,
 | 
						|
				    _QW_LOCKED) == _QW_WAITING))
 | 
						|
			break;
 | 
						|
 | 
						|
		arch_mutex_cpu_relax();
 | 
						|
	}
 | 
						|
unlock:
 | 
						|
	arch_spin_unlock(&lock->lock);
 | 
						|
}
 | 
						|
EXPORT_SYMBOL(queue_write_lock_slowpath);
 |