134 lines
		
	
	
	
		
			3.6 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
		
		
			
		
	
	
			134 lines
		
	
	
	
		
			3.6 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
| 
								 | 
							
								/*
							 | 
						||
| 
								 | 
							
								 * Queue read/write lock
							 | 
						||
| 
								 | 
							
								 *
							 | 
						||
| 
								 | 
							
								 * This program is free software; you can redistribute it and/or modify
							 | 
						||
| 
								 | 
							
								 * it under the terms of the GNU General Public License as published by
							 | 
						||
| 
								 | 
							
								 * the Free Software Foundation; either version 2 of the License, or
							 | 
						||
| 
								 | 
							
								 * (at your option) any later version.
							 | 
						||
| 
								 | 
							
								 *
							 | 
						||
| 
								 | 
							
								 * This program is distributed in the hope that it will be useful,
							 | 
						||
| 
								 | 
							
								 * but WITHOUT ANY WARRANTY; without even the implied warranty of
							 | 
						||
| 
								 | 
							
								 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
							 | 
						||
| 
								 | 
							
								 * GNU General Public License for more details.
							 | 
						||
| 
								 | 
							
								 *
							 | 
						||
| 
								 | 
							
								 * (C) Copyright 2013-2014 Hewlett-Packard Development Company, L.P.
							 | 
						||
| 
								 | 
							
								 *
							 | 
						||
| 
								 | 
							
								 * Authors: Waiman Long <waiman.long@hp.com>
							 | 
						||
| 
								 | 
							
								 */
							 | 
						||
| 
								 | 
							
								#include <linux/smp.h>
							 | 
						||
| 
								 | 
							
								#include <linux/bug.h>
							 | 
						||
| 
								 | 
							
								#include <linux/cpumask.h>
							 | 
						||
| 
								 | 
							
								#include <linux/percpu.h>
							 | 
						||
| 
								 | 
							
								#include <linux/hardirq.h>
							 | 
						||
| 
								 | 
							
								#include <linux/mutex.h>
							 | 
						||
| 
								 | 
							
								#include <asm/qrwlock.h>
							 | 
						||
| 
								 | 
							
								
							 | 
						||
| 
								 | 
							
								/**
							 | 
						||
| 
								 | 
							
								 * rspin_until_writer_unlock - inc reader count & spin until writer is gone
							 | 
						||
| 
								 | 
							
								 * @lock  : Pointer to queue rwlock structure
							 | 
						||
| 
								 | 
							
								 * @writer: Current queue rwlock writer status byte
							 | 
						||
| 
								 | 
							
								 *
							 | 
						||
| 
								 | 
							
								 * In interrupt context or at the head of the queue, the reader will just
							 | 
						||
| 
								 | 
							
								 * increment the reader count & wait until the writer releases the lock.
							 | 
						||
| 
								 | 
							
								 */
							 | 
						||
| 
								 | 
							
								static __always_inline void
							 | 
						||
| 
								 | 
							
								rspin_until_writer_unlock(struct qrwlock *lock, u32 cnts)
							 | 
						||
| 
								 | 
							
								{
							 | 
						||
| 
								 | 
							
									while ((cnts & _QW_WMASK) == _QW_LOCKED) {
							 | 
						||
| 
								 | 
							
										arch_mutex_cpu_relax();
							 | 
						||
| 
								 | 
							
										cnts = smp_load_acquire((u32 *)&lock->cnts);
							 | 
						||
| 
								 | 
							
									}
							 | 
						||
| 
								 | 
							
								}
							 | 
						||
| 
								 | 
							
								
							 | 
						||
| 
								 | 
							
								/**
							 | 
						||
| 
								 | 
							
								 * queue_read_lock_slowpath - acquire read lock of a queue rwlock
							 | 
						||
| 
								 | 
							
								 * @lock: Pointer to queue rwlock structure
							 | 
						||
| 
								 | 
							
								 */
							 | 
						||
| 
								 | 
							
								void queue_read_lock_slowpath(struct qrwlock *lock)
							 | 
						||
| 
								 | 
							
								{
							 | 
						||
| 
								 | 
							
									u32 cnts;
							 | 
						||
| 
								 | 
							
								
							 | 
						||
| 
								 | 
							
									/*
							 | 
						||
| 
								 | 
							
									 * Readers come here when they cannot get the lock without waiting
							 | 
						||
| 
								 | 
							
									 */
							 | 
						||
| 
								 | 
							
									if (unlikely(in_interrupt())) {
							 | 
						||
| 
								 | 
							
										/*
							 | 
						||
| 
								 | 
							
										 * Readers in interrupt context will spin until the lock is
							 | 
						||
| 
								 | 
							
										 * available without waiting in the queue.
							 | 
						||
| 
								 | 
							
										 */
							 | 
						||
| 
								 | 
							
										cnts = smp_load_acquire((u32 *)&lock->cnts);
							 | 
						||
| 
								 | 
							
										rspin_until_writer_unlock(lock, cnts);
							 | 
						||
| 
								 | 
							
										return;
							 | 
						||
| 
								 | 
							
									}
							 | 
						||
| 
								 | 
							
									atomic_sub(_QR_BIAS, &lock->cnts);
							 | 
						||
| 
								 | 
							
								
							 | 
						||
| 
								 | 
							
									/*
							 | 
						||
| 
								 | 
							
									 * Put the reader into the wait queue
							 | 
						||
| 
								 | 
							
									 */
							 | 
						||
| 
								 | 
							
									arch_spin_lock(&lock->lock);
							 | 
						||
| 
								 | 
							
								
							 | 
						||
| 
								 | 
							
									/*
							 | 
						||
| 
								 | 
							
									 * At the head of the wait queue now, wait until the writer state
							 | 
						||
| 
								 | 
							
									 * goes to 0 and then try to increment the reader count and get
							 | 
						||
| 
								 | 
							
									 * the lock. It is possible that an incoming writer may steal the
							 | 
						||
| 
								 | 
							
									 * lock in the interim, so it is necessary to check the writer byte
							 | 
						||
| 
								 | 
							
									 * to make sure that the write lock isn't taken.
							 | 
						||
| 
								 | 
							
									 */
							 | 
						||
| 
								 | 
							
									while (atomic_read(&lock->cnts) & _QW_WMASK)
							 | 
						||
| 
								 | 
							
										arch_mutex_cpu_relax();
							 | 
						||
| 
								 | 
							
								
							 | 
						||
| 
								 | 
							
									cnts = atomic_add_return(_QR_BIAS, &lock->cnts) - _QR_BIAS;
							 | 
						||
| 
								 | 
							
									rspin_until_writer_unlock(lock, cnts);
							 | 
						||
| 
								 | 
							
								
							 | 
						||
| 
								 | 
							
									/*
							 | 
						||
| 
								 | 
							
									 * Signal the next one in queue to become queue head
							 | 
						||
| 
								 | 
							
									 */
							 | 
						||
| 
								 | 
							
									arch_spin_unlock(&lock->lock);
							 | 
						||
| 
								 | 
							
								}
							 | 
						||
| 
								 | 
							
								EXPORT_SYMBOL(queue_read_lock_slowpath);
							 | 
						||
| 
								 | 
							
								
							 | 
						||
| 
								 | 
							
								/**
							 | 
						||
| 
								 | 
							
								 * queue_write_lock_slowpath - acquire write lock of a queue rwlock
							 | 
						||
| 
								 | 
							
								 * @lock : Pointer to queue rwlock structure
							 | 
						||
| 
								 | 
							
								 */
							 | 
						||
| 
								 | 
							
								void queue_write_lock_slowpath(struct qrwlock *lock)
							 | 
						||
| 
								 | 
							
								{
							 | 
						||
| 
								 | 
							
									u32 cnts;
							 | 
						||
| 
								 | 
							
								
							 | 
						||
| 
								 | 
							
									/* Put the writer into the wait queue */
							 | 
						||
| 
								 | 
							
									arch_spin_lock(&lock->lock);
							 | 
						||
| 
								 | 
							
								
							 | 
						||
| 
								 | 
							
									/* Try to acquire the lock directly if no reader is present */
							 | 
						||
| 
								 | 
							
									if (!atomic_read(&lock->cnts) &&
							 | 
						||
| 
								 | 
							
									    (atomic_cmpxchg(&lock->cnts, 0, _QW_LOCKED) == 0))
							 | 
						||
| 
								 | 
							
										goto unlock;
							 | 
						||
| 
								 | 
							
								
							 | 
						||
| 
								 | 
							
									/*
							 | 
						||
| 
								 | 
							
									 * Set the waiting flag to notify readers that a writer is pending,
							 | 
						||
| 
								 | 
							
									 * or wait for a previous writer to go away.
							 | 
						||
| 
								 | 
							
									 */
							 | 
						||
| 
								 | 
							
									for (;;) {
							 | 
						||
| 
								 | 
							
										cnts = atomic_read(&lock->cnts);
							 | 
						||
| 
								 | 
							
										if (!(cnts & _QW_WMASK) &&
							 | 
						||
| 
								 | 
							
										    (atomic_cmpxchg(&lock->cnts, cnts,
							 | 
						||
| 
								 | 
							
												    cnts | _QW_WAITING) == cnts))
							 | 
						||
| 
								 | 
							
											break;
							 | 
						||
| 
								 | 
							
								
							 | 
						||
| 
								 | 
							
										arch_mutex_cpu_relax();
							 | 
						||
| 
								 | 
							
									}
							 | 
						||
| 
								 | 
							
								
							 | 
						||
| 
								 | 
							
									/* When no more readers, set the locked flag */
							 | 
						||
| 
								 | 
							
									for (;;) {
							 | 
						||
| 
								 | 
							
										cnts = atomic_read(&lock->cnts);
							 | 
						||
| 
								 | 
							
										if ((cnts == _QW_WAITING) &&
							 | 
						||
| 
								 | 
							
										    (atomic_cmpxchg(&lock->cnts, _QW_WAITING,
							 | 
						||
| 
								 | 
							
												    _QW_LOCKED) == _QW_WAITING))
							 | 
						||
| 
								 | 
							
											break;
							 | 
						||
| 
								 | 
							
								
							 | 
						||
| 
								 | 
							
										arch_mutex_cpu_relax();
							 | 
						||
| 
								 | 
							
									}
							 | 
						||
| 
								 | 
							
								unlock:
							 | 
						||
| 
								 | 
							
									arch_spin_unlock(&lock->lock);
							 | 
						||
| 
								 | 
							
								}
							 | 
						||
| 
								 | 
							
								EXPORT_SYMBOL(queue_write_lock_slowpath);
							 |