Adapt to the new API. We plan to remove old cpumask APIs later. Thus this patch converts them into the new one. Signed-off-by: KOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com> Cc: David Howells <dhowells@redhat.com> Cc: Koichi Yasutake <yasutake.koichi@jp.panasonic.com> Cc: Hugh Dickins <hughd@google.com> Cc: Chris Metcalf <cmetcalf@tilera.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
		
			
				
	
	
		
			105 lines
		
	
	
	
		
			2.9 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			105 lines
		
	
	
	
		
			2.9 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
/* SMP global caching code
 | 
						|
 *
 | 
						|
 * Copyright (C) 2010 Red Hat, Inc. All Rights Reserved.
 | 
						|
 * Written by David Howells (dhowells@redhat.com)
 | 
						|
 *
 | 
						|
 * This program is free software; you can redistribute it and/or
 | 
						|
 * modify it under the terms of the GNU General Public Licence
 | 
						|
 * as published by the Free Software Foundation; either version
 | 
						|
 * 2 of the Licence, or (at your option) any later version.
 | 
						|
 */
 | 
						|
#include <linux/module.h>
 | 
						|
#include <linux/mm.h>
 | 
						|
#include <linux/mman.h>
 | 
						|
#include <linux/threads.h>
 | 
						|
#include <linux/interrupt.h>
 | 
						|
#include <asm/page.h>
 | 
						|
#include <asm/pgtable.h>
 | 
						|
#include <asm/processor.h>
 | 
						|
#include <asm/cacheflush.h>
 | 
						|
#include <asm/io.h>
 | 
						|
#include <asm/uaccess.h>
 | 
						|
#include <asm/smp.h>
 | 
						|
#include "cache-smp.h"
 | 
						|
 | 
						|
DEFINE_SPINLOCK(smp_cache_lock);
 | 
						|
static unsigned long smp_cache_mask;
 | 
						|
static unsigned long smp_cache_start;
 | 
						|
static unsigned long smp_cache_end;
 | 
						|
static cpumask_t smp_cache_ipi_map;		/* Bitmask of cache IPI done CPUs */
 | 
						|
 | 
						|
/**
 | 
						|
 * smp_cache_interrupt - Handle IPI request to flush caches.
 | 
						|
 *
 | 
						|
 * Handle a request delivered by IPI to flush the current CPU's
 | 
						|
 * caches.  The parameters are stored in smp_cache_*.
 | 
						|
 */
 | 
						|
void smp_cache_interrupt(void)
 | 
						|
{
 | 
						|
	unsigned long opr_mask = smp_cache_mask;
 | 
						|
 | 
						|
	switch ((enum smp_dcache_ops)(opr_mask & SMP_DCACHE_OP_MASK)) {
 | 
						|
	case SMP_DCACHE_NOP:
 | 
						|
		break;
 | 
						|
	case SMP_DCACHE_INV:
 | 
						|
		mn10300_local_dcache_inv();
 | 
						|
		break;
 | 
						|
	case SMP_DCACHE_INV_RANGE:
 | 
						|
		mn10300_local_dcache_inv_range(smp_cache_start, smp_cache_end);
 | 
						|
		break;
 | 
						|
	case SMP_DCACHE_FLUSH:
 | 
						|
		mn10300_local_dcache_flush();
 | 
						|
		break;
 | 
						|
	case SMP_DCACHE_FLUSH_RANGE:
 | 
						|
		mn10300_local_dcache_flush_range(smp_cache_start,
 | 
						|
						 smp_cache_end);
 | 
						|
		break;
 | 
						|
	case SMP_DCACHE_FLUSH_INV:
 | 
						|
		mn10300_local_dcache_flush_inv();
 | 
						|
		break;
 | 
						|
	case SMP_DCACHE_FLUSH_INV_RANGE:
 | 
						|
		mn10300_local_dcache_flush_inv_range(smp_cache_start,
 | 
						|
						     smp_cache_end);
 | 
						|
		break;
 | 
						|
	}
 | 
						|
 | 
						|
	switch ((enum smp_icache_ops)(opr_mask & SMP_ICACHE_OP_MASK)) {
 | 
						|
	case SMP_ICACHE_NOP:
 | 
						|
		break;
 | 
						|
	case SMP_ICACHE_INV:
 | 
						|
		mn10300_local_icache_inv();
 | 
						|
		break;
 | 
						|
	case SMP_ICACHE_INV_RANGE:
 | 
						|
		mn10300_local_icache_inv_range(smp_cache_start, smp_cache_end);
 | 
						|
		break;
 | 
						|
	}
 | 
						|
 | 
						|
	cpumask_clear_cpu(smp_processor_id(), &smp_cache_ipi_map);
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
 * smp_cache_call - Issue an IPI to request the other CPUs flush caches
 | 
						|
 * @opr_mask: Cache operation flags
 | 
						|
 * @start: Start address of request
 | 
						|
 * @end: End address of request
 | 
						|
 *
 | 
						|
 * Send cache flush IPI to other CPUs.  This invokes smp_cache_interrupt()
 | 
						|
 * above on those other CPUs and then waits for them to finish.
 | 
						|
 *
 | 
						|
 * The caller must hold smp_cache_lock.
 | 
						|
 */
 | 
						|
void smp_cache_call(unsigned long opr_mask,
 | 
						|
		    unsigned long start, unsigned long end)
 | 
						|
{
 | 
						|
	smp_cache_mask = opr_mask;
 | 
						|
	smp_cache_start = start;
 | 
						|
	smp_cache_end = end;
 | 
						|
	cpumask_copy(&smp_cache_ipi_map, cpu_online_mask);
 | 
						|
	cpumask_clear_cpu(smp_processor_id(), &smp_cache_ipi_map);
 | 
						|
 | 
						|
	send_IPI_allbutself(FLUSH_CACHE_IPI);
 | 
						|
 | 
						|
	while (!cpumask_empty(&smp_cache_ipi_map))
 | 
						|
		/* nothing. lockup detection does not belong here */
 | 
						|
		mb();
 | 
						|
}
 |