Pass clocksource pointer to the read() callback for clocksources. This allows us to share the callback between multiple instances. [hugh@veritas.com: fix powerpc build of clocksource pass clocksource mods] [akpm@linux-foundation.org: cleanup] Signed-off-by: Magnus Damm <damm@igel.co.jp> Acked-by: John Stultz <johnstul@us.ibm.com> Cc: Thomas Gleixner <tglx@linutronix.de> Signed-off-by: Hugh Dickins <hugh@veritas.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
		
			
				
	
	
		
			297 lines
		
	
	
	
		
			7 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			297 lines
		
	
	
	
		
			7 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
 * arch/sh/kernel/timers/timer-tmu.c - TMU Timer Support
 | 
						|
 *
 | 
						|
 *  Copyright (C) 2005 - 2007  Paul Mundt
 | 
						|
 *
 | 
						|
 * TMU handling code hacked out of arch/sh/kernel/time.c
 | 
						|
 *
 | 
						|
 *  Copyright (C) 1999  Tetsuya Okada & Niibe Yutaka
 | 
						|
 *  Copyright (C) 2000  Philipp Rumpf <prumpf@tux.org>
 | 
						|
 *  Copyright (C) 2002, 2003, 2004  Paul Mundt
 | 
						|
 *  Copyright (C) 2002  M. R. Brown  <mrbrown@linux-sh.org>
 | 
						|
 *
 | 
						|
 * This file is subject to the terms and conditions of the GNU General Public
 | 
						|
 * License.  See the file "COPYING" in the main directory of this archive
 | 
						|
 * for more details.
 | 
						|
 */
 | 
						|
#include <linux/init.h>
 | 
						|
#include <linux/kernel.h>
 | 
						|
#include <linux/interrupt.h>
 | 
						|
#include <linux/seqlock.h>
 | 
						|
#include <linux/clockchips.h>
 | 
						|
#include <asm/timer.h>
 | 
						|
#include <asm/rtc.h>
 | 
						|
#include <asm/io.h>
 | 
						|
#include <asm/irq.h>
 | 
						|
#include <asm/clock.h>
 | 
						|
 | 
						|
#define TMU_TOCR_INIT	0x00
 | 
						|
#define TMU_TCR_INIT	0x0020
 | 
						|
 | 
						|
#define TMU0		(0)
 | 
						|
#define TMU1		(1)
 | 
						|
 | 
						|
static inline void _tmu_start(int tmu_num)
 | 
						|
{
 | 
						|
	ctrl_outb(ctrl_inb(TMU_012_TSTR) | (0x1<<tmu_num), TMU_012_TSTR);
 | 
						|
}
 | 
						|
 | 
						|
static inline void _tmu_set_irq(int tmu_num, int enabled)
 | 
						|
{
 | 
						|
	register unsigned long tmu_tcr = TMU0_TCR + (0xc*tmu_num);
 | 
						|
	ctrl_outw( (enabled ? ctrl_inw(tmu_tcr) | (1<<5) : ctrl_inw(tmu_tcr) & ~(1<<5)), tmu_tcr);
 | 
						|
}
 | 
						|
 | 
						|
static inline void _tmu_stop(int tmu_num)
 | 
						|
{
 | 
						|
	ctrl_outb(ctrl_inb(TMU_012_TSTR) & ~(0x1<<tmu_num), TMU_012_TSTR);
 | 
						|
}
 | 
						|
 | 
						|
static inline void _tmu_clear_status(int tmu_num)
 | 
						|
{
 | 
						|
	register unsigned long tmu_tcr = TMU0_TCR + (0xc*tmu_num);
 | 
						|
	/* Clear UNF bit */
 | 
						|
	ctrl_outw(ctrl_inw(tmu_tcr) & ~0x100, tmu_tcr);
 | 
						|
}
 | 
						|
 | 
						|
static inline unsigned long _tmu_read(int tmu_num)
 | 
						|
{
 | 
						|
        return ctrl_inl(TMU0_TCNT+0xC*tmu_num);
 | 
						|
}
 | 
						|
 | 
						|
static int tmu_timer_start(void)
 | 
						|
{
 | 
						|
	_tmu_start(TMU0);
 | 
						|
	_tmu_start(TMU1);
 | 
						|
	_tmu_set_irq(TMU0,1);
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static int tmu_timer_stop(void)
 | 
						|
{
 | 
						|
	_tmu_stop(TMU0);
 | 
						|
	_tmu_stop(TMU1);
 | 
						|
	_tmu_clear_status(TMU0);
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * also when the module_clk is scaled the TMU1
 | 
						|
 * will show the same frequency
 | 
						|
 */
 | 
						|
static int tmus_are_scaled;
 | 
						|
 | 
						|
static cycle_t tmu_timer_read(struct clocksource *cs)
 | 
						|
{
 | 
						|
	return ((cycle_t)(~_tmu_read(TMU1)))<<tmus_are_scaled;
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
static unsigned long tmu_latest_interval[3];
 | 
						|
static void tmu_timer_set_interval(int tmu_num, unsigned long interval, unsigned int reload)
 | 
						|
{
 | 
						|
	unsigned long tmu_tcnt = TMU0_TCNT + tmu_num*0xC;
 | 
						|
	unsigned long tmu_tcor = TMU0_TCOR + tmu_num*0xC;
 | 
						|
 | 
						|
	_tmu_stop(tmu_num);
 | 
						|
 | 
						|
	ctrl_outl(interval, tmu_tcnt);
 | 
						|
	tmu_latest_interval[tmu_num] = interval;
 | 
						|
 | 
						|
	/*
 | 
						|
	 * TCNT reloads from TCOR on underflow, clear it if we don't
 | 
						|
	 * intend to auto-reload
 | 
						|
	 */
 | 
						|
	ctrl_outl( reload ? interval : 0 , tmu_tcor);
 | 
						|
 | 
						|
	_tmu_start(tmu_num);
 | 
						|
}
 | 
						|
 | 
						|
static int tmu_set_next_event(unsigned long cycles,
 | 
						|
			      struct clock_event_device *evt)
 | 
						|
{
 | 
						|
	tmu_timer_set_interval(TMU0,cycles, evt->mode == CLOCK_EVT_MODE_PERIODIC);
 | 
						|
	_tmu_set_irq(TMU0,1);
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static void tmu_set_mode(enum clock_event_mode mode,
 | 
						|
			 struct clock_event_device *evt)
 | 
						|
{
 | 
						|
	switch (mode) {
 | 
						|
	case CLOCK_EVT_MODE_PERIODIC:
 | 
						|
		ctrl_outl(tmu_latest_interval[TMU0], TMU0_TCOR);
 | 
						|
		break;
 | 
						|
	case CLOCK_EVT_MODE_ONESHOT:
 | 
						|
		ctrl_outl(0, TMU0_TCOR);
 | 
						|
		break;
 | 
						|
	case CLOCK_EVT_MODE_UNUSED:
 | 
						|
	case CLOCK_EVT_MODE_SHUTDOWN:
 | 
						|
	case CLOCK_EVT_MODE_RESUME:
 | 
						|
		break;
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
static struct clock_event_device tmu0_clockevent = {
 | 
						|
	.name		= "tmu0",
 | 
						|
	.shift		= 32,
 | 
						|
	.features	= CLOCK_EVT_FEAT_PERIODIC | CLOCK_EVT_FEAT_ONESHOT,
 | 
						|
	.set_mode	= tmu_set_mode,
 | 
						|
	.set_next_event	= tmu_set_next_event,
 | 
						|
};
 | 
						|
 | 
						|
static irqreturn_t tmu_timer_interrupt(int irq, void *dummy)
 | 
						|
{
 | 
						|
	struct clock_event_device *evt = &tmu0_clockevent;
 | 
						|
	_tmu_clear_status(TMU0);
 | 
						|
	_tmu_set_irq(TMU0,tmu0_clockevent.mode != CLOCK_EVT_MODE_ONESHOT);
 | 
						|
 | 
						|
	switch (tmu0_clockevent.mode) {
 | 
						|
	case CLOCK_EVT_MODE_ONESHOT:
 | 
						|
	case CLOCK_EVT_MODE_PERIODIC:
 | 
						|
		evt->event_handler(evt);
 | 
						|
		break;
 | 
						|
	default:
 | 
						|
		break;
 | 
						|
	}
 | 
						|
 | 
						|
	return IRQ_HANDLED;
 | 
						|
}
 | 
						|
 | 
						|
static struct irqaction tmu0_irq = {
 | 
						|
	.name		= "periodic/oneshot timer",
 | 
						|
	.handler	= tmu_timer_interrupt,
 | 
						|
	.flags		= IRQF_DISABLED | IRQF_TIMER | IRQF_IRQPOLL,
 | 
						|
};
 | 
						|
 | 
						|
static void __init tmu_clk_init(struct clk *clk)
 | 
						|
{
 | 
						|
	u8 divisor  = TMU_TCR_INIT & 0x7;
 | 
						|
	int tmu_num = clk->name[3]-'0';
 | 
						|
	ctrl_outw(TMU_TCR_INIT, TMU0_TCR+(tmu_num*0xC));
 | 
						|
	clk->rate = clk_get_rate(clk->parent) / (4 << (divisor << 1));
 | 
						|
}
 | 
						|
 | 
						|
static void tmu_clk_recalc(struct clk *clk)
 | 
						|
{
 | 
						|
	int tmu_num = clk->name[3]-'0';
 | 
						|
	unsigned long prev_rate = clk_get_rate(clk);
 | 
						|
	unsigned long flags;
 | 
						|
	u8 divisor = ctrl_inw(TMU0_TCR+tmu_num*0xC) & 0x7;
 | 
						|
	clk->rate  = clk_get_rate(clk->parent) / (4 << (divisor << 1));
 | 
						|
 | 
						|
	if(prev_rate==clk_get_rate(clk))
 | 
						|
		return;
 | 
						|
 | 
						|
	if(tmu_num)
 | 
						|
		return; /* No more work on TMU1 */
 | 
						|
 | 
						|
	local_irq_save(flags);
 | 
						|
	tmus_are_scaled = (prev_rate > clk->rate);
 | 
						|
 | 
						|
	_tmu_stop(TMU0);
 | 
						|
 | 
						|
	tmu0_clockevent.mult = div_sc(clk->rate, NSEC_PER_SEC,
 | 
						|
				tmu0_clockevent.shift);
 | 
						|
	tmu0_clockevent.max_delta_ns =
 | 
						|
			clockevent_delta2ns(-1, &tmu0_clockevent);
 | 
						|
	tmu0_clockevent.min_delta_ns =
 | 
						|
			clockevent_delta2ns(1, &tmu0_clockevent);
 | 
						|
 | 
						|
	if (tmus_are_scaled)
 | 
						|
		tmu_latest_interval[TMU0] >>= 1;
 | 
						|
	else
 | 
						|
		tmu_latest_interval[TMU0] <<= 1;
 | 
						|
 | 
						|
	tmu_timer_set_interval(TMU0,
 | 
						|
		tmu_latest_interval[TMU0],
 | 
						|
		tmu0_clockevent.mode == CLOCK_EVT_MODE_PERIODIC);
 | 
						|
 | 
						|
	_tmu_start(TMU0);
 | 
						|
 | 
						|
	local_irq_restore(flags);
 | 
						|
}
 | 
						|
 | 
						|
static struct clk_ops tmu_clk_ops = {
 | 
						|
	.init		= tmu_clk_init,
 | 
						|
	.recalc		= tmu_clk_recalc,
 | 
						|
};
 | 
						|
 | 
						|
static struct clk tmu0_clk = {
 | 
						|
	.name		= "tmu0_clk",
 | 
						|
	.ops		= &tmu_clk_ops,
 | 
						|
};
 | 
						|
 | 
						|
static struct clk tmu1_clk = {
 | 
						|
	.name		= "tmu1_clk",
 | 
						|
	.ops		= &tmu_clk_ops,
 | 
						|
};
 | 
						|
 | 
						|
static int tmu_timer_init(void)
 | 
						|
{
 | 
						|
	unsigned long interval;
 | 
						|
	unsigned long frequency;
 | 
						|
 | 
						|
	setup_irq(CONFIG_SH_TIMER_IRQ, &tmu0_irq);
 | 
						|
 | 
						|
	tmu0_clk.parent = clk_get(NULL, "module_clk");
 | 
						|
	tmu1_clk.parent = clk_get(NULL, "module_clk");
 | 
						|
 | 
						|
	tmu_timer_stop();
 | 
						|
 | 
						|
#if !defined(CONFIG_CPU_SUBTYPE_SH7720) && \
 | 
						|
    !defined(CONFIG_CPU_SUBTYPE_SH7721) && \
 | 
						|
    !defined(CONFIG_CPU_SUBTYPE_SH7760) && \
 | 
						|
    !defined(CONFIG_CPU_SUBTYPE_SH7785) && \
 | 
						|
    !defined(CONFIG_CPU_SUBTYPE_SH7786) && \
 | 
						|
    !defined(CONFIG_CPU_SUBTYPE_SHX3)
 | 
						|
	ctrl_outb(TMU_TOCR_INIT, TMU_TOCR);
 | 
						|
#endif
 | 
						|
 | 
						|
	clk_register(&tmu0_clk);
 | 
						|
	clk_register(&tmu1_clk);
 | 
						|
	clk_enable(&tmu0_clk);
 | 
						|
	clk_enable(&tmu1_clk);
 | 
						|
 | 
						|
	frequency = clk_get_rate(&tmu0_clk);
 | 
						|
	interval = (frequency + HZ / 2) / HZ;
 | 
						|
 | 
						|
	tmu_timer_set_interval(TMU0,interval, 1);
 | 
						|
	tmu_timer_set_interval(TMU1,~0,1);
 | 
						|
 | 
						|
	_tmu_start(TMU1);
 | 
						|
 | 
						|
	clocksource_sh.rating = 200;
 | 
						|
	clocksource_sh.mask = CLOCKSOURCE_MASK(32);
 | 
						|
	clocksource_sh.read = tmu_timer_read;
 | 
						|
	clocksource_sh.shift = 10;
 | 
						|
	clocksource_sh.mult = clocksource_hz2mult(clk_get_rate(&tmu1_clk),
 | 
						|
						  clocksource_sh.shift);
 | 
						|
	clocksource_sh.flags = CLOCK_SOURCE_IS_CONTINUOUS;
 | 
						|
	clocksource_register(&clocksource_sh);
 | 
						|
 | 
						|
	tmu0_clockevent.mult = div_sc(frequency, NSEC_PER_SEC,
 | 
						|
				      tmu0_clockevent.shift);
 | 
						|
	tmu0_clockevent.max_delta_ns =
 | 
						|
			clockevent_delta2ns(-1, &tmu0_clockevent);
 | 
						|
	tmu0_clockevent.min_delta_ns =
 | 
						|
			clockevent_delta2ns(1, &tmu0_clockevent);
 | 
						|
 | 
						|
	tmu0_clockevent.cpumask = cpumask_of(0);
 | 
						|
	tmu0_clockevent.rating = 100;
 | 
						|
 | 
						|
	clockevents_register_device(&tmu0_clockevent);
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static struct sys_timer_ops tmu_timer_ops = {
 | 
						|
	.init		= tmu_timer_init,
 | 
						|
	.start		= tmu_timer_start,
 | 
						|
	.stop		= tmu_timer_stop,
 | 
						|
};
 | 
						|
 | 
						|
struct sys_timer tmu_timer = {
 | 
						|
	.name	= "tmu",
 | 
						|
	.ops	= &tmu_timer_ops,
 | 
						|
};
 |