421 lines
		
	
	
	
		
			9.4 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
		
		
			
		
	
	
			421 lines
		
	
	
	
		
			9.4 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
|   | /*
 | |||
|  |  *	linux/arch/alpha/kernel/sys_titan.c | |||
|  |  * | |||
|  |  *	Copyright (C) 1995 David A Rusling | |||
|  |  *	Copyright (C) 1996, 1999 Jay A Estabrook | |||
|  |  *	Copyright (C) 1998, 1999 Richard Henderson | |||
|  |  *      Copyright (C) 1999, 2000 Jeff Wiedemeier | |||
|  |  * | |||
|  |  * Code supporting TITAN systems (EV6+TITAN), currently: | |||
|  |  *      Privateer | |||
|  |  *	Falcon | |||
|  |  *	Granite | |||
|  |  */ | |||
|  | 
 | |||
|  | #include <linux/config.h>
 | |||
|  | #include <linux/kernel.h>
 | |||
|  | #include <linux/types.h>
 | |||
|  | #include <linux/mm.h>
 | |||
|  | #include <linux/sched.h>
 | |||
|  | #include <linux/pci.h>
 | |||
|  | #include <linux/init.h>
 | |||
|  | #include <linux/bitops.h>
 | |||
|  | 
 | |||
|  | #include <asm/ptrace.h>
 | |||
|  | #include <asm/system.h>
 | |||
|  | #include <asm/dma.h>
 | |||
|  | #include <asm/irq.h>
 | |||
|  | #include <asm/mmu_context.h>
 | |||
|  | #include <asm/io.h>
 | |||
|  | #include <asm/pgtable.h>
 | |||
|  | #include <asm/core_titan.h>
 | |||
|  | #include <asm/hwrpb.h>
 | |||
|  | #include <asm/tlbflush.h>
 | |||
|  | 
 | |||
|  | #include "proto.h"
 | |||
|  | #include "irq_impl.h"
 | |||
|  | #include "pci_impl.h"
 | |||
|  | #include "machvec_impl.h"
 | |||
|  | #include "err_impl.h"
 | |||
|  | 
 | |||
|  |  | |||
|  | /*
 | |||
|  |  * Titan generic | |||
|  |  */ | |||
|  | 
 | |||
|  | /*
 | |||
|  |  * Titan supports up to 4 CPUs | |||
|  |  */ | |||
|  | static unsigned long titan_cpu_irq_affinity[4] = { ~0UL, ~0UL, ~0UL, ~0UL }; | |||
|  | 
 | |||
|  | /*
 | |||
|  |  * Mask is set (1) if enabled | |||
|  |  */ | |||
|  | static unsigned long titan_cached_irq_mask; | |||
|  | 
 | |||
|  | /*
 | |||
|  |  * Need SMP-safe access to interrupt CSRs | |||
|  |  */ | |||
|  | DEFINE_SPINLOCK(titan_irq_lock); | |||
|  | 
 | |||
|  | static void | |||
|  | titan_update_irq_hw(unsigned long mask) | |||
|  | { | |||
|  | 	register titan_cchip *cchip = TITAN_cchip; | |||
|  | 	unsigned long isa_enable = 1UL << 55; | |||
|  | 	register int bcpu = boot_cpuid; | |||
|  | 
 | |||
|  | #ifdef CONFIG_SMP
 | |||
|  | 	cpumask_t cpm = cpu_present_mask; | |||
|  | 	volatile unsigned long *dim0, *dim1, *dim2, *dim3; | |||
|  | 	unsigned long mask0, mask1, mask2, mask3, dummy; | |||
|  | 
 | |||
|  | 	mask &= ~isa_enable; | |||
|  | 	mask0 = mask & titan_cpu_irq_affinity[0]; | |||
|  | 	mask1 = mask & titan_cpu_irq_affinity[1]; | |||
|  | 	mask2 = mask & titan_cpu_irq_affinity[2]; | |||
|  | 	mask3 = mask & titan_cpu_irq_affinity[3]; | |||
|  | 
 | |||
|  | 	if (bcpu == 0) mask0 |= isa_enable; | |||
|  | 	else if (bcpu == 1) mask1 |= isa_enable; | |||
|  | 	else if (bcpu == 2) mask2 |= isa_enable; | |||
|  | 	else mask3 |= isa_enable; | |||
|  | 
 | |||
|  | 	dim0 = &cchip->dim0.csr; | |||
|  | 	dim1 = &cchip->dim1.csr; | |||
|  | 	dim2 = &cchip->dim2.csr; | |||
|  | 	dim3 = &cchip->dim3.csr; | |||
|  | 	if (!cpu_isset(0, cpm)) dim0 = &dummy; | |||
|  | 	if (!cpu_isset(1, cpm)) dim1 = &dummy; | |||
|  | 	if (!cpu_isset(2, cpm)) dim2 = &dummy; | |||
|  | 	if (!cpu_isset(3, cpm)) dim3 = &dummy; | |||
|  | 
 | |||
|  | 	*dim0 = mask0; | |||
|  | 	*dim1 = mask1; | |||
|  | 	*dim2 = mask2; | |||
|  | 	*dim3 = mask3; | |||
|  | 	mb(); | |||
|  | 	*dim0; | |||
|  | 	*dim1; | |||
|  | 	*dim2; | |||
|  | 	*dim3; | |||
|  | #else
 | |||
|  | 	volatile unsigned long *dimB; | |||
|  | 	dimB = &cchip->dim0.csr; | |||
|  | 	if (bcpu == 1) dimB = &cchip->dim1.csr; | |||
|  | 	else if (bcpu == 2) dimB = &cchip->dim2.csr; | |||
|  | 	else if (bcpu == 3) dimB = &cchip->dim3.csr; | |||
|  | 
 | |||
|  | 	*dimB = mask | isa_enable; | |||
|  | 	mb(); | |||
|  | 	*dimB; | |||
|  | #endif
 | |||
|  | } | |||
|  | 
 | |||
|  | static inline void | |||
|  | titan_enable_irq(unsigned int irq) | |||
|  | { | |||
|  | 	spin_lock(&titan_irq_lock); | |||
|  | 	titan_cached_irq_mask |= 1UL << (irq - 16); | |||
|  | 	titan_update_irq_hw(titan_cached_irq_mask); | |||
|  | 	spin_unlock(&titan_irq_lock); | |||
|  | } | |||
|  | 
 | |||
|  | static inline void | |||
|  | titan_disable_irq(unsigned int irq) | |||
|  | { | |||
|  | 	spin_lock(&titan_irq_lock); | |||
|  | 	titan_cached_irq_mask &= ~(1UL << (irq - 16)); | |||
|  | 	titan_update_irq_hw(titan_cached_irq_mask); | |||
|  | 	spin_unlock(&titan_irq_lock); | |||
|  | } | |||
|  | 
 | |||
|  | static unsigned int | |||
|  | titan_startup_irq(unsigned int irq) | |||
|  | { | |||
|  | 	titan_enable_irq(irq); | |||
|  | 	return 0;	/* never anything pending */ | |||
|  | } | |||
|  | 
 | |||
|  | static void | |||
|  | titan_end_irq(unsigned int irq) | |||
|  | { | |||
|  | 	if (!(irq_desc[irq].status & (IRQ_DISABLED|IRQ_INPROGRESS))) | |||
|  | 		titan_enable_irq(irq); | |||
|  | } | |||
|  | 
 | |||
|  | static void | |||
|  | titan_cpu_set_irq_affinity(unsigned int irq, cpumask_t affinity) | |||
|  | { | |||
|  | 	int cpu; | |||
|  | 
 | |||
|  | 	for (cpu = 0; cpu < 4; cpu++) { | |||
|  | 		if (cpu_isset(cpu, affinity)) | |||
|  | 			titan_cpu_irq_affinity[cpu] |= 1UL << irq; | |||
|  | 		else | |||
|  | 			titan_cpu_irq_affinity[cpu] &= ~(1UL << irq); | |||
|  | 	} | |||
|  | 
 | |||
|  | } | |||
|  | 
 | |||
|  | static void | |||
|  | titan_set_irq_affinity(unsigned int irq, cpumask_t affinity) | |||
|  | {  | |||
|  | 	spin_lock(&titan_irq_lock); | |||
|  | 	titan_cpu_set_irq_affinity(irq - 16, affinity); | |||
|  | 	titan_update_irq_hw(titan_cached_irq_mask); | |||
|  | 	spin_unlock(&titan_irq_lock); | |||
|  | } | |||
|  | 
 | |||
|  | static void | |||
|  | titan_device_interrupt(unsigned long vector, struct pt_regs * regs) | |||
|  | { | |||
|  | 	printk("titan_device_interrupt: NOT IMPLEMENTED YET!! \n"); | |||
|  | } | |||
|  | 
 | |||
|  | static void  | |||
|  | titan_srm_device_interrupt(unsigned long vector, struct pt_regs * regs) | |||
|  | { | |||
|  | 	int irq; | |||
|  | 
 | |||
|  | 	irq = (vector - 0x800) >> 4; | |||
|  | 	handle_irq(irq, regs); | |||
|  | } | |||
|  | 
 | |||
|  | 
 | |||
|  | static void __init | |||
|  | init_titan_irqs(struct hw_interrupt_type * ops, int imin, int imax) | |||
|  | { | |||
|  | 	long i; | |||
|  | 	for (i = imin; i <= imax; ++i) { | |||
|  | 		irq_desc[i].status = IRQ_DISABLED | IRQ_LEVEL; | |||
|  | 		irq_desc[i].handler = ops; | |||
|  | 	} | |||
|  | } | |||
|  | 
 | |||
|  | static struct hw_interrupt_type titan_irq_type = { | |||
|  |        .typename       = "TITAN", | |||
|  |        .startup        = titan_startup_irq, | |||
|  |        .shutdown       = titan_disable_irq, | |||
|  |        .enable         = titan_enable_irq, | |||
|  |        .disable        = titan_disable_irq, | |||
|  |        .ack            = titan_disable_irq, | |||
|  |        .end            = titan_end_irq, | |||
|  |        .set_affinity   = titan_set_irq_affinity, | |||
|  | }; | |||
|  | 
 | |||
|  | static irqreturn_t | |||
|  | titan_intr_nop(int irq, void *dev_id, struct pt_regs *regs)                     | |||
|  | { | |||
|  |       /*
 | |||
|  |        * This is a NOP interrupt handler for the purposes of | |||
|  |        * event counting -- just return. | |||
|  |        */                                                                      | |||
|  |        return IRQ_HANDLED; | |||
|  | } | |||
|  | 
 | |||
|  | static void __init | |||
|  | titan_init_irq(void) | |||
|  | { | |||
|  | 	if (alpha_using_srm && !alpha_mv.device_interrupt) | |||
|  | 		alpha_mv.device_interrupt = titan_srm_device_interrupt; | |||
|  | 	if (!alpha_mv.device_interrupt) | |||
|  | 		alpha_mv.device_interrupt = titan_device_interrupt; | |||
|  | 
 | |||
|  | 	titan_update_irq_hw(0); | |||
|  | 
 | |||
|  | 	init_titan_irqs(&titan_irq_type, 16, 63 + 16); | |||
|  | } | |||
|  |    | |||
|  | static void __init | |||
|  | titan_legacy_init_irq(void) | |||
|  | { | |||
|  | 	/* init the legacy dma controller */ | |||
|  | 	outb(0, DMA1_RESET_REG); | |||
|  | 	outb(0, DMA2_RESET_REG); | |||
|  | 	outb(DMA_MODE_CASCADE, DMA2_MODE_REG); | |||
|  | 	outb(0, DMA2_MASK_REG); | |||
|  | 
 | |||
|  | 	/* init the legacy irq controller */ | |||
|  | 	init_i8259a_irqs(); | |||
|  | 
 | |||
|  | 	/* init the titan irqs */ | |||
|  | 	titan_init_irq(); | |||
|  | } | |||
|  | 
 | |||
|  | void | |||
|  | titan_dispatch_irqs(u64 mask, struct pt_regs *regs) | |||
|  | { | |||
|  | 	unsigned long vector; | |||
|  | 
 | |||
|  | 	/*
 | |||
|  | 	 * Mask down to those interrupts which are enable on this processor | |||
|  | 	 */ | |||
|  | 	mask &= titan_cpu_irq_affinity[smp_processor_id()]; | |||
|  | 
 | |||
|  | 	/*
 | |||
|  | 	 * Dispatch all requested interrupts  | |||
|  | 	 */ | |||
|  | 	while (mask) { | |||
|  | 		/* convert to SRM vector... priority is <63> -> <0> */ | |||
|  | 		__asm__("ctlz %1, %0" : "=r"(vector) : "r"(mask)); | |||
|  | 		vector = 63 - vector; | |||
|  | 		mask &= ~(1UL << vector);	/* clear it out 	 */ | |||
|  | 		vector = 0x900 + (vector << 4);	/* convert to SRM vector */ | |||
|  | 		 | |||
|  | 		/* dispatch it */ | |||
|  | 		alpha_mv.device_interrupt(vector, regs); | |||
|  | 	} | |||
|  | } | |||
|  |    | |||
|  |  | |||
|  | /*
 | |||
|  |  * Titan Family | |||
|  |  */ | |||
|  | static void __init | |||
|  | titan_late_init(void) | |||
|  | { | |||
|  | 	/*
 | |||
|  | 	 * Enable the system error interrupts. These interrupts are  | |||
|  | 	 * all reported to the kernel as machine checks, so the handler | |||
|  | 	 * is a nop so it can be called to count the individual events. | |||
|  | 	 */ | |||
|  | 	request_irq(63+16, titan_intr_nop, SA_INTERRUPT,  | |||
|  | 		    "CChip Error", NULL); | |||
|  | 	request_irq(62+16, titan_intr_nop, SA_INTERRUPT,  | |||
|  | 		    "PChip 0 H_Error", NULL); | |||
|  | 	request_irq(61+16, titan_intr_nop, SA_INTERRUPT,  | |||
|  | 		    "PChip 1 H_Error", NULL); | |||
|  | 	request_irq(60+16, titan_intr_nop, SA_INTERRUPT,  | |||
|  | 		    "PChip 0 C_Error", NULL); | |||
|  | 	request_irq(59+16, titan_intr_nop, SA_INTERRUPT,  | |||
|  | 		    "PChip 1 C_Error", NULL); | |||
|  | 
 | |||
|  | 	/* 
 | |||
|  | 	 * Register our error handlers. | |||
|  | 	 */ | |||
|  | 	titan_register_error_handlers(); | |||
|  | 
 | |||
|  | 	/*
 | |||
|  | 	 * Check if the console left us any error logs. | |||
|  | 	 */ | |||
|  | 	cdl_check_console_data_log(); | |||
|  | 
 | |||
|  | } | |||
|  | 
 | |||
|  | static int __devinit | |||
|  | titan_map_irq(struct pci_dev *dev, u8 slot, u8 pin) | |||
|  | { | |||
|  | 	u8 intline; | |||
|  | 	int irq; | |||
|  | 
 | |||
|  |  	/* Get the current intline.  */ | |||
|  | 	pci_read_config_byte(dev, PCI_INTERRUPT_LINE, &intline); | |||
|  | 	irq = intline; | |||
|  | 
 | |||
|  |  	/* Is it explicitly routed through ISA?  */ | |||
|  |  	if ((irq & 0xF0) == 0xE0) | |||
|  |  		return irq; | |||
|  |   | |||
|  |  	/* Offset by 16 to make room for ISA interrupts 0 - 15.  */ | |||
|  |  	return irq + 16; | |||
|  | } | |||
|  | 
 | |||
|  | static void __init | |||
|  | titan_init_pci(void) | |||
|  | { | |||
|  |  	/*
 | |||
|  |  	 * This isn't really the right place, but there's some init | |||
|  |  	 * that needs to be done after everything is basically up. | |||
|  |  	 */ | |||
|  |  	titan_late_init(); | |||
|  |   | |||
|  | 	pci_probe_only = 1; | |||
|  | 	common_init_pci(); | |||
|  | 	SMC669_Init(0); | |||
|  | #ifdef CONFIG_VGA_HOSE
 | |||
|  | 	locate_and_init_vga(NULL); | |||
|  | #endif
 | |||
|  | } | |||
|  | 
 | |||
|  |  | |||
|  | /*
 | |||
|  |  * Privateer | |||
|  |  */ | |||
|  | static void __init | |||
|  | privateer_init_pci(void) | |||
|  | { | |||
|  | 	/*
 | |||
|  | 	 * Hook a couple of extra err interrupts that the | |||
|  | 	 * common titan code won't. | |||
|  | 	 */ | |||
|  | 	request_irq(53+16, titan_intr_nop, SA_INTERRUPT,  | |||
|  | 		    "NMI", NULL); | |||
|  | 	request_irq(50+16, titan_intr_nop, SA_INTERRUPT,  | |||
|  | 		    "Temperature Warning", NULL); | |||
|  | 
 | |||
|  | 	/*
 | |||
|  | 	 * Finish with the common version. | |||
|  | 	 */ | |||
|  | 	return titan_init_pci(); | |||
|  | } | |||
|  | 
 | |||
|  |  | |||
|  | /*
 | |||
|  |  * The System Vectors. | |||
|  |  */ | |||
|  | struct alpha_machine_vector titan_mv __initmv = { | |||
|  | 	.vector_name		= "TITAN", | |||
|  | 	DO_EV6_MMU, | |||
|  | 	DO_DEFAULT_RTC, | |||
|  | 	DO_TITAN_IO, | |||
|  | 	.machine_check		= titan_machine_check, | |||
|  | 	.max_isa_dma_address	= ALPHA_MAX_ISA_DMA_ADDRESS, | |||
|  | 	.min_io_address		= DEFAULT_IO_BASE, | |||
|  | 	.min_mem_address	= DEFAULT_MEM_BASE, | |||
|  | 	.pci_dac_offset		= TITAN_DAC_OFFSET, | |||
|  | 
 | |||
|  | 	.nr_irqs		= 80,	/* 64 + 16 */ | |||
|  | 	/* device_interrupt will be filled in by titan_init_irq */ | |||
|  | 
 | |||
|  | 	.agp_info		= titan_agp_info, | |||
|  | 
 | |||
|  | 	.init_arch		= titan_init_arch, | |||
|  | 	.init_irq		= titan_legacy_init_irq, | |||
|  | 	.init_rtc		= common_init_rtc, | |||
|  | 	.init_pci		= titan_init_pci, | |||
|  | 
 | |||
|  | 	.kill_arch		= titan_kill_arch, | |||
|  | 	.pci_map_irq		= titan_map_irq, | |||
|  | 	.pci_swizzle		= common_swizzle, | |||
|  | }; | |||
|  | ALIAS_MV(titan) | |||
|  | 
 | |||
|  | struct alpha_machine_vector privateer_mv __initmv = { | |||
|  | 	.vector_name		= "PRIVATEER", | |||
|  | 	DO_EV6_MMU, | |||
|  | 	DO_DEFAULT_RTC, | |||
|  | 	DO_TITAN_IO, | |||
|  | 	.machine_check		= privateer_machine_check, | |||
|  | 	.max_isa_dma_address	= ALPHA_MAX_ISA_DMA_ADDRESS, | |||
|  | 	.min_io_address		= DEFAULT_IO_BASE, | |||
|  | 	.min_mem_address	= DEFAULT_MEM_BASE, | |||
|  | 	.pci_dac_offset		= TITAN_DAC_OFFSET, | |||
|  | 
 | |||
|  | 	.nr_irqs		= 80,	/* 64 + 16 */ | |||
|  | 	/* device_interrupt will be filled in by titan_init_irq */ | |||
|  | 
 | |||
|  | 	.agp_info		= titan_agp_info, | |||
|  | 
 | |||
|  | 	.init_arch		= titan_init_arch, | |||
|  | 	.init_irq		= titan_legacy_init_irq, | |||
|  | 	.init_rtc		= common_init_rtc, | |||
|  | 	.init_pci		= privateer_init_pci, | |||
|  | 
 | |||
|  | 	.kill_arch		= titan_kill_arch, | |||
|  | 	.pci_map_irq		= titan_map_irq, | |||
|  | 	.pci_swizzle		= common_swizzle, | |||
|  | }; | |||
|  | /* No alpha_mv alias for privateer since we compile it 
 | |||
|  |    in unconditionally with titan; setup_arch knows how to cope. */ |