The SSR.MD status amongst other things are already made available, which can be used for encoding a more precise fault code value. Signed-off-by: Paul Mundt <lethal@linux-sh.org>
		
			
				
	
	
		
			166 lines
		
	
	
	
		
			4.7 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			166 lines
		
	
	
	
		
			4.7 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
 * The SH64 TLB miss.
 | 
						|
 *
 | 
						|
 * Original code from fault.c
 | 
						|
 * Copyright (C) 2000, 2001  Paolo Alberelli
 | 
						|
 *
 | 
						|
 * Fast PTE->TLB refill path
 | 
						|
 * Copyright (C) 2003 Richard.Curnow@superh.com
 | 
						|
 *
 | 
						|
 * IMPORTANT NOTES :
 | 
						|
 * The do_fast_page_fault function is called from a context in entry.S
 | 
						|
 * where very few registers have been saved.  In particular, the code in
 | 
						|
 * this file must be compiled not to use ANY caller-save registers that
 | 
						|
 * are not part of the restricted save set.  Also, it means that code in
 | 
						|
 * this file must not make calls to functions elsewhere in the kernel, or
 | 
						|
 * else the excepting context will see corruption in its caller-save
 | 
						|
 * registers.  Plus, the entry.S save area is non-reentrant, so this code
 | 
						|
 * has to run with SR.BL==1, i.e. no interrupts taken inside it and panic
 | 
						|
 * on any exception.
 | 
						|
 *
 | 
						|
 * This file is subject to the terms and conditions of the GNU General Public
 | 
						|
 * License.  See the file "COPYING" in the main directory of this archive
 | 
						|
 * for more details.
 | 
						|
 */
 | 
						|
#include <linux/signal.h>
 | 
						|
#include <linux/sched.h>
 | 
						|
#include <linux/kernel.h>
 | 
						|
#include <linux/errno.h>
 | 
						|
#include <linux/string.h>
 | 
						|
#include <linux/types.h>
 | 
						|
#include <linux/ptrace.h>
 | 
						|
#include <linux/mman.h>
 | 
						|
#include <linux/mm.h>
 | 
						|
#include <linux/smp.h>
 | 
						|
#include <linux/interrupt.h>
 | 
						|
#include <linux/kprobes.h>
 | 
						|
#include <asm/tlb.h>
 | 
						|
#include <asm/io.h>
 | 
						|
#include <asm/uaccess.h>
 | 
						|
#include <asm/pgalloc.h>
 | 
						|
#include <asm/mmu_context.h>
 | 
						|
 | 
						|
static int handle_tlbmiss(unsigned long long protection_flags,
 | 
						|
			  unsigned long address)
 | 
						|
{
 | 
						|
	pgd_t *pgd;
 | 
						|
	pud_t *pud;
 | 
						|
	pmd_t *pmd;
 | 
						|
	pte_t *pte;
 | 
						|
	pte_t entry;
 | 
						|
 | 
						|
	if (is_vmalloc_addr((void *)address)) {
 | 
						|
		pgd = pgd_offset_k(address);
 | 
						|
	} else {
 | 
						|
		if (unlikely(address >= TASK_SIZE || !current->mm))
 | 
						|
			return 1;
 | 
						|
 | 
						|
		pgd = pgd_offset(current->mm, address);
 | 
						|
	}
 | 
						|
 | 
						|
	pud = pud_offset(pgd, address);
 | 
						|
	if (pud_none(*pud) || !pud_present(*pud))
 | 
						|
		return 1;
 | 
						|
 | 
						|
	pmd = pmd_offset(pud, address);
 | 
						|
	if (pmd_none(*pmd) || !pmd_present(*pmd))
 | 
						|
		return 1;
 | 
						|
 | 
						|
	pte = pte_offset_kernel(pmd, address);
 | 
						|
	entry = *pte;
 | 
						|
	if (pte_none(entry) || !pte_present(entry))
 | 
						|
		return 1;
 | 
						|
 | 
						|
	/*
 | 
						|
	 * If the page doesn't have sufficient protection bits set to
 | 
						|
	 * service the kind of fault being handled, there's not much
 | 
						|
	 * point doing the TLB refill.  Punt the fault to the general
 | 
						|
	 * handler.
 | 
						|
	 */
 | 
						|
	if ((pte_val(entry) & protection_flags) != protection_flags)
 | 
						|
		return 1;
 | 
						|
 | 
						|
	update_mmu_cache(NULL, address, pte);
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * Put all this information into one structure so that everything is just
 | 
						|
 * arithmetic relative to a single base address.  This reduces the number
 | 
						|
 * of movi/shori pairs needed just to load addresses of static data.
 | 
						|
 */
 | 
						|
struct expevt_lookup {
 | 
						|
	unsigned short protection_flags[8];
 | 
						|
	unsigned char  is_text_access[8];
 | 
						|
	unsigned char  is_write_access[8];
 | 
						|
};
 | 
						|
 | 
						|
#define PRU (1<<9)
 | 
						|
#define PRW (1<<8)
 | 
						|
#define PRX (1<<7)
 | 
						|
#define PRR (1<<6)
 | 
						|
 | 
						|
/* Sized as 8 rather than 4 to allow checking the PTE's PRU bit against whether
 | 
						|
   the fault happened in user mode or privileged mode. */
 | 
						|
static struct expevt_lookup expevt_lookup_table = {
 | 
						|
	.protection_flags = {PRX, PRX, 0, 0, PRR, PRR, PRW, PRW},
 | 
						|
	.is_text_access   = {1,   1,   0, 0, 0,   0,   0,   0}
 | 
						|
};
 | 
						|
 | 
						|
static inline unsigned int
 | 
						|
expevt_to_fault_code(unsigned long expevt)
 | 
						|
{
 | 
						|
	if (expevt == 0xa40)
 | 
						|
		return FAULT_CODE_ITLB;
 | 
						|
	else if (expevt == 0x060)
 | 
						|
		return FAULT_CODE_WRITE;
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
   This routine handles page faults that can be serviced just by refilling a
 | 
						|
   TLB entry from an existing page table entry.  (This case represents a very
 | 
						|
   large majority of page faults.) Return 1 if the fault was successfully
 | 
						|
   handled.  Return 0 if the fault could not be handled.  (This leads into the
 | 
						|
   general fault handling in fault.c which deals with mapping file-backed
 | 
						|
   pages, stack growth, segmentation faults, swapping etc etc)
 | 
						|
 */
 | 
						|
asmlinkage int __kprobes
 | 
						|
do_fast_page_fault(unsigned long long ssr_md, unsigned long long expevt,
 | 
						|
		   unsigned long address)
 | 
						|
{
 | 
						|
	unsigned long long protection_flags;
 | 
						|
	unsigned long long index;
 | 
						|
	unsigned long long expevt4;
 | 
						|
	unsigned int fault_code;
 | 
						|
 | 
						|
	/* The next few lines implement a way of hashing EXPEVT into a
 | 
						|
	 * small array index which can be used to lookup parameters
 | 
						|
	 * specific to the type of TLBMISS being handled.
 | 
						|
	 *
 | 
						|
	 * Note:
 | 
						|
	 *	ITLBMISS has EXPEVT==0xa40
 | 
						|
	 *	RTLBMISS has EXPEVT==0x040
 | 
						|
	 *	WTLBMISS has EXPEVT==0x060
 | 
						|
	 */
 | 
						|
	expevt4 = (expevt >> 4);
 | 
						|
	/* TODO : xor ssr_md into this expression too. Then we can check
 | 
						|
	 * that PRU is set when it needs to be. */
 | 
						|
	index = expevt4 ^ (expevt4 >> 5);
 | 
						|
	index &= 7;
 | 
						|
 | 
						|
	fault_code = expevt_to_fault_code(expevt);
 | 
						|
 | 
						|
	protection_flags = expevt_lookup_table.protection_flags[index];
 | 
						|
 | 
						|
	if (expevt_lookup_table.is_text_access[index])
 | 
						|
		fault_code |= FAULT_CODE_ITLB;
 | 
						|
	if (!ssr_md)
 | 
						|
		fault_code |= FAULT_CODE_USER;
 | 
						|
 | 
						|
	set_thread_fault_code(fault_code);
 | 
						|
 | 
						|
	return handle_tlbmiss(protection_flags, address);
 | 
						|
}
 |