1071 lines
		
	
	
	
		
			24 KiB
			
		
	
	
	
		
			ArmAsm
		
	
	
	
	
	
			
		
		
	
	
			1071 lines
		
	
	
	
		
			24 KiB
			
		
	
	
	
		
			ArmAsm
		
	
	
	
	
	
/*
 | 
						|
 *  PARISC TLB and cache flushing support
 | 
						|
 *  Copyright (C) 2000-2001 Hewlett-Packard (John Marvin)
 | 
						|
 *  Copyright (C) 2001 Matthew Wilcox (willy at parisc-linux.org)
 | 
						|
 *  Copyright (C) 2002 Richard Hirst (rhirst with parisc-linux.org)
 | 
						|
 *
 | 
						|
 *    This program is free software; you can redistribute it and/or modify
 | 
						|
 *    it under the terms of the GNU General Public License as published by
 | 
						|
 *    the Free Software Foundation; either version 2, or (at your option)
 | 
						|
 *    any later version.
 | 
						|
 *
 | 
						|
 *    This program is distributed in the hope that it will be useful,
 | 
						|
 *    but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
						|
 *    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | 
						|
 *    GNU General Public License for more details.
 | 
						|
 *
 | 
						|
 *    You should have received a copy of the GNU General Public License
 | 
						|
 *    along with this program; if not, write to the Free Software
 | 
						|
 *    Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307  USA
 | 
						|
 */
 | 
						|
 | 
						|
/*
 | 
						|
 * NOTE: fdc,fic, and pdc instructions that use base register modification
 | 
						|
 *       should only use index and base registers that are not shadowed,
 | 
						|
 *       so that the fast path emulation in the non access miss handler
 | 
						|
 *       can be used.
 | 
						|
 */
 | 
						|
 | 
						|
#ifdef CONFIG_64BIT
 | 
						|
	.level	2.0w
 | 
						|
#else
 | 
						|
	.level	2.0
 | 
						|
#endif
 | 
						|
 | 
						|
#include <asm/psw.h>
 | 
						|
#include <asm/assembly.h>
 | 
						|
#include <asm/pgtable.h>
 | 
						|
#include <asm/cache.h>
 | 
						|
#include <linux/linkage.h>
 | 
						|
 | 
						|
	.text
 | 
						|
	.align	128
 | 
						|
 | 
						|
ENTRY(flush_tlb_all_local)
 | 
						|
	.proc
 | 
						|
	.callinfo NO_CALLS
 | 
						|
	.entry
 | 
						|
 | 
						|
	/*
 | 
						|
	 * The pitlbe and pdtlbe instructions should only be used to
 | 
						|
	 * flush the entire tlb. Also, there needs to be no intervening
 | 
						|
	 * tlb operations, e.g. tlb misses, so the operation needs
 | 
						|
	 * to happen in real mode with all interruptions disabled.
 | 
						|
	 */
 | 
						|
 | 
						|
	/* pcxt_ssm_bug	- relied upon translation! PA 2.0 Arch. F-4 and F-5 */
 | 
						|
	rsm		PSW_SM_I, %r19		/* save I-bit state */
 | 
						|
	load32		PA(1f), %r1
 | 
						|
	nop
 | 
						|
	nop
 | 
						|
	nop
 | 
						|
	nop
 | 
						|
	nop
 | 
						|
 | 
						|
	rsm		PSW_SM_Q, %r0		/* prep to load iia queue */
 | 
						|
	mtctl		%r0, %cr17		/* Clear IIASQ tail */
 | 
						|
	mtctl		%r0, %cr17		/* Clear IIASQ head */
 | 
						|
	mtctl		%r1, %cr18		/* IIAOQ head */
 | 
						|
	ldo		4(%r1), %r1
 | 
						|
	mtctl		%r1, %cr18		/* IIAOQ tail */
 | 
						|
	load32		REAL_MODE_PSW, %r1
 | 
						|
	mtctl           %r1, %ipsw
 | 
						|
	rfi
 | 
						|
	nop
 | 
						|
 | 
						|
1:      load32		PA(cache_info), %r1
 | 
						|
 | 
						|
	/* Flush Instruction Tlb */
 | 
						|
 | 
						|
	LDREG		ITLB_SID_BASE(%r1), %r20
 | 
						|
	LDREG		ITLB_SID_STRIDE(%r1), %r21
 | 
						|
	LDREG		ITLB_SID_COUNT(%r1), %r22
 | 
						|
	LDREG		ITLB_OFF_BASE(%r1), %arg0
 | 
						|
	LDREG		ITLB_OFF_STRIDE(%r1), %arg1
 | 
						|
	LDREG		ITLB_OFF_COUNT(%r1), %arg2
 | 
						|
	LDREG		ITLB_LOOP(%r1), %arg3
 | 
						|
 | 
						|
	addib,COND(=)		-1, %arg3, fitoneloop	/* Preadjust and test */
 | 
						|
	movb,<,n	%arg3, %r31, fitdone	/* If loop < 0, skip */
 | 
						|
	copy		%arg0, %r28		/* Init base addr */
 | 
						|
 | 
						|
fitmanyloop:					/* Loop if LOOP >= 2 */
 | 
						|
	mtsp		%r20, %sr1
 | 
						|
	add		%r21, %r20, %r20	/* increment space */
 | 
						|
	copy		%arg2, %r29		/* Init middle loop count */
 | 
						|
 | 
						|
fitmanymiddle:					/* Loop if LOOP >= 2 */
 | 
						|
	addib,COND(>)		-1, %r31, fitmanymiddle	/* Adjusted inner loop decr */
 | 
						|
	pitlbe		0(%sr1, %r28)
 | 
						|
	pitlbe,m	%arg1(%sr1, %r28)	/* Last pitlbe and addr adjust */
 | 
						|
	addib,COND(>)		-1, %r29, fitmanymiddle	/* Middle loop decr */
 | 
						|
	copy		%arg3, %r31		/* Re-init inner loop count */
 | 
						|
 | 
						|
	movb,tr		%arg0, %r28, fitmanyloop /* Re-init base addr */
 | 
						|
	addib,COND(<=),n	-1, %r22, fitdone	/* Outer loop count decr */
 | 
						|
 | 
						|
fitoneloop:					/* Loop if LOOP = 1 */
 | 
						|
	mtsp		%r20, %sr1
 | 
						|
	copy		%arg0, %r28		/* init base addr */
 | 
						|
	copy		%arg2, %r29		/* init middle loop count */
 | 
						|
 | 
						|
fitonemiddle:					/* Loop if LOOP = 1 */
 | 
						|
	addib,COND(>)		-1, %r29, fitonemiddle	/* Middle loop count decr */
 | 
						|
	pitlbe,m	%arg1(%sr1, %r28)	/* pitlbe for one loop */
 | 
						|
 | 
						|
	addib,COND(>)		-1, %r22, fitoneloop	/* Outer loop count decr */
 | 
						|
	add		%r21, %r20, %r20		/* increment space */
 | 
						|
 | 
						|
fitdone:
 | 
						|
 | 
						|
	/* Flush Data Tlb */
 | 
						|
 | 
						|
	LDREG		DTLB_SID_BASE(%r1), %r20
 | 
						|
	LDREG		DTLB_SID_STRIDE(%r1), %r21
 | 
						|
	LDREG		DTLB_SID_COUNT(%r1), %r22
 | 
						|
	LDREG		DTLB_OFF_BASE(%r1), %arg0
 | 
						|
	LDREG		DTLB_OFF_STRIDE(%r1), %arg1
 | 
						|
	LDREG		DTLB_OFF_COUNT(%r1), %arg2
 | 
						|
	LDREG		DTLB_LOOP(%r1), %arg3
 | 
						|
 | 
						|
	addib,COND(=)		-1, %arg3, fdtoneloop	/* Preadjust and test */
 | 
						|
	movb,<,n	%arg3, %r31, fdtdone	/* If loop < 0, skip */
 | 
						|
	copy		%arg0, %r28		/* Init base addr */
 | 
						|
 | 
						|
fdtmanyloop:					/* Loop if LOOP >= 2 */
 | 
						|
	mtsp		%r20, %sr1
 | 
						|
	add		%r21, %r20, %r20	/* increment space */
 | 
						|
	copy		%arg2, %r29		/* Init middle loop count */
 | 
						|
 | 
						|
fdtmanymiddle:					/* Loop if LOOP >= 2 */
 | 
						|
	addib,COND(>)		-1, %r31, fdtmanymiddle	/* Adjusted inner loop decr */
 | 
						|
	pdtlbe		0(%sr1, %r28)
 | 
						|
	pdtlbe,m	%arg1(%sr1, %r28)	/* Last pdtlbe and addr adjust */
 | 
						|
	addib,COND(>)		-1, %r29, fdtmanymiddle	/* Middle loop decr */
 | 
						|
	copy		%arg3, %r31		/* Re-init inner loop count */
 | 
						|
 | 
						|
	movb,tr		%arg0, %r28, fdtmanyloop /* Re-init base addr */
 | 
						|
	addib,COND(<=),n	-1, %r22,fdtdone	/* Outer loop count decr */
 | 
						|
 | 
						|
fdtoneloop:					/* Loop if LOOP = 1 */
 | 
						|
	mtsp		%r20, %sr1
 | 
						|
	copy		%arg0, %r28		/* init base addr */
 | 
						|
	copy		%arg2, %r29		/* init middle loop count */
 | 
						|
 | 
						|
fdtonemiddle:					/* Loop if LOOP = 1 */
 | 
						|
	addib,COND(>)		-1, %r29, fdtonemiddle	/* Middle loop count decr */
 | 
						|
	pdtlbe,m	%arg1(%sr1, %r28)	/* pdtlbe for one loop */
 | 
						|
 | 
						|
	addib,COND(>)		-1, %r22, fdtoneloop	/* Outer loop count decr */
 | 
						|
	add		%r21, %r20, %r20	/* increment space */
 | 
						|
 | 
						|
 | 
						|
fdtdone:
 | 
						|
	/*
 | 
						|
	 * Switch back to virtual mode
 | 
						|
	 */
 | 
						|
	/* pcxt_ssm_bug */
 | 
						|
	rsm		PSW_SM_I, %r0
 | 
						|
	load32		2f, %r1
 | 
						|
	nop
 | 
						|
	nop
 | 
						|
	nop
 | 
						|
	nop
 | 
						|
	nop
 | 
						|
 | 
						|
	rsm		PSW_SM_Q, %r0		/* prep to load iia queue */
 | 
						|
	mtctl		%r0, %cr17		/* Clear IIASQ tail */
 | 
						|
	mtctl		%r0, %cr17		/* Clear IIASQ head */
 | 
						|
	mtctl		%r1, %cr18		/* IIAOQ head */
 | 
						|
	ldo		4(%r1), %r1
 | 
						|
	mtctl		%r1, %cr18		/* IIAOQ tail */
 | 
						|
	load32		KERNEL_PSW, %r1
 | 
						|
	or		%r1, %r19, %r1	/* I-bit to state on entry */
 | 
						|
	mtctl		%r1, %ipsw	/* restore I-bit (entire PSW) */
 | 
						|
	rfi
 | 
						|
	nop
 | 
						|
 | 
						|
2:      bv		%r0(%r2)
 | 
						|
	nop
 | 
						|
 | 
						|
	.exit
 | 
						|
	.procend
 | 
						|
ENDPROC(flush_tlb_all_local)
 | 
						|
 | 
						|
	.import cache_info,data
 | 
						|
 | 
						|
ENTRY(flush_instruction_cache_local)
 | 
						|
	.proc
 | 
						|
	.callinfo NO_CALLS
 | 
						|
	.entry
 | 
						|
 | 
						|
	mtsp		%r0, %sr1
 | 
						|
	load32		cache_info, %r1
 | 
						|
 | 
						|
	/* Flush Instruction Cache */
 | 
						|
 | 
						|
	LDREG		ICACHE_BASE(%r1), %arg0
 | 
						|
	LDREG		ICACHE_STRIDE(%r1), %arg1
 | 
						|
	LDREG		ICACHE_COUNT(%r1), %arg2
 | 
						|
	LDREG		ICACHE_LOOP(%r1), %arg3
 | 
						|
	rsm             PSW_SM_I, %r22		/* No mmgt ops during loop*/
 | 
						|
	addib,COND(=)		-1, %arg3, fioneloop	/* Preadjust and test */
 | 
						|
	movb,<,n	%arg3, %r31, fisync	/* If loop < 0, do sync */
 | 
						|
 | 
						|
fimanyloop:					/* Loop if LOOP >= 2 */
 | 
						|
	addib,COND(>)		-1, %r31, fimanyloop	/* Adjusted inner loop decr */
 | 
						|
	fice            %r0(%sr1, %arg0)
 | 
						|
	fice,m		%arg1(%sr1, %arg0)	/* Last fice and addr adjust */
 | 
						|
	movb,tr		%arg3, %r31, fimanyloop	/* Re-init inner loop count */
 | 
						|
	addib,COND(<=),n	-1, %arg2, fisync	/* Outer loop decr */
 | 
						|
 | 
						|
fioneloop:					/* Loop if LOOP = 1 */
 | 
						|
	addib,COND(>)		-1, %arg2, fioneloop	/* Outer loop count decr */
 | 
						|
	fice,m		%arg1(%sr1, %arg0)	/* Fice for one loop */
 | 
						|
 | 
						|
fisync:
 | 
						|
	sync
 | 
						|
	mtsm		%r22			/* restore I-bit */
 | 
						|
	bv		%r0(%r2)
 | 
						|
	nop
 | 
						|
	.exit
 | 
						|
 | 
						|
	.procend
 | 
						|
ENDPROC(flush_instruction_cache_local)
 | 
						|
 | 
						|
 | 
						|
	.import cache_info, data
 | 
						|
ENTRY(flush_data_cache_local)
 | 
						|
	.proc
 | 
						|
	.callinfo NO_CALLS
 | 
						|
	.entry
 | 
						|
 | 
						|
	mtsp		%r0, %sr1
 | 
						|
	load32 		cache_info, %r1
 | 
						|
 | 
						|
	/* Flush Data Cache */
 | 
						|
 | 
						|
	LDREG		DCACHE_BASE(%r1), %arg0
 | 
						|
	LDREG		DCACHE_STRIDE(%r1), %arg1
 | 
						|
	LDREG		DCACHE_COUNT(%r1), %arg2
 | 
						|
	LDREG		DCACHE_LOOP(%r1), %arg3
 | 
						|
	rsm		PSW_SM_I, %r22
 | 
						|
	addib,COND(=)		-1, %arg3, fdoneloop	/* Preadjust and test */
 | 
						|
	movb,<,n	%arg3, %r31, fdsync	/* If loop < 0, do sync */
 | 
						|
 | 
						|
fdmanyloop:					/* Loop if LOOP >= 2 */
 | 
						|
	addib,COND(>)		-1, %r31, fdmanyloop	/* Adjusted inner loop decr */
 | 
						|
	fdce		%r0(%sr1, %arg0)
 | 
						|
	fdce,m		%arg1(%sr1, %arg0)	/* Last fdce and addr adjust */
 | 
						|
	movb,tr		%arg3, %r31, fdmanyloop	/* Re-init inner loop count */
 | 
						|
	addib,COND(<=),n	-1, %arg2, fdsync	/* Outer loop decr */
 | 
						|
 | 
						|
fdoneloop:					/* Loop if LOOP = 1 */
 | 
						|
	addib,COND(>)		-1, %arg2, fdoneloop	/* Outer loop count decr */
 | 
						|
	fdce,m		%arg1(%sr1, %arg0)	/* Fdce for one loop */
 | 
						|
 | 
						|
fdsync:
 | 
						|
	syncdma
 | 
						|
	sync
 | 
						|
	mtsm		%r22			/* restore I-bit */
 | 
						|
	bv		%r0(%r2)
 | 
						|
	nop
 | 
						|
	.exit
 | 
						|
 | 
						|
	.procend
 | 
						|
ENDPROC(flush_data_cache_local)
 | 
						|
 | 
						|
	.align	16
 | 
						|
 | 
						|
ENTRY(copy_user_page_asm)
 | 
						|
	.proc
 | 
						|
	.callinfo NO_CALLS
 | 
						|
	.entry
 | 
						|
 | 
						|
#ifdef CONFIG_64BIT
 | 
						|
	/* PA8x00 CPUs can consume 2 loads or 1 store per cycle.
 | 
						|
	 * Unroll the loop by hand and arrange insn appropriately.
 | 
						|
	 * GCC probably can do this just as well.
 | 
						|
	 */
 | 
						|
 | 
						|
	ldd		0(%r25), %r19
 | 
						|
	ldi		(PAGE_SIZE / 128), %r1
 | 
						|
 | 
						|
	ldw		64(%r25), %r0		/* prefetch 1 cacheline ahead */
 | 
						|
	ldw		128(%r25), %r0		/* prefetch 2 */
 | 
						|
 | 
						|
1:	ldd		8(%r25), %r20
 | 
						|
	ldw		192(%r25), %r0		/* prefetch 3 */
 | 
						|
	ldw		256(%r25), %r0		/* prefetch 4 */
 | 
						|
 | 
						|
	ldd		16(%r25), %r21
 | 
						|
	ldd		24(%r25), %r22
 | 
						|
	std		%r19, 0(%r26)
 | 
						|
	std		%r20, 8(%r26)
 | 
						|
 | 
						|
	ldd		32(%r25), %r19
 | 
						|
	ldd		40(%r25), %r20
 | 
						|
	std		%r21, 16(%r26)
 | 
						|
	std		%r22, 24(%r26)
 | 
						|
 | 
						|
	ldd		48(%r25), %r21
 | 
						|
	ldd		56(%r25), %r22
 | 
						|
	std		%r19, 32(%r26)
 | 
						|
	std		%r20, 40(%r26)
 | 
						|
 | 
						|
	ldd		64(%r25), %r19
 | 
						|
	ldd		72(%r25), %r20
 | 
						|
	std		%r21, 48(%r26)
 | 
						|
	std		%r22, 56(%r26)
 | 
						|
 | 
						|
	ldd		80(%r25), %r21
 | 
						|
	ldd		88(%r25), %r22
 | 
						|
	std		%r19, 64(%r26)
 | 
						|
	std		%r20, 72(%r26)
 | 
						|
 | 
						|
	ldd		 96(%r25), %r19
 | 
						|
	ldd		104(%r25), %r20
 | 
						|
	std		%r21, 80(%r26)
 | 
						|
	std		%r22, 88(%r26)
 | 
						|
 | 
						|
	ldd		112(%r25), %r21
 | 
						|
	ldd		120(%r25), %r22
 | 
						|
	std		%r19, 96(%r26)
 | 
						|
	std		%r20, 104(%r26)
 | 
						|
 | 
						|
	ldo		128(%r25), %r25
 | 
						|
	std		%r21, 112(%r26)
 | 
						|
	std		%r22, 120(%r26)
 | 
						|
	ldo		128(%r26), %r26
 | 
						|
 | 
						|
	/* conditional branches nullify on forward taken branch, and on
 | 
						|
	 * non-taken backward branch. Note that .+4 is a backwards branch.
 | 
						|
	 * The ldd should only get executed if the branch is taken.
 | 
						|
	 */
 | 
						|
	addib,COND(>),n	-1, %r1, 1b		/* bundle 10 */
 | 
						|
	ldd		0(%r25), %r19		/* start next loads */
 | 
						|
 | 
						|
#else
 | 
						|
 | 
						|
	/*
 | 
						|
	 * This loop is optimized for PCXL/PCXL2 ldw/ldw and stw/stw
 | 
						|
	 * bundles (very restricted rules for bundling).
 | 
						|
	 * Note that until (if) we start saving
 | 
						|
	 * the full 64 bit register values on interrupt, we can't
 | 
						|
	 * use ldd/std on a 32 bit kernel.
 | 
						|
	 */
 | 
						|
	ldw		0(%r25), %r19
 | 
						|
	ldi		(PAGE_SIZE / 64), %r1
 | 
						|
 | 
						|
1:
 | 
						|
	ldw		4(%r25), %r20
 | 
						|
	ldw		8(%r25), %r21
 | 
						|
	ldw		12(%r25), %r22
 | 
						|
	stw		%r19, 0(%r26)
 | 
						|
	stw		%r20, 4(%r26)
 | 
						|
	stw		%r21, 8(%r26)
 | 
						|
	stw		%r22, 12(%r26)
 | 
						|
	ldw		16(%r25), %r19
 | 
						|
	ldw		20(%r25), %r20
 | 
						|
	ldw		24(%r25), %r21
 | 
						|
	ldw		28(%r25), %r22
 | 
						|
	stw		%r19, 16(%r26)
 | 
						|
	stw		%r20, 20(%r26)
 | 
						|
	stw		%r21, 24(%r26)
 | 
						|
	stw		%r22, 28(%r26)
 | 
						|
	ldw		32(%r25), %r19
 | 
						|
	ldw		36(%r25), %r20
 | 
						|
	ldw		40(%r25), %r21
 | 
						|
	ldw		44(%r25), %r22
 | 
						|
	stw		%r19, 32(%r26)
 | 
						|
	stw		%r20, 36(%r26)
 | 
						|
	stw		%r21, 40(%r26)
 | 
						|
	stw		%r22, 44(%r26)
 | 
						|
	ldw		48(%r25), %r19
 | 
						|
	ldw		52(%r25), %r20
 | 
						|
	ldw		56(%r25), %r21
 | 
						|
	ldw		60(%r25), %r22
 | 
						|
	stw		%r19, 48(%r26)
 | 
						|
	stw		%r20, 52(%r26)
 | 
						|
	ldo		64(%r25), %r25
 | 
						|
	stw		%r21, 56(%r26)
 | 
						|
	stw		%r22, 60(%r26)
 | 
						|
	ldo		64(%r26), %r26
 | 
						|
	addib,COND(>),n	-1, %r1, 1b
 | 
						|
	ldw		0(%r25), %r19
 | 
						|
#endif
 | 
						|
	bv		%r0(%r2)
 | 
						|
	nop
 | 
						|
	.exit
 | 
						|
 | 
						|
	.procend
 | 
						|
ENDPROC(copy_user_page_asm)
 | 
						|
 | 
						|
/*
 | 
						|
 * NOTE: Code in clear_user_page has a hard coded dependency on the
 | 
						|
 *       maximum alias boundary being 4 Mb. We've been assured by the
 | 
						|
 *       parisc chip designers that there will not ever be a parisc
 | 
						|
 *       chip with a larger alias boundary (Never say never :-) ).
 | 
						|
 *
 | 
						|
 *       Subtle: the dtlb miss handlers support the temp alias region by
 | 
						|
 *       "knowing" that if a dtlb miss happens within the temp alias
 | 
						|
 *       region it must have occurred while in clear_user_page. Since
 | 
						|
 *       this routine makes use of processor local translations, we
 | 
						|
 *       don't want to insert them into the kernel page table. Instead,
 | 
						|
 *       we load up some general registers (they need to be registers
 | 
						|
 *       which aren't shadowed) with the physical page numbers (preshifted
 | 
						|
 *       for tlb insertion) needed to insert the translations. When we
 | 
						|
 *       miss on the translation, the dtlb miss handler inserts the
 | 
						|
 *       translation into the tlb using these values:
 | 
						|
 *
 | 
						|
 *          %r26 physical page (shifted for tlb insert) of "to" translation
 | 
						|
 *          %r23 physical page (shifted for tlb insert) of "from" translation
 | 
						|
 */
 | 
						|
 | 
						|
#if 0
 | 
						|
 | 
						|
	/*
 | 
						|
	 * We can't do this since copy_user_page is used to bring in
 | 
						|
	 * file data that might have instructions. Since the data would
 | 
						|
	 * then need to be flushed out so the i-fetch can see it, it
 | 
						|
	 * makes more sense to just copy through the kernel translation
 | 
						|
	 * and flush it.
 | 
						|
	 *
 | 
						|
	 * I'm still keeping this around because it may be possible to
 | 
						|
	 * use it if more information is passed into copy_user_page().
 | 
						|
	 * Have to do some measurements to see if it is worthwhile to
 | 
						|
	 * lobby for such a change.
 | 
						|
	 */
 | 
						|
 | 
						|
ENTRY(copy_user_page_asm)
 | 
						|
	.proc
 | 
						|
	.callinfo NO_CALLS
 | 
						|
	.entry
 | 
						|
 | 
						|
	ldil		L%(__PAGE_OFFSET), %r1
 | 
						|
	sub		%r26, %r1, %r26
 | 
						|
	sub		%r25, %r1, %r23		/* move physical addr into non shadowed reg */
 | 
						|
 | 
						|
	ldil		L%(TMPALIAS_MAP_START), %r28
 | 
						|
	/* FIXME for different page sizes != 4k */
 | 
						|
#ifdef CONFIG_64BIT
 | 
						|
	extrd,u		%r26,56,32, %r26		/* convert phys addr to tlb insert format */
 | 
						|
	extrd,u		%r23,56,32, %r23		/* convert phys addr to tlb insert format */
 | 
						|
	depd		%r24,63,22, %r28		/* Form aliased virtual address 'to' */
 | 
						|
	depdi		0, 63,12, %r28		/* Clear any offset bits */
 | 
						|
	copy		%r28, %r29
 | 
						|
	depdi		1, 41,1, %r29		/* Form aliased virtual address 'from' */
 | 
						|
#else
 | 
						|
	extrw,u		%r26, 24,25, %r26	/* convert phys addr to tlb insert format */
 | 
						|
	extrw,u		%r23, 24,25, %r23	/* convert phys addr to tlb insert format */
 | 
						|
	depw		%r24, 31,22, %r28	/* Form aliased virtual address 'to' */
 | 
						|
	depwi		0, 31,12, %r28		/* Clear any offset bits */
 | 
						|
	copy		%r28, %r29
 | 
						|
	depwi		1, 9,1, %r29		/* Form aliased virtual address 'from' */
 | 
						|
#endif
 | 
						|
 | 
						|
	/* Purge any old translations */
 | 
						|
 | 
						|
	pdtlb		0(%r28)
 | 
						|
	pdtlb		0(%r29)
 | 
						|
 | 
						|
	ldi		64, %r1
 | 
						|
 | 
						|
	/*
 | 
						|
	 * This loop is optimized for PCXL/PCXL2 ldw/ldw and stw/stw
 | 
						|
	 * bundles (very restricted rules for bundling). It probably
 | 
						|
	 * does OK on PCXU and better, but we could do better with
 | 
						|
	 * ldd/std instructions. Note that until (if) we start saving
 | 
						|
	 * the full 64 bit register values on interrupt, we can't
 | 
						|
	 * use ldd/std on a 32 bit kernel.
 | 
						|
	 */
 | 
						|
 | 
						|
 | 
						|
1:
 | 
						|
	ldw		0(%r29), %r19
 | 
						|
	ldw		4(%r29), %r20
 | 
						|
	ldw		8(%r29), %r21
 | 
						|
	ldw		12(%r29), %r22
 | 
						|
	stw		%r19, 0(%r28)
 | 
						|
	stw		%r20, 4(%r28)
 | 
						|
	stw		%r21, 8(%r28)
 | 
						|
	stw		%r22, 12(%r28)
 | 
						|
	ldw		16(%r29), %r19
 | 
						|
	ldw		20(%r29), %r20
 | 
						|
	ldw		24(%r29), %r21
 | 
						|
	ldw		28(%r29), %r22
 | 
						|
	stw		%r19, 16(%r28)
 | 
						|
	stw		%r20, 20(%r28)
 | 
						|
	stw		%r21, 24(%r28)
 | 
						|
	stw		%r22, 28(%r28)
 | 
						|
	ldw		32(%r29), %r19
 | 
						|
	ldw		36(%r29), %r20
 | 
						|
	ldw		40(%r29), %r21
 | 
						|
	ldw		44(%r29), %r22
 | 
						|
	stw		%r19, 32(%r28)
 | 
						|
	stw		%r20, 36(%r28)
 | 
						|
	stw		%r21, 40(%r28)
 | 
						|
	stw		%r22, 44(%r28)
 | 
						|
	ldw		48(%r29), %r19
 | 
						|
	ldw		52(%r29), %r20
 | 
						|
	ldw		56(%r29), %r21
 | 
						|
	ldw		60(%r29), %r22
 | 
						|
	stw		%r19, 48(%r28)
 | 
						|
	stw		%r20, 52(%r28)
 | 
						|
	stw		%r21, 56(%r28)
 | 
						|
	stw		%r22, 60(%r28)
 | 
						|
	ldo		64(%r28), %r28
 | 
						|
	addib,COND(>)		-1, %r1,1b
 | 
						|
	ldo		64(%r29), %r29
 | 
						|
 | 
						|
	bv		%r0(%r2)
 | 
						|
	nop
 | 
						|
	.exit
 | 
						|
 | 
						|
	.procend
 | 
						|
ENDPROC(copy_user_page_asm)
 | 
						|
#endif
 | 
						|
 | 
						|
ENTRY(__clear_user_page_asm)
 | 
						|
	.proc
 | 
						|
	.callinfo NO_CALLS
 | 
						|
	.entry
 | 
						|
 | 
						|
	tophys_r1	%r26
 | 
						|
 | 
						|
	ldil		L%(TMPALIAS_MAP_START), %r28
 | 
						|
#ifdef CONFIG_64BIT
 | 
						|
#if (TMPALIAS_MAP_START >= 0x80000000)
 | 
						|
	depdi		0, 31,32, %r28		/* clear any sign extension */
 | 
						|
	/* FIXME: page size dependend */
 | 
						|
#endif
 | 
						|
	extrd,u		%r26, 56,32, %r26	/* convert phys addr to tlb insert format */
 | 
						|
	depd		%r25, 63,22, %r28	/* Form aliased virtual address 'to' */
 | 
						|
	depdi		0, 63,12, %r28		/* Clear any offset bits */
 | 
						|
#else
 | 
						|
	extrw,u		%r26, 24,25, %r26	/* convert phys addr to tlb insert format */
 | 
						|
	depw		%r25, 31,22, %r28	/* Form aliased virtual address 'to' */
 | 
						|
	depwi		0, 31,12, %r28		/* Clear any offset bits */
 | 
						|
#endif
 | 
						|
 | 
						|
	/* Purge any old translation */
 | 
						|
 | 
						|
	pdtlb		0(%r28)
 | 
						|
 | 
						|
#ifdef CONFIG_64BIT
 | 
						|
	ldi		(PAGE_SIZE / 128), %r1
 | 
						|
 | 
						|
	/* PREFETCH (Write) has not (yet) been proven to help here */
 | 
						|
	/* #define	PREFETCHW_OP	ldd		256(%0), %r0 */
 | 
						|
 | 
						|
1:	std		%r0, 0(%r28)
 | 
						|
	std		%r0, 8(%r28)
 | 
						|
	std		%r0, 16(%r28)
 | 
						|
	std		%r0, 24(%r28)
 | 
						|
	std		%r0, 32(%r28)
 | 
						|
	std		%r0, 40(%r28)
 | 
						|
	std		%r0, 48(%r28)
 | 
						|
	std		%r0, 56(%r28)
 | 
						|
	std		%r0, 64(%r28)
 | 
						|
	std		%r0, 72(%r28)
 | 
						|
	std		%r0, 80(%r28)
 | 
						|
	std		%r0, 88(%r28)
 | 
						|
	std		%r0, 96(%r28)
 | 
						|
	std		%r0, 104(%r28)
 | 
						|
	std		%r0, 112(%r28)
 | 
						|
	std		%r0, 120(%r28)
 | 
						|
	addib,COND(>)		-1, %r1, 1b
 | 
						|
	ldo		128(%r28), %r28
 | 
						|
 | 
						|
#else	/* ! CONFIG_64BIT */
 | 
						|
	ldi		(PAGE_SIZE / 64), %r1
 | 
						|
 | 
						|
1:
 | 
						|
	stw		%r0, 0(%r28)
 | 
						|
	stw		%r0, 4(%r28)
 | 
						|
	stw		%r0, 8(%r28)
 | 
						|
	stw		%r0, 12(%r28)
 | 
						|
	stw		%r0, 16(%r28)
 | 
						|
	stw		%r0, 20(%r28)
 | 
						|
	stw		%r0, 24(%r28)
 | 
						|
	stw		%r0, 28(%r28)
 | 
						|
	stw		%r0, 32(%r28)
 | 
						|
	stw		%r0, 36(%r28)
 | 
						|
	stw		%r0, 40(%r28)
 | 
						|
	stw		%r0, 44(%r28)
 | 
						|
	stw		%r0, 48(%r28)
 | 
						|
	stw		%r0, 52(%r28)
 | 
						|
	stw		%r0, 56(%r28)
 | 
						|
	stw		%r0, 60(%r28)
 | 
						|
	addib,COND(>)		-1, %r1, 1b
 | 
						|
	ldo		64(%r28), %r28
 | 
						|
#endif	/* CONFIG_64BIT */
 | 
						|
 | 
						|
	bv		%r0(%r2)
 | 
						|
	nop
 | 
						|
	.exit
 | 
						|
 | 
						|
	.procend
 | 
						|
ENDPROC(__clear_user_page_asm)
 | 
						|
 | 
						|
ENTRY(flush_kernel_dcache_page_asm)
 | 
						|
	.proc
 | 
						|
	.callinfo NO_CALLS
 | 
						|
	.entry
 | 
						|
 | 
						|
	ldil		L%dcache_stride, %r1
 | 
						|
	ldw		R%dcache_stride(%r1), %r23
 | 
						|
 | 
						|
#ifdef CONFIG_64BIT
 | 
						|
	depdi,z		1, 63-PAGE_SHIFT,1, %r25
 | 
						|
#else
 | 
						|
	depwi,z		1, 31-PAGE_SHIFT,1, %r25
 | 
						|
#endif
 | 
						|
	add		%r26, %r25, %r25
 | 
						|
	sub		%r25, %r23, %r25
 | 
						|
 | 
						|
 | 
						|
1:      fdc,m		%r23(%r26)
 | 
						|
	fdc,m		%r23(%r26)
 | 
						|
	fdc,m		%r23(%r26)
 | 
						|
	fdc,m		%r23(%r26)
 | 
						|
	fdc,m		%r23(%r26)
 | 
						|
	fdc,m		%r23(%r26)
 | 
						|
	fdc,m		%r23(%r26)
 | 
						|
	fdc,m		%r23(%r26)
 | 
						|
	fdc,m		%r23(%r26)
 | 
						|
	fdc,m		%r23(%r26)
 | 
						|
	fdc,m		%r23(%r26)
 | 
						|
	fdc,m		%r23(%r26)
 | 
						|
	fdc,m		%r23(%r26)
 | 
						|
	fdc,m		%r23(%r26)
 | 
						|
	fdc,m		%r23(%r26)
 | 
						|
	cmpb,COND(<<)		%r26, %r25,1b
 | 
						|
	fdc,m		%r23(%r26)
 | 
						|
 | 
						|
	sync
 | 
						|
	bv		%r0(%r2)
 | 
						|
	nop
 | 
						|
	.exit
 | 
						|
 | 
						|
	.procend
 | 
						|
ENDPROC(flush_kernel_dcache_page_asm)
 | 
						|
	
 | 
						|
ENTRY(flush_user_dcache_page)
 | 
						|
	.proc
 | 
						|
	.callinfo NO_CALLS
 | 
						|
	.entry
 | 
						|
 | 
						|
	ldil		L%dcache_stride, %r1
 | 
						|
	ldw		R%dcache_stride(%r1), %r23
 | 
						|
 | 
						|
#ifdef CONFIG_64BIT
 | 
						|
	depdi,z		1,63-PAGE_SHIFT,1, %r25
 | 
						|
#else
 | 
						|
	depwi,z		1,31-PAGE_SHIFT,1, %r25
 | 
						|
#endif
 | 
						|
	add		%r26, %r25, %r25
 | 
						|
	sub		%r25, %r23, %r25
 | 
						|
 | 
						|
 | 
						|
1:      fdc,m		%r23(%sr3, %r26)
 | 
						|
	fdc,m		%r23(%sr3, %r26)
 | 
						|
	fdc,m		%r23(%sr3, %r26)
 | 
						|
	fdc,m		%r23(%sr3, %r26)
 | 
						|
	fdc,m		%r23(%sr3, %r26)
 | 
						|
	fdc,m		%r23(%sr3, %r26)
 | 
						|
	fdc,m		%r23(%sr3, %r26)
 | 
						|
	fdc,m		%r23(%sr3, %r26)
 | 
						|
	fdc,m		%r23(%sr3, %r26)
 | 
						|
	fdc,m		%r23(%sr3, %r26)
 | 
						|
	fdc,m		%r23(%sr3, %r26)
 | 
						|
	fdc,m		%r23(%sr3, %r26)
 | 
						|
	fdc,m		%r23(%sr3, %r26)
 | 
						|
	fdc,m		%r23(%sr3, %r26)
 | 
						|
	fdc,m		%r23(%sr3, %r26)
 | 
						|
	cmpb,COND(<<)		%r26, %r25,1b
 | 
						|
	fdc,m		%r23(%sr3, %r26)
 | 
						|
 | 
						|
	sync
 | 
						|
	bv		%r0(%r2)
 | 
						|
	nop
 | 
						|
	.exit
 | 
						|
 | 
						|
	.procend
 | 
						|
ENDPROC(flush_user_dcache_page)
 | 
						|
 | 
						|
ENTRY(flush_user_icache_page)
 | 
						|
	.proc
 | 
						|
	.callinfo NO_CALLS
 | 
						|
	.entry
 | 
						|
 | 
						|
	ldil		L%dcache_stride, %r1
 | 
						|
	ldw		R%dcache_stride(%r1), %r23
 | 
						|
 | 
						|
#ifdef CONFIG_64BIT
 | 
						|
	depdi,z		1, 63-PAGE_SHIFT,1, %r25
 | 
						|
#else
 | 
						|
	depwi,z		1, 31-PAGE_SHIFT,1, %r25
 | 
						|
#endif
 | 
						|
	add		%r26, %r25, %r25
 | 
						|
	sub		%r25, %r23, %r25
 | 
						|
 | 
						|
 | 
						|
1:      fic,m		%r23(%sr3, %r26)
 | 
						|
	fic,m		%r23(%sr3, %r26)
 | 
						|
	fic,m		%r23(%sr3, %r26)
 | 
						|
	fic,m		%r23(%sr3, %r26)
 | 
						|
	fic,m		%r23(%sr3, %r26)
 | 
						|
	fic,m		%r23(%sr3, %r26)
 | 
						|
	fic,m		%r23(%sr3, %r26)
 | 
						|
	fic,m		%r23(%sr3, %r26)
 | 
						|
	fic,m		%r23(%sr3, %r26)
 | 
						|
	fic,m		%r23(%sr3, %r26)
 | 
						|
	fic,m		%r23(%sr3, %r26)
 | 
						|
	fic,m		%r23(%sr3, %r26)
 | 
						|
	fic,m		%r23(%sr3, %r26)
 | 
						|
	fic,m		%r23(%sr3, %r26)
 | 
						|
	fic,m		%r23(%sr3, %r26)
 | 
						|
	cmpb,COND(<<)		%r26, %r25,1b
 | 
						|
	fic,m		%r23(%sr3, %r26)
 | 
						|
 | 
						|
	sync
 | 
						|
	bv		%r0(%r2)
 | 
						|
	nop
 | 
						|
	.exit
 | 
						|
 | 
						|
	.procend
 | 
						|
ENDPROC(flush_user_icache_page)
 | 
						|
 | 
						|
 | 
						|
ENTRY(purge_kernel_dcache_page)
 | 
						|
	.proc
 | 
						|
	.callinfo NO_CALLS
 | 
						|
	.entry
 | 
						|
 | 
						|
	ldil		L%dcache_stride, %r1
 | 
						|
	ldw		R%dcache_stride(%r1), %r23
 | 
						|
 | 
						|
#ifdef CONFIG_64BIT
 | 
						|
	depdi,z		1, 63-PAGE_SHIFT,1, %r25
 | 
						|
#else
 | 
						|
	depwi,z		1, 31-PAGE_SHIFT,1, %r25
 | 
						|
#endif
 | 
						|
	add		%r26, %r25, %r25
 | 
						|
	sub		%r25, %r23, %r25
 | 
						|
 | 
						|
1:      pdc,m		%r23(%r26)
 | 
						|
	pdc,m		%r23(%r26)
 | 
						|
	pdc,m		%r23(%r26)
 | 
						|
	pdc,m		%r23(%r26)
 | 
						|
	pdc,m		%r23(%r26)
 | 
						|
	pdc,m		%r23(%r26)
 | 
						|
	pdc,m		%r23(%r26)
 | 
						|
	pdc,m		%r23(%r26)
 | 
						|
	pdc,m		%r23(%r26)
 | 
						|
	pdc,m		%r23(%r26)
 | 
						|
	pdc,m		%r23(%r26)
 | 
						|
	pdc,m		%r23(%r26)
 | 
						|
	pdc,m		%r23(%r26)
 | 
						|
	pdc,m		%r23(%r26)
 | 
						|
	pdc,m		%r23(%r26)
 | 
						|
	cmpb,COND(<<)		%r26, %r25, 1b
 | 
						|
	pdc,m		%r23(%r26)
 | 
						|
 | 
						|
	sync
 | 
						|
	bv		%r0(%r2)
 | 
						|
	nop
 | 
						|
	.exit
 | 
						|
 | 
						|
	.procend
 | 
						|
ENDPROC(purge_kernel_dcache_page)
 | 
						|
 | 
						|
#if 0
 | 
						|
	/* Currently not used, but it still is a possible alternate
 | 
						|
	 * solution.
 | 
						|
	 */
 | 
						|
 | 
						|
ENTRY(flush_alias_page)
 | 
						|
	.proc
 | 
						|
	.callinfo NO_CALLS
 | 
						|
	.entry
 | 
						|
 | 
						|
	tophys_r1		%r26
 | 
						|
 | 
						|
	ldil		L%(TMPALIAS_MAP_START), %r28
 | 
						|
#ifdef CONFIG_64BIT
 | 
						|
	extrd,u		%r26, 56,32, %r26	/* convert phys addr to tlb insert format */
 | 
						|
	depd		%r25, 63,22, %r28	/* Form aliased virtual address 'to' */
 | 
						|
	depdi		0, 63,12, %r28		/* Clear any offset bits */
 | 
						|
#else
 | 
						|
	extrw,u		%r26, 24,25, %r26	/* convert phys addr to tlb insert format */
 | 
						|
	depw		%r25, 31,22, %r28	/* Form aliased virtual address 'to' */
 | 
						|
	depwi		0, 31,12, %r28		/* Clear any offset bits */
 | 
						|
#endif
 | 
						|
 | 
						|
	/* Purge any old translation */
 | 
						|
 | 
						|
	pdtlb		0(%r28)
 | 
						|
 | 
						|
	ldil		L%dcache_stride, %r1
 | 
						|
	ldw		R%dcache_stride(%r1), %r23
 | 
						|
 | 
						|
#ifdef CONFIG_64BIT
 | 
						|
	depdi,z		1, 63-PAGE_SHIFT,1, %r29
 | 
						|
#else
 | 
						|
	depwi,z		1, 31-PAGE_SHIFT,1, %r29
 | 
						|
#endif
 | 
						|
	add		%r28, %r29, %r29
 | 
						|
	sub		%r29, %r23, %r29
 | 
						|
 | 
						|
1:      fdc,m		%r23(%r28)
 | 
						|
	fdc,m		%r23(%r28)
 | 
						|
	fdc,m		%r23(%r28)
 | 
						|
	fdc,m		%r23(%r28)
 | 
						|
	fdc,m		%r23(%r28)
 | 
						|
	fdc,m		%r23(%r28)
 | 
						|
	fdc,m		%r23(%r28)
 | 
						|
	fdc,m		%r23(%r28)
 | 
						|
	fdc,m		%r23(%r28)
 | 
						|
	fdc,m		%r23(%r28)
 | 
						|
	fdc,m		%r23(%r28)
 | 
						|
	fdc,m		%r23(%r28)
 | 
						|
	fdc,m		%r23(%r28)
 | 
						|
	fdc,m		%r23(%r28)
 | 
						|
	fdc,m		%r23(%r28)
 | 
						|
	cmpb,COND(<<)		%r28, %r29, 1b
 | 
						|
	fdc,m		%r23(%r28)
 | 
						|
 | 
						|
	sync
 | 
						|
	bv		%r0(%r2)
 | 
						|
	nop
 | 
						|
	.exit
 | 
						|
 | 
						|
	.procend
 | 
						|
#endif
 | 
						|
 | 
						|
	.export flush_user_dcache_range_asm
 | 
						|
 | 
						|
flush_user_dcache_range_asm:
 | 
						|
	.proc
 | 
						|
	.callinfo NO_CALLS
 | 
						|
	.entry
 | 
						|
 | 
						|
	ldil		L%dcache_stride, %r1
 | 
						|
	ldw		R%dcache_stride(%r1), %r23
 | 
						|
	ldo		-1(%r23), %r21
 | 
						|
	ANDCM		%r26, %r21, %r26
 | 
						|
 | 
						|
1:      cmpb,COND(<<),n	%r26, %r25, 1b
 | 
						|
	fdc,m		%r23(%sr3, %r26)
 | 
						|
 | 
						|
	sync
 | 
						|
	bv		%r0(%r2)
 | 
						|
	nop
 | 
						|
	.exit
 | 
						|
 | 
						|
	.procend
 | 
						|
ENDPROC(flush_alias_page)
 | 
						|
 | 
						|
ENTRY(flush_kernel_dcache_range_asm)
 | 
						|
	.proc
 | 
						|
	.callinfo NO_CALLS
 | 
						|
	.entry
 | 
						|
 | 
						|
	ldil		L%dcache_stride, %r1
 | 
						|
	ldw		R%dcache_stride(%r1), %r23
 | 
						|
	ldo		-1(%r23), %r21
 | 
						|
	ANDCM		%r26, %r21, %r26
 | 
						|
 | 
						|
1:      cmpb,COND(<<),n	%r26, %r25,1b
 | 
						|
	fdc,m		%r23(%r26)
 | 
						|
 | 
						|
	sync
 | 
						|
	syncdma
 | 
						|
	bv		%r0(%r2)
 | 
						|
	nop
 | 
						|
	.exit
 | 
						|
 | 
						|
	.procend
 | 
						|
ENDPROC(flush_kernel_dcache_range_asm)
 | 
						|
 | 
						|
ENTRY(flush_user_icache_range_asm)
 | 
						|
	.proc
 | 
						|
	.callinfo NO_CALLS
 | 
						|
	.entry
 | 
						|
 | 
						|
	ldil		L%icache_stride, %r1
 | 
						|
	ldw		R%icache_stride(%r1), %r23
 | 
						|
	ldo		-1(%r23), %r21
 | 
						|
	ANDCM		%r26, %r21, %r26
 | 
						|
 | 
						|
1:      cmpb,COND(<<),n	%r26, %r25,1b
 | 
						|
	fic,m		%r23(%sr3, %r26)
 | 
						|
 | 
						|
	sync
 | 
						|
	bv		%r0(%r2)
 | 
						|
	nop
 | 
						|
	.exit
 | 
						|
 | 
						|
	.procend
 | 
						|
ENDPROC(flush_user_icache_range_asm)
 | 
						|
 | 
						|
ENTRY(flush_kernel_icache_page)
 | 
						|
	.proc
 | 
						|
	.callinfo NO_CALLS
 | 
						|
	.entry
 | 
						|
 | 
						|
	ldil		L%icache_stride, %r1
 | 
						|
	ldw		R%icache_stride(%r1), %r23
 | 
						|
 | 
						|
#ifdef CONFIG_64BIT
 | 
						|
	depdi,z		1, 63-PAGE_SHIFT,1, %r25
 | 
						|
#else
 | 
						|
	depwi,z		1, 31-PAGE_SHIFT,1, %r25
 | 
						|
#endif
 | 
						|
	add		%r26, %r25, %r25
 | 
						|
	sub		%r25, %r23, %r25
 | 
						|
 | 
						|
 | 
						|
1:      fic,m		%r23(%sr4, %r26)
 | 
						|
	fic,m		%r23(%sr4, %r26)
 | 
						|
	fic,m		%r23(%sr4, %r26)
 | 
						|
	fic,m		%r23(%sr4, %r26)
 | 
						|
	fic,m		%r23(%sr4, %r26)
 | 
						|
	fic,m		%r23(%sr4, %r26)
 | 
						|
	fic,m		%r23(%sr4, %r26)
 | 
						|
	fic,m		%r23(%sr4, %r26)
 | 
						|
	fic,m		%r23(%sr4, %r26)
 | 
						|
	fic,m		%r23(%sr4, %r26)
 | 
						|
	fic,m		%r23(%sr4, %r26)
 | 
						|
	fic,m		%r23(%sr4, %r26)
 | 
						|
	fic,m		%r23(%sr4, %r26)
 | 
						|
	fic,m		%r23(%sr4, %r26)
 | 
						|
	fic,m		%r23(%sr4, %r26)
 | 
						|
	cmpb,COND(<<)		%r26, %r25, 1b
 | 
						|
	fic,m		%r23(%sr4, %r26)
 | 
						|
 | 
						|
	sync
 | 
						|
	bv		%r0(%r2)
 | 
						|
	nop
 | 
						|
	.exit
 | 
						|
 | 
						|
	.procend
 | 
						|
ENDPROC(flush_kernel_icache_page)
 | 
						|
 | 
						|
ENTRY(flush_kernel_icache_range_asm)
 | 
						|
	.proc
 | 
						|
	.callinfo NO_CALLS
 | 
						|
	.entry
 | 
						|
 | 
						|
	ldil		L%icache_stride, %r1
 | 
						|
	ldw		R%icache_stride(%r1), %r23
 | 
						|
	ldo		-1(%r23), %r21
 | 
						|
	ANDCM		%r26, %r21, %r26
 | 
						|
 | 
						|
1:      cmpb,COND(<<),n	%r26, %r25, 1b
 | 
						|
	fic,m		%r23(%sr4, %r26)
 | 
						|
 | 
						|
	sync
 | 
						|
	bv		%r0(%r2)
 | 
						|
	nop
 | 
						|
	.exit
 | 
						|
	.procend
 | 
						|
ENDPROC(flush_kernel_icache_range_asm)
 | 
						|
 | 
						|
	/* align should cover use of rfi in disable_sr_hashing_asm and
 | 
						|
	 * srdis_done.
 | 
						|
	 */
 | 
						|
	.align	256
 | 
						|
ENTRY(disable_sr_hashing_asm)
 | 
						|
	.proc
 | 
						|
	.callinfo NO_CALLS
 | 
						|
	.entry
 | 
						|
 | 
						|
	/*
 | 
						|
	 * Switch to real mode
 | 
						|
	 */
 | 
						|
	/* pcxt_ssm_bug */
 | 
						|
	rsm		PSW_SM_I, %r0
 | 
						|
	load32		PA(1f), %r1
 | 
						|
	nop
 | 
						|
	nop
 | 
						|
	nop
 | 
						|
	nop
 | 
						|
	nop
 | 
						|
 | 
						|
	rsm		PSW_SM_Q, %r0		/* prep to load iia queue */
 | 
						|
	mtctl		%r0, %cr17		/* Clear IIASQ tail */
 | 
						|
	mtctl		%r0, %cr17		/* Clear IIASQ head */
 | 
						|
	mtctl		%r1, %cr18		/* IIAOQ head */
 | 
						|
	ldo		4(%r1), %r1
 | 
						|
	mtctl		%r1, %cr18		/* IIAOQ tail */
 | 
						|
	load32		REAL_MODE_PSW, %r1
 | 
						|
	mtctl		%r1, %ipsw
 | 
						|
	rfi
 | 
						|
	nop
 | 
						|
 | 
						|
1:      cmpib,=,n	SRHASH_PCXST, %r26,srdis_pcxs
 | 
						|
	cmpib,=,n	SRHASH_PCXL, %r26,srdis_pcxl
 | 
						|
	cmpib,=,n	SRHASH_PA20, %r26,srdis_pa20
 | 
						|
	b,n		srdis_done
 | 
						|
 | 
						|
srdis_pcxs:
 | 
						|
 | 
						|
	/* Disable Space Register Hashing for PCXS,PCXT,PCXT' */
 | 
						|
 | 
						|
	.word		0x141c1a00		/* mfdiag %dr0, %r28 */
 | 
						|
	.word		0x141c1a00		/* must issue twice */
 | 
						|
	depwi		0,18,1, %r28		/* Clear DHE (dcache hash enable) */
 | 
						|
	depwi		0,20,1, %r28		/* Clear IHE (icache hash enable) */
 | 
						|
	.word		0x141c1600		/* mtdiag %r28, %dr0 */
 | 
						|
	.word		0x141c1600		/* must issue twice */
 | 
						|
	b,n		srdis_done
 | 
						|
 | 
						|
srdis_pcxl:
 | 
						|
 | 
						|
	/* Disable Space Register Hashing for PCXL */
 | 
						|
 | 
						|
	.word		0x141c0600		/* mfdiag %dr0, %r28 */
 | 
						|
	depwi           0,28,2, %r28		/* Clear DHASH_EN & IHASH_EN */
 | 
						|
	.word		0x141c0240		/* mtdiag %r28, %dr0 */
 | 
						|
	b,n		srdis_done
 | 
						|
 | 
						|
srdis_pa20:
 | 
						|
 | 
						|
	/* Disable Space Register Hashing for PCXU,PCXU+,PCXW,PCXW+,PCXW2 */
 | 
						|
 | 
						|
	.word		0x144008bc		/* mfdiag %dr2, %r28 */
 | 
						|
	depdi		0, 54,1, %r28		/* clear DIAG_SPHASH_ENAB (bit 54) */
 | 
						|
	.word		0x145c1840		/* mtdiag %r28, %dr2 */
 | 
						|
 | 
						|
 | 
						|
srdis_done:
 | 
						|
	/* Switch back to virtual mode */
 | 
						|
	rsm		PSW_SM_I, %r0		/* prep to load iia queue */
 | 
						|
	load32 	   	2f, %r1
 | 
						|
	nop
 | 
						|
	nop
 | 
						|
	nop
 | 
						|
	nop
 | 
						|
	nop
 | 
						|
 | 
						|
	rsm		PSW_SM_Q, %r0		/* prep to load iia queue */
 | 
						|
	mtctl		%r0, %cr17		/* Clear IIASQ tail */
 | 
						|
	mtctl		%r0, %cr17		/* Clear IIASQ head */
 | 
						|
	mtctl		%r1, %cr18		/* IIAOQ head */
 | 
						|
	ldo		4(%r1), %r1
 | 
						|
	mtctl		%r1, %cr18		/* IIAOQ tail */
 | 
						|
	load32		KERNEL_PSW, %r1
 | 
						|
	mtctl		%r1, %ipsw
 | 
						|
	rfi
 | 
						|
	nop
 | 
						|
 | 
						|
2:      bv		%r0(%r2)
 | 
						|
	nop
 | 
						|
	.exit
 | 
						|
 | 
						|
	.procend
 | 
						|
ENDPROC(disable_sr_hashing_asm)
 | 
						|
 | 
						|
	.end
 |