 70dff4d90a
			
		
	
	
	70dff4d90a
	
	
	
		
			
			Commit d96cc3d1ec "MIPS: Add microMIPS MSA support." attempted to use
the value of a macro within an inline asm statement but instead emitted
a comment leading to the cfcmsa & ctcmsa instructions being omitted. Fix
that by passing CFC_MSA_INSN & CTC_MSA_INSN as arguments to the asm
statements.
Signed-off-by: Paul Burton <paul.burton@imgtec.com>
Cc: linux-mips@linux-mips.org
Patchwork: https://patchwork.linux-mips.org/patch/7305/
Signed-off-by: Ralf Baechle <ralf@linux-mips.org>
		
	
			
		
			
				
	
	
		
			217 lines
		
	
	
	
		
			5.9 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			217 lines
		
	
	
	
		
			5.9 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * Copyright (C) 2013 Imagination Technologies
 | |
|  * Author: Paul Burton <paul.burton@imgtec.com>
 | |
|  *
 | |
|  * This program is free software; you can redistribute it and/or modify it
 | |
|  * under the terms of the GNU General Public License as published by the
 | |
|  * Free Software Foundation;  either version 2 of the  License, or (at your
 | |
|  * option) any later version.
 | |
|  */
 | |
| #ifndef _ASM_MSA_H
 | |
| #define _ASM_MSA_H
 | |
| 
 | |
| #include <asm/mipsregs.h>
 | |
| 
 | |
| #ifndef __ASSEMBLY__
 | |
| 
 | |
| extern void _save_msa(struct task_struct *);
 | |
| extern void _restore_msa(struct task_struct *);
 | |
| extern void _init_msa_upper(void);
 | |
| 
 | |
| static inline void enable_msa(void)
 | |
| {
 | |
| 	if (cpu_has_msa) {
 | |
| 		set_c0_config5(MIPS_CONF5_MSAEN);
 | |
| 		enable_fpu_hazard();
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static inline void disable_msa(void)
 | |
| {
 | |
| 	if (cpu_has_msa) {
 | |
| 		clear_c0_config5(MIPS_CONF5_MSAEN);
 | |
| 		disable_fpu_hazard();
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static inline int is_msa_enabled(void)
 | |
| {
 | |
| 	if (!cpu_has_msa)
 | |
| 		return 0;
 | |
| 
 | |
| 	return read_c0_config5() & MIPS_CONF5_MSAEN;
 | |
| }
 | |
| 
 | |
| static inline int thread_msa_context_live(void)
 | |
| {
 | |
| 	/*
 | |
| 	 * Check cpu_has_msa only if it's a constant. This will allow the
 | |
| 	 * compiler to optimise out code for CPUs without MSA without adding
 | |
| 	 * an extra redundant check for CPUs with MSA.
 | |
| 	 */
 | |
| 	if (__builtin_constant_p(cpu_has_msa) && !cpu_has_msa)
 | |
| 		return 0;
 | |
| 
 | |
| 	return test_thread_flag(TIF_MSA_CTX_LIVE);
 | |
| }
 | |
| 
 | |
| static inline void save_msa(struct task_struct *t)
 | |
| {
 | |
| 	if (cpu_has_msa)
 | |
| 		_save_msa(t);
 | |
| }
 | |
| 
 | |
| static inline void restore_msa(struct task_struct *t)
 | |
| {
 | |
| 	if (cpu_has_msa)
 | |
| 		_restore_msa(t);
 | |
| }
 | |
| 
 | |
| #ifdef TOOLCHAIN_SUPPORTS_MSA
 | |
| 
 | |
| #define __BUILD_MSA_CTL_REG(name, cs)				\
 | |
| static inline unsigned int read_msa_##name(void)		\
 | |
| {								\
 | |
| 	unsigned int reg;					\
 | |
| 	__asm__ __volatile__(					\
 | |
| 	"	.set	push\n"					\
 | |
| 	"	.set	msa\n"					\
 | |
| 	"	cfcmsa	%0, $" #cs "\n"				\
 | |
| 	"	.set	pop\n"					\
 | |
| 	: "=r"(reg));						\
 | |
| 	return reg;						\
 | |
| }								\
 | |
| 								\
 | |
| static inline void write_msa_##name(unsigned int val)		\
 | |
| {								\
 | |
| 	__asm__ __volatile__(					\
 | |
| 	"	.set	push\n"					\
 | |
| 	"	.set	msa\n"					\
 | |
| 	"	ctcmsa	$" #cs ", %0\n"				\
 | |
| 	"	.set	pop\n"					\
 | |
| 	: : "r"(val));						\
 | |
| }
 | |
| 
 | |
| #else /* !TOOLCHAIN_SUPPORTS_MSA */
 | |
| 
 | |
| /*
 | |
|  * Define functions using .word for the c[ft]cmsa instructions in order to
 | |
|  * allow compilation with toolchains that do not support MSA. Once all
 | |
|  * toolchains in use support MSA these can be removed.
 | |
|  */
 | |
| #ifdef CONFIG_CPU_MICROMIPS
 | |
| #define CFC_MSA_INSN	0x587e0056
 | |
| #define CTC_MSA_INSN	0x583e0816
 | |
| #else
 | |
| #define CFC_MSA_INSN	0x787e0059
 | |
| #define CTC_MSA_INSN	0x783e0819
 | |
| #endif
 | |
| 
 | |
| #define __BUILD_MSA_CTL_REG(name, cs)				\
 | |
| static inline unsigned int read_msa_##name(void)		\
 | |
| {								\
 | |
| 	unsigned int reg;					\
 | |
| 	__asm__ __volatile__(					\
 | |
| 	"	.set	push\n"					\
 | |
| 	"	.set	noat\n"					\
 | |
| 	"	.insn\n"					\
 | |
| 	"	.word	%1 | (" #cs " << 11)\n"			\
 | |
| 	"	move	%0, $1\n"				\
 | |
| 	"	.set	pop\n"					\
 | |
| 	: "=r"(reg) : "i"(CFC_MSA_INSN));			\
 | |
| 	return reg;						\
 | |
| }								\
 | |
| 								\
 | |
| static inline void write_msa_##name(unsigned int val)		\
 | |
| {								\
 | |
| 	__asm__ __volatile__(					\
 | |
| 	"	.set	push\n"					\
 | |
| 	"	.set	noat\n"					\
 | |
| 	"	move	$1, %0\n"				\
 | |
| 	"	.insn\n"					\
 | |
| 	"	.word	%1 | (" #cs " << 6)\n"			\
 | |
| 	"	.set	pop\n"					\
 | |
| 	: : "r"(val), "i"(CTC_MSA_INSN));			\
 | |
| }
 | |
| 
 | |
| #endif /* !TOOLCHAIN_SUPPORTS_MSA */
 | |
| 
 | |
| __BUILD_MSA_CTL_REG(ir, 0)
 | |
| __BUILD_MSA_CTL_REG(csr, 1)
 | |
| __BUILD_MSA_CTL_REG(access, 2)
 | |
| __BUILD_MSA_CTL_REG(save, 3)
 | |
| __BUILD_MSA_CTL_REG(modify, 4)
 | |
| __BUILD_MSA_CTL_REG(request, 5)
 | |
| __BUILD_MSA_CTL_REG(map, 6)
 | |
| __BUILD_MSA_CTL_REG(unmap, 7)
 | |
| 
 | |
| #endif /* !__ASSEMBLY__ */
 | |
| 
 | |
| #define MSA_IR		0
 | |
| #define MSA_CSR		1
 | |
| #define MSA_ACCESS	2
 | |
| #define MSA_SAVE	3
 | |
| #define MSA_MODIFY	4
 | |
| #define MSA_REQUEST	5
 | |
| #define MSA_MAP		6
 | |
| #define MSA_UNMAP	7
 | |
| 
 | |
| /* MSA Implementation Register (MSAIR) */
 | |
| #define MSA_IR_REVB		0
 | |
| #define MSA_IR_REVF		(_ULCAST_(0xff) << MSA_IR_REVB)
 | |
| #define MSA_IR_PROCB		8
 | |
| #define MSA_IR_PROCF		(_ULCAST_(0xff) << MSA_IR_PROCB)
 | |
| #define MSA_IR_WRPB		16
 | |
| #define MSA_IR_WRPF		(_ULCAST_(0x1) << MSA_IR_WRPB)
 | |
| 
 | |
| /* MSA Control & Status Register (MSACSR) */
 | |
| #define MSA_CSR_RMB		0
 | |
| #define MSA_CSR_RMF		(_ULCAST_(0x3) << MSA_CSR_RMB)
 | |
| #define MSA_CSR_RM_NEAREST	0
 | |
| #define MSA_CSR_RM_TO_ZERO	1
 | |
| #define MSA_CSR_RM_TO_POS	2
 | |
| #define MSA_CSR_RM_TO_NEG	3
 | |
| #define MSA_CSR_FLAGSB		2
 | |
| #define MSA_CSR_FLAGSF		(_ULCAST_(0x1f) << MSA_CSR_FLAGSB)
 | |
| #define MSA_CSR_FLAGS_IB	2
 | |
| #define MSA_CSR_FLAGS_IF	(_ULCAST_(0x1) << MSA_CSR_FLAGS_IB)
 | |
| #define MSA_CSR_FLAGS_UB	3
 | |
| #define MSA_CSR_FLAGS_UF	(_ULCAST_(0x1) << MSA_CSR_FLAGS_UB)
 | |
| #define MSA_CSR_FLAGS_OB	4
 | |
| #define MSA_CSR_FLAGS_OF	(_ULCAST_(0x1) << MSA_CSR_FLAGS_OB)
 | |
| #define MSA_CSR_FLAGS_ZB	5
 | |
| #define MSA_CSR_FLAGS_ZF	(_ULCAST_(0x1) << MSA_CSR_FLAGS_ZB)
 | |
| #define MSA_CSR_FLAGS_VB	6
 | |
| #define MSA_CSR_FLAGS_VF	(_ULCAST_(0x1) << MSA_CSR_FLAGS_VB)
 | |
| #define MSA_CSR_ENABLESB	7
 | |
| #define MSA_CSR_ENABLESF	(_ULCAST_(0x1f) << MSA_CSR_ENABLESB)
 | |
| #define MSA_CSR_ENABLES_IB	7
 | |
| #define MSA_CSR_ENABLES_IF	(_ULCAST_(0x1) << MSA_CSR_ENABLES_IB)
 | |
| #define MSA_CSR_ENABLES_UB	8
 | |
| #define MSA_CSR_ENABLES_UF	(_ULCAST_(0x1) << MSA_CSR_ENABLES_UB)
 | |
| #define MSA_CSR_ENABLES_OB	9
 | |
| #define MSA_CSR_ENABLES_OF	(_ULCAST_(0x1) << MSA_CSR_ENABLES_OB)
 | |
| #define MSA_CSR_ENABLES_ZB	10
 | |
| #define MSA_CSR_ENABLES_ZF	(_ULCAST_(0x1) << MSA_CSR_ENABLES_ZB)
 | |
| #define MSA_CSR_ENABLES_VB	11
 | |
| #define MSA_CSR_ENABLES_VF	(_ULCAST_(0x1) << MSA_CSR_ENABLES_VB)
 | |
| #define MSA_CSR_CAUSEB		12
 | |
| #define MSA_CSR_CAUSEF		(_ULCAST_(0x3f) << MSA_CSR_CAUSEB)
 | |
| #define MSA_CSR_CAUSE_IB	12
 | |
| #define MSA_CSR_CAUSE_IF	(_ULCAST_(0x1) << MSA_CSR_CAUSE_IB)
 | |
| #define MSA_CSR_CAUSE_UB	13
 | |
| #define MSA_CSR_CAUSE_UF	(_ULCAST_(0x1) << MSA_CSR_CAUSE_UB)
 | |
| #define MSA_CSR_CAUSE_OB	14
 | |
| #define MSA_CSR_CAUSE_OF	(_ULCAST_(0x1) << MSA_CSR_CAUSE_OB)
 | |
| #define MSA_CSR_CAUSE_ZB	15
 | |
| #define MSA_CSR_CAUSE_ZF	(_ULCAST_(0x1) << MSA_CSR_CAUSE_ZB)
 | |
| #define MSA_CSR_CAUSE_VB	16
 | |
| #define MSA_CSR_CAUSE_VF	(_ULCAST_(0x1) << MSA_CSR_CAUSE_VB)
 | |
| #define MSA_CSR_CAUSE_EB	17
 | |
| #define MSA_CSR_CAUSE_EF	(_ULCAST_(0x1) << MSA_CSR_CAUSE_EB)
 | |
| #define MSA_CSR_NXB		18
 | |
| #define MSA_CSR_NXF		(_ULCAST_(0x1) << MSA_CSR_NXB)
 | |
| #define MSA_CSR_FSB		24
 | |
| #define MSA_CSR_FSF		(_ULCAST_(0x1) << MSA_CSR_FSB)
 | |
| 
 | |
| #endif /* _ASM_MSA_H */
 |