 7056741fd9
			
		
	
	
	7056741fd9
	
	
	
		
			
			Optimize RAID6 recovery functions to take advantage of the 256-bit YMM integer instructions introduced in AVX2. The patch was tested and benchmarked before submission. However hardware is not yet released so benchmark numbers cannot be reported. Acked-by: "H. Peter Anvin" <hpa@zytor.com> Signed-off-by: Jim Kukunas <james.t.kukunas@linux.intel.com> Signed-off-by: NeilBrown <neilb@suse.de>
		
			
				
	
	
		
			70 lines
		
	
	
	
		
			2 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			70 lines
		
	
	
	
		
			2 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
| /* ----------------------------------------------------------------------- *
 | |
|  *
 | |
|  *   Copyright 2002-2004 H. Peter Anvin - All Rights Reserved
 | |
|  *
 | |
|  *   This program is free software; you can redistribute it and/or modify
 | |
|  *   it under the terms of the GNU General Public License as published by
 | |
|  *   the Free Software Foundation, Inc., 53 Temple Place Ste 330,
 | |
|  *   Boston MA 02111-1307, USA; either version 2 of the License, or
 | |
|  *   (at your option) any later version; incorporated herein by reference.
 | |
|  *
 | |
|  * ----------------------------------------------------------------------- */
 | |
| 
 | |
| /*
 | |
|  * raid6/x86.h
 | |
|  *
 | |
|  * Definitions common to x86 and x86-64 RAID-6 code only
 | |
|  */
 | |
| 
 | |
| #ifndef LINUX_RAID_RAID6X86_H
 | |
| #define LINUX_RAID_RAID6X86_H
 | |
| 
 | |
| #if (defined(__i386__) || defined(__x86_64__)) && !defined(__arch_um__)
 | |
| 
 | |
| #ifdef __KERNEL__ /* Real code */
 | |
| 
 | |
| #include <asm/i387.h>
 | |
| 
 | |
| #else /* Dummy code for user space testing */
 | |
| 
 | |
| static inline void kernel_fpu_begin(void)
 | |
| {
 | |
| }
 | |
| 
 | |
| static inline void kernel_fpu_end(void)
 | |
| {
 | |
| }
 | |
| 
 | |
| #define __aligned(x) __attribute__((aligned(x)))
 | |
| 
 | |
| #define X86_FEATURE_MMX		(0*32+23) /* Multimedia Extensions */
 | |
| #define X86_FEATURE_FXSR	(0*32+24) /* FXSAVE and FXRSTOR instructions
 | |
| 					   * (fast save and restore) */
 | |
| #define X86_FEATURE_XMM		(0*32+25) /* Streaming SIMD Extensions */
 | |
| #define X86_FEATURE_XMM2	(0*32+26) /* Streaming SIMD Extensions-2 */
 | |
| #define X86_FEATURE_XMM3	(4*32+ 0) /* "pni" SSE-3 */
 | |
| #define X86_FEATURE_SSSE3	(4*32+ 9) /* Supplemental SSE-3 */
 | |
| #define X86_FEATURE_AVX	(4*32+28) /* Advanced Vector Extensions */
 | |
| #define X86_FEATURE_AVX2        (9*32+ 5) /* AVX2 instructions */
 | |
| #define X86_FEATURE_MMXEXT	(1*32+22) /* AMD MMX extensions */
 | |
| 
 | |
| /* Should work well enough on modern CPUs for testing */
 | |
| static inline int boot_cpu_has(int flag)
 | |
| {
 | |
| 	u32 eax, ebx, ecx, edx;
 | |
| 
 | |
| 	eax = (flag & 0x100) ? 7 :
 | |
| 		(flag & 0x20) ? 0x80000001 : 1;
 | |
| 	ecx = 0;
 | |
| 
 | |
| 	asm volatile("cpuid"
 | |
| 		     : "+a" (eax), "=b" (ebx), "=d" (edx), "+c" (ecx));
 | |
| 
 | |
| 	return ((flag & 0x100 ? ebx :
 | |
| 		(flag & 0x80) ? ecx : edx) >> (flag & 31)) & 1;
 | |
| }
 | |
| 
 | |
| #endif /* ndef __KERNEL__ */
 | |
| 
 | |
| #endif
 | |
| #endif
 |