Original port to early 2.6 kernel using TI COFF toolchain. Brought up to date by Mark Salter <msalter@redhat.com> Signed-off-by: Aurelien Jacquiot <a-jacquiot@ti.com> Signed-off-by: Mark Salter <msalter@redhat.com> Acked-by: Arnd Bergmann <arnd@arndb.de>
		
			
				
	
	
		
			170 lines
		
	
	
	
		
			4.6 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			170 lines
		
	
	
	
		
			4.6 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
 *  Port on Texas Instruments TMS320C6x architecture
 | 
						|
 *
 | 
						|
 *  Copyright (C) 2004, 2009, 2010 Texas Instruments Incorporated
 | 
						|
 *  Author: Aurelien Jacquiot (aurelien.jacquiot@jaluna.com)
 | 
						|
 *  Rewritten for 2.6.3x: Mark Salter <msalter@redhat.com>
 | 
						|
 *
 | 
						|
 *  This program is free software; you can redistribute it and/or modify
 | 
						|
 *  it under the terms of the GNU General Public License version 2 as
 | 
						|
 *  published by the Free Software Foundation.
 | 
						|
 */
 | 
						|
#ifndef _ASM_C6X_UNALIGNED_H
 | 
						|
#define _ASM_C6X_UNALIGNED_H
 | 
						|
 | 
						|
#include <linux/swab.h>
 | 
						|
 | 
						|
/*
 | 
						|
 * The C64x+ can do unaligned word and dword accesses in hardware
 | 
						|
 * using special load/store instructions.
 | 
						|
 */
 | 
						|
 | 
						|
static inline u16 get_unaligned_le16(const void *p)
 | 
						|
{
 | 
						|
	const u8 *_p = p;
 | 
						|
	return _p[0] | _p[1] << 8;
 | 
						|
}
 | 
						|
 | 
						|
static inline u16 get_unaligned_be16(const void *p)
 | 
						|
{
 | 
						|
	const u8 *_p = p;
 | 
						|
	return _p[0] << 8 | _p[1];
 | 
						|
}
 | 
						|
 | 
						|
static inline void put_unaligned_le16(u16 val, void *p)
 | 
						|
{
 | 
						|
	u8 *_p = p;
 | 
						|
	_p[0] = val;
 | 
						|
	_p[1] = val >> 8;
 | 
						|
}
 | 
						|
 | 
						|
static inline void put_unaligned_be16(u16 val, void *p)
 | 
						|
{
 | 
						|
	u8 *_p = p;
 | 
						|
	_p[0] = val >> 8;
 | 
						|
	_p[1] = val;
 | 
						|
}
 | 
						|
 | 
						|
static inline u32 get_unaligned32(const void *p)
 | 
						|
{
 | 
						|
	u32 val = (u32) p;
 | 
						|
	asm (" ldnw	.d1t1	*%0,%0\n"
 | 
						|
	     " nop     4\n"
 | 
						|
	     : "+a"(val));
 | 
						|
	return val;
 | 
						|
}
 | 
						|
 | 
						|
static inline void put_unaligned32(u32 val, void *p)
 | 
						|
{
 | 
						|
	asm volatile (" stnw	.d2t1	%0,*%1\n"
 | 
						|
		      : : "a"(val), "b"(p) : "memory");
 | 
						|
}
 | 
						|
 | 
						|
static inline u64 get_unaligned64(const void *p)
 | 
						|
{
 | 
						|
	u64 val;
 | 
						|
	asm volatile (" ldndw	.d1t1	*%1,%0\n"
 | 
						|
		      " nop     4\n"
 | 
						|
		      : "=a"(val) : "a"(p));
 | 
						|
	return val;
 | 
						|
}
 | 
						|
 | 
						|
static inline void put_unaligned64(u64 val, const void *p)
 | 
						|
{
 | 
						|
	asm volatile (" stndw	.d2t1	%0,*%1\n"
 | 
						|
		      : : "a"(val), "b"(p) : "memory");
 | 
						|
}
 | 
						|
 | 
						|
#ifdef CONFIG_CPU_BIG_ENDIAN
 | 
						|
 | 
						|
#define get_unaligned_le32(p)	 __swab32(get_unaligned32(p))
 | 
						|
#define get_unaligned_le64(p)	 __swab64(get_unaligned64(p))
 | 
						|
#define get_unaligned_be32(p)	 get_unaligned32(p)
 | 
						|
#define get_unaligned_be64(p)	 get_unaligned64(p)
 | 
						|
#define put_unaligned_le32(v, p) put_unaligned32(__swab32(v), (p))
 | 
						|
#define put_unaligned_le64(v, p) put_unaligned64(__swab64(v), (p))
 | 
						|
#define put_unaligned_be32(v, p) put_unaligned32((v), (p))
 | 
						|
#define put_unaligned_be64(v, p) put_unaligned64((v), (p))
 | 
						|
#define get_unaligned	__get_unaligned_be
 | 
						|
#define put_unaligned	__put_unaligned_be
 | 
						|
 | 
						|
#else
 | 
						|
 | 
						|
#define get_unaligned_le32(p)	 get_unaligned32(p)
 | 
						|
#define get_unaligned_le64(p)	 get_unaligned64(p)
 | 
						|
#define get_unaligned_be32(p)	 __swab32(get_unaligned32(p))
 | 
						|
#define get_unaligned_be64(p)	 __swab64(get_unaligned64(p))
 | 
						|
#define put_unaligned_le32(v, p) put_unaligned32((v), (p))
 | 
						|
#define put_unaligned_le64(v, p) put_unaligned64((v), (p))
 | 
						|
#define put_unaligned_be32(v, p) put_unaligned32(__swab32(v), (p))
 | 
						|
#define put_unaligned_be64(v, p) put_unaligned64(__swab64(v), (p))
 | 
						|
#define get_unaligned	__get_unaligned_le
 | 
						|
#define put_unaligned	__put_unaligned_le
 | 
						|
 | 
						|
#endif
 | 
						|
 | 
						|
/*
 | 
						|
 * Cause a link-time error if we try an unaligned access other than
 | 
						|
 * 1,2,4 or 8 bytes long
 | 
						|
 */
 | 
						|
extern int __bad_unaligned_access_size(void);
 | 
						|
 | 
						|
#define __get_unaligned_le(ptr) (typeof(*(ptr)))({			\
 | 
						|
	sizeof(*(ptr)) == 1 ? *(ptr) :					\
 | 
						|
	  (sizeof(*(ptr)) == 2 ? get_unaligned_le16((ptr)) :		\
 | 
						|
	     (sizeof(*(ptr)) == 4 ? get_unaligned_le32((ptr)) :		\
 | 
						|
		(sizeof(*(ptr)) == 8 ? get_unaligned_le64((ptr)) :	\
 | 
						|
		   __bad_unaligned_access_size())));			\
 | 
						|
	})
 | 
						|
 | 
						|
#define __get_unaligned_be(ptr) (__force typeof(*(ptr)))({	\
 | 
						|
	sizeof(*(ptr)) == 1 ? *(ptr) :					\
 | 
						|
	  (sizeof(*(ptr)) == 2 ? get_unaligned_be16((ptr)) :		\
 | 
						|
	     (sizeof(*(ptr)) == 4 ? get_unaligned_be32((ptr)) :		\
 | 
						|
		(sizeof(*(ptr)) == 8 ? get_unaligned_be64((ptr)) :	\
 | 
						|
		   __bad_unaligned_access_size())));			\
 | 
						|
	})
 | 
						|
 | 
						|
#define __put_unaligned_le(val, ptr) ({					\
 | 
						|
	void *__gu_p = (ptr);						\
 | 
						|
	switch (sizeof(*(ptr))) {					\
 | 
						|
	case 1:								\
 | 
						|
		*(u8 *)__gu_p = (__force u8)(val);			\
 | 
						|
		break;							\
 | 
						|
	case 2:								\
 | 
						|
		put_unaligned_le16((__force u16)(val), __gu_p);		\
 | 
						|
		break;							\
 | 
						|
	case 4:								\
 | 
						|
		put_unaligned_le32((__force u32)(val), __gu_p);		\
 | 
						|
		break;							\
 | 
						|
	case 8:								\
 | 
						|
		put_unaligned_le64((__force u64)(val), __gu_p);		\
 | 
						|
		break;							\
 | 
						|
	default:							\
 | 
						|
		__bad_unaligned_access_size();				\
 | 
						|
		break;							\
 | 
						|
	}								\
 | 
						|
	(void)0; })
 | 
						|
 | 
						|
#define __put_unaligned_be(val, ptr) ({					\
 | 
						|
	void *__gu_p = (ptr);						\
 | 
						|
	switch (sizeof(*(ptr))) {					\
 | 
						|
	case 1:								\
 | 
						|
		*(u8 *)__gu_p = (__force u8)(val);			\
 | 
						|
		break;							\
 | 
						|
	case 2:								\
 | 
						|
		put_unaligned_be16((__force u16)(val), __gu_p);		\
 | 
						|
		break;							\
 | 
						|
	case 4:								\
 | 
						|
		put_unaligned_be32((__force u32)(val), __gu_p);		\
 | 
						|
		break;							\
 | 
						|
	case 8:								\
 | 
						|
		put_unaligned_be64((__force u64)(val), __gu_p);		\
 | 
						|
		break;							\
 | 
						|
	default:							\
 | 
						|
		__bad_unaligned_access_size();				\
 | 
						|
		break;							\
 | 
						|
	}								\
 | 
						|
	(void)0; })
 | 
						|
 | 
						|
#endif /* _ASM_C6X_UNALIGNED_H */
 |