This moves the espfix64 logic into native_iret. To make this work, it gets rid of the native patch for INTERRUPT_RETURN: INTERRUPT_RETURN on native kernels is now 'jmp native_iret'. This changes the 16-bit SS behavior on Xen from OOPSing to leaking some bits of the Xen hypervisor's RSP (I think). [ hpa: this is a nonzero cost on native, but probably not enough to measure. Xen needs to fix this in their own code, probably doing something equivalent to espfix64. ] Signed-off-by: Andy Lutomirski <luto@amacapital.net> Link: http://lkml.kernel.org/r/7b8f1d8ef6597cb16ae004a43c56980a7de3cf94.1406129132.git.luto@amacapital.net Signed-off-by: H. Peter Anvin <hpa@linux.intel.com> Cc: <stable@vger.kernel.org>
		
			
				
	
	
		
			209 lines
		
	
	
	
		
			4 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			209 lines
		
	
	
	
		
			4 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
#ifndef _X86_IRQFLAGS_H_
 | 
						|
#define _X86_IRQFLAGS_H_
 | 
						|
 | 
						|
#include <asm/processor-flags.h>
 | 
						|
 | 
						|
#ifndef __ASSEMBLY__
 | 
						|
/*
 | 
						|
 * Interrupt control:
 | 
						|
 */
 | 
						|
 | 
						|
static inline unsigned long native_save_fl(void)
 | 
						|
{
 | 
						|
	unsigned long flags;
 | 
						|
 | 
						|
	/*
 | 
						|
	 * "=rm" is safe here, because "pop" adjusts the stack before
 | 
						|
	 * it evaluates its effective address -- this is part of the
 | 
						|
	 * documented behavior of the "pop" instruction.
 | 
						|
	 */
 | 
						|
	asm volatile("# __raw_save_flags\n\t"
 | 
						|
		     "pushf ; pop %0"
 | 
						|
		     : "=rm" (flags)
 | 
						|
		     : /* no input */
 | 
						|
		     : "memory");
 | 
						|
 | 
						|
	return flags;
 | 
						|
}
 | 
						|
 | 
						|
static inline void native_restore_fl(unsigned long flags)
 | 
						|
{
 | 
						|
	asm volatile("push %0 ; popf"
 | 
						|
		     : /* no output */
 | 
						|
		     :"g" (flags)
 | 
						|
		     :"memory", "cc");
 | 
						|
}
 | 
						|
 | 
						|
static inline void native_irq_disable(void)
 | 
						|
{
 | 
						|
	asm volatile("cli": : :"memory");
 | 
						|
}
 | 
						|
 | 
						|
static inline void native_irq_enable(void)
 | 
						|
{
 | 
						|
	asm volatile("sti": : :"memory");
 | 
						|
}
 | 
						|
 | 
						|
static inline void native_safe_halt(void)
 | 
						|
{
 | 
						|
	asm volatile("sti; hlt": : :"memory");
 | 
						|
}
 | 
						|
 | 
						|
static inline void native_halt(void)
 | 
						|
{
 | 
						|
	asm volatile("hlt": : :"memory");
 | 
						|
}
 | 
						|
 | 
						|
#endif
 | 
						|
 | 
						|
#ifdef CONFIG_PARAVIRT
 | 
						|
#include <asm/paravirt.h>
 | 
						|
#else
 | 
						|
#ifndef __ASSEMBLY__
 | 
						|
#include <linux/types.h>
 | 
						|
 | 
						|
static inline notrace unsigned long arch_local_save_flags(void)
 | 
						|
{
 | 
						|
	return native_save_fl();
 | 
						|
}
 | 
						|
 | 
						|
static inline notrace void arch_local_irq_restore(unsigned long flags)
 | 
						|
{
 | 
						|
	native_restore_fl(flags);
 | 
						|
}
 | 
						|
 | 
						|
static inline notrace void arch_local_irq_disable(void)
 | 
						|
{
 | 
						|
	native_irq_disable();
 | 
						|
}
 | 
						|
 | 
						|
static inline notrace void arch_local_irq_enable(void)
 | 
						|
{
 | 
						|
	native_irq_enable();
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * Used in the idle loop; sti takes one instruction cycle
 | 
						|
 * to complete:
 | 
						|
 */
 | 
						|
static inline void arch_safe_halt(void)
 | 
						|
{
 | 
						|
	native_safe_halt();
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * Used when interrupts are already enabled or to
 | 
						|
 * shutdown the processor:
 | 
						|
 */
 | 
						|
static inline void halt(void)
 | 
						|
{
 | 
						|
	native_halt();
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * For spinlocks, etc:
 | 
						|
 */
 | 
						|
static inline notrace unsigned long arch_local_irq_save(void)
 | 
						|
{
 | 
						|
	unsigned long flags = arch_local_save_flags();
 | 
						|
	arch_local_irq_disable();
 | 
						|
	return flags;
 | 
						|
}
 | 
						|
#else
 | 
						|
 | 
						|
#define ENABLE_INTERRUPTS(x)	sti
 | 
						|
#define DISABLE_INTERRUPTS(x)	cli
 | 
						|
 | 
						|
#ifdef CONFIG_X86_64
 | 
						|
#define SWAPGS	swapgs
 | 
						|
/*
 | 
						|
 * Currently paravirt can't handle swapgs nicely when we
 | 
						|
 * don't have a stack we can rely on (such as a user space
 | 
						|
 * stack).  So we either find a way around these or just fault
 | 
						|
 * and emulate if a guest tries to call swapgs directly.
 | 
						|
 *
 | 
						|
 * Either way, this is a good way to document that we don't
 | 
						|
 * have a reliable stack. x86_64 only.
 | 
						|
 */
 | 
						|
#define SWAPGS_UNSAFE_STACK	swapgs
 | 
						|
 | 
						|
#define PARAVIRT_ADJUST_EXCEPTION_FRAME	/*  */
 | 
						|
 | 
						|
#define INTERRUPT_RETURN	jmp native_iret
 | 
						|
#define USERGS_SYSRET64				\
 | 
						|
	swapgs;					\
 | 
						|
	sysretq;
 | 
						|
#define USERGS_SYSRET32				\
 | 
						|
	swapgs;					\
 | 
						|
	sysretl
 | 
						|
#define ENABLE_INTERRUPTS_SYSEXIT32		\
 | 
						|
	swapgs;					\
 | 
						|
	sti;					\
 | 
						|
	sysexit
 | 
						|
 | 
						|
#else
 | 
						|
#define INTERRUPT_RETURN		iret
 | 
						|
#define ENABLE_INTERRUPTS_SYSEXIT	sti; sysexit
 | 
						|
#define GET_CR0_INTO_EAX		movl %cr0, %eax
 | 
						|
#endif
 | 
						|
 | 
						|
 | 
						|
#endif /* __ASSEMBLY__ */
 | 
						|
#endif /* CONFIG_PARAVIRT */
 | 
						|
 | 
						|
#ifndef __ASSEMBLY__
 | 
						|
static inline int arch_irqs_disabled_flags(unsigned long flags)
 | 
						|
{
 | 
						|
	return !(flags & X86_EFLAGS_IF);
 | 
						|
}
 | 
						|
 | 
						|
static inline int arch_irqs_disabled(void)
 | 
						|
{
 | 
						|
	unsigned long flags = arch_local_save_flags();
 | 
						|
 | 
						|
	return arch_irqs_disabled_flags(flags);
 | 
						|
}
 | 
						|
 | 
						|
#else
 | 
						|
 | 
						|
#ifdef CONFIG_X86_64
 | 
						|
#define ARCH_LOCKDEP_SYS_EXIT		call lockdep_sys_exit_thunk
 | 
						|
#define ARCH_LOCKDEP_SYS_EXIT_IRQ	\
 | 
						|
	TRACE_IRQS_ON; \
 | 
						|
	sti; \
 | 
						|
	SAVE_REST; \
 | 
						|
	LOCKDEP_SYS_EXIT; \
 | 
						|
	RESTORE_REST; \
 | 
						|
	cli; \
 | 
						|
	TRACE_IRQS_OFF;
 | 
						|
 | 
						|
#else
 | 
						|
#define ARCH_LOCKDEP_SYS_EXIT			\
 | 
						|
	pushl %eax;				\
 | 
						|
	pushl %ecx;				\
 | 
						|
	pushl %edx;				\
 | 
						|
	call lockdep_sys_exit;			\
 | 
						|
	popl %edx;				\
 | 
						|
	popl %ecx;				\
 | 
						|
	popl %eax;
 | 
						|
 | 
						|
#define ARCH_LOCKDEP_SYS_EXIT_IRQ
 | 
						|
#endif
 | 
						|
 | 
						|
#ifdef CONFIG_TRACE_IRQFLAGS
 | 
						|
#  define TRACE_IRQS_ON		call trace_hardirqs_on_thunk;
 | 
						|
#  define TRACE_IRQS_OFF	call trace_hardirqs_off_thunk;
 | 
						|
#else
 | 
						|
#  define TRACE_IRQS_ON
 | 
						|
#  define TRACE_IRQS_OFF
 | 
						|
#endif
 | 
						|
#ifdef CONFIG_DEBUG_LOCK_ALLOC
 | 
						|
#  define LOCKDEP_SYS_EXIT	ARCH_LOCKDEP_SYS_EXIT
 | 
						|
#  define LOCKDEP_SYS_EXIT_IRQ	ARCH_LOCKDEP_SYS_EXIT_IRQ
 | 
						|
# else
 | 
						|
#  define LOCKDEP_SYS_EXIT
 | 
						|
#  define LOCKDEP_SYS_EXIT_IRQ
 | 
						|
# endif
 | 
						|
 | 
						|
#endif /* __ASSEMBLY__ */
 | 
						|
#endif
 |