Pull x86 vdso updates from Ingo Molnar: "Various vDSO updates from Andy Lutomirski, mostly cleanups and reorganization to improve maintainability, but also some micro-optimizations and robustization changes" * 'x86-vdso-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip: x86_64/vsyscall: Restore orig_ax after vsyscall seccomp x86_64: Add a comment explaining the TASK_SIZE_MAX guard page x86_64,vsyscall: Make vsyscall emulation configurable x86_64, vsyscall: Rewrite comment and clean up headers in vsyscall code x86_64, vsyscall: Turn vsyscalls all the way off when vsyscall==none x86,vdso: Use LSL unconditionally for vgetcpu x86: vdso: Fix build with older gcc x86_64/vdso: Clean up vgetcpu init and merge the vdso initcalls x86_64/vdso: Remove jiffies from the vvar page x86/vdso: Make the PER_CPU segment 32 bits x86/vdso: Make the PER_CPU segment start out accessed x86/vdso: Change the PER_CPU segment to use struct desc_struct x86_64/vdso: Move getcpu code from vsyscall_64.c to vdso/vma.c x86_64/vsyscall: Move all of the gate_area code to vsyscall_64.c
		
			
				
	
	
		
			170 lines
		
	
	
	
		
			4.9 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			170 lines
		
	
	
	
		
			4.9 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
 * fixmap.h: compile-time virtual memory allocation
 | 
						|
 *
 | 
						|
 * This file is subject to the terms and conditions of the GNU General Public
 | 
						|
 * License.  See the file "COPYING" in the main directory of this archive
 | 
						|
 * for more details.
 | 
						|
 *
 | 
						|
 * Copyright (C) 1998 Ingo Molnar
 | 
						|
 *
 | 
						|
 * Support of BIGMEM added by Gerhard Wichert, Siemens AG, July 1999
 | 
						|
 * x86_32 and x86_64 integration by Gustavo F. Padovan, February 2009
 | 
						|
 */
 | 
						|
 | 
						|
#ifndef _ASM_X86_FIXMAP_H
 | 
						|
#define _ASM_X86_FIXMAP_H
 | 
						|
 | 
						|
#ifndef __ASSEMBLY__
 | 
						|
#include <linux/kernel.h>
 | 
						|
#include <asm/acpi.h>
 | 
						|
#include <asm/apicdef.h>
 | 
						|
#include <asm/page.h>
 | 
						|
#include <asm/pvclock.h>
 | 
						|
#ifdef CONFIG_X86_32
 | 
						|
#include <linux/threads.h>
 | 
						|
#include <asm/kmap_types.h>
 | 
						|
#else
 | 
						|
#include <uapi/asm/vsyscall.h>
 | 
						|
#endif
 | 
						|
 | 
						|
/*
 | 
						|
 * We can't declare FIXADDR_TOP as variable for x86_64 because vsyscall
 | 
						|
 * uses fixmaps that relies on FIXADDR_TOP for proper address calculation.
 | 
						|
 * Because of this, FIXADDR_TOP x86 integration was left as later work.
 | 
						|
 */
 | 
						|
#ifdef CONFIG_X86_32
 | 
						|
/* used by vmalloc.c, vsyscall.lds.S.
 | 
						|
 *
 | 
						|
 * Leave one empty page between vmalloc'ed areas and
 | 
						|
 * the start of the fixmap.
 | 
						|
 */
 | 
						|
extern unsigned long __FIXADDR_TOP;
 | 
						|
#define FIXADDR_TOP	((unsigned long)__FIXADDR_TOP)
 | 
						|
#else
 | 
						|
#define FIXADDR_TOP	(round_up(VSYSCALL_ADDR + PAGE_SIZE, 1<<PMD_SHIFT) - \
 | 
						|
			 PAGE_SIZE)
 | 
						|
#endif
 | 
						|
 | 
						|
 | 
						|
/*
 | 
						|
 * Here we define all the compile-time 'special' virtual
 | 
						|
 * addresses. The point is to have a constant address at
 | 
						|
 * compile time, but to set the physical address only
 | 
						|
 * in the boot process.
 | 
						|
 * for x86_32: We allocate these special addresses
 | 
						|
 * from the end of virtual memory (0xfffff000) backwards.
 | 
						|
 * Also this lets us do fail-safe vmalloc(), we
 | 
						|
 * can guarantee that these special addresses and
 | 
						|
 * vmalloc()-ed addresses never overlap.
 | 
						|
 *
 | 
						|
 * These 'compile-time allocated' memory buffers are
 | 
						|
 * fixed-size 4k pages (or larger if used with an increment
 | 
						|
 * higher than 1). Use set_fixmap(idx,phys) to associate
 | 
						|
 * physical memory with fixmap indices.
 | 
						|
 *
 | 
						|
 * TLB entries of such buffers will not be flushed across
 | 
						|
 * task switches.
 | 
						|
 */
 | 
						|
enum fixed_addresses {
 | 
						|
#ifdef CONFIG_X86_32
 | 
						|
	FIX_HOLE,
 | 
						|
#else
 | 
						|
#ifdef CONFIG_X86_VSYSCALL_EMULATION
 | 
						|
	VSYSCALL_PAGE = (FIXADDR_TOP - VSYSCALL_ADDR) >> PAGE_SHIFT,
 | 
						|
#endif
 | 
						|
#ifdef CONFIG_PARAVIRT_CLOCK
 | 
						|
	PVCLOCK_FIXMAP_BEGIN,
 | 
						|
	PVCLOCK_FIXMAP_END = PVCLOCK_FIXMAP_BEGIN+PVCLOCK_VSYSCALL_NR_PAGES-1,
 | 
						|
#endif
 | 
						|
#endif
 | 
						|
	FIX_DBGP_BASE,
 | 
						|
	FIX_EARLYCON_MEM_BASE,
 | 
						|
#ifdef CONFIG_PROVIDE_OHCI1394_DMA_INIT
 | 
						|
	FIX_OHCI1394_BASE,
 | 
						|
#endif
 | 
						|
#ifdef CONFIG_X86_LOCAL_APIC
 | 
						|
	FIX_APIC_BASE,	/* local (CPU) APIC) -- required for SMP or not */
 | 
						|
#endif
 | 
						|
#ifdef CONFIG_X86_IO_APIC
 | 
						|
	FIX_IO_APIC_BASE_0,
 | 
						|
	FIX_IO_APIC_BASE_END = FIX_IO_APIC_BASE_0 + MAX_IO_APICS - 1,
 | 
						|
#endif
 | 
						|
	FIX_RO_IDT,	/* Virtual mapping for read-only IDT */
 | 
						|
#ifdef CONFIG_X86_32
 | 
						|
	FIX_KMAP_BEGIN,	/* reserved pte's for temporary kernel mappings */
 | 
						|
	FIX_KMAP_END = FIX_KMAP_BEGIN+(KM_TYPE_NR*NR_CPUS)-1,
 | 
						|
#ifdef CONFIG_PCI_MMCONFIG
 | 
						|
	FIX_PCIE_MCFG,
 | 
						|
#endif
 | 
						|
#endif
 | 
						|
#ifdef CONFIG_PARAVIRT
 | 
						|
	FIX_PARAVIRT_BOOTMAP,
 | 
						|
#endif
 | 
						|
	FIX_TEXT_POKE1,	/* reserve 2 pages for text_poke() */
 | 
						|
	FIX_TEXT_POKE0, /* first page is last, because allocation is backward */
 | 
						|
#ifdef	CONFIG_X86_INTEL_MID
 | 
						|
	FIX_LNW_VRTC,
 | 
						|
#endif
 | 
						|
	__end_of_permanent_fixed_addresses,
 | 
						|
 | 
						|
	/*
 | 
						|
	 * 512 temporary boot-time mappings, used by early_ioremap(),
 | 
						|
	 * before ioremap() is functional.
 | 
						|
	 *
 | 
						|
	 * If necessary we round it up to the next 512 pages boundary so
 | 
						|
	 * that we can have a single pgd entry and a single pte table:
 | 
						|
	 */
 | 
						|
#define NR_FIX_BTMAPS		64
 | 
						|
#define FIX_BTMAPS_SLOTS	8
 | 
						|
#define TOTAL_FIX_BTMAPS	(NR_FIX_BTMAPS * FIX_BTMAPS_SLOTS)
 | 
						|
	FIX_BTMAP_END =
 | 
						|
	 (__end_of_permanent_fixed_addresses ^
 | 
						|
	  (__end_of_permanent_fixed_addresses + TOTAL_FIX_BTMAPS - 1)) &
 | 
						|
	 -PTRS_PER_PTE
 | 
						|
	 ? __end_of_permanent_fixed_addresses + TOTAL_FIX_BTMAPS -
 | 
						|
	   (__end_of_permanent_fixed_addresses & (TOTAL_FIX_BTMAPS - 1))
 | 
						|
	 : __end_of_permanent_fixed_addresses,
 | 
						|
	FIX_BTMAP_BEGIN = FIX_BTMAP_END + TOTAL_FIX_BTMAPS - 1,
 | 
						|
#ifdef CONFIG_X86_32
 | 
						|
	FIX_WP_TEST,
 | 
						|
#endif
 | 
						|
#ifdef CONFIG_INTEL_TXT
 | 
						|
	FIX_TBOOT_BASE,
 | 
						|
#endif
 | 
						|
	__end_of_fixed_addresses
 | 
						|
};
 | 
						|
 | 
						|
 | 
						|
extern void reserve_top_address(unsigned long reserve);
 | 
						|
 | 
						|
#define FIXADDR_SIZE	(__end_of_permanent_fixed_addresses << PAGE_SHIFT)
 | 
						|
#define FIXADDR_START		(FIXADDR_TOP - FIXADDR_SIZE)
 | 
						|
 | 
						|
extern int fixmaps_set;
 | 
						|
 | 
						|
extern pte_t *kmap_pte;
 | 
						|
extern pgprot_t kmap_prot;
 | 
						|
extern pte_t *pkmap_page_table;
 | 
						|
 | 
						|
void __native_set_fixmap(enum fixed_addresses idx, pte_t pte);
 | 
						|
void native_set_fixmap(enum fixed_addresses idx,
 | 
						|
		       phys_addr_t phys, pgprot_t flags);
 | 
						|
 | 
						|
#ifndef CONFIG_PARAVIRT
 | 
						|
static inline void __set_fixmap(enum fixed_addresses idx,
 | 
						|
				phys_addr_t phys, pgprot_t flags)
 | 
						|
{
 | 
						|
	native_set_fixmap(idx, phys, flags);
 | 
						|
}
 | 
						|
#endif
 | 
						|
 | 
						|
#include <asm-generic/fixmap.h>
 | 
						|
 | 
						|
#define __late_set_fixmap(idx, phys, flags) __set_fixmap(idx, phys, flags)
 | 
						|
#define __late_clear_fixmap(idx) __set_fixmap(idx, 0, __pgprot(0))
 | 
						|
 | 
						|
void __early_set_fixmap(enum fixed_addresses idx,
 | 
						|
			phys_addr_t phys, pgprot_t flags);
 | 
						|
 | 
						|
#endif /* !__ASSEMBLY__ */
 | 
						|
#endif /* _ASM_X86_FIXMAP_H */
 |