 b9085bcbf5
			
		
	
	
	b9085bcbf5
	
	
	
		
			
			Common: Optional support for adding a small amount of polling on each HLT
 instruction executed in the guest (or equivalent for other architectures).
 This can improve latency up to 50% on some scenarios (e.g. O_DSYNC writes
 or TCP_RR netperf tests).  This also has to be enabled manually for now,
 but the plan is to auto-tune this in the future.
 
 ARM/ARM64: the highlights are support for GICv3 emulation and dirty page
 tracking
 
 s390: several optimizations and bugfixes.  Also a first: a feature
 exposed by KVM (UUID and long guest name in /proc/sysinfo) before
 it is available in IBM's hypervisor! :)
 
 MIPS: Bugfixes.
 
 x86: Support for PML (page modification logging, a new feature in
 Broadwell Xeons that speeds up dirty page tracking), nested virtualization
 improvements (nested APICv---a nice optimization), usual round of emulation
 fixes.  There is also a new option to reduce latency of the TSC deadline
 timer in the guest; this needs to be tuned manually.
 
 Some commits are common between this pull and Catalin's; I see you
 have already included his tree.
 
 ARM has other conflicts where functions are added in the same place
 by 3.19-rc and 3.20 patches.  These are not large though, and entirely
 within KVM.
 -----BEGIN PGP SIGNATURE-----
 Version: GnuPG v2.0.22 (GNU/Linux)
 
 iQEcBAABAgAGBQJU28rkAAoJEL/70l94x66DXqQH/1TDOfJIjW7P2kb0Sw7Fy1wi
 cEX1KO/VFxAqc8R0E/0Wb55CXyPjQJM6xBXuFr5cUDaIjQ8ULSktL4pEwXyyv/s5
 DBDkN65mriry2w5VuEaRLVcuX9Wy+tqLQXWNkEySfyb4uhZChWWHvKEcgw5SqCyg
 NlpeHurYESIoNyov3jWqvBjr4OmaQENyv7t2c6q5ErIgG02V+iCux5QGbphM2IC9
 LFtPKxoqhfeB2xFxTOIt8HJiXrZNwflsTejIlCl/NSEiDVLLxxHCxK2tWK/tUXMn
 JfLD9ytXBWtNMwInvtFm4fPmDouv2VDyR0xnK2db+/axsJZnbxqjGu1um4Dqbak=
 =7gdx
 -----END PGP SIGNATURE-----
Merge tag 'for-linus' of git://git.kernel.org/pub/scm/virt/kvm/kvm
Pull KVM update from Paolo Bonzini:
 "Fairly small update, but there are some interesting new features.
  Common:
     Optional support for adding a small amount of polling on each HLT
     instruction executed in the guest (or equivalent for other
     architectures).  This can improve latency up to 50% on some
     scenarios (e.g. O_DSYNC writes or TCP_RR netperf tests).  This
     also has to be enabled manually for now, but the plan is to
     auto-tune this in the future.
  ARM/ARM64:
     The highlights are support for GICv3 emulation and dirty page
     tracking
  s390:
     Several optimizations and bugfixes.  Also a first: a feature
     exposed by KVM (UUID and long guest name in /proc/sysinfo) before
     it is available in IBM's hypervisor! :)
  MIPS:
     Bugfixes.
  x86:
     Support for PML (page modification logging, a new feature in
     Broadwell Xeons that speeds up dirty page tracking), nested
     virtualization improvements (nested APICv---a nice optimization),
     usual round of emulation fixes.
     There is also a new option to reduce latency of the TSC deadline
     timer in the guest; this needs to be tuned manually.
     Some commits are common between this pull and Catalin's; I see you
     have already included his tree.
  Powerpc:
     Nothing yet.
     The KVM/PPC changes will come in through the PPC maintainers,
     because I haven't received them yet and I might end up being
     offline for some part of next week"
* tag 'for-linus' of git://git.kernel.org/pub/scm/virt/kvm/kvm: (130 commits)
  KVM: ia64: drop kvm.h from installed user headers
  KVM: x86: fix build with !CONFIG_SMP
  KVM: x86: emulate: correct page fault error code for NoWrite instructions
  KVM: Disable compat ioctl for s390
  KVM: s390: add cpu model support
  KVM: s390: use facilities and cpu_id per KVM
  KVM: s390/CPACF: Choose crypto control block format
  s390/kernel: Update /proc/sysinfo file with Extended Name and UUID
  KVM: s390: reenable LPP facility
  KVM: s390: floating irqs: fix user triggerable endless loop
  kvm: add halt_poll_ns module parameter
  kvm: remove KVM_MMIO_SIZE
  KVM: MIPS: Don't leak FPU/DSP to guest
  KVM: MIPS: Disable HTW while in guest
  KVM: nVMX: Enable nested posted interrupt processing
  KVM: nVMX: Enable nested virtual interrupt delivery
  KVM: nVMX: Enable nested apic register virtualization
  KVM: nVMX: Make nested control MSRs per-cpu
  KVM: nVMX: Enable nested virtualize x2apic mode
  KVM: nVMX: Prepare for using hardware MSR bitmap
  ...
		
	
			
		
			
				
	
	
		
			166 lines
		
	
	
	
		
			8.1 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			166 lines
		
	
	
	
		
			8.1 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * Based on arch/arm/kernel/asm-offsets.c
 | |
|  *
 | |
|  * Copyright (C) 1995-2003 Russell King
 | |
|  *               2001-2002 Keith Owens
 | |
|  * Copyright (C) 2012 ARM Ltd.
 | |
|  *
 | |
|  * This program is free software; you can redistribute it and/or modify
 | |
|  * it under the terms of the GNU General Public License version 2 as
 | |
|  * published by the Free Software Foundation.
 | |
|  *
 | |
|  * This program is distributed in the hope that it will be useful,
 | |
|  * but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
|  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | |
|  * GNU General Public License for more details.
 | |
|  *
 | |
|  * You should have received a copy of the GNU General Public License
 | |
|  * along with this program.  If not, see <http://www.gnu.org/licenses/>.
 | |
|  */
 | |
| 
 | |
| #include <linux/sched.h>
 | |
| #include <linux/mm.h>
 | |
| #include <linux/dma-mapping.h>
 | |
| #include <linux/kvm_host.h>
 | |
| #include <asm/thread_info.h>
 | |
| #include <asm/memory.h>
 | |
| #include <asm/cputable.h>
 | |
| #include <asm/smp_plat.h>
 | |
| #include <asm/suspend.h>
 | |
| #include <asm/vdso_datapage.h>
 | |
| #include <linux/kbuild.h>
 | |
| 
 | |
| int main(void)
 | |
| {
 | |
|   DEFINE(TSK_ACTIVE_MM,		offsetof(struct task_struct, active_mm));
 | |
|   BLANK();
 | |
|   DEFINE(TI_FLAGS,		offsetof(struct thread_info, flags));
 | |
|   DEFINE(TI_PREEMPT,		offsetof(struct thread_info, preempt_count));
 | |
|   DEFINE(TI_ADDR_LIMIT,		offsetof(struct thread_info, addr_limit));
 | |
|   DEFINE(TI_TASK,		offsetof(struct thread_info, task));
 | |
|   DEFINE(TI_EXEC_DOMAIN,	offsetof(struct thread_info, exec_domain));
 | |
|   DEFINE(TI_CPU,		offsetof(struct thread_info, cpu));
 | |
|   BLANK();
 | |
|   DEFINE(THREAD_CPU_CONTEXT,	offsetof(struct task_struct, thread.cpu_context));
 | |
|   BLANK();
 | |
|   DEFINE(S_X0,			offsetof(struct pt_regs, regs[0]));
 | |
|   DEFINE(S_X1,			offsetof(struct pt_regs, regs[1]));
 | |
|   DEFINE(S_X2,			offsetof(struct pt_regs, regs[2]));
 | |
|   DEFINE(S_X3,			offsetof(struct pt_regs, regs[3]));
 | |
|   DEFINE(S_X4,			offsetof(struct pt_regs, regs[4]));
 | |
|   DEFINE(S_X5,			offsetof(struct pt_regs, regs[5]));
 | |
|   DEFINE(S_X6,			offsetof(struct pt_regs, regs[6]));
 | |
|   DEFINE(S_X7,			offsetof(struct pt_regs, regs[7]));
 | |
|   DEFINE(S_LR,			offsetof(struct pt_regs, regs[30]));
 | |
|   DEFINE(S_SP,			offsetof(struct pt_regs, sp));
 | |
| #ifdef CONFIG_COMPAT
 | |
|   DEFINE(S_COMPAT_SP,		offsetof(struct pt_regs, compat_sp));
 | |
| #endif
 | |
|   DEFINE(S_PSTATE,		offsetof(struct pt_regs, pstate));
 | |
|   DEFINE(S_PC,			offsetof(struct pt_regs, pc));
 | |
|   DEFINE(S_ORIG_X0,		offsetof(struct pt_regs, orig_x0));
 | |
|   DEFINE(S_SYSCALLNO,		offsetof(struct pt_regs, syscallno));
 | |
|   DEFINE(S_FRAME_SIZE,		sizeof(struct pt_regs));
 | |
|   BLANK();
 | |
|   DEFINE(MM_CONTEXT_ID,		offsetof(struct mm_struct, context.id));
 | |
|   BLANK();
 | |
|   DEFINE(VMA_VM_MM,		offsetof(struct vm_area_struct, vm_mm));
 | |
|   DEFINE(VMA_VM_FLAGS,		offsetof(struct vm_area_struct, vm_flags));
 | |
|   BLANK();
 | |
|   DEFINE(VM_EXEC,	       	VM_EXEC);
 | |
|   BLANK();
 | |
|   DEFINE(PAGE_SZ,	       	PAGE_SIZE);
 | |
|   BLANK();
 | |
|   DEFINE(CPU_INFO_SZ,		sizeof(struct cpu_info));
 | |
|   DEFINE(CPU_INFO_SETUP,	offsetof(struct cpu_info, cpu_setup));
 | |
|   BLANK();
 | |
|   DEFINE(DMA_BIDIRECTIONAL,	DMA_BIDIRECTIONAL);
 | |
|   DEFINE(DMA_TO_DEVICE,		DMA_TO_DEVICE);
 | |
|   DEFINE(DMA_FROM_DEVICE,	DMA_FROM_DEVICE);
 | |
|   BLANK();
 | |
|   DEFINE(CLOCK_REALTIME,	CLOCK_REALTIME);
 | |
|   DEFINE(CLOCK_MONOTONIC,	CLOCK_MONOTONIC);
 | |
|   DEFINE(CLOCK_REALTIME_RES,	MONOTONIC_RES_NSEC);
 | |
|   DEFINE(CLOCK_REALTIME_COARSE,	CLOCK_REALTIME_COARSE);
 | |
|   DEFINE(CLOCK_MONOTONIC_COARSE,CLOCK_MONOTONIC_COARSE);
 | |
|   DEFINE(CLOCK_COARSE_RES,	LOW_RES_NSEC);
 | |
|   DEFINE(NSEC_PER_SEC,		NSEC_PER_SEC);
 | |
|   BLANK();
 | |
|   DEFINE(VDSO_CS_CYCLE_LAST,	offsetof(struct vdso_data, cs_cycle_last));
 | |
|   DEFINE(VDSO_XTIME_CLK_SEC,	offsetof(struct vdso_data, xtime_clock_sec));
 | |
|   DEFINE(VDSO_XTIME_CLK_NSEC,	offsetof(struct vdso_data, xtime_clock_nsec));
 | |
|   DEFINE(VDSO_XTIME_CRS_SEC,	offsetof(struct vdso_data, xtime_coarse_sec));
 | |
|   DEFINE(VDSO_XTIME_CRS_NSEC,	offsetof(struct vdso_data, xtime_coarse_nsec));
 | |
|   DEFINE(VDSO_WTM_CLK_SEC,	offsetof(struct vdso_data, wtm_clock_sec));
 | |
|   DEFINE(VDSO_WTM_CLK_NSEC,	offsetof(struct vdso_data, wtm_clock_nsec));
 | |
|   DEFINE(VDSO_TB_SEQ_COUNT,	offsetof(struct vdso_data, tb_seq_count));
 | |
|   DEFINE(VDSO_CS_MULT,		offsetof(struct vdso_data, cs_mult));
 | |
|   DEFINE(VDSO_CS_SHIFT,		offsetof(struct vdso_data, cs_shift));
 | |
|   DEFINE(VDSO_TZ_MINWEST,	offsetof(struct vdso_data, tz_minuteswest));
 | |
|   DEFINE(VDSO_TZ_DSTTIME,	offsetof(struct vdso_data, tz_dsttime));
 | |
|   DEFINE(VDSO_USE_SYSCALL,	offsetof(struct vdso_data, use_syscall));
 | |
|   BLANK();
 | |
|   DEFINE(TVAL_TV_SEC,		offsetof(struct timeval, tv_sec));
 | |
|   DEFINE(TVAL_TV_USEC,		offsetof(struct timeval, tv_usec));
 | |
|   DEFINE(TSPEC_TV_SEC,		offsetof(struct timespec, tv_sec));
 | |
|   DEFINE(TSPEC_TV_NSEC,		offsetof(struct timespec, tv_nsec));
 | |
|   BLANK();
 | |
|   DEFINE(TZ_MINWEST,		offsetof(struct timezone, tz_minuteswest));
 | |
|   DEFINE(TZ_DSTTIME,		offsetof(struct timezone, tz_dsttime));
 | |
|   BLANK();
 | |
| #ifdef CONFIG_KVM_ARM_HOST
 | |
|   DEFINE(VCPU_CONTEXT,		offsetof(struct kvm_vcpu, arch.ctxt));
 | |
|   DEFINE(CPU_GP_REGS,		offsetof(struct kvm_cpu_context, gp_regs));
 | |
|   DEFINE(CPU_USER_PT_REGS,	offsetof(struct kvm_regs, regs));
 | |
|   DEFINE(CPU_FP_REGS,		offsetof(struct kvm_regs, fp_regs));
 | |
|   DEFINE(CPU_SP_EL1,		offsetof(struct kvm_regs, sp_el1));
 | |
|   DEFINE(CPU_ELR_EL1,		offsetof(struct kvm_regs, elr_el1));
 | |
|   DEFINE(CPU_SPSR,		offsetof(struct kvm_regs, spsr));
 | |
|   DEFINE(CPU_SYSREGS,		offsetof(struct kvm_cpu_context, sys_regs));
 | |
|   DEFINE(VCPU_ESR_EL2,		offsetof(struct kvm_vcpu, arch.fault.esr_el2));
 | |
|   DEFINE(VCPU_FAR_EL2,		offsetof(struct kvm_vcpu, arch.fault.far_el2));
 | |
|   DEFINE(VCPU_HPFAR_EL2,	offsetof(struct kvm_vcpu, arch.fault.hpfar_el2));
 | |
|   DEFINE(VCPU_DEBUG_FLAGS,	offsetof(struct kvm_vcpu, arch.debug_flags));
 | |
|   DEFINE(VCPU_HCR_EL2,		offsetof(struct kvm_vcpu, arch.hcr_el2));
 | |
|   DEFINE(VCPU_IRQ_LINES,	offsetof(struct kvm_vcpu, arch.irq_lines));
 | |
|   DEFINE(VCPU_HOST_CONTEXT,	offsetof(struct kvm_vcpu, arch.host_cpu_context));
 | |
|   DEFINE(VCPU_TIMER_CNTV_CTL,	offsetof(struct kvm_vcpu, arch.timer_cpu.cntv_ctl));
 | |
|   DEFINE(VCPU_TIMER_CNTV_CVAL,	offsetof(struct kvm_vcpu, arch.timer_cpu.cntv_cval));
 | |
|   DEFINE(KVM_TIMER_CNTVOFF,	offsetof(struct kvm, arch.timer.cntvoff));
 | |
|   DEFINE(KVM_TIMER_ENABLED,	offsetof(struct kvm, arch.timer.enabled));
 | |
|   DEFINE(VCPU_KVM,		offsetof(struct kvm_vcpu, kvm));
 | |
|   DEFINE(VCPU_VGIC_CPU,		offsetof(struct kvm_vcpu, arch.vgic_cpu));
 | |
|   DEFINE(VGIC_SAVE_FN,		offsetof(struct vgic_sr_vectors, save_vgic));
 | |
|   DEFINE(VGIC_RESTORE_FN,	offsetof(struct vgic_sr_vectors, restore_vgic));
 | |
|   DEFINE(VGIC_SR_VECTOR_SZ,	sizeof(struct vgic_sr_vectors));
 | |
|   DEFINE(VGIC_V2_CPU_HCR,	offsetof(struct vgic_cpu, vgic_v2.vgic_hcr));
 | |
|   DEFINE(VGIC_V2_CPU_VMCR,	offsetof(struct vgic_cpu, vgic_v2.vgic_vmcr));
 | |
|   DEFINE(VGIC_V2_CPU_MISR,	offsetof(struct vgic_cpu, vgic_v2.vgic_misr));
 | |
|   DEFINE(VGIC_V2_CPU_EISR,	offsetof(struct vgic_cpu, vgic_v2.vgic_eisr));
 | |
|   DEFINE(VGIC_V2_CPU_ELRSR,	offsetof(struct vgic_cpu, vgic_v2.vgic_elrsr));
 | |
|   DEFINE(VGIC_V2_CPU_APR,	offsetof(struct vgic_cpu, vgic_v2.vgic_apr));
 | |
|   DEFINE(VGIC_V2_CPU_LR,	offsetof(struct vgic_cpu, vgic_v2.vgic_lr));
 | |
|   DEFINE(VGIC_V3_CPU_SRE,	offsetof(struct vgic_cpu, vgic_v3.vgic_sre));
 | |
|   DEFINE(VGIC_V3_CPU_HCR,	offsetof(struct vgic_cpu, vgic_v3.vgic_hcr));
 | |
|   DEFINE(VGIC_V3_CPU_VMCR,	offsetof(struct vgic_cpu, vgic_v3.vgic_vmcr));
 | |
|   DEFINE(VGIC_V3_CPU_MISR,	offsetof(struct vgic_cpu, vgic_v3.vgic_misr));
 | |
|   DEFINE(VGIC_V3_CPU_EISR,	offsetof(struct vgic_cpu, vgic_v3.vgic_eisr));
 | |
|   DEFINE(VGIC_V3_CPU_ELRSR,	offsetof(struct vgic_cpu, vgic_v3.vgic_elrsr));
 | |
|   DEFINE(VGIC_V3_CPU_AP0R,	offsetof(struct vgic_cpu, vgic_v3.vgic_ap0r));
 | |
|   DEFINE(VGIC_V3_CPU_AP1R,	offsetof(struct vgic_cpu, vgic_v3.vgic_ap1r));
 | |
|   DEFINE(VGIC_V3_CPU_LR,	offsetof(struct vgic_cpu, vgic_v3.vgic_lr));
 | |
|   DEFINE(VGIC_CPU_NR_LR,	offsetof(struct vgic_cpu, nr_lr));
 | |
|   DEFINE(KVM_VTTBR,		offsetof(struct kvm, arch.vttbr));
 | |
|   DEFINE(KVM_VGIC_VCTRL,	offsetof(struct kvm, arch.vgic.vctrl_base));
 | |
| #endif
 | |
| #ifdef CONFIG_CPU_PM
 | |
|   DEFINE(CPU_SUSPEND_SZ,	sizeof(struct cpu_suspend_ctx));
 | |
|   DEFINE(CPU_CTX_SP,		offsetof(struct cpu_suspend_ctx, sp));
 | |
|   DEFINE(MPIDR_HASH_MASK,	offsetof(struct mpidr_hash, mask));
 | |
|   DEFINE(MPIDR_HASH_SHIFTS,	offsetof(struct mpidr_hash, shift_aff));
 | |
|   DEFINE(SLEEP_SAVE_SP_SZ,	sizeof(struct sleep_save_sp));
 | |
|   DEFINE(SLEEP_SAVE_SP_PHYS,	offsetof(struct sleep_save_sp, save_ptr_stash_phys));
 | |
|   DEFINE(SLEEP_SAVE_SP_VIRT,	offsetof(struct sleep_save_sp, save_ptr_stash));
 | |
| #endif
 | |
|   return 0;
 | |
| }
 |