2 * Copyright (C) 2012 - Virtual Open Systems and Columbia University
3 * Author: Christoffer Dall <c.dall@virtualopensystems.com>
5 * This program is free software; you can redistribute it and/or modify
6 * it under the terms of the GNU General Public License, version 2, as
7 * published by the Free Software Foundation.
9 * This program is distributed in the hope that it will be useful,
10 * but WITHOUT ANY WARRANTY; without even the implied warranty of
11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
12 * GNU General Public License for more details.
14 * You should have received a copy of the GNU General Public License
15 * along with this program; if not, write to the Free Software
16 * Foundation, 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
19 #ifndef __ARM_KVM_HOST_H__
20 #define __ARM_KVM_HOST_H__
22 #include <linux/types.h>
23 #include <linux/kvm_types.h>
25 #include <asm/kvm_asm.h>
26 #include <asm/kvm_mmio.h>
27 #include <asm/fpstate.h>
28 #include <kvm/arm_arch_timer.h>
30 #define __KVM_HAVE_ARCH_INTC_INITIALIZED
32 #define KVM_USER_MEM_SLOTS 32
33 #define KVM_PRIVATE_MEM_SLOTS 4
34 #define KVM_COALESCED_MMIO_PAGE_OFFSET 1
35 #define KVM_HAVE_ONE_REG
37 #define KVM_VCPU_MAX_FEATURES 2
39 #include <kvm/arm_vgic.h>
41 #define KVM_MAX_VCPUS VGIC_V2_MAX_CPUS
43 u32
*kvm_vcpu_reg(struct kvm_vcpu
*vcpu
, u8 reg_num
, u32 mode
);
44 int __attribute_const__
kvm_target_cpu(void);
45 int kvm_reset_vcpu(struct kvm_vcpu
*vcpu
);
46 void kvm_reset_coprocs(struct kvm_vcpu
*vcpu
);
49 /* VTTBR value associated with below pgd and vmid */
53 struct arch_timer_kvm timer
;
56 * Anything that is not used directly from assembly code goes
60 /* The VMID generation used for the virt. memory system */
64 /* Stage-2 page table */
67 /* Interrupt controller */
68 struct vgic_dist vgic
;
72 #define KVM_NR_MEM_OBJS 40
75 * We don't want allocation failures within the mmu code, so we preallocate
76 * enough memory for a single page fault in a cache.
78 struct kvm_mmu_memory_cache
{
80 void *objects
[KVM_NR_MEM_OBJS
];
83 struct kvm_vcpu_fault_info
{
84 u32 hsr
; /* Hyp Syndrome Register */
85 u32 hxfar
; /* Hyp Data/Inst. Fault Address Register */
86 u32 hpfar
; /* Hyp IPA Fault Address Register */
87 u32 hyp_pc
; /* PC when exception was taken from Hyp mode */
90 typedef struct vfp_hard_struct kvm_cpu_context_t
;
92 struct kvm_vcpu_arch
{
95 int target
; /* Processor target */
96 DECLARE_BITMAP(features
, KVM_VCPU_MAX_FEATURES
);
98 /* System control coprocessor (cp15) */
99 u32 cp15
[NR_CP15_REGS
];
101 /* The CPU type we expose to the VM */
104 /* HYP trapping configuration */
107 /* Interrupt related fields */
108 u32 irq_lines
; /* IRQ and FIQ levels */
110 /* Exception Information */
111 struct kvm_vcpu_fault_info fault
;
113 /* Floating point registers (VFP and Advanced SIMD/NEON) */
114 struct vfp_hard_struct vfp_guest
;
116 /* Host FP context */
117 kvm_cpu_context_t
*host_cpu_context
;
120 struct vgic_cpu vgic_cpu
;
121 struct arch_timer_cpu timer_cpu
;
124 * Anything that is not used directly from assembly code goes
128 /* Don't run the guest on this vcpu */
131 /* IO related fields */
132 struct kvm_decode mmio_decode
;
134 /* Cache some mmu pages needed inside spinlock regions */
135 struct kvm_mmu_memory_cache mmu_page_cache
;
137 /* Detect first run of a vcpu */
142 u32 remote_tlb_flush
;
145 struct kvm_vcpu_stat
{
146 u32 halt_successful_poll
;
147 u32 halt_attempted_poll
;
151 int kvm_vcpu_preferred_target(struct kvm_vcpu_init
*init
);
152 unsigned long kvm_arm_num_regs(struct kvm_vcpu
*vcpu
);
153 int kvm_arm_copy_reg_indices(struct kvm_vcpu
*vcpu
, u64 __user
*indices
);
154 int kvm_arm_get_reg(struct kvm_vcpu
*vcpu
, const struct kvm_one_reg
*reg
);
155 int kvm_arm_set_reg(struct kvm_vcpu
*vcpu
, const struct kvm_one_reg
*reg
);
156 u64
kvm_call_hyp(void *hypfn
, ...);
157 void force_vm_exit(const cpumask_t
*mask
);
159 #define KVM_ARCH_WANT_MMU_NOTIFIER
160 int kvm_unmap_hva(struct kvm
*kvm
, unsigned long hva
);
161 int kvm_unmap_hva_range(struct kvm
*kvm
,
162 unsigned long start
, unsigned long end
);
163 void kvm_set_spte_hva(struct kvm
*kvm
, unsigned long hva
, pte_t pte
);
165 unsigned long kvm_arm_num_regs(struct kvm_vcpu
*vcpu
);
166 int kvm_arm_copy_reg_indices(struct kvm_vcpu
*vcpu
, u64 __user
*indices
);
167 int kvm_age_hva(struct kvm
*kvm
, unsigned long start
, unsigned long end
);
168 int kvm_test_age_hva(struct kvm
*kvm
, unsigned long hva
);
170 /* We do not have shadow page tables, hence the empty hooks */
171 static inline void kvm_arch_mmu_notifier_invalidate_page(struct kvm
*kvm
,
172 unsigned long address
)
176 struct kvm_vcpu
*kvm_arm_get_running_vcpu(void);
177 struct kvm_vcpu __percpu
**kvm_get_running_vcpus(void);
179 int kvm_arm_copy_coproc_indices(struct kvm_vcpu
*vcpu
, u64 __user
*uindices
);
180 unsigned long kvm_arm_num_coproc_regs(struct kvm_vcpu
*vcpu
);
181 int kvm_arm_coproc_get_reg(struct kvm_vcpu
*vcpu
, const struct kvm_one_reg
*);
182 int kvm_arm_coproc_set_reg(struct kvm_vcpu
*vcpu
, const struct kvm_one_reg
*);
184 int handle_exit(struct kvm_vcpu
*vcpu
, struct kvm_run
*run
,
185 int exception_index
);
187 static inline void __cpu_init_hyp_mode(phys_addr_t boot_pgd_ptr
,
189 unsigned long hyp_stack_ptr
,
190 unsigned long vector_ptr
)
193 * Call initialization code, and switch to the full blown HYP
194 * code. The init code doesn't need to preserve these
195 * registers as r0-r3 are already callee saved according to
197 * Note that we slightly misuse the prototype by casing the
198 * stack pointer to a void *.
200 * We don't have enough registers to perform the full init in
201 * one go. Install the boot PGD first, and then install the
202 * runtime PGD, stack pointer and vectors. The PGDs are always
203 * passed as the third argument, in order to be passed into
204 * r2-r3 to the init code (yes, this is compliant with the
208 kvm_call_hyp(NULL
, 0, boot_pgd_ptr
);
210 kvm_call_hyp((void*)hyp_stack_ptr
, vector_ptr
, pgd_ptr
);
213 static inline int kvm_arch_dev_ioctl_check_extension(long ext
)
218 int kvm_perf_init(void);
219 int kvm_perf_teardown(void);
221 void kvm_mmu_wp_memory_region(struct kvm
*kvm
, int slot
);
223 struct kvm_vcpu
*kvm_mpidr_to_vcpu(struct kvm
*kvm
, unsigned long mpidr
);
225 static inline void kvm_arch_hardware_disable(void) {}
226 static inline void kvm_arch_hardware_unsetup(void) {}
227 static inline void kvm_arch_sync_events(struct kvm
*kvm
) {}
228 static inline void kvm_arch_vcpu_uninit(struct kvm_vcpu
*vcpu
) {}
229 static inline void kvm_arch_sched_in(struct kvm_vcpu
*vcpu
, int cpu
) {}
231 static inline void kvm_arm_init_debug(void) {}
232 static inline void kvm_arm_setup_debug(struct kvm_vcpu
*vcpu
) {}
233 static inline void kvm_arm_clear_debug(struct kvm_vcpu
*vcpu
) {}
234 static inline void kvm_arm_reset_debug_ptr(struct kvm_vcpu
*vcpu
) {}
236 #endif /* __ARM_KVM_HOST_H__ */