]>
Commit | Line | Data |
---|---|---|
749cf76c CD |
1 | /* |
2 | * Copyright (C) 2012 - Virtual Open Systems and Columbia University | |
3 | * Author: Christoffer Dall <c.dall@virtualopensystems.com> | |
4 | * | |
5 | * This program is free software; you can redistribute it and/or modify | |
6 | * it under the terms of the GNU General Public License, version 2, as | |
7 | * published by the Free Software Foundation. | |
8 | * | |
9 | * This program is distributed in the hope that it will be useful, | |
10 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | |
11 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | |
12 | * GNU General Public License for more details. | |
13 | * | |
14 | * You should have received a copy of the GNU General Public License | |
15 | * along with this program; if not, write to the Free Software | |
16 | * Foundation, 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. | |
17 | */ | |
18 | ||
19 | #ifndef __ARM_KVM_HOST_H__ | |
20 | #define __ARM_KVM_HOST_H__ | |
21 | ||
65647300 PB |
22 | #include <linux/types.h> |
23 | #include <linux/kvm_types.h> | |
749cf76c CD |
24 | #include <asm/kvm.h> |
25 | #include <asm/kvm_asm.h> | |
45e96ea6 | 26 | #include <asm/kvm_mmio.h> |
f7ed45be | 27 | #include <asm/fpstate.h> |
7275acdf | 28 | #include <kvm/arm_arch_timer.h> |
749cf76c | 29 | |
c1426e4c EA |
30 | #define __KVM_HAVE_ARCH_INTC_INITIALIZED |
31 | ||
2b5e1e47 | 32 | #define KVM_USER_MEM_SLOTS 32 |
1138245c | 33 | #define KVM_HAVE_ONE_REG |
920552b2 | 34 | #define KVM_HALT_POLL_NS_DEFAULT 500000 |
749cf76c | 35 | |
7d0f84aa | 36 | #define KVM_VCPU_MAX_FEATURES 2 |
749cf76c | 37 | |
7275acdf | 38 | #include <kvm/arm_vgic.h> |
1a89dd91 | 39 | |
acda5430 VM |
40 | |
41 | #ifdef CONFIG_ARM_GIC_V3 | |
42 | #define KVM_MAX_VCPUS VGIC_V3_MAX_CPUS | |
43 | #else | |
ef748917 | 44 | #define KVM_MAX_VCPUS VGIC_V2_MAX_CPUS |
acda5430 | 45 | #endif |
ef748917 | 46 | |
930f7fd6 | 47 | #define KVM_REQ_VCPU_EXIT (8 | KVM_REQUEST_NO_WAKEUP) |
b13216cf | 48 | |
749cf76c | 49 | u32 *kvm_vcpu_reg(struct kvm_vcpu *vcpu, u8 reg_num, u32 mode); |
6951e48b | 50 | int __attribute_const__ kvm_target_cpu(void); |
749cf76c CD |
51 | int kvm_reset_vcpu(struct kvm_vcpu *vcpu); |
52 | void kvm_reset_coprocs(struct kvm_vcpu *vcpu); | |
53 | ||
54 | struct kvm_arch { | |
55 | /* VTTBR value associated with below pgd and vmid */ | |
56 | u64 vttbr; | |
57 | ||
94d0e598 MZ |
58 | /* The last vcpu id that ran on each physical CPU */ |
59 | int __percpu *last_vcpu_ran; | |
60 | ||
749cf76c CD |
61 | /* |
62 | * Anything that is not used directly from assembly code goes | |
63 | * here. | |
64 | */ | |
65 | ||
66 | /* The VMID generation used for the virt. memory system */ | |
67 | u64 vmid_gen; | |
68 | u32 vmid; | |
69 | ||
70 | /* Stage-2 page table */ | |
71 | pgd_t *pgd; | |
1a89dd91 MZ |
72 | |
73 | /* Interrupt controller */ | |
74 | struct vgic_dist vgic; | |
3caa2d8c | 75 | int max_vcpus; |
749cf76c CD |
76 | }; |
77 | ||
78 | #define KVM_NR_MEM_OBJS 40 | |
79 | ||
80 | /* | |
81 | * We don't want allocation failures within the mmu code, so we preallocate | |
82 | * enough memory for a single page fault in a cache. | |
83 | */ | |
84 | struct kvm_mmu_memory_cache { | |
85 | int nobjs; | |
86 | void *objects[KVM_NR_MEM_OBJS]; | |
87 | }; | |
88 | ||
7393b599 MZ |
89 | struct kvm_vcpu_fault_info { |
90 | u32 hsr; /* Hyp Syndrome Register */ | |
91 | u32 hxfar; /* Hyp Data/Inst. Fault Address Register */ | |
92 | u32 hpfar; /* Hyp IPA Fault Address Register */ | |
7393b599 MZ |
93 | }; |
94 | ||
4448932f MZ |
95 | /* |
96 | * 0 is reserved as an invalid value. | |
97 | * Order should be kept in sync with the save/restore code. | |
98 | */ | |
99 | enum vcpu_sysreg { | |
100 | __INVALID_SYSREG__, | |
101 | c0_MPIDR, /* MultiProcessor ID Register */ | |
102 | c0_CSSELR, /* Cache Size Selection Register */ | |
103 | c1_SCTLR, /* System Control Register */ | |
104 | c1_ACTLR, /* Auxiliary Control Register */ | |
105 | c1_CPACR, /* Coprocessor Access Control */ | |
106 | c2_TTBR0, /* Translation Table Base Register 0 */ | |
107 | c2_TTBR0_high, /* TTBR0 top 32 bits */ | |
108 | c2_TTBR1, /* Translation Table Base Register 1 */ | |
109 | c2_TTBR1_high, /* TTBR1 top 32 bits */ | |
110 | c2_TTBCR, /* Translation Table Base Control R. */ | |
111 | c3_DACR, /* Domain Access Control Register */ | |
112 | c5_DFSR, /* Data Fault Status Register */ | |
113 | c5_IFSR, /* Instruction Fault Status Register */ | |
114 | c5_ADFSR, /* Auxilary Data Fault Status R */ | |
115 | c5_AIFSR, /* Auxilary Instrunction Fault Status R */ | |
116 | c6_DFAR, /* Data Fault Address Register */ | |
117 | c6_IFAR, /* Instruction Fault Address Register */ | |
118 | c7_PAR, /* Physical Address Register */ | |
119 | c7_PAR_high, /* PAR top 32 bits */ | |
120 | c9_L2CTLR, /* Cortex A15/A7 L2 Control Register */ | |
121 | c10_PRRR, /* Primary Region Remap Register */ | |
122 | c10_NMRR, /* Normal Memory Remap Register */ | |
123 | c12_VBAR, /* Vector Base Address Register */ | |
124 | c13_CID, /* Context ID Register */ | |
125 | c13_TID_URW, /* Thread ID, User R/W */ | |
126 | c13_TID_URO, /* Thread ID, User R/O */ | |
127 | c13_TID_PRIV, /* Thread ID, Privileged */ | |
128 | c14_CNTKCTL, /* Timer Control Register (PL1) */ | |
129 | c10_AMAIR0, /* Auxilary Memory Attribute Indirection Reg0 */ | |
130 | c10_AMAIR1, /* Auxilary Memory Attribute Indirection Reg1 */ | |
131 | NR_CP15_REGS /* Number of regs (incl. invalid) */ | |
132 | }; | |
133 | ||
0ca5565d | 134 | struct kvm_cpu_context { |
c2a8dab5 | 135 | struct kvm_regs gp_regs; |
0ca5565d | 136 | struct vfp_hard_struct vfp; |
fb32a52a | 137 | u32 cp15[NR_CP15_REGS]; |
0ca5565d MZ |
138 | }; |
139 | ||
140 | typedef struct kvm_cpu_context kvm_cpu_context_t; | |
9c7a6432 | 141 | |
749cf76c | 142 | struct kvm_vcpu_arch { |
0ca5565d MZ |
143 | struct kvm_cpu_context ctxt; |
144 | ||
749cf76c CD |
145 | int target; /* Processor target */ |
146 | DECLARE_BITMAP(features, KVM_VCPU_MAX_FEATURES); | |
147 | ||
749cf76c CD |
148 | /* The CPU type we expose to the VM */ |
149 | u32 midr; | |
150 | ||
ac30a11e MZ |
151 | /* HYP trapping configuration */ |
152 | u32 hcr; | |
153 | ||
154 | /* Interrupt related fields */ | |
155 | u32 irq_lines; /* IRQ and FIQ levels */ | |
156 | ||
749cf76c | 157 | /* Exception Information */ |
7393b599 | 158 | struct kvm_vcpu_fault_info fault; |
749cf76c | 159 | |
3de50da6 MZ |
160 | /* Host FP context */ |
161 | kvm_cpu_context_t *host_cpu_context; | |
f7ed45be | 162 | |
1a89dd91 MZ |
163 | /* VGIC state */ |
164 | struct vgic_cpu vgic_cpu; | |
53e72406 | 165 | struct arch_timer_cpu timer_cpu; |
1a89dd91 | 166 | |
f7ed45be CD |
167 | /* |
168 | * Anything that is not used directly from assembly code goes | |
169 | * here. | |
170 | */ | |
5b3e5e5b | 171 | |
3781528e EA |
172 | /* vcpu power-off state */ |
173 | bool power_off; | |
aa024c2f | 174 | |
3b92830a EA |
175 | /* Don't run the guest (internal implementation need) */ |
176 | bool pause; | |
177 | ||
45e96ea6 CD |
178 | /* IO related fields */ |
179 | struct kvm_decode mmio_decode; | |
180 | ||
749cf76c CD |
181 | /* Cache some mmu pages needed inside spinlock regions */ |
182 | struct kvm_mmu_memory_cache mmu_page_cache; | |
f7ed45be CD |
183 | |
184 | /* Detect first run of a vcpu */ | |
185 | bool has_run_once; | |
749cf76c CD |
186 | }; |
187 | ||
188 | struct kvm_vm_stat { | |
8a7e75d4 | 189 | ulong remote_tlb_flush; |
749cf76c CD |
190 | }; |
191 | ||
192 | struct kvm_vcpu_stat { | |
8a7e75d4 SJS |
193 | u64 halt_successful_poll; |
194 | u64 halt_attempted_poll; | |
195 | u64 halt_poll_invalid; | |
196 | u64 halt_wakeup; | |
197 | u64 hvc_exit_stat; | |
b19e6892 AT |
198 | u64 wfe_exit_stat; |
199 | u64 wfi_exit_stat; | |
200 | u64 mmio_exit_user; | |
201 | u64 mmio_exit_kernel; | |
202 | u64 exits; | |
749cf76c CD |
203 | }; |
204 | ||
fb32a52a MZ |
205 | #define vcpu_cp15(v,r) (v)->arch.ctxt.cp15[r] |
206 | ||
4a6fee80 | 207 | int kvm_vcpu_preferred_target(struct kvm_vcpu_init *init); |
749cf76c CD |
208 | unsigned long kvm_arm_num_regs(struct kvm_vcpu *vcpu); |
209 | int kvm_arm_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *indices); | |
749cf76c CD |
210 | int kvm_arm_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg); |
211 | int kvm_arm_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg); | |
b57cd6f6 | 212 | unsigned long kvm_call_hyp(void *hypfn, ...); |
f7ed45be | 213 | void force_vm_exit(const cpumask_t *mask); |
d5d8184d CD |
214 | |
215 | #define KVM_ARCH_WANT_MMU_NOTIFIER | |
d5d8184d CD |
216 | int kvm_unmap_hva(struct kvm *kvm, unsigned long hva); |
217 | int kvm_unmap_hva_range(struct kvm *kvm, | |
218 | unsigned long start, unsigned long end); | |
219 | void kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte); | |
220 | ||
1138245c CD |
221 | unsigned long kvm_arm_num_regs(struct kvm_vcpu *vcpu); |
222 | int kvm_arm_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *indices); | |
35307b9a MZ |
223 | int kvm_age_hva(struct kvm *kvm, unsigned long start, unsigned long end); |
224 | int kvm_test_age_hva(struct kvm *kvm, unsigned long hva); | |
1138245c | 225 | |
d5d8184d | 226 | /* We do not have shadow page tables, hence the empty hooks */ |
fe71557a TC |
227 | static inline void kvm_arch_mmu_notifier_invalidate_page(struct kvm *kvm, |
228 | unsigned long address) | |
229 | { | |
230 | } | |
231 | ||
1638a12d MZ |
232 | struct kvm_vcpu *kvm_arm_get_running_vcpu(void); |
233 | struct kvm_vcpu __percpu **kvm_get_running_vcpus(void); | |
b13216cf CD |
234 | void kvm_arm_halt_guest(struct kvm *kvm); |
235 | void kvm_arm_resume_guest(struct kvm *kvm); | |
35a2d585 CD |
236 | void kvm_arm_halt_vcpu(struct kvm_vcpu *vcpu); |
237 | void kvm_arm_resume_vcpu(struct kvm_vcpu *vcpu); | |
1638a12d MZ |
238 | |
239 | int kvm_arm_copy_coproc_indices(struct kvm_vcpu *vcpu, u64 __user *uindices); | |
240 | unsigned long kvm_arm_num_coproc_regs(struct kvm_vcpu *vcpu); | |
1638a12d MZ |
241 | int kvm_arm_coproc_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *); |
242 | int kvm_arm_coproc_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *); | |
243 | ||
3414bbff MZ |
244 | int handle_exit(struct kvm_vcpu *vcpu, struct kvm_run *run, |
245 | int exception_index); | |
246 | ||
12fda812 | 247 | static inline void __cpu_init_hyp_mode(phys_addr_t pgd_ptr, |
e7858c58 MZ |
248 | unsigned long hyp_stack_ptr, |
249 | unsigned long vector_ptr) | |
250 | { | |
e7858c58 | 251 | /* |
5a677ce0 MZ |
252 | * Call initialization code, and switch to the full blown HYP |
253 | * code. The init code doesn't need to preserve these | |
254 | * registers as r0-r3 are already callee saved according to | |
255 | * the AAPCS. | |
cd602a37 | 256 | * Note that we slightly misuse the prototype by casting the |
5a677ce0 | 257 | * stack pointer to a void *. |
5a677ce0 | 258 | |
cd602a37 MZ |
259 | * The PGDs are always passed as the third argument, in order |
260 | * to be passed into r2-r3 to the init code (yes, this is | |
261 | * compliant with the PCS!). | |
262 | */ | |
5a677ce0 MZ |
263 | |
264 | kvm_call_hyp((void*)hyp_stack_ptr, vector_ptr, pgd_ptr); | |
e7858c58 MZ |
265 | } |
266 | ||
35a2491a MZ |
267 | static inline void __cpu_init_stage2(void) |
268 | { | |
d4c7688c | 269 | kvm_call_hyp(__init_stage2_translation); |
35a2491a MZ |
270 | } |
271 | ||
e537ecd7 MZ |
272 | static inline void __cpu_reset_hyp_mode(unsigned long vector_ptr, |
273 | phys_addr_t phys_idmap_start) | |
67f69197 | 274 | { |
e537ecd7 | 275 | kvm_call_hyp((void *)virt_to_idmap(__kvm_hyp_reset), vector_ptr); |
67f69197 AT |
276 | } |
277 | ||
b46f01ce | 278 | static inline int kvm_arch_dev_ioctl_check_extension(struct kvm *kvm, long ext) |
17b1e31f MZ |
279 | { |
280 | return 0; | |
281 | } | |
282 | ||
210552c1 MZ |
283 | int kvm_perf_init(void); |
284 | int kvm_perf_teardown(void); | |
285 | ||
c6473555 MS |
286 | void kvm_mmu_wp_memory_region(struct kvm *kvm, int slot); |
287 | ||
4429fc64 AP |
288 | struct kvm_vcpu *kvm_mpidr_to_vcpu(struct kvm *kvm, unsigned long mpidr); |
289 | ||
0865e636 RK |
290 | static inline void kvm_arch_hardware_unsetup(void) {} |
291 | static inline void kvm_arch_sync_events(struct kvm *kvm) {} | |
292 | static inline void kvm_arch_vcpu_uninit(struct kvm_vcpu *vcpu) {} | |
293 | static inline void kvm_arch_sched_in(struct kvm_vcpu *vcpu, int cpu) {} | |
3491caf2 | 294 | static inline void kvm_arch_vcpu_block_finish(struct kvm_vcpu *vcpu) {} |
0865e636 | 295 | |
56c7f5e7 AB |
296 | static inline void kvm_arm_init_debug(void) {} |
297 | static inline void kvm_arm_setup_debug(struct kvm_vcpu *vcpu) {} | |
298 | static inline void kvm_arm_clear_debug(struct kvm_vcpu *vcpu) {} | |
84e690bf | 299 | static inline void kvm_arm_reset_debug_ptr(struct kvm_vcpu *vcpu) {} |
bb0c70bc SZ |
300 | static inline int kvm_arm_vcpu_arch_set_attr(struct kvm_vcpu *vcpu, |
301 | struct kvm_device_attr *attr) | |
302 | { | |
303 | return -ENXIO; | |
304 | } | |
305 | static inline int kvm_arm_vcpu_arch_get_attr(struct kvm_vcpu *vcpu, | |
306 | struct kvm_device_attr *attr) | |
307 | { | |
308 | return -ENXIO; | |
309 | } | |
310 | static inline int kvm_arm_vcpu_arch_has_attr(struct kvm_vcpu *vcpu, | |
311 | struct kvm_device_attr *attr) | |
312 | { | |
313 | return -ENXIO; | |
314 | } | |
56c7f5e7 | 315 | |
749cf76c | 316 | #endif /* __ARM_KVM_HOST_H__ */ |