]>
Commit | Line | Data |
---|---|---|
749cf76c CD |
1 | /* |
2 | * Copyright (C) 2012 - Virtual Open Systems and Columbia University | |
3 | * Author: Christoffer Dall <c.dall@virtualopensystems.com> | |
4 | * | |
5 | * This program is free software; you can redistribute it and/or modify | |
6 | * it under the terms of the GNU General Public License, version 2, as | |
7 | * published by the Free Software Foundation. | |
8 | * | |
9 | * This program is distributed in the hope that it will be useful, | |
10 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | |
11 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | |
12 | * GNU General Public License for more details. | |
13 | * | |
14 | * You should have received a copy of the GNU General Public License | |
15 | * along with this program; if not, write to the Free Software | |
16 | * Foundation, 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. | |
17 | */ | |
18 | ||
19 | #ifndef __ARM_KVM_HOST_H__ | |
20 | #define __ARM_KVM_HOST_H__ | |
21 | ||
65647300 PB |
22 | #include <linux/types.h> |
23 | #include <linux/kvm_types.h> | |
749cf76c CD |
24 | #include <asm/kvm.h> |
25 | #include <asm/kvm_asm.h> | |
45e96ea6 | 26 | #include <asm/kvm_mmio.h> |
f7ed45be | 27 | #include <asm/fpstate.h> |
7275acdf | 28 | #include <kvm/arm_arch_timer.h> |
749cf76c | 29 | |
c1426e4c EA |
30 | #define __KVM_HAVE_ARCH_INTC_INITIALIZED |
31 | ||
2b5e1e47 | 32 | #define KVM_USER_MEM_SLOTS 32 |
749cf76c CD |
33 | #define KVM_PRIVATE_MEM_SLOTS 4 |
34 | #define KVM_COALESCED_MMIO_PAGE_OFFSET 1 | |
1138245c | 35 | #define KVM_HAVE_ONE_REG |
920552b2 | 36 | #define KVM_HALT_POLL_NS_DEFAULT 500000 |
749cf76c | 37 | |
7d0f84aa | 38 | #define KVM_VCPU_MAX_FEATURES 2 |
749cf76c | 39 | |
7275acdf | 40 | #include <kvm/arm_vgic.h> |
1a89dd91 | 41 | |
ef748917 ML |
42 | #define KVM_MAX_VCPUS VGIC_V2_MAX_CPUS |
43 | ||
b13216cf CD |
44 | #define KVM_REQ_VCPU_EXIT 8 |
45 | ||
749cf76c | 46 | u32 *kvm_vcpu_reg(struct kvm_vcpu *vcpu, u8 reg_num, u32 mode); |
6951e48b | 47 | int __attribute_const__ kvm_target_cpu(void); |
749cf76c CD |
48 | int kvm_reset_vcpu(struct kvm_vcpu *vcpu); |
49 | void kvm_reset_coprocs(struct kvm_vcpu *vcpu); | |
50 | ||
51 | struct kvm_arch { | |
52 | /* VTTBR value associated with below pgd and vmid */ | |
53 | u64 vttbr; | |
54 | ||
53e72406 MZ |
55 | /* Timer */ |
56 | struct arch_timer_kvm timer; | |
57 | ||
749cf76c CD |
58 | /* |
59 | * Anything that is not used directly from assembly code goes | |
60 | * here. | |
61 | */ | |
62 | ||
63 | /* The VMID generation used for the virt. memory system */ | |
64 | u64 vmid_gen; | |
65 | u32 vmid; | |
66 | ||
67 | /* Stage-2 page table */ | |
68 | pgd_t *pgd; | |
1a89dd91 MZ |
69 | |
70 | /* Interrupt controller */ | |
71 | struct vgic_dist vgic; | |
3caa2d8c | 72 | int max_vcpus; |
749cf76c CD |
73 | }; |
74 | ||
75 | #define KVM_NR_MEM_OBJS 40 | |
76 | ||
77 | /* | |
78 | * We don't want allocation failures within the mmu code, so we preallocate | |
79 | * enough memory for a single page fault in a cache. | |
80 | */ | |
81 | struct kvm_mmu_memory_cache { | |
82 | int nobjs; | |
83 | void *objects[KVM_NR_MEM_OBJS]; | |
84 | }; | |
85 | ||
7393b599 MZ |
86 | struct kvm_vcpu_fault_info { |
87 | u32 hsr; /* Hyp Syndrome Register */ | |
88 | u32 hxfar; /* Hyp Data/Inst. Fault Address Register */ | |
89 | u32 hpfar; /* Hyp IPA Fault Address Register */ | |
7393b599 MZ |
90 | }; |
91 | ||
4448932f MZ |
92 | /* |
93 | * 0 is reserved as an invalid value. | |
94 | * Order should be kept in sync with the save/restore code. | |
95 | */ | |
96 | enum vcpu_sysreg { | |
97 | __INVALID_SYSREG__, | |
98 | c0_MPIDR, /* MultiProcessor ID Register */ | |
99 | c0_CSSELR, /* Cache Size Selection Register */ | |
100 | c1_SCTLR, /* System Control Register */ | |
101 | c1_ACTLR, /* Auxiliary Control Register */ | |
102 | c1_CPACR, /* Coprocessor Access Control */ | |
103 | c2_TTBR0, /* Translation Table Base Register 0 */ | |
104 | c2_TTBR0_high, /* TTBR0 top 32 bits */ | |
105 | c2_TTBR1, /* Translation Table Base Register 1 */ | |
106 | c2_TTBR1_high, /* TTBR1 top 32 bits */ | |
107 | c2_TTBCR, /* Translation Table Base Control R. */ | |
108 | c3_DACR, /* Domain Access Control Register */ | |
109 | c5_DFSR, /* Data Fault Status Register */ | |
110 | c5_IFSR, /* Instruction Fault Status Register */ | |
111 | c5_ADFSR, /* Auxilary Data Fault Status R */ | |
112 | c5_AIFSR, /* Auxilary Instrunction Fault Status R */ | |
113 | c6_DFAR, /* Data Fault Address Register */ | |
114 | c6_IFAR, /* Instruction Fault Address Register */ | |
115 | c7_PAR, /* Physical Address Register */ | |
116 | c7_PAR_high, /* PAR top 32 bits */ | |
117 | c9_L2CTLR, /* Cortex A15/A7 L2 Control Register */ | |
118 | c10_PRRR, /* Primary Region Remap Register */ | |
119 | c10_NMRR, /* Normal Memory Remap Register */ | |
120 | c12_VBAR, /* Vector Base Address Register */ | |
121 | c13_CID, /* Context ID Register */ | |
122 | c13_TID_URW, /* Thread ID, User R/W */ | |
123 | c13_TID_URO, /* Thread ID, User R/O */ | |
124 | c13_TID_PRIV, /* Thread ID, Privileged */ | |
125 | c14_CNTKCTL, /* Timer Control Register (PL1) */ | |
126 | c10_AMAIR0, /* Auxilary Memory Attribute Indirection Reg0 */ | |
127 | c10_AMAIR1, /* Auxilary Memory Attribute Indirection Reg1 */ | |
128 | NR_CP15_REGS /* Number of regs (incl. invalid) */ | |
129 | }; | |
130 | ||
0ca5565d | 131 | struct kvm_cpu_context { |
c2a8dab5 | 132 | struct kvm_regs gp_regs; |
0ca5565d | 133 | struct vfp_hard_struct vfp; |
fb32a52a | 134 | u32 cp15[NR_CP15_REGS]; |
0ca5565d MZ |
135 | }; |
136 | ||
137 | typedef struct kvm_cpu_context kvm_cpu_context_t; | |
9c7a6432 | 138 | |
749cf76c | 139 | struct kvm_vcpu_arch { |
0ca5565d MZ |
140 | struct kvm_cpu_context ctxt; |
141 | ||
749cf76c CD |
142 | int target; /* Processor target */ |
143 | DECLARE_BITMAP(features, KVM_VCPU_MAX_FEATURES); | |
144 | ||
749cf76c CD |
145 | /* The CPU type we expose to the VM */ |
146 | u32 midr; | |
147 | ||
ac30a11e MZ |
148 | /* HYP trapping configuration */ |
149 | u32 hcr; | |
150 | ||
151 | /* Interrupt related fields */ | |
152 | u32 irq_lines; /* IRQ and FIQ levels */ | |
153 | ||
749cf76c | 154 | /* Exception Information */ |
7393b599 | 155 | struct kvm_vcpu_fault_info fault; |
749cf76c | 156 | |
3de50da6 MZ |
157 | /* Host FP context */ |
158 | kvm_cpu_context_t *host_cpu_context; | |
f7ed45be | 159 | |
1a89dd91 MZ |
160 | /* VGIC state */ |
161 | struct vgic_cpu vgic_cpu; | |
53e72406 | 162 | struct arch_timer_cpu timer_cpu; |
1a89dd91 | 163 | |
f7ed45be CD |
164 | /* |
165 | * Anything that is not used directly from assembly code goes | |
166 | * here. | |
167 | */ | |
5b3e5e5b | 168 | |
3781528e EA |
169 | /* vcpu power-off state */ |
170 | bool power_off; | |
aa024c2f | 171 | |
3b92830a EA |
172 | /* Don't run the guest (internal implementation need) */ |
173 | bool pause; | |
174 | ||
45e96ea6 CD |
175 | /* IO related fields */ |
176 | struct kvm_decode mmio_decode; | |
177 | ||
749cf76c CD |
178 | /* Cache some mmu pages needed inside spinlock regions */ |
179 | struct kvm_mmu_memory_cache mmu_page_cache; | |
f7ed45be CD |
180 | |
181 | /* Detect first run of a vcpu */ | |
182 | bool has_run_once; | |
749cf76c CD |
183 | }; |
184 | ||
185 | struct kvm_vm_stat { | |
8a7e75d4 | 186 | ulong remote_tlb_flush; |
749cf76c CD |
187 | }; |
188 | ||
189 | struct kvm_vcpu_stat { | |
8a7e75d4 SJS |
190 | u64 halt_successful_poll; |
191 | u64 halt_attempted_poll; | |
192 | u64 halt_poll_invalid; | |
193 | u64 halt_wakeup; | |
194 | u64 hvc_exit_stat; | |
b19e6892 AT |
195 | u64 wfe_exit_stat; |
196 | u64 wfi_exit_stat; | |
197 | u64 mmio_exit_user; | |
198 | u64 mmio_exit_kernel; | |
199 | u64 exits; | |
749cf76c CD |
200 | }; |
201 | ||
fb32a52a MZ |
202 | #define vcpu_cp15(v,r) (v)->arch.ctxt.cp15[r] |
203 | ||
4a6fee80 | 204 | int kvm_vcpu_preferred_target(struct kvm_vcpu_init *init); |
749cf76c CD |
205 | unsigned long kvm_arm_num_regs(struct kvm_vcpu *vcpu); |
206 | int kvm_arm_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *indices); | |
749cf76c CD |
207 | int kvm_arm_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg); |
208 | int kvm_arm_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg); | |
b57cd6f6 | 209 | unsigned long kvm_call_hyp(void *hypfn, ...); |
f7ed45be | 210 | void force_vm_exit(const cpumask_t *mask); |
d5d8184d CD |
211 | |
212 | #define KVM_ARCH_WANT_MMU_NOTIFIER | |
d5d8184d CD |
213 | int kvm_unmap_hva(struct kvm *kvm, unsigned long hva); |
214 | int kvm_unmap_hva_range(struct kvm *kvm, | |
215 | unsigned long start, unsigned long end); | |
216 | void kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte); | |
217 | ||
1138245c CD |
218 | unsigned long kvm_arm_num_regs(struct kvm_vcpu *vcpu); |
219 | int kvm_arm_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *indices); | |
35307b9a MZ |
220 | int kvm_age_hva(struct kvm *kvm, unsigned long start, unsigned long end); |
221 | int kvm_test_age_hva(struct kvm *kvm, unsigned long hva); | |
1138245c | 222 | |
d5d8184d | 223 | /* We do not have shadow page tables, hence the empty hooks */ |
fe71557a TC |
224 | static inline void kvm_arch_mmu_notifier_invalidate_page(struct kvm *kvm, |
225 | unsigned long address) | |
226 | { | |
227 | } | |
228 | ||
1638a12d MZ |
229 | struct kvm_vcpu *kvm_arm_get_running_vcpu(void); |
230 | struct kvm_vcpu __percpu **kvm_get_running_vcpus(void); | |
b13216cf CD |
231 | void kvm_arm_halt_guest(struct kvm *kvm); |
232 | void kvm_arm_resume_guest(struct kvm *kvm); | |
35a2d585 CD |
233 | void kvm_arm_halt_vcpu(struct kvm_vcpu *vcpu); |
234 | void kvm_arm_resume_vcpu(struct kvm_vcpu *vcpu); | |
1638a12d MZ |
235 | |
236 | int kvm_arm_copy_coproc_indices(struct kvm_vcpu *vcpu, u64 __user *uindices); | |
237 | unsigned long kvm_arm_num_coproc_regs(struct kvm_vcpu *vcpu); | |
1638a12d MZ |
238 | int kvm_arm_coproc_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *); |
239 | int kvm_arm_coproc_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *); | |
240 | ||
3414bbff MZ |
241 | int handle_exit(struct kvm_vcpu *vcpu, struct kvm_run *run, |
242 | int exception_index); | |
243 | ||
12fda812 | 244 | static inline void __cpu_init_hyp_mode(phys_addr_t pgd_ptr, |
e7858c58 MZ |
245 | unsigned long hyp_stack_ptr, |
246 | unsigned long vector_ptr) | |
247 | { | |
e7858c58 | 248 | /* |
5a677ce0 MZ |
249 | * Call initialization code, and switch to the full blown HYP |
250 | * code. The init code doesn't need to preserve these | |
251 | * registers as r0-r3 are already callee saved according to | |
252 | * the AAPCS. | |
cd602a37 | 253 | * Note that we slightly misuse the prototype by casting the |
5a677ce0 | 254 | * stack pointer to a void *. |
5a677ce0 | 255 | |
cd602a37 MZ |
256 | * The PGDs are always passed as the third argument, in order |
257 | * to be passed into r2-r3 to the init code (yes, this is | |
258 | * compliant with the PCS!). | |
259 | */ | |
5a677ce0 MZ |
260 | |
261 | kvm_call_hyp((void*)hyp_stack_ptr, vector_ptr, pgd_ptr); | |
e7858c58 MZ |
262 | } |
263 | ||
35a2491a MZ |
264 | static inline void __cpu_init_stage2(void) |
265 | { | |
d4c7688c | 266 | kvm_call_hyp(__init_stage2_translation); |
35a2491a MZ |
267 | } |
268 | ||
e537ecd7 MZ |
269 | static inline void __cpu_reset_hyp_mode(unsigned long vector_ptr, |
270 | phys_addr_t phys_idmap_start) | |
67f69197 | 271 | { |
e537ecd7 | 272 | kvm_call_hyp((void *)virt_to_idmap(__kvm_hyp_reset), vector_ptr); |
67f69197 AT |
273 | } |
274 | ||
b46f01ce | 275 | static inline int kvm_arch_dev_ioctl_check_extension(struct kvm *kvm, long ext) |
17b1e31f MZ |
276 | { |
277 | return 0; | |
278 | } | |
279 | ||
210552c1 MZ |
280 | int kvm_perf_init(void); |
281 | int kvm_perf_teardown(void); | |
282 | ||
c6473555 MS |
283 | void kvm_mmu_wp_memory_region(struct kvm *kvm, int slot); |
284 | ||
4429fc64 AP |
285 | struct kvm_vcpu *kvm_mpidr_to_vcpu(struct kvm *kvm, unsigned long mpidr); |
286 | ||
0865e636 RK |
287 | static inline void kvm_arch_hardware_unsetup(void) {} |
288 | static inline void kvm_arch_sync_events(struct kvm *kvm) {} | |
289 | static inline void kvm_arch_vcpu_uninit(struct kvm_vcpu *vcpu) {} | |
290 | static inline void kvm_arch_sched_in(struct kvm_vcpu *vcpu, int cpu) {} | |
3491caf2 | 291 | static inline void kvm_arch_vcpu_block_finish(struct kvm_vcpu *vcpu) {} |
0865e636 | 292 | |
56c7f5e7 AB |
293 | static inline void kvm_arm_init_debug(void) {} |
294 | static inline void kvm_arm_setup_debug(struct kvm_vcpu *vcpu) {} | |
295 | static inline void kvm_arm_clear_debug(struct kvm_vcpu *vcpu) {} | |
84e690bf | 296 | static inline void kvm_arm_reset_debug_ptr(struct kvm_vcpu *vcpu) {} |
bb0c70bc SZ |
297 | static inline int kvm_arm_vcpu_arch_set_attr(struct kvm_vcpu *vcpu, |
298 | struct kvm_device_attr *attr) | |
299 | { | |
300 | return -ENXIO; | |
301 | } | |
302 | static inline int kvm_arm_vcpu_arch_get_attr(struct kvm_vcpu *vcpu, | |
303 | struct kvm_device_attr *attr) | |
304 | { | |
305 | return -ENXIO; | |
306 | } | |
307 | static inline int kvm_arm_vcpu_arch_has_attr(struct kvm_vcpu *vcpu, | |
308 | struct kvm_device_attr *attr) | |
309 | { | |
310 | return -ENXIO; | |
311 | } | |
56c7f5e7 | 312 | |
749cf76c | 313 | #endif /* __ARM_KVM_HOST_H__ */ |