]>
Commit | Line | Data |
---|---|---|
4f8d6632 MZ |
1 | /* |
2 | * Copyright (C) 2012,2013 - ARM Ltd | |
3 | * Author: Marc Zyngier <marc.zyngier@arm.com> | |
4 | * | |
5 | * Derived from arch/arm/include/asm/kvm_host.h: | |
6 | * Copyright (C) 2012 - Virtual Open Systems and Columbia University | |
7 | * Author: Christoffer Dall <c.dall@virtualopensystems.com> | |
8 | * | |
9 | * This program is free software; you can redistribute it and/or modify | |
10 | * it under the terms of the GNU General Public License version 2 as | |
11 | * published by the Free Software Foundation. | |
12 | * | |
13 | * This program is distributed in the hope that it will be useful, | |
14 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | |
15 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | |
16 | * GNU General Public License for more details. | |
17 | * | |
18 | * You should have received a copy of the GNU General Public License | |
19 | * along with this program. If not, see <http://www.gnu.org/licenses/>. | |
20 | */ | |
21 | ||
22 | #ifndef __ARM64_KVM_HOST_H__ | |
23 | #define __ARM64_KVM_HOST_H__ | |
24 | ||
25 | #include <asm/kvm.h> | |
26 | #include <asm/kvm_asm.h> | |
27 | #include <asm/kvm_mmio.h> | |
28 | ||
29 | #define KVM_MAX_VCPUS 4 | |
30 | #define KVM_USER_MEM_SLOTS 32 | |
31 | #define KVM_PRIVATE_MEM_SLOTS 4 | |
32 | #define KVM_COALESCED_MMIO_PAGE_OFFSET 1 | |
33 | ||
34 | #include <kvm/arm_vgic.h> | |
35 | #include <kvm/arm_arch_timer.h> | |
36 | ||
0d854a60 | 37 | #define KVM_VCPU_MAX_FEATURES 2 |
4f8d6632 | 38 | |
4f8d6632 MZ |
39 | struct kvm_vcpu; |
40 | int kvm_target_cpu(void); | |
41 | int kvm_reset_vcpu(struct kvm_vcpu *vcpu); | |
42 | int kvm_arch_dev_ioctl_check_extension(long ext); | |
43 | ||
44 | struct kvm_arch { | |
45 | /* The VMID generation used for the virt. memory system */ | |
46 | u64 vmid_gen; | |
47 | u32 vmid; | |
48 | ||
49 | /* 1-level 2nd stage table and lock */ | |
50 | spinlock_t pgd_lock; | |
51 | pgd_t *pgd; | |
52 | ||
53 | /* VTTBR value associated with above pgd and vmid */ | |
54 | u64 vttbr; | |
55 | ||
56 | /* Interrupt controller */ | |
57 | struct vgic_dist vgic; | |
58 | ||
59 | /* Timer */ | |
60 | struct arch_timer_kvm timer; | |
61 | }; | |
62 | ||
63 | #define KVM_NR_MEM_OBJS 40 | |
64 | ||
65 | /* | |
66 | * We don't want allocation failures within the mmu code, so we preallocate | |
67 | * enough memory for a single page fault in a cache. | |
68 | */ | |
69 | struct kvm_mmu_memory_cache { | |
70 | int nobjs; | |
71 | void *objects[KVM_NR_MEM_OBJS]; | |
72 | }; | |
73 | ||
74 | struct kvm_vcpu_fault_info { | |
75 | u32 esr_el2; /* Hyp Syndrom Register */ | |
76 | u64 far_el2; /* Hyp Fault Address Register */ | |
77 | u64 hpfar_el2; /* Hyp IPA Fault Address Register */ | |
78 | }; | |
79 | ||
80 | struct kvm_cpu_context { | |
81 | struct kvm_regs gp_regs; | |
40033a61 MZ |
82 | union { |
83 | u64 sys_regs[NR_SYS_REGS]; | |
84 | u32 cp15[NR_CP15_REGS]; | |
85 | }; | |
4f8d6632 MZ |
86 | }; |
87 | ||
88 | typedef struct kvm_cpu_context kvm_cpu_context_t; | |
89 | ||
90 | struct kvm_vcpu_arch { | |
91 | struct kvm_cpu_context ctxt; | |
92 | ||
93 | /* HYP configuration */ | |
94 | u64 hcr_el2; | |
95 | ||
96 | /* Exception Information */ | |
97 | struct kvm_vcpu_fault_info fault; | |
98 | ||
99 | /* Pointer to host CPU context */ | |
100 | kvm_cpu_context_t *host_cpu_context; | |
101 | ||
102 | /* VGIC state */ | |
103 | struct vgic_cpu vgic_cpu; | |
104 | struct arch_timer_cpu timer_cpu; | |
105 | ||
106 | /* | |
107 | * Anything that is not used directly from assembly code goes | |
108 | * here. | |
109 | */ | |
110 | /* dcache set/way operation pending */ | |
111 | int last_pcpu; | |
112 | cpumask_t require_dcache_flush; | |
113 | ||
114 | /* Don't run the guest */ | |
115 | bool pause; | |
116 | ||
117 | /* IO related fields */ | |
118 | struct kvm_decode mmio_decode; | |
119 | ||
120 | /* Interrupt related fields */ | |
121 | u64 irq_lines; /* IRQ and FIQ levels */ | |
122 | ||
123 | /* Cache some mmu pages needed inside spinlock regions */ | |
124 | struct kvm_mmu_memory_cache mmu_page_cache; | |
125 | ||
126 | /* Target CPU and feature flags */ | |
6c8c0c4d | 127 | int target; |
4f8d6632 MZ |
128 | DECLARE_BITMAP(features, KVM_VCPU_MAX_FEATURES); |
129 | ||
130 | /* Detect first run of a vcpu */ | |
131 | bool has_run_once; | |
132 | }; | |
133 | ||
134 | #define vcpu_gp_regs(v) (&(v)->arch.ctxt.gp_regs) | |
135 | #define vcpu_sys_reg(v,r) ((v)->arch.ctxt.sys_regs[(r)]) | |
136 | #define vcpu_cp15(v,r) ((v)->arch.ctxt.cp15[(r)]) | |
137 | ||
138 | struct kvm_vm_stat { | |
139 | u32 remote_tlb_flush; | |
140 | }; | |
141 | ||
142 | struct kvm_vcpu_stat { | |
143 | u32 halt_wakeup; | |
144 | }; | |
145 | ||
146 | struct kvm_vcpu_init; | |
147 | int kvm_vcpu_set_target(struct kvm_vcpu *vcpu, | |
148 | const struct kvm_vcpu_init *init); | |
473bdc0e | 149 | int kvm_vcpu_preferred_target(struct kvm_vcpu_init *init); |
4f8d6632 MZ |
150 | unsigned long kvm_arm_num_regs(struct kvm_vcpu *vcpu); |
151 | int kvm_arm_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *indices); | |
152 | struct kvm_one_reg; | |
153 | int kvm_arm_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg); | |
154 | int kvm_arm_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg); | |
155 | ||
156 | #define KVM_ARCH_WANT_MMU_NOTIFIER | |
157 | struct kvm; | |
158 | int kvm_unmap_hva(struct kvm *kvm, unsigned long hva); | |
159 | int kvm_unmap_hva_range(struct kvm *kvm, | |
160 | unsigned long start, unsigned long end); | |
161 | void kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte); | |
162 | ||
163 | /* We do not have shadow page tables, hence the empty hooks */ | |
164 | static inline int kvm_age_hva(struct kvm *kvm, unsigned long hva) | |
165 | { | |
166 | return 0; | |
167 | } | |
168 | ||
169 | static inline int kvm_test_age_hva(struct kvm *kvm, unsigned long hva) | |
170 | { | |
171 | return 0; | |
172 | } | |
173 | ||
174 | struct kvm_vcpu *kvm_arm_get_running_vcpu(void); | |
175 | struct kvm_vcpu __percpu **kvm_get_running_vcpus(void); | |
176 | ||
177 | u64 kvm_call_hyp(void *hypfn, ...); | |
178 | ||
179 | int handle_exit(struct kvm_vcpu *vcpu, struct kvm_run *run, | |
180 | int exception_index); | |
181 | ||
182 | int kvm_perf_init(void); | |
183 | int kvm_perf_teardown(void); | |
184 | ||
092bd143 MZ |
185 | static inline void __cpu_init_hyp_mode(phys_addr_t boot_pgd_ptr, |
186 | phys_addr_t pgd_ptr, | |
187 | unsigned long hyp_stack_ptr, | |
188 | unsigned long vector_ptr) | |
189 | { | |
190 | /* | |
191 | * Call initialization code, and switch to the full blown | |
192 | * HYP code. | |
193 | */ | |
194 | kvm_call_hyp((void *)boot_pgd_ptr, pgd_ptr, | |
195 | hyp_stack_ptr, vector_ptr); | |
196 | } | |
197 | ||
4f8d6632 | 198 | #endif /* __ARM64_KVM_HOST_H__ */ |