1 /* SPDX-License-Identifier: GPL-2.0-only */
3 * tools/testing/selftests/kvm/include/kvm_util.h
5 * Copyright (C) 2018, Google LLC.
7 #ifndef SELFTEST_KVM_UTIL_H
8 #define SELFTEST_KVM_UTIL_H
10 #include "test_util.h"
13 #include "linux/list.h"
14 #include "linux/kvm.h"
15 #include <sys/ioctl.h>
17 #include "sparsebit.h"
19 #define KVM_DEV_PATH "/dev/kvm"
20 #define KVM_MAX_VCPUS 512
23 * Callers of kvm_util only have an incomplete/opaque description of the
24 * structure kvm_util is using to maintain the state of a VM.
28 typedef uint64_t vm_paddr_t
; /* Virtual Machine (Guest) physical address */
29 typedef uint64_t vm_vaddr_t
; /* Virtual Machine (Guest) virtual address */
31 /* Minimum allocated guest virtual and physical addresses */
32 #define KVM_UTIL_MIN_VADDR 0x2000
33 #define KVM_GUEST_PAGE_TABLE_MIN_PADDR 0x180000
35 #define DEFAULT_GUEST_PHY_PAGES 512
36 #define DEFAULT_GUEST_STACK_VADDR_MIN 0xab6000
37 #define DEFAULT_STACK_PGS 5
46 VM_MODE_PXXV48_4K
, /* For 48bits VA but ANY bits PA */
52 #if defined(__aarch64__)
54 #define VM_MODE_DEFAULT VM_MODE_P40V48_4K
55 #define MIN_PAGE_SHIFT 12U
56 #define ptes_per_page(page_size) ((page_size) / 8)
58 #elif defined(__x86_64__)
60 #define VM_MODE_DEFAULT VM_MODE_PXXV48_4K
61 #define MIN_PAGE_SHIFT 12U
62 #define ptes_per_page(page_size) ((page_size) / 8)
64 #elif defined(__s390x__)
66 #define VM_MODE_DEFAULT VM_MODE_P44V64_4K
67 #define MIN_PAGE_SHIFT 12U
68 #define ptes_per_page(page_size) ((page_size) / 16)
72 #if defined(__x86_64__)
73 unsigned long vm_compute_max_gfn(struct kvm_vm
*vm
);
75 static inline unsigned long vm_compute_max_gfn(struct kvm_vm
*vm
)
77 return ((1ULL << vm
->pa_bits
) >> vm
->page_shift
) - 1;
81 #define MIN_PAGE_SIZE (1U << MIN_PAGE_SHIFT)
82 #define PTES_PER_MIN_PAGE ptes_per_page(MIN_PAGE_SIZE)
84 struct vm_guest_mode_params
{
87 unsigned int page_size
;
88 unsigned int page_shift
;
90 extern const struct vm_guest_mode_params vm_guest_mode_params
[];
92 int open_kvm_dev_path_or_exit(void);
93 int kvm_check_cap(long cap
);
94 int vm_enable_cap(struct kvm_vm
*vm
, struct kvm_enable_cap
*cap
);
95 int vcpu_enable_cap(struct kvm_vm
*vm
, uint32_t vcpu_id
,
96 struct kvm_enable_cap
*cap
);
97 void vm_enable_dirty_ring(struct kvm_vm
*vm
, uint32_t ring_size
);
98 const char *vm_guest_mode_string(uint32_t i
);
100 struct kvm_vm
*vm_create(enum vm_guest_mode mode
, uint64_t phy_pages
, int perm
);
101 void kvm_vm_free(struct kvm_vm
*vmp
);
102 void kvm_vm_restart(struct kvm_vm
*vmp
, int perm
);
103 void kvm_vm_release(struct kvm_vm
*vmp
);
104 void kvm_vm_get_dirty_log(struct kvm_vm
*vm
, int slot
, void *log
);
105 void kvm_vm_clear_dirty_log(struct kvm_vm
*vm
, int slot
, void *log
,
106 uint64_t first_page
, uint32_t num_pages
);
107 uint32_t kvm_vm_reset_dirty_ring(struct kvm_vm
*vm
);
109 int kvm_memcmp_hva_gva(void *hva
, struct kvm_vm
*vm
, const vm_vaddr_t gva
,
112 void kvm_vm_elf_load(struct kvm_vm
*vm
, const char *filename
);
114 void vm_dump(FILE *stream
, struct kvm_vm
*vm
, uint8_t indent
);
120 * stream - Output FILE stream
121 * vm - Virtual Machine
123 * indent - Left margin indent amount
129 * Dumps the current state of the VCPU specified by @vcpuid, within the VM
130 * given by @vm, to the FILE stream given by @stream.
132 void vcpu_dump(FILE *stream
, struct kvm_vm
*vm
, uint32_t vcpuid
,
135 void vm_create_irqchip(struct kvm_vm
*vm
);
137 void vm_userspace_mem_region_add(struct kvm_vm
*vm
,
138 enum vm_mem_backing_src_type src_type
,
139 uint64_t guest_paddr
, uint32_t slot
, uint64_t npages
,
142 void vcpu_ioctl(struct kvm_vm
*vm
, uint32_t vcpuid
, unsigned long ioctl
,
144 int _vcpu_ioctl(struct kvm_vm
*vm
, uint32_t vcpuid
, unsigned long ioctl
,
146 void vm_ioctl(struct kvm_vm
*vm
, unsigned long ioctl
, void *arg
);
147 int _vm_ioctl(struct kvm_vm
*vm
, unsigned long cmd
, void *arg
);
148 void kvm_ioctl(struct kvm_vm
*vm
, unsigned long ioctl
, void *arg
);
149 int _kvm_ioctl(struct kvm_vm
*vm
, unsigned long ioctl
, void *arg
);
150 void vm_mem_region_set_flags(struct kvm_vm
*vm
, uint32_t slot
, uint32_t flags
);
151 void vm_mem_region_move(struct kvm_vm
*vm
, uint32_t slot
, uint64_t new_gpa
);
152 void vm_mem_region_delete(struct kvm_vm
*vm
, uint32_t slot
);
153 void vm_vcpu_add(struct kvm_vm
*vm
, uint32_t vcpuid
);
154 vm_vaddr_t
vm_vaddr_alloc(struct kvm_vm
*vm
, size_t sz
, vm_vaddr_t vaddr_min
);
155 vm_vaddr_t
vm_vaddr_alloc_pages(struct kvm_vm
*vm
, int nr_pages
);
156 vm_vaddr_t
vm_vaddr_alloc_page(struct kvm_vm
*vm
);
158 void virt_map(struct kvm_vm
*vm
, uint64_t vaddr
, uint64_t paddr
,
159 unsigned int npages
);
160 void *addr_gpa2hva(struct kvm_vm
*vm
, vm_paddr_t gpa
);
161 void *addr_gva2hva(struct kvm_vm
*vm
, vm_vaddr_t gva
);
162 vm_paddr_t
addr_hva2gpa(struct kvm_vm
*vm
, void *hva
);
163 void *addr_gpa2alias(struct kvm_vm
*vm
, vm_paddr_t gpa
);
166 * Address Guest Virtual to Guest Physical
169 * vm - Virtual Machine
170 * gva - VM virtual address
175 * Equivalent VM physical address
177 * Returns the VM physical address of the translated VM virtual
178 * address given by @gva.
180 vm_paddr_t
addr_gva2gpa(struct kvm_vm
*vm
, vm_vaddr_t gva
);
182 struct kvm_run
*vcpu_state(struct kvm_vm
*vm
, uint32_t vcpuid
);
183 void vcpu_run(struct kvm_vm
*vm
, uint32_t vcpuid
);
184 int _vcpu_run(struct kvm_vm
*vm
, uint32_t vcpuid
);
185 int vcpu_get_fd(struct kvm_vm
*vm
, uint32_t vcpuid
);
186 void vcpu_run_complete_io(struct kvm_vm
*vm
, uint32_t vcpuid
);
187 void vcpu_set_guest_debug(struct kvm_vm
*vm
, uint32_t vcpuid
,
188 struct kvm_guest_debug
*debug
);
189 void vcpu_set_mp_state(struct kvm_vm
*vm
, uint32_t vcpuid
,
190 struct kvm_mp_state
*mp_state
);
191 struct kvm_reg_list
*vcpu_get_reg_list(struct kvm_vm
*vm
, uint32_t vcpuid
);
192 void vcpu_regs_get(struct kvm_vm
*vm
, uint32_t vcpuid
, struct kvm_regs
*regs
);
193 void vcpu_regs_set(struct kvm_vm
*vm
, uint32_t vcpuid
, struct kvm_regs
*regs
);
199 * vm - Virtual Machine
201 * num - number of arguments
202 * ... - arguments, each of type uint64_t
208 * Sets the first @num function input registers of the VCPU with @vcpuid,
209 * per the C calling convention of the architecture, to the values given
210 * as variable args. Each of the variable args is expected to be of type
211 * uint64_t. The maximum @num can be is specific to the architecture.
213 void vcpu_args_set(struct kvm_vm
*vm
, uint32_t vcpuid
, unsigned int num
, ...);
215 void vcpu_sregs_get(struct kvm_vm
*vm
, uint32_t vcpuid
,
216 struct kvm_sregs
*sregs
);
217 void vcpu_sregs_set(struct kvm_vm
*vm
, uint32_t vcpuid
,
218 struct kvm_sregs
*sregs
);
219 int _vcpu_sregs_set(struct kvm_vm
*vm
, uint32_t vcpuid
,
220 struct kvm_sregs
*sregs
);
221 void vcpu_fpu_get(struct kvm_vm
*vm
, uint32_t vcpuid
,
222 struct kvm_fpu
*fpu
);
223 void vcpu_fpu_set(struct kvm_vm
*vm
, uint32_t vcpuid
,
224 struct kvm_fpu
*fpu
);
225 void vcpu_get_reg(struct kvm_vm
*vm
, uint32_t vcpuid
, struct kvm_one_reg
*reg
);
226 void vcpu_set_reg(struct kvm_vm
*vm
, uint32_t vcpuid
, struct kvm_one_reg
*reg
);
227 #ifdef __KVM_HAVE_VCPU_EVENTS
228 void vcpu_events_get(struct kvm_vm
*vm
, uint32_t vcpuid
,
229 struct kvm_vcpu_events
*events
);
230 void vcpu_events_set(struct kvm_vm
*vm
, uint32_t vcpuid
,
231 struct kvm_vcpu_events
*events
);
234 void vcpu_nested_state_get(struct kvm_vm
*vm
, uint32_t vcpuid
,
235 struct kvm_nested_state
*state
);
236 int vcpu_nested_state_set(struct kvm_vm
*vm
, uint32_t vcpuid
,
237 struct kvm_nested_state
*state
, bool ignore_error
);
239 void *vcpu_map_dirty_ring(struct kvm_vm
*vm
, uint32_t vcpuid
);
241 int _kvm_device_check_attr(int dev_fd
, uint32_t group
, uint64_t attr
);
242 int kvm_device_check_attr(int dev_fd
, uint32_t group
, uint64_t attr
);
243 int _kvm_create_device(struct kvm_vm
*vm
, uint64_t type
, bool test
, int *fd
);
244 int kvm_create_device(struct kvm_vm
*vm
, uint64_t type
, bool test
);
245 int _kvm_device_access(int dev_fd
, uint32_t group
, uint64_t attr
,
246 void *val
, bool write
);
247 int kvm_device_access(int dev_fd
, uint32_t group
, uint64_t attr
,
248 void *val
, bool write
);
250 const char *exit_reason_str(unsigned int exit_reason
);
252 void virt_pgd_alloc(struct kvm_vm
*vm
);
255 * VM Virtual Page Map
258 * vm - Virtual Machine
259 * vaddr - VM Virtual Address
260 * paddr - VM Physical Address
261 * memslot - Memory region slot for new virtual translation tables
267 * Within @vm, creates a virtual translation for the page starting
268 * at @vaddr to the page starting at @paddr.
270 void virt_pg_map(struct kvm_vm
*vm
, uint64_t vaddr
, uint64_t paddr
);
272 vm_paddr_t
vm_phy_page_alloc(struct kvm_vm
*vm
, vm_paddr_t paddr_min
,
274 vm_paddr_t
vm_phy_pages_alloc(struct kvm_vm
*vm
, size_t num
,
275 vm_paddr_t paddr_min
, uint32_t memslot
);
276 vm_paddr_t
vm_alloc_page_table(struct kvm_vm
*vm
);
279 * Create a VM with reasonable defaults
282 * vcpuid - The id of the single VCPU to add to the VM.
283 * extra_mem_pages - The number of extra pages to add (this will
284 * decide how much extra space we will need to
285 * setup the page tables using memslot 0)
286 * guest_code - The vCPU's entry point
291 * Pointer to opaque structure that describes the created VM.
293 struct kvm_vm
*vm_create_default(uint32_t vcpuid
, uint64_t extra_mem_pages
,
296 /* Same as vm_create_default, but can be used for more than one vcpu */
297 struct kvm_vm
*vm_create_default_with_vcpus(uint32_t nr_vcpus
, uint64_t extra_mem_pages
,
298 uint32_t num_percpu_pages
, void *guest_code
,
301 /* Like vm_create_default_with_vcpus, but accepts mode and slot0 memory as a parameter */
302 struct kvm_vm
*vm_create_with_vcpus(enum vm_guest_mode mode
, uint32_t nr_vcpus
,
303 uint64_t slot0_mem_pages
, uint64_t extra_mem_pages
,
304 uint32_t num_percpu_pages
, void *guest_code
,
308 * Adds a vCPU with reasonable defaults (e.g. a stack)
311 * vm - Virtual Machine
312 * vcpuid - The id of the VCPU to add to the VM.
313 * guest_code - The vCPU's entry point
315 void vm_vcpu_add_default(struct kvm_vm
*vm
, uint32_t vcpuid
, void *guest_code
);
317 bool vm_is_unrestricted_guest(struct kvm_vm
*vm
);
319 unsigned int vm_get_page_size(struct kvm_vm
*vm
);
320 unsigned int vm_get_page_shift(struct kvm_vm
*vm
);
321 uint64_t vm_get_max_gfn(struct kvm_vm
*vm
);
322 int vm_get_fd(struct kvm_vm
*vm
);
324 unsigned int vm_calc_num_guest_pages(enum vm_guest_mode mode
, size_t size
);
325 unsigned int vm_num_host_pages(enum vm_guest_mode mode
, unsigned int num_guest_pages
);
326 unsigned int vm_num_guest_pages(enum vm_guest_mode mode
, unsigned int num_host_pages
);
327 static inline unsigned int
328 vm_adjust_num_guest_pages(enum vm_guest_mode mode
, unsigned int num_guest_pages
)
331 n
= vm_num_guest_pages(mode
, vm_num_host_pages(mode
, num_guest_pages
));
333 /* s390 requires 1M aligned guest sizes */
334 n
= (n
+ 255) & ~255;
339 struct kvm_userspace_memory_region
*
340 kvm_userspace_memory_region_find(struct kvm_vm
*vm
, uint64_t start
,
343 struct kvm_dirty_log
*
344 allocate_kvm_dirty_log(struct kvm_userspace_memory_region
*region
);
346 int vm_create_device(struct kvm_vm
*vm
, struct kvm_create_device
*cd
);
348 #define sync_global_to_guest(vm, g) ({ \
349 typeof(g) *_p = addr_gva2hva(vm, (vm_vaddr_t)&(g)); \
350 memcpy(_p, &(g), sizeof(g)); \
353 #define sync_global_from_guest(vm, g) ({ \
354 typeof(g) *_p = addr_gva2hva(vm, (vm_vaddr_t)&(g)); \
355 memcpy(&(g), _p, sizeof(g)); \
358 void assert_on_unhandled_exception(struct kvm_vm
*vm
, uint32_t vcpuid
);
369 #define UCALL_MAX_ARGS 6
373 uint64_t args
[UCALL_MAX_ARGS
];
376 void ucall_init(struct kvm_vm
*vm
, void *arg
);
377 void ucall_uninit(struct kvm_vm
*vm
);
378 void ucall(uint64_t cmd
, int nargs
, ...);
379 uint64_t get_ucall(struct kvm_vm
*vm
, uint32_t vcpu_id
, struct ucall
*uc
);
381 #define GUEST_SYNC_ARGS(stage, arg1, arg2, arg3, arg4) \
382 ucall(UCALL_SYNC, 6, "hello", stage, arg1, arg2, arg3, arg4)
383 #define GUEST_SYNC(stage) ucall(UCALL_SYNC, 2, "hello", stage)
384 #define GUEST_DONE() ucall(UCALL_DONE, 0)
385 #define __GUEST_ASSERT(_condition, _condstr, _nargs, _args...) do { \
387 ucall(UCALL_ABORT, 2 + _nargs, \
388 "Failed guest assert: " \
389 _condstr, __LINE__, _args); \
392 #define GUEST_ASSERT(_condition) \
393 __GUEST_ASSERT(_condition, #_condition, 0, 0)
395 #define GUEST_ASSERT_1(_condition, arg1) \
396 __GUEST_ASSERT(_condition, #_condition, 1, (arg1))
398 #define GUEST_ASSERT_2(_condition, arg1, arg2) \
399 __GUEST_ASSERT(_condition, #_condition, 2, (arg1), (arg2))
401 #define GUEST_ASSERT_3(_condition, arg1, arg2, arg3) \
402 __GUEST_ASSERT(_condition, #_condition, 3, (arg1), (arg2), (arg3))
404 #define GUEST_ASSERT_4(_condition, arg1, arg2, arg3, arg4) \
405 __GUEST_ASSERT(_condition, #_condition, 4, (arg1), (arg2), (arg3), (arg4))
407 #define GUEST_ASSERT_EQ(a, b) __GUEST_ASSERT((a) == (b), #a " == " #b, 2, a, b)
409 int vm_get_stats_fd(struct kvm_vm
*vm
);
410 int vcpu_get_stats_fd(struct kvm_vm
*vm
, uint32_t vcpuid
);
412 #endif /* SELFTEST_KVM_UTIL_H */