1 /* SPDX-License-Identifier: GPL-2.0-only */
3 * tools/testing/selftests/kvm/include/kvm_util.h
5 * Copyright (C) 2018, Google LLC.
7 #ifndef SELFTEST_KVM_UTIL_H
8 #define SELFTEST_KVM_UTIL_H
10 #include "test_util.h"
13 #include "linux/list.h"
14 #include "linux/kvm.h"
15 #include <sys/ioctl.h>
17 #include "sparsebit.h"
19 #define KVM_DEV_PATH "/dev/kvm"
20 #define KVM_MAX_VCPUS 512
23 * Callers of kvm_util only have an incomplete/opaque description of the
24 * structure kvm_util is using to maintain the state of a VM.
28 typedef uint64_t vm_paddr_t
; /* Virtual Machine (Guest) physical address */
29 typedef uint64_t vm_vaddr_t
; /* Virtual Machine (Guest) virtual address */
31 /* Minimum allocated guest virtual and physical addresses */
32 #define KVM_UTIL_MIN_VADDR 0x2000
33 #define KVM_GUEST_PAGE_TABLE_MIN_PADDR 0x180000
35 #define DEFAULT_GUEST_PHY_PAGES 512
36 #define DEFAULT_GUEST_STACK_VADDR_MIN 0xab6000
37 #define DEFAULT_STACK_PGS 5
46 VM_MODE_PXXV48_4K
, /* For 48bits VA but ANY bits PA */
51 #if defined(__aarch64__)
53 #define VM_MODE_DEFAULT VM_MODE_P40V48_4K
54 #define MIN_PAGE_SHIFT 12U
55 #define ptes_per_page(page_size) ((page_size) / 8)
57 #elif defined(__x86_64__)
59 #define VM_MODE_DEFAULT VM_MODE_PXXV48_4K
60 #define MIN_PAGE_SHIFT 12U
61 #define ptes_per_page(page_size) ((page_size) / 8)
63 #elif defined(__s390x__)
65 #define VM_MODE_DEFAULT VM_MODE_P47V64_4K
66 #define MIN_PAGE_SHIFT 12U
67 #define ptes_per_page(page_size) ((page_size) / 16)
71 #define MIN_PAGE_SIZE (1U << MIN_PAGE_SHIFT)
72 #define PTES_PER_MIN_PAGE ptes_per_page(MIN_PAGE_SIZE)
74 struct vm_guest_mode_params
{
77 unsigned int page_size
;
78 unsigned int page_shift
;
80 extern const struct vm_guest_mode_params vm_guest_mode_params
[];
82 int open_kvm_dev_path_or_exit(void);
83 int kvm_check_cap(long cap
);
84 int vm_enable_cap(struct kvm_vm
*vm
, struct kvm_enable_cap
*cap
);
85 int vcpu_enable_cap(struct kvm_vm
*vm
, uint32_t vcpu_id
,
86 struct kvm_enable_cap
*cap
);
87 void vm_enable_dirty_ring(struct kvm_vm
*vm
, uint32_t ring_size
);
88 const char *vm_guest_mode_string(uint32_t i
);
90 struct kvm_vm
*vm_create(enum vm_guest_mode mode
, uint64_t phy_pages
, int perm
);
91 void kvm_vm_free(struct kvm_vm
*vmp
);
92 void kvm_vm_restart(struct kvm_vm
*vmp
, int perm
);
93 void kvm_vm_release(struct kvm_vm
*vmp
);
94 void kvm_vm_get_dirty_log(struct kvm_vm
*vm
, int slot
, void *log
);
95 void kvm_vm_clear_dirty_log(struct kvm_vm
*vm
, int slot
, void *log
,
96 uint64_t first_page
, uint32_t num_pages
);
97 uint32_t kvm_vm_reset_dirty_ring(struct kvm_vm
*vm
);
99 int kvm_memcmp_hva_gva(void *hva
, struct kvm_vm
*vm
, const vm_vaddr_t gva
,
102 void kvm_vm_elf_load(struct kvm_vm
*vm
, const char *filename
);
104 void vm_dump(FILE *stream
, struct kvm_vm
*vm
, uint8_t indent
);
110 * stream - Output FILE stream
111 * vm - Virtual Machine
113 * indent - Left margin indent amount
119 * Dumps the current state of the VCPU specified by @vcpuid, within the VM
120 * given by @vm, to the FILE stream given by @stream.
122 void vcpu_dump(FILE *stream
, struct kvm_vm
*vm
, uint32_t vcpuid
,
125 void vm_create_irqchip(struct kvm_vm
*vm
);
127 void vm_userspace_mem_region_add(struct kvm_vm
*vm
,
128 enum vm_mem_backing_src_type src_type
,
129 uint64_t guest_paddr
, uint32_t slot
, uint64_t npages
,
132 void vcpu_ioctl(struct kvm_vm
*vm
, uint32_t vcpuid
, unsigned long ioctl
,
134 int _vcpu_ioctl(struct kvm_vm
*vm
, uint32_t vcpuid
, unsigned long ioctl
,
136 void vm_ioctl(struct kvm_vm
*vm
, unsigned long ioctl
, void *arg
);
137 int _vm_ioctl(struct kvm_vm
*vm
, unsigned long cmd
, void *arg
);
138 void kvm_ioctl(struct kvm_vm
*vm
, unsigned long ioctl
, void *arg
);
139 int _kvm_ioctl(struct kvm_vm
*vm
, unsigned long ioctl
, void *arg
);
140 void vm_mem_region_set_flags(struct kvm_vm
*vm
, uint32_t slot
, uint32_t flags
);
141 void vm_mem_region_move(struct kvm_vm
*vm
, uint32_t slot
, uint64_t new_gpa
);
142 void vm_mem_region_delete(struct kvm_vm
*vm
, uint32_t slot
);
143 void vm_vcpu_add(struct kvm_vm
*vm
, uint32_t vcpuid
);
144 vm_vaddr_t
vm_vaddr_alloc(struct kvm_vm
*vm
, size_t sz
, vm_vaddr_t vaddr_min
);
145 vm_vaddr_t
vm_vaddr_alloc_pages(struct kvm_vm
*vm
, int nr_pages
);
146 vm_vaddr_t
vm_vaddr_alloc_page(struct kvm_vm
*vm
);
148 void virt_map(struct kvm_vm
*vm
, uint64_t vaddr
, uint64_t paddr
,
149 unsigned int npages
);
150 void *addr_gpa2hva(struct kvm_vm
*vm
, vm_paddr_t gpa
);
151 void *addr_gva2hva(struct kvm_vm
*vm
, vm_vaddr_t gva
);
152 vm_paddr_t
addr_hva2gpa(struct kvm_vm
*vm
, void *hva
);
153 void *addr_gpa2alias(struct kvm_vm
*vm
, vm_paddr_t gpa
);
156 * Address Guest Virtual to Guest Physical
159 * vm - Virtual Machine
160 * gva - VM virtual address
165 * Equivalent VM physical address
167 * Returns the VM physical address of the translated VM virtual
168 * address given by @gva.
170 vm_paddr_t
addr_gva2gpa(struct kvm_vm
*vm
, vm_vaddr_t gva
);
172 struct kvm_run
*vcpu_state(struct kvm_vm
*vm
, uint32_t vcpuid
);
173 void vcpu_run(struct kvm_vm
*vm
, uint32_t vcpuid
);
174 int _vcpu_run(struct kvm_vm
*vm
, uint32_t vcpuid
);
175 int vcpu_get_fd(struct kvm_vm
*vm
, uint32_t vcpuid
);
176 void vcpu_run_complete_io(struct kvm_vm
*vm
, uint32_t vcpuid
);
177 void vcpu_set_guest_debug(struct kvm_vm
*vm
, uint32_t vcpuid
,
178 struct kvm_guest_debug
*debug
);
179 void vcpu_set_mp_state(struct kvm_vm
*vm
, uint32_t vcpuid
,
180 struct kvm_mp_state
*mp_state
);
181 struct kvm_reg_list
*vcpu_get_reg_list(struct kvm_vm
*vm
, uint32_t vcpuid
);
182 void vcpu_regs_get(struct kvm_vm
*vm
, uint32_t vcpuid
, struct kvm_regs
*regs
);
183 void vcpu_regs_set(struct kvm_vm
*vm
, uint32_t vcpuid
, struct kvm_regs
*regs
);
189 * vm - Virtual Machine
191 * num - number of arguments
192 * ... - arguments, each of type uint64_t
198 * Sets the first @num function input registers of the VCPU with @vcpuid,
199 * per the C calling convention of the architecture, to the values given
200 * as variable args. Each of the variable args is expected to be of type
201 * uint64_t. The maximum @num can be is specific to the architecture.
203 void vcpu_args_set(struct kvm_vm
*vm
, uint32_t vcpuid
, unsigned int num
, ...);
205 void vcpu_sregs_get(struct kvm_vm
*vm
, uint32_t vcpuid
,
206 struct kvm_sregs
*sregs
);
207 void vcpu_sregs_set(struct kvm_vm
*vm
, uint32_t vcpuid
,
208 struct kvm_sregs
*sregs
);
209 int _vcpu_sregs_set(struct kvm_vm
*vm
, uint32_t vcpuid
,
210 struct kvm_sregs
*sregs
);
211 void vcpu_fpu_get(struct kvm_vm
*vm
, uint32_t vcpuid
,
212 struct kvm_fpu
*fpu
);
213 void vcpu_fpu_set(struct kvm_vm
*vm
, uint32_t vcpuid
,
214 struct kvm_fpu
*fpu
);
215 void vcpu_get_reg(struct kvm_vm
*vm
, uint32_t vcpuid
, struct kvm_one_reg
*reg
);
216 void vcpu_set_reg(struct kvm_vm
*vm
, uint32_t vcpuid
, struct kvm_one_reg
*reg
);
217 #ifdef __KVM_HAVE_VCPU_EVENTS
218 void vcpu_events_get(struct kvm_vm
*vm
, uint32_t vcpuid
,
219 struct kvm_vcpu_events
*events
);
220 void vcpu_events_set(struct kvm_vm
*vm
, uint32_t vcpuid
,
221 struct kvm_vcpu_events
*events
);
224 void vcpu_nested_state_get(struct kvm_vm
*vm
, uint32_t vcpuid
,
225 struct kvm_nested_state
*state
);
226 int vcpu_nested_state_set(struct kvm_vm
*vm
, uint32_t vcpuid
,
227 struct kvm_nested_state
*state
, bool ignore_error
);
229 void *vcpu_map_dirty_ring(struct kvm_vm
*vm
, uint32_t vcpuid
);
231 int _kvm_device_check_attr(int dev_fd
, uint32_t group
, uint64_t attr
);
232 int kvm_device_check_attr(int dev_fd
, uint32_t group
, uint64_t attr
);
233 int _kvm_create_device(struct kvm_vm
*vm
, uint64_t type
, bool test
, int *fd
);
234 int kvm_create_device(struct kvm_vm
*vm
, uint64_t type
, bool test
);
235 int _kvm_device_access(int dev_fd
, uint32_t group
, uint64_t attr
,
236 void *val
, bool write
);
237 int kvm_device_access(int dev_fd
, uint32_t group
, uint64_t attr
,
238 void *val
, bool write
);
240 const char *exit_reason_str(unsigned int exit_reason
);
242 void virt_pgd_alloc(struct kvm_vm
*vm
);
245 * VM Virtual Page Map
248 * vm - Virtual Machine
249 * vaddr - VM Virtual Address
250 * paddr - VM Physical Address
251 * memslot - Memory region slot for new virtual translation tables
257 * Within @vm, creates a virtual translation for the page starting
258 * at @vaddr to the page starting at @paddr.
260 void virt_pg_map(struct kvm_vm
*vm
, uint64_t vaddr
, uint64_t paddr
);
262 vm_paddr_t
vm_phy_page_alloc(struct kvm_vm
*vm
, vm_paddr_t paddr_min
,
264 vm_paddr_t
vm_phy_pages_alloc(struct kvm_vm
*vm
, size_t num
,
265 vm_paddr_t paddr_min
, uint32_t memslot
);
266 vm_paddr_t
vm_alloc_page_table(struct kvm_vm
*vm
);
269 * Create a VM with reasonable defaults
272 * vcpuid - The id of the single VCPU to add to the VM.
273 * extra_mem_pages - The number of extra pages to add (this will
274 * decide how much extra space we will need to
275 * setup the page tables using memslot 0)
276 * guest_code - The vCPU's entry point
281 * Pointer to opaque structure that describes the created VM.
283 struct kvm_vm
*vm_create_default(uint32_t vcpuid
, uint64_t extra_mem_pages
,
286 /* Same as vm_create_default, but can be used for more than one vcpu */
287 struct kvm_vm
*vm_create_default_with_vcpus(uint32_t nr_vcpus
, uint64_t extra_mem_pages
,
288 uint32_t num_percpu_pages
, void *guest_code
,
291 /* Like vm_create_default_with_vcpus, but accepts mode and slot0 memory as a parameter */
292 struct kvm_vm
*vm_create_with_vcpus(enum vm_guest_mode mode
, uint32_t nr_vcpus
,
293 uint64_t slot0_mem_pages
, uint64_t extra_mem_pages
,
294 uint32_t num_percpu_pages
, void *guest_code
,
298 * Adds a vCPU with reasonable defaults (e.g. a stack)
301 * vm - Virtual Machine
302 * vcpuid - The id of the VCPU to add to the VM.
303 * guest_code - The vCPU's entry point
305 void vm_vcpu_add_default(struct kvm_vm
*vm
, uint32_t vcpuid
, void *guest_code
);
307 bool vm_is_unrestricted_guest(struct kvm_vm
*vm
);
309 unsigned int vm_get_page_size(struct kvm_vm
*vm
);
310 unsigned int vm_get_page_shift(struct kvm_vm
*vm
);
311 uint64_t vm_get_max_gfn(struct kvm_vm
*vm
);
312 int vm_get_fd(struct kvm_vm
*vm
);
314 unsigned int vm_calc_num_guest_pages(enum vm_guest_mode mode
, size_t size
);
315 unsigned int vm_num_host_pages(enum vm_guest_mode mode
, unsigned int num_guest_pages
);
316 unsigned int vm_num_guest_pages(enum vm_guest_mode mode
, unsigned int num_host_pages
);
317 static inline unsigned int
318 vm_adjust_num_guest_pages(enum vm_guest_mode mode
, unsigned int num_guest_pages
)
321 n
= vm_num_guest_pages(mode
, vm_num_host_pages(mode
, num_guest_pages
));
323 /* s390 requires 1M aligned guest sizes */
324 n
= (n
+ 255) & ~255;
329 struct kvm_userspace_memory_region
*
330 kvm_userspace_memory_region_find(struct kvm_vm
*vm
, uint64_t start
,
333 struct kvm_dirty_log
*
334 allocate_kvm_dirty_log(struct kvm_userspace_memory_region
*region
);
336 int vm_create_device(struct kvm_vm
*vm
, struct kvm_create_device
*cd
);
338 #define sync_global_to_guest(vm, g) ({ \
339 typeof(g) *_p = addr_gva2hva(vm, (vm_vaddr_t)&(g)); \
340 memcpy(_p, &(g), sizeof(g)); \
343 #define sync_global_from_guest(vm, g) ({ \
344 typeof(g) *_p = addr_gva2hva(vm, (vm_vaddr_t)&(g)); \
345 memcpy(&(g), _p, sizeof(g)); \
348 void assert_on_unhandled_exception(struct kvm_vm
*vm
, uint32_t vcpuid
);
358 #define UCALL_MAX_ARGS 6
362 uint64_t args
[UCALL_MAX_ARGS
];
365 void ucall_init(struct kvm_vm
*vm
, void *arg
);
366 void ucall_uninit(struct kvm_vm
*vm
);
367 void ucall(uint64_t cmd
, int nargs
, ...);
368 uint64_t get_ucall(struct kvm_vm
*vm
, uint32_t vcpu_id
, struct ucall
*uc
);
370 #define GUEST_SYNC_ARGS(stage, arg1, arg2, arg3, arg4) \
371 ucall(UCALL_SYNC, 6, "hello", stage, arg1, arg2, arg3, arg4)
372 #define GUEST_SYNC(stage) ucall(UCALL_SYNC, 2, "hello", stage)
373 #define GUEST_DONE() ucall(UCALL_DONE, 0)
374 #define __GUEST_ASSERT(_condition, _nargs, _args...) do { \
376 ucall(UCALL_ABORT, 2 + _nargs, \
377 "Failed guest assert: " \
378 #_condition, __LINE__, _args); \
381 #define GUEST_ASSERT(_condition) \
382 __GUEST_ASSERT((_condition), 0, 0)
384 #define GUEST_ASSERT_1(_condition, arg1) \
385 __GUEST_ASSERT((_condition), 1, (arg1))
387 #define GUEST_ASSERT_2(_condition, arg1, arg2) \
388 __GUEST_ASSERT((_condition), 2, (arg1), (arg2))
390 #define GUEST_ASSERT_3(_condition, arg1, arg2, arg3) \
391 __GUEST_ASSERT((_condition), 3, (arg1), (arg2), (arg3))
393 #define GUEST_ASSERT_4(_condition, arg1, arg2, arg3, arg4) \
394 __GUEST_ASSERT((_condition), 4, (arg1), (arg2), (arg3), (arg4))
396 int vm_get_stats_fd(struct kvm_vm
*vm
);
397 int vcpu_get_stats_fd(struct kvm_vm
*vm
, uint32_t vcpuid
);
399 #endif /* SELFTEST_KVM_UTIL_H */