]> git.proxmox.com Git - mirror_qemu.git/blob - tests/qtest/fuzz/generic_fuzz.c
Merge tag 'for-upstream' of https://gitlab.com/bonzini/qemu into staging
[mirror_qemu.git] / tests / qtest / fuzz / generic_fuzz.c
1 /*
2 * Generic Virtual-Device Fuzzing Target
3 *
4 * Copyright Red Hat Inc., 2020
5 *
6 * Authors:
7 * Alexander Bulekov <alxndr@bu.edu>
8 *
9 * This work is licensed under the terms of the GNU GPL, version 2 or later.
10 * See the COPYING file in the top-level directory.
11 */
12
13 #include "qemu/osdep.h"
14
15 #include <wordexp.h>
16
17 #include "hw/core/cpu.h"
18 #include "tests/qtest/libqtest.h"
19 #include "tests/qtest/libqos/pci-pc.h"
20 #include "fuzz.h"
21 #include "fork_fuzz.h"
22 #include "string.h"
23 #include "exec/memory.h"
24 #include "exec/ramblock.h"
25 #include "hw/qdev-core.h"
26 #include "hw/pci/pci.h"
27 #include "hw/boards.h"
28 #include "generic_fuzz_configs.h"
29 #include "hw/mem/sparse-mem.h"
30
31 /*
32 * SEPARATOR is used to separate "operations" in the fuzz input
33 */
34 #define SEPARATOR "FUZZ"
35
36 enum cmds {
37 OP_IN,
38 OP_OUT,
39 OP_READ,
40 OP_WRITE,
41 OP_PCI_READ,
42 OP_PCI_WRITE,
43 OP_DISABLE_PCI,
44 OP_ADD_DMA_PATTERN,
45 OP_CLEAR_DMA_PATTERNS,
46 OP_CLOCK_STEP,
47 };
48
49 #define DEFAULT_TIMEOUT_US 100000
50 #define USEC_IN_SEC 1000000000
51
52 #define MAX_DMA_FILL_SIZE 0x10000
53
54 #define PCI_HOST_BRIDGE_CFG 0xcf8
55 #define PCI_HOST_BRIDGE_DATA 0xcfc
56
57 typedef struct {
58 ram_addr_t addr;
59 ram_addr_t size; /* The number of bytes until the end of the I/O region */
60 } address_range;
61
62 static useconds_t timeout = DEFAULT_TIMEOUT_US;
63
64 static bool qtest_log_enabled;
65
66 MemoryRegion *sparse_mem_mr;
67
68 /*
69 * A pattern used to populate a DMA region or perform a memwrite. This is
70 * useful for e.g. populating tables of unique addresses.
71 * Example {.index = 1; .stride = 2; .len = 3; .data = "\x00\x01\x02"}
72 * Renders as: 00 01 02 00 03 02 00 05 02 00 07 02 ...
73 */
74 typedef struct {
75 uint8_t index; /* Index of a byte to increment by stride */
76 uint8_t stride; /* Increment each index'th byte by this amount */
77 size_t len;
78 const uint8_t *data;
79 } pattern;
80
81 /* Avoid filling the same DMA region between MMIO/PIO commands ? */
82 static bool avoid_double_fetches;
83
84 static QTestState *qts_global; /* Need a global for the DMA callback */
85
86 /*
87 * List of memory regions that are children of QOM objects specified by the
88 * user for fuzzing.
89 */
90 static GHashTable *fuzzable_memoryregions;
91 static GPtrArray *fuzzable_pci_devices;
92
93 struct get_io_cb_info {
94 int index;
95 int found;
96 address_range result;
97 };
98
99 static bool get_io_address_cb(Int128 start, Int128 size,
100 const MemoryRegion *mr,
101 hwaddr offset_in_region,
102 void *opaque)
103 {
104 struct get_io_cb_info *info = opaque;
105 if (g_hash_table_lookup(fuzzable_memoryregions, mr)) {
106 if (info->index == 0) {
107 info->result.addr = (ram_addr_t)start;
108 info->result.size = (ram_addr_t)size;
109 info->found = 1;
110 return true;
111 }
112 info->index--;
113 }
114 return false;
115 }
116
117 /*
118 * List of dma regions populated since the last fuzzing command. Used to ensure
119 * that we only write to each DMA address once, to avoid race conditions when
120 * building reproducers.
121 */
122 static GArray *dma_regions;
123
124 static GArray *dma_patterns;
125 static int dma_pattern_index;
126 static bool pci_disabled;
127
128 /*
129 * Allocate a block of memory and populate it with a pattern.
130 */
131 static void *pattern_alloc(pattern p, size_t len)
132 {
133 int i;
134 uint8_t *buf = g_malloc(len);
135 uint8_t sum = 0;
136
137 for (i = 0; i < len; ++i) {
138 buf[i] = p.data[i % p.len];
139 if ((i % p.len) == p.index) {
140 buf[i] += sum;
141 sum += p.stride;
142 }
143 }
144 return buf;
145 }
146
147 static int fuzz_memory_access_size(MemoryRegion *mr, unsigned l, hwaddr addr)
148 {
149 unsigned access_size_max = mr->ops->valid.max_access_size;
150
151 /*
152 * Regions are assumed to support 1-4 byte accesses unless
153 * otherwise specified.
154 */
155 if (access_size_max == 0) {
156 access_size_max = 4;
157 }
158
159 /* Bound the maximum access by the alignment of the address. */
160 if (!mr->ops->impl.unaligned) {
161 unsigned align_size_max = addr & -addr;
162 if (align_size_max != 0 && align_size_max < access_size_max) {
163 access_size_max = align_size_max;
164 }
165 }
166
167 /* Don't attempt accesses larger than the maximum. */
168 if (l > access_size_max) {
169 l = access_size_max;
170 }
171 l = pow2floor(l);
172
173 return l;
174 }
175
176 /*
177 * Call-back for functions that perform DMA reads from guest memory. Confirm
178 * that the region has not already been populated since the last loop in
179 * generic_fuzz(), avoiding potential race-conditions, which we don't have
180 * a good way for reproducing right now.
181 */
182 void fuzz_dma_read_cb(size_t addr, size_t len, MemoryRegion *mr)
183 {
184 /* Are we in the generic-fuzzer or are we using another fuzz-target? */
185 if (!qts_global) {
186 return;
187 }
188
189 /*
190 * Return immediately if:
191 * - We have no DMA patterns defined
192 * - The length of the DMA read request is zero
193 * - The DMA read is hitting an MR other than the machine's main RAM
194 * - The DMA request hits past the bounds of our RAM
195 */
196 if (dma_patterns->len == 0
197 || len == 0
198 || (mr != current_machine->ram && mr != sparse_mem_mr)) {
199 return;
200 }
201
202 /*
203 * If we overlap with any existing dma_regions, split the range and only
204 * populate the non-overlapping parts.
205 */
206 address_range region;
207 bool double_fetch = false;
208 for (int i = 0;
209 i < dma_regions->len && (avoid_double_fetches || qtest_log_enabled);
210 ++i) {
211 region = g_array_index(dma_regions, address_range, i);
212 if (addr < region.addr + region.size && addr + len > region.addr) {
213 double_fetch = true;
214 if (addr < region.addr
215 && avoid_double_fetches) {
216 fuzz_dma_read_cb(addr, region.addr - addr, mr);
217 }
218 if (addr + len > region.addr + region.size
219 && avoid_double_fetches) {
220 fuzz_dma_read_cb(region.addr + region.size,
221 addr + len - (region.addr + region.size), mr);
222 }
223 return;
224 }
225 }
226
227 /* Cap the length of the DMA access to something reasonable */
228 len = MIN(len, MAX_DMA_FILL_SIZE);
229
230 address_range ar = {addr, len};
231 g_array_append_val(dma_regions, ar);
232 pattern p = g_array_index(dma_patterns, pattern, dma_pattern_index);
233 void *buf_base = pattern_alloc(p, ar.size);
234 void *buf = buf_base;
235 hwaddr l, addr1;
236 MemoryRegion *mr1;
237 while (len > 0) {
238 l = len;
239 mr1 = address_space_translate(first_cpu->as,
240 addr, &addr1, &l, true,
241 MEMTXATTRS_UNSPECIFIED);
242
243 /*
244 * If mr1 isn't RAM, address_space_translate doesn't update l. Use
245 * fuzz_memory_access_size to identify the number of bytes that it
246 * is safe to write without accidentally writing to another
247 * MemoryRegion.
248 */
249 if (!memory_region_is_ram(mr1)) {
250 l = fuzz_memory_access_size(mr1, l, addr1);
251 }
252 if (memory_region_is_ram(mr1) ||
253 memory_region_is_romd(mr1) ||
254 mr1 == sparse_mem_mr) {
255 /* ROM/RAM case */
256 if (qtest_log_enabled) {
257 /*
258 * With QTEST_LOG, use a normal, slow QTest memwrite. Prefix the log
259 * that will be written by qtest.c with a DMA tag, so we can reorder
260 * the resulting QTest trace so the DMA fills precede the last PIO/MMIO
261 * command.
262 */
263 fprintf(stderr, "[DMA] ");
264 if (double_fetch) {
265 fprintf(stderr, "[DOUBLE-FETCH] ");
266 }
267 fflush(stderr);
268 }
269 qtest_memwrite(qts_global, addr, buf, l);
270 }
271 len -= l;
272 buf += l;
273 addr += l;
274
275 }
276 g_free(buf_base);
277
278 /* Increment the index of the pattern for the next DMA access */
279 dma_pattern_index = (dma_pattern_index + 1) % dma_patterns->len;
280 }
281
282 /*
283 * Here we want to convert a fuzzer-provided [io-region-index, offset] to
284 * a physical address. To do this, we iterate over all of the matched
285 * MemoryRegions. Check whether each region exists within the particular io
286 * space. Return the absolute address of the offset within the index'th region
287 * that is a subregion of the io_space and the distance until the end of the
288 * memory region.
289 */
290 static bool get_io_address(address_range *result, AddressSpace *as,
291 uint8_t index,
292 uint32_t offset) {
293 FlatView *view;
294 view = as->current_map;
295 g_assert(view);
296 struct get_io_cb_info cb_info = {};
297
298 cb_info.index = index;
299
300 /*
301 * Loop around the FlatView until we match "index" number of
302 * fuzzable_memoryregions, or until we know that there are no matching
303 * memory_regions.
304 */
305 do {
306 flatview_for_each_range(view, get_io_address_cb , &cb_info);
307 } while (cb_info.index != index && !cb_info.found);
308
309 *result = cb_info.result;
310 if (result->size) {
311 offset = offset % result->size;
312 result->addr += offset;
313 result->size -= offset;
314 }
315 return cb_info.found;
316 }
317
318 static bool get_pio_address(address_range *result,
319 uint8_t index, uint16_t offset)
320 {
321 /*
322 * PIO BARs can be set past the maximum port address (0xFFFF). Thus, result
323 * can contain an addr that extends past the PIO space. When we pass this
324 * address to qtest_in/qtest_out, it is cast to a uint16_t, so we might end
325 * up fuzzing a completely different MemoryRegion/Device. Therefore, check
326 * that the address here is within the PIO space limits.
327 */
328 bool found = get_io_address(result, &address_space_io, index, offset);
329 return result->addr <= 0xFFFF ? found : false;
330 }
331
332 static bool get_mmio_address(address_range *result,
333 uint8_t index, uint32_t offset)
334 {
335 return get_io_address(result, &address_space_memory, index, offset);
336 }
337
338 static void op_in(QTestState *s, const unsigned char * data, size_t len)
339 {
340 enum Sizes {Byte, Word, Long, end_sizes};
341 struct {
342 uint8_t size;
343 uint8_t base;
344 uint16_t offset;
345 } a;
346 address_range abs;
347
348 if (len < sizeof(a)) {
349 return;
350 }
351 memcpy(&a, data, sizeof(a));
352 if (get_pio_address(&abs, a.base, a.offset) == 0) {
353 return;
354 }
355
356 switch (a.size %= end_sizes) {
357 case Byte:
358 qtest_inb(s, abs.addr);
359 break;
360 case Word:
361 if (abs.size >= 2) {
362 qtest_inw(s, abs.addr);
363 }
364 break;
365 case Long:
366 if (abs.size >= 4) {
367 qtest_inl(s, abs.addr);
368 }
369 break;
370 }
371 }
372
373 static void op_out(QTestState *s, const unsigned char * data, size_t len)
374 {
375 enum Sizes {Byte, Word, Long, end_sizes};
376 struct {
377 uint8_t size;
378 uint8_t base;
379 uint16_t offset;
380 uint32_t value;
381 } a;
382 address_range abs;
383
384 if (len < sizeof(a)) {
385 return;
386 }
387 memcpy(&a, data, sizeof(a));
388
389 if (get_pio_address(&abs, a.base, a.offset) == 0) {
390 return;
391 }
392
393 switch (a.size %= end_sizes) {
394 case Byte:
395 qtest_outb(s, abs.addr, a.value & 0xFF);
396 break;
397 case Word:
398 if (abs.size >= 2) {
399 qtest_outw(s, abs.addr, a.value & 0xFFFF);
400 }
401 break;
402 case Long:
403 if (abs.size >= 4) {
404 qtest_outl(s, abs.addr, a.value);
405 }
406 break;
407 }
408 }
409
410 static void op_read(QTestState *s, const unsigned char * data, size_t len)
411 {
412 enum Sizes {Byte, Word, Long, Quad, end_sizes};
413 struct {
414 uint8_t size;
415 uint8_t base;
416 uint32_t offset;
417 } a;
418 address_range abs;
419
420 if (len < sizeof(a)) {
421 return;
422 }
423 memcpy(&a, data, sizeof(a));
424
425 if (get_mmio_address(&abs, a.base, a.offset) == 0) {
426 return;
427 }
428
429 switch (a.size %= end_sizes) {
430 case Byte:
431 qtest_readb(s, abs.addr);
432 break;
433 case Word:
434 if (abs.size >= 2) {
435 qtest_readw(s, abs.addr);
436 }
437 break;
438 case Long:
439 if (abs.size >= 4) {
440 qtest_readl(s, abs.addr);
441 }
442 break;
443 case Quad:
444 if (abs.size >= 8) {
445 qtest_readq(s, abs.addr);
446 }
447 break;
448 }
449 }
450
451 static void op_write(QTestState *s, const unsigned char * data, size_t len)
452 {
453 enum Sizes {Byte, Word, Long, Quad, end_sizes};
454 struct {
455 uint8_t size;
456 uint8_t base;
457 uint32_t offset;
458 uint64_t value;
459 } a;
460 address_range abs;
461
462 if (len < sizeof(a)) {
463 return;
464 }
465 memcpy(&a, data, sizeof(a));
466
467 if (get_mmio_address(&abs, a.base, a.offset) == 0) {
468 return;
469 }
470
471 switch (a.size %= end_sizes) {
472 case Byte:
473 qtest_writeb(s, abs.addr, a.value & 0xFF);
474 break;
475 case Word:
476 if (abs.size >= 2) {
477 qtest_writew(s, abs.addr, a.value & 0xFFFF);
478 }
479 break;
480 case Long:
481 if (abs.size >= 4) {
482 qtest_writel(s, abs.addr, a.value & 0xFFFFFFFF);
483 }
484 break;
485 case Quad:
486 if (abs.size >= 8) {
487 qtest_writeq(s, abs.addr, a.value);
488 }
489 break;
490 }
491 }
492
493 static void op_pci_read(QTestState *s, const unsigned char * data, size_t len)
494 {
495 enum Sizes {Byte, Word, Long, end_sizes};
496 struct {
497 uint8_t size;
498 uint8_t base;
499 uint8_t offset;
500 } a;
501 if (len < sizeof(a) || fuzzable_pci_devices->len == 0 || pci_disabled) {
502 return;
503 }
504 memcpy(&a, data, sizeof(a));
505 PCIDevice *dev = g_ptr_array_index(fuzzable_pci_devices,
506 a.base % fuzzable_pci_devices->len);
507 int devfn = dev->devfn;
508 qtest_outl(s, PCI_HOST_BRIDGE_CFG, (1U << 31) | (devfn << 8) | a.offset);
509 switch (a.size %= end_sizes) {
510 case Byte:
511 qtest_inb(s, PCI_HOST_BRIDGE_DATA);
512 break;
513 case Word:
514 qtest_inw(s, PCI_HOST_BRIDGE_DATA);
515 break;
516 case Long:
517 qtest_inl(s, PCI_HOST_BRIDGE_DATA);
518 break;
519 }
520 }
521
522 static void op_pci_write(QTestState *s, const unsigned char * data, size_t len)
523 {
524 enum Sizes {Byte, Word, Long, end_sizes};
525 struct {
526 uint8_t size;
527 uint8_t base;
528 uint8_t offset;
529 uint32_t value;
530 } a;
531 if (len < sizeof(a) || fuzzable_pci_devices->len == 0 || pci_disabled) {
532 return;
533 }
534 memcpy(&a, data, sizeof(a));
535 PCIDevice *dev = g_ptr_array_index(fuzzable_pci_devices,
536 a.base % fuzzable_pci_devices->len);
537 int devfn = dev->devfn;
538 qtest_outl(s, PCI_HOST_BRIDGE_CFG, (1U << 31) | (devfn << 8) | a.offset);
539 switch (a.size %= end_sizes) {
540 case Byte:
541 qtest_outb(s, PCI_HOST_BRIDGE_DATA, a.value & 0xFF);
542 break;
543 case Word:
544 qtest_outw(s, PCI_HOST_BRIDGE_DATA, a.value & 0xFFFF);
545 break;
546 case Long:
547 qtest_outl(s, PCI_HOST_BRIDGE_DATA, a.value & 0xFFFFFFFF);
548 break;
549 }
550 }
551
552 static void op_add_dma_pattern(QTestState *s,
553 const unsigned char *data, size_t len)
554 {
555 struct {
556 /*
557 * index and stride can be used to increment the index-th byte of the
558 * pattern by the value stride, for each loop of the pattern.
559 */
560 uint8_t index;
561 uint8_t stride;
562 } a;
563
564 if (len < sizeof(a) + 1) {
565 return;
566 }
567 memcpy(&a, data, sizeof(a));
568 pattern p = {a.index, a.stride, len - sizeof(a), data + sizeof(a)};
569 p.index = a.index % p.len;
570 g_array_append_val(dma_patterns, p);
571 return;
572 }
573
574 static void op_clear_dma_patterns(QTestState *s,
575 const unsigned char *data, size_t len)
576 {
577 g_array_set_size(dma_patterns, 0);
578 dma_pattern_index = 0;
579 }
580
581 static void op_clock_step(QTestState *s, const unsigned char *data, size_t len)
582 {
583 qtest_clock_step_next(s);
584 }
585
586 static void op_disable_pci(QTestState *s, const unsigned char *data, size_t len)
587 {
588 pci_disabled = true;
589 }
590
591 static void handle_timeout(int sig)
592 {
593 if (qtest_log_enabled) {
594 fprintf(stderr, "[Timeout]\n");
595 fflush(stderr);
596 }
597
598 /*
599 * If there is a crash, libfuzzer/ASAN forks a child to run an
600 * "llvm-symbolizer" process for printing out a pretty stacktrace. It
601 * communicates with this child using a pipe. If we timeout+Exit, while
602 * libfuzzer is still communicating with the llvm-symbolizer child, we will
603 * be left with an orphan llvm-symbolizer process. Sometimes, this appears
604 * to lead to a deadlock in the forkserver. Use waitpid to check if there
605 * are any waitable children. If so, exit out of the signal-handler, and
606 * let libfuzzer finish communicating with the child, and exit, on its own.
607 */
608 if (waitpid(-1, NULL, WNOHANG) == 0) {
609 return;
610 }
611
612 _Exit(0);
613 }
614
615 /*
616 * Here, we interpret random bytes from the fuzzer, as a sequence of commands.
617 * Some commands can be variable-width, so we use a separator, SEPARATOR, to
618 * specify the boundaries between commands. SEPARATOR is used to separate
619 * "operations" in the fuzz input. Why use a separator, instead of just using
620 * the operations' length to identify operation boundaries?
621 * 1. This is a simple way to support variable-length operations
622 * 2. This adds "stability" to the input.
623 * For example take the input "AbBcgDefg", where there is no separator and
624 * Opcodes are capitalized.
625 * Simply, by removing the first byte, we end up with a very different
626 * sequence:
627 * BbcGdefg...
628 * By adding a separator, we avoid this problem:
629 * Ab SEP Bcg SEP Defg -> B SEP Bcg SEP Defg
630 * Since B uses two additional bytes as operands, the first "B" will be
631 * ignored. The fuzzer actively tries to reduce inputs, so such unused
632 * bytes are likely to be pruned, eventually.
633 *
634 * SEPARATOR is trivial for the fuzzer to discover when using ASan. Optionally,
635 * SEPARATOR can be manually specified as a dictionary value (see libfuzzer's
636 * -dict), though this should not be necessary.
637 *
638 * As a result, the stream of bytes is converted into a sequence of commands.
639 * In a simplified example where SEPARATOR is 0xFF:
640 * 00 01 02 FF 03 04 05 06 FF 01 FF ...
641 * becomes this sequence of commands:
642 * 00 01 02 -> op00 (0102) -> in (0102, 2)
643 * 03 04 05 06 -> op03 (040506) -> write (040506, 3)
644 * 01 -> op01 (-,0) -> out (-,0)
645 * ...
646 *
647 * Note here that it is the job of the individual opcode functions to check
648 * that enough data was provided. I.e. in the last command out (,0), out needs
649 * to check that there is not enough data provided to select an address/value
650 * for the operation.
651 */
652 static void generic_fuzz(QTestState *s, const unsigned char *Data, size_t Size)
653 {
654 void (*ops[]) (QTestState *s, const unsigned char* , size_t) = {
655 [OP_IN] = op_in,
656 [OP_OUT] = op_out,
657 [OP_READ] = op_read,
658 [OP_WRITE] = op_write,
659 [OP_PCI_READ] = op_pci_read,
660 [OP_PCI_WRITE] = op_pci_write,
661 [OP_DISABLE_PCI] = op_disable_pci,
662 [OP_ADD_DMA_PATTERN] = op_add_dma_pattern,
663 [OP_CLEAR_DMA_PATTERNS] = op_clear_dma_patterns,
664 [OP_CLOCK_STEP] = op_clock_step,
665 };
666 const unsigned char *cmd = Data;
667 const unsigned char *nextcmd;
668 size_t cmd_len;
669 uint8_t op;
670
671 if (fork() == 0) {
672 struct sigaction sact;
673 struct itimerval timer;
674 sigset_t set;
675 /*
676 * Sometimes the fuzzer will find inputs that take quite a long time to
677 * process. Often times, these inputs do not result in new coverage.
678 * Even if these inputs might be interesting, they can slow down the
679 * fuzzer, overall. Set a timeout for each command to avoid hurting
680 * performance, too much
681 */
682 if (timeout) {
683
684 sigemptyset(&sact.sa_mask);
685 sact.sa_flags = SA_NODEFER;
686 sact.sa_handler = handle_timeout;
687 sigaction(SIGALRM, &sact, NULL);
688
689 sigemptyset(&set);
690 sigaddset(&set, SIGALRM);
691 pthread_sigmask(SIG_UNBLOCK, &set, NULL);
692
693 memset(&timer, 0, sizeof(timer));
694 timer.it_value.tv_sec = timeout / USEC_IN_SEC;
695 timer.it_value.tv_usec = timeout % USEC_IN_SEC;
696 }
697
698 op_clear_dma_patterns(s, NULL, 0);
699 pci_disabled = false;
700
701 while (cmd && Size) {
702 /* Reset the timeout, each time we run a new command */
703 if (timeout) {
704 setitimer(ITIMER_REAL, &timer, NULL);
705 }
706
707 /* Get the length until the next command or end of input */
708 nextcmd = memmem(cmd, Size, SEPARATOR, strlen(SEPARATOR));
709 cmd_len = nextcmd ? nextcmd - cmd : Size;
710
711 if (cmd_len > 0) {
712 /* Interpret the first byte of the command as an opcode */
713 op = *cmd % (sizeof(ops) / sizeof((ops)[0]));
714 ops[op](s, cmd + 1, cmd_len - 1);
715
716 /* Run the main loop */
717 flush_events(s);
718 }
719 /* Advance to the next command */
720 cmd = nextcmd ? nextcmd + sizeof(SEPARATOR) - 1 : nextcmd;
721 Size = Size - (cmd_len + sizeof(SEPARATOR) - 1);
722 g_array_set_size(dma_regions, 0);
723 }
724 _Exit(0);
725 } else {
726 flush_events(s);
727 wait(0);
728 }
729 }
730
731 static void usage(void)
732 {
733 printf("Please specify the following environment variables:\n");
734 printf("QEMU_FUZZ_ARGS= the command line arguments passed to qemu\n");
735 printf("QEMU_FUZZ_OBJECTS= "
736 "a space separated list of QOM type names for objects to fuzz\n");
737 printf("Optionally: QEMU_AVOID_DOUBLE_FETCH= "
738 "Try to avoid racy DMA double fetch bugs? %d by default\n",
739 avoid_double_fetches);
740 printf("Optionally: QEMU_FUZZ_TIMEOUT= Specify a custom timeout (us). "
741 "0 to disable. %d by default\n", timeout);
742 exit(0);
743 }
744
745 static int locate_fuzz_memory_regions(Object *child, void *opaque)
746 {
747 MemoryRegion *mr;
748 if (object_dynamic_cast(child, TYPE_MEMORY_REGION)) {
749 mr = MEMORY_REGION(child);
750 if ((memory_region_is_ram(mr) ||
751 memory_region_is_ram_device(mr) ||
752 memory_region_is_rom(mr)) == false) {
753 /*
754 * We don't want duplicate pointers to the same MemoryRegion, so
755 * try to remove copies of the pointer, before adding it.
756 */
757 g_hash_table_insert(fuzzable_memoryregions, mr, (gpointer)true);
758 }
759 }
760 return 0;
761 }
762
763 static int locate_fuzz_objects(Object *child, void *opaque)
764 {
765 GString *type_name;
766 GString *path_name;
767 char *pattern = opaque;
768
769 type_name = g_string_new(object_get_typename(child));
770 g_string_ascii_down(type_name);
771 if (g_pattern_match_simple(pattern, type_name->str)) {
772 /* Find and save ptrs to any child MemoryRegions */
773 object_child_foreach_recursive(child, locate_fuzz_memory_regions, NULL);
774
775 /*
776 * We matched an object. If its a PCI device, store a pointer to it so
777 * we can map BARs and fuzz its config space.
778 */
779 if (object_dynamic_cast(OBJECT(child), TYPE_PCI_DEVICE)) {
780 /*
781 * Don't want duplicate pointers to the same PCIDevice, so remove
782 * copies of the pointer, before adding it.
783 */
784 g_ptr_array_remove_fast(fuzzable_pci_devices, PCI_DEVICE(child));
785 g_ptr_array_add(fuzzable_pci_devices, PCI_DEVICE(child));
786 }
787 } else if (object_dynamic_cast(OBJECT(child), TYPE_MEMORY_REGION)) {
788 path_name = g_string_new(object_get_canonical_path_component(child));
789 g_string_ascii_down(path_name);
790 if (g_pattern_match_simple(pattern, path_name->str)) {
791 MemoryRegion *mr;
792 mr = MEMORY_REGION(child);
793 if ((memory_region_is_ram(mr) ||
794 memory_region_is_ram_device(mr) ||
795 memory_region_is_rom(mr)) == false) {
796 g_hash_table_insert(fuzzable_memoryregions, mr, (gpointer)true);
797 }
798 }
799 g_string_free(path_name, true);
800 }
801 g_string_free(type_name, true);
802 return 0;
803 }
804
805
806 static void pci_enum(gpointer pcidev, gpointer bus)
807 {
808 PCIDevice *dev = pcidev;
809 QPCIDevice *qdev;
810 int i;
811
812 qdev = qpci_device_find(bus, dev->devfn);
813 g_assert(qdev != NULL);
814 for (i = 0; i < 6; i++) {
815 if (dev->io_regions[i].size) {
816 qpci_iomap(qdev, i, NULL);
817 }
818 }
819 qpci_device_enable(qdev);
820 g_free(qdev);
821 }
822
823 static void generic_pre_fuzz(QTestState *s)
824 {
825 GHashTableIter iter;
826 MemoryRegion *mr;
827 QPCIBus *pcibus;
828 char **result;
829 GString *name_pattern;
830
831 if (!getenv("QEMU_FUZZ_OBJECTS")) {
832 usage();
833 }
834 if (getenv("QTEST_LOG")) {
835 qtest_log_enabled = 1;
836 }
837 if (getenv("QEMU_AVOID_DOUBLE_FETCH")) {
838 avoid_double_fetches = 1;
839 }
840 if (getenv("QEMU_FUZZ_TIMEOUT")) {
841 timeout = g_ascii_strtoll(getenv("QEMU_FUZZ_TIMEOUT"), NULL, 0);
842 }
843 qts_global = s;
844
845 /*
846 * Create a special device that we can use to back DMA buffers at very
847 * high memory addresses
848 */
849 sparse_mem_mr = sparse_mem_init(0, UINT64_MAX);
850
851 dma_regions = g_array_new(false, false, sizeof(address_range));
852 dma_patterns = g_array_new(false, false, sizeof(pattern));
853
854 fuzzable_memoryregions = g_hash_table_new(NULL, NULL);
855 fuzzable_pci_devices = g_ptr_array_new();
856
857 result = g_strsplit(getenv("QEMU_FUZZ_OBJECTS"), " ", -1);
858 for (int i = 0; result[i] != NULL; i++) {
859 name_pattern = g_string_new(result[i]);
860 /*
861 * Make the pattern lowercase. We do the same for all the MemoryRegion
862 * and Type names so the configs are case-insensitive.
863 */
864 g_string_ascii_down(name_pattern);
865 printf("Matching objects by name %s\n", result[i]);
866 object_child_foreach_recursive(qdev_get_machine(),
867 locate_fuzz_objects,
868 name_pattern->str);
869 g_string_free(name_pattern, true);
870 }
871 g_strfreev(result);
872 printf("This process will try to fuzz the following MemoryRegions:\n");
873
874 g_hash_table_iter_init(&iter, fuzzable_memoryregions);
875 while (g_hash_table_iter_next(&iter, (gpointer)&mr, NULL)) {
876 printf(" * %s (size 0x%" PRIx64 ")\n",
877 object_get_canonical_path_component(&(mr->parent_obj)),
878 memory_region_size(mr));
879 }
880
881 if (!g_hash_table_size(fuzzable_memoryregions)) {
882 printf("No fuzzable memory regions found...\n");
883 exit(1);
884 }
885
886 pcibus = qpci_new_pc(s, NULL);
887 g_ptr_array_foreach(fuzzable_pci_devices, pci_enum, pcibus);
888 qpci_free_pc(pcibus);
889
890 counter_shm_init();
891 }
892
893 /*
894 * When libfuzzer gives us two inputs to combine, return a new input with the
895 * following structure:
896 *
897 * Input 1 (data1)
898 * SEPARATOR
899 * Clear out the DMA Patterns
900 * SEPARATOR
901 * Disable the pci_read/write instructions
902 * SEPARATOR
903 * Input 2 (data2)
904 *
905 * The idea is to collate the core behaviors of the two inputs.
906 * For example:
907 * Input 1: maps a device's BARs, sets up three DMA patterns, and triggers
908 * device functionality A
909 * Input 2: maps a device's BARs, sets up one DMA pattern, and triggers device
910 * functionality B
911 *
912 * This function attempts to produce an input that:
913 * Ouptut: maps a device's BARs, set up three DMA patterns, triggers
914 * functionality A device, replaces the DMA patterns with a single
915 * patten, and triggers device functionality B.
916 */
917 static size_t generic_fuzz_crossover(const uint8_t *data1, size_t size1, const
918 uint8_t *data2, size_t size2, uint8_t *out,
919 size_t max_out_size, unsigned int seed)
920 {
921 size_t copy_len = 0, size = 0;
922
923 /* Check that we have enough space for data1 and at least part of data2 */
924 if (max_out_size <= size1 + strlen(SEPARATOR) * 3 + 2) {
925 return 0;
926 }
927
928 /* Copy_Len in the first input */
929 copy_len = size1;
930 memcpy(out + size, data1, copy_len);
931 size += copy_len;
932 max_out_size -= copy_len;
933
934 /* Append a separator */
935 copy_len = strlen(SEPARATOR);
936 memcpy(out + size, SEPARATOR, copy_len);
937 size += copy_len;
938 max_out_size -= copy_len;
939
940 /* Clear out the DMA Patterns */
941 copy_len = 1;
942 if (copy_len) {
943 out[size] = OP_CLEAR_DMA_PATTERNS;
944 }
945 size += copy_len;
946 max_out_size -= copy_len;
947
948 /* Append a separator */
949 copy_len = strlen(SEPARATOR);
950 memcpy(out + size, SEPARATOR, copy_len);
951 size += copy_len;
952 max_out_size -= copy_len;
953
954 /* Disable PCI ops. Assume data1 took care of setting up PCI */
955 copy_len = 1;
956 if (copy_len) {
957 out[size] = OP_DISABLE_PCI;
958 }
959 size += copy_len;
960 max_out_size -= copy_len;
961
962 /* Append a separator */
963 copy_len = strlen(SEPARATOR);
964 memcpy(out + size, SEPARATOR, copy_len);
965 size += copy_len;
966 max_out_size -= copy_len;
967
968 /* Copy_Len over the second input */
969 copy_len = MIN(size2, max_out_size);
970 memcpy(out + size, data2, copy_len);
971 size += copy_len;
972 max_out_size -= copy_len;
973
974 return size;
975 }
976
977
978 static GString *generic_fuzz_cmdline(FuzzTarget *t)
979 {
980 GString *cmd_line = g_string_new(TARGET_NAME);
981 if (!getenv("QEMU_FUZZ_ARGS")) {
982 usage();
983 }
984 g_string_append_printf(cmd_line, " -display none \
985 -machine accel=qtest, \
986 -m 512M %s ", getenv("QEMU_FUZZ_ARGS"));
987 return cmd_line;
988 }
989
990 static GString *generic_fuzz_predefined_config_cmdline(FuzzTarget *t)
991 {
992 gchar *args;
993 const generic_fuzz_config *config;
994 g_assert(t->opaque);
995
996 config = t->opaque;
997 setenv("QEMU_AVOID_DOUBLE_FETCH", "1", 1);
998 if (config->argfunc) {
999 args = config->argfunc();
1000 setenv("QEMU_FUZZ_ARGS", args, 1);
1001 g_free(args);
1002 } else {
1003 g_assert_nonnull(config->args);
1004 setenv("QEMU_FUZZ_ARGS", config->args, 1);
1005 }
1006 setenv("QEMU_FUZZ_OBJECTS", config->objects, 1);
1007 return generic_fuzz_cmdline(t);
1008 }
1009
1010 static void register_generic_fuzz_targets(void)
1011 {
1012 fuzz_add_target(&(FuzzTarget){
1013 .name = "generic-fuzz",
1014 .description = "Fuzz based on any qemu command-line args. ",
1015 .get_init_cmdline = generic_fuzz_cmdline,
1016 .pre_fuzz = generic_pre_fuzz,
1017 .fuzz = generic_fuzz,
1018 .crossover = generic_fuzz_crossover
1019 });
1020
1021 GString *name;
1022 const generic_fuzz_config *config;
1023
1024 for (int i = 0;
1025 i < sizeof(predefined_configs) / sizeof(generic_fuzz_config);
1026 i++) {
1027 config = predefined_configs + i;
1028 name = g_string_new("generic-fuzz");
1029 g_string_append_printf(name, "-%s", config->name);
1030 fuzz_add_target(&(FuzzTarget){
1031 .name = name->str,
1032 .description = "Predefined generic-fuzz config.",
1033 .get_init_cmdline = generic_fuzz_predefined_config_cmdline,
1034 .pre_fuzz = generic_pre_fuzz,
1035 .fuzz = generic_fuzz,
1036 .crossover = generic_fuzz_crossover,
1037 .opaque = (void *)config
1038 });
1039 }
1040 }
1041
1042 fuzz_target_init(register_generic_fuzz_targets);