]> git.proxmox.com Git - qemu.git/blob - exec-obsolete.h
blockdev: preserve readonly and snapshot states across media changes
[qemu.git] / exec-obsolete.h
1 /*
2 * Declarations for obsolete exec.c functions
3 *
4 * Copyright 2011 Red Hat, Inc. and/or its affiliates
5 *
6 * Authors:
7 * Avi Kivity <avi@redhat.com>
8 *
9 * This work is licensed under the terms of the GNU GPL, version 2 or
10 * later. See the COPYING file in the top-level directory.
11 *
12 */
13
14 /*
15 * This header is for use by exec.c and memory.c ONLY. Do not include it.
16 * The functions declared here will be removed soon.
17 */
18
19 #ifndef EXEC_OBSOLETE_H
20 #define EXEC_OBSOLETE_H
21
22 #ifndef WANT_EXEC_OBSOLETE
23 #error Do not include exec-obsolete.h
24 #endif
25
26 #ifndef CONFIG_USER_ONLY
27
28 ram_addr_t qemu_ram_alloc_from_ptr(ram_addr_t size, void *host,
29 MemoryRegion *mr);
30 ram_addr_t qemu_ram_alloc(ram_addr_t size, MemoryRegion *mr);
31 void qemu_ram_free(ram_addr_t addr);
32 void qemu_ram_free_from_ptr(ram_addr_t addr);
33
34 struct MemoryRegion;
35 struct MemoryRegionSection;
36 void cpu_register_physical_memory_log(struct MemoryRegionSection *section,
37 bool readonly);
38
39 void qemu_register_coalesced_mmio(target_phys_addr_t addr, ram_addr_t size);
40 void qemu_unregister_coalesced_mmio(target_phys_addr_t addr, ram_addr_t size);
41
42 int cpu_physical_memory_set_dirty_tracking(int enable);
43
44 #define VGA_DIRTY_FLAG 0x01
45 #define CODE_DIRTY_FLAG 0x02
46 #define MIGRATION_DIRTY_FLAG 0x08
47
48 static inline int cpu_physical_memory_get_dirty_flags(ram_addr_t addr)
49 {
50 return ram_list.phys_dirty[addr >> TARGET_PAGE_BITS];
51 }
52
53 /* read dirty bit (return 0 or 1) */
54 static inline int cpu_physical_memory_is_dirty(ram_addr_t addr)
55 {
56 return cpu_physical_memory_get_dirty_flags(addr) == 0xff;
57 }
58
59 static inline int cpu_physical_memory_get_dirty(ram_addr_t start,
60 ram_addr_t length,
61 int dirty_flags)
62 {
63 int ret = 0;
64 ram_addr_t addr, end;
65
66 end = TARGET_PAGE_ALIGN(start + length);
67 start &= TARGET_PAGE_MASK;
68 for (addr = start; addr < end; addr += TARGET_PAGE_SIZE) {
69 ret |= cpu_physical_memory_get_dirty_flags(addr) & dirty_flags;
70 }
71 return ret;
72 }
73
74 static inline int cpu_physical_memory_set_dirty_flags(ram_addr_t addr,
75 int dirty_flags)
76 {
77 if ((dirty_flags & MIGRATION_DIRTY_FLAG) &&
78 !cpu_physical_memory_get_dirty(addr, TARGET_PAGE_SIZE,
79 MIGRATION_DIRTY_FLAG)) {
80 ram_list.dirty_pages++;
81 }
82 return ram_list.phys_dirty[addr >> TARGET_PAGE_BITS] |= dirty_flags;
83 }
84
85 static inline void cpu_physical_memory_set_dirty(ram_addr_t addr)
86 {
87 cpu_physical_memory_set_dirty_flags(addr, 0xff);
88 }
89
90 static inline int cpu_physical_memory_clear_dirty_flags(ram_addr_t addr,
91 int dirty_flags)
92 {
93 int mask = ~dirty_flags;
94
95 if ((dirty_flags & MIGRATION_DIRTY_FLAG) &&
96 cpu_physical_memory_get_dirty(addr, TARGET_PAGE_SIZE,
97 MIGRATION_DIRTY_FLAG)) {
98 ram_list.dirty_pages--;
99 }
100 return ram_list.phys_dirty[addr >> TARGET_PAGE_BITS] &= mask;
101 }
102
103 static inline void cpu_physical_memory_set_dirty_range(ram_addr_t start,
104 ram_addr_t length,
105 int dirty_flags)
106 {
107 ram_addr_t addr, end;
108
109 end = TARGET_PAGE_ALIGN(start + length);
110 start &= TARGET_PAGE_MASK;
111 for (addr = start; addr < end; addr += TARGET_PAGE_SIZE) {
112 cpu_physical_memory_set_dirty_flags(addr, dirty_flags);
113 }
114 }
115
116 static inline void cpu_physical_memory_mask_dirty_range(ram_addr_t start,
117 ram_addr_t length,
118 int dirty_flags)
119 {
120 ram_addr_t addr, end;
121
122 end = TARGET_PAGE_ALIGN(start + length);
123 start &= TARGET_PAGE_MASK;
124 for (addr = start; addr < end; addr += TARGET_PAGE_SIZE) {
125 cpu_physical_memory_clear_dirty_flags(addr, dirty_flags);
126 }
127 }
128
129 void cpu_physical_memory_reset_dirty(ram_addr_t start, ram_addr_t end,
130 int dirty_flags);
131
132 extern const IORangeOps memory_region_iorange_ops;
133
134 #endif
135
136 #endif