]> git.proxmox.com Git - mirror_qemu.git/blame - util/cacheflush.c
bsd-user: add arm target build
[mirror_qemu.git] / util / cacheflush.c
CommitLineData
084cfca1
RH
1/*
2 * Flush the host cpu caches.
3 *
4 * This work is licensed under the terms of the GNU GPL, version 2 or later.
5 * See the COPYING file in the top-level directory.
6 */
7
8#include "qemu/osdep.h"
9#include "qemu/cacheflush.h"
664a7973 10#include "qemu/bitops.h"
084cfca1
RH
11
12
13#if defined(__i386__) || defined(__x86_64__) || defined(__s390__)
14
15/* Caches are coherent and do not require flushing; symbol inline. */
16
664a7973
RH
17#elif defined(__aarch64__)
18
19#ifdef CONFIG_DARWIN
20/* Apple does not expose CTR_EL0, so we must use system interfaces. */
21extern void sys_icache_invalidate(void *start, size_t len);
22extern void sys_dcache_flush(void *start, size_t len);
23void flush_idcache_range(uintptr_t rx, uintptr_t rw, size_t len)
24{
25 sys_dcache_flush((void *)rw, len);
26 sys_icache_invalidate((void *)rx, len);
27}
28#else
29
30/*
31 * TODO: unify this with cacheinfo.c.
32 * We want to save the whole contents of CTR_EL0, so that we
33 * have more than the linesize, but also IDC and DIC.
34 */
acd15fc2 35static uint64_t save_ctr_el0;
664a7973
RH
36static void __attribute__((constructor)) init_ctr_el0(void)
37{
38 asm volatile("mrs\t%0, ctr_el0" : "=r"(save_ctr_el0));
39}
40
41/*
42 * This is a copy of gcc's __aarch64_sync_cache_range, modified
43 * to fit this three-operand interface.
44 */
45void flush_idcache_range(uintptr_t rx, uintptr_t rw, size_t len)
46{
47 const unsigned CTR_IDC = 1u << 28;
48 const unsigned CTR_DIC = 1u << 29;
acd15fc2
GQ
49 const uint64_t ctr_el0 = save_ctr_el0;
50 const uintptr_t icache_lsize = 4 << extract64(ctr_el0, 0, 4);
51 const uintptr_t dcache_lsize = 4 << extract64(ctr_el0, 16, 4);
664a7973
RH
52 uintptr_t p;
53
54 /*
55 * If CTR_EL0.IDC is enabled, Data cache clean to the Point of Unification
56 * is not required for instruction to data coherence.
57 */
58 if (!(ctr_el0 & CTR_IDC)) {
59 /*
60 * Loop over the address range, clearing one cache line at once.
61 * Data cache must be flushed to unification first to make sure
62 * the instruction cache fetches the updated data.
63 */
64 for (p = rw & -dcache_lsize; p < rw + len; p += dcache_lsize) {
65 asm volatile("dc\tcvau, %0" : : "r" (p) : "memory");
66 }
67 asm volatile("dsb\tish" : : : "memory");
68 }
69
70 /*
71 * If CTR_EL0.DIC is enabled, Instruction cache cleaning to the Point
72 * of Unification is not required for instruction to data coherence.
73 */
74 if (!(ctr_el0 & CTR_DIC)) {
75 for (p = rx & -icache_lsize; p < rx + len; p += icache_lsize) {
76 asm volatile("ic\tivau, %0" : : "r"(p) : "memory");
77 }
78 asm volatile ("dsb\tish" : : : "memory");
79 }
80
81 asm volatile("isb" : : : "memory");
82}
83#endif /* CONFIG_DARWIN */
84
084cfca1
RH
85#elif defined(__mips__)
86
87#ifdef __OpenBSD__
88#include <machine/sysarch.h>
89#else
90#include <sys/cachectl.h>
91#endif
92
1da8de39 93void flush_idcache_range(uintptr_t rx, uintptr_t rw, size_t len)
084cfca1 94{
1da8de39
RH
95 if (rx != rw) {
96 cacheflush((void *)rw, len, DCACHE);
97 }
98 cacheflush((void *)rx, len, ICACHE);
084cfca1
RH
99}
100
101#elif defined(__powerpc__)
102
1da8de39 103void flush_idcache_range(uintptr_t rx, uintptr_t rw, size_t len)
084cfca1 104{
1da8de39 105 uintptr_t p, b, e;
084cfca1
RH
106 size_t dsize = qemu_dcache_linesize;
107 size_t isize = qemu_icache_linesize;
108
1da8de39
RH
109 b = rw & ~(dsize - 1);
110 e = (rw + len + dsize - 1) & ~(dsize - 1);
111 for (p = b; p < e; p += dsize) {
084cfca1
RH
112 asm volatile ("dcbst 0,%0" : : "r"(p) : "memory");
113 }
114 asm volatile ("sync" : : : "memory");
115
1da8de39
RH
116 b = rx & ~(isize - 1);
117 e = (rx + len + isize - 1) & ~(isize - 1);
118 for (p = b; p < e; p += isize) {
084cfca1
RH
119 asm volatile ("icbi 0,%0" : : "r"(p) : "memory");
120 }
121 asm volatile ("sync" : : : "memory");
122 asm volatile ("isync" : : : "memory");
123}
124
125#elif defined(__sparc__)
126
1da8de39 127void flush_idcache_range(uintptr_t rx, uintptr_t rw, size_t len)
084cfca1 128{
1da8de39
RH
129 /* No additional data flush to the RW virtual address required. */
130 uintptr_t p, end = (rx + len + 7) & -8;
131 for (p = rx & -8; p < end; p += 8) {
084cfca1
RH
132 __asm__ __volatile__("flush\t%0" : : "r" (p));
133 }
134}
135
136#else
137
1da8de39 138void flush_idcache_range(uintptr_t rx, uintptr_t rw, size_t len)
084cfca1 139{
1da8de39
RH
140 if (rw != rx) {
141 __builtin___clear_cache((char *)rw, (char *)rw + len);
142 }
143 __builtin___clear_cache((char *)rx, (char *)rx + len);
084cfca1
RH
144}
145
146#endif