]> git.proxmox.com Git - mirror_qemu.git/blob - target/arm/machine.c
fc4a4a406442d7bb160db7146b7fa74136959846
[mirror_qemu.git] / target / arm / machine.c
1 #include "qemu/osdep.h"
2 #include "cpu.h"
3 #include "qemu/error-report.h"
4 #include "sysemu/kvm.h"
5 #include "sysemu/tcg.h"
6 #include "kvm_arm.h"
7 #include "internals.h"
8 #include "migration/cpu.h"
9
10 static bool vfp_needed(void *opaque)
11 {
12 ARMCPU *cpu = opaque;
13
14 return (arm_feature(&cpu->env, ARM_FEATURE_AARCH64)
15 ? cpu_isar_feature(aa64_fp_simd, cpu)
16 : cpu_isar_feature(aa32_vfp_simd, cpu));
17 }
18
19 static int get_fpscr(QEMUFile *f, void *opaque, size_t size,
20 const VMStateField *field)
21 {
22 ARMCPU *cpu = opaque;
23 CPUARMState *env = &cpu->env;
24 uint32_t val = qemu_get_be32(f);
25
26 vfp_set_fpscr(env, val);
27 return 0;
28 }
29
30 static int put_fpscr(QEMUFile *f, void *opaque, size_t size,
31 const VMStateField *field, JSONWriter *vmdesc)
32 {
33 ARMCPU *cpu = opaque;
34 CPUARMState *env = &cpu->env;
35
36 qemu_put_be32(f, vfp_get_fpscr(env));
37 return 0;
38 }
39
40 static const VMStateInfo vmstate_fpscr = {
41 .name = "fpscr",
42 .get = get_fpscr,
43 .put = put_fpscr,
44 };
45
46 static const VMStateDescription vmstate_vfp = {
47 .name = "cpu/vfp",
48 .version_id = 3,
49 .minimum_version_id = 3,
50 .needed = vfp_needed,
51 .fields = (VMStateField[]) {
52 /* For compatibility, store Qn out of Zn here. */
53 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[0].d, ARMCPU, 0, 2),
54 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[1].d, ARMCPU, 0, 2),
55 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[2].d, ARMCPU, 0, 2),
56 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[3].d, ARMCPU, 0, 2),
57 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[4].d, ARMCPU, 0, 2),
58 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[5].d, ARMCPU, 0, 2),
59 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[6].d, ARMCPU, 0, 2),
60 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[7].d, ARMCPU, 0, 2),
61 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[8].d, ARMCPU, 0, 2),
62 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[9].d, ARMCPU, 0, 2),
63 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[10].d, ARMCPU, 0, 2),
64 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[11].d, ARMCPU, 0, 2),
65 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[12].d, ARMCPU, 0, 2),
66 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[13].d, ARMCPU, 0, 2),
67 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[14].d, ARMCPU, 0, 2),
68 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[15].d, ARMCPU, 0, 2),
69 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[16].d, ARMCPU, 0, 2),
70 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[17].d, ARMCPU, 0, 2),
71 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[18].d, ARMCPU, 0, 2),
72 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[19].d, ARMCPU, 0, 2),
73 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[20].d, ARMCPU, 0, 2),
74 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[21].d, ARMCPU, 0, 2),
75 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[22].d, ARMCPU, 0, 2),
76 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[23].d, ARMCPU, 0, 2),
77 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[24].d, ARMCPU, 0, 2),
78 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[25].d, ARMCPU, 0, 2),
79 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[26].d, ARMCPU, 0, 2),
80 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[27].d, ARMCPU, 0, 2),
81 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[28].d, ARMCPU, 0, 2),
82 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[29].d, ARMCPU, 0, 2),
83 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[30].d, ARMCPU, 0, 2),
84 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[31].d, ARMCPU, 0, 2),
85
86 /* The xregs array is a little awkward because element 1 (FPSCR)
87 * requires a specific accessor, so we have to split it up in
88 * the vmstate:
89 */
90 VMSTATE_UINT32(env.vfp.xregs[0], ARMCPU),
91 VMSTATE_UINT32_SUB_ARRAY(env.vfp.xregs, ARMCPU, 2, 14),
92 {
93 .name = "fpscr",
94 .version_id = 0,
95 .size = sizeof(uint32_t),
96 .info = &vmstate_fpscr,
97 .flags = VMS_SINGLE,
98 .offset = 0,
99 },
100 VMSTATE_END_OF_LIST()
101 }
102 };
103
104 static bool iwmmxt_needed(void *opaque)
105 {
106 ARMCPU *cpu = opaque;
107 CPUARMState *env = &cpu->env;
108
109 return arm_feature(env, ARM_FEATURE_IWMMXT);
110 }
111
112 static const VMStateDescription vmstate_iwmmxt = {
113 .name = "cpu/iwmmxt",
114 .version_id = 1,
115 .minimum_version_id = 1,
116 .needed = iwmmxt_needed,
117 .fields = (VMStateField[]) {
118 VMSTATE_UINT64_ARRAY(env.iwmmxt.regs, ARMCPU, 16),
119 VMSTATE_UINT32_ARRAY(env.iwmmxt.cregs, ARMCPU, 16),
120 VMSTATE_END_OF_LIST()
121 }
122 };
123
124 #ifdef TARGET_AARCH64
125 /* The expression ARM_MAX_VQ - 2 is 0 for pure AArch32 build,
126 * and ARMPredicateReg is actively empty. This triggers errors
127 * in the expansion of the VMSTATE macros.
128 */
129
130 static bool sve_needed(void *opaque)
131 {
132 ARMCPU *cpu = opaque;
133
134 return cpu_isar_feature(aa64_sve, cpu);
135 }
136
137 /* The first two words of each Zreg is stored in VFP state. */
138 static const VMStateDescription vmstate_zreg_hi_reg = {
139 .name = "cpu/sve/zreg_hi",
140 .version_id = 1,
141 .minimum_version_id = 1,
142 .fields = (VMStateField[]) {
143 VMSTATE_UINT64_SUB_ARRAY(d, ARMVectorReg, 2, ARM_MAX_VQ - 2),
144 VMSTATE_END_OF_LIST()
145 }
146 };
147
148 static const VMStateDescription vmstate_preg_reg = {
149 .name = "cpu/sve/preg",
150 .version_id = 1,
151 .minimum_version_id = 1,
152 .fields = (VMStateField[]) {
153 VMSTATE_UINT64_ARRAY(p, ARMPredicateReg, 2 * ARM_MAX_VQ / 8),
154 VMSTATE_END_OF_LIST()
155 }
156 };
157
158 static const VMStateDescription vmstate_sve = {
159 .name = "cpu/sve",
160 .version_id = 1,
161 .minimum_version_id = 1,
162 .needed = sve_needed,
163 .fields = (VMStateField[]) {
164 VMSTATE_STRUCT_ARRAY(env.vfp.zregs, ARMCPU, 32, 0,
165 vmstate_zreg_hi_reg, ARMVectorReg),
166 VMSTATE_STRUCT_ARRAY(env.vfp.pregs, ARMCPU, 17, 0,
167 vmstate_preg_reg, ARMPredicateReg),
168 VMSTATE_END_OF_LIST()
169 }
170 };
171
172 static const VMStateDescription vmstate_vreg = {
173 .name = "vreg",
174 .version_id = 1,
175 .minimum_version_id = 1,
176 .fields = (VMStateField[]) {
177 VMSTATE_UINT64_ARRAY(d, ARMVectorReg, ARM_MAX_VQ * 2),
178 VMSTATE_END_OF_LIST()
179 }
180 };
181
182 static bool za_needed(void *opaque)
183 {
184 ARMCPU *cpu = opaque;
185
186 /*
187 * When ZA storage is disabled, its contents are discarded.
188 * It will be zeroed when ZA storage is re-enabled.
189 */
190 return FIELD_EX64(cpu->env.svcr, SVCR, ZA);
191 }
192
193 static const VMStateDescription vmstate_za = {
194 .name = "cpu/sme",
195 .version_id = 1,
196 .minimum_version_id = 1,
197 .needed = za_needed,
198 .fields = (VMStateField[]) {
199 VMSTATE_STRUCT_ARRAY(env.zarray, ARMCPU, ARM_MAX_VQ * 16, 0,
200 vmstate_vreg, ARMVectorReg),
201 VMSTATE_END_OF_LIST()
202 }
203 };
204 #endif /* AARCH64 */
205
206 static bool serror_needed(void *opaque)
207 {
208 ARMCPU *cpu = opaque;
209 CPUARMState *env = &cpu->env;
210
211 return env->serror.pending != 0;
212 }
213
214 static const VMStateDescription vmstate_serror = {
215 .name = "cpu/serror",
216 .version_id = 1,
217 .minimum_version_id = 1,
218 .needed = serror_needed,
219 .fields = (VMStateField[]) {
220 VMSTATE_UINT8(env.serror.pending, ARMCPU),
221 VMSTATE_UINT8(env.serror.has_esr, ARMCPU),
222 VMSTATE_UINT64(env.serror.esr, ARMCPU),
223 VMSTATE_END_OF_LIST()
224 }
225 };
226
227 static bool irq_line_state_needed(void *opaque)
228 {
229 return true;
230 }
231
232 static const VMStateDescription vmstate_irq_line_state = {
233 .name = "cpu/irq-line-state",
234 .version_id = 1,
235 .minimum_version_id = 1,
236 .needed = irq_line_state_needed,
237 .fields = (VMStateField[]) {
238 VMSTATE_UINT32(env.irq_line_state, ARMCPU),
239 VMSTATE_END_OF_LIST()
240 }
241 };
242
243 static bool m_needed(void *opaque)
244 {
245 ARMCPU *cpu = opaque;
246 CPUARMState *env = &cpu->env;
247
248 return arm_feature(env, ARM_FEATURE_M);
249 }
250
251 static const VMStateDescription vmstate_m_faultmask_primask = {
252 .name = "cpu/m/faultmask-primask",
253 .version_id = 1,
254 .minimum_version_id = 1,
255 .needed = m_needed,
256 .fields = (VMStateField[]) {
257 VMSTATE_UINT32(env.v7m.faultmask[M_REG_NS], ARMCPU),
258 VMSTATE_UINT32(env.v7m.primask[M_REG_NS], ARMCPU),
259 VMSTATE_END_OF_LIST()
260 }
261 };
262
263 /* CSSELR is in a subsection because we didn't implement it previously.
264 * Migration from an old implementation will leave it at zero, which
265 * is OK since the only CPUs in the old implementation make the
266 * register RAZ/WI.
267 * Since there was no version of QEMU which implemented the CSSELR for
268 * just non-secure, we transfer both banks here rather than putting
269 * the secure banked version in the m-security subsection.
270 */
271 static bool csselr_vmstate_validate(void *opaque, int version_id)
272 {
273 ARMCPU *cpu = opaque;
274
275 return cpu->env.v7m.csselr[M_REG_NS] <= R_V7M_CSSELR_INDEX_MASK
276 && cpu->env.v7m.csselr[M_REG_S] <= R_V7M_CSSELR_INDEX_MASK;
277 }
278
279 static bool m_csselr_needed(void *opaque)
280 {
281 ARMCPU *cpu = opaque;
282
283 return !arm_v7m_csselr_razwi(cpu);
284 }
285
286 static const VMStateDescription vmstate_m_csselr = {
287 .name = "cpu/m/csselr",
288 .version_id = 1,
289 .minimum_version_id = 1,
290 .needed = m_csselr_needed,
291 .fields = (VMStateField[]) {
292 VMSTATE_UINT32_ARRAY(env.v7m.csselr, ARMCPU, M_REG_NUM_BANKS),
293 VMSTATE_VALIDATE("CSSELR is valid", csselr_vmstate_validate),
294 VMSTATE_END_OF_LIST()
295 }
296 };
297
298 static const VMStateDescription vmstate_m_scr = {
299 .name = "cpu/m/scr",
300 .version_id = 1,
301 .minimum_version_id = 1,
302 .needed = m_needed,
303 .fields = (VMStateField[]) {
304 VMSTATE_UINT32(env.v7m.scr[M_REG_NS], ARMCPU),
305 VMSTATE_END_OF_LIST()
306 }
307 };
308
309 static const VMStateDescription vmstate_m_other_sp = {
310 .name = "cpu/m/other-sp",
311 .version_id = 1,
312 .minimum_version_id = 1,
313 .needed = m_needed,
314 .fields = (VMStateField[]) {
315 VMSTATE_UINT32(env.v7m.other_sp, ARMCPU),
316 VMSTATE_END_OF_LIST()
317 }
318 };
319
320 static bool m_v8m_needed(void *opaque)
321 {
322 ARMCPU *cpu = opaque;
323 CPUARMState *env = &cpu->env;
324
325 return arm_feature(env, ARM_FEATURE_M) && arm_feature(env, ARM_FEATURE_V8);
326 }
327
328 static const VMStateDescription vmstate_m_v8m = {
329 .name = "cpu/m/v8m",
330 .version_id = 1,
331 .minimum_version_id = 1,
332 .needed = m_v8m_needed,
333 .fields = (VMStateField[]) {
334 VMSTATE_UINT32_ARRAY(env.v7m.msplim, ARMCPU, M_REG_NUM_BANKS),
335 VMSTATE_UINT32_ARRAY(env.v7m.psplim, ARMCPU, M_REG_NUM_BANKS),
336 VMSTATE_END_OF_LIST()
337 }
338 };
339
340 static const VMStateDescription vmstate_m_fp = {
341 .name = "cpu/m/fp",
342 .version_id = 1,
343 .minimum_version_id = 1,
344 .needed = vfp_needed,
345 .fields = (VMStateField[]) {
346 VMSTATE_UINT32_ARRAY(env.v7m.fpcar, ARMCPU, M_REG_NUM_BANKS),
347 VMSTATE_UINT32_ARRAY(env.v7m.fpccr, ARMCPU, M_REG_NUM_BANKS),
348 VMSTATE_UINT32_ARRAY(env.v7m.fpdscr, ARMCPU, M_REG_NUM_BANKS),
349 VMSTATE_UINT32_ARRAY(env.v7m.cpacr, ARMCPU, M_REG_NUM_BANKS),
350 VMSTATE_UINT32(env.v7m.nsacr, ARMCPU),
351 VMSTATE_END_OF_LIST()
352 }
353 };
354
355 static bool mve_needed(void *opaque)
356 {
357 ARMCPU *cpu = opaque;
358
359 return cpu_isar_feature(aa32_mve, cpu);
360 }
361
362 static const VMStateDescription vmstate_m_mve = {
363 .name = "cpu/m/mve",
364 .version_id = 1,
365 .minimum_version_id = 1,
366 .needed = mve_needed,
367 .fields = (VMStateField[]) {
368 VMSTATE_UINT32(env.v7m.vpr, ARMCPU),
369 VMSTATE_UINT32(env.v7m.ltpsize, ARMCPU),
370 VMSTATE_END_OF_LIST()
371 },
372 };
373
374 static const VMStateDescription vmstate_m = {
375 .name = "cpu/m",
376 .version_id = 4,
377 .minimum_version_id = 4,
378 .needed = m_needed,
379 .fields = (VMStateField[]) {
380 VMSTATE_UINT32(env.v7m.vecbase[M_REG_NS], ARMCPU),
381 VMSTATE_UINT32(env.v7m.basepri[M_REG_NS], ARMCPU),
382 VMSTATE_UINT32(env.v7m.control[M_REG_NS], ARMCPU),
383 VMSTATE_UINT32(env.v7m.ccr[M_REG_NS], ARMCPU),
384 VMSTATE_UINT32(env.v7m.cfsr[M_REG_NS], ARMCPU),
385 VMSTATE_UINT32(env.v7m.hfsr, ARMCPU),
386 VMSTATE_UINT32(env.v7m.dfsr, ARMCPU),
387 VMSTATE_UINT32(env.v7m.mmfar[M_REG_NS], ARMCPU),
388 VMSTATE_UINT32(env.v7m.bfar, ARMCPU),
389 VMSTATE_UINT32(env.v7m.mpu_ctrl[M_REG_NS], ARMCPU),
390 VMSTATE_INT32(env.v7m.exception, ARMCPU),
391 VMSTATE_END_OF_LIST()
392 },
393 .subsections = (const VMStateDescription*[]) {
394 &vmstate_m_faultmask_primask,
395 &vmstate_m_csselr,
396 &vmstate_m_scr,
397 &vmstate_m_other_sp,
398 &vmstate_m_v8m,
399 &vmstate_m_fp,
400 &vmstate_m_mve,
401 NULL
402 }
403 };
404
405 static bool thumb2ee_needed(void *opaque)
406 {
407 ARMCPU *cpu = opaque;
408 CPUARMState *env = &cpu->env;
409
410 return arm_feature(env, ARM_FEATURE_THUMB2EE);
411 }
412
413 static const VMStateDescription vmstate_thumb2ee = {
414 .name = "cpu/thumb2ee",
415 .version_id = 1,
416 .minimum_version_id = 1,
417 .needed = thumb2ee_needed,
418 .fields = (VMStateField[]) {
419 VMSTATE_UINT32(env.teecr, ARMCPU),
420 VMSTATE_UINT32(env.teehbr, ARMCPU),
421 VMSTATE_END_OF_LIST()
422 }
423 };
424
425 static bool pmsav7_needed(void *opaque)
426 {
427 ARMCPU *cpu = opaque;
428 CPUARMState *env = &cpu->env;
429
430 return arm_feature(env, ARM_FEATURE_PMSA) &&
431 arm_feature(env, ARM_FEATURE_V7) &&
432 !arm_feature(env, ARM_FEATURE_V8);
433 }
434
435 static bool pmsav7_rgnr_vmstate_validate(void *opaque, int version_id)
436 {
437 ARMCPU *cpu = opaque;
438
439 return cpu->env.pmsav7.rnr[M_REG_NS] < cpu->pmsav7_dregion;
440 }
441
442 static const VMStateDescription vmstate_pmsav7 = {
443 .name = "cpu/pmsav7",
444 .version_id = 1,
445 .minimum_version_id = 1,
446 .needed = pmsav7_needed,
447 .fields = (VMStateField[]) {
448 VMSTATE_VARRAY_UINT32(env.pmsav7.drbar, ARMCPU, pmsav7_dregion, 0,
449 vmstate_info_uint32, uint32_t),
450 VMSTATE_VARRAY_UINT32(env.pmsav7.drsr, ARMCPU, pmsav7_dregion, 0,
451 vmstate_info_uint32, uint32_t),
452 VMSTATE_VARRAY_UINT32(env.pmsav7.dracr, ARMCPU, pmsav7_dregion, 0,
453 vmstate_info_uint32, uint32_t),
454 VMSTATE_VALIDATE("rgnr is valid", pmsav7_rgnr_vmstate_validate),
455 VMSTATE_END_OF_LIST()
456 }
457 };
458
459 static bool pmsav7_rnr_needed(void *opaque)
460 {
461 ARMCPU *cpu = opaque;
462 CPUARMState *env = &cpu->env;
463
464 /* For R profile cores pmsav7.rnr is migrated via the cpreg
465 * "RGNR" definition in helper.h. For M profile we have to
466 * migrate it separately.
467 */
468 return arm_feature(env, ARM_FEATURE_M);
469 }
470
471 static const VMStateDescription vmstate_pmsav7_rnr = {
472 .name = "cpu/pmsav7-rnr",
473 .version_id = 1,
474 .minimum_version_id = 1,
475 .needed = pmsav7_rnr_needed,
476 .fields = (VMStateField[]) {
477 VMSTATE_UINT32(env.pmsav7.rnr[M_REG_NS], ARMCPU),
478 VMSTATE_END_OF_LIST()
479 }
480 };
481
482 static bool pmsav8_needed(void *opaque)
483 {
484 ARMCPU *cpu = opaque;
485 CPUARMState *env = &cpu->env;
486
487 return arm_feature(env, ARM_FEATURE_PMSA) &&
488 arm_feature(env, ARM_FEATURE_V8);
489 }
490
491 static bool pmsav8r_needed(void *opaque)
492 {
493 ARMCPU *cpu = opaque;
494 CPUARMState *env = &cpu->env;
495
496 return arm_feature(env, ARM_FEATURE_PMSA) &&
497 arm_feature(env, ARM_FEATURE_V8) &&
498 !arm_feature(env, ARM_FEATURE_M);
499 }
500
501 static const VMStateDescription vmstate_pmsav8r = {
502 .name = "cpu/pmsav8/pmsav8r",
503 .version_id = 1,
504 .minimum_version_id = 1,
505 .needed = pmsav8r_needed,
506 .fields = (VMStateField[]) {
507 VMSTATE_VARRAY_UINT32(env.pmsav8.hprbar, ARMCPU,
508 pmsav8r_hdregion, 0, vmstate_info_uint32, uint32_t),
509 VMSTATE_VARRAY_UINT32(env.pmsav8.hprlar, ARMCPU,
510 pmsav8r_hdregion, 0, vmstate_info_uint32, uint32_t),
511 VMSTATE_END_OF_LIST()
512 },
513 };
514
515 static const VMStateDescription vmstate_pmsav8 = {
516 .name = "cpu/pmsav8",
517 .version_id = 1,
518 .minimum_version_id = 1,
519 .needed = pmsav8_needed,
520 .fields = (VMStateField[]) {
521 VMSTATE_VARRAY_UINT32(env.pmsav8.rbar[M_REG_NS], ARMCPU, pmsav7_dregion,
522 0, vmstate_info_uint32, uint32_t),
523 VMSTATE_VARRAY_UINT32(env.pmsav8.rlar[M_REG_NS], ARMCPU, pmsav7_dregion,
524 0, vmstate_info_uint32, uint32_t),
525 VMSTATE_UINT32(env.pmsav8.mair0[M_REG_NS], ARMCPU),
526 VMSTATE_UINT32(env.pmsav8.mair1[M_REG_NS], ARMCPU),
527 VMSTATE_END_OF_LIST()
528 },
529 .subsections = (const VMStateDescription * []) {
530 &vmstate_pmsav8r,
531 NULL
532 }
533 };
534
535 static bool s_rnr_vmstate_validate(void *opaque, int version_id)
536 {
537 ARMCPU *cpu = opaque;
538
539 return cpu->env.pmsav7.rnr[M_REG_S] < cpu->pmsav7_dregion;
540 }
541
542 static bool sau_rnr_vmstate_validate(void *opaque, int version_id)
543 {
544 ARMCPU *cpu = opaque;
545
546 return cpu->env.sau.rnr < cpu->sau_sregion;
547 }
548
549 static bool m_security_needed(void *opaque)
550 {
551 ARMCPU *cpu = opaque;
552 CPUARMState *env = &cpu->env;
553
554 return arm_feature(env, ARM_FEATURE_M_SECURITY);
555 }
556
557 static const VMStateDescription vmstate_m_security = {
558 .name = "cpu/m-security",
559 .version_id = 1,
560 .minimum_version_id = 1,
561 .needed = m_security_needed,
562 .fields = (VMStateField[]) {
563 VMSTATE_UINT32(env.v7m.secure, ARMCPU),
564 VMSTATE_UINT32(env.v7m.other_ss_msp, ARMCPU),
565 VMSTATE_UINT32(env.v7m.other_ss_psp, ARMCPU),
566 VMSTATE_UINT32(env.v7m.basepri[M_REG_S], ARMCPU),
567 VMSTATE_UINT32(env.v7m.primask[M_REG_S], ARMCPU),
568 VMSTATE_UINT32(env.v7m.faultmask[M_REG_S], ARMCPU),
569 VMSTATE_UINT32(env.v7m.control[M_REG_S], ARMCPU),
570 VMSTATE_UINT32(env.v7m.vecbase[M_REG_S], ARMCPU),
571 VMSTATE_UINT32(env.pmsav8.mair0[M_REG_S], ARMCPU),
572 VMSTATE_UINT32(env.pmsav8.mair1[M_REG_S], ARMCPU),
573 VMSTATE_VARRAY_UINT32(env.pmsav8.rbar[M_REG_S], ARMCPU, pmsav7_dregion,
574 0, vmstate_info_uint32, uint32_t),
575 VMSTATE_VARRAY_UINT32(env.pmsav8.rlar[M_REG_S], ARMCPU, pmsav7_dregion,
576 0, vmstate_info_uint32, uint32_t),
577 VMSTATE_UINT32(env.pmsav7.rnr[M_REG_S], ARMCPU),
578 VMSTATE_VALIDATE("secure MPU_RNR is valid", s_rnr_vmstate_validate),
579 VMSTATE_UINT32(env.v7m.mpu_ctrl[M_REG_S], ARMCPU),
580 VMSTATE_UINT32(env.v7m.ccr[M_REG_S], ARMCPU),
581 VMSTATE_UINT32(env.v7m.mmfar[M_REG_S], ARMCPU),
582 VMSTATE_UINT32(env.v7m.cfsr[M_REG_S], ARMCPU),
583 VMSTATE_UINT32(env.v7m.sfsr, ARMCPU),
584 VMSTATE_UINT32(env.v7m.sfar, ARMCPU),
585 VMSTATE_VARRAY_UINT32(env.sau.rbar, ARMCPU, sau_sregion, 0,
586 vmstate_info_uint32, uint32_t),
587 VMSTATE_VARRAY_UINT32(env.sau.rlar, ARMCPU, sau_sregion, 0,
588 vmstate_info_uint32, uint32_t),
589 VMSTATE_UINT32(env.sau.rnr, ARMCPU),
590 VMSTATE_VALIDATE("SAU_RNR is valid", sau_rnr_vmstate_validate),
591 VMSTATE_UINT32(env.sau.ctrl, ARMCPU),
592 VMSTATE_UINT32(env.v7m.scr[M_REG_S], ARMCPU),
593 /* AIRCR is not secure-only, but our implementation is R/O if the
594 * security extension is unimplemented, so we migrate it here.
595 */
596 VMSTATE_UINT32(env.v7m.aircr, ARMCPU),
597 VMSTATE_END_OF_LIST()
598 }
599 };
600
601 static int get_cpsr(QEMUFile *f, void *opaque, size_t size,
602 const VMStateField *field)
603 {
604 ARMCPU *cpu = opaque;
605 CPUARMState *env = &cpu->env;
606 uint32_t val = qemu_get_be32(f);
607
608 if (arm_feature(env, ARM_FEATURE_M)) {
609 if (val & XPSR_EXCP) {
610 /* This is a CPSR format value from an older QEMU. (We can tell
611 * because values transferred in XPSR format always have zero
612 * for the EXCP field, and CPSR format will always have bit 4
613 * set in CPSR_M.) Rearrange it into XPSR format. The significant
614 * differences are that the T bit is not in the same place, the
615 * primask/faultmask info may be in the CPSR I and F bits, and
616 * we do not want the mode bits.
617 * We know that this cleanup happened before v8M, so there
618 * is no complication with banked primask/faultmask.
619 */
620 uint32_t newval = val;
621
622 assert(!arm_feature(env, ARM_FEATURE_M_SECURITY));
623
624 newval &= (CPSR_NZCV | CPSR_Q | CPSR_IT | CPSR_GE);
625 if (val & CPSR_T) {
626 newval |= XPSR_T;
627 }
628 /* If the I or F bits are set then this is a migration from
629 * an old QEMU which still stored the M profile FAULTMASK
630 * and PRIMASK in env->daif. For a new QEMU, the data is
631 * transferred using the vmstate_m_faultmask_primask subsection.
632 */
633 if (val & CPSR_F) {
634 env->v7m.faultmask[M_REG_NS] = 1;
635 }
636 if (val & CPSR_I) {
637 env->v7m.primask[M_REG_NS] = 1;
638 }
639 val = newval;
640 }
641 /* Ignore the low bits, they are handled by vmstate_m. */
642 xpsr_write(env, val, ~XPSR_EXCP);
643 return 0;
644 }
645
646 env->aarch64 = ((val & PSTATE_nRW) == 0);
647
648 if (is_a64(env)) {
649 pstate_write(env, val);
650 return 0;
651 }
652
653 cpsr_write(env, val, 0xffffffff, CPSRWriteRaw);
654 return 0;
655 }
656
657 static int put_cpsr(QEMUFile *f, void *opaque, size_t size,
658 const VMStateField *field, JSONWriter *vmdesc)
659 {
660 ARMCPU *cpu = opaque;
661 CPUARMState *env = &cpu->env;
662 uint32_t val;
663
664 if (arm_feature(env, ARM_FEATURE_M)) {
665 /* The low 9 bits are v7m.exception, which is handled by vmstate_m. */
666 val = xpsr_read(env) & ~XPSR_EXCP;
667 } else if (is_a64(env)) {
668 val = pstate_read(env);
669 } else {
670 val = cpsr_read(env);
671 }
672
673 qemu_put_be32(f, val);
674 return 0;
675 }
676
677 static const VMStateInfo vmstate_cpsr = {
678 .name = "cpsr",
679 .get = get_cpsr,
680 .put = put_cpsr,
681 };
682
683 static int get_power(QEMUFile *f, void *opaque, size_t size,
684 const VMStateField *field)
685 {
686 ARMCPU *cpu = opaque;
687 bool powered_off = qemu_get_byte(f);
688 cpu->power_state = powered_off ? PSCI_OFF : PSCI_ON;
689 return 0;
690 }
691
692 static int put_power(QEMUFile *f, void *opaque, size_t size,
693 const VMStateField *field, JSONWriter *vmdesc)
694 {
695 ARMCPU *cpu = opaque;
696
697 /* Migration should never happen while we transition power states */
698
699 if (cpu->power_state == PSCI_ON ||
700 cpu->power_state == PSCI_OFF) {
701 bool powered_off = (cpu->power_state == PSCI_OFF) ? true : false;
702 qemu_put_byte(f, powered_off);
703 return 0;
704 } else {
705 return 1;
706 }
707 }
708
709 static const VMStateInfo vmstate_powered_off = {
710 .name = "powered_off",
711 .get = get_power,
712 .put = put_power,
713 };
714
715 static int cpu_pre_save(void *opaque)
716 {
717 ARMCPU *cpu = opaque;
718
719 if (!kvm_enabled()) {
720 pmu_op_start(&cpu->env);
721 }
722
723 if (kvm_enabled()) {
724 if (!write_kvmstate_to_list(cpu)) {
725 /* This should never fail */
726 g_assert_not_reached();
727 }
728
729 /*
730 * kvm_arm_cpu_pre_save() must be called after
731 * write_kvmstate_to_list()
732 */
733 kvm_arm_cpu_pre_save(cpu);
734 } else {
735 if (!write_cpustate_to_list(cpu, false)) {
736 /* This should never fail. */
737 g_assert_not_reached();
738 }
739 }
740
741 cpu->cpreg_vmstate_array_len = cpu->cpreg_array_len;
742 memcpy(cpu->cpreg_vmstate_indexes, cpu->cpreg_indexes,
743 cpu->cpreg_array_len * sizeof(uint64_t));
744 memcpy(cpu->cpreg_vmstate_values, cpu->cpreg_values,
745 cpu->cpreg_array_len * sizeof(uint64_t));
746
747 return 0;
748 }
749
750 static int cpu_post_save(void *opaque)
751 {
752 ARMCPU *cpu = opaque;
753
754 if (!kvm_enabled()) {
755 pmu_op_finish(&cpu->env);
756 }
757
758 return 0;
759 }
760
761 static int cpu_pre_load(void *opaque)
762 {
763 ARMCPU *cpu = opaque;
764 CPUARMState *env = &cpu->env;
765
766 /*
767 * Pre-initialize irq_line_state to a value that's never valid as
768 * real data, so cpu_post_load() can tell whether we've seen the
769 * irq-line-state subsection in the incoming migration state.
770 */
771 env->irq_line_state = UINT32_MAX;
772
773 if (!kvm_enabled()) {
774 pmu_op_start(&cpu->env);
775 }
776
777 return 0;
778 }
779
780 static int cpu_post_load(void *opaque, int version_id)
781 {
782 ARMCPU *cpu = opaque;
783 CPUARMState *env = &cpu->env;
784 int i, v;
785
786 /*
787 * Handle migration compatibility from old QEMU which didn't
788 * send the irq-line-state subsection. A QEMU without it did not
789 * implement the HCR_EL2.{VI,VF} bits as generating interrupts,
790 * so for TCG the line state matches the bits set in cs->interrupt_request.
791 * For KVM the line state is not stored in cs->interrupt_request
792 * and so this will leave irq_line_state as 0, but this is OK because
793 * we only need to care about it for TCG.
794 */
795 if (env->irq_line_state == UINT32_MAX) {
796 CPUState *cs = CPU(cpu);
797
798 env->irq_line_state = cs->interrupt_request &
799 (CPU_INTERRUPT_HARD | CPU_INTERRUPT_FIQ |
800 CPU_INTERRUPT_VIRQ | CPU_INTERRUPT_VFIQ);
801 }
802
803 /* Update the values list from the incoming migration data.
804 * Anything in the incoming data which we don't know about is
805 * a migration failure; anything we know about but the incoming
806 * data doesn't specify retains its current (reset) value.
807 * The indexes list remains untouched -- we only inspect the
808 * incoming migration index list so we can match the values array
809 * entries with the right slots in our own values array.
810 */
811
812 for (i = 0, v = 0; i < cpu->cpreg_array_len
813 && v < cpu->cpreg_vmstate_array_len; i++) {
814 if (cpu->cpreg_vmstate_indexes[v] > cpu->cpreg_indexes[i]) {
815 /* register in our list but not incoming : skip it */
816 continue;
817 }
818 if (cpu->cpreg_vmstate_indexes[v] < cpu->cpreg_indexes[i]) {
819 /* register in their list but not ours: fail migration */
820 return -1;
821 }
822 /* matching register, copy the value over */
823 cpu->cpreg_values[i] = cpu->cpreg_vmstate_values[v];
824 v++;
825 }
826
827 if (kvm_enabled()) {
828 if (!write_list_to_kvmstate(cpu, KVM_PUT_FULL_STATE)) {
829 return -1;
830 }
831 /* Note that it's OK for the TCG side not to know about
832 * every register in the list; KVM is authoritative if
833 * we're using it.
834 */
835 write_list_to_cpustate(cpu);
836 kvm_arm_cpu_post_load(cpu);
837 } else {
838 if (!write_list_to_cpustate(cpu)) {
839 return -1;
840 }
841 }
842
843 /*
844 * Misaligned thumb pc is architecturally impossible. Fail the
845 * incoming migration. For TCG it would trigger the assert in
846 * thumb_tr_translate_insn().
847 */
848 if (!is_a64(env) && env->thumb && (env->regs[15] & 1)) {
849 return -1;
850 }
851
852 if (tcg_enabled()) {
853 hw_breakpoint_update_all(cpu);
854 hw_watchpoint_update_all(cpu);
855 }
856
857 /*
858 * TCG gen_update_fp_context() relies on the invariant that
859 * FPDSCR.LTPSIZE is constant 4 for M-profile with the LOB extension;
860 * forbid bogus incoming data with some other value.
861 */
862 if (arm_feature(env, ARM_FEATURE_M) && cpu_isar_feature(aa32_lob, cpu)) {
863 if (extract32(env->v7m.fpdscr[M_REG_NS],
864 FPCR_LTPSIZE_SHIFT, FPCR_LTPSIZE_LENGTH) != 4 ||
865 extract32(env->v7m.fpdscr[M_REG_S],
866 FPCR_LTPSIZE_SHIFT, FPCR_LTPSIZE_LENGTH) != 4) {
867 return -1;
868 }
869 }
870
871 if (!kvm_enabled()) {
872 pmu_op_finish(&cpu->env);
873 }
874
875 if (tcg_enabled()) {
876 arm_rebuild_hflags(&cpu->env);
877 }
878
879 return 0;
880 }
881
882 const VMStateDescription vmstate_arm_cpu = {
883 .name = "cpu",
884 .version_id = 22,
885 .minimum_version_id = 22,
886 .pre_save = cpu_pre_save,
887 .post_save = cpu_post_save,
888 .pre_load = cpu_pre_load,
889 .post_load = cpu_post_load,
890 .fields = (VMStateField[]) {
891 VMSTATE_UINT32_ARRAY(env.regs, ARMCPU, 16),
892 VMSTATE_UINT64_ARRAY(env.xregs, ARMCPU, 32),
893 VMSTATE_UINT64(env.pc, ARMCPU),
894 {
895 .name = "cpsr",
896 .version_id = 0,
897 .size = sizeof(uint32_t),
898 .info = &vmstate_cpsr,
899 .flags = VMS_SINGLE,
900 .offset = 0,
901 },
902 VMSTATE_UINT32(env.spsr, ARMCPU),
903 VMSTATE_UINT64_ARRAY(env.banked_spsr, ARMCPU, 8),
904 VMSTATE_UINT32_ARRAY(env.banked_r13, ARMCPU, 8),
905 VMSTATE_UINT32_ARRAY(env.banked_r14, ARMCPU, 8),
906 VMSTATE_UINT32_ARRAY(env.usr_regs, ARMCPU, 5),
907 VMSTATE_UINT32_ARRAY(env.fiq_regs, ARMCPU, 5),
908 VMSTATE_UINT64_ARRAY(env.elr_el, ARMCPU, 4),
909 VMSTATE_UINT64_ARRAY(env.sp_el, ARMCPU, 4),
910 /* The length-check must come before the arrays to avoid
911 * incoming data possibly overflowing the array.
912 */
913 VMSTATE_INT32_POSITIVE_LE(cpreg_vmstate_array_len, ARMCPU),
914 VMSTATE_VARRAY_INT32(cpreg_vmstate_indexes, ARMCPU,
915 cpreg_vmstate_array_len,
916 0, vmstate_info_uint64, uint64_t),
917 VMSTATE_VARRAY_INT32(cpreg_vmstate_values, ARMCPU,
918 cpreg_vmstate_array_len,
919 0, vmstate_info_uint64, uint64_t),
920 VMSTATE_UINT64(env.exclusive_addr, ARMCPU),
921 VMSTATE_UINT64(env.exclusive_val, ARMCPU),
922 VMSTATE_UINT64(env.exclusive_high, ARMCPU),
923 VMSTATE_UNUSED(sizeof(uint64_t)),
924 VMSTATE_UINT32(env.exception.syndrome, ARMCPU),
925 VMSTATE_UINT32(env.exception.fsr, ARMCPU),
926 VMSTATE_UINT64(env.exception.vaddress, ARMCPU),
927 VMSTATE_TIMER_PTR(gt_timer[GTIMER_PHYS], ARMCPU),
928 VMSTATE_TIMER_PTR(gt_timer[GTIMER_VIRT], ARMCPU),
929 {
930 .name = "power_state",
931 .version_id = 0,
932 .size = sizeof(bool),
933 .info = &vmstate_powered_off,
934 .flags = VMS_SINGLE,
935 .offset = 0,
936 },
937 VMSTATE_END_OF_LIST()
938 },
939 .subsections = (const VMStateDescription*[]) {
940 &vmstate_vfp,
941 &vmstate_iwmmxt,
942 &vmstate_m,
943 &vmstate_thumb2ee,
944 /* pmsav7_rnr must come before pmsav7 so that we have the
945 * region number before we test it in the VMSTATE_VALIDATE
946 * in vmstate_pmsav7.
947 */
948 &vmstate_pmsav7_rnr,
949 &vmstate_pmsav7,
950 &vmstate_pmsav8,
951 &vmstate_m_security,
952 #ifdef TARGET_AARCH64
953 &vmstate_sve,
954 &vmstate_za,
955 #endif
956 &vmstate_serror,
957 &vmstate_irq_line_state,
958 NULL
959 }
960 };