]> git.proxmox.com Git - mirror_qemu.git/blob - target/arm/machine.c
Merge tag 'for-upstream' of https://repo.or.cz/qemu/kevin into staging
[mirror_qemu.git] / target / arm / machine.c
1 #include "qemu/osdep.h"
2 #include "cpu.h"
3 #include "qemu/error-report.h"
4 #include "sysemu/kvm.h"
5 #include "kvm_arm.h"
6 #include "internals.h"
7 #include "migration/cpu.h"
8
9 static bool vfp_needed(void *opaque)
10 {
11 ARMCPU *cpu = opaque;
12
13 return (arm_feature(&cpu->env, ARM_FEATURE_AARCH64)
14 ? cpu_isar_feature(aa64_fp_simd, cpu)
15 : cpu_isar_feature(aa32_vfp_simd, cpu));
16 }
17
18 static int get_fpscr(QEMUFile *f, void *opaque, size_t size,
19 const VMStateField *field)
20 {
21 ARMCPU *cpu = opaque;
22 CPUARMState *env = &cpu->env;
23 uint32_t val = qemu_get_be32(f);
24
25 vfp_set_fpscr(env, val);
26 return 0;
27 }
28
29 static int put_fpscr(QEMUFile *f, void *opaque, size_t size,
30 const VMStateField *field, JSONWriter *vmdesc)
31 {
32 ARMCPU *cpu = opaque;
33 CPUARMState *env = &cpu->env;
34
35 qemu_put_be32(f, vfp_get_fpscr(env));
36 return 0;
37 }
38
39 static const VMStateInfo vmstate_fpscr = {
40 .name = "fpscr",
41 .get = get_fpscr,
42 .put = put_fpscr,
43 };
44
45 static const VMStateDescription vmstate_vfp = {
46 .name = "cpu/vfp",
47 .version_id = 3,
48 .minimum_version_id = 3,
49 .needed = vfp_needed,
50 .fields = (VMStateField[]) {
51 /* For compatibility, store Qn out of Zn here. */
52 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[0].d, ARMCPU, 0, 2),
53 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[1].d, ARMCPU, 0, 2),
54 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[2].d, ARMCPU, 0, 2),
55 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[3].d, ARMCPU, 0, 2),
56 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[4].d, ARMCPU, 0, 2),
57 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[5].d, ARMCPU, 0, 2),
58 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[6].d, ARMCPU, 0, 2),
59 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[7].d, ARMCPU, 0, 2),
60 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[8].d, ARMCPU, 0, 2),
61 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[9].d, ARMCPU, 0, 2),
62 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[10].d, ARMCPU, 0, 2),
63 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[11].d, ARMCPU, 0, 2),
64 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[12].d, ARMCPU, 0, 2),
65 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[13].d, ARMCPU, 0, 2),
66 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[14].d, ARMCPU, 0, 2),
67 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[15].d, ARMCPU, 0, 2),
68 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[16].d, ARMCPU, 0, 2),
69 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[17].d, ARMCPU, 0, 2),
70 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[18].d, ARMCPU, 0, 2),
71 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[19].d, ARMCPU, 0, 2),
72 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[20].d, ARMCPU, 0, 2),
73 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[21].d, ARMCPU, 0, 2),
74 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[22].d, ARMCPU, 0, 2),
75 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[23].d, ARMCPU, 0, 2),
76 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[24].d, ARMCPU, 0, 2),
77 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[25].d, ARMCPU, 0, 2),
78 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[26].d, ARMCPU, 0, 2),
79 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[27].d, ARMCPU, 0, 2),
80 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[28].d, ARMCPU, 0, 2),
81 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[29].d, ARMCPU, 0, 2),
82 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[30].d, ARMCPU, 0, 2),
83 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[31].d, ARMCPU, 0, 2),
84
85 /* The xregs array is a little awkward because element 1 (FPSCR)
86 * requires a specific accessor, so we have to split it up in
87 * the vmstate:
88 */
89 VMSTATE_UINT32(env.vfp.xregs[0], ARMCPU),
90 VMSTATE_UINT32_SUB_ARRAY(env.vfp.xregs, ARMCPU, 2, 14),
91 {
92 .name = "fpscr",
93 .version_id = 0,
94 .size = sizeof(uint32_t),
95 .info = &vmstate_fpscr,
96 .flags = VMS_SINGLE,
97 .offset = 0,
98 },
99 VMSTATE_END_OF_LIST()
100 }
101 };
102
103 static bool iwmmxt_needed(void *opaque)
104 {
105 ARMCPU *cpu = opaque;
106 CPUARMState *env = &cpu->env;
107
108 return arm_feature(env, ARM_FEATURE_IWMMXT);
109 }
110
111 static const VMStateDescription vmstate_iwmmxt = {
112 .name = "cpu/iwmmxt",
113 .version_id = 1,
114 .minimum_version_id = 1,
115 .needed = iwmmxt_needed,
116 .fields = (VMStateField[]) {
117 VMSTATE_UINT64_ARRAY(env.iwmmxt.regs, ARMCPU, 16),
118 VMSTATE_UINT32_ARRAY(env.iwmmxt.cregs, ARMCPU, 16),
119 VMSTATE_END_OF_LIST()
120 }
121 };
122
123 #ifdef TARGET_AARCH64
124 /* The expression ARM_MAX_VQ - 2 is 0 for pure AArch32 build,
125 * and ARMPredicateReg is actively empty. This triggers errors
126 * in the expansion of the VMSTATE macros.
127 */
128
129 static bool sve_needed(void *opaque)
130 {
131 ARMCPU *cpu = opaque;
132
133 return cpu_isar_feature(aa64_sve, cpu);
134 }
135
136 /* The first two words of each Zreg is stored in VFP state. */
137 static const VMStateDescription vmstate_zreg_hi_reg = {
138 .name = "cpu/sve/zreg_hi",
139 .version_id = 1,
140 .minimum_version_id = 1,
141 .fields = (VMStateField[]) {
142 VMSTATE_UINT64_SUB_ARRAY(d, ARMVectorReg, 2, ARM_MAX_VQ - 2),
143 VMSTATE_END_OF_LIST()
144 }
145 };
146
147 static const VMStateDescription vmstate_preg_reg = {
148 .name = "cpu/sve/preg",
149 .version_id = 1,
150 .minimum_version_id = 1,
151 .fields = (VMStateField[]) {
152 VMSTATE_UINT64_ARRAY(p, ARMPredicateReg, 2 * ARM_MAX_VQ / 8),
153 VMSTATE_END_OF_LIST()
154 }
155 };
156
157 static const VMStateDescription vmstate_sve = {
158 .name = "cpu/sve",
159 .version_id = 1,
160 .minimum_version_id = 1,
161 .needed = sve_needed,
162 .fields = (VMStateField[]) {
163 VMSTATE_STRUCT_ARRAY(env.vfp.zregs, ARMCPU, 32, 0,
164 vmstate_zreg_hi_reg, ARMVectorReg),
165 VMSTATE_STRUCT_ARRAY(env.vfp.pregs, ARMCPU, 17, 0,
166 vmstate_preg_reg, ARMPredicateReg),
167 VMSTATE_END_OF_LIST()
168 }
169 };
170
171 static const VMStateDescription vmstate_vreg = {
172 .name = "vreg",
173 .version_id = 1,
174 .minimum_version_id = 1,
175 .fields = (VMStateField[]) {
176 VMSTATE_UINT64_ARRAY(d, ARMVectorReg, ARM_MAX_VQ * 2),
177 VMSTATE_END_OF_LIST()
178 }
179 };
180
181 static bool za_needed(void *opaque)
182 {
183 ARMCPU *cpu = opaque;
184
185 /*
186 * When ZA storage is disabled, its contents are discarded.
187 * It will be zeroed when ZA storage is re-enabled.
188 */
189 return FIELD_EX64(cpu->env.svcr, SVCR, ZA);
190 }
191
192 static const VMStateDescription vmstate_za = {
193 .name = "cpu/sme",
194 .version_id = 1,
195 .minimum_version_id = 1,
196 .needed = za_needed,
197 .fields = (VMStateField[]) {
198 VMSTATE_STRUCT_ARRAY(env.zarray, ARMCPU, ARM_MAX_VQ * 16, 0,
199 vmstate_vreg, ARMVectorReg),
200 VMSTATE_END_OF_LIST()
201 }
202 };
203 #endif /* AARCH64 */
204
205 static bool serror_needed(void *opaque)
206 {
207 ARMCPU *cpu = opaque;
208 CPUARMState *env = &cpu->env;
209
210 return env->serror.pending != 0;
211 }
212
213 static const VMStateDescription vmstate_serror = {
214 .name = "cpu/serror",
215 .version_id = 1,
216 .minimum_version_id = 1,
217 .needed = serror_needed,
218 .fields = (VMStateField[]) {
219 VMSTATE_UINT8(env.serror.pending, ARMCPU),
220 VMSTATE_UINT8(env.serror.has_esr, ARMCPU),
221 VMSTATE_UINT64(env.serror.esr, ARMCPU),
222 VMSTATE_END_OF_LIST()
223 }
224 };
225
226 static bool irq_line_state_needed(void *opaque)
227 {
228 return true;
229 }
230
231 static const VMStateDescription vmstate_irq_line_state = {
232 .name = "cpu/irq-line-state",
233 .version_id = 1,
234 .minimum_version_id = 1,
235 .needed = irq_line_state_needed,
236 .fields = (VMStateField[]) {
237 VMSTATE_UINT32(env.irq_line_state, ARMCPU),
238 VMSTATE_END_OF_LIST()
239 }
240 };
241
242 static bool m_needed(void *opaque)
243 {
244 ARMCPU *cpu = opaque;
245 CPUARMState *env = &cpu->env;
246
247 return arm_feature(env, ARM_FEATURE_M);
248 }
249
250 static const VMStateDescription vmstate_m_faultmask_primask = {
251 .name = "cpu/m/faultmask-primask",
252 .version_id = 1,
253 .minimum_version_id = 1,
254 .needed = m_needed,
255 .fields = (VMStateField[]) {
256 VMSTATE_UINT32(env.v7m.faultmask[M_REG_NS], ARMCPU),
257 VMSTATE_UINT32(env.v7m.primask[M_REG_NS], ARMCPU),
258 VMSTATE_END_OF_LIST()
259 }
260 };
261
262 /* CSSELR is in a subsection because we didn't implement it previously.
263 * Migration from an old implementation will leave it at zero, which
264 * is OK since the only CPUs in the old implementation make the
265 * register RAZ/WI.
266 * Since there was no version of QEMU which implemented the CSSELR for
267 * just non-secure, we transfer both banks here rather than putting
268 * the secure banked version in the m-security subsection.
269 */
270 static bool csselr_vmstate_validate(void *opaque, int version_id)
271 {
272 ARMCPU *cpu = opaque;
273
274 return cpu->env.v7m.csselr[M_REG_NS] <= R_V7M_CSSELR_INDEX_MASK
275 && cpu->env.v7m.csselr[M_REG_S] <= R_V7M_CSSELR_INDEX_MASK;
276 }
277
278 static bool m_csselr_needed(void *opaque)
279 {
280 ARMCPU *cpu = opaque;
281
282 return !arm_v7m_csselr_razwi(cpu);
283 }
284
285 static const VMStateDescription vmstate_m_csselr = {
286 .name = "cpu/m/csselr",
287 .version_id = 1,
288 .minimum_version_id = 1,
289 .needed = m_csselr_needed,
290 .fields = (VMStateField[]) {
291 VMSTATE_UINT32_ARRAY(env.v7m.csselr, ARMCPU, M_REG_NUM_BANKS),
292 VMSTATE_VALIDATE("CSSELR is valid", csselr_vmstate_validate),
293 VMSTATE_END_OF_LIST()
294 }
295 };
296
297 static const VMStateDescription vmstate_m_scr = {
298 .name = "cpu/m/scr",
299 .version_id = 1,
300 .minimum_version_id = 1,
301 .needed = m_needed,
302 .fields = (VMStateField[]) {
303 VMSTATE_UINT32(env.v7m.scr[M_REG_NS], ARMCPU),
304 VMSTATE_END_OF_LIST()
305 }
306 };
307
308 static const VMStateDescription vmstate_m_other_sp = {
309 .name = "cpu/m/other-sp",
310 .version_id = 1,
311 .minimum_version_id = 1,
312 .needed = m_needed,
313 .fields = (VMStateField[]) {
314 VMSTATE_UINT32(env.v7m.other_sp, ARMCPU),
315 VMSTATE_END_OF_LIST()
316 }
317 };
318
319 static bool m_v8m_needed(void *opaque)
320 {
321 ARMCPU *cpu = opaque;
322 CPUARMState *env = &cpu->env;
323
324 return arm_feature(env, ARM_FEATURE_M) && arm_feature(env, ARM_FEATURE_V8);
325 }
326
327 static const VMStateDescription vmstate_m_v8m = {
328 .name = "cpu/m/v8m",
329 .version_id = 1,
330 .minimum_version_id = 1,
331 .needed = m_v8m_needed,
332 .fields = (VMStateField[]) {
333 VMSTATE_UINT32_ARRAY(env.v7m.msplim, ARMCPU, M_REG_NUM_BANKS),
334 VMSTATE_UINT32_ARRAY(env.v7m.psplim, ARMCPU, M_REG_NUM_BANKS),
335 VMSTATE_END_OF_LIST()
336 }
337 };
338
339 static const VMStateDescription vmstate_m_fp = {
340 .name = "cpu/m/fp",
341 .version_id = 1,
342 .minimum_version_id = 1,
343 .needed = vfp_needed,
344 .fields = (VMStateField[]) {
345 VMSTATE_UINT32_ARRAY(env.v7m.fpcar, ARMCPU, M_REG_NUM_BANKS),
346 VMSTATE_UINT32_ARRAY(env.v7m.fpccr, ARMCPU, M_REG_NUM_BANKS),
347 VMSTATE_UINT32_ARRAY(env.v7m.fpdscr, ARMCPU, M_REG_NUM_BANKS),
348 VMSTATE_UINT32_ARRAY(env.v7m.cpacr, ARMCPU, M_REG_NUM_BANKS),
349 VMSTATE_UINT32(env.v7m.nsacr, ARMCPU),
350 VMSTATE_END_OF_LIST()
351 }
352 };
353
354 static bool mve_needed(void *opaque)
355 {
356 ARMCPU *cpu = opaque;
357
358 return cpu_isar_feature(aa32_mve, cpu);
359 }
360
361 static const VMStateDescription vmstate_m_mve = {
362 .name = "cpu/m/mve",
363 .version_id = 1,
364 .minimum_version_id = 1,
365 .needed = mve_needed,
366 .fields = (VMStateField[]) {
367 VMSTATE_UINT32(env.v7m.vpr, ARMCPU),
368 VMSTATE_UINT32(env.v7m.ltpsize, ARMCPU),
369 VMSTATE_END_OF_LIST()
370 },
371 };
372
373 static const VMStateDescription vmstate_m = {
374 .name = "cpu/m",
375 .version_id = 4,
376 .minimum_version_id = 4,
377 .needed = m_needed,
378 .fields = (VMStateField[]) {
379 VMSTATE_UINT32(env.v7m.vecbase[M_REG_NS], ARMCPU),
380 VMSTATE_UINT32(env.v7m.basepri[M_REG_NS], ARMCPU),
381 VMSTATE_UINT32(env.v7m.control[M_REG_NS], ARMCPU),
382 VMSTATE_UINT32(env.v7m.ccr[M_REG_NS], ARMCPU),
383 VMSTATE_UINT32(env.v7m.cfsr[M_REG_NS], ARMCPU),
384 VMSTATE_UINT32(env.v7m.hfsr, ARMCPU),
385 VMSTATE_UINT32(env.v7m.dfsr, ARMCPU),
386 VMSTATE_UINT32(env.v7m.mmfar[M_REG_NS], ARMCPU),
387 VMSTATE_UINT32(env.v7m.bfar, ARMCPU),
388 VMSTATE_UINT32(env.v7m.mpu_ctrl[M_REG_NS], ARMCPU),
389 VMSTATE_INT32(env.v7m.exception, ARMCPU),
390 VMSTATE_END_OF_LIST()
391 },
392 .subsections = (const VMStateDescription*[]) {
393 &vmstate_m_faultmask_primask,
394 &vmstate_m_csselr,
395 &vmstate_m_scr,
396 &vmstate_m_other_sp,
397 &vmstate_m_v8m,
398 &vmstate_m_fp,
399 &vmstate_m_mve,
400 NULL
401 }
402 };
403
404 static bool thumb2ee_needed(void *opaque)
405 {
406 ARMCPU *cpu = opaque;
407 CPUARMState *env = &cpu->env;
408
409 return arm_feature(env, ARM_FEATURE_THUMB2EE);
410 }
411
412 static const VMStateDescription vmstate_thumb2ee = {
413 .name = "cpu/thumb2ee",
414 .version_id = 1,
415 .minimum_version_id = 1,
416 .needed = thumb2ee_needed,
417 .fields = (VMStateField[]) {
418 VMSTATE_UINT32(env.teecr, ARMCPU),
419 VMSTATE_UINT32(env.teehbr, ARMCPU),
420 VMSTATE_END_OF_LIST()
421 }
422 };
423
424 static bool pmsav7_needed(void *opaque)
425 {
426 ARMCPU *cpu = opaque;
427 CPUARMState *env = &cpu->env;
428
429 return arm_feature(env, ARM_FEATURE_PMSA) &&
430 arm_feature(env, ARM_FEATURE_V7) &&
431 !arm_feature(env, ARM_FEATURE_V8);
432 }
433
434 static bool pmsav7_rgnr_vmstate_validate(void *opaque, int version_id)
435 {
436 ARMCPU *cpu = opaque;
437
438 return cpu->env.pmsav7.rnr[M_REG_NS] < cpu->pmsav7_dregion;
439 }
440
441 static const VMStateDescription vmstate_pmsav7 = {
442 .name = "cpu/pmsav7",
443 .version_id = 1,
444 .minimum_version_id = 1,
445 .needed = pmsav7_needed,
446 .fields = (VMStateField[]) {
447 VMSTATE_VARRAY_UINT32(env.pmsav7.drbar, ARMCPU, pmsav7_dregion, 0,
448 vmstate_info_uint32, uint32_t),
449 VMSTATE_VARRAY_UINT32(env.pmsav7.drsr, ARMCPU, pmsav7_dregion, 0,
450 vmstate_info_uint32, uint32_t),
451 VMSTATE_VARRAY_UINT32(env.pmsav7.dracr, ARMCPU, pmsav7_dregion, 0,
452 vmstate_info_uint32, uint32_t),
453 VMSTATE_VALIDATE("rgnr is valid", pmsav7_rgnr_vmstate_validate),
454 VMSTATE_END_OF_LIST()
455 }
456 };
457
458 static bool pmsav7_rnr_needed(void *opaque)
459 {
460 ARMCPU *cpu = opaque;
461 CPUARMState *env = &cpu->env;
462
463 /* For R profile cores pmsav7.rnr is migrated via the cpreg
464 * "RGNR" definition in helper.h. For M profile we have to
465 * migrate it separately.
466 */
467 return arm_feature(env, ARM_FEATURE_M);
468 }
469
470 static const VMStateDescription vmstate_pmsav7_rnr = {
471 .name = "cpu/pmsav7-rnr",
472 .version_id = 1,
473 .minimum_version_id = 1,
474 .needed = pmsav7_rnr_needed,
475 .fields = (VMStateField[]) {
476 VMSTATE_UINT32(env.pmsav7.rnr[M_REG_NS], ARMCPU),
477 VMSTATE_END_OF_LIST()
478 }
479 };
480
481 static bool pmsav8_needed(void *opaque)
482 {
483 ARMCPU *cpu = opaque;
484 CPUARMState *env = &cpu->env;
485
486 return arm_feature(env, ARM_FEATURE_PMSA) &&
487 arm_feature(env, ARM_FEATURE_V8);
488 }
489
490 static bool pmsav8r_needed(void *opaque)
491 {
492 ARMCPU *cpu = opaque;
493 CPUARMState *env = &cpu->env;
494
495 return arm_feature(env, ARM_FEATURE_PMSA) &&
496 arm_feature(env, ARM_FEATURE_V8) &&
497 !arm_feature(env, ARM_FEATURE_M);
498 }
499
500 static const VMStateDescription vmstate_pmsav8r = {
501 .name = "cpu/pmsav8/pmsav8r",
502 .version_id = 1,
503 .minimum_version_id = 1,
504 .needed = pmsav8r_needed,
505 .fields = (VMStateField[]) {
506 VMSTATE_VARRAY_UINT32(env.pmsav8.hprbar, ARMCPU,
507 pmsav8r_hdregion, 0, vmstate_info_uint32, uint32_t),
508 VMSTATE_VARRAY_UINT32(env.pmsav8.hprlar, ARMCPU,
509 pmsav8r_hdregion, 0, vmstate_info_uint32, uint32_t),
510 VMSTATE_END_OF_LIST()
511 },
512 };
513
514 static const VMStateDescription vmstate_pmsav8 = {
515 .name = "cpu/pmsav8",
516 .version_id = 1,
517 .minimum_version_id = 1,
518 .needed = pmsav8_needed,
519 .fields = (VMStateField[]) {
520 VMSTATE_VARRAY_UINT32(env.pmsav8.rbar[M_REG_NS], ARMCPU, pmsav7_dregion,
521 0, vmstate_info_uint32, uint32_t),
522 VMSTATE_VARRAY_UINT32(env.pmsav8.rlar[M_REG_NS], ARMCPU, pmsav7_dregion,
523 0, vmstate_info_uint32, uint32_t),
524 VMSTATE_UINT32(env.pmsav8.mair0[M_REG_NS], ARMCPU),
525 VMSTATE_UINT32(env.pmsav8.mair1[M_REG_NS], ARMCPU),
526 VMSTATE_END_OF_LIST()
527 },
528 .subsections = (const VMStateDescription * []) {
529 &vmstate_pmsav8r,
530 NULL
531 }
532 };
533
534 static bool s_rnr_vmstate_validate(void *opaque, int version_id)
535 {
536 ARMCPU *cpu = opaque;
537
538 return cpu->env.pmsav7.rnr[M_REG_S] < cpu->pmsav7_dregion;
539 }
540
541 static bool sau_rnr_vmstate_validate(void *opaque, int version_id)
542 {
543 ARMCPU *cpu = opaque;
544
545 return cpu->env.sau.rnr < cpu->sau_sregion;
546 }
547
548 static bool m_security_needed(void *opaque)
549 {
550 ARMCPU *cpu = opaque;
551 CPUARMState *env = &cpu->env;
552
553 return arm_feature(env, ARM_FEATURE_M_SECURITY);
554 }
555
556 static const VMStateDescription vmstate_m_security = {
557 .name = "cpu/m-security",
558 .version_id = 1,
559 .minimum_version_id = 1,
560 .needed = m_security_needed,
561 .fields = (VMStateField[]) {
562 VMSTATE_UINT32(env.v7m.secure, ARMCPU),
563 VMSTATE_UINT32(env.v7m.other_ss_msp, ARMCPU),
564 VMSTATE_UINT32(env.v7m.other_ss_psp, ARMCPU),
565 VMSTATE_UINT32(env.v7m.basepri[M_REG_S], ARMCPU),
566 VMSTATE_UINT32(env.v7m.primask[M_REG_S], ARMCPU),
567 VMSTATE_UINT32(env.v7m.faultmask[M_REG_S], ARMCPU),
568 VMSTATE_UINT32(env.v7m.control[M_REG_S], ARMCPU),
569 VMSTATE_UINT32(env.v7m.vecbase[M_REG_S], ARMCPU),
570 VMSTATE_UINT32(env.pmsav8.mair0[M_REG_S], ARMCPU),
571 VMSTATE_UINT32(env.pmsav8.mair1[M_REG_S], ARMCPU),
572 VMSTATE_VARRAY_UINT32(env.pmsav8.rbar[M_REG_S], ARMCPU, pmsav7_dregion,
573 0, vmstate_info_uint32, uint32_t),
574 VMSTATE_VARRAY_UINT32(env.pmsav8.rlar[M_REG_S], ARMCPU, pmsav7_dregion,
575 0, vmstate_info_uint32, uint32_t),
576 VMSTATE_UINT32(env.pmsav7.rnr[M_REG_S], ARMCPU),
577 VMSTATE_VALIDATE("secure MPU_RNR is valid", s_rnr_vmstate_validate),
578 VMSTATE_UINT32(env.v7m.mpu_ctrl[M_REG_S], ARMCPU),
579 VMSTATE_UINT32(env.v7m.ccr[M_REG_S], ARMCPU),
580 VMSTATE_UINT32(env.v7m.mmfar[M_REG_S], ARMCPU),
581 VMSTATE_UINT32(env.v7m.cfsr[M_REG_S], ARMCPU),
582 VMSTATE_UINT32(env.v7m.sfsr, ARMCPU),
583 VMSTATE_UINT32(env.v7m.sfar, ARMCPU),
584 VMSTATE_VARRAY_UINT32(env.sau.rbar, ARMCPU, sau_sregion, 0,
585 vmstate_info_uint32, uint32_t),
586 VMSTATE_VARRAY_UINT32(env.sau.rlar, ARMCPU, sau_sregion, 0,
587 vmstate_info_uint32, uint32_t),
588 VMSTATE_UINT32(env.sau.rnr, ARMCPU),
589 VMSTATE_VALIDATE("SAU_RNR is valid", sau_rnr_vmstate_validate),
590 VMSTATE_UINT32(env.sau.ctrl, ARMCPU),
591 VMSTATE_UINT32(env.v7m.scr[M_REG_S], ARMCPU),
592 /* AIRCR is not secure-only, but our implementation is R/O if the
593 * security extension is unimplemented, so we migrate it here.
594 */
595 VMSTATE_UINT32(env.v7m.aircr, ARMCPU),
596 VMSTATE_END_OF_LIST()
597 }
598 };
599
600 static int get_cpsr(QEMUFile *f, void *opaque, size_t size,
601 const VMStateField *field)
602 {
603 ARMCPU *cpu = opaque;
604 CPUARMState *env = &cpu->env;
605 uint32_t val = qemu_get_be32(f);
606
607 if (arm_feature(env, ARM_FEATURE_M)) {
608 if (val & XPSR_EXCP) {
609 /* This is a CPSR format value from an older QEMU. (We can tell
610 * because values transferred in XPSR format always have zero
611 * for the EXCP field, and CPSR format will always have bit 4
612 * set in CPSR_M.) Rearrange it into XPSR format. The significant
613 * differences are that the T bit is not in the same place, the
614 * primask/faultmask info may be in the CPSR I and F bits, and
615 * we do not want the mode bits.
616 * We know that this cleanup happened before v8M, so there
617 * is no complication with banked primask/faultmask.
618 */
619 uint32_t newval = val;
620
621 assert(!arm_feature(env, ARM_FEATURE_M_SECURITY));
622
623 newval &= (CPSR_NZCV | CPSR_Q | CPSR_IT | CPSR_GE);
624 if (val & CPSR_T) {
625 newval |= XPSR_T;
626 }
627 /* If the I or F bits are set then this is a migration from
628 * an old QEMU which still stored the M profile FAULTMASK
629 * and PRIMASK in env->daif. For a new QEMU, the data is
630 * transferred using the vmstate_m_faultmask_primask subsection.
631 */
632 if (val & CPSR_F) {
633 env->v7m.faultmask[M_REG_NS] = 1;
634 }
635 if (val & CPSR_I) {
636 env->v7m.primask[M_REG_NS] = 1;
637 }
638 val = newval;
639 }
640 /* Ignore the low bits, they are handled by vmstate_m. */
641 xpsr_write(env, val, ~XPSR_EXCP);
642 return 0;
643 }
644
645 env->aarch64 = ((val & PSTATE_nRW) == 0);
646
647 if (is_a64(env)) {
648 pstate_write(env, val);
649 return 0;
650 }
651
652 cpsr_write(env, val, 0xffffffff, CPSRWriteRaw);
653 return 0;
654 }
655
656 static int put_cpsr(QEMUFile *f, void *opaque, size_t size,
657 const VMStateField *field, JSONWriter *vmdesc)
658 {
659 ARMCPU *cpu = opaque;
660 CPUARMState *env = &cpu->env;
661 uint32_t val;
662
663 if (arm_feature(env, ARM_FEATURE_M)) {
664 /* The low 9 bits are v7m.exception, which is handled by vmstate_m. */
665 val = xpsr_read(env) & ~XPSR_EXCP;
666 } else if (is_a64(env)) {
667 val = pstate_read(env);
668 } else {
669 val = cpsr_read(env);
670 }
671
672 qemu_put_be32(f, val);
673 return 0;
674 }
675
676 static const VMStateInfo vmstate_cpsr = {
677 .name = "cpsr",
678 .get = get_cpsr,
679 .put = put_cpsr,
680 };
681
682 static int get_power(QEMUFile *f, void *opaque, size_t size,
683 const VMStateField *field)
684 {
685 ARMCPU *cpu = opaque;
686 bool powered_off = qemu_get_byte(f);
687 cpu->power_state = powered_off ? PSCI_OFF : PSCI_ON;
688 return 0;
689 }
690
691 static int put_power(QEMUFile *f, void *opaque, size_t size,
692 const VMStateField *field, JSONWriter *vmdesc)
693 {
694 ARMCPU *cpu = opaque;
695
696 /* Migration should never happen while we transition power states */
697
698 if (cpu->power_state == PSCI_ON ||
699 cpu->power_state == PSCI_OFF) {
700 bool powered_off = (cpu->power_state == PSCI_OFF) ? true : false;
701 qemu_put_byte(f, powered_off);
702 return 0;
703 } else {
704 return 1;
705 }
706 }
707
708 static const VMStateInfo vmstate_powered_off = {
709 .name = "powered_off",
710 .get = get_power,
711 .put = put_power,
712 };
713
714 static int cpu_pre_save(void *opaque)
715 {
716 ARMCPU *cpu = opaque;
717
718 if (!kvm_enabled()) {
719 pmu_op_start(&cpu->env);
720 }
721
722 if (kvm_enabled()) {
723 if (!write_kvmstate_to_list(cpu)) {
724 /* This should never fail */
725 g_assert_not_reached();
726 }
727
728 /*
729 * kvm_arm_cpu_pre_save() must be called after
730 * write_kvmstate_to_list()
731 */
732 kvm_arm_cpu_pre_save(cpu);
733 } else {
734 if (!write_cpustate_to_list(cpu, false)) {
735 /* This should never fail. */
736 g_assert_not_reached();
737 }
738 }
739
740 cpu->cpreg_vmstate_array_len = cpu->cpreg_array_len;
741 memcpy(cpu->cpreg_vmstate_indexes, cpu->cpreg_indexes,
742 cpu->cpreg_array_len * sizeof(uint64_t));
743 memcpy(cpu->cpreg_vmstate_values, cpu->cpreg_values,
744 cpu->cpreg_array_len * sizeof(uint64_t));
745
746 return 0;
747 }
748
749 static int cpu_post_save(void *opaque)
750 {
751 ARMCPU *cpu = opaque;
752
753 if (!kvm_enabled()) {
754 pmu_op_finish(&cpu->env);
755 }
756
757 return 0;
758 }
759
760 static int cpu_pre_load(void *opaque)
761 {
762 ARMCPU *cpu = opaque;
763 CPUARMState *env = &cpu->env;
764
765 /*
766 * Pre-initialize irq_line_state to a value that's never valid as
767 * real data, so cpu_post_load() can tell whether we've seen the
768 * irq-line-state subsection in the incoming migration state.
769 */
770 env->irq_line_state = UINT32_MAX;
771
772 if (!kvm_enabled()) {
773 pmu_op_start(&cpu->env);
774 }
775
776 return 0;
777 }
778
779 static int cpu_post_load(void *opaque, int version_id)
780 {
781 ARMCPU *cpu = opaque;
782 CPUARMState *env = &cpu->env;
783 int i, v;
784
785 /*
786 * Handle migration compatibility from old QEMU which didn't
787 * send the irq-line-state subsection. A QEMU without it did not
788 * implement the HCR_EL2.{VI,VF} bits as generating interrupts,
789 * so for TCG the line state matches the bits set in cs->interrupt_request.
790 * For KVM the line state is not stored in cs->interrupt_request
791 * and so this will leave irq_line_state as 0, but this is OK because
792 * we only need to care about it for TCG.
793 */
794 if (env->irq_line_state == UINT32_MAX) {
795 CPUState *cs = CPU(cpu);
796
797 env->irq_line_state = cs->interrupt_request &
798 (CPU_INTERRUPT_HARD | CPU_INTERRUPT_FIQ |
799 CPU_INTERRUPT_VIRQ | CPU_INTERRUPT_VFIQ);
800 }
801
802 /* Update the values list from the incoming migration data.
803 * Anything in the incoming data which we don't know about is
804 * a migration failure; anything we know about but the incoming
805 * data doesn't specify retains its current (reset) value.
806 * The indexes list remains untouched -- we only inspect the
807 * incoming migration index list so we can match the values array
808 * entries with the right slots in our own values array.
809 */
810
811 for (i = 0, v = 0; i < cpu->cpreg_array_len
812 && v < cpu->cpreg_vmstate_array_len; i++) {
813 if (cpu->cpreg_vmstate_indexes[v] > cpu->cpreg_indexes[i]) {
814 /* register in our list but not incoming : skip it */
815 continue;
816 }
817 if (cpu->cpreg_vmstate_indexes[v] < cpu->cpreg_indexes[i]) {
818 /* register in their list but not ours: fail migration */
819 return -1;
820 }
821 /* matching register, copy the value over */
822 cpu->cpreg_values[i] = cpu->cpreg_vmstate_values[v];
823 v++;
824 }
825
826 if (kvm_enabled()) {
827 if (!write_list_to_kvmstate(cpu, KVM_PUT_FULL_STATE)) {
828 return -1;
829 }
830 /* Note that it's OK for the TCG side not to know about
831 * every register in the list; KVM is authoritative if
832 * we're using it.
833 */
834 write_list_to_cpustate(cpu);
835 kvm_arm_cpu_post_load(cpu);
836 } else {
837 if (!write_list_to_cpustate(cpu)) {
838 return -1;
839 }
840 }
841
842 /*
843 * Misaligned thumb pc is architecturally impossible. Fail the
844 * incoming migration. For TCG it would trigger the assert in
845 * thumb_tr_translate_insn().
846 */
847 if (!is_a64(env) && env->thumb && (env->regs[15] & 1)) {
848 return -1;
849 }
850
851 hw_breakpoint_update_all(cpu);
852 hw_watchpoint_update_all(cpu);
853
854 /*
855 * TCG gen_update_fp_context() relies on the invariant that
856 * FPDSCR.LTPSIZE is constant 4 for M-profile with the LOB extension;
857 * forbid bogus incoming data with some other value.
858 */
859 if (arm_feature(env, ARM_FEATURE_M) && cpu_isar_feature(aa32_lob, cpu)) {
860 if (extract32(env->v7m.fpdscr[M_REG_NS],
861 FPCR_LTPSIZE_SHIFT, FPCR_LTPSIZE_LENGTH) != 4 ||
862 extract32(env->v7m.fpdscr[M_REG_S],
863 FPCR_LTPSIZE_SHIFT, FPCR_LTPSIZE_LENGTH) != 4) {
864 return -1;
865 }
866 }
867
868 if (!kvm_enabled()) {
869 pmu_op_finish(&cpu->env);
870 }
871 arm_rebuild_hflags(&cpu->env);
872
873 return 0;
874 }
875
876 const VMStateDescription vmstate_arm_cpu = {
877 .name = "cpu",
878 .version_id = 22,
879 .minimum_version_id = 22,
880 .pre_save = cpu_pre_save,
881 .post_save = cpu_post_save,
882 .pre_load = cpu_pre_load,
883 .post_load = cpu_post_load,
884 .fields = (VMStateField[]) {
885 VMSTATE_UINT32_ARRAY(env.regs, ARMCPU, 16),
886 VMSTATE_UINT64_ARRAY(env.xregs, ARMCPU, 32),
887 VMSTATE_UINT64(env.pc, ARMCPU),
888 {
889 .name = "cpsr",
890 .version_id = 0,
891 .size = sizeof(uint32_t),
892 .info = &vmstate_cpsr,
893 .flags = VMS_SINGLE,
894 .offset = 0,
895 },
896 VMSTATE_UINT32(env.spsr, ARMCPU),
897 VMSTATE_UINT64_ARRAY(env.banked_spsr, ARMCPU, 8),
898 VMSTATE_UINT32_ARRAY(env.banked_r13, ARMCPU, 8),
899 VMSTATE_UINT32_ARRAY(env.banked_r14, ARMCPU, 8),
900 VMSTATE_UINT32_ARRAY(env.usr_regs, ARMCPU, 5),
901 VMSTATE_UINT32_ARRAY(env.fiq_regs, ARMCPU, 5),
902 VMSTATE_UINT64_ARRAY(env.elr_el, ARMCPU, 4),
903 VMSTATE_UINT64_ARRAY(env.sp_el, ARMCPU, 4),
904 /* The length-check must come before the arrays to avoid
905 * incoming data possibly overflowing the array.
906 */
907 VMSTATE_INT32_POSITIVE_LE(cpreg_vmstate_array_len, ARMCPU),
908 VMSTATE_VARRAY_INT32(cpreg_vmstate_indexes, ARMCPU,
909 cpreg_vmstate_array_len,
910 0, vmstate_info_uint64, uint64_t),
911 VMSTATE_VARRAY_INT32(cpreg_vmstate_values, ARMCPU,
912 cpreg_vmstate_array_len,
913 0, vmstate_info_uint64, uint64_t),
914 VMSTATE_UINT64(env.exclusive_addr, ARMCPU),
915 VMSTATE_UINT64(env.exclusive_val, ARMCPU),
916 VMSTATE_UINT64(env.exclusive_high, ARMCPU),
917 VMSTATE_UNUSED(sizeof(uint64_t)),
918 VMSTATE_UINT32(env.exception.syndrome, ARMCPU),
919 VMSTATE_UINT32(env.exception.fsr, ARMCPU),
920 VMSTATE_UINT64(env.exception.vaddress, ARMCPU),
921 VMSTATE_TIMER_PTR(gt_timer[GTIMER_PHYS], ARMCPU),
922 VMSTATE_TIMER_PTR(gt_timer[GTIMER_VIRT], ARMCPU),
923 {
924 .name = "power_state",
925 .version_id = 0,
926 .size = sizeof(bool),
927 .info = &vmstate_powered_off,
928 .flags = VMS_SINGLE,
929 .offset = 0,
930 },
931 VMSTATE_END_OF_LIST()
932 },
933 .subsections = (const VMStateDescription*[]) {
934 &vmstate_vfp,
935 &vmstate_iwmmxt,
936 &vmstate_m,
937 &vmstate_thumb2ee,
938 /* pmsav7_rnr must come before pmsav7 so that we have the
939 * region number before we test it in the VMSTATE_VALIDATE
940 * in vmstate_pmsav7.
941 */
942 &vmstate_pmsav7_rnr,
943 &vmstate_pmsav7,
944 &vmstate_pmsav8,
945 &vmstate_m_security,
946 #ifdef TARGET_AARCH64
947 &vmstate_sve,
948 &vmstate_za,
949 #endif
950 &vmstate_serror,
951 &vmstate_irq_line_state,
952 NULL
953 }
954 };