]> git.proxmox.com Git - mirror_qemu.git/blob - target/arm/machine.c
Merge remote-tracking branch 'remotes/jasowang/tags/net-pull-request' into staging
[mirror_qemu.git] / target / arm / machine.c
1 #include "qemu/osdep.h"
2 #include "cpu.h"
3 #include "qemu/error-report.h"
4 #include "sysemu/kvm.h"
5 #include "kvm_arm.h"
6 #include "internals.h"
7 #include "migration/cpu.h"
8
9 static bool vfp_needed(void *opaque)
10 {
11 ARMCPU *cpu = opaque;
12
13 return (arm_feature(&cpu->env, ARM_FEATURE_AARCH64)
14 ? cpu_isar_feature(aa64_fp_simd, cpu)
15 : cpu_isar_feature(aa32_vfp_simd, cpu));
16 }
17
18 static int get_fpscr(QEMUFile *f, void *opaque, size_t size,
19 const VMStateField *field)
20 {
21 ARMCPU *cpu = opaque;
22 CPUARMState *env = &cpu->env;
23 uint32_t val = qemu_get_be32(f);
24
25 vfp_set_fpscr(env, val);
26 return 0;
27 }
28
29 static int put_fpscr(QEMUFile *f, void *opaque, size_t size,
30 const VMStateField *field, JSONWriter *vmdesc)
31 {
32 ARMCPU *cpu = opaque;
33 CPUARMState *env = &cpu->env;
34
35 qemu_put_be32(f, vfp_get_fpscr(env));
36 return 0;
37 }
38
39 static const VMStateInfo vmstate_fpscr = {
40 .name = "fpscr",
41 .get = get_fpscr,
42 .put = put_fpscr,
43 };
44
45 static const VMStateDescription vmstate_vfp = {
46 .name = "cpu/vfp",
47 .version_id = 3,
48 .minimum_version_id = 3,
49 .needed = vfp_needed,
50 .fields = (VMStateField[]) {
51 /* For compatibility, store Qn out of Zn here. */
52 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[0].d, ARMCPU, 0, 2),
53 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[1].d, ARMCPU, 0, 2),
54 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[2].d, ARMCPU, 0, 2),
55 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[3].d, ARMCPU, 0, 2),
56 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[4].d, ARMCPU, 0, 2),
57 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[5].d, ARMCPU, 0, 2),
58 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[6].d, ARMCPU, 0, 2),
59 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[7].d, ARMCPU, 0, 2),
60 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[8].d, ARMCPU, 0, 2),
61 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[9].d, ARMCPU, 0, 2),
62 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[10].d, ARMCPU, 0, 2),
63 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[11].d, ARMCPU, 0, 2),
64 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[12].d, ARMCPU, 0, 2),
65 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[13].d, ARMCPU, 0, 2),
66 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[14].d, ARMCPU, 0, 2),
67 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[15].d, ARMCPU, 0, 2),
68 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[16].d, ARMCPU, 0, 2),
69 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[17].d, ARMCPU, 0, 2),
70 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[18].d, ARMCPU, 0, 2),
71 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[19].d, ARMCPU, 0, 2),
72 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[20].d, ARMCPU, 0, 2),
73 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[21].d, ARMCPU, 0, 2),
74 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[22].d, ARMCPU, 0, 2),
75 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[23].d, ARMCPU, 0, 2),
76 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[24].d, ARMCPU, 0, 2),
77 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[25].d, ARMCPU, 0, 2),
78 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[26].d, ARMCPU, 0, 2),
79 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[27].d, ARMCPU, 0, 2),
80 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[28].d, ARMCPU, 0, 2),
81 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[29].d, ARMCPU, 0, 2),
82 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[30].d, ARMCPU, 0, 2),
83 VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[31].d, ARMCPU, 0, 2),
84
85 /* The xregs array is a little awkward because element 1 (FPSCR)
86 * requires a specific accessor, so we have to split it up in
87 * the vmstate:
88 */
89 VMSTATE_UINT32(env.vfp.xregs[0], ARMCPU),
90 VMSTATE_UINT32_SUB_ARRAY(env.vfp.xregs, ARMCPU, 2, 14),
91 {
92 .name = "fpscr",
93 .version_id = 0,
94 .size = sizeof(uint32_t),
95 .info = &vmstate_fpscr,
96 .flags = VMS_SINGLE,
97 .offset = 0,
98 },
99 VMSTATE_END_OF_LIST()
100 }
101 };
102
103 static bool iwmmxt_needed(void *opaque)
104 {
105 ARMCPU *cpu = opaque;
106 CPUARMState *env = &cpu->env;
107
108 return arm_feature(env, ARM_FEATURE_IWMMXT);
109 }
110
111 static const VMStateDescription vmstate_iwmmxt = {
112 .name = "cpu/iwmmxt",
113 .version_id = 1,
114 .minimum_version_id = 1,
115 .needed = iwmmxt_needed,
116 .fields = (VMStateField[]) {
117 VMSTATE_UINT64_ARRAY(env.iwmmxt.regs, ARMCPU, 16),
118 VMSTATE_UINT32_ARRAY(env.iwmmxt.cregs, ARMCPU, 16),
119 VMSTATE_END_OF_LIST()
120 }
121 };
122
123 #ifdef TARGET_AARCH64
124 /* The expression ARM_MAX_VQ - 2 is 0 for pure AArch32 build,
125 * and ARMPredicateReg is actively empty. This triggers errors
126 * in the expansion of the VMSTATE macros.
127 */
128
129 static bool sve_needed(void *opaque)
130 {
131 ARMCPU *cpu = opaque;
132
133 return cpu_isar_feature(aa64_sve, cpu);
134 }
135
136 /* The first two words of each Zreg is stored in VFP state. */
137 static const VMStateDescription vmstate_zreg_hi_reg = {
138 .name = "cpu/sve/zreg_hi",
139 .version_id = 1,
140 .minimum_version_id = 1,
141 .fields = (VMStateField[]) {
142 VMSTATE_UINT64_SUB_ARRAY(d, ARMVectorReg, 2, ARM_MAX_VQ - 2),
143 VMSTATE_END_OF_LIST()
144 }
145 };
146
147 static const VMStateDescription vmstate_preg_reg = {
148 .name = "cpu/sve/preg",
149 .version_id = 1,
150 .minimum_version_id = 1,
151 .fields = (VMStateField[]) {
152 VMSTATE_UINT64_ARRAY(p, ARMPredicateReg, 2 * ARM_MAX_VQ / 8),
153 VMSTATE_END_OF_LIST()
154 }
155 };
156
157 static const VMStateDescription vmstate_sve = {
158 .name = "cpu/sve",
159 .version_id = 1,
160 .minimum_version_id = 1,
161 .needed = sve_needed,
162 .fields = (VMStateField[]) {
163 VMSTATE_STRUCT_ARRAY(env.vfp.zregs, ARMCPU, 32, 0,
164 vmstate_zreg_hi_reg, ARMVectorReg),
165 VMSTATE_STRUCT_ARRAY(env.vfp.pregs, ARMCPU, 17, 0,
166 vmstate_preg_reg, ARMPredicateReg),
167 VMSTATE_END_OF_LIST()
168 }
169 };
170 #endif /* AARCH64 */
171
172 static bool serror_needed(void *opaque)
173 {
174 ARMCPU *cpu = opaque;
175 CPUARMState *env = &cpu->env;
176
177 return env->serror.pending != 0;
178 }
179
180 static const VMStateDescription vmstate_serror = {
181 .name = "cpu/serror",
182 .version_id = 1,
183 .minimum_version_id = 1,
184 .needed = serror_needed,
185 .fields = (VMStateField[]) {
186 VMSTATE_UINT8(env.serror.pending, ARMCPU),
187 VMSTATE_UINT8(env.serror.has_esr, ARMCPU),
188 VMSTATE_UINT64(env.serror.esr, ARMCPU),
189 VMSTATE_END_OF_LIST()
190 }
191 };
192
193 static bool irq_line_state_needed(void *opaque)
194 {
195 return true;
196 }
197
198 static const VMStateDescription vmstate_irq_line_state = {
199 .name = "cpu/irq-line-state",
200 .version_id = 1,
201 .minimum_version_id = 1,
202 .needed = irq_line_state_needed,
203 .fields = (VMStateField[]) {
204 VMSTATE_UINT32(env.irq_line_state, ARMCPU),
205 VMSTATE_END_OF_LIST()
206 }
207 };
208
209 static bool m_needed(void *opaque)
210 {
211 ARMCPU *cpu = opaque;
212 CPUARMState *env = &cpu->env;
213
214 return arm_feature(env, ARM_FEATURE_M);
215 }
216
217 static const VMStateDescription vmstate_m_faultmask_primask = {
218 .name = "cpu/m/faultmask-primask",
219 .version_id = 1,
220 .minimum_version_id = 1,
221 .needed = m_needed,
222 .fields = (VMStateField[]) {
223 VMSTATE_UINT32(env.v7m.faultmask[M_REG_NS], ARMCPU),
224 VMSTATE_UINT32(env.v7m.primask[M_REG_NS], ARMCPU),
225 VMSTATE_END_OF_LIST()
226 }
227 };
228
229 /* CSSELR is in a subsection because we didn't implement it previously.
230 * Migration from an old implementation will leave it at zero, which
231 * is OK since the only CPUs in the old implementation make the
232 * register RAZ/WI.
233 * Since there was no version of QEMU which implemented the CSSELR for
234 * just non-secure, we transfer both banks here rather than putting
235 * the secure banked version in the m-security subsection.
236 */
237 static bool csselr_vmstate_validate(void *opaque, int version_id)
238 {
239 ARMCPU *cpu = opaque;
240
241 return cpu->env.v7m.csselr[M_REG_NS] <= R_V7M_CSSELR_INDEX_MASK
242 && cpu->env.v7m.csselr[M_REG_S] <= R_V7M_CSSELR_INDEX_MASK;
243 }
244
245 static bool m_csselr_needed(void *opaque)
246 {
247 ARMCPU *cpu = opaque;
248
249 return !arm_v7m_csselr_razwi(cpu);
250 }
251
252 static const VMStateDescription vmstate_m_csselr = {
253 .name = "cpu/m/csselr",
254 .version_id = 1,
255 .minimum_version_id = 1,
256 .needed = m_csselr_needed,
257 .fields = (VMStateField[]) {
258 VMSTATE_UINT32_ARRAY(env.v7m.csselr, ARMCPU, M_REG_NUM_BANKS),
259 VMSTATE_VALIDATE("CSSELR is valid", csselr_vmstate_validate),
260 VMSTATE_END_OF_LIST()
261 }
262 };
263
264 static const VMStateDescription vmstate_m_scr = {
265 .name = "cpu/m/scr",
266 .version_id = 1,
267 .minimum_version_id = 1,
268 .needed = m_needed,
269 .fields = (VMStateField[]) {
270 VMSTATE_UINT32(env.v7m.scr[M_REG_NS], ARMCPU),
271 VMSTATE_END_OF_LIST()
272 }
273 };
274
275 static const VMStateDescription vmstate_m_other_sp = {
276 .name = "cpu/m/other-sp",
277 .version_id = 1,
278 .minimum_version_id = 1,
279 .needed = m_needed,
280 .fields = (VMStateField[]) {
281 VMSTATE_UINT32(env.v7m.other_sp, ARMCPU),
282 VMSTATE_END_OF_LIST()
283 }
284 };
285
286 static bool m_v8m_needed(void *opaque)
287 {
288 ARMCPU *cpu = opaque;
289 CPUARMState *env = &cpu->env;
290
291 return arm_feature(env, ARM_FEATURE_M) && arm_feature(env, ARM_FEATURE_V8);
292 }
293
294 static const VMStateDescription vmstate_m_v8m = {
295 .name = "cpu/m/v8m",
296 .version_id = 1,
297 .minimum_version_id = 1,
298 .needed = m_v8m_needed,
299 .fields = (VMStateField[]) {
300 VMSTATE_UINT32_ARRAY(env.v7m.msplim, ARMCPU, M_REG_NUM_BANKS),
301 VMSTATE_UINT32_ARRAY(env.v7m.psplim, ARMCPU, M_REG_NUM_BANKS),
302 VMSTATE_END_OF_LIST()
303 }
304 };
305
306 static const VMStateDescription vmstate_m_fp = {
307 .name = "cpu/m/fp",
308 .version_id = 1,
309 .minimum_version_id = 1,
310 .needed = vfp_needed,
311 .fields = (VMStateField[]) {
312 VMSTATE_UINT32_ARRAY(env.v7m.fpcar, ARMCPU, M_REG_NUM_BANKS),
313 VMSTATE_UINT32_ARRAY(env.v7m.fpccr, ARMCPU, M_REG_NUM_BANKS),
314 VMSTATE_UINT32_ARRAY(env.v7m.fpdscr, ARMCPU, M_REG_NUM_BANKS),
315 VMSTATE_UINT32_ARRAY(env.v7m.cpacr, ARMCPU, M_REG_NUM_BANKS),
316 VMSTATE_UINT32(env.v7m.nsacr, ARMCPU),
317 VMSTATE_END_OF_LIST()
318 }
319 };
320
321 static bool mve_needed(void *opaque)
322 {
323 ARMCPU *cpu = opaque;
324
325 return cpu_isar_feature(aa32_mve, cpu);
326 }
327
328 static const VMStateDescription vmstate_m_mve = {
329 .name = "cpu/m/mve",
330 .version_id = 1,
331 .minimum_version_id = 1,
332 .needed = mve_needed,
333 .fields = (VMStateField[]) {
334 VMSTATE_UINT32(env.v7m.vpr, ARMCPU),
335 VMSTATE_UINT32(env.v7m.ltpsize, ARMCPU),
336 VMSTATE_END_OF_LIST()
337 },
338 };
339
340 static const VMStateDescription vmstate_m = {
341 .name = "cpu/m",
342 .version_id = 4,
343 .minimum_version_id = 4,
344 .needed = m_needed,
345 .fields = (VMStateField[]) {
346 VMSTATE_UINT32(env.v7m.vecbase[M_REG_NS], ARMCPU),
347 VMSTATE_UINT32(env.v7m.basepri[M_REG_NS], ARMCPU),
348 VMSTATE_UINT32(env.v7m.control[M_REG_NS], ARMCPU),
349 VMSTATE_UINT32(env.v7m.ccr[M_REG_NS], ARMCPU),
350 VMSTATE_UINT32(env.v7m.cfsr[M_REG_NS], ARMCPU),
351 VMSTATE_UINT32(env.v7m.hfsr, ARMCPU),
352 VMSTATE_UINT32(env.v7m.dfsr, ARMCPU),
353 VMSTATE_UINT32(env.v7m.mmfar[M_REG_NS], ARMCPU),
354 VMSTATE_UINT32(env.v7m.bfar, ARMCPU),
355 VMSTATE_UINT32(env.v7m.mpu_ctrl[M_REG_NS], ARMCPU),
356 VMSTATE_INT32(env.v7m.exception, ARMCPU),
357 VMSTATE_END_OF_LIST()
358 },
359 .subsections = (const VMStateDescription*[]) {
360 &vmstate_m_faultmask_primask,
361 &vmstate_m_csselr,
362 &vmstate_m_scr,
363 &vmstate_m_other_sp,
364 &vmstate_m_v8m,
365 &vmstate_m_fp,
366 &vmstate_m_mve,
367 NULL
368 }
369 };
370
371 static bool thumb2ee_needed(void *opaque)
372 {
373 ARMCPU *cpu = opaque;
374 CPUARMState *env = &cpu->env;
375
376 return arm_feature(env, ARM_FEATURE_THUMB2EE);
377 }
378
379 static const VMStateDescription vmstate_thumb2ee = {
380 .name = "cpu/thumb2ee",
381 .version_id = 1,
382 .minimum_version_id = 1,
383 .needed = thumb2ee_needed,
384 .fields = (VMStateField[]) {
385 VMSTATE_UINT32(env.teecr, ARMCPU),
386 VMSTATE_UINT32(env.teehbr, ARMCPU),
387 VMSTATE_END_OF_LIST()
388 }
389 };
390
391 static bool pmsav7_needed(void *opaque)
392 {
393 ARMCPU *cpu = opaque;
394 CPUARMState *env = &cpu->env;
395
396 return arm_feature(env, ARM_FEATURE_PMSA) &&
397 arm_feature(env, ARM_FEATURE_V7) &&
398 !arm_feature(env, ARM_FEATURE_V8);
399 }
400
401 static bool pmsav7_rgnr_vmstate_validate(void *opaque, int version_id)
402 {
403 ARMCPU *cpu = opaque;
404
405 return cpu->env.pmsav7.rnr[M_REG_NS] < cpu->pmsav7_dregion;
406 }
407
408 static const VMStateDescription vmstate_pmsav7 = {
409 .name = "cpu/pmsav7",
410 .version_id = 1,
411 .minimum_version_id = 1,
412 .needed = pmsav7_needed,
413 .fields = (VMStateField[]) {
414 VMSTATE_VARRAY_UINT32(env.pmsav7.drbar, ARMCPU, pmsav7_dregion, 0,
415 vmstate_info_uint32, uint32_t),
416 VMSTATE_VARRAY_UINT32(env.pmsav7.drsr, ARMCPU, pmsav7_dregion, 0,
417 vmstate_info_uint32, uint32_t),
418 VMSTATE_VARRAY_UINT32(env.pmsav7.dracr, ARMCPU, pmsav7_dregion, 0,
419 vmstate_info_uint32, uint32_t),
420 VMSTATE_VALIDATE("rgnr is valid", pmsav7_rgnr_vmstate_validate),
421 VMSTATE_END_OF_LIST()
422 }
423 };
424
425 static bool pmsav7_rnr_needed(void *opaque)
426 {
427 ARMCPU *cpu = opaque;
428 CPUARMState *env = &cpu->env;
429
430 /* For R profile cores pmsav7.rnr is migrated via the cpreg
431 * "RGNR" definition in helper.h. For M profile we have to
432 * migrate it separately.
433 */
434 return arm_feature(env, ARM_FEATURE_M);
435 }
436
437 static const VMStateDescription vmstate_pmsav7_rnr = {
438 .name = "cpu/pmsav7-rnr",
439 .version_id = 1,
440 .minimum_version_id = 1,
441 .needed = pmsav7_rnr_needed,
442 .fields = (VMStateField[]) {
443 VMSTATE_UINT32(env.pmsav7.rnr[M_REG_NS], ARMCPU),
444 VMSTATE_END_OF_LIST()
445 }
446 };
447
448 static bool pmsav8_needed(void *opaque)
449 {
450 ARMCPU *cpu = opaque;
451 CPUARMState *env = &cpu->env;
452
453 return arm_feature(env, ARM_FEATURE_PMSA) &&
454 arm_feature(env, ARM_FEATURE_V8);
455 }
456
457 static const VMStateDescription vmstate_pmsav8 = {
458 .name = "cpu/pmsav8",
459 .version_id = 1,
460 .minimum_version_id = 1,
461 .needed = pmsav8_needed,
462 .fields = (VMStateField[]) {
463 VMSTATE_VARRAY_UINT32(env.pmsav8.rbar[M_REG_NS], ARMCPU, pmsav7_dregion,
464 0, vmstate_info_uint32, uint32_t),
465 VMSTATE_VARRAY_UINT32(env.pmsav8.rlar[M_REG_NS], ARMCPU, pmsav7_dregion,
466 0, vmstate_info_uint32, uint32_t),
467 VMSTATE_UINT32(env.pmsav8.mair0[M_REG_NS], ARMCPU),
468 VMSTATE_UINT32(env.pmsav8.mair1[M_REG_NS], ARMCPU),
469 VMSTATE_END_OF_LIST()
470 }
471 };
472
473 static bool s_rnr_vmstate_validate(void *opaque, int version_id)
474 {
475 ARMCPU *cpu = opaque;
476
477 return cpu->env.pmsav7.rnr[M_REG_S] < cpu->pmsav7_dregion;
478 }
479
480 static bool sau_rnr_vmstate_validate(void *opaque, int version_id)
481 {
482 ARMCPU *cpu = opaque;
483
484 return cpu->env.sau.rnr < cpu->sau_sregion;
485 }
486
487 static bool m_security_needed(void *opaque)
488 {
489 ARMCPU *cpu = opaque;
490 CPUARMState *env = &cpu->env;
491
492 return arm_feature(env, ARM_FEATURE_M_SECURITY);
493 }
494
495 static const VMStateDescription vmstate_m_security = {
496 .name = "cpu/m-security",
497 .version_id = 1,
498 .minimum_version_id = 1,
499 .needed = m_security_needed,
500 .fields = (VMStateField[]) {
501 VMSTATE_UINT32(env.v7m.secure, ARMCPU),
502 VMSTATE_UINT32(env.v7m.other_ss_msp, ARMCPU),
503 VMSTATE_UINT32(env.v7m.other_ss_psp, ARMCPU),
504 VMSTATE_UINT32(env.v7m.basepri[M_REG_S], ARMCPU),
505 VMSTATE_UINT32(env.v7m.primask[M_REG_S], ARMCPU),
506 VMSTATE_UINT32(env.v7m.faultmask[M_REG_S], ARMCPU),
507 VMSTATE_UINT32(env.v7m.control[M_REG_S], ARMCPU),
508 VMSTATE_UINT32(env.v7m.vecbase[M_REG_S], ARMCPU),
509 VMSTATE_UINT32(env.pmsav8.mair0[M_REG_S], ARMCPU),
510 VMSTATE_UINT32(env.pmsav8.mair1[M_REG_S], ARMCPU),
511 VMSTATE_VARRAY_UINT32(env.pmsav8.rbar[M_REG_S], ARMCPU, pmsav7_dregion,
512 0, vmstate_info_uint32, uint32_t),
513 VMSTATE_VARRAY_UINT32(env.pmsav8.rlar[M_REG_S], ARMCPU, pmsav7_dregion,
514 0, vmstate_info_uint32, uint32_t),
515 VMSTATE_UINT32(env.pmsav7.rnr[M_REG_S], ARMCPU),
516 VMSTATE_VALIDATE("secure MPU_RNR is valid", s_rnr_vmstate_validate),
517 VMSTATE_UINT32(env.v7m.mpu_ctrl[M_REG_S], ARMCPU),
518 VMSTATE_UINT32(env.v7m.ccr[M_REG_S], ARMCPU),
519 VMSTATE_UINT32(env.v7m.mmfar[M_REG_S], ARMCPU),
520 VMSTATE_UINT32(env.v7m.cfsr[M_REG_S], ARMCPU),
521 VMSTATE_UINT32(env.v7m.sfsr, ARMCPU),
522 VMSTATE_UINT32(env.v7m.sfar, ARMCPU),
523 VMSTATE_VARRAY_UINT32(env.sau.rbar, ARMCPU, sau_sregion, 0,
524 vmstate_info_uint32, uint32_t),
525 VMSTATE_VARRAY_UINT32(env.sau.rlar, ARMCPU, sau_sregion, 0,
526 vmstate_info_uint32, uint32_t),
527 VMSTATE_UINT32(env.sau.rnr, ARMCPU),
528 VMSTATE_VALIDATE("SAU_RNR is valid", sau_rnr_vmstate_validate),
529 VMSTATE_UINT32(env.sau.ctrl, ARMCPU),
530 VMSTATE_UINT32(env.v7m.scr[M_REG_S], ARMCPU),
531 /* AIRCR is not secure-only, but our implementation is R/O if the
532 * security extension is unimplemented, so we migrate it here.
533 */
534 VMSTATE_UINT32(env.v7m.aircr, ARMCPU),
535 VMSTATE_END_OF_LIST()
536 }
537 };
538
539 static int get_cpsr(QEMUFile *f, void *opaque, size_t size,
540 const VMStateField *field)
541 {
542 ARMCPU *cpu = opaque;
543 CPUARMState *env = &cpu->env;
544 uint32_t val = qemu_get_be32(f);
545
546 if (arm_feature(env, ARM_FEATURE_M)) {
547 if (val & XPSR_EXCP) {
548 /* This is a CPSR format value from an older QEMU. (We can tell
549 * because values transferred in XPSR format always have zero
550 * for the EXCP field, and CPSR format will always have bit 4
551 * set in CPSR_M.) Rearrange it into XPSR format. The significant
552 * differences are that the T bit is not in the same place, the
553 * primask/faultmask info may be in the CPSR I and F bits, and
554 * we do not want the mode bits.
555 * We know that this cleanup happened before v8M, so there
556 * is no complication with banked primask/faultmask.
557 */
558 uint32_t newval = val;
559
560 assert(!arm_feature(env, ARM_FEATURE_M_SECURITY));
561
562 newval &= (CPSR_NZCV | CPSR_Q | CPSR_IT | CPSR_GE);
563 if (val & CPSR_T) {
564 newval |= XPSR_T;
565 }
566 /* If the I or F bits are set then this is a migration from
567 * an old QEMU which still stored the M profile FAULTMASK
568 * and PRIMASK in env->daif. For a new QEMU, the data is
569 * transferred using the vmstate_m_faultmask_primask subsection.
570 */
571 if (val & CPSR_F) {
572 env->v7m.faultmask[M_REG_NS] = 1;
573 }
574 if (val & CPSR_I) {
575 env->v7m.primask[M_REG_NS] = 1;
576 }
577 val = newval;
578 }
579 /* Ignore the low bits, they are handled by vmstate_m. */
580 xpsr_write(env, val, ~XPSR_EXCP);
581 return 0;
582 }
583
584 env->aarch64 = ((val & PSTATE_nRW) == 0);
585
586 if (is_a64(env)) {
587 pstate_write(env, val);
588 return 0;
589 }
590
591 cpsr_write(env, val, 0xffffffff, CPSRWriteRaw);
592 return 0;
593 }
594
595 static int put_cpsr(QEMUFile *f, void *opaque, size_t size,
596 const VMStateField *field, JSONWriter *vmdesc)
597 {
598 ARMCPU *cpu = opaque;
599 CPUARMState *env = &cpu->env;
600 uint32_t val;
601
602 if (arm_feature(env, ARM_FEATURE_M)) {
603 /* The low 9 bits are v7m.exception, which is handled by vmstate_m. */
604 val = xpsr_read(env) & ~XPSR_EXCP;
605 } else if (is_a64(env)) {
606 val = pstate_read(env);
607 } else {
608 val = cpsr_read(env);
609 }
610
611 qemu_put_be32(f, val);
612 return 0;
613 }
614
615 static const VMStateInfo vmstate_cpsr = {
616 .name = "cpsr",
617 .get = get_cpsr,
618 .put = put_cpsr,
619 };
620
621 static int get_power(QEMUFile *f, void *opaque, size_t size,
622 const VMStateField *field)
623 {
624 ARMCPU *cpu = opaque;
625 bool powered_off = qemu_get_byte(f);
626 cpu->power_state = powered_off ? PSCI_OFF : PSCI_ON;
627 return 0;
628 }
629
630 static int put_power(QEMUFile *f, void *opaque, size_t size,
631 const VMStateField *field, JSONWriter *vmdesc)
632 {
633 ARMCPU *cpu = opaque;
634
635 /* Migration should never happen while we transition power states */
636
637 if (cpu->power_state == PSCI_ON ||
638 cpu->power_state == PSCI_OFF) {
639 bool powered_off = (cpu->power_state == PSCI_OFF) ? true : false;
640 qemu_put_byte(f, powered_off);
641 return 0;
642 } else {
643 return 1;
644 }
645 }
646
647 static const VMStateInfo vmstate_powered_off = {
648 .name = "powered_off",
649 .get = get_power,
650 .put = put_power,
651 };
652
653 static int cpu_pre_save(void *opaque)
654 {
655 ARMCPU *cpu = opaque;
656
657 if (!kvm_enabled()) {
658 pmu_op_start(&cpu->env);
659 }
660
661 if (kvm_enabled()) {
662 if (!write_kvmstate_to_list(cpu)) {
663 /* This should never fail */
664 abort();
665 }
666
667 /*
668 * kvm_arm_cpu_pre_save() must be called after
669 * write_kvmstate_to_list()
670 */
671 kvm_arm_cpu_pre_save(cpu);
672 } else {
673 if (!write_cpustate_to_list(cpu, false)) {
674 /* This should never fail. */
675 abort();
676 }
677 }
678
679 cpu->cpreg_vmstate_array_len = cpu->cpreg_array_len;
680 memcpy(cpu->cpreg_vmstate_indexes, cpu->cpreg_indexes,
681 cpu->cpreg_array_len * sizeof(uint64_t));
682 memcpy(cpu->cpreg_vmstate_values, cpu->cpreg_values,
683 cpu->cpreg_array_len * sizeof(uint64_t));
684
685 return 0;
686 }
687
688 static int cpu_post_save(void *opaque)
689 {
690 ARMCPU *cpu = opaque;
691
692 if (!kvm_enabled()) {
693 pmu_op_finish(&cpu->env);
694 }
695
696 return 0;
697 }
698
699 static int cpu_pre_load(void *opaque)
700 {
701 ARMCPU *cpu = opaque;
702 CPUARMState *env = &cpu->env;
703
704 /*
705 * Pre-initialize irq_line_state to a value that's never valid as
706 * real data, so cpu_post_load() can tell whether we've seen the
707 * irq-line-state subsection in the incoming migration state.
708 */
709 env->irq_line_state = UINT32_MAX;
710
711 if (!kvm_enabled()) {
712 pmu_op_start(&cpu->env);
713 }
714
715 return 0;
716 }
717
718 static int cpu_post_load(void *opaque, int version_id)
719 {
720 ARMCPU *cpu = opaque;
721 CPUARMState *env = &cpu->env;
722 int i, v;
723
724 /*
725 * Handle migration compatibility from old QEMU which didn't
726 * send the irq-line-state subsection. A QEMU without it did not
727 * implement the HCR_EL2.{VI,VF} bits as generating interrupts,
728 * so for TCG the line state matches the bits set in cs->interrupt_request.
729 * For KVM the line state is not stored in cs->interrupt_request
730 * and so this will leave irq_line_state as 0, but this is OK because
731 * we only need to care about it for TCG.
732 */
733 if (env->irq_line_state == UINT32_MAX) {
734 CPUState *cs = CPU(cpu);
735
736 env->irq_line_state = cs->interrupt_request &
737 (CPU_INTERRUPT_HARD | CPU_INTERRUPT_FIQ |
738 CPU_INTERRUPT_VIRQ | CPU_INTERRUPT_VFIQ);
739 }
740
741 /* Update the values list from the incoming migration data.
742 * Anything in the incoming data which we don't know about is
743 * a migration failure; anything we know about but the incoming
744 * data doesn't specify retains its current (reset) value.
745 * The indexes list remains untouched -- we only inspect the
746 * incoming migration index list so we can match the values array
747 * entries with the right slots in our own values array.
748 */
749
750 for (i = 0, v = 0; i < cpu->cpreg_array_len
751 && v < cpu->cpreg_vmstate_array_len; i++) {
752 if (cpu->cpreg_vmstate_indexes[v] > cpu->cpreg_indexes[i]) {
753 /* register in our list but not incoming : skip it */
754 continue;
755 }
756 if (cpu->cpreg_vmstate_indexes[v] < cpu->cpreg_indexes[i]) {
757 /* register in their list but not ours: fail migration */
758 return -1;
759 }
760 /* matching register, copy the value over */
761 cpu->cpreg_values[i] = cpu->cpreg_vmstate_values[v];
762 v++;
763 }
764
765 if (kvm_enabled()) {
766 if (!write_list_to_kvmstate(cpu, KVM_PUT_FULL_STATE)) {
767 return -1;
768 }
769 /* Note that it's OK for the TCG side not to know about
770 * every register in the list; KVM is authoritative if
771 * we're using it.
772 */
773 write_list_to_cpustate(cpu);
774 kvm_arm_cpu_post_load(cpu);
775 } else {
776 if (!write_list_to_cpustate(cpu)) {
777 return -1;
778 }
779 }
780
781 hw_breakpoint_update_all(cpu);
782 hw_watchpoint_update_all(cpu);
783
784 if (!kvm_enabled()) {
785 pmu_op_finish(&cpu->env);
786 }
787 arm_rebuild_hflags(&cpu->env);
788
789 return 0;
790 }
791
792 const VMStateDescription vmstate_arm_cpu = {
793 .name = "cpu",
794 .version_id = 22,
795 .minimum_version_id = 22,
796 .pre_save = cpu_pre_save,
797 .post_save = cpu_post_save,
798 .pre_load = cpu_pre_load,
799 .post_load = cpu_post_load,
800 .fields = (VMStateField[]) {
801 VMSTATE_UINT32_ARRAY(env.regs, ARMCPU, 16),
802 VMSTATE_UINT64_ARRAY(env.xregs, ARMCPU, 32),
803 VMSTATE_UINT64(env.pc, ARMCPU),
804 {
805 .name = "cpsr",
806 .version_id = 0,
807 .size = sizeof(uint32_t),
808 .info = &vmstate_cpsr,
809 .flags = VMS_SINGLE,
810 .offset = 0,
811 },
812 VMSTATE_UINT32(env.spsr, ARMCPU),
813 VMSTATE_UINT64_ARRAY(env.banked_spsr, ARMCPU, 8),
814 VMSTATE_UINT32_ARRAY(env.banked_r13, ARMCPU, 8),
815 VMSTATE_UINT32_ARRAY(env.banked_r14, ARMCPU, 8),
816 VMSTATE_UINT32_ARRAY(env.usr_regs, ARMCPU, 5),
817 VMSTATE_UINT32_ARRAY(env.fiq_regs, ARMCPU, 5),
818 VMSTATE_UINT64_ARRAY(env.elr_el, ARMCPU, 4),
819 VMSTATE_UINT64_ARRAY(env.sp_el, ARMCPU, 4),
820 /* The length-check must come before the arrays to avoid
821 * incoming data possibly overflowing the array.
822 */
823 VMSTATE_INT32_POSITIVE_LE(cpreg_vmstate_array_len, ARMCPU),
824 VMSTATE_VARRAY_INT32(cpreg_vmstate_indexes, ARMCPU,
825 cpreg_vmstate_array_len,
826 0, vmstate_info_uint64, uint64_t),
827 VMSTATE_VARRAY_INT32(cpreg_vmstate_values, ARMCPU,
828 cpreg_vmstate_array_len,
829 0, vmstate_info_uint64, uint64_t),
830 VMSTATE_UINT64(env.exclusive_addr, ARMCPU),
831 VMSTATE_UINT64(env.exclusive_val, ARMCPU),
832 VMSTATE_UINT64(env.exclusive_high, ARMCPU),
833 VMSTATE_UNUSED(sizeof(uint64_t)),
834 VMSTATE_UINT32(env.exception.syndrome, ARMCPU),
835 VMSTATE_UINT32(env.exception.fsr, ARMCPU),
836 VMSTATE_UINT64(env.exception.vaddress, ARMCPU),
837 VMSTATE_TIMER_PTR(gt_timer[GTIMER_PHYS], ARMCPU),
838 VMSTATE_TIMER_PTR(gt_timer[GTIMER_VIRT], ARMCPU),
839 {
840 .name = "power_state",
841 .version_id = 0,
842 .size = sizeof(bool),
843 .info = &vmstate_powered_off,
844 .flags = VMS_SINGLE,
845 .offset = 0,
846 },
847 VMSTATE_END_OF_LIST()
848 },
849 .subsections = (const VMStateDescription*[]) {
850 &vmstate_vfp,
851 &vmstate_iwmmxt,
852 &vmstate_m,
853 &vmstate_thumb2ee,
854 /* pmsav7_rnr must come before pmsav7 so that we have the
855 * region number before we test it in the VMSTATE_VALIDATE
856 * in vmstate_pmsav7.
857 */
858 &vmstate_pmsav7_rnr,
859 &vmstate_pmsav7,
860 &vmstate_pmsav8,
861 &vmstate_m_security,
862 #ifdef TARGET_AARCH64
863 &vmstate_sve,
864 #endif
865 &vmstate_serror,
866 &vmstate_irq_line_state,
867 NULL
868 }
869 };