]> git.proxmox.com Git - mirror_qemu.git/blame - target/arm/tcg/translate-m-nocp.c
target/arm: Tidy helpers for translation
[mirror_qemu.git] / target / arm / tcg / translate-m-nocp.c
CommitLineData
9a5071ab
PM
1/*
2 * ARM translation: M-profile NOCP special-case instructions
3 *
4 * Copyright (c) 2020 Linaro, Ltd.
5 *
6 * This library is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU Lesser General Public
8 * License as published by the Free Software Foundation; either
9 * version 2.1 of the License, or (at your option) any later version.
10 *
11 * This library is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14 * Lesser General Public License for more details.
15 *
16 * You should have received a copy of the GNU Lesser General Public
17 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
18 */
19
20#include "qemu/osdep.h"
9a5071ab
PM
21#include "translate.h"
22#include "translate-a32.h"
23
24#include "decode-m-nocp.c.inc"
25
26/*
27 * Decode VLLDM and VLSTM are nonstandard because:
28 * * if there is no FPU then these insns must NOP in
29 * Secure state and UNDEF in Nonsecure state
30 * * if there is an FPU then these insns do not have
31 * the usual behaviour that vfp_access_check() provides of
32 * being controlled by CPACR/NSACR enable bits or the
33 * lazy-stacking logic.
34 */
35static bool trans_VLLDM_VLSTM(DisasContext *s, arg_VLLDM_VLSTM *a)
36{
37 TCGv_i32 fptr;
38
39 if (!arm_dc_feature(s, ARM_FEATURE_M) ||
40 !arm_dc_feature(s, ARM_FEATURE_V8)) {
41 return false;
42 }
43
44 if (a->op) {
45 /*
46 * T2 encoding ({D0-D31} reglist): v8.1M and up. We choose not
47 * to take the IMPDEF option to make memory accesses to the stack
48 * slots that correspond to the D16-D31 registers (discarding
49 * read data and writing UNKNOWN values), so for us the T2
50 * encoding behaves identically to the T1 encoding.
51 */
52 if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
53 return false;
54 }
55 } else {
56 /*
57 * T1 encoding ({D0-D15} reglist); undef if we have 32 Dregs.
58 * This is currently architecturally impossible, but we add the
59 * check to stay in line with the pseudocode. Note that we must
60 * emit code for the UNDEF so it takes precedence over the NOCP.
61 */
62 if (dc_isar_feature(aa32_simd_r32, s)) {
63 unallocated_encoding(s);
64 return true;
65 }
66 }
67
68 /*
69 * If not secure, UNDEF. We must emit code for this
70 * rather than returning false so that this takes
71 * precedence over the m-nocp.decode NOCP fallback.
72 */
73 if (!s->v8m_secure) {
74 unallocated_encoding(s);
75 return true;
76 }
5138bd01
PM
77
78 s->eci_handled = true;
79
9a5071ab
PM
80 /* If no fpu, NOP. */
81 if (!dc_isar_feature(aa32_vfp, s)) {
5138bd01 82 clear_eci_state(s);
9a5071ab
PM
83 return true;
84 }
85
86 fptr = load_reg(s, a->rn);
87 if (a->l) {
88 gen_helper_v7m_vlldm(cpu_env, fptr);
89 } else {
90 gen_helper_v7m_vlstm(cpu_env, fptr);
91 }
9a5071ab 92
5138bd01
PM
93 clear_eci_state(s);
94
26702213
PM
95 /*
96 * End the TB, because we have updated FP control bits,
97 * and possibly VPR or LTPSIZE.
98 */
9a5071ab
PM
99 s->base.is_jmp = DISAS_UPDATE_EXIT;
100 return true;
101}
102
103static bool trans_VSCCLRM(DisasContext *s, arg_VSCCLRM *a)
104{
105 int btmreg, topreg;
106 TCGv_i64 zero;
107 TCGv_i32 aspen, sfpa;
108
109 if (!dc_isar_feature(aa32_m_sec_state, s)) {
110 /* Before v8.1M, fall through in decode to NOCP check */
111 return false;
112 }
113
114 /* Explicitly UNDEF because this takes precedence over NOCP */
115 if (!arm_dc_feature(s, ARM_FEATURE_M_MAIN) || !s->v8m_secure) {
116 unallocated_encoding(s);
117 return true;
118 }
119
5138bd01
PM
120 s->eci_handled = true;
121
9a5071ab
PM
122 if (!dc_isar_feature(aa32_vfp_simd, s)) {
123 /* NOP if we have neither FP nor MVE */
5138bd01 124 clear_eci_state(s);
9a5071ab
PM
125 return true;
126 }
127
128 /*
129 * If FPCCR.ASPEN != 0 && CONTROL_S.SFPA == 0 then there is no
130 * active floating point context so we must NOP (without doing
131 * any lazy state preservation or the NOCP check).
132 */
133 aspen = load_cpu_field(v7m.fpccr[M_REG_S]);
134 sfpa = load_cpu_field(v7m.control[M_REG_S]);
135 tcg_gen_andi_i32(aspen, aspen, R_V7M_FPCCR_ASPEN_MASK);
136 tcg_gen_xori_i32(aspen, aspen, R_V7M_FPCCR_ASPEN_MASK);
137 tcg_gen_andi_i32(sfpa, sfpa, R_V7M_CONTROL_SFPA_MASK);
138 tcg_gen_or_i32(sfpa, sfpa, aspen);
139 arm_gen_condlabel(s);
abb80995 140 tcg_gen_brcondi_i32(TCG_COND_EQ, sfpa, 0, s->condlabel.label);
9a5071ab
PM
141
142 if (s->fp_excp_el != 0) {
55086e62 143 gen_exception_insn_el(s, 0, EXCP_NOCP,
8c5d24dc 144 syn_uncategorized(), s->fp_excp_el);
9a5071ab
PM
145 return true;
146 }
147
148 topreg = a->vd + a->imm - 1;
149 btmreg = a->vd;
150
151 /* Convert to Sreg numbers if the insn specified in Dregs */
152 if (a->size == 3) {
153 topreg = topreg * 2 + 1;
154 btmreg *= 2;
155 }
156
157 if (topreg > 63 || (topreg > 31 && !(topreg & 1))) {
158 /* UNPREDICTABLE: we choose to undef */
159 unallocated_encoding(s);
160 return true;
161 }
162
163 /* Silently ignore requests to clear D16-D31 if they don't exist */
164 if (topreg > 31 && !dc_isar_feature(aa32_simd_r32, s)) {
165 topreg = 31;
166 }
167
168 if (!vfp_access_check(s)) {
169 return true;
170 }
171
172 /* Zero the Sregs from btmreg to topreg inclusive. */
01d90db5 173 zero = tcg_constant_i64(0);
9a5071ab
PM
174 if (btmreg & 1) {
175 write_neon_element64(zero, btmreg >> 1, 1, MO_32);
176 btmreg++;
177 }
178 for (; btmreg + 1 <= topreg; btmreg += 2) {
179 write_neon_element64(zero, btmreg >> 1, 0, MO_64);
180 }
181 if (btmreg == topreg) {
182 write_neon_element64(zero, btmreg >> 1, 0, MO_32);
183 btmreg++;
184 }
185 assert(btmreg == topreg + 1);
375256a8 186 if (dc_isar_feature(aa32_mve, s)) {
01d90db5 187 store_cpu_field(tcg_constant_i32(0), v7m.vpr);
375256a8 188 }
5138bd01
PM
189
190 clear_eci_state(s);
9a5071ab
PM
191 return true;
192}
193
fa856736
PM
194/*
195 * M-profile provides two different sets of instructions that can
196 * access floating point system registers: VMSR/VMRS (which move
197 * to/from a general purpose register) and VLDR/VSTR sysreg (which
198 * move directly to/from memory). In some cases there are also side
199 * effects which must happen after any write to memory (which could
200 * cause an exception). So we implement the common logic for the
201 * sysreg access in gen_M_fp_sysreg_write() and gen_M_fp_sysreg_read(),
202 * which take pointers to callback functions which will perform the
203 * actual "read/write general purpose register" and "read/write
204 * memory" operations.
205 */
206
207/*
208 * Emit code to store the sysreg to its final destination; frees the
e494cd0a
PM
209 * TCG temp 'value' it is passed. do_access is true to do the store,
210 * and false to skip it and only perform side-effects like base
211 * register writeback.
fa856736 212 */
e494cd0a
PM
213typedef void fp_sysreg_storefn(DisasContext *s, void *opaque, TCGv_i32 value,
214 bool do_access);
fa856736
PM
215/*
216 * Emit code to load the value to be copied to the sysreg; returns
e494cd0a
PM
217 * a new TCG temporary. do_access is true to do the store,
218 * and false to skip it and only perform side-effects like base
219 * register writeback.
fa856736 220 */
e494cd0a
PM
221typedef TCGv_i32 fp_sysreg_loadfn(DisasContext *s, void *opaque,
222 bool do_access);
fa856736
PM
223
224/* Common decode/access checks for fp sysreg read/write */
225typedef enum FPSysRegCheckResult {
226 FPSysRegCheckFailed, /* caller should return false */
227 FPSysRegCheckDone, /* caller should return true */
228 FPSysRegCheckContinue, /* caller should continue generating code */
229} FPSysRegCheckResult;
230
231static FPSysRegCheckResult fp_sysreg_checks(DisasContext *s, int regno)
232{
233 if (!dc_isar_feature(aa32_fpsp_v2, s) && !dc_isar_feature(aa32_mve, s)) {
234 return FPSysRegCheckFailed;
235 }
236
237 switch (regno) {
238 case ARM_VFP_FPSCR:
239 case QEMU_VFP_FPSCR_NZCV:
240 break;
241 case ARM_VFP_FPSCR_NZCVQC:
242 if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
243 return FPSysRegCheckFailed;
244 }
245 break;
246 case ARM_VFP_FPCXT_S:
247 case ARM_VFP_FPCXT_NS:
248 if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
249 return FPSysRegCheckFailed;
250 }
251 if (!s->v8m_secure) {
252 return FPSysRegCheckFailed;
253 }
254 break;
255 case ARM_VFP_VPR:
256 case ARM_VFP_P0:
257 if (!dc_isar_feature(aa32_mve, s)) {
258 return FPSysRegCheckFailed;
259 }
260 break;
261 default:
262 return FPSysRegCheckFailed;
263 }
264
265 /*
266 * FPCXT_NS is a special case: it has specific handling for
267 * "current FP state is inactive", and must do the PreserveFPState()
268 * but not the usual full set of actions done by ExecuteFPCheck().
269 * So we don't call vfp_access_check() and the callers must handle this.
270 */
271 if (regno != ARM_VFP_FPCXT_NS && !vfp_access_check(s)) {
272 return FPSysRegCheckDone;
273 }
274 return FPSysRegCheckContinue;
275}
276
277static void gen_branch_fpInactive(DisasContext *s, TCGCond cond,
278 TCGLabel *label)
279{
280 /*
281 * FPCXT_NS is a special case: it has specific handling for
282 * "current FP state is inactive", and must do the PreserveFPState()
283 * but not the usual full set of actions done by ExecuteFPCheck().
284 * We don't have a TB flag that matches the fpInactive check, so we
285 * do it at runtime as we don't expect FPCXT_NS accesses to be frequent.
286 *
287 * Emit code that checks fpInactive and does a conditional
288 * branch to label based on it:
289 * if cond is TCG_COND_NE then branch if fpInactive != 0 (ie if inactive)
290 * if cond is TCG_COND_EQ then branch if fpInactive == 0 (ie if active)
291 */
292 assert(cond == TCG_COND_EQ || cond == TCG_COND_NE);
293
294 /* fpInactive = FPCCR_NS.ASPEN == 1 && CONTROL.FPCA == 0 */
295 TCGv_i32 aspen, fpca;
296 aspen = load_cpu_field(v7m.fpccr[M_REG_NS]);
297 fpca = load_cpu_field(v7m.control[M_REG_S]);
298 tcg_gen_andi_i32(aspen, aspen, R_V7M_FPCCR_ASPEN_MASK);
299 tcg_gen_xori_i32(aspen, aspen, R_V7M_FPCCR_ASPEN_MASK);
300 tcg_gen_andi_i32(fpca, fpca, R_V7M_CONTROL_FPCA_MASK);
301 tcg_gen_or_i32(fpca, fpca, aspen);
302 tcg_gen_brcondi_i32(tcg_invert_cond(cond), fpca, 0, label);
fa856736
PM
303}
304
305static bool gen_M_fp_sysreg_write(DisasContext *s, int regno,
306 fp_sysreg_loadfn *loadfn,
307 void *opaque)
308{
309 /* Do a write to an M-profile floating point system register */
310 TCGv_i32 tmp;
311 TCGLabel *lab_end = NULL;
312
313 switch (fp_sysreg_checks(s, regno)) {
314 case FPSysRegCheckFailed:
315 return false;
316 case FPSysRegCheckDone:
317 return true;
318 case FPSysRegCheckContinue:
319 break;
320 }
321
322 switch (regno) {
323 case ARM_VFP_FPSCR:
e494cd0a 324 tmp = loadfn(s, opaque, true);
fa856736 325 gen_helper_vfp_set_fpscr(cpu_env, tmp);
fa856736
PM
326 gen_lookup_tb(s);
327 break;
328 case ARM_VFP_FPSCR_NZCVQC:
329 {
330 TCGv_i32 fpscr;
e494cd0a 331 tmp = loadfn(s, opaque, true);
fa856736
PM
332 if (dc_isar_feature(aa32_mve, s)) {
333 /* QC is only present for MVE; otherwise RES0 */
334 TCGv_i32 qc = tcg_temp_new_i32();
335 tcg_gen_andi_i32(qc, tmp, FPCR_QC);
336 /*
337 * The 4 vfp.qc[] fields need only be "zero" vs "non-zero";
338 * here writing the same value into all elements is simplest.
339 */
340 tcg_gen_gvec_dup_i32(MO_32, offsetof(CPUARMState, vfp.qc),
341 16, 16, qc);
342 }
343 tcg_gen_andi_i32(tmp, tmp, FPCR_NZCV_MASK);
344 fpscr = load_cpu_field(vfp.xregs[ARM_VFP_FPSCR]);
345 tcg_gen_andi_i32(fpscr, fpscr, ~FPCR_NZCV_MASK);
346 tcg_gen_or_i32(fpscr, fpscr, tmp);
347 store_cpu_field(fpscr, vfp.xregs[ARM_VFP_FPSCR]);
fa856736
PM
348 break;
349 }
350 case ARM_VFP_FPCXT_NS:
e494cd0a
PM
351 {
352 TCGLabel *lab_active = gen_new_label();
353
fa856736 354 lab_end = gen_new_label();
e494cd0a
PM
355 gen_branch_fpInactive(s, TCG_COND_EQ, lab_active);
356 /*
357 * fpInactive case: write is a NOP, so only do side effects
358 * like register writeback before we branch to end
359 */
360 loadfn(s, opaque, false);
361 tcg_gen_br(lab_end);
362
363 gen_set_label(lab_active);
fa856736
PM
364 /*
365 * !fpInactive: if FPU disabled, take NOCP exception;
366 * otherwise PreserveFPState(), and then FPCXT_NS writes
367 * behave the same as FPCXT_S writes.
368 */
88137f78 369 if (!vfp_access_check_m(s, true)) {
fa856736
PM
370 /*
371 * This was only a conditional exception, so override
8c5d24dc 372 * gen_exception_insn_el()'s default to DISAS_NORETURN
fa856736
PM
373 */
374 s->base.is_jmp = DISAS_NEXT;
375 break;
376 }
e494cd0a
PM
377 }
378 /* fall through */
fa856736
PM
379 case ARM_VFP_FPCXT_S:
380 {
381 TCGv_i32 sfpa, control;
382 /*
383 * Set FPSCR and CONTROL.SFPA from value; the new FPSCR takes
384 * bits [27:0] from value and zeroes bits [31:28].
385 */
e494cd0a 386 tmp = loadfn(s, opaque, true);
fa856736
PM
387 sfpa = tcg_temp_new_i32();
388 tcg_gen_shri_i32(sfpa, tmp, 31);
389 control = load_cpu_field(v7m.control[M_REG_S]);
390 tcg_gen_deposit_i32(control, control, sfpa,
391 R_V7M_CONTROL_SFPA_SHIFT, 1);
392 store_cpu_field(control, v7m.control[M_REG_S]);
393 tcg_gen_andi_i32(tmp, tmp, ~FPCR_NZCV_MASK);
394 gen_helper_vfp_set_fpscr(cpu_env, tmp);
26702213 395 s->base.is_jmp = DISAS_UPDATE_NOCHAIN;
fa856736
PM
396 break;
397 }
398 case ARM_VFP_VPR:
399 /* Behaves as NOP if not privileged */
400 if (IS_USER(s)) {
e494cd0a 401 loadfn(s, opaque, false);
fa856736
PM
402 break;
403 }
e494cd0a 404 tmp = loadfn(s, opaque, true);
fa856736 405 store_cpu_field(tmp, v7m.vpr);
26702213 406 s->base.is_jmp = DISAS_UPDATE_NOCHAIN;
fa856736
PM
407 break;
408 case ARM_VFP_P0:
409 {
410 TCGv_i32 vpr;
e494cd0a 411 tmp = loadfn(s, opaque, true);
fa856736
PM
412 vpr = load_cpu_field(v7m.vpr);
413 tcg_gen_deposit_i32(vpr, vpr, tmp,
414 R_V7M_VPR_P0_SHIFT, R_V7M_VPR_P0_LENGTH);
415 store_cpu_field(vpr, v7m.vpr);
26702213 416 s->base.is_jmp = DISAS_UPDATE_NOCHAIN;
fa856736
PM
417 break;
418 }
419 default:
420 g_assert_not_reached();
421 }
422 if (lab_end) {
423 gen_set_label(lab_end);
424 }
425 return true;
426}
427
428static bool gen_M_fp_sysreg_read(DisasContext *s, int regno,
429 fp_sysreg_storefn *storefn,
430 void *opaque)
431{
432 /* Do a read from an M-profile floating point system register */
433 TCGv_i32 tmp;
434 TCGLabel *lab_end = NULL;
435 bool lookup_tb = false;
436
437 switch (fp_sysreg_checks(s, regno)) {
438 case FPSysRegCheckFailed:
439 return false;
440 case FPSysRegCheckDone:
441 return true;
442 case FPSysRegCheckContinue:
443 break;
444 }
445
446 if (regno == ARM_VFP_FPSCR_NZCVQC && !dc_isar_feature(aa32_mve, s)) {
447 /* QC is RES0 without MVE, so NZCVQC simplifies to NZCV */
448 regno = QEMU_VFP_FPSCR_NZCV;
449 }
450
451 switch (regno) {
452 case ARM_VFP_FPSCR:
453 tmp = tcg_temp_new_i32();
454 gen_helper_vfp_get_fpscr(tmp, cpu_env);
e494cd0a 455 storefn(s, opaque, tmp, true);
fa856736
PM
456 break;
457 case ARM_VFP_FPSCR_NZCVQC:
458 tmp = tcg_temp_new_i32();
459 gen_helper_vfp_get_fpscr(tmp, cpu_env);
460 tcg_gen_andi_i32(tmp, tmp, FPCR_NZCVQC_MASK);
e494cd0a 461 storefn(s, opaque, tmp, true);
fa856736
PM
462 break;
463 case QEMU_VFP_FPSCR_NZCV:
464 /*
465 * Read just NZCV; this is a special case to avoid the
466 * helper call for the "VMRS to CPSR.NZCV" insn.
467 */
468 tmp = load_cpu_field(vfp.xregs[ARM_VFP_FPSCR]);
469 tcg_gen_andi_i32(tmp, tmp, FPCR_NZCV_MASK);
e494cd0a 470 storefn(s, opaque, tmp, true);
fa856736
PM
471 break;
472 case ARM_VFP_FPCXT_S:
473 {
474 TCGv_i32 control, sfpa, fpscr;
475 /* Bits [27:0] from FPSCR, bit [31] from CONTROL.SFPA */
476 tmp = tcg_temp_new_i32();
477 sfpa = tcg_temp_new_i32();
478 gen_helper_vfp_get_fpscr(tmp, cpu_env);
479 tcg_gen_andi_i32(tmp, tmp, ~FPCR_NZCV_MASK);
480 control = load_cpu_field(v7m.control[M_REG_S]);
481 tcg_gen_andi_i32(sfpa, control, R_V7M_CONTROL_SFPA_MASK);
482 tcg_gen_shli_i32(sfpa, sfpa, 31 - R_V7M_CONTROL_SFPA_SHIFT);
483 tcg_gen_or_i32(tmp, tmp, sfpa);
fa856736
PM
484 /*
485 * Store result before updating FPSCR etc, in case
486 * it is a memory write which causes an exception.
487 */
e494cd0a 488 storefn(s, opaque, tmp, true);
fa856736
PM
489 /*
490 * Now we must reset FPSCR from FPDSCR_NS, and clear
491 * CONTROL.SFPA; so we'll end the TB here.
492 */
493 tcg_gen_andi_i32(control, control, ~R_V7M_CONTROL_SFPA_MASK);
494 store_cpu_field(control, v7m.control[M_REG_S]);
495 fpscr = load_cpu_field(v7m.fpdscr[M_REG_NS]);
496 gen_helper_vfp_set_fpscr(cpu_env, fpscr);
fa856736
PM
497 lookup_tb = true;
498 break;
499 }
500 case ARM_VFP_FPCXT_NS:
501 {
01d90db5 502 TCGv_i32 control, sfpa, fpscr, fpdscr;
fa856736
PM
503 TCGLabel *lab_active = gen_new_label();
504
505 lookup_tb = true;
506
507 gen_branch_fpInactive(s, TCG_COND_EQ, lab_active);
508 /* fpInactive case: reads as FPDSCR_NS */
509 TCGv_i32 tmp = load_cpu_field(v7m.fpdscr[M_REG_NS]);
e494cd0a 510 storefn(s, opaque, tmp, true);
fa856736
PM
511 lab_end = gen_new_label();
512 tcg_gen_br(lab_end);
513
514 gen_set_label(lab_active);
515 /*
516 * !fpInactive: if FPU disabled, take NOCP exception;
517 * otherwise PreserveFPState(), and then FPCXT_NS
518 * reads the same as FPCXT_S.
519 */
88137f78 520 if (!vfp_access_check_m(s, true)) {
fa856736
PM
521 /*
522 * This was only a conditional exception, so override
8c5d24dc 523 * gen_exception_insn_el()'s default to DISAS_NORETURN
fa856736
PM
524 */
525 s->base.is_jmp = DISAS_NEXT;
526 break;
527 }
fa856736
PM
528 tmp = tcg_temp_new_i32();
529 sfpa = tcg_temp_new_i32();
530 fpscr = tcg_temp_new_i32();
531 gen_helper_vfp_get_fpscr(fpscr, cpu_env);
532 tcg_gen_andi_i32(tmp, fpscr, ~FPCR_NZCV_MASK);
533 control = load_cpu_field(v7m.control[M_REG_S]);
534 tcg_gen_andi_i32(sfpa, control, R_V7M_CONTROL_SFPA_MASK);
535 tcg_gen_shli_i32(sfpa, sfpa, 31 - R_V7M_CONTROL_SFPA_SHIFT);
536 tcg_gen_or_i32(tmp, tmp, sfpa);
fa856736 537 /* Store result before updating FPSCR, in case it faults */
e494cd0a 538 storefn(s, opaque, tmp, true);
fa856736
PM
539 /* If SFPA is zero then set FPSCR from FPDSCR_NS */
540 fpdscr = load_cpu_field(v7m.fpdscr[M_REG_NS]);
01d90db5
RH
541 tcg_gen_movcond_i32(TCG_COND_EQ, fpscr, sfpa, tcg_constant_i32(0),
542 fpdscr, fpscr);
fa856736 543 gen_helper_vfp_set_fpscr(cpu_env, fpscr);
fa856736
PM
544 break;
545 }
546 case ARM_VFP_VPR:
547 /* Behaves as NOP if not privileged */
548 if (IS_USER(s)) {
e494cd0a 549 storefn(s, opaque, NULL, false);
fa856736
PM
550 break;
551 }
552 tmp = load_cpu_field(v7m.vpr);
e494cd0a 553 storefn(s, opaque, tmp, true);
fa856736
PM
554 break;
555 case ARM_VFP_P0:
556 tmp = load_cpu_field(v7m.vpr);
557 tcg_gen_extract_i32(tmp, tmp, R_V7M_VPR_P0_SHIFT, R_V7M_VPR_P0_LENGTH);
e494cd0a 558 storefn(s, opaque, tmp, true);
fa856736
PM
559 break;
560 default:
561 g_assert_not_reached();
562 }
563
564 if (lab_end) {
565 gen_set_label(lab_end);
566 }
567 if (lookup_tb) {
568 gen_lookup_tb(s);
569 }
570 return true;
571}
572
e494cd0a
PM
573static void fp_sysreg_to_gpr(DisasContext *s, void *opaque, TCGv_i32 value,
574 bool do_access)
fa856736
PM
575{
576 arg_VMSR_VMRS *a = opaque;
577
e494cd0a
PM
578 if (!do_access) {
579 return;
580 }
581
fa856736
PM
582 if (a->rt == 15) {
583 /* Set the 4 flag bits in the CPSR */
584 gen_set_nzcv(value);
fa856736
PM
585 } else {
586 store_reg(s, a->rt, value);
587 }
588}
589
e494cd0a 590static TCGv_i32 gpr_to_fp_sysreg(DisasContext *s, void *opaque, bool do_access)
fa856736
PM
591{
592 arg_VMSR_VMRS *a = opaque;
593
e494cd0a
PM
594 if (!do_access) {
595 return NULL;
596 }
fa856736
PM
597 return load_reg(s, a->rt);
598}
599
600static bool trans_VMSR_VMRS(DisasContext *s, arg_VMSR_VMRS *a)
601{
602 /*
603 * Accesses to R15 are UNPREDICTABLE; we choose to undef.
604 * FPSCR -> r15 is a special case which writes to the PSR flags;
605 * set a->reg to a special value to tell gen_M_fp_sysreg_read()
606 * we only care about the top 4 bits of FPSCR there.
607 */
608 if (a->rt == 15) {
609 if (a->l && a->reg == ARM_VFP_FPSCR) {
610 a->reg = QEMU_VFP_FPSCR_NZCV;
611 } else {
612 return false;
613 }
614 }
615
616 if (a->l) {
617 /* VMRS, move FP system register to gp register */
618 return gen_M_fp_sysreg_read(s, a->reg, fp_sysreg_to_gpr, a);
619 } else {
620 /* VMSR, move gp register to FP system register */
621 return gen_M_fp_sysreg_write(s, a->reg, gpr_to_fp_sysreg, a);
622 }
623}
624
e494cd0a
PM
625static void fp_sysreg_to_memory(DisasContext *s, void *opaque, TCGv_i32 value,
626 bool do_access)
fa856736
PM
627{
628 arg_vldr_sysreg *a = opaque;
629 uint32_t offset = a->imm;
630 TCGv_i32 addr;
631
632 if (!a->a) {
633 offset = -offset;
634 }
635
e494cd0a
PM
636 if (!do_access && !a->w) {
637 return;
638 }
639
fa856736
PM
640 addr = load_reg(s, a->rn);
641 if (a->p) {
642 tcg_gen_addi_i32(addr, addr, offset);
643 }
644
645 if (s->v8m_stackcheck && a->rn == 13 && a->w) {
646 gen_helper_v8m_stackcheck(cpu_env, addr);
647 }
648
e494cd0a
PM
649 if (do_access) {
650 gen_aa32_st_i32(s, value, addr, get_mem_index(s),
651 MO_UL | MO_ALIGN | s->be_data);
e494cd0a 652 }
fa856736
PM
653
654 if (a->w) {
655 /* writeback */
656 if (!a->p) {
657 tcg_gen_addi_i32(addr, addr, offset);
658 }
659 store_reg(s, a->rn, addr);
fa856736
PM
660 }
661}
662
e494cd0a
PM
663static TCGv_i32 memory_to_fp_sysreg(DisasContext *s, void *opaque,
664 bool do_access)
fa856736
PM
665{
666 arg_vldr_sysreg *a = opaque;
667 uint32_t offset = a->imm;
668 TCGv_i32 addr;
e494cd0a 669 TCGv_i32 value = NULL;
fa856736
PM
670
671 if (!a->a) {
672 offset = -offset;
673 }
674
e494cd0a
PM
675 if (!do_access && !a->w) {
676 return NULL;
677 }
678
fa856736
PM
679 addr = load_reg(s, a->rn);
680 if (a->p) {
681 tcg_gen_addi_i32(addr, addr, offset);
682 }
683
684 if (s->v8m_stackcheck && a->rn == 13 && a->w) {
685 gen_helper_v8m_stackcheck(cpu_env, addr);
686 }
687
e494cd0a
PM
688 if (do_access) {
689 value = tcg_temp_new_i32();
690 gen_aa32_ld_i32(s, value, addr, get_mem_index(s),
691 MO_UL | MO_ALIGN | s->be_data);
692 }
fa856736
PM
693
694 if (a->w) {
695 /* writeback */
696 if (!a->p) {
697 tcg_gen_addi_i32(addr, addr, offset);
698 }
699 store_reg(s, a->rn, addr);
fa856736
PM
700 }
701 return value;
702}
703
704static bool trans_VLDR_sysreg(DisasContext *s, arg_vldr_sysreg *a)
705{
706 if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
707 return false;
708 }
709 if (a->rn == 15) {
710 return false;
711 }
712 return gen_M_fp_sysreg_write(s, a->reg, memory_to_fp_sysreg, a);
713}
714
715static bool trans_VSTR_sysreg(DisasContext *s, arg_vldr_sysreg *a)
716{
717 if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
718 return false;
719 }
720 if (a->rn == 15) {
721 return false;
722 }
723 return gen_M_fp_sysreg_read(s, a->reg, fp_sysreg_to_memory, a);
724}
725
9a5071ab
PM
726static bool trans_NOCP(DisasContext *s, arg_nocp *a)
727{
728 /*
729 * Handle M-profile early check for disabled coprocessor:
730 * all we need to do here is emit the NOCP exception if
731 * the coprocessor is disabled. Otherwise we return false
732 * and the real VFP/etc decode will handle the insn.
733 */
734 assert(arm_dc_feature(s, ARM_FEATURE_M));
735
736 if (a->cp == 11) {
737 a->cp = 10;
738 }
739 if (arm_dc_feature(s, ARM_FEATURE_V8_1M) &&
740 (a->cp == 8 || a->cp == 9 || a->cp == 14 || a->cp == 15)) {
741 /* in v8.1M cp 8, 9, 14, 15 also are governed by the cp10 enable */
742 a->cp = 10;
743 }
744
745 if (a->cp != 10) {
55086e62 746 gen_exception_insn(s, 0, EXCP_NOCP, syn_uncategorized());
9a5071ab
PM
747 return true;
748 }
749
750 if (s->fp_excp_el != 0) {
55086e62 751 gen_exception_insn_el(s, 0, EXCP_NOCP,
8c5d24dc 752 syn_uncategorized(), s->fp_excp_el);
9a5071ab
PM
753 return true;
754 }
755
756 return false;
757}
758
759static bool trans_NOCP_8_1(DisasContext *s, arg_nocp *a)
760{
761 /* This range needs a coprocessor check for v8.1M and later only */
762 if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
763 return false;
764 }
765 return trans_NOCP(s, a);
766}