4 * Copyright (c) 2016-2017 Sagar Karandikar, sagark@eecs.berkeley.edu
5 * Copyright (c) 2017-2018 SiFive, Inc.
7 * This program is free software; you can redistribute it and/or modify it
8 * under the terms and conditions of the GNU General Public License,
9 * version 2 or later, as published by the Free Software Foundation.
11 * This program is distributed in the hope it will be useful, but WITHOUT
12 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
13 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
16 * You should have received a copy of the GNU General Public License along with
17 * this program. If not, see <http://www.gnu.org/licenses/>.
20 #include "qemu/osdep.h"
21 #include "qemu/qemu-print.h"
22 #include "qemu/ctype.h"
25 #include "internals.h"
26 #include "exec/exec-all.h"
27 #include "qapi/error.h"
28 #include "qemu/error-report.h"
29 #include "hw/qdev-properties.h"
30 #include "migration/vmstate.h"
31 #include "fpu/softfloat-helpers.h"
32 #include "sysemu/kvm.h"
33 #include "kvm_riscv.h"
35 /* RISC-V CPU definitions */
37 static const char riscv_exts
[26] = "IEMAFDQCLBJTPVNSUHKORWXYZG";
39 const char * const riscv_int_regnames
[] = {
40 "x0/zero", "x1/ra", "x2/sp", "x3/gp", "x4/tp", "x5/t0", "x6/t1",
41 "x7/t2", "x8/s0", "x9/s1", "x10/a0", "x11/a1", "x12/a2", "x13/a3",
42 "x14/a4", "x15/a5", "x16/a6", "x17/a7", "x18/s2", "x19/s3", "x20/s4",
43 "x21/s5", "x22/s6", "x23/s7", "x24/s8", "x25/s9", "x26/s10", "x27/s11",
44 "x28/t3", "x29/t4", "x30/t5", "x31/t6"
47 const char * const riscv_int_regnamesh
[] = {
48 "x0h/zeroh", "x1h/rah", "x2h/sph", "x3h/gph", "x4h/tph", "x5h/t0h",
49 "x6h/t1h", "x7h/t2h", "x8h/s0h", "x9h/s1h", "x10h/a0h", "x11h/a1h",
50 "x12h/a2h", "x13h/a3h", "x14h/a4h", "x15h/a5h", "x16h/a6h", "x17h/a7h",
51 "x18h/s2h", "x19h/s3h", "x20h/s4h", "x21h/s5h", "x22h/s6h", "x23h/s7h",
52 "x24h/s8h", "x25h/s9h", "x26h/s10h", "x27h/s11h", "x28h/t3h", "x29h/t4h",
53 "x30h/t5h", "x31h/t6h"
56 const char * const riscv_fpr_regnames
[] = {
57 "f0/ft0", "f1/ft1", "f2/ft2", "f3/ft3", "f4/ft4", "f5/ft5",
58 "f6/ft6", "f7/ft7", "f8/fs0", "f9/fs1", "f10/fa0", "f11/fa1",
59 "f12/fa2", "f13/fa3", "f14/fa4", "f15/fa5", "f16/fa6", "f17/fa7",
60 "f18/fs2", "f19/fs3", "f20/fs4", "f21/fs5", "f22/fs6", "f23/fs7",
61 "f24/fs8", "f25/fs9", "f26/fs10", "f27/fs11", "f28/ft8", "f29/ft9",
62 "f30/ft10", "f31/ft11"
65 static const char * const riscv_excp_names
[] = {
68 "illegal_instruction",
86 "guest_exec_page_fault",
87 "guest_load_page_fault",
89 "guest_store_page_fault",
92 static const char * const riscv_intr_names
[] = {
111 const char *riscv_cpu_get_trap_name(target_ulong cause
, bool async
)
114 return (cause
< ARRAY_SIZE(riscv_intr_names
)) ?
115 riscv_intr_names
[cause
] : "(unknown)";
117 return (cause
< ARRAY_SIZE(riscv_excp_names
)) ?
118 riscv_excp_names
[cause
] : "(unknown)";
122 static void set_misa(CPURISCVState
*env
, RISCVMXL mxl
, uint32_t ext
)
124 env
->misa_mxl_max
= env
->misa_mxl
= mxl
;
125 env
->misa_ext_mask
= env
->misa_ext
= ext
;
128 static void set_priv_version(CPURISCVState
*env
, int priv_ver
)
130 env
->priv_ver
= priv_ver
;
133 static void set_vext_version(CPURISCVState
*env
, int vext_ver
)
135 env
->vext_ver
= vext_ver
;
138 static void set_feature(CPURISCVState
*env
, int feature
)
140 env
->features
|= (1ULL << feature
);
143 static void set_resetvec(CPURISCVState
*env
, target_ulong resetvec
)
145 #ifndef CONFIG_USER_ONLY
146 env
->resetvec
= resetvec
;
150 static void riscv_any_cpu_init(Object
*obj
)
152 CPURISCVState
*env
= &RISCV_CPU(obj
)->env
;
153 #if defined(TARGET_RISCV32)
154 set_misa(env
, MXL_RV32
, RVI
| RVM
| RVA
| RVF
| RVD
| RVC
| RVU
);
155 #elif defined(TARGET_RISCV64)
156 set_misa(env
, MXL_RV64
, RVI
| RVM
| RVA
| RVF
| RVD
| RVC
| RVU
);
158 set_priv_version(env
, PRIV_VERSION_1_11_0
);
161 #if defined(TARGET_RISCV64)
162 static void rv64_base_cpu_init(Object
*obj
)
164 CPURISCVState
*env
= &RISCV_CPU(obj
)->env
;
165 /* We set this in the realise function */
166 set_misa(env
, MXL_RV64
, 0);
169 static void rv64_sifive_u_cpu_init(Object
*obj
)
171 CPURISCVState
*env
= &RISCV_CPU(obj
)->env
;
172 set_misa(env
, MXL_RV64
, RVI
| RVM
| RVA
| RVF
| RVD
| RVC
| RVS
| RVU
);
173 set_priv_version(env
, PRIV_VERSION_1_10_0
);
176 static void rv64_sifive_e_cpu_init(Object
*obj
)
178 CPURISCVState
*env
= &RISCV_CPU(obj
)->env
;
179 set_misa(env
, MXL_RV64
, RVI
| RVM
| RVA
| RVC
| RVU
);
180 set_priv_version(env
, PRIV_VERSION_1_10_0
);
181 qdev_prop_set_bit(DEVICE(obj
), "mmu", false);
184 static void rv128_base_cpu_init(Object
*obj
)
186 if (qemu_tcg_mttcg_enabled()) {
187 /* Missing 128-bit aligned atomics */
188 error_report("128-bit RISC-V currently does not work with Multi "
189 "Threaded TCG. Please use: -accel tcg,thread=single");
192 CPURISCVState
*env
= &RISCV_CPU(obj
)->env
;
193 /* We set this in the realise function */
194 set_misa(env
, MXL_RV128
, 0);
197 static void rv32_base_cpu_init(Object
*obj
)
199 CPURISCVState
*env
= &RISCV_CPU(obj
)->env
;
200 /* We set this in the realise function */
201 set_misa(env
, MXL_RV32
, 0);
204 static void rv32_sifive_u_cpu_init(Object
*obj
)
206 CPURISCVState
*env
= &RISCV_CPU(obj
)->env
;
207 set_misa(env
, MXL_RV32
, RVI
| RVM
| RVA
| RVF
| RVD
| RVC
| RVS
| RVU
);
208 set_priv_version(env
, PRIV_VERSION_1_10_0
);
211 static void rv32_sifive_e_cpu_init(Object
*obj
)
213 CPURISCVState
*env
= &RISCV_CPU(obj
)->env
;
214 set_misa(env
, MXL_RV32
, RVI
| RVM
| RVA
| RVC
| RVU
);
215 set_priv_version(env
, PRIV_VERSION_1_10_0
);
216 qdev_prop_set_bit(DEVICE(obj
), "mmu", false);
219 static void rv32_ibex_cpu_init(Object
*obj
)
221 CPURISCVState
*env
= &RISCV_CPU(obj
)->env
;
222 set_misa(env
, MXL_RV32
, RVI
| RVM
| RVC
| RVU
);
223 set_priv_version(env
, PRIV_VERSION_1_10_0
);
224 qdev_prop_set_bit(DEVICE(obj
), "mmu", false);
225 qdev_prop_set_bit(DEVICE(obj
), "x-epmp", true);
228 static void rv32_imafcu_nommu_cpu_init(Object
*obj
)
230 CPURISCVState
*env
= &RISCV_CPU(obj
)->env
;
231 set_misa(env
, MXL_RV32
, RVI
| RVM
| RVA
| RVF
| RVC
| RVU
);
232 set_priv_version(env
, PRIV_VERSION_1_10_0
);
233 set_resetvec(env
, DEFAULT_RSTVEC
);
234 qdev_prop_set_bit(DEVICE(obj
), "mmu", false);
238 #if defined(CONFIG_KVM)
239 static void riscv_host_cpu_init(Object
*obj
)
241 CPURISCVState
*env
= &RISCV_CPU(obj
)->env
;
242 #if defined(TARGET_RISCV32)
243 set_misa(env
, MXL_RV32
, 0);
244 #elif defined(TARGET_RISCV64)
245 set_misa(env
, MXL_RV64
, 0);
250 static ObjectClass
*riscv_cpu_class_by_name(const char *cpu_model
)
256 cpuname
= g_strsplit(cpu_model
, ",", 1);
257 typename
= g_strdup_printf(RISCV_CPU_TYPE_NAME("%s"), cpuname
[0]);
258 oc
= object_class_by_name(typename
);
261 if (!oc
|| !object_class_dynamic_cast(oc
, TYPE_RISCV_CPU
) ||
262 object_class_is_abstract(oc
)) {
268 static void riscv_cpu_dump_state(CPUState
*cs
, FILE *f
, int flags
)
270 RISCVCPU
*cpu
= RISCV_CPU(cs
);
271 CPURISCVState
*env
= &cpu
->env
;
274 #if !defined(CONFIG_USER_ONLY)
275 if (riscv_has_ext(env
, RVH
)) {
276 qemu_fprintf(f
, " %s %d\n", "V = ", riscv_cpu_virt_enabled(env
));
279 qemu_fprintf(f
, " %s " TARGET_FMT_lx
"\n", "pc ", env
->pc
);
280 #ifndef CONFIG_USER_ONLY
282 static const int dump_csrs
[] = {
319 for (int i
= 0; i
< ARRAY_SIZE(dump_csrs
); ++i
) {
320 int csrno
= dump_csrs
[i
];
321 target_ulong val
= 0;
322 RISCVException res
= riscv_csrrw_debug(env
, csrno
, &val
, 0, 0);
325 * Rely on the smode, hmode, etc, predicates within csr.c
326 * to do the filtering of the registers that are present.
328 if (res
== RISCV_EXCP_NONE
) {
329 qemu_fprintf(f
, " %-8s " TARGET_FMT_lx
"\n",
330 csr_ops
[csrno
].name
, val
);
336 for (i
= 0; i
< 32; i
++) {
337 qemu_fprintf(f
, " %-8s " TARGET_FMT_lx
,
338 riscv_int_regnames
[i
], env
->gpr
[i
]);
340 qemu_fprintf(f
, "\n");
343 if (flags
& CPU_DUMP_FPU
) {
344 for (i
= 0; i
< 32; i
++) {
345 qemu_fprintf(f
, " %-8s %016" PRIx64
,
346 riscv_fpr_regnames
[i
], env
->fpr
[i
]);
348 qemu_fprintf(f
, "\n");
354 static void riscv_cpu_set_pc(CPUState
*cs
, vaddr value
)
356 RISCVCPU
*cpu
= RISCV_CPU(cs
);
357 CPURISCVState
*env
= &cpu
->env
;
359 if (env
->xl
== MXL_RV32
) {
360 env
->pc
= (int32_t)value
;
366 static void riscv_cpu_synchronize_from_tb(CPUState
*cs
,
367 const TranslationBlock
*tb
)
369 RISCVCPU
*cpu
= RISCV_CPU(cs
);
370 CPURISCVState
*env
= &cpu
->env
;
371 RISCVMXL xl
= FIELD_EX32(tb
->flags
, TB_FLAGS
, XL
);
373 if (xl
== MXL_RV32
) {
374 env
->pc
= (int32_t)tb
->pc
;
380 static bool riscv_cpu_has_work(CPUState
*cs
)
382 #ifndef CONFIG_USER_ONLY
383 RISCVCPU
*cpu
= RISCV_CPU(cs
);
384 CPURISCVState
*env
= &cpu
->env
;
386 * Definition of the WFI instruction requires it to ignore the privilege
387 * mode and delegation registers, but respect individual enables
389 return (env
->mip
& env
->mie
) != 0;
395 void restore_state_to_opc(CPURISCVState
*env
, TranslationBlock
*tb
,
398 RISCVMXL xl
= FIELD_EX32(tb
->flags
, TB_FLAGS
, XL
);
399 if (xl
== MXL_RV32
) {
400 env
->pc
= (int32_t)data
[0];
406 static void riscv_cpu_reset(DeviceState
*dev
)
408 CPUState
*cs
= CPU(dev
);
409 RISCVCPU
*cpu
= RISCV_CPU(cs
);
410 RISCVCPUClass
*mcc
= RISCV_CPU_GET_CLASS(cpu
);
411 CPURISCVState
*env
= &cpu
->env
;
413 mcc
->parent_reset(dev
);
414 #ifndef CONFIG_USER_ONLY
415 env
->misa_mxl
= env
->misa_mxl_max
;
417 env
->mstatus
&= ~(MSTATUS_MIE
| MSTATUS_MPRV
);
418 if (env
->misa_mxl
> MXL_RV32
) {
420 * The reset status of SXL/UXL is undefined, but mstatus is WARL
421 * and we must ensure that the value after init is valid for read.
423 env
->mstatus
= set_field(env
->mstatus
, MSTATUS64_SXL
, env
->misa_mxl
);
424 env
->mstatus
= set_field(env
->mstatus
, MSTATUS64_UXL
, env
->misa_mxl
);
425 if (riscv_has_ext(env
, RVH
)) {
426 env
->vsstatus
= set_field(env
->vsstatus
,
427 MSTATUS64_SXL
, env
->misa_mxl
);
428 env
->vsstatus
= set_field(env
->vsstatus
,
429 MSTATUS64_UXL
, env
->misa_mxl
);
430 env
->mstatus_hs
= set_field(env
->mstatus_hs
,
431 MSTATUS64_SXL
, env
->misa_mxl
);
432 env
->mstatus_hs
= set_field(env
->mstatus_hs
,
433 MSTATUS64_UXL
, env
->misa_mxl
);
437 env
->pc
= env
->resetvec
;
438 env
->two_stage_lookup
= false;
439 /* mmte is supposed to have pm.current hardwired to 1 */
440 env
->mmte
|= (PM_EXT_INITIAL
| MMTE_M_PM_CURRENT
);
442 env
->xl
= riscv_cpu_mxl(env
);
443 riscv_cpu_update_mask(env
);
444 cs
->exception_index
= RISCV_EXCP_NONE
;
446 set_default_nan_mode(1, &env
->fp_status
);
448 #ifndef CONFIG_USER_ONLY
450 kvm_riscv_reset_vcpu(cpu
);
455 static void riscv_cpu_disas_set_info(CPUState
*s
, disassemble_info
*info
)
457 RISCVCPU
*cpu
= RISCV_CPU(s
);
459 switch (riscv_cpu_mxl(&cpu
->env
)) {
461 info
->print_insn
= print_insn_riscv32
;
464 info
->print_insn
= print_insn_riscv64
;
467 info
->print_insn
= print_insn_riscv128
;
470 g_assert_not_reached();
474 static void riscv_cpu_realize(DeviceState
*dev
, Error
**errp
)
476 CPUState
*cs
= CPU(dev
);
477 RISCVCPU
*cpu
= RISCV_CPU(dev
);
478 CPURISCVState
*env
= &cpu
->env
;
479 RISCVCPUClass
*mcc
= RISCV_CPU_GET_CLASS(dev
);
480 CPUClass
*cc
= CPU_CLASS(mcc
);
481 int priv_version
= 0;
482 Error
*local_err
= NULL
;
484 cpu_exec_realizefn(cs
, &local_err
);
485 if (local_err
!= NULL
) {
486 error_propagate(errp
, local_err
);
490 if (cpu
->cfg
.priv_spec
) {
491 if (!g_strcmp0(cpu
->cfg
.priv_spec
, "v1.11.0")) {
492 priv_version
= PRIV_VERSION_1_11_0
;
493 } else if (!g_strcmp0(cpu
->cfg
.priv_spec
, "v1.10.0")) {
494 priv_version
= PRIV_VERSION_1_10_0
;
497 "Unsupported privilege spec version '%s'",
504 set_priv_version(env
, priv_version
);
505 } else if (!env
->priv_ver
) {
506 set_priv_version(env
, PRIV_VERSION_1_11_0
);
510 set_feature(env
, RISCV_FEATURE_MMU
);
514 set_feature(env
, RISCV_FEATURE_PMP
);
517 * Enhanced PMP should only be available
518 * on harts with PMP support
521 set_feature(env
, RISCV_FEATURE_EPMP
);
525 set_resetvec(env
, cpu
->cfg
.resetvec
);
527 /* Validate that MISA_MXL is set properly. */
528 switch (env
->misa_mxl_max
) {
529 #ifdef TARGET_RISCV64
531 cc
->gdb_core_xml_file
= "riscv-64bit-cpu.xml";
537 cc
->gdb_core_xml_file
= "riscv-32bit-cpu.xml";
540 g_assert_not_reached();
542 assert(env
->misa_mxl_max
== env
->misa_mxl
);
544 /* If only MISA_EXT is unset for misa, then set it from properties */
545 if (env
->misa_ext
== 0) {
548 /* Do some ISA extension error checking */
549 if (cpu
->cfg
.ext_i
&& cpu
->cfg
.ext_e
) {
551 "I and E extensions are incompatible");
555 if (!cpu
->cfg
.ext_i
&& !cpu
->cfg
.ext_e
) {
557 "Either I or E extension must be set");
561 if (cpu
->cfg
.ext_g
&& !(cpu
->cfg
.ext_i
& cpu
->cfg
.ext_m
&
562 cpu
->cfg
.ext_a
& cpu
->cfg
.ext_f
&
564 warn_report("Setting G will also set IMAFD");
565 cpu
->cfg
.ext_i
= true;
566 cpu
->cfg
.ext_m
= true;
567 cpu
->cfg
.ext_a
= true;
568 cpu
->cfg
.ext_f
= true;
569 cpu
->cfg
.ext_d
= true;
572 /* Set the ISA extensions, checks should have happened above */
573 if (cpu
->cfg
.ext_i
) {
576 if (cpu
->cfg
.ext_e
) {
579 if (cpu
->cfg
.ext_m
) {
582 if (cpu
->cfg
.ext_a
) {
585 if (cpu
->cfg
.ext_f
) {
588 if (cpu
->cfg
.ext_d
) {
591 if (cpu
->cfg
.ext_c
) {
594 if (cpu
->cfg
.ext_s
) {
597 if (cpu
->cfg
.ext_u
) {
600 if (cpu
->cfg
.ext_h
) {
603 if (cpu
->cfg
.ext_v
) {
604 int vext_version
= VEXT_VERSION_1_00_0
;
606 if (!is_power_of_2(cpu
->cfg
.vlen
)) {
608 "Vector extension VLEN must be power of 2");
611 if (cpu
->cfg
.vlen
> RV_VLEN_MAX
|| cpu
->cfg
.vlen
< 128) {
613 "Vector extension implementation only supports VLEN "
614 "in the range [128, %d]", RV_VLEN_MAX
);
617 if (!is_power_of_2(cpu
->cfg
.elen
)) {
619 "Vector extension ELEN must be power of 2");
622 if (cpu
->cfg
.elen
> 64 || cpu
->cfg
.vlen
< 8) {
624 "Vector extension implementation only supports ELEN "
625 "in the range [8, 64]");
628 if (cpu
->cfg
.vext_spec
) {
629 if (!g_strcmp0(cpu
->cfg
.vext_spec
, "v1.0")) {
630 vext_version
= VEXT_VERSION_1_00_0
;
633 "Unsupported vector spec version '%s'",
638 qemu_log("vector version is not specified, "
639 "use the default value v1.0\n");
641 set_vext_version(env
, vext_version
);
643 if ((cpu
->cfg
.ext_zve32f
|| cpu
->cfg
.ext_zve64f
) && !cpu
->cfg
.ext_f
) {
644 error_setg(errp
, "Zve32f/Zve64f extension depends upon RVF.");
647 if (cpu
->cfg
.ext_j
) {
651 set_misa(env
, env
->misa_mxl
, ext
);
654 riscv_cpu_register_gdb_regs_for_features(cs
);
659 mcc
->parent_realize(dev
, errp
);
662 #ifndef CONFIG_USER_ONLY
663 static void riscv_cpu_set_irq(void *opaque
, int irq
, int level
)
665 RISCVCPU
*cpu
= RISCV_CPU(opaque
);
681 kvm_riscv_set_irq(cpu
, irq
, level
);
683 riscv_cpu_update_mip(cpu
, 1 << irq
, BOOL_TO_MASK(level
));
687 g_assert_not_reached();
690 #endif /* CONFIG_USER_ONLY */
692 static void riscv_cpu_init(Object
*obj
)
694 RISCVCPU
*cpu
= RISCV_CPU(obj
);
696 cpu_set_cpustate_pointers(cpu
);
698 #ifndef CONFIG_USER_ONLY
699 qdev_init_gpio_in(DEVICE(cpu
), riscv_cpu_set_irq
, 12);
700 #endif /* CONFIG_USER_ONLY */
703 static Property riscv_cpu_properties
[] = {
704 /* Defaults for standard extensions */
705 DEFINE_PROP_BOOL("i", RISCVCPU
, cfg
.ext_i
, true),
706 DEFINE_PROP_BOOL("e", RISCVCPU
, cfg
.ext_e
, false),
707 DEFINE_PROP_BOOL("g", RISCVCPU
, cfg
.ext_g
, true),
708 DEFINE_PROP_BOOL("m", RISCVCPU
, cfg
.ext_m
, true),
709 DEFINE_PROP_BOOL("a", RISCVCPU
, cfg
.ext_a
, true),
710 DEFINE_PROP_BOOL("f", RISCVCPU
, cfg
.ext_f
, true),
711 DEFINE_PROP_BOOL("d", RISCVCPU
, cfg
.ext_d
, true),
712 DEFINE_PROP_BOOL("c", RISCVCPU
, cfg
.ext_c
, true),
713 DEFINE_PROP_BOOL("s", RISCVCPU
, cfg
.ext_s
, true),
714 DEFINE_PROP_BOOL("u", RISCVCPU
, cfg
.ext_u
, true),
715 DEFINE_PROP_BOOL("v", RISCVCPU
, cfg
.ext_v
, false),
716 DEFINE_PROP_BOOL("h", RISCVCPU
, cfg
.ext_h
, true),
717 DEFINE_PROP_BOOL("Counters", RISCVCPU
, cfg
.ext_counters
, true),
718 DEFINE_PROP_BOOL("Zifencei", RISCVCPU
, cfg
.ext_ifencei
, true),
719 DEFINE_PROP_BOOL("Zicsr", RISCVCPU
, cfg
.ext_icsr
, true),
720 DEFINE_PROP_BOOL("Zfh", RISCVCPU
, cfg
.ext_zfh
, false),
721 DEFINE_PROP_BOOL("Zfhmin", RISCVCPU
, cfg
.ext_zfhmin
, false),
722 DEFINE_PROP_BOOL("Zve32f", RISCVCPU
, cfg
.ext_zve32f
, false),
723 DEFINE_PROP_BOOL("Zve64f", RISCVCPU
, cfg
.ext_zve64f
, false),
724 DEFINE_PROP_BOOL("mmu", RISCVCPU
, cfg
.mmu
, true),
725 DEFINE_PROP_BOOL("pmp", RISCVCPU
, cfg
.pmp
, true),
727 DEFINE_PROP_STRING("priv_spec", RISCVCPU
, cfg
.priv_spec
),
728 DEFINE_PROP_STRING("vext_spec", RISCVCPU
, cfg
.vext_spec
),
729 DEFINE_PROP_UINT16("vlen", RISCVCPU
, cfg
.vlen
, 128),
730 DEFINE_PROP_UINT16("elen", RISCVCPU
, cfg
.elen
, 64),
732 DEFINE_PROP_BOOL("zba", RISCVCPU
, cfg
.ext_zba
, true),
733 DEFINE_PROP_BOOL("zbb", RISCVCPU
, cfg
.ext_zbb
, true),
734 DEFINE_PROP_BOOL("zbc", RISCVCPU
, cfg
.ext_zbc
, true),
735 DEFINE_PROP_BOOL("zbs", RISCVCPU
, cfg
.ext_zbs
, true),
737 /* These are experimental so mark with 'x-' */
738 DEFINE_PROP_BOOL("x-j", RISCVCPU
, cfg
.ext_j
, false),
740 DEFINE_PROP_BOOL("x-epmp", RISCVCPU
, cfg
.epmp
, false),
742 DEFINE_PROP_UINT64("resetvec", RISCVCPU
, cfg
.resetvec
, DEFAULT_RSTVEC
),
743 DEFINE_PROP_END_OF_LIST(),
746 static gchar
*riscv_gdb_arch_name(CPUState
*cs
)
748 RISCVCPU
*cpu
= RISCV_CPU(cs
);
749 CPURISCVState
*env
= &cpu
->env
;
751 switch (riscv_cpu_mxl(env
)) {
753 return g_strdup("riscv:rv32");
756 return g_strdup("riscv:rv64");
758 g_assert_not_reached();
762 static const char *riscv_gdb_get_dynamic_xml(CPUState
*cs
, const char *xmlname
)
764 RISCVCPU
*cpu
= RISCV_CPU(cs
);
766 if (strcmp(xmlname
, "riscv-csr.xml") == 0) {
767 return cpu
->dyn_csr_xml
;
768 } else if (strcmp(xmlname
, "riscv-vector.xml") == 0) {
769 return cpu
->dyn_vreg_xml
;
775 #ifndef CONFIG_USER_ONLY
776 #include "hw/core/sysemu-cpu-ops.h"
778 static const struct SysemuCPUOps riscv_sysemu_ops
= {
779 .get_phys_page_debug
= riscv_cpu_get_phys_page_debug
,
780 .write_elf64_note
= riscv_cpu_write_elf64_note
,
781 .write_elf32_note
= riscv_cpu_write_elf32_note
,
782 .legacy_vmsd
= &vmstate_riscv_cpu
,
786 #include "hw/core/tcg-cpu-ops.h"
788 static const struct TCGCPUOps riscv_tcg_ops
= {
789 .initialize
= riscv_translate_init
,
790 .synchronize_from_tb
= riscv_cpu_synchronize_from_tb
,
792 #ifndef CONFIG_USER_ONLY
793 .tlb_fill
= riscv_cpu_tlb_fill
,
794 .cpu_exec_interrupt
= riscv_cpu_exec_interrupt
,
795 .do_interrupt
= riscv_cpu_do_interrupt
,
796 .do_transaction_failed
= riscv_cpu_do_transaction_failed
,
797 .do_unaligned_access
= riscv_cpu_do_unaligned_access
,
798 #endif /* !CONFIG_USER_ONLY */
801 static void riscv_cpu_class_init(ObjectClass
*c
, void *data
)
803 RISCVCPUClass
*mcc
= RISCV_CPU_CLASS(c
);
804 CPUClass
*cc
= CPU_CLASS(c
);
805 DeviceClass
*dc
= DEVICE_CLASS(c
);
807 device_class_set_parent_realize(dc
, riscv_cpu_realize
,
808 &mcc
->parent_realize
);
810 device_class_set_parent_reset(dc
, riscv_cpu_reset
, &mcc
->parent_reset
);
812 cc
->class_by_name
= riscv_cpu_class_by_name
;
813 cc
->has_work
= riscv_cpu_has_work
;
814 cc
->dump_state
= riscv_cpu_dump_state
;
815 cc
->set_pc
= riscv_cpu_set_pc
;
816 cc
->gdb_read_register
= riscv_cpu_gdb_read_register
;
817 cc
->gdb_write_register
= riscv_cpu_gdb_write_register
;
818 cc
->gdb_num_core_regs
= 33;
819 cc
->gdb_stop_before_watchpoint
= true;
820 cc
->disas_set_info
= riscv_cpu_disas_set_info
;
821 #ifndef CONFIG_USER_ONLY
822 cc
->sysemu_ops
= &riscv_sysemu_ops
;
824 cc
->gdb_arch_name
= riscv_gdb_arch_name
;
825 cc
->gdb_get_dynamic_xml
= riscv_gdb_get_dynamic_xml
;
826 cc
->tcg_ops
= &riscv_tcg_ops
;
828 device_class_set_props(dc
, riscv_cpu_properties
);
831 char *riscv_isa_string(RISCVCPU
*cpu
)
834 const size_t maxlen
= sizeof("rv128") + sizeof(riscv_exts
) + 1;
835 char *isa_str
= g_new(char, maxlen
);
836 char *p
= isa_str
+ snprintf(isa_str
, maxlen
, "rv%d", TARGET_LONG_BITS
);
837 for (i
= 0; i
< sizeof(riscv_exts
); i
++) {
838 if (cpu
->env
.misa_ext
& RV(riscv_exts
[i
])) {
839 *p
++ = qemu_tolower(riscv_exts
[i
]);
846 static gint
riscv_cpu_list_compare(gconstpointer a
, gconstpointer b
)
848 ObjectClass
*class_a
= (ObjectClass
*)a
;
849 ObjectClass
*class_b
= (ObjectClass
*)b
;
850 const char *name_a
, *name_b
;
852 name_a
= object_class_get_name(class_a
);
853 name_b
= object_class_get_name(class_b
);
854 return strcmp(name_a
, name_b
);
857 static void riscv_cpu_list_entry(gpointer data
, gpointer user_data
)
859 const char *typename
= object_class_get_name(OBJECT_CLASS(data
));
860 int len
= strlen(typename
) - strlen(RISCV_CPU_TYPE_SUFFIX
);
862 qemu_printf("%.*s\n", len
, typename
);
865 void riscv_cpu_list(void)
869 list
= object_class_get_list(TYPE_RISCV_CPU
, false);
870 list
= g_slist_sort(list
, riscv_cpu_list_compare
);
871 g_slist_foreach(list
, riscv_cpu_list_entry
, NULL
);
875 #define DEFINE_CPU(type_name, initfn) \
878 .parent = TYPE_RISCV_CPU, \
879 .instance_init = initfn \
882 static const TypeInfo riscv_cpu_type_infos
[] = {
884 .name
= TYPE_RISCV_CPU
,
886 .instance_size
= sizeof(RISCVCPU
),
887 .instance_align
= __alignof__(RISCVCPU
),
888 .instance_init
= riscv_cpu_init
,
890 .class_size
= sizeof(RISCVCPUClass
),
891 .class_init
= riscv_cpu_class_init
,
893 DEFINE_CPU(TYPE_RISCV_CPU_ANY
, riscv_any_cpu_init
),
894 #if defined(CONFIG_KVM)
895 DEFINE_CPU(TYPE_RISCV_CPU_HOST
, riscv_host_cpu_init
),
897 #if defined(TARGET_RISCV32)
898 DEFINE_CPU(TYPE_RISCV_CPU_BASE32
, rv32_base_cpu_init
),
899 DEFINE_CPU(TYPE_RISCV_CPU_IBEX
, rv32_ibex_cpu_init
),
900 DEFINE_CPU(TYPE_RISCV_CPU_SIFIVE_E31
, rv32_sifive_e_cpu_init
),
901 DEFINE_CPU(TYPE_RISCV_CPU_SIFIVE_E34
, rv32_imafcu_nommu_cpu_init
),
902 DEFINE_CPU(TYPE_RISCV_CPU_SIFIVE_U34
, rv32_sifive_u_cpu_init
),
903 #elif defined(TARGET_RISCV64)
904 DEFINE_CPU(TYPE_RISCV_CPU_BASE64
, rv64_base_cpu_init
),
905 DEFINE_CPU(TYPE_RISCV_CPU_SIFIVE_E51
, rv64_sifive_e_cpu_init
),
906 DEFINE_CPU(TYPE_RISCV_CPU_SIFIVE_U54
, rv64_sifive_u_cpu_init
),
907 DEFINE_CPU(TYPE_RISCV_CPU_SHAKTI_C
, rv64_sifive_u_cpu_init
),
908 DEFINE_CPU(TYPE_RISCV_CPU_BASE128
, rv128_base_cpu_init
),
912 DEFINE_TYPES(riscv_cpu_type_infos
)