4 * Copyright (c) 2016-2017 Sagar Karandikar, sagark@eecs.berkeley.edu
5 * Copyright (c) 2017-2018 SiFive, Inc.
7 * This program is free software; you can redistribute it and/or modify it
8 * under the terms and conditions of the GNU General Public License,
9 * version 2 or later, as published by the Free Software Foundation.
11 * This program is distributed in the hope it will be useful, but WITHOUT
12 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
13 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
16 * You should have received a copy of the GNU General Public License along with
17 * this program. If not, see <http://www.gnu.org/licenses/>.
23 #include "hw/core/cpu.h"
24 #include "hw/registerfields.h"
25 #include "exec/cpu-defs.h"
26 #include "qemu/cpu-float.h"
27 #include "qom/object.h"
28 #include "qemu/int128.h"
30 #include "qapi/qapi-types-common.h"
32 #define TCG_GUEST_DEFAULT_MO 0
35 * RISC-V-specific extra insn start words:
36 * 1: Original instruction opcode
38 #define TARGET_INSN_START_EXTRA_WORDS 1
40 #define TYPE_RISCV_CPU "riscv-cpu"
42 #define RISCV_CPU_TYPE_SUFFIX "-" TYPE_RISCV_CPU
43 #define RISCV_CPU_TYPE_NAME(name) (name RISCV_CPU_TYPE_SUFFIX)
44 #define CPU_RESOLVING_TYPE TYPE_RISCV_CPU
46 #define TYPE_RISCV_CPU_ANY RISCV_CPU_TYPE_NAME("any")
47 #define TYPE_RISCV_CPU_BASE32 RISCV_CPU_TYPE_NAME("rv32")
48 #define TYPE_RISCV_CPU_BASE64 RISCV_CPU_TYPE_NAME("rv64")
49 #define TYPE_RISCV_CPU_BASE128 RISCV_CPU_TYPE_NAME("x-rv128")
50 #define TYPE_RISCV_CPU_IBEX RISCV_CPU_TYPE_NAME("lowrisc-ibex")
51 #define TYPE_RISCV_CPU_SHAKTI_C RISCV_CPU_TYPE_NAME("shakti-c")
52 #define TYPE_RISCV_CPU_SIFIVE_E31 RISCV_CPU_TYPE_NAME("sifive-e31")
53 #define TYPE_RISCV_CPU_SIFIVE_E34 RISCV_CPU_TYPE_NAME("sifive-e34")
54 #define TYPE_RISCV_CPU_SIFIVE_E51 RISCV_CPU_TYPE_NAME("sifive-e51")
55 #define TYPE_RISCV_CPU_SIFIVE_U34 RISCV_CPU_TYPE_NAME("sifive-u34")
56 #define TYPE_RISCV_CPU_SIFIVE_U54 RISCV_CPU_TYPE_NAME("sifive-u54")
57 #define TYPE_RISCV_CPU_THEAD_C906 RISCV_CPU_TYPE_NAME("thead-c906")
58 #define TYPE_RISCV_CPU_HOST RISCV_CPU_TYPE_NAME("host")
60 #if defined(TARGET_RISCV32)
61 # define TYPE_RISCV_CPU_BASE TYPE_RISCV_CPU_BASE32
62 #elif defined(TARGET_RISCV64)
63 # define TYPE_RISCV_CPU_BASE TYPE_RISCV_CPU_BASE64
66 #define RV(x) ((target_ulong)1 << (x - 'A'))
69 * Consider updating register_cpu_props() when adding
73 #define RVE RV('E') /* E and I are mutually exclusive */
86 /* Privileged specification version */
88 PRIV_VERSION_1_10_0
= 0,
93 #define VEXT_VERSION_1_00_0 0x00010000
99 TRANSLATE_G_STAGE_FAIL
102 #define MMU_USER_IDX 3
104 #define MAX_RISCV_PMPS (16)
106 typedef struct CPUArchState CPURISCVState
;
108 #if !defined(CONFIG_USER_ONLY)
113 #define RV_VLEN_MAX 1024
114 #define RV_MAX_MHPMEVENTS 32
115 #define RV_MAX_MHPMCOUNTERS 32
117 FIELD(VTYPE
, VLMUL
, 0, 3)
118 FIELD(VTYPE
, VSEW
, 3, 3)
119 FIELD(VTYPE
, VTA
, 6, 1)
120 FIELD(VTYPE
, VMA
, 7, 1)
121 FIELD(VTYPE
, VEDIV
, 8, 2)
122 FIELD(VTYPE
, RESERVED
, 10, sizeof(target_ulong
) * 8 - 11)
124 typedef struct PMUCTRState
{
125 /* Current value of a counter */
126 target_ulong mhpmcounter_val
;
127 /* Current value of a counter in RV32 */
128 target_ulong mhpmcounterh_val
;
129 /* Snapshot values of counter */
130 target_ulong mhpmcounter_prev
;
131 /* Snapshort value of a counter in RV32 */
132 target_ulong mhpmcounterh_prev
;
134 /* Value beyond UINT32_MAX/UINT64_MAX before overflow interrupt trigger */
135 target_ulong irq_overflow_left
;
138 struct CPUArchState
{
139 target_ulong gpr
[32];
140 target_ulong gprh
[32]; /* 64 top bits of the 128-bit registers */
142 /* vector coprocessor state. */
143 uint64_t vreg
[32 * RV_VLEN_MAX
/ 64] QEMU_ALIGNED(16);
152 target_ulong load_res
;
153 target_ulong load_val
;
155 /* Floating-Point state */
156 uint64_t fpr
[32]; /* assume both F and D extensions */
158 float_status fp_status
;
160 target_ulong badaddr
;
163 target_ulong guest_phys_fault_addr
;
165 target_ulong priv_ver
;
166 target_ulong bext_ver
;
167 target_ulong vext_ver
;
169 /* RISCVMXL, but uint32_t for vmstate migration */
170 uint32_t misa_mxl
; /* current mxl */
171 uint32_t misa_mxl_max
; /* max mxl for this cpu */
172 uint32_t misa_ext
; /* current extensions */
173 uint32_t misa_ext_mask
; /* max ext for this cpu */
174 uint32_t xl
; /* current xlen */
176 /* 128-bit helpers upper part return value */
181 #ifdef CONFIG_USER_ONLY
185 #ifndef CONFIG_USER_ONLY
187 /* This contains QEMU specific information about the virt state. */
192 target_ulong mhartid
;
194 * For RV32 this is 32-bit mstatus and 32-bit mstatush.
195 * For RV64 this is a 64-bit mstatus.
201 * MIP contains the software writable version of SEIP ORed with the
202 * external interrupt value. The MIP register is always up-to-date.
203 * To keep track of the current source, we also save booleans of the values
214 target_ulong satp
; /* since: priv-1.10.0 */
216 target_ulong medeleg
;
225 target_ulong mtval
; /* since: priv-1.10.0 */
227 /* Machine and Supervisor interrupt priorities */
232 target_ulong miselect
;
233 target_ulong siselect
;
235 /* Hypervisor CSRs */
236 target_ulong hstatus
;
237 target_ulong hedeleg
;
239 target_ulong hcounteren
;
247 /* Hypervisor controlled virtual interrupt priorities */
251 /* Upper 64-bits of 128-bit CSRs */
257 * For RV32 this is 32-bit vsstatus and 32-bit vsstatush.
258 * For RV64 this is a 64-bit vsstatus.
262 target_ulong vsscratch
;
264 target_ulong vscause
;
268 /* AIA VS-mode CSRs */
269 target_ulong vsiselect
;
275 target_ulong stvec_hs
;
276 target_ulong sscratch_hs
;
277 target_ulong sepc_hs
;
278 target_ulong scause_hs
;
279 target_ulong stval_hs
;
280 target_ulong satp_hs
;
284 * Signals whether the current exception occurred with two-stage address
285 * translation active.
287 bool two_stage_lookup
;
289 * Signals whether the current exception occurred while doing two-stage
290 * address translation for the VS-stage page table walk.
292 bool two_stage_indirect_lookup
;
294 target_ulong scounteren
;
295 target_ulong mcounteren
;
297 target_ulong mcountinhibit
;
299 /* PMU counter state */
300 PMUCTRState pmu_ctrs
[RV_MAX_MHPMCOUNTERS
];
302 /* PMU event selector configured values. First three are unused */
303 target_ulong mhpmevent_val
[RV_MAX_MHPMEVENTS
];
305 /* PMU event selector configured values for RV32 */
306 target_ulong mhpmeventh_val
[RV_MAX_MHPMEVENTS
];
308 target_ulong sscratch
;
309 target_ulong mscratch
;
316 /* physical memory protection */
317 pmp_table_t pmp_state
;
318 target_ulong mseccfg
;
321 target_ulong trigger_cur
;
322 target_ulong tdata1
[RV_MAX_TRIGGERS
];
323 target_ulong tdata2
[RV_MAX_TRIGGERS
];
324 target_ulong tdata3
[RV_MAX_TRIGGERS
];
325 struct CPUBreakpoint
*cpu_breakpoint
[RV_MAX_TRIGGERS
];
326 struct CPUWatchpoint
*cpu_watchpoint
[RV_MAX_TRIGGERS
];
327 QEMUTimer
*itrigger_timer
[RV_MAX_TRIGGERS
];
329 bool itrigger_enabled
;
331 /* machine specific rdtime callback */
332 uint64_t (*rdtime_fn
)(void *);
335 /* machine specific AIA ireg read-modify-write callback */
336 #define AIA_MAKE_IREG(__isel, __priv, __virt, __vgein, __xlen) \
337 ((((__xlen) & 0xff) << 24) | \
338 (((__vgein) & 0x3f) << 20) | \
339 (((__virt) & 0x1) << 18) | \
340 (((__priv) & 0x3) << 16) | \
342 #define AIA_IREG_ISEL(__ireg) ((__ireg) & 0xffff)
343 #define AIA_IREG_PRIV(__ireg) (((__ireg) >> 16) & 0x3)
344 #define AIA_IREG_VIRT(__ireg) (((__ireg) >> 18) & 0x1)
345 #define AIA_IREG_VGEIN(__ireg) (((__ireg) >> 20) & 0x3f)
346 #define AIA_IREG_XLEN(__ireg) (((__ireg) >> 24) & 0xff)
347 int (*aia_ireg_rmw_fn
[4])(void *arg
, target_ulong reg
,
348 target_ulong
*val
, target_ulong new_val
, target_ulong write_mask
);
349 void *aia_ireg_rmw_fn_arg
[4];
351 /* True if in debugger mode. */
355 * CSRs for PointerMasking extension
358 target_ulong mpmmask
;
359 target_ulong mpmbase
;
360 target_ulong spmmask
;
361 target_ulong spmbase
;
362 target_ulong upmmask
;
363 target_ulong upmbase
;
365 /* CSRs for execution enviornment configuration */
367 uint64_t mstateen
[SMSTATEEN_MAX_COUNT
];
368 uint64_t hstateen
[SMSTATEEN_MAX_COUNT
];
369 uint64_t sstateen
[SMSTATEEN_MAX_COUNT
];
370 target_ulong senvcfg
;
373 target_ulong cur_pmmask
;
374 target_ulong cur_pmbase
;
376 /* Fields from here on are preserved across CPU reset. */
377 QEMUTimer
*stimer
; /* Internal timer for S-mode interrupt */
378 QEMUTimer
*vstimer
; /* Internal timer for VS-mode interrupt */
385 bool kvm_timer_dirty
;
386 uint64_t kvm_timer_time
;
387 uint64_t kvm_timer_compare
;
388 uint64_t kvm_timer_state
;
389 uint64_t kvm_timer_frequency
;
392 OBJECT_DECLARE_CPU_TYPE(RISCVCPU
, RISCVCPUClass
, RISCV_CPU
)
396 * @parent_realize: The parent class' realize handler.
397 * @parent_phases: The parent class' reset phase handlers.
401 struct RISCVCPUClass
{
403 CPUClass parent_class
;
405 DeviceRealize parent_realize
;
406 ResettablePhases parent_phases
;
410 * map is a 16-bit bitmap: the most significant set bit in map is the maximum
411 * satp mode that is supported. It may be chosen by the user and must respect
412 * what qemu implements (valid_1_10_32/64) and what the hw is capable of
413 * (supported bitmap below).
415 * init is a 16-bit bitmap used to make sure the user selected a correct
416 * configuration as per the specification.
418 * supported is a 16-bit bitmap used to reflect the hw capabilities.
421 uint16_t map
, init
, supported
;
424 struct RISCVCPUConfig
{
467 bool ext_zihintpause
;
497 /* Vendor-specific custom extensions */
502 bool ext_xtheadcondmov
;
503 bool ext_xtheadfmemidx
;
506 bool ext_xtheadmemidx
;
507 bool ext_xtheadmempair
;
509 bool ext_XVentanaCondOps
;
518 uint16_t cbom_blocksize
;
519 uint16_t cboz_blocksize
;
526 bool short_isa_string
;
528 #ifndef CONFIG_USER_ONLY
529 RISCVSATPMap satp_mode
;
533 typedef struct RISCVCPUConfig RISCVCPUConfig
;
537 * @env: #CPURISCVState
545 CPUNegativeOffsetState neg
;
551 /* Configuration Settings */
554 QEMUTimer
*pmu_timer
;
555 /* A bitmask of Available programmable counters */
556 uint32_t pmu_avail_ctrs
;
557 /* Mapping of events to counters */
558 GHashTable
*pmu_event_ctr_map
;
561 static inline int riscv_has_ext(CPURISCVState
*env
, target_ulong ext
)
563 return (env
->misa_ext
& ext
) != 0;
566 #include "cpu_user.h"
568 extern const char * const riscv_int_regnames
[];
569 extern const char * const riscv_int_regnamesh
[];
570 extern const char * const riscv_fpr_regnames
[];
572 const char *riscv_cpu_get_trap_name(target_ulong cause
, bool async
);
573 void riscv_cpu_do_interrupt(CPUState
*cpu
);
574 int riscv_cpu_write_elf64_note(WriteCoreDumpFunction f
, CPUState
*cs
,
575 int cpuid
, DumpState
*s
);
576 int riscv_cpu_write_elf32_note(WriteCoreDumpFunction f
, CPUState
*cs
,
577 int cpuid
, DumpState
*s
);
578 int riscv_cpu_gdb_read_register(CPUState
*cpu
, GByteArray
*buf
, int reg
);
579 int riscv_cpu_gdb_write_register(CPUState
*cpu
, uint8_t *buf
, int reg
);
580 int riscv_cpu_hviprio_index2irq(int index
, int *out_irq
, int *out_rdzero
);
581 uint8_t riscv_cpu_default_priority(int irq
);
582 uint64_t riscv_cpu_all_pending(CPURISCVState
*env
);
583 int riscv_cpu_mirq_pending(CPURISCVState
*env
);
584 int riscv_cpu_sirq_pending(CPURISCVState
*env
);
585 int riscv_cpu_vsirq_pending(CPURISCVState
*env
);
586 bool riscv_cpu_fp_enabled(CPURISCVState
*env
);
587 target_ulong
riscv_cpu_get_geilen(CPURISCVState
*env
);
588 void riscv_cpu_set_geilen(CPURISCVState
*env
, target_ulong geilen
);
589 bool riscv_cpu_vector_enabled(CPURISCVState
*env
);
590 void riscv_cpu_set_virt_enabled(CPURISCVState
*env
, bool enable
);
591 bool riscv_cpu_two_stage_lookup(int mmu_idx
);
592 int riscv_cpu_mmu_index(CPURISCVState
*env
, bool ifetch
);
593 G_NORETURN
void riscv_cpu_do_unaligned_access(CPUState
*cs
, vaddr addr
,
594 MMUAccessType access_type
, int mmu_idx
,
596 bool riscv_cpu_tlb_fill(CPUState
*cs
, vaddr address
, int size
,
597 MMUAccessType access_type
, int mmu_idx
,
598 bool probe
, uintptr_t retaddr
);
599 char *riscv_isa_string(RISCVCPU
*cpu
);
600 void riscv_cpu_list(void);
602 #define cpu_list riscv_cpu_list
603 #define cpu_mmu_index riscv_cpu_mmu_index
605 #ifndef CONFIG_USER_ONLY
606 void riscv_cpu_do_transaction_failed(CPUState
*cs
, hwaddr physaddr
,
607 vaddr addr
, unsigned size
,
608 MMUAccessType access_type
,
609 int mmu_idx
, MemTxAttrs attrs
,
610 MemTxResult response
, uintptr_t retaddr
);
611 hwaddr
riscv_cpu_get_phys_page_debug(CPUState
*cpu
, vaddr addr
);
612 bool riscv_cpu_exec_interrupt(CPUState
*cs
, int interrupt_request
);
613 void riscv_cpu_swap_hypervisor_regs(CPURISCVState
*env
);
614 int riscv_cpu_claim_interrupts(RISCVCPU
*cpu
, uint64_t interrupts
);
615 uint64_t riscv_cpu_update_mip(CPURISCVState
*env
, uint64_t mask
,
617 #define BOOL_TO_MASK(x) (-!!(x)) /* helper for riscv_cpu_update_mip value */
618 void riscv_cpu_set_rdtime_fn(CPURISCVState
*env
, uint64_t (*fn
)(void *),
620 void riscv_cpu_set_aia_ireg_rmw_fn(CPURISCVState
*env
, uint32_t priv
,
621 int (*rmw_fn
)(void *arg
,
624 target_ulong new_val
,
625 target_ulong write_mask
),
628 RISCVException
smstateen_acc_ok(CPURISCVState
*env
, int index
, uint64_t bit
);
630 void riscv_cpu_set_mode(CPURISCVState
*env
, target_ulong newpriv
);
632 void riscv_translate_init(void);
633 G_NORETURN
void riscv_raise_exception(CPURISCVState
*env
,
634 uint32_t exception
, uintptr_t pc
);
636 target_ulong
riscv_cpu_get_fflags(CPURISCVState
*env
);
637 void riscv_cpu_set_fflags(CPURISCVState
*env
, target_ulong
);
639 #define TB_FLAGS_PRIV_MMU_MASK 3
640 #define TB_FLAGS_PRIV_HYP_ACCESS_MASK (1 << 2)
641 #define TB_FLAGS_MSTATUS_FS MSTATUS_FS
642 #define TB_FLAGS_MSTATUS_VS MSTATUS_VS
644 #include "exec/cpu-all.h"
646 FIELD(TB_FLAGS
, MEM_IDX
, 0, 3)
647 FIELD(TB_FLAGS
, LMUL
, 3, 3)
648 FIELD(TB_FLAGS
, SEW
, 6, 3)
649 /* Skip MSTATUS_VS (0x600) bits */
650 FIELD(TB_FLAGS
, VL_EQ_VLMAX
, 11, 1)
651 FIELD(TB_FLAGS
, VILL
, 12, 1)
652 /* Skip MSTATUS_FS (0x6000) bits */
653 /* Is a Hypervisor instruction load/store allowed? */
654 FIELD(TB_FLAGS
, HLSX
, 15, 1)
655 FIELD(TB_FLAGS
, MSTATUS_HS_FS
, 16, 2)
656 FIELD(TB_FLAGS
, MSTATUS_HS_VS
, 18, 2)
657 /* The combination of MXL/SXL/UXL that applies to the current cpu mode. */
658 FIELD(TB_FLAGS
, XL
, 20, 2)
659 /* If PointerMasking should be applied */
660 FIELD(TB_FLAGS
, PM_MASK_ENABLED
, 22, 1)
661 FIELD(TB_FLAGS
, PM_BASE_ENABLED
, 23, 1)
662 FIELD(TB_FLAGS
, VTA
, 24, 1)
663 FIELD(TB_FLAGS
, VMA
, 25, 1)
664 /* Native debug itrigger */
665 FIELD(TB_FLAGS
, ITRIGGER
, 26, 1)
667 #ifdef TARGET_RISCV32
668 #define riscv_cpu_mxl(env) ((void)(env), MXL_RV32)
670 static inline RISCVMXL
riscv_cpu_mxl(CPURISCVState
*env
)
672 return env
->misa_mxl
;
675 #define riscv_cpu_mxl_bits(env) (1UL << (4 + riscv_cpu_mxl(env)))
677 static inline const RISCVCPUConfig
*riscv_cpu_cfg(CPURISCVState
*env
)
679 return &env_archcpu(env
)->cfg
;
682 #if defined(TARGET_RISCV32)
683 #define cpu_recompute_xl(env) ((void)(env), MXL_RV32)
685 static inline RISCVMXL
cpu_recompute_xl(CPURISCVState
*env
)
687 RISCVMXL xl
= env
->misa_mxl
;
688 #if !defined(CONFIG_USER_ONLY)
690 * When emulating a 32-bit-only cpu, use RV32.
691 * When emulating a 64-bit cpu, and MXL has been reduced to RV32,
692 * MSTATUSH doesn't have UXL/SXL, therefore XLEN cannot be widened
693 * back to RV64 for lower privs.
695 if (xl
!= MXL_RV32
) {
700 xl
= get_field(env
->mstatus
, MSTATUS64_UXL
);
702 default: /* PRV_S | PRV_H */
703 xl
= get_field(env
->mstatus
, MSTATUS64_SXL
);
712 static inline int riscv_cpu_xlen(CPURISCVState
*env
)
714 return 16 << env
->xl
;
717 #ifdef TARGET_RISCV32
718 #define riscv_cpu_sxl(env) ((void)(env), MXL_RV32)
720 static inline RISCVMXL
riscv_cpu_sxl(CPURISCVState
*env
)
722 #ifdef CONFIG_USER_ONLY
723 return env
->misa_mxl
;
725 return get_field(env
->mstatus
, MSTATUS64_SXL
);
731 * Encode LMUL to lmul as follows:
742 * then, we can calculate VLMAX = vlen >> (vsew + 3 - lmul)
743 * e.g. vlen = 256 bits, SEW = 16, LMUL = 1/8
744 * => VLMAX = vlen >> (1 + 3 - (-3))
748 static inline uint32_t vext_get_vlmax(RISCVCPU
*cpu
, target_ulong vtype
)
750 uint8_t sew
= FIELD_EX64(vtype
, VTYPE
, VSEW
);
751 int8_t lmul
= sextract32(FIELD_EX64(vtype
, VTYPE
, VLMUL
), 0, 3);
752 return cpu
->cfg
.vlen
>> (sew
+ 3 - lmul
);
755 void cpu_get_tb_cpu_state(CPURISCVState
*env
, target_ulong
*pc
,
756 target_ulong
*cs_base
, uint32_t *pflags
);
758 void riscv_cpu_update_mask(CPURISCVState
*env
);
760 RISCVException
riscv_csrrw(CPURISCVState
*env
, int csrno
,
761 target_ulong
*ret_value
,
762 target_ulong new_value
, target_ulong write_mask
);
763 RISCVException
riscv_csrrw_debug(CPURISCVState
*env
, int csrno
,
764 target_ulong
*ret_value
,
765 target_ulong new_value
,
766 target_ulong write_mask
);
768 static inline void riscv_csr_write(CPURISCVState
*env
, int csrno
,
771 riscv_csrrw(env
, csrno
, NULL
, val
, MAKE_64BIT_MASK(0, TARGET_LONG_BITS
));
774 static inline target_ulong
riscv_csr_read(CPURISCVState
*env
, int csrno
)
776 target_ulong val
= 0;
777 riscv_csrrw(env
, csrno
, &val
, 0, 0);
781 typedef RISCVException (*riscv_csr_predicate_fn
)(CPURISCVState
*env
,
783 typedef RISCVException (*riscv_csr_read_fn
)(CPURISCVState
*env
, int csrno
,
784 target_ulong
*ret_value
);
785 typedef RISCVException (*riscv_csr_write_fn
)(CPURISCVState
*env
, int csrno
,
786 target_ulong new_value
);
787 typedef RISCVException (*riscv_csr_op_fn
)(CPURISCVState
*env
, int csrno
,
788 target_ulong
*ret_value
,
789 target_ulong new_value
,
790 target_ulong write_mask
);
792 RISCVException
riscv_csrrw_i128(CPURISCVState
*env
, int csrno
,
794 Int128 new_value
, Int128 write_mask
);
796 typedef RISCVException (*riscv_csr_read128_fn
)(CPURISCVState
*env
, int csrno
,
798 typedef RISCVException (*riscv_csr_write128_fn
)(CPURISCVState
*env
, int csrno
,
803 riscv_csr_predicate_fn predicate
;
804 riscv_csr_read_fn read
;
805 riscv_csr_write_fn write
;
807 riscv_csr_read128_fn read128
;
808 riscv_csr_write128_fn write128
;
809 /* The default priv spec version should be PRIV_VERSION_1_10_0 (i.e 0) */
810 uint32_t min_priv_ver
;
811 } riscv_csr_operations
;
813 /* CSR function table constants */
815 CSR_TABLE_SIZE
= 0x1000
819 * The event id are encoded based on the encoding specified in the
820 * SBI specification v0.3
823 enum riscv_pmu_event_idx
{
824 RISCV_PMU_EVENT_HW_CPU_CYCLES
= 0x01,
825 RISCV_PMU_EVENT_HW_INSTRUCTIONS
= 0x02,
826 RISCV_PMU_EVENT_CACHE_DTLB_READ_MISS
= 0x10019,
827 RISCV_PMU_EVENT_CACHE_DTLB_WRITE_MISS
= 0x1001B,
828 RISCV_PMU_EVENT_CACHE_ITLB_PREFETCH_MISS
= 0x10021,
831 /* CSR function table */
832 extern riscv_csr_operations csr_ops
[CSR_TABLE_SIZE
];
834 extern const bool valid_vm_1_10_32
[], valid_vm_1_10_64
[];
836 void riscv_get_csr_ops(int csrno
, riscv_csr_operations
*ops
);
837 void riscv_set_csr_ops(int csrno
, riscv_csr_operations
*ops
);
839 void riscv_cpu_register_gdb_regs_for_features(CPUState
*cs
);
841 uint8_t satp_mode_max_from_map(uint32_t map
);
842 const char *satp_mode_str(uint8_t satp_mode
, bool is_32_bit
);
844 #endif /* RISCV_CPU_H */