]>
Commit | Line | Data |
---|---|---|
2874c5fd | 1 | /* SPDX-License-Identifier: GPL-2.0-or-later */ |
7f65afb9 PB |
2 | /* |
3 | * Copyright (C) 2013 Imagination Technologies | |
fb615d61 | 4 | * Author: Paul Burton <paul.burton@mips.com> |
7f65afb9 PB |
5 | */ |
6 | #ifndef _ASM_MSA_H | |
7 | #define _ASM_MSA_H | |
8 | ||
9 | #include <asm/mipsregs.h> | |
10 | ||
558155a0 PB |
11 | #ifndef __ASSEMBLY__ |
12 | ||
6b35e114 PB |
13 | #include <asm/inst.h> |
14 | ||
1db1af84 PB |
15 | extern void _save_msa(struct task_struct *); |
16 | extern void _restore_msa(struct task_struct *); | |
c9017757 | 17 | extern void _init_msa_upper(void); |
1db1af84 | 18 | |
6b35e114 PB |
19 | extern void read_msa_wr_b(unsigned idx, union fpureg *to); |
20 | extern void read_msa_wr_h(unsigned idx, union fpureg *to); | |
21 | extern void read_msa_wr_w(unsigned idx, union fpureg *to); | |
22 | extern void read_msa_wr_d(unsigned idx, union fpureg *to); | |
23 | ||
24 | /** | |
25 | * read_msa_wr() - Read a single MSA vector register | |
26 | * @idx: The index of the vector register to read | |
27 | * @to: The FPU register union to store the registers value in | |
28 | * @fmt: The format of the data in the vector register | |
29 | * | |
30 | * Read the value of MSA vector register idx into the FPU register | |
31 | * union to, using the format fmt. | |
32 | */ | |
33 | static inline void read_msa_wr(unsigned idx, union fpureg *to, | |
34 | enum msa_2b_fmt fmt) | |
35 | { | |
36 | switch (fmt) { | |
37 | case msa_fmt_b: | |
38 | read_msa_wr_b(idx, to); | |
39 | break; | |
40 | ||
41 | case msa_fmt_h: | |
42 | read_msa_wr_h(idx, to); | |
43 | break; | |
44 | ||
45 | case msa_fmt_w: | |
46 | read_msa_wr_w(idx, to); | |
47 | break; | |
48 | ||
49 | case msa_fmt_d: | |
50 | read_msa_wr_d(idx, to); | |
51 | break; | |
52 | ||
53 | default: | |
54 | BUG(); | |
55 | } | |
56 | } | |
57 | ||
58 | extern void write_msa_wr_b(unsigned idx, union fpureg *from); | |
59 | extern void write_msa_wr_h(unsigned idx, union fpureg *from); | |
60 | extern void write_msa_wr_w(unsigned idx, union fpureg *from); | |
61 | extern void write_msa_wr_d(unsigned idx, union fpureg *from); | |
62 | ||
63 | /** | |
64 | * write_msa_wr() - Write a single MSA vector register | |
65 | * @idx: The index of the vector register to write | |
66 | * @from: The FPU register union to take the registers value from | |
67 | * @fmt: The format of the data in the vector register | |
68 | * | |
69 | * Write the value from the FPU register union from into MSA vector | |
70 | * register idx, using the format fmt. | |
71 | */ | |
72 | static inline void write_msa_wr(unsigned idx, union fpureg *from, | |
73 | enum msa_2b_fmt fmt) | |
74 | { | |
75 | switch (fmt) { | |
76 | case msa_fmt_b: | |
77 | write_msa_wr_b(idx, from); | |
78 | break; | |
79 | ||
80 | case msa_fmt_h: | |
81 | write_msa_wr_h(idx, from); | |
82 | break; | |
83 | ||
84 | case msa_fmt_w: | |
85 | write_msa_wr_w(idx, from); | |
86 | break; | |
87 | ||
88 | case msa_fmt_d: | |
89 | write_msa_wr_d(idx, from); | |
90 | break; | |
91 | ||
92 | default: | |
93 | BUG(); | |
94 | } | |
95 | } | |
96 | ||
7f65afb9 PB |
97 | static inline void enable_msa(void) |
98 | { | |
99 | if (cpu_has_msa) { | |
100 | set_c0_config5(MIPS_CONF5_MSAEN); | |
101 | enable_fpu_hazard(); | |
102 | } | |
103 | } | |
104 | ||
105 | static inline void disable_msa(void) | |
106 | { | |
107 | if (cpu_has_msa) { | |
108 | clear_c0_config5(MIPS_CONF5_MSAEN); | |
109 | disable_fpu_hazard(); | |
110 | } | |
111 | } | |
112 | ||
113 | static inline int is_msa_enabled(void) | |
114 | { | |
115 | if (!cpu_has_msa) | |
116 | return 0; | |
117 | ||
118 | return read_c0_config5() & MIPS_CONF5_MSAEN; | |
119 | } | |
120 | ||
1db1af84 PB |
121 | static inline int thread_msa_context_live(void) |
122 | { | |
123 | /* | |
124 | * Check cpu_has_msa only if it's a constant. This will allow the | |
125 | * compiler to optimise out code for CPUs without MSA without adding | |
126 | * an extra redundant check for CPUs with MSA. | |
127 | */ | |
128 | if (__builtin_constant_p(cpu_has_msa) && !cpu_has_msa) | |
129 | return 0; | |
130 | ||
131 | return test_thread_flag(TIF_MSA_CTX_LIVE); | |
132 | } | |
133 | ||
134 | static inline void save_msa(struct task_struct *t) | |
135 | { | |
136 | if (cpu_has_msa) | |
137 | _save_msa(t); | |
138 | } | |
139 | ||
140 | static inline void restore_msa(struct task_struct *t) | |
141 | { | |
142 | if (cpu_has_msa) | |
143 | _restore_msa(t); | |
144 | } | |
145 | ||
e49d3848 MR |
146 | static inline void init_msa_upper(void) |
147 | { | |
148 | /* | |
149 | * Check cpu_has_msa only if it's a constant. This will allow the | |
150 | * compiler to optimise out code for CPUs without MSA without adding | |
151 | * an extra redundant check for CPUs with MSA. | |
152 | */ | |
153 | if (__builtin_constant_p(cpu_has_msa) && !cpu_has_msa) | |
154 | return; | |
155 | ||
156 | _init_msa_upper(); | |
157 | } | |
158 | ||
93738d48 JH |
159 | #ifndef TOOLCHAIN_SUPPORTS_MSA |
160 | /* | |
161 | * Define assembler macros using .word for the c[ft]cmsa instructions in order | |
162 | * to allow compilation with toolchains that do not support MSA. Once all | |
163 | * toolchains in use support MSA these can be removed. | |
164 | */ | |
165 | _ASM_MACRO_2R(cfcmsa, rd, cs, | |
166 | _ASM_INSN_IF_MIPS(0x787e0019 | __cs << 11 | __rd << 6) | |
167 | _ASM_INSN32_IF_MM(0x587e0016 | __cs << 11 | __rd << 6)); | |
168 | _ASM_MACRO_2R(ctcmsa, cd, rs, | |
169 | _ASM_INSN_IF_MIPS(0x783e0019 | __rs << 11 | __cd << 6) | |
170 | _ASM_INSN32_IF_MM(0x583e0016 | __rs << 11 | __cd << 6)); | |
171 | #define _ASM_SET_MSA "" | |
172 | #else /* TOOLCHAIN_SUPPORTS_MSA */ | |
173 | #define _ASM_SET_MSA ".set\tfp=64\n\t" \ | |
174 | ".set\tmsa\n\t" | |
175 | #endif | |
7f65afb9 PB |
176 | |
177 | #define __BUILD_MSA_CTL_REG(name, cs) \ | |
178 | static inline unsigned int read_msa_##name(void) \ | |
179 | { \ | |
180 | unsigned int reg; \ | |
181 | __asm__ __volatile__( \ | |
182 | " .set push\n" \ | |
93738d48 | 183 | _ASM_SET_MSA \ |
7f65afb9 PB |
184 | " cfcmsa %0, $" #cs "\n" \ |
185 | " .set pop\n" \ | |
186 | : "=r"(reg)); \ | |
187 | return reg; \ | |
188 | } \ | |
189 | \ | |
190 | static inline void write_msa_##name(unsigned int val) \ | |
191 | { \ | |
192 | __asm__ __volatile__( \ | |
193 | " .set push\n" \ | |
93738d48 | 194 | _ASM_SET_MSA \ |
25198235 | 195 | " ctcmsa $" #cs ", %0\n" \ |
7f65afb9 PB |
196 | " .set pop\n" \ |
197 | : : "r"(val)); \ | |
198 | } | |
199 | ||
7f65afb9 PB |
200 | __BUILD_MSA_CTL_REG(ir, 0) |
201 | __BUILD_MSA_CTL_REG(csr, 1) | |
202 | __BUILD_MSA_CTL_REG(access, 2) | |
203 | __BUILD_MSA_CTL_REG(save, 3) | |
204 | __BUILD_MSA_CTL_REG(modify, 4) | |
205 | __BUILD_MSA_CTL_REG(request, 5) | |
206 | __BUILD_MSA_CTL_REG(map, 6) | |
207 | __BUILD_MSA_CTL_REG(unmap, 7) | |
208 | ||
558155a0 PB |
209 | #endif /* !__ASSEMBLY__ */ |
210 | ||
211 | #define MSA_IR 0 | |
212 | #define MSA_CSR 1 | |
213 | #define MSA_ACCESS 2 | |
214 | #define MSA_SAVE 3 | |
215 | #define MSA_MODIFY 4 | |
216 | #define MSA_REQUEST 5 | |
217 | #define MSA_MAP 6 | |
218 | #define MSA_UNMAP 7 | |
219 | ||
7f65afb9 PB |
220 | /* MSA Implementation Register (MSAIR) */ |
221 | #define MSA_IR_REVB 0 | |
222 | #define MSA_IR_REVF (_ULCAST_(0xff) << MSA_IR_REVB) | |
223 | #define MSA_IR_PROCB 8 | |
224 | #define MSA_IR_PROCF (_ULCAST_(0xff) << MSA_IR_PROCB) | |
225 | #define MSA_IR_WRPB 16 | |
226 | #define MSA_IR_WRPF (_ULCAST_(0x1) << MSA_IR_WRPB) | |
227 | ||
228 | /* MSA Control & Status Register (MSACSR) */ | |
229 | #define MSA_CSR_RMB 0 | |
230 | #define MSA_CSR_RMF (_ULCAST_(0x3) << MSA_CSR_RMB) | |
231 | #define MSA_CSR_RM_NEAREST 0 | |
232 | #define MSA_CSR_RM_TO_ZERO 1 | |
233 | #define MSA_CSR_RM_TO_POS 2 | |
234 | #define MSA_CSR_RM_TO_NEG 3 | |
235 | #define MSA_CSR_FLAGSB 2 | |
236 | #define MSA_CSR_FLAGSF (_ULCAST_(0x1f) << MSA_CSR_FLAGSB) | |
237 | #define MSA_CSR_FLAGS_IB 2 | |
238 | #define MSA_CSR_FLAGS_IF (_ULCAST_(0x1) << MSA_CSR_FLAGS_IB) | |
239 | #define MSA_CSR_FLAGS_UB 3 | |
240 | #define MSA_CSR_FLAGS_UF (_ULCAST_(0x1) << MSA_CSR_FLAGS_UB) | |
241 | #define MSA_CSR_FLAGS_OB 4 | |
242 | #define MSA_CSR_FLAGS_OF (_ULCAST_(0x1) << MSA_CSR_FLAGS_OB) | |
243 | #define MSA_CSR_FLAGS_ZB 5 | |
244 | #define MSA_CSR_FLAGS_ZF (_ULCAST_(0x1) << MSA_CSR_FLAGS_ZB) | |
245 | #define MSA_CSR_FLAGS_VB 6 | |
246 | #define MSA_CSR_FLAGS_VF (_ULCAST_(0x1) << MSA_CSR_FLAGS_VB) | |
247 | #define MSA_CSR_ENABLESB 7 | |
248 | #define MSA_CSR_ENABLESF (_ULCAST_(0x1f) << MSA_CSR_ENABLESB) | |
249 | #define MSA_CSR_ENABLES_IB 7 | |
250 | #define MSA_CSR_ENABLES_IF (_ULCAST_(0x1) << MSA_CSR_ENABLES_IB) | |
251 | #define MSA_CSR_ENABLES_UB 8 | |
252 | #define MSA_CSR_ENABLES_UF (_ULCAST_(0x1) << MSA_CSR_ENABLES_UB) | |
253 | #define MSA_CSR_ENABLES_OB 9 | |
254 | #define MSA_CSR_ENABLES_OF (_ULCAST_(0x1) << MSA_CSR_ENABLES_OB) | |
255 | #define MSA_CSR_ENABLES_ZB 10 | |
256 | #define MSA_CSR_ENABLES_ZF (_ULCAST_(0x1) << MSA_CSR_ENABLES_ZB) | |
257 | #define MSA_CSR_ENABLES_VB 11 | |
258 | #define MSA_CSR_ENABLES_VF (_ULCAST_(0x1) << MSA_CSR_ENABLES_VB) | |
259 | #define MSA_CSR_CAUSEB 12 | |
260 | #define MSA_CSR_CAUSEF (_ULCAST_(0x3f) << MSA_CSR_CAUSEB) | |
261 | #define MSA_CSR_CAUSE_IB 12 | |
262 | #define MSA_CSR_CAUSE_IF (_ULCAST_(0x1) << MSA_CSR_CAUSE_IB) | |
263 | #define MSA_CSR_CAUSE_UB 13 | |
264 | #define MSA_CSR_CAUSE_UF (_ULCAST_(0x1) << MSA_CSR_CAUSE_UB) | |
265 | #define MSA_CSR_CAUSE_OB 14 | |
266 | #define MSA_CSR_CAUSE_OF (_ULCAST_(0x1) << MSA_CSR_CAUSE_OB) | |
267 | #define MSA_CSR_CAUSE_ZB 15 | |
268 | #define MSA_CSR_CAUSE_ZF (_ULCAST_(0x1) << MSA_CSR_CAUSE_ZB) | |
269 | #define MSA_CSR_CAUSE_VB 16 | |
270 | #define MSA_CSR_CAUSE_VF (_ULCAST_(0x1) << MSA_CSR_CAUSE_VB) | |
271 | #define MSA_CSR_CAUSE_EB 17 | |
272 | #define MSA_CSR_CAUSE_EF (_ULCAST_(0x1) << MSA_CSR_CAUSE_EB) | |
273 | #define MSA_CSR_NXB 18 | |
274 | #define MSA_CSR_NXF (_ULCAST_(0x1) << MSA_CSR_NXB) | |
275 | #define MSA_CSR_FSB 24 | |
276 | #define MSA_CSR_FSF (_ULCAST_(0x1) << MSA_CSR_FSB) | |
277 | ||
278 | #endif /* _ASM_MSA_H */ |