]>
git.proxmox.com Git - mirror_qemu.git/blob - target/hexagon/gen_tcg_hvx.h
2 * Copyright(c) 2019-2023 Qualcomm Innovation Center, Inc. All Rights Reserved.
4 * This program is free software; you can redistribute it and/or modify
5 * it under the terms of the GNU General Public License as published by
6 * the Free Software Foundation; either version 2 of the License, or
7 * (at your option) any later version.
9 * This program is distributed in the hope that it will be useful,
10 * but WITHOUT ANY WARRANTY; without even the implied warranty of
11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
12 * GNU General Public License for more details.
14 * You should have received a copy of the GNU General Public License
15 * along with this program; if not, see <http://www.gnu.org/licenses/>.
18 #ifndef HEXAGON_GEN_TCG_HVX_H
19 #define HEXAGON_GEN_TCG_HVX_H
22 * Histogram instructions
24 * Note that these instructions operate directly on the vector registers
25 * and therefore happen after commit.
27 * The generate_<tag> function is called twice
28 * The first time is during the normal TCG generation
29 * ctx->pre_commit is true
30 * In the masked cases, we save the mask to the qtmp temporary
31 * Otherwise, there is nothing to do
32 * The second call is at the end of gen_commit_packet
33 * ctx->pre_commit is false
34 * Generate the call to the helper
37 static inline void assert_vhist_tmp(DisasContext
*ctx
)
39 /* vhist instructions require exactly one .tmp to be defined */
40 g_assert(ctx
->tmp_vregs_idx
== 1);
43 #define fGEN_TCG_V6_vhist(SHORTCODE) \
44 if (!ctx->pre_commit) { \
45 assert_vhist_tmp(ctx); \
46 gen_helper_vhist(cpu_env); \
48 #define fGEN_TCG_V6_vhistq(SHORTCODE) \
50 if (ctx->pre_commit) { \
51 intptr_t dstoff = offsetof(CPUHexagonState, qtmp); \
52 tcg_gen_gvec_mov(MO_64, dstoff, QvV_off, \
53 sizeof(MMVector), sizeof(MMVector)); \
55 assert_vhist_tmp(ctx); \
56 gen_helper_vhistq(cpu_env); \
59 #define fGEN_TCG_V6_vwhist256(SHORTCODE) \
60 if (!ctx->pre_commit) { \
61 assert_vhist_tmp(ctx); \
62 gen_helper_vwhist256(cpu_env); \
64 #define fGEN_TCG_V6_vwhist256q(SHORTCODE) \
66 if (ctx->pre_commit) { \
67 intptr_t dstoff = offsetof(CPUHexagonState, qtmp); \
68 tcg_gen_gvec_mov(MO_64, dstoff, QvV_off, \
69 sizeof(MMVector), sizeof(MMVector)); \
71 assert_vhist_tmp(ctx); \
72 gen_helper_vwhist256q(cpu_env); \
75 #define fGEN_TCG_V6_vwhist256_sat(SHORTCODE) \
76 if (!ctx->pre_commit) { \
77 assert_vhist_tmp(ctx); \
78 gen_helper_vwhist256_sat(cpu_env); \
80 #define fGEN_TCG_V6_vwhist256q_sat(SHORTCODE) \
82 if (ctx->pre_commit) { \
83 intptr_t dstoff = offsetof(CPUHexagonState, qtmp); \
84 tcg_gen_gvec_mov(MO_64, dstoff, QvV_off, \
85 sizeof(MMVector), sizeof(MMVector)); \
87 assert_vhist_tmp(ctx); \
88 gen_helper_vwhist256q_sat(cpu_env); \
91 #define fGEN_TCG_V6_vwhist128(SHORTCODE) \
92 if (!ctx->pre_commit) { \
93 assert_vhist_tmp(ctx); \
94 gen_helper_vwhist128(cpu_env); \
96 #define fGEN_TCG_V6_vwhist128q(SHORTCODE) \
98 if (ctx->pre_commit) { \
99 intptr_t dstoff = offsetof(CPUHexagonState, qtmp); \
100 tcg_gen_gvec_mov(MO_64, dstoff, QvV_off, \
101 sizeof(MMVector), sizeof(MMVector)); \
103 assert_vhist_tmp(ctx); \
104 gen_helper_vwhist128q(cpu_env); \
107 #define fGEN_TCG_V6_vwhist128m(SHORTCODE) \
108 if (!ctx->pre_commit) { \
109 TCGv tcgv_uiV = tcg_constant_tl(uiV); \
110 assert_vhist_tmp(ctx); \
111 gen_helper_vwhist128m(cpu_env, tcgv_uiV); \
113 #define fGEN_TCG_V6_vwhist128qm(SHORTCODE) \
115 if (ctx->pre_commit) { \
116 intptr_t dstoff = offsetof(CPUHexagonState, qtmp); \
117 tcg_gen_gvec_mov(MO_64, dstoff, QvV_off, \
118 sizeof(MMVector), sizeof(MMVector)); \
120 TCGv tcgv_uiV = tcg_constant_tl(uiV); \
121 assert_vhist_tmp(ctx); \
122 gen_helper_vwhist128qm(cpu_env, tcgv_uiV); \
127 #define fGEN_TCG_V6_vassign(SHORTCODE) \
128 tcg_gen_gvec_mov(MO_64, VdV_off, VuV_off, \
129 sizeof(MMVector), sizeof(MMVector))
131 /* Vector conditional move */
132 #define fGEN_TCG_VEC_CMOV(PRED) \
134 TCGv lsb = tcg_temp_new(); \
135 TCGLabel *false_label = gen_new_label(); \
136 tcg_gen_andi_tl(lsb, PsV, 1); \
137 tcg_gen_brcondi_tl(TCG_COND_NE, lsb, PRED, false_label); \
138 tcg_gen_gvec_mov(MO_64, VdV_off, VuV_off, \
139 sizeof(MMVector), sizeof(MMVector)); \
140 gen_set_label(false_label); \
144 /* Vector conditional move (true) */
145 #define fGEN_TCG_V6_vcmov(SHORTCODE) \
148 /* Vector conditional move (false) */
149 #define fGEN_TCG_V6_vncmov(SHORTCODE) \
152 /* Vector add - various forms */
153 #define fGEN_TCG_V6_vaddb(SHORTCODE) \
154 tcg_gen_gvec_add(MO_8, VdV_off, VuV_off, VvV_off, \
155 sizeof(MMVector), sizeof(MMVector))
157 #define fGEN_TCG_V6_vaddh(SHORTCYDE) \
158 tcg_gen_gvec_add(MO_16, VdV_off, VuV_off, VvV_off, \
159 sizeof(MMVector), sizeof(MMVector))
161 #define fGEN_TCG_V6_vaddw(SHORTCODE) \
162 tcg_gen_gvec_add(MO_32, VdV_off, VuV_off, VvV_off, \
163 sizeof(MMVector), sizeof(MMVector))
165 #define fGEN_TCG_V6_vaddb_dv(SHORTCODE) \
166 tcg_gen_gvec_add(MO_8, VddV_off, VuuV_off, VvvV_off, \
167 sizeof(MMVector) * 2, sizeof(MMVector) * 2)
169 #define fGEN_TCG_V6_vaddh_dv(SHORTCYDE) \
170 tcg_gen_gvec_add(MO_16, VddV_off, VuuV_off, VvvV_off, \
171 sizeof(MMVector) * 2, sizeof(MMVector) * 2)
173 #define fGEN_TCG_V6_vaddw_dv(SHORTCODE) \
174 tcg_gen_gvec_add(MO_32, VddV_off, VuuV_off, VvvV_off, \
175 sizeof(MMVector) * 2, sizeof(MMVector) * 2)
177 /* Vector sub - various forms */
178 #define fGEN_TCG_V6_vsubb(SHORTCODE) \
179 tcg_gen_gvec_sub(MO_8, VdV_off, VuV_off, VvV_off, \
180 sizeof(MMVector), sizeof(MMVector))
182 #define fGEN_TCG_V6_vsubh(SHORTCODE) \
183 tcg_gen_gvec_sub(MO_16, VdV_off, VuV_off, VvV_off, \
184 sizeof(MMVector), sizeof(MMVector))
186 #define fGEN_TCG_V6_vsubw(SHORTCODE) \
187 tcg_gen_gvec_sub(MO_32, VdV_off, VuV_off, VvV_off, \
188 sizeof(MMVector), sizeof(MMVector))
190 #define fGEN_TCG_V6_vsubb_dv(SHORTCODE) \
191 tcg_gen_gvec_sub(MO_8, VddV_off, VuuV_off, VvvV_off, \
192 sizeof(MMVector) * 2, sizeof(MMVector) * 2)
194 #define fGEN_TCG_V6_vsubh_dv(SHORTCODE) \
195 tcg_gen_gvec_sub(MO_16, VddV_off, VuuV_off, VvvV_off, \
196 sizeof(MMVector) * 2, sizeof(MMVector) * 2)
198 #define fGEN_TCG_V6_vsubw_dv(SHORTCODE) \
199 tcg_gen_gvec_sub(MO_32, VddV_off, VuuV_off, VvvV_off, \
200 sizeof(MMVector) * 2, sizeof(MMVector) * 2)
202 /* Vector shift right - various forms */
203 #define fGEN_TCG_V6_vasrh(SHORTCODE) \
205 TCGv shift = tcg_temp_new(); \
206 tcg_gen_andi_tl(shift, RtV, 15); \
207 tcg_gen_gvec_sars(MO_16, VdV_off, VuV_off, shift, \
208 sizeof(MMVector), sizeof(MMVector)); \
211 #define fGEN_TCG_V6_vasrh_acc(SHORTCODE) \
213 intptr_t tmpoff = offsetof(CPUHexagonState, vtmp); \
214 TCGv shift = tcg_temp_new(); \
215 tcg_gen_andi_tl(shift, RtV, 15); \
216 tcg_gen_gvec_sars(MO_16, tmpoff, VuV_off, shift, \
217 sizeof(MMVector), sizeof(MMVector)); \
218 tcg_gen_gvec_add(MO_16, VxV_off, VxV_off, tmpoff, \
219 sizeof(MMVector), sizeof(MMVector)); \
222 #define fGEN_TCG_V6_vasrw(SHORTCODE) \
224 TCGv shift = tcg_temp_new(); \
225 tcg_gen_andi_tl(shift, RtV, 31); \
226 tcg_gen_gvec_sars(MO_32, VdV_off, VuV_off, shift, \
227 sizeof(MMVector), sizeof(MMVector)); \
230 #define fGEN_TCG_V6_vasrw_acc(SHORTCODE) \
232 intptr_t tmpoff = offsetof(CPUHexagonState, vtmp); \
233 TCGv shift = tcg_temp_new(); \
234 tcg_gen_andi_tl(shift, RtV, 31); \
235 tcg_gen_gvec_sars(MO_32, tmpoff, VuV_off, shift, \
236 sizeof(MMVector), sizeof(MMVector)); \
237 tcg_gen_gvec_add(MO_32, VxV_off, VxV_off, tmpoff, \
238 sizeof(MMVector), sizeof(MMVector)); \
241 #define fGEN_TCG_V6_vlsrb(SHORTCODE) \
243 TCGv shift = tcg_temp_new(); \
244 tcg_gen_andi_tl(shift, RtV, 7); \
245 tcg_gen_gvec_shrs(MO_8, VdV_off, VuV_off, shift, \
246 sizeof(MMVector), sizeof(MMVector)); \
249 #define fGEN_TCG_V6_vlsrh(SHORTCODE) \
251 TCGv shift = tcg_temp_new(); \
252 tcg_gen_andi_tl(shift, RtV, 15); \
253 tcg_gen_gvec_shrs(MO_16, VdV_off, VuV_off, shift, \
254 sizeof(MMVector), sizeof(MMVector)); \
257 #define fGEN_TCG_V6_vlsrw(SHORTCODE) \
259 TCGv shift = tcg_temp_new(); \
260 tcg_gen_andi_tl(shift, RtV, 31); \
261 tcg_gen_gvec_shrs(MO_32, VdV_off, VuV_off, shift, \
262 sizeof(MMVector), sizeof(MMVector)); \
265 /* Vector shift left - various forms */
266 #define fGEN_TCG_V6_vaslb(SHORTCODE) \
268 TCGv shift = tcg_temp_new(); \
269 tcg_gen_andi_tl(shift, RtV, 7); \
270 tcg_gen_gvec_shls(MO_8, VdV_off, VuV_off, shift, \
271 sizeof(MMVector), sizeof(MMVector)); \
274 #define fGEN_TCG_V6_vaslh(SHORTCODE) \
276 TCGv shift = tcg_temp_new(); \
277 tcg_gen_andi_tl(shift, RtV, 15); \
278 tcg_gen_gvec_shls(MO_16, VdV_off, VuV_off, shift, \
279 sizeof(MMVector), sizeof(MMVector)); \
282 #define fGEN_TCG_V6_vaslh_acc(SHORTCODE) \
284 intptr_t tmpoff = offsetof(CPUHexagonState, vtmp); \
285 TCGv shift = tcg_temp_new(); \
286 tcg_gen_andi_tl(shift, RtV, 15); \
287 tcg_gen_gvec_shls(MO_16, tmpoff, VuV_off, shift, \
288 sizeof(MMVector), sizeof(MMVector)); \
289 tcg_gen_gvec_add(MO_16, VxV_off, VxV_off, tmpoff, \
290 sizeof(MMVector), sizeof(MMVector)); \
293 #define fGEN_TCG_V6_vaslw(SHORTCODE) \
295 TCGv shift = tcg_temp_new(); \
296 tcg_gen_andi_tl(shift, RtV, 31); \
297 tcg_gen_gvec_shls(MO_32, VdV_off, VuV_off, shift, \
298 sizeof(MMVector), sizeof(MMVector)); \
301 #define fGEN_TCG_V6_vaslw_acc(SHORTCODE) \
303 intptr_t tmpoff = offsetof(CPUHexagonState, vtmp); \
304 TCGv shift = tcg_temp_new(); \
305 tcg_gen_andi_tl(shift, RtV, 31); \
306 tcg_gen_gvec_shls(MO_32, tmpoff, VuV_off, shift, \
307 sizeof(MMVector), sizeof(MMVector)); \
308 tcg_gen_gvec_add(MO_32, VxV_off, VxV_off, tmpoff, \
309 sizeof(MMVector), sizeof(MMVector)); \
312 /* Vector max - various forms */
313 #define fGEN_TCG_V6_vmaxw(SHORTCODE) \
314 tcg_gen_gvec_smax(MO_32, VdV_off, VuV_off, VvV_off, \
315 sizeof(MMVector), sizeof(MMVector))
316 #define fGEN_TCG_V6_vmaxh(SHORTCODE) \
317 tcg_gen_gvec_smax(MO_16, VdV_off, VuV_off, VvV_off, \
318 sizeof(MMVector), sizeof(MMVector))
319 #define fGEN_TCG_V6_vmaxuh(SHORTCODE) \
320 tcg_gen_gvec_umax(MO_16, VdV_off, VuV_off, VvV_off, \
321 sizeof(MMVector), sizeof(MMVector))
322 #define fGEN_TCG_V6_vmaxb(SHORTCODE) \
323 tcg_gen_gvec_smax(MO_8, VdV_off, VuV_off, VvV_off, \
324 sizeof(MMVector), sizeof(MMVector))
325 #define fGEN_TCG_V6_vmaxub(SHORTCODE) \
326 tcg_gen_gvec_umax(MO_8, VdV_off, VuV_off, VvV_off, \
327 sizeof(MMVector), sizeof(MMVector))
329 /* Vector min - various forms */
330 #define fGEN_TCG_V6_vminw(SHORTCODE) \
331 tcg_gen_gvec_smin(MO_32, VdV_off, VuV_off, VvV_off, \
332 sizeof(MMVector), sizeof(MMVector))
333 #define fGEN_TCG_V6_vminh(SHORTCODE) \
334 tcg_gen_gvec_smin(MO_16, VdV_off, VuV_off, VvV_off, \
335 sizeof(MMVector), sizeof(MMVector))
336 #define fGEN_TCG_V6_vminuh(SHORTCODE) \
337 tcg_gen_gvec_umin(MO_16, VdV_off, VuV_off, VvV_off, \
338 sizeof(MMVector), sizeof(MMVector))
339 #define fGEN_TCG_V6_vminb(SHORTCODE) \
340 tcg_gen_gvec_smin(MO_8, VdV_off, VuV_off, VvV_off, \
341 sizeof(MMVector), sizeof(MMVector))
342 #define fGEN_TCG_V6_vminub(SHORTCODE) \
343 tcg_gen_gvec_umin(MO_8, VdV_off, VuV_off, VvV_off, \
344 sizeof(MMVector), sizeof(MMVector))
346 /* Vector logical ops */
347 #define fGEN_TCG_V6_vxor(SHORTCODE) \
348 tcg_gen_gvec_xor(MO_64, VdV_off, VuV_off, VvV_off, \
349 sizeof(MMVector), sizeof(MMVector))
351 #define fGEN_TCG_V6_vand(SHORTCODE) \
352 tcg_gen_gvec_and(MO_64, VdV_off, VuV_off, VvV_off, \
353 sizeof(MMVector), sizeof(MMVector))
355 #define fGEN_TCG_V6_vor(SHORTCODE) \
356 tcg_gen_gvec_or(MO_64, VdV_off, VuV_off, VvV_off, \
357 sizeof(MMVector), sizeof(MMVector))
359 #define fGEN_TCG_V6_vnot(SHORTCODE) \
360 tcg_gen_gvec_not(MO_64, VdV_off, VuV_off, \
361 sizeof(MMVector), sizeof(MMVector))
363 /* Q register logical ops */
364 #define fGEN_TCG_V6_pred_or(SHORTCODE) \
365 tcg_gen_gvec_or(MO_64, QdV_off, QsV_off, QtV_off, \
366 sizeof(MMQReg), sizeof(MMQReg))
368 #define fGEN_TCG_V6_pred_and(SHORTCODE) \
369 tcg_gen_gvec_and(MO_64, QdV_off, QsV_off, QtV_off, \
370 sizeof(MMQReg), sizeof(MMQReg))
372 #define fGEN_TCG_V6_pred_xor(SHORTCODE) \
373 tcg_gen_gvec_xor(MO_64, QdV_off, QsV_off, QtV_off, \
374 sizeof(MMQReg), sizeof(MMQReg))
376 #define fGEN_TCG_V6_pred_or_n(SHORTCODE) \
377 tcg_gen_gvec_orc(MO_64, QdV_off, QsV_off, QtV_off, \
378 sizeof(MMQReg), sizeof(MMQReg))
380 #define fGEN_TCG_V6_pred_and_n(SHORTCODE) \
381 tcg_gen_gvec_andc(MO_64, QdV_off, QsV_off, QtV_off, \
382 sizeof(MMQReg), sizeof(MMQReg))
384 #define fGEN_TCG_V6_pred_not(SHORTCODE) \
385 tcg_gen_gvec_not(MO_64, QdV_off, QsV_off, \
386 sizeof(MMQReg), sizeof(MMQReg))
388 /* Vector compares */
389 #define fGEN_TCG_VEC_CMP(COND, TYPE, SIZE) \
391 intptr_t tmpoff = offsetof(CPUHexagonState, vtmp); \
392 tcg_gen_gvec_cmp(COND, TYPE, tmpoff, VuV_off, VvV_off, \
393 sizeof(MMVector), sizeof(MMVector)); \
394 vec_to_qvec(SIZE, QdV_off, tmpoff); \
397 #define fGEN_TCG_V6_vgtw(SHORTCODE) \
398 fGEN_TCG_VEC_CMP(TCG_COND_GT, MO_32, 4)
399 #define fGEN_TCG_V6_vgth(SHORTCODE) \
400 fGEN_TCG_VEC_CMP(TCG_COND_GT, MO_16, 2)
401 #define fGEN_TCG_V6_vgtb(SHORTCODE) \
402 fGEN_TCG_VEC_CMP(TCG_COND_GT, MO_8, 1)
404 #define fGEN_TCG_V6_vgtuw(SHORTCODE) \
405 fGEN_TCG_VEC_CMP(TCG_COND_GTU, MO_32, 4)
406 #define fGEN_TCG_V6_vgtuh(SHORTCODE) \
407 fGEN_TCG_VEC_CMP(TCG_COND_GTU, MO_16, 2)
408 #define fGEN_TCG_V6_vgtub(SHORTCODE) \
409 fGEN_TCG_VEC_CMP(TCG_COND_GTU, MO_8, 1)
411 #define fGEN_TCG_V6_veqw(SHORTCODE) \
412 fGEN_TCG_VEC_CMP(TCG_COND_EQ, MO_32, 4)
413 #define fGEN_TCG_V6_veqh(SHORTCODE) \
414 fGEN_TCG_VEC_CMP(TCG_COND_EQ, MO_16, 2)
415 #define fGEN_TCG_V6_veqb(SHORTCODE) \
416 fGEN_TCG_VEC_CMP(TCG_COND_EQ, MO_8, 1)
418 #define fGEN_TCG_VEC_CMP_OP(COND, TYPE, SIZE, OP) \
420 intptr_t tmpoff = offsetof(CPUHexagonState, vtmp); \
421 intptr_t qoff = offsetof(CPUHexagonState, qtmp); \
422 tcg_gen_gvec_cmp(COND, TYPE, tmpoff, VuV_off, VvV_off, \
423 sizeof(MMVector), sizeof(MMVector)); \
424 vec_to_qvec(SIZE, qoff, tmpoff); \
425 OP(MO_64, QxV_off, QxV_off, qoff, sizeof(MMQReg), sizeof(MMQReg)); \
428 #define fGEN_TCG_V6_vgtw_and(SHORTCODE) \
429 fGEN_TCG_VEC_CMP_OP(TCG_COND_GT, MO_32, 4, tcg_gen_gvec_and)
430 #define fGEN_TCG_V6_vgtw_or(SHORTCODE) \
431 fGEN_TCG_VEC_CMP_OP(TCG_COND_GT, MO_32, 4, tcg_gen_gvec_or)
432 #define fGEN_TCG_V6_vgtw_xor(SHORTCODE) \
433 fGEN_TCG_VEC_CMP_OP(TCG_COND_GT, MO_32, 4, tcg_gen_gvec_xor)
435 #define fGEN_TCG_V6_vgtuw_and(SHORTCODE) \
436 fGEN_TCG_VEC_CMP_OP(TCG_COND_GTU, MO_32, 4, tcg_gen_gvec_and)
437 #define fGEN_TCG_V6_vgtuw_or(SHORTCODE) \
438 fGEN_TCG_VEC_CMP_OP(TCG_COND_GTU, MO_32, 4, tcg_gen_gvec_or)
439 #define fGEN_TCG_V6_vgtuw_xor(SHORTCODE) \
440 fGEN_TCG_VEC_CMP_OP(TCG_COND_GTU, MO_32, 4, tcg_gen_gvec_xor)
442 #define fGEN_TCG_V6_vgth_and(SHORTCODE) \
443 fGEN_TCG_VEC_CMP_OP(TCG_COND_GT, MO_16, 2, tcg_gen_gvec_and)
444 #define fGEN_TCG_V6_vgth_or(SHORTCODE) \
445 fGEN_TCG_VEC_CMP_OP(TCG_COND_GT, MO_16, 2, tcg_gen_gvec_or)
446 #define fGEN_TCG_V6_vgth_xor(SHORTCODE) \
447 fGEN_TCG_VEC_CMP_OP(TCG_COND_GT, MO_16, 2, tcg_gen_gvec_xor)
449 #define fGEN_TCG_V6_vgtuh_and(SHORTCODE) \
450 fGEN_TCG_VEC_CMP_OP(TCG_COND_GTU, MO_16, 2, tcg_gen_gvec_and)
451 #define fGEN_TCG_V6_vgtuh_or(SHORTCODE) \
452 fGEN_TCG_VEC_CMP_OP(TCG_COND_GTU, MO_16, 2, tcg_gen_gvec_or)
453 #define fGEN_TCG_V6_vgtuh_xor(SHORTCODE) \
454 fGEN_TCG_VEC_CMP_OP(TCG_COND_GTU, MO_16, 2, tcg_gen_gvec_xor)
456 #define fGEN_TCG_V6_vgtb_and(SHORTCODE) \
457 fGEN_TCG_VEC_CMP_OP(TCG_COND_GT, MO_8, 1, tcg_gen_gvec_and)
458 #define fGEN_TCG_V6_vgtb_or(SHORTCODE) \
459 fGEN_TCG_VEC_CMP_OP(TCG_COND_GT, MO_8, 1, tcg_gen_gvec_or)
460 #define fGEN_TCG_V6_vgtb_xor(SHORTCODE) \
461 fGEN_TCG_VEC_CMP_OP(TCG_COND_GT, MO_8, 1, tcg_gen_gvec_xor)
463 #define fGEN_TCG_V6_vgtub_and(SHORTCODE) \
464 fGEN_TCG_VEC_CMP_OP(TCG_COND_GTU, MO_8, 1, tcg_gen_gvec_and)
465 #define fGEN_TCG_V6_vgtub_or(SHORTCODE) \
466 fGEN_TCG_VEC_CMP_OP(TCG_COND_GTU, MO_8, 1, tcg_gen_gvec_or)
467 #define fGEN_TCG_V6_vgtub_xor(SHORTCODE) \
468 fGEN_TCG_VEC_CMP_OP(TCG_COND_GTU, MO_8, 1, tcg_gen_gvec_xor)
470 #define fGEN_TCG_V6_veqw_and(SHORTCODE) \
471 fGEN_TCG_VEC_CMP_OP(TCG_COND_EQ, MO_32, 4, tcg_gen_gvec_and)
472 #define fGEN_TCG_V6_veqw_or(SHORTCODE) \
473 fGEN_TCG_VEC_CMP_OP(TCG_COND_EQ, MO_32, 4, tcg_gen_gvec_or)
474 #define fGEN_TCG_V6_veqw_xor(SHORTCODE) \
475 fGEN_TCG_VEC_CMP_OP(TCG_COND_EQ, MO_32, 4, tcg_gen_gvec_xor)
477 #define fGEN_TCG_V6_veqh_and(SHORTCODE) \
478 fGEN_TCG_VEC_CMP_OP(TCG_COND_EQ, MO_16, 2, tcg_gen_gvec_and)
479 #define fGEN_TCG_V6_veqh_or(SHORTCODE) \
480 fGEN_TCG_VEC_CMP_OP(TCG_COND_EQ, MO_16, 2, tcg_gen_gvec_or)
481 #define fGEN_TCG_V6_veqh_xor(SHORTCODE) \
482 fGEN_TCG_VEC_CMP_OP(TCG_COND_EQ, MO_16, 2, tcg_gen_gvec_xor)
484 #define fGEN_TCG_V6_veqb_and(SHORTCODE) \
485 fGEN_TCG_VEC_CMP_OP(TCG_COND_EQ, MO_8, 1, tcg_gen_gvec_and)
486 #define fGEN_TCG_V6_veqb_or(SHORTCODE) \
487 fGEN_TCG_VEC_CMP_OP(TCG_COND_EQ, MO_8, 1, tcg_gen_gvec_or)
488 #define fGEN_TCG_V6_veqb_xor(SHORTCODE) \
489 fGEN_TCG_VEC_CMP_OP(TCG_COND_EQ, MO_8, 1, tcg_gen_gvec_xor)
491 /* Vector splat - various forms */
492 #define fGEN_TCG_V6_lvsplatw(SHORTCODE) \
493 tcg_gen_gvec_dup_i32(MO_32, VdV_off, \
494 sizeof(MMVector), sizeof(MMVector), RtV)
496 #define fGEN_TCG_V6_lvsplath(SHORTCODE) \
497 tcg_gen_gvec_dup_i32(MO_16, VdV_off, \
498 sizeof(MMVector), sizeof(MMVector), RtV)
500 #define fGEN_TCG_V6_lvsplatb(SHORTCODE) \
501 tcg_gen_gvec_dup_i32(MO_8, VdV_off, \
502 sizeof(MMVector), sizeof(MMVector), RtV)
504 /* Vector absolute value - various forms */
505 #define fGEN_TCG_V6_vabsb(SHORTCODE) \
506 tcg_gen_gvec_abs(MO_8, VdV_off, VuV_off, \
507 sizeof(MMVector), sizeof(MMVector))
509 #define fGEN_TCG_V6_vabsh(SHORTCODE) \
510 tcg_gen_gvec_abs(MO_16, VdV_off, VuV_off, \
511 sizeof(MMVector), sizeof(MMVector))
513 #define fGEN_TCG_V6_vabsw(SHORTCODE) \
514 tcg_gen_gvec_abs(MO_32, VdV_off, VuV_off, \
515 sizeof(MMVector), sizeof(MMVector))
518 #define fGEN_TCG_V6_vL32b_pi(SHORTCODE) SHORTCODE
519 #define fGEN_TCG_V6_vL32Ub_pi(SHORTCODE) SHORTCODE
520 #define fGEN_TCG_V6_vL32b_cur_pi(SHORTCODE) SHORTCODE
521 #define fGEN_TCG_V6_vL32b_tmp_pi(SHORTCODE) SHORTCODE
522 #define fGEN_TCG_V6_vL32b_nt_pi(SHORTCODE) SHORTCODE
523 #define fGEN_TCG_V6_vL32b_nt_cur_pi(SHORTCODE) SHORTCODE
524 #define fGEN_TCG_V6_vL32b_nt_tmp_pi(SHORTCODE) SHORTCODE
525 #define fGEN_TCG_V6_vL32b_ai(SHORTCODE) SHORTCODE
526 #define fGEN_TCG_V6_vL32Ub_ai(SHORTCODE) SHORTCODE
527 #define fGEN_TCG_V6_vL32b_cur_ai(SHORTCODE) SHORTCODE
528 #define fGEN_TCG_V6_vL32b_tmp_ai(SHORTCODE) SHORTCODE
529 #define fGEN_TCG_V6_vL32b_nt_ai(SHORTCODE) SHORTCODE
530 #define fGEN_TCG_V6_vL32b_nt_cur_ai(SHORTCODE) SHORTCODE
531 #define fGEN_TCG_V6_vL32b_nt_tmp_ai(SHORTCODE) SHORTCODE
532 #define fGEN_TCG_V6_vL32b_ppu(SHORTCODE) SHORTCODE
533 #define fGEN_TCG_V6_vL32Ub_ppu(SHORTCODE) SHORTCODE
534 #define fGEN_TCG_V6_vL32b_cur_ppu(SHORTCODE) SHORTCODE
535 #define fGEN_TCG_V6_vL32b_tmp_ppu(SHORTCODE) SHORTCODE
536 #define fGEN_TCG_V6_vL32b_nt_ppu(SHORTCODE) SHORTCODE
537 #define fGEN_TCG_V6_vL32b_nt_cur_ppu(SHORTCODE) SHORTCODE
538 #define fGEN_TCG_V6_vL32b_nt_tmp_ppu(SHORTCODE) SHORTCODE
540 /* Predicated vector loads */
541 #define fGEN_TCG_PRED_VEC_LOAD(GET_EA, PRED, DSTOFF, INC) \
543 TCGv LSB = tcg_temp_new(); \
544 TCGLabel *false_label = gen_new_label(); \
547 tcg_gen_brcondi_tl(TCG_COND_EQ, LSB, 0, false_label); \
548 gen_vreg_load(ctx, DSTOFF, EA, true); \
550 gen_set_label(false_label); \
553 #define fGEN_TCG_PRED_VEC_LOAD_pred_pi \
554 fGEN_TCG_PRED_VEC_LOAD(fLSBOLD(PvV), \
557 fPM_I(RxV, siV * sizeof(MMVector)))
558 #define fGEN_TCG_PRED_VEC_LOAD_npred_pi \
559 fGEN_TCG_PRED_VEC_LOAD(fLSBOLDNOT(PvV), \
562 fPM_I(RxV, siV * sizeof(MMVector)))
564 #define fGEN_TCG_V6_vL32b_pred_pi(SHORTCODE) \
565 fGEN_TCG_PRED_VEC_LOAD_pred_pi
566 #define fGEN_TCG_V6_vL32b_npred_pi(SHORTCODE) \
567 fGEN_TCG_PRED_VEC_LOAD_npred_pi
568 #define fGEN_TCG_V6_vL32b_cur_pred_pi(SHORTCODE) \
569 fGEN_TCG_PRED_VEC_LOAD_pred_pi
570 #define fGEN_TCG_V6_vL32b_cur_npred_pi(SHORTCODE) \
571 fGEN_TCG_PRED_VEC_LOAD_npred_pi
572 #define fGEN_TCG_V6_vL32b_tmp_pred_pi(SHORTCODE) \
573 fGEN_TCG_PRED_VEC_LOAD_pred_pi
574 #define fGEN_TCG_V6_vL32b_tmp_npred_pi(SHORTCODE) \
575 fGEN_TCG_PRED_VEC_LOAD_npred_pi
576 #define fGEN_TCG_V6_vL32b_nt_pred_pi(SHORTCODE) \
577 fGEN_TCG_PRED_VEC_LOAD_pred_pi
578 #define fGEN_TCG_V6_vL32b_nt_npred_pi(SHORTCODE) \
579 fGEN_TCG_PRED_VEC_LOAD_npred_pi
580 #define fGEN_TCG_V6_vL32b_nt_cur_pred_pi(SHORTCODE) \
581 fGEN_TCG_PRED_VEC_LOAD_pred_pi
582 #define fGEN_TCG_V6_vL32b_nt_cur_npred_pi(SHORTCODE) \
583 fGEN_TCG_PRED_VEC_LOAD_npred_pi
584 #define fGEN_TCG_V6_vL32b_nt_tmp_pred_pi(SHORTCODE) \
585 fGEN_TCG_PRED_VEC_LOAD_pred_pi
586 #define fGEN_TCG_V6_vL32b_nt_tmp_npred_pi(SHORTCODE) \
587 fGEN_TCG_PRED_VEC_LOAD_npred_pi
589 #define fGEN_TCG_PRED_VEC_LOAD_pred_ai \
590 fGEN_TCG_PRED_VEC_LOAD(fLSBOLD(PvV), \
591 fEA_RI(RtV, siV * sizeof(MMVector)), \
594 #define fGEN_TCG_PRED_VEC_LOAD_npred_ai \
595 fGEN_TCG_PRED_VEC_LOAD(fLSBOLDNOT(PvV), \
596 fEA_RI(RtV, siV * sizeof(MMVector)), \
600 #define fGEN_TCG_V6_vL32b_pred_ai(SHORTCODE) \
601 fGEN_TCG_PRED_VEC_LOAD_pred_ai
602 #define fGEN_TCG_V6_vL32b_npred_ai(SHORTCODE) \
603 fGEN_TCG_PRED_VEC_LOAD_npred_ai
604 #define fGEN_TCG_V6_vL32b_cur_pred_ai(SHORTCODE) \
605 fGEN_TCG_PRED_VEC_LOAD_pred_ai
606 #define fGEN_TCG_V6_vL32b_cur_npred_ai(SHORTCODE) \
607 fGEN_TCG_PRED_VEC_LOAD_npred_ai
608 #define fGEN_TCG_V6_vL32b_tmp_pred_ai(SHORTCODE) \
609 fGEN_TCG_PRED_VEC_LOAD_pred_ai
610 #define fGEN_TCG_V6_vL32b_tmp_npred_ai(SHORTCODE) \
611 fGEN_TCG_PRED_VEC_LOAD_npred_ai
612 #define fGEN_TCG_V6_vL32b_nt_pred_ai(SHORTCODE) \
613 fGEN_TCG_PRED_VEC_LOAD_pred_ai
614 #define fGEN_TCG_V6_vL32b_nt_npred_ai(SHORTCODE) \
615 fGEN_TCG_PRED_VEC_LOAD_npred_ai
616 #define fGEN_TCG_V6_vL32b_nt_cur_pred_ai(SHORTCODE) \
617 fGEN_TCG_PRED_VEC_LOAD_pred_ai
618 #define fGEN_TCG_V6_vL32b_nt_cur_npred_ai(SHORTCODE) \
619 fGEN_TCG_PRED_VEC_LOAD_npred_ai
620 #define fGEN_TCG_V6_vL32b_nt_tmp_pred_ai(SHORTCODE) \
621 fGEN_TCG_PRED_VEC_LOAD_pred_ai
622 #define fGEN_TCG_V6_vL32b_nt_tmp_npred_ai(SHORTCODE) \
623 fGEN_TCG_PRED_VEC_LOAD_npred_ai
625 #define fGEN_TCG_PRED_VEC_LOAD_pred_ppu \
626 fGEN_TCG_PRED_VEC_LOAD(fLSBOLD(PvV), \
630 #define fGEN_TCG_PRED_VEC_LOAD_npred_ppu \
631 fGEN_TCG_PRED_VEC_LOAD(fLSBOLDNOT(PvV), \
636 #define fGEN_TCG_V6_vL32b_pred_ppu(SHORTCODE) \
637 fGEN_TCG_PRED_VEC_LOAD_pred_ppu
638 #define fGEN_TCG_V6_vL32b_npred_ppu(SHORTCODE) \
639 fGEN_TCG_PRED_VEC_LOAD_npred_ppu
640 #define fGEN_TCG_V6_vL32b_cur_pred_ppu(SHORTCODE) \
641 fGEN_TCG_PRED_VEC_LOAD_pred_ppu
642 #define fGEN_TCG_V6_vL32b_cur_npred_ppu(SHORTCODE) \
643 fGEN_TCG_PRED_VEC_LOAD_npred_ppu
644 #define fGEN_TCG_V6_vL32b_tmp_pred_ppu(SHORTCODE) \
645 fGEN_TCG_PRED_VEC_LOAD_pred_ppu
646 #define fGEN_TCG_V6_vL32b_tmp_npred_ppu(SHORTCODE) \
647 fGEN_TCG_PRED_VEC_LOAD_npred_ppu
648 #define fGEN_TCG_V6_vL32b_nt_pred_ppu(SHORTCODE) \
649 fGEN_TCG_PRED_VEC_LOAD_pred_ppu
650 #define fGEN_TCG_V6_vL32b_nt_npred_ppu(SHORTCODE) \
651 fGEN_TCG_PRED_VEC_LOAD_npred_ppu
652 #define fGEN_TCG_V6_vL32b_nt_cur_pred_ppu(SHORTCODE) \
653 fGEN_TCG_PRED_VEC_LOAD_pred_ppu
654 #define fGEN_TCG_V6_vL32b_nt_cur_npred_ppu(SHORTCODE) \
655 fGEN_TCG_PRED_VEC_LOAD_npred_ppu
656 #define fGEN_TCG_V6_vL32b_nt_tmp_pred_ppu(SHORTCODE) \
657 fGEN_TCG_PRED_VEC_LOAD_pred_ppu
658 #define fGEN_TCG_V6_vL32b_nt_tmp_npred_ppu(SHORTCODE) \
659 fGEN_TCG_PRED_VEC_LOAD_npred_ppu
662 #define fGEN_TCG_V6_vS32b_pi(SHORTCODE) SHORTCODE
663 #define fGEN_TCG_V6_vS32Ub_pi(SHORTCODE) SHORTCODE
664 #define fGEN_TCG_V6_vS32b_nt_pi(SHORTCODE) SHORTCODE
665 #define fGEN_TCG_V6_vS32b_ai(SHORTCODE) SHORTCODE
666 #define fGEN_TCG_V6_vS32Ub_ai(SHORTCODE) SHORTCODE
667 #define fGEN_TCG_V6_vS32b_nt_ai(SHORTCODE) SHORTCODE
668 #define fGEN_TCG_V6_vS32b_ppu(SHORTCODE) SHORTCODE
669 #define fGEN_TCG_V6_vS32Ub_ppu(SHORTCODE) SHORTCODE
670 #define fGEN_TCG_V6_vS32b_nt_ppu(SHORTCODE) SHORTCODE
672 /* New value vector stores */
673 #define fGEN_TCG_NEWVAL_VEC_STORE(GET_EA, INC) \
676 gen_vreg_store(ctx, EA, OsN_off, insn->slot, true); \
680 #define fGEN_TCG_NEWVAL_VEC_STORE_pi \
681 fGEN_TCG_NEWVAL_VEC_STORE(fEA_REG(RxV), fPM_I(RxV, siV * sizeof(MMVector)))
683 #define fGEN_TCG_V6_vS32b_new_pi(SHORTCODE) \
684 fGEN_TCG_NEWVAL_VEC_STORE_pi
685 #define fGEN_TCG_V6_vS32b_nt_new_pi(SHORTCODE) \
686 fGEN_TCG_NEWVAL_VEC_STORE_pi
688 #define fGEN_TCG_NEWVAL_VEC_STORE_ai \
689 fGEN_TCG_NEWVAL_VEC_STORE(fEA_RI(RtV, siV * sizeof(MMVector)), \
692 #define fGEN_TCG_V6_vS32b_new_ai(SHORTCODE) \
693 fGEN_TCG_NEWVAL_VEC_STORE_ai
694 #define fGEN_TCG_V6_vS32b_nt_new_ai(SHORTCODE) \
695 fGEN_TCG_NEWVAL_VEC_STORE_ai
697 #define fGEN_TCG_NEWVAL_VEC_STORE_ppu \
698 fGEN_TCG_NEWVAL_VEC_STORE(fEA_REG(RxV), fPM_M(RxV, MuV))
700 #define fGEN_TCG_V6_vS32b_new_ppu(SHORTCODE) \
701 fGEN_TCG_NEWVAL_VEC_STORE_ppu
702 #define fGEN_TCG_V6_vS32b_nt_new_ppu(SHORTCODE) \
703 fGEN_TCG_NEWVAL_VEC_STORE_ppu
705 /* Predicated vector stores */
706 #define fGEN_TCG_PRED_VEC_STORE(GET_EA, PRED, SRCOFF, ALIGN, INC) \
708 TCGv LSB = tcg_temp_new(); \
709 TCGLabel *false_label = gen_new_label(); \
712 tcg_gen_brcondi_tl(TCG_COND_EQ, LSB, 0, false_label); \
713 gen_vreg_store(ctx, EA, SRCOFF, insn->slot, ALIGN); \
715 gen_set_label(false_label); \
718 #define fGEN_TCG_PRED_VEC_STORE_pred_pi(ALIGN) \
719 fGEN_TCG_PRED_VEC_STORE(fLSBOLD(PvV), \
722 fPM_I(RxV, siV * sizeof(MMVector)))
723 #define fGEN_TCG_PRED_VEC_STORE_npred_pi(ALIGN) \
724 fGEN_TCG_PRED_VEC_STORE(fLSBOLDNOT(PvV), \
727 fPM_I(RxV, siV * sizeof(MMVector)))
728 #define fGEN_TCG_PRED_VEC_STORE_new_pred_pi \
729 fGEN_TCG_PRED_VEC_STORE(fLSBOLD(PvV), \
732 fPM_I(RxV, siV * sizeof(MMVector)))
733 #define fGEN_TCG_PRED_VEC_STORE_new_npred_pi \
734 fGEN_TCG_PRED_VEC_STORE(fLSBOLDNOT(PvV), \
737 fPM_I(RxV, siV * sizeof(MMVector)))
739 #define fGEN_TCG_V6_vS32b_pred_pi(SHORTCODE) \
740 fGEN_TCG_PRED_VEC_STORE_pred_pi(true)
741 #define fGEN_TCG_V6_vS32b_npred_pi(SHORTCODE) \
742 fGEN_TCG_PRED_VEC_STORE_npred_pi(true)
743 #define fGEN_TCG_V6_vS32Ub_pred_pi(SHORTCODE) \
744 fGEN_TCG_PRED_VEC_STORE_pred_pi(false)
745 #define fGEN_TCG_V6_vS32Ub_npred_pi(SHORTCODE) \
746 fGEN_TCG_PRED_VEC_STORE_npred_pi(false)
747 #define fGEN_TCG_V6_vS32b_nt_pred_pi(SHORTCODE) \
748 fGEN_TCG_PRED_VEC_STORE_pred_pi(true)
749 #define fGEN_TCG_V6_vS32b_nt_npred_pi(SHORTCODE) \
750 fGEN_TCG_PRED_VEC_STORE_npred_pi(true)
751 #define fGEN_TCG_V6_vS32b_new_pred_pi(SHORTCODE) \
752 fGEN_TCG_PRED_VEC_STORE_new_pred_pi
753 #define fGEN_TCG_V6_vS32b_new_npred_pi(SHORTCODE) \
754 fGEN_TCG_PRED_VEC_STORE_new_npred_pi
755 #define fGEN_TCG_V6_vS32b_nt_new_pred_pi(SHORTCODE) \
756 fGEN_TCG_PRED_VEC_STORE_new_pred_pi
757 #define fGEN_TCG_V6_vS32b_nt_new_npred_pi(SHORTCODE) \
758 fGEN_TCG_PRED_VEC_STORE_new_npred_pi
760 #define fGEN_TCG_PRED_VEC_STORE_pred_ai(ALIGN) \
761 fGEN_TCG_PRED_VEC_STORE(fLSBOLD(PvV), \
762 fEA_RI(RtV, siV * sizeof(MMVector)), \
765 #define fGEN_TCG_PRED_VEC_STORE_npred_ai(ALIGN) \
766 fGEN_TCG_PRED_VEC_STORE(fLSBOLDNOT(PvV), \
767 fEA_RI(RtV, siV * sizeof(MMVector)), \
770 #define fGEN_TCG_PRED_VEC_STORE_new_pred_ai \
771 fGEN_TCG_PRED_VEC_STORE(fLSBOLD(PvV), \
772 fEA_RI(RtV, siV * sizeof(MMVector)), \
775 #define fGEN_TCG_PRED_VEC_STORE_new_npred_ai \
776 fGEN_TCG_PRED_VEC_STORE(fLSBOLDNOT(PvV), \
777 fEA_RI(RtV, siV * sizeof(MMVector)), \
781 #define fGEN_TCG_V6_vS32b_pred_ai(SHORTCODE) \
782 fGEN_TCG_PRED_VEC_STORE_pred_ai(true)
783 #define fGEN_TCG_V6_vS32b_npred_ai(SHORTCODE) \
784 fGEN_TCG_PRED_VEC_STORE_npred_ai(true)
785 #define fGEN_TCG_V6_vS32Ub_pred_ai(SHORTCODE) \
786 fGEN_TCG_PRED_VEC_STORE_pred_ai(false)
787 #define fGEN_TCG_V6_vS32Ub_npred_ai(SHORTCODE) \
788 fGEN_TCG_PRED_VEC_STORE_npred_ai(false)
789 #define fGEN_TCG_V6_vS32b_nt_pred_ai(SHORTCODE) \
790 fGEN_TCG_PRED_VEC_STORE_pred_ai(true)
791 #define fGEN_TCG_V6_vS32b_nt_npred_ai(SHORTCODE) \
792 fGEN_TCG_PRED_VEC_STORE_npred_ai(true)
793 #define fGEN_TCG_V6_vS32b_new_pred_ai(SHORTCODE) \
794 fGEN_TCG_PRED_VEC_STORE_new_pred_ai
795 #define fGEN_TCG_V6_vS32b_new_npred_ai(SHORTCODE) \
796 fGEN_TCG_PRED_VEC_STORE_new_npred_ai
797 #define fGEN_TCG_V6_vS32b_nt_new_pred_ai(SHORTCODE) \
798 fGEN_TCG_PRED_VEC_STORE_new_pred_ai
799 #define fGEN_TCG_V6_vS32b_nt_new_npred_ai(SHORTCODE) \
800 fGEN_TCG_PRED_VEC_STORE_new_npred_ai
802 #define fGEN_TCG_PRED_VEC_STORE_pred_ppu(ALIGN) \
803 fGEN_TCG_PRED_VEC_STORE(fLSBOLD(PvV), \
807 #define fGEN_TCG_PRED_VEC_STORE_npred_ppu(ALIGN) \
808 fGEN_TCG_PRED_VEC_STORE(fLSBOLDNOT(PvV), \
812 #define fGEN_TCG_PRED_VEC_STORE_new_pred_ppu \
813 fGEN_TCG_PRED_VEC_STORE(fLSBOLD(PvV), \
817 #define fGEN_TCG_PRED_VEC_STORE_new_npred_ppu \
818 fGEN_TCG_PRED_VEC_STORE(fLSBOLDNOT(PvV), \
823 #define fGEN_TCG_V6_vS32b_pred_ppu(SHORTCODE) \
824 fGEN_TCG_PRED_VEC_STORE_pred_ppu(true)
825 #define fGEN_TCG_V6_vS32b_npred_ppu(SHORTCODE) \
826 fGEN_TCG_PRED_VEC_STORE_npred_ppu(true)
827 #define fGEN_TCG_V6_vS32Ub_pred_ppu(SHORTCODE) \
828 fGEN_TCG_PRED_VEC_STORE_pred_ppu(false)
829 #define fGEN_TCG_V6_vS32Ub_npred_ppu(SHORTCODE) \
830 fGEN_TCG_PRED_VEC_STORE_npred_ppu(false)
831 #define fGEN_TCG_V6_vS32b_nt_pred_ppu(SHORTCODE) \
832 fGEN_TCG_PRED_VEC_STORE_pred_ppu(true)
833 #define fGEN_TCG_V6_vS32b_nt_npred_ppu(SHORTCODE) \
834 fGEN_TCG_PRED_VEC_STORE_npred_ppu(true)
835 #define fGEN_TCG_V6_vS32b_new_pred_ppu(SHORTCODE) \
836 fGEN_TCG_PRED_VEC_STORE_new_pred_ppu
837 #define fGEN_TCG_V6_vS32b_new_npred_ppu(SHORTCODE) \
838 fGEN_TCG_PRED_VEC_STORE_new_npred_ppu
839 #define fGEN_TCG_V6_vS32b_nt_new_pred_ppu(SHORTCODE) \
840 fGEN_TCG_PRED_VEC_STORE_new_pred_ppu
841 #define fGEN_TCG_V6_vS32b_nt_new_npred_ppu(SHORTCODE) \
842 fGEN_TCG_PRED_VEC_STORE_new_npred_ppu
844 /* Masked vector stores */
845 #define fGEN_TCG_V6_vS32b_qpred_pi(SHORTCODE) SHORTCODE
846 #define fGEN_TCG_V6_vS32b_nt_qpred_pi(SHORTCODE) SHORTCODE
847 #define fGEN_TCG_V6_vS32b_qpred_ai(SHORTCODE) SHORTCODE
848 #define fGEN_TCG_V6_vS32b_nt_qpred_ai(SHORTCODE) SHORTCODE
849 #define fGEN_TCG_V6_vS32b_qpred_ppu(SHORTCODE) SHORTCODE
850 #define fGEN_TCG_V6_vS32b_nt_qpred_ppu(SHORTCODE) SHORTCODE
851 #define fGEN_TCG_V6_vS32b_nqpred_pi(SHORTCODE) SHORTCODE
852 #define fGEN_TCG_V6_vS32b_nt_nqpred_pi(SHORTCODE) SHORTCODE
853 #define fGEN_TCG_V6_vS32b_nqpred_ai(SHORTCODE) SHORTCODE
854 #define fGEN_TCG_V6_vS32b_nt_nqpred_ai(SHORTCODE) SHORTCODE
855 #define fGEN_TCG_V6_vS32b_nqpred_ppu(SHORTCODE) SHORTCODE
856 #define fGEN_TCG_V6_vS32b_nt_nqpred_ppu(SHORTCODE) SHORTCODE
858 /* Store release not modelled in qemu, but need to suppress compiler warnings */
859 #define fGEN_TCG_V6_vS32b_srls_pi(SHORTCODE) \
863 #define fGEN_TCG_V6_vS32b_srls_ai(SHORTCODE) \
868 #define fGEN_TCG_V6_vS32b_srls_ppu(SHORTCODE) \