]> git.proxmox.com Git - mirror_qemu.git/blob - target/s390x/tcg/misc_helper.c
system/cpus: rename qemu_mutex_lock_iothread() to bql_lock()
[mirror_qemu.git] / target / s390x / tcg / misc_helper.c
1 /*
2 * S/390 misc helper routines
3 *
4 * Copyright (c) 2009 Ulrich Hecht
5 * Copyright (c) 2009 Alexander Graf
6 *
7 * This library is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU Lesser General Public
9 * License as published by the Free Software Foundation; either
10 * version 2.1 of the License, or (at your option) any later version.
11 *
12 * This library is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15 * Lesser General Public License for more details.
16 *
17 * You should have received a copy of the GNU Lesser General Public
18 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
19 */
20
21 #include "qemu/osdep.h"
22 #include "qemu/cutils.h"
23 #include "cpu.h"
24 #include "s390x-internal.h"
25 #include "qemu/host-utils.h"
26 #include "exec/helper-proto.h"
27 #include "qemu/timer.h"
28 #include "exec/exec-all.h"
29 #include "exec/cpu_ldst.h"
30 #include "qapi/error.h"
31 #include "tcg_s390x.h"
32 #include "s390-tod.h"
33
34 #if !defined(CONFIG_USER_ONLY)
35 #include "sysemu/cpus.h"
36 #include "sysemu/sysemu.h"
37 #include "hw/s390x/ebcdic.h"
38 #include "hw/s390x/s390-virtio-hcall.h"
39 #include "hw/s390x/sclp.h"
40 #include "hw/s390x/s390_flic.h"
41 #include "hw/s390x/ioinst.h"
42 #include "hw/s390x/s390-pci-inst.h"
43 #include "hw/boards.h"
44 #include "hw/s390x/tod.h"
45 #endif
46
47 /* #define DEBUG_HELPER */
48 #ifdef DEBUG_HELPER
49 #define HELPER_LOG(x...) qemu_log(x)
50 #else
51 #define HELPER_LOG(x...)
52 #endif
53
54 /* Raise an exception statically from a TB. */
55 void HELPER(exception)(CPUS390XState *env, uint32_t excp)
56 {
57 CPUState *cs = env_cpu(env);
58
59 HELPER_LOG("%s: exception %d\n", __func__, excp);
60 cs->exception_index = excp;
61 cpu_loop_exit(cs);
62 }
63
64 /* Store CPU Timer (also used for EXTRACT CPU TIME) */
65 uint64_t HELPER(stpt)(CPUS390XState *env)
66 {
67 #if defined(CONFIG_USER_ONLY)
68 /*
69 * Fake a descending CPU timer. We could get negative values here,
70 * but we don't care as it is up to the OS when to process that
71 * interrupt and reset to > 0.
72 */
73 return UINT64_MAX - (uint64_t)cpu_get_host_ticks();
74 #else
75 return time2tod(env->cputm - qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL));
76 #endif
77 }
78
79 /* Store Clock */
80 uint64_t HELPER(stck)(CPUS390XState *env)
81 {
82 #ifdef CONFIG_USER_ONLY
83 struct timespec ts;
84 uint64_t ns;
85
86 clock_gettime(CLOCK_REALTIME, &ts);
87 ns = ts.tv_sec * NANOSECONDS_PER_SECOND + ts.tv_nsec;
88
89 return TOD_UNIX_EPOCH + time2tod(ns);
90 #else
91 S390TODState *td = s390_get_todstate();
92 S390TODClass *tdc = S390_TOD_GET_CLASS(td);
93 S390TOD tod;
94
95 tdc->get(td, &tod, &error_abort);
96 return tod.low;
97 #endif
98 }
99
100 #ifndef CONFIG_USER_ONLY
101 /* SCLP service call */
102 uint32_t HELPER(servc)(CPUS390XState *env, uint64_t r1, uint64_t r2)
103 {
104 bql_lock();
105 int r = sclp_service_call(env_archcpu(env), r1, r2);
106 bql_unlock();
107 if (r < 0) {
108 tcg_s390_program_interrupt(env, -r, GETPC());
109 }
110 return r;
111 }
112
113 void HELPER(diag)(CPUS390XState *env, uint32_t r1, uint32_t r3, uint32_t num)
114 {
115 uint64_t r;
116
117 switch (num) {
118 case 0x500:
119 /* KVM hypercall */
120 bql_lock();
121 r = s390_virtio_hypercall(env);
122 bql_unlock();
123 break;
124 case 0x44:
125 /* yield */
126 r = 0;
127 break;
128 case 0x308:
129 /* ipl */
130 bql_lock();
131 handle_diag_308(env, r1, r3, GETPC());
132 bql_unlock();
133 r = 0;
134 break;
135 case 0x288:
136 /* time bomb (watchdog) */
137 r = handle_diag_288(env, r1, r3);
138 break;
139 default:
140 r = -1;
141 break;
142 }
143
144 if (r) {
145 tcg_s390_program_interrupt(env, PGM_SPECIFICATION, GETPC());
146 }
147 }
148
149 /* Set Prefix */
150 void HELPER(spx)(CPUS390XState *env, uint64_t a1)
151 {
152 const uint32_t prefix = a1 & 0x7fffe000;
153 const uint32_t old_prefix = env->psa;
154 CPUState *cs = env_cpu(env);
155
156 if (prefix == old_prefix) {
157 return;
158 }
159 /*
160 * Since prefix got aligned to 8k and memory increments are a multiple of
161 * 8k checking the first page is sufficient
162 */
163 if (!mmu_absolute_addr_valid(prefix, true)) {
164 tcg_s390_program_interrupt(env, PGM_ADDRESSING, GETPC());
165 }
166
167 env->psa = prefix;
168 HELPER_LOG("prefix: %#x\n", prefix);
169 tlb_flush_page(cs, 0);
170 tlb_flush_page(cs, TARGET_PAGE_SIZE);
171 if (prefix != 0) {
172 tlb_flush_page(cs, prefix);
173 tlb_flush_page(cs, prefix + TARGET_PAGE_SIZE);
174 }
175 if (old_prefix != 0) {
176 tlb_flush_page(cs, old_prefix);
177 tlb_flush_page(cs, old_prefix + TARGET_PAGE_SIZE);
178 }
179 }
180
181 static void update_ckc_timer(CPUS390XState *env)
182 {
183 S390TODState *td = s390_get_todstate();
184 uint64_t time;
185
186 /* stop the timer and remove pending CKC IRQs */
187 timer_del(env->tod_timer);
188 g_assert(bql_locked());
189 env->pending_int &= ~INTERRUPT_EXT_CLOCK_COMPARATOR;
190
191 /* the tod has to exceed the ckc, this can never happen if ckc is all 1's */
192 if (env->ckc == -1ULL) {
193 return;
194 }
195
196 /* difference between origins */
197 time = env->ckc - td->base.low;
198
199 /* nanoseconds */
200 time = tod2time(time);
201
202 timer_mod(env->tod_timer, time);
203 }
204
205 /* Set Clock Comparator */
206 void HELPER(sckc)(CPUS390XState *env, uint64_t ckc)
207 {
208 env->ckc = ckc;
209
210 bql_lock();
211 update_ckc_timer(env);
212 bql_unlock();
213 }
214
215 void tcg_s390_tod_updated(CPUState *cs, run_on_cpu_data opaque)
216 {
217 S390CPU *cpu = S390_CPU(cs);
218
219 update_ckc_timer(&cpu->env);
220 }
221
222 /* Set Clock */
223 uint32_t HELPER(sck)(CPUS390XState *env, uint64_t tod_low)
224 {
225 S390TODState *td = s390_get_todstate();
226 S390TODClass *tdc = S390_TOD_GET_CLASS(td);
227 S390TOD tod = {
228 .high = 0,
229 .low = tod_low,
230 };
231
232 bql_lock();
233 tdc->set(td, &tod, &error_abort);
234 bql_unlock();
235 return 0;
236 }
237
238 /* Set Tod Programmable Field */
239 void HELPER(sckpf)(CPUS390XState *env, uint64_t r0)
240 {
241 uint32_t val = r0;
242
243 if (val & 0xffff0000) {
244 tcg_s390_program_interrupt(env, PGM_SPECIFICATION, GETPC());
245 }
246 env->todpr = val;
247 }
248
249 /* Store Clock Comparator */
250 uint64_t HELPER(stckc)(CPUS390XState *env)
251 {
252 return env->ckc;
253 }
254
255 /* Set CPU Timer */
256 void HELPER(spt)(CPUS390XState *env, uint64_t time)
257 {
258 if (time == -1ULL) {
259 return;
260 }
261
262 /* nanoseconds */
263 time = tod2time(time);
264
265 env->cputm = qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) + time;
266
267 timer_mod(env->cpu_timer, env->cputm);
268 }
269
270 /* Store System Information */
271 uint32_t HELPER(stsi)(CPUS390XState *env, uint64_t a0, uint64_t r0, uint64_t r1)
272 {
273 const uintptr_t ra = GETPC();
274 const uint32_t sel1 = r0 & STSI_R0_SEL1_MASK;
275 const uint32_t sel2 = r1 & STSI_R1_SEL2_MASK;
276 const MachineState *ms = MACHINE(qdev_get_machine());
277 uint16_t total_cpus = 0, conf_cpus = 0, reserved_cpus = 0;
278 S390CPU *cpu = env_archcpu(env);
279 SysIB sysib = { };
280 int i, cc = 0;
281
282 if ((r0 & STSI_R0_FC_MASK) > STSI_R0_FC_LEVEL_3) {
283 /* invalid function code: no other checks are performed */
284 return 3;
285 }
286
287 if ((r0 & STSI_R0_RESERVED_MASK) || (r1 & STSI_R1_RESERVED_MASK)) {
288 tcg_s390_program_interrupt(env, PGM_SPECIFICATION, ra);
289 }
290
291 if ((r0 & STSI_R0_FC_MASK) == STSI_R0_FC_CURRENT) {
292 /* query the current level: no further checks are performed */
293 env->regs[0] = STSI_R0_FC_LEVEL_3;
294 return 0;
295 }
296
297 if (a0 & ~TARGET_PAGE_MASK) {
298 tcg_s390_program_interrupt(env, PGM_SPECIFICATION, ra);
299 }
300
301 /* count the cpus and split them into configured and reserved ones */
302 for (i = 0; i < ms->possible_cpus->len; i++) {
303 total_cpus++;
304 if (ms->possible_cpus->cpus[i].cpu) {
305 conf_cpus++;
306 } else {
307 reserved_cpus++;
308 }
309 }
310
311 /*
312 * In theory, we could report Level 1 / Level 2 as current. However,
313 * the Linux kernel will detect this as running under LPAR and assume
314 * that we have a sclp linemode console (which is always present on
315 * LPAR, but not the default for QEMU), therefore not displaying boot
316 * messages and making booting a Linux kernel under TCG harder.
317 *
318 * For now we fake the same SMP configuration on all levels.
319 *
320 * TODO: We could later make the level configurable via the machine
321 * and change defaults (linemode console) based on machine type
322 * and accelerator.
323 */
324 switch (r0 & STSI_R0_FC_MASK) {
325 case STSI_R0_FC_LEVEL_1:
326 if ((sel1 == 1) && (sel2 == 1)) {
327 /* Basic Machine Configuration */
328 char type[5] = {};
329
330 ebcdic_put(sysib.sysib_111.manuf, "QEMU ", 16);
331 /* same as machine type number in STORE CPU ID, but in EBCDIC */
332 snprintf(type, ARRAY_SIZE(type), "%X", cpu->model->def->type);
333 ebcdic_put(sysib.sysib_111.type, type, 4);
334 /* model number (not stored in STORE CPU ID for z/Architecture) */
335 ebcdic_put(sysib.sysib_111.model, "QEMU ", 16);
336 ebcdic_put(sysib.sysib_111.sequence, "QEMU ", 16);
337 ebcdic_put(sysib.sysib_111.plant, "QEMU", 4);
338 } else if ((sel1 == 2) && (sel2 == 1)) {
339 /* Basic Machine CPU */
340 ebcdic_put(sysib.sysib_121.sequence, "QEMUQEMUQEMUQEMU", 16);
341 ebcdic_put(sysib.sysib_121.plant, "QEMU", 4);
342 sysib.sysib_121.cpu_addr = cpu_to_be16(env->core_id);
343 } else if ((sel1 == 2) && (sel2 == 2)) {
344 /* Basic Machine CPUs */
345 sysib.sysib_122.capability = cpu_to_be32(0x443afc29);
346 sysib.sysib_122.total_cpus = cpu_to_be16(total_cpus);
347 sysib.sysib_122.conf_cpus = cpu_to_be16(conf_cpus);
348 sysib.sysib_122.reserved_cpus = cpu_to_be16(reserved_cpus);
349 } else {
350 cc = 3;
351 }
352 break;
353 case STSI_R0_FC_LEVEL_2:
354 if ((sel1 == 2) && (sel2 == 1)) {
355 /* LPAR CPU */
356 ebcdic_put(sysib.sysib_221.sequence, "QEMUQEMUQEMUQEMU", 16);
357 ebcdic_put(sysib.sysib_221.plant, "QEMU", 4);
358 sysib.sysib_221.cpu_addr = cpu_to_be16(env->core_id);
359 } else if ((sel1 == 2) && (sel2 == 2)) {
360 /* LPAR CPUs */
361 sysib.sysib_222.lcpuc = 0x80; /* dedicated */
362 sysib.sysib_222.total_cpus = cpu_to_be16(total_cpus);
363 sysib.sysib_222.conf_cpus = cpu_to_be16(conf_cpus);
364 sysib.sysib_222.reserved_cpus = cpu_to_be16(reserved_cpus);
365 ebcdic_put(sysib.sysib_222.name, "QEMU ", 8);
366 sysib.sysib_222.caf = cpu_to_be32(1000);
367 sysib.sysib_222.dedicated_cpus = cpu_to_be16(conf_cpus);
368 } else {
369 cc = 3;
370 }
371 break;
372 case STSI_R0_FC_LEVEL_3:
373 if ((sel1 == 2) && (sel2 == 2)) {
374 /* VM CPUs */
375 sysib.sysib_322.count = 1;
376 sysib.sysib_322.vm[0].total_cpus = cpu_to_be16(total_cpus);
377 sysib.sysib_322.vm[0].conf_cpus = cpu_to_be16(conf_cpus);
378 sysib.sysib_322.vm[0].reserved_cpus = cpu_to_be16(reserved_cpus);
379 sysib.sysib_322.vm[0].caf = cpu_to_be32(1000);
380 /* Linux kernel uses this to distinguish us from z/VM */
381 ebcdic_put(sysib.sysib_322.vm[0].cpi, "KVM/Linux ", 16);
382 sysib.sysib_322.vm[0].ext_name_encoding = 2; /* UTF-8 */
383
384 /* If our VM has a name, use the real name */
385 if (qemu_name) {
386 memset(sysib.sysib_322.vm[0].name, 0x40,
387 sizeof(sysib.sysib_322.vm[0].name));
388 ebcdic_put(sysib.sysib_322.vm[0].name, qemu_name,
389 MIN(sizeof(sysib.sysib_322.vm[0].name),
390 strlen(qemu_name)));
391 strpadcpy((char *)sysib.sysib_322.ext_names[0],
392 sizeof(sysib.sysib_322.ext_names[0]),
393 qemu_name, '\0');
394
395 } else {
396 ebcdic_put(sysib.sysib_322.vm[0].name, "TCGguest", 8);
397 strcpy((char *)sysib.sysib_322.ext_names[0], "TCGguest");
398 }
399
400 /* add the uuid */
401 memcpy(sysib.sysib_322.vm[0].uuid, &qemu_uuid,
402 sizeof(sysib.sysib_322.vm[0].uuid));
403 } else {
404 cc = 3;
405 }
406 break;
407 }
408
409 if (cc == 0) {
410 if (s390_cpu_virt_mem_write(cpu, a0, 0, &sysib, sizeof(sysib))) {
411 s390_cpu_virt_mem_handle_exc(cpu, ra);
412 }
413 }
414
415 return cc;
416 }
417
418 uint32_t HELPER(sigp)(CPUS390XState *env, uint64_t order_code, uint32_t r1,
419 uint32_t r3)
420 {
421 int cc;
422
423 /* TODO: needed to inject interrupts - push further down */
424 bql_lock();
425 cc = handle_sigp(env, order_code & SIGP_ORDER_MASK, r1, r3);
426 bql_unlock();
427
428 return cc;
429 }
430 #endif
431
432 #ifndef CONFIG_USER_ONLY
433 void HELPER(xsch)(CPUS390XState *env, uint64_t r1)
434 {
435 S390CPU *cpu = env_archcpu(env);
436 bql_lock();
437 ioinst_handle_xsch(cpu, r1, GETPC());
438 bql_unlock();
439 }
440
441 void HELPER(csch)(CPUS390XState *env, uint64_t r1)
442 {
443 S390CPU *cpu = env_archcpu(env);
444 bql_lock();
445 ioinst_handle_csch(cpu, r1, GETPC());
446 bql_unlock();
447 }
448
449 void HELPER(hsch)(CPUS390XState *env, uint64_t r1)
450 {
451 S390CPU *cpu = env_archcpu(env);
452 bql_lock();
453 ioinst_handle_hsch(cpu, r1, GETPC());
454 bql_unlock();
455 }
456
457 void HELPER(msch)(CPUS390XState *env, uint64_t r1, uint64_t inst)
458 {
459 S390CPU *cpu = env_archcpu(env);
460 bql_lock();
461 ioinst_handle_msch(cpu, r1, inst >> 16, GETPC());
462 bql_unlock();
463 }
464
465 void HELPER(rchp)(CPUS390XState *env, uint64_t r1)
466 {
467 S390CPU *cpu = env_archcpu(env);
468 bql_lock();
469 ioinst_handle_rchp(cpu, r1, GETPC());
470 bql_unlock();
471 }
472
473 void HELPER(rsch)(CPUS390XState *env, uint64_t r1)
474 {
475 S390CPU *cpu = env_archcpu(env);
476 bql_lock();
477 ioinst_handle_rsch(cpu, r1, GETPC());
478 bql_unlock();
479 }
480
481 void HELPER(sal)(CPUS390XState *env, uint64_t r1)
482 {
483 S390CPU *cpu = env_archcpu(env);
484
485 bql_lock();
486 ioinst_handle_sal(cpu, r1, GETPC());
487 bql_unlock();
488 }
489
490 void HELPER(schm)(CPUS390XState *env, uint64_t r1, uint64_t r2, uint64_t inst)
491 {
492 S390CPU *cpu = env_archcpu(env);
493
494 bql_lock();
495 ioinst_handle_schm(cpu, r1, r2, inst >> 16, GETPC());
496 bql_unlock();
497 }
498
499 void HELPER(ssch)(CPUS390XState *env, uint64_t r1, uint64_t inst)
500 {
501 S390CPU *cpu = env_archcpu(env);
502 bql_lock();
503 ioinst_handle_ssch(cpu, r1, inst >> 16, GETPC());
504 bql_unlock();
505 }
506
507 void HELPER(stcrw)(CPUS390XState *env, uint64_t inst)
508 {
509 S390CPU *cpu = env_archcpu(env);
510
511 bql_lock();
512 ioinst_handle_stcrw(cpu, inst >> 16, GETPC());
513 bql_unlock();
514 }
515
516 void HELPER(stsch)(CPUS390XState *env, uint64_t r1, uint64_t inst)
517 {
518 S390CPU *cpu = env_archcpu(env);
519 bql_lock();
520 ioinst_handle_stsch(cpu, r1, inst >> 16, GETPC());
521 bql_unlock();
522 }
523
524 uint32_t HELPER(tpi)(CPUS390XState *env, uint64_t addr)
525 {
526 const uintptr_t ra = GETPC();
527 S390CPU *cpu = env_archcpu(env);
528 QEMUS390FLICState *flic = s390_get_qemu_flic(s390_get_flic());
529 QEMUS390FlicIO *io = NULL;
530 LowCore *lowcore;
531
532 if (addr & 0x3) {
533 tcg_s390_program_interrupt(env, PGM_SPECIFICATION, ra);
534 }
535
536 bql_lock();
537 io = qemu_s390_flic_dequeue_io(flic, env->cregs[6]);
538 if (!io) {
539 bql_unlock();
540 return 0;
541 }
542
543 if (addr) {
544 struct {
545 uint16_t id;
546 uint16_t nr;
547 uint32_t parm;
548 } intc = {
549 .id = cpu_to_be16(io->id),
550 .nr = cpu_to_be16(io->nr),
551 .parm = cpu_to_be32(io->parm),
552 };
553
554 if (s390_cpu_virt_mem_write(cpu, addr, 0, &intc, sizeof(intc))) {
555 /* writing failed, reinject and properly clean up */
556 s390_io_interrupt(io->id, io->nr, io->parm, io->word);
557 bql_unlock();
558 g_free(io);
559 s390_cpu_virt_mem_handle_exc(cpu, ra);
560 return 0;
561 }
562 } else {
563 /* no protection applies */
564 lowcore = cpu_map_lowcore(env);
565 lowcore->subchannel_id = cpu_to_be16(io->id);
566 lowcore->subchannel_nr = cpu_to_be16(io->nr);
567 lowcore->io_int_parm = cpu_to_be32(io->parm);
568 lowcore->io_int_word = cpu_to_be32(io->word);
569 cpu_unmap_lowcore(lowcore);
570 }
571
572 g_free(io);
573 bql_unlock();
574 return 1;
575 }
576
577 void HELPER(tsch)(CPUS390XState *env, uint64_t r1, uint64_t inst)
578 {
579 S390CPU *cpu = env_archcpu(env);
580 bql_lock();
581 ioinst_handle_tsch(cpu, r1, inst >> 16, GETPC());
582 bql_unlock();
583 }
584
585 void HELPER(chsc)(CPUS390XState *env, uint64_t inst)
586 {
587 S390CPU *cpu = env_archcpu(env);
588 bql_lock();
589 ioinst_handle_chsc(cpu, inst >> 16, GETPC());
590 bql_unlock();
591 }
592 #endif
593
594 #ifndef CONFIG_USER_ONLY
595 void HELPER(per_check_exception)(CPUS390XState *env)
596 {
597 if (env->per_perc_atmid) {
598 tcg_s390_program_interrupt(env, PGM_PER, GETPC());
599 }
600 }
601
602 /* Check if an address is within the PER starting address and the PER
603 ending address. The address range might loop. */
604 static inline bool get_per_in_range(CPUS390XState *env, uint64_t addr)
605 {
606 if (env->cregs[10] <= env->cregs[11]) {
607 return env->cregs[10] <= addr && addr <= env->cregs[11];
608 } else {
609 return env->cregs[10] <= addr || addr <= env->cregs[11];
610 }
611 }
612
613 void HELPER(per_branch)(CPUS390XState *env, uint64_t from, uint64_t to)
614 {
615 if ((env->cregs[9] & PER_CR9_EVENT_BRANCH)) {
616 if (!(env->cregs[9] & PER_CR9_CONTROL_BRANCH_ADDRESS)
617 || get_per_in_range(env, to)) {
618 env->per_address = from;
619 env->per_perc_atmid = PER_CODE_EVENT_BRANCH | get_per_atmid(env);
620 }
621 }
622 }
623
624 void HELPER(per_ifetch)(CPUS390XState *env, uint64_t addr)
625 {
626 if ((env->cregs[9] & PER_CR9_EVENT_IFETCH) && get_per_in_range(env, addr)) {
627 env->per_address = addr;
628 env->per_perc_atmid = PER_CODE_EVENT_IFETCH | get_per_atmid(env);
629
630 /* If the instruction has to be nullified, trigger the
631 exception immediately. */
632 if (env->cregs[9] & PER_CR9_EVENT_NULLIFICATION) {
633 CPUState *cs = env_cpu(env);
634
635 env->per_perc_atmid |= PER_CODE_EVENT_NULLIFICATION;
636 env->int_pgm_code = PGM_PER;
637 env->int_pgm_ilen = get_ilen(cpu_ldub_code(env, addr));
638
639 cs->exception_index = EXCP_PGM;
640 cpu_loop_exit(cs);
641 }
642 }
643 }
644
645 void HELPER(per_store_real)(CPUS390XState *env)
646 {
647 if ((env->cregs[9] & PER_CR9_EVENT_STORE) &&
648 (env->cregs[9] & PER_CR9_EVENT_STORE_REAL)) {
649 /* PSW is saved just before calling the helper. */
650 env->per_address = env->psw.addr;
651 env->per_perc_atmid = PER_CODE_EVENT_STORE_REAL | get_per_atmid(env);
652 }
653 }
654 #endif
655
656 static uint8_t stfl_bytes[2048];
657 static unsigned int used_stfl_bytes;
658
659 static void prepare_stfl(void)
660 {
661 static bool initialized;
662 int i;
663
664 /* racy, but we don't care, the same values are always written */
665 if (initialized) {
666 return;
667 }
668
669 s390_get_feat_block(S390_FEAT_TYPE_STFL, stfl_bytes);
670 for (i = 0; i < sizeof(stfl_bytes); i++) {
671 if (stfl_bytes[i]) {
672 used_stfl_bytes = i + 1;
673 }
674 }
675 initialized = true;
676 }
677
678 #ifndef CONFIG_USER_ONLY
679 void HELPER(stfl)(CPUS390XState *env)
680 {
681 LowCore *lowcore;
682
683 lowcore = cpu_map_lowcore(env);
684 prepare_stfl();
685 memcpy(&lowcore->stfl_fac_list, stfl_bytes, sizeof(lowcore->stfl_fac_list));
686 cpu_unmap_lowcore(lowcore);
687 }
688 #endif
689
690 uint32_t HELPER(stfle)(CPUS390XState *env, uint64_t addr)
691 {
692 const uintptr_t ra = GETPC();
693 const int count_bytes = ((env->regs[0] & 0xff) + 1) * 8;
694 int max_bytes;
695 int i;
696
697 if (addr & 0x7) {
698 tcg_s390_program_interrupt(env, PGM_SPECIFICATION, ra);
699 }
700
701 prepare_stfl();
702 max_bytes = ROUND_UP(used_stfl_bytes, 8);
703
704 /*
705 * The PoP says that doublewords beyond the highest-numbered facility
706 * bit may or may not be stored. However, existing hardware appears to
707 * not store the words, and existing software depend on that.
708 */
709 for (i = 0; i < MIN(count_bytes, max_bytes); ++i) {
710 cpu_stb_data_ra(env, addr + i, stfl_bytes[i], ra);
711 }
712
713 env->regs[0] = deposit64(env->regs[0], 0, 8, (max_bytes / 8) - 1);
714 return count_bytes >= max_bytes ? 0 : 3;
715 }
716
717 #ifndef CONFIG_USER_ONLY
718 /*
719 * Note: we ignore any return code of the functions called for the pci
720 * instructions, as the only time they return !0 is when the stub is
721 * called, and in that case we didn't even offer the zpci facility.
722 * The only exception is SIC, where program checks need to be handled
723 * by the caller.
724 */
725 void HELPER(clp)(CPUS390XState *env, uint32_t r2)
726 {
727 S390CPU *cpu = env_archcpu(env);
728
729 bql_lock();
730 clp_service_call(cpu, r2, GETPC());
731 bql_unlock();
732 }
733
734 void HELPER(pcilg)(CPUS390XState *env, uint32_t r1, uint32_t r2)
735 {
736 S390CPU *cpu = env_archcpu(env);
737
738 bql_lock();
739 pcilg_service_call(cpu, r1, r2, GETPC());
740 bql_unlock();
741 }
742
743 void HELPER(pcistg)(CPUS390XState *env, uint32_t r1, uint32_t r2)
744 {
745 S390CPU *cpu = env_archcpu(env);
746
747 bql_lock();
748 pcistg_service_call(cpu, r1, r2, GETPC());
749 bql_unlock();
750 }
751
752 void HELPER(stpcifc)(CPUS390XState *env, uint32_t r1, uint64_t fiba,
753 uint32_t ar)
754 {
755 S390CPU *cpu = env_archcpu(env);
756
757 bql_lock();
758 stpcifc_service_call(cpu, r1, fiba, ar, GETPC());
759 bql_unlock();
760 }
761
762 void HELPER(sic)(CPUS390XState *env, uint64_t r1, uint64_t r3)
763 {
764 S390CPU *cpu = env_archcpu(env);
765 int r;
766
767 bql_lock();
768 r = css_do_sic(cpu, (r3 >> 27) & 0x7, r1 & 0xffff);
769 bql_unlock();
770 /* css_do_sic() may actually return a PGM_xxx value to inject */
771 if (r) {
772 tcg_s390_program_interrupt(env, -r, GETPC());
773 }
774 }
775
776 void HELPER(rpcit)(CPUS390XState *env, uint32_t r1, uint32_t r2)
777 {
778 S390CPU *cpu = env_archcpu(env);
779
780 bql_lock();
781 rpcit_service_call(cpu, r1, r2, GETPC());
782 bql_unlock();
783 }
784
785 void HELPER(pcistb)(CPUS390XState *env, uint32_t r1, uint32_t r3,
786 uint64_t gaddr, uint32_t ar)
787 {
788 S390CPU *cpu = env_archcpu(env);
789
790 bql_lock();
791 pcistb_service_call(cpu, r1, r3, gaddr, ar, GETPC());
792 bql_unlock();
793 }
794
795 void HELPER(mpcifc)(CPUS390XState *env, uint32_t r1, uint64_t fiba,
796 uint32_t ar)
797 {
798 S390CPU *cpu = env_archcpu(env);
799
800 bql_lock();
801 mpcifc_service_call(cpu, r1, fiba, ar, GETPC());
802 bql_unlock();
803 }
804 #endif