]> git.proxmox.com Git - mirror_qemu.git/blob - hw/intc/ioapic.c
Include hw/qdev-properties.h less
[mirror_qemu.git] / hw / intc / ioapic.c
1 /*
2 * ioapic.c IOAPIC emulation logic
3 *
4 * Copyright (c) 2004-2005 Fabrice Bellard
5 *
6 * Split the ioapic logic from apic.c
7 * Xiantao Zhang <xiantao.zhang@intel.com>
8 *
9 * This library is free software; you can redistribute it and/or
10 * modify it under the terms of the GNU Lesser General Public
11 * License as published by the Free Software Foundation; either
12 * version 2 of the License, or (at your option) any later version.
13 *
14 * This library is distributed in the hope that it will be useful,
15 * but WITHOUT ANY WARRANTY; without even the implied warranty of
16 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
17 * Lesser General Public License for more details.
18 *
19 * You should have received a copy of the GNU Lesser General Public
20 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
21 */
22
23 #include "qemu/osdep.h"
24 #include "qapi/error.h"
25 #include "monitor/monitor.h"
26 #include "hw/i386/pc.h"
27 #include "hw/i386/apic.h"
28 #include "hw/i386/ioapic.h"
29 #include "hw/i386/ioapic_internal.h"
30 #include "hw/pci/msi.h"
31 #include "hw/qdev-properties.h"
32 #include "sysemu/kvm.h"
33 #include "hw/i386/apic-msidef.h"
34 #include "hw/i386/x86-iommu.h"
35 #include "trace.h"
36
37 #define APIC_DELIVERY_MODE_SHIFT 8
38 #define APIC_POLARITY_SHIFT 14
39 #define APIC_TRIG_MODE_SHIFT 15
40
41 static IOAPICCommonState *ioapics[MAX_IOAPICS];
42
43 /* global variable from ioapic_common.c */
44 extern int ioapic_no;
45
46 struct ioapic_entry_info {
47 /* fields parsed from IOAPIC entries */
48 uint8_t masked;
49 uint8_t trig_mode;
50 uint16_t dest_idx;
51 uint8_t dest_mode;
52 uint8_t delivery_mode;
53 uint8_t vector;
54
55 /* MSI message generated from above parsed fields */
56 uint32_t addr;
57 uint32_t data;
58 };
59
60 static void ioapic_entry_parse(uint64_t entry, struct ioapic_entry_info *info)
61 {
62 memset(info, 0, sizeof(*info));
63 info->masked = (entry >> IOAPIC_LVT_MASKED_SHIFT) & 1;
64 info->trig_mode = (entry >> IOAPIC_LVT_TRIGGER_MODE_SHIFT) & 1;
65 /*
66 * By default, this would be dest_id[8] + reserved[8]. When IR
67 * is enabled, this would be interrupt_index[15] +
68 * interrupt_format[1]. This field never means anything, but
69 * only used to generate corresponding MSI.
70 */
71 info->dest_idx = (entry >> IOAPIC_LVT_DEST_IDX_SHIFT) & 0xffff;
72 info->dest_mode = (entry >> IOAPIC_LVT_DEST_MODE_SHIFT) & 1;
73 info->delivery_mode = (entry >> IOAPIC_LVT_DELIV_MODE_SHIFT) \
74 & IOAPIC_DM_MASK;
75 if (info->delivery_mode == IOAPIC_DM_EXTINT) {
76 info->vector = pic_read_irq(isa_pic);
77 } else {
78 info->vector = entry & IOAPIC_VECTOR_MASK;
79 }
80
81 info->addr = APIC_DEFAULT_ADDRESS | \
82 (info->dest_idx << MSI_ADDR_DEST_IDX_SHIFT) | \
83 (info->dest_mode << MSI_ADDR_DEST_MODE_SHIFT);
84 info->data = (info->vector << MSI_DATA_VECTOR_SHIFT) | \
85 (info->trig_mode << MSI_DATA_TRIGGER_SHIFT) | \
86 (info->delivery_mode << MSI_DATA_DELIVERY_MODE_SHIFT);
87 }
88
89 static void ioapic_service(IOAPICCommonState *s)
90 {
91 AddressSpace *ioapic_as = PC_MACHINE(qdev_get_machine())->ioapic_as;
92 struct ioapic_entry_info info;
93 uint8_t i;
94 uint32_t mask;
95 uint64_t entry;
96
97 for (i = 0; i < IOAPIC_NUM_PINS; i++) {
98 mask = 1 << i;
99 if (s->irr & mask) {
100 int coalesce = 0;
101
102 entry = s->ioredtbl[i];
103 ioapic_entry_parse(entry, &info);
104 if (!info.masked) {
105 if (info.trig_mode == IOAPIC_TRIGGER_EDGE) {
106 s->irr &= ~mask;
107 } else {
108 coalesce = s->ioredtbl[i] & IOAPIC_LVT_REMOTE_IRR;
109 trace_ioapic_set_remote_irr(i);
110 s->ioredtbl[i] |= IOAPIC_LVT_REMOTE_IRR;
111 }
112
113 if (coalesce) {
114 /* We are level triggered interrupts, and the
115 * guest should be still working on previous one,
116 * so skip it. */
117 continue;
118 }
119
120 #ifdef CONFIG_KVM
121 if (kvm_irqchip_is_split()) {
122 if (info.trig_mode == IOAPIC_TRIGGER_EDGE) {
123 kvm_set_irq(kvm_state, i, 1);
124 kvm_set_irq(kvm_state, i, 0);
125 } else {
126 kvm_set_irq(kvm_state, i, 1);
127 }
128 continue;
129 }
130 #endif
131
132 /* No matter whether IR is enabled, we translate
133 * the IOAPIC message into a MSI one, and its
134 * address space will decide whether we need a
135 * translation. */
136 stl_le_phys(ioapic_as, info.addr, info.data);
137 }
138 }
139 }
140 }
141
142 #define SUCCESSIVE_IRQ_MAX_COUNT 10000
143
144 static void delayed_ioapic_service_cb(void *opaque)
145 {
146 IOAPICCommonState *s = opaque;
147
148 ioapic_service(s);
149 }
150
151 static void ioapic_set_irq(void *opaque, int vector, int level)
152 {
153 IOAPICCommonState *s = opaque;
154
155 /* ISA IRQs map to GSI 1-1 except for IRQ0 which maps
156 * to GSI 2. GSI maps to ioapic 1-1. This is not
157 * the cleanest way of doing it but it should work. */
158
159 trace_ioapic_set_irq(vector, level);
160 ioapic_stat_update_irq(s, vector, level);
161 if (vector == 0) {
162 vector = 2;
163 }
164 if (vector < IOAPIC_NUM_PINS) {
165 uint32_t mask = 1 << vector;
166 uint64_t entry = s->ioredtbl[vector];
167
168 if (((entry >> IOAPIC_LVT_TRIGGER_MODE_SHIFT) & 1) ==
169 IOAPIC_TRIGGER_LEVEL) {
170 /* level triggered */
171 if (level) {
172 s->irr |= mask;
173 if (!(entry & IOAPIC_LVT_REMOTE_IRR)) {
174 ioapic_service(s);
175 }
176 } else {
177 s->irr &= ~mask;
178 }
179 } else {
180 /* According to the 82093AA manual, we must ignore edge requests
181 * if the input pin is masked. */
182 if (level && !(entry & IOAPIC_LVT_MASKED)) {
183 s->irr |= mask;
184 ioapic_service(s);
185 }
186 }
187 }
188 }
189
190 static void ioapic_update_kvm_routes(IOAPICCommonState *s)
191 {
192 #ifdef CONFIG_KVM
193 int i;
194
195 if (kvm_irqchip_is_split()) {
196 for (i = 0; i < IOAPIC_NUM_PINS; i++) {
197 MSIMessage msg;
198 struct ioapic_entry_info info;
199 ioapic_entry_parse(s->ioredtbl[i], &info);
200 if (!info.masked) {
201 msg.address = info.addr;
202 msg.data = info.data;
203 kvm_irqchip_update_msi_route(kvm_state, i, msg, NULL);
204 }
205 }
206 kvm_irqchip_commit_routes(kvm_state);
207 }
208 #endif
209 }
210
211 #ifdef CONFIG_KVM
212 static void ioapic_iec_notifier(void *private, bool global,
213 uint32_t index, uint32_t mask)
214 {
215 IOAPICCommonState *s = (IOAPICCommonState *)private;
216 /* For simplicity, we just update all the routes */
217 ioapic_update_kvm_routes(s);
218 }
219 #endif
220
221 void ioapic_eoi_broadcast(int vector)
222 {
223 IOAPICCommonState *s;
224 uint64_t entry;
225 int i, n;
226
227 trace_ioapic_eoi_broadcast(vector);
228
229 for (i = 0; i < MAX_IOAPICS; i++) {
230 s = ioapics[i];
231 if (!s) {
232 continue;
233 }
234 for (n = 0; n < IOAPIC_NUM_PINS; n++) {
235 entry = s->ioredtbl[n];
236
237 if ((entry & IOAPIC_VECTOR_MASK) != vector ||
238 ((entry >> IOAPIC_LVT_TRIGGER_MODE_SHIFT) & 1) != IOAPIC_TRIGGER_LEVEL) {
239 continue;
240 }
241
242 if (!(entry & IOAPIC_LVT_REMOTE_IRR)) {
243 continue;
244 }
245
246 trace_ioapic_clear_remote_irr(n, vector);
247 s->ioredtbl[n] = entry & ~IOAPIC_LVT_REMOTE_IRR;
248
249 if (!(entry & IOAPIC_LVT_MASKED) && (s->irr & (1 << n))) {
250 ++s->irq_eoi[n];
251 if (s->irq_eoi[n] >= SUCCESSIVE_IRQ_MAX_COUNT) {
252 /*
253 * Real hardware does not deliver the interrupt immediately
254 * during eoi broadcast, and this lets a buggy guest make
255 * slow progress even if it does not correctly handle a
256 * level-triggered interrupt. Emulate this behavior if we
257 * detect an interrupt storm.
258 */
259 s->irq_eoi[n] = 0;
260 timer_mod_anticipate(s->delayed_ioapic_service_timer,
261 qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) +
262 NANOSECONDS_PER_SECOND / 100);
263 trace_ioapic_eoi_delayed_reassert(n);
264 } else {
265 ioapic_service(s);
266 }
267 } else {
268 s->irq_eoi[n] = 0;
269 }
270 }
271 }
272 }
273
274 static uint64_t
275 ioapic_mem_read(void *opaque, hwaddr addr, unsigned int size)
276 {
277 IOAPICCommonState *s = opaque;
278 int index;
279 uint32_t val = 0;
280
281 addr &= 0xff;
282
283 switch (addr) {
284 case IOAPIC_IOREGSEL:
285 val = s->ioregsel;
286 break;
287 case IOAPIC_IOWIN:
288 if (size != 4) {
289 break;
290 }
291 switch (s->ioregsel) {
292 case IOAPIC_REG_ID:
293 case IOAPIC_REG_ARB:
294 val = s->id << IOAPIC_ID_SHIFT;
295 break;
296 case IOAPIC_REG_VER:
297 val = s->version |
298 ((IOAPIC_NUM_PINS - 1) << IOAPIC_VER_ENTRIES_SHIFT);
299 break;
300 default:
301 index = (s->ioregsel - IOAPIC_REG_REDTBL_BASE) >> 1;
302 if (index >= 0 && index < IOAPIC_NUM_PINS) {
303 if (s->ioregsel & 1) {
304 val = s->ioredtbl[index] >> 32;
305 } else {
306 val = s->ioredtbl[index] & 0xffffffff;
307 }
308 }
309 }
310 break;
311 }
312
313 trace_ioapic_mem_read(addr, s->ioregsel, size, val);
314
315 return val;
316 }
317
318 /*
319 * This is to satisfy the hack in Linux kernel. One hack of it is to
320 * simulate clearing the Remote IRR bit of IOAPIC entry using the
321 * following:
322 *
323 * "For IO-APIC's with EOI register, we use that to do an explicit EOI.
324 * Otherwise, we simulate the EOI message manually by changing the trigger
325 * mode to edge and then back to level, with RTE being masked during
326 * this."
327 *
328 * (See linux kernel __eoi_ioapic_pin() comment in commit c0205701)
329 *
330 * This is based on the assumption that, Remote IRR bit will be
331 * cleared by IOAPIC hardware when configured as edge-triggered
332 * interrupts.
333 *
334 * Without this, level-triggered interrupts in IR mode might fail to
335 * work correctly.
336 */
337 static inline void
338 ioapic_fix_edge_remote_irr(uint64_t *entry)
339 {
340 if (!(*entry & IOAPIC_LVT_TRIGGER_MODE)) {
341 /* Edge-triggered interrupts, make sure remote IRR is zero */
342 *entry &= ~((uint64_t)IOAPIC_LVT_REMOTE_IRR);
343 }
344 }
345
346 static void
347 ioapic_mem_write(void *opaque, hwaddr addr, uint64_t val,
348 unsigned int size)
349 {
350 IOAPICCommonState *s = opaque;
351 int index;
352
353 addr &= 0xff;
354 trace_ioapic_mem_write(addr, s->ioregsel, size, val);
355
356 switch (addr) {
357 case IOAPIC_IOREGSEL:
358 s->ioregsel = val;
359 break;
360 case IOAPIC_IOWIN:
361 if (size != 4) {
362 break;
363 }
364 switch (s->ioregsel) {
365 case IOAPIC_REG_ID:
366 s->id = (val >> IOAPIC_ID_SHIFT) & IOAPIC_ID_MASK;
367 break;
368 case IOAPIC_REG_VER:
369 case IOAPIC_REG_ARB:
370 break;
371 default:
372 index = (s->ioregsel - IOAPIC_REG_REDTBL_BASE) >> 1;
373 if (index >= 0 && index < IOAPIC_NUM_PINS) {
374 uint64_t ro_bits = s->ioredtbl[index] & IOAPIC_RO_BITS;
375 if (s->ioregsel & 1) {
376 s->ioredtbl[index] &= 0xffffffff;
377 s->ioredtbl[index] |= (uint64_t)val << 32;
378 } else {
379 s->ioredtbl[index] &= ~0xffffffffULL;
380 s->ioredtbl[index] |= val;
381 }
382 /* restore RO bits */
383 s->ioredtbl[index] &= IOAPIC_RW_BITS;
384 s->ioredtbl[index] |= ro_bits;
385 s->irq_eoi[index] = 0;
386 ioapic_fix_edge_remote_irr(&s->ioredtbl[index]);
387 ioapic_service(s);
388 }
389 }
390 break;
391 case IOAPIC_EOI:
392 /* Explicit EOI is only supported for IOAPIC version 0x20 */
393 if (size != 4 || s->version != 0x20) {
394 break;
395 }
396 ioapic_eoi_broadcast(val);
397 break;
398 }
399
400 ioapic_update_kvm_routes(s);
401 }
402
403 static const MemoryRegionOps ioapic_io_ops = {
404 .read = ioapic_mem_read,
405 .write = ioapic_mem_write,
406 .endianness = DEVICE_NATIVE_ENDIAN,
407 };
408
409 static void ioapic_machine_done_notify(Notifier *notifier, void *data)
410 {
411 #ifdef CONFIG_KVM
412 IOAPICCommonState *s = container_of(notifier, IOAPICCommonState,
413 machine_done);
414
415 if (kvm_irqchip_is_split()) {
416 X86IOMMUState *iommu = x86_iommu_get_default();
417 if (iommu) {
418 /* Register this IOAPIC with IOMMU IEC notifier, so that
419 * when there are IR invalidates, we can be notified to
420 * update kernel IR cache. */
421 x86_iommu_iec_register_notifier(iommu, ioapic_iec_notifier, s);
422 }
423 }
424 #endif
425 }
426
427 #define IOAPIC_VER_DEF 0x20
428
429 static void ioapic_realize(DeviceState *dev, Error **errp)
430 {
431 IOAPICCommonState *s = IOAPIC_COMMON(dev);
432
433 if (s->version != 0x11 && s->version != 0x20) {
434 error_setg(errp, "IOAPIC only supports version 0x11 or 0x20 "
435 "(default: 0x%x).", IOAPIC_VER_DEF);
436 return;
437 }
438
439 memory_region_init_io(&s->io_memory, OBJECT(s), &ioapic_io_ops, s,
440 "ioapic", 0x1000);
441
442 s->delayed_ioapic_service_timer =
443 timer_new_ns(QEMU_CLOCK_VIRTUAL, delayed_ioapic_service_cb, s);
444
445 qdev_init_gpio_in(dev, ioapic_set_irq, IOAPIC_NUM_PINS);
446
447 ioapics[ioapic_no] = s;
448 s->machine_done.notify = ioapic_machine_done_notify;
449 qemu_add_machine_init_done_notifier(&s->machine_done);
450 }
451
452 static void ioapic_unrealize(DeviceState *dev, Error **errp)
453 {
454 IOAPICCommonState *s = IOAPIC_COMMON(dev);
455
456 timer_del(s->delayed_ioapic_service_timer);
457 timer_free(s->delayed_ioapic_service_timer);
458 }
459
460 static Property ioapic_properties[] = {
461 DEFINE_PROP_UINT8("version", IOAPICCommonState, version, IOAPIC_VER_DEF),
462 DEFINE_PROP_END_OF_LIST(),
463 };
464
465 static void ioapic_class_init(ObjectClass *klass, void *data)
466 {
467 IOAPICCommonClass *k = IOAPIC_COMMON_CLASS(klass);
468 DeviceClass *dc = DEVICE_CLASS(klass);
469
470 k->realize = ioapic_realize;
471 k->unrealize = ioapic_unrealize;
472 /*
473 * If APIC is in kernel, we need to update the kernel cache after
474 * migration, otherwise first 24 gsi routes will be invalid.
475 */
476 k->post_load = ioapic_update_kvm_routes;
477 dc->reset = ioapic_reset_common;
478 dc->props = ioapic_properties;
479 }
480
481 static const TypeInfo ioapic_info = {
482 .name = TYPE_IOAPIC,
483 .parent = TYPE_IOAPIC_COMMON,
484 .instance_size = sizeof(IOAPICCommonState),
485 .class_init = ioapic_class_init,
486 };
487
488 static void ioapic_register_types(void)
489 {
490 type_register_static(&ioapic_info);
491 }
492
493 type_init(ioapic_register_types)