]>
Commit | Line | Data |
---|---|---|
1fd4f2a5 ED |
1 | /* |
2 | * Copyright (C) 2001 MandrakeSoft S.A. | |
221d059d | 3 | * Copyright 2010 Red Hat, Inc. and/or its affiliates. |
1fd4f2a5 ED |
4 | * |
5 | * MandrakeSoft S.A. | |
6 | * 43, rue d'Aboukir | |
7 | * 75002 Paris - France | |
8 | * http://www.linux-mandrake.com/ | |
9 | * http://www.mandrakesoft.com/ | |
10 | * | |
11 | * This library is free software; you can redistribute it and/or | |
12 | * modify it under the terms of the GNU Lesser General Public | |
13 | * License as published by the Free Software Foundation; either | |
14 | * version 2 of the License, or (at your option) any later version. | |
15 | * | |
16 | * This library is distributed in the hope that it will be useful, | |
17 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | |
18 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU | |
19 | * Lesser General Public License for more details. | |
20 | * | |
21 | * You should have received a copy of the GNU Lesser General Public | |
22 | * License along with this library; if not, write to the Free Software | |
23 | * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA | |
24 | * | |
25 | * Yunhong Jiang <yunhong.jiang@intel.com> | |
26 | * Yaozu (Eddie) Dong <eddie.dong@intel.com> | |
27 | * Based on Xen 3.1 code. | |
28 | */ | |
29 | ||
edf88417 | 30 | #include <linux/kvm_host.h> |
1fd4f2a5 ED |
31 | #include <linux/kvm.h> |
32 | #include <linux/mm.h> | |
33 | #include <linux/highmem.h> | |
34 | #include <linux/smp.h> | |
35 | #include <linux/hrtimer.h> | |
36 | #include <linux/io.h> | |
5a0e3ad6 | 37 | #include <linux/slab.h> |
c7c9c56c | 38 | #include <linux/export.h> |
8c86405f | 39 | #include <linux/nospec.h> |
1fd4f2a5 | 40 | #include <asm/processor.h> |
1fd4f2a5 ED |
41 | #include <asm/page.h> |
42 | #include <asm/current.h> | |
1000ff8d | 43 | #include <trace/events/kvm.h> |
82470196 ZX |
44 | |
45 | #include "ioapic.h" | |
46 | #include "lapic.h" | |
f5244726 | 47 | #include "irq.h" |
82470196 | 48 | |
0b10a1c8 | 49 | static int ioapic_service(struct kvm_ioapic *vioapic, int irq, |
aa2fbe6d | 50 | bool line_status); |
1fd4f2a5 | 51 | |
f458d039 SS |
52 | static void kvm_ioapic_update_eoi_one(struct kvm_vcpu *vcpu, |
53 | struct kvm_ioapic *ioapic, | |
54 | int trigger_mode, | |
55 | int pin); | |
56 | ||
1fd4f2a5 ED |
57 | static unsigned long ioapic_read_indirect(struct kvm_ioapic *ioapic, |
58 | unsigned long addr, | |
59 | unsigned long length) | |
60 | { | |
61 | unsigned long result = 0; | |
62 | ||
63 | switch (ioapic->ioregsel) { | |
64 | case IOAPIC_REG_VERSION: | |
65 | result = ((((IOAPIC_NUM_PINS - 1) & 0xff) << 16) | |
66 | | (IOAPIC_VERSION_ID & 0xff)); | |
67 | break; | |
68 | ||
69 | case IOAPIC_REG_APIC_ID: | |
70 | case IOAPIC_REG_ARB_ID: | |
71 | result = ((ioapic->id & 0xf) << 24); | |
72 | break; | |
73 | ||
74 | default: | |
75 | { | |
76 | u32 redir_index = (ioapic->ioregsel - 0x10) >> 1; | |
8c86405f | 77 | u64 redir_content = ~0ULL; |
1fd4f2a5 | 78 | |
8c86405f MP |
79 | if (redir_index < IOAPIC_NUM_PINS) { |
80 | u32 index = array_index_nospec( | |
81 | redir_index, IOAPIC_NUM_PINS); | |
82 | ||
83 | redir_content = ioapic->redirtbl[index].bits; | |
84 | } | |
1fd4f2a5 | 85 | |
1fd4f2a5 ED |
86 | result = (ioapic->ioregsel & 0x1) ? |
87 | (redir_content >> 32) & 0xffffffff : | |
88 | redir_content & 0xffffffff; | |
89 | break; | |
90 | } | |
91 | } | |
92 | ||
93 | return result; | |
94 | } | |
95 | ||
10606919 YZ |
96 | static void rtc_irq_eoi_tracking_reset(struct kvm_ioapic *ioapic) |
97 | { | |
98 | ioapic->rtc_status.pending_eoi = 0; | |
f9de4db3 | 99 | bitmap_zero(ioapic->rtc_status.dest_map.map, KVM_MAX_VCPU_ID); |
10606919 YZ |
100 | } |
101 | ||
4009b249 PB |
102 | static void kvm_rtc_eoi_tracking_restore_all(struct kvm_ioapic *ioapic); |
103 | ||
104 | static void rtc_status_pending_eoi_check_valid(struct kvm_ioapic *ioapic) | |
105 | { | |
106 | if (WARN_ON(ioapic->rtc_status.pending_eoi < 0)) | |
107 | kvm_rtc_eoi_tracking_restore_all(ioapic); | |
108 | } | |
109 | ||
10606919 YZ |
110 | static void __rtc_irq_eoi_tracking_restore_one(struct kvm_vcpu *vcpu) |
111 | { | |
112 | bool new_val, old_val; | |
113 | struct kvm_ioapic *ioapic = vcpu->kvm->arch.vioapic; | |
b0eaf450 | 114 | struct dest_map *dest_map = &ioapic->rtc_status.dest_map; |
10606919 YZ |
115 | union kvm_ioapic_redirect_entry *e; |
116 | ||
117 | e = &ioapic->redirtbl[RTC_GSI]; | |
5c69d5c1 PX |
118 | if (!kvm_apic_match_dest(vcpu, NULL, APIC_DEST_NOSHORT, |
119 | e->fields.dest_id, | |
120 | kvm_lapic_irq_dest_mode(!!e->fields.dest_mode))) | |
10606919 YZ |
121 | return; |
122 | ||
123 | new_val = kvm_apic_pending_eoi(vcpu, e->fields.vector); | |
b0eaf450 | 124 | old_val = test_bit(vcpu->vcpu_id, dest_map->map); |
10606919 YZ |
125 | |
126 | if (new_val == old_val) | |
127 | return; | |
128 | ||
129 | if (new_val) { | |
b0eaf450 PB |
130 | __set_bit(vcpu->vcpu_id, dest_map->map); |
131 | dest_map->vectors[vcpu->vcpu_id] = e->fields.vector; | |
10606919 YZ |
132 | ioapic->rtc_status.pending_eoi++; |
133 | } else { | |
b0eaf450 | 134 | __clear_bit(vcpu->vcpu_id, dest_map->map); |
10606919 | 135 | ioapic->rtc_status.pending_eoi--; |
4009b249 | 136 | rtc_status_pending_eoi_check_valid(ioapic); |
10606919 | 137 | } |
10606919 YZ |
138 | } |
139 | ||
140 | void kvm_rtc_eoi_tracking_restore_one(struct kvm_vcpu *vcpu) | |
141 | { | |
142 | struct kvm_ioapic *ioapic = vcpu->kvm->arch.vioapic; | |
143 | ||
144 | spin_lock(&ioapic->lock); | |
145 | __rtc_irq_eoi_tracking_restore_one(vcpu); | |
146 | spin_unlock(&ioapic->lock); | |
147 | } | |
148 | ||
149 | static void kvm_rtc_eoi_tracking_restore_all(struct kvm_ioapic *ioapic) | |
150 | { | |
151 | struct kvm_vcpu *vcpu; | |
152 | int i; | |
153 | ||
154 | if (RTC_GSI >= IOAPIC_NUM_PINS) | |
155 | return; | |
156 | ||
157 | rtc_irq_eoi_tracking_reset(ioapic); | |
158 | kvm_for_each_vcpu(i, vcpu, ioapic->kvm) | |
159 | __rtc_irq_eoi_tracking_restore_one(vcpu); | |
160 | } | |
161 | ||
1ec2405c SS |
162 | static void rtc_irq_eoi(struct kvm_ioapic *ioapic, struct kvm_vcpu *vcpu, |
163 | int vector) | |
2c2bf011 | 164 | { |
1ec2405c SS |
165 | struct dest_map *dest_map = &ioapic->rtc_status.dest_map; |
166 | ||
167 | /* RTC special handling */ | |
168 | if (test_bit(vcpu->vcpu_id, dest_map->map) && | |
169 | (vector == dest_map->vectors[vcpu->vcpu_id]) && | |
170 | (test_and_clear_bit(vcpu->vcpu_id, | |
171 | ioapic->rtc_status.dest_map.map))) { | |
2c2bf011 | 172 | --ioapic->rtc_status.pending_eoi; |
4009b249 PB |
173 | rtc_status_pending_eoi_check_valid(ioapic); |
174 | } | |
2c2bf011 YZ |
175 | } |
176 | ||
177 | static bool rtc_irq_check_coalesced(struct kvm_ioapic *ioapic) | |
178 | { | |
179 | if (ioapic->rtc_status.pending_eoi > 0) | |
180 | return true; /* coalesced */ | |
181 | ||
182 | return false; | |
183 | } | |
184 | ||
f458d039 SS |
185 | static void ioapic_lazy_update_eoi(struct kvm_ioapic *ioapic, int irq) |
186 | { | |
187 | int i; | |
188 | struct kvm_vcpu *vcpu; | |
189 | union kvm_ioapic_redirect_entry *entry = &ioapic->redirtbl[irq]; | |
190 | ||
191 | kvm_for_each_vcpu(i, vcpu, ioapic->kvm) { | |
192 | if (!kvm_apic_match_dest(vcpu, NULL, APIC_DEST_NOSHORT, | |
193 | entry->fields.dest_id, | |
194 | entry->fields.dest_mode) || | |
195 | kvm_apic_pending_eoi(vcpu, entry->fields.vector)) | |
196 | continue; | |
197 | ||
198 | /* | |
199 | * If no longer has pending EOI in LAPICs, update | |
77377064 | 200 | * EOI for this vector. |
f458d039 SS |
201 | */ |
202 | rtc_irq_eoi(ioapic, vcpu, entry->fields.vector); | |
f458d039 SS |
203 | break; |
204 | } | |
205 | } | |
206 | ||
44847dea PB |
207 | static int ioapic_set_irq(struct kvm_ioapic *ioapic, unsigned int irq, |
208 | int irq_level, bool line_status) | |
209 | { | |
210 | union kvm_ioapic_redirect_entry entry; | |
211 | u32 mask = 1 << irq; | |
212 | u32 old_irr; | |
213 | int edge, ret; | |
214 | ||
215 | entry = ioapic->redirtbl[irq]; | |
216 | edge = (entry.fields.trig_mode == IOAPIC_EDGE_TRIG); | |
217 | ||
218 | if (!irq_level) { | |
219 | ioapic->irr &= ~mask; | |
220 | ret = 1; | |
221 | goto out; | |
222 | } | |
223 | ||
f458d039 | 224 | /* |
8be8f932 PB |
225 | * AMD SVM AVIC accelerate EOI write iff the interrupt is edge |
226 | * triggered, in which case the in-kernel IOAPIC will not be able | |
227 | * to receive the EOI. In this case, we do a lazy update of the | |
228 | * pending EOI when trying to set IOAPIC irq. | |
f458d039 | 229 | */ |
8be8f932 | 230 | if (edge && kvm_apicv_activated(ioapic->kvm)) |
f458d039 SS |
231 | ioapic_lazy_update_eoi(ioapic, irq); |
232 | ||
44847dea PB |
233 | /* |
234 | * Return 0 for coalesced interrupts; for edge-triggered interrupts, | |
235 | * this only happens if a previous edge has not been delivered due | |
00116795 | 236 | * to masking. For level interrupts, the remote_irr field tells |
44847dea PB |
237 | * us if the interrupt is waiting for an EOI. |
238 | * | |
239 | * RTC is special: it is edge-triggered, but userspace likes to know | |
240 | * if it has been already ack-ed via EOI because coalesced RTC | |
241 | * interrupts lead to time drift in Windows guests. So we track | |
242 | * EOI manually for the RTC interrupt. | |
243 | */ | |
244 | if (irq == RTC_GSI && line_status && | |
245 | rtc_irq_check_coalesced(ioapic)) { | |
246 | ret = 0; | |
247 | goto out; | |
248 | } | |
249 | ||
250 | old_irr = ioapic->irr; | |
251 | ioapic->irr |= mask; | |
7d225368 | 252 | if (edge) { |
5bda6eed | 253 | ioapic->irr_delivered &= ~mask; |
7d225368 NL |
254 | if (old_irr == ioapic->irr) { |
255 | ret = 0; | |
256 | goto out; | |
257 | } | |
44847dea PB |
258 | } |
259 | ||
260 | ret = ioapic_service(ioapic, irq, line_status); | |
261 | ||
262 | out: | |
263 | trace_kvm_ioapic_set_irq(entry.bits, irq, ret == 0); | |
264 | return ret; | |
265 | } | |
266 | ||
673f7b42 PB |
267 | static void kvm_ioapic_inject_all(struct kvm_ioapic *ioapic, unsigned long irr) |
268 | { | |
269 | u32 idx; | |
270 | ||
271 | rtc_irq_eoi_tracking_reset(ioapic); | |
272 | for_each_set_bit(idx, &irr, IOAPIC_NUM_PINS) | |
273 | ioapic_set_irq(ioapic, idx, 1, true); | |
274 | ||
275 | kvm_rtc_eoi_tracking_restore_all(ioapic); | |
276 | } | |
277 | ||
278 | ||
6308630b | 279 | void kvm_ioapic_scan_entry(struct kvm_vcpu *vcpu, ulong *ioapic_handled_vectors) |
c7c9c56c YZ |
280 | { |
281 | struct kvm_ioapic *ioapic = vcpu->kvm->arch.vioapic; | |
4d99ba89 | 282 | struct dest_map *dest_map = &ioapic->rtc_status.dest_map; |
c7c9c56c | 283 | union kvm_ioapic_redirect_entry *e; |
c7c9c56c YZ |
284 | int index; |
285 | ||
286 | spin_lock(&ioapic->lock); | |
4d99ba89 JR |
287 | |
288 | /* Make sure we see any missing RTC EOI */ | |
289 | if (test_bit(vcpu->vcpu_id, dest_map->map)) | |
290 | __set_bit(dest_map->vectors[vcpu->vcpu_id], | |
291 | ioapic_handled_vectors); | |
292 | ||
c7c9c56c YZ |
293 | for (index = 0; index < IOAPIC_NUM_PINS; index++) { |
294 | e = &ioapic->redirtbl[index]; | |
0f6c0a74 PB |
295 | if (e->fields.trig_mode == IOAPIC_LEVEL_TRIG || |
296 | kvm_irq_has_notifier(ioapic->kvm, KVM_IRQCHIP_IOAPIC, index) || | |
297 | index == RTC_GSI) { | |
5c69d5c1 PX |
298 | u16 dm = kvm_lapic_irq_dest_mode(!!e->fields.dest_mode); |
299 | ||
300 | if (kvm_apic_match_dest(vcpu, NULL, APIC_DEST_NOSHORT, | |
301 | e->fields.dest_id, dm) || | |
0fc5a36d | 302 | kvm_apic_pending_eoi(vcpu, e->fields.vector)) |
cf9e65b7 | 303 | __set_bit(e->fields.vector, |
6308630b | 304 | ioapic_handled_vectors); |
c7c9c56c YZ |
305 | } |
306 | } | |
307 | spin_unlock(&ioapic->lock); | |
308 | } | |
c7c9c56c | 309 | |
993225ad | 310 | void kvm_arch_post_irq_ack_notifier_list_update(struct kvm *kvm) |
c7c9c56c | 311 | { |
0bceb15a | 312 | if (!ioapic_in_kernel(kvm)) |
c7c9c56c | 313 | return; |
3d81bc7e | 314 | kvm_make_scan_ioapic_request(kvm); |
c7c9c56c YZ |
315 | } |
316 | ||
1fd4f2a5 ED |
317 | static void ioapic_write_indirect(struct kvm_ioapic *ioapic, u32 val) |
318 | { | |
319 | unsigned index; | |
75858a84 | 320 | bool mask_before, mask_after; |
70f93dae | 321 | union kvm_ioapic_redirect_entry *e; |
7ee30bc1 | 322 | int old_remote_irr, old_delivery_status, old_dest_id, old_dest_mode; |
2f9b68f5 | 323 | DECLARE_BITMAP(vcpu_bitmap, KVM_MAX_VCPUS); |
1fd4f2a5 ED |
324 | |
325 | switch (ioapic->ioregsel) { | |
326 | case IOAPIC_REG_VERSION: | |
327 | /* Writes are ignored. */ | |
328 | break; | |
329 | ||
330 | case IOAPIC_REG_APIC_ID: | |
331 | ioapic->id = (val >> 24) & 0xf; | |
332 | break; | |
333 | ||
334 | case IOAPIC_REG_ARB_ID: | |
335 | break; | |
336 | ||
337 | default: | |
338 | index = (ioapic->ioregsel - 0x10) >> 1; | |
339 | ||
1fd4f2a5 ED |
340 | if (index >= IOAPIC_NUM_PINS) |
341 | return; | |
67056455 | 342 | index = array_index_nospec(index, IOAPIC_NUM_PINS); |
70f93dae GN |
343 | e = &ioapic->redirtbl[index]; |
344 | mask_before = e->fields.mask; | |
b200dded NL |
345 | /* Preserve read-only fields */ |
346 | old_remote_irr = e->fields.remote_irr; | |
347 | old_delivery_status = e->fields.delivery_status; | |
7ee30bc1 NNL |
348 | old_dest_id = e->fields.dest_id; |
349 | old_dest_mode = e->fields.dest_mode; | |
1fd4f2a5 | 350 | if (ioapic->ioregsel & 1) { |
70f93dae GN |
351 | e->bits &= 0xffffffff; |
352 | e->bits |= (u64) val << 32; | |
1fd4f2a5 | 353 | } else { |
70f93dae GN |
354 | e->bits &= ~0xffffffffULL; |
355 | e->bits |= (u32) val; | |
1fd4f2a5 | 356 | } |
b200dded NL |
357 | e->fields.remote_irr = old_remote_irr; |
358 | e->fields.delivery_status = old_delivery_status; | |
a8bfec29 NL |
359 | |
360 | /* | |
361 | * Some OSes (Linux, Xen) assume that Remote IRR bit will | |
362 | * be cleared by IOAPIC hardware when the entry is configured | |
363 | * as edge-triggered. This behavior is used to simulate an | |
364 | * explicit EOI on IOAPICs that don't have the EOI register. | |
365 | */ | |
366 | if (e->fields.trig_mode == IOAPIC_EDGE_TRIG) | |
367 | e->fields.remote_irr = 0; | |
368 | ||
70f93dae | 369 | mask_after = e->fields.mask; |
75858a84 | 370 | if (mask_before != mask_after) |
4a994358 | 371 | kvm_fire_mask_notifiers(ioapic->kvm, KVM_IRQCHIP_IOAPIC, index, mask_after); |
70f93dae | 372 | if (e->fields.trig_mode == IOAPIC_LEVEL_TRIG |
b4a2f5e7 | 373 | && ioapic->irr & (1 << index)) |
aa2fbe6d | 374 | ioapic_service(ioapic, index, false); |
7ee30bc1 NNL |
375 | if (e->fields.delivery_mode == APIC_DM_FIXED) { |
376 | struct kvm_lapic_irq irq; | |
377 | ||
7ee30bc1 NNL |
378 | irq.vector = e->fields.vector; |
379 | irq.delivery_mode = e->fields.delivery_mode << 8; | |
c96001c5 PX |
380 | irq.dest_mode = |
381 | kvm_lapic_irq_dest_mode(!!e->fields.dest_mode); | |
0c22056f NNL |
382 | irq.level = false; |
383 | irq.trig_mode = e->fields.trig_mode; | |
384 | irq.shorthand = APIC_DEST_NOSHORT; | |
385 | irq.dest_id = e->fields.dest_id; | |
386 | irq.msi_redir_hint = false; | |
2f9b68f5 | 387 | bitmap_zero(vcpu_bitmap, KVM_MAX_VCPUS); |
7ee30bc1 | 388 | kvm_bitmap_or_dest_vcpus(ioapic->kvm, &irq, |
2f9b68f5 | 389 | vcpu_bitmap); |
7ee30bc1 NNL |
390 | if (old_dest_mode != e->fields.dest_mode || |
391 | old_dest_id != e->fields.dest_id) { | |
392 | /* | |
393 | * Update vcpu_bitmap with vcpus specified in | |
394 | * the previous request as well. This is done to | |
395 | * keep ioapic_handled_vectors synchronized. | |
396 | */ | |
397 | irq.dest_id = old_dest_id; | |
c96001c5 PX |
398 | irq.dest_mode = |
399 | kvm_lapic_irq_dest_mode( | |
400 | !!e->fields.dest_mode); | |
7ee30bc1 | 401 | kvm_bitmap_or_dest_vcpus(ioapic->kvm, &irq, |
2f9b68f5 | 402 | vcpu_bitmap); |
7ee30bc1 NNL |
403 | } |
404 | kvm_make_scan_ioapic_request_mask(ioapic->kvm, | |
2f9b68f5 | 405 | vcpu_bitmap); |
7ee30bc1 NNL |
406 | } else { |
407 | kvm_make_scan_ioapic_request(ioapic->kvm); | |
408 | } | |
1fd4f2a5 ED |
409 | break; |
410 | } | |
411 | } | |
412 | ||
0b10a1c8 | 413 | static int ioapic_service(struct kvm_ioapic *ioapic, int irq, bool line_status) |
a53c17d2 | 414 | { |
58c2dde1 GN |
415 | union kvm_ioapic_redirect_entry *entry = &ioapic->redirtbl[irq]; |
416 | struct kvm_lapic_irq irqe; | |
2c2bf011 | 417 | int ret; |
a53c17d2 | 418 | |
da3fe7bd NL |
419 | if (entry->fields.mask || |
420 | (entry->fields.trig_mode == IOAPIC_LEVEL_TRIG && | |
421 | entry->fields.remote_irr)) | |
0b10a1c8 PB |
422 | return -1; |
423 | ||
58c2dde1 GN |
424 | irqe.dest_id = entry->fields.dest_id; |
425 | irqe.vector = entry->fields.vector; | |
c96001c5 | 426 | irqe.dest_mode = kvm_lapic_irq_dest_mode(!!entry->fields.dest_mode); |
58c2dde1 GN |
427 | irqe.trig_mode = entry->fields.trig_mode; |
428 | irqe.delivery_mode = entry->fields.delivery_mode << 8; | |
429 | irqe.level = 1; | |
150a84fe | 430 | irqe.shorthand = APIC_DEST_NOSHORT; |
93bbf0b8 | 431 | irqe.msi_redir_hint = false; |
a53c17d2 | 432 | |
0bc830b0 | 433 | if (irqe.trig_mode == IOAPIC_EDGE_TRIG) |
5bda6eed | 434 | ioapic->irr_delivered |= 1 << irq; |
0bc830b0 | 435 | |
2c2bf011 | 436 | if (irq == RTC_GSI && line_status) { |
4009b249 PB |
437 | /* |
438 | * pending_eoi cannot ever become negative (see | |
439 | * rtc_status_pending_eoi_check_valid) and the caller | |
440 | * ensures that it is only called if it is >= zero, namely | |
441 | * if rtc_irq_check_coalesced returns false). | |
442 | */ | |
2c2bf011 YZ |
443 | BUG_ON(ioapic->rtc_status.pending_eoi != 0); |
444 | ret = kvm_irq_delivery_to_apic(ioapic->kvm, NULL, &irqe, | |
9e4aabe2 | 445 | &ioapic->rtc_status.dest_map); |
5678de3f | 446 | ioapic->rtc_status.pending_eoi = (ret < 0 ? 0 : ret); |
2c2bf011 YZ |
447 | } else |
448 | ret = kvm_irq_delivery_to_apic(ioapic->kvm, NULL, &irqe, NULL); | |
449 | ||
0b10a1c8 PB |
450 | if (ret && irqe.trig_mode == IOAPIC_LEVEL_TRIG) |
451 | entry->fields.remote_irr = 1; | |
452 | ||
2c2bf011 | 453 | return ret; |
a53c17d2 GN |
454 | } |
455 | ||
1a577b72 | 456 | int kvm_ioapic_set_irq(struct kvm_ioapic *ioapic, int irq, int irq_source_id, |
aa2fbe6d | 457 | int level, bool line_status) |
1fd4f2a5 | 458 | { |
28a6fdab MT |
459 | int ret, irq_level; |
460 | ||
461 | BUG_ON(irq < 0 || irq >= IOAPIC_NUM_PINS); | |
1fd4f2a5 | 462 | |
46a47b1e | 463 | spin_lock(&ioapic->lock); |
28a6fdab MT |
464 | irq_level = __kvm_irq_line_state(&ioapic->irq_states[irq], |
465 | irq_source_id, level); | |
44847dea | 466 | ret = ioapic_set_irq(ioapic, irq, irq_level, line_status); |
2c2bf011 | 467 | |
46a47b1e | 468 | spin_unlock(&ioapic->lock); |
eba0226b | 469 | |
4925663a | 470 | return ret; |
1fd4f2a5 ED |
471 | } |
472 | ||
1a577b72 MT |
473 | void kvm_ioapic_clear_all(struct kvm_ioapic *ioapic, int irq_source_id) |
474 | { | |
475 | int i; | |
476 | ||
477 | spin_lock(&ioapic->lock); | |
478 | for (i = 0; i < KVM_IOAPIC_NUM_PINS; i++) | |
479 | __clear_bit(irq_source_id, &ioapic->irq_states[i]); | |
480 | spin_unlock(&ioapic->lock); | |
481 | } | |
482 | ||
184564ef ZH |
483 | static void kvm_ioapic_eoi_inject_work(struct work_struct *work) |
484 | { | |
485 | int i; | |
486 | struct kvm_ioapic *ioapic = container_of(work, struct kvm_ioapic, | |
487 | eoi_inject.work); | |
488 | spin_lock(&ioapic->lock); | |
489 | for (i = 0; i < IOAPIC_NUM_PINS; i++) { | |
490 | union kvm_ioapic_redirect_entry *ent = &ioapic->redirtbl[i]; | |
491 | ||
492 | if (ent->fields.trig_mode != IOAPIC_LEVEL_TRIG) | |
493 | continue; | |
494 | ||
495 | if (ioapic->irr & (1 << i) && !ent->fields.remote_irr) | |
496 | ioapic_service(ioapic, i, false); | |
497 | } | |
498 | spin_unlock(&ioapic->lock); | |
499 | } | |
500 | ||
501 | #define IOAPIC_SUCCESSIVE_IRQ_MAX_COUNT 10000 | |
1ec2405c SS |
502 | static void kvm_ioapic_update_eoi_one(struct kvm_vcpu *vcpu, |
503 | struct kvm_ioapic *ioapic, | |
504 | int trigger_mode, | |
505 | int pin) | |
1fd4f2a5 | 506 | { |
c806a6ad | 507 | struct kvm_lapic *apic = vcpu->arch.apic; |
1ec2405c | 508 | union kvm_ioapic_redirect_entry *ent = &ioapic->redirtbl[pin]; |
eba0226b | 509 | |
1ec2405c SS |
510 | /* |
511 | * We are dropping lock while calling ack notifiers because ack | |
512 | * notifier callbacks for assigned devices call into IOAPIC | |
513 | * recursively. Since remote_irr is cleared only after call | |
514 | * to notifiers if the same vector will be delivered while lock | |
515 | * is dropped it will be put into irr and will be delivered | |
516 | * after ack notifier returns. | |
517 | */ | |
518 | spin_unlock(&ioapic->lock); | |
519 | kvm_notify_acked_irq(ioapic->kvm, KVM_IRQCHIP_IOAPIC, pin); | |
520 | spin_lock(&ioapic->lock); | |
eba0226b | 521 | |
1ec2405c SS |
522 | if (trigger_mode != IOAPIC_LEVEL_TRIG || |
523 | kvm_lapic_get_reg(apic, APIC_SPIV) & APIC_SPIV_DIRECTED_EOI) | |
524 | return; | |
f5244726 | 525 | |
1ec2405c SS |
526 | ASSERT(ent->fields.trig_mode == IOAPIC_LEVEL_TRIG); |
527 | ent->fields.remote_irr = 0; | |
528 | if (!ent->fields.mask && (ioapic->irr & (1 << pin))) { | |
529 | ++ioapic->irq_eoi[pin]; | |
530 | if (ioapic->irq_eoi[pin] == IOAPIC_SUCCESSIVE_IRQ_MAX_COUNT) { | |
531 | /* | |
532 | * Real hardware does not deliver the interrupt | |
533 | * immediately during eoi broadcast, and this | |
534 | * lets a buggy guest make slow progress | |
535 | * even if it does not correctly handle a | |
536 | * level-triggered interrupt. Emulate this | |
537 | * behavior if we detect an interrupt storm. | |
538 | */ | |
539 | schedule_delayed_work(&ioapic->eoi_inject, HZ / 100); | |
540 | ioapic->irq_eoi[pin] = 0; | |
541 | trace_kvm_ioapic_delayed_eoi_inj(ent->bits); | |
184564ef | 542 | } else { |
1ec2405c | 543 | ioapic_service(ioapic, pin, false); |
184564ef | 544 | } |
1ec2405c SS |
545 | } else { |
546 | ioapic->irq_eoi[pin] = 0; | |
f5244726 | 547 | } |
1fd4f2a5 ED |
548 | } |
549 | ||
1fcc7890 | 550 | void kvm_ioapic_update_eoi(struct kvm_vcpu *vcpu, int vector, int trigger_mode) |
4fa6b9c5 | 551 | { |
1ec2405c | 552 | int i; |
1fcc7890 | 553 | struct kvm_ioapic *ioapic = vcpu->kvm->arch.vioapic; |
4fa6b9c5 | 554 | |
46a47b1e | 555 | spin_lock(&ioapic->lock); |
1ec2405c SS |
556 | rtc_irq_eoi(ioapic, vcpu, vector); |
557 | for (i = 0; i < IOAPIC_NUM_PINS; i++) { | |
558 | union kvm_ioapic_redirect_entry *ent = &ioapic->redirtbl[i]; | |
559 | ||
560 | if (ent->fields.vector != vector) | |
561 | continue; | |
562 | kvm_ioapic_update_eoi_one(vcpu, ioapic, trigger_mode, i); | |
563 | } | |
46a47b1e | 564 | spin_unlock(&ioapic->lock); |
4fa6b9c5 AK |
565 | } |
566 | ||
d76685c4 GH |
567 | static inline struct kvm_ioapic *to_ioapic(struct kvm_io_device *dev) |
568 | { | |
569 | return container_of(dev, struct kvm_ioapic, dev); | |
570 | } | |
571 | ||
bda9020e | 572 | static inline int ioapic_in_range(struct kvm_ioapic *ioapic, gpa_t addr) |
1fd4f2a5 | 573 | { |
1fd4f2a5 ED |
574 | return ((addr >= ioapic->base_address && |
575 | (addr < ioapic->base_address + IOAPIC_MEM_LENGTH))); | |
576 | } | |
577 | ||
e32edf4f NN |
578 | static int ioapic_mmio_read(struct kvm_vcpu *vcpu, struct kvm_io_device *this, |
579 | gpa_t addr, int len, void *val) | |
1fd4f2a5 | 580 | { |
d76685c4 | 581 | struct kvm_ioapic *ioapic = to_ioapic(this); |
1fd4f2a5 | 582 | u32 result; |
bda9020e MT |
583 | if (!ioapic_in_range(ioapic, addr)) |
584 | return -EOPNOTSUPP; | |
1fd4f2a5 | 585 | |
1fd4f2a5 ED |
586 | ASSERT(!(addr & 0xf)); /* check alignment */ |
587 | ||
588 | addr &= 0xff; | |
46a47b1e | 589 | spin_lock(&ioapic->lock); |
1fd4f2a5 ED |
590 | switch (addr) { |
591 | case IOAPIC_REG_SELECT: | |
592 | result = ioapic->ioregsel; | |
593 | break; | |
594 | ||
595 | case IOAPIC_REG_WINDOW: | |
596 | result = ioapic_read_indirect(ioapic, addr, len); | |
597 | break; | |
598 | ||
599 | default: | |
600 | result = 0; | |
601 | break; | |
602 | } | |
46a47b1e | 603 | spin_unlock(&ioapic->lock); |
eba0226b | 604 | |
1fd4f2a5 ED |
605 | switch (len) { |
606 | case 8: | |
607 | *(u64 *) val = result; | |
608 | break; | |
609 | case 1: | |
610 | case 2: | |
611 | case 4: | |
612 | memcpy(val, (char *)&result, len); | |
613 | break; | |
614 | default: | |
615 | printk(KERN_WARNING "ioapic: wrong length %d\n", len); | |
616 | } | |
bda9020e | 617 | return 0; |
1fd4f2a5 ED |
618 | } |
619 | ||
e32edf4f NN |
620 | static int ioapic_mmio_write(struct kvm_vcpu *vcpu, struct kvm_io_device *this, |
621 | gpa_t addr, int len, const void *val) | |
1fd4f2a5 | 622 | { |
d76685c4 | 623 | struct kvm_ioapic *ioapic = to_ioapic(this); |
1fd4f2a5 | 624 | u32 data; |
bda9020e MT |
625 | if (!ioapic_in_range(ioapic, addr)) |
626 | return -EOPNOTSUPP; | |
1fd4f2a5 | 627 | |
1fd4f2a5 | 628 | ASSERT(!(addr & 0xf)); /* check alignment */ |
60eead79 | 629 | |
d77fe635 JS |
630 | switch (len) { |
631 | case 8: | |
632 | case 4: | |
1fd4f2a5 | 633 | data = *(u32 *) val; |
d77fe635 JS |
634 | break; |
635 | case 2: | |
636 | data = *(u16 *) val; | |
637 | break; | |
638 | case 1: | |
639 | data = *(u8 *) val; | |
640 | break; | |
641 | default: | |
1fd4f2a5 | 642 | printk(KERN_WARNING "ioapic: Unsupported size %d\n", len); |
eba0226b | 643 | return 0; |
1fd4f2a5 ED |
644 | } |
645 | ||
646 | addr &= 0xff; | |
46a47b1e | 647 | spin_lock(&ioapic->lock); |
1fd4f2a5 ED |
648 | switch (addr) { |
649 | case IOAPIC_REG_SELECT: | |
d77fe635 | 650 | ioapic->ioregsel = data & 0xFF; /* 8-bit register */ |
1fd4f2a5 ED |
651 | break; |
652 | ||
653 | case IOAPIC_REG_WINDOW: | |
654 | ioapic_write_indirect(ioapic, data); | |
655 | break; | |
656 | ||
657 | default: | |
658 | break; | |
659 | } | |
46a47b1e | 660 | spin_unlock(&ioapic->lock); |
bda9020e | 661 | return 0; |
1fd4f2a5 ED |
662 | } |
663 | ||
7940876e | 664 | static void kvm_ioapic_reset(struct kvm_ioapic *ioapic) |
8c392696 ED |
665 | { |
666 | int i; | |
667 | ||
184564ef | 668 | cancel_delayed_work_sync(&ioapic->eoi_inject); |
8c392696 ED |
669 | for (i = 0; i < IOAPIC_NUM_PINS; i++) |
670 | ioapic->redirtbl[i].fields.mask = 1; | |
671 | ioapic->base_address = IOAPIC_DEFAULT_BASE_ADDRESS; | |
672 | ioapic->ioregsel = 0; | |
673 | ioapic->irr = 0; | |
5bda6eed | 674 | ioapic->irr_delivered = 0; |
8c392696 | 675 | ioapic->id = 0; |
8678654e | 676 | memset(ioapic->irq_eoi, 0x00, sizeof(ioapic->irq_eoi)); |
10606919 | 677 | rtc_irq_eoi_tracking_reset(ioapic); |
8c392696 ED |
678 | } |
679 | ||
d76685c4 GH |
680 | static const struct kvm_io_device_ops ioapic_mmio_ops = { |
681 | .read = ioapic_mmio_read, | |
682 | .write = ioapic_mmio_write, | |
d76685c4 GH |
683 | }; |
684 | ||
1fd4f2a5 ED |
685 | int kvm_ioapic_init(struct kvm *kvm) |
686 | { | |
687 | struct kvm_ioapic *ioapic; | |
090b7aff | 688 | int ret; |
1fd4f2a5 | 689 | |
254272ce | 690 | ioapic = kzalloc(sizeof(struct kvm_ioapic), GFP_KERNEL_ACCOUNT); |
1fd4f2a5 ED |
691 | if (!ioapic) |
692 | return -ENOMEM; | |
46a47b1e | 693 | spin_lock_init(&ioapic->lock); |
184564ef | 694 | INIT_DELAYED_WORK(&ioapic->eoi_inject, kvm_ioapic_eoi_inject_work); |
d7deeeb0 | 695 | kvm->arch.vioapic = ioapic; |
8c392696 | 696 | kvm_ioapic_reset(ioapic); |
d76685c4 | 697 | kvm_iodevice_init(&ioapic->dev, &ioapic_mmio_ops); |
1fd4f2a5 | 698 | ioapic->kvm = kvm; |
79fac95e | 699 | mutex_lock(&kvm->slots_lock); |
743eeb0b SL |
700 | ret = kvm_io_bus_register_dev(kvm, KVM_MMIO_BUS, ioapic->base_address, |
701 | IOAPIC_MEM_LENGTH, &ioapic->dev); | |
79fac95e | 702 | mutex_unlock(&kvm->slots_lock); |
1ae77bad WY |
703 | if (ret < 0) { |
704 | kvm->arch.vioapic = NULL; | |
090b7aff | 705 | kfree(ioapic); |
1ae77bad | 706 | } |
090b7aff GH |
707 | |
708 | return ret; | |
1fd4f2a5 | 709 | } |
75858a84 | 710 | |
72bb2fcd WY |
711 | void kvm_ioapic_destroy(struct kvm *kvm) |
712 | { | |
713 | struct kvm_ioapic *ioapic = kvm->arch.vioapic; | |
714 | ||
950712eb PX |
715 | if (!ioapic) |
716 | return; | |
717 | ||
184564ef | 718 | cancel_delayed_work_sync(&ioapic->eoi_inject); |
49f520b9 | 719 | mutex_lock(&kvm->slots_lock); |
d90e3a35 | 720 | kvm_io_bus_unregister_dev(kvm, KVM_MMIO_BUS, &ioapic->dev); |
49f520b9 | 721 | mutex_unlock(&kvm->slots_lock); |
d90e3a35 JL |
722 | kvm->arch.vioapic = NULL; |
723 | kfree(ioapic); | |
72bb2fcd WY |
724 | } |
725 | ||
33392b49 | 726 | void kvm_get_ioapic(struct kvm *kvm, struct kvm_ioapic_state *state) |
eba0226b | 727 | { |
0191e92d | 728 | struct kvm_ioapic *ioapic = kvm->arch.vioapic; |
eba0226b | 729 | |
46a47b1e | 730 | spin_lock(&ioapic->lock); |
eba0226b | 731 | memcpy(state, ioapic, sizeof(struct kvm_ioapic_state)); |
5bda6eed | 732 | state->irr &= ~ioapic->irr_delivered; |
46a47b1e | 733 | spin_unlock(&ioapic->lock); |
eba0226b GN |
734 | } |
735 | ||
33392b49 | 736 | void kvm_set_ioapic(struct kvm *kvm, struct kvm_ioapic_state *state) |
eba0226b | 737 | { |
0191e92d | 738 | struct kvm_ioapic *ioapic = kvm->arch.vioapic; |
eba0226b | 739 | |
46a47b1e | 740 | spin_lock(&ioapic->lock); |
eba0226b | 741 | memcpy(ioapic, state, sizeof(struct kvm_ioapic_state)); |
673f7b42 | 742 | ioapic->irr = 0; |
5bda6eed | 743 | ioapic->irr_delivered = 0; |
ca8ab3f8 | 744 | kvm_make_scan_ioapic_request(kvm); |
673f7b42 | 745 | kvm_ioapic_inject_all(ioapic, state->irr); |
46a47b1e | 746 | spin_unlock(&ioapic->lock); |
eba0226b | 747 | } |