]>
Commit | Line | Data |
---|---|---|
b2441318 | 1 | // SPDX-License-Identifier: GPL-2.0 |
5f94c174 LV |
2 | /* |
3 | * KVM coalesced MMIO | |
4 | * | |
5 | * Copyright (c) 2008 Bull S.A.S. | |
221d059d | 6 | * Copyright 2009 Red Hat, Inc. and/or its affiliates. |
5f94c174 LV |
7 | * |
8 | * Author: Laurent Vivier <Laurent.Vivier@bull.net> | |
9 | * | |
10 | */ | |
11 | ||
af669ac6 | 12 | #include <kvm/iodev.h> |
5f94c174 LV |
13 | |
14 | #include <linux/kvm_host.h> | |
5a0e3ad6 | 15 | #include <linux/slab.h> |
5f94c174 LV |
16 | #include <linux/kvm.h> |
17 | ||
18 | #include "coalesced_mmio.h" | |
19 | ||
d76685c4 GH |
20 | static inline struct kvm_coalesced_mmio_dev *to_mmio(struct kvm_io_device *dev) |
21 | { | |
22 | return container_of(dev, struct kvm_coalesced_mmio_dev, dev); | |
23 | } | |
24 | ||
bda9020e MT |
25 | static int coalesced_mmio_in_range(struct kvm_coalesced_mmio_dev *dev, |
26 | gpa_t addr, int len) | |
5f94c174 | 27 | { |
2b3c246a SL |
28 | /* is it in a batchable area ? |
29 | * (addr,len) is fully included in | |
30 | * (zone->addr, zone->size) | |
31 | */ | |
1a214246 DC |
32 | if (len < 0) |
33 | return 0; | |
34 | if (addr + len < addr) | |
35 | return 0; | |
36 | if (addr < dev->zone.addr) | |
37 | return 0; | |
38 | if (addr + len > dev->zone.addr + dev->zone.size) | |
39 | return 0; | |
40 | return 1; | |
5f94c174 LV |
41 | } |
42 | ||
e1a5725f | 43 | static int coalesced_mmio_has_room(struct kvm_coalesced_mmio_dev *dev, u32 last) |
c298125f SL |
44 | { |
45 | struct kvm_coalesced_mmio_ring *ring; | |
46 | unsigned avail; | |
47 | ||
48 | /* Are we able to batch it ? */ | |
49 | ||
50 | /* last is the first free entry | |
51 | * check if we don't meet the first used entry | |
52 | * there is always one unused entry in the buffer | |
53 | */ | |
54 | ring = dev->kvm->coalesced_mmio_ring; | |
e1a5725f | 55 | avail = (ring->first - last - 1) % KVM_COALESCED_MMIO_MAX; |
c298125f SL |
56 | if (avail == 0) { |
57 | /* full */ | |
58 | return 0; | |
59 | } | |
60 | ||
61 | return 1; | |
62 | } | |
63 | ||
e32edf4f NN |
64 | static int coalesced_mmio_write(struct kvm_vcpu *vcpu, |
65 | struct kvm_io_device *this, gpa_t addr, | |
66 | int len, const void *val) | |
5f94c174 | 67 | { |
d76685c4 | 68 | struct kvm_coalesced_mmio_dev *dev = to_mmio(this); |
5f94c174 | 69 | struct kvm_coalesced_mmio_ring *ring = dev->kvm->coalesced_mmio_ring; |
e1a5725f | 70 | __u32 insert; |
c298125f | 71 | |
bda9020e MT |
72 | if (!coalesced_mmio_in_range(dev, addr, len)) |
73 | return -EOPNOTSUPP; | |
5f94c174 | 74 | |
2b3c246a | 75 | spin_lock(&dev->kvm->ring_lock); |
5f94c174 | 76 | |
e1a5725f MD |
77 | insert = READ_ONCE(ring->last); |
78 | if (!coalesced_mmio_has_room(dev, insert) || | |
79 | insert >= KVM_COALESCED_MMIO_MAX) { | |
2b3c246a | 80 | spin_unlock(&dev->kvm->ring_lock); |
c298125f SL |
81 | return -EOPNOTSUPP; |
82 | } | |
83 | ||
5f94c174 LV |
84 | /* copy data in first free entry of the ring */ |
85 | ||
e1a5725f MD |
86 | ring->coalesced_mmio[insert].phys_addr = addr; |
87 | ring->coalesced_mmio[insert].len = len; | |
88 | memcpy(ring->coalesced_mmio[insert].data, val, len); | |
5f94c174 | 89 | smp_wmb(); |
e1a5725f | 90 | ring->last = (insert + 1) % KVM_COALESCED_MMIO_MAX; |
2b3c246a | 91 | spin_unlock(&dev->kvm->ring_lock); |
bda9020e | 92 | return 0; |
5f94c174 LV |
93 | } |
94 | ||
95 | static void coalesced_mmio_destructor(struct kvm_io_device *this) | |
96 | { | |
d76685c4 | 97 | struct kvm_coalesced_mmio_dev *dev = to_mmio(this); |
787a660a | 98 | |
2b3c246a SL |
99 | list_del(&dev->list); |
100 | ||
787a660a | 101 | kfree(dev); |
5f94c174 LV |
102 | } |
103 | ||
d76685c4 GH |
104 | static const struct kvm_io_device_ops coalesced_mmio_ops = { |
105 | .write = coalesced_mmio_write, | |
d76685c4 GH |
106 | .destructor = coalesced_mmio_destructor, |
107 | }; | |
108 | ||
5f94c174 LV |
109 | int kvm_coalesced_mmio_init(struct kvm *kvm) |
110 | { | |
980da6ce | 111 | struct page *page; |
090b7aff | 112 | int ret; |
5f94c174 | 113 | |
980da6ce AK |
114 | ret = -ENOMEM; |
115 | page = alloc_page(GFP_KERNEL | __GFP_ZERO); | |
116 | if (!page) | |
117 | goto out_err; | |
980da6ce | 118 | |
2b3c246a SL |
119 | ret = 0; |
120 | kvm->coalesced_mmio_ring = page_address(page); | |
980da6ce | 121 | |
2b3c246a SL |
122 | /* |
123 | * We're using this spinlock to sync access to the coalesced ring. | |
124 | * The list doesn't need it's own lock since device registration and | |
125 | * unregistration should only happen when kvm->slots_lock is held. | |
126 | */ | |
127 | spin_lock_init(&kvm->ring_lock); | |
128 | INIT_LIST_HEAD(&kvm->coalesced_zones); | |
090b7aff | 129 | |
980da6ce | 130 | out_err: |
090b7aff | 131 | return ret; |
5f94c174 LV |
132 | } |
133 | ||
980da6ce AK |
134 | void kvm_coalesced_mmio_free(struct kvm *kvm) |
135 | { | |
136 | if (kvm->coalesced_mmio_ring) | |
137 | free_page((unsigned long)kvm->coalesced_mmio_ring); | |
138 | } | |
139 | ||
5f94c174 | 140 | int kvm_vm_ioctl_register_coalesced_mmio(struct kvm *kvm, |
43db6697 | 141 | struct kvm_coalesced_mmio_zone *zone) |
5f94c174 | 142 | { |
2b3c246a SL |
143 | int ret; |
144 | struct kvm_coalesced_mmio_dev *dev; | |
5f94c174 | 145 | |
2b3c246a SL |
146 | dev = kzalloc(sizeof(struct kvm_coalesced_mmio_dev), GFP_KERNEL); |
147 | if (!dev) | |
148 | return -ENOMEM; | |
149 | ||
150 | kvm_iodevice_init(&dev->dev, &coalesced_mmio_ops); | |
151 | dev->kvm = kvm; | |
152 | dev->zone = *zone; | |
5f94c174 | 153 | |
79fac95e | 154 | mutex_lock(&kvm->slots_lock); |
743eeb0b SL |
155 | ret = kvm_io_bus_register_dev(kvm, KVM_MMIO_BUS, zone->addr, |
156 | zone->size, &dev->dev); | |
2b3c246a SL |
157 | if (ret < 0) |
158 | goto out_free_dev; | |
159 | list_add_tail(&dev->list, &kvm->coalesced_zones); | |
160 | mutex_unlock(&kvm->slots_lock); | |
5f94c174 | 161 | |
aac5c422 | 162 | return 0; |
5f94c174 | 163 | |
2b3c246a | 164 | out_free_dev: |
79fac95e | 165 | mutex_unlock(&kvm->slots_lock); |
2b3c246a SL |
166 | kfree(dev); |
167 | ||
aac5c422 | 168 | return ret; |
5f94c174 LV |
169 | } |
170 | ||
171 | int kvm_vm_ioctl_unregister_coalesced_mmio(struct kvm *kvm, | |
172 | struct kvm_coalesced_mmio_zone *zone) | |
173 | { | |
2b3c246a | 174 | struct kvm_coalesced_mmio_dev *dev, *tmp; |
5f94c174 | 175 | |
79fac95e | 176 | mutex_lock(&kvm->slots_lock); |
5f94c174 | 177 | |
2b3c246a SL |
178 | list_for_each_entry_safe(dev, tmp, &kvm->coalesced_zones, list) |
179 | if (coalesced_mmio_in_range(dev, zone->addr, zone->size)) { | |
180 | kvm_io_bus_unregister_dev(kvm, KVM_MMIO_BUS, &dev->dev); | |
181 | kvm_iodevice_destructor(&dev->dev); | |
5f94c174 | 182 | } |
5f94c174 | 183 | |
79fac95e | 184 | mutex_unlock(&kvm->slots_lock); |
5f94c174 LV |
185 | |
186 | return 0; | |
187 | } |