]> git.proxmox.com Git - mirror_ubuntu-zesty-kernel.git/blame - drivers/vhost/test.c
vhost-scsi: Always access vq->private_data under vq mutex
[mirror_ubuntu-zesty-kernel.git] / drivers / vhost / test.c
CommitLineData
71ccc212
MT
1/* Copyright (C) 2009 Red Hat, Inc.
2 * Author: Michael S. Tsirkin <mst@redhat.com>
3 *
4 * This work is licensed under the terms of the GNU GPL, version 2.
5 *
6 * test virtio server in host kernel.
7 */
8
9#include <linux/compat.h>
10#include <linux/eventfd.h>
11#include <linux/vhost.h>
12#include <linux/miscdevice.h>
13#include <linux/module.h>
14#include <linux/mutex.h>
15#include <linux/workqueue.h>
16#include <linux/rcupdate.h>
17#include <linux/file.h>
18#include <linux/slab.h>
19
20#include "test.h"
6ac1afbf 21#include "vhost.h"
71ccc212
MT
22
23/* Max number of bytes transferred before requeueing the job.
24 * Using this limit prevents one virtqueue from starving others. */
25#define VHOST_TEST_WEIGHT 0x80000
26
27enum {
28 VHOST_TEST_VQ = 0,
29 VHOST_TEST_VQ_MAX = 1,
30};
31
32struct vhost_test {
33 struct vhost_dev dev;
34 struct vhost_virtqueue vqs[VHOST_TEST_VQ_MAX];
35};
36
37/* Expects to be always run from workqueue - which acts as
38 * read-size critical section for our kind of RCU. */
39static void handle_vq(struct vhost_test *n)
40{
09a34c84 41 struct vhost_virtqueue *vq = &n->vqs[VHOST_TEST_VQ];
71ccc212
MT
42 unsigned out, in;
43 int head;
44 size_t len, total_len = 0;
45 void *private;
46
09a34c84
MT
47 mutex_lock(&vq->mutex);
48 private = vq->private_data;
49 if (!private) {
50 mutex_unlock(&vq->mutex);
71ccc212 51 return;
09a34c84 52 }
71ccc212 53
8ea8cf89 54 vhost_disable_notify(&n->dev, vq);
71ccc212
MT
55
56 for (;;) {
57 head = vhost_get_vq_desc(&n->dev, vq, vq->iov,
58 ARRAY_SIZE(vq->iov),
59 &out, &in,
60 NULL, NULL);
61 /* On error, stop handling until the next kick. */
62 if (unlikely(head < 0))
63 break;
64 /* Nothing new? Wait for eventfd to tell us they refilled. */
65 if (head == vq->num) {
8ea8cf89
MT
66 if (unlikely(vhost_enable_notify(&n->dev, vq))) {
67 vhost_disable_notify(&n->dev, vq);
71ccc212
MT
68 continue;
69 }
70 break;
71 }
72 if (in) {
73 vq_err(vq, "Unexpected descriptor format for TX: "
74 "out %d, int %d\n", out, in);
75 break;
76 }
77 len = iov_length(vq->iov, out);
78 /* Sanity check */
79 if (!len) {
80 vq_err(vq, "Unexpected 0 len for TX\n");
81 break;
82 }
83 vhost_add_used_and_signal(&n->dev, vq, head, 0);
84 total_len += len;
85 if (unlikely(total_len >= VHOST_TEST_WEIGHT)) {
86 vhost_poll_queue(&vq->poll);
87 break;
88 }
89 }
90
91 mutex_unlock(&vq->mutex);
92}
93
94static void handle_vq_kick(struct vhost_work *work)
95{
96 struct vhost_virtqueue *vq = container_of(work, struct vhost_virtqueue,
97 poll.work);
98 struct vhost_test *n = container_of(vq->dev, struct vhost_test, dev);
99
100 handle_vq(n);
101}
102
103static int vhost_test_open(struct inode *inode, struct file *f)
104{
105 struct vhost_test *n = kmalloc(sizeof *n, GFP_KERNEL);
106 struct vhost_dev *dev;
09a34c84 107 struct vhost_virtqueue **vqs;
71ccc212
MT
108 int r;
109
110 if (!n)
111 return -ENOMEM;
09a34c84
MT
112 vqs = kmalloc(VHOST_TEST_VQ_MAX * sizeof(*vqs), GFP_KERNEL);
113 if (!vqs) {
114 kfree(n);
115 return -ENOMEM;
116 }
71ccc212
MT
117
118 dev = &n->dev;
09a34c84 119 vqs[VHOST_TEST_VQ] = &n->vqs[VHOST_TEST_VQ];
71ccc212 120 n->vqs[VHOST_TEST_VQ].handle_kick = handle_vq_kick;
09a34c84 121 r = vhost_dev_init(dev, vqs, VHOST_TEST_VQ_MAX);
71ccc212 122 if (r < 0) {
09a34c84 123 kfree(vqs);
71ccc212
MT
124 kfree(n);
125 return r;
126 }
127
128 f->private_data = n;
129
130 return 0;
131}
132
133static void *vhost_test_stop_vq(struct vhost_test *n,
134 struct vhost_virtqueue *vq)
135{
136 void *private;
137
138 mutex_lock(&vq->mutex);
09a34c84
MT
139 private = vq->private_data;
140 vq->private_data = NULL;
71ccc212
MT
141 mutex_unlock(&vq->mutex);
142 return private;
143}
144
145static void vhost_test_stop(struct vhost_test *n, void **privatep)
146{
147 *privatep = vhost_test_stop_vq(n, n->vqs + VHOST_TEST_VQ);
148}
149
150static void vhost_test_flush_vq(struct vhost_test *n, int index)
151{
09a34c84 152 vhost_poll_flush(&n->vqs[index].poll);
71ccc212
MT
153}
154
155static void vhost_test_flush(struct vhost_test *n)
156{
157 vhost_test_flush_vq(n, VHOST_TEST_VQ);
158}
159
160static int vhost_test_release(struct inode *inode, struct file *f)
161{
162 struct vhost_test *n = f->private_data;
163 void *private;
164
165 vhost_test_stop(n, &private);
166 vhost_test_flush(n);
5e7045b0 167 vhost_dev_cleanup(&n->dev, false);
71ccc212
MT
168 /* We do an extra flush before freeing memory,
169 * since jobs can re-queue themselves. */
170 vhost_test_flush(n);
171 kfree(n);
172 return 0;
173}
174
175static long vhost_test_run(struct vhost_test *n, int test)
176{
177 void *priv, *oldpriv;
178 struct vhost_virtqueue *vq;
179 int r, index;
180
181 if (test < 0 || test > 1)
182 return -EINVAL;
183
184 mutex_lock(&n->dev.mutex);
185 r = vhost_dev_check_owner(&n->dev);
186 if (r)
187 goto err;
188
189 for (index = 0; index < n->dev.nvqs; ++index) {
190 /* Verify that ring has been setup correctly. */
191 if (!vhost_vq_access_ok(&n->vqs[index])) {
192 r = -EFAULT;
193 goto err;
194 }
195 }
196
197 for (index = 0; index < n->dev.nvqs; ++index) {
198 vq = n->vqs + index;
199 mutex_lock(&vq->mutex);
200 priv = test ? n : NULL;
201
202 /* start polling new socket */
203 oldpriv = rcu_dereference_protected(vq->private_data,
204 lockdep_is_held(&vq->mutex));
205 rcu_assign_pointer(vq->private_data, priv);
206
f59281da
JW
207 r = vhost_init_used(&n->vqs[index]);
208
71ccc212
MT
209 mutex_unlock(&vq->mutex);
210
f59281da
JW
211 if (r)
212 goto err;
213
71ccc212
MT
214 if (oldpriv) {
215 vhost_test_flush_vq(n, index);
216 }
217 }
218
219 mutex_unlock(&n->dev.mutex);
220 return 0;
221
222err:
223 mutex_unlock(&n->dev.mutex);
224 return r;
225}
226
227static long vhost_test_reset_owner(struct vhost_test *n)
228{
229 void *priv = NULL;
230 long err;
150b9e51
MT
231 struct vhost_memory *memory;
232
71ccc212
MT
233 mutex_lock(&n->dev.mutex);
234 err = vhost_dev_check_owner(&n->dev);
235 if (err)
236 goto done;
150b9e51
MT
237 memory = vhost_dev_reset_owner_prepare();
238 if (!memory) {
239 err = -ENOMEM;
240 goto done;
241 }
71ccc212
MT
242 vhost_test_stop(n, &priv);
243 vhost_test_flush(n);
150b9e51 244 vhost_dev_reset_owner(&n->dev, memory);
71ccc212
MT
245done:
246 mutex_unlock(&n->dev.mutex);
247 return err;
248}
249
250static int vhost_test_set_features(struct vhost_test *n, u64 features)
251{
252 mutex_lock(&n->dev.mutex);
253 if ((features & (1 << VHOST_F_LOG_ALL)) &&
254 !vhost_log_access_ok(&n->dev)) {
255 mutex_unlock(&n->dev.mutex);
256 return -EFAULT;
257 }
258 n->dev.acked_features = features;
259 smp_wmb();
260 vhost_test_flush(n);
261 mutex_unlock(&n->dev.mutex);
262 return 0;
263}
264
265static long vhost_test_ioctl(struct file *f, unsigned int ioctl,
266 unsigned long arg)
267{
268 struct vhost_test *n = f->private_data;
269 void __user *argp = (void __user *)arg;
270 u64 __user *featurep = argp;
271 int test;
272 u64 features;
273 int r;
274 switch (ioctl) {
275 case VHOST_TEST_RUN:
276 if (copy_from_user(&test, argp, sizeof test))
277 return -EFAULT;
278 return vhost_test_run(n, test);
279 case VHOST_GET_FEATURES:
09a34c84 280 features = VHOST_FEATURES;
71ccc212
MT
281 if (copy_to_user(featurep, &features, sizeof features))
282 return -EFAULT;
283 return 0;
284 case VHOST_SET_FEATURES:
285 if (copy_from_user(&features, featurep, sizeof features))
286 return -EFAULT;
09a34c84 287 if (features & ~VHOST_FEATURES)
71ccc212
MT
288 return -EOPNOTSUPP;
289 return vhost_test_set_features(n, features);
290 case VHOST_RESET_OWNER:
291 return vhost_test_reset_owner(n);
292 default:
293 mutex_lock(&n->dev.mutex);
73640c99
MT
294 r = vhost_dev_ioctl(&n->dev, ioctl, argp);
295 if (r == -ENOIOCTLCMD)
296 r = vhost_vring_ioctl(&n->dev, ioctl, argp);
71ccc212
MT
297 vhost_test_flush(n);
298 mutex_unlock(&n->dev.mutex);
299 return r;
300 }
301}
302
303#ifdef CONFIG_COMPAT
304static long vhost_test_compat_ioctl(struct file *f, unsigned int ioctl,
305 unsigned long arg)
306{
307 return vhost_test_ioctl(f, ioctl, (unsigned long)compat_ptr(arg));
308}
309#endif
310
311static const struct file_operations vhost_test_fops = {
312 .owner = THIS_MODULE,
313 .release = vhost_test_release,
314 .unlocked_ioctl = vhost_test_ioctl,
315#ifdef CONFIG_COMPAT
316 .compat_ioctl = vhost_test_compat_ioctl,
317#endif
318 .open = vhost_test_open,
319 .llseek = noop_llseek,
320};
321
322static struct miscdevice vhost_test_misc = {
323 MISC_DYNAMIC_MINOR,
324 "vhost-test",
325 &vhost_test_fops,
326};
327
328static int vhost_test_init(void)
329{
330 return misc_register(&vhost_test_misc);
331}
332module_init(vhost_test_init);
333
334static void vhost_test_exit(void)
335{
336 misc_deregister(&vhost_test_misc);
337}
338module_exit(vhost_test_exit);
339
340MODULE_VERSION("0.0.1");
341MODULE_LICENSE("GPL v2");
342MODULE_AUTHOR("Michael S. Tsirkin");
343MODULE_DESCRIPTION("Host kernel side for virtio simulator");