Merge branch 'for-3.11/core' of git://git.kernel.dk/linux-block
[deliverable/linux.git] / drivers / vfio / vfio.c
1 /*
2 * VFIO core
3 *
4 * Copyright (C) 2012 Red Hat, Inc. All rights reserved.
5 * Author: Alex Williamson <alex.williamson@redhat.com>
6 *
7 * This program is free software; you can redistribute it and/or modify
8 * it under the terms of the GNU General Public License version 2 as
9 * published by the Free Software Foundation.
10 *
11 * Derived from original vfio:
12 * Copyright 2010 Cisco Systems, Inc. All rights reserved.
13 * Author: Tom Lyon, pugs@cisco.com
14 */
15
16 #include <linux/cdev.h>
17 #include <linux/compat.h>
18 #include <linux/device.h>
19 #include <linux/file.h>
20 #include <linux/anon_inodes.h>
21 #include <linux/fs.h>
22 #include <linux/idr.h>
23 #include <linux/iommu.h>
24 #include <linux/list.h>
25 #include <linux/module.h>
26 #include <linux/mutex.h>
27 #include <linux/rwsem.h>
28 #include <linux/sched.h>
29 #include <linux/slab.h>
30 #include <linux/stat.h>
31 #include <linux/string.h>
32 #include <linux/uaccess.h>
33 #include <linux/vfio.h>
34 #include <linux/wait.h>
35
36 #define DRIVER_VERSION "0.3"
37 #define DRIVER_AUTHOR "Alex Williamson <alex.williamson@redhat.com>"
38 #define DRIVER_DESC "VFIO - User Level meta-driver"
39
40 static struct vfio {
41 struct class *class;
42 struct list_head iommu_drivers_list;
43 struct mutex iommu_drivers_lock;
44 struct list_head group_list;
45 struct idr group_idr;
46 struct mutex group_lock;
47 struct cdev group_cdev;
48 struct device *dev;
49 dev_t devt;
50 struct cdev cdev;
51 wait_queue_head_t release_q;
52 } vfio;
53
54 struct vfio_iommu_driver {
55 const struct vfio_iommu_driver_ops *ops;
56 struct list_head vfio_next;
57 };
58
59 struct vfio_container {
60 struct kref kref;
61 struct list_head group_list;
62 struct rw_semaphore group_lock;
63 struct vfio_iommu_driver *iommu_driver;
64 void *iommu_data;
65 };
66
67 struct vfio_group {
68 struct kref kref;
69 int minor;
70 atomic_t container_users;
71 struct iommu_group *iommu_group;
72 struct vfio_container *container;
73 struct list_head device_list;
74 struct mutex device_lock;
75 struct device *dev;
76 struct notifier_block nb;
77 struct list_head vfio_next;
78 struct list_head container_next;
79 atomic_t opened;
80 };
81
82 struct vfio_device {
83 struct kref kref;
84 struct device *dev;
85 const struct vfio_device_ops *ops;
86 struct vfio_group *group;
87 struct list_head group_next;
88 void *device_data;
89 };
90
91 /**
92 * IOMMU driver registration
93 */
94 int vfio_register_iommu_driver(const struct vfio_iommu_driver_ops *ops)
95 {
96 struct vfio_iommu_driver *driver, *tmp;
97
98 driver = kzalloc(sizeof(*driver), GFP_KERNEL);
99 if (!driver)
100 return -ENOMEM;
101
102 driver->ops = ops;
103
104 mutex_lock(&vfio.iommu_drivers_lock);
105
106 /* Check for duplicates */
107 list_for_each_entry(tmp, &vfio.iommu_drivers_list, vfio_next) {
108 if (tmp->ops == ops) {
109 mutex_unlock(&vfio.iommu_drivers_lock);
110 kfree(driver);
111 return -EINVAL;
112 }
113 }
114
115 list_add(&driver->vfio_next, &vfio.iommu_drivers_list);
116
117 mutex_unlock(&vfio.iommu_drivers_lock);
118
119 return 0;
120 }
121 EXPORT_SYMBOL_GPL(vfio_register_iommu_driver);
122
123 void vfio_unregister_iommu_driver(const struct vfio_iommu_driver_ops *ops)
124 {
125 struct vfio_iommu_driver *driver;
126
127 mutex_lock(&vfio.iommu_drivers_lock);
128 list_for_each_entry(driver, &vfio.iommu_drivers_list, vfio_next) {
129 if (driver->ops == ops) {
130 list_del(&driver->vfio_next);
131 mutex_unlock(&vfio.iommu_drivers_lock);
132 kfree(driver);
133 return;
134 }
135 }
136 mutex_unlock(&vfio.iommu_drivers_lock);
137 }
138 EXPORT_SYMBOL_GPL(vfio_unregister_iommu_driver);
139
140 /**
141 * Group minor allocation/free - both called with vfio.group_lock held
142 */
143 static int vfio_alloc_group_minor(struct vfio_group *group)
144 {
145 /* index 0 is used by /dev/vfio/vfio */
146 return idr_alloc(&vfio.group_idr, group, 1, MINORMASK + 1, GFP_KERNEL);
147 }
148
149 static void vfio_free_group_minor(int minor)
150 {
151 idr_remove(&vfio.group_idr, minor);
152 }
153
154 static int vfio_iommu_group_notifier(struct notifier_block *nb,
155 unsigned long action, void *data);
156 static void vfio_group_get(struct vfio_group *group);
157
158 /**
159 * Container objects - containers are created when /dev/vfio/vfio is
160 * opened, but their lifecycle extends until the last user is done, so
161 * it's freed via kref. Must support container/group/device being
162 * closed in any order.
163 */
164 static void vfio_container_get(struct vfio_container *container)
165 {
166 kref_get(&container->kref);
167 }
168
169 static void vfio_container_release(struct kref *kref)
170 {
171 struct vfio_container *container;
172 container = container_of(kref, struct vfio_container, kref);
173
174 kfree(container);
175 }
176
177 static void vfio_container_put(struct vfio_container *container)
178 {
179 kref_put(&container->kref, vfio_container_release);
180 }
181
182 static void vfio_group_unlock_and_free(struct vfio_group *group)
183 {
184 mutex_unlock(&vfio.group_lock);
185 /*
186 * Unregister outside of lock. A spurious callback is harmless now
187 * that the group is no longer in vfio.group_list.
188 */
189 iommu_group_unregister_notifier(group->iommu_group, &group->nb);
190 kfree(group);
191 }
192
193 /**
194 * Group objects - create, release, get, put, search
195 */
196 static struct vfio_group *vfio_create_group(struct iommu_group *iommu_group)
197 {
198 struct vfio_group *group, *tmp;
199 struct device *dev;
200 int ret, minor;
201
202 group = kzalloc(sizeof(*group), GFP_KERNEL);
203 if (!group)
204 return ERR_PTR(-ENOMEM);
205
206 kref_init(&group->kref);
207 INIT_LIST_HEAD(&group->device_list);
208 mutex_init(&group->device_lock);
209 atomic_set(&group->container_users, 0);
210 atomic_set(&group->opened, 0);
211 group->iommu_group = iommu_group;
212
213 group->nb.notifier_call = vfio_iommu_group_notifier;
214
215 /*
216 * blocking notifiers acquire a rwsem around registering and hold
217 * it around callback. Therefore, need to register outside of
218 * vfio.group_lock to avoid A-B/B-A contention. Our callback won't
219 * do anything unless it can find the group in vfio.group_list, so
220 * no harm in registering early.
221 */
222 ret = iommu_group_register_notifier(iommu_group, &group->nb);
223 if (ret) {
224 kfree(group);
225 return ERR_PTR(ret);
226 }
227
228 mutex_lock(&vfio.group_lock);
229
230 minor = vfio_alloc_group_minor(group);
231 if (minor < 0) {
232 vfio_group_unlock_and_free(group);
233 return ERR_PTR(minor);
234 }
235
236 /* Did we race creating this group? */
237 list_for_each_entry(tmp, &vfio.group_list, vfio_next) {
238 if (tmp->iommu_group == iommu_group) {
239 vfio_group_get(tmp);
240 vfio_free_group_minor(minor);
241 vfio_group_unlock_and_free(group);
242 return tmp;
243 }
244 }
245
246 dev = device_create(vfio.class, NULL, MKDEV(MAJOR(vfio.devt), minor),
247 group, "%d", iommu_group_id(iommu_group));
248 if (IS_ERR(dev)) {
249 vfio_free_group_minor(minor);
250 vfio_group_unlock_and_free(group);
251 return (struct vfio_group *)dev; /* ERR_PTR */
252 }
253
254 group->minor = minor;
255 group->dev = dev;
256
257 list_add(&group->vfio_next, &vfio.group_list);
258
259 mutex_unlock(&vfio.group_lock);
260
261 return group;
262 }
263
264 /* called with vfio.group_lock held */
265 static void vfio_group_release(struct kref *kref)
266 {
267 struct vfio_group *group = container_of(kref, struct vfio_group, kref);
268
269 WARN_ON(!list_empty(&group->device_list));
270
271 device_destroy(vfio.class, MKDEV(MAJOR(vfio.devt), group->minor));
272 list_del(&group->vfio_next);
273 vfio_free_group_minor(group->minor);
274 vfio_group_unlock_and_free(group);
275 }
276
277 static void vfio_group_put(struct vfio_group *group)
278 {
279 kref_put_mutex(&group->kref, vfio_group_release, &vfio.group_lock);
280 }
281
282 /* Assume group_lock or group reference is held */
283 static void vfio_group_get(struct vfio_group *group)
284 {
285 kref_get(&group->kref);
286 }
287
288 /*
289 * Not really a try as we will sleep for mutex, but we need to make
290 * sure the group pointer is valid under lock and get a reference.
291 */
292 static struct vfio_group *vfio_group_try_get(struct vfio_group *group)
293 {
294 struct vfio_group *target = group;
295
296 mutex_lock(&vfio.group_lock);
297 list_for_each_entry(group, &vfio.group_list, vfio_next) {
298 if (group == target) {
299 vfio_group_get(group);
300 mutex_unlock(&vfio.group_lock);
301 return group;
302 }
303 }
304 mutex_unlock(&vfio.group_lock);
305
306 return NULL;
307 }
308
309 static
310 struct vfio_group *vfio_group_get_from_iommu(struct iommu_group *iommu_group)
311 {
312 struct vfio_group *group;
313
314 mutex_lock(&vfio.group_lock);
315 list_for_each_entry(group, &vfio.group_list, vfio_next) {
316 if (group->iommu_group == iommu_group) {
317 vfio_group_get(group);
318 mutex_unlock(&vfio.group_lock);
319 return group;
320 }
321 }
322 mutex_unlock(&vfio.group_lock);
323
324 return NULL;
325 }
326
327 static struct vfio_group *vfio_group_get_from_minor(int minor)
328 {
329 struct vfio_group *group;
330
331 mutex_lock(&vfio.group_lock);
332 group = idr_find(&vfio.group_idr, minor);
333 if (!group) {
334 mutex_unlock(&vfio.group_lock);
335 return NULL;
336 }
337 vfio_group_get(group);
338 mutex_unlock(&vfio.group_lock);
339
340 return group;
341 }
342
343 /**
344 * Device objects - create, release, get, put, search
345 */
346 static
347 struct vfio_device *vfio_group_create_device(struct vfio_group *group,
348 struct device *dev,
349 const struct vfio_device_ops *ops,
350 void *device_data)
351 {
352 struct vfio_device *device;
353 int ret;
354
355 device = kzalloc(sizeof(*device), GFP_KERNEL);
356 if (!device)
357 return ERR_PTR(-ENOMEM);
358
359 kref_init(&device->kref);
360 device->dev = dev;
361 device->group = group;
362 device->ops = ops;
363 device->device_data = device_data;
364
365 ret = dev_set_drvdata(dev, device);
366 if (ret) {
367 kfree(device);
368 return ERR_PTR(ret);
369 }
370
371 /* No need to get group_lock, caller has group reference */
372 vfio_group_get(group);
373
374 mutex_lock(&group->device_lock);
375 list_add(&device->group_next, &group->device_list);
376 mutex_unlock(&group->device_lock);
377
378 return device;
379 }
380
381 static void vfio_device_release(struct kref *kref)
382 {
383 struct vfio_device *device = container_of(kref,
384 struct vfio_device, kref);
385 struct vfio_group *group = device->group;
386
387 list_del(&device->group_next);
388 mutex_unlock(&group->device_lock);
389
390 dev_set_drvdata(device->dev, NULL);
391
392 kfree(device);
393
394 /* vfio_del_group_dev may be waiting for this device */
395 wake_up(&vfio.release_q);
396 }
397
398 /* Device reference always implies a group reference */
399 void vfio_device_put(struct vfio_device *device)
400 {
401 struct vfio_group *group = device->group;
402 kref_put_mutex(&device->kref, vfio_device_release, &group->device_lock);
403 vfio_group_put(group);
404 }
405 EXPORT_SYMBOL_GPL(vfio_device_put);
406
407 static void vfio_device_get(struct vfio_device *device)
408 {
409 vfio_group_get(device->group);
410 kref_get(&device->kref);
411 }
412
413 static struct vfio_device *vfio_group_get_device(struct vfio_group *group,
414 struct device *dev)
415 {
416 struct vfio_device *device;
417
418 mutex_lock(&group->device_lock);
419 list_for_each_entry(device, &group->device_list, group_next) {
420 if (device->dev == dev) {
421 vfio_device_get(device);
422 mutex_unlock(&group->device_lock);
423 return device;
424 }
425 }
426 mutex_unlock(&group->device_lock);
427 return NULL;
428 }
429
430 /*
431 * Whitelist some drivers that we know are safe (no dma) or just sit on
432 * a device. It's not always practical to leave a device within a group
433 * driverless as it could get re-bound to something unsafe.
434 */
435 static const char * const vfio_driver_whitelist[] = { "pci-stub", "pcieport" };
436
437 static bool vfio_whitelisted_driver(struct device_driver *drv)
438 {
439 int i;
440
441 for (i = 0; i < ARRAY_SIZE(vfio_driver_whitelist); i++) {
442 if (!strcmp(drv->name, vfio_driver_whitelist[i]))
443 return true;
444 }
445
446 return false;
447 }
448
449 /*
450 * A vfio group is viable for use by userspace if all devices are either
451 * driver-less or bound to a vfio or whitelisted driver. We test the
452 * latter by the existence of a struct vfio_device matching the dev.
453 */
454 static int vfio_dev_viable(struct device *dev, void *data)
455 {
456 struct vfio_group *group = data;
457 struct vfio_device *device;
458 struct device_driver *drv = ACCESS_ONCE(dev->driver);
459
460 if (!drv || vfio_whitelisted_driver(drv))
461 return 0;
462
463 device = vfio_group_get_device(group, dev);
464 if (device) {
465 vfio_device_put(device);
466 return 0;
467 }
468
469 return -EINVAL;
470 }
471
472 /**
473 * Async device support
474 */
475 static int vfio_group_nb_add_dev(struct vfio_group *group, struct device *dev)
476 {
477 struct vfio_device *device;
478
479 /* Do we already know about it? We shouldn't */
480 device = vfio_group_get_device(group, dev);
481 if (WARN_ON_ONCE(device)) {
482 vfio_device_put(device);
483 return 0;
484 }
485
486 /* Nothing to do for idle groups */
487 if (!atomic_read(&group->container_users))
488 return 0;
489
490 /* TODO Prevent device auto probing */
491 WARN("Device %s added to live group %d!\n", dev_name(dev),
492 iommu_group_id(group->iommu_group));
493
494 return 0;
495 }
496
497 static int vfio_group_nb_del_dev(struct vfio_group *group, struct device *dev)
498 {
499 struct vfio_device *device;
500
501 /*
502 * Expect to fall out here. If a device was in use, it would
503 * have been bound to a vfio sub-driver, which would have blocked
504 * in .remove at vfio_del_group_dev. Sanity check that we no
505 * longer track the device, so it's safe to remove.
506 */
507 device = vfio_group_get_device(group, dev);
508 if (likely(!device))
509 return 0;
510
511 WARN("Device %s removed from live group %d!\n", dev_name(dev),
512 iommu_group_id(group->iommu_group));
513
514 vfio_device_put(device);
515 return 0;
516 }
517
518 static int vfio_group_nb_verify(struct vfio_group *group, struct device *dev)
519 {
520 /* We don't care what happens when the group isn't in use */
521 if (!atomic_read(&group->container_users))
522 return 0;
523
524 return vfio_dev_viable(dev, group);
525 }
526
527 static int vfio_iommu_group_notifier(struct notifier_block *nb,
528 unsigned long action, void *data)
529 {
530 struct vfio_group *group = container_of(nb, struct vfio_group, nb);
531 struct device *dev = data;
532
533 /*
534 * Need to go through a group_lock lookup to get a reference or
535 * we risk racing a group being removed. Leave a WARN_ON for
536 * debuging, but if the group no longer exists, a spurious notify
537 * is harmless.
538 */
539 group = vfio_group_try_get(group);
540 if (WARN_ON(!group))
541 return NOTIFY_OK;
542
543 switch (action) {
544 case IOMMU_GROUP_NOTIFY_ADD_DEVICE:
545 vfio_group_nb_add_dev(group, dev);
546 break;
547 case IOMMU_GROUP_NOTIFY_DEL_DEVICE:
548 vfio_group_nb_del_dev(group, dev);
549 break;
550 case IOMMU_GROUP_NOTIFY_BIND_DRIVER:
551 pr_debug("%s: Device %s, group %d binding to driver\n",
552 __func__, dev_name(dev),
553 iommu_group_id(group->iommu_group));
554 break;
555 case IOMMU_GROUP_NOTIFY_BOUND_DRIVER:
556 pr_debug("%s: Device %s, group %d bound to driver %s\n",
557 __func__, dev_name(dev),
558 iommu_group_id(group->iommu_group), dev->driver->name);
559 BUG_ON(vfio_group_nb_verify(group, dev));
560 break;
561 case IOMMU_GROUP_NOTIFY_UNBIND_DRIVER:
562 pr_debug("%s: Device %s, group %d unbinding from driver %s\n",
563 __func__, dev_name(dev),
564 iommu_group_id(group->iommu_group), dev->driver->name);
565 break;
566 case IOMMU_GROUP_NOTIFY_UNBOUND_DRIVER:
567 pr_debug("%s: Device %s, group %d unbound from driver\n",
568 __func__, dev_name(dev),
569 iommu_group_id(group->iommu_group));
570 /*
571 * XXX An unbound device in a live group is ok, but we'd
572 * really like to avoid the above BUG_ON by preventing other
573 * drivers from binding to it. Once that occurs, we have to
574 * stop the system to maintain isolation. At a minimum, we'd
575 * want a toggle to disable driver auto probe for this device.
576 */
577 break;
578 }
579
580 vfio_group_put(group);
581 return NOTIFY_OK;
582 }
583
584 /**
585 * VFIO driver API
586 */
587 int vfio_add_group_dev(struct device *dev,
588 const struct vfio_device_ops *ops, void *device_data)
589 {
590 struct iommu_group *iommu_group;
591 struct vfio_group *group;
592 struct vfio_device *device;
593
594 iommu_group = iommu_group_get(dev);
595 if (!iommu_group)
596 return -EINVAL;
597
598 group = vfio_group_get_from_iommu(iommu_group);
599 if (!group) {
600 group = vfio_create_group(iommu_group);
601 if (IS_ERR(group)) {
602 iommu_group_put(iommu_group);
603 return PTR_ERR(group);
604 }
605 }
606
607 device = vfio_group_get_device(group, dev);
608 if (device) {
609 WARN(1, "Device %s already exists on group %d\n",
610 dev_name(dev), iommu_group_id(iommu_group));
611 vfio_device_put(device);
612 vfio_group_put(group);
613 iommu_group_put(iommu_group);
614 return -EBUSY;
615 }
616
617 device = vfio_group_create_device(group, dev, ops, device_data);
618 if (IS_ERR(device)) {
619 vfio_group_put(group);
620 iommu_group_put(iommu_group);
621 return PTR_ERR(device);
622 }
623
624 /*
625 * Added device holds reference to iommu_group and vfio_device
626 * (which in turn holds reference to vfio_group). Drop extra
627 * group reference used while acquiring device.
628 */
629 vfio_group_put(group);
630
631 return 0;
632 }
633 EXPORT_SYMBOL_GPL(vfio_add_group_dev);
634
635 /**
636 * Get a reference to the vfio_device for a device that is known to
637 * be bound to a vfio driver. The driver implicitly holds a
638 * vfio_device reference between vfio_add_group_dev and
639 * vfio_del_group_dev. We can therefore use drvdata to increment
640 * that reference from the struct device. This additional
641 * reference must be released by calling vfio_device_put.
642 */
643 struct vfio_device *vfio_device_get_from_dev(struct device *dev)
644 {
645 struct vfio_device *device = dev_get_drvdata(dev);
646
647 vfio_device_get(device);
648
649 return device;
650 }
651 EXPORT_SYMBOL_GPL(vfio_device_get_from_dev);
652
653 /*
654 * Caller must hold a reference to the vfio_device
655 */
656 void *vfio_device_data(struct vfio_device *device)
657 {
658 return device->device_data;
659 }
660 EXPORT_SYMBOL_GPL(vfio_device_data);
661
662 /* Given a referenced group, check if it contains the device */
663 static bool vfio_dev_present(struct vfio_group *group, struct device *dev)
664 {
665 struct vfio_device *device;
666
667 device = vfio_group_get_device(group, dev);
668 if (!device)
669 return false;
670
671 vfio_device_put(device);
672 return true;
673 }
674
675 /*
676 * Decrement the device reference count and wait for the device to be
677 * removed. Open file descriptors for the device... */
678 void *vfio_del_group_dev(struct device *dev)
679 {
680 struct vfio_device *device = dev_get_drvdata(dev);
681 struct vfio_group *group = device->group;
682 struct iommu_group *iommu_group = group->iommu_group;
683 void *device_data = device->device_data;
684
685 /*
686 * The group exists so long as we have a device reference. Get
687 * a group reference and use it to scan for the device going away.
688 */
689 vfio_group_get(group);
690
691 vfio_device_put(device);
692
693 /* TODO send a signal to encourage this to be released */
694 wait_event(vfio.release_q, !vfio_dev_present(group, dev));
695
696 vfio_group_put(group);
697
698 iommu_group_put(iommu_group);
699
700 return device_data;
701 }
702 EXPORT_SYMBOL_GPL(vfio_del_group_dev);
703
704 /**
705 * VFIO base fd, /dev/vfio/vfio
706 */
707 static long vfio_ioctl_check_extension(struct vfio_container *container,
708 unsigned long arg)
709 {
710 struct vfio_iommu_driver *driver;
711 long ret = 0;
712
713 down_read(&container->group_lock);
714
715 driver = container->iommu_driver;
716
717 switch (arg) {
718 /* No base extensions yet */
719 default:
720 /*
721 * If no driver is set, poll all registered drivers for
722 * extensions and return the first positive result. If
723 * a driver is already set, further queries will be passed
724 * only to that driver.
725 */
726 if (!driver) {
727 mutex_lock(&vfio.iommu_drivers_lock);
728 list_for_each_entry(driver, &vfio.iommu_drivers_list,
729 vfio_next) {
730 if (!try_module_get(driver->ops->owner))
731 continue;
732
733 ret = driver->ops->ioctl(NULL,
734 VFIO_CHECK_EXTENSION,
735 arg);
736 module_put(driver->ops->owner);
737 if (ret > 0)
738 break;
739 }
740 mutex_unlock(&vfio.iommu_drivers_lock);
741 } else
742 ret = driver->ops->ioctl(container->iommu_data,
743 VFIO_CHECK_EXTENSION, arg);
744 }
745
746 up_read(&container->group_lock);
747
748 return ret;
749 }
750
751 /* hold write lock on container->group_lock */
752 static int __vfio_container_attach_groups(struct vfio_container *container,
753 struct vfio_iommu_driver *driver,
754 void *data)
755 {
756 struct vfio_group *group;
757 int ret = -ENODEV;
758
759 list_for_each_entry(group, &container->group_list, container_next) {
760 ret = driver->ops->attach_group(data, group->iommu_group);
761 if (ret)
762 goto unwind;
763 }
764
765 return ret;
766
767 unwind:
768 list_for_each_entry_continue_reverse(group, &container->group_list,
769 container_next) {
770 driver->ops->detach_group(data, group->iommu_group);
771 }
772
773 return ret;
774 }
775
776 static long vfio_ioctl_set_iommu(struct vfio_container *container,
777 unsigned long arg)
778 {
779 struct vfio_iommu_driver *driver;
780 long ret = -ENODEV;
781
782 down_write(&container->group_lock);
783
784 /*
785 * The container is designed to be an unprivileged interface while
786 * the group can be assigned to specific users. Therefore, only by
787 * adding a group to a container does the user get the privilege of
788 * enabling the iommu, which may allocate finite resources. There
789 * is no unset_iommu, but by removing all the groups from a container,
790 * the container is deprivileged and returns to an unset state.
791 */
792 if (list_empty(&container->group_list) || container->iommu_driver) {
793 up_write(&container->group_lock);
794 return -EINVAL;
795 }
796
797 mutex_lock(&vfio.iommu_drivers_lock);
798 list_for_each_entry(driver, &vfio.iommu_drivers_list, vfio_next) {
799 void *data;
800
801 if (!try_module_get(driver->ops->owner))
802 continue;
803
804 /*
805 * The arg magic for SET_IOMMU is the same as CHECK_EXTENSION,
806 * so test which iommu driver reported support for this
807 * extension and call open on them. We also pass them the
808 * magic, allowing a single driver to support multiple
809 * interfaces if they'd like.
810 */
811 if (driver->ops->ioctl(NULL, VFIO_CHECK_EXTENSION, arg) <= 0) {
812 module_put(driver->ops->owner);
813 continue;
814 }
815
816 /* module reference holds the driver we're working on */
817 mutex_unlock(&vfio.iommu_drivers_lock);
818
819 data = driver->ops->open(arg);
820 if (IS_ERR(data)) {
821 ret = PTR_ERR(data);
822 module_put(driver->ops->owner);
823 goto skip_drivers_unlock;
824 }
825
826 ret = __vfio_container_attach_groups(container, driver, data);
827 if (!ret) {
828 container->iommu_driver = driver;
829 container->iommu_data = data;
830 } else {
831 driver->ops->release(data);
832 module_put(driver->ops->owner);
833 }
834
835 goto skip_drivers_unlock;
836 }
837
838 mutex_unlock(&vfio.iommu_drivers_lock);
839 skip_drivers_unlock:
840 up_write(&container->group_lock);
841
842 return ret;
843 }
844
845 static long vfio_fops_unl_ioctl(struct file *filep,
846 unsigned int cmd, unsigned long arg)
847 {
848 struct vfio_container *container = filep->private_data;
849 struct vfio_iommu_driver *driver;
850 void *data;
851 long ret = -EINVAL;
852
853 if (!container)
854 return ret;
855
856 switch (cmd) {
857 case VFIO_GET_API_VERSION:
858 ret = VFIO_API_VERSION;
859 break;
860 case VFIO_CHECK_EXTENSION:
861 ret = vfio_ioctl_check_extension(container, arg);
862 break;
863 case VFIO_SET_IOMMU:
864 ret = vfio_ioctl_set_iommu(container, arg);
865 break;
866 default:
867 down_read(&container->group_lock);
868
869 driver = container->iommu_driver;
870 data = container->iommu_data;
871
872 if (driver) /* passthrough all unrecognized ioctls */
873 ret = driver->ops->ioctl(data, cmd, arg);
874
875 up_read(&container->group_lock);
876 }
877
878 return ret;
879 }
880
881 #ifdef CONFIG_COMPAT
882 static long vfio_fops_compat_ioctl(struct file *filep,
883 unsigned int cmd, unsigned long arg)
884 {
885 arg = (unsigned long)compat_ptr(arg);
886 return vfio_fops_unl_ioctl(filep, cmd, arg);
887 }
888 #endif /* CONFIG_COMPAT */
889
890 static int vfio_fops_open(struct inode *inode, struct file *filep)
891 {
892 struct vfio_container *container;
893
894 container = kzalloc(sizeof(*container), GFP_KERNEL);
895 if (!container)
896 return -ENOMEM;
897
898 INIT_LIST_HEAD(&container->group_list);
899 init_rwsem(&container->group_lock);
900 kref_init(&container->kref);
901
902 filep->private_data = container;
903
904 return 0;
905 }
906
907 static int vfio_fops_release(struct inode *inode, struct file *filep)
908 {
909 struct vfio_container *container = filep->private_data;
910
911 filep->private_data = NULL;
912
913 vfio_container_put(container);
914
915 return 0;
916 }
917
918 /*
919 * Once an iommu driver is set, we optionally pass read/write/mmap
920 * on to the driver, allowing management interfaces beyond ioctl.
921 */
922 static ssize_t vfio_fops_read(struct file *filep, char __user *buf,
923 size_t count, loff_t *ppos)
924 {
925 struct vfio_container *container = filep->private_data;
926 struct vfio_iommu_driver *driver;
927 ssize_t ret = -EINVAL;
928
929 down_read(&container->group_lock);
930
931 driver = container->iommu_driver;
932 if (likely(driver && driver->ops->read))
933 ret = driver->ops->read(container->iommu_data,
934 buf, count, ppos);
935
936 up_read(&container->group_lock);
937
938 return ret;
939 }
940
941 static ssize_t vfio_fops_write(struct file *filep, const char __user *buf,
942 size_t count, loff_t *ppos)
943 {
944 struct vfio_container *container = filep->private_data;
945 struct vfio_iommu_driver *driver;
946 ssize_t ret = -EINVAL;
947
948 down_read(&container->group_lock);
949
950 driver = container->iommu_driver;
951 if (likely(driver && driver->ops->write))
952 ret = driver->ops->write(container->iommu_data,
953 buf, count, ppos);
954
955 up_read(&container->group_lock);
956
957 return ret;
958 }
959
960 static int vfio_fops_mmap(struct file *filep, struct vm_area_struct *vma)
961 {
962 struct vfio_container *container = filep->private_data;
963 struct vfio_iommu_driver *driver;
964 int ret = -EINVAL;
965
966 down_read(&container->group_lock);
967
968 driver = container->iommu_driver;
969 if (likely(driver && driver->ops->mmap))
970 ret = driver->ops->mmap(container->iommu_data, vma);
971
972 up_read(&container->group_lock);
973
974 return ret;
975 }
976
977 static const struct file_operations vfio_fops = {
978 .owner = THIS_MODULE,
979 .open = vfio_fops_open,
980 .release = vfio_fops_release,
981 .read = vfio_fops_read,
982 .write = vfio_fops_write,
983 .unlocked_ioctl = vfio_fops_unl_ioctl,
984 #ifdef CONFIG_COMPAT
985 .compat_ioctl = vfio_fops_compat_ioctl,
986 #endif
987 .mmap = vfio_fops_mmap,
988 };
989
990 /**
991 * VFIO Group fd, /dev/vfio/$GROUP
992 */
993 static void __vfio_group_unset_container(struct vfio_group *group)
994 {
995 struct vfio_container *container = group->container;
996 struct vfio_iommu_driver *driver;
997
998 down_write(&container->group_lock);
999
1000 driver = container->iommu_driver;
1001 if (driver)
1002 driver->ops->detach_group(container->iommu_data,
1003 group->iommu_group);
1004
1005 group->container = NULL;
1006 list_del(&group->container_next);
1007
1008 /* Detaching the last group deprivileges a container, remove iommu */
1009 if (driver && list_empty(&container->group_list)) {
1010 driver->ops->release(container->iommu_data);
1011 module_put(driver->ops->owner);
1012 container->iommu_driver = NULL;
1013 container->iommu_data = NULL;
1014 }
1015
1016 up_write(&container->group_lock);
1017
1018 vfio_container_put(container);
1019 }
1020
1021 /*
1022 * VFIO_GROUP_UNSET_CONTAINER should fail if there are other users or
1023 * if there was no container to unset. Since the ioctl is called on
1024 * the group, we know that still exists, therefore the only valid
1025 * transition here is 1->0.
1026 */
1027 static int vfio_group_unset_container(struct vfio_group *group)
1028 {
1029 int users = atomic_cmpxchg(&group->container_users, 1, 0);
1030
1031 if (!users)
1032 return -EINVAL;
1033 if (users != 1)
1034 return -EBUSY;
1035
1036 __vfio_group_unset_container(group);
1037
1038 return 0;
1039 }
1040
1041 /*
1042 * When removing container users, anything that removes the last user
1043 * implicitly removes the group from the container. That is, if the
1044 * group file descriptor is closed, as well as any device file descriptors,
1045 * the group is free.
1046 */
1047 static void vfio_group_try_dissolve_container(struct vfio_group *group)
1048 {
1049 if (0 == atomic_dec_if_positive(&group->container_users))
1050 __vfio_group_unset_container(group);
1051 }
1052
1053 static int vfio_group_set_container(struct vfio_group *group, int container_fd)
1054 {
1055 struct fd f;
1056 struct vfio_container *container;
1057 struct vfio_iommu_driver *driver;
1058 int ret = 0;
1059
1060 if (atomic_read(&group->container_users))
1061 return -EINVAL;
1062
1063 f = fdget(container_fd);
1064 if (!f.file)
1065 return -EBADF;
1066
1067 /* Sanity check, is this really our fd? */
1068 if (f.file->f_op != &vfio_fops) {
1069 fdput(f);
1070 return -EINVAL;
1071 }
1072
1073 container = f.file->private_data;
1074 WARN_ON(!container); /* fget ensures we don't race vfio_release */
1075
1076 down_write(&container->group_lock);
1077
1078 driver = container->iommu_driver;
1079 if (driver) {
1080 ret = driver->ops->attach_group(container->iommu_data,
1081 group->iommu_group);
1082 if (ret)
1083 goto unlock_out;
1084 }
1085
1086 group->container = container;
1087 list_add(&group->container_next, &container->group_list);
1088
1089 /* Get a reference on the container and mark a user within the group */
1090 vfio_container_get(container);
1091 atomic_inc(&group->container_users);
1092
1093 unlock_out:
1094 up_write(&container->group_lock);
1095 fdput(f);
1096 return ret;
1097 }
1098
1099 static bool vfio_group_viable(struct vfio_group *group)
1100 {
1101 return (iommu_group_for_each_dev(group->iommu_group,
1102 group, vfio_dev_viable) == 0);
1103 }
1104
1105 static const struct file_operations vfio_device_fops;
1106
1107 static int vfio_group_get_device_fd(struct vfio_group *group, char *buf)
1108 {
1109 struct vfio_device *device;
1110 struct file *filep;
1111 int ret = -ENODEV;
1112
1113 if (0 == atomic_read(&group->container_users) ||
1114 !group->container->iommu_driver || !vfio_group_viable(group))
1115 return -EINVAL;
1116
1117 mutex_lock(&group->device_lock);
1118 list_for_each_entry(device, &group->device_list, group_next) {
1119 if (strcmp(dev_name(device->dev), buf))
1120 continue;
1121
1122 ret = device->ops->open(device->device_data);
1123 if (ret)
1124 break;
1125 /*
1126 * We can't use anon_inode_getfd() because we need to modify
1127 * the f_mode flags directly to allow more than just ioctls
1128 */
1129 ret = get_unused_fd();
1130 if (ret < 0) {
1131 device->ops->release(device->device_data);
1132 break;
1133 }
1134
1135 filep = anon_inode_getfile("[vfio-device]", &vfio_device_fops,
1136 device, O_RDWR);
1137 if (IS_ERR(filep)) {
1138 put_unused_fd(ret);
1139 ret = PTR_ERR(filep);
1140 device->ops->release(device->device_data);
1141 break;
1142 }
1143
1144 /*
1145 * TODO: add an anon_inode interface to do this.
1146 * Appears to be missing by lack of need rather than
1147 * explicitly prevented. Now there's need.
1148 */
1149 filep->f_mode |= (FMODE_LSEEK | FMODE_PREAD | FMODE_PWRITE);
1150
1151 vfio_device_get(device);
1152 atomic_inc(&group->container_users);
1153
1154 fd_install(ret, filep);
1155 break;
1156 }
1157 mutex_unlock(&group->device_lock);
1158
1159 return ret;
1160 }
1161
1162 static long vfio_group_fops_unl_ioctl(struct file *filep,
1163 unsigned int cmd, unsigned long arg)
1164 {
1165 struct vfio_group *group = filep->private_data;
1166 long ret = -ENOTTY;
1167
1168 switch (cmd) {
1169 case VFIO_GROUP_GET_STATUS:
1170 {
1171 struct vfio_group_status status;
1172 unsigned long minsz;
1173
1174 minsz = offsetofend(struct vfio_group_status, flags);
1175
1176 if (copy_from_user(&status, (void __user *)arg, minsz))
1177 return -EFAULT;
1178
1179 if (status.argsz < minsz)
1180 return -EINVAL;
1181
1182 status.flags = 0;
1183
1184 if (vfio_group_viable(group))
1185 status.flags |= VFIO_GROUP_FLAGS_VIABLE;
1186
1187 if (group->container)
1188 status.flags |= VFIO_GROUP_FLAGS_CONTAINER_SET;
1189
1190 if (copy_to_user((void __user *)arg, &status, minsz))
1191 return -EFAULT;
1192
1193 ret = 0;
1194 break;
1195 }
1196 case VFIO_GROUP_SET_CONTAINER:
1197 {
1198 int fd;
1199
1200 if (get_user(fd, (int __user *)arg))
1201 return -EFAULT;
1202
1203 if (fd < 0)
1204 return -EINVAL;
1205
1206 ret = vfio_group_set_container(group, fd);
1207 break;
1208 }
1209 case VFIO_GROUP_UNSET_CONTAINER:
1210 ret = vfio_group_unset_container(group);
1211 break;
1212 case VFIO_GROUP_GET_DEVICE_FD:
1213 {
1214 char *buf;
1215
1216 buf = strndup_user((const char __user *)arg, PAGE_SIZE);
1217 if (IS_ERR(buf))
1218 return PTR_ERR(buf);
1219
1220 ret = vfio_group_get_device_fd(group, buf);
1221 kfree(buf);
1222 break;
1223 }
1224 }
1225
1226 return ret;
1227 }
1228
1229 #ifdef CONFIG_COMPAT
1230 static long vfio_group_fops_compat_ioctl(struct file *filep,
1231 unsigned int cmd, unsigned long arg)
1232 {
1233 arg = (unsigned long)compat_ptr(arg);
1234 return vfio_group_fops_unl_ioctl(filep, cmd, arg);
1235 }
1236 #endif /* CONFIG_COMPAT */
1237
1238 static int vfio_group_fops_open(struct inode *inode, struct file *filep)
1239 {
1240 struct vfio_group *group;
1241 int opened;
1242
1243 group = vfio_group_get_from_minor(iminor(inode));
1244 if (!group)
1245 return -ENODEV;
1246
1247 /* Do we need multiple instances of the group open? Seems not. */
1248 opened = atomic_cmpxchg(&group->opened, 0, 1);
1249 if (opened) {
1250 vfio_group_put(group);
1251 return -EBUSY;
1252 }
1253
1254 /* Is something still in use from a previous open? */
1255 if (group->container) {
1256 atomic_dec(&group->opened);
1257 vfio_group_put(group);
1258 return -EBUSY;
1259 }
1260
1261 filep->private_data = group;
1262
1263 return 0;
1264 }
1265
1266 static int vfio_group_fops_release(struct inode *inode, struct file *filep)
1267 {
1268 struct vfio_group *group = filep->private_data;
1269
1270 filep->private_data = NULL;
1271
1272 vfio_group_try_dissolve_container(group);
1273
1274 atomic_dec(&group->opened);
1275
1276 vfio_group_put(group);
1277
1278 return 0;
1279 }
1280
1281 static const struct file_operations vfio_group_fops = {
1282 .owner = THIS_MODULE,
1283 .unlocked_ioctl = vfio_group_fops_unl_ioctl,
1284 #ifdef CONFIG_COMPAT
1285 .compat_ioctl = vfio_group_fops_compat_ioctl,
1286 #endif
1287 .open = vfio_group_fops_open,
1288 .release = vfio_group_fops_release,
1289 };
1290
1291 /**
1292 * VFIO Device fd
1293 */
1294 static int vfio_device_fops_release(struct inode *inode, struct file *filep)
1295 {
1296 struct vfio_device *device = filep->private_data;
1297
1298 device->ops->release(device->device_data);
1299
1300 vfio_group_try_dissolve_container(device->group);
1301
1302 vfio_device_put(device);
1303
1304 return 0;
1305 }
1306
1307 static long vfio_device_fops_unl_ioctl(struct file *filep,
1308 unsigned int cmd, unsigned long arg)
1309 {
1310 struct vfio_device *device = filep->private_data;
1311
1312 if (unlikely(!device->ops->ioctl))
1313 return -EINVAL;
1314
1315 return device->ops->ioctl(device->device_data, cmd, arg);
1316 }
1317
1318 static ssize_t vfio_device_fops_read(struct file *filep, char __user *buf,
1319 size_t count, loff_t *ppos)
1320 {
1321 struct vfio_device *device = filep->private_data;
1322
1323 if (unlikely(!device->ops->read))
1324 return -EINVAL;
1325
1326 return device->ops->read(device->device_data, buf, count, ppos);
1327 }
1328
1329 static ssize_t vfio_device_fops_write(struct file *filep,
1330 const char __user *buf,
1331 size_t count, loff_t *ppos)
1332 {
1333 struct vfio_device *device = filep->private_data;
1334
1335 if (unlikely(!device->ops->write))
1336 return -EINVAL;
1337
1338 return device->ops->write(device->device_data, buf, count, ppos);
1339 }
1340
1341 static int vfio_device_fops_mmap(struct file *filep, struct vm_area_struct *vma)
1342 {
1343 struct vfio_device *device = filep->private_data;
1344
1345 if (unlikely(!device->ops->mmap))
1346 return -EINVAL;
1347
1348 return device->ops->mmap(device->device_data, vma);
1349 }
1350
1351 #ifdef CONFIG_COMPAT
1352 static long vfio_device_fops_compat_ioctl(struct file *filep,
1353 unsigned int cmd, unsigned long arg)
1354 {
1355 arg = (unsigned long)compat_ptr(arg);
1356 return vfio_device_fops_unl_ioctl(filep, cmd, arg);
1357 }
1358 #endif /* CONFIG_COMPAT */
1359
1360 static const struct file_operations vfio_device_fops = {
1361 .owner = THIS_MODULE,
1362 .release = vfio_device_fops_release,
1363 .read = vfio_device_fops_read,
1364 .write = vfio_device_fops_write,
1365 .unlocked_ioctl = vfio_device_fops_unl_ioctl,
1366 #ifdef CONFIG_COMPAT
1367 .compat_ioctl = vfio_device_fops_compat_ioctl,
1368 #endif
1369 .mmap = vfio_device_fops_mmap,
1370 };
1371
1372 /**
1373 * Module/class support
1374 */
1375 static char *vfio_devnode(struct device *dev, umode_t *mode)
1376 {
1377 if (mode && (MINOR(dev->devt) == 0))
1378 *mode = S_IRUGO | S_IWUGO;
1379
1380 return kasprintf(GFP_KERNEL, "vfio/%s", dev_name(dev));
1381 }
1382
1383 static int __init vfio_init(void)
1384 {
1385 int ret;
1386
1387 idr_init(&vfio.group_idr);
1388 mutex_init(&vfio.group_lock);
1389 mutex_init(&vfio.iommu_drivers_lock);
1390 INIT_LIST_HEAD(&vfio.group_list);
1391 INIT_LIST_HEAD(&vfio.iommu_drivers_list);
1392 init_waitqueue_head(&vfio.release_q);
1393
1394 vfio.class = class_create(THIS_MODULE, "vfio");
1395 if (IS_ERR(vfio.class)) {
1396 ret = PTR_ERR(vfio.class);
1397 goto err_class;
1398 }
1399
1400 vfio.class->devnode = vfio_devnode;
1401
1402 ret = alloc_chrdev_region(&vfio.devt, 0, MINORMASK, "vfio");
1403 if (ret)
1404 goto err_base_chrdev;
1405
1406 cdev_init(&vfio.cdev, &vfio_fops);
1407 ret = cdev_add(&vfio.cdev, vfio.devt, 1);
1408 if (ret)
1409 goto err_base_cdev;
1410
1411 vfio.dev = device_create(vfio.class, NULL, vfio.devt, NULL, "vfio");
1412 if (IS_ERR(vfio.dev)) {
1413 ret = PTR_ERR(vfio.dev);
1414 goto err_base_dev;
1415 }
1416
1417 /* /dev/vfio/$GROUP */
1418 cdev_init(&vfio.group_cdev, &vfio_group_fops);
1419 ret = cdev_add(&vfio.group_cdev,
1420 MKDEV(MAJOR(vfio.devt), 1), MINORMASK - 1);
1421 if (ret)
1422 goto err_groups_cdev;
1423
1424 pr_info(DRIVER_DESC " version: " DRIVER_VERSION "\n");
1425
1426 /*
1427 * Attempt to load known iommu-drivers. This gives us a working
1428 * environment without the user needing to explicitly load iommu
1429 * drivers.
1430 */
1431 request_module_nowait("vfio_iommu_type1");
1432 request_module_nowait("vfio_iommu_spapr_tce");
1433
1434 return 0;
1435
1436 err_groups_cdev:
1437 device_destroy(vfio.class, vfio.devt);
1438 err_base_dev:
1439 cdev_del(&vfio.cdev);
1440 err_base_cdev:
1441 unregister_chrdev_region(vfio.devt, MINORMASK);
1442 err_base_chrdev:
1443 class_destroy(vfio.class);
1444 vfio.class = NULL;
1445 err_class:
1446 return ret;
1447 }
1448
1449 static void __exit vfio_cleanup(void)
1450 {
1451 WARN_ON(!list_empty(&vfio.group_list));
1452
1453 idr_destroy(&vfio.group_idr);
1454 cdev_del(&vfio.group_cdev);
1455 device_destroy(vfio.class, vfio.devt);
1456 cdev_del(&vfio.cdev);
1457 unregister_chrdev_region(vfio.devt, MINORMASK);
1458 class_destroy(vfio.class);
1459 vfio.class = NULL;
1460 }
1461
1462 module_init(vfio_init);
1463 module_exit(vfio_cleanup);
1464
1465 MODULE_VERSION(DRIVER_VERSION);
1466 MODULE_LICENSE("GPL v2");
1467 MODULE_AUTHOR(DRIVER_AUTHOR);
1468 MODULE_DESCRIPTION(DRIVER_DESC);
This page took 0.083594 seconds and 5 git commands to generate.