blob: c488da5db7c7f1b0bda8e3b7559971ebfba0c442 [file] [log] [blame]
Alex Williamsoncba33452012-07-31 08:16:22 -06001/*
2 * VFIO core
3 *
4 * Copyright (C) 2012 Red Hat, Inc. All rights reserved.
5 * Author: Alex Williamson <alex.williamson@redhat.com>
6 *
7 * This program is free software; you can redistribute it and/or modify
8 * it under the terms of the GNU General Public License version 2 as
9 * published by the Free Software Foundation.
10 *
11 * Derived from original vfio:
12 * Copyright 2010 Cisco Systems, Inc. All rights reserved.
13 * Author: Tom Lyon, pugs@cisco.com
14 */
15
16#include <linux/cdev.h>
17#include <linux/compat.h>
18#include <linux/device.h>
19#include <linux/file.h>
20#include <linux/anon_inodes.h>
21#include <linux/fs.h>
22#include <linux/idr.h>
23#include <linux/iommu.h>
24#include <linux/list.h>
25#include <linux/module.h>
26#include <linux/mutex.h>
Alex Williamson9587f442013-04-25 16:12:38 -060027#include <linux/rwsem.h>
Alex Williamsoncba33452012-07-31 08:16:22 -060028#include <linux/sched.h>
29#include <linux/slab.h>
Alex Williamson664e9382013-04-30 15:42:28 -060030#include <linux/stat.h>
Alex Williamsoncba33452012-07-31 08:16:22 -060031#include <linux/string.h>
32#include <linux/uaccess.h>
33#include <linux/vfio.h>
34#include <linux/wait.h>
35
36#define DRIVER_VERSION "0.3"
37#define DRIVER_AUTHOR "Alex Williamson <alex.williamson@redhat.com>"
38#define DRIVER_DESC "VFIO - User Level meta-driver"
39
40static struct vfio {
41 struct class *class;
42 struct list_head iommu_drivers_list;
43 struct mutex iommu_drivers_lock;
44 struct list_head group_list;
45 struct idr group_idr;
46 struct mutex group_lock;
47 struct cdev group_cdev;
48 struct device *dev;
49 dev_t devt;
50 struct cdev cdev;
51 wait_queue_head_t release_q;
52} vfio;
53
54struct vfio_iommu_driver {
55 const struct vfio_iommu_driver_ops *ops;
56 struct list_head vfio_next;
57};
58
59struct vfio_container {
60 struct kref kref;
61 struct list_head group_list;
Alex Williamson9587f442013-04-25 16:12:38 -060062 struct rw_semaphore group_lock;
Alex Williamsoncba33452012-07-31 08:16:22 -060063 struct vfio_iommu_driver *iommu_driver;
64 void *iommu_data;
65};
66
67struct vfio_group {
68 struct kref kref;
69 int minor;
70 atomic_t container_users;
71 struct iommu_group *iommu_group;
72 struct vfio_container *container;
73 struct list_head device_list;
74 struct mutex device_lock;
75 struct device *dev;
76 struct notifier_block nb;
77 struct list_head vfio_next;
78 struct list_head container_next;
Alex Williamson6d6768c2013-06-25 16:06:54 -060079 atomic_t opened;
Alex Williamsoncba33452012-07-31 08:16:22 -060080};
81
82struct vfio_device {
83 struct kref kref;
84 struct device *dev;
85 const struct vfio_device_ops *ops;
86 struct vfio_group *group;
87 struct list_head group_next;
88 void *device_data;
89};
90
91/**
92 * IOMMU driver registration
93 */
94int vfio_register_iommu_driver(const struct vfio_iommu_driver_ops *ops)
95{
96 struct vfio_iommu_driver *driver, *tmp;
97
98 driver = kzalloc(sizeof(*driver), GFP_KERNEL);
99 if (!driver)
100 return -ENOMEM;
101
102 driver->ops = ops;
103
104 mutex_lock(&vfio.iommu_drivers_lock);
105
106 /* Check for duplicates */
107 list_for_each_entry(tmp, &vfio.iommu_drivers_list, vfio_next) {
108 if (tmp->ops == ops) {
109 mutex_unlock(&vfio.iommu_drivers_lock);
110 kfree(driver);
111 return -EINVAL;
112 }
113 }
114
115 list_add(&driver->vfio_next, &vfio.iommu_drivers_list);
116
117 mutex_unlock(&vfio.iommu_drivers_lock);
118
119 return 0;
120}
121EXPORT_SYMBOL_GPL(vfio_register_iommu_driver);
122
123void vfio_unregister_iommu_driver(const struct vfio_iommu_driver_ops *ops)
124{
125 struct vfio_iommu_driver *driver;
126
127 mutex_lock(&vfio.iommu_drivers_lock);
128 list_for_each_entry(driver, &vfio.iommu_drivers_list, vfio_next) {
129 if (driver->ops == ops) {
130 list_del(&driver->vfio_next);
131 mutex_unlock(&vfio.iommu_drivers_lock);
132 kfree(driver);
133 return;
134 }
135 }
136 mutex_unlock(&vfio.iommu_drivers_lock);
137}
138EXPORT_SYMBOL_GPL(vfio_unregister_iommu_driver);
139
140/**
141 * Group minor allocation/free - both called with vfio.group_lock held
142 */
143static int vfio_alloc_group_minor(struct vfio_group *group)
144{
Alex Williamsoncba33452012-07-31 08:16:22 -0600145 /* index 0 is used by /dev/vfio/vfio */
Tejun Heoa1c36b12013-02-27 17:04:48 -0800146 return idr_alloc(&vfio.group_idr, group, 1, MINORMASK + 1, GFP_KERNEL);
Alex Williamsoncba33452012-07-31 08:16:22 -0600147}
148
149static void vfio_free_group_minor(int minor)
150{
151 idr_remove(&vfio.group_idr, minor);
152}
153
154static int vfio_iommu_group_notifier(struct notifier_block *nb,
155 unsigned long action, void *data);
156static void vfio_group_get(struct vfio_group *group);
157
158/**
159 * Container objects - containers are created when /dev/vfio/vfio is
160 * opened, but their lifecycle extends until the last user is done, so
161 * it's freed via kref. Must support container/group/device being
162 * closed in any order.
163 */
164static void vfio_container_get(struct vfio_container *container)
165{
166 kref_get(&container->kref);
167}
168
169static void vfio_container_release(struct kref *kref)
170{
171 struct vfio_container *container;
172 container = container_of(kref, struct vfio_container, kref);
173
174 kfree(container);
175}
176
177static void vfio_container_put(struct vfio_container *container)
178{
179 kref_put(&container->kref, vfio_container_release);
180}
181
Jiang Liu9df7b252012-12-07 13:43:50 -0700182static void vfio_group_unlock_and_free(struct vfio_group *group)
183{
184 mutex_unlock(&vfio.group_lock);
185 /*
186 * Unregister outside of lock. A spurious callback is harmless now
187 * that the group is no longer in vfio.group_list.
188 */
189 iommu_group_unregister_notifier(group->iommu_group, &group->nb);
190 kfree(group);
191}
192
Alex Williamsoncba33452012-07-31 08:16:22 -0600193/**
194 * Group objects - create, release, get, put, search
195 */
196static struct vfio_group *vfio_create_group(struct iommu_group *iommu_group)
197{
198 struct vfio_group *group, *tmp;
199 struct device *dev;
200 int ret, minor;
201
202 group = kzalloc(sizeof(*group), GFP_KERNEL);
203 if (!group)
204 return ERR_PTR(-ENOMEM);
205
206 kref_init(&group->kref);
207 INIT_LIST_HEAD(&group->device_list);
208 mutex_init(&group->device_lock);
209 atomic_set(&group->container_users, 0);
Alex Williamson6d6768c2013-06-25 16:06:54 -0600210 atomic_set(&group->opened, 0);
Alex Williamsoncba33452012-07-31 08:16:22 -0600211 group->iommu_group = iommu_group;
212
213 group->nb.notifier_call = vfio_iommu_group_notifier;
214
215 /*
216 * blocking notifiers acquire a rwsem around registering and hold
217 * it around callback. Therefore, need to register outside of
218 * vfio.group_lock to avoid A-B/B-A contention. Our callback won't
219 * do anything unless it can find the group in vfio.group_list, so
220 * no harm in registering early.
221 */
222 ret = iommu_group_register_notifier(iommu_group, &group->nb);
223 if (ret) {
224 kfree(group);
225 return ERR_PTR(ret);
226 }
227
228 mutex_lock(&vfio.group_lock);
229
230 minor = vfio_alloc_group_minor(group);
231 if (minor < 0) {
Jiang Liu9df7b252012-12-07 13:43:50 -0700232 vfio_group_unlock_and_free(group);
Alex Williamsoncba33452012-07-31 08:16:22 -0600233 return ERR_PTR(minor);
234 }
235
236 /* Did we race creating this group? */
237 list_for_each_entry(tmp, &vfio.group_list, vfio_next) {
238 if (tmp->iommu_group == iommu_group) {
239 vfio_group_get(tmp);
240 vfio_free_group_minor(minor);
Jiang Liu9df7b252012-12-07 13:43:50 -0700241 vfio_group_unlock_and_free(group);
Alex Williamsoncba33452012-07-31 08:16:22 -0600242 return tmp;
243 }
244 }
245
246 dev = device_create(vfio.class, NULL, MKDEV(MAJOR(vfio.devt), minor),
247 group, "%d", iommu_group_id(iommu_group));
248 if (IS_ERR(dev)) {
249 vfio_free_group_minor(minor);
Jiang Liu9df7b252012-12-07 13:43:50 -0700250 vfio_group_unlock_and_free(group);
Alex Williamsoncba33452012-07-31 08:16:22 -0600251 return (struct vfio_group *)dev; /* ERR_PTR */
252 }
253
254 group->minor = minor;
255 group->dev = dev;
256
257 list_add(&group->vfio_next, &vfio.group_list);
258
259 mutex_unlock(&vfio.group_lock);
260
261 return group;
262}
263
Al Viro6d2cd3c2012-08-17 21:27:32 -0400264/* called with vfio.group_lock held */
Alex Williamsoncba33452012-07-31 08:16:22 -0600265static void vfio_group_release(struct kref *kref)
266{
267 struct vfio_group *group = container_of(kref, struct vfio_group, kref);
268
269 WARN_ON(!list_empty(&group->device_list));
270
271 device_destroy(vfio.class, MKDEV(MAJOR(vfio.devt), group->minor));
272 list_del(&group->vfio_next);
273 vfio_free_group_minor(group->minor);
Jiang Liu9df7b252012-12-07 13:43:50 -0700274 vfio_group_unlock_and_free(group);
Alex Williamsoncba33452012-07-31 08:16:22 -0600275}
276
277static void vfio_group_put(struct vfio_group *group)
278{
Al Viro6d2cd3c2012-08-17 21:27:32 -0400279 kref_put_mutex(&group->kref, vfio_group_release, &vfio.group_lock);
Alex Williamsoncba33452012-07-31 08:16:22 -0600280}
281
282/* Assume group_lock or group reference is held */
283static void vfio_group_get(struct vfio_group *group)
284{
285 kref_get(&group->kref);
286}
287
288/*
289 * Not really a try as we will sleep for mutex, but we need to make
290 * sure the group pointer is valid under lock and get a reference.
291 */
292static struct vfio_group *vfio_group_try_get(struct vfio_group *group)
293{
294 struct vfio_group *target = group;
295
296 mutex_lock(&vfio.group_lock);
297 list_for_each_entry(group, &vfio.group_list, vfio_next) {
298 if (group == target) {
299 vfio_group_get(group);
300 mutex_unlock(&vfio.group_lock);
301 return group;
302 }
303 }
304 mutex_unlock(&vfio.group_lock);
305
306 return NULL;
307}
308
309static
310struct vfio_group *vfio_group_get_from_iommu(struct iommu_group *iommu_group)
311{
312 struct vfio_group *group;
313
314 mutex_lock(&vfio.group_lock);
315 list_for_each_entry(group, &vfio.group_list, vfio_next) {
316 if (group->iommu_group == iommu_group) {
317 vfio_group_get(group);
318 mutex_unlock(&vfio.group_lock);
319 return group;
320 }
321 }
322 mutex_unlock(&vfio.group_lock);
323
324 return NULL;
325}
326
327static struct vfio_group *vfio_group_get_from_minor(int minor)
328{
329 struct vfio_group *group;
330
331 mutex_lock(&vfio.group_lock);
332 group = idr_find(&vfio.group_idr, minor);
333 if (!group) {
334 mutex_unlock(&vfio.group_lock);
335 return NULL;
336 }
337 vfio_group_get(group);
338 mutex_unlock(&vfio.group_lock);
339
340 return group;
341}
342
343/**
344 * Device objects - create, release, get, put, search
345 */
346static
347struct vfio_device *vfio_group_create_device(struct vfio_group *group,
348 struct device *dev,
349 const struct vfio_device_ops *ops,
350 void *device_data)
351{
352 struct vfio_device *device;
353 int ret;
354
355 device = kzalloc(sizeof(*device), GFP_KERNEL);
356 if (!device)
357 return ERR_PTR(-ENOMEM);
358
359 kref_init(&device->kref);
360 device->dev = dev;
361 device->group = group;
362 device->ops = ops;
363 device->device_data = device_data;
364
365 ret = dev_set_drvdata(dev, device);
366 if (ret) {
367 kfree(device);
368 return ERR_PTR(ret);
369 }
370
371 /* No need to get group_lock, caller has group reference */
372 vfio_group_get(group);
373
374 mutex_lock(&group->device_lock);
375 list_add(&device->group_next, &group->device_list);
376 mutex_unlock(&group->device_lock);
377
378 return device;
379}
380
381static void vfio_device_release(struct kref *kref)
382{
383 struct vfio_device *device = container_of(kref,
384 struct vfio_device, kref);
385 struct vfio_group *group = device->group;
386
Alex Williamsoncba33452012-07-31 08:16:22 -0600387 list_del(&device->group_next);
388 mutex_unlock(&group->device_lock);
389
390 dev_set_drvdata(device->dev, NULL);
391
392 kfree(device);
393
394 /* vfio_del_group_dev may be waiting for this device */
395 wake_up(&vfio.release_q);
396}
397
398/* Device reference always implies a group reference */
Vijay Mohan Pandarathil44f50712013-03-11 09:28:44 -0600399void vfio_device_put(struct vfio_device *device)
Alex Williamsoncba33452012-07-31 08:16:22 -0600400{
Al Viro934ad4c2012-08-17 19:49:09 -0400401 struct vfio_group *group = device->group;
Al Viro90b12532012-08-17 21:29:06 -0400402 kref_put_mutex(&device->kref, vfio_device_release, &group->device_lock);
Al Viro934ad4c2012-08-17 19:49:09 -0400403 vfio_group_put(group);
Alex Williamsoncba33452012-07-31 08:16:22 -0600404}
Vijay Mohan Pandarathil44f50712013-03-11 09:28:44 -0600405EXPORT_SYMBOL_GPL(vfio_device_put);
Alex Williamsoncba33452012-07-31 08:16:22 -0600406
407static void vfio_device_get(struct vfio_device *device)
408{
409 vfio_group_get(device->group);
410 kref_get(&device->kref);
411}
412
413static struct vfio_device *vfio_group_get_device(struct vfio_group *group,
414 struct device *dev)
415{
416 struct vfio_device *device;
417
418 mutex_lock(&group->device_lock);
419 list_for_each_entry(device, &group->device_list, group_next) {
420 if (device->dev == dev) {
421 vfio_device_get(device);
422 mutex_unlock(&group->device_lock);
423 return device;
424 }
425 }
426 mutex_unlock(&group->device_lock);
427 return NULL;
428}
429
430/*
431 * Whitelist some drivers that we know are safe (no dma) or just sit on
432 * a device. It's not always practical to leave a device within a group
433 * driverless as it could get re-bound to something unsafe.
434 */
Alex Williamson2b489a42013-02-14 14:02:13 -0700435static const char * const vfio_driver_whitelist[] = { "pci-stub", "pcieport" };
Alex Williamsoncba33452012-07-31 08:16:22 -0600436
437static bool vfio_whitelisted_driver(struct device_driver *drv)
438{
439 int i;
440
441 for (i = 0; i < ARRAY_SIZE(vfio_driver_whitelist); i++) {
442 if (!strcmp(drv->name, vfio_driver_whitelist[i]))
443 return true;
444 }
445
446 return false;
447}
448
449/*
450 * A vfio group is viable for use by userspace if all devices are either
451 * driver-less or bound to a vfio or whitelisted driver. We test the
452 * latter by the existence of a struct vfio_device matching the dev.
453 */
454static int vfio_dev_viable(struct device *dev, void *data)
455{
456 struct vfio_group *group = data;
457 struct vfio_device *device;
Jiang Liude2b3ee2012-12-07 13:43:50 -0700458 struct device_driver *drv = ACCESS_ONCE(dev->driver);
Alex Williamsoncba33452012-07-31 08:16:22 -0600459
Jiang Liude2b3ee2012-12-07 13:43:50 -0700460 if (!drv || vfio_whitelisted_driver(drv))
Alex Williamsoncba33452012-07-31 08:16:22 -0600461 return 0;
462
463 device = vfio_group_get_device(group, dev);
464 if (device) {
465 vfio_device_put(device);
466 return 0;
467 }
468
469 return -EINVAL;
470}
471
472/**
473 * Async device support
474 */
475static int vfio_group_nb_add_dev(struct vfio_group *group, struct device *dev)
476{
477 struct vfio_device *device;
478
479 /* Do we already know about it? We shouldn't */
480 device = vfio_group_get_device(group, dev);
481 if (WARN_ON_ONCE(device)) {
482 vfio_device_put(device);
483 return 0;
484 }
485
486 /* Nothing to do for idle groups */
487 if (!atomic_read(&group->container_users))
488 return 0;
489
490 /* TODO Prevent device auto probing */
491 WARN("Device %s added to live group %d!\n", dev_name(dev),
492 iommu_group_id(group->iommu_group));
493
494 return 0;
495}
496
497static int vfio_group_nb_del_dev(struct vfio_group *group, struct device *dev)
498{
499 struct vfio_device *device;
500
501 /*
502 * Expect to fall out here. If a device was in use, it would
503 * have been bound to a vfio sub-driver, which would have blocked
504 * in .remove at vfio_del_group_dev. Sanity check that we no
505 * longer track the device, so it's safe to remove.
506 */
507 device = vfio_group_get_device(group, dev);
508 if (likely(!device))
509 return 0;
510
511 WARN("Device %s removed from live group %d!\n", dev_name(dev),
512 iommu_group_id(group->iommu_group));
513
514 vfio_device_put(device);
515 return 0;
516}
517
518static int vfio_group_nb_verify(struct vfio_group *group, struct device *dev)
519{
520 /* We don't care what happens when the group isn't in use */
521 if (!atomic_read(&group->container_users))
522 return 0;
523
524 return vfio_dev_viable(dev, group);
525}
526
527static int vfio_iommu_group_notifier(struct notifier_block *nb,
528 unsigned long action, void *data)
529{
530 struct vfio_group *group = container_of(nb, struct vfio_group, nb);
531 struct device *dev = data;
532
533 /*
534 * Need to go through a group_lock lookup to get a reference or
535 * we risk racing a group being removed. Leave a WARN_ON for
536 * debuging, but if the group no longer exists, a spurious notify
537 * is harmless.
538 */
539 group = vfio_group_try_get(group);
540 if (WARN_ON(!group))
541 return NOTIFY_OK;
542
543 switch (action) {
544 case IOMMU_GROUP_NOTIFY_ADD_DEVICE:
545 vfio_group_nb_add_dev(group, dev);
546 break;
547 case IOMMU_GROUP_NOTIFY_DEL_DEVICE:
548 vfio_group_nb_del_dev(group, dev);
549 break;
550 case IOMMU_GROUP_NOTIFY_BIND_DRIVER:
551 pr_debug("%s: Device %s, group %d binding to driver\n",
552 __func__, dev_name(dev),
553 iommu_group_id(group->iommu_group));
554 break;
555 case IOMMU_GROUP_NOTIFY_BOUND_DRIVER:
556 pr_debug("%s: Device %s, group %d bound to driver %s\n",
557 __func__, dev_name(dev),
558 iommu_group_id(group->iommu_group), dev->driver->name);
559 BUG_ON(vfio_group_nb_verify(group, dev));
560 break;
561 case IOMMU_GROUP_NOTIFY_UNBIND_DRIVER:
562 pr_debug("%s: Device %s, group %d unbinding from driver %s\n",
563 __func__, dev_name(dev),
564 iommu_group_id(group->iommu_group), dev->driver->name);
565 break;
566 case IOMMU_GROUP_NOTIFY_UNBOUND_DRIVER:
567 pr_debug("%s: Device %s, group %d unbound from driver\n",
568 __func__, dev_name(dev),
569 iommu_group_id(group->iommu_group));
570 /*
571 * XXX An unbound device in a live group is ok, but we'd
572 * really like to avoid the above BUG_ON by preventing other
573 * drivers from binding to it. Once that occurs, we have to
574 * stop the system to maintain isolation. At a minimum, we'd
575 * want a toggle to disable driver auto probe for this device.
576 */
577 break;
578 }
579
580 vfio_group_put(group);
581 return NOTIFY_OK;
582}
583
584/**
585 * VFIO driver API
586 */
587int vfio_add_group_dev(struct device *dev,
588 const struct vfio_device_ops *ops, void *device_data)
589{
590 struct iommu_group *iommu_group;
591 struct vfio_group *group;
592 struct vfio_device *device;
593
594 iommu_group = iommu_group_get(dev);
595 if (!iommu_group)
596 return -EINVAL;
597
598 group = vfio_group_get_from_iommu(iommu_group);
599 if (!group) {
600 group = vfio_create_group(iommu_group);
601 if (IS_ERR(group)) {
602 iommu_group_put(iommu_group);
603 return PTR_ERR(group);
604 }
605 }
606
607 device = vfio_group_get_device(group, dev);
608 if (device) {
609 WARN(1, "Device %s already exists on group %d\n",
610 dev_name(dev), iommu_group_id(iommu_group));
611 vfio_device_put(device);
612 vfio_group_put(group);
613 iommu_group_put(iommu_group);
614 return -EBUSY;
615 }
616
617 device = vfio_group_create_device(group, dev, ops, device_data);
618 if (IS_ERR(device)) {
619 vfio_group_put(group);
620 iommu_group_put(iommu_group);
621 return PTR_ERR(device);
622 }
623
624 /*
625 * Added device holds reference to iommu_group and vfio_device
626 * (which in turn holds reference to vfio_group). Drop extra
627 * group reference used while acquiring device.
628 */
629 vfio_group_put(group);
630
631 return 0;
632}
633EXPORT_SYMBOL_GPL(vfio_add_group_dev);
634
Vijay Mohan Pandarathil44f50712013-03-11 09:28:44 -0600635/**
636 * Get a reference to the vfio_device for a device that is known to
637 * be bound to a vfio driver. The driver implicitly holds a
638 * vfio_device reference between vfio_add_group_dev and
639 * vfio_del_group_dev. We can therefore use drvdata to increment
640 * that reference from the struct device. This additional
641 * reference must be released by calling vfio_device_put.
642 */
643struct vfio_device *vfio_device_get_from_dev(struct device *dev)
644{
645 struct vfio_device *device = dev_get_drvdata(dev);
646
647 vfio_device_get(device);
648
649 return device;
650}
651EXPORT_SYMBOL_GPL(vfio_device_get_from_dev);
652
653/*
654 * Caller must hold a reference to the vfio_device
655 */
656void *vfio_device_data(struct vfio_device *device)
657{
658 return device->device_data;
659}
660EXPORT_SYMBOL_GPL(vfio_device_data);
661
Alex Williamsone014e942013-02-14 14:02:13 -0700662/* Given a referenced group, check if it contains the device */
663static bool vfio_dev_present(struct vfio_group *group, struct device *dev)
Alex Williamsoncba33452012-07-31 08:16:22 -0600664{
Alex Williamsoncba33452012-07-31 08:16:22 -0600665 struct vfio_device *device;
666
Alex Williamsoncba33452012-07-31 08:16:22 -0600667 device = vfio_group_get_device(group, dev);
Alex Williamsone014e942013-02-14 14:02:13 -0700668 if (!device)
Alex Williamsoncba33452012-07-31 08:16:22 -0600669 return false;
Alex Williamsoncba33452012-07-31 08:16:22 -0600670
671 vfio_device_put(device);
Alex Williamsoncba33452012-07-31 08:16:22 -0600672 return true;
673}
674
675/*
676 * Decrement the device reference count and wait for the device to be
677 * removed. Open file descriptors for the device... */
678void *vfio_del_group_dev(struct device *dev)
679{
680 struct vfio_device *device = dev_get_drvdata(dev);
681 struct vfio_group *group = device->group;
682 struct iommu_group *iommu_group = group->iommu_group;
683 void *device_data = device->device_data;
684
Alex Williamsone014e942013-02-14 14:02:13 -0700685 /*
686 * The group exists so long as we have a device reference. Get
687 * a group reference and use it to scan for the device going away.
688 */
689 vfio_group_get(group);
690
Alex Williamsoncba33452012-07-31 08:16:22 -0600691 vfio_device_put(device);
692
693 /* TODO send a signal to encourage this to be released */
Alex Williamsone014e942013-02-14 14:02:13 -0700694 wait_event(vfio.release_q, !vfio_dev_present(group, dev));
695
696 vfio_group_put(group);
Alex Williamsoncba33452012-07-31 08:16:22 -0600697
698 iommu_group_put(iommu_group);
699
700 return device_data;
701}
702EXPORT_SYMBOL_GPL(vfio_del_group_dev);
703
704/**
705 * VFIO base fd, /dev/vfio/vfio
706 */
707static long vfio_ioctl_check_extension(struct vfio_container *container,
708 unsigned long arg)
709{
Alex Williamson0b43c082013-04-29 08:41:36 -0600710 struct vfio_iommu_driver *driver;
Alex Williamsoncba33452012-07-31 08:16:22 -0600711 long ret = 0;
712
Alex Williamson0b43c082013-04-29 08:41:36 -0600713 down_read(&container->group_lock);
714
715 driver = container->iommu_driver;
716
Alex Williamsoncba33452012-07-31 08:16:22 -0600717 switch (arg) {
718 /* No base extensions yet */
719 default:
720 /*
721 * If no driver is set, poll all registered drivers for
722 * extensions and return the first positive result. If
723 * a driver is already set, further queries will be passed
724 * only to that driver.
725 */
726 if (!driver) {
727 mutex_lock(&vfio.iommu_drivers_lock);
728 list_for_each_entry(driver, &vfio.iommu_drivers_list,
729 vfio_next) {
730 if (!try_module_get(driver->ops->owner))
731 continue;
732
733 ret = driver->ops->ioctl(NULL,
734 VFIO_CHECK_EXTENSION,
735 arg);
736 module_put(driver->ops->owner);
737 if (ret > 0)
738 break;
739 }
740 mutex_unlock(&vfio.iommu_drivers_lock);
741 } else
742 ret = driver->ops->ioctl(container->iommu_data,
743 VFIO_CHECK_EXTENSION, arg);
744 }
745
Alex Williamson0b43c082013-04-29 08:41:36 -0600746 up_read(&container->group_lock);
747
Alex Williamsoncba33452012-07-31 08:16:22 -0600748 return ret;
749}
750
Alex Williamson9587f442013-04-25 16:12:38 -0600751/* hold write lock on container->group_lock */
Alex Williamsoncba33452012-07-31 08:16:22 -0600752static int __vfio_container_attach_groups(struct vfio_container *container,
753 struct vfio_iommu_driver *driver,
754 void *data)
755{
756 struct vfio_group *group;
757 int ret = -ENODEV;
758
759 list_for_each_entry(group, &container->group_list, container_next) {
760 ret = driver->ops->attach_group(data, group->iommu_group);
761 if (ret)
762 goto unwind;
763 }
764
765 return ret;
766
767unwind:
768 list_for_each_entry_continue_reverse(group, &container->group_list,
769 container_next) {
770 driver->ops->detach_group(data, group->iommu_group);
771 }
772
773 return ret;
774}
775
776static long vfio_ioctl_set_iommu(struct vfio_container *container,
777 unsigned long arg)
778{
779 struct vfio_iommu_driver *driver;
780 long ret = -ENODEV;
781
Alex Williamson9587f442013-04-25 16:12:38 -0600782 down_write(&container->group_lock);
Alex Williamsoncba33452012-07-31 08:16:22 -0600783
784 /*
785 * The container is designed to be an unprivileged interface while
786 * the group can be assigned to specific users. Therefore, only by
787 * adding a group to a container does the user get the privilege of
788 * enabling the iommu, which may allocate finite resources. There
789 * is no unset_iommu, but by removing all the groups from a container,
790 * the container is deprivileged and returns to an unset state.
791 */
792 if (list_empty(&container->group_list) || container->iommu_driver) {
Alex Williamson9587f442013-04-25 16:12:38 -0600793 up_write(&container->group_lock);
Alex Williamsoncba33452012-07-31 08:16:22 -0600794 return -EINVAL;
795 }
796
797 mutex_lock(&vfio.iommu_drivers_lock);
798 list_for_each_entry(driver, &vfio.iommu_drivers_list, vfio_next) {
799 void *data;
800
801 if (!try_module_get(driver->ops->owner))
802 continue;
803
804 /*
805 * The arg magic for SET_IOMMU is the same as CHECK_EXTENSION,
806 * so test which iommu driver reported support for this
807 * extension and call open on them. We also pass them the
808 * magic, allowing a single driver to support multiple
809 * interfaces if they'd like.
810 */
811 if (driver->ops->ioctl(NULL, VFIO_CHECK_EXTENSION, arg) <= 0) {
812 module_put(driver->ops->owner);
813 continue;
814 }
815
816 /* module reference holds the driver we're working on */
817 mutex_unlock(&vfio.iommu_drivers_lock);
818
819 data = driver->ops->open(arg);
820 if (IS_ERR(data)) {
821 ret = PTR_ERR(data);
822 module_put(driver->ops->owner);
823 goto skip_drivers_unlock;
824 }
825
826 ret = __vfio_container_attach_groups(container, driver, data);
827 if (!ret) {
828 container->iommu_driver = driver;
829 container->iommu_data = data;
830 } else {
831 driver->ops->release(data);
832 module_put(driver->ops->owner);
833 }
834
835 goto skip_drivers_unlock;
836 }
837
838 mutex_unlock(&vfio.iommu_drivers_lock);
839skip_drivers_unlock:
Alex Williamson9587f442013-04-25 16:12:38 -0600840 up_write(&container->group_lock);
Alex Williamsoncba33452012-07-31 08:16:22 -0600841
842 return ret;
843}
844
845static long vfio_fops_unl_ioctl(struct file *filep,
846 unsigned int cmd, unsigned long arg)
847{
848 struct vfio_container *container = filep->private_data;
849 struct vfio_iommu_driver *driver;
850 void *data;
851 long ret = -EINVAL;
852
853 if (!container)
854 return ret;
855
Alex Williamsoncba33452012-07-31 08:16:22 -0600856 switch (cmd) {
857 case VFIO_GET_API_VERSION:
858 ret = VFIO_API_VERSION;
859 break;
860 case VFIO_CHECK_EXTENSION:
861 ret = vfio_ioctl_check_extension(container, arg);
862 break;
863 case VFIO_SET_IOMMU:
864 ret = vfio_ioctl_set_iommu(container, arg);
865 break;
866 default:
Alex Williamson0b43c082013-04-29 08:41:36 -0600867 down_read(&container->group_lock);
868
869 driver = container->iommu_driver;
870 data = container->iommu_data;
871
Alex Williamsoncba33452012-07-31 08:16:22 -0600872 if (driver) /* passthrough all unrecognized ioctls */
873 ret = driver->ops->ioctl(data, cmd, arg);
Alex Williamson0b43c082013-04-29 08:41:36 -0600874
875 up_read(&container->group_lock);
Alex Williamsoncba33452012-07-31 08:16:22 -0600876 }
877
878 return ret;
879}
880
881#ifdef CONFIG_COMPAT
882static long vfio_fops_compat_ioctl(struct file *filep,
883 unsigned int cmd, unsigned long arg)
884{
885 arg = (unsigned long)compat_ptr(arg);
886 return vfio_fops_unl_ioctl(filep, cmd, arg);
887}
888#endif /* CONFIG_COMPAT */
889
890static int vfio_fops_open(struct inode *inode, struct file *filep)
891{
892 struct vfio_container *container;
893
894 container = kzalloc(sizeof(*container), GFP_KERNEL);
895 if (!container)
896 return -ENOMEM;
897
898 INIT_LIST_HEAD(&container->group_list);
Alex Williamson9587f442013-04-25 16:12:38 -0600899 init_rwsem(&container->group_lock);
Alex Williamsoncba33452012-07-31 08:16:22 -0600900 kref_init(&container->kref);
901
902 filep->private_data = container;
903
904 return 0;
905}
906
907static int vfio_fops_release(struct inode *inode, struct file *filep)
908{
909 struct vfio_container *container = filep->private_data;
910
911 filep->private_data = NULL;
912
913 vfio_container_put(container);
914
915 return 0;
916}
917
918/*
919 * Once an iommu driver is set, we optionally pass read/write/mmap
920 * on to the driver, allowing management interfaces beyond ioctl.
921 */
922static ssize_t vfio_fops_read(struct file *filep, char __user *buf,
923 size_t count, loff_t *ppos)
924{
925 struct vfio_container *container = filep->private_data;
Alex Williamson0b43c082013-04-29 08:41:36 -0600926 struct vfio_iommu_driver *driver;
927 ssize_t ret = -EINVAL;
Alex Williamsoncba33452012-07-31 08:16:22 -0600928
Alex Williamson0b43c082013-04-29 08:41:36 -0600929 down_read(&container->group_lock);
Alex Williamsoncba33452012-07-31 08:16:22 -0600930
Alex Williamson0b43c082013-04-29 08:41:36 -0600931 driver = container->iommu_driver;
932 if (likely(driver && driver->ops->read))
933 ret = driver->ops->read(container->iommu_data,
934 buf, count, ppos);
935
936 up_read(&container->group_lock);
937
938 return ret;
Alex Williamsoncba33452012-07-31 08:16:22 -0600939}
940
941static ssize_t vfio_fops_write(struct file *filep, const char __user *buf,
942 size_t count, loff_t *ppos)
943{
944 struct vfio_container *container = filep->private_data;
Alex Williamson0b43c082013-04-29 08:41:36 -0600945 struct vfio_iommu_driver *driver;
946 ssize_t ret = -EINVAL;
Alex Williamsoncba33452012-07-31 08:16:22 -0600947
Alex Williamson0b43c082013-04-29 08:41:36 -0600948 down_read(&container->group_lock);
Alex Williamsoncba33452012-07-31 08:16:22 -0600949
Alex Williamson0b43c082013-04-29 08:41:36 -0600950 driver = container->iommu_driver;
951 if (likely(driver && driver->ops->write))
952 ret = driver->ops->write(container->iommu_data,
953 buf, count, ppos);
954
955 up_read(&container->group_lock);
956
957 return ret;
Alex Williamsoncba33452012-07-31 08:16:22 -0600958}
959
960static int vfio_fops_mmap(struct file *filep, struct vm_area_struct *vma)
961{
962 struct vfio_container *container = filep->private_data;
Alex Williamson0b43c082013-04-29 08:41:36 -0600963 struct vfio_iommu_driver *driver;
964 int ret = -EINVAL;
Alex Williamsoncba33452012-07-31 08:16:22 -0600965
Alex Williamson0b43c082013-04-29 08:41:36 -0600966 down_read(&container->group_lock);
Alex Williamsoncba33452012-07-31 08:16:22 -0600967
Alex Williamson0b43c082013-04-29 08:41:36 -0600968 driver = container->iommu_driver;
969 if (likely(driver && driver->ops->mmap))
970 ret = driver->ops->mmap(container->iommu_data, vma);
971
972 up_read(&container->group_lock);
973
974 return ret;
Alex Williamsoncba33452012-07-31 08:16:22 -0600975}
976
977static const struct file_operations vfio_fops = {
978 .owner = THIS_MODULE,
979 .open = vfio_fops_open,
980 .release = vfio_fops_release,
981 .read = vfio_fops_read,
982 .write = vfio_fops_write,
983 .unlocked_ioctl = vfio_fops_unl_ioctl,
984#ifdef CONFIG_COMPAT
985 .compat_ioctl = vfio_fops_compat_ioctl,
986#endif
987 .mmap = vfio_fops_mmap,
988};
989
990/**
991 * VFIO Group fd, /dev/vfio/$GROUP
992 */
993static void __vfio_group_unset_container(struct vfio_group *group)
994{
995 struct vfio_container *container = group->container;
996 struct vfio_iommu_driver *driver;
997
Alex Williamson9587f442013-04-25 16:12:38 -0600998 down_write(&container->group_lock);
Alex Williamsoncba33452012-07-31 08:16:22 -0600999
1000 driver = container->iommu_driver;
1001 if (driver)
1002 driver->ops->detach_group(container->iommu_data,
1003 group->iommu_group);
1004
1005 group->container = NULL;
1006 list_del(&group->container_next);
1007
1008 /* Detaching the last group deprivileges a container, remove iommu */
1009 if (driver && list_empty(&container->group_list)) {
1010 driver->ops->release(container->iommu_data);
1011 module_put(driver->ops->owner);
1012 container->iommu_driver = NULL;
1013 container->iommu_data = NULL;
1014 }
1015
Alex Williamson9587f442013-04-25 16:12:38 -06001016 up_write(&container->group_lock);
Alex Williamsoncba33452012-07-31 08:16:22 -06001017
1018 vfio_container_put(container);
1019}
1020
1021/*
1022 * VFIO_GROUP_UNSET_CONTAINER should fail if there are other users or
1023 * if there was no container to unset. Since the ioctl is called on
1024 * the group, we know that still exists, therefore the only valid
1025 * transition here is 1->0.
1026 */
1027static int vfio_group_unset_container(struct vfio_group *group)
1028{
1029 int users = atomic_cmpxchg(&group->container_users, 1, 0);
1030
1031 if (!users)
1032 return -EINVAL;
1033 if (users != 1)
1034 return -EBUSY;
1035
1036 __vfio_group_unset_container(group);
1037
1038 return 0;
1039}
1040
1041/*
1042 * When removing container users, anything that removes the last user
1043 * implicitly removes the group from the container. That is, if the
1044 * group file descriptor is closed, as well as any device file descriptors,
1045 * the group is free.
1046 */
1047static void vfio_group_try_dissolve_container(struct vfio_group *group)
1048{
1049 if (0 == atomic_dec_if_positive(&group->container_users))
1050 __vfio_group_unset_container(group);
1051}
1052
1053static int vfio_group_set_container(struct vfio_group *group, int container_fd)
1054{
Al Viro2903ff02012-08-28 12:52:22 -04001055 struct fd f;
Alex Williamsoncba33452012-07-31 08:16:22 -06001056 struct vfio_container *container;
1057 struct vfio_iommu_driver *driver;
Al Viro2903ff02012-08-28 12:52:22 -04001058 int ret = 0;
Alex Williamsoncba33452012-07-31 08:16:22 -06001059
1060 if (atomic_read(&group->container_users))
1061 return -EINVAL;
1062
Al Viro2903ff02012-08-28 12:52:22 -04001063 f = fdget(container_fd);
1064 if (!f.file)
Alex Williamsoncba33452012-07-31 08:16:22 -06001065 return -EBADF;
1066
1067 /* Sanity check, is this really our fd? */
Al Viro2903ff02012-08-28 12:52:22 -04001068 if (f.file->f_op != &vfio_fops) {
1069 fdput(f);
Alex Williamsoncba33452012-07-31 08:16:22 -06001070 return -EINVAL;
1071 }
1072
Al Viro2903ff02012-08-28 12:52:22 -04001073 container = f.file->private_data;
Alex Williamsoncba33452012-07-31 08:16:22 -06001074 WARN_ON(!container); /* fget ensures we don't race vfio_release */
1075
Alex Williamson9587f442013-04-25 16:12:38 -06001076 down_write(&container->group_lock);
Alex Williamsoncba33452012-07-31 08:16:22 -06001077
1078 driver = container->iommu_driver;
1079 if (driver) {
1080 ret = driver->ops->attach_group(container->iommu_data,
1081 group->iommu_group);
1082 if (ret)
1083 goto unlock_out;
1084 }
1085
1086 group->container = container;
1087 list_add(&group->container_next, &container->group_list);
1088
1089 /* Get a reference on the container and mark a user within the group */
1090 vfio_container_get(container);
1091 atomic_inc(&group->container_users);
1092
1093unlock_out:
Alex Williamson9587f442013-04-25 16:12:38 -06001094 up_write(&container->group_lock);
Al Viro2903ff02012-08-28 12:52:22 -04001095 fdput(f);
Alex Williamsoncba33452012-07-31 08:16:22 -06001096 return ret;
1097}
1098
1099static bool vfio_group_viable(struct vfio_group *group)
1100{
1101 return (iommu_group_for_each_dev(group->iommu_group,
1102 group, vfio_dev_viable) == 0);
1103}
1104
1105static const struct file_operations vfio_device_fops;
1106
1107static int vfio_group_get_device_fd(struct vfio_group *group, char *buf)
1108{
1109 struct vfio_device *device;
1110 struct file *filep;
1111 int ret = -ENODEV;
1112
1113 if (0 == atomic_read(&group->container_users) ||
1114 !group->container->iommu_driver || !vfio_group_viable(group))
1115 return -EINVAL;
1116
1117 mutex_lock(&group->device_lock);
1118 list_for_each_entry(device, &group->device_list, group_next) {
1119 if (strcmp(dev_name(device->dev), buf))
1120 continue;
1121
1122 ret = device->ops->open(device->device_data);
1123 if (ret)
1124 break;
1125 /*
1126 * We can't use anon_inode_getfd() because we need to modify
1127 * the f_mode flags directly to allow more than just ioctls
1128 */
1129 ret = get_unused_fd();
1130 if (ret < 0) {
1131 device->ops->release(device->device_data);
1132 break;
1133 }
1134
1135 filep = anon_inode_getfile("[vfio-device]", &vfio_device_fops,
1136 device, O_RDWR);
1137 if (IS_ERR(filep)) {
1138 put_unused_fd(ret);
1139 ret = PTR_ERR(filep);
1140 device->ops->release(device->device_data);
1141 break;
1142 }
1143
1144 /*
1145 * TODO: add an anon_inode interface to do this.
1146 * Appears to be missing by lack of need rather than
1147 * explicitly prevented. Now there's need.
1148 */
1149 filep->f_mode |= (FMODE_LSEEK | FMODE_PREAD | FMODE_PWRITE);
1150
Alex Williamsoncba33452012-07-31 08:16:22 -06001151 vfio_device_get(device);
1152 atomic_inc(&group->container_users);
Al Viro31605de2012-08-17 21:32:56 -04001153
1154 fd_install(ret, filep);
Alex Williamsoncba33452012-07-31 08:16:22 -06001155 break;
1156 }
1157 mutex_unlock(&group->device_lock);
1158
1159 return ret;
1160}
1161
1162static long vfio_group_fops_unl_ioctl(struct file *filep,
1163 unsigned int cmd, unsigned long arg)
1164{
1165 struct vfio_group *group = filep->private_data;
1166 long ret = -ENOTTY;
1167
1168 switch (cmd) {
1169 case VFIO_GROUP_GET_STATUS:
1170 {
1171 struct vfio_group_status status;
1172 unsigned long minsz;
1173
1174 minsz = offsetofend(struct vfio_group_status, flags);
1175
1176 if (copy_from_user(&status, (void __user *)arg, minsz))
1177 return -EFAULT;
1178
1179 if (status.argsz < minsz)
1180 return -EINVAL;
1181
1182 status.flags = 0;
1183
1184 if (vfio_group_viable(group))
1185 status.flags |= VFIO_GROUP_FLAGS_VIABLE;
1186
1187 if (group->container)
1188 status.flags |= VFIO_GROUP_FLAGS_CONTAINER_SET;
1189
1190 if (copy_to_user((void __user *)arg, &status, minsz))
1191 return -EFAULT;
1192
1193 ret = 0;
1194 break;
1195 }
1196 case VFIO_GROUP_SET_CONTAINER:
1197 {
1198 int fd;
1199
1200 if (get_user(fd, (int __user *)arg))
1201 return -EFAULT;
1202
1203 if (fd < 0)
1204 return -EINVAL;
1205
1206 ret = vfio_group_set_container(group, fd);
1207 break;
1208 }
1209 case VFIO_GROUP_UNSET_CONTAINER:
1210 ret = vfio_group_unset_container(group);
1211 break;
1212 case VFIO_GROUP_GET_DEVICE_FD:
1213 {
1214 char *buf;
1215
1216 buf = strndup_user((const char __user *)arg, PAGE_SIZE);
1217 if (IS_ERR(buf))
1218 return PTR_ERR(buf);
1219
1220 ret = vfio_group_get_device_fd(group, buf);
1221 kfree(buf);
1222 break;
1223 }
1224 }
1225
1226 return ret;
1227}
1228
1229#ifdef CONFIG_COMPAT
1230static long vfio_group_fops_compat_ioctl(struct file *filep,
1231 unsigned int cmd, unsigned long arg)
1232{
1233 arg = (unsigned long)compat_ptr(arg);
1234 return vfio_group_fops_unl_ioctl(filep, cmd, arg);
1235}
1236#endif /* CONFIG_COMPAT */
1237
1238static int vfio_group_fops_open(struct inode *inode, struct file *filep)
1239{
1240 struct vfio_group *group;
Alex Williamson6d6768c2013-06-25 16:06:54 -06001241 int opened;
Alex Williamsoncba33452012-07-31 08:16:22 -06001242
1243 group = vfio_group_get_from_minor(iminor(inode));
1244 if (!group)
1245 return -ENODEV;
1246
Alex Williamson6d6768c2013-06-25 16:06:54 -06001247 /* Do we need multiple instances of the group open? Seems not. */
1248 opened = atomic_cmpxchg(&group->opened, 0, 1);
1249 if (opened) {
1250 vfio_group_put(group);
1251 return -EBUSY;
1252 }
1253
1254 /* Is something still in use from a previous open? */
Alex Williamsoncba33452012-07-31 08:16:22 -06001255 if (group->container) {
Alex Williamson6d6768c2013-06-25 16:06:54 -06001256 atomic_dec(&group->opened);
Alex Williamsoncba33452012-07-31 08:16:22 -06001257 vfio_group_put(group);
1258 return -EBUSY;
1259 }
1260
1261 filep->private_data = group;
1262
1263 return 0;
1264}
1265
1266static int vfio_group_fops_release(struct inode *inode, struct file *filep)
1267{
1268 struct vfio_group *group = filep->private_data;
1269
1270 filep->private_data = NULL;
1271
1272 vfio_group_try_dissolve_container(group);
1273
Alex Williamson6d6768c2013-06-25 16:06:54 -06001274 atomic_dec(&group->opened);
1275
Alex Williamsoncba33452012-07-31 08:16:22 -06001276 vfio_group_put(group);
1277
1278 return 0;
1279}
1280
1281static const struct file_operations vfio_group_fops = {
1282 .owner = THIS_MODULE,
1283 .unlocked_ioctl = vfio_group_fops_unl_ioctl,
1284#ifdef CONFIG_COMPAT
1285 .compat_ioctl = vfio_group_fops_compat_ioctl,
1286#endif
1287 .open = vfio_group_fops_open,
1288 .release = vfio_group_fops_release,
1289};
1290
1291/**
1292 * VFIO Device fd
1293 */
1294static int vfio_device_fops_release(struct inode *inode, struct file *filep)
1295{
1296 struct vfio_device *device = filep->private_data;
1297
1298 device->ops->release(device->device_data);
1299
1300 vfio_group_try_dissolve_container(device->group);
1301
1302 vfio_device_put(device);
1303
1304 return 0;
1305}
1306
1307static long vfio_device_fops_unl_ioctl(struct file *filep,
1308 unsigned int cmd, unsigned long arg)
1309{
1310 struct vfio_device *device = filep->private_data;
1311
1312 if (unlikely(!device->ops->ioctl))
1313 return -EINVAL;
1314
1315 return device->ops->ioctl(device->device_data, cmd, arg);
1316}
1317
1318static ssize_t vfio_device_fops_read(struct file *filep, char __user *buf,
1319 size_t count, loff_t *ppos)
1320{
1321 struct vfio_device *device = filep->private_data;
1322
1323 if (unlikely(!device->ops->read))
1324 return -EINVAL;
1325
1326 return device->ops->read(device->device_data, buf, count, ppos);
1327}
1328
1329static ssize_t vfio_device_fops_write(struct file *filep,
1330 const char __user *buf,
1331 size_t count, loff_t *ppos)
1332{
1333 struct vfio_device *device = filep->private_data;
1334
1335 if (unlikely(!device->ops->write))
1336 return -EINVAL;
1337
1338 return device->ops->write(device->device_data, buf, count, ppos);
1339}
1340
1341static int vfio_device_fops_mmap(struct file *filep, struct vm_area_struct *vma)
1342{
1343 struct vfio_device *device = filep->private_data;
1344
1345 if (unlikely(!device->ops->mmap))
1346 return -EINVAL;
1347
1348 return device->ops->mmap(device->device_data, vma);
1349}
1350
1351#ifdef CONFIG_COMPAT
1352static long vfio_device_fops_compat_ioctl(struct file *filep,
1353 unsigned int cmd, unsigned long arg)
1354{
1355 arg = (unsigned long)compat_ptr(arg);
1356 return vfio_device_fops_unl_ioctl(filep, cmd, arg);
1357}
1358#endif /* CONFIG_COMPAT */
1359
1360static const struct file_operations vfio_device_fops = {
1361 .owner = THIS_MODULE,
1362 .release = vfio_device_fops_release,
1363 .read = vfio_device_fops_read,
1364 .write = vfio_device_fops_write,
1365 .unlocked_ioctl = vfio_device_fops_unl_ioctl,
1366#ifdef CONFIG_COMPAT
1367 .compat_ioctl = vfio_device_fops_compat_ioctl,
1368#endif
1369 .mmap = vfio_device_fops_mmap,
1370};
1371
1372/**
1373 * Module/class support
1374 */
1375static char *vfio_devnode(struct device *dev, umode_t *mode)
1376{
Alexey Kardashevskiy9a6aa272013-06-05 08:54:16 -06001377 if (mode && (MINOR(dev->devt) == 0))
Alex Williamson664e9382013-04-30 15:42:28 -06001378 *mode = S_IRUGO | S_IWUGO;
1379
Alex Williamsoncba33452012-07-31 08:16:22 -06001380 return kasprintf(GFP_KERNEL, "vfio/%s", dev_name(dev));
1381}
1382
1383static int __init vfio_init(void)
1384{
1385 int ret;
1386
1387 idr_init(&vfio.group_idr);
1388 mutex_init(&vfio.group_lock);
1389 mutex_init(&vfio.iommu_drivers_lock);
1390 INIT_LIST_HEAD(&vfio.group_list);
1391 INIT_LIST_HEAD(&vfio.iommu_drivers_list);
1392 init_waitqueue_head(&vfio.release_q);
1393
1394 vfio.class = class_create(THIS_MODULE, "vfio");
1395 if (IS_ERR(vfio.class)) {
1396 ret = PTR_ERR(vfio.class);
1397 goto err_class;
1398 }
1399
1400 vfio.class->devnode = vfio_devnode;
1401
1402 ret = alloc_chrdev_region(&vfio.devt, 0, MINORMASK, "vfio");
1403 if (ret)
1404 goto err_base_chrdev;
1405
1406 cdev_init(&vfio.cdev, &vfio_fops);
1407 ret = cdev_add(&vfio.cdev, vfio.devt, 1);
1408 if (ret)
1409 goto err_base_cdev;
1410
1411 vfio.dev = device_create(vfio.class, NULL, vfio.devt, NULL, "vfio");
1412 if (IS_ERR(vfio.dev)) {
1413 ret = PTR_ERR(vfio.dev);
1414 goto err_base_dev;
1415 }
1416
1417 /* /dev/vfio/$GROUP */
1418 cdev_init(&vfio.group_cdev, &vfio_group_fops);
1419 ret = cdev_add(&vfio.group_cdev,
1420 MKDEV(MAJOR(vfio.devt), 1), MINORMASK - 1);
1421 if (ret)
1422 goto err_groups_cdev;
1423
1424 pr_info(DRIVER_DESC " version: " DRIVER_VERSION "\n");
1425
Alex Williamson73fa0d12012-07-31 08:16:23 -06001426 /*
1427 * Attempt to load known iommu-drivers. This gives us a working
1428 * environment without the user needing to explicitly load iommu
1429 * drivers.
1430 */
1431 request_module_nowait("vfio_iommu_type1");
Alexey Kardashevskiy5ffd2292013-05-21 13:33:10 +10001432 request_module_nowait("vfio_iommu_spapr_tce");
Alex Williamson73fa0d12012-07-31 08:16:23 -06001433
Alex Williamsoncba33452012-07-31 08:16:22 -06001434 return 0;
1435
1436err_groups_cdev:
1437 device_destroy(vfio.class, vfio.devt);
1438err_base_dev:
1439 cdev_del(&vfio.cdev);
1440err_base_cdev:
1441 unregister_chrdev_region(vfio.devt, MINORMASK);
1442err_base_chrdev:
1443 class_destroy(vfio.class);
1444 vfio.class = NULL;
1445err_class:
1446 return ret;
1447}
1448
1449static void __exit vfio_cleanup(void)
1450{
1451 WARN_ON(!list_empty(&vfio.group_list));
1452
1453 idr_destroy(&vfio.group_idr);
1454 cdev_del(&vfio.group_cdev);
1455 device_destroy(vfio.class, vfio.devt);
1456 cdev_del(&vfio.cdev);
1457 unregister_chrdev_region(vfio.devt, MINORMASK);
1458 class_destroy(vfio.class);
1459 vfio.class = NULL;
1460}
1461
1462module_init(vfio_init);
1463module_exit(vfio_cleanup);
1464
1465MODULE_VERSION(DRIVER_VERSION);
1466MODULE_LICENSE("GPL v2");
1467MODULE_AUTHOR(DRIVER_AUTHOR);
1468MODULE_DESCRIPTION(DRIVER_DESC);