Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 1 | /* |
| 2 | * V4L2 asynchronous subdevice registration API |
| 3 | * |
| 4 | * Copyright (C) 2012-2013, Guennadi Liakhovetski <g.liakhovetski@gmx.de> |
| 5 | * |
| 6 | * This program is free software; you can redistribute it and/or modify |
| 7 | * it under the terms of the GNU General Public License version 2 as |
| 8 | * published by the Free Software Foundation. |
| 9 | */ |
| 10 | |
| 11 | #include <linux/device.h> |
| 12 | #include <linux/err.h> |
| 13 | #include <linux/i2c.h> |
| 14 | #include <linux/list.h> |
| 15 | #include <linux/module.h> |
| 16 | #include <linux/mutex.h> |
| 17 | #include <linux/platform_device.h> |
| 18 | #include <linux/slab.h> |
| 19 | #include <linux/types.h> |
| 20 | |
| 21 | #include <media/v4l2-async.h> |
| 22 | #include <media/v4l2-device.h> |
| 23 | #include <media/v4l2-subdev.h> |
| 24 | |
Sakari Ailus | 8621765 | 2015-06-11 12:18:01 -0700 | [diff] [blame] | 25 | static bool match_i2c(struct v4l2_subdev *sd, struct v4l2_async_subdev *asd) |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 26 | { |
Guennadi Liakhovetski | fe05e14 | 2013-06-24 05:13:51 -0300 | [diff] [blame] | 27 | #if IS_ENABLED(CONFIG_I2C) |
Sakari Ailus | 8621765 | 2015-06-11 12:18:01 -0700 | [diff] [blame] | 28 | struct i2c_client *client = i2c_verify_client(sd->dev); |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 29 | return client && |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 30 | asd->match.i2c.adapter_id == client->adapter->nr && |
| 31 | asd->match.i2c.address == client->addr; |
Guennadi Liakhovetski | fe05e14 | 2013-06-24 05:13:51 -0300 | [diff] [blame] | 32 | #else |
| 33 | return false; |
| 34 | #endif |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 35 | } |
| 36 | |
Sakari Ailus | 8621765 | 2015-06-11 12:18:01 -0700 | [diff] [blame] | 37 | static bool match_devname(struct v4l2_subdev *sd, |
| 38 | struct v4l2_async_subdev *asd) |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 39 | { |
Sakari Ailus | 8621765 | 2015-06-11 12:18:01 -0700 | [diff] [blame] | 40 | return !strcmp(asd->match.device_name.name, dev_name(sd->dev)); |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 41 | } |
| 42 | |
Sakari Ailus | 8621765 | 2015-06-11 12:18:01 -0700 | [diff] [blame] | 43 | static bool match_of(struct v4l2_subdev *sd, struct v4l2_async_subdev *asd) |
Sylwester Nawrocki | e7359f8 | 2013-07-19 12:21:29 -0300 | [diff] [blame] | 44 | { |
Javi Merino | d2180e0 | 2016-12-05 08:09:56 -0200 | [diff] [blame^] | 45 | return !of_node_cmp(of_node_full_name(sd->of_node), |
| 46 | of_node_full_name(asd->match.of.node)); |
Sakari Ailus | 8621765 | 2015-06-11 12:18:01 -0700 | [diff] [blame] | 47 | } |
| 48 | |
| 49 | static bool match_custom(struct v4l2_subdev *sd, struct v4l2_async_subdev *asd) |
| 50 | { |
| 51 | if (!asd->match.custom.match) |
| 52 | /* Match always */ |
| 53 | return true; |
| 54 | |
| 55 | return asd->match.custom.match(sd->dev, asd); |
Sylwester Nawrocki | e7359f8 | 2013-07-19 12:21:29 -0300 | [diff] [blame] | 56 | } |
| 57 | |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 58 | static LIST_HEAD(subdev_list); |
| 59 | static LIST_HEAD(notifier_list); |
| 60 | static DEFINE_MUTEX(list_lock); |
| 61 | |
| 62 | static struct v4l2_async_subdev *v4l2_async_belongs(struct v4l2_async_notifier *notifier, |
Sylwester Nawrocki | b426b3a | 2013-07-22 08:01:33 -0300 | [diff] [blame] | 63 | struct v4l2_subdev *sd) |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 64 | { |
Sakari Ailus | 8621765 | 2015-06-11 12:18:01 -0700 | [diff] [blame] | 65 | bool (*match)(struct v4l2_subdev *, struct v4l2_async_subdev *); |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 66 | struct v4l2_async_subdev *asd; |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 67 | |
| 68 | list_for_each_entry(asd, ¬ifier->waiting, list) { |
| 69 | /* bus_type has been verified valid before */ |
Sylwester Nawrocki | cfca764 | 2013-07-19 12:14:46 -0300 | [diff] [blame] | 70 | switch (asd->match_type) { |
| 71 | case V4L2_ASYNC_MATCH_CUSTOM: |
Sakari Ailus | 8621765 | 2015-06-11 12:18:01 -0700 | [diff] [blame] | 72 | match = match_custom; |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 73 | break; |
Sylwester Nawrocki | cfca764 | 2013-07-19 12:14:46 -0300 | [diff] [blame] | 74 | case V4L2_ASYNC_MATCH_DEVNAME: |
| 75 | match = match_devname; |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 76 | break; |
Sylwester Nawrocki | cfca764 | 2013-07-19 12:14:46 -0300 | [diff] [blame] | 77 | case V4L2_ASYNC_MATCH_I2C: |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 78 | match = match_i2c; |
| 79 | break; |
Sylwester Nawrocki | e7359f8 | 2013-07-19 12:21:29 -0300 | [diff] [blame] | 80 | case V4L2_ASYNC_MATCH_OF: |
| 81 | match = match_of; |
| 82 | break; |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 83 | default: |
| 84 | /* Cannot happen, unless someone breaks us */ |
| 85 | WARN_ON(true); |
| 86 | return NULL; |
| 87 | } |
| 88 | |
| 89 | /* match cannot be NULL here */ |
Sakari Ailus | 8621765 | 2015-06-11 12:18:01 -0700 | [diff] [blame] | 90 | if (match(sd, asd)) |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 91 | return asd; |
| 92 | } |
| 93 | |
| 94 | return NULL; |
| 95 | } |
| 96 | |
| 97 | static int v4l2_async_test_notify(struct v4l2_async_notifier *notifier, |
Sylwester Nawrocki | b426b3a | 2013-07-22 08:01:33 -0300 | [diff] [blame] | 98 | struct v4l2_subdev *sd, |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 99 | struct v4l2_async_subdev *asd) |
| 100 | { |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 101 | int ret; |
| 102 | |
| 103 | /* Remove from the waiting list */ |
| 104 | list_del(&asd->list); |
Sylwester Nawrocki | b426b3a | 2013-07-22 08:01:33 -0300 | [diff] [blame] | 105 | sd->asd = asd; |
| 106 | sd->notifier = notifier; |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 107 | |
| 108 | if (notifier->bound) { |
| 109 | ret = notifier->bound(notifier, sd, asd); |
| 110 | if (ret < 0) |
| 111 | return ret; |
| 112 | } |
| 113 | /* Move from the global subdevice list to notifier's done */ |
Sylwester Nawrocki | b426b3a | 2013-07-22 08:01:33 -0300 | [diff] [blame] | 114 | list_move(&sd->async_list, ¬ifier->done); |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 115 | |
| 116 | ret = v4l2_device_register_subdev(notifier->v4l2_dev, sd); |
| 117 | if (ret < 0) { |
| 118 | if (notifier->unbind) |
| 119 | notifier->unbind(notifier, sd, asd); |
| 120 | return ret; |
| 121 | } |
| 122 | |
| 123 | if (list_empty(¬ifier->waiting) && notifier->complete) |
| 124 | return notifier->complete(notifier); |
| 125 | |
| 126 | return 0; |
| 127 | } |
| 128 | |
Sylwester Nawrocki | b426b3a | 2013-07-22 08:01:33 -0300 | [diff] [blame] | 129 | static void v4l2_async_cleanup(struct v4l2_subdev *sd) |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 130 | { |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 131 | v4l2_device_unregister_subdev(sd); |
Sylwester Nawrocki | b426b3a | 2013-07-22 08:01:33 -0300 | [diff] [blame] | 132 | /* Subdevice driver will reprobe and put the subdev back onto the list */ |
| 133 | list_del_init(&sd->async_list); |
| 134 | sd->asd = NULL; |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 135 | sd->dev = NULL; |
| 136 | } |
| 137 | |
| 138 | int v4l2_async_notifier_register(struct v4l2_device *v4l2_dev, |
| 139 | struct v4l2_async_notifier *notifier) |
| 140 | { |
Sylwester Nawrocki | b426b3a | 2013-07-22 08:01:33 -0300 | [diff] [blame] | 141 | struct v4l2_subdev *sd, *tmp; |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 142 | struct v4l2_async_subdev *asd; |
| 143 | int i; |
| 144 | |
| 145 | if (!notifier->num_subdevs || notifier->num_subdevs > V4L2_MAX_SUBDEVS) |
| 146 | return -EINVAL; |
| 147 | |
| 148 | notifier->v4l2_dev = v4l2_dev; |
| 149 | INIT_LIST_HEAD(¬ifier->waiting); |
| 150 | INIT_LIST_HEAD(¬ifier->done); |
| 151 | |
| 152 | for (i = 0; i < notifier->num_subdevs; i++) { |
Sylwester Nawrocki | e8419d0 | 2013-07-19 12:31:10 -0300 | [diff] [blame] | 153 | asd = notifier->subdevs[i]; |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 154 | |
Sylwester Nawrocki | cfca764 | 2013-07-19 12:14:46 -0300 | [diff] [blame] | 155 | switch (asd->match_type) { |
| 156 | case V4L2_ASYNC_MATCH_CUSTOM: |
| 157 | case V4L2_ASYNC_MATCH_DEVNAME: |
| 158 | case V4L2_ASYNC_MATCH_I2C: |
Sylwester Nawrocki | e7359f8 | 2013-07-19 12:21:29 -0300 | [diff] [blame] | 159 | case V4L2_ASYNC_MATCH_OF: |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 160 | break; |
| 161 | default: |
| 162 | dev_err(notifier->v4l2_dev ? notifier->v4l2_dev->dev : NULL, |
Sylwester Nawrocki | cfca764 | 2013-07-19 12:14:46 -0300 | [diff] [blame] | 163 | "Invalid match type %u on %p\n", |
| 164 | asd->match_type, asd); |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 165 | return -EINVAL; |
| 166 | } |
| 167 | list_add_tail(&asd->list, ¬ifier->waiting); |
| 168 | } |
| 169 | |
| 170 | mutex_lock(&list_lock); |
| 171 | |
| 172 | /* Keep also completed notifiers on the list */ |
| 173 | list_add(¬ifier->list, ¬ifier_list); |
| 174 | |
Sylwester Nawrocki | b426b3a | 2013-07-22 08:01:33 -0300 | [diff] [blame] | 175 | list_for_each_entry_safe(sd, tmp, &subdev_list, async_list) { |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 176 | int ret; |
| 177 | |
Sylwester Nawrocki | b426b3a | 2013-07-22 08:01:33 -0300 | [diff] [blame] | 178 | asd = v4l2_async_belongs(notifier, sd); |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 179 | if (!asd) |
| 180 | continue; |
| 181 | |
Sylwester Nawrocki | b426b3a | 2013-07-22 08:01:33 -0300 | [diff] [blame] | 182 | ret = v4l2_async_test_notify(notifier, sd, asd); |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 183 | if (ret < 0) { |
| 184 | mutex_unlock(&list_lock); |
| 185 | return ret; |
| 186 | } |
| 187 | } |
| 188 | |
| 189 | mutex_unlock(&list_lock); |
| 190 | |
| 191 | return 0; |
| 192 | } |
| 193 | EXPORT_SYMBOL(v4l2_async_notifier_register); |
| 194 | |
| 195 | void v4l2_async_notifier_unregister(struct v4l2_async_notifier *notifier) |
| 196 | { |
Sylwester Nawrocki | b426b3a | 2013-07-22 08:01:33 -0300 | [diff] [blame] | 197 | struct v4l2_subdev *sd, *tmp; |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 198 | unsigned int notif_n_subdev = notifier->num_subdevs; |
| 199 | unsigned int n_subdev = min(notif_n_subdev, V4L2_MAX_SUBDEVS); |
Mauro Carvalho Chehab | 24e9a47 | 2013-11-02 06:20:16 -0300 | [diff] [blame] | 200 | struct device **dev; |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 201 | int i = 0; |
| 202 | |
Laurent Pinchart | 8e3fbfe | 2013-07-03 07:49:06 -0300 | [diff] [blame] | 203 | if (!notifier->v4l2_dev) |
| 204 | return; |
| 205 | |
Markus Elfring | f9e9c06 | 2016-12-26 15:14:33 -0200 | [diff] [blame] | 206 | dev = kmalloc_array(n_subdev, sizeof(*dev), GFP_KERNEL); |
Mauro Carvalho Chehab | 24e9a47 | 2013-11-02 06:20:16 -0300 | [diff] [blame] | 207 | if (!dev) { |
| 208 | dev_err(notifier->v4l2_dev->dev, |
| 209 | "Failed to allocate device cache!\n"); |
| 210 | } |
| 211 | |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 212 | mutex_lock(&list_lock); |
| 213 | |
| 214 | list_del(¬ifier->list); |
| 215 | |
Sylwester Nawrocki | ceedcc4 | 2013-07-31 13:10:18 -0300 | [diff] [blame] | 216 | list_for_each_entry_safe(sd, tmp, ¬ifier->done, async_list) { |
Mauro Carvalho Chehab | 24e9a47 | 2013-11-02 06:20:16 -0300 | [diff] [blame] | 217 | struct device *d; |
| 218 | |
| 219 | d = get_device(sd->dev); |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 220 | |
Sylwester Nawrocki | b426b3a | 2013-07-22 08:01:33 -0300 | [diff] [blame] | 221 | v4l2_async_cleanup(sd); |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 222 | |
| 223 | /* If we handled USB devices, we'd have to lock the parent too */ |
Mauro Carvalho Chehab | 24e9a47 | 2013-11-02 06:20:16 -0300 | [diff] [blame] | 224 | device_release_driver(d); |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 225 | |
| 226 | if (notifier->unbind) |
Sylwester Nawrocki | b426b3a | 2013-07-22 08:01:33 -0300 | [diff] [blame] | 227 | notifier->unbind(notifier, sd, sd->asd); |
Mauro Carvalho Chehab | 24e9a47 | 2013-11-02 06:20:16 -0300 | [diff] [blame] | 228 | |
| 229 | /* |
| 230 | * Store device at the device cache, in order to call |
| 231 | * put_device() on the final step |
| 232 | */ |
| 233 | if (dev) |
| 234 | dev[i++] = d; |
| 235 | else |
| 236 | put_device(d); |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 237 | } |
| 238 | |
| 239 | mutex_unlock(&list_lock); |
| 240 | |
Mauro Carvalho Chehab | 24e9a47 | 2013-11-02 06:20:16 -0300 | [diff] [blame] | 241 | /* |
| 242 | * Call device_attach() to reprobe devices |
| 243 | * |
| 244 | * NOTE: If dev allocation fails, i is 0, and the whole loop won't be |
| 245 | * executed. |
| 246 | */ |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 247 | while (i--) { |
| 248 | struct device *d = dev[i]; |
| 249 | |
| 250 | if (d && device_attach(d) < 0) { |
| 251 | const char *name = "(none)"; |
| 252 | int lock = device_trylock(d); |
| 253 | |
| 254 | if (lock && d->driver) |
| 255 | name = d->driver->name; |
| 256 | dev_err(d, "Failed to re-probe to %s\n", name); |
| 257 | if (lock) |
| 258 | device_unlock(d); |
| 259 | } |
| 260 | put_device(d); |
| 261 | } |
Mauro Carvalho Chehab | 24e9a47 | 2013-11-02 06:20:16 -0300 | [diff] [blame] | 262 | kfree(dev); |
Laurent Pinchart | 8e3fbfe | 2013-07-03 07:49:06 -0300 | [diff] [blame] | 263 | |
| 264 | notifier->v4l2_dev = NULL; |
| 265 | |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 266 | /* |
| 267 | * Don't care about the waiting list, it is initialised and populated |
| 268 | * upon notifier registration. |
| 269 | */ |
| 270 | } |
| 271 | EXPORT_SYMBOL(v4l2_async_notifier_unregister); |
| 272 | |
| 273 | int v4l2_async_register_subdev(struct v4l2_subdev *sd) |
| 274 | { |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 275 | struct v4l2_async_notifier *notifier; |
| 276 | |
Sakari Ailus | 8621765 | 2015-06-11 12:18:01 -0700 | [diff] [blame] | 277 | /* |
| 278 | * No reference taken. The reference is held by the device |
| 279 | * (struct v4l2_subdev.dev), and async sub-device does not |
| 280 | * exist independently of the device at any point of time. |
| 281 | */ |
| 282 | if (!sd->of_node && sd->dev) |
| 283 | sd->of_node = sd->dev->of_node; |
| 284 | |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 285 | mutex_lock(&list_lock); |
| 286 | |
Sylwester Nawrocki | b426b3a | 2013-07-22 08:01:33 -0300 | [diff] [blame] | 287 | INIT_LIST_HEAD(&sd->async_list); |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 288 | |
| 289 | list_for_each_entry(notifier, ¬ifier_list, list) { |
Sylwester Nawrocki | b426b3a | 2013-07-22 08:01:33 -0300 | [diff] [blame] | 290 | struct v4l2_async_subdev *asd = v4l2_async_belongs(notifier, sd); |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 291 | if (asd) { |
Sylwester Nawrocki | b426b3a | 2013-07-22 08:01:33 -0300 | [diff] [blame] | 292 | int ret = v4l2_async_test_notify(notifier, sd, asd); |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 293 | mutex_unlock(&list_lock); |
| 294 | return ret; |
| 295 | } |
| 296 | } |
| 297 | |
| 298 | /* None matched, wait for hot-plugging */ |
Sylwester Nawrocki | b426b3a | 2013-07-22 08:01:33 -0300 | [diff] [blame] | 299 | list_add(&sd->async_list, &subdev_list); |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 300 | |
| 301 | mutex_unlock(&list_lock); |
| 302 | |
| 303 | return 0; |
| 304 | } |
| 305 | EXPORT_SYMBOL(v4l2_async_register_subdev); |
| 306 | |
| 307 | void v4l2_async_unregister_subdev(struct v4l2_subdev *sd) |
| 308 | { |
Sylwester Nawrocki | b426b3a | 2013-07-22 08:01:33 -0300 | [diff] [blame] | 309 | struct v4l2_async_notifier *notifier = sd->notifier; |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 310 | |
Sylwester Nawrocki | b426b3a | 2013-07-22 08:01:33 -0300 | [diff] [blame] | 311 | if (!sd->asd) { |
| 312 | if (!list_empty(&sd->async_list)) |
| 313 | v4l2_async_cleanup(sd); |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 314 | return; |
| 315 | } |
| 316 | |
| 317 | mutex_lock(&list_lock); |
| 318 | |
Sylwester Nawrocki | b426b3a | 2013-07-22 08:01:33 -0300 | [diff] [blame] | 319 | list_add(&sd->asd->list, ¬ifier->waiting); |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 320 | |
Sylwester Nawrocki | b426b3a | 2013-07-22 08:01:33 -0300 | [diff] [blame] | 321 | v4l2_async_cleanup(sd); |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 322 | |
| 323 | if (notifier->unbind) |
Sylwester Nawrocki | b426b3a | 2013-07-22 08:01:33 -0300 | [diff] [blame] | 324 | notifier->unbind(notifier, sd, sd->asd); |
Guennadi Liakhovetski | e9e3104 | 2013-01-08 07:06:31 -0300 | [diff] [blame] | 325 | |
| 326 | mutex_unlock(&list_lock); |
| 327 | } |
| 328 | EXPORT_SYMBOL(v4l2_async_unregister_subdev); |