blob: 1d5b3fc62160ff95afddba819e9b91eb2cad26a0 [file] [log] [blame]
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -06001/*
2 * Copyright (C) 2003-2008 Takahiro Hirofuchi
3 *
4 * This is free software; you can redistribute it and/or modify
5 * it under the terms of the GNU General Public License as published by
6 * the Free Software Foundation; either version 2 of the License, or
7 * (at your option) any later version.
8 *
9 * This is distributed in the hope that it will be useful,
10 * but WITHOUT ANY WARRANTY; without even the implied warranty of
11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
12 * GNU General Public License for more details.
13 *
14 * You should have received a copy of the GNU General Public License
15 * along with this program; if not, write to the Free Software
16 * Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307,
17 * USA.
18 */
19
matt mooney7aaacb42011-05-11 22:33:43 -070020#include <asm/byteorder.h>
Arnd Bergmann9720b4b2011-03-02 00:13:05 +010021#include <linux/kthread.h>
matt mooney7aaacb42011-05-11 22:33:43 -070022#include <linux/usb.h>
23#include <linux/usb/hcd.h>
Tejun Heo5a0e3ad2010-03-24 17:04:11 +090024
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -060025#include "usbip_common.h"
26#include "stub.h"
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -060027
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -060028static int is_clear_halt_cmd(struct urb *urb)
29{
30 struct usb_ctrlrequest *req;
31
32 req = (struct usb_ctrlrequest *) urb->setup_packet;
33
34 return (req->bRequest == USB_REQ_CLEAR_FEATURE) &&
35 (req->bRequestType == USB_RECIP_ENDPOINT) &&
36 (req->wValue == USB_ENDPOINT_HALT);
37}
38
39static int is_set_interface_cmd(struct urb *urb)
40{
41 struct usb_ctrlrequest *req;
42
43 req = (struct usb_ctrlrequest *) urb->setup_packet;
44
45 return (req->bRequest == USB_REQ_SET_INTERFACE) &&
matt mooney64f338e2011-05-06 03:47:43 -070046 (req->bRequestType == USB_RECIP_INTERFACE);
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -060047}
48
49static int is_set_configuration_cmd(struct urb *urb)
50{
51 struct usb_ctrlrequest *req;
52
53 req = (struct usb_ctrlrequest *) urb->setup_packet;
54
55 return (req->bRequest == USB_REQ_SET_CONFIGURATION) &&
matt mooney64f338e2011-05-06 03:47:43 -070056 (req->bRequestType == USB_RECIP_DEVICE);
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -060057}
58
59static int is_reset_device_cmd(struct urb *urb)
60{
61 struct usb_ctrlrequest *req;
62 __u16 value;
63 __u16 index;
64
65 req = (struct usb_ctrlrequest *) urb->setup_packet;
66 value = le16_to_cpu(req->wValue);
67 index = le16_to_cpu(req->wIndex);
68
69 if ((req->bRequest == USB_REQ_SET_FEATURE) &&
matt mooney64f338e2011-05-06 03:47:43 -070070 (req->bRequestType == USB_RT_PORT) &&
71 (value == USB_PORT_FEAT_RESET)) {
Brian G. Merrellb8868e42009-07-21 00:46:13 -060072 usbip_dbg_stub_rx("reset_device_cmd, port %u\n", index);
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -060073 return 1;
74 } else
75 return 0;
76}
77
78static int tweak_clear_halt_cmd(struct urb *urb)
79{
80 struct usb_ctrlrequest *req;
81 int target_endp;
82 int target_dir;
83 int target_pipe;
84 int ret;
85
86 req = (struct usb_ctrlrequest *) urb->setup_packet;
87
88 /*
89 * The stalled endpoint is specified in the wIndex value. The endpoint
90 * of the urb is the target of this clear_halt request (i.e., control
91 * endpoint).
92 */
93 target_endp = le16_to_cpu(req->wIndex) & 0x000f;
94
95 /* the stalled endpoint direction is IN or OUT?. USB_DIR_IN is 0x80. */
96 target_dir = le16_to_cpu(req->wIndex) & 0x0080;
97
98 if (target_dir)
99 target_pipe = usb_rcvctrlpipe(urb->dev, target_endp);
100 else
101 target_pipe = usb_sndctrlpipe(urb->dev, target_endp);
102
103 ret = usb_clear_halt(urb->dev, target_pipe);
104 if (ret < 0)
matt mooney1a4b6f62011-05-19 16:47:32 -0700105 dev_err(&urb->dev->dev, "usb_clear_halt error: devnum %d endp "
106 "%d ret %d\n", urb->dev->devnum, target_endp, ret);
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600107 else
matt mooney1a4b6f62011-05-19 16:47:32 -0700108 dev_info(&urb->dev->dev, "usb_clear_halt done: devnum %d endp "
109 "%d\n", urb->dev->devnum, target_endp);
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600110
111 return ret;
112}
113
114static int tweak_set_interface_cmd(struct urb *urb)
115{
116 struct usb_ctrlrequest *req;
117 __u16 alternate;
118 __u16 interface;
119 int ret;
120
121 req = (struct usb_ctrlrequest *) urb->setup_packet;
122 alternate = le16_to_cpu(req->wValue);
123 interface = le16_to_cpu(req->wIndex);
124
matt mooney64f338e2011-05-06 03:47:43 -0700125 usbip_dbg_stub_rx("set_interface: inf %u alt %u\n",
126 interface, alternate);
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600127
128 ret = usb_set_interface(urb->dev, interface, alternate);
129 if (ret < 0)
matt mooney1a4b6f62011-05-19 16:47:32 -0700130 dev_err(&urb->dev->dev, "usb_set_interface error: inf %u alt "
131 "%u ret %d\n", interface, alternate, ret);
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600132 else
matt mooney1a4b6f62011-05-19 16:47:32 -0700133 dev_info(&urb->dev->dev, "usb_set_interface done: inf %u alt "
134 "%u\n", interface, alternate);
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600135
136 return ret;
137}
138
139static int tweak_set_configuration_cmd(struct urb *urb)
140{
141 struct usb_ctrlrequest *req;
142 __u16 config;
143
144 req = (struct usb_ctrlrequest *) urb->setup_packet;
145 config = le16_to_cpu(req->wValue);
146
147 /*
148 * I have never seen a multi-config device. Very rare.
149 * For most devices, this will be called to choose a default
150 * configuration only once in an initialization phase.
151 *
152 * set_configuration may change a device configuration and its device
153 * drivers will be unbound and assigned for a new device configuration.
154 * This means this usbip driver will be also unbound when called, then
155 * eventually reassigned to the device as far as driver matching
156 * condition is kept.
157 *
158 * Unfortunatelly, an existing usbip connection will be dropped
159 * due to this driver unbinding. So, skip here.
160 * A user may need to set a special configuration value before
161 * exporting the device.
162 */
matt mooney1a4b6f62011-05-19 16:47:32 -0700163 dev_info(&urb->dev->dev, "usb_set_configuration %d to %s... skip!\n",
164 config, dev_name(&urb->dev->dev));
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600165
166 return 0;
167 /* return usb_driver_set_configuration(urb->dev, config); */
168}
169
170static int tweak_reset_device_cmd(struct urb *urb)
171{
Arjan Melsd2dd0b02011-04-05 20:26:11 +0200172 struct stub_priv *priv = (struct stub_priv *) urb->context;
173 struct stub_device *sdev = priv->sdev;
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600174
matt mooney1a4b6f62011-05-19 16:47:32 -0700175 dev_info(&urb->dev->dev, "usb_queue_reset_device\n");
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600176
Arjan Melsd2dd0b02011-04-05 20:26:11 +0200177 /*
Akshay Joshi553a1a52011-08-17 15:58:31 -0400178 * With the implementation of pre_reset and post_reset the driver no
Arjan Melsd3ac0772011-05-20 23:25:46 +0200179 * longer unbinds. This allows the use of synchronous reset.
Arjan Melsd2dd0b02011-04-05 20:26:11 +0200180 */
Arjan Melsd3ac0772011-05-20 23:25:46 +0200181
Akshay Joshi97c451c2011-08-17 15:58:32 -0400182 if (usb_lock_device_for_reset(sdev->udev, sdev->interface) < 0) {
Arjan Melsd3ac0772011-05-20 23:25:46 +0200183 dev_err(&urb->dev->dev, "could not obtain lock to reset device\n");
184 return 0;
185 }
186 usb_reset_device(sdev->udev);
187 usb_unlock_device(sdev->udev);
188
Arjan Melsd2dd0b02011-04-05 20:26:11 +0200189 return 0;
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600190}
191
192/*
193 * clear_halt, set_interface, and set_configuration require special tricks.
194 */
195static void tweak_special_requests(struct urb *urb)
196{
197 if (!urb || !urb->setup_packet)
198 return;
199
200 if (usb_pipetype(urb->pipe) != PIPE_CONTROL)
201 return;
202
203 if (is_clear_halt_cmd(urb))
204 /* tweak clear_halt */
205 tweak_clear_halt_cmd(urb);
206
207 else if (is_set_interface_cmd(urb))
208 /* tweak set_interface */
209 tweak_set_interface_cmd(urb);
210
211 else if (is_set_configuration_cmd(urb))
212 /* tweak set_configuration */
213 tweak_set_configuration_cmd(urb);
214
215 else if (is_reset_device_cmd(urb))
216 tweak_reset_device_cmd(urb);
217 else
Brian G. Merrellb8868e42009-07-21 00:46:13 -0600218 usbip_dbg_stub_rx("no need to tweak\n");
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600219}
220
221/*
222 * stub_recv_unlink() unlinks the URB by a call to usb_unlink_urb().
223 * By unlinking the urb asynchronously, stub_rx can continuously
224 * process coming urbs. Even if the urb is unlinked, its completion
225 * handler will be called and stub_tx will send a return pdu.
226 *
227 * See also comments about unlinking strategy in vhci_hcd.c.
228 */
229static int stub_recv_cmd_unlink(struct stub_device *sdev,
matt mooney64f338e2011-05-06 03:47:43 -0700230 struct usbip_header *pdu)
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600231{
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600232 unsigned long flags;
233
234 struct stub_priv *priv;
235
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600236 spin_lock_irqsave(&sdev->priv_lock, flags);
237
Alexander Beregalov88a1e902008-12-07 05:32:46 +0300238 list_for_each_entry(priv, &sdev->priv_init, list) {
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600239 if (priv->seqnum == pdu->u.cmd_unlink.seqnum) {
240 int ret;
241
242 dev_info(&priv->urb->dev->dev, "unlink urb %p\n",
243 priv->urb);
244
245 /*
246 * This matched urb is not completed yet (i.e., be in
247 * flight in usb hcd hardware/driver). Now we are
248 * cancelling it. The unlinking flag means that we are
249 * now not going to return the normal result pdu of a
250 * submission request, but going to return a result pdu
251 * of the unlink request.
252 */
253 priv->unlinking = 1;
254
255 /*
256 * In the case that unlinking flag is on, prev->seqnum
257 * is changed from the seqnum of the cancelling urb to
258 * the seqnum of the unlink request. This will be used
259 * to make the result pdu of the unlink request.
260 */
261 priv->seqnum = pdu->base.seqnum;
262
263 spin_unlock_irqrestore(&sdev->priv_lock, flags);
264
265 /*
266 * usb_unlink_urb() is now out of spinlocking to avoid
267 * spinlock recursion since stub_complete() is
268 * sometimes called in this context but not in the
269 * interrupt context. If stub_complete() is executed
270 * before we call usb_unlink_urb(), usb_unlink_urb()
271 * will return an error value. In this case, stub_tx
272 * will return the result pdu of this unlink request
273 * though submission is completed and actual unlinking
274 * is not executed. OK?
275 */
276 /* In the above case, urb->status is not -ECONNRESET,
277 * so a driver in a client host will know the failure
278 * of the unlink request ?
279 */
280 ret = usb_unlink_urb(priv->urb);
281 if (ret != -EINPROGRESS)
282 dev_err(&priv->urb->dev->dev,
283 "failed to unlink a urb %p, ret %d\n",
284 priv->urb, ret);
285 return 0;
286 }
287 }
288
Brian G. Merrellb8868e42009-07-21 00:46:13 -0600289 usbip_dbg_stub_rx("seqnum %d is not pending\n",
matt mooney64f338e2011-05-06 03:47:43 -0700290 pdu->u.cmd_unlink.seqnum);
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600291
292 /*
293 * The urb of the unlink target is not found in priv_init queue. It was
294 * already completed and its results is/was going to be sent by a
295 * CMD_RET pdu. In this case, usb_unlink_urb() is not needed. We only
296 * return the completeness of this unlink request to vhci_hcd.
297 */
298 stub_enqueue_ret_unlink(sdev, pdu->base.seqnum, 0);
299
300 spin_unlock_irqrestore(&sdev->priv_lock, flags);
301
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600302 return 0;
303}
304
305static int valid_request(struct stub_device *sdev, struct usbip_header *pdu)
306{
307 struct usbip_device *ud = &sdev->ud;
Márton Németh9ba422b2011-05-24 23:19:18 +0200308 int valid = 0;
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600309
310 if (pdu->base.devid == sdev->devid) {
311 spin_lock(&ud->lock);
312 if (ud->status == SDEV_ST_USED) {
313 /* A request is valid. */
Márton Németh9ba422b2011-05-24 23:19:18 +0200314 valid = 1;
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600315 }
316 spin_unlock(&ud->lock);
317 }
318
Márton Németh9ba422b2011-05-24 23:19:18 +0200319 return valid;
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600320}
321
322static struct stub_priv *stub_priv_alloc(struct stub_device *sdev,
323 struct usbip_header *pdu)
324{
325 struct stub_priv *priv;
326 struct usbip_device *ud = &sdev->ud;
327 unsigned long flags;
328
329 spin_lock_irqsave(&sdev->priv_lock, flags);
330
Wei Yongjune68f2842009-02-06 11:08:58 +0800331 priv = kmem_cache_zalloc(stub_priv_cache, GFP_ATOMIC);
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600332 if (!priv) {
333 dev_err(&sdev->interface->dev, "alloc stub_priv\n");
334 spin_unlock_irqrestore(&sdev->priv_lock, flags);
335 usbip_event_add(ud, SDEV_EVENT_ERROR_MALLOC);
336 return NULL;
337 }
338
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600339 priv->seqnum = pdu->base.seqnum;
340 priv->sdev = sdev;
341
342 /*
343 * After a stub_priv is linked to a list_head,
344 * our error handler can free allocated data.
345 */
346 list_add_tail(&priv->list, &sdev->priv_init);
347
348 spin_unlock_irqrestore(&sdev->priv_lock, flags);
349
350 return priv;
351}
352
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600353static int get_pipe(struct stub_device *sdev, int epnum, int dir)
354{
Max Vozeler2d8f4592011-01-12 15:01:59 +0200355 struct usb_device *udev = sdev->udev;
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600356 struct usb_host_endpoint *ep;
357 struct usb_endpoint_descriptor *epd = NULL;
358
Endre Kollarab30f122010-07-27 12:39:45 +0200359 if (dir == USBIP_DIR_IN)
360 ep = udev->ep_in[epnum & 0x7f];
361 else
362 ep = udev->ep_out[epnum & 0x7f];
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600363 if (!ep) {
364 dev_err(&sdev->interface->dev, "no such endpoint?, %d\n",
365 epnum);
366 BUG();
367 }
368
369 epd = &ep->desc;
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600370 if (usb_endpoint_xfer_control(epd)) {
371 if (dir == USBIP_DIR_OUT)
372 return usb_sndctrlpipe(udev, epnum);
373 else
374 return usb_rcvctrlpipe(udev, epnum);
375 }
376
377 if (usb_endpoint_xfer_bulk(epd)) {
378 if (dir == USBIP_DIR_OUT)
379 return usb_sndbulkpipe(udev, epnum);
380 else
381 return usb_rcvbulkpipe(udev, epnum);
382 }
383
384 if (usb_endpoint_xfer_int(epd)) {
385 if (dir == USBIP_DIR_OUT)
386 return usb_sndintpipe(udev, epnum);
387 else
388 return usb_rcvintpipe(udev, epnum);
389 }
390
391 if (usb_endpoint_xfer_isoc(epd)) {
392 if (dir == USBIP_DIR_OUT)
393 return usb_sndisocpipe(udev, epnum);
394 else
395 return usb_rcvisocpipe(udev, epnum);
396 }
397
398 /* NOT REACHED */
399 dev_err(&sdev->interface->dev, "get pipe, epnum %d\n", epnum);
400 return 0;
401}
402
Endre Kollarb7a937e2010-07-27 12:40:00 +0200403static void masking_bogus_flags(struct urb *urb)
404{
405 int xfertype;
406 struct usb_device *dev;
407 struct usb_host_endpoint *ep;
408 int is_out;
409 unsigned int allowed;
410
411 if (!urb || urb->hcpriv || !urb->complete)
412 return;
413 dev = urb->dev;
414 if ((!dev) || (dev->state < USB_STATE_UNAUTHENTICATED))
415 return;
416
417 ep = (usb_pipein(urb->pipe) ? dev->ep_in : dev->ep_out)
matt mooney64f338e2011-05-06 03:47:43 -0700418 [usb_pipeendpoint(urb->pipe)];
Endre Kollarb7a937e2010-07-27 12:40:00 +0200419 if (!ep)
420 return;
421
422 xfertype = usb_endpoint_type(&ep->desc);
423 if (xfertype == USB_ENDPOINT_XFER_CONTROL) {
424 struct usb_ctrlrequest *setup =
matt mooney64f338e2011-05-06 03:47:43 -0700425 (struct usb_ctrlrequest *) urb->setup_packet;
Endre Kollarb7a937e2010-07-27 12:40:00 +0200426
427 if (!setup)
428 return;
429 is_out = !(setup->bRequestType & USB_DIR_IN) ||
matt mooney64f338e2011-05-06 03:47:43 -0700430 !setup->wLength;
Endre Kollarb7a937e2010-07-27 12:40:00 +0200431 } else {
432 is_out = usb_endpoint_dir_out(&ep->desc);
433 }
434
435 /* enforce simple/standard policy */
Greg Kroah-Hartman39248652010-07-27 11:12:21 -0700436 allowed = (URB_NO_TRANSFER_DMA_MAP | URB_NO_INTERRUPT |
437 URB_DIR_MASK | URB_FREE_BUFFER);
Endre Kollarb7a937e2010-07-27 12:40:00 +0200438 switch (xfertype) {
439 case USB_ENDPOINT_XFER_BULK:
440 if (is_out)
441 allowed |= URB_ZERO_PACKET;
442 /* FALLTHROUGH */
443 case USB_ENDPOINT_XFER_CONTROL:
444 allowed |= URB_NO_FSBR; /* only affects UHCI */
445 /* FALLTHROUGH */
446 default: /* all non-iso endpoints */
447 if (!is_out)
448 allowed |= URB_SHORT_NOT_OK;
449 break;
450 case USB_ENDPOINT_XFER_ISOC:
451 allowed |= URB_ISO_ASAP;
452 break;
453 }
454 urb->transfer_flags &= allowed;
455}
456
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600457static void stub_recv_cmd_submit(struct stub_device *sdev,
458 struct usbip_header *pdu)
459{
460 int ret;
461 struct stub_priv *priv;
462 struct usbip_device *ud = &sdev->ud;
Max Vozeler2d8f4592011-01-12 15:01:59 +0200463 struct usb_device *udev = sdev->udev;
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600464 int pipe = get_pipe(sdev, pdu->base.ep, pdu->base.direction);
465
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600466 priv = stub_priv_alloc(sdev, pdu);
467 if (!priv)
468 return;
469
470 /* setup a urb */
471 if (usb_pipeisoc(pipe))
472 priv->urb = usb_alloc_urb(pdu->u.cmd_submit.number_of_packets,
matt mooney64f338e2011-05-06 03:47:43 -0700473 GFP_KERNEL);
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600474 else
475 priv->urb = usb_alloc_urb(0, GFP_KERNEL);
476
477 if (!priv->urb) {
478 dev_err(&sdev->interface->dev, "malloc urb\n");
479 usbip_event_add(ud, SDEV_EVENT_ERROR_MALLOC);
480 return;
481 }
482
483 /* set priv->urb->transfer_buffer */
484 if (pdu->u.cmd_submit.transfer_buffer_length > 0) {
485 priv->urb->transfer_buffer =
486 kzalloc(pdu->u.cmd_submit.transfer_buffer_length,
matt mooney64f338e2011-05-06 03:47:43 -0700487 GFP_KERNEL);
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600488 if (!priv->urb->transfer_buffer) {
489 dev_err(&sdev->interface->dev, "malloc x_buff\n");
490 usbip_event_add(ud, SDEV_EVENT_ERROR_MALLOC);
491 return;
492 }
493 }
494
495 /* set priv->urb->setup_packet */
Julia Lawall94002c02010-05-15 23:21:43 +0200496 priv->urb->setup_packet = kmemdup(&pdu->u.cmd_submit.setup, 8,
497 GFP_KERNEL);
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600498 if (!priv->urb->setup_packet) {
499 dev_err(&sdev->interface->dev, "allocate setup_packet\n");
500 usbip_event_add(ud, SDEV_EVENT_ERROR_MALLOC);
501 return;
502 }
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600503
504 /* set other members from the base header of pdu */
505 priv->urb->context = (void *) priv;
506 priv->urb->dev = udev;
507 priv->urb->pipe = pipe;
508 priv->urb->complete = stub_complete;
509
510 usbip_pack_pdu(pdu, priv->urb, USBIP_CMD_SUBMIT, 0);
511
512
513 if (usbip_recv_xbuff(ud, priv->urb) < 0)
514 return;
515
516 if (usbip_recv_iso(ud, priv->urb) < 0)
517 return;
518
519 /* no need to submit an intercepted request, but harmless? */
520 tweak_special_requests(priv->urb);
521
Endre Kollarb7a937e2010-07-27 12:40:00 +0200522 masking_bogus_flags(priv->urb);
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600523 /* urb is now ready to submit */
524 ret = usb_submit_urb(priv->urb, GFP_KERNEL);
525
526 if (ret == 0)
Brian G. Merrellb8868e42009-07-21 00:46:13 -0600527 usbip_dbg_stub_rx("submit urb ok, seqnum %u\n",
matt mooney64f338e2011-05-06 03:47:43 -0700528 pdu->base.seqnum);
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600529 else {
530 dev_err(&sdev->interface->dev, "submit_urb error, %d\n", ret);
531 usbip_dump_header(pdu);
532 usbip_dump_urb(priv->urb);
533
534 /*
535 * Pessimistic.
536 * This connection will be discarded.
537 */
538 usbip_event_add(ud, SDEV_EVENT_ERROR_SUBMIT);
539 }
540
Brian G. Merrellb8868e42009-07-21 00:46:13 -0600541 usbip_dbg_stub_rx("Leave\n");
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600542 return;
543}
544
545/* recv a pdu */
546static void stub_rx_pdu(struct usbip_device *ud)
547{
548 int ret;
549 struct usbip_header pdu;
550 struct stub_device *sdev = container_of(ud, struct stub_device, ud);
551 struct device *dev = &sdev->interface->dev;
552
Brian G. Merrellb8868e42009-07-21 00:46:13 -0600553 usbip_dbg_stub_rx("Enter\n");
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600554
555 memset(&pdu, 0, sizeof(pdu));
556
557 /* 1. receive a pdu header */
Bart Westgeest5a08c522011-12-19 17:44:11 -0500558 ret = usbip_recv(ud->tcp_socket, &pdu, sizeof(pdu));
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600559 if (ret != sizeof(pdu)) {
560 dev_err(dev, "recv a header, %d\n", ret);
561 usbip_event_add(ud, SDEV_EVENT_ERROR_TCP);
562 return;
563 }
564
565 usbip_header_correct_endian(&pdu, 0);
566
Brian G. Merrellb8868e42009-07-21 00:46:13 -0600567 if (usbip_dbg_flag_stub_rx)
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600568 usbip_dump_header(&pdu);
569
570 if (!valid_request(sdev, &pdu)) {
571 dev_err(dev, "recv invalid request\n");
572 usbip_event_add(ud, SDEV_EVENT_ERROR_TCP);
573 return;
574 }
575
576 switch (pdu.base.command) {
577 case USBIP_CMD_UNLINK:
578 stub_recv_cmd_unlink(sdev, &pdu);
579 break;
580
581 case USBIP_CMD_SUBMIT:
582 stub_recv_cmd_submit(sdev, &pdu);
583 break;
584
585 default:
586 /* NOTREACHED */
587 dev_err(dev, "unknown pdu\n");
588 usbip_event_add(ud, SDEV_EVENT_ERROR_TCP);
matt mooney49aecef2011-05-06 03:47:54 -0700589 break;
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600590 }
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600591}
592
Arnd Bergmann9720b4b2011-03-02 00:13:05 +0100593int stub_rx_loop(void *data)
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600594{
Arnd Bergmann9720b4b2011-03-02 00:13:05 +0100595 struct usbip_device *ud = data;
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600596
Arnd Bergmann9720b4b2011-03-02 00:13:05 +0100597 while (!kthread_should_stop()) {
Brian G. Merrellb8868e42009-07-21 00:46:13 -0600598 if (usbip_event_happened(ud))
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600599 break;
600
601 stub_rx_pdu(ud);
602 }
matt mooney64f338e2011-05-06 03:47:43 -0700603
Arnd Bergmann9720b4b2011-03-02 00:13:05 +0100604 return 0;
Takahiro Hirofuchi4d7b5c72008-07-09 14:56:51 -0600605}