blob: 3a4a664bab44c44583ce696e8918146d504e5ae2 [file] [log] [blame]
David Brownellc1dca562008-06-19 17:51:44 -07001/*
2 * u_serial.c - utilities for USB gadget "serial port"/TTY support
3 *
4 * Copyright (C) 2003 Al Borchers (alborchers@steinerpoint.com)
5 * Copyright (C) 2008 David Brownell
6 * Copyright (C) 2008 by Nokia Corporation
7 *
8 * This code also borrows from usbserial.c, which is
9 * Copyright (C) 1999 - 2002 Greg Kroah-Hartman (greg@kroah.com)
10 * Copyright (C) 2000 Peter Berger (pberger@brimson.com)
11 * Copyright (C) 2000 Al Borchers (alborchers@steinerpoint.com)
12 *
13 * This software is distributed under the terms of the GNU General
14 * Public License ("GPL") as published by the Free Software Foundation,
15 * either version 2 of that License or (at your option) any later version.
16 */
17
18/* #define VERBOSE_DEBUG */
19
20#include <linux/kernel.h>
stephane duverger1e413942010-06-29 16:57:25 +020021#include <linux/sched.h>
David Brownellc1dca562008-06-19 17:51:44 -070022#include <linux/interrupt.h>
23#include <linux/device.h>
24#include <linux/delay.h>
25#include <linux/tty.h>
26#include <linux/tty_flip.h>
Tejun Heo5a0e3ad2010-03-24 17:04:11 +090027#include <linux/slab.h>
David Brownellc1dca562008-06-19 17:51:44 -070028
29#include "u_serial.h"
30
31
32/*
33 * This component encapsulates the TTY layer glue needed to provide basic
34 * "serial port" functionality through the USB gadget stack. Each such
35 * port is exposed through a /dev/ttyGS* node.
36 *
37 * After initialization (gserial_setup), these TTY port devices stay
38 * available until they are removed (gserial_cleanup). Each one may be
39 * connected to a USB function (gserial_connect), or disconnected (with
40 * gserial_disconnect) when the USB host issues a config change event.
41 * Data can only flow when the port is connected to the host.
42 *
43 * A given TTY port can be made available in multiple configurations.
44 * For example, each one might expose a ttyGS0 node which provides a
45 * login application. In one case that might use CDC ACM interface 0,
46 * while another configuration might use interface 3 for that. The
47 * work to handle that (including descriptor management) is not part
48 * of this component.
49 *
50 * Configurations may expose more than one TTY port. For example, if
51 * ttyGS0 provides login service, then ttyGS1 might provide dialer access
52 * for a telephone or fax link. And ttyGS2 might be something that just
53 * needs a simple byte stream interface for some messaging protocol that
54 * is managed in userspace ... OBEX, PTP, and MTP have been mentioned.
55 */
56
David Brownell937ef732008-07-07 12:16:08 -070057#define PREFIX "ttyGS"
58
David Brownellc1dca562008-06-19 17:51:44 -070059/*
60 * gserial is the lifecycle interface, used by USB functions
61 * gs_port is the I/O nexus, used by the tty driver
62 * tty_struct links to the tty/filesystem framework
63 *
64 * gserial <---> gs_port ... links will be null when the USB link is
David Brownell1f1ba112008-08-06 18:49:57 -070065 * inactive; managed by gserial_{connect,disconnect}(). each gserial
66 * instance can wrap its own USB control protocol.
David Brownellc1dca562008-06-19 17:51:44 -070067 * gserial->ioport == usb_ep->driver_data ... gs_port
68 * gs_port->port_usb ... gserial
69 *
70 * gs_port <---> tty_struct ... links will be null when the TTY file
71 * isn't opened; managed by gs_open()/gs_close()
72 * gserial->port_tty ... tty_struct
73 * tty_struct->driver_data ... gserial
74 */
75
76/* RX and TX queues can buffer QUEUE_SIZE packets before they hit the
77 * next layer of buffering. For TX that's a circular buffer; for RX
78 * consider it a NOP. A third layer is provided by the TTY code.
79 */
80#define QUEUE_SIZE 16
81#define WRITE_BUF_SIZE 8192 /* TX only */
82
83/* circular buffer */
84struct gs_buf {
85 unsigned buf_size;
86 char *buf_buf;
87 char *buf_get;
88 char *buf_put;
89};
90
91/*
92 * The port structure holds info for each port, one for each minor number
93 * (and thus for each /dev/ node).
94 */
95struct gs_port {
96 spinlock_t port_lock; /* guard port_* access */
97
98 struct gserial *port_usb;
99 struct tty_struct *port_tty;
100
101 unsigned open_count;
102 bool openclose; /* open/close in progress */
103 u8 port_num;
104
105 wait_queue_head_t close_wait; /* wait for last close */
106
107 struct list_head read_pool;
Jim Sung28609d42010-11-04 18:47:51 -0700108 int read_started;
109 int read_allocated;
David Brownell937ef732008-07-07 12:16:08 -0700110 struct list_head read_queue;
111 unsigned n_read;
David Brownellc1dca562008-06-19 17:51:44 -0700112 struct tasklet_struct push;
113
114 struct list_head write_pool;
Jim Sung28609d42010-11-04 18:47:51 -0700115 int write_started;
116 int write_allocated;
David Brownellc1dca562008-06-19 17:51:44 -0700117 struct gs_buf port_write_buf;
118 wait_queue_head_t drain_wait; /* wait while writes drain */
119
120 /* REVISIT this state ... */
121 struct usb_cdc_line_coding port_line_coding; /* 8-N-1 etc */
122};
123
124/* increase N_PORTS if you need more */
125#define N_PORTS 4
126static struct portmaster {
127 struct mutex lock; /* protect open/close */
128 struct gs_port *port;
129} ports[N_PORTS];
130static unsigned n_ports;
131
132#define GS_CLOSE_TIMEOUT 15 /* seconds */
133
134
135
136#ifdef VERBOSE_DEBUG
137#define pr_vdebug(fmt, arg...) \
138 pr_debug(fmt, ##arg)
139#else
140#define pr_vdebug(fmt, arg...) \
141 ({ if (0) pr_debug(fmt, ##arg); })
142#endif
143
144/*-------------------------------------------------------------------------*/
145
146/* Circular Buffer */
147
148/*
149 * gs_buf_alloc
150 *
151 * Allocate a circular buffer and all associated memory.
152 */
153static int gs_buf_alloc(struct gs_buf *gb, unsigned size)
154{
155 gb->buf_buf = kmalloc(size, GFP_KERNEL);
156 if (gb->buf_buf == NULL)
157 return -ENOMEM;
158
159 gb->buf_size = size;
160 gb->buf_put = gb->buf_buf;
161 gb->buf_get = gb->buf_buf;
162
163 return 0;
164}
165
166/*
167 * gs_buf_free
168 *
169 * Free the buffer and all associated memory.
170 */
171static void gs_buf_free(struct gs_buf *gb)
172{
173 kfree(gb->buf_buf);
174 gb->buf_buf = NULL;
175}
176
177/*
178 * gs_buf_clear
179 *
180 * Clear out all data in the circular buffer.
181 */
182static void gs_buf_clear(struct gs_buf *gb)
183{
184 gb->buf_get = gb->buf_put;
185 /* equivalent to a get of all data available */
186}
187
188/*
189 * gs_buf_data_avail
190 *
David Brownell1f1ba112008-08-06 18:49:57 -0700191 * Return the number of bytes of data written into the circular
David Brownellc1dca562008-06-19 17:51:44 -0700192 * buffer.
193 */
194static unsigned gs_buf_data_avail(struct gs_buf *gb)
195{
196 return (gb->buf_size + gb->buf_put - gb->buf_get) % gb->buf_size;
197}
198
199/*
200 * gs_buf_space_avail
201 *
202 * Return the number of bytes of space available in the circular
203 * buffer.
204 */
205static unsigned gs_buf_space_avail(struct gs_buf *gb)
206{
207 return (gb->buf_size + gb->buf_get - gb->buf_put - 1) % gb->buf_size;
208}
209
210/*
211 * gs_buf_put
212 *
213 * Copy data data from a user buffer and put it into the circular buffer.
214 * Restrict to the amount of space available.
215 *
216 * Return the number of bytes copied.
217 */
218static unsigned
219gs_buf_put(struct gs_buf *gb, const char *buf, unsigned count)
220{
221 unsigned len;
222
223 len = gs_buf_space_avail(gb);
224 if (count > len)
225 count = len;
226
227 if (count == 0)
228 return 0;
229
230 len = gb->buf_buf + gb->buf_size - gb->buf_put;
231 if (count > len) {
232 memcpy(gb->buf_put, buf, len);
233 memcpy(gb->buf_buf, buf+len, count - len);
234 gb->buf_put = gb->buf_buf + count - len;
235 } else {
236 memcpy(gb->buf_put, buf, count);
237 if (count < len)
238 gb->buf_put += count;
239 else /* count == len */
240 gb->buf_put = gb->buf_buf;
241 }
242
243 return count;
244}
245
246/*
247 * gs_buf_get
248 *
249 * Get data from the circular buffer and copy to the given buffer.
250 * Restrict to the amount of data available.
251 *
252 * Return the number of bytes copied.
253 */
254static unsigned
255gs_buf_get(struct gs_buf *gb, char *buf, unsigned count)
256{
257 unsigned len;
258
259 len = gs_buf_data_avail(gb);
260 if (count > len)
261 count = len;
262
263 if (count == 0)
264 return 0;
265
266 len = gb->buf_buf + gb->buf_size - gb->buf_get;
267 if (count > len) {
268 memcpy(buf, gb->buf_get, len);
269 memcpy(buf+len, gb->buf_buf, count - len);
270 gb->buf_get = gb->buf_buf + count - len;
271 } else {
272 memcpy(buf, gb->buf_get, count);
273 if (count < len)
274 gb->buf_get += count;
275 else /* count == len */
276 gb->buf_get = gb->buf_buf;
277 }
278
279 return count;
280}
281
282/*-------------------------------------------------------------------------*/
283
284/* I/O glue between TTY (upper) and USB function (lower) driver layers */
285
286/*
287 * gs_alloc_req
288 *
289 * Allocate a usb_request and its buffer. Returns a pointer to the
290 * usb_request or NULL if there is an error.
291 */
David Brownell1f1ba112008-08-06 18:49:57 -0700292struct usb_request *
David Brownellc1dca562008-06-19 17:51:44 -0700293gs_alloc_req(struct usb_ep *ep, unsigned len, gfp_t kmalloc_flags)
294{
295 struct usb_request *req;
296
297 req = usb_ep_alloc_request(ep, kmalloc_flags);
298
299 if (req != NULL) {
300 req->length = len;
301 req->buf = kmalloc(len, kmalloc_flags);
302 if (req->buf == NULL) {
303 usb_ep_free_request(ep, req);
304 return NULL;
305 }
306 }
307
308 return req;
309}
310
311/*
312 * gs_free_req
313 *
314 * Free a usb_request and its buffer.
315 */
David Brownell1f1ba112008-08-06 18:49:57 -0700316void gs_free_req(struct usb_ep *ep, struct usb_request *req)
David Brownellc1dca562008-06-19 17:51:44 -0700317{
318 kfree(req->buf);
319 usb_ep_free_request(ep, req);
320}
321
322/*
323 * gs_send_packet
324 *
325 * If there is data to send, a packet is built in the given
326 * buffer and the size is returned. If there is no data to
327 * send, 0 is returned.
328 *
329 * Called with port_lock held.
330 */
331static unsigned
332gs_send_packet(struct gs_port *port, char *packet, unsigned size)
333{
334 unsigned len;
335
336 len = gs_buf_data_avail(&port->port_write_buf);
337 if (len < size)
338 size = len;
339 if (size != 0)
340 size = gs_buf_get(&port->port_write_buf, packet, size);
341 return size;
342}
343
344/*
345 * gs_start_tx
346 *
347 * This function finds available write requests, calls
348 * gs_send_packet to fill these packets with data, and
349 * continues until either there are no more write requests
350 * available or no more data to send. This function is
351 * run whenever data arrives or write requests are available.
352 *
353 * Context: caller owns port_lock; port_usb is non-null.
354 */
355static int gs_start_tx(struct gs_port *port)
356/*
357__releases(&port->port_lock)
358__acquires(&port->port_lock)
359*/
360{
361 struct list_head *pool = &port->write_pool;
362 struct usb_ep *in = port->port_usb->in;
363 int status = 0;
364 bool do_tty_wake = false;
365
366 while (!list_empty(pool)) {
367 struct usb_request *req;
368 int len;
369
Jim Sung28609d42010-11-04 18:47:51 -0700370 if (port->write_started >= QUEUE_SIZE)
371 break;
372
David Brownellc1dca562008-06-19 17:51:44 -0700373 req = list_entry(pool->next, struct usb_request, list);
374 len = gs_send_packet(port, req->buf, in->maxpacket);
375 if (len == 0) {
376 wake_up_interruptible(&port->drain_wait);
377 break;
378 }
379 do_tty_wake = true;
380
381 req->length = len;
382 list_del(&req->list);
Daniel Glöckner2e251342009-05-28 12:53:24 +0200383 req->zero = (gs_buf_data_avail(&port->port_write_buf) == 0);
David Brownellc1dca562008-06-19 17:51:44 -0700384
David Brownell937ef732008-07-07 12:16:08 -0700385 pr_vdebug(PREFIX "%d: tx len=%d, 0x%02x 0x%02x 0x%02x ...\n",
386 port->port_num, len, *((u8 *)req->buf),
David Brownellc1dca562008-06-19 17:51:44 -0700387 *((u8 *)req->buf+1), *((u8 *)req->buf+2));
David Brownellc1dca562008-06-19 17:51:44 -0700388
389 /* Drop lock while we call out of driver; completions
390 * could be issued while we do so. Disconnection may
391 * happen too; maybe immediately before we queue this!
392 *
393 * NOTE that we may keep sending data for a while after
394 * the TTY closed (dev->ioport->port_tty is NULL).
395 */
396 spin_unlock(&port->port_lock);
397 status = usb_ep_queue(in, req, GFP_ATOMIC);
398 spin_lock(&port->port_lock);
399
400 if (status) {
401 pr_debug("%s: %s %s err %d\n",
402 __func__, "queue", in->name, status);
403 list_add(&req->list, pool);
404 break;
405 }
406
Jim Sung28609d42010-11-04 18:47:51 -0700407 port->write_started++;
408
David Brownellc1dca562008-06-19 17:51:44 -0700409 /* abort immediately after disconnect */
410 if (!port->port_usb)
411 break;
412 }
413
414 if (do_tty_wake && port->port_tty)
415 tty_wakeup(port->port_tty);
416 return status;
417}
418
David Brownellc1dca562008-06-19 17:51:44 -0700419/*
420 * Context: caller owns port_lock, and port_usb is set
421 */
422static unsigned gs_start_rx(struct gs_port *port)
423/*
424__releases(&port->port_lock)
425__acquires(&port->port_lock)
426*/
427{
428 struct list_head *pool = &port->read_pool;
429 struct usb_ep *out = port->port_usb->out;
David Brownellc1dca562008-06-19 17:51:44 -0700430
431 while (!list_empty(pool)) {
432 struct usb_request *req;
433 int status;
434 struct tty_struct *tty;
435
David Brownell937ef732008-07-07 12:16:08 -0700436 /* no more rx if closed */
David Brownellc1dca562008-06-19 17:51:44 -0700437 tty = port->port_tty;
David Brownell937ef732008-07-07 12:16:08 -0700438 if (!tty)
David Brownellc1dca562008-06-19 17:51:44 -0700439 break;
440
Jim Sung28609d42010-11-04 18:47:51 -0700441 if (port->read_started >= QUEUE_SIZE)
442 break;
443
David Brownellc1dca562008-06-19 17:51:44 -0700444 req = list_entry(pool->next, struct usb_request, list);
445 list_del(&req->list);
446 req->length = out->maxpacket;
447
448 /* drop lock while we call out; the controller driver
449 * may need to call us back (e.g. for disconnect)
450 */
451 spin_unlock(&port->port_lock);
452 status = usb_ep_queue(out, req, GFP_ATOMIC);
453 spin_lock(&port->port_lock);
454
455 if (status) {
456 pr_debug("%s: %s %s err %d\n",
457 __func__, "queue", out->name, status);
458 list_add(&req->list, pool);
459 break;
460 }
Jim Sung28609d42010-11-04 18:47:51 -0700461 port->read_started++;
David Brownellc1dca562008-06-19 17:51:44 -0700462
463 /* abort immediately after disconnect */
464 if (!port->port_usb)
465 break;
466 }
Jim Sung28609d42010-11-04 18:47:51 -0700467 return port->read_started;
David Brownellc1dca562008-06-19 17:51:44 -0700468}
469
David Brownell937ef732008-07-07 12:16:08 -0700470/*
471 * RX tasklet takes data out of the RX queue and hands it up to the TTY
472 * layer until it refuses to take any more data (or is throttled back).
473 * Then it issues reads for any further data.
474 *
475 * If the RX queue becomes full enough that no usb_request is queued,
476 * the OUT endpoint may begin NAKing as soon as its FIFO fills up.
477 * So QUEUE_SIZE packets plus however many the FIFO holds (usually two)
478 * can be buffered before the TTY layer's buffers (currently 64 KB).
479 */
480static void gs_rx_push(unsigned long _port)
481{
482 struct gs_port *port = (void *)_port;
483 struct tty_struct *tty;
484 struct list_head *queue = &port->read_queue;
485 bool disconnect = false;
486 bool do_push = false;
487
488 /* hand any queued data to the tty */
489 spin_lock_irq(&port->port_lock);
490 tty = port->port_tty;
491 while (!list_empty(queue)) {
492 struct usb_request *req;
493
494 req = list_first_entry(queue, struct usb_request, list);
495
496 /* discard data if tty was closed */
497 if (!tty)
498 goto recycle;
499
500 /* leave data queued if tty was rx throttled */
501 if (test_bit(TTY_THROTTLED, &tty->flags))
502 break;
503
504 switch (req->status) {
505 case -ESHUTDOWN:
506 disconnect = true;
507 pr_vdebug(PREFIX "%d: shutdown\n", port->port_num);
508 break;
509
510 default:
511 /* presumably a transient fault */
512 pr_warning(PREFIX "%d: unexpected RX status %d\n",
513 port->port_num, req->status);
514 /* FALLTHROUGH */
515 case 0:
516 /* normal completion */
517 break;
518 }
519
520 /* push data to (open) tty */
521 if (req->actual) {
522 char *packet = req->buf;
523 unsigned size = req->actual;
524 unsigned n;
525 int count;
526
527 /* we may have pushed part of this packet already... */
528 n = port->n_read;
529 if (n) {
530 packet += n;
531 size -= n;
532 }
533
534 count = tty_insert_flip_string(tty, packet, size);
535 if (count)
536 do_push = true;
537 if (count != size) {
538 /* stop pushing; TTY layer can't handle more */
539 port->n_read += count;
540 pr_vdebug(PREFIX "%d: rx block %d/%d\n",
541 port->port_num,
542 count, req->actual);
543 break;
544 }
545 port->n_read = 0;
546 }
547recycle:
548 list_move(&req->list, &port->read_pool);
Jim Sung28609d42010-11-04 18:47:51 -0700549 port->read_started--;
David Brownell937ef732008-07-07 12:16:08 -0700550 }
551
Jon Povey44a0c012010-06-14 19:41:04 +0900552 /* Push from tty to ldisc; without low_latency set this is handled by
553 * a workqueue, so we won't get callbacks and can hold port_lock
David Brownell937ef732008-07-07 12:16:08 -0700554 */
Shaun Silk50238292011-09-26 11:26:43 +1000555 if (tty && do_push)
David Brownell937ef732008-07-07 12:16:08 -0700556 tty_flip_buffer_push(tty);
David Brownell937ef732008-07-07 12:16:08 -0700557
558
559 /* We want our data queue to become empty ASAP, keeping data
560 * in the tty and ldisc (not here). If we couldn't push any
561 * this time around, there may be trouble unless there's an
562 * implicit tty_unthrottle() call on its way...
563 *
564 * REVISIT we should probably add a timer to keep the tasklet
565 * from starving ... but it's not clear that case ever happens.
566 */
567 if (!list_empty(queue) && tty) {
568 if (!test_bit(TTY_THROTTLED, &tty->flags)) {
569 if (do_push)
570 tasklet_schedule(&port->push);
571 else
572 pr_warning(PREFIX "%d: RX not scheduled?\n",
573 port->port_num);
574 }
575 }
576
577 /* If we're still connected, refill the USB RX queue. */
578 if (!disconnect && port->port_usb)
579 gs_start_rx(port);
580
581 spin_unlock_irq(&port->port_lock);
582}
583
David Brownellc1dca562008-06-19 17:51:44 -0700584static void gs_read_complete(struct usb_ep *ep, struct usb_request *req)
585{
David Brownellc1dca562008-06-19 17:51:44 -0700586 struct gs_port *port = ep->driver_data;
587
David Brownell937ef732008-07-07 12:16:08 -0700588 /* Queue all received data until the tty layer is ready for it. */
David Brownellc1dca562008-06-19 17:51:44 -0700589 spin_lock(&port->port_lock);
David Brownell937ef732008-07-07 12:16:08 -0700590 list_add_tail(&req->list, &port->read_queue);
591 tasklet_schedule(&port->push);
David Brownellc1dca562008-06-19 17:51:44 -0700592 spin_unlock(&port->port_lock);
593}
594
595static void gs_write_complete(struct usb_ep *ep, struct usb_request *req)
596{
597 struct gs_port *port = ep->driver_data;
598
599 spin_lock(&port->port_lock);
600 list_add(&req->list, &port->write_pool);
Jim Sung28609d42010-11-04 18:47:51 -0700601 port->write_started--;
David Brownellc1dca562008-06-19 17:51:44 -0700602
603 switch (req->status) {
604 default:
605 /* presumably a transient fault */
606 pr_warning("%s: unexpected %s status %d\n",
607 __func__, ep->name, req->status);
608 /* FALL THROUGH */
609 case 0:
610 /* normal completion */
611 gs_start_tx(port);
612 break;
613
614 case -ESHUTDOWN:
615 /* disconnect */
616 pr_vdebug("%s: %s shutdown\n", __func__, ep->name);
617 break;
618 }
619
620 spin_unlock(&port->port_lock);
621}
622
Jim Sung28609d42010-11-04 18:47:51 -0700623static void gs_free_requests(struct usb_ep *ep, struct list_head *head,
624 int *allocated)
David Brownellc1dca562008-06-19 17:51:44 -0700625{
626 struct usb_request *req;
627
628 while (!list_empty(head)) {
629 req = list_entry(head->next, struct usb_request, list);
630 list_del(&req->list);
631 gs_free_req(ep, req);
Jim Sung28609d42010-11-04 18:47:51 -0700632 if (allocated)
633 (*allocated)--;
David Brownellc1dca562008-06-19 17:51:44 -0700634 }
635}
636
637static int gs_alloc_requests(struct usb_ep *ep, struct list_head *head,
Jim Sung28609d42010-11-04 18:47:51 -0700638 void (*fn)(struct usb_ep *, struct usb_request *),
639 int *allocated)
David Brownellc1dca562008-06-19 17:51:44 -0700640{
641 int i;
642 struct usb_request *req;
Jim Sung28609d42010-11-04 18:47:51 -0700643 int n = allocated ? QUEUE_SIZE - *allocated : QUEUE_SIZE;
David Brownellc1dca562008-06-19 17:51:44 -0700644
645 /* Pre-allocate up to QUEUE_SIZE transfers, but if we can't
646 * do quite that many this time, don't fail ... we just won't
647 * be as speedy as we might otherwise be.
648 */
Jim Sung28609d42010-11-04 18:47:51 -0700649 for (i = 0; i < n; i++) {
David Brownellc1dca562008-06-19 17:51:44 -0700650 req = gs_alloc_req(ep, ep->maxpacket, GFP_ATOMIC);
651 if (!req)
652 return list_empty(head) ? -ENOMEM : 0;
653 req->complete = fn;
654 list_add_tail(&req->list, head);
Jim Sung28609d42010-11-04 18:47:51 -0700655 if (allocated)
656 (*allocated)++;
David Brownellc1dca562008-06-19 17:51:44 -0700657 }
658 return 0;
659}
660
661/**
662 * gs_start_io - start USB I/O streams
663 * @dev: encapsulates endpoints to use
664 * Context: holding port_lock; port_tty and port_usb are non-null
665 *
666 * We only start I/O when something is connected to both sides of
667 * this port. If nothing is listening on the host side, we may
668 * be pointlessly filling up our TX buffers and FIFO.
669 */
670static int gs_start_io(struct gs_port *port)
671{
672 struct list_head *head = &port->read_pool;
673 struct usb_ep *ep = port->port_usb->out;
674 int status;
675 unsigned started;
676
677 /* Allocate RX and TX I/O buffers. We can't easily do this much
678 * earlier (with GFP_KERNEL) because the requests are coupled to
679 * endpoints, as are the packet sizes we'll be using. Different
680 * configurations may use different endpoints with a given port;
681 * and high speed vs full speed changes packet sizes too.
682 */
Jim Sung28609d42010-11-04 18:47:51 -0700683 status = gs_alloc_requests(ep, head, gs_read_complete,
684 &port->read_allocated);
David Brownellc1dca562008-06-19 17:51:44 -0700685 if (status)
686 return status;
687
688 status = gs_alloc_requests(port->port_usb->in, &port->write_pool,
Jim Sung28609d42010-11-04 18:47:51 -0700689 gs_write_complete, &port->write_allocated);
David Brownellc1dca562008-06-19 17:51:44 -0700690 if (status) {
Jim Sung28609d42010-11-04 18:47:51 -0700691 gs_free_requests(ep, head, &port->read_allocated);
David Brownellc1dca562008-06-19 17:51:44 -0700692 return status;
693 }
694
695 /* queue read requests */
David Brownell937ef732008-07-07 12:16:08 -0700696 port->n_read = 0;
David Brownellc1dca562008-06-19 17:51:44 -0700697 started = gs_start_rx(port);
698
699 /* unblock any pending writes into our circular buffer */
700 if (started) {
701 tty_wakeup(port->port_tty);
702 } else {
Jim Sung28609d42010-11-04 18:47:51 -0700703 gs_free_requests(ep, head, &port->read_allocated);
704 gs_free_requests(port->port_usb->in, &port->write_pool,
705 &port->write_allocated);
David Brownell937ef732008-07-07 12:16:08 -0700706 status = -EIO;
David Brownellc1dca562008-06-19 17:51:44 -0700707 }
708
David Brownell937ef732008-07-07 12:16:08 -0700709 return status;
David Brownellc1dca562008-06-19 17:51:44 -0700710}
711
712/*-------------------------------------------------------------------------*/
713
714/* TTY Driver */
715
716/*
717 * gs_open sets up the link between a gs_port and its associated TTY.
718 * That link is broken *only* by TTY close(), and all driver methods
719 * know that.
720 */
721static int gs_open(struct tty_struct *tty, struct file *file)
722{
723 int port_num = tty->index;
724 struct gs_port *port;
725 int status;
726
727 if (port_num < 0 || port_num >= n_ports)
728 return -ENXIO;
729
730 do {
731 mutex_lock(&ports[port_num].lock);
732 port = ports[port_num].port;
733 if (!port)
734 status = -ENODEV;
735 else {
736 spin_lock_irq(&port->port_lock);
737
738 /* already open? Great. */
739 if (port->open_count) {
740 status = 0;
741 port->open_count++;
742
743 /* currently opening/closing? wait ... */
744 } else if (port->openclose) {
745 status = -EBUSY;
746
747 /* ... else we do the work */
748 } else {
749 status = -EAGAIN;
750 port->openclose = true;
751 }
752 spin_unlock_irq(&port->port_lock);
753 }
754 mutex_unlock(&ports[port_num].lock);
755
756 switch (status) {
757 default:
758 /* fully handled */
759 return status;
760 case -EAGAIN:
761 /* must do the work */
762 break;
763 case -EBUSY:
764 /* wait for EAGAIN task to finish */
765 msleep(1);
766 /* REVISIT could have a waitchannel here, if
767 * concurrent open performance is important
768 */
769 break;
770 }
771 } while (status != -EAGAIN);
772
773 /* Do the "real open" */
774 spin_lock_irq(&port->port_lock);
775
776 /* allocate circular buffer on first open */
777 if (port->port_write_buf.buf_buf == NULL) {
778
779 spin_unlock_irq(&port->port_lock);
780 status = gs_buf_alloc(&port->port_write_buf, WRITE_BUF_SIZE);
781 spin_lock_irq(&port->port_lock);
782
783 if (status) {
784 pr_debug("gs_open: ttyGS%d (%p,%p) no buffer\n",
785 port->port_num, tty, file);
786 port->openclose = false;
787 goto exit_unlock_port;
788 }
789 }
790
791 /* REVISIT if REMOVED (ports[].port NULL), abort the open
792 * to let rmmod work faster (but this way isn't wrong).
793 */
794
795 /* REVISIT maybe wait for "carrier detect" */
796
797 tty->driver_data = port;
798 port->port_tty = tty;
799
800 port->open_count = 1;
801 port->openclose = false;
802
David Brownellc1dca562008-06-19 17:51:44 -0700803 /* if connected, start the I/O stream */
804 if (port->port_usb) {
David Brownell1f1ba112008-08-06 18:49:57 -0700805 struct gserial *gser = port->port_usb;
806
David Brownellc1dca562008-06-19 17:51:44 -0700807 pr_debug("gs_open: start ttyGS%d\n", port->port_num);
808 gs_start_io(port);
809
David Brownell1f1ba112008-08-06 18:49:57 -0700810 if (gser->connect)
811 gser->connect(gser);
David Brownellc1dca562008-06-19 17:51:44 -0700812 }
813
814 pr_debug("gs_open: ttyGS%d (%p,%p)\n", port->port_num, tty, file);
815
816 status = 0;
817
818exit_unlock_port:
819 spin_unlock_irq(&port->port_lock);
820 return status;
821}
822
823static int gs_writes_finished(struct gs_port *p)
824{
825 int cond;
826
827 /* return true on disconnect or empty buffer */
828 spin_lock_irq(&p->port_lock);
829 cond = (p->port_usb == NULL) || !gs_buf_data_avail(&p->port_write_buf);
830 spin_unlock_irq(&p->port_lock);
831
832 return cond;
833}
834
835static void gs_close(struct tty_struct *tty, struct file *file)
836{
837 struct gs_port *port = tty->driver_data;
David Brownell1f1ba112008-08-06 18:49:57 -0700838 struct gserial *gser;
David Brownellc1dca562008-06-19 17:51:44 -0700839
840 spin_lock_irq(&port->port_lock);
841
842 if (port->open_count != 1) {
843 if (port->open_count == 0)
844 WARN_ON(1);
845 else
846 --port->open_count;
847 goto exit;
848 }
849
850 pr_debug("gs_close: ttyGS%d (%p,%p) ...\n", port->port_num, tty, file);
851
852 /* mark port as closing but in use; we can drop port lock
853 * and sleep if necessary
854 */
855 port->openclose = true;
856 port->open_count = 0;
857
David Brownell1f1ba112008-08-06 18:49:57 -0700858 gser = port->port_usb;
859 if (gser && gser->disconnect)
860 gser->disconnect(gser);
David Brownellc1dca562008-06-19 17:51:44 -0700861
862 /* wait for circular write buffer to drain, disconnect, or at
863 * most GS_CLOSE_TIMEOUT seconds; then discard the rest
864 */
David Brownell1f1ba112008-08-06 18:49:57 -0700865 if (gs_buf_data_avail(&port->port_write_buf) > 0 && gser) {
David Brownellc1dca562008-06-19 17:51:44 -0700866 spin_unlock_irq(&port->port_lock);
867 wait_event_interruptible_timeout(port->drain_wait,
868 gs_writes_finished(port),
869 GS_CLOSE_TIMEOUT * HZ);
870 spin_lock_irq(&port->port_lock);
David Brownell1f1ba112008-08-06 18:49:57 -0700871 gser = port->port_usb;
David Brownellc1dca562008-06-19 17:51:44 -0700872 }
873
874 /* Iff we're disconnected, there can be no I/O in flight so it's
875 * ok to free the circular buffer; else just scrub it. And don't
876 * let the push tasklet fire again until we're re-opened.
877 */
David Brownell1f1ba112008-08-06 18:49:57 -0700878 if (gser == NULL)
David Brownellc1dca562008-06-19 17:51:44 -0700879 gs_buf_free(&port->port_write_buf);
880 else
881 gs_buf_clear(&port->port_write_buf);
882
David Brownellc1dca562008-06-19 17:51:44 -0700883 tty->driver_data = NULL;
884 port->port_tty = NULL;
885
886 port->openclose = false;
887
888 pr_debug("gs_close: ttyGS%d (%p,%p) done!\n",
889 port->port_num, tty, file);
890
891 wake_up_interruptible(&port->close_wait);
892exit:
893 spin_unlock_irq(&port->port_lock);
894}
895
896static int gs_write(struct tty_struct *tty, const unsigned char *buf, int count)
897{
898 struct gs_port *port = tty->driver_data;
899 unsigned long flags;
900 int status;
901
902 pr_vdebug("gs_write: ttyGS%d (%p) writing %d bytes\n",
903 port->port_num, tty, count);
904
905 spin_lock_irqsave(&port->port_lock, flags);
906 if (count)
907 count = gs_buf_put(&port->port_write_buf, buf, count);
908 /* treat count == 0 as flush_chars() */
909 if (port->port_usb)
910 status = gs_start_tx(port);
911 spin_unlock_irqrestore(&port->port_lock, flags);
912
913 return count;
914}
915
916static int gs_put_char(struct tty_struct *tty, unsigned char ch)
917{
918 struct gs_port *port = tty->driver_data;
919 unsigned long flags;
920 int status;
921
922 pr_vdebug("gs_put_char: (%d,%p) char=0x%x, called from %p\n",
923 port->port_num, tty, ch, __builtin_return_address(0));
924
925 spin_lock_irqsave(&port->port_lock, flags);
926 status = gs_buf_put(&port->port_write_buf, &ch, 1);
927 spin_unlock_irqrestore(&port->port_lock, flags);
928
929 return status;
930}
931
932static void gs_flush_chars(struct tty_struct *tty)
933{
934 struct gs_port *port = tty->driver_data;
935 unsigned long flags;
936
937 pr_vdebug("gs_flush_chars: (%d,%p)\n", port->port_num, tty);
938
939 spin_lock_irqsave(&port->port_lock, flags);
940 if (port->port_usb)
941 gs_start_tx(port);
942 spin_unlock_irqrestore(&port->port_lock, flags);
943}
944
945static int gs_write_room(struct tty_struct *tty)
946{
947 struct gs_port *port = tty->driver_data;
948 unsigned long flags;
949 int room = 0;
950
951 spin_lock_irqsave(&port->port_lock, flags);
952 if (port->port_usb)
953 room = gs_buf_space_avail(&port->port_write_buf);
954 spin_unlock_irqrestore(&port->port_lock, flags);
955
956 pr_vdebug("gs_write_room: (%d,%p) room=%d\n",
957 port->port_num, tty, room);
958
959 return room;
960}
961
962static int gs_chars_in_buffer(struct tty_struct *tty)
963{
964 struct gs_port *port = tty->driver_data;
965 unsigned long flags;
966 int chars = 0;
967
968 spin_lock_irqsave(&port->port_lock, flags);
969 chars = gs_buf_data_avail(&port->port_write_buf);
970 spin_unlock_irqrestore(&port->port_lock, flags);
971
972 pr_vdebug("gs_chars_in_buffer: (%d,%p) chars=%d\n",
973 port->port_num, tty, chars);
974
975 return chars;
976}
977
978/* undo side effects of setting TTY_THROTTLED */
979static void gs_unthrottle(struct tty_struct *tty)
980{
981 struct gs_port *port = tty->driver_data;
982 unsigned long flags;
David Brownellc1dca562008-06-19 17:51:44 -0700983
984 spin_lock_irqsave(&port->port_lock, flags);
David Brownell937ef732008-07-07 12:16:08 -0700985 if (port->port_usb) {
986 /* Kickstart read queue processing. We don't do xon/xoff,
987 * rts/cts, or other handshaking with the host, but if the
988 * read queue backs up enough we'll be NAKing OUT packets.
989 */
990 tasklet_schedule(&port->push);
991 pr_vdebug(PREFIX "%d: unthrottle\n", port->port_num);
992 }
David Brownellc1dca562008-06-19 17:51:44 -0700993 spin_unlock_irqrestore(&port->port_lock, flags);
David Brownellc1dca562008-06-19 17:51:44 -0700994}
995
David Brownell1f1ba112008-08-06 18:49:57 -0700996static int gs_break_ctl(struct tty_struct *tty, int duration)
997{
998 struct gs_port *port = tty->driver_data;
999 int status = 0;
1000 struct gserial *gser;
1001
1002 pr_vdebug("gs_break_ctl: ttyGS%d, send break (%d) \n",
1003 port->port_num, duration);
1004
1005 spin_lock_irq(&port->port_lock);
1006 gser = port->port_usb;
1007 if (gser && gser->send_break)
1008 status = gser->send_break(gser, duration);
1009 spin_unlock_irq(&port->port_lock);
1010
1011 return status;
1012}
1013
David Brownellc1dca562008-06-19 17:51:44 -07001014static const struct tty_operations gs_tty_ops = {
1015 .open = gs_open,
1016 .close = gs_close,
1017 .write = gs_write,
1018 .put_char = gs_put_char,
1019 .flush_chars = gs_flush_chars,
1020 .write_room = gs_write_room,
1021 .chars_in_buffer = gs_chars_in_buffer,
1022 .unthrottle = gs_unthrottle,
David Brownell1f1ba112008-08-06 18:49:57 -07001023 .break_ctl = gs_break_ctl,
David Brownellc1dca562008-06-19 17:51:44 -07001024};
1025
1026/*-------------------------------------------------------------------------*/
1027
1028static struct tty_driver *gs_tty_driver;
1029
1030static int __init
1031gs_port_alloc(unsigned port_num, struct usb_cdc_line_coding *coding)
1032{
1033 struct gs_port *port;
1034
1035 port = kzalloc(sizeof(struct gs_port), GFP_KERNEL);
1036 if (port == NULL)
1037 return -ENOMEM;
1038
1039 spin_lock_init(&port->port_lock);
1040 init_waitqueue_head(&port->close_wait);
1041 init_waitqueue_head(&port->drain_wait);
1042
1043 tasklet_init(&port->push, gs_rx_push, (unsigned long) port);
1044
1045 INIT_LIST_HEAD(&port->read_pool);
David Brownell937ef732008-07-07 12:16:08 -07001046 INIT_LIST_HEAD(&port->read_queue);
David Brownellc1dca562008-06-19 17:51:44 -07001047 INIT_LIST_HEAD(&port->write_pool);
1048
1049 port->port_num = port_num;
1050 port->port_line_coding = *coding;
1051
1052 ports[port_num].port = port;
1053
1054 return 0;
1055}
1056
1057/**
1058 * gserial_setup - initialize TTY driver for one or more ports
1059 * @g: gadget to associate with these ports
1060 * @count: how many ports to support
1061 * Context: may sleep
1062 *
1063 * The TTY stack needs to know in advance how many devices it should
1064 * plan to manage. Use this call to set up the ports you will be
1065 * exporting through USB. Later, connect them to functions based
1066 * on what configuration is activated by the USB host; and disconnect
1067 * them as appropriate.
1068 *
1069 * An example would be a two-configuration device in which both
1070 * configurations expose port 0, but through different functions.
1071 * One configuration could even expose port 1 while the other
1072 * one doesn't.
1073 *
1074 * Returns negative errno or zero.
1075 */
1076int __init gserial_setup(struct usb_gadget *g, unsigned count)
1077{
1078 unsigned i;
1079 struct usb_cdc_line_coding coding;
1080 int status;
1081
1082 if (count == 0 || count > N_PORTS)
1083 return -EINVAL;
1084
1085 gs_tty_driver = alloc_tty_driver(count);
1086 if (!gs_tty_driver)
1087 return -ENOMEM;
1088
1089 gs_tty_driver->owner = THIS_MODULE;
1090 gs_tty_driver->driver_name = "g_serial";
David Brownell937ef732008-07-07 12:16:08 -07001091 gs_tty_driver->name = PREFIX;
David Brownellc1dca562008-06-19 17:51:44 -07001092 /* uses dynamically assigned dev_t values */
1093
1094 gs_tty_driver->type = TTY_DRIVER_TYPE_SERIAL;
1095 gs_tty_driver->subtype = SERIAL_TYPE_NORMAL;
1096 gs_tty_driver->flags = TTY_DRIVER_REAL_RAW | TTY_DRIVER_DYNAMIC_DEV;
1097 gs_tty_driver->init_termios = tty_std_termios;
1098
1099 /* 9600-8-N-1 ... matches defaults expected by "usbser.sys" on
1100 * MS-Windows. Otherwise, most of these flags shouldn't affect
1101 * anything unless we were to actually hook up to a serial line.
1102 */
1103 gs_tty_driver->init_termios.c_cflag =
1104 B9600 | CS8 | CREAD | HUPCL | CLOCAL;
1105 gs_tty_driver->init_termios.c_ispeed = 9600;
1106 gs_tty_driver->init_termios.c_ospeed = 9600;
1107
Harvey Harrison551509d2009-02-11 14:11:36 -08001108 coding.dwDTERate = cpu_to_le32(9600);
David Brownellc1dca562008-06-19 17:51:44 -07001109 coding.bCharFormat = 8;
1110 coding.bParityType = USB_CDC_NO_PARITY;
1111 coding.bDataBits = USB_CDC_1_STOP_BITS;
1112
1113 tty_set_operations(gs_tty_driver, &gs_tty_ops);
1114
1115 /* make devices be openable */
1116 for (i = 0; i < count; i++) {
1117 mutex_init(&ports[i].lock);
1118 status = gs_port_alloc(i, &coding);
1119 if (status) {
1120 count = i;
1121 goto fail;
1122 }
1123 }
1124 n_ports = count;
1125
1126 /* export the driver ... */
1127 status = tty_register_driver(gs_tty_driver);
1128 if (status) {
David Brownellc1dca562008-06-19 17:51:44 -07001129 pr_err("%s: cannot register, err %d\n",
1130 __func__, status);
1131 goto fail;
1132 }
1133
1134 /* ... and sysfs class devices, so mdev/udev make /dev/ttyGS* */
1135 for (i = 0; i < count; i++) {
1136 struct device *tty_dev;
1137
1138 tty_dev = tty_register_device(gs_tty_driver, i, &g->dev);
1139 if (IS_ERR(tty_dev))
1140 pr_warning("%s: no classdev for port %d, err %ld\n",
1141 __func__, i, PTR_ERR(tty_dev));
1142 }
1143
1144 pr_debug("%s: registered %d ttyGS* device%s\n", __func__,
1145 count, (count == 1) ? "" : "s");
1146
1147 return status;
1148fail:
1149 while (count--)
1150 kfree(ports[count].port);
1151 put_tty_driver(gs_tty_driver);
1152 gs_tty_driver = NULL;
1153 return status;
1154}
1155
1156static int gs_closed(struct gs_port *port)
1157{
1158 int cond;
1159
1160 spin_lock_irq(&port->port_lock);
1161 cond = (port->open_count == 0) && !port->openclose;
1162 spin_unlock_irq(&port->port_lock);
1163 return cond;
1164}
1165
1166/**
1167 * gserial_cleanup - remove TTY-over-USB driver and devices
1168 * Context: may sleep
1169 *
1170 * This is called to free all resources allocated by @gserial_setup().
1171 * Accordingly, it may need to wait until some open /dev/ files have
1172 * closed.
1173 *
1174 * The caller must have issued @gserial_disconnect() for any ports
1175 * that had previously been connected, so that there is never any
1176 * I/O pending when it's called.
1177 */
1178void gserial_cleanup(void)
1179{
1180 unsigned i;
1181 struct gs_port *port;
1182
David Brownellac90e362008-07-01 13:18:20 -07001183 if (!gs_tty_driver)
1184 return;
1185
David Brownellc1dca562008-06-19 17:51:44 -07001186 /* start sysfs and /dev/ttyGS* node removal */
1187 for (i = 0; i < n_ports; i++)
1188 tty_unregister_device(gs_tty_driver, i);
1189
1190 for (i = 0; i < n_ports; i++) {
1191 /* prevent new opens */
1192 mutex_lock(&ports[i].lock);
1193 port = ports[i].port;
1194 ports[i].port = NULL;
1195 mutex_unlock(&ports[i].lock);
1196
David Brownell937ef732008-07-07 12:16:08 -07001197 tasklet_kill(&port->push);
1198
David Brownellc1dca562008-06-19 17:51:44 -07001199 /* wait for old opens to finish */
1200 wait_event(port->close_wait, gs_closed(port));
1201
1202 WARN_ON(port->port_usb != NULL);
1203
1204 kfree(port);
1205 }
1206 n_ports = 0;
1207
1208 tty_unregister_driver(gs_tty_driver);
Jon Poveyb23097b2010-06-14 19:42:10 +09001209 put_tty_driver(gs_tty_driver);
David Brownellc1dca562008-06-19 17:51:44 -07001210 gs_tty_driver = NULL;
1211
1212 pr_debug("%s: cleaned up ttyGS* support\n", __func__);
1213}
1214
1215/**
1216 * gserial_connect - notify TTY I/O glue that USB link is active
1217 * @gser: the function, set up with endpoints and descriptors
1218 * @port_num: which port is active
1219 * Context: any (usually from irq)
1220 *
1221 * This is called activate endpoints and let the TTY layer know that
1222 * the connection is active ... not unlike "carrier detect". It won't
1223 * necessarily start I/O queues; unless the TTY is held open by any
1224 * task, there would be no point. However, the endpoints will be
1225 * activated so the USB host can perform I/O, subject to basic USB
1226 * hardware flow control.
1227 *
1228 * Caller needs to have set up the endpoints and USB function in @dev
1229 * before calling this, as well as the appropriate (speed-specific)
1230 * endpoint descriptors, and also have set up the TTY driver by calling
1231 * @gserial_setup().
1232 *
1233 * Returns negative errno or zero.
1234 * On success, ep->driver_data will be overwritten.
1235 */
1236int gserial_connect(struct gserial *gser, u8 port_num)
1237{
1238 struct gs_port *port;
1239 unsigned long flags;
1240 int status;
1241
1242 if (!gs_tty_driver || port_num >= n_ports)
1243 return -ENXIO;
1244
1245 /* we "know" gserial_cleanup() hasn't been called */
1246 port = ports[port_num].port;
1247
1248 /* activate the endpoints */
Tatyana Brokhman72c973d2011-06-28 16:33:48 +03001249 status = usb_ep_enable(gser->in);
David Brownellc1dca562008-06-19 17:51:44 -07001250 if (status < 0)
1251 return status;
1252 gser->in->driver_data = port;
1253
Tatyana Brokhman72c973d2011-06-28 16:33:48 +03001254 status = usb_ep_enable(gser->out);
David Brownellc1dca562008-06-19 17:51:44 -07001255 if (status < 0)
1256 goto fail_out;
1257 gser->out->driver_data = port;
1258
1259 /* then tell the tty glue that I/O can work */
1260 spin_lock_irqsave(&port->port_lock, flags);
1261 gser->ioport = port;
1262 port->port_usb = gser;
1263
1264 /* REVISIT unclear how best to handle this state...
1265 * we don't really couple it with the Linux TTY.
1266 */
1267 gser->port_line_coding = port->port_line_coding;
1268
1269 /* REVISIT if waiting on "carrier detect", signal. */
1270
David Brownell1f1ba112008-08-06 18:49:57 -07001271 /* if it's already open, start I/O ... and notify the serial
1272 * protocol about open/close status (connect/disconnect).
David Brownellc1dca562008-06-19 17:51:44 -07001273 */
David Brownellc1dca562008-06-19 17:51:44 -07001274 if (port->open_count) {
1275 pr_debug("gserial_connect: start ttyGS%d\n", port->port_num);
1276 gs_start_io(port);
David Brownell1f1ba112008-08-06 18:49:57 -07001277 if (gser->connect)
1278 gser->connect(gser);
1279 } else {
1280 if (gser->disconnect)
1281 gser->disconnect(gser);
David Brownellc1dca562008-06-19 17:51:44 -07001282 }
1283
1284 spin_unlock_irqrestore(&port->port_lock, flags);
1285
1286 return status;
1287
1288fail_out:
1289 usb_ep_disable(gser->in);
1290 gser->in->driver_data = NULL;
1291 return status;
1292}
1293
1294/**
1295 * gserial_disconnect - notify TTY I/O glue that USB link is inactive
1296 * @gser: the function, on which gserial_connect() was called
1297 * Context: any (usually from irq)
1298 *
1299 * This is called to deactivate endpoints and let the TTY layer know
1300 * that the connection went inactive ... not unlike "hangup".
1301 *
1302 * On return, the state is as if gserial_connect() had never been called;
1303 * there is no active USB I/O on these endpoints.
1304 */
1305void gserial_disconnect(struct gserial *gser)
1306{
1307 struct gs_port *port = gser->ioport;
1308 unsigned long flags;
1309
1310 if (!port)
1311 return;
1312
1313 /* tell the TTY glue not to do I/O here any more */
1314 spin_lock_irqsave(&port->port_lock, flags);
1315
1316 /* REVISIT as above: how best to track this? */
1317 port->port_line_coding = gser->port_line_coding;
1318
1319 port->port_usb = NULL;
1320 gser->ioport = NULL;
1321 if (port->open_count > 0 || port->openclose) {
1322 wake_up_interruptible(&port->drain_wait);
1323 if (port->port_tty)
1324 tty_hangup(port->port_tty);
1325 }
1326 spin_unlock_irqrestore(&port->port_lock, flags);
1327
1328 /* disable endpoints, aborting down any active I/O */
1329 usb_ep_disable(gser->out);
1330 gser->out->driver_data = NULL;
1331
1332 usb_ep_disable(gser->in);
1333 gser->in->driver_data = NULL;
1334
1335 /* finally, free any unused/unusable I/O buffers */
1336 spin_lock_irqsave(&port->port_lock, flags);
1337 if (port->open_count == 0 && !port->openclose)
1338 gs_buf_free(&port->port_write_buf);
Jim Sung28609d42010-11-04 18:47:51 -07001339 gs_free_requests(gser->out, &port->read_pool, NULL);
1340 gs_free_requests(gser->out, &port->read_queue, NULL);
1341 gs_free_requests(gser->in, &port->write_pool, NULL);
1342
1343 port->read_allocated = port->read_started =
1344 port->write_allocated = port->write_started = 0;
1345
David Brownellc1dca562008-06-19 17:51:44 -07001346 spin_unlock_irqrestore(&port->port_lock, flags);
1347}