blob: f7771d86ad6c8d19e2016773c1b5ba4b2f3f33eb [file] [log] [blame]
David Brownellc1dca562008-06-19 17:51:44 -07001/*
2 * u_serial.c - utilities for USB gadget "serial port"/TTY support
3 *
4 * Copyright (C) 2003 Al Borchers (alborchers@steinerpoint.com)
5 * Copyright (C) 2008 David Brownell
6 * Copyright (C) 2008 by Nokia Corporation
7 *
8 * This code also borrows from usbserial.c, which is
9 * Copyright (C) 1999 - 2002 Greg Kroah-Hartman (greg@kroah.com)
10 * Copyright (C) 2000 Peter Berger (pberger@brimson.com)
11 * Copyright (C) 2000 Al Borchers (alborchers@steinerpoint.com)
12 *
13 * This software is distributed under the terms of the GNU General
14 * Public License ("GPL") as published by the Free Software Foundation,
15 * either version 2 of that License or (at your option) any later version.
16 */
17
18/* #define VERBOSE_DEBUG */
19
20#include <linux/kernel.h>
stephane duverger1e413942010-06-29 16:57:25 +020021#include <linux/sched.h>
David Brownellc1dca562008-06-19 17:51:44 -070022#include <linux/interrupt.h>
23#include <linux/device.h>
24#include <linux/delay.h>
25#include <linux/tty.h>
26#include <linux/tty_flip.h>
Tejun Heo5a0e3ad2010-03-24 17:04:11 +090027#include <linux/slab.h>
Paul Gortmakerf940fcd2011-05-27 09:56:31 -040028#include <linux/export.h>
Sebastian Andrzej Siewior3249ca22012-12-23 21:10:04 +010029#include <linux/module.h>
David Brownellc1dca562008-06-19 17:51:44 -070030
31#include "u_serial.h"
32
33
34/*
35 * This component encapsulates the TTY layer glue needed to provide basic
36 * "serial port" functionality through the USB gadget stack. Each such
37 * port is exposed through a /dev/ttyGS* node.
38 *
Sebastian Andrzej Siewior19b10a82012-12-23 21:10:06 +010039 * After this module has been loaded, the individual TTY port can be requested
40 * (gserial_alloc_line()) and it will stay available until they are removed
41 * (gserial_free_line()). Each one may be connected to a USB function
42 * (gserial_connect), or disconnected (with gserial_disconnect) when the USB
43 * host issues a config change event. Data can only flow when the port is
44 * connected to the host.
David Brownellc1dca562008-06-19 17:51:44 -070045 *
46 * A given TTY port can be made available in multiple configurations.
47 * For example, each one might expose a ttyGS0 node which provides a
48 * login application. In one case that might use CDC ACM interface 0,
49 * while another configuration might use interface 3 for that. The
50 * work to handle that (including descriptor management) is not part
51 * of this component.
52 *
53 * Configurations may expose more than one TTY port. For example, if
54 * ttyGS0 provides login service, then ttyGS1 might provide dialer access
55 * for a telephone or fax link. And ttyGS2 might be something that just
56 * needs a simple byte stream interface for some messaging protocol that
57 * is managed in userspace ... OBEX, PTP, and MTP have been mentioned.
Richard Leitnerc572a212014-08-21 08:57:28 +020058 *
59 *
David Brownellc1dca562008-06-19 17:51:44 -070060 * gserial is the lifecycle interface, used by USB functions
61 * gs_port is the I/O nexus, used by the tty driver
62 * tty_struct links to the tty/filesystem framework
63 *
64 * gserial <---> gs_port ... links will be null when the USB link is
David Brownell1f1ba112008-08-06 18:49:57 -070065 * inactive; managed by gserial_{connect,disconnect}(). each gserial
66 * instance can wrap its own USB control protocol.
David Brownellc1dca562008-06-19 17:51:44 -070067 * gserial->ioport == usb_ep->driver_data ... gs_port
68 * gs_port->port_usb ... gserial
69 *
70 * gs_port <---> tty_struct ... links will be null when the TTY file
71 * isn't opened; managed by gs_open()/gs_close()
72 * gserial->port_tty ... tty_struct
73 * tty_struct->driver_data ... gserial
74 */
75
76/* RX and TX queues can buffer QUEUE_SIZE packets before they hit the
77 * next layer of buffering. For TX that's a circular buffer; for RX
78 * consider it a NOP. A third layer is provided by the TTY code.
79 */
80#define QUEUE_SIZE 16
81#define WRITE_BUF_SIZE 8192 /* TX only */
82
83/* circular buffer */
84struct gs_buf {
85 unsigned buf_size;
86 char *buf_buf;
87 char *buf_get;
88 char *buf_put;
89};
90
91/*
92 * The port structure holds info for each port, one for each minor number
93 * (and thus for each /dev/ node).
94 */
95struct gs_port {
Jiri Slaby266e37e2012-04-02 13:54:47 +020096 struct tty_port port;
David Brownellc1dca562008-06-19 17:51:44 -070097 spinlock_t port_lock; /* guard port_* access */
98
99 struct gserial *port_usb;
David Brownellc1dca562008-06-19 17:51:44 -0700100
David Brownellc1dca562008-06-19 17:51:44 -0700101 bool openclose; /* open/close in progress */
102 u8 port_num;
103
David Brownellc1dca562008-06-19 17:51:44 -0700104 struct list_head read_pool;
Jim Sung28609d402010-11-04 18:47:51 -0700105 int read_started;
106 int read_allocated;
David Brownell937ef732008-07-07 12:16:08 -0700107 struct list_head read_queue;
108 unsigned n_read;
David Brownellc1dca562008-06-19 17:51:44 -0700109 struct tasklet_struct push;
110
111 struct list_head write_pool;
Jim Sung28609d402010-11-04 18:47:51 -0700112 int write_started;
113 int write_allocated;
David Brownellc1dca562008-06-19 17:51:44 -0700114 struct gs_buf port_write_buf;
115 wait_queue_head_t drain_wait; /* wait while writes drain */
Philip Oberstaller3e9d3d22015-03-27 17:42:18 +0100116 bool write_busy;
Peter Hurleyb140dfe2015-10-10 16:00:53 -0400117 wait_queue_head_t close_wait;
David Brownellc1dca562008-06-19 17:51:44 -0700118
119 /* REVISIT this state ... */
120 struct usb_cdc_line_coding port_line_coding; /* 8-N-1 etc */
121};
122
David Brownellc1dca562008-06-19 17:51:44 -0700123static struct portmaster {
124 struct mutex lock; /* protect open/close */
125 struct gs_port *port;
Sebastian Andrzej Siewior19b10a82012-12-23 21:10:06 +0100126} ports[MAX_U_SERIAL_PORTS];
David Brownellc1dca562008-06-19 17:51:44 -0700127
128#define GS_CLOSE_TIMEOUT 15 /* seconds */
129
130
131
132#ifdef VERBOSE_DEBUG
Andrzej Pietrasiewiczea0e6272012-06-11 11:13:15 +0200133#ifndef pr_vdebug
David Brownellc1dca562008-06-19 17:51:44 -0700134#define pr_vdebug(fmt, arg...) \
135 pr_debug(fmt, ##arg)
Andrzej Pietrasiewiczea0e6272012-06-11 11:13:15 +0200136#endif /* pr_vdebug */
David Brownellc1dca562008-06-19 17:51:44 -0700137#else
Bo Shen273daf22013-03-13 16:54:07 +0800138#ifndef pr_vdebug
David Brownellc1dca562008-06-19 17:51:44 -0700139#define pr_vdebug(fmt, arg...) \
140 ({ if (0) pr_debug(fmt, ##arg); })
Andrzej Pietrasiewiczea0e6272012-06-11 11:13:15 +0200141#endif /* pr_vdebug */
David Brownellc1dca562008-06-19 17:51:44 -0700142#endif
143
144/*-------------------------------------------------------------------------*/
145
146/* Circular Buffer */
147
148/*
149 * gs_buf_alloc
150 *
151 * Allocate a circular buffer and all associated memory.
152 */
153static int gs_buf_alloc(struct gs_buf *gb, unsigned size)
154{
155 gb->buf_buf = kmalloc(size, GFP_KERNEL);
156 if (gb->buf_buf == NULL)
157 return -ENOMEM;
158
159 gb->buf_size = size;
160 gb->buf_put = gb->buf_buf;
161 gb->buf_get = gb->buf_buf;
162
163 return 0;
164}
165
166/*
167 * gs_buf_free
168 *
169 * Free the buffer and all associated memory.
170 */
171static void gs_buf_free(struct gs_buf *gb)
172{
173 kfree(gb->buf_buf);
174 gb->buf_buf = NULL;
175}
176
177/*
178 * gs_buf_clear
179 *
180 * Clear out all data in the circular buffer.
181 */
182static void gs_buf_clear(struct gs_buf *gb)
183{
184 gb->buf_get = gb->buf_put;
185 /* equivalent to a get of all data available */
186}
187
188/*
189 * gs_buf_data_avail
190 *
David Brownell1f1ba112008-08-06 18:49:57 -0700191 * Return the number of bytes of data written into the circular
David Brownellc1dca562008-06-19 17:51:44 -0700192 * buffer.
193 */
194static unsigned gs_buf_data_avail(struct gs_buf *gb)
195{
196 return (gb->buf_size + gb->buf_put - gb->buf_get) % gb->buf_size;
197}
198
199/*
200 * gs_buf_space_avail
201 *
202 * Return the number of bytes of space available in the circular
203 * buffer.
204 */
205static unsigned gs_buf_space_avail(struct gs_buf *gb)
206{
207 return (gb->buf_size + gb->buf_get - gb->buf_put - 1) % gb->buf_size;
208}
209
210/*
211 * gs_buf_put
212 *
213 * Copy data data from a user buffer and put it into the circular buffer.
214 * Restrict to the amount of space available.
215 *
216 * Return the number of bytes copied.
217 */
218static unsigned
219gs_buf_put(struct gs_buf *gb, const char *buf, unsigned count)
220{
221 unsigned len;
222
223 len = gs_buf_space_avail(gb);
224 if (count > len)
225 count = len;
226
227 if (count == 0)
228 return 0;
229
230 len = gb->buf_buf + gb->buf_size - gb->buf_put;
231 if (count > len) {
232 memcpy(gb->buf_put, buf, len);
233 memcpy(gb->buf_buf, buf+len, count - len);
234 gb->buf_put = gb->buf_buf + count - len;
235 } else {
236 memcpy(gb->buf_put, buf, count);
237 if (count < len)
238 gb->buf_put += count;
239 else /* count == len */
240 gb->buf_put = gb->buf_buf;
241 }
242
243 return count;
244}
245
246/*
247 * gs_buf_get
248 *
249 * Get data from the circular buffer and copy to the given buffer.
250 * Restrict to the amount of data available.
251 *
252 * Return the number of bytes copied.
253 */
254static unsigned
255gs_buf_get(struct gs_buf *gb, char *buf, unsigned count)
256{
257 unsigned len;
258
259 len = gs_buf_data_avail(gb);
260 if (count > len)
261 count = len;
262
263 if (count == 0)
264 return 0;
265
266 len = gb->buf_buf + gb->buf_size - gb->buf_get;
267 if (count > len) {
268 memcpy(buf, gb->buf_get, len);
269 memcpy(buf+len, gb->buf_buf, count - len);
270 gb->buf_get = gb->buf_buf + count - len;
271 } else {
272 memcpy(buf, gb->buf_get, count);
273 if (count < len)
274 gb->buf_get += count;
275 else /* count == len */
276 gb->buf_get = gb->buf_buf;
277 }
278
279 return count;
280}
281
282/*-------------------------------------------------------------------------*/
283
284/* I/O glue between TTY (upper) and USB function (lower) driver layers */
285
286/*
287 * gs_alloc_req
288 *
289 * Allocate a usb_request and its buffer. Returns a pointer to the
290 * usb_request or NULL if there is an error.
291 */
David Brownell1f1ba112008-08-06 18:49:57 -0700292struct usb_request *
David Brownellc1dca562008-06-19 17:51:44 -0700293gs_alloc_req(struct usb_ep *ep, unsigned len, gfp_t kmalloc_flags)
294{
295 struct usb_request *req;
296
297 req = usb_ep_alloc_request(ep, kmalloc_flags);
298
299 if (req != NULL) {
300 req->length = len;
301 req->buf = kmalloc(len, kmalloc_flags);
302 if (req->buf == NULL) {
303 usb_ep_free_request(ep, req);
304 return NULL;
305 }
306 }
307
308 return req;
309}
Sebastian Andrzej Siewior3249ca22012-12-23 21:10:04 +0100310EXPORT_SYMBOL_GPL(gs_alloc_req);
David Brownellc1dca562008-06-19 17:51:44 -0700311
312/*
313 * gs_free_req
314 *
315 * Free a usb_request and its buffer.
316 */
David Brownell1f1ba112008-08-06 18:49:57 -0700317void gs_free_req(struct usb_ep *ep, struct usb_request *req)
David Brownellc1dca562008-06-19 17:51:44 -0700318{
319 kfree(req->buf);
320 usb_ep_free_request(ep, req);
321}
Sebastian Andrzej Siewior3249ca22012-12-23 21:10:04 +0100322EXPORT_SYMBOL_GPL(gs_free_req);
David Brownellc1dca562008-06-19 17:51:44 -0700323
324/*
325 * gs_send_packet
326 *
327 * If there is data to send, a packet is built in the given
328 * buffer and the size is returned. If there is no data to
329 * send, 0 is returned.
330 *
331 * Called with port_lock held.
332 */
333static unsigned
334gs_send_packet(struct gs_port *port, char *packet, unsigned size)
335{
336 unsigned len;
337
338 len = gs_buf_data_avail(&port->port_write_buf);
339 if (len < size)
340 size = len;
341 if (size != 0)
342 size = gs_buf_get(&port->port_write_buf, packet, size);
343 return size;
344}
345
346/*
347 * gs_start_tx
348 *
349 * This function finds available write requests, calls
350 * gs_send_packet to fill these packets with data, and
351 * continues until either there are no more write requests
352 * available or no more data to send. This function is
353 * run whenever data arrives or write requests are available.
354 *
355 * Context: caller owns port_lock; port_usb is non-null.
356 */
357static int gs_start_tx(struct gs_port *port)
358/*
359__releases(&port->port_lock)
360__acquires(&port->port_lock)
361*/
362{
363 struct list_head *pool = &port->write_pool;
364 struct usb_ep *in = port->port_usb->in;
365 int status = 0;
366 bool do_tty_wake = false;
367
Philip Oberstaller3e9d3d22015-03-27 17:42:18 +0100368 while (!port->write_busy && !list_empty(pool)) {
David Brownellc1dca562008-06-19 17:51:44 -0700369 struct usb_request *req;
370 int len;
371
Jim Sung28609d402010-11-04 18:47:51 -0700372 if (port->write_started >= QUEUE_SIZE)
373 break;
374
David Brownellc1dca562008-06-19 17:51:44 -0700375 req = list_entry(pool->next, struct usb_request, list);
376 len = gs_send_packet(port, req->buf, in->maxpacket);
377 if (len == 0) {
378 wake_up_interruptible(&port->drain_wait);
379 break;
380 }
381 do_tty_wake = true;
382
383 req->length = len;
384 list_del(&req->list);
Daniel Glöckner2e251342009-05-28 12:53:24 +0200385 req->zero = (gs_buf_data_avail(&port->port_write_buf) == 0);
David Brownellc1dca562008-06-19 17:51:44 -0700386
Richard Leitnerc572a212014-08-21 08:57:28 +0200387 pr_vdebug("ttyGS%d: tx len=%d, 0x%02x 0x%02x 0x%02x ...\n",
388 port->port_num, len, *((u8 *)req->buf),
389 *((u8 *)req->buf+1), *((u8 *)req->buf+2));
David Brownellc1dca562008-06-19 17:51:44 -0700390
391 /* Drop lock while we call out of driver; completions
392 * could be issued while we do so. Disconnection may
393 * happen too; maybe immediately before we queue this!
394 *
395 * NOTE that we may keep sending data for a while after
396 * the TTY closed (dev->ioport->port_tty is NULL).
397 */
Philip Oberstaller3e9d3d22015-03-27 17:42:18 +0100398 port->write_busy = true;
David Brownellc1dca562008-06-19 17:51:44 -0700399 spin_unlock(&port->port_lock);
400 status = usb_ep_queue(in, req, GFP_ATOMIC);
401 spin_lock(&port->port_lock);
Philip Oberstaller3e9d3d22015-03-27 17:42:18 +0100402 port->write_busy = false;
David Brownellc1dca562008-06-19 17:51:44 -0700403
404 if (status) {
405 pr_debug("%s: %s %s err %d\n",
406 __func__, "queue", in->name, status);
407 list_add(&req->list, pool);
408 break;
409 }
410
Jim Sung28609d402010-11-04 18:47:51 -0700411 port->write_started++;
412
David Brownellc1dca562008-06-19 17:51:44 -0700413 /* abort immediately after disconnect */
414 if (!port->port_usb)
415 break;
416 }
417
Jiri Slaby35f95fd2012-04-02 13:54:48 +0200418 if (do_tty_wake && port->port.tty)
419 tty_wakeup(port->port.tty);
David Brownellc1dca562008-06-19 17:51:44 -0700420 return status;
421}
422
David Brownellc1dca562008-06-19 17:51:44 -0700423/*
424 * Context: caller owns port_lock, and port_usb is set
425 */
426static unsigned gs_start_rx(struct gs_port *port)
427/*
428__releases(&port->port_lock)
429__acquires(&port->port_lock)
430*/
431{
432 struct list_head *pool = &port->read_pool;
433 struct usb_ep *out = port->port_usb->out;
David Brownellc1dca562008-06-19 17:51:44 -0700434
435 while (!list_empty(pool)) {
436 struct usb_request *req;
437 int status;
438 struct tty_struct *tty;
439
David Brownell937ef732008-07-07 12:16:08 -0700440 /* no more rx if closed */
Jiri Slaby35f95fd2012-04-02 13:54:48 +0200441 tty = port->port.tty;
David Brownell937ef732008-07-07 12:16:08 -0700442 if (!tty)
David Brownellc1dca562008-06-19 17:51:44 -0700443 break;
444
Jim Sung28609d402010-11-04 18:47:51 -0700445 if (port->read_started >= QUEUE_SIZE)
446 break;
447
David Brownellc1dca562008-06-19 17:51:44 -0700448 req = list_entry(pool->next, struct usb_request, list);
449 list_del(&req->list);
450 req->length = out->maxpacket;
451
452 /* drop lock while we call out; the controller driver
453 * may need to call us back (e.g. for disconnect)
454 */
455 spin_unlock(&port->port_lock);
456 status = usb_ep_queue(out, req, GFP_ATOMIC);
457 spin_lock(&port->port_lock);
458
459 if (status) {
460 pr_debug("%s: %s %s err %d\n",
461 __func__, "queue", out->name, status);
462 list_add(&req->list, pool);
463 break;
464 }
Jim Sung28609d402010-11-04 18:47:51 -0700465 port->read_started++;
David Brownellc1dca562008-06-19 17:51:44 -0700466
467 /* abort immediately after disconnect */
468 if (!port->port_usb)
469 break;
470 }
Jim Sung28609d402010-11-04 18:47:51 -0700471 return port->read_started;
David Brownellc1dca562008-06-19 17:51:44 -0700472}
473
David Brownell937ef732008-07-07 12:16:08 -0700474/*
475 * RX tasklet takes data out of the RX queue and hands it up to the TTY
476 * layer until it refuses to take any more data (or is throttled back).
477 * Then it issues reads for any further data.
478 *
479 * If the RX queue becomes full enough that no usb_request is queued,
480 * the OUT endpoint may begin NAKing as soon as its FIFO fills up.
481 * So QUEUE_SIZE packets plus however many the FIFO holds (usually two)
482 * can be buffered before the TTY layer's buffers (currently 64 KB).
483 */
484static void gs_rx_push(unsigned long _port)
485{
486 struct gs_port *port = (void *)_port;
487 struct tty_struct *tty;
488 struct list_head *queue = &port->read_queue;
489 bool disconnect = false;
490 bool do_push = false;
491
492 /* hand any queued data to the tty */
493 spin_lock_irq(&port->port_lock);
Jiri Slaby35f95fd2012-04-02 13:54:48 +0200494 tty = port->port.tty;
David Brownell937ef732008-07-07 12:16:08 -0700495 while (!list_empty(queue)) {
496 struct usb_request *req;
497
498 req = list_first_entry(queue, struct usb_request, list);
499
David Brownell937ef732008-07-07 12:16:08 -0700500 /* leave data queued if tty was rx throttled */
Jiri Slaby05c7cd32013-01-03 15:53:04 +0100501 if (tty && test_bit(TTY_THROTTLED, &tty->flags))
David Brownell937ef732008-07-07 12:16:08 -0700502 break;
503
504 switch (req->status) {
505 case -ESHUTDOWN:
506 disconnect = true;
Richard Leitnerc572a212014-08-21 08:57:28 +0200507 pr_vdebug("ttyGS%d: shutdown\n", port->port_num);
David Brownell937ef732008-07-07 12:16:08 -0700508 break;
509
510 default:
511 /* presumably a transient fault */
Richard Leitnerc572a212014-08-21 08:57:28 +0200512 pr_warn("ttyGS%d: unexpected RX status %d\n",
513 port->port_num, req->status);
David Brownell937ef732008-07-07 12:16:08 -0700514 /* FALLTHROUGH */
515 case 0:
516 /* normal completion */
517 break;
518 }
519
520 /* push data to (open) tty */
521 if (req->actual) {
522 char *packet = req->buf;
523 unsigned size = req->actual;
524 unsigned n;
525 int count;
526
527 /* we may have pushed part of this packet already... */
528 n = port->n_read;
529 if (n) {
530 packet += n;
531 size -= n;
532 }
533
Jiri Slaby05c7cd32013-01-03 15:53:04 +0100534 count = tty_insert_flip_string(&port->port, packet,
535 size);
David Brownell937ef732008-07-07 12:16:08 -0700536 if (count)
537 do_push = true;
538 if (count != size) {
539 /* stop pushing; TTY layer can't handle more */
540 port->n_read += count;
Richard Leitnerc572a212014-08-21 08:57:28 +0200541 pr_vdebug("ttyGS%d: rx block %d/%d\n",
542 port->port_num, count, req->actual);
David Brownell937ef732008-07-07 12:16:08 -0700543 break;
544 }
545 port->n_read = 0;
546 }
Jiri Slaby05c7cd32013-01-03 15:53:04 +0100547
David Brownell937ef732008-07-07 12:16:08 -0700548 list_move(&req->list, &port->read_pool);
Jim Sung28609d402010-11-04 18:47:51 -0700549 port->read_started--;
David Brownell937ef732008-07-07 12:16:08 -0700550 }
551
Peter Hurleya9c3f682014-02-22 07:31:21 -0500552 /* Push from tty to ldisc; this is handled by a workqueue,
553 * so we won't get callbacks and can hold port_lock
David Brownell937ef732008-07-07 12:16:08 -0700554 */
Jiri Slaby2e124b42013-01-03 15:53:06 +0100555 if (do_push)
556 tty_flip_buffer_push(&port->port);
David Brownell937ef732008-07-07 12:16:08 -0700557
558
559 /* We want our data queue to become empty ASAP, keeping data
560 * in the tty and ldisc (not here). If we couldn't push any
561 * this time around, there may be trouble unless there's an
562 * implicit tty_unthrottle() call on its way...
563 *
564 * REVISIT we should probably add a timer to keep the tasklet
565 * from starving ... but it's not clear that case ever happens.
566 */
567 if (!list_empty(queue) && tty) {
568 if (!test_bit(TTY_THROTTLED, &tty->flags)) {
569 if (do_push)
570 tasklet_schedule(&port->push);
571 else
Richard Leitnerc572a212014-08-21 08:57:28 +0200572 pr_warn("ttyGS%d: RX not scheduled?\n",
David Brownell937ef732008-07-07 12:16:08 -0700573 port->port_num);
574 }
575 }
576
577 /* If we're still connected, refill the USB RX queue. */
578 if (!disconnect && port->port_usb)
579 gs_start_rx(port);
580
581 spin_unlock_irq(&port->port_lock);
582}
583
David Brownellc1dca562008-06-19 17:51:44 -0700584static void gs_read_complete(struct usb_ep *ep, struct usb_request *req)
585{
David Brownellc1dca562008-06-19 17:51:44 -0700586 struct gs_port *port = ep->driver_data;
587
David Brownell937ef732008-07-07 12:16:08 -0700588 /* Queue all received data until the tty layer is ready for it. */
David Brownellc1dca562008-06-19 17:51:44 -0700589 spin_lock(&port->port_lock);
David Brownell937ef732008-07-07 12:16:08 -0700590 list_add_tail(&req->list, &port->read_queue);
591 tasklet_schedule(&port->push);
David Brownellc1dca562008-06-19 17:51:44 -0700592 spin_unlock(&port->port_lock);
593}
594
595static void gs_write_complete(struct usb_ep *ep, struct usb_request *req)
596{
597 struct gs_port *port = ep->driver_data;
598
599 spin_lock(&port->port_lock);
600 list_add(&req->list, &port->write_pool);
Jim Sung28609d402010-11-04 18:47:51 -0700601 port->write_started--;
David Brownellc1dca562008-06-19 17:51:44 -0700602
603 switch (req->status) {
604 default:
605 /* presumably a transient fault */
606 pr_warning("%s: unexpected %s status %d\n",
607 __func__, ep->name, req->status);
608 /* FALL THROUGH */
609 case 0:
610 /* normal completion */
611 gs_start_tx(port);
612 break;
613
614 case -ESHUTDOWN:
615 /* disconnect */
616 pr_vdebug("%s: %s shutdown\n", __func__, ep->name);
617 break;
618 }
619
620 spin_unlock(&port->port_lock);
621}
622
Jim Sung28609d402010-11-04 18:47:51 -0700623static void gs_free_requests(struct usb_ep *ep, struct list_head *head,
624 int *allocated)
David Brownellc1dca562008-06-19 17:51:44 -0700625{
626 struct usb_request *req;
627
628 while (!list_empty(head)) {
629 req = list_entry(head->next, struct usb_request, list);
630 list_del(&req->list);
631 gs_free_req(ep, req);
Jim Sung28609d402010-11-04 18:47:51 -0700632 if (allocated)
633 (*allocated)--;
David Brownellc1dca562008-06-19 17:51:44 -0700634 }
635}
636
637static int gs_alloc_requests(struct usb_ep *ep, struct list_head *head,
Jim Sung28609d402010-11-04 18:47:51 -0700638 void (*fn)(struct usb_ep *, struct usb_request *),
639 int *allocated)
David Brownellc1dca562008-06-19 17:51:44 -0700640{
641 int i;
642 struct usb_request *req;
Jim Sung28609d402010-11-04 18:47:51 -0700643 int n = allocated ? QUEUE_SIZE - *allocated : QUEUE_SIZE;
David Brownellc1dca562008-06-19 17:51:44 -0700644
645 /* Pre-allocate up to QUEUE_SIZE transfers, but if we can't
646 * do quite that many this time, don't fail ... we just won't
647 * be as speedy as we might otherwise be.
648 */
Jim Sung28609d402010-11-04 18:47:51 -0700649 for (i = 0; i < n; i++) {
David Brownellc1dca562008-06-19 17:51:44 -0700650 req = gs_alloc_req(ep, ep->maxpacket, GFP_ATOMIC);
651 if (!req)
652 return list_empty(head) ? -ENOMEM : 0;
653 req->complete = fn;
654 list_add_tail(&req->list, head);
Jim Sung28609d402010-11-04 18:47:51 -0700655 if (allocated)
656 (*allocated)++;
David Brownellc1dca562008-06-19 17:51:44 -0700657 }
658 return 0;
659}
660
661/**
662 * gs_start_io - start USB I/O streams
663 * @dev: encapsulates endpoints to use
664 * Context: holding port_lock; port_tty and port_usb are non-null
665 *
666 * We only start I/O when something is connected to both sides of
667 * this port. If nothing is listening on the host side, we may
668 * be pointlessly filling up our TX buffers and FIFO.
669 */
670static int gs_start_io(struct gs_port *port)
671{
672 struct list_head *head = &port->read_pool;
673 struct usb_ep *ep = port->port_usb->out;
674 int status;
675 unsigned started;
676
677 /* Allocate RX and TX I/O buffers. We can't easily do this much
678 * earlier (with GFP_KERNEL) because the requests are coupled to
679 * endpoints, as are the packet sizes we'll be using. Different
680 * configurations may use different endpoints with a given port;
681 * and high speed vs full speed changes packet sizes too.
682 */
Jim Sung28609d402010-11-04 18:47:51 -0700683 status = gs_alloc_requests(ep, head, gs_read_complete,
684 &port->read_allocated);
David Brownellc1dca562008-06-19 17:51:44 -0700685 if (status)
686 return status;
687
688 status = gs_alloc_requests(port->port_usb->in, &port->write_pool,
Jim Sung28609d402010-11-04 18:47:51 -0700689 gs_write_complete, &port->write_allocated);
David Brownellc1dca562008-06-19 17:51:44 -0700690 if (status) {
Jim Sung28609d402010-11-04 18:47:51 -0700691 gs_free_requests(ep, head, &port->read_allocated);
David Brownellc1dca562008-06-19 17:51:44 -0700692 return status;
693 }
694
695 /* queue read requests */
David Brownell937ef732008-07-07 12:16:08 -0700696 port->n_read = 0;
David Brownellc1dca562008-06-19 17:51:44 -0700697 started = gs_start_rx(port);
698
699 /* unblock any pending writes into our circular buffer */
700 if (started) {
Jiri Slaby35f95fd2012-04-02 13:54:48 +0200701 tty_wakeup(port->port.tty);
David Brownellc1dca562008-06-19 17:51:44 -0700702 } else {
Jim Sung28609d402010-11-04 18:47:51 -0700703 gs_free_requests(ep, head, &port->read_allocated);
704 gs_free_requests(port->port_usb->in, &port->write_pool,
705 &port->write_allocated);
David Brownell937ef732008-07-07 12:16:08 -0700706 status = -EIO;
David Brownellc1dca562008-06-19 17:51:44 -0700707 }
708
David Brownell937ef732008-07-07 12:16:08 -0700709 return status;
David Brownellc1dca562008-06-19 17:51:44 -0700710}
711
712/*-------------------------------------------------------------------------*/
713
714/* TTY Driver */
715
716/*
717 * gs_open sets up the link between a gs_port and its associated TTY.
718 * That link is broken *only* by TTY close(), and all driver methods
719 * know that.
720 */
721static int gs_open(struct tty_struct *tty, struct file *file)
722{
723 int port_num = tty->index;
724 struct gs_port *port;
725 int status;
726
David Brownellc1dca562008-06-19 17:51:44 -0700727 do {
728 mutex_lock(&ports[port_num].lock);
729 port = ports[port_num].port;
730 if (!port)
731 status = -ENODEV;
732 else {
733 spin_lock_irq(&port->port_lock);
734
735 /* already open? Great. */
Jiri Slaby266e37e2012-04-02 13:54:47 +0200736 if (port->port.count) {
David Brownellc1dca562008-06-19 17:51:44 -0700737 status = 0;
Jiri Slaby266e37e2012-04-02 13:54:47 +0200738 port->port.count++;
David Brownellc1dca562008-06-19 17:51:44 -0700739
740 /* currently opening/closing? wait ... */
741 } else if (port->openclose) {
742 status = -EBUSY;
743
744 /* ... else we do the work */
745 } else {
746 status = -EAGAIN;
747 port->openclose = true;
748 }
749 spin_unlock_irq(&port->port_lock);
750 }
751 mutex_unlock(&ports[port_num].lock);
752
753 switch (status) {
754 default:
755 /* fully handled */
756 return status;
757 case -EAGAIN:
758 /* must do the work */
759 break;
760 case -EBUSY:
761 /* wait for EAGAIN task to finish */
762 msleep(1);
763 /* REVISIT could have a waitchannel here, if
764 * concurrent open performance is important
765 */
766 break;
767 }
768 } while (status != -EAGAIN);
769
770 /* Do the "real open" */
771 spin_lock_irq(&port->port_lock);
772
773 /* allocate circular buffer on first open */
774 if (port->port_write_buf.buf_buf == NULL) {
775
776 spin_unlock_irq(&port->port_lock);
777 status = gs_buf_alloc(&port->port_write_buf, WRITE_BUF_SIZE);
778 spin_lock_irq(&port->port_lock);
779
780 if (status) {
781 pr_debug("gs_open: ttyGS%d (%p,%p) no buffer\n",
782 port->port_num, tty, file);
783 port->openclose = false;
784 goto exit_unlock_port;
785 }
786 }
787
788 /* REVISIT if REMOVED (ports[].port NULL), abort the open
789 * to let rmmod work faster (but this way isn't wrong).
790 */
791
792 /* REVISIT maybe wait for "carrier detect" */
793
794 tty->driver_data = port;
Jiri Slaby35f95fd2012-04-02 13:54:48 +0200795 port->port.tty = tty;
David Brownellc1dca562008-06-19 17:51:44 -0700796
Jiri Slaby266e37e2012-04-02 13:54:47 +0200797 port->port.count = 1;
David Brownellc1dca562008-06-19 17:51:44 -0700798 port->openclose = false;
799
David Brownellc1dca562008-06-19 17:51:44 -0700800 /* if connected, start the I/O stream */
801 if (port->port_usb) {
David Brownell1f1ba112008-08-06 18:49:57 -0700802 struct gserial *gser = port->port_usb;
803
David Brownellc1dca562008-06-19 17:51:44 -0700804 pr_debug("gs_open: start ttyGS%d\n", port->port_num);
805 gs_start_io(port);
806
David Brownell1f1ba112008-08-06 18:49:57 -0700807 if (gser->connect)
808 gser->connect(gser);
David Brownellc1dca562008-06-19 17:51:44 -0700809 }
810
811 pr_debug("gs_open: ttyGS%d (%p,%p)\n", port->port_num, tty, file);
812
813 status = 0;
814
815exit_unlock_port:
816 spin_unlock_irq(&port->port_lock);
817 return status;
818}
819
820static int gs_writes_finished(struct gs_port *p)
821{
822 int cond;
823
824 /* return true on disconnect or empty buffer */
825 spin_lock_irq(&p->port_lock);
826 cond = (p->port_usb == NULL) || !gs_buf_data_avail(&p->port_write_buf);
827 spin_unlock_irq(&p->port_lock);
828
829 return cond;
830}
831
832static void gs_close(struct tty_struct *tty, struct file *file)
833{
834 struct gs_port *port = tty->driver_data;
David Brownell1f1ba112008-08-06 18:49:57 -0700835 struct gserial *gser;
David Brownellc1dca562008-06-19 17:51:44 -0700836
837 spin_lock_irq(&port->port_lock);
838
Jiri Slaby266e37e2012-04-02 13:54:47 +0200839 if (port->port.count != 1) {
840 if (port->port.count == 0)
David Brownellc1dca562008-06-19 17:51:44 -0700841 WARN_ON(1);
842 else
Jiri Slaby266e37e2012-04-02 13:54:47 +0200843 --port->port.count;
David Brownellc1dca562008-06-19 17:51:44 -0700844 goto exit;
845 }
846
847 pr_debug("gs_close: ttyGS%d (%p,%p) ...\n", port->port_num, tty, file);
848
849 /* mark port as closing but in use; we can drop port lock
850 * and sleep if necessary
851 */
852 port->openclose = true;
Jiri Slaby266e37e2012-04-02 13:54:47 +0200853 port->port.count = 0;
David Brownellc1dca562008-06-19 17:51:44 -0700854
David Brownell1f1ba112008-08-06 18:49:57 -0700855 gser = port->port_usb;
856 if (gser && gser->disconnect)
857 gser->disconnect(gser);
David Brownellc1dca562008-06-19 17:51:44 -0700858
859 /* wait for circular write buffer to drain, disconnect, or at
860 * most GS_CLOSE_TIMEOUT seconds; then discard the rest
861 */
David Brownell1f1ba112008-08-06 18:49:57 -0700862 if (gs_buf_data_avail(&port->port_write_buf) > 0 && gser) {
David Brownellc1dca562008-06-19 17:51:44 -0700863 spin_unlock_irq(&port->port_lock);
864 wait_event_interruptible_timeout(port->drain_wait,
865 gs_writes_finished(port),
866 GS_CLOSE_TIMEOUT * HZ);
867 spin_lock_irq(&port->port_lock);
David Brownell1f1ba112008-08-06 18:49:57 -0700868 gser = port->port_usb;
David Brownellc1dca562008-06-19 17:51:44 -0700869 }
870
871 /* Iff we're disconnected, there can be no I/O in flight so it's
872 * ok to free the circular buffer; else just scrub it. And don't
873 * let the push tasklet fire again until we're re-opened.
874 */
David Brownell1f1ba112008-08-06 18:49:57 -0700875 if (gser == NULL)
David Brownellc1dca562008-06-19 17:51:44 -0700876 gs_buf_free(&port->port_write_buf);
877 else
878 gs_buf_clear(&port->port_write_buf);
879
Jiri Slaby35f95fd2012-04-02 13:54:48 +0200880 port->port.tty = NULL;
David Brownellc1dca562008-06-19 17:51:44 -0700881
882 port->openclose = false;
883
884 pr_debug("gs_close: ttyGS%d (%p,%p) done!\n",
885 port->port_num, tty, file);
886
Peter Hurleyb140dfe2015-10-10 16:00:53 -0400887 wake_up(&port->close_wait);
David Brownellc1dca562008-06-19 17:51:44 -0700888exit:
889 spin_unlock_irq(&port->port_lock);
890}
891
892static int gs_write(struct tty_struct *tty, const unsigned char *buf, int count)
893{
894 struct gs_port *port = tty->driver_data;
895 unsigned long flags;
896 int status;
897
898 pr_vdebug("gs_write: ttyGS%d (%p) writing %d bytes\n",
899 port->port_num, tty, count);
900
901 spin_lock_irqsave(&port->port_lock, flags);
902 if (count)
903 count = gs_buf_put(&port->port_write_buf, buf, count);
904 /* treat count == 0 as flush_chars() */
905 if (port->port_usb)
906 status = gs_start_tx(port);
907 spin_unlock_irqrestore(&port->port_lock, flags);
908
909 return count;
910}
911
912static int gs_put_char(struct tty_struct *tty, unsigned char ch)
913{
914 struct gs_port *port = tty->driver_data;
915 unsigned long flags;
916 int status;
917
Scott Wood16d9efa2015-03-12 16:46:01 -0500918 pr_vdebug("gs_put_char: (%d,%p) char=0x%x, called from %ps\n",
David Brownellc1dca562008-06-19 17:51:44 -0700919 port->port_num, tty, ch, __builtin_return_address(0));
920
921 spin_lock_irqsave(&port->port_lock, flags);
922 status = gs_buf_put(&port->port_write_buf, &ch, 1);
923 spin_unlock_irqrestore(&port->port_lock, flags);
924
925 return status;
926}
927
928static void gs_flush_chars(struct tty_struct *tty)
929{
930 struct gs_port *port = tty->driver_data;
931 unsigned long flags;
932
933 pr_vdebug("gs_flush_chars: (%d,%p)\n", port->port_num, tty);
934
935 spin_lock_irqsave(&port->port_lock, flags);
936 if (port->port_usb)
937 gs_start_tx(port);
938 spin_unlock_irqrestore(&port->port_lock, flags);
939}
940
941static int gs_write_room(struct tty_struct *tty)
942{
943 struct gs_port *port = tty->driver_data;
944 unsigned long flags;
945 int room = 0;
946
947 spin_lock_irqsave(&port->port_lock, flags);
948 if (port->port_usb)
949 room = gs_buf_space_avail(&port->port_write_buf);
950 spin_unlock_irqrestore(&port->port_lock, flags);
951
952 pr_vdebug("gs_write_room: (%d,%p) room=%d\n",
953 port->port_num, tty, room);
954
955 return room;
956}
957
958static int gs_chars_in_buffer(struct tty_struct *tty)
959{
960 struct gs_port *port = tty->driver_data;
961 unsigned long flags;
962 int chars = 0;
963
964 spin_lock_irqsave(&port->port_lock, flags);
965 chars = gs_buf_data_avail(&port->port_write_buf);
966 spin_unlock_irqrestore(&port->port_lock, flags);
967
968 pr_vdebug("gs_chars_in_buffer: (%d,%p) chars=%d\n",
969 port->port_num, tty, chars);
970
971 return chars;
972}
973
974/* undo side effects of setting TTY_THROTTLED */
975static void gs_unthrottle(struct tty_struct *tty)
976{
977 struct gs_port *port = tty->driver_data;
978 unsigned long flags;
David Brownellc1dca562008-06-19 17:51:44 -0700979
980 spin_lock_irqsave(&port->port_lock, flags);
David Brownell937ef732008-07-07 12:16:08 -0700981 if (port->port_usb) {
982 /* Kickstart read queue processing. We don't do xon/xoff,
983 * rts/cts, or other handshaking with the host, but if the
984 * read queue backs up enough we'll be NAKing OUT packets.
985 */
986 tasklet_schedule(&port->push);
Richard Leitnerc572a212014-08-21 08:57:28 +0200987 pr_vdebug("ttyGS%d: unthrottle\n", port->port_num);
David Brownell937ef732008-07-07 12:16:08 -0700988 }
David Brownellc1dca562008-06-19 17:51:44 -0700989 spin_unlock_irqrestore(&port->port_lock, flags);
David Brownellc1dca562008-06-19 17:51:44 -0700990}
991
David Brownell1f1ba112008-08-06 18:49:57 -0700992static int gs_break_ctl(struct tty_struct *tty, int duration)
993{
994 struct gs_port *port = tty->driver_data;
995 int status = 0;
996 struct gserial *gser;
997
998 pr_vdebug("gs_break_ctl: ttyGS%d, send break (%d) \n",
999 port->port_num, duration);
1000
1001 spin_lock_irq(&port->port_lock);
1002 gser = port->port_usb;
1003 if (gser && gser->send_break)
1004 status = gser->send_break(gser, duration);
1005 spin_unlock_irq(&port->port_lock);
1006
1007 return status;
1008}
1009
David Brownellc1dca562008-06-19 17:51:44 -07001010static const struct tty_operations gs_tty_ops = {
1011 .open = gs_open,
1012 .close = gs_close,
1013 .write = gs_write,
1014 .put_char = gs_put_char,
1015 .flush_chars = gs_flush_chars,
1016 .write_room = gs_write_room,
1017 .chars_in_buffer = gs_chars_in_buffer,
1018 .unthrottle = gs_unthrottle,
David Brownell1f1ba112008-08-06 18:49:57 -07001019 .break_ctl = gs_break_ctl,
David Brownellc1dca562008-06-19 17:51:44 -07001020};
1021
1022/*-------------------------------------------------------------------------*/
1023
1024static struct tty_driver *gs_tty_driver;
1025
Benoit Gobyc3c04b22012-05-10 10:08:01 +02001026static int
David Brownellc1dca562008-06-19 17:51:44 -07001027gs_port_alloc(unsigned port_num, struct usb_cdc_line_coding *coding)
1028{
1029 struct gs_port *port;
Sebastian Andrzej Siewior19b10a82012-12-23 21:10:06 +01001030 int ret = 0;
1031
1032 mutex_lock(&ports[port_num].lock);
1033 if (ports[port_num].port) {
1034 ret = -EBUSY;
1035 goto out;
1036 }
David Brownellc1dca562008-06-19 17:51:44 -07001037
1038 port = kzalloc(sizeof(struct gs_port), GFP_KERNEL);
Sebastian Andrzej Siewior19b10a82012-12-23 21:10:06 +01001039 if (port == NULL) {
1040 ret = -ENOMEM;
1041 goto out;
1042 }
David Brownellc1dca562008-06-19 17:51:44 -07001043
Jiri Slaby266e37e2012-04-02 13:54:47 +02001044 tty_port_init(&port->port);
David Brownellc1dca562008-06-19 17:51:44 -07001045 spin_lock_init(&port->port_lock);
David Brownellc1dca562008-06-19 17:51:44 -07001046 init_waitqueue_head(&port->drain_wait);
Peter Hurleyb140dfe2015-10-10 16:00:53 -04001047 init_waitqueue_head(&port->close_wait);
David Brownellc1dca562008-06-19 17:51:44 -07001048
1049 tasklet_init(&port->push, gs_rx_push, (unsigned long) port);
1050
1051 INIT_LIST_HEAD(&port->read_pool);
David Brownell937ef732008-07-07 12:16:08 -07001052 INIT_LIST_HEAD(&port->read_queue);
David Brownellc1dca562008-06-19 17:51:44 -07001053 INIT_LIST_HEAD(&port->write_pool);
1054
1055 port->port_num = port_num;
1056 port->port_line_coding = *coding;
1057
1058 ports[port_num].port = port;
Sebastian Andrzej Siewior19b10a82012-12-23 21:10:06 +01001059out:
1060 mutex_unlock(&ports[port_num].lock);
1061 return ret;
David Brownellc1dca562008-06-19 17:51:44 -07001062}
1063
1064static int gs_closed(struct gs_port *port)
1065{
1066 int cond;
1067
1068 spin_lock_irq(&port->port_lock);
Jiri Slaby266e37e2012-04-02 13:54:47 +02001069 cond = (port->port.count == 0) && !port->openclose;
David Brownellc1dca562008-06-19 17:51:44 -07001070 spin_unlock_irq(&port->port_lock);
1071 return cond;
1072}
1073
Sebastian Andrzej Siewior19b10a82012-12-23 21:10:06 +01001074static void gserial_free_port(struct gs_port *port)
David Brownellc1dca562008-06-19 17:51:44 -07001075{
Sebastian Andrzej Siewior19b10a82012-12-23 21:10:06 +01001076 tasklet_kill(&port->push);
1077 /* wait for old opens to finish */
Peter Hurleyb140dfe2015-10-10 16:00:53 -04001078 wait_event(port->close_wait, gs_closed(port));
Sebastian Andrzej Siewior19b10a82012-12-23 21:10:06 +01001079 WARN_ON(port->port_usb != NULL);
1080 tty_port_destroy(&port->port);
1081 kfree(port);
1082}
1083
1084void gserial_free_line(unsigned char port_num)
1085{
David Brownellc1dca562008-06-19 17:51:44 -07001086 struct gs_port *port;
1087
Sebastian Andrzej Siewior19b10a82012-12-23 21:10:06 +01001088 mutex_lock(&ports[port_num].lock);
1089 if (WARN_ON(!ports[port_num].port)) {
1090 mutex_unlock(&ports[port_num].lock);
David Brownellac90e362008-07-01 13:18:20 -07001091 return;
David Brownellc1dca562008-06-19 17:51:44 -07001092 }
Sebastian Andrzej Siewior19b10a82012-12-23 21:10:06 +01001093 port = ports[port_num].port;
1094 ports[port_num].port = NULL;
1095 mutex_unlock(&ports[port_num].lock);
David Brownellc1dca562008-06-19 17:51:44 -07001096
Sebastian Andrzej Siewior19b10a82012-12-23 21:10:06 +01001097 gserial_free_port(port);
1098 tty_unregister_device(gs_tty_driver, port_num);
David Brownellc1dca562008-06-19 17:51:44 -07001099}
Sebastian Andrzej Siewior19b10a82012-12-23 21:10:06 +01001100EXPORT_SYMBOL_GPL(gserial_free_line);
1101
1102int gserial_alloc_line(unsigned char *line_num)
1103{
1104 struct usb_cdc_line_coding coding;
1105 struct device *tty_dev;
1106 int ret;
1107 int port_num;
1108
1109 coding.dwDTERate = cpu_to_le32(9600);
1110 coding.bCharFormat = 8;
1111 coding.bParityType = USB_CDC_NO_PARITY;
1112 coding.bDataBits = USB_CDC_1_STOP_BITS;
1113
1114 for (port_num = 0; port_num < MAX_U_SERIAL_PORTS; port_num++) {
1115 ret = gs_port_alloc(port_num, &coding);
1116 if (ret == -EBUSY)
1117 continue;
1118 if (ret)
1119 return ret;
1120 break;
1121 }
1122 if (ret)
1123 return ret;
1124
1125 /* ... and sysfs class devices, so mdev/udev make /dev/ttyGS* */
1126
1127 tty_dev = tty_port_register_device(&ports[port_num].port->port,
1128 gs_tty_driver, port_num, NULL);
1129 if (IS_ERR(tty_dev)) {
1130 struct gs_port *port;
1131 pr_err("%s: failed to register tty for port %d, err %ld\n",
1132 __func__, port_num, PTR_ERR(tty_dev));
1133
1134 ret = PTR_ERR(tty_dev);
1135 port = ports[port_num].port;
1136 ports[port_num].port = NULL;
1137 gserial_free_port(port);
1138 goto err;
1139 }
1140 *line_num = port_num;
1141err:
1142 return ret;
1143}
1144EXPORT_SYMBOL_GPL(gserial_alloc_line);
David Brownellc1dca562008-06-19 17:51:44 -07001145
1146/**
1147 * gserial_connect - notify TTY I/O glue that USB link is active
1148 * @gser: the function, set up with endpoints and descriptors
1149 * @port_num: which port is active
1150 * Context: any (usually from irq)
1151 *
1152 * This is called activate endpoints and let the TTY layer know that
1153 * the connection is active ... not unlike "carrier detect". It won't
1154 * necessarily start I/O queues; unless the TTY is held open by any
1155 * task, there would be no point. However, the endpoints will be
1156 * activated so the USB host can perform I/O, subject to basic USB
1157 * hardware flow control.
1158 *
1159 * Caller needs to have set up the endpoints and USB function in @dev
1160 * before calling this, as well as the appropriate (speed-specific)
Sebastian Andrzej Siewior19b10a82012-12-23 21:10:06 +01001161 * endpoint descriptors, and also have allocate @port_num by calling
1162 * @gserial_alloc_line().
David Brownellc1dca562008-06-19 17:51:44 -07001163 *
1164 * Returns negative errno or zero.
1165 * On success, ep->driver_data will be overwritten.
1166 */
1167int gserial_connect(struct gserial *gser, u8 port_num)
1168{
1169 struct gs_port *port;
1170 unsigned long flags;
1171 int status;
1172
Sebastian Andrzej Siewior19b10a82012-12-23 21:10:06 +01001173 if (port_num >= MAX_U_SERIAL_PORTS)
David Brownellc1dca562008-06-19 17:51:44 -07001174 return -ENXIO;
1175
David Brownellc1dca562008-06-19 17:51:44 -07001176 port = ports[port_num].port;
Sebastian Andrzej Siewior19b10a82012-12-23 21:10:06 +01001177 if (!port) {
1178 pr_err("serial line %d not allocated.\n", port_num);
1179 return -EINVAL;
1180 }
1181 if (port->port_usb) {
1182 pr_err("serial line %d is in use.\n", port_num);
1183 return -EBUSY;
1184 }
David Brownellc1dca562008-06-19 17:51:44 -07001185
1186 /* activate the endpoints */
Tatyana Brokhman72c973d2011-06-28 16:33:48 +03001187 status = usb_ep_enable(gser->in);
David Brownellc1dca562008-06-19 17:51:44 -07001188 if (status < 0)
1189 return status;
1190 gser->in->driver_data = port;
1191
Tatyana Brokhman72c973d2011-06-28 16:33:48 +03001192 status = usb_ep_enable(gser->out);
David Brownellc1dca562008-06-19 17:51:44 -07001193 if (status < 0)
1194 goto fail_out;
1195 gser->out->driver_data = port;
1196
1197 /* then tell the tty glue that I/O can work */
1198 spin_lock_irqsave(&port->port_lock, flags);
1199 gser->ioport = port;
1200 port->port_usb = gser;
1201
1202 /* REVISIT unclear how best to handle this state...
1203 * we don't really couple it with the Linux TTY.
1204 */
1205 gser->port_line_coding = port->port_line_coding;
1206
1207 /* REVISIT if waiting on "carrier detect", signal. */
1208
David Brownell1f1ba112008-08-06 18:49:57 -07001209 /* if it's already open, start I/O ... and notify the serial
1210 * protocol about open/close status (connect/disconnect).
David Brownellc1dca562008-06-19 17:51:44 -07001211 */
Jiri Slaby266e37e2012-04-02 13:54:47 +02001212 if (port->port.count) {
David Brownellc1dca562008-06-19 17:51:44 -07001213 pr_debug("gserial_connect: start ttyGS%d\n", port->port_num);
1214 gs_start_io(port);
David Brownell1f1ba112008-08-06 18:49:57 -07001215 if (gser->connect)
1216 gser->connect(gser);
1217 } else {
1218 if (gser->disconnect)
1219 gser->disconnect(gser);
David Brownellc1dca562008-06-19 17:51:44 -07001220 }
1221
1222 spin_unlock_irqrestore(&port->port_lock, flags);
1223
1224 return status;
1225
1226fail_out:
1227 usb_ep_disable(gser->in);
David Brownellc1dca562008-06-19 17:51:44 -07001228 return status;
1229}
Sebastian Andrzej Siewior3249ca22012-12-23 21:10:04 +01001230EXPORT_SYMBOL_GPL(gserial_connect);
David Brownellc1dca562008-06-19 17:51:44 -07001231/**
1232 * gserial_disconnect - notify TTY I/O glue that USB link is inactive
1233 * @gser: the function, on which gserial_connect() was called
1234 * Context: any (usually from irq)
1235 *
1236 * This is called to deactivate endpoints and let the TTY layer know
1237 * that the connection went inactive ... not unlike "hangup".
1238 *
1239 * On return, the state is as if gserial_connect() had never been called;
1240 * there is no active USB I/O on these endpoints.
1241 */
1242void gserial_disconnect(struct gserial *gser)
1243{
1244 struct gs_port *port = gser->ioport;
1245 unsigned long flags;
1246
1247 if (!port)
1248 return;
1249
1250 /* tell the TTY glue not to do I/O here any more */
1251 spin_lock_irqsave(&port->port_lock, flags);
1252
1253 /* REVISIT as above: how best to track this? */
1254 port->port_line_coding = gser->port_line_coding;
1255
1256 port->port_usb = NULL;
1257 gser->ioport = NULL;
Jiri Slaby266e37e2012-04-02 13:54:47 +02001258 if (port->port.count > 0 || port->openclose) {
David Brownellc1dca562008-06-19 17:51:44 -07001259 wake_up_interruptible(&port->drain_wait);
Jiri Slaby35f95fd2012-04-02 13:54:48 +02001260 if (port->port.tty)
1261 tty_hangup(port->port.tty);
David Brownellc1dca562008-06-19 17:51:44 -07001262 }
1263 spin_unlock_irqrestore(&port->port_lock, flags);
1264
1265 /* disable endpoints, aborting down any active I/O */
1266 usb_ep_disable(gser->out);
David Brownellc1dca562008-06-19 17:51:44 -07001267 usb_ep_disable(gser->in);
David Brownellc1dca562008-06-19 17:51:44 -07001268
1269 /* finally, free any unused/unusable I/O buffers */
1270 spin_lock_irqsave(&port->port_lock, flags);
Jiri Slaby266e37e2012-04-02 13:54:47 +02001271 if (port->port.count == 0 && !port->openclose)
David Brownellc1dca562008-06-19 17:51:44 -07001272 gs_buf_free(&port->port_write_buf);
Jim Sung28609d402010-11-04 18:47:51 -07001273 gs_free_requests(gser->out, &port->read_pool, NULL);
1274 gs_free_requests(gser->out, &port->read_queue, NULL);
1275 gs_free_requests(gser->in, &port->write_pool, NULL);
1276
1277 port->read_allocated = port->read_started =
1278 port->write_allocated = port->write_started = 0;
1279
David Brownellc1dca562008-06-19 17:51:44 -07001280 spin_unlock_irqrestore(&port->port_lock, flags);
1281}
Sebastian Andrzej Siewior3249ca22012-12-23 21:10:04 +01001282EXPORT_SYMBOL_GPL(gserial_disconnect);
1283
Felipe Balbi38b3ad52013-01-18 13:18:44 +02001284static int userial_init(void)
Sebastian Andrzej Siewior19b10a82012-12-23 21:10:06 +01001285{
1286 unsigned i;
1287 int status;
1288
1289 gs_tty_driver = alloc_tty_driver(MAX_U_SERIAL_PORTS);
1290 if (!gs_tty_driver)
1291 return -ENOMEM;
1292
1293 gs_tty_driver->driver_name = "g_serial";
Richard Leitnerc572a212014-08-21 08:57:28 +02001294 gs_tty_driver->name = "ttyGS";
Sebastian Andrzej Siewior19b10a82012-12-23 21:10:06 +01001295 /* uses dynamically assigned dev_t values */
1296
1297 gs_tty_driver->type = TTY_DRIVER_TYPE_SERIAL;
1298 gs_tty_driver->subtype = SERIAL_TYPE_NORMAL;
1299 gs_tty_driver->flags = TTY_DRIVER_REAL_RAW | TTY_DRIVER_DYNAMIC_DEV;
1300 gs_tty_driver->init_termios = tty_std_termios;
1301
1302 /* 9600-8-N-1 ... matches defaults expected by "usbser.sys" on
1303 * MS-Windows. Otherwise, most of these flags shouldn't affect
1304 * anything unless we were to actually hook up to a serial line.
1305 */
1306 gs_tty_driver->init_termios.c_cflag =
1307 B9600 | CS8 | CREAD | HUPCL | CLOCAL;
1308 gs_tty_driver->init_termios.c_ispeed = 9600;
1309 gs_tty_driver->init_termios.c_ospeed = 9600;
1310
1311 tty_set_operations(gs_tty_driver, &gs_tty_ops);
1312 for (i = 0; i < MAX_U_SERIAL_PORTS; i++)
1313 mutex_init(&ports[i].lock);
1314
1315 /* export the driver ... */
1316 status = tty_register_driver(gs_tty_driver);
1317 if (status) {
1318 pr_err("%s: cannot register, err %d\n",
1319 __func__, status);
1320 goto fail;
1321 }
1322
1323 pr_debug("%s: registered %d ttyGS* device%s\n", __func__,
1324 MAX_U_SERIAL_PORTS,
1325 (MAX_U_SERIAL_PORTS == 1) ? "" : "s");
1326
1327 return status;
1328fail:
1329 put_tty_driver(gs_tty_driver);
1330 gs_tty_driver = NULL;
1331 return status;
1332}
1333module_init(userial_init);
1334
1335static void userial_cleanup(void)
1336{
1337 tty_unregister_driver(gs_tty_driver);
1338 put_tty_driver(gs_tty_driver);
1339 gs_tty_driver = NULL;
1340}
1341module_exit(userial_cleanup);
1342
Sebastian Andrzej Siewior3249ca22012-12-23 21:10:04 +01001343MODULE_LICENSE("GPL");