blob: 83aabea259d7113d82d94d870bd04ebffd64b633 [file] [log] [blame]
Daniel Macke5779992010-03-04 19:46:13 +01001/*
2 * This program is free software; you can redistribute it and/or modify
3 * it under the terms of the GNU General Public License as published by
4 * the Free Software Foundation; either version 2 of the License, or
5 * (at your option) any later version.
6 *
7 * This program is distributed in the hope that it will be useful,
8 * but WITHOUT ANY WARRANTY; without even the implied warranty of
9 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
10 * GNU General Public License for more details.
11 *
12 * You should have received a copy of the GNU General Public License
13 * along with this program; if not, write to the Free Software
14 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
15 *
16 */
17
Daniel Mackc731bc92011-09-14 12:46:57 +020018#include <linux/gfp.h>
19#include <linux/init.h>
Takashi Iwai80c8a2a2012-01-09 11:37:20 +010020#include <linux/ratelimit.h>
Daniel Mackc731bc92011-09-14 12:46:57 +020021#include <linux/usb.h>
22#include <linux/usb/audio.h>
Daniel Mack8fdff6a2012-04-12 13:51:11 +020023#include <linux/slab.h>
Daniel Mackc731bc92011-09-14 12:46:57 +020024
25#include <sound/core.h>
26#include <sound/pcm.h>
Daniel Mack8fdff6a2012-04-12 13:51:11 +020027#include <sound/pcm_params.h>
Daniel Mackc731bc92011-09-14 12:46:57 +020028
29#include "usbaudio.h"
30#include "helper.h"
31#include "card.h"
32#include "endpoint.h"
33#include "pcm.h"
Daniel Mack2b58fd52012-09-04 10:23:07 +020034#include "quirks.h"
Daniel Mackc731bc92011-09-14 12:46:57 +020035
Daniel Mack8fdff6a2012-04-12 13:51:11 +020036#define EP_FLAG_RUNNING 1
Takashi Iwaif58161b2012-11-08 08:52:45 +010037#define EP_FLAG_STOPPING 2
Daniel Mack8fdff6a2012-04-12 13:51:11 +020038
Daniel Mackc731bc92011-09-14 12:46:57 +020039/*
Daniel Mack94c27212012-04-12 13:51:15 +020040 * snd_usb_endpoint is a model that abstracts everything related to an
41 * USB endpoint and its streaming.
42 *
43 * There are functions to activate and deactivate the streaming URBs and
Daniel Mack07a5e9d2012-04-24 19:31:24 +020044 * optional callbacks to let the pcm logic handle the actual content of the
Daniel Mack94c27212012-04-12 13:51:15 +020045 * packets for playback and record. Thus, the bus streaming and the audio
46 * handlers are fully decoupled.
47 *
Daniel Mack07a5e9d2012-04-24 19:31:24 +020048 * There are two different types of endpoints in audio applications.
Daniel Mack94c27212012-04-12 13:51:15 +020049 *
50 * SND_USB_ENDPOINT_TYPE_DATA handles full audio data payload for both
51 * inbound and outbound traffic.
52 *
Daniel Mack07a5e9d2012-04-24 19:31:24 +020053 * SND_USB_ENDPOINT_TYPE_SYNC endpoints are for inbound traffic only and
54 * expect the payload to carry Q10.14 / Q16.16 formatted sync information
55 * (3 or 4 bytes).
Daniel Mack94c27212012-04-12 13:51:15 +020056 *
Daniel Mack07a5e9d2012-04-24 19:31:24 +020057 * Each endpoint has to be configured prior to being used by calling
58 * snd_usb_endpoint_set_params().
Daniel Mack94c27212012-04-12 13:51:15 +020059 *
60 * The model incorporates a reference counting, so that multiple users
61 * can call snd_usb_endpoint_start() and snd_usb_endpoint_stop(), and
62 * only the first user will effectively start the URBs, and only the last
Daniel Mack07a5e9d2012-04-24 19:31:24 +020063 * one to stop it will tear the URBs down again.
Daniel Mack94c27212012-04-12 13:51:15 +020064 */
65
66/*
Daniel Mackc731bc92011-09-14 12:46:57 +020067 * convert a sampling rate into our full speed format (fs/1000 in Q16.16)
68 * this will overflow at approx 524 kHz
69 */
70static inline unsigned get_usb_full_speed_rate(unsigned int rate)
71{
72 return ((rate << 13) + 62) / 125;
73}
74
75/*
76 * convert a sampling rate into USB high speed format (fs/8000 in Q16.16)
77 * this will overflow at approx 4 MHz
78 */
79static inline unsigned get_usb_high_speed_rate(unsigned int rate)
80{
81 return ((rate << 10) + 62) / 125;
82}
83
84/*
Daniel Mackc731bc92011-09-14 12:46:57 +020085 * release a urb data
86 */
87static void release_urb_ctx(struct snd_urb_ctx *u)
88{
Daniel Mackd399ff92012-04-12 13:51:13 +020089 if (u->buffer_size)
90 usb_free_coherent(u->ep->chip->dev, u->buffer_size,
91 u->urb->transfer_buffer,
92 u->urb->transfer_dma);
93 usb_free_urb(u->urb);
94 u->urb = NULL;
Daniel Mackc731bc92011-09-14 12:46:57 +020095}
96
97static const char *usb_error_string(int err)
98{
99 switch (err) {
100 case -ENODEV:
101 return "no device";
102 case -ENOENT:
103 return "endpoint not enabled";
104 case -EPIPE:
105 return "endpoint stalled";
106 case -ENOSPC:
107 return "not enough bandwidth";
108 case -ESHUTDOWN:
109 return "device disabled";
110 case -EHOSTUNREACH:
111 return "device suspended";
112 case -EINVAL:
113 case -EAGAIN:
114 case -EFBIG:
115 case -EMSGSIZE:
116 return "internal error";
117 default:
118 return "unknown error";
119 }
120}
121
Daniel Mack94c27212012-04-12 13:51:15 +0200122/**
123 * snd_usb_endpoint_implicit_feedback_sink: Report endpoint usage type
124 *
Daniel Mack07a5e9d2012-04-24 19:31:24 +0200125 * @ep: The snd_usb_endpoint
Daniel Mack94c27212012-04-12 13:51:15 +0200126 *
127 * Determine whether an endpoint is driven by an implicit feedback
128 * data endpoint source.
129 */
Eldad Zack98ae4722013-04-03 23:18:52 +0200130int snd_usb_endpoint_implicit_feedback_sink(struct snd_usb_endpoint *ep)
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200131{
132 return ep->sync_master &&
133 ep->sync_master->type == SND_USB_ENDPOINT_TYPE_DATA &&
134 ep->type == SND_USB_ENDPOINT_TYPE_DATA &&
135 usb_pipeout(ep->pipe);
136}
137
Daniel Mack94c27212012-04-12 13:51:15 +0200138/*
139 * For streaming based on information derived from sync endpoints,
140 * prepare_outbound_urb_sizes() will call next_packet_size() to
141 * determine the number of samples to be sent in the next packet.
142 *
143 * For implicit feedback, next_packet_size() is unused.
144 */
Daniel Mack245baf92012-08-30 18:52:30 +0200145int snd_usb_endpoint_next_packet_size(struct snd_usb_endpoint *ep)
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200146{
147 unsigned long flags;
148 int ret;
149
150 if (ep->fill_max)
151 return ep->maxframesize;
152
153 spin_lock_irqsave(&ep->lock, flags);
154 ep->phase = (ep->phase & 0xffff)
155 + (ep->freqm << ep->datainterval);
156 ret = min(ep->phase >> 16, ep->maxframesize);
157 spin_unlock_irqrestore(&ep->lock, flags);
158
159 return ret;
160}
161
162static void retire_outbound_urb(struct snd_usb_endpoint *ep,
163 struct snd_urb_ctx *urb_ctx)
164{
165 if (ep->retire_data_urb)
166 ep->retire_data_urb(ep->data_subs, urb_ctx->urb);
167}
168
169static void retire_inbound_urb(struct snd_usb_endpoint *ep,
170 struct snd_urb_ctx *urb_ctx)
171{
172 struct urb *urb = urb_ctx->urb;
173
Daniel Mack2b58fd52012-09-04 10:23:07 +0200174 if (unlikely(ep->skip_packets > 0)) {
175 ep->skip_packets--;
176 return;
177 }
178
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200179 if (ep->sync_slave)
180 snd_usb_handle_sync_urb(ep->sync_slave, ep, urb);
181
182 if (ep->retire_data_urb)
183 ep->retire_data_urb(ep->data_subs, urb);
184}
185
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200186/*
187 * Prepare a PLAYBACK urb for submission to the bus.
188 */
189static void prepare_outbound_urb(struct snd_usb_endpoint *ep,
190 struct snd_urb_ctx *ctx)
191{
192 int i;
193 struct urb *urb = ctx->urb;
194 unsigned char *cp = urb->transfer_buffer;
195
196 urb->dev = ep->chip->dev; /* we need to set this at each time */
197
198 switch (ep->type) {
199 case SND_USB_ENDPOINT_TYPE_DATA:
200 if (ep->prepare_data_urb) {
201 ep->prepare_data_urb(ep->data_subs, urb);
202 } else {
203 /* no data provider, so send silence */
204 unsigned int offs = 0;
205 for (i = 0; i < ctx->packets; ++i) {
Daniel Mack8dce30c2012-09-27 10:26:01 +0200206 int counts;
207
208 if (ctx->packet_size[i])
209 counts = ctx->packet_size[i];
210 else
211 counts = snd_usb_endpoint_next_packet_size(ep);
212
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200213 urb->iso_frame_desc[i].offset = offs * ep->stride;
214 urb->iso_frame_desc[i].length = counts * ep->stride;
215 offs += counts;
216 }
217
218 urb->number_of_packets = ctx->packets;
219 urb->transfer_buffer_length = offs * ep->stride;
220 memset(urb->transfer_buffer, ep->silence_value,
221 offs * ep->stride);
222 }
223 break;
224
225 case SND_USB_ENDPOINT_TYPE_SYNC:
226 if (snd_usb_get_speed(ep->chip->dev) >= USB_SPEED_HIGH) {
227 /*
228 * fill the length and offset of each urb descriptor.
229 * the fixed 12.13 frequency is passed as 16.16 through the pipe.
230 */
231 urb->iso_frame_desc[0].length = 4;
232 urb->iso_frame_desc[0].offset = 0;
233 cp[0] = ep->freqn;
234 cp[1] = ep->freqn >> 8;
235 cp[2] = ep->freqn >> 16;
236 cp[3] = ep->freqn >> 24;
237 } else {
238 /*
239 * fill the length and offset of each urb descriptor.
240 * the fixed 10.14 frequency is passed through the pipe.
241 */
242 urb->iso_frame_desc[0].length = 3;
243 urb->iso_frame_desc[0].offset = 0;
244 cp[0] = ep->freqn >> 2;
245 cp[1] = ep->freqn >> 10;
246 cp[2] = ep->freqn >> 18;
247 }
248
249 break;
250 }
251}
252
253/*
254 * Prepare a CAPTURE or SYNC urb for submission to the bus.
255 */
256static inline void prepare_inbound_urb(struct snd_usb_endpoint *ep,
257 struct snd_urb_ctx *urb_ctx)
258{
259 int i, offs;
260 struct urb *urb = urb_ctx->urb;
261
262 urb->dev = ep->chip->dev; /* we need to set this at each time */
263
264 switch (ep->type) {
265 case SND_USB_ENDPOINT_TYPE_DATA:
266 offs = 0;
267 for (i = 0; i < urb_ctx->packets; i++) {
268 urb->iso_frame_desc[i].offset = offs;
269 urb->iso_frame_desc[i].length = ep->curpacksize;
270 offs += ep->curpacksize;
271 }
272
273 urb->transfer_buffer_length = offs;
274 urb->number_of_packets = urb_ctx->packets;
275 break;
276
277 case SND_USB_ENDPOINT_TYPE_SYNC:
278 urb->iso_frame_desc[0].length = min(4u, ep->syncmaxsize);
279 urb->iso_frame_desc[0].offset = 0;
280 break;
281 }
282}
283
Daniel Mack94c27212012-04-12 13:51:15 +0200284/*
Daniel Mack07a5e9d2012-04-24 19:31:24 +0200285 * Send output urbs that have been prepared previously. URBs are dequeued
Daniel Mack94c27212012-04-12 13:51:15 +0200286 * from ep->ready_playback_urbs and in case there there aren't any available
287 * or there are no packets that have been prepared, this function does
288 * nothing.
289 *
Daniel Mack07a5e9d2012-04-24 19:31:24 +0200290 * The reason why the functionality of sending and preparing URBs is separated
291 * is that host controllers don't guarantee the order in which they return
292 * inbound and outbound packets to their submitters.
Daniel Mack94c27212012-04-12 13:51:15 +0200293 *
294 * This function is only used for implicit feedback endpoints. For endpoints
Daniel Mack07a5e9d2012-04-24 19:31:24 +0200295 * driven by dedicated sync endpoints, URBs are immediately re-submitted
296 * from their completion handler.
Daniel Mack94c27212012-04-12 13:51:15 +0200297 */
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200298static void queue_pending_output_urbs(struct snd_usb_endpoint *ep)
299{
300 while (test_bit(EP_FLAG_RUNNING, &ep->flags)) {
301
302 unsigned long flags;
Andrew Morton68853fa2012-04-24 08:10:10 +0200303 struct snd_usb_packet_info *uninitialized_var(packet);
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200304 struct snd_urb_ctx *ctx = NULL;
305 struct urb *urb;
306 int err, i;
307
308 spin_lock_irqsave(&ep->lock, flags);
309 if (ep->next_packet_read_pos != ep->next_packet_write_pos) {
310 packet = ep->next_packet + ep->next_packet_read_pos;
311 ep->next_packet_read_pos++;
312 ep->next_packet_read_pos %= MAX_URBS;
313
314 /* take URB out of FIFO */
315 if (!list_empty(&ep->ready_playback_urbs))
316 ctx = list_first_entry(&ep->ready_playback_urbs,
317 struct snd_urb_ctx, ready_list);
318 }
319 spin_unlock_irqrestore(&ep->lock, flags);
320
321 if (ctx == NULL)
322 return;
323
324 list_del_init(&ctx->ready_list);
325 urb = ctx->urb;
326
327 /* copy over the length information */
328 for (i = 0; i < packet->packets; i++)
329 ctx->packet_size[i] = packet->packet_size[i];
330
Daniel Mack94c27212012-04-12 13:51:15 +0200331 /* call the data handler to fill in playback data */
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200332 prepare_outbound_urb(ep, ctx);
333
334 err = usb_submit_urb(ctx->urb, GFP_ATOMIC);
335 if (err < 0)
336 snd_printk(KERN_ERR "Unable to submit urb #%d: %d (urb %p)\n",
337 ctx->index, err, ctx->urb);
338 else
339 set_bit(ctx->index, &ep->active_mask);
340 }
341}
342
343/*
344 * complete callback for urbs
345 */
346static void snd_complete_urb(struct urb *urb)
347{
348 struct snd_urb_ctx *ctx = urb->context;
349 struct snd_usb_endpoint *ep = ctx->ep;
350 int err;
351
352 if (unlikely(urb->status == -ENOENT || /* unlinked */
353 urb->status == -ENODEV || /* device removed */
354 urb->status == -ECONNRESET || /* unlinked */
355 urb->status == -ESHUTDOWN || /* device disabled */
356 ep->chip->shutdown)) /* device disconnected */
357 goto exit_clear;
358
359 if (usb_pipeout(ep->pipe)) {
360 retire_outbound_urb(ep, ctx);
361 /* can be stopped during retire callback */
362 if (unlikely(!test_bit(EP_FLAG_RUNNING, &ep->flags)))
363 goto exit_clear;
364
Eldad Zack98ae4722013-04-03 23:18:52 +0200365 if (snd_usb_endpoint_implicit_feedback_sink(ep)) {
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200366 unsigned long flags;
367
368 spin_lock_irqsave(&ep->lock, flags);
369 list_add_tail(&ctx->ready_list, &ep->ready_playback_urbs);
370 spin_unlock_irqrestore(&ep->lock, flags);
371 queue_pending_output_urbs(ep);
372
373 goto exit_clear;
374 }
375
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200376 prepare_outbound_urb(ep, ctx);
377 } else {
378 retire_inbound_urb(ep, ctx);
379 /* can be stopped during retire callback */
380 if (unlikely(!test_bit(EP_FLAG_RUNNING, &ep->flags)))
381 goto exit_clear;
382
383 prepare_inbound_urb(ep, ctx);
384 }
385
386 err = usb_submit_urb(urb, GFP_ATOMIC);
387 if (err == 0)
388 return;
389
390 snd_printk(KERN_ERR "cannot submit urb (err = %d)\n", err);
391 //snd_pcm_stop(substream, SNDRV_PCM_STATE_XRUN);
392
393exit_clear:
394 clear_bit(ctx->index, &ep->active_mask);
395}
396
Daniel Mack94c27212012-04-12 13:51:15 +0200397/**
Daniel Mack07a5e9d2012-04-24 19:31:24 +0200398 * snd_usb_add_endpoint: Add an endpoint to an USB audio chip
Daniel Mack94c27212012-04-12 13:51:15 +0200399 *
400 * @chip: The chip
401 * @alts: The USB host interface
402 * @ep_num: The number of the endpoint to use
403 * @direction: SNDRV_PCM_STREAM_PLAYBACK or SNDRV_PCM_STREAM_CAPTURE
404 * @type: SND_USB_ENDPOINT_TYPE_DATA or SND_USB_ENDPOINT_TYPE_SYNC
405 *
406 * If the requested endpoint has not been added to the given chip before,
407 * a new instance is created. Otherwise, a pointer to the previoulsy
408 * created instance is returned. In case of any error, NULL is returned.
409 *
410 * New endpoints will be added to chip->ep_list and must be freed by
411 * calling snd_usb_endpoint_free().
412 */
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200413struct snd_usb_endpoint *snd_usb_add_endpoint(struct snd_usb_audio *chip,
414 struct usb_host_interface *alts,
415 int ep_num, int direction, int type)
416{
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200417 struct snd_usb_endpoint *ep;
Daniel Mack68e67f42012-07-12 13:08:40 +0200418 int is_playback = direction == SNDRV_PCM_STREAM_PLAYBACK;
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200419
Eldad Zacke7e58df2013-08-03 10:51:15 +0200420 if (WARN_ON(!alts))
421 return NULL;
422
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200423 mutex_lock(&chip->mutex);
424
Eldad Zack88766f02013-04-03 23:18:49 +0200425 list_for_each_entry(ep, &chip->ep_list, list) {
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200426 if (ep->ep_num == ep_num &&
427 ep->iface == alts->desc.bInterfaceNumber &&
Eldad Zackdf23a242013-10-06 22:31:13 +0200428 ep->altsetting == alts->desc.bAlternateSetting) {
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200429 snd_printdd(KERN_DEBUG "Re-using EP %x in iface %d,%d @%p\n",
Eldad Zackdf23a242013-10-06 22:31:13 +0200430 ep_num, ep->iface, ep->altsetting, ep);
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200431 goto __exit_unlock;
432 }
433 }
434
435 snd_printdd(KERN_DEBUG "Creating new %s %s endpoint #%x\n",
436 is_playback ? "playback" : "capture",
437 type == SND_USB_ENDPOINT_TYPE_DATA ? "data" : "sync",
438 ep_num);
439
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200440 ep = kzalloc(sizeof(*ep), GFP_KERNEL);
441 if (!ep)
442 goto __exit_unlock;
443
444 ep->chip = chip;
445 spin_lock_init(&ep->lock);
446 ep->type = type;
447 ep->ep_num = ep_num;
448 ep->iface = alts->desc.bInterfaceNumber;
Eldad Zackdf23a242013-10-06 22:31:13 +0200449 ep->altsetting = alts->desc.bAlternateSetting;
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200450 INIT_LIST_HEAD(&ep->ready_playback_urbs);
451 ep_num &= USB_ENDPOINT_NUMBER_MASK;
452
453 if (is_playback)
454 ep->pipe = usb_sndisocpipe(chip->dev, ep_num);
455 else
456 ep->pipe = usb_rcvisocpipe(chip->dev, ep_num);
457
458 if (type == SND_USB_ENDPOINT_TYPE_SYNC) {
459 if (get_endpoint(alts, 1)->bLength >= USB_DT_ENDPOINT_AUDIO_SIZE &&
460 get_endpoint(alts, 1)->bRefresh >= 1 &&
461 get_endpoint(alts, 1)->bRefresh <= 9)
462 ep->syncinterval = get_endpoint(alts, 1)->bRefresh;
463 else if (snd_usb_get_speed(chip->dev) == USB_SPEED_FULL)
464 ep->syncinterval = 1;
465 else if (get_endpoint(alts, 1)->bInterval >= 1 &&
466 get_endpoint(alts, 1)->bInterval <= 16)
467 ep->syncinterval = get_endpoint(alts, 1)->bInterval - 1;
468 else
469 ep->syncinterval = 3;
470
471 ep->syncmaxsize = le16_to_cpu(get_endpoint(alts, 1)->wMaxPacketSize);
472 }
473
474 list_add_tail(&ep->list, &chip->ep_list);
475
476__exit_unlock:
477 mutex_unlock(&chip->mutex);
478
479 return ep;
480}
481
482/*
483 * wait until all urbs are processed.
484 */
485static int wait_clear_urbs(struct snd_usb_endpoint *ep)
486{
487 unsigned long end_time = jiffies + msecs_to_jiffies(1000);
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200488 int alive;
489
490 do {
Joe Perches190006f2012-11-16 23:35:16 -0800491 alive = bitmap_weight(&ep->active_mask, ep->nurbs);
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200492 if (!alive)
493 break;
494
495 schedule_timeout_uninterruptible(1);
496 } while (time_before(jiffies, end_time));
497
498 if (alive)
499 snd_printk(KERN_ERR "timeout: still %d active urbs on EP #%x\n",
500 alive, ep->ep_num);
Takashi Iwaif58161b2012-11-08 08:52:45 +0100501 clear_bit(EP_FLAG_STOPPING, &ep->flags);
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200502
503 return 0;
504}
505
Takashi Iwaif58161b2012-11-08 08:52:45 +0100506/* sync the pending stop operation;
507 * this function itself doesn't trigger the stop operation
508 */
509void snd_usb_endpoint_sync_pending_stop(struct snd_usb_endpoint *ep)
510{
511 if (ep && test_bit(EP_FLAG_STOPPING, &ep->flags))
512 wait_clear_urbs(ep);
513}
514
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200515/*
516 * unlink active urbs.
517 */
Takashi Iwaiccc16962012-11-21 08:22:52 +0100518static int deactivate_urbs(struct snd_usb_endpoint *ep, bool force)
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200519{
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200520 unsigned int i;
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200521
522 if (!force && ep->chip->shutdown) /* to be sure... */
523 return -EBADFD;
524
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200525 clear_bit(EP_FLAG_RUNNING, &ep->flags);
526
527 INIT_LIST_HEAD(&ep->ready_playback_urbs);
528 ep->next_packet_read_pos = 0;
529 ep->next_packet_write_pos = 0;
530
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200531 for (i = 0; i < ep->nurbs; i++) {
532 if (test_bit(i, &ep->active_mask)) {
533 if (!test_and_set_bit(i, &ep->unlink_mask)) {
534 struct urb *u = ep->urb[i].urb;
Takashi Iwaiccc16962012-11-21 08:22:52 +0100535 usb_unlink_urb(u);
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200536 }
537 }
538 }
539
540 return 0;
541}
542
543/*
544 * release an endpoint's urbs
545 */
546static void release_urbs(struct snd_usb_endpoint *ep, int force)
547{
548 int i;
549
550 /* route incoming urbs to nirvana */
551 ep->retire_data_urb = NULL;
552 ep->prepare_data_urb = NULL;
553
554 /* stop urbs */
Takashi Iwaiccc16962012-11-21 08:22:52 +0100555 deactivate_urbs(ep, force);
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200556 wait_clear_urbs(ep);
557
558 for (i = 0; i < ep->nurbs; i++)
559 release_urb_ctx(&ep->urb[i]);
560
561 if (ep->syncbuf)
562 usb_free_coherent(ep->chip->dev, SYNC_URBS * 4,
563 ep->syncbuf, ep->sync_dma);
564
565 ep->syncbuf = NULL;
566 ep->nurbs = 0;
567}
568
Daniel Mack94c27212012-04-12 13:51:15 +0200569/*
570 * configure a data endpoint
571 */
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200572static int data_ep_set_params(struct snd_usb_endpoint *ep,
Dylan Reid35ec7aa22012-09-18 09:49:47 -0700573 snd_pcm_format_t pcm_format,
574 unsigned int channels,
575 unsigned int period_bytes,
Alan Stern976b6c02013-09-24 15:51:58 -0400576 unsigned int frames_per_period,
577 unsigned int periods_per_buffer,
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200578 struct audioformat *fmt,
579 struct snd_usb_endpoint *sync_ep)
580{
Alan Stern976b6c02013-09-24 15:51:58 -0400581 unsigned int maxsize, minsize, packs_per_ms, max_packs_per_urb;
582 unsigned int max_packs_per_period, urbs_per_period, urb_packs;
583 unsigned int max_urbs, i;
Dylan Reid35ec7aa22012-09-18 09:49:47 -0700584 int frame_bits = snd_pcm_format_physical_width(pcm_format) * channels;
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200585
Daniel Mackd24f5062013-04-17 00:01:38 +0800586 if (pcm_format == SNDRV_PCM_FORMAT_DSD_U16_LE && fmt->dsd_dop) {
587 /*
588 * When operating in DSD DOP mode, the size of a sample frame
589 * in hardware differs from the actual physical format width
590 * because we need to make room for the DOP markers.
591 */
592 frame_bits += channels << 3;
593 }
594
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200595 ep->datainterval = fmt->datainterval;
596 ep->stride = frame_bits >> 3;
Dylan Reid35ec7aa22012-09-18 09:49:47 -0700597 ep->silence_value = pcm_format == SNDRV_PCM_FORMAT_U8 ? 0x80 : 0;
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200598
Clemens Ladisch57e6dae2013-08-08 11:24:55 +0200599 /* assume max. frequency is 25% higher than nominal */
600 ep->freqmax = ep->freqn + (ep->freqn >> 2);
601 maxsize = ((ep->freqmax + 0xffff) * (frame_bits >> 3))
602 >> (16 - ep->datainterval);
603 /* but wMaxPacketSize might reduce this */
604 if (ep->maxpacksize && ep->maxpacksize < maxsize) {
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200605 /* whatever fits into a max. size packet */
606 maxsize = ep->maxpacksize;
607 ep->freqmax = (maxsize / (frame_bits >> 3))
608 << (16 - ep->datainterval);
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200609 }
610
611 if (ep->fill_max)
612 ep->curpacksize = ep->maxpacksize;
613 else
614 ep->curpacksize = maxsize;
615
Alan Stern976b6c02013-09-24 15:51:58 -0400616 if (snd_usb_get_speed(ep->chip->dev) != USB_SPEED_FULL) {
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200617 packs_per_ms = 8 >> ep->datainterval;
Alan Stern976b6c02013-09-24 15:51:58 -0400618 max_packs_per_urb = MAX_PACKS_HS;
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200619 } else {
Alan Stern976b6c02013-09-24 15:51:58 -0400620 packs_per_ms = 1;
621 max_packs_per_urb = MAX_PACKS;
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200622 }
Eldad Zack98ae4722013-04-03 23:18:52 +0200623 if (sync_ep && !snd_usb_endpoint_implicit_feedback_sink(ep))
Alan Stern976b6c02013-09-24 15:51:58 -0400624 max_packs_per_urb = min(max_packs_per_urb,
625 1U << sync_ep->syncinterval);
626 max_packs_per_urb = max(1u, max_packs_per_urb >> ep->datainterval);
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200627
Alan Stern976b6c02013-09-24 15:51:58 -0400628 /*
629 * Capture endpoints need to use small URBs because there's no way
630 * to tell in advance where the next period will end, and we don't
631 * want the next URB to complete much after the period ends.
632 *
633 * Playback endpoints with implicit sync much use the same parameters
634 * as their corresponding capture endpoint.
635 */
636 if (usb_pipein(ep->pipe) ||
637 snd_usb_endpoint_implicit_feedback_sink(ep)) {
638
Thomas Pugliesea93455e2013-11-26 13:58:15 -0600639 urb_packs = packs_per_ms;
640 /*
641 * Wireless devices can poll at a max rate of once per 4ms.
642 * For dataintervals less than 5, increase the packet count to
643 * allow the host controller to use bursting to fill in the
644 * gaps.
645 */
646 if (snd_usb_get_speed(ep->chip->dev) == USB_SPEED_WIRELESS) {
647 int interval = ep->datainterval;
648 while (interval < 5) {
649 urb_packs <<= 1;
650 ++interval;
651 }
652 }
Alan Stern976b6c02013-09-24 15:51:58 -0400653 /* make capture URBs <= 1 ms and smaller than a period */
Thomas Pugliesea93455e2013-11-26 13:58:15 -0600654 urb_packs = min(max_packs_per_urb, urb_packs);
Alan Stern976b6c02013-09-24 15:51:58 -0400655 while (urb_packs > 1 && urb_packs * maxsize >= period_bytes)
656 urb_packs >>= 1;
657 ep->nurbs = MAX_URBS;
658
659 /*
660 * Playback endpoints without implicit sync are adjusted so that
661 * a period fits as evenly as possible in the smallest number of
662 * URBs. The total number of URBs is adjusted to the size of the
663 * ALSA buffer, subject to the MAX_URBS and MAX_QUEUE limits.
664 */
665 } else {
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200666 /* determine how small a packet can be */
Alan Stern976b6c02013-09-24 15:51:58 -0400667 minsize = (ep->freqn >> (16 - ep->datainterval)) *
668 (frame_bits >> 3);
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200669 /* with sync from device, assume it can be 12% lower */
670 if (sync_ep)
671 minsize -= minsize >> 3;
672 minsize = max(minsize, 1u);
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200673
Alan Stern976b6c02013-09-24 15:51:58 -0400674 /* how many packets will contain an entire ALSA period? */
675 max_packs_per_period = DIV_ROUND_UP(period_bytes, minsize);
676
677 /* how many URBs will contain a period? */
678 urbs_per_period = DIV_ROUND_UP(max_packs_per_period,
679 max_packs_per_urb);
680 /* how many packets are needed in each URB? */
681 urb_packs = DIV_ROUND_UP(max_packs_per_period, urbs_per_period);
682
683 /* limit the number of frames in a single URB */
684 ep->max_urb_frames = DIV_ROUND_UP(frames_per_period,
685 urbs_per_period);
686
687 /* try to use enough URBs to contain an entire ALSA buffer */
688 max_urbs = min((unsigned) MAX_URBS,
689 MAX_QUEUE * packs_per_ms / urb_packs);
690 ep->nurbs = min(max_urbs, urbs_per_period * periods_per_buffer);
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200691 }
692
693 /* allocate and initialize data urbs */
694 for (i = 0; i < ep->nurbs; i++) {
695 struct snd_urb_ctx *u = &ep->urb[i];
696 u->index = i;
697 u->ep = ep;
Alan Stern976b6c02013-09-24 15:51:58 -0400698 u->packets = urb_packs;
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200699 u->buffer_size = maxsize * u->packets;
700
701 if (fmt->fmt_type == UAC_FORMAT_TYPE_II)
702 u->packets++; /* for transfer delimiter */
703 u->urb = usb_alloc_urb(u->packets, GFP_KERNEL);
704 if (!u->urb)
705 goto out_of_memory;
706
707 u->urb->transfer_buffer =
708 usb_alloc_coherent(ep->chip->dev, u->buffer_size,
709 GFP_KERNEL, &u->urb->transfer_dma);
710 if (!u->urb->transfer_buffer)
711 goto out_of_memory;
712 u->urb->pipe = ep->pipe;
Clemens Ladischc75c5ab2013-04-27 12:10:32 +0200713 u->urb->transfer_flags = URB_NO_TRANSFER_DMA_MAP;
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200714 u->urb->interval = 1 << ep->datainterval;
715 u->urb->context = u;
716 u->urb->complete = snd_complete_urb;
717 INIT_LIST_HEAD(&u->ready_list);
718 }
719
720 return 0;
721
722out_of_memory:
723 release_urbs(ep, 0);
724 return -ENOMEM;
725}
726
Daniel Mack94c27212012-04-12 13:51:15 +0200727/*
728 * configure a sync endpoint
729 */
Eldad Zack93721032013-10-06 22:31:06 +0200730static int sync_ep_set_params(struct snd_usb_endpoint *ep)
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200731{
732 int i;
733
734 ep->syncbuf = usb_alloc_coherent(ep->chip->dev, SYNC_URBS * 4,
735 GFP_KERNEL, &ep->sync_dma);
736 if (!ep->syncbuf)
737 return -ENOMEM;
738
739 for (i = 0; i < SYNC_URBS; i++) {
740 struct snd_urb_ctx *u = &ep->urb[i];
741 u->index = i;
742 u->ep = ep;
743 u->packets = 1;
744 u->urb = usb_alloc_urb(1, GFP_KERNEL);
745 if (!u->urb)
746 goto out_of_memory;
747 u->urb->transfer_buffer = ep->syncbuf + i * 4;
748 u->urb->transfer_dma = ep->sync_dma + i * 4;
749 u->urb->transfer_buffer_length = 4;
750 u->urb->pipe = ep->pipe;
Clemens Ladischc75c5ab2013-04-27 12:10:32 +0200751 u->urb->transfer_flags = URB_NO_TRANSFER_DMA_MAP;
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200752 u->urb->number_of_packets = 1;
753 u->urb->interval = 1 << ep->syncinterval;
754 u->urb->context = u;
755 u->urb->complete = snd_complete_urb;
756 }
757
758 ep->nurbs = SYNC_URBS;
759
760 return 0;
761
762out_of_memory:
763 release_urbs(ep, 0);
764 return -ENOMEM;
765}
766
Daniel Mack94c27212012-04-12 13:51:15 +0200767/**
Daniel Mack07a5e9d2012-04-24 19:31:24 +0200768 * snd_usb_endpoint_set_params: configure an snd_usb_endpoint
Daniel Mack94c27212012-04-12 13:51:15 +0200769 *
Daniel Mack07a5e9d2012-04-24 19:31:24 +0200770 * @ep: the snd_usb_endpoint to configure
Dylan Reid35ec7aa22012-09-18 09:49:47 -0700771 * @pcm_format: the audio fomat.
772 * @channels: the number of audio channels.
773 * @period_bytes: the number of bytes in one alsa period.
Alan Stern976b6c02013-09-24 15:51:58 -0400774 * @period_frames: the number of frames in one alsa period.
775 * @buffer_periods: the number of periods in one alsa buffer.
Dylan Reid35ec7aa22012-09-18 09:49:47 -0700776 * @rate: the frame rate.
Daniel Mack07a5e9d2012-04-24 19:31:24 +0200777 * @fmt: the USB audio format information
778 * @sync_ep: the sync endpoint to use, if any
Daniel Mack94c27212012-04-12 13:51:15 +0200779 *
Daniel Mack07a5e9d2012-04-24 19:31:24 +0200780 * Determine the number of URBs to be used on this endpoint.
Daniel Mack94c27212012-04-12 13:51:15 +0200781 * An endpoint must be configured before it can be started.
782 * An endpoint that is already running can not be reconfigured.
783 */
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200784int snd_usb_endpoint_set_params(struct snd_usb_endpoint *ep,
Dylan Reid35ec7aa22012-09-18 09:49:47 -0700785 snd_pcm_format_t pcm_format,
786 unsigned int channels,
787 unsigned int period_bytes,
Alan Stern976b6c02013-09-24 15:51:58 -0400788 unsigned int period_frames,
789 unsigned int buffer_periods,
Dylan Reid35ec7aa22012-09-18 09:49:47 -0700790 unsigned int rate,
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200791 struct audioformat *fmt,
792 struct snd_usb_endpoint *sync_ep)
793{
794 int err;
795
796 if (ep->use_count != 0) {
797 snd_printk(KERN_WARNING "Unable to change format on ep #%x: already in use\n",
798 ep->ep_num);
799 return -EBUSY;
800 }
801
802 /* release old buffers, if any */
803 release_urbs(ep, 0);
804
805 ep->datainterval = fmt->datainterval;
806 ep->maxpacksize = fmt->maxpacksize;
Takashi Iwai85f71932012-04-13 12:41:54 +0200807 ep->fill_max = !!(fmt->attributes & UAC_EP_CS_ATTR_FILL_MAX);
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200808
809 if (snd_usb_get_speed(ep->chip->dev) == USB_SPEED_FULL)
Dylan Reid35ec7aa22012-09-18 09:49:47 -0700810 ep->freqn = get_usb_full_speed_rate(rate);
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200811 else
Dylan Reid35ec7aa22012-09-18 09:49:47 -0700812 ep->freqn = get_usb_high_speed_rate(rate);
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200813
814 /* calculate the frequency in 16.16 format */
815 ep->freqm = ep->freqn;
816 ep->freqshift = INT_MIN;
817
818 ep->phase = 0;
819
820 switch (ep->type) {
821 case SND_USB_ENDPOINT_TYPE_DATA:
Dylan Reid35ec7aa22012-09-18 09:49:47 -0700822 err = data_ep_set_params(ep, pcm_format, channels,
Alan Stern976b6c02013-09-24 15:51:58 -0400823 period_bytes, period_frames,
824 buffer_periods, fmt, sync_ep);
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200825 break;
826 case SND_USB_ENDPOINT_TYPE_SYNC:
Eldad Zack93721032013-10-06 22:31:06 +0200827 err = sync_ep_set_params(ep);
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200828 break;
829 default:
830 err = -EINVAL;
831 }
832
833 snd_printdd(KERN_DEBUG "Setting params for ep #%x (type %d, %d urbs), ret=%d\n",
834 ep->ep_num, ep->type, ep->nurbs, err);
835
836 return err;
837}
838
Daniel Mack94c27212012-04-12 13:51:15 +0200839/**
840 * snd_usb_endpoint_start: start an snd_usb_endpoint
841 *
Daniel Mack015618b2012-08-29 13:17:05 +0200842 * @ep: the endpoint to start
843 * @can_sleep: flag indicating whether the operation is executed in
844 * non-atomic context
Daniel Mack94c27212012-04-12 13:51:15 +0200845 *
846 * A call to this function will increment the use count of the endpoint.
Daniel Mack07a5e9d2012-04-24 19:31:24 +0200847 * In case it is not already running, the URBs for this endpoint will be
Daniel Mack94c27212012-04-12 13:51:15 +0200848 * submitted. Otherwise, this function does nothing.
849 *
850 * Must be balanced to calls of snd_usb_endpoint_stop().
851 *
852 * Returns an error if the URB submission failed, 0 in all other cases.
853 */
Takashi Iwaia9bb3622012-11-20 18:32:06 +0100854int snd_usb_endpoint_start(struct snd_usb_endpoint *ep, bool can_sleep)
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200855{
856 int err;
857 unsigned int i;
858
859 if (ep->chip->shutdown)
860 return -EBADFD;
861
862 /* already running? */
863 if (++ep->use_count != 1)
864 return 0;
865
Daniel Mack015618b2012-08-29 13:17:05 +0200866 /* just to be sure */
Takashi Iwaiccc16962012-11-21 08:22:52 +0100867 deactivate_urbs(ep, false);
Daniel Mack015618b2012-08-29 13:17:05 +0200868 if (can_sleep)
869 wait_clear_urbs(ep);
870
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200871 ep->active_mask = 0;
872 ep->unlink_mask = 0;
873 ep->phase = 0;
874
Daniel Mack2b58fd52012-09-04 10:23:07 +0200875 snd_usb_endpoint_start_quirk(ep);
876
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200877 /*
878 * If this endpoint has a data endpoint as implicit feedback source,
879 * don't start the urbs here. Instead, mark them all as available,
Daniel Mack07a5e9d2012-04-24 19:31:24 +0200880 * wait for the record urbs to return and queue the playback urbs
881 * from that context.
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200882 */
883
884 set_bit(EP_FLAG_RUNNING, &ep->flags);
885
Eldad Zack98ae4722013-04-03 23:18:52 +0200886 if (snd_usb_endpoint_implicit_feedback_sink(ep)) {
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200887 for (i = 0; i < ep->nurbs; i++) {
888 struct snd_urb_ctx *ctx = ep->urb + i;
889 list_add_tail(&ctx->ready_list, &ep->ready_playback_urbs);
890 }
891
892 return 0;
893 }
894
895 for (i = 0; i < ep->nurbs; i++) {
896 struct urb *urb = ep->urb[i].urb;
897
898 if (snd_BUG_ON(!urb))
899 goto __error;
900
901 if (usb_pipeout(ep->pipe)) {
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200902 prepare_outbound_urb(ep, urb->context);
903 } else {
904 prepare_inbound_urb(ep, urb->context);
905 }
906
907 err = usb_submit_urb(urb, GFP_ATOMIC);
908 if (err < 0) {
909 snd_printk(KERN_ERR "cannot submit urb %d, error %d: %s\n",
910 i, err, usb_error_string(err));
911 goto __error;
912 }
913 set_bit(i, &ep->active_mask);
914 }
915
916 return 0;
917
918__error:
919 clear_bit(EP_FLAG_RUNNING, &ep->flags);
920 ep->use_count--;
Takashi Iwaiccc16962012-11-21 08:22:52 +0100921 deactivate_urbs(ep, false);
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200922 return -EPIPE;
923}
924
Daniel Mack94c27212012-04-12 13:51:15 +0200925/**
926 * snd_usb_endpoint_stop: stop an snd_usb_endpoint
927 *
928 * @ep: the endpoint to stop (may be NULL)
929 *
930 * A call to this function will decrement the use count of the endpoint.
931 * In case the last user has requested the endpoint stop, the URBs will
Daniel Mack07a5e9d2012-04-24 19:31:24 +0200932 * actually be deactivated.
Daniel Mack94c27212012-04-12 13:51:15 +0200933 *
934 * Must be balanced to calls of snd_usb_endpoint_start().
Takashi Iwaib2eb9502012-11-21 08:30:48 +0100935 *
936 * The caller needs to synchronize the pending stop operation via
937 * snd_usb_endpoint_sync_pending_stop().
Daniel Mack94c27212012-04-12 13:51:15 +0200938 */
Takashi Iwaib2eb9502012-11-21 08:30:48 +0100939void snd_usb_endpoint_stop(struct snd_usb_endpoint *ep)
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200940{
941 if (!ep)
942 return;
943
944 if (snd_BUG_ON(ep->use_count == 0))
945 return;
946
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200947 if (--ep->use_count == 0) {
Takashi Iwaiccc16962012-11-21 08:22:52 +0100948 deactivate_urbs(ep, false);
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200949 ep->data_subs = NULL;
950 ep->sync_slave = NULL;
951 ep->retire_data_urb = NULL;
952 ep->prepare_data_urb = NULL;
Takashi Iwaib2eb9502012-11-21 08:30:48 +0100953 set_bit(EP_FLAG_STOPPING, &ep->flags);
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200954 }
955}
956
Daniel Mack94c27212012-04-12 13:51:15 +0200957/**
Daniel Mack94c27212012-04-12 13:51:15 +0200958 * snd_usb_endpoint_deactivate: deactivate an snd_usb_endpoint
959 *
960 * @ep: the endpoint to deactivate
961 *
Eldad Zack9b7c5522013-10-06 22:31:10 +0200962 * If the endpoint is not currently in use, this functions will
963 * deactivate its associated URBs.
Daniel Mack94c27212012-04-12 13:51:15 +0200964 *
965 * In case of any active users, this functions does nothing.
Daniel Mack94c27212012-04-12 13:51:15 +0200966 */
Eldad Zack9b7c5522013-10-06 22:31:10 +0200967void snd_usb_endpoint_deactivate(struct snd_usb_endpoint *ep)
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200968{
969 if (!ep)
Eldad Zack9b7c5522013-10-06 22:31:10 +0200970 return;
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200971
972 if (ep->use_count != 0)
Eldad Zack9b7c5522013-10-06 22:31:10 +0200973 return;
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200974
Eldad Zack239b9f72013-10-06 22:31:09 +0200975 deactivate_urbs(ep, true);
976 wait_clear_urbs(ep);
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200977}
978
Daniel Mack07a5e9d2012-04-24 19:31:24 +0200979/**
980 * snd_usb_endpoint_free: Free the resources of an snd_usb_endpoint
Daniel Mack94c27212012-04-12 13:51:15 +0200981 *
982 * @ep: the list header of the endpoint to free
983 *
984 * This function does not care for the endpoint's use count but will tear
985 * down all the streaming URBs immediately and free all resources.
986 */
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200987void snd_usb_endpoint_free(struct list_head *head)
988{
989 struct snd_usb_endpoint *ep;
990
991 ep = list_entry(head, struct snd_usb_endpoint, list);
992 release_urbs(ep, 1);
993 kfree(ep);
994}
995
Daniel Mack94c27212012-04-12 13:51:15 +0200996/**
997 * snd_usb_handle_sync_urb: parse an USB sync packet
Daniel Mack8fdff6a2012-04-12 13:51:11 +0200998 *
Daniel Mack94c27212012-04-12 13:51:15 +0200999 * @ep: the endpoint to handle the packet
1000 * @sender: the sending endpoint
1001 * @urb: the received packet
1002 *
1003 * This function is called from the context of an endpoint that received
1004 * the packet and is used to let another endpoint object handle the payload.
Daniel Mack8fdff6a2012-04-12 13:51:11 +02001005 */
1006void snd_usb_handle_sync_urb(struct snd_usb_endpoint *ep,
1007 struct snd_usb_endpoint *sender,
1008 const struct urb *urb)
1009{
1010 int shift;
1011 unsigned int f;
1012 unsigned long flags;
1013
1014 snd_BUG_ON(ep == sender);
1015
Daniel Mack94c27212012-04-12 13:51:15 +02001016 /*
1017 * In case the endpoint is operating in implicit feedback mode, prepare
Daniel Mack07a5e9d2012-04-24 19:31:24 +02001018 * a new outbound URB that has the same layout as the received packet
1019 * and add it to the list of pending urbs. queue_pending_output_urbs()
1020 * will take care of them later.
Daniel Mack94c27212012-04-12 13:51:15 +02001021 */
Eldad Zack98ae4722013-04-03 23:18:52 +02001022 if (snd_usb_endpoint_implicit_feedback_sink(ep) &&
Daniel Mack8fdff6a2012-04-12 13:51:11 +02001023 ep->use_count != 0) {
1024
1025 /* implicit feedback case */
1026 int i, bytes = 0;
1027 struct snd_urb_ctx *in_ctx;
1028 struct snd_usb_packet_info *out_packet;
1029
1030 in_ctx = urb->context;
1031
1032 /* Count overall packet size */
1033 for (i = 0; i < in_ctx->packets; i++)
1034 if (urb->iso_frame_desc[i].status == 0)
1035 bytes += urb->iso_frame_desc[i].actual_length;
1036
1037 /*
1038 * skip empty packets. At least M-Audio's Fast Track Ultra stops
1039 * streaming once it received a 0-byte OUT URB
1040 */
1041 if (bytes == 0)
1042 return;
1043
1044 spin_lock_irqsave(&ep->lock, flags);
1045 out_packet = ep->next_packet + ep->next_packet_write_pos;
1046
1047 /*
1048 * Iterate through the inbound packet and prepare the lengths
1049 * for the output packet. The OUT packet we are about to send
Eldad Zack28acb122012-11-28 23:55:34 +01001050 * will have the same amount of payload bytes per stride as the
1051 * IN packet we just received. Since the actual size is scaled
1052 * by the stride, use the sender stride to calculate the length
1053 * in case the number of channels differ between the implicitly
1054 * fed-back endpoint and the synchronizing endpoint.
Daniel Mack8fdff6a2012-04-12 13:51:11 +02001055 */
1056
1057 out_packet->packets = in_ctx->packets;
1058 for (i = 0; i < in_ctx->packets; i++) {
1059 if (urb->iso_frame_desc[i].status == 0)
1060 out_packet->packet_size[i] =
Eldad Zack28acb122012-11-28 23:55:34 +01001061 urb->iso_frame_desc[i].actual_length / sender->stride;
Daniel Mack8fdff6a2012-04-12 13:51:11 +02001062 else
1063 out_packet->packet_size[i] = 0;
1064 }
1065
1066 ep->next_packet_write_pos++;
1067 ep->next_packet_write_pos %= MAX_URBS;
1068 spin_unlock_irqrestore(&ep->lock, flags);
1069 queue_pending_output_urbs(ep);
1070
1071 return;
1072 }
1073
Daniel Mack94c27212012-04-12 13:51:15 +02001074 /*
1075 * process after playback sync complete
1076 *
1077 * Full speed devices report feedback values in 10.14 format as samples
1078 * per frame, high speed devices in 16.16 format as samples per
1079 * microframe.
1080 *
1081 * Because the Audio Class 1 spec was written before USB 2.0, many high
1082 * speed devices use a wrong interpretation, some others use an
1083 * entirely different format.
1084 *
1085 * Therefore, we cannot predict what format any particular device uses
1086 * and must detect it automatically.
1087 */
Daniel Mack8fdff6a2012-04-12 13:51:11 +02001088
1089 if (urb->iso_frame_desc[0].status != 0 ||
1090 urb->iso_frame_desc[0].actual_length < 3)
1091 return;
1092
1093 f = le32_to_cpup(urb->transfer_buffer);
1094 if (urb->iso_frame_desc[0].actual_length == 3)
1095 f &= 0x00ffffff;
1096 else
1097 f &= 0x0fffffff;
1098
1099 if (f == 0)
1100 return;
1101
1102 if (unlikely(ep->freqshift == INT_MIN)) {
1103 /*
1104 * The first time we see a feedback value, determine its format
1105 * by shifting it left or right until it matches the nominal
1106 * frequency value. This assumes that the feedback does not
1107 * differ from the nominal value more than +50% or -25%.
1108 */
1109 shift = 0;
1110 while (f < ep->freqn - ep->freqn / 4) {
1111 f <<= 1;
1112 shift++;
1113 }
1114 while (f > ep->freqn + ep->freqn / 2) {
1115 f >>= 1;
1116 shift--;
1117 }
1118 ep->freqshift = shift;
1119 } else if (ep->freqshift >= 0)
1120 f <<= ep->freqshift;
1121 else
1122 f >>= -ep->freqshift;
1123
1124 if (likely(f >= ep->freqn - ep->freqn / 8 && f <= ep->freqmax)) {
1125 /*
1126 * If the frequency looks valid, set it.
1127 * This value is referred to in prepare_playback_urb().
1128 */
1129 spin_lock_irqsave(&ep->lock, flags);
1130 ep->freqm = f;
1131 spin_unlock_irqrestore(&ep->lock, flags);
1132 } else {
1133 /*
1134 * Out of range; maybe the shift value is wrong.
1135 * Reset it so that we autodetect again the next time.
1136 */
1137 ep->freqshift = INT_MIN;
1138 }
1139}
1140