]> git.proxmox.com Git - mirror_ubuntu-zesty-kernel.git/blob - drivers/staging/ozwpan/ozpd.c
staging: board: Migrate away from __pm_genpd_name_add_device()
[mirror_ubuntu-zesty-kernel.git] / drivers / staging / ozwpan / ozpd.c
1 /* -----------------------------------------------------------------------------
2 * Copyright (c) 2011 Ozmo Inc
3 * Released under the GNU General Public License Version 2 (GPLv2).
4 * -----------------------------------------------------------------------------
5 */
6
7 #include <linux/module.h>
8 #include <linux/timer.h>
9 #include <linux/sched.h>
10 #include <linux/netdevice.h>
11 #include <linux/etherdevice.h>
12 #include <linux/errno.h>
13 #include "ozdbg.h"
14 #include "ozprotocol.h"
15 #include "ozeltbuf.h"
16 #include "ozpd.h"
17 #include "ozproto.h"
18 #include "ozcdev.h"
19 #include "ozusbsvc.h"
20 #include <asm/unaligned.h>
21 #include <linux/uaccess.h>
22 #include <net/psnap.h>
23
24 static struct oz_tx_frame *oz_tx_frame_alloc(struct oz_pd *pd);
25 static void oz_tx_frame_free(struct oz_pd *pd, struct oz_tx_frame *f);
26 static void oz_tx_isoc_free(struct oz_pd *pd, struct oz_tx_frame *f);
27 static struct sk_buff *oz_build_frame(struct oz_pd *pd, struct oz_tx_frame *f);
28 static int oz_send_isoc_frame(struct oz_pd *pd);
29 static void oz_retire_frame(struct oz_pd *pd, struct oz_tx_frame *f);
30 static void oz_isoc_stream_free(struct oz_isoc_stream *st);
31 static int oz_send_next_queued_frame(struct oz_pd *pd, int more_data);
32 static void oz_isoc_destructor(struct sk_buff *skb);
33
34 /*
35 * Counts the uncompleted isoc frames submitted to netcard.
36 */
37 static atomic_t g_submitted_isoc = ATOMIC_INIT(0);
38
39 /* Application handler functions.
40 */
41 static const struct oz_app_if g_app_if[OZ_NB_APPS] = {
42 [OZ_APPID_USB] = {
43 .init = oz_usb_init,
44 .term = oz_usb_term,
45 .start = oz_usb_start,
46 .stop = oz_usb_stop,
47 .rx = oz_usb_rx,
48 .heartbeat = oz_usb_heartbeat,
49 .farewell = oz_usb_farewell,
50 },
51 [OZ_APPID_SERIAL] = {
52 .init = oz_cdev_init,
53 .term = oz_cdev_term,
54 .start = oz_cdev_start,
55 .stop = oz_cdev_stop,
56 .rx = oz_cdev_rx,
57 },
58 };
59
60
61 /*
62 * Context: softirq or process
63 */
64 void oz_pd_set_state(struct oz_pd *pd, unsigned state)
65 {
66 pd->state = state;
67 switch (state) {
68 case OZ_PD_S_IDLE:
69 oz_pd_dbg(pd, ON, "PD State: OZ_PD_S_IDLE\n");
70 break;
71 case OZ_PD_S_CONNECTED:
72 oz_pd_dbg(pd, ON, "PD State: OZ_PD_S_CONNECTED\n");
73 break;
74 case OZ_PD_S_STOPPED:
75 oz_pd_dbg(pd, ON, "PD State: OZ_PD_S_STOPPED\n");
76 break;
77 case OZ_PD_S_SLEEP:
78 oz_pd_dbg(pd, ON, "PD State: OZ_PD_S_SLEEP\n");
79 break;
80 }
81 }
82
83 /*
84 * Context: softirq or process
85 */
86 void oz_pd_get(struct oz_pd *pd)
87 {
88 atomic_inc(&pd->ref_count);
89 }
90
91 /*
92 * Context: softirq or process
93 */
94 void oz_pd_put(struct oz_pd *pd)
95 {
96 if (atomic_dec_and_test(&pd->ref_count))
97 oz_pd_destroy(pd);
98 }
99
100 /*
101 * Context: softirq-serialized
102 */
103 struct oz_pd *oz_pd_alloc(const u8 *mac_addr)
104 {
105 struct oz_pd *pd;
106 int i;
107
108 pd = kzalloc(sizeof(struct oz_pd), GFP_ATOMIC);
109 if (!pd)
110 return NULL;
111
112 atomic_set(&pd->ref_count, 2);
113 for (i = 0; i < OZ_NB_APPS; i++)
114 spin_lock_init(&pd->app_lock[i]);
115 pd->last_rx_pkt_num = 0xffffffff;
116 oz_pd_set_state(pd, OZ_PD_S_IDLE);
117 pd->max_tx_size = OZ_MAX_TX_SIZE;
118 ether_addr_copy(pd->mac_addr, mac_addr);
119 oz_elt_buf_init(&pd->elt_buff);
120 spin_lock_init(&pd->tx_frame_lock);
121 INIT_LIST_HEAD(&pd->tx_queue);
122 INIT_LIST_HEAD(&pd->farewell_list);
123 pd->last_sent_frame = &pd->tx_queue;
124 spin_lock_init(&pd->stream_lock);
125 INIT_LIST_HEAD(&pd->stream_list);
126 tasklet_init(&pd->heartbeat_tasklet, oz_pd_heartbeat_handler,
127 (unsigned long)pd);
128 tasklet_init(&pd->timeout_tasklet, oz_pd_timeout_handler,
129 (unsigned long)pd);
130 hrtimer_init(&pd->heartbeat, CLOCK_MONOTONIC, HRTIMER_MODE_REL);
131 hrtimer_init(&pd->timeout, CLOCK_MONOTONIC, HRTIMER_MODE_REL);
132 pd->heartbeat.function = oz_pd_heartbeat_event;
133 pd->timeout.function = oz_pd_timeout_event;
134
135 return pd;
136 }
137
138 /*
139 * Context: softirq or process
140 */
141 static void oz_pd_free(struct work_struct *work)
142 {
143 struct list_head *e, *n;
144 struct oz_pd *pd;
145
146 oz_pd_dbg(pd, ON, "Destroying PD\n");
147 pd = container_of(work, struct oz_pd, workitem);
148 /*Disable timer tasklets*/
149 tasklet_kill(&pd->heartbeat_tasklet);
150 tasklet_kill(&pd->timeout_tasklet);
151
152 /* Free streams, queued tx frames and farewells. */
153
154 list_for_each_safe(e, n, &pd->stream_list)
155 oz_isoc_stream_free(list_entry(e, struct oz_isoc_stream, link));
156
157 list_for_each_safe(e, n, &pd->tx_queue) {
158 struct oz_tx_frame *f = list_entry(e, struct oz_tx_frame, link);
159
160 if (f->skb != NULL)
161 kfree_skb(f->skb);
162 oz_retire_frame(pd, f);
163 }
164
165 oz_elt_buf_term(&pd->elt_buff);
166
167 list_for_each_safe(e, n, &pd->farewell_list)
168 kfree(list_entry(e, struct oz_farewell, link));
169
170 if (pd->net_dev)
171 dev_put(pd->net_dev);
172 kfree(pd);
173 }
174
175 /*
176 * Context: softirq or Process
177 */
178 void oz_pd_destroy(struct oz_pd *pd)
179 {
180 if (hrtimer_active(&pd->timeout))
181 hrtimer_cancel(&pd->timeout);
182 if (hrtimer_active(&pd->heartbeat))
183 hrtimer_cancel(&pd->heartbeat);
184
185 INIT_WORK(&pd->workitem, oz_pd_free);
186 if (!schedule_work(&pd->workitem))
187 oz_pd_dbg(pd, ON, "failed to schedule workitem\n");
188 }
189
190 /*
191 * Context: softirq-serialized
192 */
193 int oz_services_start(struct oz_pd *pd, u16 apps, int resume)
194 {
195 int i, rc = 0;
196
197 oz_pd_dbg(pd, ON, "%s: (0x%x) resume(%d)\n", __func__, apps, resume);
198 for (i = 0; i < OZ_NB_APPS; i++) {
199 if (g_app_if[i].start && (apps & (1 << i))) {
200 if (g_app_if[i].start(pd, resume)) {
201 rc = -1;
202 oz_pd_dbg(pd, ON,
203 "Unable to start service %d\n", i);
204 break;
205 }
206 spin_lock_bh(&g_polling_lock);
207 pd->total_apps |= (1 << i);
208 if (resume)
209 pd->paused_apps &= ~(1 << i);
210 spin_unlock_bh(&g_polling_lock);
211 }
212 }
213 return rc;
214 }
215
216 /*
217 * Context: softirq or process
218 */
219 void oz_services_stop(struct oz_pd *pd, u16 apps, int pause)
220 {
221 int i;
222
223 oz_pd_dbg(pd, ON, "%s: (0x%x) pause(%d)\n", __func__, apps, pause);
224 for (i = 0; i < OZ_NB_APPS; i++) {
225 if (g_app_if[i].stop && (apps & (1 << i))) {
226 spin_lock_bh(&g_polling_lock);
227 if (pause) {
228 pd->paused_apps |= (1 << i);
229 } else {
230 pd->total_apps &= ~(1 << i);
231 pd->paused_apps &= ~(1 << i);
232 }
233 spin_unlock_bh(&g_polling_lock);
234 g_app_if[i].stop(pd, pause);
235 }
236 }
237 }
238
239 /*
240 * Context: softirq
241 */
242 void oz_pd_heartbeat(struct oz_pd *pd, u16 apps)
243 {
244 int i, more = 0;
245
246 for (i = 0; i < OZ_NB_APPS; i++) {
247 if (g_app_if[i].heartbeat && (apps & (1 << i))) {
248 if (g_app_if[i].heartbeat(pd))
249 more = 1;
250 }
251 }
252 if ((!more) && (hrtimer_active(&pd->heartbeat)))
253 hrtimer_cancel(&pd->heartbeat);
254 if (pd->mode & OZ_F_ISOC_ANYTIME) {
255 int count = 8;
256
257 while (count-- && (oz_send_isoc_frame(pd) >= 0))
258 ;
259 }
260 }
261
262 /*
263 * Context: softirq or process
264 */
265 void oz_pd_stop(struct oz_pd *pd)
266 {
267 u16 stop_apps;
268
269 oz_dbg(ON, "oz_pd_stop() State = 0x%x\n", pd->state);
270 oz_pd_indicate_farewells(pd);
271 spin_lock_bh(&g_polling_lock);
272 stop_apps = pd->total_apps;
273 pd->total_apps = 0;
274 pd->paused_apps = 0;
275 spin_unlock_bh(&g_polling_lock);
276 oz_services_stop(pd, stop_apps, 0);
277 spin_lock_bh(&g_polling_lock);
278 oz_pd_set_state(pd, OZ_PD_S_STOPPED);
279 /* Remove from PD list.*/
280 list_del(&pd->link);
281 spin_unlock_bh(&g_polling_lock);
282 oz_dbg(ON, "pd ref count = %d\n", atomic_read(&pd->ref_count));
283 oz_pd_put(pd);
284 }
285
286 /*
287 * Context: softirq
288 */
289 int oz_pd_sleep(struct oz_pd *pd)
290 {
291 int do_stop = 0;
292 u16 stop_apps;
293
294 spin_lock_bh(&g_polling_lock);
295 if (pd->state & (OZ_PD_S_SLEEP | OZ_PD_S_STOPPED)) {
296 spin_unlock_bh(&g_polling_lock);
297 return 0;
298 }
299 if (pd->keep_alive && pd->session_id)
300 oz_pd_set_state(pd, OZ_PD_S_SLEEP);
301 else
302 do_stop = 1;
303
304 stop_apps = pd->total_apps;
305 spin_unlock_bh(&g_polling_lock);
306 if (do_stop) {
307 oz_pd_stop(pd);
308 } else {
309 oz_services_stop(pd, stop_apps, 1);
310 oz_timer_add(pd, OZ_TIMER_STOP, pd->keep_alive);
311 }
312 return do_stop;
313 }
314
315 /*
316 * Context: softirq
317 */
318 static struct oz_tx_frame *oz_tx_frame_alloc(struct oz_pd *pd)
319 {
320 struct oz_tx_frame *f;
321
322 f = kmem_cache_alloc(oz_tx_frame_cache, GFP_ATOMIC);
323 if (f) {
324 f->total_size = sizeof(struct oz_hdr);
325 INIT_LIST_HEAD(&f->link);
326 INIT_LIST_HEAD(&f->elt_list);
327 }
328 return f;
329 }
330
331 /*
332 * Context: softirq or process
333 */
334 static void oz_tx_isoc_free(struct oz_pd *pd, struct oz_tx_frame *f)
335 {
336 pd->nb_queued_isoc_frames--;
337 list_del_init(&f->link);
338
339 kmem_cache_free(oz_tx_frame_cache, f);
340
341 oz_dbg(TX_FRAMES, "Releasing ISOC Frame isoc_nb= %d\n",
342 pd->nb_queued_isoc_frames);
343 }
344
345 /*
346 * Context: softirq or process
347 */
348 static void oz_tx_frame_free(struct oz_pd *pd, struct oz_tx_frame *f)
349 {
350 kmem_cache_free(oz_tx_frame_cache, f);
351 }
352
353 /*
354 * Context: softirq-serialized
355 */
356 static void oz_set_more_bit(struct sk_buff *skb)
357 {
358 struct oz_hdr *oz_hdr = (struct oz_hdr *)skb_network_header(skb);
359
360 oz_hdr->control |= OZ_F_MORE_DATA;
361 }
362
363 /*
364 * Context: softirq-serialized
365 */
366 static void oz_set_last_pkt_nb(struct oz_pd *pd, struct sk_buff *skb)
367 {
368 struct oz_hdr *oz_hdr = (struct oz_hdr *)skb_network_header(skb);
369
370 oz_hdr->last_pkt_num = pd->trigger_pkt_num & OZ_LAST_PN_MASK;
371 }
372
373 /*
374 * Context: softirq
375 */
376 int oz_prepare_frame(struct oz_pd *pd, int empty)
377 {
378 struct oz_tx_frame *f;
379
380 if ((pd->mode & OZ_MODE_MASK) != OZ_MODE_TRIGGERED)
381 return -1;
382 if (pd->nb_queued_frames >= OZ_MAX_QUEUED_FRAMES)
383 return -1;
384 if (!empty && !oz_are_elts_available(&pd->elt_buff))
385 return -1;
386 f = oz_tx_frame_alloc(pd);
387 if (f == NULL)
388 return -1;
389 f->skb = NULL;
390 f->hdr.control =
391 (OZ_PROTOCOL_VERSION<<OZ_VERSION_SHIFT) | OZ_F_ACK_REQUESTED;
392 ++pd->last_tx_pkt_num;
393 put_unaligned(cpu_to_le32(pd->last_tx_pkt_num), &f->hdr.pkt_num);
394 if (empty == 0) {
395 oz_select_elts_for_tx(&pd->elt_buff, 0, &f->total_size,
396 pd->max_tx_size, &f->elt_list);
397 }
398 spin_lock(&pd->tx_frame_lock);
399 list_add_tail(&f->link, &pd->tx_queue);
400 pd->nb_queued_frames++;
401 spin_unlock(&pd->tx_frame_lock);
402 return 0;
403 }
404
405 /*
406 * Context: softirq-serialized
407 */
408 static struct sk_buff *oz_build_frame(struct oz_pd *pd, struct oz_tx_frame *f)
409 {
410 struct sk_buff *skb;
411 struct net_device *dev = pd->net_dev;
412 struct oz_hdr *oz_hdr;
413 struct oz_elt *elt;
414 struct oz_elt_info *ei;
415
416 /* Allocate skb with enough space for the lower layers as well
417 * as the space we need.
418 */
419 skb = alloc_skb(f->total_size + OZ_ALLOCATED_SPACE(dev), GFP_ATOMIC);
420 if (skb == NULL)
421 return NULL;
422 /* Reserve the head room for lower layers.
423 */
424 skb_reserve(skb, LL_RESERVED_SPACE(dev));
425 skb_reset_network_header(skb);
426 skb->dev = dev;
427 skb->protocol = htons(OZ_ETHERTYPE);
428 if (dev_hard_header(skb, dev, OZ_ETHERTYPE, pd->mac_addr,
429 dev->dev_addr, skb->len) < 0)
430 goto fail;
431 /* Push the tail to the end of the area we are going to copy to.
432 */
433 oz_hdr = (struct oz_hdr *)skb_put(skb, f->total_size);
434 f->hdr.last_pkt_num = pd->trigger_pkt_num & OZ_LAST_PN_MASK;
435 memcpy(oz_hdr, &f->hdr, sizeof(struct oz_hdr));
436 /* Copy the elements into the frame body.
437 */
438 elt = (struct oz_elt *)(oz_hdr+1);
439 list_for_each_entry(ei, &f->elt_list, link) {
440 memcpy(elt, ei->data, ei->length);
441 elt = oz_next_elt(elt);
442 }
443 return skb;
444 fail:
445 kfree_skb(skb);
446 return NULL;
447 }
448
449 /*
450 * Context: softirq or process
451 */
452 static void oz_retire_frame(struct oz_pd *pd, struct oz_tx_frame *f)
453 {
454 struct oz_elt_info *ei, *n;
455
456 list_for_each_entry_safe(ei, n, &f->elt_list, link) {
457 list_del_init(&ei->link);
458 if (ei->callback)
459 ei->callback(pd, ei->context);
460 spin_lock_bh(&pd->elt_buff.lock);
461 oz_elt_info_free(&pd->elt_buff, ei);
462 spin_unlock_bh(&pd->elt_buff.lock);
463 }
464 oz_tx_frame_free(pd, f);
465 }
466
467 /*
468 * Context: softirq-serialized
469 */
470 static int oz_send_next_queued_frame(struct oz_pd *pd, int more_data)
471 {
472 struct sk_buff *skb;
473 struct oz_tx_frame *f;
474 struct list_head *e;
475
476 spin_lock(&pd->tx_frame_lock);
477 e = pd->last_sent_frame->next;
478 if (e == &pd->tx_queue) {
479 spin_unlock(&pd->tx_frame_lock);
480 return -1;
481 }
482 f = list_entry(e, struct oz_tx_frame, link);
483
484 if (f->skb != NULL) {
485 skb = f->skb;
486 oz_tx_isoc_free(pd, f);
487 spin_unlock(&pd->tx_frame_lock);
488 if (more_data)
489 oz_set_more_bit(skb);
490 oz_set_last_pkt_nb(pd, skb);
491 if ((int)atomic_read(&g_submitted_isoc) <
492 OZ_MAX_SUBMITTED_ISOC) {
493 if (dev_queue_xmit(skb) < 0) {
494 oz_dbg(TX_FRAMES, "Dropping ISOC Frame\n");
495 return -1;
496 }
497 atomic_inc(&g_submitted_isoc);
498 oz_dbg(TX_FRAMES, "Sending ISOC Frame, nb_isoc= %d\n",
499 pd->nb_queued_isoc_frames);
500 return 0;
501 }
502 kfree_skb(skb);
503 oz_dbg(TX_FRAMES, "Dropping ISOC Frame>\n");
504 return -1;
505 }
506
507 pd->last_sent_frame = e;
508 skb = oz_build_frame(pd, f);
509 spin_unlock(&pd->tx_frame_lock);
510 if (!skb)
511 return -1;
512 if (more_data)
513 oz_set_more_bit(skb);
514 oz_dbg(TX_FRAMES, "TX frame PN=0x%x\n", f->hdr.pkt_num);
515 if (dev_queue_xmit(skb) < 0)
516 return -1;
517
518 return 0;
519 }
520
521 /*
522 * Context: softirq-serialized
523 */
524 void oz_send_queued_frames(struct oz_pd *pd, int backlog)
525 {
526 while (oz_prepare_frame(pd, 0) >= 0)
527 backlog++;
528
529 switch (pd->mode & (OZ_F_ISOC_NO_ELTS | OZ_F_ISOC_ANYTIME)) {
530
531 case OZ_F_ISOC_NO_ELTS: {
532 backlog += pd->nb_queued_isoc_frames;
533 if (backlog <= 0)
534 goto out;
535 if (backlog > OZ_MAX_SUBMITTED_ISOC)
536 backlog = OZ_MAX_SUBMITTED_ISOC;
537 break;
538 }
539 case OZ_NO_ELTS_ANYTIME: {
540 if ((backlog <= 0) && (pd->isoc_sent == 0))
541 goto out;
542 break;
543 }
544 default: {
545 if (backlog <= 0)
546 goto out;
547 break;
548 }
549 }
550 while (backlog--) {
551 if (oz_send_next_queued_frame(pd, backlog) < 0)
552 break;
553 }
554 return;
555
556 out: oz_prepare_frame(pd, 1);
557 oz_send_next_queued_frame(pd, 0);
558 }
559
560 /*
561 * Context: softirq
562 */
563 static int oz_send_isoc_frame(struct oz_pd *pd)
564 {
565 struct sk_buff *skb;
566 struct net_device *dev = pd->net_dev;
567 struct oz_hdr *oz_hdr;
568 struct oz_elt *elt;
569 struct oz_elt_info *ei;
570 LIST_HEAD(list);
571 int total_size = sizeof(struct oz_hdr);
572
573 oz_select_elts_for_tx(&pd->elt_buff, 1, &total_size,
574 pd->max_tx_size, &list);
575 if (list_empty(&list))
576 return 0;
577 skb = alloc_skb(total_size + OZ_ALLOCATED_SPACE(dev), GFP_ATOMIC);
578 if (skb == NULL) {
579 oz_dbg(ON, "Cannot alloc skb\n");
580 oz_elt_info_free_chain(&pd->elt_buff, &list);
581 return -1;
582 }
583 skb_reserve(skb, LL_RESERVED_SPACE(dev));
584 skb_reset_network_header(skb);
585 skb->dev = dev;
586 skb->protocol = htons(OZ_ETHERTYPE);
587 if (dev_hard_header(skb, dev, OZ_ETHERTYPE, pd->mac_addr,
588 dev->dev_addr, skb->len) < 0) {
589 kfree_skb(skb);
590 return -1;
591 }
592 oz_hdr = (struct oz_hdr *)skb_put(skb, total_size);
593 oz_hdr->control = (OZ_PROTOCOL_VERSION<<OZ_VERSION_SHIFT) | OZ_F_ISOC;
594 oz_hdr->last_pkt_num = pd->trigger_pkt_num & OZ_LAST_PN_MASK;
595 elt = (struct oz_elt *)(oz_hdr+1);
596
597 list_for_each_entry(ei, &list, link) {
598 memcpy(elt, ei->data, ei->length);
599 elt = oz_next_elt(elt);
600 }
601 dev_queue_xmit(skb);
602 oz_elt_info_free_chain(&pd->elt_buff, &list);
603 return 0;
604 }
605
606 /*
607 * Context: softirq-serialized
608 */
609 void oz_retire_tx_frames(struct oz_pd *pd, u8 lpn)
610 {
611 struct oz_tx_frame *f, *tmp = NULL;
612 u8 diff;
613 u32 pkt_num;
614
615 LIST_HEAD(list);
616
617 spin_lock(&pd->tx_frame_lock);
618 list_for_each_entry(f, &pd->tx_queue, link) {
619 pkt_num = le32_to_cpu(get_unaligned(&f->hdr.pkt_num));
620 diff = (lpn - (pkt_num & OZ_LAST_PN_MASK)) & OZ_LAST_PN_MASK;
621 if ((diff > OZ_LAST_PN_HALF_CYCLE) || (pkt_num == 0))
622 break;
623 oz_dbg(TX_FRAMES, "Releasing pkt_num= %u, nb= %d\n",
624 pkt_num, pd->nb_queued_frames);
625 tmp = f;
626 pd->nb_queued_frames--;
627 }
628 if (tmp)
629 list_cut_position(&list, &pd->tx_queue, &tmp->link);
630 pd->last_sent_frame = &pd->tx_queue;
631 spin_unlock(&pd->tx_frame_lock);
632
633 list_for_each_entry_safe(f, tmp, &list, link)
634 oz_retire_frame(pd, f);
635 }
636
637 /*
638 * Precondition: stream_lock must be held.
639 * Context: softirq
640 */
641 static struct oz_isoc_stream *pd_stream_find(struct oz_pd *pd, u8 ep_num)
642 {
643 struct oz_isoc_stream *st;
644
645 list_for_each_entry(st, &pd->stream_list, link) {
646 if (st->ep_num == ep_num)
647 return st;
648 }
649 return NULL;
650 }
651
652 /*
653 * Context: softirq
654 */
655 int oz_isoc_stream_create(struct oz_pd *pd, u8 ep_num)
656 {
657 struct oz_isoc_stream *st;
658
659 st = kzalloc(sizeof(struct oz_isoc_stream), GFP_ATOMIC);
660 if (!st)
661 return -ENOMEM;
662 st->ep_num = ep_num;
663 spin_lock_bh(&pd->stream_lock);
664 if (!pd_stream_find(pd, ep_num)) {
665 list_add(&st->link, &pd->stream_list);
666 st = NULL;
667 }
668 spin_unlock_bh(&pd->stream_lock);
669 kfree(st);
670 return 0;
671 }
672
673 /*
674 * Context: softirq or process
675 */
676 static void oz_isoc_stream_free(struct oz_isoc_stream *st)
677 {
678 kfree_skb(st->skb);
679 kfree(st);
680 }
681
682 /*
683 * Context: softirq
684 */
685 int oz_isoc_stream_delete(struct oz_pd *pd, u8 ep_num)
686 {
687 struct oz_isoc_stream *st;
688
689 spin_lock_bh(&pd->stream_lock);
690 st = pd_stream_find(pd, ep_num);
691 if (st)
692 list_del(&st->link);
693 spin_unlock_bh(&pd->stream_lock);
694 if (st)
695 oz_isoc_stream_free(st);
696 return 0;
697 }
698
699 /*
700 * Context: any
701 */
702 static void oz_isoc_destructor(struct sk_buff *skb)
703 {
704 atomic_dec(&g_submitted_isoc);
705 }
706
707 /*
708 * Context: softirq
709 */
710 int oz_send_isoc_unit(struct oz_pd *pd, u8 ep_num, const u8 *data, int len)
711 {
712 struct net_device *dev = pd->net_dev;
713 struct oz_isoc_stream *st;
714 u8 nb_units = 0;
715 struct sk_buff *skb = NULL;
716 struct oz_hdr *oz_hdr = NULL;
717 int size = 0;
718
719 spin_lock_bh(&pd->stream_lock);
720 st = pd_stream_find(pd, ep_num);
721 if (st) {
722 skb = st->skb;
723 st->skb = NULL;
724 nb_units = st->nb_units;
725 st->nb_units = 0;
726 oz_hdr = st->oz_hdr;
727 size = st->size;
728 }
729 spin_unlock_bh(&pd->stream_lock);
730 if (!st)
731 return 0;
732 if (!skb) {
733 /* Allocate enough space for max size frame. */
734 skb = alloc_skb(pd->max_tx_size + OZ_ALLOCATED_SPACE(dev),
735 GFP_ATOMIC);
736 if (skb == NULL)
737 return 0;
738 /* Reserve the head room for lower layers. */
739 skb_reserve(skb, LL_RESERVED_SPACE(dev));
740 skb_reset_network_header(skb);
741 skb->dev = dev;
742 skb->protocol = htons(OZ_ETHERTYPE);
743 /* For audio packet set priority to AC_VO */
744 skb->priority = 0x7;
745 size = sizeof(struct oz_hdr) + sizeof(struct oz_isoc_large);
746 oz_hdr = (struct oz_hdr *)skb_put(skb, size);
747 }
748 memcpy(skb_put(skb, len), data, len);
749 size += len;
750 if (++nb_units < pd->ms_per_isoc) {
751 spin_lock_bh(&pd->stream_lock);
752 st->skb = skb;
753 st->nb_units = nb_units;
754 st->oz_hdr = oz_hdr;
755 st->size = size;
756 spin_unlock_bh(&pd->stream_lock);
757 } else {
758 struct oz_hdr oz;
759 struct oz_isoc_large iso;
760
761 spin_lock_bh(&pd->stream_lock);
762 iso.frame_number = st->frame_num;
763 st->frame_num += nb_units;
764 spin_unlock_bh(&pd->stream_lock);
765 oz.control =
766 (OZ_PROTOCOL_VERSION<<OZ_VERSION_SHIFT) | OZ_F_ISOC;
767 oz.last_pkt_num = pd->trigger_pkt_num & OZ_LAST_PN_MASK;
768 oz.pkt_num = 0;
769 iso.endpoint = ep_num;
770 iso.format = OZ_DATA_F_ISOC_LARGE;
771 iso.ms_data = nb_units;
772 memcpy(oz_hdr, &oz, sizeof(oz));
773 memcpy(oz_hdr+1, &iso, sizeof(iso));
774 if (dev_hard_header(skb, dev, OZ_ETHERTYPE, pd->mac_addr,
775 dev->dev_addr, skb->len) < 0)
776 goto out;
777
778 skb->destructor = oz_isoc_destructor;
779 /*Queue for Xmit if mode is not ANYTIME*/
780 if (!(pd->mode & OZ_F_ISOC_ANYTIME)) {
781 struct oz_tx_frame *isoc_unit = NULL;
782 int nb = pd->nb_queued_isoc_frames;
783
784 if (nb >= pd->isoc_latency) {
785 struct oz_tx_frame *f;
786
787 oz_dbg(TX_FRAMES, "Dropping ISOC Unit nb= %d\n",
788 nb);
789 spin_lock(&pd->tx_frame_lock);
790 list_for_each_entry(f, &pd->tx_queue, link) {
791 if (f->skb != NULL) {
792 oz_tx_isoc_free(pd, f);
793 break;
794 }
795 }
796 spin_unlock(&pd->tx_frame_lock);
797 }
798 isoc_unit = oz_tx_frame_alloc(pd);
799 if (isoc_unit == NULL)
800 goto out;
801 isoc_unit->hdr = oz;
802 isoc_unit->skb = skb;
803 spin_lock_bh(&pd->tx_frame_lock);
804 list_add_tail(&isoc_unit->link, &pd->tx_queue);
805 pd->nb_queued_isoc_frames++;
806 spin_unlock_bh(&pd->tx_frame_lock);
807 oz_dbg(TX_FRAMES,
808 "Added ISOC Frame to Tx Queue isoc_nb= %d, nb= %d\n",
809 pd->nb_queued_isoc_frames, pd->nb_queued_frames);
810 return 0;
811 }
812
813 /*In ANYTIME mode Xmit unit immediately*/
814 if (atomic_read(&g_submitted_isoc) < OZ_MAX_SUBMITTED_ISOC) {
815 atomic_inc(&g_submitted_isoc);
816 if (dev_queue_xmit(skb) < 0)
817 return -1;
818 return 0;
819 }
820
821 out: kfree_skb(skb);
822 return -1;
823
824 }
825 return 0;
826 }
827
828 /*
829 * Context: process
830 */
831 void oz_apps_init(void)
832 {
833 int i;
834
835 for (i = 0; i < OZ_NB_APPS; i++) {
836 if (g_app_if[i].init)
837 g_app_if[i].init();
838 }
839 }
840
841 /*
842 * Context: process
843 */
844 void oz_apps_term(void)
845 {
846 int i;
847
848 /* Terminate all the apps. */
849 for (i = 0; i < OZ_NB_APPS; i++) {
850 if (g_app_if[i].term)
851 g_app_if[i].term();
852 }
853 }
854
855 /*
856 * Context: softirq-serialized
857 */
858 void oz_handle_app_elt(struct oz_pd *pd, u8 app_id, struct oz_elt *elt)
859 {
860 if (app_id < OZ_NB_APPS && g_app_if[app_id].rx)
861 g_app_if[app_id].rx(pd, elt);
862 }
863
864 /*
865 * Context: softirq or process
866 */
867 void oz_pd_indicate_farewells(struct oz_pd *pd)
868 {
869 struct oz_farewell *f;
870 const struct oz_app_if *ai = &g_app_if[OZ_APPID_USB];
871
872 while (1) {
873 spin_lock_bh(&g_polling_lock);
874 if (list_empty(&pd->farewell_list)) {
875 spin_unlock_bh(&g_polling_lock);
876 break;
877 }
878 f = list_first_entry(&pd->farewell_list,
879 struct oz_farewell, link);
880 list_del(&f->link);
881 spin_unlock_bh(&g_polling_lock);
882 if (ai->farewell)
883 ai->farewell(pd, f->ep_num, f->report, f->len);
884 kfree(f);
885 }
886 }