]> git.proxmox.com Git - mirror_ubuntu-zesty-kernel.git/blob - drivers/staging/ozwpan/ozproto.c
pinctrl: sirf: move sgpio lock into state container
[mirror_ubuntu-zesty-kernel.git] / drivers / staging / ozwpan / ozproto.c
1 /* -----------------------------------------------------------------------------
2 * Copyright (c) 2011 Ozmo Inc
3 * Released under the GNU General Public License Version 2 (GPLv2).
4 * -----------------------------------------------------------------------------
5 */
6
7 #include <linux/module.h>
8 #include <linux/timer.h>
9 #include <linux/sched.h>
10 #include <linux/netdevice.h>
11 #include <linux/etherdevice.h>
12 #include <linux/errno.h>
13 #include <linux/ieee80211.h>
14 #include <linux/slab.h>
15 #include "ozdbg.h"
16 #include "ozprotocol.h"
17 #include "ozeltbuf.h"
18 #include "ozpd.h"
19 #include "ozproto.h"
20 #include "ozusbsvc.h"
21
22 #include "ozappif.h"
23 #include <asm/unaligned.h>
24 #include <linux/uaccess.h>
25 #include <net/psnap.h>
26
27 #define OZ_CF_CONN_SUCCESS 1
28 #define OZ_CF_CONN_FAILURE 2
29
30 #define OZ_DO_STOP 1
31 #define OZ_DO_SLEEP 2
32
33 struct oz_binding {
34 struct packet_type ptype;
35 char name[OZ_MAX_BINDING_LEN];
36 struct list_head link;
37 };
38
39 /*
40 * External variable
41 */
42
43 DEFINE_SPINLOCK(g_polling_lock);
44 /*
45 * Static external variables.
46 */
47 static LIST_HEAD(g_pd_list);
48 static LIST_HEAD(g_binding);
49 static DEFINE_SPINLOCK(g_binding_lock);
50 static struct sk_buff_head g_rx_queue;
51 static u8 g_session_id;
52 static u16 g_apps = 0x1;
53 static int g_processing_rx;
54
55 struct kmem_cache *oz_elt_info_cache;
56 struct kmem_cache *oz_tx_frame_cache;
57
58 /*
59 * Context: softirq-serialized
60 */
61 static u8 oz_get_new_session_id(u8 exclude)
62 {
63 if (++g_session_id == 0)
64 g_session_id = 1;
65 if (g_session_id == exclude) {
66 if (++g_session_id == 0)
67 g_session_id = 1;
68 }
69 return g_session_id;
70 }
71
72 /*
73 * Context: softirq-serialized
74 */
75 static void oz_send_conn_rsp(struct oz_pd *pd, u8 status)
76 {
77 struct sk_buff *skb;
78 struct net_device *dev = pd->net_dev;
79 struct oz_hdr *oz_hdr;
80 struct oz_elt *elt;
81 struct oz_elt_connect_rsp *body;
82
83 int sz = sizeof(struct oz_hdr) + sizeof(struct oz_elt) +
84 sizeof(struct oz_elt_connect_rsp);
85 skb = alloc_skb(sz + OZ_ALLOCATED_SPACE(dev), GFP_ATOMIC);
86 if (skb == NULL)
87 return;
88 skb_reserve(skb, LL_RESERVED_SPACE(dev));
89 skb_reset_network_header(skb);
90 oz_hdr = (struct oz_hdr *)skb_put(skb, sz);
91 elt = (struct oz_elt *)(oz_hdr+1);
92 body = (struct oz_elt_connect_rsp *)(elt+1);
93 skb->dev = dev;
94 skb->protocol = htons(OZ_ETHERTYPE);
95 /* Fill in device header */
96 if (dev_hard_header(skb, dev, OZ_ETHERTYPE, pd->mac_addr,
97 dev->dev_addr, skb->len) < 0) {
98 kfree_skb(skb);
99 return;
100 }
101 oz_hdr->control = (OZ_PROTOCOL_VERSION<<OZ_VERSION_SHIFT);
102 oz_hdr->last_pkt_num = 0;
103 put_unaligned(0, &oz_hdr->pkt_num);
104 elt->type = OZ_ELT_CONNECT_RSP;
105 elt->length = sizeof(struct oz_elt_connect_rsp);
106 memset(body, 0, sizeof(struct oz_elt_connect_rsp));
107 body->status = status;
108 if (status == 0) {
109 body->mode = pd->mode;
110 body->session_id = pd->session_id;
111 put_unaligned(cpu_to_le16(pd->total_apps), &body->apps);
112 }
113 oz_dbg(ON, "TX: OZ_ELT_CONNECT_RSP %d", status);
114 dev_queue_xmit(skb);
115 }
116
117 /*
118 * Context: softirq-serialized
119 */
120 static void pd_set_keepalive(struct oz_pd *pd, u8 kalive)
121 {
122 unsigned long keep_alive = kalive & OZ_KALIVE_VALUE_MASK;
123
124 switch (kalive & OZ_KALIVE_TYPE_MASK) {
125 case OZ_KALIVE_SPECIAL:
126 pd->keep_alive = keep_alive * 1000*60*60*24*20;
127 break;
128 case OZ_KALIVE_SECS:
129 pd->keep_alive = keep_alive*1000;
130 break;
131 case OZ_KALIVE_MINS:
132 pd->keep_alive = keep_alive*1000*60;
133 break;
134 case OZ_KALIVE_HOURS:
135 pd->keep_alive = keep_alive*1000*60*60;
136 break;
137 default:
138 pd->keep_alive = 0;
139 }
140 oz_dbg(ON, "Keepalive = %lu mSec\n", pd->keep_alive);
141 }
142
143 /*
144 * Context: softirq-serialized
145 */
146 static void pd_set_presleep(struct oz_pd *pd, u8 presleep, u8 start_timer)
147 {
148 if (presleep)
149 pd->presleep = presleep*100;
150 else
151 pd->presleep = OZ_PRESLEEP_TOUT;
152 if (start_timer) {
153 spin_unlock(&g_polling_lock);
154 oz_timer_add(pd, OZ_TIMER_TOUT, pd->presleep);
155 spin_lock(&g_polling_lock);
156 }
157 oz_dbg(ON, "Presleep time = %lu mSec\n", pd->presleep);
158 }
159
160 /*
161 * Context: softirq-serialized
162 */
163 static struct oz_pd *oz_connect_req(struct oz_pd *cur_pd, struct oz_elt *elt,
164 const u8 *pd_addr, struct net_device *net_dev)
165 {
166 struct oz_pd *pd;
167 struct oz_elt_connect_req *body =
168 (struct oz_elt_connect_req *)(elt+1);
169 u8 rsp_status = OZ_STATUS_SUCCESS;
170 u8 stop_needed = 0;
171 u16 new_apps = g_apps;
172 struct net_device *old_net_dev = NULL;
173 struct oz_pd *free_pd = NULL;
174
175 if (cur_pd) {
176 pd = cur_pd;
177 spin_lock_bh(&g_polling_lock);
178 } else {
179 struct oz_pd *pd2 = NULL;
180 struct list_head *e;
181
182 pd = oz_pd_alloc(pd_addr);
183 if (pd == NULL)
184 return NULL;
185 getnstimeofday(&pd->last_rx_timestamp);
186 spin_lock_bh(&g_polling_lock);
187 list_for_each(e, &g_pd_list) {
188 pd2 = list_entry(e, struct oz_pd, link);
189 if (ether_addr_equal(pd2->mac_addr, pd_addr)) {
190 free_pd = pd;
191 pd = pd2;
192 break;
193 }
194 }
195 if (pd != pd2)
196 list_add_tail(&pd->link, &g_pd_list);
197 }
198 if (pd == NULL) {
199 spin_unlock_bh(&g_polling_lock);
200 return NULL;
201 }
202 if (pd->net_dev != net_dev) {
203 old_net_dev = pd->net_dev;
204 dev_hold(net_dev);
205 pd->net_dev = net_dev;
206 }
207 oz_dbg(ON, "Host vendor: %d\n", body->host_vendor);
208 pd->max_tx_size = OZ_MAX_TX_SIZE;
209 pd->mode = body->mode;
210 pd->pd_info = body->pd_info;
211 if (pd->mode & OZ_F_ISOC_NO_ELTS) {
212 pd->ms_per_isoc = body->ms_per_isoc;
213 if (!pd->ms_per_isoc)
214 pd->ms_per_isoc = 4;
215
216 switch (body->ms_isoc_latency & OZ_LATENCY_MASK) {
217 case OZ_ONE_MS_LATENCY:
218 pd->isoc_latency = (body->ms_isoc_latency &
219 ~OZ_LATENCY_MASK) / pd->ms_per_isoc;
220 break;
221 case OZ_TEN_MS_LATENCY:
222 pd->isoc_latency = ((body->ms_isoc_latency &
223 ~OZ_LATENCY_MASK) * 10) / pd->ms_per_isoc;
224 break;
225 default:
226 pd->isoc_latency = OZ_MAX_TX_QUEUE_ISOC;
227 }
228 }
229 if (body->max_len_div16)
230 pd->max_tx_size = ((u16)body->max_len_div16)<<4;
231 oz_dbg(ON, "Max frame:%u Ms per isoc:%u\n",
232 pd->max_tx_size, pd->ms_per_isoc);
233 pd->max_stream_buffering = 3*1024;
234 pd->pulse_period = OZ_QUANTUM;
235 pd_set_presleep(pd, body->presleep, 0);
236 pd_set_keepalive(pd, body->keep_alive);
237
238 new_apps &= le16_to_cpu(get_unaligned(&body->apps));
239 if ((new_apps & 0x1) && (body->session_id)) {
240 if (pd->session_id) {
241 if (pd->session_id != body->session_id) {
242 rsp_status = OZ_STATUS_SESSION_MISMATCH;
243 goto done;
244 }
245 } else {
246 new_apps &= ~0x1; /* Resume not permitted */
247 pd->session_id =
248 oz_get_new_session_id(body->session_id);
249 }
250 } else {
251 if (pd->session_id && !body->session_id) {
252 rsp_status = OZ_STATUS_SESSION_TEARDOWN;
253 stop_needed = 1;
254 } else {
255 new_apps &= ~0x1; /* Resume not permitted */
256 pd->session_id =
257 oz_get_new_session_id(body->session_id);
258 }
259 }
260 done:
261 if (rsp_status == OZ_STATUS_SUCCESS) {
262 u16 start_apps = new_apps & ~pd->total_apps & ~0x1;
263 u16 stop_apps = pd->total_apps & ~new_apps & ~0x1;
264 u16 resume_apps = new_apps & pd->paused_apps & ~0x1;
265
266 spin_unlock_bh(&g_polling_lock);
267 oz_pd_set_state(pd, OZ_PD_S_CONNECTED);
268 oz_dbg(ON, "new_apps=0x%x total_apps=0x%x paused_apps=0x%x\n",
269 new_apps, pd->total_apps, pd->paused_apps);
270 if (start_apps) {
271 if (oz_services_start(pd, start_apps, 0))
272 rsp_status = OZ_STATUS_TOO_MANY_PDS;
273 }
274 if (resume_apps)
275 if (oz_services_start(pd, resume_apps, 1))
276 rsp_status = OZ_STATUS_TOO_MANY_PDS;
277 if (stop_apps)
278 oz_services_stop(pd, stop_apps, 0);
279 oz_pd_request_heartbeat(pd);
280 } else {
281 spin_unlock_bh(&g_polling_lock);
282 }
283 oz_send_conn_rsp(pd, rsp_status);
284 if (rsp_status != OZ_STATUS_SUCCESS) {
285 if (stop_needed)
286 oz_pd_stop(pd);
287 oz_pd_put(pd);
288 pd = NULL;
289 }
290 if (old_net_dev)
291 dev_put(old_net_dev);
292 if (free_pd)
293 oz_pd_destroy(free_pd);
294 return pd;
295 }
296
297 /*
298 * Context: softirq-serialized
299 */
300 static void oz_add_farewell(struct oz_pd *pd, u8 ep_num, u8 index,
301 const u8 *report, u8 len)
302 {
303 struct oz_farewell *f;
304 struct oz_farewell *f2;
305 int found = 0;
306
307 f = kmalloc(sizeof(struct oz_farewell) + len, GFP_ATOMIC);
308 if (!f)
309 return;
310 f->ep_num = ep_num;
311 f->index = index;
312 f->len = len;
313 memcpy(f->report, report, len);
314 oz_dbg(ON, "RX: Adding farewell report\n");
315 spin_lock(&g_polling_lock);
316 list_for_each_entry(f2, &pd->farewell_list, link) {
317 if ((f2->ep_num == ep_num) && (f2->index == index)) {
318 found = 1;
319 list_del(&f2->link);
320 break;
321 }
322 }
323 list_add_tail(&f->link, &pd->farewell_list);
324 spin_unlock(&g_polling_lock);
325 if (found)
326 kfree(f2);
327 }
328
329 /*
330 * Context: softirq-serialized
331 */
332 static void oz_rx_frame(struct sk_buff *skb)
333 {
334 u8 *mac_hdr;
335 u8 *src_addr;
336 struct oz_elt *elt;
337 int length;
338 struct oz_pd *pd = NULL;
339 struct oz_hdr *oz_hdr = (struct oz_hdr *)skb_network_header(skb);
340 struct timespec current_time;
341 int dup = 0;
342 u32 pkt_num;
343
344 oz_dbg(RX_FRAMES, "RX frame PN=0x%x LPN=0x%x control=0x%x\n",
345 oz_hdr->pkt_num, oz_hdr->last_pkt_num, oz_hdr->control);
346 mac_hdr = skb_mac_header(skb);
347 src_addr = &mac_hdr[ETH_ALEN];
348 length = skb->len;
349
350 /* Check the version field */
351 if (oz_get_prot_ver(oz_hdr->control) != OZ_PROTOCOL_VERSION) {
352 oz_dbg(ON, "Incorrect protocol version: %d\n",
353 oz_get_prot_ver(oz_hdr->control));
354 goto done;
355 }
356
357 pkt_num = le32_to_cpu(get_unaligned(&oz_hdr->pkt_num));
358
359 pd = oz_pd_find(src_addr);
360 if (pd) {
361 if (!(pd->state & OZ_PD_S_CONNECTED))
362 oz_pd_set_state(pd, OZ_PD_S_CONNECTED);
363 getnstimeofday(&current_time);
364 if ((current_time.tv_sec != pd->last_rx_timestamp.tv_sec) ||
365 (pd->presleep < MSEC_PER_SEC)) {
366 oz_timer_add(pd, OZ_TIMER_TOUT, pd->presleep);
367 pd->last_rx_timestamp = current_time;
368 }
369 if (pkt_num != pd->last_rx_pkt_num) {
370 pd->last_rx_pkt_num = pkt_num;
371 } else {
372 dup = 1;
373 oz_dbg(ON, "Duplicate frame\n");
374 }
375 }
376
377 if (pd && !dup && ((pd->mode & OZ_MODE_MASK) == OZ_MODE_TRIGGERED)) {
378 oz_dbg(RX_FRAMES, "Received TRIGGER Frame\n");
379 pd->last_sent_frame = &pd->tx_queue;
380 if (oz_hdr->control & OZ_F_ACK) {
381 /* Retire completed frames */
382 oz_retire_tx_frames(pd, oz_hdr->last_pkt_num);
383 }
384 if ((oz_hdr->control & OZ_F_ACK_REQUESTED) &&
385 (pd->state == OZ_PD_S_CONNECTED)) {
386 int backlog = pd->nb_queued_frames;
387
388 pd->trigger_pkt_num = pkt_num;
389 /* Send queued frames */
390 oz_send_queued_frames(pd, backlog);
391 }
392 }
393
394 length -= sizeof(struct oz_hdr);
395 elt = (struct oz_elt *)((u8 *)oz_hdr + sizeof(struct oz_hdr));
396
397 while (length >= sizeof(struct oz_elt)) {
398 length -= sizeof(struct oz_elt) + elt->length;
399 if (length < 0)
400 break;
401 switch (elt->type) {
402 case OZ_ELT_CONNECT_REQ:
403 oz_dbg(ON, "RX: OZ_ELT_CONNECT_REQ\n");
404 pd = oz_connect_req(pd, elt, src_addr, skb->dev);
405 break;
406 case OZ_ELT_DISCONNECT:
407 oz_dbg(ON, "RX: OZ_ELT_DISCONNECT\n");
408 if (pd)
409 oz_pd_sleep(pd);
410 break;
411 case OZ_ELT_UPDATE_PARAM_REQ: {
412 struct oz_elt_update_param *body =
413 (struct oz_elt_update_param *)(elt + 1);
414 oz_dbg(ON, "RX: OZ_ELT_UPDATE_PARAM_REQ\n");
415 if (pd && (pd->state & OZ_PD_S_CONNECTED)) {
416 spin_lock(&g_polling_lock);
417 pd_set_keepalive(pd, body->keepalive);
418 pd_set_presleep(pd, body->presleep, 1);
419 spin_unlock(&g_polling_lock);
420 }
421 }
422 break;
423 case OZ_ELT_FAREWELL_REQ: {
424 struct oz_elt_farewell *body =
425 (struct oz_elt_farewell *)(elt + 1);
426 oz_dbg(ON, "RX: OZ_ELT_FAREWELL_REQ\n");
427 oz_add_farewell(pd, body->ep_num,
428 body->index, body->report,
429 elt->length + 1 - sizeof(*body));
430 }
431 break;
432 case OZ_ELT_APP_DATA:
433 if (pd && (pd->state & OZ_PD_S_CONNECTED)) {
434 struct oz_app_hdr *app_hdr =
435 (struct oz_app_hdr *)(elt+1);
436 if (dup)
437 break;
438 oz_handle_app_elt(pd, app_hdr->app_id, elt);
439 }
440 break;
441 default:
442 oz_dbg(ON, "RX: Unknown elt %02x\n", elt->type);
443 }
444 elt = oz_next_elt(elt);
445 }
446 done:
447 if (pd)
448 oz_pd_put(pd);
449 consume_skb(skb);
450 }
451
452 /*
453 * Context: process
454 */
455 void oz_protocol_term(void)
456 {
457 struct oz_binding *b, *t;
458
459 /* Walk the list of bindings and remove each one.
460 */
461 spin_lock_bh(&g_binding_lock);
462 list_for_each_entry_safe(b, t, &g_binding, link) {
463 list_del(&b->link);
464 spin_unlock_bh(&g_binding_lock);
465 dev_remove_pack(&b->ptype);
466 if (b->ptype.dev)
467 dev_put(b->ptype.dev);
468 kfree(b);
469 spin_lock_bh(&g_binding_lock);
470 }
471 spin_unlock_bh(&g_binding_lock);
472 /* Walk the list of PDs and stop each one. This causes the PD to be
473 * removed from the list so we can just pull each one from the head
474 * of the list.
475 */
476 spin_lock_bh(&g_polling_lock);
477 while (!list_empty(&g_pd_list)) {
478 struct oz_pd *pd =
479 list_first_entry(&g_pd_list, struct oz_pd, link);
480 oz_pd_get(pd);
481 spin_unlock_bh(&g_polling_lock);
482 oz_pd_stop(pd);
483 oz_pd_put(pd);
484 spin_lock_bh(&g_polling_lock);
485 }
486 spin_unlock_bh(&g_polling_lock);
487 oz_dbg(ON, "Protocol stopped\n");
488
489 kmem_cache_destroy(oz_tx_frame_cache);
490 kmem_cache_destroy(oz_elt_info_cache);
491 }
492
493 /*
494 * Context: softirq
495 */
496 void oz_pd_heartbeat_handler(unsigned long data)
497 {
498 struct oz_pd *pd = (struct oz_pd *)data;
499 u16 apps = 0;
500
501 spin_lock_bh(&g_polling_lock);
502 if (pd->state & OZ_PD_S_CONNECTED)
503 apps = pd->total_apps;
504 spin_unlock_bh(&g_polling_lock);
505 if (apps)
506 oz_pd_heartbeat(pd, apps);
507 oz_pd_put(pd);
508 }
509
510 /*
511 * Context: softirq
512 */
513 void oz_pd_timeout_handler(unsigned long data)
514 {
515 int type;
516 struct oz_pd *pd = (struct oz_pd *)data;
517
518 spin_lock_bh(&g_polling_lock);
519 type = pd->timeout_type;
520 spin_unlock_bh(&g_polling_lock);
521 switch (type) {
522 case OZ_TIMER_TOUT:
523 oz_pd_sleep(pd);
524 break;
525 case OZ_TIMER_STOP:
526 oz_pd_stop(pd);
527 break;
528 }
529 oz_pd_put(pd);
530 }
531
532 /*
533 * Context: Interrupt
534 */
535 enum hrtimer_restart oz_pd_heartbeat_event(struct hrtimer *timer)
536 {
537 struct oz_pd *pd;
538
539 pd = container_of(timer, struct oz_pd, heartbeat);
540 hrtimer_forward_now(timer, ktime_set(pd->pulse_period /
541 MSEC_PER_SEC, (pd->pulse_period % MSEC_PER_SEC) * NSEC_PER_MSEC));
542 oz_pd_get(pd);
543 tasklet_schedule(&pd->heartbeat_tasklet);
544 return HRTIMER_RESTART;
545 }
546
547 /*
548 * Context: Interrupt
549 */
550 enum hrtimer_restart oz_pd_timeout_event(struct hrtimer *timer)
551 {
552 struct oz_pd *pd;
553
554 pd = container_of(timer, struct oz_pd, timeout);
555 oz_pd_get(pd);
556 tasklet_schedule(&pd->timeout_tasklet);
557 return HRTIMER_NORESTART;
558 }
559
560 /*
561 * Context: softirq or process
562 */
563 void oz_timer_add(struct oz_pd *pd, int type, unsigned long due_time)
564 {
565 spin_lock_bh(&g_polling_lock);
566 switch (type) {
567 case OZ_TIMER_TOUT:
568 case OZ_TIMER_STOP:
569 if (hrtimer_active(&pd->timeout)) {
570 hrtimer_set_expires(&pd->timeout, ktime_set(due_time /
571 MSEC_PER_SEC, (due_time % MSEC_PER_SEC) *
572 NSEC_PER_MSEC));
573 hrtimer_start_expires(&pd->timeout, HRTIMER_MODE_REL);
574 } else {
575 hrtimer_start(&pd->timeout, ktime_set(due_time /
576 MSEC_PER_SEC, (due_time % MSEC_PER_SEC) *
577 NSEC_PER_MSEC), HRTIMER_MODE_REL);
578 }
579 pd->timeout_type = type;
580 break;
581 case OZ_TIMER_HEARTBEAT:
582 if (!hrtimer_active(&pd->heartbeat))
583 hrtimer_start(&pd->heartbeat, ktime_set(due_time /
584 MSEC_PER_SEC, (due_time % MSEC_PER_SEC) *
585 NSEC_PER_MSEC), HRTIMER_MODE_REL);
586 break;
587 }
588 spin_unlock_bh(&g_polling_lock);
589 }
590
591 /*
592 * Context: softirq or process
593 */
594 void oz_pd_request_heartbeat(struct oz_pd *pd)
595 {
596 oz_timer_add(pd, OZ_TIMER_HEARTBEAT, pd->pulse_period > 0 ?
597 pd->pulse_period : OZ_QUANTUM);
598 }
599
600 /*
601 * Context: softirq or process
602 */
603 struct oz_pd *oz_pd_find(const u8 *mac_addr)
604 {
605 struct oz_pd *pd;
606
607 spin_lock_bh(&g_polling_lock);
608 list_for_each_entry(pd, &g_pd_list, link) {
609 if (ether_addr_equal(pd->mac_addr, mac_addr)) {
610 oz_pd_get(pd);
611 spin_unlock_bh(&g_polling_lock);
612 return pd;
613 }
614 }
615 spin_unlock_bh(&g_polling_lock);
616 return NULL;
617 }
618
619 /*
620 * Context: process
621 */
622 void oz_app_enable(int app_id, int enable)
623 {
624 if (app_id < OZ_NB_APPS) {
625 spin_lock_bh(&g_polling_lock);
626 if (enable)
627 g_apps |= (1<<app_id);
628 else
629 g_apps &= ~(1<<app_id);
630 spin_unlock_bh(&g_polling_lock);
631 }
632 }
633
634 /*
635 * Context: softirq
636 */
637 static int oz_pkt_recv(struct sk_buff *skb, struct net_device *dev,
638 struct packet_type *pt, struct net_device *orig_dev)
639 {
640 skb = skb_share_check(skb, GFP_ATOMIC);
641 if (skb == NULL)
642 return 0;
643 spin_lock_bh(&g_rx_queue.lock);
644 if (g_processing_rx) {
645 /* We already hold the lock so use __ variant.
646 */
647 __skb_queue_head(&g_rx_queue, skb);
648 spin_unlock_bh(&g_rx_queue.lock);
649 } else {
650 g_processing_rx = 1;
651 do {
652
653 spin_unlock_bh(&g_rx_queue.lock);
654 oz_rx_frame(skb);
655 spin_lock_bh(&g_rx_queue.lock);
656 if (skb_queue_empty(&g_rx_queue)) {
657 g_processing_rx = 0;
658 spin_unlock_bh(&g_rx_queue.lock);
659 break;
660 }
661 /* We already hold the lock so use __ variant.
662 */
663 skb = __skb_dequeue(&g_rx_queue);
664 } while (1);
665 }
666 return 0;
667 }
668
669 /*
670 * Context: process
671 */
672 void oz_binding_add(const char *net_dev)
673 {
674 struct oz_binding *binding;
675
676 binding = kzalloc(sizeof(struct oz_binding), GFP_KERNEL);
677 if (!binding)
678 return;
679
680 binding->ptype.type = htons(OZ_ETHERTYPE);
681 binding->ptype.func = oz_pkt_recv;
682 if (net_dev && *net_dev) {
683 memcpy(binding->name, net_dev, OZ_MAX_BINDING_LEN);
684 oz_dbg(ON, "Adding binding: %s\n", net_dev);
685 binding->ptype.dev = dev_get_by_name(&init_net, net_dev);
686 if (binding->ptype.dev == NULL) {
687 oz_dbg(ON, "Netdev %s not found\n", net_dev);
688 kfree(binding);
689 return;
690 }
691 }
692 dev_add_pack(&binding->ptype);
693 spin_lock_bh(&g_binding_lock);
694 list_add_tail(&binding->link, &g_binding);
695 spin_unlock_bh(&g_binding_lock);
696 }
697
698 /*
699 * Context: process
700 */
701 static void pd_stop_all_for_device(struct net_device *net_dev)
702 {
703 LIST_HEAD(h);
704 struct oz_pd *pd;
705 struct oz_pd *n;
706
707 spin_lock_bh(&g_polling_lock);
708 list_for_each_entry_safe(pd, n, &g_pd_list, link) {
709 if (pd->net_dev == net_dev) {
710 list_move(&pd->link, &h);
711 oz_pd_get(pd);
712 }
713 }
714 spin_unlock_bh(&g_polling_lock);
715 while (!list_empty(&h)) {
716 pd = list_first_entry(&h, struct oz_pd, link);
717 oz_pd_stop(pd);
718 oz_pd_put(pd);
719 }
720 }
721
722 /*
723 * Context: process
724 */
725 void oz_binding_remove(const char *net_dev)
726 {
727 struct oz_binding *binding;
728 int found = 0;
729
730 oz_dbg(ON, "Removing binding: %s\n", net_dev);
731 spin_lock_bh(&g_binding_lock);
732 list_for_each_entry(binding, &g_binding, link) {
733 if (strncmp(binding->name, net_dev, OZ_MAX_BINDING_LEN) == 0) {
734 oz_dbg(ON, "Binding '%s' found\n", net_dev);
735 found = 1;
736 break;
737 }
738 }
739 spin_unlock_bh(&g_binding_lock);
740 if (found) {
741 dev_remove_pack(&binding->ptype);
742 if (binding->ptype.dev) {
743 dev_put(binding->ptype.dev);
744 pd_stop_all_for_device(binding->ptype.dev);
745 }
746 list_del(&binding->link);
747 kfree(binding);
748 }
749 }
750
751 /*
752 * Context: process
753 */
754 static char *oz_get_next_device_name(char *s, char *dname, int max_size)
755 {
756 while (*s == ',')
757 s++;
758 while (*s && (*s != ',') && max_size > 1) {
759 *dname++ = *s++;
760 max_size--;
761 }
762 *dname = 0;
763 return s;
764 }
765
766 /*
767 * Context: process
768 */
769 int oz_protocol_init(char *devs)
770 {
771 oz_elt_info_cache = KMEM_CACHE(oz_elt_info, 0);
772 if (!oz_elt_info_cache)
773 return -ENOMEM;
774
775 oz_tx_frame_cache = KMEM_CACHE(oz_tx_frame, 0);
776 if (!oz_tx_frame_cache) {
777 kmem_cache_destroy(oz_elt_info_cache);
778 return -ENOMEM;
779 }
780
781 skb_queue_head_init(&g_rx_queue);
782 if (devs[0] == '*') {
783 oz_binding_add(NULL);
784 } else {
785 char d[32];
786
787 while (*devs) {
788 devs = oz_get_next_device_name(devs, d, sizeof(d));
789 if (d[0])
790 oz_binding_add(d);
791 }
792 }
793 return 0;
794 }
795
796 /*
797 * Context: process
798 */
799 int oz_get_pd_list(struct oz_mac_addr *addr, int max_count)
800 {
801 struct oz_pd *pd;
802 int count = 0;
803
804 spin_lock_bh(&g_polling_lock);
805 list_for_each_entry(pd, &g_pd_list, link) {
806 if (count >= max_count)
807 break;
808 ether_addr_copy((u8 *)&addr[count++], pd->mac_addr);
809 }
810 spin_unlock_bh(&g_polling_lock);
811 return count;
812 }
813