]> git.proxmox.com Git - mirror_ubuntu-focal-kernel.git/blame - drivers/net/wireless/brcm80211/brcmfmac/msgbuf.c
brcmfmac: (clean) Rename sdio related files.
[mirror_ubuntu-focal-kernel.git] / drivers / net / wireless / brcm80211 / brcmfmac / msgbuf.c
CommitLineData
9a1bb602
HM
1/* Copyright (c) 2014 Broadcom Corporation
2 *
3 * Permission to use, copy, modify, and/or distribute this software for any
4 * purpose with or without fee is hereby granted, provided that the above
5 * copyright notice and this permission notice appear in all copies.
6 *
7 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
8 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
9 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
10 * SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
11 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION
12 * OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN
13 * CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
14 */
15
16/*******************************************************************************
17 * Communicates with the dongle by using dcmd codes.
18 * For certain dcmd codes, the dongle interprets string data from the host.
19 ******************************************************************************/
20
21#include <linux/types.h>
22#include <linux/netdevice.h>
23
24#include <brcmu_utils.h>
25#include <brcmu_wifi.h>
26
27#include "dhd.h"
a8e8ed34 28#include "debug.h"
9a1bb602
HM
29#include "proto.h"
30#include "msgbuf.h"
31#include "commonring.h"
32#include "flowring.h"
d14f78b9 33#include "bus.h"
9a1bb602
HM
34#include "tracepoint.h"
35
36
37#define MSGBUF_IOCTL_RESP_TIMEOUT 2000
38
39#define MSGBUF_TYPE_GEN_STATUS 0x1
40#define MSGBUF_TYPE_RING_STATUS 0x2
41#define MSGBUF_TYPE_FLOW_RING_CREATE 0x3
42#define MSGBUF_TYPE_FLOW_RING_CREATE_CMPLT 0x4
43#define MSGBUF_TYPE_FLOW_RING_DELETE 0x5
44#define MSGBUF_TYPE_FLOW_RING_DELETE_CMPLT 0x6
45#define MSGBUF_TYPE_FLOW_RING_FLUSH 0x7
46#define MSGBUF_TYPE_FLOW_RING_FLUSH_CMPLT 0x8
47#define MSGBUF_TYPE_IOCTLPTR_REQ 0x9
48#define MSGBUF_TYPE_IOCTLPTR_REQ_ACK 0xA
49#define MSGBUF_TYPE_IOCTLRESP_BUF_POST 0xB
50#define MSGBUF_TYPE_IOCTL_CMPLT 0xC
51#define MSGBUF_TYPE_EVENT_BUF_POST 0xD
52#define MSGBUF_TYPE_WL_EVENT 0xE
53#define MSGBUF_TYPE_TX_POST 0xF
54#define MSGBUF_TYPE_TX_STATUS 0x10
55#define MSGBUF_TYPE_RXBUF_POST 0x11
56#define MSGBUF_TYPE_RX_CMPLT 0x12
57#define MSGBUF_TYPE_LPBK_DMAXFER 0x13
58#define MSGBUF_TYPE_LPBK_DMAXFER_CMPLT 0x14
59
60#define NR_TX_PKTIDS 2048
61#define NR_RX_PKTIDS 1024
62
63#define BRCMF_IOCTL_REQ_PKTID 0xFFFE
64
65#define BRCMF_MSGBUF_MAX_PKT_SIZE 2048
66#define BRCMF_MSGBUF_RXBUFPOST_THRESHOLD 32
67#define BRCMF_MSGBUF_MAX_IOCTLRESPBUF_POST 8
68#define BRCMF_MSGBUF_MAX_EVENTBUF_POST 8
69
70#define BRCMF_MSGBUF_PKT_FLAGS_FRAME_802_3 0x01
71#define BRCMF_MSGBUF_PKT_FLAGS_PRIO_SHIFT 5
72
73#define BRCMF_MSGBUF_TX_FLUSH_CNT1 32
74#define BRCMF_MSGBUF_TX_FLUSH_CNT2 96
75
76
77struct msgbuf_common_hdr {
78 u8 msgtype;
79 u8 ifidx;
80 u8 flags;
81 u8 rsvd0;
82 __le32 request_id;
83};
84
85struct msgbuf_buf_addr {
86 __le32 low_addr;
87 __le32 high_addr;
88};
89
90struct msgbuf_ioctl_req_hdr {
91 struct msgbuf_common_hdr msg;
92 __le32 cmd;
93 __le16 trans_id;
94 __le16 input_buf_len;
95 __le16 output_buf_len;
96 __le16 rsvd0[3];
97 struct msgbuf_buf_addr req_buf_addr;
98 __le32 rsvd1[2];
99};
100
101struct msgbuf_tx_msghdr {
102 struct msgbuf_common_hdr msg;
103 u8 txhdr[ETH_HLEN];
104 u8 flags;
105 u8 seg_cnt;
106 struct msgbuf_buf_addr metadata_buf_addr;
107 struct msgbuf_buf_addr data_buf_addr;
108 __le16 metadata_buf_len;
109 __le16 data_len;
110 __le32 rsvd0;
111};
112
113struct msgbuf_rx_bufpost {
114 struct msgbuf_common_hdr msg;
115 __le16 metadata_buf_len;
116 __le16 data_buf_len;
117 __le32 rsvd0;
118 struct msgbuf_buf_addr metadata_buf_addr;
119 struct msgbuf_buf_addr data_buf_addr;
120};
121
122struct msgbuf_rx_ioctl_resp_or_event {
123 struct msgbuf_common_hdr msg;
124 __le16 host_buf_len;
125 __le16 rsvd0[3];
126 struct msgbuf_buf_addr host_buf_addr;
127 __le32 rsvd1[4];
128};
129
130struct msgbuf_completion_hdr {
131 __le16 status;
132 __le16 flow_ring_id;
133};
134
135struct msgbuf_rx_event {
136 struct msgbuf_common_hdr msg;
137 struct msgbuf_completion_hdr compl_hdr;
138 __le16 event_data_len;
139 __le16 seqnum;
140 __le16 rsvd0[4];
141};
142
143struct msgbuf_ioctl_resp_hdr {
144 struct msgbuf_common_hdr msg;
145 struct msgbuf_completion_hdr compl_hdr;
146 __le16 resp_len;
147 __le16 trans_id;
148 __le32 cmd;
149 __le32 rsvd0;
150};
151
152struct msgbuf_tx_status {
153 struct msgbuf_common_hdr msg;
154 struct msgbuf_completion_hdr compl_hdr;
155 __le16 metadata_len;
156 __le16 tx_status;
157};
158
159struct msgbuf_rx_complete {
160 struct msgbuf_common_hdr msg;
161 struct msgbuf_completion_hdr compl_hdr;
162 __le16 metadata_len;
163 __le16 data_len;
164 __le16 data_offset;
165 __le16 flags;
166 __le32 rx_status_0;
167 __le32 rx_status_1;
168 __le32 rsvd0;
169};
170
171struct msgbuf_tx_flowring_create_req {
172 struct msgbuf_common_hdr msg;
173 u8 da[ETH_ALEN];
174 u8 sa[ETH_ALEN];
175 u8 tid;
176 u8 if_flags;
177 __le16 flow_ring_id;
178 u8 tc;
179 u8 priority;
180 __le16 int_vector;
181 __le16 max_items;
182 __le16 len_item;
183 struct msgbuf_buf_addr flow_ring_addr;
184};
185
186struct msgbuf_tx_flowring_delete_req {
187 struct msgbuf_common_hdr msg;
188 __le16 flow_ring_id;
189 __le16 reason;
190 __le32 rsvd0[7];
191};
192
193struct msgbuf_flowring_create_resp {
194 struct msgbuf_common_hdr msg;
195 struct msgbuf_completion_hdr compl_hdr;
196 __le32 rsvd0[3];
197};
198
199struct msgbuf_flowring_delete_resp {
200 struct msgbuf_common_hdr msg;
201 struct msgbuf_completion_hdr compl_hdr;
202 __le32 rsvd0[3];
203};
204
205struct msgbuf_flowring_flush_resp {
206 struct msgbuf_common_hdr msg;
207 struct msgbuf_completion_hdr compl_hdr;
208 __le32 rsvd0[3];
209};
210
3ba06610
HM
211struct brcmf_msgbuf_work_item {
212 struct list_head queue;
213 u32 flowid;
214 int ifidx;
215 u8 sa[ETH_ALEN];
216 u8 da[ETH_ALEN];
217};
218
9a1bb602
HM
219struct brcmf_msgbuf {
220 struct brcmf_pub *drvr;
221
222 struct brcmf_commonring **commonrings;
223 struct brcmf_commonring **flowrings;
224 dma_addr_t *flowring_dma_handle;
225 u16 nrof_flowrings;
226
227 u16 rx_dataoffset;
228 u32 max_rxbufpost;
229 u16 rx_metadata_offset;
230 u32 rxbufpost;
231
232 u32 max_ioctlrespbuf;
233 u32 cur_ioctlrespbuf;
234 u32 max_eventbuf;
235 u32 cur_eventbuf;
236
237 void *ioctbuf;
238 dma_addr_t ioctbuf_handle;
239 u32 ioctbuf_phys_hi;
240 u32 ioctbuf_phys_lo;
ff0a6230 241 int ioctl_resp_status;
9a1bb602
HM
242 u32 ioctl_resp_ret_len;
243 u32 ioctl_resp_pktid;
244
245 u16 data_seq_no;
246 u16 ioctl_seq_no;
247 u32 reqid;
248 wait_queue_head_t ioctl_resp_wait;
249 bool ctl_completed;
250
251 struct brcmf_msgbuf_pktids *tx_pktids;
252 struct brcmf_msgbuf_pktids *rx_pktids;
253 struct brcmf_flowring *flow;
254
255 struct workqueue_struct *txflow_wq;
256 struct work_struct txflow_work;
257 unsigned long *flow_map;
258 unsigned long *txstatus_done_map;
3ba06610
HM
259
260 struct work_struct flowring_work;
261 spinlock_t flowring_work_lock;
262 struct list_head work_queue;
9a1bb602
HM
263};
264
265struct brcmf_msgbuf_pktid {
266 atomic_t allocated;
267 u16 data_offset;
268 struct sk_buff *skb;
269 dma_addr_t physaddr;
270};
271
272struct brcmf_msgbuf_pktids {
273 u32 array_size;
274 u32 last_allocated_idx;
275 enum dma_data_direction direction;
276 struct brcmf_msgbuf_pktid *array;
277};
278
279
280/* dma flushing needs implementation for mips and arm platforms. Should
281 * be put in util. Note, this is not real flushing. It is virtual non
282 * cached memory. Only write buffers should have to be drained. Though
283 * this may be different depending on platform......
284 */
285#define brcmf_dma_flush(addr, len)
286#define brcmf_dma_invalidate_cache(addr, len)
287
288
289static void brcmf_msgbuf_rxbuf_ioctlresp_post(struct brcmf_msgbuf *msgbuf);
290
291
292static struct brcmf_msgbuf_pktids *
293brcmf_msgbuf_init_pktids(u32 nr_array_entries,
294 enum dma_data_direction direction)
295{
296 struct brcmf_msgbuf_pktid *array;
297 struct brcmf_msgbuf_pktids *pktids;
298
3ba06610 299 array = kcalloc(nr_array_entries, sizeof(*array), GFP_KERNEL);
9a1bb602
HM
300 if (!array)
301 return NULL;
302
3ba06610 303 pktids = kzalloc(sizeof(*pktids), GFP_KERNEL);
9a1bb602
HM
304 if (!pktids) {
305 kfree(array);
306 return NULL;
307 }
308 pktids->array = array;
309 pktids->array_size = nr_array_entries;
310
311 return pktids;
312}
313
314
315static int
316brcmf_msgbuf_alloc_pktid(struct device *dev,
317 struct brcmf_msgbuf_pktids *pktids,
318 struct sk_buff *skb, u16 data_offset,
319 dma_addr_t *physaddr, u32 *idx)
320{
321 struct brcmf_msgbuf_pktid *array;
322 u32 count;
323
324 array = pktids->array;
325
326 *physaddr = dma_map_single(dev, skb->data + data_offset,
327 skb->len - data_offset, pktids->direction);
328
329 if (dma_mapping_error(dev, *physaddr)) {
330 brcmf_err("dma_map_single failed !!\n");
331 return -ENOMEM;
332 }
333
334 *idx = pktids->last_allocated_idx;
335
336 count = 0;
337 do {
338 (*idx)++;
339 if (*idx == pktids->array_size)
340 *idx = 0;
341 if (array[*idx].allocated.counter == 0)
342 if (atomic_cmpxchg(&array[*idx].allocated, 0, 1) == 0)
343 break;
344 count++;
345 } while (count < pktids->array_size);
346
347 if (count == pktids->array_size)
348 return -ENOMEM;
349
350 array[*idx].data_offset = data_offset;
351 array[*idx].physaddr = *physaddr;
352 array[*idx].skb = skb;
353
354 pktids->last_allocated_idx = *idx;
355
356 return 0;
357}
358
359
360static struct sk_buff *
361brcmf_msgbuf_get_pktid(struct device *dev, struct brcmf_msgbuf_pktids *pktids,
362 u32 idx)
363{
364 struct brcmf_msgbuf_pktid *pktid;
365 struct sk_buff *skb;
366
367 if (idx >= pktids->array_size) {
368 brcmf_err("Invalid packet id %d (max %d)\n", idx,
369 pktids->array_size);
370 return NULL;
371 }
372 if (pktids->array[idx].allocated.counter) {
373 pktid = &pktids->array[idx];
374 dma_unmap_single(dev, pktid->physaddr,
375 pktid->skb->len - pktid->data_offset,
376 pktids->direction);
377 skb = pktid->skb;
378 pktid->allocated.counter = 0;
379 return skb;
380 } else {
381 brcmf_err("Invalid packet id %d (not in use)\n", idx);
382 }
383
384 return NULL;
385}
386
387
388static void
389brcmf_msgbuf_release_array(struct device *dev,
390 struct brcmf_msgbuf_pktids *pktids)
391{
392 struct brcmf_msgbuf_pktid *array;
393 struct brcmf_msgbuf_pktid *pktid;
394 u32 count;
395
396 array = pktids->array;
397 count = 0;
398 do {
399 if (array[count].allocated.counter) {
400 pktid = &array[count];
401 dma_unmap_single(dev, pktid->physaddr,
402 pktid->skb->len - pktid->data_offset,
403 pktids->direction);
404 brcmu_pkt_buf_free_skb(pktid->skb);
405 }
406 count++;
407 } while (count < pktids->array_size);
408
409 kfree(array);
410 kfree(pktids);
411}
412
413
414static void brcmf_msgbuf_release_pktids(struct brcmf_msgbuf *msgbuf)
415{
416 if (msgbuf->rx_pktids)
417 brcmf_msgbuf_release_array(msgbuf->drvr->bus_if->dev,
418 msgbuf->rx_pktids);
419 if (msgbuf->tx_pktids)
420 brcmf_msgbuf_release_array(msgbuf->drvr->bus_if->dev,
421 msgbuf->tx_pktids);
422}
423
424
425static int brcmf_msgbuf_tx_ioctl(struct brcmf_pub *drvr, int ifidx,
426 uint cmd, void *buf, uint len)
427{
428 struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd;
429 struct brcmf_commonring *commonring;
430 struct msgbuf_ioctl_req_hdr *request;
431 u16 buf_len;
432 void *ret_ptr;
433 int err;
434
435 commonring = msgbuf->commonrings[BRCMF_H2D_MSGRING_CONTROL_SUBMIT];
436 brcmf_commonring_lock(commonring);
437 ret_ptr = brcmf_commonring_reserve_for_write(commonring);
438 if (!ret_ptr) {
439 brcmf_err("Failed to reserve space in commonring\n");
440 brcmf_commonring_unlock(commonring);
441 return -ENOMEM;
442 }
443
444 msgbuf->reqid++;
445
446 request = (struct msgbuf_ioctl_req_hdr *)ret_ptr;
447 request->msg.msgtype = MSGBUF_TYPE_IOCTLPTR_REQ;
448 request->msg.ifidx = (u8)ifidx;
449 request->msg.flags = 0;
450 request->msg.request_id = cpu_to_le32(BRCMF_IOCTL_REQ_PKTID);
451 request->cmd = cpu_to_le32(cmd);
452 request->output_buf_len = cpu_to_le16(len);
453 request->trans_id = cpu_to_le16(msgbuf->reqid);
454
455 buf_len = min_t(u16, len, BRCMF_TX_IOCTL_MAX_MSG_SIZE);
456 request->input_buf_len = cpu_to_le16(buf_len);
457 request->req_buf_addr.high_addr = cpu_to_le32(msgbuf->ioctbuf_phys_hi);
458 request->req_buf_addr.low_addr = cpu_to_le32(msgbuf->ioctbuf_phys_lo);
459 if (buf)
460 memcpy(msgbuf->ioctbuf, buf, buf_len);
461 else
462 memset(msgbuf->ioctbuf, 0, buf_len);
463 brcmf_dma_flush(ioctl_buf, buf_len);
464
465 err = brcmf_commonring_write_complete(commonring);
466 brcmf_commonring_unlock(commonring);
467
468 return err;
469}
470
471
472static int brcmf_msgbuf_ioctl_resp_wait(struct brcmf_msgbuf *msgbuf)
473{
474 return wait_event_timeout(msgbuf->ioctl_resp_wait,
475 msgbuf->ctl_completed,
476 msecs_to_jiffies(MSGBUF_IOCTL_RESP_TIMEOUT));
477}
478
479
480static void brcmf_msgbuf_ioctl_resp_wake(struct brcmf_msgbuf *msgbuf)
481{
482 if (waitqueue_active(&msgbuf->ioctl_resp_wait)) {
483 msgbuf->ctl_completed = true;
484 wake_up(&msgbuf->ioctl_resp_wait);
485 }
486}
487
488
489static int brcmf_msgbuf_query_dcmd(struct brcmf_pub *drvr, int ifidx,
490 uint cmd, void *buf, uint len)
491{
492 struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd;
493 struct sk_buff *skb = NULL;
494 int timeout;
495 int err;
496
497 brcmf_dbg(MSGBUF, "ifidx=%d, cmd=%d, len=%d\n", ifidx, cmd, len);
498 msgbuf->ctl_completed = false;
499 err = brcmf_msgbuf_tx_ioctl(drvr, ifidx, cmd, buf, len);
500 if (err)
501 return err;
502
503 timeout = brcmf_msgbuf_ioctl_resp_wait(msgbuf);
504 if (!timeout) {
505 brcmf_err("Timeout on response for query command\n");
506 return -EIO;
507 }
508
509 skb = brcmf_msgbuf_get_pktid(msgbuf->drvr->bus_if->dev,
510 msgbuf->rx_pktids,
511 msgbuf->ioctl_resp_pktid);
512 if (msgbuf->ioctl_resp_ret_len != 0) {
513 if (!skb) {
514 brcmf_err("Invalid packet id idx recv'd %d\n",
515 msgbuf->ioctl_resp_pktid);
516 return -EBADF;
517 }
518 memcpy(buf, skb->data, (len < msgbuf->ioctl_resp_ret_len) ?
519 len : msgbuf->ioctl_resp_ret_len);
520 }
521 if (skb)
522 brcmu_pkt_buf_free_skb(skb);
523
524 return msgbuf->ioctl_resp_status;
525}
526
527
528static int brcmf_msgbuf_set_dcmd(struct brcmf_pub *drvr, int ifidx,
529 uint cmd, void *buf, uint len)
530{
531 return brcmf_msgbuf_query_dcmd(drvr, ifidx, cmd, buf, len);
532}
533
534
535static int brcmf_msgbuf_hdrpull(struct brcmf_pub *drvr, bool do_fws,
536 u8 *ifidx, struct sk_buff *skb)
537{
538 return -ENODEV;
539}
540
541
542static void
543brcmf_msgbuf_remove_flowring(struct brcmf_msgbuf *msgbuf, u16 flowid)
544{
545 u32 dma_sz;
546 void *dma_buf;
547
548 brcmf_dbg(MSGBUF, "Removing flowring %d\n", flowid);
549
550 dma_sz = BRCMF_H2D_TXFLOWRING_MAX_ITEM * BRCMF_H2D_TXFLOWRING_ITEMSIZE;
551 dma_buf = msgbuf->flowrings[flowid]->buf_addr;
552 dma_free_coherent(msgbuf->drvr->bus_if->dev, dma_sz, dma_buf,
553 msgbuf->flowring_dma_handle[flowid]);
554
555 brcmf_flowring_delete(msgbuf->flow, flowid);
556}
557
558
3ba06610
HM
559static struct brcmf_msgbuf_work_item *
560brcmf_msgbuf_dequeue_work(struct brcmf_msgbuf *msgbuf)
561{
562 struct brcmf_msgbuf_work_item *work = NULL;
563 ulong flags;
564
565 spin_lock_irqsave(&msgbuf->flowring_work_lock, flags);
566 if (!list_empty(&msgbuf->work_queue)) {
567 work = list_first_entry(&msgbuf->work_queue,
568 struct brcmf_msgbuf_work_item, queue);
569 list_del(&work->queue);
570 }
571 spin_unlock_irqrestore(&msgbuf->flowring_work_lock, flags);
572
573 return work;
574}
575
576
577static u32
578brcmf_msgbuf_flowring_create_worker(struct brcmf_msgbuf *msgbuf,
579 struct brcmf_msgbuf_work_item *work)
9a1bb602
HM
580{
581 struct msgbuf_tx_flowring_create_req *create;
9a1bb602
HM
582 struct brcmf_commonring *commonring;
583 void *ret_ptr;
584 u32 flowid;
585 void *dma_buf;
586 u32 dma_sz;
587 long long address;
588 int err;
589
3ba06610 590 flowid = work->flowid;
9a1bb602 591 dma_sz = BRCMF_H2D_TXFLOWRING_MAX_ITEM * BRCMF_H2D_TXFLOWRING_ITEMSIZE;
9a1bb602
HM
592 dma_buf = dma_alloc_coherent(msgbuf->drvr->bus_if->dev, dma_sz,
593 &msgbuf->flowring_dma_handle[flowid],
3ba06610 594 GFP_KERNEL);
9a1bb602
HM
595 if (!dma_buf) {
596 brcmf_err("dma_alloc_coherent failed\n");
597 brcmf_flowring_delete(msgbuf->flow, flowid);
598 return BRCMF_FLOWRING_INVALID_ID;
599 }
600
601 brcmf_commonring_config(msgbuf->flowrings[flowid],
602 BRCMF_H2D_TXFLOWRING_MAX_ITEM,
603 BRCMF_H2D_TXFLOWRING_ITEMSIZE, dma_buf);
604
605 commonring = msgbuf->commonrings[BRCMF_H2D_MSGRING_CONTROL_SUBMIT];
606 brcmf_commonring_lock(commonring);
607 ret_ptr = brcmf_commonring_reserve_for_write(commonring);
608 if (!ret_ptr) {
609 brcmf_err("Failed to reserve space in commonring\n");
610 brcmf_commonring_unlock(commonring);
611 brcmf_msgbuf_remove_flowring(msgbuf, flowid);
612 return BRCMF_FLOWRING_INVALID_ID;
613 }
614
615 create = (struct msgbuf_tx_flowring_create_req *)ret_ptr;
616 create->msg.msgtype = MSGBUF_TYPE_FLOW_RING_CREATE;
3ba06610 617 create->msg.ifidx = work->ifidx;
9a1bb602
HM
618 create->msg.request_id = 0;
619 create->tid = brcmf_flowring_tid(msgbuf->flow, flowid);
620 create->flow_ring_id = cpu_to_le16(flowid +
621 BRCMF_NROF_H2D_COMMON_MSGRINGS);
3ba06610
HM
622 memcpy(create->sa, work->sa, ETH_ALEN);
623 memcpy(create->da, work->da, ETH_ALEN);
9a1bb602
HM
624 address = (long long)(long)msgbuf->flowring_dma_handle[flowid];
625 create->flow_ring_addr.high_addr = cpu_to_le32(address >> 32);
626 create->flow_ring_addr.low_addr = cpu_to_le32(address & 0xffffffff);
627 create->max_items = cpu_to_le16(BRCMF_H2D_TXFLOWRING_MAX_ITEM);
628 create->len_item = cpu_to_le16(BRCMF_H2D_TXFLOWRING_ITEMSIZE);
629
630 brcmf_dbg(MSGBUF, "Send Flow Create Req flow ID %d for peer %pM prio %d ifindex %d\n",
3ba06610 631 flowid, work->da, create->tid, work->ifidx);
9a1bb602
HM
632
633 err = brcmf_commonring_write_complete(commonring);
634 brcmf_commonring_unlock(commonring);
635 if (err) {
636 brcmf_err("Failed to write commonring\n");
637 brcmf_msgbuf_remove_flowring(msgbuf, flowid);
638 return BRCMF_FLOWRING_INVALID_ID;
639 }
640
641 return flowid;
642}
643
644
3ba06610
HM
645static void brcmf_msgbuf_flowring_worker(struct work_struct *work)
646{
647 struct brcmf_msgbuf *msgbuf;
648 struct brcmf_msgbuf_work_item *create;
649
650 msgbuf = container_of(work, struct brcmf_msgbuf, flowring_work);
651
652 while ((create = brcmf_msgbuf_dequeue_work(msgbuf))) {
653 brcmf_msgbuf_flowring_create_worker(msgbuf, create);
654 kfree(create);
655 }
656}
657
658
659static u32 brcmf_msgbuf_flowring_create(struct brcmf_msgbuf *msgbuf, int ifidx,
660 struct sk_buff *skb)
661{
662 struct brcmf_msgbuf_work_item *create;
663 struct ethhdr *eh = (struct ethhdr *)(skb->data);
664 u32 flowid;
665 ulong flags;
666
667 create = kzalloc(sizeof(*create), GFP_ATOMIC);
668 if (create == NULL)
669 return BRCMF_FLOWRING_INVALID_ID;
670
671 flowid = brcmf_flowring_create(msgbuf->flow, eh->h_dest,
672 skb->priority, ifidx);
673 if (flowid == BRCMF_FLOWRING_INVALID_ID) {
674 kfree(create);
675 return flowid;
676 }
677
678 create->flowid = flowid;
679 create->ifidx = ifidx;
680 memcpy(create->sa, eh->h_source, ETH_ALEN);
681 memcpy(create->da, eh->h_dest, ETH_ALEN);
682
683 spin_lock_irqsave(&msgbuf->flowring_work_lock, flags);
684 list_add_tail(&create->queue, &msgbuf->work_queue);
685 spin_unlock_irqrestore(&msgbuf->flowring_work_lock, flags);
686 schedule_work(&msgbuf->flowring_work);
687
688 return flowid;
689}
690
691
9a1bb602
HM
692static void brcmf_msgbuf_txflow(struct brcmf_msgbuf *msgbuf, u8 flowid)
693{
694 struct brcmf_flowring *flow = msgbuf->flow;
695 struct brcmf_commonring *commonring;
696 void *ret_ptr;
697 u32 count;
698 struct sk_buff *skb;
699 dma_addr_t physaddr;
700 u32 pktid;
701 struct msgbuf_tx_msghdr *tx_msghdr;
702 long long address;
703
704 commonring = msgbuf->flowrings[flowid];
705 if (!brcmf_commonring_write_available(commonring))
706 return;
707
708 brcmf_commonring_lock(commonring);
709
710 count = BRCMF_MSGBUF_TX_FLUSH_CNT2 - BRCMF_MSGBUF_TX_FLUSH_CNT1;
711 while (brcmf_flowring_qlen(flow, flowid)) {
712 skb = brcmf_flowring_dequeue(flow, flowid);
713 if (skb == NULL) {
714 brcmf_err("No SKB, but qlen %d\n",
715 brcmf_flowring_qlen(flow, flowid));
716 break;
717 }
718 skb_orphan(skb);
719 if (brcmf_msgbuf_alloc_pktid(msgbuf->drvr->bus_if->dev,
720 msgbuf->tx_pktids, skb, ETH_HLEN,
721 &physaddr, &pktid)) {
722 brcmf_flowring_reinsert(flow, flowid, skb);
723 brcmf_err("No PKTID available !!\n");
724 break;
725 }
726 ret_ptr = brcmf_commonring_reserve_for_write(commonring);
727 if (!ret_ptr) {
728 brcmf_msgbuf_get_pktid(msgbuf->drvr->bus_if->dev,
729 msgbuf->tx_pktids, pktid);
730 brcmf_flowring_reinsert(flow, flowid, skb);
731 break;
732 }
733 count++;
734
735 tx_msghdr = (struct msgbuf_tx_msghdr *)ret_ptr;
736
737 tx_msghdr->msg.msgtype = MSGBUF_TYPE_TX_POST;
738 tx_msghdr->msg.request_id = cpu_to_le32(pktid);
739 tx_msghdr->msg.ifidx = brcmf_flowring_ifidx_get(flow, flowid);
740 tx_msghdr->flags = BRCMF_MSGBUF_PKT_FLAGS_FRAME_802_3;
741 tx_msghdr->flags |= (skb->priority & 0x07) <<
742 BRCMF_MSGBUF_PKT_FLAGS_PRIO_SHIFT;
743 tx_msghdr->seg_cnt = 1;
744 memcpy(tx_msghdr->txhdr, skb->data, ETH_HLEN);
745 tx_msghdr->data_len = cpu_to_le16(skb->len - ETH_HLEN);
746 address = (long long)(long)physaddr;
747 tx_msghdr->data_buf_addr.high_addr = cpu_to_le32(address >> 32);
748 tx_msghdr->data_buf_addr.low_addr =
749 cpu_to_le32(address & 0xffffffff);
750 tx_msghdr->metadata_buf_len = 0;
751 tx_msghdr->metadata_buf_addr.high_addr = 0;
752 tx_msghdr->metadata_buf_addr.low_addr = 0;
753 if (count >= BRCMF_MSGBUF_TX_FLUSH_CNT2) {
754 brcmf_commonring_write_complete(commonring);
755 count = 0;
756 }
757 }
758 if (count)
759 brcmf_commonring_write_complete(commonring);
760 brcmf_commonring_unlock(commonring);
761}
762
763
764static void brcmf_msgbuf_txflow_worker(struct work_struct *worker)
765{
766 struct brcmf_msgbuf *msgbuf;
767 u32 flowid;
768
769 msgbuf = container_of(worker, struct brcmf_msgbuf, txflow_work);
770 for_each_set_bit(flowid, msgbuf->flow_map, msgbuf->nrof_flowrings) {
771 clear_bit(flowid, msgbuf->flow_map);
772 brcmf_msgbuf_txflow(msgbuf, flowid);
773 }
774}
775
776
777static int brcmf_msgbuf_schedule_txdata(struct brcmf_msgbuf *msgbuf, u32 flowid)
778{
779 set_bit(flowid, msgbuf->flow_map);
780 queue_work(msgbuf->txflow_wq, &msgbuf->txflow_work);
781
782 return 0;
783}
784
785
786static int brcmf_msgbuf_txdata(struct brcmf_pub *drvr, int ifidx,
787 u8 offset, struct sk_buff *skb)
788{
789 struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd;
790 struct brcmf_flowring *flow = msgbuf->flow;
791 struct ethhdr *eh = (struct ethhdr *)(skb->data);
792 u32 flowid;
793
794 flowid = brcmf_flowring_lookup(flow, eh->h_dest, skb->priority, ifidx);
795 if (flowid == BRCMF_FLOWRING_INVALID_ID) {
796 flowid = brcmf_msgbuf_flowring_create(msgbuf, ifidx, skb);
797 if (flowid == BRCMF_FLOWRING_INVALID_ID)
798 return -ENOMEM;
799 }
800 brcmf_flowring_enqueue(flow, flowid, skb);
801 brcmf_msgbuf_schedule_txdata(msgbuf, flowid);
802
803 return 0;
804}
805
806
807static void
808brcmf_msgbuf_configure_addr_mode(struct brcmf_pub *drvr, int ifidx,
809 enum proto_addr_mode addr_mode)
810{
811 struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd;
812
813 brcmf_flowring_configure_addr_mode(msgbuf->flow, ifidx, addr_mode);
814}
815
816
817static void
818brcmf_msgbuf_delete_peer(struct brcmf_pub *drvr, int ifidx, u8 peer[ETH_ALEN])
819{
820 struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd;
821
822 brcmf_flowring_delete_peer(msgbuf->flow, ifidx, peer);
823}
824
825
70b7d94b
HM
826static void
827brcmf_msgbuf_add_tdls_peer(struct brcmf_pub *drvr, int ifidx, u8 peer[ETH_ALEN])
828{
829 struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd;
830
831 brcmf_flowring_add_tdls_peer(msgbuf->flow, ifidx, peer);
832}
833
834
9a1bb602
HM
835static void
836brcmf_msgbuf_process_ioctl_complete(struct brcmf_msgbuf *msgbuf, void *buf)
837{
838 struct msgbuf_ioctl_resp_hdr *ioctl_resp;
839
840 ioctl_resp = (struct msgbuf_ioctl_resp_hdr *)buf;
841
ff0a6230
HM
842 msgbuf->ioctl_resp_status =
843 (s16)le16_to_cpu(ioctl_resp->compl_hdr.status);
9a1bb602
HM
844 msgbuf->ioctl_resp_ret_len = le16_to_cpu(ioctl_resp->resp_len);
845 msgbuf->ioctl_resp_pktid = le32_to_cpu(ioctl_resp->msg.request_id);
846
847 brcmf_msgbuf_ioctl_resp_wake(msgbuf);
848
849 if (msgbuf->cur_ioctlrespbuf)
850 msgbuf->cur_ioctlrespbuf--;
851 brcmf_msgbuf_rxbuf_ioctlresp_post(msgbuf);
852}
853
854
855static void
856brcmf_msgbuf_process_txstatus(struct brcmf_msgbuf *msgbuf, void *buf)
857{
858 struct msgbuf_tx_status *tx_status;
859 u32 idx;
860 struct sk_buff *skb;
861 u16 flowid;
862
863 tx_status = (struct msgbuf_tx_status *)buf;
864 idx = le32_to_cpu(tx_status->msg.request_id);
865 flowid = le16_to_cpu(tx_status->compl_hdr.flow_ring_id);
866 flowid -= BRCMF_NROF_H2D_COMMON_MSGRINGS;
867 skb = brcmf_msgbuf_get_pktid(msgbuf->drvr->bus_if->dev,
868 msgbuf->tx_pktids, idx);
869 if (!skb) {
870 brcmf_err("Invalid packet id idx recv'd %d\n", idx);
871 return;
872 }
873
874 set_bit(flowid, msgbuf->txstatus_done_map);
875
876 brcmf_txfinalize(msgbuf->drvr, skb, tx_status->msg.ifidx, true);
877}
878
879
880static u32 brcmf_msgbuf_rxbuf_data_post(struct brcmf_msgbuf *msgbuf, u32 count)
881{
882 struct brcmf_commonring *commonring;
883 void *ret_ptr;
884 struct sk_buff *skb;
885 u16 alloced;
886 u32 pktlen;
887 dma_addr_t physaddr;
888 struct msgbuf_rx_bufpost *rx_bufpost;
889 long long address;
890 u32 pktid;
891 u32 i;
892
893 commonring = msgbuf->commonrings[BRCMF_H2D_MSGRING_RXPOST_SUBMIT];
894 ret_ptr = brcmf_commonring_reserve_for_write_multiple(commonring,
895 count,
896 &alloced);
897 if (!ret_ptr) {
898 brcmf_err("Failed to reserve space in commonring\n");
899 return 0;
900 }
901
902 for (i = 0; i < alloced; i++) {
903 rx_bufpost = (struct msgbuf_rx_bufpost *)ret_ptr;
904 memset(rx_bufpost, 0, sizeof(*rx_bufpost));
905
906 skb = brcmu_pkt_buf_get_skb(BRCMF_MSGBUF_MAX_PKT_SIZE);
907
908 if (skb == NULL) {
909 brcmf_err("Failed to alloc SKB\n");
910 brcmf_commonring_write_cancel(commonring, alloced - i);
911 break;
912 }
913
914 pktlen = skb->len;
915 if (brcmf_msgbuf_alloc_pktid(msgbuf->drvr->bus_if->dev,
916 msgbuf->rx_pktids, skb, 0,
917 &physaddr, &pktid)) {
918 dev_kfree_skb_any(skb);
919 brcmf_err("No PKTID available !!\n");
920 brcmf_commonring_write_cancel(commonring, alloced - i);
921 break;
922 }
923
924 if (msgbuf->rx_metadata_offset) {
925 address = (long long)(long)physaddr;
926 rx_bufpost->metadata_buf_len =
927 cpu_to_le16(msgbuf->rx_metadata_offset);
928 rx_bufpost->metadata_buf_addr.high_addr =
929 cpu_to_le32(address >> 32);
930 rx_bufpost->metadata_buf_addr.low_addr =
931 cpu_to_le32(address & 0xffffffff);
932
933 skb_pull(skb, msgbuf->rx_metadata_offset);
934 pktlen = skb->len;
935 physaddr += msgbuf->rx_metadata_offset;
936 }
937 rx_bufpost->msg.msgtype = MSGBUF_TYPE_RXBUF_POST;
938 rx_bufpost->msg.request_id = cpu_to_le32(pktid);
939
940 address = (long long)(long)physaddr;
941 rx_bufpost->data_buf_len = cpu_to_le16((u16)pktlen);
942 rx_bufpost->data_buf_addr.high_addr =
943 cpu_to_le32(address >> 32);
944 rx_bufpost->data_buf_addr.low_addr =
945 cpu_to_le32(address & 0xffffffff);
946
947 ret_ptr += brcmf_commonring_len_item(commonring);
948 }
949
950 if (i)
951 brcmf_commonring_write_complete(commonring);
952
953 return i;
954}
955
956
957static void
958brcmf_msgbuf_rxbuf_data_fill(struct brcmf_msgbuf *msgbuf)
959{
960 u32 fillbufs;
961 u32 retcount;
962
963 fillbufs = msgbuf->max_rxbufpost - msgbuf->rxbufpost;
964
965 while (fillbufs) {
966 retcount = brcmf_msgbuf_rxbuf_data_post(msgbuf, fillbufs);
967 if (!retcount)
968 break;
969 msgbuf->rxbufpost += retcount;
970 fillbufs -= retcount;
971 }
972}
973
974
975static void
976brcmf_msgbuf_update_rxbufpost_count(struct brcmf_msgbuf *msgbuf, u16 rxcnt)
977{
978 msgbuf->rxbufpost -= rxcnt;
979 if (msgbuf->rxbufpost <= (msgbuf->max_rxbufpost -
980 BRCMF_MSGBUF_RXBUFPOST_THRESHOLD))
981 brcmf_msgbuf_rxbuf_data_fill(msgbuf);
982}
983
984
985static u32
986brcmf_msgbuf_rxbuf_ctrl_post(struct brcmf_msgbuf *msgbuf, bool event_buf,
987 u32 count)
988{
989 struct brcmf_commonring *commonring;
990 void *ret_ptr;
991 struct sk_buff *skb;
992 u16 alloced;
993 u32 pktlen;
994 dma_addr_t physaddr;
995 struct msgbuf_rx_ioctl_resp_or_event *rx_bufpost;
996 long long address;
997 u32 pktid;
998 u32 i;
999
1000 commonring = msgbuf->commonrings[BRCMF_H2D_MSGRING_CONTROL_SUBMIT];
1001 brcmf_commonring_lock(commonring);
1002 ret_ptr = brcmf_commonring_reserve_for_write_multiple(commonring,
1003 count,
1004 &alloced);
1005 if (!ret_ptr) {
1006 brcmf_err("Failed to reserve space in commonring\n");
1007 brcmf_commonring_unlock(commonring);
1008 return 0;
1009 }
1010
1011 for (i = 0; i < alloced; i++) {
1012 rx_bufpost = (struct msgbuf_rx_ioctl_resp_or_event *)ret_ptr;
1013 memset(rx_bufpost, 0, sizeof(*rx_bufpost));
1014
1015 skb = brcmu_pkt_buf_get_skb(BRCMF_MSGBUF_MAX_PKT_SIZE);
1016
1017 if (skb == NULL) {
1018 brcmf_err("Failed to alloc SKB\n");
1019 brcmf_commonring_write_cancel(commonring, alloced - i);
1020 break;
1021 }
1022
1023 pktlen = skb->len;
1024 if (brcmf_msgbuf_alloc_pktid(msgbuf->drvr->bus_if->dev,
1025 msgbuf->rx_pktids, skb, 0,
1026 &physaddr, &pktid)) {
1027 dev_kfree_skb_any(skb);
1028 brcmf_err("No PKTID available !!\n");
1029 brcmf_commonring_write_cancel(commonring, alloced - i);
1030 break;
1031 }
1032 if (event_buf)
1033 rx_bufpost->msg.msgtype = MSGBUF_TYPE_EVENT_BUF_POST;
1034 else
1035 rx_bufpost->msg.msgtype =
1036 MSGBUF_TYPE_IOCTLRESP_BUF_POST;
1037 rx_bufpost->msg.request_id = cpu_to_le32(pktid);
1038
1039 address = (long long)(long)physaddr;
1040 rx_bufpost->host_buf_len = cpu_to_le16((u16)pktlen);
1041 rx_bufpost->host_buf_addr.high_addr =
1042 cpu_to_le32(address >> 32);
1043 rx_bufpost->host_buf_addr.low_addr =
1044 cpu_to_le32(address & 0xffffffff);
1045
1046 ret_ptr += brcmf_commonring_len_item(commonring);
1047 }
1048
1049 if (i)
1050 brcmf_commonring_write_complete(commonring);
1051
1052 brcmf_commonring_unlock(commonring);
1053
1054 return i;
1055}
1056
1057
1058static void brcmf_msgbuf_rxbuf_ioctlresp_post(struct brcmf_msgbuf *msgbuf)
1059{
1060 u32 count;
1061
1062 count = msgbuf->max_ioctlrespbuf - msgbuf->cur_ioctlrespbuf;
1063 count = brcmf_msgbuf_rxbuf_ctrl_post(msgbuf, false, count);
1064 msgbuf->cur_ioctlrespbuf += count;
1065}
1066
1067
1068static void brcmf_msgbuf_rxbuf_event_post(struct brcmf_msgbuf *msgbuf)
1069{
1070 u32 count;
1071
1072 count = msgbuf->max_eventbuf - msgbuf->cur_eventbuf;
1073 count = brcmf_msgbuf_rxbuf_ctrl_post(msgbuf, true, count);
1074 msgbuf->cur_eventbuf += count;
1075}
1076
1077
1078static void
1079brcmf_msgbuf_rx_skb(struct brcmf_msgbuf *msgbuf, struct sk_buff *skb,
1080 u8 ifidx)
1081{
1082 struct brcmf_if *ifp;
1083
1084 ifp = msgbuf->drvr->iflist[ifidx];
1085 if (!ifp || !ifp->ndev) {
1086 brcmu_pkt_buf_free_skb(skb);
1087 return;
1088 }
1089 brcmf_netif_rx(ifp, skb);
1090}
1091
1092
1093static void brcmf_msgbuf_process_event(struct brcmf_msgbuf *msgbuf, void *buf)
1094{
1095 struct msgbuf_rx_event *event;
1096 u32 idx;
1097 u16 buflen;
1098 struct sk_buff *skb;
1099
1100 event = (struct msgbuf_rx_event *)buf;
1101 idx = le32_to_cpu(event->msg.request_id);
1102 buflen = le16_to_cpu(event->event_data_len);
1103
1104 if (msgbuf->cur_eventbuf)
1105 msgbuf->cur_eventbuf--;
1106 brcmf_msgbuf_rxbuf_event_post(msgbuf);
1107
1108 skb = brcmf_msgbuf_get_pktid(msgbuf->drvr->bus_if->dev,
1109 msgbuf->rx_pktids, idx);
1110 if (!skb)
1111 return;
1112
1113 if (msgbuf->rx_dataoffset)
1114 skb_pull(skb, msgbuf->rx_dataoffset);
1115
1116 skb_trim(skb, buflen);
1117
1118 brcmf_msgbuf_rx_skb(msgbuf, skb, event->msg.ifidx);
1119}
1120
1121
1122static void
1123brcmf_msgbuf_process_rx_complete(struct brcmf_msgbuf *msgbuf, void *buf)
1124{
1125 struct msgbuf_rx_complete *rx_complete;
1126 struct sk_buff *skb;
1127 u16 data_offset;
1128 u16 buflen;
1129 u32 idx;
1130
1131 brcmf_msgbuf_update_rxbufpost_count(msgbuf, 1);
1132
1133 rx_complete = (struct msgbuf_rx_complete *)buf;
1134 data_offset = le16_to_cpu(rx_complete->data_offset);
1135 buflen = le16_to_cpu(rx_complete->data_len);
1136 idx = le32_to_cpu(rx_complete->msg.request_id);
1137
1138 skb = brcmf_msgbuf_get_pktid(msgbuf->drvr->bus_if->dev,
1139 msgbuf->rx_pktids, idx);
1140
1141 if (data_offset)
1142 skb_pull(skb, data_offset);
1143 else if (msgbuf->rx_dataoffset)
1144 skb_pull(skb, msgbuf->rx_dataoffset);
1145
1146 skb_trim(skb, buflen);
1147
1148 brcmf_msgbuf_rx_skb(msgbuf, skb, rx_complete->msg.ifidx);
1149}
1150
1151
1152static void
1153brcmf_msgbuf_process_flow_ring_create_response(struct brcmf_msgbuf *msgbuf,
1154 void *buf)
1155{
1156 struct msgbuf_flowring_create_resp *flowring_create_resp;
1157 u16 status;
1158 u16 flowid;
1159
1160 flowring_create_resp = (struct msgbuf_flowring_create_resp *)buf;
1161
1162 flowid = le16_to_cpu(flowring_create_resp->compl_hdr.flow_ring_id);
1163 flowid -= BRCMF_NROF_H2D_COMMON_MSGRINGS;
1164 status = le16_to_cpu(flowring_create_resp->compl_hdr.status);
1165
1166 if (status) {
1167 brcmf_err("Flowring creation failed, code %d\n", status);
1168 brcmf_msgbuf_remove_flowring(msgbuf, flowid);
1169 return;
1170 }
1171 brcmf_dbg(MSGBUF, "Flowring %d Create response status %d\n", flowid,
1172 status);
1173
1174 brcmf_flowring_open(msgbuf->flow, flowid);
1175
1176 brcmf_msgbuf_schedule_txdata(msgbuf, flowid);
1177}
1178
1179
1180static void
1181brcmf_msgbuf_process_flow_ring_delete_response(struct brcmf_msgbuf *msgbuf,
1182 void *buf)
1183{
1184 struct msgbuf_flowring_delete_resp *flowring_delete_resp;
1185 u16 status;
1186 u16 flowid;
1187
1188 flowring_delete_resp = (struct msgbuf_flowring_delete_resp *)buf;
1189
1190 flowid = le16_to_cpu(flowring_delete_resp->compl_hdr.flow_ring_id);
1191 flowid -= BRCMF_NROF_H2D_COMMON_MSGRINGS;
1192 status = le16_to_cpu(flowring_delete_resp->compl_hdr.status);
1193
1194 if (status) {
1195 brcmf_err("Flowring deletion failed, code %d\n", status);
1196 brcmf_flowring_delete(msgbuf->flow, flowid);
1197 return;
1198 }
1199 brcmf_dbg(MSGBUF, "Flowring %d Delete response status %d\n", flowid,
1200 status);
1201
1202 brcmf_msgbuf_remove_flowring(msgbuf, flowid);
1203}
1204
1205
1206static void brcmf_msgbuf_process_msgtype(struct brcmf_msgbuf *msgbuf, void *buf)
1207{
1208 struct msgbuf_common_hdr *msg;
1209
1210 msg = (struct msgbuf_common_hdr *)buf;
1211 switch (msg->msgtype) {
1212 case MSGBUF_TYPE_FLOW_RING_CREATE_CMPLT:
1213 brcmf_dbg(MSGBUF, "MSGBUF_TYPE_FLOW_RING_CREATE_CMPLT\n");
1214 brcmf_msgbuf_process_flow_ring_create_response(msgbuf, buf);
1215 break;
1216 case MSGBUF_TYPE_FLOW_RING_DELETE_CMPLT:
1217 brcmf_dbg(MSGBUF, "MSGBUF_TYPE_FLOW_RING_DELETE_CMPLT\n");
1218 brcmf_msgbuf_process_flow_ring_delete_response(msgbuf, buf);
1219 break;
1220 case MSGBUF_TYPE_IOCTLPTR_REQ_ACK:
1221 brcmf_dbg(MSGBUF, "MSGBUF_TYPE_IOCTLPTR_REQ_ACK\n");
1222 break;
1223 case MSGBUF_TYPE_IOCTL_CMPLT:
1224 brcmf_dbg(MSGBUF, "MSGBUF_TYPE_IOCTL_CMPLT\n");
1225 brcmf_msgbuf_process_ioctl_complete(msgbuf, buf);
1226 break;
1227 case MSGBUF_TYPE_WL_EVENT:
1228 brcmf_dbg(MSGBUF, "MSGBUF_TYPE_WL_EVENT\n");
1229 brcmf_msgbuf_process_event(msgbuf, buf);
1230 break;
1231 case MSGBUF_TYPE_TX_STATUS:
1232 brcmf_dbg(MSGBUF, "MSGBUF_TYPE_TX_STATUS\n");
1233 brcmf_msgbuf_process_txstatus(msgbuf, buf);
1234 break;
1235 case MSGBUF_TYPE_RX_CMPLT:
1236 brcmf_dbg(MSGBUF, "MSGBUF_TYPE_RX_CMPLT\n");
1237 brcmf_msgbuf_process_rx_complete(msgbuf, buf);
1238 break;
1239 default:
1240 brcmf_err("Unsupported msgtype %d\n", msg->msgtype);
1241 break;
1242 }
1243}
1244
1245
1246static void brcmf_msgbuf_process_rx(struct brcmf_msgbuf *msgbuf,
1247 struct brcmf_commonring *commonring)
1248{
1249 void *buf;
1250 u16 count;
1251
1252again:
1253 buf = brcmf_commonring_get_read_ptr(commonring, &count);
1254 if (buf == NULL)
1255 return;
1256
1257 while (count) {
1258 brcmf_msgbuf_process_msgtype(msgbuf,
1259 buf + msgbuf->rx_dataoffset);
1260 buf += brcmf_commonring_len_item(commonring);
1261 count--;
1262 }
1263 brcmf_commonring_read_complete(commonring);
1264
1265 if (commonring->r_ptr == 0)
1266 goto again;
1267}
1268
1269
1270int brcmf_proto_msgbuf_rx_trigger(struct device *dev)
1271{
1272 struct brcmf_bus *bus_if = dev_get_drvdata(dev);
1273 struct brcmf_pub *drvr = bus_if->drvr;
1274 struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd;
1275 void *buf;
1276 u32 flowid;
1277
1278 buf = msgbuf->commonrings[BRCMF_D2H_MSGRING_RX_COMPLETE];
1279 brcmf_msgbuf_process_rx(msgbuf, buf);
1280 buf = msgbuf->commonrings[BRCMF_D2H_MSGRING_TX_COMPLETE];
1281 brcmf_msgbuf_process_rx(msgbuf, buf);
1282 buf = msgbuf->commonrings[BRCMF_D2H_MSGRING_CONTROL_COMPLETE];
1283 brcmf_msgbuf_process_rx(msgbuf, buf);
1284
1285 for_each_set_bit(flowid, msgbuf->txstatus_done_map,
1286 msgbuf->nrof_flowrings) {
1287 clear_bit(flowid, msgbuf->txstatus_done_map);
1288 if (brcmf_flowring_qlen(msgbuf->flow, flowid))
1289 brcmf_msgbuf_schedule_txdata(msgbuf, flowid);
1290 }
1291
1292 return 0;
1293}
1294
1295
1296void brcmf_msgbuf_delete_flowring(struct brcmf_pub *drvr, u8 flowid)
1297{
1298 struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd;
1299 struct msgbuf_tx_flowring_delete_req *delete;
1300 struct brcmf_commonring *commonring;
1301 void *ret_ptr;
1302 u8 ifidx;
1303 int err;
1304
1305 commonring = msgbuf->commonrings[BRCMF_H2D_MSGRING_CONTROL_SUBMIT];
1306 brcmf_commonring_lock(commonring);
1307 ret_ptr = brcmf_commonring_reserve_for_write(commonring);
1308 if (!ret_ptr) {
1309 brcmf_err("FW unaware, flowring will be removed !!\n");
1310 brcmf_commonring_unlock(commonring);
1311 brcmf_msgbuf_remove_flowring(msgbuf, flowid);
1312 return;
1313 }
1314
1315 delete = (struct msgbuf_tx_flowring_delete_req *)ret_ptr;
1316
1317 ifidx = brcmf_flowring_ifidx_get(msgbuf->flow, flowid);
1318
1319 delete->msg.msgtype = MSGBUF_TYPE_FLOW_RING_DELETE;
1320 delete->msg.ifidx = ifidx;
1321 delete->msg.request_id = 0;
1322
1323 delete->flow_ring_id = cpu_to_le16(flowid +
1324 BRCMF_NROF_H2D_COMMON_MSGRINGS);
1325 delete->reason = 0;
1326
1327 brcmf_dbg(MSGBUF, "Send Flow Delete Req flow ID %d, ifindex %d\n",
1328 flowid, ifidx);
1329
1330 err = brcmf_commonring_write_complete(commonring);
1331 brcmf_commonring_unlock(commonring);
1332 if (err) {
1333 brcmf_err("Failed to submit RING_DELETE, flowring will be removed\n");
1334 brcmf_msgbuf_remove_flowring(msgbuf, flowid);
1335 }
1336}
1337
1338
1339int brcmf_proto_msgbuf_attach(struct brcmf_pub *drvr)
1340{
1341 struct brcmf_bus_msgbuf *if_msgbuf;
1342 struct brcmf_msgbuf *msgbuf;
1343 long long address;
1344 u32 count;
1345
1346 if_msgbuf = drvr->bus_if->msgbuf;
3ba06610 1347 msgbuf = kzalloc(sizeof(*msgbuf), GFP_KERNEL);
9a1bb602
HM
1348 if (!msgbuf)
1349 goto fail;
1350
1351 msgbuf->txflow_wq = create_singlethread_workqueue("msgbuf_txflow");
1352 if (msgbuf->txflow_wq == NULL) {
1353 brcmf_err("workqueue creation failed\n");
1354 goto fail;
1355 }
1356 INIT_WORK(&msgbuf->txflow_work, brcmf_msgbuf_txflow_worker);
1357 count = BITS_TO_LONGS(if_msgbuf->nrof_flowrings);
3ba06610 1358 msgbuf->flow_map = kzalloc(count, GFP_KERNEL);
9a1bb602
HM
1359 if (!msgbuf->flow_map)
1360 goto fail;
1361
3ba06610 1362 msgbuf->txstatus_done_map = kzalloc(count, GFP_KERNEL);
9a1bb602
HM
1363 if (!msgbuf->txstatus_done_map)
1364 goto fail;
1365
1366 msgbuf->drvr = drvr;
1367 msgbuf->ioctbuf = dma_alloc_coherent(drvr->bus_if->dev,
1368 BRCMF_TX_IOCTL_MAX_MSG_SIZE,
1369 &msgbuf->ioctbuf_handle,
3ba06610 1370 GFP_KERNEL);
9a1bb602
HM
1371 if (!msgbuf->ioctbuf)
1372 goto fail;
1373 address = (long long)(long)msgbuf->ioctbuf_handle;
1374 msgbuf->ioctbuf_phys_hi = address >> 32;
1375 msgbuf->ioctbuf_phys_lo = address & 0xffffffff;
1376
1377 drvr->proto->hdrpull = brcmf_msgbuf_hdrpull;
1378 drvr->proto->query_dcmd = brcmf_msgbuf_query_dcmd;
1379 drvr->proto->set_dcmd = brcmf_msgbuf_set_dcmd;
1380 drvr->proto->txdata = brcmf_msgbuf_txdata;
1381 drvr->proto->configure_addr_mode = brcmf_msgbuf_configure_addr_mode;
1382 drvr->proto->delete_peer = brcmf_msgbuf_delete_peer;
70b7d94b 1383 drvr->proto->add_tdls_peer = brcmf_msgbuf_add_tdls_peer;
9a1bb602
HM
1384 drvr->proto->pd = msgbuf;
1385
1386 init_waitqueue_head(&msgbuf->ioctl_resp_wait);
1387
1388 msgbuf->commonrings =
1389 (struct brcmf_commonring **)if_msgbuf->commonrings;
1390 msgbuf->flowrings = (struct brcmf_commonring **)if_msgbuf->flowrings;
1391 msgbuf->nrof_flowrings = if_msgbuf->nrof_flowrings;
1392 msgbuf->flowring_dma_handle = kzalloc(msgbuf->nrof_flowrings *
3ba06610 1393 sizeof(*msgbuf->flowring_dma_handle), GFP_KERNEL);
2d116b88
AS
1394 if (!msgbuf->flowring_dma_handle)
1395 goto fail;
9a1bb602
HM
1396
1397 msgbuf->rx_dataoffset = if_msgbuf->rx_dataoffset;
1398 msgbuf->max_rxbufpost = if_msgbuf->max_rxbufpost;
1399
1400 msgbuf->max_ioctlrespbuf = BRCMF_MSGBUF_MAX_IOCTLRESPBUF_POST;
1401 msgbuf->max_eventbuf = BRCMF_MSGBUF_MAX_EVENTBUF_POST;
1402
1403 msgbuf->tx_pktids = brcmf_msgbuf_init_pktids(NR_TX_PKTIDS,
1404 DMA_TO_DEVICE);
1405 if (!msgbuf->tx_pktids)
1406 goto fail;
1407 msgbuf->rx_pktids = brcmf_msgbuf_init_pktids(NR_RX_PKTIDS,
1408 DMA_FROM_DEVICE);
1409 if (!msgbuf->rx_pktids)
1410 goto fail;
1411
1412 msgbuf->flow = brcmf_flowring_attach(drvr->bus_if->dev,
1413 if_msgbuf->nrof_flowrings);
1414 if (!msgbuf->flow)
1415 goto fail;
1416
1417
1418 brcmf_dbg(MSGBUF, "Feeding buffers, rx data %d, rx event %d, rx ioctl resp %d\n",
1419 msgbuf->max_rxbufpost, msgbuf->max_eventbuf,
1420 msgbuf->max_ioctlrespbuf);
1421 count = 0;
1422 do {
1423 brcmf_msgbuf_rxbuf_data_fill(msgbuf);
1424 if (msgbuf->max_rxbufpost != msgbuf->rxbufpost)
1425 msleep(10);
1426 else
1427 break;
1428 count++;
1429 } while (count < 10);
1430 brcmf_msgbuf_rxbuf_event_post(msgbuf);
1431 brcmf_msgbuf_rxbuf_ioctlresp_post(msgbuf);
1432
3ba06610
HM
1433 INIT_WORK(&msgbuf->flowring_work, brcmf_msgbuf_flowring_worker);
1434 spin_lock_init(&msgbuf->flowring_work_lock);
1435 INIT_LIST_HEAD(&msgbuf->work_queue);
1436
9a1bb602
HM
1437 return 0;
1438
1439fail:
1440 if (msgbuf) {
1441 kfree(msgbuf->flow_map);
1442 kfree(msgbuf->txstatus_done_map);
1443 brcmf_msgbuf_release_pktids(msgbuf);
2d116b88 1444 kfree(msgbuf->flowring_dma_handle);
9a1bb602
HM
1445 if (msgbuf->ioctbuf)
1446 dma_free_coherent(drvr->bus_if->dev,
1447 BRCMF_TX_IOCTL_MAX_MSG_SIZE,
1448 msgbuf->ioctbuf,
1449 msgbuf->ioctbuf_handle);
1450 kfree(msgbuf);
1451 }
1452 return -ENOMEM;
1453}
1454
1455
1456void brcmf_proto_msgbuf_detach(struct brcmf_pub *drvr)
1457{
1458 struct brcmf_msgbuf *msgbuf;
3ba06610 1459 struct brcmf_msgbuf_work_item *work;
9a1bb602
HM
1460
1461 brcmf_dbg(TRACE, "Enter\n");
1462 if (drvr->proto->pd) {
1463 msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd;
3ba06610
HM
1464 cancel_work_sync(&msgbuf->flowring_work);
1465 while (!list_empty(&msgbuf->work_queue)) {
1466 work = list_first_entry(&msgbuf->work_queue,
1467 struct brcmf_msgbuf_work_item,
1468 queue);
1469 list_del(&work->queue);
1470 kfree(work);
1471 }
9a1bb602
HM
1472 kfree(msgbuf->flow_map);
1473 kfree(msgbuf->txstatus_done_map);
1474 if (msgbuf->txflow_wq)
1475 destroy_workqueue(msgbuf->txflow_wq);
1476
1477 brcmf_flowring_detach(msgbuf->flow);
1478 dma_free_coherent(drvr->bus_if->dev,
1479 BRCMF_TX_IOCTL_MAX_MSG_SIZE,
1480 msgbuf->ioctbuf, msgbuf->ioctbuf_handle);
1481 brcmf_msgbuf_release_pktids(msgbuf);
2d116b88 1482 kfree(msgbuf->flowring_dma_handle);
9a1bb602
HM
1483 kfree(msgbuf);
1484 drvr->proto->pd = NULL;
1485 }
1486}