2 * Datapath implementation for ST-Ericsson CW1200 mac80211 drivers
4 * Copyright (c) 2010, ST-Ericsson
5 * Author: Dmitry Tarnyagin <dmitry.tarnyagin@lockless.no>
7 * This program is free software; you can redistribute it and/or modify
8 * it under the terms of the GNU General Public License version 2 as
9 * published by the Free Software Foundation.
12 #include <net/mac80211.h>
13 #include <linux/etherdevice.h>
14 #include <linux/skbuff.h>
22 #define CW1200_INVALID_RATE_ID (0xFF)
24 static int cw1200_handle_action_rx(struct cw1200_common
*priv
,
26 static const struct ieee80211_rate
*
27 cw1200_get_tx_rate(const struct cw1200_common
*priv
,
28 const struct ieee80211_tx_rate
*rate
);
30 /* ******************************************************************** */
31 /* TX queue lock / unlock */
33 static inline void cw1200_tx_queues_lock(struct cw1200_common
*priv
)
36 for (i
= 0; i
< 4; ++i
)
37 cw1200_queue_lock(&priv
->tx_queue
[i
]);
40 static inline void cw1200_tx_queues_unlock(struct cw1200_common
*priv
)
43 for (i
= 0; i
< 4; ++i
)
44 cw1200_queue_unlock(&priv
->tx_queue
[i
]);
47 /* ******************************************************************** */
48 /* TX policy cache implementation */
50 static void tx_policy_dump(struct tx_policy
*policy
)
52 pr_debug("[TX policy] %.1X%.1X%.1X%.1X%.1X%.1X%.1X%.1X %.1X%.1X%.1X%.1X%.1X%.1X%.1X%.1X %.1X%.1X%.1X%.1X%.1X%.1X%.1X%.1X: %d\n",
53 policy
->raw
[0] & 0x0F, policy
->raw
[0] >> 4,
54 policy
->raw
[1] & 0x0F, policy
->raw
[1] >> 4,
55 policy
->raw
[2] & 0x0F, policy
->raw
[2] >> 4,
56 policy
->raw
[3] & 0x0F, policy
->raw
[3] >> 4,
57 policy
->raw
[4] & 0x0F, policy
->raw
[4] >> 4,
58 policy
->raw
[5] & 0x0F, policy
->raw
[5] >> 4,
59 policy
->raw
[6] & 0x0F, policy
->raw
[6] >> 4,
60 policy
->raw
[7] & 0x0F, policy
->raw
[7] >> 4,
61 policy
->raw
[8] & 0x0F, policy
->raw
[8] >> 4,
62 policy
->raw
[9] & 0x0F, policy
->raw
[9] >> 4,
63 policy
->raw
[10] & 0x0F, policy
->raw
[10] >> 4,
64 policy
->raw
[11] & 0x0F, policy
->raw
[11] >> 4,
68 static void tx_policy_build(const struct cw1200_common
*priv
,
69 /* [out] */ struct tx_policy
*policy
,
70 struct ieee80211_tx_rate
*rates
, size_t count
)
73 unsigned limit
= priv
->short_frame_max_tx_count
;
75 BUG_ON(rates
[0].idx
< 0);
76 memset(policy
, 0, sizeof(*policy
));
78 /* Sort rates in descending order. */
79 for (i
= 1; i
< count
; ++i
) {
80 if (rates
[i
].idx
< 0) {
84 if (rates
[i
].idx
> rates
[i
- 1].idx
) {
85 struct ieee80211_tx_rate tmp
= rates
[i
- 1];
86 rates
[i
- 1] = rates
[i
];
91 /* Eliminate duplicates. */
92 total
= rates
[0].count
;
93 for (i
= 0, j
= 1; j
< count
; ++j
) {
94 if (rates
[j
].idx
== rates
[i
].idx
) {
95 rates
[i
].count
+= rates
[j
].count
;
96 } else if (rates
[j
].idx
> rates
[i
].idx
) {
103 total
+= rates
[j
].count
;
107 /* Re-fill policy trying to keep every requested rate and with
108 * respect to the global max tx retransmission count.
113 for (i
= 0; i
< count
; ++i
) {
114 int left
= count
- i
- 1;
115 if (rates
[i
].count
> limit
- left
)
116 rates
[i
].count
= limit
- left
;
117 limit
-= rates
[i
].count
;
121 /* HACK!!! Device has problems (at least) switching from
122 * 54Mbps CTS to 1Mbps. This switch takes enormous amount
123 * of time (100-200 ms), leading to valuable throughput drop.
124 * As a workaround, additional g-rates are injected to the
127 if (count
== 2 && !(rates
[0].flags
& IEEE80211_TX_RC_MCS
) &&
128 rates
[0].idx
> 4 && rates
[0].count
> 2 &&
130 int mid_rate
= (rates
[0].idx
+ 4) >> 1;
132 /* Decrease number of retries for the initial rate */
136 /* Keep fallback rate at 1Mbps. */
139 /* Inject 1 transmission on lowest g-rate */
142 rates
[2].flags
= rates
[1].flags
;
144 /* Inject 1 transmission on mid-rate */
145 rates
[1].idx
= mid_rate
;
148 /* Fallback to 1 Mbps is a really bad thing,
149 * so let's try to increase probability of
150 * successful transmission on the lowest g rate
153 if (rates
[0].count
>= 3) {
158 /* Adjust amount of rates defined */
161 /* Keep fallback rate at 1Mbps. */
164 /* Inject 2 transmissions on lowest g-rate */
168 /* Adjust amount of rates defined */
173 policy
->defined
= cw1200_get_tx_rate(priv
, &rates
[0])->hw_value
+ 1;
175 for (i
= 0; i
< count
; ++i
) {
176 register unsigned rateid
, off
, shift
, retries
;
178 rateid
= cw1200_get_tx_rate(priv
, &rates
[i
])->hw_value
;
179 off
= rateid
>> 3; /* eq. rateid / 8 */
180 shift
= (rateid
& 0x07) << 2; /* eq. (rateid % 8) * 4 */
182 retries
= rates
[i
].count
;
183 if (retries
> 0x0F) {
184 rates
[i
].count
= 0x0f;
187 policy
->tbl
[off
] |= __cpu_to_le32(retries
<< shift
);
188 policy
->retry_count
+= retries
;
191 pr_debug("[TX policy] Policy (%zu): %d:%d, %d:%d, %d:%d, %d:%d\n",
193 rates
[0].idx
, rates
[0].count
,
194 rates
[1].idx
, rates
[1].count
,
195 rates
[2].idx
, rates
[2].count
,
196 rates
[3].idx
, rates
[3].count
);
199 static inline bool tx_policy_is_equal(const struct tx_policy
*wanted
,
200 const struct tx_policy
*cached
)
202 size_t count
= wanted
->defined
>> 1;
203 if (wanted
->defined
> cached
->defined
)
206 if (memcmp(wanted
->raw
, cached
->raw
, count
))
209 if (wanted
->defined
& 1) {
210 if ((wanted
->raw
[count
] & 0x0F) != (cached
->raw
[count
] & 0x0F))
216 static int tx_policy_find(struct tx_policy_cache
*cache
,
217 const struct tx_policy
*wanted
)
219 /* O(n) complexity. Not so good, but there's only 8 entries in
221 * Also lru helps to reduce search time.
223 struct tx_policy_cache_entry
*it
;
224 /* First search for policy in "used" list */
225 list_for_each_entry(it
, &cache
->used
, link
) {
226 if (tx_policy_is_equal(wanted
, &it
->policy
))
227 return it
- cache
->cache
;
229 /* Then - in "free list" */
230 list_for_each_entry(it
, &cache
->free
, link
) {
231 if (tx_policy_is_equal(wanted
, &it
->policy
))
232 return it
- cache
->cache
;
237 static inline void tx_policy_use(struct tx_policy_cache
*cache
,
238 struct tx_policy_cache_entry
*entry
)
240 ++entry
->policy
.usage_count
;
241 list_move(&entry
->link
, &cache
->used
);
244 static inline int tx_policy_release(struct tx_policy_cache
*cache
,
245 struct tx_policy_cache_entry
*entry
)
247 int ret
= --entry
->policy
.usage_count
;
249 list_move(&entry
->link
, &cache
->free
);
253 void tx_policy_clean(struct cw1200_common
*priv
)
256 struct tx_policy_cache
*cache
= &priv
->tx_policy_cache
;
257 struct tx_policy_cache_entry
*entry
;
259 cw1200_tx_queues_lock(priv
);
260 spin_lock_bh(&cache
->lock
);
261 locked
= list_empty(&cache
->free
);
263 for (idx
= 0; idx
< TX_POLICY_CACHE_SIZE
; idx
++) {
264 entry
= &cache
->cache
[idx
];
265 /* Policy usage count should be 0 at this time as all queues
268 if (WARN_ON(entry
->policy
.usage_count
)) {
269 entry
->policy
.usage_count
= 0;
270 list_move(&entry
->link
, &cache
->free
);
272 memset(&entry
->policy
, 0, sizeof(entry
->policy
));
275 cw1200_tx_queues_unlock(priv
);
277 cw1200_tx_queues_unlock(priv
);
278 spin_unlock_bh(&cache
->lock
);
281 /* ******************************************************************** */
282 /* External TX policy cache API */
284 void tx_policy_init(struct cw1200_common
*priv
)
286 struct tx_policy_cache
*cache
= &priv
->tx_policy_cache
;
289 memset(cache
, 0, sizeof(*cache
));
291 spin_lock_init(&cache
->lock
);
292 INIT_LIST_HEAD(&cache
->used
);
293 INIT_LIST_HEAD(&cache
->free
);
295 for (i
= 0; i
< TX_POLICY_CACHE_SIZE
; ++i
)
296 list_add(&cache
->cache
[i
].link
, &cache
->free
);
299 static int tx_policy_get(struct cw1200_common
*priv
,
300 struct ieee80211_tx_rate
*rates
,
301 size_t count
, bool *renew
)
304 struct tx_policy_cache
*cache
= &priv
->tx_policy_cache
;
305 struct tx_policy wanted
;
307 tx_policy_build(priv
, &wanted
, rates
, count
);
309 spin_lock_bh(&cache
->lock
);
310 if (WARN_ON_ONCE(list_empty(&cache
->free
))) {
311 spin_unlock_bh(&cache
->lock
);
312 return CW1200_INVALID_RATE_ID
;
314 idx
= tx_policy_find(cache
, &wanted
);
316 pr_debug("[TX policy] Used TX policy: %d\n", idx
);
319 struct tx_policy_cache_entry
*entry
;
321 /* If policy is not found create a new one
322 * using the oldest entry in "free" list
324 entry
= list_entry(cache
->free
.prev
,
325 struct tx_policy_cache_entry
, link
);
326 entry
->policy
= wanted
;
327 idx
= entry
- cache
->cache
;
328 pr_debug("[TX policy] New TX policy: %d\n", idx
);
329 tx_policy_dump(&entry
->policy
);
331 tx_policy_use(cache
, &cache
->cache
[idx
]);
332 if (list_empty(&cache
->free
)) {
333 /* Lock TX queues. */
334 cw1200_tx_queues_lock(priv
);
336 spin_unlock_bh(&cache
->lock
);
340 static void tx_policy_put(struct cw1200_common
*priv
, int idx
)
343 struct tx_policy_cache
*cache
= &priv
->tx_policy_cache
;
345 spin_lock_bh(&cache
->lock
);
346 locked
= list_empty(&cache
->free
);
347 usage
= tx_policy_release(cache
, &cache
->cache
[idx
]);
348 if (locked
&& !usage
) {
349 /* Unlock TX queues. */
350 cw1200_tx_queues_unlock(priv
);
352 spin_unlock_bh(&cache
->lock
);
355 static int tx_policy_upload(struct cw1200_common
*priv
)
357 struct tx_policy_cache
*cache
= &priv
->tx_policy_cache
;
359 struct wsm_set_tx_rate_retry_policy arg
= {
362 spin_lock_bh(&cache
->lock
);
364 /* Upload only modified entries. */
365 for (i
= 0; i
< TX_POLICY_CACHE_SIZE
; ++i
) {
366 struct tx_policy
*src
= &cache
->cache
[i
].policy
;
367 if (src
->retry_count
&& !src
->uploaded
) {
368 struct wsm_tx_rate_retry_policy
*dst
=
371 dst
->short_retries
= priv
->short_frame_max_tx_count
;
372 dst
->long_retries
= priv
->long_frame_max_tx_count
;
374 dst
->flags
= WSM_TX_RATE_POLICY_FLAG_TERMINATE_WHEN_FINISHED
|
375 WSM_TX_RATE_POLICY_FLAG_COUNT_INITIAL_TRANSMIT
;
376 memcpy(dst
->rate_count_indices
, src
->tbl
,
377 sizeof(dst
->rate_count_indices
));
382 spin_unlock_bh(&cache
->lock
);
383 cw1200_debug_tx_cache_miss(priv
);
384 pr_debug("[TX policy] Upload %d policies\n", arg
.num
);
385 return wsm_set_tx_rate_retry_policy(priv
, &arg
);
388 void tx_policy_upload_work(struct work_struct
*work
)
390 struct cw1200_common
*priv
=
391 container_of(work
, struct cw1200_common
, tx_policy_upload_work
);
393 pr_debug("[TX] TX policy upload.\n");
394 tx_policy_upload(priv
);
397 cw1200_tx_queues_unlock(priv
);
400 /* ******************************************************************** */
401 /* cw1200 TX implementation */
403 struct cw1200_txinfo
{
406 struct ieee80211_tx_info
*tx_info
;
407 const struct ieee80211_rate
*rate
;
408 struct ieee80211_hdr
*hdr
;
411 struct cw1200_sta_priv
*sta_priv
;
412 struct ieee80211_sta
*sta
;
413 struct cw1200_txpriv txpriv
;
416 u32
cw1200_rate_mask_to_wsm(struct cw1200_common
*priv
, u32 rates
)
420 for (i
= 0; i
< 32; ++i
) {
422 ret
|= BIT(priv
->rates
[i
].hw_value
);
427 static const struct ieee80211_rate
*
428 cw1200_get_tx_rate(const struct cw1200_common
*priv
,
429 const struct ieee80211_tx_rate
*rate
)
433 if (rate
->flags
& IEEE80211_TX_RC_MCS
)
434 return &priv
->mcs_rates
[rate
->idx
];
435 return &priv
->hw
->wiphy
->bands
[priv
->channel
->band
]->
440 cw1200_tx_h_calc_link_ids(struct cw1200_common
*priv
,
441 struct cw1200_txinfo
*t
)
443 if (t
->sta
&& t
->sta_priv
->link_id
)
444 t
->txpriv
.raw_link_id
=
446 t
->sta_priv
->link_id
;
447 else if (priv
->mode
!= NL80211_IFTYPE_AP
)
448 t
->txpriv
.raw_link_id
=
449 t
->txpriv
.link_id
= 0;
450 else if (is_multicast_ether_addr(t
->da
)) {
451 if (priv
->enable_beacon
) {
452 t
->txpriv
.raw_link_id
= 0;
453 t
->txpriv
.link_id
= CW1200_LINK_ID_AFTER_DTIM
;
455 t
->txpriv
.raw_link_id
= 0;
456 t
->txpriv
.link_id
= 0;
459 t
->txpriv
.link_id
= cw1200_find_link_id(priv
, t
->da
);
460 if (!t
->txpriv
.link_id
)
461 t
->txpriv
.link_id
= cw1200_alloc_link_id(priv
, t
->da
);
462 if (!t
->txpriv
.link_id
) {
463 wiphy_err(priv
->hw
->wiphy
,
464 "No more link IDs available.\n");
467 t
->txpriv
.raw_link_id
= t
->txpriv
.link_id
;
469 if (t
->txpriv
.raw_link_id
)
470 priv
->link_id_db
[t
->txpriv
.raw_link_id
- 1].timestamp
=
472 if (t
->sta
&& (t
->sta
->uapsd_queues
& BIT(t
->queue
)))
473 t
->txpriv
.link_id
= CW1200_LINK_ID_UAPSD
;
478 cw1200_tx_h_pm(struct cw1200_common
*priv
,
479 struct cw1200_txinfo
*t
)
481 if (ieee80211_is_auth(t
->hdr
->frame_control
)) {
482 u32 mask
= ~BIT(t
->txpriv
.raw_link_id
);
483 spin_lock_bh(&priv
->ps_state_lock
);
484 priv
->sta_asleep_mask
&= mask
;
485 priv
->pspoll_mask
&= mask
;
486 spin_unlock_bh(&priv
->ps_state_lock
);
491 cw1200_tx_h_calc_tid(struct cw1200_common
*priv
,
492 struct cw1200_txinfo
*t
)
494 if (ieee80211_is_data_qos(t
->hdr
->frame_control
)) {
495 u8
*qos
= ieee80211_get_qos_ctl(t
->hdr
);
496 t
->txpriv
.tid
= qos
[0] & IEEE80211_QOS_CTL_TID_MASK
;
497 } else if (ieee80211_is_data(t
->hdr
->frame_control
)) {
503 cw1200_tx_h_crypt(struct cw1200_common
*priv
,
504 struct cw1200_txinfo
*t
)
506 if (!t
->tx_info
->control
.hw_key
||
507 !ieee80211_has_protected(t
->hdr
->frame_control
))
510 t
->hdrlen
+= t
->tx_info
->control
.hw_key
->iv_len
;
511 skb_put(t
->skb
, t
->tx_info
->control
.hw_key
->icv_len
);
513 if (t
->tx_info
->control
.hw_key
->cipher
== WLAN_CIPHER_SUITE_TKIP
)
514 skb_put(t
->skb
, 8); /* MIC space */
520 cw1200_tx_h_align(struct cw1200_common
*priv
,
521 struct cw1200_txinfo
*t
,
524 size_t offset
= (size_t)t
->skb
->data
& 3;
530 wiphy_err(priv
->hw
->wiphy
,
531 "Bug: attempt to transmit a frame with wrong alignment: %zu\n",
536 if (skb_headroom(t
->skb
) < offset
) {
537 wiphy_err(priv
->hw
->wiphy
,
538 "Bug: no space allocated for DMA alignment. headroom: %d\n",
539 skb_headroom(t
->skb
));
542 skb_push(t
->skb
, offset
);
544 t
->txpriv
.offset
+= offset
;
545 *flags
|= WSM_TX_2BYTES_SHIFT
;
546 cw1200_debug_tx_align(priv
);
551 cw1200_tx_h_action(struct cw1200_common
*priv
,
552 struct cw1200_txinfo
*t
)
554 struct ieee80211_mgmt
*mgmt
=
555 (struct ieee80211_mgmt
*)t
->hdr
;
556 if (ieee80211_is_action(t
->hdr
->frame_control
) &&
557 mgmt
->u
.action
.category
== WLAN_CATEGORY_BACK
)
564 static struct wsm_tx
*
565 cw1200_tx_h_wsm(struct cw1200_common
*priv
,
566 struct cw1200_txinfo
*t
)
570 if (skb_headroom(t
->skb
) < sizeof(struct wsm_tx
)) {
571 wiphy_err(priv
->hw
->wiphy
,
572 "Bug: no space allocated for WSM header. headroom: %d\n",
573 skb_headroom(t
->skb
));
577 wsm
= (struct wsm_tx
*)skb_push(t
->skb
, sizeof(struct wsm_tx
));
578 t
->txpriv
.offset
+= sizeof(struct wsm_tx
);
579 memset(wsm
, 0, sizeof(*wsm
));
580 wsm
->hdr
.len
= __cpu_to_le16(t
->skb
->len
);
581 wsm
->hdr
.id
= __cpu_to_le16(0x0004);
582 wsm
->queue_id
= wsm_queue_id_to_wsm(t
->queue
);
586 /* BT Coex specific handling */
588 cw1200_tx_h_bt(struct cw1200_common
*priv
,
589 struct cw1200_txinfo
*t
,
594 if (!priv
->bt_present
)
597 if (ieee80211_is_nullfunc(t
->hdr
->frame_control
)) {
598 priority
= WSM_EPTA_PRIORITY_MGT
;
599 } else if (ieee80211_is_data(t
->hdr
->frame_control
)) {
600 /* Skip LLC SNAP header (+6) */
601 u8
*payload
= &t
->skb
->data
[t
->hdrlen
];
602 __be16
*ethertype
= (__be16
*)&payload
[6];
603 if (be16_to_cpu(*ethertype
) == ETH_P_PAE
)
604 priority
= WSM_EPTA_PRIORITY_EAPOL
;
605 } else if (ieee80211_is_assoc_req(t
->hdr
->frame_control
) ||
606 ieee80211_is_reassoc_req(t
->hdr
->frame_control
)) {
607 struct ieee80211_mgmt
*mgt_frame
=
608 (struct ieee80211_mgmt
*)t
->hdr
;
610 if (le16_to_cpu(mgt_frame
->u
.assoc_req
.listen_interval
) <
611 priv
->listen_interval
) {
612 pr_debug("Modified Listen Interval to %d from %d\n",
613 priv
->listen_interval
,
614 mgt_frame
->u
.assoc_req
.listen_interval
);
615 /* Replace listen interval derieved from
616 * the one read from SDD
618 mgt_frame
->u
.assoc_req
.listen_interval
= cpu_to_le16(priv
->listen_interval
);
623 if (ieee80211_is_action(t
->hdr
->frame_control
))
624 priority
= WSM_EPTA_PRIORITY_ACTION
;
625 else if (ieee80211_is_mgmt(t
->hdr
->frame_control
))
626 priority
= WSM_EPTA_PRIORITY_MGT
;
627 else if ((wsm
->queue_id
== WSM_QUEUE_VOICE
))
628 priority
= WSM_EPTA_PRIORITY_VOICE
;
629 else if ((wsm
->queue_id
== WSM_QUEUE_VIDEO
))
630 priority
= WSM_EPTA_PRIORITY_VIDEO
;
632 priority
= WSM_EPTA_PRIORITY_DATA
;
635 pr_debug("[TX] EPTA priority %d.\n", priority
);
637 wsm
->flags
|= priority
<< 1;
641 cw1200_tx_h_rate_policy(struct cw1200_common
*priv
,
642 struct cw1200_txinfo
*t
,
645 bool tx_policy_renew
= false;
647 t
->txpriv
.rate_id
= tx_policy_get(priv
,
648 t
->tx_info
->control
.rates
, IEEE80211_TX_MAX_RATES
,
650 if (t
->txpriv
.rate_id
== CW1200_INVALID_RATE_ID
)
653 wsm
->flags
|= t
->txpriv
.rate_id
<< 4;
655 t
->rate
= cw1200_get_tx_rate(priv
,
656 &t
->tx_info
->control
.rates
[0]),
657 wsm
->max_tx_rate
= t
->rate
->hw_value
;
658 if (t
->rate
->flags
& IEEE80211_TX_RC_MCS
) {
659 if (cw1200_ht_greenfield(&priv
->ht_info
))
660 wsm
->ht_tx_parameters
|=
661 __cpu_to_le32(WSM_HT_TX_GREENFIELD
);
663 wsm
->ht_tx_parameters
|=
664 __cpu_to_le32(WSM_HT_TX_MIXED
);
667 if (tx_policy_renew
) {
668 pr_debug("[TX] TX policy renew.\n");
669 /* It's not so optimal to stop TX queues every now and then.
670 * Better to reimplement task scheduling with
673 wsm_lock_tx_async(priv
);
674 cw1200_tx_queues_lock(priv
);
675 if (queue_work(priv
->workqueue
,
676 &priv
->tx_policy_upload_work
) <= 0) {
677 cw1200_tx_queues_unlock(priv
);
685 cw1200_tx_h_pm_state(struct cw1200_common
*priv
,
686 struct cw1200_txinfo
*t
)
688 int was_buffered
= 1;
690 if (t
->txpriv
.link_id
== CW1200_LINK_ID_AFTER_DTIM
&&
691 !priv
->buffered_multicasts
) {
692 priv
->buffered_multicasts
= true;
693 if (priv
->sta_asleep_mask
)
694 queue_work(priv
->workqueue
,
695 &priv
->multicast_start_work
);
698 if (t
->txpriv
.raw_link_id
&& t
->txpriv
.tid
< CW1200_MAX_TID
)
699 was_buffered
= priv
->link_id_db
[t
->txpriv
.raw_link_id
- 1].buffered
[t
->txpriv
.tid
]++;
701 return !was_buffered
;
704 /* ******************************************************************** */
706 void cw1200_tx(struct ieee80211_hw
*dev
,
707 struct ieee80211_tx_control
*control
,
710 struct cw1200_common
*priv
= dev
->priv
;
711 struct cw1200_txinfo t
= {
713 .queue
= skb_get_queue_mapping(skb
),
714 .tx_info
= IEEE80211_SKB_CB(skb
),
715 .hdr
= (struct ieee80211_hdr
*)skb
->data
,
716 .txpriv
.tid
= CW1200_MAX_TID
,
717 .txpriv
.rate_id
= CW1200_INVALID_RATE_ID
,
719 struct ieee80211_sta
*sta
;
728 t
.hdrlen
= ieee80211_hdrlen(t
.hdr
->frame_control
);
729 t
.da
= ieee80211_get_DA(t
.hdr
);
731 t
.sta
= control
->sta
;
732 t
.sta_priv
= (struct cw1200_sta_priv
*)&t
.sta
->drv_priv
;
735 if (WARN_ON(t
.queue
>= 4))
738 ret
= cw1200_tx_h_calc_link_ids(priv
, &t
);
742 pr_debug("[TX] TX %d bytes (queue: %d, link_id: %d (%d)).\n",
743 skb
->len
, t
.queue
, t
.txpriv
.link_id
,
744 t
.txpriv
.raw_link_id
);
746 cw1200_tx_h_pm(priv
, &t
);
747 cw1200_tx_h_calc_tid(priv
, &t
);
748 ret
= cw1200_tx_h_crypt(priv
, &t
);
751 ret
= cw1200_tx_h_align(priv
, &t
, &flags
);
754 ret
= cw1200_tx_h_action(priv
, &t
);
757 wsm
= cw1200_tx_h_wsm(priv
, &t
);
763 cw1200_tx_h_bt(priv
, &t
, wsm
);
764 ret
= cw1200_tx_h_rate_policy(priv
, &t
, wsm
);
769 sta
= rcu_dereference(t
.sta
);
771 spin_lock_bh(&priv
->ps_state_lock
);
773 tid_update
= cw1200_tx_h_pm_state(priv
, &t
);
774 BUG_ON(cw1200_queue_put(&priv
->tx_queue
[t
.queue
],
777 spin_unlock_bh(&priv
->ps_state_lock
);
779 if (tid_update
&& sta
)
780 ieee80211_sta_set_buffered(sta
, t
.txpriv
.tid
, true);
784 cw1200_bh_wakeup(priv
);
789 cw1200_skb_dtor(priv
, skb
, &t
.txpriv
);
793 /* ******************************************************************** */
795 static int cw1200_handle_action_rx(struct cw1200_common
*priv
,
798 struct ieee80211_mgmt
*mgmt
= (void *)skb
->data
;
800 /* Filter block ACK negotiation: fully controlled by firmware */
801 if (mgmt
->u
.action
.category
== WLAN_CATEGORY_BACK
)
807 static int cw1200_handle_pspoll(struct cw1200_common
*priv
,
810 struct ieee80211_sta
*sta
;
811 struct ieee80211_pspoll
*pspoll
= (struct ieee80211_pspoll
*)skb
->data
;
817 if (priv
->join_status
!= CW1200_JOIN_STATUS_AP
)
819 if (memcmp(priv
->vif
->addr
, pspoll
->bssid
, ETH_ALEN
))
823 sta
= ieee80211_find_sta(priv
->vif
, pspoll
->ta
);
825 struct cw1200_sta_priv
*sta_priv
;
826 sta_priv
= (struct cw1200_sta_priv
*)&sta
->drv_priv
;
827 link_id
= sta_priv
->link_id
;
828 pspoll_mask
= BIT(sta_priv
->link_id
);
834 priv
->pspoll_mask
|= pspoll_mask
;
837 /* Do not report pspols if data for given link id is queued already. */
838 for (i
= 0; i
< 4; ++i
) {
839 if (cw1200_queue_get_num_queued(&priv
->tx_queue
[i
],
841 cw1200_bh_wakeup(priv
);
846 pr_debug("[RX] PSPOLL: %s\n", drop
? "local" : "fwd");
851 /* ******************************************************************** */
853 void cw1200_tx_confirm_cb(struct cw1200_common
*priv
,
855 struct wsm_tx_confirm
*arg
)
857 u8 queue_id
= cw1200_queue_get_queue_id(arg
->packet_id
);
858 struct cw1200_queue
*queue
= &priv
->tx_queue
[queue_id
];
860 const struct cw1200_txpriv
*txpriv
;
862 pr_debug("[TX] TX confirm: %d, %d.\n",
863 arg
->status
, arg
->ack_failures
);
865 if (priv
->mode
== NL80211_IFTYPE_UNSPECIFIED
) {
866 /* STA is stopped. */
870 if (WARN_ON(queue_id
>= 4))
874 pr_debug("TX failed: %d.\n", arg
->status
);
876 if ((arg
->status
== WSM_REQUEUE
) &&
877 (arg
->flags
& WSM_TX_STATUS_REQUEUE
)) {
878 /* "Requeue" means "implicit suspend" */
879 struct wsm_suspend_resume suspend
= {
882 .multicast
= !link_id
,
884 cw1200_suspend_resume(priv
, &suspend
);
885 wiphy_warn(priv
->hw
->wiphy
, "Requeue for link_id %d (try %d). STAs asleep: 0x%.8X\n",
887 cw1200_queue_get_generation(arg
->packet_id
) + 1,
888 priv
->sta_asleep_mask
);
889 cw1200_queue_requeue(queue
, arg
->packet_id
);
890 spin_lock_bh(&priv
->ps_state_lock
);
892 priv
->buffered_multicasts
= true;
893 if (priv
->sta_asleep_mask
) {
894 queue_work(priv
->workqueue
,
895 &priv
->multicast_start_work
);
898 spin_unlock_bh(&priv
->ps_state_lock
);
899 } else if (!cw1200_queue_get_skb(queue
, arg
->packet_id
,
901 struct ieee80211_tx_info
*tx
= IEEE80211_SKB_CB(skb
);
902 int tx_count
= arg
->ack_failures
;
906 if (cw1200_ht_greenfield(&priv
->ht_info
))
907 ht_flags
|= IEEE80211_TX_RC_GREEN_FIELD
;
909 spin_lock(&priv
->bss_loss_lock
);
910 if (priv
->bss_loss_state
&&
911 arg
->packet_id
== priv
->bss_loss_confirm_id
) {
913 /* Recovery failed */
914 __cw1200_cqm_bssloss_sm(priv
, 0, 0, 1);
916 /* Recovery succeeded */
917 __cw1200_cqm_bssloss_sm(priv
, 0, 1, 0);
920 spin_unlock(&priv
->bss_loss_lock
);
923 tx
->flags
|= IEEE80211_TX_STAT_ACK
;
925 cw1200_debug_txed(priv
);
926 if (arg
->flags
& WSM_TX_STATUS_AGGREGATION
) {
927 /* Do not report aggregation to mac80211:
928 * it confuses minstrel a lot.
930 /* tx->flags |= IEEE80211_TX_STAT_AMPDU; */
931 cw1200_debug_txed_agg(priv
);
938 for (i
= 0; i
< IEEE80211_TX_MAX_RATES
; ++i
) {
939 if (tx
->status
.rates
[i
].count
>= tx_count
) {
940 tx
->status
.rates
[i
].count
= tx_count
;
943 tx_count
-= tx
->status
.rates
[i
].count
;
944 if (tx
->status
.rates
[i
].flags
& IEEE80211_TX_RC_MCS
)
945 tx
->status
.rates
[i
].flags
|= ht_flags
;
948 for (++i
; i
< IEEE80211_TX_MAX_RATES
; ++i
) {
949 tx
->status
.rates
[i
].count
= 0;
950 tx
->status
.rates
[i
].idx
= -1;
953 /* Pull off any crypto trailers that we added on */
954 if (tx
->control
.hw_key
) {
955 skb_trim(skb
, skb
->len
- tx
->control
.hw_key
->icv_len
);
956 if (tx
->control
.hw_key
->cipher
== WLAN_CIPHER_SUITE_TKIP
)
957 skb_trim(skb
, skb
->len
- 8); /* MIC space */
959 cw1200_queue_remove(queue
, arg
->packet_id
);
961 /* XXX TODO: Only wake if there are pending transmits.. */
962 cw1200_bh_wakeup(priv
);
965 static void cw1200_notify_buffered_tx(struct cw1200_common
*priv
,
966 struct sk_buff
*skb
, int link_id
, int tid
)
968 struct ieee80211_sta
*sta
;
969 struct ieee80211_hdr
*hdr
;
971 u8 still_buffered
= 0;
973 if (link_id
&& tid
< CW1200_MAX_TID
) {
974 buffered
= priv
->link_id_db
975 [link_id
- 1].buffered
;
977 spin_lock_bh(&priv
->ps_state_lock
);
978 if (!WARN_ON(!buffered
[tid
]))
979 still_buffered
= --buffered
[tid
];
980 spin_unlock_bh(&priv
->ps_state_lock
);
982 if (!still_buffered
&& tid
< CW1200_MAX_TID
) {
983 hdr
= (struct ieee80211_hdr
*)skb
->data
;
985 sta
= ieee80211_find_sta(priv
->vif
, hdr
->addr1
);
987 ieee80211_sta_set_buffered(sta
, tid
, false);
993 void cw1200_skb_dtor(struct cw1200_common
*priv
,
995 const struct cw1200_txpriv
*txpriv
)
997 skb_pull(skb
, txpriv
->offset
);
998 if (txpriv
->rate_id
!= CW1200_INVALID_RATE_ID
) {
999 cw1200_notify_buffered_tx(priv
, skb
,
1000 txpriv
->raw_link_id
, txpriv
->tid
);
1001 tx_policy_put(priv
, txpriv
->rate_id
);
1003 ieee80211_tx_status(priv
->hw
, skb
);
1006 void cw1200_rx_cb(struct cw1200_common
*priv
,
1009 struct sk_buff
**skb_p
)
1011 struct sk_buff
*skb
= *skb_p
;
1012 struct ieee80211_rx_status
*hdr
= IEEE80211_SKB_RXCB(skb
);
1013 struct ieee80211_hdr
*frame
= (struct ieee80211_hdr
*)skb
->data
;
1014 struct ieee80211_mgmt
*mgmt
= (struct ieee80211_mgmt
*)skb
->data
;
1015 struct cw1200_link_entry
*entry
= NULL
;
1016 unsigned long grace_period
;
1018 bool early_data
= false;
1019 bool p2p
= priv
->vif
&& priv
->vif
->p2p
;
1023 if (priv
->mode
== NL80211_IFTYPE_UNSPECIFIED
) {
1024 /* STA is stopped. */
1028 if (link_id
&& link_id
<= CW1200_MAX_STA_IN_AP_MODE
) {
1029 entry
= &priv
->link_id_db
[link_id
- 1];
1030 if (entry
->status
== CW1200_LINK_SOFT
&&
1031 ieee80211_is_data(frame
->frame_control
))
1033 entry
->timestamp
= jiffies
;
1035 ieee80211_is_action(frame
->frame_control
) &&
1036 (mgmt
->u
.action
.category
== WLAN_CATEGORY_PUBLIC
)) {
1037 pr_debug("[RX] Going to MAP&RESET link ID\n");
1038 WARN_ON(work_pending(&priv
->linkid_reset_work
));
1039 memcpy(&priv
->action_frame_sa
[0],
1040 ieee80211_get_SA(frame
), ETH_ALEN
);
1041 priv
->action_linkid
= 0;
1042 schedule_work(&priv
->linkid_reset_work
);
1045 if (link_id
&& p2p
&&
1046 ieee80211_is_action(frame
->frame_control
) &&
1047 (mgmt
->u
.action
.category
== WLAN_CATEGORY_PUBLIC
)) {
1048 /* Link ID already exists for the ACTION frame.
1051 WARN_ON(work_pending(&priv
->linkid_reset_work
));
1052 memcpy(&priv
->action_frame_sa
[0],
1053 ieee80211_get_SA(frame
), ETH_ALEN
);
1054 priv
->action_linkid
= link_id
;
1055 schedule_work(&priv
->linkid_reset_work
);
1058 if (arg
->status
== WSM_STATUS_MICFAILURE
) {
1059 pr_debug("[RX] MIC failure.\n");
1060 hdr
->flag
|= RX_FLAG_MMIC_ERROR
;
1061 } else if (arg
->status
== WSM_STATUS_NO_KEY_FOUND
) {
1062 pr_debug("[RX] No key found.\n");
1065 pr_debug("[RX] Receive failure: %d.\n",
1071 if (skb
->len
< sizeof(struct ieee80211_pspoll
)) {
1072 wiphy_warn(priv
->hw
->wiphy
, "Mailformed SDU rx'ed. Size is lesser than IEEE header.\n");
1076 if (ieee80211_is_pspoll(frame
->frame_control
))
1077 if (cw1200_handle_pspoll(priv
, skb
))
1080 hdr
->band
= ((arg
->channel_number
& 0xff00) ||
1081 (arg
->channel_number
> 14)) ?
1082 IEEE80211_BAND_5GHZ
: IEEE80211_BAND_2GHZ
;
1083 hdr
->freq
= ieee80211_channel_to_frequency(
1084 arg
->channel_number
,
1087 if (arg
->rx_rate
>= 14) {
1088 hdr
->flag
|= RX_FLAG_HT
;
1089 hdr
->rate_idx
= arg
->rx_rate
- 14;
1090 } else if (arg
->rx_rate
>= 4) {
1091 hdr
->rate_idx
= arg
->rx_rate
- 2;
1093 hdr
->rate_idx
= arg
->rx_rate
;
1096 hdr
->signal
= (s8
)arg
->rcpi_rssi
;
1099 hdrlen
= ieee80211_hdrlen(frame
->frame_control
);
1101 if (WSM_RX_STATUS_ENCRYPTION(arg
->flags
)) {
1102 size_t iv_len
= 0, icv_len
= 0;
1104 hdr
->flag
|= RX_FLAG_DECRYPTED
| RX_FLAG_IV_STRIPPED
;
1106 /* Oops... There is no fast way to ask mac80211 about
1107 * IV/ICV lengths. Even defineas are not exposed.
1109 switch (WSM_RX_STATUS_ENCRYPTION(arg
->flags
)) {
1110 case WSM_RX_STATUS_WEP
:
1111 iv_len
= 4 /* WEP_IV_LEN */;
1112 icv_len
= 4 /* WEP_ICV_LEN */;
1114 case WSM_RX_STATUS_TKIP
:
1115 iv_len
= 8 /* TKIP_IV_LEN */;
1116 icv_len
= 4 /* TKIP_ICV_LEN */
1117 + 8 /*MICHAEL_MIC_LEN*/;
1118 hdr
->flag
|= RX_FLAG_MMIC_STRIPPED
;
1120 case WSM_RX_STATUS_AES
:
1121 iv_len
= 8 /* CCMP_HDR_LEN */;
1122 icv_len
= 8 /* CCMP_MIC_LEN */;
1124 case WSM_RX_STATUS_WAPI
:
1125 iv_len
= 18 /* WAPI_HDR_LEN */;
1126 icv_len
= 16 /* WAPI_MIC_LEN */;
1129 pr_warn("Unknown encryption type %d\n",
1130 WSM_RX_STATUS_ENCRYPTION(arg
->flags
));
1134 /* Firmware strips ICV in case of MIC failure. */
1135 if (arg
->status
== WSM_STATUS_MICFAILURE
)
1138 if (skb
->len
< hdrlen
+ iv_len
+ icv_len
) {
1139 wiphy_warn(priv
->hw
->wiphy
, "Malformed SDU rx'ed. Size is lesser than crypto headers.\n");
1143 /* Remove IV, ICV and MIC */
1144 skb_trim(skb
, skb
->len
- icv_len
);
1145 memmove(skb
->data
+ iv_len
, skb
->data
, hdrlen
);
1146 skb_pull(skb
, iv_len
);
1149 /* Remove TSF from the end of frame */
1150 if (arg
->flags
& WSM_RX_STATUS_TSF_INCLUDED
) {
1151 memcpy(&hdr
->mactime
, skb
->data
+ skb
->len
- 8, 8);
1152 hdr
->mactime
= le64_to_cpu(hdr
->mactime
);
1154 skb_trim(skb
, skb
->len
- 8);
1159 cw1200_debug_rxed(priv
);
1160 if (arg
->flags
& WSM_RX_STATUS_AGGREGATE
)
1161 cw1200_debug_rxed_agg(priv
);
1163 if (ieee80211_is_action(frame
->frame_control
) &&
1164 (arg
->flags
& WSM_RX_STATUS_ADDRESS1
)) {
1165 if (cw1200_handle_action_rx(priv
, skb
))
1167 } else if (ieee80211_is_beacon(frame
->frame_control
) &&
1168 !arg
->status
&& priv
->vif
&&
1169 !memcmp(ieee80211_get_SA(frame
), priv
->vif
->bss_conf
.bssid
,
1172 u8
*ies
= ((struct ieee80211_mgmt
*)
1173 (skb
->data
))->u
.beacon
.variable
;
1174 size_t ies_len
= skb
->len
- (ies
- (u8
*)(skb
->data
));
1176 tim_ie
= cfg80211_find_ie(WLAN_EID_TIM
, ies
, ies_len
);
1178 struct ieee80211_tim_ie
*tim
=
1179 (struct ieee80211_tim_ie
*)&tim_ie
[2];
1181 if (priv
->join_dtim_period
!= tim
->dtim_period
) {
1182 priv
->join_dtim_period
= tim
->dtim_period
;
1183 queue_work(priv
->workqueue
,
1184 &priv
->set_beacon_wakeup_period_work
);
1188 /* Disable beacon filter once we're associated... */
1189 if (priv
->disable_beacon_filter
&&
1190 (priv
->vif
->bss_conf
.assoc
||
1191 priv
->vif
->bss_conf
.ibss_joined
)) {
1192 priv
->disable_beacon_filter
= false;
1193 queue_work(priv
->workqueue
,
1194 &priv
->update_filtering_work
);
1198 /* Stay awake after frame is received to give
1199 * userspace chance to react and acquire appropriate
1202 if (ieee80211_is_auth(frame
->frame_control
))
1203 grace_period
= 5 * HZ
;
1204 else if (ieee80211_is_deauth(frame
->frame_control
))
1205 grace_period
= 5 * HZ
;
1207 grace_period
= 1 * HZ
;
1208 cw1200_pm_stay_awake(&priv
->pm_state
, grace_period
);
1211 spin_lock_bh(&priv
->ps_state_lock
);
1212 /* Double-check status with lock held */
1213 if (entry
->status
== CW1200_LINK_SOFT
)
1214 skb_queue_tail(&entry
->rx_queue
, skb
);
1216 ieee80211_rx_irqsafe(priv
->hw
, skb
);
1217 spin_unlock_bh(&priv
->ps_state_lock
);
1219 ieee80211_rx_irqsafe(priv
->hw
, skb
);
1226 /* TODO: update failure counters */
1230 /* ******************************************************************** */
1233 int cw1200_alloc_key(struct cw1200_common
*priv
)
1237 idx
= ffs(~priv
->key_map
) - 1;
1238 if (idx
< 0 || idx
> WSM_KEY_MAX_INDEX
)
1241 priv
->key_map
|= BIT(idx
);
1242 priv
->keys
[idx
].index
= idx
;
1246 void cw1200_free_key(struct cw1200_common
*priv
, int idx
)
1248 BUG_ON(!(priv
->key_map
& BIT(idx
)));
1249 memset(&priv
->keys
[idx
], 0, sizeof(priv
->keys
[idx
]));
1250 priv
->key_map
&= ~BIT(idx
);
1253 void cw1200_free_keys(struct cw1200_common
*priv
)
1255 memset(&priv
->keys
, 0, sizeof(priv
->keys
));
1259 int cw1200_upload_keys(struct cw1200_common
*priv
)
1262 for (idx
= 0; idx
<= WSM_KEY_MAX_INDEX
; ++idx
)
1263 if (priv
->key_map
& BIT(idx
)) {
1264 ret
= wsm_add_key(priv
, &priv
->keys
[idx
]);
1271 /* Workaround for WFD test case 6.1.10 */
1272 void cw1200_link_id_reset(struct work_struct
*work
)
1274 struct cw1200_common
*priv
=
1275 container_of(work
, struct cw1200_common
, linkid_reset_work
);
1278 if (!priv
->action_linkid
) {
1279 /* In GO mode we can receive ACTION frames without a linkID */
1280 temp_linkid
= cw1200_alloc_link_id(priv
,
1281 &priv
->action_frame_sa
[0]);
1282 WARN_ON(!temp_linkid
);
1284 /* Make sure we execute the WQ */
1285 flush_workqueue(priv
->workqueue
);
1286 /* Release the link ID */
1287 spin_lock_bh(&priv
->ps_state_lock
);
1288 priv
->link_id_db
[temp_linkid
- 1].prev_status
=
1289 priv
->link_id_db
[temp_linkid
- 1].status
;
1290 priv
->link_id_db
[temp_linkid
- 1].status
=
1292 spin_unlock_bh(&priv
->ps_state_lock
);
1293 wsm_lock_tx_async(priv
);
1294 if (queue_work(priv
->workqueue
,
1295 &priv
->link_id_work
) <= 0)
1296 wsm_unlock_tx(priv
);
1299 spin_lock_bh(&priv
->ps_state_lock
);
1300 priv
->link_id_db
[priv
->action_linkid
- 1].prev_status
=
1301 priv
->link_id_db
[priv
->action_linkid
- 1].status
;
1302 priv
->link_id_db
[priv
->action_linkid
- 1].status
=
1303 CW1200_LINK_RESET_REMAP
;
1304 spin_unlock_bh(&priv
->ps_state_lock
);
1305 wsm_lock_tx_async(priv
);
1306 if (queue_work(priv
->workqueue
, &priv
->link_id_work
) <= 0)
1307 wsm_unlock_tx(priv
);
1308 flush_workqueue(priv
->workqueue
);
1312 int cw1200_find_link_id(struct cw1200_common
*priv
, const u8
*mac
)
1315 spin_lock_bh(&priv
->ps_state_lock
);
1316 for (i
= 0; i
< CW1200_MAX_STA_IN_AP_MODE
; ++i
) {
1317 if (!memcmp(mac
, priv
->link_id_db
[i
].mac
, ETH_ALEN
) &&
1318 priv
->link_id_db
[i
].status
) {
1319 priv
->link_id_db
[i
].timestamp
= jiffies
;
1324 spin_unlock_bh(&priv
->ps_state_lock
);
1328 int cw1200_alloc_link_id(struct cw1200_common
*priv
, const u8
*mac
)
1331 unsigned long max_inactivity
= 0;
1332 unsigned long now
= jiffies
;
1334 spin_lock_bh(&priv
->ps_state_lock
);
1335 for (i
= 0; i
< CW1200_MAX_STA_IN_AP_MODE
; ++i
) {
1336 if (!priv
->link_id_db
[i
].status
) {
1339 } else if (priv
->link_id_db
[i
].status
!= CW1200_LINK_HARD
&&
1340 !priv
->tx_queue_stats
.link_map_cache
[i
+ 1]) {
1341 unsigned long inactivity
=
1342 now
- priv
->link_id_db
[i
].timestamp
;
1343 if (inactivity
< max_inactivity
)
1345 max_inactivity
= inactivity
;
1350 struct cw1200_link_entry
*entry
= &priv
->link_id_db
[ret
- 1];
1351 pr_debug("[AP] STA added, link_id: %d\n", ret
);
1352 entry
->status
= CW1200_LINK_RESERVE
;
1353 memcpy(&entry
->mac
, mac
, ETH_ALEN
);
1354 memset(&entry
->buffered
, 0, CW1200_MAX_TID
);
1355 skb_queue_head_init(&entry
->rx_queue
);
1356 wsm_lock_tx_async(priv
);
1357 if (queue_work(priv
->workqueue
, &priv
->link_id_work
) <= 0)
1358 wsm_unlock_tx(priv
);
1360 wiphy_info(priv
->hw
->wiphy
,
1361 "[AP] Early: no more link IDs available.\n");
1364 spin_unlock_bh(&priv
->ps_state_lock
);
1368 void cw1200_link_id_work(struct work_struct
*work
)
1370 struct cw1200_common
*priv
=
1371 container_of(work
, struct cw1200_common
, link_id_work
);
1373 cw1200_link_id_gc_work(&priv
->link_id_gc_work
.work
);
1374 wsm_unlock_tx(priv
);
1377 void cw1200_link_id_gc_work(struct work_struct
*work
)
1379 struct cw1200_common
*priv
=
1380 container_of(work
, struct cw1200_common
, link_id_gc_work
.work
);
1381 struct wsm_reset reset
= {
1382 .reset_statistics
= false,
1384 struct wsm_map_link map_link
= {
1387 unsigned long now
= jiffies
;
1388 unsigned long next_gc
= -1;
1394 if (priv
->join_status
!= CW1200_JOIN_STATUS_AP
)
1398 spin_lock_bh(&priv
->ps_state_lock
);
1399 for (i
= 0; i
< CW1200_MAX_STA_IN_AP_MODE
; ++i
) {
1402 if (priv
->link_id_db
[i
].status
== CW1200_LINK_RESERVE
||
1403 (priv
->link_id_db
[i
].status
== CW1200_LINK_HARD
&&
1404 !(priv
->link_id_map
& mask
))) {
1405 if (priv
->link_id_map
& mask
) {
1406 priv
->sta_asleep_mask
&= ~mask
;
1407 priv
->pspoll_mask
&= ~mask
;
1410 priv
->link_id_map
|= mask
;
1411 if (priv
->link_id_db
[i
].status
!= CW1200_LINK_HARD
)
1412 priv
->link_id_db
[i
].status
= CW1200_LINK_SOFT
;
1413 memcpy(map_link
.mac_addr
, priv
->link_id_db
[i
].mac
,
1415 spin_unlock_bh(&priv
->ps_state_lock
);
1417 reset
.link_id
= i
+ 1;
1418 wsm_reset(priv
, &reset
);
1420 map_link
.link_id
= i
+ 1;
1421 wsm_map_link(priv
, &map_link
);
1422 next_gc
= min(next_gc
, CW1200_LINK_ID_GC_TIMEOUT
);
1423 spin_lock_bh(&priv
->ps_state_lock
);
1424 } else if (priv
->link_id_db
[i
].status
== CW1200_LINK_SOFT
) {
1425 ttl
= priv
->link_id_db
[i
].timestamp
- now
+
1426 CW1200_LINK_ID_GC_TIMEOUT
;
1429 priv
->link_id_db
[i
].status
= CW1200_LINK_OFF
;
1430 priv
->link_id_map
&= ~mask
;
1431 priv
->sta_asleep_mask
&= ~mask
;
1432 priv
->pspoll_mask
&= ~mask
;
1433 memset(map_link
.mac_addr
, 0, ETH_ALEN
);
1434 spin_unlock_bh(&priv
->ps_state_lock
);
1435 reset
.link_id
= i
+ 1;
1436 wsm_reset(priv
, &reset
);
1437 spin_lock_bh(&priv
->ps_state_lock
);
1439 next_gc
= min_t(unsigned long, next_gc
, ttl
);
1441 } else if (priv
->link_id_db
[i
].status
== CW1200_LINK_RESET
||
1442 priv
->link_id_db
[i
].status
==
1443 CW1200_LINK_RESET_REMAP
) {
1444 int status
= priv
->link_id_db
[i
].status
;
1445 priv
->link_id_db
[i
].status
=
1446 priv
->link_id_db
[i
].prev_status
;
1447 priv
->link_id_db
[i
].timestamp
= now
;
1448 reset
.link_id
= i
+ 1;
1449 spin_unlock_bh(&priv
->ps_state_lock
);
1450 wsm_reset(priv
, &reset
);
1451 if (status
== CW1200_LINK_RESET_REMAP
) {
1452 memcpy(map_link
.mac_addr
,
1453 priv
->link_id_db
[i
].mac
,
1455 map_link
.link_id
= i
+ 1;
1456 wsm_map_link(priv
, &map_link
);
1457 next_gc
= min(next_gc
,
1458 CW1200_LINK_ID_GC_TIMEOUT
);
1460 spin_lock_bh(&priv
->ps_state_lock
);
1463 skb_queue_purge(&priv
->link_id_db
[i
].rx_queue
);
1464 pr_debug("[AP] STA removed, link_id: %d\n",
1468 spin_unlock_bh(&priv
->ps_state_lock
);
1470 queue_delayed_work(priv
->workqueue
,
1471 &priv
->link_id_gc_work
, next_gc
);
1472 wsm_unlock_tx(priv
);