1 /******************************************************************************
3 * This file is provided under a dual BSD/GPLv2 license. When using or
4 * redistributing this file, you may do so under either license.
8 * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved.
9 * Copyright(c) 2013 - 2014 Intel Mobile Communications GmbH
11 * This program is free software; you can redistribute it and/or modify
12 * it under the terms of version 2 of the GNU General Public License as
13 * published by the Free Software Foundation.
15 * This program is distributed in the hope that it will be useful, but
16 * WITHOUT ANY WARRANTY; without even the implied warranty of
17 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
18 * General Public License for more details.
20 * You should have received a copy of the GNU General Public License
21 * along with this program; if not, write to the Free Software
22 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110,
25 * The full GNU General Public License is included in this distribution
26 * in the file called COPYING.
28 * Contact Information:
29 * Intel Linux Wireless <ilw@linux.intel.com>
30 * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
34 * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved.
35 * Copyright(c) 2013 - 2014 Intel Mobile Communications GmbH
36 * All rights reserved.
38 * Redistribution and use in source and binary forms, with or without
39 * modification, are permitted provided that the following conditions
42 * * Redistributions of source code must retain the above copyright
43 * notice, this list of conditions and the following disclaimer.
44 * * Redistributions in binary form must reproduce the above copyright
45 * notice, this list of conditions and the following disclaimer in
46 * the documentation and/or other materials provided with the
48 * * Neither the name Intel Corporation nor the names of its
49 * contributors may be used to endorse or promote products derived
50 * from this software without specific prior written permission.
52 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
53 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
54 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
55 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
56 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
57 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
58 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
59 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
60 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
61 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
62 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
64 *****************************************************************************/
65 #include <linux/kernel.h>
66 #include <linux/slab.h>
67 #include <linux/skbuff.h>
68 #include <linux/netdevice.h>
69 #include <linux/etherdevice.h>
71 #include <linux/if_arp.h>
72 #include <linux/devcoredump.h>
73 #include <net/mac80211.h>
74 #include <net/ieee80211_radiotap.h>
77 #include "iwl-op-mode.h"
81 #include "time-event.h"
82 #include "iwl-eeprom-parse.h"
83 #include "fw-api-scan.h"
84 #include "iwl-phy-db.h"
86 #include "iwl-fw-error-dump.h"
89 #include "iwl-nvm-parse.h"
91 static const struct ieee80211_iface_limit iwl_mvm_limits
[] = {
94 .types
= BIT(NL80211_IFTYPE_STATION
),
98 .types
= BIT(NL80211_IFTYPE_AP
) |
99 BIT(NL80211_IFTYPE_P2P_CLIENT
) |
100 BIT(NL80211_IFTYPE_P2P_GO
),
104 .types
= BIT(NL80211_IFTYPE_P2P_DEVICE
),
108 static const struct ieee80211_iface_combination iwl_mvm_iface_combinations
[] = {
110 .num_different_channels
= 2,
112 .limits
= iwl_mvm_limits
,
113 .n_limits
= ARRAY_SIZE(iwl_mvm_limits
),
117 #ifdef CONFIG_PM_SLEEP
118 static const struct nl80211_wowlan_tcp_data_token_feature
119 iwl_mvm_wowlan_tcp_token_feature
= {
122 .bufsize
= IWL_WOWLAN_REMOTE_WAKE_MAX_TOKENS
,
125 static const struct wiphy_wowlan_tcp_support iwl_mvm_wowlan_tcp_support
= {
126 .tok
= &iwl_mvm_wowlan_tcp_token_feature
,
127 .data_payload_max
= IWL_WOWLAN_TCP_MAX_PACKET_LEN
-
128 sizeof(struct ethhdr
) -
129 sizeof(struct iphdr
) -
130 sizeof(struct tcphdr
),
131 .data_interval_max
= 65535, /* __le16 in API */
132 .wake_payload_max
= IWL_WOWLAN_REMOTE_WAKE_MAX_PACKET_LEN
-
133 sizeof(struct ethhdr
) -
134 sizeof(struct iphdr
) -
135 sizeof(struct tcphdr
),
140 #ifdef CONFIG_IWLWIFI_BCAST_FILTERING
142 * Use the reserved field to indicate magic values.
143 * these values will only be used internally by the driver,
144 * and won't make it to the fw (reserved will be 0).
145 * BC_FILTER_MAGIC_IP - configure the val of this attribute to
146 * be the vif's ip address. in case there is not a single
147 * ip address (0, or more than 1), this attribute will
149 * BC_FILTER_MAGIC_MAC - set the val of this attribute to
150 * the LSB bytes of the vif's mac address
153 BC_FILTER_MAGIC_NONE
= 0,
158 static const struct iwl_fw_bcast_filter iwl_mvm_default_bcast_filters
[] = {
162 .frame_type
= BCAST_FILTER_FRAME_TYPE_ALL
,
165 /* frame type - arp, hw type - ethernet */
167 BCAST_FILTER_OFFSET_PAYLOAD_START
,
168 .offset
= sizeof(rfc1042_header
),
169 .val
= cpu_to_be32(0x08060001),
170 .mask
= cpu_to_be32(0xffffffff),
175 BCAST_FILTER_OFFSET_PAYLOAD_START
,
176 .offset
= sizeof(rfc1042_header
) + 2 +
177 sizeof(struct arphdr
) +
178 ETH_ALEN
+ sizeof(__be32
) +
180 .mask
= cpu_to_be32(0xffffffff),
181 /* mark it as special field */
182 .reserved1
= cpu_to_le16(BC_FILTER_MAGIC_IP
),
187 /* dhcp offer bcast */
189 .frame_type
= BCAST_FILTER_FRAME_TYPE_IPV4
,
192 /* udp dest port - 68 (bootp client)*/
193 .offset_type
= BCAST_FILTER_OFFSET_IP_END
,
194 .offset
= offsetof(struct udphdr
, dest
),
195 .val
= cpu_to_be32(0x00440000),
196 .mask
= cpu_to_be32(0xffff0000),
199 /* dhcp - lsb bytes of client hw address */
200 .offset_type
= BCAST_FILTER_OFFSET_IP_END
,
202 .mask
= cpu_to_be32(0xffffffff),
203 /* mark it as special field */
204 .reserved1
= cpu_to_le16(BC_FILTER_MAGIC_MAC
),
208 /* last filter must be empty */
213 void iwl_mvm_ref(struct iwl_mvm
*mvm
, enum iwl_mvm_ref_type ref_type
)
215 if (!iwl_mvm_is_d0i3_supported(mvm
))
218 IWL_DEBUG_RPM(mvm
, "Take mvm reference - type %d\n", ref_type
);
219 spin_lock_bh(&mvm
->refs_lock
);
220 mvm
->refs
[ref_type
]++;
221 spin_unlock_bh(&mvm
->refs_lock
);
222 iwl_trans_ref(mvm
->trans
);
225 void iwl_mvm_unref(struct iwl_mvm
*mvm
, enum iwl_mvm_ref_type ref_type
)
227 if (!iwl_mvm_is_d0i3_supported(mvm
))
230 IWL_DEBUG_RPM(mvm
, "Leave mvm reference - type %d\n", ref_type
);
231 spin_lock_bh(&mvm
->refs_lock
);
232 WARN_ON(!mvm
->refs
[ref_type
]--);
233 spin_unlock_bh(&mvm
->refs_lock
);
234 iwl_trans_unref(mvm
->trans
);
237 static void iwl_mvm_unref_all_except(struct iwl_mvm
*mvm
,
238 enum iwl_mvm_ref_type except_ref
)
242 if (!iwl_mvm_is_d0i3_supported(mvm
))
245 spin_lock_bh(&mvm
->refs_lock
);
246 for (i
= 0; i
< IWL_MVM_REF_COUNT
; i
++) {
247 if (except_ref
== i
|| !mvm
->refs
[i
])
250 IWL_DEBUG_RPM(mvm
, "Cleanup: remove mvm ref type %d (%d)\n",
252 for (j
= 0; j
< mvm
->refs
[i
]; j
++)
253 iwl_trans_unref(mvm
->trans
);
256 spin_unlock_bh(&mvm
->refs_lock
);
259 bool iwl_mvm_ref_taken(struct iwl_mvm
*mvm
)
264 if (!iwl_mvm_is_d0i3_supported(mvm
))
267 spin_lock_bh(&mvm
->refs_lock
);
268 for (i
= 0; i
< IWL_MVM_REF_COUNT
; i
++) {
274 spin_unlock_bh(&mvm
->refs_lock
);
279 int iwl_mvm_ref_sync(struct iwl_mvm
*mvm
, enum iwl_mvm_ref_type ref_type
)
281 iwl_mvm_ref(mvm
, ref_type
);
283 if (!wait_event_timeout(mvm
->d0i3_exit_waitq
,
284 !test_bit(IWL_MVM_STATUS_IN_D0I3
, &mvm
->status
),
287 iwl_mvm_unref(mvm
, ref_type
);
294 static void iwl_mvm_reset_phy_ctxts(struct iwl_mvm
*mvm
)
298 memset(mvm
->phy_ctxts
, 0, sizeof(mvm
->phy_ctxts
));
299 for (i
= 0; i
< NUM_PHY_CTX
; i
++) {
300 mvm
->phy_ctxts
[i
].id
= i
;
301 mvm
->phy_ctxts
[i
].ref
= 0;
305 struct ieee80211_regdomain
*iwl_mvm_get_regdomain(struct wiphy
*wiphy
,
307 enum iwl_mcc_source src_id
,
310 struct ieee80211_regdomain
*regd
= NULL
;
311 struct ieee80211_hw
*hw
= wiphy_to_ieee80211_hw(wiphy
);
312 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
313 struct iwl_mcc_update_resp
*resp
;
315 IWL_DEBUG_LAR(mvm
, "Getting regdomain data for %s from FW\n", alpha2
);
317 lockdep_assert_held(&mvm
->mutex
);
319 resp
= iwl_mvm_update_mcc(mvm
, alpha2
, src_id
);
320 if (IS_ERR_OR_NULL(resp
)) {
321 IWL_DEBUG_LAR(mvm
, "Could not get update from FW %d\n",
327 *changed
= (resp
->status
== MCC_RESP_NEW_CHAN_PROFILE
);
329 regd
= iwl_parse_nvm_mcc_info(mvm
->trans
->dev
, mvm
->cfg
,
330 __le32_to_cpu(resp
->n_channels
),
332 __le16_to_cpu(resp
->mcc
));
333 /* Store the return source id */
334 src_id
= resp
->source_id
;
336 if (IS_ERR_OR_NULL(regd
)) {
337 IWL_DEBUG_LAR(mvm
, "Could not get parse update from FW %d\n",
342 IWL_DEBUG_LAR(mvm
, "setting alpha2 from FW to %s (0x%x, 0x%x) src=%d\n",
343 regd
->alpha2
, regd
->alpha2
[0], regd
->alpha2
[1], src_id
);
344 mvm
->lar_regdom_set
= true;
345 mvm
->mcc_src
= src_id
;
351 void iwl_mvm_update_changed_regdom(struct iwl_mvm
*mvm
)
354 struct ieee80211_regdomain
*regd
;
356 if (!iwl_mvm_is_lar_supported(mvm
))
359 regd
= iwl_mvm_get_current_regdomain(mvm
, &changed
);
360 if (!IS_ERR_OR_NULL(regd
)) {
361 /* only update the regulatory core if changed */
363 regulatory_set_wiphy_regd(mvm
->hw
->wiphy
, regd
);
369 struct ieee80211_regdomain
*iwl_mvm_get_current_regdomain(struct iwl_mvm
*mvm
,
372 return iwl_mvm_get_regdomain(mvm
->hw
->wiphy
, "ZZ",
373 iwl_mvm_is_wifi_mcc_supported(mvm
) ?
374 MCC_SOURCE_GET_CURRENT
:
375 MCC_SOURCE_OLD_FW
, changed
);
378 int iwl_mvm_init_fw_regd(struct iwl_mvm
*mvm
)
380 enum iwl_mcc_source used_src
;
381 struct ieee80211_regdomain
*regd
;
384 const struct ieee80211_regdomain
*r
=
385 rtnl_dereference(mvm
->hw
->wiphy
->regd
);
390 /* save the last source in case we overwrite it below */
391 used_src
= mvm
->mcc_src
;
392 if (iwl_mvm_is_wifi_mcc_supported(mvm
)) {
393 /* Notify the firmware we support wifi location updates */
394 regd
= iwl_mvm_get_current_regdomain(mvm
, NULL
);
395 if (!IS_ERR_OR_NULL(regd
))
399 /* Now set our last stored MCC and source */
400 regd
= iwl_mvm_get_regdomain(mvm
->hw
->wiphy
, r
->alpha2
, used_src
,
402 if (IS_ERR_OR_NULL(regd
))
405 /* update cfg80211 if the regdomain was changed */
407 ret
= regulatory_set_wiphy_regd_sync_rtnl(mvm
->hw
->wiphy
, regd
);
415 int iwl_mvm_mac_setup_register(struct iwl_mvm
*mvm
)
417 struct ieee80211_hw
*hw
= mvm
->hw
;
420 /* Tell mac80211 our characteristics */
421 hw
->flags
= IEEE80211_HW_SIGNAL_DBM
|
422 IEEE80211_HW_SPECTRUM_MGMT
|
423 IEEE80211_HW_REPORTS_TX_ACK_STATUS
|
424 IEEE80211_HW_QUEUE_CONTROL
|
425 IEEE80211_HW_WANT_MONITOR_VIF
|
426 IEEE80211_HW_SUPPORTS_PS
|
427 IEEE80211_HW_SUPPORTS_DYNAMIC_PS
|
428 IEEE80211_HW_AMPDU_AGGREGATION
|
429 IEEE80211_HW_TIMING_BEACON_ONLY
|
430 IEEE80211_HW_CONNECTION_MONITOR
|
431 IEEE80211_HW_CHANCTX_STA_CSA
|
432 IEEE80211_HW_SUPPORTS_CLONED_SKBS
;
434 hw
->queues
= mvm
->first_agg_queue
;
435 hw
->offchannel_tx_hw_queue
= IWL_MVM_OFFCHANNEL_QUEUE
;
436 hw
->radiotap_mcs_details
|= IEEE80211_RADIOTAP_MCS_HAVE_FEC
|
437 IEEE80211_RADIOTAP_MCS_HAVE_STBC
;
438 hw
->radiotap_vht_details
|= IEEE80211_RADIOTAP_VHT_KNOWN_STBC
|
439 IEEE80211_RADIOTAP_VHT_KNOWN_BEAMFORMED
;
440 hw
->rate_control_algorithm
= "iwl-mvm-rs";
441 hw
->uapsd_queues
= IWL_MVM_UAPSD_QUEUES
;
442 hw
->uapsd_max_sp_len
= IWL_UAPSD_MAX_SP
;
445 * Enable 11w if advertised by firmware and software crypto
446 * is not enabled (as the firmware will interpret some mgmt
447 * packets, so enabling it with software crypto isn't safe)
449 if (mvm
->fw
->ucode_capa
.flags
& IWL_UCODE_TLV_FLAGS_MFP
&&
450 !iwlwifi_mod_params
.sw_crypto
)
451 hw
->flags
|= IEEE80211_HW_MFP_CAPABLE
;
453 hw
->flags
|= IEEE80211_HW_SINGLE_SCAN_ON_ALL_BANDS
;
454 hw
->wiphy
->features
|=
455 NL80211_FEATURE_SCHED_SCAN_RANDOM_MAC_ADDR
|
456 NL80211_FEATURE_SCAN_RANDOM_MAC_ADDR
;
458 hw
->sta_data_size
= sizeof(struct iwl_mvm_sta
);
459 hw
->vif_data_size
= sizeof(struct iwl_mvm_vif
);
460 hw
->chanctx_data_size
= sizeof(u16
);
462 hw
->wiphy
->interface_modes
= BIT(NL80211_IFTYPE_STATION
) |
463 BIT(NL80211_IFTYPE_P2P_CLIENT
) |
464 BIT(NL80211_IFTYPE_AP
) |
465 BIT(NL80211_IFTYPE_P2P_GO
) |
466 BIT(NL80211_IFTYPE_P2P_DEVICE
) |
467 BIT(NL80211_IFTYPE_ADHOC
);
469 hw
->wiphy
->flags
|= WIPHY_FLAG_IBSS_RSN
;
470 hw
->wiphy
->regulatory_flags
|= REGULATORY_ENABLE_RELAX_NO_IR
;
471 if (iwl_mvm_is_lar_supported(mvm
))
472 hw
->wiphy
->regulatory_flags
|= REGULATORY_WIPHY_SELF_MANAGED
;
474 hw
->wiphy
->regulatory_flags
|= REGULATORY_CUSTOM_REG
|
475 REGULATORY_DISABLE_BEACON_HINTS
;
477 if (mvm
->fw
->ucode_capa
.flags
& IWL_UCODE_TLV_FLAGS_GO_UAPSD
)
478 hw
->wiphy
->flags
|= WIPHY_FLAG_AP_UAPSD
;
480 hw
->wiphy
->flags
|= WIPHY_FLAG_HAS_CHANNEL_SWITCH
;
482 hw
->wiphy
->iface_combinations
= iwl_mvm_iface_combinations
;
483 hw
->wiphy
->n_iface_combinations
=
484 ARRAY_SIZE(iwl_mvm_iface_combinations
);
486 hw
->wiphy
->max_remain_on_channel_duration
= 10000;
487 hw
->max_listen_interval
= IWL_CONN_MAX_LISTEN_INTERVAL
;
488 /* we can compensate an offset of up to 3 channels = 15 MHz */
489 hw
->wiphy
->max_adj_channel_rssi_comp
= 3 * 5;
491 /* Extract MAC address */
492 memcpy(mvm
->addresses
[0].addr
, mvm
->nvm_data
->hw_addr
, ETH_ALEN
);
493 hw
->wiphy
->addresses
= mvm
->addresses
;
494 hw
->wiphy
->n_addresses
= 1;
496 /* Extract additional MAC addresses if available */
497 num_mac
= (mvm
->nvm_data
->n_hw_addrs
> 1) ?
498 min(IWL_MVM_MAX_ADDRESSES
, mvm
->nvm_data
->n_hw_addrs
) : 1;
500 for (i
= 1; i
< num_mac
; i
++) {
501 memcpy(mvm
->addresses
[i
].addr
, mvm
->addresses
[i
-1].addr
,
503 mvm
->addresses
[i
].addr
[5]++;
504 hw
->wiphy
->n_addresses
++;
507 iwl_mvm_reset_phy_ctxts(mvm
);
509 hw
->wiphy
->max_scan_ie_len
= iwl_mvm_max_scan_ie_len(mvm
, false);
511 hw
->wiphy
->max_scan_ssids
= PROBE_OPTION_MAX
;
513 if (mvm
->nvm_data
->bands
[IEEE80211_BAND_2GHZ
].n_channels
)
514 hw
->wiphy
->bands
[IEEE80211_BAND_2GHZ
] =
515 &mvm
->nvm_data
->bands
[IEEE80211_BAND_2GHZ
];
516 if (mvm
->nvm_data
->bands
[IEEE80211_BAND_5GHZ
].n_channels
) {
517 hw
->wiphy
->bands
[IEEE80211_BAND_5GHZ
] =
518 &mvm
->nvm_data
->bands
[IEEE80211_BAND_5GHZ
];
520 if ((mvm
->fw
->ucode_capa
.capa
[0] &
521 IWL_UCODE_TLV_CAPA_BEAMFORMER
) &&
522 (mvm
->fw
->ucode_capa
.api
[0] &
523 IWL_UCODE_TLV_API_LQ_SS_PARAMS
))
524 hw
->wiphy
->bands
[IEEE80211_BAND_5GHZ
]->vht_cap
.cap
|=
525 IEEE80211_VHT_CAP_SU_BEAMFORMER_CAPABLE
;
528 hw
->wiphy
->hw_version
= mvm
->trans
->hw_id
;
530 if (iwlmvm_mod_params
.power_scheme
!= IWL_POWER_SCHEME_CAM
)
531 hw
->wiphy
->flags
|= WIPHY_FLAG_PS_ON_BY_DEFAULT
;
533 hw
->wiphy
->flags
&= ~WIPHY_FLAG_PS_ON_BY_DEFAULT
;
535 if (IWL_UCODE_API(mvm
->fw
->ucode_ver
) >= 10) {
536 hw
->wiphy
->flags
|= WIPHY_FLAG_SUPPORTS_SCHED_SCAN
;
537 hw
->wiphy
->max_sched_scan_ssids
= PROBE_OPTION_MAX
;
538 hw
->wiphy
->max_match_sets
= IWL_SCAN_MAX_PROFILES
;
539 /* we create the 802.11 header and zero length SSID IE. */
540 hw
->wiphy
->max_sched_scan_ie_len
=
541 SCAN_OFFLOAD_PROBE_REQ_SIZE
- 24 - 2;
544 hw
->wiphy
->features
|= NL80211_FEATURE_P2P_GO_CTWIN
|
545 NL80211_FEATURE_LOW_PRIORITY_SCAN
|
546 NL80211_FEATURE_P2P_GO_OPPPS
|
547 NL80211_FEATURE_DYNAMIC_SMPS
|
548 NL80211_FEATURE_STATIC_SMPS
|
549 NL80211_FEATURE_SUPPORTS_WMM_ADMISSION
;
551 if (mvm
->fw
->ucode_capa
.capa
[0] &
552 IWL_UCODE_TLV_CAPA_TXPOWER_INSERTION_SUPPORT
)
553 hw
->wiphy
->features
|= NL80211_FEATURE_TX_POWER_INSERTION
;
554 if (mvm
->fw
->ucode_capa
.capa
[0] &
555 IWL_UCODE_TLV_CAPA_QUIET_PERIOD_SUPPORT
)
556 hw
->wiphy
->features
|= NL80211_FEATURE_QUIET
;
558 if (mvm
->fw
->ucode_capa
.capa
[0] &
559 IWL_UCODE_TLV_CAPA_DS_PARAM_SET_IE_SUPPORT
)
560 hw
->wiphy
->features
|=
561 NL80211_FEATURE_DS_PARAM_SET_IE_IN_PROBES
;
563 if (mvm
->fw
->ucode_capa
.capa
[0] &
564 IWL_UCODE_TLV_CAPA_WFA_TPC_REP_IE_SUPPORT
)
565 hw
->wiphy
->features
|= NL80211_FEATURE_WFA_TPC_IE_IN_PROBES
;
567 mvm
->rts_threshold
= IEEE80211_MAX_RTS_THRESHOLD
;
569 /* currently FW API supports only one optional cipher scheme */
570 if (mvm
->fw
->cs
[0].cipher
) {
571 mvm
->hw
->n_cipher_schemes
= 1;
572 mvm
->hw
->cipher_schemes
= &mvm
->fw
->cs
[0];
575 #ifdef CONFIG_PM_SLEEP
576 if (iwl_mvm_is_d0i3_supported(mvm
) &&
577 device_can_wakeup(mvm
->trans
->dev
)) {
578 mvm
->wowlan
.flags
= WIPHY_WOWLAN_ANY
;
579 hw
->wiphy
->wowlan
= &mvm
->wowlan
;
582 if (mvm
->fw
->img
[IWL_UCODE_WOWLAN
].sec
[0].len
&&
583 mvm
->trans
->ops
->d3_suspend
&&
584 mvm
->trans
->ops
->d3_resume
&&
585 device_can_wakeup(mvm
->trans
->dev
)) {
586 mvm
->wowlan
.flags
|= WIPHY_WOWLAN_MAGIC_PKT
|
587 WIPHY_WOWLAN_DISCONNECT
|
588 WIPHY_WOWLAN_EAP_IDENTITY_REQ
|
589 WIPHY_WOWLAN_RFKILL_RELEASE
|
590 WIPHY_WOWLAN_NET_DETECT
;
591 if (!iwlwifi_mod_params
.sw_crypto
)
592 mvm
->wowlan
.flags
|= WIPHY_WOWLAN_SUPPORTS_GTK_REKEY
|
593 WIPHY_WOWLAN_GTK_REKEY_FAILURE
|
594 WIPHY_WOWLAN_4WAY_HANDSHAKE
;
596 mvm
->wowlan
.n_patterns
= IWL_WOWLAN_MAX_PATTERNS
;
597 mvm
->wowlan
.pattern_min_len
= IWL_WOWLAN_MIN_PATTERN_LEN
;
598 mvm
->wowlan
.pattern_max_len
= IWL_WOWLAN_MAX_PATTERN_LEN
;
599 mvm
->wowlan
.max_nd_match_sets
= IWL_SCAN_MAX_PROFILES
;
600 mvm
->wowlan
.tcp
= &iwl_mvm_wowlan_tcp_support
;
601 hw
->wiphy
->wowlan
= &mvm
->wowlan
;
605 #ifdef CONFIG_IWLWIFI_BCAST_FILTERING
606 /* assign default bcast filtering configuration */
607 mvm
->bcast_filters
= iwl_mvm_default_bcast_filters
;
610 ret
= iwl_mvm_leds_init(mvm
);
614 if (mvm
->fw
->ucode_capa
.capa
[0] & IWL_UCODE_TLV_CAPA_TDLS_SUPPORT
) {
615 IWL_DEBUG_TDLS(mvm
, "TDLS supported\n");
616 hw
->wiphy
->flags
|= WIPHY_FLAG_SUPPORTS_TDLS
;
619 if (mvm
->fw
->ucode_capa
.capa
[0] &
620 IWL_UCODE_TLV_CAPA_TDLS_CHANNEL_SWITCH
) {
621 IWL_DEBUG_TDLS(mvm
, "TDLS channel switch supported\n");
622 hw
->wiphy
->features
|= NL80211_FEATURE_TDLS_CHANNEL_SWITCH
;
625 ret
= ieee80211_register_hw(mvm
->hw
);
627 iwl_mvm_leds_exit(mvm
);
632 static bool iwl_mvm_defer_tx(struct iwl_mvm
*mvm
,
633 struct ieee80211_sta
*sta
,
636 struct iwl_mvm_sta
*mvmsta
;
640 * double check the IN_D0I3 flag both before and after
641 * taking the spinlock, in order to prevent taking
642 * the spinlock when not needed.
644 if (likely(!test_bit(IWL_MVM_STATUS_IN_D0I3
, &mvm
->status
)))
647 spin_lock(&mvm
->d0i3_tx_lock
);
649 * testing the flag again ensures the skb dequeue
650 * loop (on d0i3 exit) hasn't run yet.
652 if (!test_bit(IWL_MVM_STATUS_IN_D0I3
, &mvm
->status
))
655 mvmsta
= iwl_mvm_sta_from_mac80211(sta
);
656 if (mvmsta
->sta_id
== IWL_MVM_STATION_COUNT
||
657 mvmsta
->sta_id
!= mvm
->d0i3_ap_sta_id
)
660 __skb_queue_tail(&mvm
->d0i3_tx
, skb
);
661 ieee80211_stop_queues(mvm
->hw
);
664 iwl_mvm_ref(mvm
, IWL_MVM_REF_TX
);
665 iwl_mvm_unref(mvm
, IWL_MVM_REF_TX
);
669 spin_unlock(&mvm
->d0i3_tx_lock
);
673 static void iwl_mvm_mac_tx(struct ieee80211_hw
*hw
,
674 struct ieee80211_tx_control
*control
,
677 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
678 struct ieee80211_sta
*sta
= control
->sta
;
679 struct ieee80211_tx_info
*info
= IEEE80211_SKB_CB(skb
);
680 struct ieee80211_hdr
*hdr
= (void *)skb
->data
;
682 if (iwl_mvm_is_radio_killed(mvm
)) {
683 IWL_DEBUG_DROP(mvm
, "Dropping - RF/CT KILL\n");
687 if (IEEE80211_SKB_CB(skb
)->hw_queue
== IWL_MVM_OFFCHANNEL_QUEUE
&&
688 !test_bit(IWL_MVM_STATUS_ROC_RUNNING
, &mvm
->status
) &&
689 !test_bit(IWL_MVM_STATUS_ROC_AUX_RUNNING
, &mvm
->status
))
692 /* treat non-bufferable MMPDUs as broadcast if sta is sleeping */
693 if (unlikely(info
->flags
& IEEE80211_TX_CTL_NO_PS_BUFFER
&&
694 ieee80211_is_mgmt(hdr
->frame_control
) &&
695 !ieee80211_is_deauth(hdr
->frame_control
) &&
696 !ieee80211_is_disassoc(hdr
->frame_control
) &&
697 !ieee80211_is_action(hdr
->frame_control
)))
701 if (iwl_mvm_defer_tx(mvm
, sta
, skb
))
703 if (iwl_mvm_tx_skb(mvm
, skb
, sta
))
708 if (iwl_mvm_tx_skb_non_sta(mvm
, skb
))
712 ieee80211_free_txskb(hw
, skb
);
715 static inline bool iwl_enable_rx_ampdu(const struct iwl_cfg
*cfg
)
717 if (iwlwifi_mod_params
.disable_11n
& IWL_DISABLE_HT_RXAGG
)
722 static inline bool iwl_enable_tx_ampdu(const struct iwl_cfg
*cfg
)
724 if (iwlwifi_mod_params
.disable_11n
& IWL_DISABLE_HT_TXAGG
)
726 if (iwlwifi_mod_params
.disable_11n
& IWL_ENABLE_HT_TXAGG
)
729 /* enabled by default */
733 static int iwl_mvm_mac_ampdu_action(struct ieee80211_hw
*hw
,
734 struct ieee80211_vif
*vif
,
735 enum ieee80211_ampdu_mlme_action action
,
736 struct ieee80211_sta
*sta
, u16 tid
,
737 u16
*ssn
, u8 buf_size
)
739 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
741 bool tx_agg_ref
= false;
743 IWL_DEBUG_HT(mvm
, "A-MPDU action on addr %pM tid %d: action %d\n",
744 sta
->addr
, tid
, action
);
746 if (!(mvm
->nvm_data
->sku_cap_11n_enable
))
749 /* return from D0i3 before starting a new Tx aggregation */
751 case IEEE80211_AMPDU_TX_START
:
752 case IEEE80211_AMPDU_TX_STOP_CONT
:
753 case IEEE80211_AMPDU_TX_STOP_FLUSH
:
754 case IEEE80211_AMPDU_TX_STOP_FLUSH_CONT
:
755 case IEEE80211_AMPDU_TX_OPERATIONAL
:
757 * for tx start, wait synchronously until D0i3 exit to
758 * get the correct sequence number for the tid.
759 * additionally, some other ampdu actions use direct
760 * target access, which is not handled automatically
761 * by the trans layer (unlike commands), so wait for
762 * d0i3 exit in these cases as well.
764 ret
= iwl_mvm_ref_sync(mvm
, IWL_MVM_REF_TX_AGG
);
774 mutex_lock(&mvm
->mutex
);
777 case IEEE80211_AMPDU_RX_START
:
778 if (!iwl_enable_rx_ampdu(mvm
->cfg
)) {
782 ret
= iwl_mvm_sta_rx_agg(mvm
, sta
, tid
, *ssn
, true);
784 case IEEE80211_AMPDU_RX_STOP
:
785 ret
= iwl_mvm_sta_rx_agg(mvm
, sta
, tid
, 0, false);
787 case IEEE80211_AMPDU_TX_START
:
788 if (!iwl_enable_tx_ampdu(mvm
->cfg
)) {
792 ret
= iwl_mvm_sta_tx_agg_start(mvm
, vif
, sta
, tid
, ssn
);
794 case IEEE80211_AMPDU_TX_STOP_CONT
:
795 ret
= iwl_mvm_sta_tx_agg_stop(mvm
, vif
, sta
, tid
);
797 case IEEE80211_AMPDU_TX_STOP_FLUSH
:
798 case IEEE80211_AMPDU_TX_STOP_FLUSH_CONT
:
799 ret
= iwl_mvm_sta_tx_agg_flush(mvm
, vif
, sta
, tid
);
801 case IEEE80211_AMPDU_TX_OPERATIONAL
:
802 ret
= iwl_mvm_sta_tx_agg_oper(mvm
, vif
, sta
, tid
, buf_size
);
809 mutex_unlock(&mvm
->mutex
);
812 * If the tid is marked as started, we won't use it for offloaded
813 * traffic on the next D0i3 entry. It's safe to unref.
816 iwl_mvm_unref(mvm
, IWL_MVM_REF_TX_AGG
);
821 static void iwl_mvm_cleanup_iterator(void *data
, u8
*mac
,
822 struct ieee80211_vif
*vif
)
824 struct iwl_mvm
*mvm
= data
;
825 struct iwl_mvm_vif
*mvmvif
= iwl_mvm_vif_from_mac80211(vif
);
827 mvmvif
->uploaded
= false;
828 mvmvif
->ap_sta_id
= IWL_MVM_STATION_COUNT
;
830 spin_lock_bh(&mvm
->time_event_lock
);
831 iwl_mvm_te_clear_data(mvm
, &mvmvif
->time_event_data
);
832 spin_unlock_bh(&mvm
->time_event_lock
);
834 mvmvif
->phy_ctxt
= NULL
;
835 memset(&mvmvif
->bf_data
, 0, sizeof(mvmvif
->bf_data
));
838 static ssize_t
iwl_mvm_read_coredump(char *buffer
, loff_t offset
, size_t count
,
839 const void *data
, size_t datalen
)
841 const struct iwl_mvm_dump_ptrs
*dump_ptrs
= data
;
843 ssize_t bytes_read_trans
;
845 if (offset
< dump_ptrs
->op_mode_len
) {
846 bytes_read
= min_t(ssize_t
, count
,
847 dump_ptrs
->op_mode_len
- offset
);
848 memcpy(buffer
, (u8
*)dump_ptrs
->op_mode_ptr
+ offset
,
850 offset
+= bytes_read
;
859 if (!dump_ptrs
->trans_ptr
)
862 offset
-= dump_ptrs
->op_mode_len
;
863 bytes_read_trans
= min_t(ssize_t
, count
,
864 dump_ptrs
->trans_ptr
->len
- offset
);
865 memcpy(buffer
+ bytes_read
,
866 (u8
*)dump_ptrs
->trans_ptr
->data
+ offset
,
869 return bytes_read
+ bytes_read_trans
;
872 static void iwl_mvm_free_coredump(const void *data
)
874 const struct iwl_mvm_dump_ptrs
*fw_error_dump
= data
;
876 vfree(fw_error_dump
->op_mode_ptr
);
877 vfree(fw_error_dump
->trans_ptr
);
878 kfree(fw_error_dump
);
881 static void iwl_mvm_dump_fifos(struct iwl_mvm
*mvm
,
882 struct iwl_fw_error_dump_data
**dump_data
)
884 struct iwl_fw_error_dump_fifo
*fifo_hdr
;
890 if (!iwl_trans_grab_nic_access(mvm
->trans
, false, &flags
))
893 /* Pull RXF data from all RXFs */
894 for (i
= 0; i
< ARRAY_SIZE(mvm
->shared_mem_cfg
.rxfifo_size
); i
++) {
896 * Keep aside the additional offset that might be needed for
899 u32 offset_diff
= RXF_DIFF_FROM_PREV
* i
;
901 fifo_hdr
= (void *)(*dump_data
)->data
;
902 fifo_data
= (void *)fifo_hdr
->data
;
903 fifo_len
= mvm
->shared_mem_cfg
.rxfifo_size
[i
];
905 /* No need to try to read the data if the length is 0 */
909 /* Add a TLV for the RXF */
910 (*dump_data
)->type
= cpu_to_le32(IWL_FW_ERROR_DUMP_RXF
);
911 (*dump_data
)->len
= cpu_to_le32(fifo_len
+ sizeof(*fifo_hdr
));
913 fifo_hdr
->fifo_num
= cpu_to_le32(i
);
914 fifo_hdr
->available_bytes
=
915 cpu_to_le32(iwl_trans_read_prph(mvm
->trans
,
919 cpu_to_le32(iwl_trans_read_prph(mvm
->trans
,
923 cpu_to_le32(iwl_trans_read_prph(mvm
->trans
,
926 fifo_hdr
->fence_ptr
=
927 cpu_to_le32(iwl_trans_read_prph(mvm
->trans
,
930 fifo_hdr
->fence_mode
=
931 cpu_to_le32(iwl_trans_read_prph(mvm
->trans
,
936 iwl_trans_write_prph(mvm
->trans
,
937 RXF_SET_FENCE_MODE
+ offset_diff
, 0x1);
938 /* Set fence pointer to the same place like WR pointer */
939 iwl_trans_write_prph(mvm
->trans
,
940 RXF_LD_WR2FENCE
+ offset_diff
, 0x1);
941 /* Set fence offset */
942 iwl_trans_write_prph(mvm
->trans
,
943 RXF_LD_FENCE_OFFSET_ADDR
+ offset_diff
,
947 fifo_len
/= sizeof(u32
); /* Size in DWORDS */
948 for (j
= 0; j
< fifo_len
; j
++)
949 fifo_data
[j
] = iwl_trans_read_prph(mvm
->trans
,
950 RXF_FIFO_RD_FENCE_INC
+
952 *dump_data
= iwl_fw_error_next_data(*dump_data
);
955 /* Pull TXF data from all TXFs */
956 for (i
= 0; i
< ARRAY_SIZE(mvm
->shared_mem_cfg
.txfifo_size
); i
++) {
957 /* Mark the number of TXF we're pulling now */
958 iwl_trans_write_prph(mvm
->trans
, TXF_LARC_NUM
, i
);
960 fifo_hdr
= (void *)(*dump_data
)->data
;
961 fifo_data
= (void *)fifo_hdr
->data
;
962 fifo_len
= mvm
->shared_mem_cfg
.txfifo_size
[i
];
964 /* No need to try to read the data if the length is 0 */
968 /* Add a TLV for the FIFO */
969 (*dump_data
)->type
= cpu_to_le32(IWL_FW_ERROR_DUMP_TXF
);
970 (*dump_data
)->len
= cpu_to_le32(fifo_len
+ sizeof(*fifo_hdr
));
972 fifo_hdr
->fifo_num
= cpu_to_le32(i
);
973 fifo_hdr
->available_bytes
=
974 cpu_to_le32(iwl_trans_read_prph(mvm
->trans
,
977 cpu_to_le32(iwl_trans_read_prph(mvm
->trans
,
980 cpu_to_le32(iwl_trans_read_prph(mvm
->trans
,
982 fifo_hdr
->fence_ptr
=
983 cpu_to_le32(iwl_trans_read_prph(mvm
->trans
,
985 fifo_hdr
->fence_mode
=
986 cpu_to_le32(iwl_trans_read_prph(mvm
->trans
,
989 /* Set the TXF_READ_MODIFY_ADDR to TXF_WR_PTR */
990 iwl_trans_write_prph(mvm
->trans
, TXF_READ_MODIFY_ADDR
,
993 /* Dummy-read to advance the read pointer to the head */
994 iwl_trans_read_prph(mvm
->trans
, TXF_READ_MODIFY_DATA
);
997 fifo_len
/= sizeof(u32
); /* Size in DWORDS */
998 for (j
= 0; j
< fifo_len
; j
++)
999 fifo_data
[j
] = iwl_trans_read_prph(mvm
->trans
,
1000 TXF_READ_MODIFY_DATA
);
1001 *dump_data
= iwl_fw_error_next_data(*dump_data
);
1004 iwl_trans_release_nic_access(mvm
->trans
, &flags
);
1007 void iwl_mvm_free_fw_dump_desc(struct iwl_mvm
*mvm
)
1009 if (mvm
->fw_dump_desc
== &iwl_mvm_dump_desc_assert
||
1013 kfree(mvm
->fw_dump_desc
);
1014 mvm
->fw_dump_desc
= NULL
;
1017 #define IWL8260_ICCM_OFFSET 0x44000 /* Only for B-step */
1018 #define IWL8260_ICCM_LEN 0xC000 /* Only for B-step */
1020 void iwl_mvm_fw_error_dump(struct iwl_mvm
*mvm
)
1022 struct iwl_fw_error_dump_file
*dump_file
;
1023 struct iwl_fw_error_dump_data
*dump_data
;
1024 struct iwl_fw_error_dump_info
*dump_info
;
1025 struct iwl_fw_error_dump_mem
*dump_mem
;
1026 struct iwl_fw_error_dump_trigger_desc
*dump_trig
;
1027 struct iwl_mvm_dump_ptrs
*fw_error_dump
;
1028 u32 sram_len
, sram_ofs
;
1029 u32 file_len
, fifo_data_len
= 0;
1030 u32 smem_len
= mvm
->cfg
->smem_len
;
1031 u32 sram2_len
= mvm
->cfg
->dccm2_len
;
1033 lockdep_assert_held(&mvm
->mutex
);
1035 fw_error_dump
= kzalloc(sizeof(*fw_error_dump
), GFP_KERNEL
);
1039 /* SRAM - include stack CCM if driver knows the values for it */
1040 if (!mvm
->cfg
->dccm_offset
|| !mvm
->cfg
->dccm_len
) {
1041 const struct fw_img
*img
;
1043 img
= &mvm
->fw
->img
[mvm
->cur_ucode
];
1044 sram_ofs
= img
->sec
[IWL_UCODE_SECTION_DATA
].offset
;
1045 sram_len
= img
->sec
[IWL_UCODE_SECTION_DATA
].len
;
1047 sram_ofs
= mvm
->cfg
->dccm_offset
;
1048 sram_len
= mvm
->cfg
->dccm_len
;
1051 /* reading RXF/TXF sizes */
1052 if (test_bit(STATUS_FW_ERROR
, &mvm
->trans
->status
)) {
1053 struct iwl_mvm_shared_mem_cfg
*mem_cfg
= &mvm
->shared_mem_cfg
;
1058 /* Count RXF size */
1059 for (i
= 0; i
< ARRAY_SIZE(mem_cfg
->rxfifo_size
); i
++) {
1060 if (!mem_cfg
->rxfifo_size
[i
])
1063 /* Add header info */
1064 fifo_data_len
+= mem_cfg
->rxfifo_size
[i
] +
1065 sizeof(*dump_data
) +
1066 sizeof(struct iwl_fw_error_dump_fifo
);
1069 for (i
= 0; i
< ARRAY_SIZE(mem_cfg
->txfifo_size
); i
++) {
1070 if (!mem_cfg
->txfifo_size
[i
])
1073 /* Add header info */
1074 fifo_data_len
+= mem_cfg
->txfifo_size
[i
] +
1075 sizeof(*dump_data
) +
1076 sizeof(struct iwl_fw_error_dump_fifo
);
1080 file_len
= sizeof(*dump_file
) +
1081 sizeof(*dump_data
) * 2 +
1082 sram_len
+ sizeof(*dump_mem
) +
1087 * In 8000 HW family B-step include the ICCM (which resides separately)
1089 if (mvm
->cfg
->device_family
== IWL_DEVICE_FAMILY_8000
&&
1090 CSR_HW_REV_STEP(mvm
->trans
->hw_rev
) == SILICON_B_STEP
)
1091 file_len
+= sizeof(*dump_data
) + sizeof(*dump_mem
) +
1094 if (mvm
->fw_dump_desc
)
1095 file_len
+= sizeof(*dump_data
) + sizeof(*dump_trig
) +
1096 mvm
->fw_dump_desc
->len
;
1098 /* Make room for the SMEM, if it exists */
1100 file_len
+= sizeof(*dump_data
) + sizeof(*dump_mem
) + smem_len
;
1102 /* Make room for the secondary SRAM, if it exists */
1104 file_len
+= sizeof(*dump_data
) + sizeof(*dump_mem
) + sram2_len
;
1106 dump_file
= vzalloc(file_len
);
1108 kfree(fw_error_dump
);
1109 iwl_mvm_free_fw_dump_desc(mvm
);
1113 fw_error_dump
->op_mode_ptr
= dump_file
;
1115 dump_file
->barker
= cpu_to_le32(IWL_FW_ERROR_DUMP_BARKER
);
1116 dump_data
= (void *)dump_file
->data
;
1118 dump_data
->type
= cpu_to_le32(IWL_FW_ERROR_DUMP_DEV_FW_INFO
);
1119 dump_data
->len
= cpu_to_le32(sizeof(*dump_info
));
1120 dump_info
= (void *) dump_data
->data
;
1121 dump_info
->device_family
=
1122 mvm
->cfg
->device_family
== IWL_DEVICE_FAMILY_7000
?
1123 cpu_to_le32(IWL_FW_ERROR_DUMP_FAMILY_7
) :
1124 cpu_to_le32(IWL_FW_ERROR_DUMP_FAMILY_8
);
1125 dump_info
->hw_step
= cpu_to_le32(CSR_HW_REV_STEP(mvm
->trans
->hw_rev
));
1126 memcpy(dump_info
->fw_human_readable
, mvm
->fw
->human_readable
,
1127 sizeof(dump_info
->fw_human_readable
));
1128 strncpy(dump_info
->dev_human_readable
, mvm
->cfg
->name
,
1129 sizeof(dump_info
->dev_human_readable
));
1130 strncpy(dump_info
->bus_human_readable
, mvm
->dev
->bus
->name
,
1131 sizeof(dump_info
->bus_human_readable
));
1133 dump_data
= iwl_fw_error_next_data(dump_data
);
1134 /* We only dump the FIFOs if the FW is in error state */
1135 if (test_bit(STATUS_FW_ERROR
, &mvm
->trans
->status
))
1136 iwl_mvm_dump_fifos(mvm
, &dump_data
);
1138 if (mvm
->fw_dump_desc
) {
1139 dump_data
->type
= cpu_to_le32(IWL_FW_ERROR_DUMP_ERROR_INFO
);
1140 dump_data
->len
= cpu_to_le32(sizeof(*dump_trig
) +
1141 mvm
->fw_dump_desc
->len
);
1142 dump_trig
= (void *)dump_data
->data
;
1143 memcpy(dump_trig
, &mvm
->fw_dump_desc
->trig_desc
,
1144 sizeof(*dump_trig
) + mvm
->fw_dump_desc
->len
);
1146 /* now we can free this copy */
1147 iwl_mvm_free_fw_dump_desc(mvm
);
1148 dump_data
= iwl_fw_error_next_data(dump_data
);
1151 dump_data
->type
= cpu_to_le32(IWL_FW_ERROR_DUMP_MEM
);
1152 dump_data
->len
= cpu_to_le32(sram_len
+ sizeof(*dump_mem
));
1153 dump_mem
= (void *)dump_data
->data
;
1154 dump_mem
->type
= cpu_to_le32(IWL_FW_ERROR_DUMP_MEM_SRAM
);
1155 dump_mem
->offset
= cpu_to_le32(sram_ofs
);
1156 iwl_trans_read_mem_bytes(mvm
->trans
, sram_ofs
, dump_mem
->data
,
1160 dump_data
= iwl_fw_error_next_data(dump_data
);
1161 dump_data
->type
= cpu_to_le32(IWL_FW_ERROR_DUMP_MEM
);
1162 dump_data
->len
= cpu_to_le32(smem_len
+ sizeof(*dump_mem
));
1163 dump_mem
= (void *)dump_data
->data
;
1164 dump_mem
->type
= cpu_to_le32(IWL_FW_ERROR_DUMP_MEM_SMEM
);
1165 dump_mem
->offset
= cpu_to_le32(mvm
->cfg
->smem_offset
);
1166 iwl_trans_read_mem_bytes(mvm
->trans
, mvm
->cfg
->smem_offset
,
1167 dump_mem
->data
, smem_len
);
1171 dump_data
= iwl_fw_error_next_data(dump_data
);
1172 dump_data
->type
= cpu_to_le32(IWL_FW_ERROR_DUMP_MEM
);
1173 dump_data
->len
= cpu_to_le32(sram2_len
+ sizeof(*dump_mem
));
1174 dump_mem
= (void *)dump_data
->data
;
1175 dump_mem
->type
= cpu_to_le32(IWL_FW_ERROR_DUMP_MEM_SRAM
);
1176 dump_mem
->offset
= cpu_to_le32(mvm
->cfg
->dccm2_offset
);
1177 iwl_trans_read_mem_bytes(mvm
->trans
, mvm
->cfg
->dccm2_offset
,
1178 dump_mem
->data
, sram2_len
);
1181 if (mvm
->cfg
->device_family
== IWL_DEVICE_FAMILY_8000
&&
1182 CSR_HW_REV_STEP(mvm
->trans
->hw_rev
) == SILICON_B_STEP
) {
1183 dump_data
= iwl_fw_error_next_data(dump_data
);
1184 dump_data
->type
= cpu_to_le32(IWL_FW_ERROR_DUMP_MEM
);
1185 dump_data
->len
= cpu_to_le32(IWL8260_ICCM_LEN
+
1187 dump_mem
= (void *)dump_data
->data
;
1188 dump_mem
->type
= cpu_to_le32(IWL_FW_ERROR_DUMP_MEM_SRAM
);
1189 dump_mem
->offset
= cpu_to_le32(IWL8260_ICCM_OFFSET
);
1190 iwl_trans_read_mem_bytes(mvm
->trans
, IWL8260_ICCM_OFFSET
,
1191 dump_mem
->data
, IWL8260_ICCM_LEN
);
1194 fw_error_dump
->trans_ptr
= iwl_trans_dump_data(mvm
->trans
);
1195 fw_error_dump
->op_mode_len
= file_len
;
1196 if (fw_error_dump
->trans_ptr
)
1197 file_len
+= fw_error_dump
->trans_ptr
->len
;
1198 dump_file
->file_len
= cpu_to_le32(file_len
);
1200 dev_coredumpm(mvm
->trans
->dev
, THIS_MODULE
, fw_error_dump
, 0,
1201 GFP_KERNEL
, iwl_mvm_read_coredump
, iwl_mvm_free_coredump
);
1203 clear_bit(IWL_MVM_STATUS_DUMPING_FW_LOG
, &mvm
->status
);
1206 struct iwl_mvm_dump_desc iwl_mvm_dump_desc_assert
= {
1208 .type
= cpu_to_le32(FW_DBG_TRIGGER_FW_ASSERT
),
1212 static void iwl_mvm_restart_cleanup(struct iwl_mvm
*mvm
)
1214 /* clear the D3 reconfig, we only need it to avoid dumping a
1215 * firmware coredump on reconfiguration, we shouldn't do that
1216 * on D3->D0 transition
1218 if (!test_and_clear_bit(IWL_MVM_STATUS_D3_RECONFIG
, &mvm
->status
)) {
1219 mvm
->fw_dump_desc
= &iwl_mvm_dump_desc_assert
;
1220 iwl_mvm_fw_error_dump(mvm
);
1223 /* cleanup all stale references (scan, roc), but keep the
1224 * ucode_down ref until reconfig is complete
1226 iwl_mvm_unref_all_except(mvm
, IWL_MVM_REF_UCODE_DOWN
);
1228 iwl_trans_stop_device(mvm
->trans
);
1230 mvm
->scan_status
= IWL_MVM_SCAN_NONE
;
1231 mvm
->ps_disabled
= false;
1232 mvm
->calibrating
= false;
1234 /* just in case one was running */
1235 ieee80211_remain_on_channel_expired(mvm
->hw
);
1237 ieee80211_iterate_active_interfaces_atomic(
1238 mvm
->hw
, IEEE80211_IFACE_ITER_RESUME_ALL
,
1239 iwl_mvm_cleanup_iterator
, mvm
);
1241 mvm
->p2p_device_vif
= NULL
;
1242 mvm
->d0i3_ap_sta_id
= IWL_MVM_STATION_COUNT
;
1244 iwl_mvm_reset_phy_ctxts(mvm
);
1245 memset(mvm
->fw_key_table
, 0, sizeof(mvm
->fw_key_table
));
1246 memset(mvm
->sta_drained
, 0, sizeof(mvm
->sta_drained
));
1247 memset(mvm
->tfd_drained
, 0, sizeof(mvm
->tfd_drained
));
1248 memset(&mvm
->last_bt_notif
, 0, sizeof(mvm
->last_bt_notif
));
1249 memset(&mvm
->last_bt_notif_old
, 0, sizeof(mvm
->last_bt_notif_old
));
1250 memset(&mvm
->last_bt_ci_cmd
, 0, sizeof(mvm
->last_bt_ci_cmd
));
1251 memset(&mvm
->last_bt_ci_cmd_old
, 0, sizeof(mvm
->last_bt_ci_cmd_old
));
1252 memset(&mvm
->bt_ack_kill_msk
, 0, sizeof(mvm
->bt_ack_kill_msk
));
1253 memset(&mvm
->bt_cts_kill_msk
, 0, sizeof(mvm
->bt_cts_kill_msk
));
1255 ieee80211_wake_queues(mvm
->hw
);
1257 /* clear any stale d0i3 state */
1258 clear_bit(IWL_MVM_STATUS_IN_D0I3
, &mvm
->status
);
1261 mvm
->rx_ba_sessions
= 0;
1262 mvm
->fw_dbg_conf
= FW_DBG_INVALID
;
1264 /* keep statistics ticking */
1265 iwl_mvm_accu_radio_stats(mvm
);
1268 int __iwl_mvm_mac_start(struct iwl_mvm
*mvm
)
1272 lockdep_assert_held(&mvm
->mutex
);
1274 /* Clean up some internal and mac80211 state on restart */
1275 if (test_bit(IWL_MVM_STATUS_IN_HW_RESTART
, &mvm
->status
))
1276 iwl_mvm_restart_cleanup(mvm
);
1278 ret
= iwl_mvm_up(mvm
);
1280 if (ret
&& test_bit(IWL_MVM_STATUS_IN_HW_RESTART
, &mvm
->status
)) {
1281 /* Something went wrong - we need to finish some cleanup
1282 * that normally iwl_mvm_mac_restart_complete() below
1285 clear_bit(IWL_MVM_STATUS_IN_HW_RESTART
, &mvm
->status
);
1286 iwl_mvm_d0i3_enable_tx(mvm
, NULL
);
1292 static int iwl_mvm_mac_start(struct ieee80211_hw
*hw
)
1294 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
1297 /* Some hw restart cleanups must not hold the mutex */
1298 if (test_bit(IWL_MVM_STATUS_IN_HW_RESTART
, &mvm
->status
)) {
1300 * Make sure we are out of d0i3. This is needed
1301 * to make sure the reference accounting is correct
1302 * (and there is no stale d0i3_exit_work).
1304 wait_event_timeout(mvm
->d0i3_exit_waitq
,
1305 !test_bit(IWL_MVM_STATUS_IN_D0I3
,
1310 mutex_lock(&mvm
->mutex
);
1311 ret
= __iwl_mvm_mac_start(mvm
);
1312 mutex_unlock(&mvm
->mutex
);
1317 static void iwl_mvm_restart_complete(struct iwl_mvm
*mvm
)
1321 mutex_lock(&mvm
->mutex
);
1323 clear_bit(IWL_MVM_STATUS_IN_HW_RESTART
, &mvm
->status
);
1324 iwl_mvm_d0i3_enable_tx(mvm
, NULL
);
1325 ret
= iwl_mvm_update_quotas(mvm
, false, NULL
);
1327 IWL_ERR(mvm
, "Failed to update quotas after restart (%d)\n",
1330 /* allow transport/FW low power modes */
1331 iwl_mvm_unref(mvm
, IWL_MVM_REF_UCODE_DOWN
);
1334 * If we have TDLS peers, remove them. We don't know the last seqno/PN
1335 * of packets the FW sent out, so we must reconnect.
1337 iwl_mvm_teardown_tdls_peers(mvm
);
1339 mutex_unlock(&mvm
->mutex
);
1342 static void iwl_mvm_resume_complete(struct iwl_mvm
*mvm
)
1346 if (!iwl_mvm_is_d0i3_supported(mvm
))
1349 mutex_lock(&mvm
->d0i3_suspend_mutex
);
1350 __clear_bit(D0I3_DEFER_WAKEUP
, &mvm
->d0i3_suspend_flags
);
1351 exit_now
= __test_and_clear_bit(D0I3_PENDING_WAKEUP
,
1352 &mvm
->d0i3_suspend_flags
);
1353 mutex_unlock(&mvm
->d0i3_suspend_mutex
);
1356 IWL_DEBUG_RPM(mvm
, "Run deferred d0i3 exit\n");
1357 _iwl_mvm_exit_d0i3(mvm
);
1360 if (mvm
->trans
->d0i3_mode
== IWL_D0I3_MODE_ON_SUSPEND
)
1361 if (!wait_event_timeout(mvm
->d0i3_exit_waitq
,
1362 !test_bit(IWL_MVM_STATUS_IN_D0I3
,
1365 WARN_ONCE(1, "D0i3 exit on resume timed out\n");
1369 iwl_mvm_mac_reconfig_complete(struct ieee80211_hw
*hw
,
1370 enum ieee80211_reconfig_type reconfig_type
)
1372 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
1374 switch (reconfig_type
) {
1375 case IEEE80211_RECONFIG_TYPE_RESTART
:
1376 iwl_mvm_restart_complete(mvm
);
1378 case IEEE80211_RECONFIG_TYPE_SUSPEND
:
1379 iwl_mvm_resume_complete(mvm
);
1384 void __iwl_mvm_mac_stop(struct iwl_mvm
*mvm
)
1386 lockdep_assert_held(&mvm
->mutex
);
1388 /* firmware counters are obviously reset now, but we shouldn't
1389 * partially track so also clear the fw_reset_accu counters.
1391 memset(&mvm
->accu_radio_stats
, 0, sizeof(mvm
->accu_radio_stats
));
1394 * Disallow low power states when the FW is down by taking
1395 * the UCODE_DOWN ref. in case of ongoing hw restart the
1396 * ref is already taken, so don't take it again.
1398 if (!test_bit(IWL_MVM_STATUS_IN_HW_RESTART
, &mvm
->status
))
1399 iwl_mvm_ref(mvm
, IWL_MVM_REF_UCODE_DOWN
);
1401 /* async_handlers_wk is now blocked */
1404 * The work item could be running or queued if the
1405 * ROC time event stops just as we get here.
1407 cancel_work_sync(&mvm
->roc_done_wk
);
1409 iwl_trans_stop_device(mvm
->trans
);
1411 iwl_mvm_async_handlers_purge(mvm
);
1412 /* async_handlers_list is empty and will stay empty: HW is stopped */
1414 /* the fw is stopped, the aux sta is dead: clean up driver state */
1415 iwl_mvm_del_aux_sta(mvm
);
1418 * Clear IN_HW_RESTART flag when stopping the hw (as restart_complete()
1419 * won't be called in this case).
1421 clear_bit(IWL_MVM_STATUS_IN_HW_RESTART
, &mvm
->status
);
1423 /* We shouldn't have any UIDs still set. Loop over all the UIDs to
1424 * make sure there's nothing left there and warn if any is found.
1426 if (mvm
->fw
->ucode_capa
.capa
[0] & IWL_UCODE_TLV_CAPA_UMAC_SCAN
) {
1429 for (i
= 0; i
< IWL_MVM_MAX_SIMULTANEOUS_SCANS
; i
++) {
1430 if (WARN_ONCE(mvm
->scan_uid
[i
],
1431 "UMAC scan UID %d was not cleaned\n",
1433 mvm
->scan_uid
[i
] = 0;
1437 mvm
->ucode_loaded
= false;
1440 static void iwl_mvm_mac_stop(struct ieee80211_hw
*hw
)
1442 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
1444 flush_work(&mvm
->d0i3_exit_work
);
1445 flush_work(&mvm
->async_handlers_wk
);
1446 cancel_delayed_work_sync(&mvm
->fw_dump_wk
);
1447 iwl_mvm_free_fw_dump_desc(mvm
);
1449 mutex_lock(&mvm
->mutex
);
1450 __iwl_mvm_mac_stop(mvm
);
1451 mutex_unlock(&mvm
->mutex
);
1454 * The worker might have been waiting for the mutex, let it run and
1455 * discover that its list is now empty.
1457 cancel_work_sync(&mvm
->async_handlers_wk
);
1460 static struct iwl_mvm_phy_ctxt
*iwl_mvm_get_free_phy_ctxt(struct iwl_mvm
*mvm
)
1464 lockdep_assert_held(&mvm
->mutex
);
1466 for (i
= 0; i
< NUM_PHY_CTX
; i
++)
1467 if (!mvm
->phy_ctxts
[i
].ref
)
1468 return &mvm
->phy_ctxts
[i
];
1470 IWL_ERR(mvm
, "No available PHY context\n");
1474 static int iwl_mvm_set_tx_power(struct iwl_mvm
*mvm
, struct ieee80211_vif
*vif
,
1477 /* FW is in charge of regulatory enforcement */
1478 struct iwl_reduce_tx_power_cmd reduce_txpwr_cmd
= {
1479 .mac_context_id
= iwl_mvm_vif_from_mac80211(vif
)->id
,
1480 .pwr_restriction
= cpu_to_le16(tx_power
),
1483 return iwl_mvm_send_cmd_pdu(mvm
, REDUCE_TX_POWER_CMD
, 0,
1484 sizeof(reduce_txpwr_cmd
),
1488 static int iwl_mvm_mac_add_interface(struct ieee80211_hw
*hw
,
1489 struct ieee80211_vif
*vif
)
1491 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
1492 struct iwl_mvm_vif
*mvmvif
= iwl_mvm_vif_from_mac80211(vif
);
1498 * make sure D0i3 exit is completed, otherwise a target access
1499 * during tx queue configuration could be done when still in
1502 ret
= iwl_mvm_ref_sync(mvm
, IWL_MVM_REF_ADD_IF
);
1507 * Not much to do here. The stack will not allow interface
1508 * types or combinations that we didn't advertise, so we
1509 * don't really have to check the types.
1512 mutex_lock(&mvm
->mutex
);
1514 /* make sure that beacon statistics don't go backwards with FW reset */
1515 if (test_bit(IWL_MVM_STATUS_IN_HW_RESTART
, &mvm
->status
))
1516 mvmvif
->beacon_stats
.accu_num_beacons
+=
1517 mvmvif
->beacon_stats
.num_beacons
;
1519 /* Allocate resources for the MAC context, and add it to the fw */
1520 ret
= iwl_mvm_mac_ctxt_init(mvm
, vif
);
1524 /* Counting number of interfaces is needed for legacy PM */
1525 if (vif
->type
!= NL80211_IFTYPE_P2P_DEVICE
)
1529 * The AP binding flow can be done only after the beacon
1530 * template is configured (which happens only in the mac80211
1531 * start_ap() flow), and adding the broadcast station can happen
1532 * only after the binding.
1533 * In addition, since modifying the MAC before adding a bcast
1534 * station is not allowed by the FW, delay the adding of MAC context to
1535 * the point where we can also add the bcast station.
1536 * In short: there's not much we can do at this point, other than
1537 * allocating resources :)
1539 if (vif
->type
== NL80211_IFTYPE_AP
||
1540 vif
->type
== NL80211_IFTYPE_ADHOC
) {
1541 ret
= iwl_mvm_alloc_bcast_sta(mvm
, vif
);
1543 IWL_ERR(mvm
, "Failed to allocate bcast sta\n");
1547 iwl_mvm_vif_dbgfs_register(mvm
, vif
);
1551 ret
= iwl_mvm_mac_ctxt_add(mvm
, vif
);
1555 ret
= iwl_mvm_power_update_mac(mvm
);
1557 goto out_remove_mac
;
1559 /* beacon filtering */
1560 ret
= iwl_mvm_disable_beacon_filter(mvm
, vif
, 0);
1562 goto out_remove_mac
;
1564 if (!mvm
->bf_allowed_vif
&&
1565 vif
->type
== NL80211_IFTYPE_STATION
&& !vif
->p2p
) {
1566 mvm
->bf_allowed_vif
= mvmvif
;
1567 vif
->driver_flags
|= IEEE80211_VIF_BEACON_FILTER
|
1568 IEEE80211_VIF_SUPPORTS_CQM_RSSI
;
1572 * P2P_DEVICE interface does not have a channel context assigned to it,
1573 * so a dedicated PHY context is allocated to it and the corresponding
1574 * MAC context is bound to it at this stage.
1576 if (vif
->type
== NL80211_IFTYPE_P2P_DEVICE
) {
1578 mvmvif
->phy_ctxt
= iwl_mvm_get_free_phy_ctxt(mvm
);
1579 if (!mvmvif
->phy_ctxt
) {
1584 iwl_mvm_phy_ctxt_ref(mvm
, mvmvif
->phy_ctxt
);
1585 ret
= iwl_mvm_binding_add_vif(mvm
, vif
);
1589 ret
= iwl_mvm_add_bcast_sta(mvm
, vif
);
1593 /* Save a pointer to p2p device vif, so it can later be used to
1594 * update the p2p device MAC when a GO is started/stopped */
1595 mvm
->p2p_device_vif
= vif
;
1598 iwl_mvm_vif_dbgfs_register(mvm
, vif
);
1602 iwl_mvm_binding_remove_vif(mvm
, vif
);
1604 iwl_mvm_phy_ctxt_unref(mvm
, mvmvif
->phy_ctxt
);
1606 if (mvm
->bf_allowed_vif
== mvmvif
) {
1607 mvm
->bf_allowed_vif
= NULL
;
1608 vif
->driver_flags
&= ~(IEEE80211_VIF_BEACON_FILTER
|
1609 IEEE80211_VIF_SUPPORTS_CQM_RSSI
);
1612 mvmvif
->phy_ctxt
= NULL
;
1613 iwl_mvm_mac_ctxt_remove(mvm
, vif
);
1615 if (vif
->type
!= NL80211_IFTYPE_P2P_DEVICE
)
1618 iwl_mvm_mac_ctxt_release(mvm
, vif
);
1620 mutex_unlock(&mvm
->mutex
);
1622 iwl_mvm_unref(mvm
, IWL_MVM_REF_ADD_IF
);
1627 static void iwl_mvm_prepare_mac_removal(struct iwl_mvm
*mvm
,
1628 struct ieee80211_vif
*vif
)
1630 u32 tfd_msk
= iwl_mvm_mac_get_queues_mask(vif
);
1634 * mac80211 first removes all the stations of the vif and
1635 * then removes the vif. When it removes a station it also
1636 * flushes the AMPDU session. So by now, all the AMPDU sessions
1637 * of all the stations of this vif are closed, and the queues
1638 * of these AMPDU sessions are properly closed.
1639 * We still need to take care of the shared queues of the vif.
1642 mutex_lock(&mvm
->mutex
);
1643 iwl_mvm_flush_tx_path(mvm
, tfd_msk
, true);
1644 mutex_unlock(&mvm
->mutex
);
1647 * There are transports that buffer a few frames in the host.
1648 * For these, the flush above isn't enough since while we were
1649 * flushing, the transport might have sent more frames to the
1650 * device. To solve this, wait here until the transport is
1651 * empty. Technically, this could have replaced the flush
1652 * above, but flush is much faster than draining. So flush
1653 * first, and drain to make sure we have no frames in the
1654 * transport anymore.
1655 * If a station still had frames on the shared queues, it is
1656 * already marked as draining, so to complete the draining, we
1657 * just need to wait until the transport is empty.
1659 iwl_trans_wait_tx_queue_empty(mvm
->trans
, tfd_msk
);
1662 if (vif
->type
== NL80211_IFTYPE_P2P_DEVICE
) {
1664 * Flush the ROC worker which will flush the OFFCHANNEL queue.
1665 * We assume here that all the packets sent to the OFFCHANNEL
1666 * queue are sent in ROC session.
1668 flush_work(&mvm
->roc_done_wk
);
1671 * By now, all the AC queues are empty. The AGG queues are
1672 * empty too. We already got all the Tx responses for all the
1673 * packets in the queues. The drain work can have been
1674 * triggered. Flush it.
1676 flush_work(&mvm
->sta_drained_wk
);
1680 static void iwl_mvm_mac_remove_interface(struct ieee80211_hw
*hw
,
1681 struct ieee80211_vif
*vif
)
1683 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
1684 struct iwl_mvm_vif
*mvmvif
= iwl_mvm_vif_from_mac80211(vif
);
1686 iwl_mvm_prepare_mac_removal(mvm
, vif
);
1688 mutex_lock(&mvm
->mutex
);
1690 if (mvm
->bf_allowed_vif
== mvmvif
) {
1691 mvm
->bf_allowed_vif
= NULL
;
1692 vif
->driver_flags
&= ~(IEEE80211_VIF_BEACON_FILTER
|
1693 IEEE80211_VIF_SUPPORTS_CQM_RSSI
);
1696 iwl_mvm_vif_dbgfs_clean(mvm
, vif
);
1699 * For AP/GO interface, the tear down of the resources allocated to the
1700 * interface is be handled as part of the stop_ap flow.
1702 if (vif
->type
== NL80211_IFTYPE_AP
||
1703 vif
->type
== NL80211_IFTYPE_ADHOC
) {
1704 #ifdef CONFIG_NL80211_TESTMODE
1705 if (vif
== mvm
->noa_vif
) {
1706 mvm
->noa_vif
= NULL
;
1707 mvm
->noa_duration
= 0;
1710 iwl_mvm_dealloc_bcast_sta(mvm
, vif
);
1714 if (vif
->type
== NL80211_IFTYPE_P2P_DEVICE
) {
1715 mvm
->p2p_device_vif
= NULL
;
1716 iwl_mvm_rm_bcast_sta(mvm
, vif
);
1717 iwl_mvm_binding_remove_vif(mvm
, vif
);
1718 iwl_mvm_phy_ctxt_unref(mvm
, mvmvif
->phy_ctxt
);
1719 mvmvif
->phy_ctxt
= NULL
;
1722 if (mvm
->vif_count
&& vif
->type
!= NL80211_IFTYPE_P2P_DEVICE
)
1725 iwl_mvm_power_update_mac(mvm
);
1726 iwl_mvm_mac_ctxt_remove(mvm
, vif
);
1729 iwl_mvm_mac_ctxt_release(mvm
, vif
);
1730 mutex_unlock(&mvm
->mutex
);
1733 static int iwl_mvm_mac_config(struct ieee80211_hw
*hw
, u32 changed
)
1738 struct iwl_mvm_mc_iter_data
{
1739 struct iwl_mvm
*mvm
;
1743 static void iwl_mvm_mc_iface_iterator(void *_data
, u8
*mac
,
1744 struct ieee80211_vif
*vif
)
1746 struct iwl_mvm_mc_iter_data
*data
= _data
;
1747 struct iwl_mvm
*mvm
= data
->mvm
;
1748 struct iwl_mcast_filter_cmd
*cmd
= mvm
->mcast_filter_cmd
;
1751 /* if we don't have free ports, mcast frames will be dropped */
1752 if (WARN_ON_ONCE(data
->port_id
>= MAX_PORT_ID_NUM
))
1755 if (vif
->type
!= NL80211_IFTYPE_STATION
||
1756 !vif
->bss_conf
.assoc
)
1759 cmd
->port_id
= data
->port_id
++;
1760 memcpy(cmd
->bssid
, vif
->bss_conf
.bssid
, ETH_ALEN
);
1761 len
= roundup(sizeof(*cmd
) + cmd
->count
* ETH_ALEN
, 4);
1763 ret
= iwl_mvm_send_cmd_pdu(mvm
, MCAST_FILTER_CMD
, CMD_ASYNC
, len
, cmd
);
1765 IWL_ERR(mvm
, "mcast filter cmd error. ret=%d\n", ret
);
1768 static void iwl_mvm_recalc_multicast(struct iwl_mvm
*mvm
)
1770 struct iwl_mvm_mc_iter_data iter_data
= {
1774 lockdep_assert_held(&mvm
->mutex
);
1776 if (WARN_ON_ONCE(!mvm
->mcast_filter_cmd
))
1779 ieee80211_iterate_active_interfaces_atomic(
1780 mvm
->hw
, IEEE80211_IFACE_ITER_NORMAL
,
1781 iwl_mvm_mc_iface_iterator
, &iter_data
);
1784 static u64
iwl_mvm_prepare_multicast(struct ieee80211_hw
*hw
,
1785 struct netdev_hw_addr_list
*mc_list
)
1787 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
1788 struct iwl_mcast_filter_cmd
*cmd
;
1789 struct netdev_hw_addr
*addr
;
1794 addr_count
= netdev_hw_addr_list_count(mc_list
);
1795 pass_all
= addr_count
> MAX_MCAST_FILTERING_ADDRESSES
||
1796 IWL_MVM_FW_MCAST_FILTER_PASS_ALL
;
1800 len
= roundup(sizeof(*cmd
) + addr_count
* ETH_ALEN
, 4);
1801 cmd
= kzalloc(len
, GFP_ATOMIC
);
1807 return (u64
)(unsigned long)cmd
;
1810 netdev_hw_addr_list_for_each(addr
, mc_list
) {
1811 IWL_DEBUG_MAC80211(mvm
, "mcast addr (%d): %pM\n",
1812 cmd
->count
, addr
->addr
);
1813 memcpy(&cmd
->addr_list
[cmd
->count
* ETH_ALEN
],
1814 addr
->addr
, ETH_ALEN
);
1818 return (u64
)(unsigned long)cmd
;
1821 static void iwl_mvm_configure_filter(struct ieee80211_hw
*hw
,
1822 unsigned int changed_flags
,
1823 unsigned int *total_flags
,
1826 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
1827 struct iwl_mcast_filter_cmd
*cmd
= (void *)(unsigned long)multicast
;
1829 mutex_lock(&mvm
->mutex
);
1831 /* replace previous configuration */
1832 kfree(mvm
->mcast_filter_cmd
);
1833 mvm
->mcast_filter_cmd
= cmd
;
1838 iwl_mvm_recalc_multicast(mvm
);
1840 mutex_unlock(&mvm
->mutex
);
1844 #ifdef CONFIG_IWLWIFI_BCAST_FILTERING
1845 struct iwl_bcast_iter_data
{
1846 struct iwl_mvm
*mvm
;
1847 struct iwl_bcast_filter_cmd
*cmd
;
1852 iwl_mvm_set_bcast_filter(struct ieee80211_vif
*vif
,
1853 const struct iwl_fw_bcast_filter
*in_filter
,
1854 struct iwl_fw_bcast_filter
*out_filter
)
1856 struct iwl_fw_bcast_filter_attr
*attr
;
1859 memcpy(out_filter
, in_filter
, sizeof(*out_filter
));
1861 for (i
= 0; i
< ARRAY_SIZE(out_filter
->attrs
); i
++) {
1862 attr
= &out_filter
->attrs
[i
];
1867 switch (attr
->reserved1
) {
1868 case cpu_to_le16(BC_FILTER_MAGIC_IP
):
1869 if (vif
->bss_conf
.arp_addr_cnt
!= 1) {
1874 attr
->val
= vif
->bss_conf
.arp_addr_list
[0];
1876 case cpu_to_le16(BC_FILTER_MAGIC_MAC
):
1877 attr
->val
= *(__be32
*)&vif
->addr
[2];
1882 attr
->reserved1
= 0;
1883 out_filter
->num_attrs
++;
1887 static void iwl_mvm_bcast_filter_iterator(void *_data
, u8
*mac
,
1888 struct ieee80211_vif
*vif
)
1890 struct iwl_bcast_iter_data
*data
= _data
;
1891 struct iwl_mvm
*mvm
= data
->mvm
;
1892 struct iwl_bcast_filter_cmd
*cmd
= data
->cmd
;
1893 struct iwl_mvm_vif
*mvmvif
= iwl_mvm_vif_from_mac80211(vif
);
1894 struct iwl_fw_bcast_mac
*bcast_mac
;
1897 if (WARN_ON(mvmvif
->id
>= ARRAY_SIZE(cmd
->macs
)))
1900 bcast_mac
= &cmd
->macs
[mvmvif
->id
];
1903 * enable filtering only for associated stations, but not for P2P
1906 if (vif
->type
!= NL80211_IFTYPE_STATION
|| vif
->p2p
||
1907 !vif
->bss_conf
.assoc
)
1910 bcast_mac
->default_discard
= 1;
1912 /* copy all configured filters */
1913 for (i
= 0; mvm
->bcast_filters
[i
].attrs
[0].mask
; i
++) {
1915 * Make sure we don't exceed our filters limit.
1916 * if there is still a valid filter to be configured,
1917 * be on the safe side and just allow bcast for this mac.
1919 if (WARN_ON_ONCE(data
->current_filter
>=
1920 ARRAY_SIZE(cmd
->filters
))) {
1921 bcast_mac
->default_discard
= 0;
1922 bcast_mac
->attached_filters
= 0;
1926 iwl_mvm_set_bcast_filter(vif
,
1927 &mvm
->bcast_filters
[i
],
1928 &cmd
->filters
[data
->current_filter
]);
1930 /* skip current filter if it contains no attributes */
1931 if (!cmd
->filters
[data
->current_filter
].num_attrs
)
1934 /* attach the filter to current mac */
1935 bcast_mac
->attached_filters
|=
1936 cpu_to_le16(BIT(data
->current_filter
));
1938 data
->current_filter
++;
1942 bool iwl_mvm_bcast_filter_build_cmd(struct iwl_mvm
*mvm
,
1943 struct iwl_bcast_filter_cmd
*cmd
)
1945 struct iwl_bcast_iter_data iter_data
= {
1950 if (IWL_MVM_FW_BCAST_FILTER_PASS_ALL
)
1953 memset(cmd
, 0, sizeof(*cmd
));
1954 cmd
->max_bcast_filters
= ARRAY_SIZE(cmd
->filters
);
1955 cmd
->max_macs
= ARRAY_SIZE(cmd
->macs
);
1957 #ifdef CONFIG_IWLWIFI_DEBUGFS
1958 /* use debugfs filters/macs if override is configured */
1959 if (mvm
->dbgfs_bcast_filtering
.override
) {
1960 memcpy(cmd
->filters
, &mvm
->dbgfs_bcast_filtering
.cmd
.filters
,
1961 sizeof(cmd
->filters
));
1962 memcpy(cmd
->macs
, &mvm
->dbgfs_bcast_filtering
.cmd
.macs
,
1968 /* if no filters are configured, do nothing */
1969 if (!mvm
->bcast_filters
)
1972 /* configure and attach these filters for each associated sta vif */
1973 ieee80211_iterate_active_interfaces(
1974 mvm
->hw
, IEEE80211_IFACE_ITER_NORMAL
,
1975 iwl_mvm_bcast_filter_iterator
, &iter_data
);
1979 static int iwl_mvm_configure_bcast_filter(struct iwl_mvm
*mvm
,
1980 struct ieee80211_vif
*vif
)
1982 struct iwl_bcast_filter_cmd cmd
;
1984 if (!(mvm
->fw
->ucode_capa
.flags
& IWL_UCODE_TLV_FLAGS_BCAST_FILTERING
))
1987 if (!iwl_mvm_bcast_filter_build_cmd(mvm
, &cmd
))
1990 return iwl_mvm_send_cmd_pdu(mvm
, BCAST_FILTER_CMD
, 0,
1994 static inline int iwl_mvm_configure_bcast_filter(struct iwl_mvm
*mvm
,
1995 struct ieee80211_vif
*vif
)
2001 static void iwl_mvm_bss_info_changed_station(struct iwl_mvm
*mvm
,
2002 struct ieee80211_vif
*vif
,
2003 struct ieee80211_bss_conf
*bss_conf
,
2006 struct iwl_mvm_vif
*mvmvif
= iwl_mvm_vif_from_mac80211(vif
);
2010 * Re-calculate the tsf id, as the master-slave relations depend on the
2011 * beacon interval, which was not known when the station interface was
2014 if (changes
& BSS_CHANGED_ASSOC
&& bss_conf
->assoc
)
2015 iwl_mvm_mac_ctxt_recalc_tsf_id(mvm
, vif
);
2018 * If we're not associated yet, take the (new) BSSID before associating
2019 * so the firmware knows. If we're already associated, then use the old
2020 * BSSID here, and we'll send a cleared one later in the CHANGED_ASSOC
2021 * branch for disassociation below.
2023 if (changes
& BSS_CHANGED_BSSID
&& !mvmvif
->associated
)
2024 memcpy(mvmvif
->bssid
, bss_conf
->bssid
, ETH_ALEN
);
2026 ret
= iwl_mvm_mac_ctxt_changed(mvm
, vif
, false, mvmvif
->bssid
);
2028 IWL_ERR(mvm
, "failed to update MAC %pM\n", vif
->addr
);
2030 /* after sending it once, adopt mac80211 data */
2031 memcpy(mvmvif
->bssid
, bss_conf
->bssid
, ETH_ALEN
);
2032 mvmvif
->associated
= bss_conf
->assoc
;
2034 if (changes
& BSS_CHANGED_ASSOC
) {
2035 if (bss_conf
->assoc
) {
2036 /* clear statistics to get clean beacon counter */
2037 iwl_mvm_request_statistics(mvm
, true);
2038 memset(&mvmvif
->beacon_stats
, 0,
2039 sizeof(mvmvif
->beacon_stats
));
2041 /* add quota for this interface */
2042 ret
= iwl_mvm_update_quotas(mvm
, true, NULL
);
2044 IWL_ERR(mvm
, "failed to update quotas\n");
2048 if (test_bit(IWL_MVM_STATUS_IN_HW_RESTART
,
2051 * If we're restarting then the firmware will
2052 * obviously have lost synchronisation with
2053 * the AP. It will attempt to synchronise by
2054 * itself, but we can make it more reliable by
2055 * scheduling a session protection time event.
2057 * The firmware needs to receive a beacon to
2058 * catch up with synchronisation, use 110% of
2059 * the beacon interval.
2061 * Set a large maximum delay to allow for more
2062 * than a single interface.
2064 u32 dur
= (11 * vif
->bss_conf
.beacon_int
) / 10;
2065 iwl_mvm_protect_session(mvm
, vif
, dur
, dur
,
2069 iwl_mvm_sf_update(mvm
, vif
, false);
2070 iwl_mvm_power_vif_assoc(mvm
, vif
);
2072 iwl_mvm_ref(mvm
, IWL_MVM_REF_P2P_CLIENT
);
2073 iwl_mvm_update_smps(mvm
, vif
,
2074 IWL_MVM_SMPS_REQ_PROT
,
2075 IEEE80211_SMPS_DYNAMIC
);
2077 } else if (mvmvif
->ap_sta_id
!= IWL_MVM_STATION_COUNT
) {
2079 * If update fails - SF might be running in associated
2080 * mode while disassociated - which is forbidden.
2082 WARN_ONCE(iwl_mvm_sf_update(mvm
, vif
, false),
2083 "Failed to update SF upon disassociation\n");
2085 /* remove AP station now that the MAC is unassoc */
2086 ret
= iwl_mvm_rm_sta_id(mvm
, vif
, mvmvif
->ap_sta_id
);
2088 IWL_ERR(mvm
, "failed to remove AP station\n");
2090 if (mvm
->d0i3_ap_sta_id
== mvmvif
->ap_sta_id
)
2091 mvm
->d0i3_ap_sta_id
= IWL_MVM_STATION_COUNT
;
2092 mvmvif
->ap_sta_id
= IWL_MVM_STATION_COUNT
;
2093 /* remove quota for this interface */
2094 ret
= iwl_mvm_update_quotas(mvm
, false, NULL
);
2096 IWL_ERR(mvm
, "failed to update quotas\n");
2099 iwl_mvm_unref(mvm
, IWL_MVM_REF_P2P_CLIENT
);
2101 /* this will take the cleared BSSID from bss_conf */
2102 ret
= iwl_mvm_mac_ctxt_changed(mvm
, vif
, false, NULL
);
2105 "failed to update MAC %pM (clear after unassoc)\n",
2109 iwl_mvm_recalc_multicast(mvm
);
2110 iwl_mvm_configure_bcast_filter(mvm
, vif
);
2112 /* reset rssi values */
2113 mvmvif
->bf_data
.ave_beacon_signal
= 0;
2115 iwl_mvm_bt_coex_vif_change(mvm
);
2116 iwl_mvm_update_smps(mvm
, vif
, IWL_MVM_SMPS_REQ_TT
,
2117 IEEE80211_SMPS_AUTOMATIC
);
2118 } else if (changes
& BSS_CHANGED_BEACON_INFO
) {
2120 * We received a beacon _after_ association so
2121 * remove the session protection.
2123 iwl_mvm_remove_time_event(mvm
, mvmvif
,
2124 &mvmvif
->time_event_data
);
2127 if (changes
& BSS_CHANGED_BEACON_INFO
) {
2128 iwl_mvm_sf_update(mvm
, vif
, false);
2129 WARN_ON(iwl_mvm_enable_beacon_filter(mvm
, vif
, 0));
2132 if (changes
& (BSS_CHANGED_PS
| BSS_CHANGED_P2P_PS
| BSS_CHANGED_QOS
)) {
2133 ret
= iwl_mvm_power_update_mac(mvm
);
2135 IWL_ERR(mvm
, "failed to update power mode\n");
2138 if (changes
& BSS_CHANGED_TXPOWER
) {
2139 IWL_DEBUG_CALIB(mvm
, "Changing TX Power to %d\n",
2141 iwl_mvm_set_tx_power(mvm
, vif
, bss_conf
->txpower
);
2144 if (changes
& BSS_CHANGED_CQM
) {
2145 IWL_DEBUG_MAC80211(mvm
, "cqm info_changed\n");
2146 /* reset cqm events tracking */
2147 mvmvif
->bf_data
.last_cqm_event
= 0;
2148 if (mvmvif
->bf_data
.bf_enabled
) {
2149 ret
= iwl_mvm_enable_beacon_filter(mvm
, vif
, 0);
2152 "failed to update CQM thresholds\n");
2156 if (changes
& BSS_CHANGED_ARP_FILTER
) {
2157 IWL_DEBUG_MAC80211(mvm
, "arp filter changed\n");
2158 iwl_mvm_configure_bcast_filter(mvm
, vif
);
2162 static int iwl_mvm_start_ap_ibss(struct ieee80211_hw
*hw
,
2163 struct ieee80211_vif
*vif
)
2165 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
2166 struct iwl_mvm_vif
*mvmvif
= iwl_mvm_vif_from_mac80211(vif
);
2170 * iwl_mvm_mac_ctxt_add() might read directly from the device
2171 * (the system time), so make sure it is available.
2173 ret
= iwl_mvm_ref_sync(mvm
, IWL_MVM_REF_START_AP
);
2177 mutex_lock(&mvm
->mutex
);
2179 /* Send the beacon template */
2180 ret
= iwl_mvm_mac_ctxt_beacon_changed(mvm
, vif
);
2185 * Re-calculate the tsf id, as the master-slave relations depend on the
2186 * beacon interval, which was not known when the AP interface was added.
2188 if (vif
->type
== NL80211_IFTYPE_AP
)
2189 iwl_mvm_mac_ctxt_recalc_tsf_id(mvm
, vif
);
2191 /* Add the mac context */
2192 ret
= iwl_mvm_mac_ctxt_add(mvm
, vif
);
2196 /* Perform the binding */
2197 ret
= iwl_mvm_binding_add_vif(mvm
, vif
);
2201 /* Send the bcast station. At this stage the TBTT and DTIM time events
2202 * are added and applied to the scheduler */
2203 ret
= iwl_mvm_send_add_bcast_sta(mvm
, vif
);
2207 /* must be set before quota calculations */
2208 mvmvif
->ap_ibss_active
= true;
2210 /* power updated needs to be done before quotas */
2211 iwl_mvm_power_update_mac(mvm
);
2213 ret
= iwl_mvm_update_quotas(mvm
, false, NULL
);
2215 goto out_quota_failed
;
2217 /* Need to update the P2P Device MAC (only GO, IBSS is single vif) */
2218 if (vif
->p2p
&& mvm
->p2p_device_vif
)
2219 iwl_mvm_mac_ctxt_changed(mvm
, mvm
->p2p_device_vif
, false, NULL
);
2221 iwl_mvm_ref(mvm
, IWL_MVM_REF_AP_IBSS
);
2223 iwl_mvm_bt_coex_vif_change(mvm
);
2225 /* we don't support TDLS during DCM */
2226 if (iwl_mvm_phy_ctx_count(mvm
) > 1)
2227 iwl_mvm_teardown_tdls_peers(mvm
);
2232 iwl_mvm_power_update_mac(mvm
);
2233 mvmvif
->ap_ibss_active
= false;
2234 iwl_mvm_send_rm_bcast_sta(mvm
, vif
);
2236 iwl_mvm_binding_remove_vif(mvm
, vif
);
2238 iwl_mvm_mac_ctxt_remove(mvm
, vif
);
2240 mutex_unlock(&mvm
->mutex
);
2241 iwl_mvm_unref(mvm
, IWL_MVM_REF_START_AP
);
2245 static void iwl_mvm_stop_ap_ibss(struct ieee80211_hw
*hw
,
2246 struct ieee80211_vif
*vif
)
2248 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
2249 struct iwl_mvm_vif
*mvmvif
= iwl_mvm_vif_from_mac80211(vif
);
2251 iwl_mvm_prepare_mac_removal(mvm
, vif
);
2253 mutex_lock(&mvm
->mutex
);
2255 /* Handle AP stop while in CSA */
2256 if (rcu_access_pointer(mvm
->csa_vif
) == vif
) {
2257 iwl_mvm_remove_time_event(mvm
, mvmvif
,
2258 &mvmvif
->time_event_data
);
2259 RCU_INIT_POINTER(mvm
->csa_vif
, NULL
);
2262 if (rcu_access_pointer(mvm
->csa_tx_blocked_vif
) == vif
) {
2263 RCU_INIT_POINTER(mvm
->csa_tx_blocked_vif
, NULL
);
2264 mvm
->csa_tx_block_bcn_timeout
= 0;
2267 mvmvif
->ap_ibss_active
= false;
2268 mvm
->ap_last_beacon_gp2
= 0;
2270 iwl_mvm_bt_coex_vif_change(mvm
);
2272 iwl_mvm_unref(mvm
, IWL_MVM_REF_AP_IBSS
);
2274 /* Need to update the P2P Device MAC (only GO, IBSS is single vif) */
2275 if (vif
->p2p
&& mvm
->p2p_device_vif
)
2276 iwl_mvm_mac_ctxt_changed(mvm
, mvm
->p2p_device_vif
, false, NULL
);
2278 iwl_mvm_update_quotas(mvm
, false, NULL
);
2279 iwl_mvm_send_rm_bcast_sta(mvm
, vif
);
2280 iwl_mvm_binding_remove_vif(mvm
, vif
);
2282 iwl_mvm_power_update_mac(mvm
);
2284 iwl_mvm_mac_ctxt_remove(mvm
, vif
);
2286 mutex_unlock(&mvm
->mutex
);
2290 iwl_mvm_bss_info_changed_ap_ibss(struct iwl_mvm
*mvm
,
2291 struct ieee80211_vif
*vif
,
2292 struct ieee80211_bss_conf
*bss_conf
,
2295 struct iwl_mvm_vif
*mvmvif
= iwl_mvm_vif_from_mac80211(vif
);
2297 /* Changes will be applied when the AP/IBSS is started */
2298 if (!mvmvif
->ap_ibss_active
)
2301 if (changes
& (BSS_CHANGED_ERP_CTS_PROT
| BSS_CHANGED_HT
|
2302 BSS_CHANGED_BANDWIDTH
| BSS_CHANGED_QOS
) &&
2303 iwl_mvm_mac_ctxt_changed(mvm
, vif
, false, NULL
))
2304 IWL_ERR(mvm
, "failed to update MAC %pM\n", vif
->addr
);
2306 /* Need to send a new beacon template to the FW */
2307 if (changes
& BSS_CHANGED_BEACON
&&
2308 iwl_mvm_mac_ctxt_beacon_changed(mvm
, vif
))
2309 IWL_WARN(mvm
, "Failed updating beacon data\n");
2311 if (changes
& BSS_CHANGED_TXPOWER
) {
2312 IWL_DEBUG_CALIB(mvm
, "Changing TX Power to %d\n",
2314 iwl_mvm_set_tx_power(mvm
, vif
, bss_conf
->txpower
);
2319 static void iwl_mvm_bss_info_changed(struct ieee80211_hw
*hw
,
2320 struct ieee80211_vif
*vif
,
2321 struct ieee80211_bss_conf
*bss_conf
,
2324 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
2327 * iwl_mvm_bss_info_changed_station() might call
2328 * iwl_mvm_protect_session(), which reads directly from
2329 * the device (the system time), so make sure it is available.
2331 if (iwl_mvm_ref_sync(mvm
, IWL_MVM_REF_BSS_CHANGED
))
2334 mutex_lock(&mvm
->mutex
);
2336 if (changes
& BSS_CHANGED_IDLE
&& !bss_conf
->idle
)
2337 iwl_mvm_scan_offload_stop(mvm
, true);
2339 switch (vif
->type
) {
2340 case NL80211_IFTYPE_STATION
:
2341 iwl_mvm_bss_info_changed_station(mvm
, vif
, bss_conf
, changes
);
2343 case NL80211_IFTYPE_AP
:
2344 case NL80211_IFTYPE_ADHOC
:
2345 iwl_mvm_bss_info_changed_ap_ibss(mvm
, vif
, bss_conf
, changes
);
2348 /* shouldn't happen */
2352 mutex_unlock(&mvm
->mutex
);
2353 iwl_mvm_unref(mvm
, IWL_MVM_REF_BSS_CHANGED
);
2356 static int iwl_mvm_cancel_scan_wait_notif(struct iwl_mvm
*mvm
,
2357 enum iwl_scan_status scan_type
)
2360 bool wait_for_handlers
= false;
2362 mutex_lock(&mvm
->mutex
);
2364 if (mvm
->scan_status
!= scan_type
) {
2366 /* make sure there are no pending notifications */
2367 wait_for_handlers
= true;
2371 switch (scan_type
) {
2372 case IWL_MVM_SCAN_SCHED
:
2373 ret
= iwl_mvm_scan_offload_stop(mvm
, true);
2375 case IWL_MVM_SCAN_OS
:
2376 ret
= iwl_mvm_cancel_scan(mvm
);
2378 case IWL_MVM_SCAN_NONE
:
2387 wait_for_handlers
= true;
2389 mutex_unlock(&mvm
->mutex
);
2391 /* make sure we consume the completion notification */
2392 if (wait_for_handlers
)
2393 iwl_mvm_wait_for_async_handlers(mvm
);
2397 static int iwl_mvm_mac_hw_scan(struct ieee80211_hw
*hw
,
2398 struct ieee80211_vif
*vif
,
2399 struct ieee80211_scan_request
*hw_req
)
2401 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
2402 struct cfg80211_scan_request
*req
= &hw_req
->req
;
2405 if (req
->n_channels
== 0 ||
2406 req
->n_channels
> mvm
->fw
->ucode_capa
.n_scan_channels
)
2409 if (!(mvm
->fw
->ucode_capa
.capa
[0] & IWL_UCODE_TLV_CAPA_UMAC_SCAN
)) {
2410 ret
= iwl_mvm_cancel_scan_wait_notif(mvm
, IWL_MVM_SCAN_SCHED
);
2415 mutex_lock(&mvm
->mutex
);
2417 if (iwl_mvm_is_lar_supported(mvm
) && !mvm
->lar_regdom_set
) {
2418 IWL_ERR(mvm
, "scan while LAR regdomain is not set\n");
2423 if (mvm
->scan_status
!= IWL_MVM_SCAN_NONE
) {
2428 iwl_mvm_ref(mvm
, IWL_MVM_REF_SCAN
);
2430 if (mvm
->fw
->ucode_capa
.capa
[0] & IWL_UCODE_TLV_CAPA_UMAC_SCAN
)
2431 ret
= iwl_mvm_scan_umac(mvm
, vif
, hw_req
);
2433 ret
= iwl_mvm_unified_scan_lmac(mvm
, vif
, hw_req
);
2436 iwl_mvm_unref(mvm
, IWL_MVM_REF_SCAN
);
2438 mutex_unlock(&mvm
->mutex
);
2442 static void iwl_mvm_mac_cancel_hw_scan(struct ieee80211_hw
*hw
,
2443 struct ieee80211_vif
*vif
)
2445 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
2447 mutex_lock(&mvm
->mutex
);
2449 /* Due to a race condition, it's possible that mac80211 asks
2450 * us to stop a hw_scan when it's already stopped. This can
2451 * happen, for instance, if we stopped the scan ourselves,
2452 * called ieee80211_scan_completed() and the userspace called
2453 * cancel scan scan before ieee80211_scan_work() could run.
2454 * To handle that, simply return if the scan is not running.
2456 /* FIXME: for now, we ignore this race for UMAC scans, since
2457 * they don't set the scan_status.
2459 if ((mvm
->scan_status
== IWL_MVM_SCAN_OS
) ||
2460 (mvm
->fw
->ucode_capa
.capa
[0] & IWL_UCODE_TLV_CAPA_UMAC_SCAN
))
2461 iwl_mvm_cancel_scan(mvm
);
2463 mutex_unlock(&mvm
->mutex
);
2467 iwl_mvm_mac_allow_buffered_frames(struct ieee80211_hw
*hw
,
2468 struct ieee80211_sta
*sta
, u16 tids
,
2470 enum ieee80211_frame_release_type reason
,
2473 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
2475 /* Called when we need to transmit (a) frame(s) from mac80211 */
2477 iwl_mvm_sta_modify_sleep_tx_count(mvm
, sta
, reason
, num_frames
,
2478 tids
, more_data
, false);
2482 iwl_mvm_mac_release_buffered_frames(struct ieee80211_hw
*hw
,
2483 struct ieee80211_sta
*sta
, u16 tids
,
2485 enum ieee80211_frame_release_type reason
,
2488 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
2490 /* Called when we need to transmit (a) frame(s) from agg queue */
2492 iwl_mvm_sta_modify_sleep_tx_count(mvm
, sta
, reason
, num_frames
,
2493 tids
, more_data
, true);
2496 static void iwl_mvm_mac_sta_notify(struct ieee80211_hw
*hw
,
2497 struct ieee80211_vif
*vif
,
2498 enum sta_notify_cmd cmd
,
2499 struct ieee80211_sta
*sta
)
2501 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
2502 struct iwl_mvm_sta
*mvmsta
= iwl_mvm_sta_from_mac80211(sta
);
2503 unsigned long txqs
= 0, tids
= 0;
2506 spin_lock_bh(&mvmsta
->lock
);
2507 for (tid
= 0; tid
< IWL_MAX_TID_COUNT
; tid
++) {
2508 struct iwl_mvm_tid_data
*tid_data
= &mvmsta
->tid_data
[tid
];
2510 if (tid_data
->state
!= IWL_AGG_ON
&&
2511 tid_data
->state
!= IWL_EMPTYING_HW_QUEUE_DELBA
)
2514 __set_bit(tid_data
->txq_id
, &txqs
);
2516 if (iwl_mvm_tid_queued(tid_data
) == 0)
2519 __set_bit(tid
, &tids
);
2523 case STA_NOTIFY_SLEEP
:
2524 if (atomic_read(&mvm
->pending_frames
[mvmsta
->sta_id
]) > 0)
2525 ieee80211_sta_block_awake(hw
, sta
, true);
2527 for_each_set_bit(tid
, &tids
, IWL_MAX_TID_COUNT
)
2528 ieee80211_sta_set_buffered(sta
, tid
, true);
2531 iwl_trans_freeze_txq_timer(mvm
->trans
, txqs
, true);
2533 * The fw updates the STA to be asleep. Tx packets on the Tx
2534 * queues to this station will not be transmitted. The fw will
2535 * send a Tx response with TX_STATUS_FAIL_DEST_PS.
2538 case STA_NOTIFY_AWAKE
:
2539 if (WARN_ON(mvmsta
->sta_id
== IWL_MVM_STATION_COUNT
))
2543 iwl_trans_freeze_txq_timer(mvm
->trans
, txqs
, false);
2544 iwl_mvm_sta_modify_ps_wake(mvm
, sta
);
2549 spin_unlock_bh(&mvmsta
->lock
);
2552 static void iwl_mvm_sta_pre_rcu_remove(struct ieee80211_hw
*hw
,
2553 struct ieee80211_vif
*vif
,
2554 struct ieee80211_sta
*sta
)
2556 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
2557 struct iwl_mvm_sta
*mvm_sta
= iwl_mvm_sta_from_mac80211(sta
);
2560 * This is called before mac80211 does RCU synchronisation,
2561 * so here we already invalidate our internal RCU-protected
2562 * station pointer. The rest of the code will thus no longer
2563 * be able to find the station this way, and we don't rely
2564 * on further RCU synchronisation after the sta_state()
2565 * callback deleted the station.
2567 mutex_lock(&mvm
->mutex
);
2568 if (sta
== rcu_access_pointer(mvm
->fw_id_to_mac_id
[mvm_sta
->sta_id
]))
2569 rcu_assign_pointer(mvm
->fw_id_to_mac_id
[mvm_sta
->sta_id
],
2571 mutex_unlock(&mvm
->mutex
);
2574 static void iwl_mvm_check_uapsd(struct iwl_mvm
*mvm
, struct ieee80211_vif
*vif
,
2577 if (!(mvm
->fw
->ucode_capa
.flags
& IWL_UCODE_TLV_FLAGS_UAPSD_SUPPORT
))
2580 if (iwlwifi_mod_params
.uapsd_disable
) {
2581 vif
->driver_flags
&= ~IEEE80211_VIF_SUPPORTS_UAPSD
;
2585 vif
->driver_flags
|= IEEE80211_VIF_SUPPORTS_UAPSD
;
2588 static int iwl_mvm_mac_sta_state(struct ieee80211_hw
*hw
,
2589 struct ieee80211_vif
*vif
,
2590 struct ieee80211_sta
*sta
,
2591 enum ieee80211_sta_state old_state
,
2592 enum ieee80211_sta_state new_state
)
2594 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
2595 struct iwl_mvm_vif
*mvmvif
= iwl_mvm_vif_from_mac80211(vif
);
2598 IWL_DEBUG_MAC80211(mvm
, "station %pM state change %d->%d\n",
2599 sta
->addr
, old_state
, new_state
);
2601 /* this would be a mac80211 bug ... but don't crash */
2602 if (WARN_ON_ONCE(!mvmvif
->phy_ctxt
))
2605 /* if a STA is being removed, reuse its ID */
2606 flush_work(&mvm
->sta_drained_wk
);
2608 mutex_lock(&mvm
->mutex
);
2609 if (old_state
== IEEE80211_STA_NOTEXIST
&&
2610 new_state
== IEEE80211_STA_NONE
) {
2612 * Firmware bug - it'll crash if the beacon interval is less
2613 * than 16. We can't avoid connecting at all, so refuse the
2614 * station state change, this will cause mac80211 to abandon
2615 * attempts to connect to this AP, and eventually wpa_s will
2616 * blacklist the AP...
2618 if (vif
->type
== NL80211_IFTYPE_STATION
&&
2619 vif
->bss_conf
.beacon_int
< 16) {
2621 "AP %pM beacon interval is %d, refusing due to firmware bug!\n",
2622 sta
->addr
, vif
->bss_conf
.beacon_int
);
2629 iwl_mvm_tdls_sta_count(mvm
, NULL
) ==
2630 IWL_MVM_TDLS_STA_COUNT
||
2631 iwl_mvm_phy_ctx_count(mvm
) > 1)) {
2632 IWL_DEBUG_MAC80211(mvm
, "refusing TDLS sta\n");
2637 ret
= iwl_mvm_add_sta(mvm
, vif
, sta
);
2638 if (sta
->tdls
&& ret
== 0)
2639 iwl_mvm_recalc_tdls_state(mvm
, vif
, true);
2640 } else if (old_state
== IEEE80211_STA_NONE
&&
2641 new_state
== IEEE80211_STA_AUTH
) {
2643 * EBS may be disabled due to previous failures reported by FW.
2644 * Reset EBS status here assuming environment has been changed.
2646 mvm
->last_ebs_successful
= true;
2647 iwl_mvm_check_uapsd(mvm
, vif
, sta
->addr
);
2649 } else if (old_state
== IEEE80211_STA_AUTH
&&
2650 new_state
== IEEE80211_STA_ASSOC
) {
2651 ret
= iwl_mvm_update_sta(mvm
, vif
, sta
);
2653 iwl_mvm_rs_rate_init(mvm
, sta
,
2654 mvmvif
->phy_ctxt
->channel
->band
,
2656 } else if (old_state
== IEEE80211_STA_ASSOC
&&
2657 new_state
== IEEE80211_STA_AUTHORIZED
) {
2659 /* we don't support TDLS during DCM */
2660 if (iwl_mvm_phy_ctx_count(mvm
) > 1)
2661 iwl_mvm_teardown_tdls_peers(mvm
);
2663 /* enable beacon filtering */
2664 WARN_ON(iwl_mvm_enable_beacon_filter(mvm
, vif
, 0));
2666 } else if (old_state
== IEEE80211_STA_AUTHORIZED
&&
2667 new_state
== IEEE80211_STA_ASSOC
) {
2668 /* disable beacon filtering */
2669 WARN_ON(iwl_mvm_disable_beacon_filter(mvm
, vif
, 0));
2671 } else if (old_state
== IEEE80211_STA_ASSOC
&&
2672 new_state
== IEEE80211_STA_AUTH
) {
2674 } else if (old_state
== IEEE80211_STA_AUTH
&&
2675 new_state
== IEEE80211_STA_NONE
) {
2677 } else if (old_state
== IEEE80211_STA_NONE
&&
2678 new_state
== IEEE80211_STA_NOTEXIST
) {
2679 ret
= iwl_mvm_rm_sta(mvm
, vif
, sta
);
2681 iwl_mvm_recalc_tdls_state(mvm
, vif
, false);
2686 mutex_unlock(&mvm
->mutex
);
2688 if (sta
->tdls
&& ret
== 0) {
2689 if (old_state
== IEEE80211_STA_NOTEXIST
&&
2690 new_state
== IEEE80211_STA_NONE
)
2691 ieee80211_reserve_tid(sta
, IWL_MVM_TDLS_FW_TID
);
2692 else if (old_state
== IEEE80211_STA_NONE
&&
2693 new_state
== IEEE80211_STA_NOTEXIST
)
2694 ieee80211_unreserve_tid(sta
, IWL_MVM_TDLS_FW_TID
);
2700 static int iwl_mvm_mac_set_rts_threshold(struct ieee80211_hw
*hw
, u32 value
)
2702 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
2704 mvm
->rts_threshold
= value
;
2709 static void iwl_mvm_sta_rc_update(struct ieee80211_hw
*hw
,
2710 struct ieee80211_vif
*vif
,
2711 struct ieee80211_sta
*sta
, u32 changed
)
2713 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
2715 if (vif
->type
== NL80211_IFTYPE_STATION
&&
2716 changed
& IEEE80211_RC_NSS_CHANGED
)
2717 iwl_mvm_sf_update(mvm
, vif
, false);
2720 static int iwl_mvm_mac_conf_tx(struct ieee80211_hw
*hw
,
2721 struct ieee80211_vif
*vif
, u16 ac
,
2722 const struct ieee80211_tx_queue_params
*params
)
2724 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
2725 struct iwl_mvm_vif
*mvmvif
= iwl_mvm_vif_from_mac80211(vif
);
2727 mvmvif
->queue_params
[ac
] = *params
;
2730 * No need to update right away, we'll get BSS_CHANGED_QOS
2731 * The exception is P2P_DEVICE interface which needs immediate update.
2733 if (vif
->type
== NL80211_IFTYPE_P2P_DEVICE
) {
2736 mutex_lock(&mvm
->mutex
);
2737 ret
= iwl_mvm_mac_ctxt_changed(mvm
, vif
, false, NULL
);
2738 mutex_unlock(&mvm
->mutex
);
2744 static void iwl_mvm_mac_mgd_prepare_tx(struct ieee80211_hw
*hw
,
2745 struct ieee80211_vif
*vif
)
2747 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
2748 u32 duration
= min(IWL_MVM_TE_SESSION_PROTECTION_MAX_TIME_MS
,
2749 200 + vif
->bss_conf
.beacon_int
);
2750 u32 min_duration
= min(IWL_MVM_TE_SESSION_PROTECTION_MIN_TIME_MS
,
2751 100 + vif
->bss_conf
.beacon_int
);
2753 if (WARN_ON_ONCE(vif
->bss_conf
.assoc
))
2757 * iwl_mvm_protect_session() reads directly from the device
2758 * (the system time), so make sure it is available.
2760 if (iwl_mvm_ref_sync(mvm
, IWL_MVM_REF_PREPARE_TX
))
2763 mutex_lock(&mvm
->mutex
);
2764 /* Try really hard to protect the session and hear a beacon */
2765 iwl_mvm_protect_session(mvm
, vif
, duration
, min_duration
, 500, false);
2766 mutex_unlock(&mvm
->mutex
);
2768 iwl_mvm_unref(mvm
, IWL_MVM_REF_PREPARE_TX
);
2771 static int iwl_mvm_mac_sched_scan_start(struct ieee80211_hw
*hw
,
2772 struct ieee80211_vif
*vif
,
2773 struct cfg80211_sched_scan_request
*req
,
2774 struct ieee80211_scan_ies
*ies
)
2776 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
2779 if (!(mvm
->fw
->ucode_capa
.capa
[0] & IWL_UCODE_TLV_CAPA_UMAC_SCAN
)) {
2780 ret
= iwl_mvm_cancel_scan_wait_notif(mvm
, IWL_MVM_SCAN_OS
);
2785 mutex_lock(&mvm
->mutex
);
2787 if (iwl_mvm_is_lar_supported(mvm
) && !mvm
->lar_regdom_set
) {
2788 IWL_ERR(mvm
, "sched-scan while LAR regdomain is not set\n");
2793 if (!vif
->bss_conf
.idle
) {
2798 if (mvm
->scan_status
!= IWL_MVM_SCAN_NONE
) {
2803 ret
= iwl_mvm_scan_offload_start(mvm
, vif
, req
, ies
);
2805 mvm
->scan_status
= IWL_MVM_SCAN_NONE
;
2808 mutex_unlock(&mvm
->mutex
);
2812 static int iwl_mvm_mac_sched_scan_stop(struct ieee80211_hw
*hw
,
2813 struct ieee80211_vif
*vif
)
2815 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
2818 mutex_lock(&mvm
->mutex
);
2820 /* Due to a race condition, it's possible that mac80211 asks
2821 * us to stop a sched_scan when it's already stopped. This
2822 * can happen, for instance, if we stopped the scan ourselves,
2823 * called ieee80211_sched_scan_stopped() and the userspace called
2824 * stop sched scan scan before ieee80211_sched_scan_stopped_work()
2825 * could run. To handle this, simply return if the scan is
2828 /* FIXME: for now, we ignore this race for UMAC scans, since
2829 * they don't set the scan_status.
2831 if (mvm
->scan_status
!= IWL_MVM_SCAN_SCHED
&&
2832 !(mvm
->fw
->ucode_capa
.capa
[0] & IWL_UCODE_TLV_CAPA_UMAC_SCAN
)) {
2833 mutex_unlock(&mvm
->mutex
);
2837 ret
= iwl_mvm_scan_offload_stop(mvm
, false);
2838 mutex_unlock(&mvm
->mutex
);
2839 iwl_mvm_wait_for_async_handlers(mvm
);
2844 static int iwl_mvm_mac_set_key(struct ieee80211_hw
*hw
,
2845 enum set_key_cmd cmd
,
2846 struct ieee80211_vif
*vif
,
2847 struct ieee80211_sta
*sta
,
2848 struct ieee80211_key_conf
*key
)
2850 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
2853 if (iwlwifi_mod_params
.sw_crypto
) {
2854 IWL_DEBUG_MAC80211(mvm
, "leave - hwcrypto disabled\n");
2858 switch (key
->cipher
) {
2859 case WLAN_CIPHER_SUITE_TKIP
:
2860 key
->flags
|= IEEE80211_KEY_FLAG_GENERATE_MMIC
;
2862 case WLAN_CIPHER_SUITE_CCMP
:
2863 key
->flags
|= IEEE80211_KEY_FLAG_GENERATE_IV
;
2865 case WLAN_CIPHER_SUITE_AES_CMAC
:
2866 WARN_ON_ONCE(!(hw
->flags
& IEEE80211_HW_MFP_CAPABLE
));
2868 case WLAN_CIPHER_SUITE_WEP40
:
2869 case WLAN_CIPHER_SUITE_WEP104
:
2870 /* For non-client mode, only use WEP keys for TX as we probably
2871 * don't have a station yet anyway and would then have to keep
2872 * track of the keys, linking them to each of the clients/peers
2873 * as they appear. For now, don't do that, for performance WEP
2874 * offload doesn't really matter much, but we need it for some
2875 * other offload features in client mode.
2877 if (vif
->type
!= NL80211_IFTYPE_STATION
)
2881 /* currently FW supports only one optional cipher scheme */
2882 if (hw
->n_cipher_schemes
&&
2883 hw
->cipher_schemes
->cipher
== key
->cipher
)
2884 key
->flags
|= IEEE80211_KEY_FLAG_PUT_IV_SPACE
;
2889 mutex_lock(&mvm
->mutex
);
2893 if ((vif
->type
== NL80211_IFTYPE_ADHOC
||
2894 vif
->type
== NL80211_IFTYPE_AP
) && !sta
) {
2896 * GTK on AP interface is a TX-only key, return 0;
2897 * on IBSS they're per-station and because we're lazy
2898 * we don't support them for RX, so do the same.
2901 key
->hw_key_idx
= STA_KEY_IDX_INVALID
;
2905 IWL_DEBUG_MAC80211(mvm
, "set hwcrypto key\n");
2906 ret
= iwl_mvm_set_sta_key(mvm
, vif
, sta
, key
, false);
2908 IWL_WARN(mvm
, "set key failed\n");
2910 * can't add key for RX, but we don't need it
2911 * in the device for TX so still return 0
2913 key
->hw_key_idx
= STA_KEY_IDX_INVALID
;
2919 if (key
->hw_key_idx
== STA_KEY_IDX_INVALID
) {
2924 IWL_DEBUG_MAC80211(mvm
, "disable hwcrypto key\n");
2925 ret
= iwl_mvm_remove_sta_key(mvm
, vif
, sta
, key
);
2931 mutex_unlock(&mvm
->mutex
);
2935 static void iwl_mvm_mac_update_tkip_key(struct ieee80211_hw
*hw
,
2936 struct ieee80211_vif
*vif
,
2937 struct ieee80211_key_conf
*keyconf
,
2938 struct ieee80211_sta
*sta
,
2939 u32 iv32
, u16
*phase1key
)
2941 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
2943 if (keyconf
->hw_key_idx
== STA_KEY_IDX_INVALID
)
2946 iwl_mvm_update_tkip_key(mvm
, vif
, keyconf
, sta
, iv32
, phase1key
);
2950 static bool iwl_mvm_rx_aux_roc(struct iwl_notif_wait_data
*notif_wait
,
2951 struct iwl_rx_packet
*pkt
, void *data
)
2953 struct iwl_mvm
*mvm
=
2954 container_of(notif_wait
, struct iwl_mvm
, notif_wait
);
2955 struct iwl_hs20_roc_res
*resp
;
2956 int resp_len
= iwl_rx_packet_payload_len(pkt
);
2957 struct iwl_mvm_time_event_data
*te_data
= data
;
2959 if (WARN_ON(pkt
->hdr
.cmd
!= HOT_SPOT_CMD
))
2962 if (WARN_ON_ONCE(resp_len
!= sizeof(*resp
))) {
2963 IWL_ERR(mvm
, "Invalid HOT_SPOT_CMD response\n");
2967 resp
= (void *)pkt
->data
;
2970 "Aux ROC: Recieved response from ucode: status=%d uid=%d\n",
2971 resp
->status
, resp
->event_unique_id
);
2973 te_data
->uid
= le32_to_cpu(resp
->event_unique_id
);
2974 IWL_DEBUG_TE(mvm
, "TIME_EVENT_CMD response - UID = 0x%x\n",
2977 spin_lock_bh(&mvm
->time_event_lock
);
2978 list_add_tail(&te_data
->list
, &mvm
->aux_roc_te_list
);
2979 spin_unlock_bh(&mvm
->time_event_lock
);
2984 #define AUX_ROC_MAX_DELAY_ON_CHANNEL 5000
2985 static int iwl_mvm_send_aux_roc_cmd(struct iwl_mvm
*mvm
,
2986 struct ieee80211_channel
*channel
,
2987 struct ieee80211_vif
*vif
,
2990 int res
, time_reg
= DEVICE_SYSTEM_TIME_REG
;
2991 struct iwl_mvm_vif
*mvmvif
= iwl_mvm_vif_from_mac80211(vif
);
2992 struct iwl_mvm_time_event_data
*te_data
= &mvmvif
->hs_time_event_data
;
2993 static const u8 time_event_response
[] = { HOT_SPOT_CMD
};
2994 struct iwl_notification_wait wait_time_event
;
2995 struct iwl_hs20_roc_req aux_roc_req
= {
2996 .action
= cpu_to_le32(FW_CTXT_ACTION_ADD
),
2998 cpu_to_le32(FW_CMD_ID_AND_COLOR(MAC_INDEX_AUX
, 0)),
2999 .sta_id_and_color
= cpu_to_le32(mvm
->aux_sta
.sta_id
),
3000 /* Set the channel info data */
3001 .channel_info
.band
= (channel
->band
== IEEE80211_BAND_2GHZ
) ?
3002 PHY_BAND_24
: PHY_BAND_5
,
3003 .channel_info
.channel
= channel
->hw_value
,
3004 .channel_info
.width
= PHY_VHT_CHANNEL_MODE20
,
3005 /* Set the time and duration */
3006 .apply_time
= cpu_to_le32(iwl_read_prph(mvm
->trans
, time_reg
)),
3007 .apply_time_max_delay
=
3008 cpu_to_le32(MSEC_TO_TU(AUX_ROC_MAX_DELAY_ON_CHANNEL
)),
3009 .duration
= cpu_to_le32(MSEC_TO_TU(duration
)),
3012 /* Set the node address */
3013 memcpy(aux_roc_req
.node_addr
, vif
->addr
, ETH_ALEN
);
3015 lockdep_assert_held(&mvm
->mutex
);
3017 spin_lock_bh(&mvm
->time_event_lock
);
3019 if (WARN_ON(te_data
->id
== HOT_SPOT_CMD
)) {
3020 spin_unlock_bh(&mvm
->time_event_lock
);
3025 te_data
->duration
= duration
;
3026 te_data
->id
= HOT_SPOT_CMD
;
3028 spin_unlock_bh(&mvm
->time_event_lock
);
3031 * Use a notification wait, which really just processes the
3032 * command response and doesn't wait for anything, in order
3033 * to be able to process the response and get the UID inside
3034 * the RX path. Using CMD_WANT_SKB doesn't work because it
3035 * stores the buffer and then wakes up this thread, by which
3036 * time another notification (that the time event started)
3037 * might already be processed unsuccessfully.
3039 iwl_init_notification_wait(&mvm
->notif_wait
, &wait_time_event
,
3040 time_event_response
,
3041 ARRAY_SIZE(time_event_response
),
3042 iwl_mvm_rx_aux_roc
, te_data
);
3044 res
= iwl_mvm_send_cmd_pdu(mvm
, HOT_SPOT_CMD
, 0, sizeof(aux_roc_req
),
3048 IWL_ERR(mvm
, "Couldn't send HOT_SPOT_CMD: %d\n", res
);
3049 iwl_remove_notification(&mvm
->notif_wait
, &wait_time_event
);
3053 /* No need to wait for anything, so just pass 1 (0 isn't valid) */
3054 res
= iwl_wait_notification(&mvm
->notif_wait
, &wait_time_event
, 1);
3055 /* should never fail */
3060 spin_lock_bh(&mvm
->time_event_lock
);
3061 iwl_mvm_te_clear_data(mvm
, te_data
);
3062 spin_unlock_bh(&mvm
->time_event_lock
);
3068 static int iwl_mvm_roc(struct ieee80211_hw
*hw
,
3069 struct ieee80211_vif
*vif
,
3070 struct ieee80211_channel
*channel
,
3072 enum ieee80211_roc_type type
)
3074 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
3075 struct iwl_mvm_vif
*mvmvif
= iwl_mvm_vif_from_mac80211(vif
);
3076 struct cfg80211_chan_def chandef
;
3077 struct iwl_mvm_phy_ctxt
*phy_ctxt
;
3080 IWL_DEBUG_MAC80211(mvm
, "enter (%d, %d, %d)\n", channel
->hw_value
,
3083 flush_work(&mvm
->roc_done_wk
);
3085 mutex_lock(&mvm
->mutex
);
3087 switch (vif
->type
) {
3088 case NL80211_IFTYPE_STATION
:
3089 if (mvm
->fw
->ucode_capa
.capa
[0] &
3090 IWL_UCODE_TLV_CAPA_HOTSPOT_SUPPORT
) {
3091 /* Use aux roc framework (HS20) */
3092 ret
= iwl_mvm_send_aux_roc_cmd(mvm
, channel
,
3096 IWL_ERR(mvm
, "hotspot not supported\n");
3099 case NL80211_IFTYPE_P2P_DEVICE
:
3103 IWL_ERR(mvm
, "vif isn't P2P_DEVICE: %d\n", vif
->type
);
3108 for (i
= 0; i
< NUM_PHY_CTX
; i
++) {
3109 phy_ctxt
= &mvm
->phy_ctxts
[i
];
3110 if (phy_ctxt
->ref
== 0 || mvmvif
->phy_ctxt
== phy_ctxt
)
3113 if (phy_ctxt
->ref
&& channel
== phy_ctxt
->channel
) {
3115 * Unbind the P2P_DEVICE from the current PHY context,
3116 * and if the PHY context is not used remove it.
3118 ret
= iwl_mvm_binding_remove_vif(mvm
, vif
);
3119 if (WARN(ret
, "Failed unbinding P2P_DEVICE\n"))
3122 iwl_mvm_phy_ctxt_unref(mvm
, mvmvif
->phy_ctxt
);
3124 /* Bind the P2P_DEVICE to the current PHY Context */
3125 mvmvif
->phy_ctxt
= phy_ctxt
;
3127 ret
= iwl_mvm_binding_add_vif(mvm
, vif
);
3128 if (WARN(ret
, "Failed binding P2P_DEVICE\n"))
3131 iwl_mvm_phy_ctxt_ref(mvm
, mvmvif
->phy_ctxt
);
3132 goto schedule_time_event
;
3136 /* Need to update the PHY context only if the ROC channel changed */
3137 if (channel
== mvmvif
->phy_ctxt
->channel
)
3138 goto schedule_time_event
;
3140 cfg80211_chandef_create(&chandef
, channel
, NL80211_CHAN_NO_HT
);
3143 * Change the PHY context configuration as it is currently referenced
3144 * only by the P2P Device MAC
3146 if (mvmvif
->phy_ctxt
->ref
== 1) {
3147 ret
= iwl_mvm_phy_ctxt_changed(mvm
, mvmvif
->phy_ctxt
,
3153 * The PHY context is shared with other MACs. Need to remove the
3154 * P2P Device from the binding, allocate an new PHY context and
3155 * create a new binding
3157 phy_ctxt
= iwl_mvm_get_free_phy_ctxt(mvm
);
3163 ret
= iwl_mvm_phy_ctxt_changed(mvm
, phy_ctxt
, &chandef
,
3166 IWL_ERR(mvm
, "Failed to change PHY context\n");
3170 /* Unbind the P2P_DEVICE from the current PHY context */
3171 ret
= iwl_mvm_binding_remove_vif(mvm
, vif
);
3172 if (WARN(ret
, "Failed unbinding P2P_DEVICE\n"))
3175 iwl_mvm_phy_ctxt_unref(mvm
, mvmvif
->phy_ctxt
);
3177 /* Bind the P2P_DEVICE to the new allocated PHY context */
3178 mvmvif
->phy_ctxt
= phy_ctxt
;
3180 ret
= iwl_mvm_binding_add_vif(mvm
, vif
);
3181 if (WARN(ret
, "Failed binding P2P_DEVICE\n"))
3184 iwl_mvm_phy_ctxt_ref(mvm
, mvmvif
->phy_ctxt
);
3187 schedule_time_event
:
3188 /* Schedule the time events */
3189 ret
= iwl_mvm_start_p2p_roc(mvm
, vif
, duration
, type
);
3192 mutex_unlock(&mvm
->mutex
);
3193 IWL_DEBUG_MAC80211(mvm
, "leave\n");
3197 static int iwl_mvm_cancel_roc(struct ieee80211_hw
*hw
)
3199 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
3201 IWL_DEBUG_MAC80211(mvm
, "enter\n");
3203 mutex_lock(&mvm
->mutex
);
3204 iwl_mvm_stop_roc(mvm
);
3205 mutex_unlock(&mvm
->mutex
);
3207 IWL_DEBUG_MAC80211(mvm
, "leave\n");
3211 static int __iwl_mvm_add_chanctx(struct iwl_mvm
*mvm
,
3212 struct ieee80211_chanctx_conf
*ctx
)
3214 u16
*phy_ctxt_id
= (u16
*)ctx
->drv_priv
;
3215 struct iwl_mvm_phy_ctxt
*phy_ctxt
;
3218 lockdep_assert_held(&mvm
->mutex
);
3220 IWL_DEBUG_MAC80211(mvm
, "Add channel context\n");
3222 phy_ctxt
= iwl_mvm_get_free_phy_ctxt(mvm
);
3228 ret
= iwl_mvm_phy_ctxt_changed(mvm
, phy_ctxt
, &ctx
->min_def
,
3229 ctx
->rx_chains_static
,
3230 ctx
->rx_chains_dynamic
);
3232 IWL_ERR(mvm
, "Failed to add PHY context\n");
3236 iwl_mvm_phy_ctxt_ref(mvm
, phy_ctxt
);
3237 *phy_ctxt_id
= phy_ctxt
->id
;
3242 static int iwl_mvm_add_chanctx(struct ieee80211_hw
*hw
,
3243 struct ieee80211_chanctx_conf
*ctx
)
3245 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
3248 mutex_lock(&mvm
->mutex
);
3249 ret
= __iwl_mvm_add_chanctx(mvm
, ctx
);
3250 mutex_unlock(&mvm
->mutex
);
3255 static void __iwl_mvm_remove_chanctx(struct iwl_mvm
*mvm
,
3256 struct ieee80211_chanctx_conf
*ctx
)
3258 u16
*phy_ctxt_id
= (u16
*)ctx
->drv_priv
;
3259 struct iwl_mvm_phy_ctxt
*phy_ctxt
= &mvm
->phy_ctxts
[*phy_ctxt_id
];
3261 lockdep_assert_held(&mvm
->mutex
);
3263 iwl_mvm_phy_ctxt_unref(mvm
, phy_ctxt
);
3266 static void iwl_mvm_remove_chanctx(struct ieee80211_hw
*hw
,
3267 struct ieee80211_chanctx_conf
*ctx
)
3269 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
3271 mutex_lock(&mvm
->mutex
);
3272 __iwl_mvm_remove_chanctx(mvm
, ctx
);
3273 mutex_unlock(&mvm
->mutex
);
3276 static void iwl_mvm_change_chanctx(struct ieee80211_hw
*hw
,
3277 struct ieee80211_chanctx_conf
*ctx
,
3280 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
3281 u16
*phy_ctxt_id
= (u16
*)ctx
->drv_priv
;
3282 struct iwl_mvm_phy_ctxt
*phy_ctxt
= &mvm
->phy_ctxts
[*phy_ctxt_id
];
3284 if (WARN_ONCE((phy_ctxt
->ref
> 1) &&
3285 (changed
& ~(IEEE80211_CHANCTX_CHANGE_WIDTH
|
3286 IEEE80211_CHANCTX_CHANGE_RX_CHAINS
|
3287 IEEE80211_CHANCTX_CHANGE_RADAR
|
3288 IEEE80211_CHANCTX_CHANGE_MIN_WIDTH
)),
3289 "Cannot change PHY. Ref=%d, changed=0x%X\n",
3290 phy_ctxt
->ref
, changed
))
3293 mutex_lock(&mvm
->mutex
);
3294 iwl_mvm_bt_coex_vif_change(mvm
);
3295 iwl_mvm_phy_ctxt_changed(mvm
, phy_ctxt
, &ctx
->min_def
,
3296 ctx
->rx_chains_static
,
3297 ctx
->rx_chains_dynamic
);
3298 mutex_unlock(&mvm
->mutex
);
3301 static int __iwl_mvm_assign_vif_chanctx(struct iwl_mvm
*mvm
,
3302 struct ieee80211_vif
*vif
,
3303 struct ieee80211_chanctx_conf
*ctx
,
3304 bool switching_chanctx
)
3306 u16
*phy_ctxt_id
= (u16
*)ctx
->drv_priv
;
3307 struct iwl_mvm_phy_ctxt
*phy_ctxt
= &mvm
->phy_ctxts
[*phy_ctxt_id
];
3308 struct iwl_mvm_vif
*mvmvif
= iwl_mvm_vif_from_mac80211(vif
);
3311 lockdep_assert_held(&mvm
->mutex
);
3313 mvmvif
->phy_ctxt
= phy_ctxt
;
3315 switch (vif
->type
) {
3316 case NL80211_IFTYPE_AP
:
3317 /* only needed if we're switching chanctx (i.e. during CSA) */
3318 if (switching_chanctx
) {
3319 mvmvif
->ap_ibss_active
= true;
3322 case NL80211_IFTYPE_ADHOC
:
3324 * The AP binding flow is handled as part of the start_ap flow
3325 * (in bss_info_changed), similarly for IBSS.
3329 case NL80211_IFTYPE_STATION
:
3331 case NL80211_IFTYPE_MONITOR
:
3332 /* always disable PS when a monitor interface is active */
3333 mvmvif
->ps_disabled
= true;
3340 ret
= iwl_mvm_binding_add_vif(mvm
, vif
);
3345 * Power state must be updated before quotas,
3346 * otherwise fw will complain.
3348 iwl_mvm_power_update_mac(mvm
);
3350 /* Setting the quota at this stage is only required for monitor
3351 * interfaces. For the other types, the bss_info changed flow
3352 * will handle quota settings.
3354 if (vif
->type
== NL80211_IFTYPE_MONITOR
) {
3355 mvmvif
->monitor_active
= true;
3356 ret
= iwl_mvm_update_quotas(mvm
, false, NULL
);
3358 goto out_remove_binding
;
3361 /* Handle binding during CSA */
3362 if (vif
->type
== NL80211_IFTYPE_AP
) {
3363 iwl_mvm_update_quotas(mvm
, false, NULL
);
3364 iwl_mvm_mac_ctxt_changed(mvm
, vif
, false, NULL
);
3367 if (switching_chanctx
&& vif
->type
== NL80211_IFTYPE_STATION
) {
3368 u32 duration
= 2 * vif
->bss_conf
.beacon_int
;
3370 /* iwl_mvm_protect_session() reads directly from the
3371 * device (the system time), so make sure it is
3374 ret
= iwl_mvm_ref_sync(mvm
, IWL_MVM_REF_PROTECT_CSA
);
3376 goto out_remove_binding
;
3378 /* Protect the session to make sure we hear the first
3379 * beacon on the new channel.
3381 iwl_mvm_protect_session(mvm
, vif
, duration
, duration
,
3382 vif
->bss_conf
.beacon_int
/ 2,
3385 iwl_mvm_unref(mvm
, IWL_MVM_REF_PROTECT_CSA
);
3387 iwl_mvm_update_quotas(mvm
, false, NULL
);
3393 iwl_mvm_binding_remove_vif(mvm
, vif
);
3394 iwl_mvm_power_update_mac(mvm
);
3397 mvmvif
->phy_ctxt
= NULL
;
3400 static int iwl_mvm_assign_vif_chanctx(struct ieee80211_hw
*hw
,
3401 struct ieee80211_vif
*vif
,
3402 struct ieee80211_chanctx_conf
*ctx
)
3404 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
3407 mutex_lock(&mvm
->mutex
);
3408 ret
= __iwl_mvm_assign_vif_chanctx(mvm
, vif
, ctx
, false);
3409 mutex_unlock(&mvm
->mutex
);
3414 static void __iwl_mvm_unassign_vif_chanctx(struct iwl_mvm
*mvm
,
3415 struct ieee80211_vif
*vif
,
3416 struct ieee80211_chanctx_conf
*ctx
,
3417 bool switching_chanctx
)
3419 struct iwl_mvm_vif
*mvmvif
= iwl_mvm_vif_from_mac80211(vif
);
3420 struct ieee80211_vif
*disabled_vif
= NULL
;
3422 lockdep_assert_held(&mvm
->mutex
);
3424 iwl_mvm_remove_time_event(mvm
, mvmvif
, &mvmvif
->time_event_data
);
3426 switch (vif
->type
) {
3427 case NL80211_IFTYPE_ADHOC
:
3429 case NL80211_IFTYPE_MONITOR
:
3430 mvmvif
->monitor_active
= false;
3431 mvmvif
->ps_disabled
= false;
3433 case NL80211_IFTYPE_AP
:
3434 /* This part is triggered only during CSA */
3435 if (!switching_chanctx
|| !mvmvif
->ap_ibss_active
)
3438 mvmvif
->csa_countdown
= false;
3440 /* Set CS bit on all the stations */
3441 iwl_mvm_modify_all_sta_disable_tx(mvm
, mvmvif
, true);
3443 /* Save blocked iface, the timeout is set on the next beacon */
3444 rcu_assign_pointer(mvm
->csa_tx_blocked_vif
, vif
);
3446 mvmvif
->ap_ibss_active
= false;
3448 case NL80211_IFTYPE_STATION
:
3449 if (!switching_chanctx
)
3454 iwl_mvm_mac_ctxt_changed(mvm
, vif
, true, NULL
);
3460 iwl_mvm_update_quotas(mvm
, false, disabled_vif
);
3461 iwl_mvm_binding_remove_vif(mvm
, vif
);
3464 mvmvif
->phy_ctxt
= NULL
;
3465 iwl_mvm_power_update_mac(mvm
);
3468 static void iwl_mvm_unassign_vif_chanctx(struct ieee80211_hw
*hw
,
3469 struct ieee80211_vif
*vif
,
3470 struct ieee80211_chanctx_conf
*ctx
)
3472 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
3474 mutex_lock(&mvm
->mutex
);
3475 __iwl_mvm_unassign_vif_chanctx(mvm
, vif
, ctx
, false);
3476 mutex_unlock(&mvm
->mutex
);
3480 iwl_mvm_switch_vif_chanctx_swap(struct iwl_mvm
*mvm
,
3481 struct ieee80211_vif_chanctx_switch
*vifs
)
3485 mutex_lock(&mvm
->mutex
);
3486 __iwl_mvm_unassign_vif_chanctx(mvm
, vifs
[0].vif
, vifs
[0].old_ctx
, true);
3487 __iwl_mvm_remove_chanctx(mvm
, vifs
[0].old_ctx
);
3489 ret
= __iwl_mvm_add_chanctx(mvm
, vifs
[0].new_ctx
);
3491 IWL_ERR(mvm
, "failed to add new_ctx during channel switch\n");
3495 ret
= __iwl_mvm_assign_vif_chanctx(mvm
, vifs
[0].vif
, vifs
[0].new_ctx
,
3499 "failed to assign new_ctx during channel switch\n");
3503 /* we don't support TDLS during DCM - can be caused by channel switch */
3504 if (iwl_mvm_phy_ctx_count(mvm
) > 1)
3505 iwl_mvm_teardown_tdls_peers(mvm
);
3510 __iwl_mvm_remove_chanctx(mvm
, vifs
[0].new_ctx
);
3513 if (__iwl_mvm_add_chanctx(mvm
, vifs
[0].old_ctx
)) {
3514 IWL_ERR(mvm
, "failed to add old_ctx back after failure.\n");
3518 if (__iwl_mvm_assign_vif_chanctx(mvm
, vifs
[0].vif
, vifs
[0].old_ctx
,
3520 IWL_ERR(mvm
, "failed to reassign old_ctx after failure.\n");
3527 /* things keep failing, better restart the hw */
3528 iwl_mvm_nic_restart(mvm
, false);
3531 mutex_unlock(&mvm
->mutex
);
3537 iwl_mvm_switch_vif_chanctx_reassign(struct iwl_mvm
*mvm
,
3538 struct ieee80211_vif_chanctx_switch
*vifs
)
3542 mutex_lock(&mvm
->mutex
);
3543 __iwl_mvm_unassign_vif_chanctx(mvm
, vifs
[0].vif
, vifs
[0].old_ctx
, true);
3545 ret
= __iwl_mvm_assign_vif_chanctx(mvm
, vifs
[0].vif
, vifs
[0].new_ctx
,
3549 "failed to assign new_ctx during channel switch\n");
3556 if (__iwl_mvm_assign_vif_chanctx(mvm
, vifs
[0].vif
, vifs
[0].old_ctx
,
3558 IWL_ERR(mvm
, "failed to reassign old_ctx after failure.\n");
3565 /* things keep failing, better restart the hw */
3566 iwl_mvm_nic_restart(mvm
, false);
3569 mutex_unlock(&mvm
->mutex
);
3574 static int iwl_mvm_switch_vif_chanctx(struct ieee80211_hw
*hw
,
3575 struct ieee80211_vif_chanctx_switch
*vifs
,
3577 enum ieee80211_chanctx_switch_mode mode
)
3579 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
3582 /* we only support a single-vif right now */
3587 case CHANCTX_SWMODE_SWAP_CONTEXTS
:
3588 ret
= iwl_mvm_switch_vif_chanctx_swap(mvm
, vifs
);
3590 case CHANCTX_SWMODE_REASSIGN_VIF
:
3591 ret
= iwl_mvm_switch_vif_chanctx_reassign(mvm
, vifs
);
3601 static int iwl_mvm_set_tim(struct ieee80211_hw
*hw
,
3602 struct ieee80211_sta
*sta
,
3605 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
3606 struct iwl_mvm_sta
*mvm_sta
= iwl_mvm_sta_from_mac80211(sta
);
3608 if (!mvm_sta
|| !mvm_sta
->vif
) {
3609 IWL_ERR(mvm
, "Station is not associated to a vif\n");
3613 return iwl_mvm_mac_ctxt_beacon_changed(mvm
, mvm_sta
->vif
);
3616 #ifdef CONFIG_NL80211_TESTMODE
3617 static const struct nla_policy iwl_mvm_tm_policy
[IWL_MVM_TM_ATTR_MAX
+ 1] = {
3618 [IWL_MVM_TM_ATTR_CMD
] = { .type
= NLA_U32
},
3619 [IWL_MVM_TM_ATTR_NOA_DURATION
] = { .type
= NLA_U32
},
3620 [IWL_MVM_TM_ATTR_BEACON_FILTER_STATE
] = { .type
= NLA_U32
},
3623 static int __iwl_mvm_mac_testmode_cmd(struct iwl_mvm
*mvm
,
3624 struct ieee80211_vif
*vif
,
3625 void *data
, int len
)
3627 struct nlattr
*tb
[IWL_MVM_TM_ATTR_MAX
+ 1];
3631 err
= nla_parse(tb
, IWL_MVM_TM_ATTR_MAX
, data
, len
, iwl_mvm_tm_policy
);
3635 if (!tb
[IWL_MVM_TM_ATTR_CMD
])
3638 switch (nla_get_u32(tb
[IWL_MVM_TM_ATTR_CMD
])) {
3639 case IWL_MVM_TM_CMD_SET_NOA
:
3640 if (!vif
|| vif
->type
!= NL80211_IFTYPE_AP
|| !vif
->p2p
||
3641 !vif
->bss_conf
.enable_beacon
||
3642 !tb
[IWL_MVM_TM_ATTR_NOA_DURATION
])
3645 noa_duration
= nla_get_u32(tb
[IWL_MVM_TM_ATTR_NOA_DURATION
]);
3646 if (noa_duration
>= vif
->bss_conf
.beacon_int
)
3649 mvm
->noa_duration
= noa_duration
;
3652 return iwl_mvm_update_quotas(mvm
, false, NULL
);
3653 case IWL_MVM_TM_CMD_SET_BEACON_FILTER
:
3654 /* must be associated client vif - ignore authorized */
3655 if (!vif
|| vif
->type
!= NL80211_IFTYPE_STATION
||
3656 !vif
->bss_conf
.assoc
|| !vif
->bss_conf
.dtim_period
||
3657 !tb
[IWL_MVM_TM_ATTR_BEACON_FILTER_STATE
])
3660 if (nla_get_u32(tb
[IWL_MVM_TM_ATTR_BEACON_FILTER_STATE
]))
3661 return iwl_mvm_enable_beacon_filter(mvm
, vif
, 0);
3662 return iwl_mvm_disable_beacon_filter(mvm
, vif
, 0);
3668 static int iwl_mvm_mac_testmode_cmd(struct ieee80211_hw
*hw
,
3669 struct ieee80211_vif
*vif
,
3670 void *data
, int len
)
3672 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
3675 mutex_lock(&mvm
->mutex
);
3676 err
= __iwl_mvm_mac_testmode_cmd(mvm
, vif
, data
, len
);
3677 mutex_unlock(&mvm
->mutex
);
3683 static void iwl_mvm_channel_switch(struct ieee80211_hw
*hw
,
3684 struct ieee80211_vif
*vif
,
3685 struct ieee80211_channel_switch
*chsw
)
3687 /* By implementing this operation, we prevent mac80211 from
3688 * starting its own channel switch timer, so that we can call
3689 * ieee80211_chswitch_done() ourselves at the right time
3690 * (which is when the absence time event starts).
3693 IWL_DEBUG_MAC80211(IWL_MAC80211_GET_MVM(hw
),
3694 "dummy channel switch op\n");
3697 static int iwl_mvm_pre_channel_switch(struct ieee80211_hw
*hw
,
3698 struct ieee80211_vif
*vif
,
3699 struct ieee80211_channel_switch
*chsw
)
3701 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
3702 struct ieee80211_vif
*csa_vif
;
3703 struct iwl_mvm_vif
*mvmvif
= iwl_mvm_vif_from_mac80211(vif
);
3707 mutex_lock(&mvm
->mutex
);
3709 mvmvif
->csa_failed
= false;
3711 IWL_DEBUG_MAC80211(mvm
, "pre CSA to freq %d\n",
3712 chsw
->chandef
.center_freq1
);
3714 iwl_fw_dbg_trigger_simple_stop(mvm
, vif
, FW_DBG_TRIGGER_CHANNEL_SWITCH
);
3716 switch (vif
->type
) {
3717 case NL80211_IFTYPE_AP
:
3719 rcu_dereference_protected(mvm
->csa_vif
,
3720 lockdep_is_held(&mvm
->mutex
));
3721 if (WARN_ONCE(csa_vif
&& csa_vif
->csa_active
,
3722 "Another CSA is already in progress")) {
3727 rcu_assign_pointer(mvm
->csa_vif
, vif
);
3729 if (WARN_ONCE(mvmvif
->csa_countdown
,
3730 "Previous CSA countdown didn't complete")) {
3736 case NL80211_IFTYPE_STATION
:
3737 /* Schedule the time event to a bit before beacon 1,
3738 * to make sure we're in the new channel when the
3741 apply_time
= chsw
->device_timestamp
+
3742 ((vif
->bss_conf
.beacon_int
* (chsw
->count
- 1) -
3743 IWL_MVM_CHANNEL_SWITCH_TIME_CLIENT
) * 1024);
3746 iwl_mvm_csa_client_absent(mvm
, vif
);
3748 iwl_mvm_schedule_csa_period(mvm
, vif
, vif
->bss_conf
.beacon_int
,
3750 if (mvmvif
->bf_data
.bf_enabled
) {
3751 ret
= iwl_mvm_disable_beacon_filter(mvm
, vif
, 0);
3761 mvmvif
->ps_disabled
= true;
3763 ret
= iwl_mvm_power_update_ps(mvm
);
3767 /* we won't be on this channel any longer */
3768 iwl_mvm_teardown_tdls_peers(mvm
);
3771 mutex_unlock(&mvm
->mutex
);
3776 static int iwl_mvm_post_channel_switch(struct ieee80211_hw
*hw
,
3777 struct ieee80211_vif
*vif
)
3779 struct iwl_mvm_vif
*mvmvif
= iwl_mvm_vif_from_mac80211(vif
);
3780 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
3783 mutex_lock(&mvm
->mutex
);
3785 if (mvmvif
->csa_failed
) {
3786 mvmvif
->csa_failed
= false;
3791 if (vif
->type
== NL80211_IFTYPE_STATION
) {
3792 struct iwl_mvm_sta
*mvmsta
;
3794 mvmsta
= iwl_mvm_sta_from_staid_protected(mvm
,
3797 if (WARN_ON(!mvmsta
)) {
3802 iwl_mvm_sta_modify_disable_tx(mvm
, mvmsta
, false);
3804 iwl_mvm_mac_ctxt_changed(mvm
, vif
, false, NULL
);
3806 ret
= iwl_mvm_enable_beacon_filter(mvm
, vif
, 0);
3810 iwl_mvm_stop_session_protection(mvm
, vif
);
3813 mvmvif
->ps_disabled
= false;
3815 ret
= iwl_mvm_power_update_ps(mvm
);
3818 mutex_unlock(&mvm
->mutex
);
3823 static void iwl_mvm_mac_flush(struct ieee80211_hw
*hw
,
3824 struct ieee80211_vif
*vif
, u32 queues
, bool drop
)
3826 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
3827 struct iwl_mvm_vif
*mvmvif
;
3828 struct iwl_mvm_sta
*mvmsta
;
3829 struct ieee80211_sta
*sta
;
3833 if (!vif
|| vif
->type
!= NL80211_IFTYPE_STATION
)
3836 mutex_lock(&mvm
->mutex
);
3837 mvmvif
= iwl_mvm_vif_from_mac80211(vif
);
3839 /* flush the AP-station and all TDLS peers */
3840 for (i
= 0; i
< IWL_MVM_STATION_COUNT
; i
++) {
3841 sta
= rcu_dereference_protected(mvm
->fw_id_to_mac_id
[i
],
3842 lockdep_is_held(&mvm
->mutex
));
3843 if (IS_ERR_OR_NULL(sta
))
3846 mvmsta
= iwl_mvm_sta_from_mac80211(sta
);
3847 if (mvmsta
->vif
!= vif
)
3850 /* make sure only TDLS peers or the AP are flushed */
3851 WARN_ON(i
!= mvmvif
->ap_sta_id
&& !sta
->tdls
);
3853 msk
|= mvmsta
->tfd_queue_msk
;
3857 if (iwl_mvm_flush_tx_path(mvm
, msk
, true))
3858 IWL_ERR(mvm
, "flush request fail\n");
3859 mutex_unlock(&mvm
->mutex
);
3861 mutex_unlock(&mvm
->mutex
);
3863 /* this can take a while, and we may need/want other operations
3864 * to succeed while doing this, so do it without the mutex held
3866 iwl_trans_wait_tx_queue_empty(mvm
->trans
, msk
);
3870 static int iwl_mvm_mac_get_survey(struct ieee80211_hw
*hw
, int idx
,
3871 struct survey_info
*survey
)
3873 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
3876 memset(survey
, 0, sizeof(*survey
));
3878 /* only support global statistics right now */
3882 if (!(mvm
->fw
->ucode_capa
.capa
[0] &
3883 IWL_UCODE_TLV_CAPA_RADIO_BEACON_STATS
))
3886 mutex_lock(&mvm
->mutex
);
3888 if (mvm
->ucode_loaded
) {
3889 ret
= iwl_mvm_request_statistics(mvm
, false);
3894 survey
->filled
= SURVEY_INFO_TIME
|
3895 SURVEY_INFO_TIME_RX
|
3896 SURVEY_INFO_TIME_TX
|
3897 SURVEY_INFO_TIME_SCAN
;
3898 survey
->time
= mvm
->accu_radio_stats
.on_time_rf
+
3899 mvm
->radio_stats
.on_time_rf
;
3900 do_div(survey
->time
, USEC_PER_MSEC
);
3902 survey
->time_rx
= mvm
->accu_radio_stats
.rx_time
+
3903 mvm
->radio_stats
.rx_time
;
3904 do_div(survey
->time_rx
, USEC_PER_MSEC
);
3906 survey
->time_tx
= mvm
->accu_radio_stats
.tx_time
+
3907 mvm
->radio_stats
.tx_time
;
3908 do_div(survey
->time_tx
, USEC_PER_MSEC
);
3910 survey
->time_scan
= mvm
->accu_radio_stats
.on_time_scan
+
3911 mvm
->radio_stats
.on_time_scan
;
3912 do_div(survey
->time_scan
, USEC_PER_MSEC
);
3916 mutex_unlock(&mvm
->mutex
);
3920 static void iwl_mvm_mac_sta_statistics(struct ieee80211_hw
*hw
,
3921 struct ieee80211_vif
*vif
,
3922 struct ieee80211_sta
*sta
,
3923 struct station_info
*sinfo
)
3925 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
3926 struct iwl_mvm_vif
*mvmvif
= iwl_mvm_vif_from_mac80211(vif
);
3927 struct iwl_mvm_sta
*mvmsta
= iwl_mvm_sta_from_mac80211(sta
);
3929 if (!(mvm
->fw
->ucode_capa
.capa
[0] &
3930 IWL_UCODE_TLV_CAPA_RADIO_BEACON_STATS
))
3933 /* if beacon filtering isn't on mac80211 does it anyway */
3934 if (!(vif
->driver_flags
& IEEE80211_VIF_BEACON_FILTER
))
3937 if (!vif
->bss_conf
.assoc
)
3940 mutex_lock(&mvm
->mutex
);
3942 if (mvmvif
->ap_sta_id
!= mvmsta
->sta_id
)
3945 if (iwl_mvm_request_statistics(mvm
, false))
3948 sinfo
->rx_beacon
= mvmvif
->beacon_stats
.num_beacons
+
3949 mvmvif
->beacon_stats
.accu_num_beacons
;
3950 sinfo
->filled
|= BIT(NL80211_STA_INFO_BEACON_RX
);
3951 if (mvmvif
->beacon_stats
.avg_signal
) {
3952 /* firmware only reports a value after RXing a few beacons */
3953 sinfo
->rx_beacon_signal_avg
= mvmvif
->beacon_stats
.avg_signal
;
3954 sinfo
->filled
|= BIT(NL80211_STA_INFO_BEACON_SIGNAL_AVG
);
3957 mutex_unlock(&mvm
->mutex
);
3960 static void iwl_mvm_mac_event_callback(struct ieee80211_hw
*hw
,
3961 struct ieee80211_vif
*vif
,
3962 const struct ieee80211_event
*event
)
3964 #define CHECK_MLME_TRIGGER(_mvm, _trig, _buf, _cnt, _fmt...) \
3966 if ((_cnt) && --(_cnt)) \
3968 iwl_mvm_fw_dbg_collect_trig(_mvm, _trig, _fmt);\
3971 struct iwl_mvm
*mvm
= IWL_MAC80211_GET_MVM(hw
);
3972 struct iwl_fw_dbg_trigger_tlv
*trig
;
3973 struct iwl_fw_dbg_trigger_mlme
*trig_mlme
;
3975 if (!iwl_fw_dbg_trigger_enabled(mvm
->fw
, FW_DBG_TRIGGER_MLME
))
3978 if (event
->u
.mlme
.status
== MLME_SUCCESS
)
3981 trig
= iwl_fw_dbg_get_trigger(mvm
->fw
, FW_DBG_TRIGGER_MLME
);
3982 trig_mlme
= (void *)trig
->data
;
3983 if (!iwl_fw_dbg_trigger_check_stop(mvm
, vif
, trig
))
3986 if (event
->u
.mlme
.data
== ASSOC_EVENT
) {
3987 if (event
->u
.mlme
.status
== MLME_DENIED
)
3988 CHECK_MLME_TRIGGER(mvm
, trig
, buf
,
3989 trig_mlme
->stop_assoc_denied
,
3990 "DENIED ASSOC: reason %d",
3991 event
->u
.mlme
.reason
);
3992 else if (event
->u
.mlme
.status
== MLME_TIMEOUT
)
3993 CHECK_MLME_TRIGGER(mvm
, trig
, buf
,
3994 trig_mlme
->stop_assoc_timeout
,
3996 } else if (event
->u
.mlme
.data
== AUTH_EVENT
) {
3997 if (event
->u
.mlme
.status
== MLME_DENIED
)
3998 CHECK_MLME_TRIGGER(mvm
, trig
, buf
,
3999 trig_mlme
->stop_auth_denied
,
4000 "DENIED AUTH: reason %d",
4001 event
->u
.mlme
.reason
);
4002 else if (event
->u
.mlme
.status
== MLME_TIMEOUT
)
4003 CHECK_MLME_TRIGGER(mvm
, trig
, buf
,
4004 trig_mlme
->stop_auth_timeout
,
4006 } else if (event
->u
.mlme
.data
== DEAUTH_RX_EVENT
) {
4007 CHECK_MLME_TRIGGER(mvm
, trig
, buf
,
4008 trig_mlme
->stop_rx_deauth
,
4009 "DEAUTH RX %d", event
->u
.mlme
.reason
);
4010 } else if (event
->u
.mlme
.data
== DEAUTH_TX_EVENT
) {
4011 CHECK_MLME_TRIGGER(mvm
, trig
, buf
,
4012 trig_mlme
->stop_tx_deauth
,
4013 "DEAUTH TX %d", event
->u
.mlme
.reason
);
4015 #undef CHECK_MLME_TRIGGER
4018 const struct ieee80211_ops iwl_mvm_hw_ops
= {
4019 .tx
= iwl_mvm_mac_tx
,
4020 .ampdu_action
= iwl_mvm_mac_ampdu_action
,
4021 .start
= iwl_mvm_mac_start
,
4022 .reconfig_complete
= iwl_mvm_mac_reconfig_complete
,
4023 .stop
= iwl_mvm_mac_stop
,
4024 .add_interface
= iwl_mvm_mac_add_interface
,
4025 .remove_interface
= iwl_mvm_mac_remove_interface
,
4026 .config
= iwl_mvm_mac_config
,
4027 .prepare_multicast
= iwl_mvm_prepare_multicast
,
4028 .configure_filter
= iwl_mvm_configure_filter
,
4029 .bss_info_changed
= iwl_mvm_bss_info_changed
,
4030 .hw_scan
= iwl_mvm_mac_hw_scan
,
4031 .cancel_hw_scan
= iwl_mvm_mac_cancel_hw_scan
,
4032 .sta_pre_rcu_remove
= iwl_mvm_sta_pre_rcu_remove
,
4033 .sta_state
= iwl_mvm_mac_sta_state
,
4034 .sta_notify
= iwl_mvm_mac_sta_notify
,
4035 .allow_buffered_frames
= iwl_mvm_mac_allow_buffered_frames
,
4036 .release_buffered_frames
= iwl_mvm_mac_release_buffered_frames
,
4037 .set_rts_threshold
= iwl_mvm_mac_set_rts_threshold
,
4038 .sta_rc_update
= iwl_mvm_sta_rc_update
,
4039 .conf_tx
= iwl_mvm_mac_conf_tx
,
4040 .mgd_prepare_tx
= iwl_mvm_mac_mgd_prepare_tx
,
4041 .mgd_protect_tdls_discover
= iwl_mvm_mac_mgd_protect_tdls_discover
,
4042 .flush
= iwl_mvm_mac_flush
,
4043 .sched_scan_start
= iwl_mvm_mac_sched_scan_start
,
4044 .sched_scan_stop
= iwl_mvm_mac_sched_scan_stop
,
4045 .set_key
= iwl_mvm_mac_set_key
,
4046 .update_tkip_key
= iwl_mvm_mac_update_tkip_key
,
4047 .remain_on_channel
= iwl_mvm_roc
,
4048 .cancel_remain_on_channel
= iwl_mvm_cancel_roc
,
4049 .add_chanctx
= iwl_mvm_add_chanctx
,
4050 .remove_chanctx
= iwl_mvm_remove_chanctx
,
4051 .change_chanctx
= iwl_mvm_change_chanctx
,
4052 .assign_vif_chanctx
= iwl_mvm_assign_vif_chanctx
,
4053 .unassign_vif_chanctx
= iwl_mvm_unassign_vif_chanctx
,
4054 .switch_vif_chanctx
= iwl_mvm_switch_vif_chanctx
,
4056 .start_ap
= iwl_mvm_start_ap_ibss
,
4057 .stop_ap
= iwl_mvm_stop_ap_ibss
,
4058 .join_ibss
= iwl_mvm_start_ap_ibss
,
4059 .leave_ibss
= iwl_mvm_stop_ap_ibss
,
4061 .set_tim
= iwl_mvm_set_tim
,
4063 .channel_switch
= iwl_mvm_channel_switch
,
4064 .pre_channel_switch
= iwl_mvm_pre_channel_switch
,
4065 .post_channel_switch
= iwl_mvm_post_channel_switch
,
4067 .tdls_channel_switch
= iwl_mvm_tdls_channel_switch
,
4068 .tdls_cancel_channel_switch
= iwl_mvm_tdls_cancel_channel_switch
,
4069 .tdls_recv_channel_switch
= iwl_mvm_tdls_recv_channel_switch
,
4071 .event_callback
= iwl_mvm_mac_event_callback
,
4073 CFG80211_TESTMODE_CMD(iwl_mvm_mac_testmode_cmd
)
4075 #ifdef CONFIG_PM_SLEEP
4077 .suspend
= iwl_mvm_suspend
,
4078 .resume
= iwl_mvm_resume
,
4079 .set_wakeup
= iwl_mvm_set_wakeup
,
4080 .set_rekey_data
= iwl_mvm_set_rekey_data
,
4081 #if IS_ENABLED(CONFIG_IPV6)
4082 .ipv6_addr_change
= iwl_mvm_ipv6_addr_change
,
4084 .set_default_unicast_key
= iwl_mvm_set_default_unicast_key
,
4086 .get_survey
= iwl_mvm_mac_get_survey
,
4087 .sta_statistics
= iwl_mvm_mac_sta_statistics
,