]> git.proxmox.com Git - mirror_ubuntu-focal-kernel.git/blob - drivers/net/wireless/intel/iwlwifi/mvm/mvm.h
iwlwifi: mvm: remove RECONFIGURING queue state
[mirror_ubuntu-focal-kernel.git] / drivers / net / wireless / intel / iwlwifi / mvm / mvm.h
1 /******************************************************************************
2 *
3 * This file is provided under a dual BSD/GPLv2 license. When using or
4 * redistributing this file, you may do so under either license.
5 *
6 * GPL LICENSE SUMMARY
7 *
8 * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved.
9 * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH
10 * Copyright(c) 2016 - 2017 Intel Deutschland GmbH
11 * Copyright(c) 2018 Intel Corporation
12 *
13 * This program is free software; you can redistribute it and/or modify
14 * it under the terms of version 2 of the GNU General Public License as
15 * published by the Free Software Foundation.
16 *
17 * This program is distributed in the hope that it will be useful, but
18 * WITHOUT ANY WARRANTY; without even the implied warranty of
19 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
20 * General Public License for more details.
21 *
22 * The full GNU General Public License is included in this distribution
23 * in the file called COPYING.
24 *
25 * Contact Information:
26 * Intel Linux Wireless <linuxwifi@intel.com>
27 * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
28 *
29 * BSD LICENSE
30 *
31 * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved.
32 * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH
33 * Copyright(c) 2016 - 2017 Intel Deutschland GmbH
34 * Copyright(c) 2018 Intel Corporation
35 * All rights reserved.
36 *
37 * Redistribution and use in source and binary forms, with or without
38 * modification, are permitted provided that the following conditions
39 * are met:
40 *
41 * * Redistributions of source code must retain the above copyright
42 * notice, this list of conditions and the following disclaimer.
43 * * Redistributions in binary form must reproduce the above copyright
44 * notice, this list of conditions and the following disclaimer in
45 * the documentation and/or other materials provided with the
46 * distribution.
47 * * Neither the name Intel Corporation nor the names of its
48 * contributors may be used to endorse or promote products derived
49 * from this software without specific prior written permission.
50 *
51 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
52 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
53 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
54 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
55 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
56 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
57 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
58 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
59 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
60 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
61 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
62 *
63 *****************************************************************************/
64
65 #ifndef __IWL_MVM_H__
66 #define __IWL_MVM_H__
67
68 #include <linux/list.h>
69 #include <linux/spinlock.h>
70 #include <linux/leds.h>
71 #include <linux/in6.h>
72
73 #ifdef CONFIG_THERMAL
74 #include <linux/thermal.h>
75 #endif
76
77 #include "iwl-op-mode.h"
78 #include "iwl-trans.h"
79 #include "fw/notif-wait.h"
80 #include "iwl-eeprom-parse.h"
81 #include "fw/file.h"
82 #include "iwl-config.h"
83 #include "sta.h"
84 #include "fw-api.h"
85 #include "constants.h"
86 #include "tof.h"
87 #include "fw/runtime.h"
88 #include "fw/dbg.h"
89 #include "fw/acpi.h"
90 #include "iwl-nvm-parse.h"
91
92 #include <linux/average.h>
93
94 #define IWL_MVM_MAX_ADDRESSES 5
95 /* RSSI offset for WkP */
96 #define IWL_RSSI_OFFSET 50
97 #define IWL_MVM_MISSED_BEACONS_THRESHOLD 8
98 /* A TimeUnit is 1024 microsecond */
99 #define MSEC_TO_TU(_msec) (_msec*1000/1024)
100
101 /* For GO, this value represents the number of TUs before CSA "beacon
102 * 0" TBTT when the CSA time-event needs to be scheduled to start. It
103 * must be big enough to ensure that we switch in time.
104 */
105 #define IWL_MVM_CHANNEL_SWITCH_TIME_GO 40
106
107 /* For client, this value represents the number of TUs before CSA
108 * "beacon 1" TBTT, instead. This is because we don't know when the
109 * GO/AP will be in the new channel, so we switch early enough.
110 */
111 #define IWL_MVM_CHANNEL_SWITCH_TIME_CLIENT 10
112
113 /*
114 * This value (in TUs) is used to fine tune the CSA NoA end time which should
115 * be just before "beacon 0" TBTT.
116 */
117 #define IWL_MVM_CHANNEL_SWITCH_MARGIN 4
118
119 /*
120 * Number of beacons to transmit on a new channel until we unblock tx to
121 * the stations, even if we didn't identify them on a new channel
122 */
123 #define IWL_MVM_CS_UNBLOCK_TX_TIMEOUT 3
124
125 /* offchannel queue towards mac80211 */
126 #define IWL_MVM_OFFCHANNEL_QUEUE 0
127
128 extern const struct ieee80211_ops iwl_mvm_hw_ops;
129
130 /**
131 * struct iwl_mvm_mod_params - module parameters for iwlmvm
132 * @init_dbg: if true, then the NIC won't be stopped if the INIT fw asserted.
133 * We will register to mac80211 to have testmode working. The NIC must not
134 * be up'ed after the INIT fw asserted. This is useful to be able to use
135 * proprietary tools over testmode to debug the INIT fw.
136 * @tfd_q_hang_detect: enabled the detection of hung transmit queues
137 * @power_scheme: one of enum iwl_power_scheme
138 */
139 struct iwl_mvm_mod_params {
140 bool init_dbg;
141 bool tfd_q_hang_detect;
142 int power_scheme;
143 };
144 extern struct iwl_mvm_mod_params iwlmvm_mod_params;
145
146 struct iwl_mvm_phy_ctxt {
147 u16 id;
148 u16 color;
149 u32 ref;
150
151 enum nl80211_chan_width width;
152
153 /*
154 * TODO: This should probably be removed. Currently here only for rate
155 * scaling algorithm
156 */
157 struct ieee80211_channel *channel;
158 };
159
160 struct iwl_mvm_time_event_data {
161 struct ieee80211_vif *vif;
162 struct list_head list;
163 unsigned long end_jiffies;
164 u32 duration;
165 bool running;
166 u32 uid;
167
168 /*
169 * The access to the 'id' field must be done when the
170 * mvm->time_event_lock is held, as it value is used to indicate
171 * if the te is in the time event list or not (when id == TE_MAX)
172 */
173 u32 id;
174 };
175
176 /* Power management */
177
178 /**
179 * enum iwl_power_scheme
180 * @IWL_POWER_LEVEL_CAM - Continuously Active Mode
181 * @IWL_POWER_LEVEL_BPS - Balanced Power Save (default)
182 * @IWL_POWER_LEVEL_LP - Low Power
183 */
184 enum iwl_power_scheme {
185 IWL_POWER_SCHEME_CAM = 1,
186 IWL_POWER_SCHEME_BPS,
187 IWL_POWER_SCHEME_LP
188 };
189
190 #define IWL_CONN_MAX_LISTEN_INTERVAL 10
191 #define IWL_UAPSD_MAX_SP IEEE80211_WMM_IE_STA_QOSINFO_SP_ALL
192
193 #ifdef CONFIG_IWLWIFI_DEBUGFS
194 enum iwl_dbgfs_pm_mask {
195 MVM_DEBUGFS_PM_KEEP_ALIVE = BIT(0),
196 MVM_DEBUGFS_PM_SKIP_OVER_DTIM = BIT(1),
197 MVM_DEBUGFS_PM_SKIP_DTIM_PERIODS = BIT(2),
198 MVM_DEBUGFS_PM_RX_DATA_TIMEOUT = BIT(3),
199 MVM_DEBUGFS_PM_TX_DATA_TIMEOUT = BIT(4),
200 MVM_DEBUGFS_PM_LPRX_ENA = BIT(6),
201 MVM_DEBUGFS_PM_LPRX_RSSI_THRESHOLD = BIT(7),
202 MVM_DEBUGFS_PM_SNOOZE_ENABLE = BIT(8),
203 MVM_DEBUGFS_PM_UAPSD_MISBEHAVING = BIT(9),
204 MVM_DEBUGFS_PM_USE_PS_POLL = BIT(10),
205 };
206
207 struct iwl_dbgfs_pm {
208 u16 keep_alive_seconds;
209 u32 rx_data_timeout;
210 u32 tx_data_timeout;
211 bool skip_over_dtim;
212 u8 skip_dtim_periods;
213 bool lprx_ena;
214 u32 lprx_rssi_threshold;
215 bool snooze_ena;
216 bool uapsd_misbehaving;
217 bool use_ps_poll;
218 int mask;
219 };
220
221 /* beacon filtering */
222
223 enum iwl_dbgfs_bf_mask {
224 MVM_DEBUGFS_BF_ENERGY_DELTA = BIT(0),
225 MVM_DEBUGFS_BF_ROAMING_ENERGY_DELTA = BIT(1),
226 MVM_DEBUGFS_BF_ROAMING_STATE = BIT(2),
227 MVM_DEBUGFS_BF_TEMP_THRESHOLD = BIT(3),
228 MVM_DEBUGFS_BF_TEMP_FAST_FILTER = BIT(4),
229 MVM_DEBUGFS_BF_TEMP_SLOW_FILTER = BIT(5),
230 MVM_DEBUGFS_BF_ENABLE_BEACON_FILTER = BIT(6),
231 MVM_DEBUGFS_BF_DEBUG_FLAG = BIT(7),
232 MVM_DEBUGFS_BF_ESCAPE_TIMER = BIT(8),
233 MVM_DEBUGFS_BA_ESCAPE_TIMER = BIT(9),
234 MVM_DEBUGFS_BA_ENABLE_BEACON_ABORT = BIT(10),
235 };
236
237 struct iwl_dbgfs_bf {
238 u32 bf_energy_delta;
239 u32 bf_roaming_energy_delta;
240 u32 bf_roaming_state;
241 u32 bf_temp_threshold;
242 u32 bf_temp_fast_filter;
243 u32 bf_temp_slow_filter;
244 u32 bf_enable_beacon_filter;
245 u32 bf_debug_flag;
246 u32 bf_escape_timer;
247 u32 ba_escape_timer;
248 u32 ba_enable_beacon_abort;
249 int mask;
250 };
251 #endif
252
253 enum iwl_mvm_smps_type_request {
254 IWL_MVM_SMPS_REQ_BT_COEX,
255 IWL_MVM_SMPS_REQ_TT,
256 IWL_MVM_SMPS_REQ_PROT,
257 NUM_IWL_MVM_SMPS_REQ,
258 };
259
260 enum iwl_mvm_ref_type {
261 IWL_MVM_REF_UCODE_DOWN,
262 IWL_MVM_REF_SCAN,
263 IWL_MVM_REF_ROC,
264 IWL_MVM_REF_ROC_AUX,
265 IWL_MVM_REF_P2P_CLIENT,
266 IWL_MVM_REF_AP_IBSS,
267 IWL_MVM_REF_USER,
268 IWL_MVM_REF_TX,
269 IWL_MVM_REF_TX_AGG,
270 IWL_MVM_REF_ADD_IF,
271 IWL_MVM_REF_START_AP,
272 IWL_MVM_REF_BSS_CHANGED,
273 IWL_MVM_REF_PREPARE_TX,
274 IWL_MVM_REF_PROTECT_TDLS,
275 IWL_MVM_REF_CHECK_CTKILL,
276 IWL_MVM_REF_PRPH_READ,
277 IWL_MVM_REF_PRPH_WRITE,
278 IWL_MVM_REF_NMI,
279 IWL_MVM_REF_TM_CMD,
280 IWL_MVM_REF_EXIT_WORK,
281 IWL_MVM_REF_PROTECT_CSA,
282 IWL_MVM_REF_FW_DBG_COLLECT,
283 IWL_MVM_REF_INIT_UCODE,
284 IWL_MVM_REF_SENDING_CMD,
285 IWL_MVM_REF_RX,
286
287 /* update debugfs.c when changing this */
288
289 IWL_MVM_REF_COUNT,
290 };
291
292 enum iwl_bt_force_ant_mode {
293 BT_FORCE_ANT_DIS = 0,
294 BT_FORCE_ANT_AUTO,
295 BT_FORCE_ANT_BT,
296 BT_FORCE_ANT_WIFI,
297
298 BT_FORCE_ANT_MAX,
299 };
300
301 /**
302 * struct iwl_mvm_low_latency_cause - low latency set causes
303 * @LOW_LATENCY_TRAFFIC: indicates low latency traffic was detected
304 * @LOW_LATENCY_DEBUGFS: low latency mode set from debugfs
305 * @LOW_LATENCY_VCMD: low latency mode set from vendor command
306 */
307 enum iwl_mvm_low_latency_cause {
308 LOW_LATENCY_TRAFFIC = BIT(0),
309 LOW_LATENCY_DEBUGFS = BIT(1),
310 LOW_LATENCY_VCMD = BIT(2),
311 };
312
313 /**
314 * struct iwl_mvm_vif_bf_data - beacon filtering related data
315 * @bf_enabled: indicates if beacon filtering is enabled
316 * @ba_enabled: indicated if beacon abort is enabled
317 * @ave_beacon_signal: average beacon signal
318 * @last_cqm_event: rssi of the last cqm event
319 * @bt_coex_min_thold: minimum threshold for BT coex
320 * @bt_coex_max_thold: maximum threshold for BT coex
321 * @last_bt_coex_event: rssi of the last BT coex event
322 */
323 struct iwl_mvm_vif_bf_data {
324 bool bf_enabled;
325 bool ba_enabled;
326 int ave_beacon_signal;
327 int last_cqm_event;
328 int bt_coex_min_thold;
329 int bt_coex_max_thold;
330 int last_bt_coex_event;
331 };
332
333 /**
334 * struct iwl_probe_resp_data - data for NoA/CSA updates
335 * @rcu_head: used for freeing the data on update
336 * @notif: notification data
337 * @noa_len: length of NoA attribute, calculated from the notification
338 */
339 struct iwl_probe_resp_data {
340 struct rcu_head rcu_head;
341 struct iwl_probe_resp_data_notif notif;
342 int noa_len;
343 };
344
345 /**
346 * struct iwl_mvm_vif - data per Virtual Interface, it is a MAC context
347 * @id: between 0 and 3
348 * @color: to solve races upon MAC addition and removal
349 * @ap_sta_id: the sta_id of the AP - valid only if VIF type is STA
350 * @bssid: BSSID for this (client) interface
351 * @associated: indicates that we're currently associated, used only for
352 * managing the firmware state in iwl_mvm_bss_info_changed_station()
353 * @ap_assoc_sta_count: count of stations associated to us - valid only
354 * if VIF type is AP
355 * @uploaded: indicates the MAC context has been added to the device
356 * @ap_ibss_active: indicates that AP/IBSS is configured and that the interface
357 * should get quota etc.
358 * @pm_enabled - Indicate if MAC power management is allowed
359 * @monitor_active: indicates that monitor context is configured, and that the
360 * interface should get quota etc.
361 * @low_latency: indicates low latency is set, see
362 * enum &iwl_mvm_low_latency_cause for causes.
363 * @ps_disabled: indicates that this interface requires PS to be disabled
364 * @queue_params: QoS params for this MAC
365 * @bcast_sta: station used for broadcast packets. Used by the following
366 * vifs: P2P_DEVICE, GO and AP.
367 * @beacon_skb: the skb used to hold the AP/GO beacon template
368 * @smps_requests: the SMPS requests of different parts of the driver,
369 * combined on update to yield the overall request to mac80211.
370 * @beacon_stats: beacon statistics, containing the # of received beacons,
371 * # of received beacons accumulated over FW restart, and the current
372 * average signal of beacons retrieved from the firmware
373 * @csa_failed: CSA failed to schedule time event, report an error later
374 * @features: hw features active for this vif
375 * @probe_resp_data: data from FW notification to store NOA and CSA related
376 * data to be inserted into probe response.
377 */
378 struct iwl_mvm_vif {
379 struct iwl_mvm *mvm;
380 u16 id;
381 u16 color;
382 u8 ap_sta_id;
383
384 u8 bssid[ETH_ALEN];
385 bool associated;
386 u8 ap_assoc_sta_count;
387
388 u16 cab_queue;
389
390 bool uploaded;
391 bool ap_ibss_active;
392 bool pm_enabled;
393 bool monitor_active;
394 u8 low_latency;
395 bool ps_disabled;
396 struct iwl_mvm_vif_bf_data bf_data;
397
398 struct {
399 u32 num_beacons, accu_num_beacons;
400 u8 avg_signal;
401 } beacon_stats;
402
403 u32 ap_beacon_time;
404
405 enum iwl_tsf_id tsf_id;
406
407 /*
408 * QoS data from mac80211, need to store this here
409 * as mac80211 has a separate callback but we need
410 * to have the data for the MAC context
411 */
412 struct ieee80211_tx_queue_params queue_params[IEEE80211_NUM_ACS];
413 struct iwl_mvm_time_event_data time_event_data;
414 struct iwl_mvm_time_event_data hs_time_event_data;
415
416 struct iwl_mvm_int_sta bcast_sta;
417 struct iwl_mvm_int_sta mcast_sta;
418
419 /*
420 * Assigned while mac80211 has the interface in a channel context,
421 * or, for P2P Device, while it exists.
422 */
423 struct iwl_mvm_phy_ctxt *phy_ctxt;
424
425 #ifdef CONFIG_PM
426 /* WoWLAN GTK rekey data */
427 struct {
428 u8 kck[NL80211_KCK_LEN], kek[NL80211_KEK_LEN];
429 __le64 replay_ctr;
430 bool valid;
431 } rekey_data;
432
433 int tx_key_idx;
434
435 bool seqno_valid;
436 u16 seqno;
437 #endif
438
439 #if IS_ENABLED(CONFIG_IPV6)
440 /* IPv6 addresses for WoWLAN */
441 struct in6_addr target_ipv6_addrs[IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_MAX];
442 unsigned long tentative_addrs[BITS_TO_LONGS(IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_MAX)];
443 int num_target_ipv6_addrs;
444 #endif
445
446 #ifdef CONFIG_IWLWIFI_DEBUGFS
447 struct dentry *dbgfs_dir;
448 struct dentry *dbgfs_slink;
449 struct iwl_dbgfs_pm dbgfs_pm;
450 struct iwl_dbgfs_bf dbgfs_bf;
451 struct iwl_mac_power_cmd mac_pwr_cmd;
452 int dbgfs_quota_min;
453 #endif
454
455 enum ieee80211_smps_mode smps_requests[NUM_IWL_MVM_SMPS_REQ];
456
457 /* FW identified misbehaving AP */
458 u8 uapsd_misbehaving_bssid[ETH_ALEN];
459
460 struct delayed_work uapsd_nonagg_detected_wk;
461
462 /* Indicates that CSA countdown may be started */
463 bool csa_countdown;
464 bool csa_failed;
465 u16 csa_target_freq;
466
467 /* Indicates that we are waiting for a beacon on a new channel */
468 bool csa_bcn_pending;
469
470 /* TCP Checksum Offload */
471 netdev_features_t features;
472
473 struct iwl_probe_resp_data __rcu *probe_resp_data;
474 struct ieee80211_key_conf *ap_wep_key;
475 };
476
477 static inline struct iwl_mvm_vif *
478 iwl_mvm_vif_from_mac80211(struct ieee80211_vif *vif)
479 {
480 if (!vif)
481 return NULL;
482 return (void *)vif->drv_priv;
483 }
484
485 extern const u8 tid_to_mac80211_ac[];
486
487 #define IWL_MVM_SCAN_STOPPING_SHIFT 8
488
489 enum iwl_scan_status {
490 IWL_MVM_SCAN_REGULAR = BIT(0),
491 IWL_MVM_SCAN_SCHED = BIT(1),
492 IWL_MVM_SCAN_NETDETECT = BIT(2),
493
494 IWL_MVM_SCAN_STOPPING_REGULAR = BIT(8),
495 IWL_MVM_SCAN_STOPPING_SCHED = BIT(9),
496 IWL_MVM_SCAN_STOPPING_NETDETECT = BIT(10),
497
498 IWL_MVM_SCAN_REGULAR_MASK = IWL_MVM_SCAN_REGULAR |
499 IWL_MVM_SCAN_STOPPING_REGULAR,
500 IWL_MVM_SCAN_SCHED_MASK = IWL_MVM_SCAN_SCHED |
501 IWL_MVM_SCAN_STOPPING_SCHED,
502 IWL_MVM_SCAN_NETDETECT_MASK = IWL_MVM_SCAN_NETDETECT |
503 IWL_MVM_SCAN_STOPPING_NETDETECT,
504
505 IWL_MVM_SCAN_STOPPING_MASK = 0xff << IWL_MVM_SCAN_STOPPING_SHIFT,
506 IWL_MVM_SCAN_MASK = 0xff,
507 };
508
509 enum iwl_mvm_scan_type {
510 IWL_SCAN_TYPE_NOT_SET,
511 IWL_SCAN_TYPE_UNASSOC,
512 IWL_SCAN_TYPE_WILD,
513 IWL_SCAN_TYPE_MILD,
514 IWL_SCAN_TYPE_FRAGMENTED,
515 IWL_SCAN_TYPE_FAST_BALANCE,
516 };
517
518 enum iwl_mvm_sched_scan_pass_all_states {
519 SCHED_SCAN_PASS_ALL_DISABLED,
520 SCHED_SCAN_PASS_ALL_ENABLED,
521 SCHED_SCAN_PASS_ALL_FOUND,
522 };
523
524 /**
525 * struct iwl_mvm_tt_mgnt - Thermal Throttling Management structure
526 * @ct_kill_exit: worker to exit thermal kill
527 * @dynamic_smps: Is thermal throttling enabled dynamic_smps?
528 * @tx_backoff: The current thremal throttling tx backoff in uSec.
529 * @min_backoff: The minimal tx backoff due to power restrictions
530 * @params: Parameters to configure the thermal throttling algorithm.
531 * @throttle: Is thermal throttling is active?
532 */
533 struct iwl_mvm_tt_mgmt {
534 struct delayed_work ct_kill_exit;
535 bool dynamic_smps;
536 u32 tx_backoff;
537 u32 min_backoff;
538 struct iwl_tt_params params;
539 bool throttle;
540 };
541
542 #ifdef CONFIG_THERMAL
543 /**
544 *struct iwl_mvm_thermal_device - thermal zone related data
545 * @temp_trips: temperature thresholds for report
546 * @fw_trips_index: keep indexes to original array - temp_trips
547 * @tzone: thermal zone device data
548 */
549 struct iwl_mvm_thermal_device {
550 s16 temp_trips[IWL_MAX_DTS_TRIPS];
551 u8 fw_trips_index[IWL_MAX_DTS_TRIPS];
552 struct thermal_zone_device *tzone;
553 };
554
555 /*
556 * struct iwl_mvm_cooling_device
557 * @cur_state: current state
558 * @cdev: struct thermal cooling device
559 */
560 struct iwl_mvm_cooling_device {
561 u32 cur_state;
562 struct thermal_cooling_device *cdev;
563 };
564 #endif
565
566 #define IWL_MVM_NUM_LAST_FRAMES_UCODE_RATES 8
567
568 struct iwl_mvm_frame_stats {
569 u32 legacy_frames;
570 u32 ht_frames;
571 u32 vht_frames;
572 u32 bw_20_frames;
573 u32 bw_40_frames;
574 u32 bw_80_frames;
575 u32 bw_160_frames;
576 u32 sgi_frames;
577 u32 ngi_frames;
578 u32 siso_frames;
579 u32 mimo2_frames;
580 u32 agg_frames;
581 u32 ampdu_count;
582 u32 success_frames;
583 u32 fail_frames;
584 u32 last_rates[IWL_MVM_NUM_LAST_FRAMES_UCODE_RATES];
585 int last_frame_idx;
586 };
587
588 enum {
589 D0I3_DEFER_WAKEUP,
590 D0I3_PENDING_WAKEUP,
591 };
592
593 #define IWL_MVM_DEBUG_SET_TEMPERATURE_DISABLE 0xff
594 #define IWL_MVM_DEBUG_SET_TEMPERATURE_MIN -100
595 #define IWL_MVM_DEBUG_SET_TEMPERATURE_MAX 200
596
597 enum iwl_mvm_tdls_cs_state {
598 IWL_MVM_TDLS_SW_IDLE = 0,
599 IWL_MVM_TDLS_SW_REQ_SENT,
600 IWL_MVM_TDLS_SW_RESP_RCVD,
601 IWL_MVM_TDLS_SW_REQ_RCVD,
602 IWL_MVM_TDLS_SW_ACTIVE,
603 };
604
605 enum iwl_mvm_traffic_load {
606 IWL_MVM_TRAFFIC_LOW,
607 IWL_MVM_TRAFFIC_MEDIUM,
608 IWL_MVM_TRAFFIC_HIGH,
609 };
610
611 DECLARE_EWMA(rate, 16, 16)
612
613 struct iwl_mvm_tcm_mac {
614 struct {
615 u32 pkts[IEEE80211_NUM_ACS];
616 u32 airtime;
617 } tx;
618 struct {
619 u32 pkts[IEEE80211_NUM_ACS];
620 u32 airtime;
621 u32 last_ampdu_ref;
622 } rx;
623 struct {
624 /* track AP's transfer in client mode */
625 u64 rx_bytes;
626 struct ewma_rate rate;
627 bool detected;
628 } uapsd_nonagg_detect;
629 bool opened_rx_ba_sessions;
630 };
631
632 struct iwl_mvm_tcm {
633 struct delayed_work work;
634 spinlock_t lock; /* used when time elapsed */
635 unsigned long ts; /* timestamp when period ends */
636 unsigned long ll_ts;
637 unsigned long uapsd_nonagg_ts;
638 bool paused;
639 struct iwl_mvm_tcm_mac data[NUM_MAC_INDEX_DRIVER];
640 struct {
641 u32 elapsed; /* milliseconds for this TCM period */
642 u32 airtime[NUM_MAC_INDEX_DRIVER];
643 enum iwl_mvm_traffic_load load[NUM_MAC_INDEX_DRIVER];
644 enum iwl_mvm_traffic_load band_load[NUM_NL80211_BANDS];
645 enum iwl_mvm_traffic_load global_load;
646 bool low_latency[NUM_MAC_INDEX_DRIVER];
647 bool change[NUM_MAC_INDEX_DRIVER];
648 bool global_change;
649 } result;
650 };
651
652 /**
653 * struct iwl_mvm_reorder_buffer - per ra/tid/queue reorder buffer
654 * @head_sn: reorder window head sn
655 * @num_stored: number of mpdus stored in the buffer
656 * @buf_size: the reorder buffer size as set by the last addba request
657 * @queue: queue of this reorder buffer
658 * @last_amsdu: track last ASMDU SN for duplication detection
659 * @last_sub_index: track ASMDU sub frame index for duplication detection
660 * @reorder_timer: timer for frames are in the reorder buffer. For AMSDU
661 * it is the time of last received sub-frame
662 * @removed: prevent timer re-arming
663 * @valid: reordering is valid for this queue
664 * @lock: protect reorder buffer internal state
665 * @mvm: mvm pointer, needed for frame timer context
666 */
667 struct iwl_mvm_reorder_buffer {
668 u16 head_sn;
669 u16 num_stored;
670 u16 buf_size;
671 int queue;
672 u16 last_amsdu;
673 u8 last_sub_index;
674 struct timer_list reorder_timer;
675 bool removed;
676 bool valid;
677 spinlock_t lock;
678 struct iwl_mvm *mvm;
679 } ____cacheline_aligned_in_smp;
680
681 /**
682 * struct _iwl_mvm_reorder_buf_entry - reorder buffer entry per-queue/per-seqno
683 * @frames: list of skbs stored
684 * @reorder_time: time the packet was stored in the reorder buffer
685 */
686 struct _iwl_mvm_reorder_buf_entry {
687 struct sk_buff_head frames;
688 unsigned long reorder_time;
689 };
690
691 /* make this indirection to get the aligned thing */
692 struct iwl_mvm_reorder_buf_entry {
693 struct _iwl_mvm_reorder_buf_entry e;
694 }
695 #ifndef __CHECKER__
696 /* sparse doesn't like this construct: "bad integer constant expression" */
697 __aligned(roundup_pow_of_two(sizeof(struct _iwl_mvm_reorder_buf_entry)))
698 #endif
699 ;
700
701 /**
702 * struct iwl_mvm_baid_data - BA session data
703 * @sta_id: station id
704 * @tid: tid of the session
705 * @baid baid of the session
706 * @timeout: the timeout set in the addba request
707 * @entries_per_queue: # of buffers per queue, this actually gets
708 * aligned up to avoid cache line sharing between queues
709 * @last_rx: last rx jiffies, updated only if timeout passed from last update
710 * @session_timer: timer to check if BA session expired, runs at 2 * timeout
711 * @mvm: mvm pointer, needed for timer context
712 * @reorder_buf: reorder buffer, allocated per queue
713 * @reorder_buf_data: data
714 */
715 struct iwl_mvm_baid_data {
716 struct rcu_head rcu_head;
717 u8 sta_id;
718 u8 tid;
719 u8 baid;
720 u16 timeout;
721 u16 entries_per_queue;
722 unsigned long last_rx;
723 struct timer_list session_timer;
724 struct iwl_mvm_baid_data __rcu **rcu_ptr;
725 struct iwl_mvm *mvm;
726 struct iwl_mvm_reorder_buffer reorder_buf[IWL_MAX_RX_HW_QUEUES];
727 struct iwl_mvm_reorder_buf_entry entries[];
728 };
729
730 static inline struct iwl_mvm_baid_data *
731 iwl_mvm_baid_data_from_reorder_buf(struct iwl_mvm_reorder_buffer *buf)
732 {
733 return (void *)((u8 *)buf -
734 offsetof(struct iwl_mvm_baid_data, reorder_buf) -
735 sizeof(*buf) * buf->queue);
736 }
737
738 /*
739 * enum iwl_mvm_queue_status - queue status
740 * @IWL_MVM_QUEUE_FREE: the queue is not allocated nor reserved
741 * Basically, this means that this queue can be used for any purpose
742 * @IWL_MVM_QUEUE_RESERVED: queue is reserved but not yet in use
743 * This is the state of a queue that has been dedicated for some RATID
744 * (agg'd or not), but that hasn't yet gone through the actual enablement
745 * of iwl_mvm_enable_txq(), and therefore no traffic can go through it yet.
746 * Note that in this state there is no requirement to already know what TID
747 * should be used with this queue, it is just marked as a queue that will
748 * be used, and shouldn't be allocated to anyone else.
749 * @IWL_MVM_QUEUE_READY: queue is ready to be used
750 * This is the state of a queue that has been fully configured (including
751 * SCD pointers, etc), has a specific RA/TID assigned to it, and can be
752 * used to send traffic.
753 * @IWL_MVM_QUEUE_SHARED: queue is shared, or in a process of becoming shared
754 * This is a state in which a single queue serves more than one TID, all of
755 * which are not aggregated. Note that the queue is only associated to one
756 * RA.
757 * @IWL_MVM_QUEUE_INACTIVE: queue is allocated but no traffic on it
758 * This is a state of a queue that has had traffic on it, but during the
759 * last %IWL_MVM_DQA_QUEUE_TIMEOUT time period there has been no traffic on
760 * it. In this state, when a new queue is needed to be allocated but no
761 * such free queue exists, an inactive queue might be freed and given to
762 * the new RA/TID.
763 */
764 enum iwl_mvm_queue_status {
765 IWL_MVM_QUEUE_FREE,
766 IWL_MVM_QUEUE_RESERVED,
767 IWL_MVM_QUEUE_READY,
768 IWL_MVM_QUEUE_SHARED,
769 IWL_MVM_QUEUE_INACTIVE,
770 };
771
772 #define IWL_MVM_DQA_QUEUE_TIMEOUT (5 * HZ)
773 #define IWL_MVM_INVALID_QUEUE 0xFFFF
774
775 #define IWL_MVM_NUM_CIPHERS 10
776
777 struct iwl_mvm_sar_profile {
778 bool enabled;
779 u8 table[ACPI_SAR_TABLE_SIZE];
780 };
781
782 struct iwl_mvm_geo_profile {
783 u8 values[ACPI_GEO_TABLE_SIZE];
784 };
785
786 struct iwl_mvm_dqa_txq_info {
787 u8 ra_sta_id; /* The RA this queue is mapped to, if exists */
788 bool reserved; /* Is this the TXQ reserved for a STA */
789 u8 mac80211_ac; /* The mac80211 AC this queue is mapped to */
790 u8 txq_tid; /* The TID "owner" of this queue*/
791 u16 tid_bitmap; /* Bitmap of the TIDs mapped to this queue */
792 /* Timestamp for inactivation per TID of this queue */
793 unsigned long last_frame_time[IWL_MAX_TID_COUNT + 1];
794 enum iwl_mvm_queue_status status;
795 };
796
797 struct iwl_mvm {
798 /* for logger access */
799 struct device *dev;
800
801 struct iwl_trans *trans;
802 const struct iwl_fw *fw;
803 const struct iwl_cfg *cfg;
804 struct iwl_phy_db *phy_db;
805 struct ieee80211_hw *hw;
806
807 /* for protecting access to iwl_mvm */
808 struct mutex mutex;
809 struct list_head async_handlers_list;
810 spinlock_t async_handlers_lock;
811 struct work_struct async_handlers_wk;
812
813 struct work_struct roc_done_wk;
814
815 unsigned long init_status;
816
817 unsigned long status;
818
819 u32 queue_sync_cookie;
820 atomic_t queue_sync_counter;
821 /*
822 * for beacon filtering -
823 * currently only one interface can be supported
824 */
825 struct iwl_mvm_vif *bf_allowed_vif;
826
827 bool hw_registered;
828 bool calibrating;
829 u32 error_event_table[2];
830 u32 log_event_table;
831 u32 umac_error_event_table;
832 bool support_umac_log;
833
834 u32 ampdu_ref;
835 bool ampdu_toggle;
836
837 struct iwl_notif_wait_data notif_wait;
838
839 union {
840 struct mvm_statistics_rx_v3 rx_stats_v3;
841 struct mvm_statistics_rx rx_stats;
842 };
843
844 struct {
845 u64 rx_time;
846 u64 tx_time;
847 u64 on_time_rf;
848 u64 on_time_scan;
849 } radio_stats, accu_radio_stats;
850
851 u16 hw_queue_to_mac80211[IWL_MAX_TVQM_QUEUES];
852
853 struct iwl_mvm_dqa_txq_info queue_info[IWL_MAX_HW_QUEUES];
854 spinlock_t queue_info_lock; /* For syncing queue mgmt operations */
855 struct work_struct add_stream_wk; /* To add streams to queues */
856
857 atomic_t mac80211_queue_stop_count[IEEE80211_MAX_QUEUES];
858
859 const char *nvm_file_name;
860 struct iwl_nvm_data *nvm_data;
861 /* NVM sections */
862 struct iwl_nvm_section nvm_sections[NVM_MAX_NUM_SECTIONS];
863
864 struct iwl_fw_runtime fwrt;
865
866 /* EEPROM MAC addresses */
867 struct mac_address addresses[IWL_MVM_MAX_ADDRESSES];
868
869 /* data related to data path */
870 struct iwl_rx_phy_info last_phy_info;
871 struct ieee80211_sta __rcu *fw_id_to_mac_id[IWL_MVM_STATION_COUNT];
872 unsigned long sta_deferred_frames[BITS_TO_LONGS(IWL_MVM_STATION_COUNT)];
873 u8 rx_ba_sessions;
874
875 /* configured by mac80211 */
876 u32 rts_threshold;
877
878 /* Scan status, cmd (pre-allocated) and auxiliary station */
879 unsigned int scan_status;
880 void *scan_cmd;
881 struct iwl_mcast_filter_cmd *mcast_filter_cmd;
882 /* For CDB this is low band scan type, for non-CDB - type. */
883 enum iwl_mvm_scan_type scan_type;
884 enum iwl_mvm_scan_type hb_scan_type;
885
886 enum iwl_mvm_sched_scan_pass_all_states sched_scan_pass_all;
887 struct delayed_work scan_timeout_dwork;
888
889 /* max number of simultaneous scans the FW supports */
890 unsigned int max_scans;
891
892 /* UMAC scan tracking */
893 u32 scan_uid_status[IWL_MVM_MAX_UMAC_SCANS];
894
895 /* start time of last scan in TSF of the mac that requested the scan */
896 u64 scan_start;
897
898 /* the vif that requested the current scan */
899 struct iwl_mvm_vif *scan_vif;
900
901 /* rx chain antennas set through debugfs for the scan command */
902 u8 scan_rx_ant;
903
904 #ifdef CONFIG_IWLWIFI_BCAST_FILTERING
905 /* broadcast filters to configure for each associated station */
906 const struct iwl_fw_bcast_filter *bcast_filters;
907 #ifdef CONFIG_IWLWIFI_DEBUGFS
908 struct {
909 bool override;
910 struct iwl_bcast_filter_cmd cmd;
911 } dbgfs_bcast_filtering;
912 #endif
913 #endif
914
915 /* Internal station */
916 struct iwl_mvm_int_sta aux_sta;
917 struct iwl_mvm_int_sta snif_sta;
918
919 bool last_ebs_successful;
920
921 u8 scan_last_antenna_idx; /* to toggle TX between antennas */
922 u8 mgmt_last_antenna_idx;
923
924 /* last smart fifo state that was successfully sent to firmware */
925 enum iwl_sf_state sf_state;
926
927 #ifdef CONFIG_IWLWIFI_DEBUGFS
928 struct dentry *debugfs_dir;
929 u32 dbgfs_sram_offset, dbgfs_sram_len;
930 u32 dbgfs_prph_reg_addr;
931 bool disable_power_off;
932 bool disable_power_off_d3;
933
934 bool scan_iter_notif_enabled;
935
936 struct debugfs_blob_wrapper nvm_hw_blob;
937 struct debugfs_blob_wrapper nvm_sw_blob;
938 struct debugfs_blob_wrapper nvm_calib_blob;
939 struct debugfs_blob_wrapper nvm_prod_blob;
940 struct debugfs_blob_wrapper nvm_phy_sku_blob;
941
942 struct iwl_mvm_frame_stats drv_rx_stats;
943 spinlock_t drv_stats_lock;
944 u16 dbgfs_rx_phyinfo;
945 #endif
946
947 struct iwl_mvm_phy_ctxt phy_ctxts[NUM_PHY_CTX];
948
949 struct list_head time_event_list;
950 spinlock_t time_event_lock;
951
952 /*
953 * A bitmap indicating the index of the key in use. The firmware
954 * can hold 16 keys at most. Reflect this fact.
955 */
956 unsigned long fw_key_table[BITS_TO_LONGS(STA_KEY_MAX_NUM)];
957 u8 fw_key_deleted[STA_KEY_MAX_NUM];
958
959 /* references taken by the driver and spinlock protecting them */
960 spinlock_t refs_lock;
961 u8 refs[IWL_MVM_REF_COUNT];
962
963 u8 vif_count;
964
965 /* -1 for always, 0 for never, >0 for that many times */
966 s8 fw_restart;
967
968 #ifdef CONFIG_IWLWIFI_LEDS
969 struct led_classdev led;
970 #endif
971
972 struct ieee80211_vif *p2p_device_vif;
973
974 #ifdef CONFIG_PM
975 struct wiphy_wowlan_support wowlan;
976 int gtk_ivlen, gtk_icvlen, ptk_ivlen, ptk_icvlen;
977
978 /* sched scan settings for net detect */
979 struct ieee80211_scan_ies nd_ies;
980 struct cfg80211_match_set *nd_match_sets;
981 int n_nd_match_sets;
982 struct ieee80211_channel **nd_channels;
983 int n_nd_channels;
984 bool net_detect;
985 #ifdef CONFIG_IWLWIFI_DEBUGFS
986 bool d3_wake_sysassert;
987 bool d3_test_active;
988 bool store_d3_resume_sram;
989 void *d3_resume_sram;
990 u32 d3_test_pme_ptr;
991 struct ieee80211_vif *keep_vif;
992 u32 last_netdetect_scans; /* no. of scans in the last net-detect wake */
993 #endif
994 #endif
995
996 /* d0i3 */
997 u8 d0i3_ap_sta_id;
998 bool d0i3_offloading;
999 struct work_struct d0i3_exit_work;
1000 struct sk_buff_head d0i3_tx;
1001 /* protect d0i3_suspend_flags */
1002 struct mutex d0i3_suspend_mutex;
1003 unsigned long d0i3_suspend_flags;
1004 /* sync d0i3_tx queue and IWL_MVM_STATUS_IN_D0I3 status flag */
1005 spinlock_t d0i3_tx_lock;
1006 wait_queue_head_t d0i3_exit_waitq;
1007 wait_queue_head_t rx_sync_waitq;
1008
1009 /* BT-Coex */
1010 struct iwl_bt_coex_profile_notif last_bt_notif;
1011 struct iwl_bt_coex_ci_cmd last_bt_ci_cmd;
1012
1013 u8 bt_tx_prio;
1014 enum iwl_bt_force_ant_mode bt_force_ant_mode;
1015
1016 /* Aux ROC */
1017 struct list_head aux_roc_te_list;
1018
1019 /* Thermal Throttling and CTkill */
1020 struct iwl_mvm_tt_mgmt thermal_throttle;
1021 #ifdef CONFIG_THERMAL
1022 struct iwl_mvm_thermal_device tz_device;
1023 struct iwl_mvm_cooling_device cooling_dev;
1024 #endif
1025
1026 s32 temperature; /* Celsius */
1027 /*
1028 * Debug option to set the NIC temperature. This option makes the
1029 * driver think this is the actual NIC temperature, and ignore the
1030 * real temperature that is received from the fw
1031 */
1032 bool temperature_test; /* Debug test temperature is enabled */
1033
1034 unsigned long bt_coex_last_tcm_ts;
1035 struct iwl_mvm_tcm tcm;
1036
1037 u8 uapsd_noagg_bssid_write_idx;
1038 struct mac_address uapsd_noagg_bssids[IWL_MVM_UAPSD_NOAGG_BSSIDS_NUM]
1039 __aligned(2);
1040
1041 struct iwl_time_quota_cmd last_quota_cmd;
1042
1043 #ifdef CONFIG_NL80211_TESTMODE
1044 u32 noa_duration;
1045 struct ieee80211_vif *noa_vif;
1046 #endif
1047
1048 /* Tx queues */
1049 u16 aux_queue;
1050 u16 snif_queue;
1051 u16 probe_queue;
1052 u16 p2p_dev_queue;
1053
1054 /* Indicate if device power save is allowed */
1055 u8 ps_disabled; /* u8 instead of bool to ease debugfs_create_* usage */
1056 unsigned int max_amsdu_len; /* used for debugfs only */
1057
1058 struct ieee80211_vif __rcu *csa_vif;
1059 struct ieee80211_vif __rcu *csa_tx_blocked_vif;
1060 u8 csa_tx_block_bcn_timeout;
1061
1062 /* system time of last beacon (for AP/GO interface) */
1063 u32 ap_last_beacon_gp2;
1064
1065 /* indicates that we transmitted the last beacon */
1066 bool ibss_manager;
1067
1068 bool lar_regdom_set;
1069 enum iwl_mcc_source mcc_src;
1070
1071 /* TDLS channel switch data */
1072 struct {
1073 struct delayed_work dwork;
1074 enum iwl_mvm_tdls_cs_state state;
1075
1076 /*
1077 * Current cs sta - might be different from periodic cs peer
1078 * station. Value is meaningless when the cs-state is idle.
1079 */
1080 u8 cur_sta_id;
1081
1082 /* TDLS periodic channel-switch peer */
1083 struct {
1084 u8 sta_id;
1085 u8 op_class;
1086 bool initiator; /* are we the link initiator */
1087 struct cfg80211_chan_def chandef;
1088 struct sk_buff *skb; /* ch sw template */
1089 u32 ch_sw_tm_ie;
1090
1091 /* timestamp of last ch-sw request sent (GP2 time) */
1092 u32 sent_timestamp;
1093 } peer;
1094 } tdls_cs;
1095
1096
1097 u32 ciphers[IWL_MVM_NUM_CIPHERS];
1098 struct ieee80211_cipher_scheme cs[IWL_UCODE_MAX_CS];
1099 struct iwl_mvm_tof_data tof_data;
1100
1101 struct ieee80211_vif *nan_vif;
1102 #define IWL_MAX_BAID 32
1103 struct iwl_mvm_baid_data __rcu *baid_map[IWL_MAX_BAID];
1104
1105 /*
1106 * Drop beacons from other APs in AP mode when there are no connected
1107 * clients.
1108 */
1109 bool drop_bcn_ap_mode;
1110
1111 struct delayed_work cs_tx_unblock_dwork;
1112
1113 /* does a monitor vif exist (only one can exist hence bool) */
1114 bool monitor_on;
1115 #ifdef CONFIG_ACPI
1116 struct iwl_mvm_sar_profile sar_profiles[ACPI_SAR_PROFILE_NUM];
1117 struct iwl_mvm_geo_profile geo_profiles[ACPI_NUM_GEO_PROFILES];
1118 #endif
1119 };
1120
1121 /* Extract MVM priv from op_mode and _hw */
1122 #define IWL_OP_MODE_GET_MVM(_iwl_op_mode) \
1123 ((struct iwl_mvm *)(_iwl_op_mode)->op_mode_specific)
1124
1125 #define IWL_MAC80211_GET_MVM(_hw) \
1126 IWL_OP_MODE_GET_MVM((struct iwl_op_mode *)((_hw)->priv))
1127
1128 /**
1129 * enum iwl_mvm_status - MVM status bits
1130 * @IWL_MVM_STATUS_HW_RFKILL: HW RF-kill is asserted
1131 * @IWL_MVM_STATUS_HW_CTKILL: CT-kill is active
1132 * @IWL_MVM_STATUS_ROC_RUNNING: remain-on-channel is running
1133 * @IWL_MVM_STATUS_HW_RESTART_REQUESTED: HW restart was requested
1134 * @IWL_MVM_STATUS_IN_HW_RESTART: HW restart is active
1135 * @IWL_MVM_STATUS_IN_D0I3: NIC is in D0i3
1136 * @IWL_MVM_STATUS_ROC_AUX_RUNNING: AUX remain-on-channel is running
1137 * @IWL_MVM_STATUS_D3_RECONFIG: D3 reconfiguration is being done
1138 * @IWL_MVM_STATUS_FIRMWARE_RUNNING: firmware is running
1139 * @IWL_MVM_STATUS_NEED_FLUSH_P2P: need to flush P2P bcast STA
1140 */
1141 enum iwl_mvm_status {
1142 IWL_MVM_STATUS_HW_RFKILL,
1143 IWL_MVM_STATUS_HW_CTKILL,
1144 IWL_MVM_STATUS_ROC_RUNNING,
1145 IWL_MVM_STATUS_HW_RESTART_REQUESTED,
1146 IWL_MVM_STATUS_IN_HW_RESTART,
1147 IWL_MVM_STATUS_IN_D0I3,
1148 IWL_MVM_STATUS_ROC_AUX_RUNNING,
1149 IWL_MVM_STATUS_D3_RECONFIG,
1150 IWL_MVM_STATUS_FIRMWARE_RUNNING,
1151 IWL_MVM_STATUS_NEED_FLUSH_P2P,
1152 };
1153
1154 /* Keep track of completed init configuration */
1155 enum iwl_mvm_init_status {
1156 IWL_MVM_INIT_STATUS_THERMAL_INIT_COMPLETE = BIT(0),
1157 IWL_MVM_INIT_STATUS_LEDS_INIT_COMPLETE = BIT(1),
1158 IWL_MVM_INIT_STATUS_REG_HW_INIT_COMPLETE = BIT(2),
1159 IWL_MVM_INIT_STATUS_TOF_INIT_COMPLETE = BIT(3),
1160 };
1161
1162 static inline bool iwl_mvm_is_radio_killed(struct iwl_mvm *mvm)
1163 {
1164 return test_bit(IWL_MVM_STATUS_HW_RFKILL, &mvm->status) ||
1165 test_bit(IWL_MVM_STATUS_HW_CTKILL, &mvm->status);
1166 }
1167
1168 static inline bool iwl_mvm_is_radio_hw_killed(struct iwl_mvm *mvm)
1169 {
1170 return test_bit(IWL_MVM_STATUS_HW_RFKILL, &mvm->status);
1171 }
1172
1173 static inline bool iwl_mvm_firmware_running(struct iwl_mvm *mvm)
1174 {
1175 return test_bit(IWL_MVM_STATUS_FIRMWARE_RUNNING, &mvm->status);
1176 }
1177
1178 /* Must be called with rcu_read_lock() held and it can only be
1179 * released when mvmsta is not needed anymore.
1180 */
1181 static inline struct iwl_mvm_sta *
1182 iwl_mvm_sta_from_staid_rcu(struct iwl_mvm *mvm, u8 sta_id)
1183 {
1184 struct ieee80211_sta *sta;
1185
1186 if (sta_id >= ARRAY_SIZE(mvm->fw_id_to_mac_id))
1187 return NULL;
1188
1189 sta = rcu_dereference(mvm->fw_id_to_mac_id[sta_id]);
1190
1191 /* This can happen if the station has been removed right now */
1192 if (IS_ERR_OR_NULL(sta))
1193 return NULL;
1194
1195 return iwl_mvm_sta_from_mac80211(sta);
1196 }
1197
1198 static inline struct iwl_mvm_sta *
1199 iwl_mvm_sta_from_staid_protected(struct iwl_mvm *mvm, u8 sta_id)
1200 {
1201 struct ieee80211_sta *sta;
1202
1203 if (sta_id >= ARRAY_SIZE(mvm->fw_id_to_mac_id))
1204 return NULL;
1205
1206 sta = rcu_dereference_protected(mvm->fw_id_to_mac_id[sta_id],
1207 lockdep_is_held(&mvm->mutex));
1208
1209 /* This can happen if the station has been removed right now */
1210 if (IS_ERR_OR_NULL(sta))
1211 return NULL;
1212
1213 return iwl_mvm_sta_from_mac80211(sta);
1214 }
1215
1216 static inline bool iwl_mvm_is_d0i3_supported(struct iwl_mvm *mvm)
1217 {
1218 return !iwlwifi_mod_params.d0i3_disable &&
1219 fw_has_capa(&mvm->fw->ucode_capa,
1220 IWL_UCODE_TLV_CAPA_D0I3_SUPPORT);
1221 }
1222
1223 static inline bool iwl_mvm_is_adaptive_dwell_supported(struct iwl_mvm *mvm)
1224 {
1225 return fw_has_api(&mvm->fw->ucode_capa,
1226 IWL_UCODE_TLV_API_ADAPTIVE_DWELL);
1227 }
1228
1229 static inline bool iwl_mvm_is_adaptive_dwell_v2_supported(struct iwl_mvm *mvm)
1230 {
1231 return fw_has_api(&mvm->fw->ucode_capa,
1232 IWL_UCODE_TLV_API_ADAPTIVE_DWELL_V2);
1233 }
1234
1235 static inline bool iwl_mvm_is_oce_supported(struct iwl_mvm *mvm)
1236 {
1237 /* OCE should never be enabled for LMAC scan FWs */
1238 return fw_has_api(&mvm->fw->ucode_capa, IWL_UCODE_TLV_API_OCE);
1239 }
1240
1241 static inline bool iwl_mvm_is_frag_ebs_supported(struct iwl_mvm *mvm)
1242 {
1243 return fw_has_api(&mvm->fw->ucode_capa, IWL_UCODE_TLV_API_FRAG_EBS);
1244 }
1245
1246 static inline bool iwl_mvm_enter_d0i3_on_suspend(struct iwl_mvm *mvm)
1247 {
1248 /* For now we only use this mode to differentiate between
1249 * slave transports, which handle D0i3 entry in suspend by
1250 * themselves in conjunction with runtime PM D0i3. So, this
1251 * function is used to check whether we need to do anything
1252 * when entering suspend or if the transport layer has already
1253 * done it.
1254 */
1255 return (mvm->trans->system_pm_mode == IWL_PLAT_PM_MODE_D0I3) &&
1256 (mvm->trans->runtime_pm_mode != IWL_PLAT_PM_MODE_D0I3);
1257 }
1258
1259 static inline bool iwl_mvm_is_dqa_data_queue(struct iwl_mvm *mvm, u8 queue)
1260 {
1261 return (queue >= IWL_MVM_DQA_MIN_DATA_QUEUE) &&
1262 (queue <= IWL_MVM_DQA_MAX_DATA_QUEUE);
1263 }
1264
1265 static inline bool iwl_mvm_is_dqa_mgmt_queue(struct iwl_mvm *mvm, u8 queue)
1266 {
1267 return (queue >= IWL_MVM_DQA_MIN_MGMT_QUEUE) &&
1268 (queue <= IWL_MVM_DQA_MAX_MGMT_QUEUE);
1269 }
1270
1271 static inline bool iwl_mvm_is_lar_supported(struct iwl_mvm *mvm)
1272 {
1273 bool nvm_lar = mvm->nvm_data->lar_enabled;
1274 bool tlv_lar = fw_has_capa(&mvm->fw->ucode_capa,
1275 IWL_UCODE_TLV_CAPA_LAR_SUPPORT);
1276
1277 if (iwlwifi_mod_params.lar_disable)
1278 return false;
1279
1280 /*
1281 * Enable LAR only if it is supported by the FW (TLV) &&
1282 * enabled in the NVM
1283 */
1284 if (mvm->cfg->nvm_type == IWL_NVM_EXT)
1285 return nvm_lar && tlv_lar;
1286 else
1287 return tlv_lar;
1288 }
1289
1290 static inline bool iwl_mvm_is_wifi_mcc_supported(struct iwl_mvm *mvm)
1291 {
1292 return fw_has_api(&mvm->fw->ucode_capa,
1293 IWL_UCODE_TLV_API_WIFI_MCC_UPDATE) ||
1294 fw_has_capa(&mvm->fw->ucode_capa,
1295 IWL_UCODE_TLV_CAPA_LAR_MULTI_MCC);
1296 }
1297
1298 static inline bool iwl_mvm_bt_is_rrc_supported(struct iwl_mvm *mvm)
1299 {
1300 return fw_has_capa(&mvm->fw->ucode_capa,
1301 IWL_UCODE_TLV_CAPA_BT_COEX_RRC) &&
1302 IWL_MVM_BT_COEX_RRC;
1303 }
1304
1305 static inline bool iwl_mvm_is_csum_supported(struct iwl_mvm *mvm)
1306 {
1307 return fw_has_capa(&mvm->fw->ucode_capa,
1308 IWL_UCODE_TLV_CAPA_CSUM_SUPPORT) &&
1309 !IWL_MVM_HW_CSUM_DISABLE;
1310 }
1311
1312 static inline bool iwl_mvm_is_mplut_supported(struct iwl_mvm *mvm)
1313 {
1314 return fw_has_capa(&mvm->fw->ucode_capa,
1315 IWL_UCODE_TLV_CAPA_BT_MPLUT_SUPPORT) &&
1316 IWL_MVM_BT_COEX_MPLUT;
1317 }
1318
1319 static inline
1320 bool iwl_mvm_is_p2p_scm_uapsd_supported(struct iwl_mvm *mvm)
1321 {
1322 return fw_has_capa(&mvm->fw->ucode_capa,
1323 IWL_UCODE_TLV_CAPA_P2P_SCM_UAPSD) &&
1324 !(iwlwifi_mod_params.uapsd_disable &
1325 IWL_DISABLE_UAPSD_P2P_CLIENT);
1326 }
1327
1328 static inline bool iwl_mvm_has_new_rx_api(struct iwl_mvm *mvm)
1329 {
1330 return fw_has_capa(&mvm->fw->ucode_capa,
1331 IWL_UCODE_TLV_CAPA_MULTI_QUEUE_RX_SUPPORT);
1332 }
1333
1334 static inline bool iwl_mvm_has_new_tx_api(struct iwl_mvm *mvm)
1335 {
1336 /* TODO - replace with TLV once defined */
1337 return mvm->trans->cfg->use_tfh;
1338 }
1339
1340 static inline bool iwl_mvm_has_unified_ucode(struct iwl_mvm *mvm)
1341 {
1342 /* TODO - better define this */
1343 return mvm->trans->cfg->device_family >= IWL_DEVICE_FAMILY_22000;
1344 }
1345
1346 static inline bool iwl_mvm_is_cdb_supported(struct iwl_mvm *mvm)
1347 {
1348 /*
1349 * TODO:
1350 * The issue of how to determine CDB APIs and usage is still not fully
1351 * defined.
1352 * There is a compilation for CDB and non-CDB FW, but there may
1353 * be also runtime check.
1354 * For now there is a TLV for checking compilation mode, but a
1355 * runtime check will also have to be here - once defined.
1356 */
1357 return fw_has_capa(&mvm->fw->ucode_capa,
1358 IWL_UCODE_TLV_CAPA_CDB_SUPPORT);
1359 }
1360
1361 static inline bool iwl_mvm_cdb_scan_api(struct iwl_mvm *mvm)
1362 {
1363 /*
1364 * TODO: should this be the same as iwl_mvm_is_cdb_supported()?
1365 * but then there's a little bit of code in scan that won't make
1366 * any sense...
1367 */
1368 return mvm->trans->cfg->device_family >= IWL_DEVICE_FAMILY_22000;
1369 }
1370
1371 static inline bool iwl_mvm_has_new_rx_stats_api(struct iwl_mvm *mvm)
1372 {
1373 return fw_has_api(&mvm->fw->ucode_capa,
1374 IWL_UCODE_TLV_API_NEW_RX_STATS);
1375 }
1376
1377 static inline bool iwl_mvm_has_quota_low_latency(struct iwl_mvm *mvm)
1378 {
1379 return fw_has_api(&mvm->fw->ucode_capa,
1380 IWL_UCODE_TLV_API_QUOTA_LOW_LATENCY);
1381 }
1382
1383 static inline bool iwl_mvm_has_tlc_offload(const struct iwl_mvm *mvm)
1384 {
1385 return fw_has_capa(&mvm->fw->ucode_capa,
1386 IWL_UCODE_TLV_CAPA_TLC_OFFLOAD);
1387 }
1388
1389 static inline struct agg_tx_status *
1390 iwl_mvm_get_agg_status(struct iwl_mvm *mvm, void *tx_resp)
1391 {
1392 if (iwl_mvm_has_new_tx_api(mvm))
1393 return &((struct iwl_mvm_tx_resp *)tx_resp)->status;
1394 else
1395 return ((struct iwl_mvm_tx_resp_v3 *)tx_resp)->status;
1396 }
1397
1398 static inline bool iwl_mvm_is_tt_in_fw(struct iwl_mvm *mvm)
1399 {
1400 #ifdef CONFIG_THERMAL
1401 /* these two TLV are redundant since the responsibility to CT-kill by
1402 * FW happens only after we send at least one command of
1403 * temperature THs report.
1404 */
1405 return fw_has_capa(&mvm->fw->ucode_capa,
1406 IWL_UCODE_TLV_CAPA_CT_KILL_BY_FW) &&
1407 fw_has_capa(&mvm->fw->ucode_capa,
1408 IWL_UCODE_TLV_CAPA_TEMP_THS_REPORT_SUPPORT);
1409 #else /* CONFIG_THERMAL */
1410 return false;
1411 #endif /* CONFIG_THERMAL */
1412 }
1413
1414 static inline bool iwl_mvm_is_ctdp_supported(struct iwl_mvm *mvm)
1415 {
1416 return fw_has_capa(&mvm->fw->ucode_capa,
1417 IWL_UCODE_TLV_CAPA_CTDP_SUPPORT);
1418 }
1419
1420 extern const u8 iwl_mvm_ac_to_tx_fifo[];
1421 extern const u8 iwl_mvm_ac_to_gen2_tx_fifo[];
1422
1423 static inline u8 iwl_mvm_mac_ac_to_tx_fifo(struct iwl_mvm *mvm,
1424 enum ieee80211_ac_numbers ac)
1425 {
1426 return iwl_mvm_has_new_tx_api(mvm) ?
1427 iwl_mvm_ac_to_gen2_tx_fifo[ac] : iwl_mvm_ac_to_tx_fifo[ac];
1428 }
1429
1430 struct iwl_rate_info {
1431 u8 plcp; /* uCode API: IWL_RATE_6M_PLCP, etc. */
1432 u8 plcp_siso; /* uCode API: IWL_RATE_SISO_6M_PLCP, etc. */
1433 u8 plcp_mimo2; /* uCode API: IWL_RATE_MIMO2_6M_PLCP, etc. */
1434 u8 plcp_mimo3; /* uCode API: IWL_RATE_MIMO3_6M_PLCP, etc. */
1435 u8 ieee; /* MAC header: IWL_RATE_6M_IEEE, etc. */
1436 };
1437
1438 void __iwl_mvm_mac_stop(struct iwl_mvm *mvm);
1439 int __iwl_mvm_mac_start(struct iwl_mvm *mvm);
1440
1441 /******************
1442 * MVM Methods
1443 ******************/
1444 /* uCode */
1445 int iwl_run_init_mvm_ucode(struct iwl_mvm *mvm, bool read_nvm);
1446
1447 /* Utils */
1448 int iwl_mvm_legacy_rate_to_mac80211_idx(u32 rate_n_flags,
1449 enum nl80211_band band);
1450 void iwl_mvm_hwrate_to_tx_rate(u32 rate_n_flags,
1451 enum nl80211_band band,
1452 struct ieee80211_tx_rate *r);
1453 u8 iwl_mvm_mac80211_idx_to_hwrate(int rate_idx);
1454 void iwl_mvm_dump_nic_error_log(struct iwl_mvm *mvm);
1455 u8 first_antenna(u8 mask);
1456 u8 iwl_mvm_next_antenna(struct iwl_mvm *mvm, u8 valid, u8 last_idx);
1457 void iwl_mvm_get_sync_time(struct iwl_mvm *mvm, u32 *gp2, u64 *boottime);
1458
1459 /* Tx / Host Commands */
1460 int __must_check iwl_mvm_send_cmd(struct iwl_mvm *mvm,
1461 struct iwl_host_cmd *cmd);
1462 int __must_check iwl_mvm_send_cmd_pdu(struct iwl_mvm *mvm, u32 id,
1463 u32 flags, u16 len, const void *data);
1464 int __must_check iwl_mvm_send_cmd_status(struct iwl_mvm *mvm,
1465 struct iwl_host_cmd *cmd,
1466 u32 *status);
1467 int __must_check iwl_mvm_send_cmd_pdu_status(struct iwl_mvm *mvm, u32 id,
1468 u16 len, const void *data,
1469 u32 *status);
1470 int iwl_mvm_tx_skb(struct iwl_mvm *mvm, struct sk_buff *skb,
1471 struct ieee80211_sta *sta);
1472 int iwl_mvm_tx_skb_non_sta(struct iwl_mvm *mvm, struct sk_buff *skb);
1473 void iwl_mvm_set_tx_cmd(struct iwl_mvm *mvm, struct sk_buff *skb,
1474 struct iwl_tx_cmd *tx_cmd,
1475 struct ieee80211_tx_info *info, u8 sta_id);
1476 void iwl_mvm_set_tx_cmd_rate(struct iwl_mvm *mvm, struct iwl_tx_cmd *tx_cmd,
1477 struct ieee80211_tx_info *info,
1478 struct ieee80211_sta *sta, __le16 fc);
1479 #ifdef CONFIG_IWLWIFI_DEBUG
1480 const char *iwl_mvm_get_tx_fail_reason(u32 status);
1481 #else
1482 static inline const char *iwl_mvm_get_tx_fail_reason(u32 status) { return ""; }
1483 #endif
1484 int iwl_mvm_flush_tx_path(struct iwl_mvm *mvm, u32 tfd_msk, u32 flags);
1485 int iwl_mvm_flush_sta(struct iwl_mvm *mvm, void *sta, bool internal, u32 flags);
1486 int iwl_mvm_flush_sta_tids(struct iwl_mvm *mvm, u32 sta_id,
1487 u16 tids, u32 flags);
1488
1489 void iwl_mvm_async_handlers_purge(struct iwl_mvm *mvm);
1490
1491 static inline void iwl_mvm_set_tx_cmd_ccmp(struct ieee80211_tx_info *info,
1492 struct iwl_tx_cmd *tx_cmd)
1493 {
1494 struct ieee80211_key_conf *keyconf = info->control.hw_key;
1495
1496 tx_cmd->sec_ctl = TX_CMD_SEC_CCM;
1497 memcpy(tx_cmd->key, keyconf->key, keyconf->keylen);
1498 }
1499
1500 static inline void iwl_mvm_wait_for_async_handlers(struct iwl_mvm *mvm)
1501 {
1502 flush_work(&mvm->async_handlers_wk);
1503 }
1504
1505 /* Statistics */
1506 void iwl_mvm_handle_rx_statistics(struct iwl_mvm *mvm,
1507 struct iwl_rx_packet *pkt);
1508 void iwl_mvm_rx_statistics(struct iwl_mvm *mvm,
1509 struct iwl_rx_cmd_buffer *rxb);
1510 int iwl_mvm_request_statistics(struct iwl_mvm *mvm, bool clear);
1511 void iwl_mvm_accu_radio_stats(struct iwl_mvm *mvm);
1512
1513 /* NVM */
1514 int iwl_nvm_init(struct iwl_mvm *mvm);
1515 int iwl_mvm_load_nvm_to_nic(struct iwl_mvm *mvm);
1516
1517 static inline u8 iwl_mvm_get_valid_tx_ant(struct iwl_mvm *mvm)
1518 {
1519 return mvm->nvm_data && mvm->nvm_data->valid_tx_ant ?
1520 mvm->fw->valid_tx_ant & mvm->nvm_data->valid_tx_ant :
1521 mvm->fw->valid_tx_ant;
1522 }
1523
1524 static inline u8 iwl_mvm_get_valid_rx_ant(struct iwl_mvm *mvm)
1525 {
1526 return mvm->nvm_data && mvm->nvm_data->valid_rx_ant ?
1527 mvm->fw->valid_rx_ant & mvm->nvm_data->valid_rx_ant :
1528 mvm->fw->valid_rx_ant;
1529 }
1530
1531 static inline u32 iwl_mvm_get_phy_config(struct iwl_mvm *mvm)
1532 {
1533 u32 phy_config = ~(FW_PHY_CFG_TX_CHAIN |
1534 FW_PHY_CFG_RX_CHAIN);
1535 u32 valid_rx_ant = iwl_mvm_get_valid_rx_ant(mvm);
1536 u32 valid_tx_ant = iwl_mvm_get_valid_tx_ant(mvm);
1537
1538 phy_config |= valid_tx_ant << FW_PHY_CFG_TX_CHAIN_POS |
1539 valid_rx_ant << FW_PHY_CFG_RX_CHAIN_POS;
1540
1541 return mvm->fw->phy_config & phy_config;
1542 }
1543
1544 int iwl_mvm_up(struct iwl_mvm *mvm);
1545 int iwl_mvm_load_d3_fw(struct iwl_mvm *mvm);
1546
1547 int iwl_mvm_mac_setup_register(struct iwl_mvm *mvm);
1548 bool iwl_mvm_bcast_filter_build_cmd(struct iwl_mvm *mvm,
1549 struct iwl_bcast_filter_cmd *cmd);
1550
1551 /*
1552 * FW notifications / CMD responses handlers
1553 * Convention: iwl_mvm_rx_<NAME OF THE CMD>
1554 */
1555 void iwl_mvm_rx_rx_phy_cmd(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1556 void iwl_mvm_rx_rx_mpdu(struct iwl_mvm *mvm, struct napi_struct *napi,
1557 struct iwl_rx_cmd_buffer *rxb);
1558 void iwl_mvm_rx_mpdu_mq(struct iwl_mvm *mvm, struct napi_struct *napi,
1559 struct iwl_rx_cmd_buffer *rxb, int queue);
1560 void iwl_mvm_rx_frame_release(struct iwl_mvm *mvm, struct napi_struct *napi,
1561 struct iwl_rx_cmd_buffer *rxb, int queue);
1562 int iwl_mvm_notify_rx_queue(struct iwl_mvm *mvm, u32 rxq_mask,
1563 const u8 *data, u32 count);
1564 void iwl_mvm_rx_queue_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb,
1565 int queue);
1566 void iwl_mvm_rx_tx_cmd(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1567 void iwl_mvm_mfu_assert_dump_notif(struct iwl_mvm *mvm,
1568 struct iwl_rx_cmd_buffer *rxb);
1569 void iwl_mvm_rx_ba_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1570 void iwl_mvm_rx_ant_coupling_notif(struct iwl_mvm *mvm,
1571 struct iwl_rx_cmd_buffer *rxb);
1572 void iwl_mvm_rx_fw_error(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1573 void iwl_mvm_rx_card_state_notif(struct iwl_mvm *mvm,
1574 struct iwl_rx_cmd_buffer *rxb);
1575 void iwl_mvm_rx_mfuart_notif(struct iwl_mvm *mvm,
1576 struct iwl_rx_cmd_buffer *rxb);
1577 void iwl_mvm_rx_shared_mem_cfg_notif(struct iwl_mvm *mvm,
1578 struct iwl_rx_cmd_buffer *rxb);
1579
1580 /* MVM PHY */
1581 int iwl_mvm_phy_ctxt_add(struct iwl_mvm *mvm, struct iwl_mvm_phy_ctxt *ctxt,
1582 struct cfg80211_chan_def *chandef,
1583 u8 chains_static, u8 chains_dynamic);
1584 int iwl_mvm_phy_ctxt_changed(struct iwl_mvm *mvm, struct iwl_mvm_phy_ctxt *ctxt,
1585 struct cfg80211_chan_def *chandef,
1586 u8 chains_static, u8 chains_dynamic);
1587 void iwl_mvm_phy_ctxt_ref(struct iwl_mvm *mvm,
1588 struct iwl_mvm_phy_ctxt *ctxt);
1589 void iwl_mvm_phy_ctxt_unref(struct iwl_mvm *mvm,
1590 struct iwl_mvm_phy_ctxt *ctxt);
1591 int iwl_mvm_phy_ctx_count(struct iwl_mvm *mvm);
1592 u8 iwl_mvm_get_channel_width(struct cfg80211_chan_def *chandef);
1593 u8 iwl_mvm_get_ctrl_pos(struct cfg80211_chan_def *chandef);
1594
1595 /* MAC (virtual interface) programming */
1596 int iwl_mvm_mac_ctxt_init(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1597 int iwl_mvm_mac_ctxt_add(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1598 int iwl_mvm_mac_ctxt_changed(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1599 bool force_assoc_off, const u8 *bssid_override);
1600 int iwl_mvm_mac_ctxt_remove(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1601 u32 iwl_mvm_mac_get_queues_mask(struct ieee80211_vif *vif);
1602 int iwl_mvm_mac_ctxt_beacon_changed(struct iwl_mvm *mvm,
1603 struct ieee80211_vif *vif);
1604 void iwl_mvm_rx_beacon_notif(struct iwl_mvm *mvm,
1605 struct iwl_rx_cmd_buffer *rxb);
1606 void iwl_mvm_rx_missed_beacons_notif(struct iwl_mvm *mvm,
1607 struct iwl_rx_cmd_buffer *rxb);
1608 void iwl_mvm_rx_stored_beacon_notif(struct iwl_mvm *mvm,
1609 struct iwl_rx_cmd_buffer *rxb);
1610 void iwl_mvm_mu_mimo_grp_notif(struct iwl_mvm *mvm,
1611 struct iwl_rx_cmd_buffer *rxb);
1612 void iwl_mvm_sta_pm_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1613 void iwl_mvm_window_status_notif(struct iwl_mvm *mvm,
1614 struct iwl_rx_cmd_buffer *rxb);
1615 void iwl_mvm_mac_ctxt_recalc_tsf_id(struct iwl_mvm *mvm,
1616 struct ieee80211_vif *vif);
1617 unsigned long iwl_mvm_get_used_hw_queues(struct iwl_mvm *mvm,
1618 struct ieee80211_vif *exclude_vif);
1619 void iwl_mvm_probe_resp_data_notif(struct iwl_mvm *mvm,
1620 struct iwl_rx_cmd_buffer *rxb);
1621 void iwl_mvm_channel_switch_noa_notif(struct iwl_mvm *mvm,
1622 struct iwl_rx_cmd_buffer *rxb);
1623 /* Bindings */
1624 int iwl_mvm_binding_add_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1625 int iwl_mvm_binding_remove_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1626
1627 /* Quota management */
1628 static inline size_t iwl_mvm_quota_cmd_size(struct iwl_mvm *mvm)
1629 {
1630 return iwl_mvm_has_quota_low_latency(mvm) ?
1631 sizeof(struct iwl_time_quota_cmd) :
1632 sizeof(struct iwl_time_quota_cmd_v1);
1633 }
1634
1635 static inline struct iwl_time_quota_data
1636 *iwl_mvm_quota_cmd_get_quota(struct iwl_mvm *mvm,
1637 struct iwl_time_quota_cmd *cmd,
1638 int i)
1639 {
1640 struct iwl_time_quota_data_v1 *quotas;
1641
1642 if (iwl_mvm_has_quota_low_latency(mvm))
1643 return &cmd->quotas[i];
1644
1645 quotas = (struct iwl_time_quota_data_v1 *)cmd->quotas;
1646 return (struct iwl_time_quota_data *)&quotas[i];
1647 }
1648
1649 int iwl_mvm_update_quotas(struct iwl_mvm *mvm, bool force_upload,
1650 struct ieee80211_vif *disabled_vif);
1651
1652 /* Scanning */
1653 int iwl_mvm_reg_scan_start(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1654 struct cfg80211_scan_request *req,
1655 struct ieee80211_scan_ies *ies);
1656 int iwl_mvm_scan_size(struct iwl_mvm *mvm);
1657 int iwl_mvm_scan_stop(struct iwl_mvm *mvm, int type, bool notify);
1658 int iwl_mvm_max_scan_ie_len(struct iwl_mvm *mvm);
1659 void iwl_mvm_report_scan_aborted(struct iwl_mvm *mvm);
1660 void iwl_mvm_scan_timeout_wk(struct work_struct *work);
1661
1662 /* Scheduled scan */
1663 void iwl_mvm_rx_lmac_scan_complete_notif(struct iwl_mvm *mvm,
1664 struct iwl_rx_cmd_buffer *rxb);
1665 void iwl_mvm_rx_lmac_scan_iter_complete_notif(struct iwl_mvm *mvm,
1666 struct iwl_rx_cmd_buffer *rxb);
1667 int iwl_mvm_sched_scan_start(struct iwl_mvm *mvm,
1668 struct ieee80211_vif *vif,
1669 struct cfg80211_sched_scan_request *req,
1670 struct ieee80211_scan_ies *ies,
1671 int type);
1672 void iwl_mvm_rx_scan_match_found(struct iwl_mvm *mvm,
1673 struct iwl_rx_cmd_buffer *rxb);
1674
1675 /* UMAC scan */
1676 int iwl_mvm_config_scan(struct iwl_mvm *mvm);
1677 void iwl_mvm_rx_umac_scan_complete_notif(struct iwl_mvm *mvm,
1678 struct iwl_rx_cmd_buffer *rxb);
1679 void iwl_mvm_rx_umac_scan_iter_complete_notif(struct iwl_mvm *mvm,
1680 struct iwl_rx_cmd_buffer *rxb);
1681
1682 /* MVM debugfs */
1683 #ifdef CONFIG_IWLWIFI_DEBUGFS
1684 int iwl_mvm_dbgfs_register(struct iwl_mvm *mvm, struct dentry *dbgfs_dir);
1685 void iwl_mvm_vif_dbgfs_register(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1686 void iwl_mvm_vif_dbgfs_clean(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1687 #else
1688 static inline int iwl_mvm_dbgfs_register(struct iwl_mvm *mvm,
1689 struct dentry *dbgfs_dir)
1690 {
1691 return 0;
1692 }
1693 static inline void
1694 iwl_mvm_vif_dbgfs_register(struct iwl_mvm *mvm, struct ieee80211_vif *vif)
1695 {
1696 }
1697 static inline void
1698 iwl_mvm_vif_dbgfs_clean(struct iwl_mvm *mvm, struct ieee80211_vif *vif)
1699 {
1700 }
1701 #endif /* CONFIG_IWLWIFI_DEBUGFS */
1702
1703 /* rate scaling */
1704 int iwl_mvm_send_lq_cmd(struct iwl_mvm *mvm, struct iwl_lq_cmd *lq, bool sync);
1705 void iwl_mvm_update_frame_stats(struct iwl_mvm *mvm, u32 rate, bool agg);
1706 int rs_pretty_print_rate(char *buf, int bufsz, const u32 rate);
1707 void rs_update_last_rssi(struct iwl_mvm *mvm,
1708 struct iwl_mvm_sta *mvmsta,
1709 struct ieee80211_rx_status *rx_status);
1710
1711 /* power management */
1712 int iwl_mvm_power_update_device(struct iwl_mvm *mvm);
1713 int iwl_mvm_power_update_mac(struct iwl_mvm *mvm);
1714 int iwl_mvm_power_update_ps(struct iwl_mvm *mvm);
1715 int iwl_mvm_power_mac_dbgfs_read(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1716 char *buf, int bufsz);
1717
1718 void iwl_mvm_power_vif_assoc(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1719 void iwl_mvm_power_uapsd_misbehaving_ap_notif(struct iwl_mvm *mvm,
1720 struct iwl_rx_cmd_buffer *rxb);
1721
1722 #ifdef CONFIG_IWLWIFI_LEDS
1723 int iwl_mvm_leds_init(struct iwl_mvm *mvm);
1724 void iwl_mvm_leds_exit(struct iwl_mvm *mvm);
1725 void iwl_mvm_leds_sync(struct iwl_mvm *mvm);
1726 #else
1727 static inline int iwl_mvm_leds_init(struct iwl_mvm *mvm)
1728 {
1729 return 0;
1730 }
1731 static inline void iwl_mvm_leds_exit(struct iwl_mvm *mvm)
1732 {
1733 }
1734 static inline void iwl_mvm_leds_sync(struct iwl_mvm *mvm)
1735 {
1736 }
1737 #endif
1738
1739 /* D3 (WoWLAN, NetDetect) */
1740 int iwl_mvm_suspend(struct ieee80211_hw *hw, struct cfg80211_wowlan *wowlan);
1741 int iwl_mvm_resume(struct ieee80211_hw *hw);
1742 void iwl_mvm_set_wakeup(struct ieee80211_hw *hw, bool enabled);
1743 void iwl_mvm_set_rekey_data(struct ieee80211_hw *hw,
1744 struct ieee80211_vif *vif,
1745 struct cfg80211_gtk_rekey_data *data);
1746 void iwl_mvm_ipv6_addr_change(struct ieee80211_hw *hw,
1747 struct ieee80211_vif *vif,
1748 struct inet6_dev *idev);
1749 void iwl_mvm_set_default_unicast_key(struct ieee80211_hw *hw,
1750 struct ieee80211_vif *vif, int idx);
1751 extern const struct file_operations iwl_dbgfs_d3_test_ops;
1752 struct iwl_wowlan_status *iwl_mvm_send_wowlan_get_status(struct iwl_mvm *mvm);
1753 #ifdef CONFIG_PM
1754 int iwl_mvm_wowlan_config_key_params(struct iwl_mvm *mvm,
1755 struct ieee80211_vif *vif,
1756 bool host_awake,
1757 u32 cmd_flags);
1758 void iwl_mvm_d0i3_update_keys(struct iwl_mvm *mvm,
1759 struct ieee80211_vif *vif,
1760 struct iwl_wowlan_status *status);
1761 void iwl_mvm_set_last_nonqos_seq(struct iwl_mvm *mvm,
1762 struct ieee80211_vif *vif);
1763 #else
1764 static inline int iwl_mvm_wowlan_config_key_params(struct iwl_mvm *mvm,
1765 struct ieee80211_vif *vif,
1766 bool host_awake,
1767 u32 cmd_flags)
1768 {
1769 return 0;
1770 }
1771
1772 static inline void iwl_mvm_d0i3_update_keys(struct iwl_mvm *mvm,
1773 struct ieee80211_vif *vif,
1774 struct iwl_wowlan_status *status)
1775 {
1776 }
1777
1778 static inline void
1779 iwl_mvm_set_last_nonqos_seq(struct iwl_mvm *mvm, struct ieee80211_vif *vif)
1780 {
1781 }
1782 #endif
1783 void iwl_mvm_set_wowlan_qos_seq(struct iwl_mvm_sta *mvm_ap_sta,
1784 struct iwl_wowlan_config_cmd *cmd);
1785 int iwl_mvm_send_proto_offload(struct iwl_mvm *mvm,
1786 struct ieee80211_vif *vif,
1787 bool disable_offloading,
1788 bool offload_ns,
1789 u32 cmd_flags);
1790
1791 /* D0i3 */
1792 void iwl_mvm_ref(struct iwl_mvm *mvm, enum iwl_mvm_ref_type ref_type);
1793 void iwl_mvm_unref(struct iwl_mvm *mvm, enum iwl_mvm_ref_type ref_type);
1794 int iwl_mvm_ref_sync(struct iwl_mvm *mvm, enum iwl_mvm_ref_type ref_type);
1795 bool iwl_mvm_ref_taken(struct iwl_mvm *mvm);
1796
1797 #ifdef CONFIG_PM
1798 void iwl_mvm_d0i3_enable_tx(struct iwl_mvm *mvm, __le16 *qos_seq);
1799 int iwl_mvm_enter_d0i3(struct iwl_op_mode *op_mode);
1800 int iwl_mvm_exit_d0i3(struct iwl_op_mode *op_mode);
1801 int _iwl_mvm_exit_d0i3(struct iwl_mvm *mvm);
1802 #endif
1803
1804 /* BT Coex */
1805 int iwl_mvm_send_bt_init_conf(struct iwl_mvm *mvm);
1806 void iwl_mvm_rx_bt_coex_notif(struct iwl_mvm *mvm,
1807 struct iwl_rx_cmd_buffer *rxb);
1808 void iwl_mvm_bt_rssi_event(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1809 enum ieee80211_rssi_event_data);
1810 void iwl_mvm_bt_coex_vif_change(struct iwl_mvm *mvm);
1811 u16 iwl_mvm_coex_agg_time_limit(struct iwl_mvm *mvm,
1812 struct ieee80211_sta *sta);
1813 bool iwl_mvm_bt_coex_is_mimo_allowed(struct iwl_mvm *mvm,
1814 struct ieee80211_sta *sta);
1815 bool iwl_mvm_bt_coex_is_ant_avail(struct iwl_mvm *mvm, u8 ant);
1816 bool iwl_mvm_bt_coex_is_shared_ant_avail(struct iwl_mvm *mvm);
1817 bool iwl_mvm_bt_coex_is_tpc_allowed(struct iwl_mvm *mvm,
1818 enum nl80211_band band);
1819 u8 iwl_mvm_bt_coex_get_single_ant_msk(struct iwl_mvm *mvm, u8 enabled_ants);
1820 u8 iwl_mvm_bt_coex_tx_prio(struct iwl_mvm *mvm, struct ieee80211_hdr *hdr,
1821 struct ieee80211_tx_info *info, u8 ac);
1822
1823 /* beacon filtering */
1824 #ifdef CONFIG_IWLWIFI_DEBUGFS
1825 void
1826 iwl_mvm_beacon_filter_debugfs_parameters(struct ieee80211_vif *vif,
1827 struct iwl_beacon_filter_cmd *cmd);
1828 #else
1829 static inline void
1830 iwl_mvm_beacon_filter_debugfs_parameters(struct ieee80211_vif *vif,
1831 struct iwl_beacon_filter_cmd *cmd)
1832 {}
1833 #endif
1834 int iwl_mvm_update_d0i3_power_mode(struct iwl_mvm *mvm,
1835 struct ieee80211_vif *vif,
1836 bool enable, u32 flags);
1837 int iwl_mvm_enable_beacon_filter(struct iwl_mvm *mvm,
1838 struct ieee80211_vif *vif,
1839 u32 flags);
1840 int iwl_mvm_disable_beacon_filter(struct iwl_mvm *mvm,
1841 struct ieee80211_vif *vif,
1842 u32 flags);
1843 /* SMPS */
1844 void iwl_mvm_update_smps(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1845 enum iwl_mvm_smps_type_request req_type,
1846 enum ieee80211_smps_mode smps_request);
1847 bool iwl_mvm_rx_diversity_allowed(struct iwl_mvm *mvm);
1848
1849 /* Low latency */
1850 int iwl_mvm_update_low_latency(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1851 bool low_latency,
1852 enum iwl_mvm_low_latency_cause cause);
1853 /* get SystemLowLatencyMode - only needed for beacon threshold? */
1854 bool iwl_mvm_low_latency(struct iwl_mvm *mvm);
1855 bool iwl_mvm_low_latency_band(struct iwl_mvm *mvm, enum nl80211_band band);
1856
1857 /* get VMACLowLatencyMode */
1858 static inline bool iwl_mvm_vif_low_latency(struct iwl_mvm_vif *mvmvif)
1859 {
1860 /*
1861 * should this consider associated/active/... state?
1862 *
1863 * Normally low-latency should only be active on interfaces
1864 * that are active, but at least with debugfs it can also be
1865 * enabled on interfaces that aren't active. However, when
1866 * interface aren't active then they aren't added into the
1867 * binding, so this has no real impact. For now, just return
1868 * the current desired low-latency state.
1869 */
1870 return mvmvif->low_latency;
1871 }
1872
1873 static inline
1874 void iwl_mvm_vif_set_low_latency(struct iwl_mvm_vif *mvmvif, bool set,
1875 enum iwl_mvm_low_latency_cause cause)
1876 {
1877 if (set)
1878 mvmvif->low_latency |= cause;
1879 else
1880 mvmvif->low_latency &= ~cause;
1881 }
1882
1883 /* Return a bitmask with all the hw supported queues, except for the
1884 * command queue, which can't be flushed.
1885 */
1886 static inline u32 iwl_mvm_flushable_queues(struct iwl_mvm *mvm)
1887 {
1888 return ((BIT(mvm->cfg->base_params->num_of_queues) - 1) &
1889 ~BIT(IWL_MVM_DQA_CMD_QUEUE));
1890 }
1891
1892 static inline void iwl_mvm_stop_device(struct iwl_mvm *mvm)
1893 {
1894 lockdep_assert_held(&mvm->mutex);
1895 /* calling this function without using dump_start/end since at this
1896 * point we already hold the op mode mutex
1897 */
1898 iwl_fw_dbg_collect_sync(&mvm->fwrt);
1899 iwl_fw_cancel_timestamp(&mvm->fwrt);
1900 iwl_free_fw_paging(&mvm->fwrt);
1901 clear_bit(IWL_MVM_STATUS_FIRMWARE_RUNNING, &mvm->status);
1902 iwl_fw_dump_conf_clear(&mvm->fwrt);
1903 iwl_trans_stop_device(mvm->trans);
1904 }
1905
1906 /* Stop/start all mac queues in a given bitmap */
1907 void iwl_mvm_start_mac_queues(struct iwl_mvm *mvm, unsigned long mq);
1908 void iwl_mvm_stop_mac_queues(struct iwl_mvm *mvm, unsigned long mq);
1909
1910 /* Re-configure the SCD for a queue that has already been configured */
1911 int iwl_mvm_reconfig_scd(struct iwl_mvm *mvm, int queue, int fifo, int sta_id,
1912 int tid, int frame_limit, u16 ssn);
1913
1914 /* Thermal management and CT-kill */
1915 void iwl_mvm_tt_tx_backoff(struct iwl_mvm *mvm, u32 backoff);
1916 void iwl_mvm_tt_temp_changed(struct iwl_mvm *mvm, u32 temp);
1917 void iwl_mvm_temp_notif(struct iwl_mvm *mvm,
1918 struct iwl_rx_cmd_buffer *rxb);
1919 void iwl_mvm_tt_handler(struct iwl_mvm *mvm);
1920 void iwl_mvm_thermal_initialize(struct iwl_mvm *mvm, u32 min_backoff);
1921 void iwl_mvm_thermal_exit(struct iwl_mvm *mvm);
1922 void iwl_mvm_set_hw_ctkill_state(struct iwl_mvm *mvm, bool state);
1923 int iwl_mvm_get_temp(struct iwl_mvm *mvm, s32 *temp);
1924 void iwl_mvm_ct_kill_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1925 void iwl_mvm_enter_ctkill(struct iwl_mvm *mvm);
1926 int iwl_mvm_send_temp_report_ths_cmd(struct iwl_mvm *mvm);
1927 int iwl_mvm_ctdp_command(struct iwl_mvm *mvm, u32 op, u32 budget);
1928
1929 /* Location Aware Regulatory */
1930 struct iwl_mcc_update_resp *
1931 iwl_mvm_update_mcc(struct iwl_mvm *mvm, const char *alpha2,
1932 enum iwl_mcc_source src_id);
1933 int iwl_mvm_init_mcc(struct iwl_mvm *mvm);
1934 void iwl_mvm_rx_chub_update_mcc(struct iwl_mvm *mvm,
1935 struct iwl_rx_cmd_buffer *rxb);
1936 struct ieee80211_regdomain *iwl_mvm_get_regdomain(struct wiphy *wiphy,
1937 const char *alpha2,
1938 enum iwl_mcc_source src_id,
1939 bool *changed);
1940 struct ieee80211_regdomain *iwl_mvm_get_current_regdomain(struct iwl_mvm *mvm,
1941 bool *changed);
1942 int iwl_mvm_init_fw_regd(struct iwl_mvm *mvm);
1943 void iwl_mvm_update_changed_regdom(struct iwl_mvm *mvm);
1944
1945 /* smart fifo */
1946 int iwl_mvm_sf_update(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1947 bool added_vif);
1948
1949 /* TDLS */
1950
1951 /*
1952 * We use TID 4 (VI) as a FW-used-only TID when TDLS connections are present.
1953 * This TID is marked as used vs the AP and all connected TDLS peers.
1954 */
1955 #define IWL_MVM_TDLS_FW_TID 4
1956
1957 int iwl_mvm_tdls_sta_count(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1958 void iwl_mvm_teardown_tdls_peers(struct iwl_mvm *mvm);
1959 void iwl_mvm_recalc_tdls_state(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1960 bool sta_added);
1961 void iwl_mvm_mac_mgd_protect_tdls_discover(struct ieee80211_hw *hw,
1962 struct ieee80211_vif *vif);
1963 int iwl_mvm_tdls_channel_switch(struct ieee80211_hw *hw,
1964 struct ieee80211_vif *vif,
1965 struct ieee80211_sta *sta, u8 oper_class,
1966 struct cfg80211_chan_def *chandef,
1967 struct sk_buff *tmpl_skb, u32 ch_sw_tm_ie);
1968 void iwl_mvm_tdls_recv_channel_switch(struct ieee80211_hw *hw,
1969 struct ieee80211_vif *vif,
1970 struct ieee80211_tdls_ch_sw_params *params);
1971 void iwl_mvm_tdls_cancel_channel_switch(struct ieee80211_hw *hw,
1972 struct ieee80211_vif *vif,
1973 struct ieee80211_sta *sta);
1974 void iwl_mvm_rx_tdls_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1975 void iwl_mvm_tdls_ch_switch_work(struct work_struct *work);
1976
1977 void iwl_mvm_sync_rx_queues_internal(struct iwl_mvm *mvm,
1978 struct iwl_mvm_internal_rxq_notif *notif,
1979 u32 size);
1980 void iwl_mvm_reorder_timer_expired(struct timer_list *t);
1981 struct ieee80211_vif *iwl_mvm_get_bss_vif(struct iwl_mvm *mvm);
1982 bool iwl_mvm_is_vif_assoc(struct iwl_mvm *mvm);
1983
1984 #define MVM_TCM_PERIOD_MSEC 500
1985 #define MVM_TCM_PERIOD (HZ * MVM_TCM_PERIOD_MSEC / 1000)
1986 #define MVM_LL_PERIOD (10 * HZ)
1987 void iwl_mvm_tcm_work(struct work_struct *work);
1988 void iwl_mvm_recalc_tcm(struct iwl_mvm *mvm);
1989 void iwl_mvm_pause_tcm(struct iwl_mvm *mvm, bool with_cancel);
1990 void iwl_mvm_resume_tcm(struct iwl_mvm *mvm);
1991 void iwl_mvm_tcm_add_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1992 void iwl_mvm_tcm_rm_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1993 u8 iwl_mvm_tcm_load_percentage(u32 airtime, u32 elapsed);
1994
1995 void iwl_mvm_nic_restart(struct iwl_mvm *mvm, bool fw_error);
1996 unsigned int iwl_mvm_get_wd_timeout(struct iwl_mvm *mvm,
1997 struct ieee80211_vif *vif,
1998 bool tdls, bool cmd_q);
1999 void iwl_mvm_connection_loss(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
2000 const char *errmsg);
2001 void iwl_mvm_event_frame_timeout_callback(struct iwl_mvm *mvm,
2002 struct ieee80211_vif *vif,
2003 const struct ieee80211_sta *sta,
2004 u16 tid);
2005
2006 int iwl_mvm_sar_select_profile(struct iwl_mvm *mvm, int prof_a, int prof_b);
2007 int iwl_mvm_get_sar_geo_profile(struct iwl_mvm *mvm);
2008 #ifdef CONFIG_IWLWIFI_DEBUGFS
2009 void iwl_mvm_sta_add_debugfs(struct ieee80211_hw *hw,
2010 struct ieee80211_vif *vif,
2011 struct ieee80211_sta *sta,
2012 struct dentry *dir);
2013 #endif
2014
2015 #endif /* __IWL_MVM_H__ */