]> git.proxmox.com Git - mirror_ubuntu-bionic-kernel.git/blob - drivers/net/wireless/intel/iwlwifi/mvm/mvm.h
iwlwifi: refactor out paging code
[mirror_ubuntu-bionic-kernel.git] / drivers / net / wireless / intel / iwlwifi / mvm / mvm.h
1 /******************************************************************************
2 *
3 * This file is provided under a dual BSD/GPLv2 license. When using or
4 * redistributing this file, you may do so under either license.
5 *
6 * GPL LICENSE SUMMARY
7 *
8 * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved.
9 * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH
10 * Copyright(c) 2016 - 2017 Intel Deutschland GmbH
11 *
12 * This program is free software; you can redistribute it and/or modify
13 * it under the terms of version 2 of the GNU General Public License as
14 * published by the Free Software Foundation.
15 *
16 * This program is distributed in the hope that it will be useful, but
17 * WITHOUT ANY WARRANTY; without even the implied warranty of
18 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
19 * General Public License for more details.
20 *
21 * You should have received a copy of the GNU General Public License
22 * along with this program; if not, write to the Free Software
23 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110,
24 * USA
25 *
26 * The full GNU General Public License is included in this distribution
27 * in the file called COPYING.
28 *
29 * Contact Information:
30 * Intel Linux Wireless <linuxwifi@intel.com>
31 * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
32 *
33 * BSD LICENSE
34 *
35 * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved.
36 * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH
37 * Copyright(c) 2016 - 2017 Intel Deutschland GmbH
38 * All rights reserved.
39 *
40 * Redistribution and use in source and binary forms, with or without
41 * modification, are permitted provided that the following conditions
42 * are met:
43 *
44 * * Redistributions of source code must retain the above copyright
45 * notice, this list of conditions and the following disclaimer.
46 * * Redistributions in binary form must reproduce the above copyright
47 * notice, this list of conditions and the following disclaimer in
48 * the documentation and/or other materials provided with the
49 * distribution.
50 * * Neither the name Intel Corporation nor the names of its
51 * contributors may be used to endorse or promote products derived
52 * from this software without specific prior written permission.
53 *
54 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
55 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
56 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
57 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
58 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
59 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
60 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
61 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
62 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
63 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
64 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
65 *
66 *****************************************************************************/
67
68 #ifndef __IWL_MVM_H__
69 #define __IWL_MVM_H__
70
71 #include <linux/list.h>
72 #include <linux/spinlock.h>
73 #include <linux/leds.h>
74 #include <linux/in6.h>
75
76 #ifdef CONFIG_THERMAL
77 #include <linux/thermal.h>
78 #endif
79
80 #include "iwl-op-mode.h"
81 #include "iwl-trans.h"
82 #include "fw/notif-wait.h"
83 #include "iwl-eeprom-parse.h"
84 #include "fw/file.h"
85 #include "iwl-config.h"
86 #include "sta.h"
87 #include "fw-api.h"
88 #include "constants.h"
89 #include "tof.h"
90 #include "fw/runtime.h"
91
92 #define IWL_MVM_MAX_ADDRESSES 5
93 /* RSSI offset for WkP */
94 #define IWL_RSSI_OFFSET 50
95 #define IWL_MVM_MISSED_BEACONS_THRESHOLD 8
96 /* A TimeUnit is 1024 microsecond */
97 #define MSEC_TO_TU(_msec) (_msec*1000/1024)
98
99 /* For GO, this value represents the number of TUs before CSA "beacon
100 * 0" TBTT when the CSA time-event needs to be scheduled to start. It
101 * must be big enough to ensure that we switch in time.
102 */
103 #define IWL_MVM_CHANNEL_SWITCH_TIME_GO 40
104
105 /* For client, this value represents the number of TUs before CSA
106 * "beacon 1" TBTT, instead. This is because we don't know when the
107 * GO/AP will be in the new channel, so we switch early enough.
108 */
109 #define IWL_MVM_CHANNEL_SWITCH_TIME_CLIENT 10
110
111 /*
112 * This value (in TUs) is used to fine tune the CSA NoA end time which should
113 * be just before "beacon 0" TBTT.
114 */
115 #define IWL_MVM_CHANNEL_SWITCH_MARGIN 4
116
117 /*
118 * Number of beacons to transmit on a new channel until we unblock tx to
119 * the stations, even if we didn't identify them on a new channel
120 */
121 #define IWL_MVM_CS_UNBLOCK_TX_TIMEOUT 3
122
123 extern const struct ieee80211_ops iwl_mvm_hw_ops;
124
125 /**
126 * struct iwl_mvm_mod_params - module parameters for iwlmvm
127 * @init_dbg: if true, then the NIC won't be stopped if the INIT fw asserted.
128 * We will register to mac80211 to have testmode working. The NIC must not
129 * be up'ed after the INIT fw asserted. This is useful to be able to use
130 * proprietary tools over testmode to debug the INIT fw.
131 * @tfd_q_hang_detect: enabled the detection of hung transmit queues
132 * @power_scheme: one of enum iwl_power_scheme
133 */
134 struct iwl_mvm_mod_params {
135 bool init_dbg;
136 bool tfd_q_hang_detect;
137 int power_scheme;
138 };
139 extern struct iwl_mvm_mod_params iwlmvm_mod_params;
140
141 /**
142 * struct iwl_mvm_dump_ptrs - set of pointers needed for the fw-error-dump
143 *
144 * @op_mode_ptr: pointer to the buffer coming from the mvm op_mode
145 * @trans_ptr: pointer to struct %iwl_trans_dump_data which contains the
146 * transport's data.
147 * @trans_len: length of the valid data in trans_ptr
148 * @op_mode_len: length of the valid data in op_mode_ptr
149 */
150 struct iwl_mvm_dump_ptrs {
151 struct iwl_trans_dump_data *trans_ptr;
152 void *op_mode_ptr;
153 u32 op_mode_len;
154 };
155
156 /**
157 * struct iwl_mvm_dump_desc - describes the dump
158 * @len: length of trig_desc->data
159 * @trig_desc: the description of the dump
160 */
161 struct iwl_mvm_dump_desc {
162 size_t len;
163 /* must be last */
164 struct iwl_fw_error_dump_trigger_desc trig_desc;
165 };
166
167 extern const struct iwl_mvm_dump_desc iwl_mvm_dump_desc_assert;
168
169 struct iwl_mvm_phy_ctxt {
170 u16 id;
171 u16 color;
172 u32 ref;
173
174 /*
175 * TODO: This should probably be removed. Currently here only for rate
176 * scaling algorithm
177 */
178 struct ieee80211_channel *channel;
179 };
180
181 struct iwl_mvm_time_event_data {
182 struct ieee80211_vif *vif;
183 struct list_head list;
184 unsigned long end_jiffies;
185 u32 duration;
186 bool running;
187 u32 uid;
188
189 /*
190 * The access to the 'id' field must be done when the
191 * mvm->time_event_lock is held, as it value is used to indicate
192 * if the te is in the time event list or not (when id == TE_MAX)
193 */
194 u32 id;
195 };
196
197 /* Power management */
198
199 /**
200 * enum iwl_power_scheme
201 * @IWL_POWER_LEVEL_CAM - Continuously Active Mode
202 * @IWL_POWER_LEVEL_BPS - Balanced Power Save (default)
203 * @IWL_POWER_LEVEL_LP - Low Power
204 */
205 enum iwl_power_scheme {
206 IWL_POWER_SCHEME_CAM = 1,
207 IWL_POWER_SCHEME_BPS,
208 IWL_POWER_SCHEME_LP
209 };
210
211 #define IWL_CONN_MAX_LISTEN_INTERVAL 10
212 #define IWL_UAPSD_MAX_SP IEEE80211_WMM_IE_STA_QOSINFO_SP_ALL
213
214 #ifdef CONFIG_IWLWIFI_DEBUGFS
215 enum iwl_dbgfs_pm_mask {
216 MVM_DEBUGFS_PM_KEEP_ALIVE = BIT(0),
217 MVM_DEBUGFS_PM_SKIP_OVER_DTIM = BIT(1),
218 MVM_DEBUGFS_PM_SKIP_DTIM_PERIODS = BIT(2),
219 MVM_DEBUGFS_PM_RX_DATA_TIMEOUT = BIT(3),
220 MVM_DEBUGFS_PM_TX_DATA_TIMEOUT = BIT(4),
221 MVM_DEBUGFS_PM_LPRX_ENA = BIT(6),
222 MVM_DEBUGFS_PM_LPRX_RSSI_THRESHOLD = BIT(7),
223 MVM_DEBUGFS_PM_SNOOZE_ENABLE = BIT(8),
224 MVM_DEBUGFS_PM_UAPSD_MISBEHAVING = BIT(9),
225 MVM_DEBUGFS_PM_USE_PS_POLL = BIT(10),
226 };
227
228 struct iwl_dbgfs_pm {
229 u16 keep_alive_seconds;
230 u32 rx_data_timeout;
231 u32 tx_data_timeout;
232 bool skip_over_dtim;
233 u8 skip_dtim_periods;
234 bool lprx_ena;
235 u32 lprx_rssi_threshold;
236 bool snooze_ena;
237 bool uapsd_misbehaving;
238 bool use_ps_poll;
239 int mask;
240 };
241
242 /* beacon filtering */
243
244 enum iwl_dbgfs_bf_mask {
245 MVM_DEBUGFS_BF_ENERGY_DELTA = BIT(0),
246 MVM_DEBUGFS_BF_ROAMING_ENERGY_DELTA = BIT(1),
247 MVM_DEBUGFS_BF_ROAMING_STATE = BIT(2),
248 MVM_DEBUGFS_BF_TEMP_THRESHOLD = BIT(3),
249 MVM_DEBUGFS_BF_TEMP_FAST_FILTER = BIT(4),
250 MVM_DEBUGFS_BF_TEMP_SLOW_FILTER = BIT(5),
251 MVM_DEBUGFS_BF_ENABLE_BEACON_FILTER = BIT(6),
252 MVM_DEBUGFS_BF_DEBUG_FLAG = BIT(7),
253 MVM_DEBUGFS_BF_ESCAPE_TIMER = BIT(8),
254 MVM_DEBUGFS_BA_ESCAPE_TIMER = BIT(9),
255 MVM_DEBUGFS_BA_ENABLE_BEACON_ABORT = BIT(10),
256 };
257
258 struct iwl_dbgfs_bf {
259 u32 bf_energy_delta;
260 u32 bf_roaming_energy_delta;
261 u32 bf_roaming_state;
262 u32 bf_temp_threshold;
263 u32 bf_temp_fast_filter;
264 u32 bf_temp_slow_filter;
265 u32 bf_enable_beacon_filter;
266 u32 bf_debug_flag;
267 u32 bf_escape_timer;
268 u32 ba_escape_timer;
269 u32 ba_enable_beacon_abort;
270 int mask;
271 };
272 #endif
273
274 enum iwl_mvm_smps_type_request {
275 IWL_MVM_SMPS_REQ_BT_COEX,
276 IWL_MVM_SMPS_REQ_TT,
277 IWL_MVM_SMPS_REQ_PROT,
278 NUM_IWL_MVM_SMPS_REQ,
279 };
280
281 enum iwl_mvm_ref_type {
282 IWL_MVM_REF_UCODE_DOWN,
283 IWL_MVM_REF_SCAN,
284 IWL_MVM_REF_ROC,
285 IWL_MVM_REF_ROC_AUX,
286 IWL_MVM_REF_P2P_CLIENT,
287 IWL_MVM_REF_AP_IBSS,
288 IWL_MVM_REF_USER,
289 IWL_MVM_REF_TX,
290 IWL_MVM_REF_TX_AGG,
291 IWL_MVM_REF_ADD_IF,
292 IWL_MVM_REF_START_AP,
293 IWL_MVM_REF_BSS_CHANGED,
294 IWL_MVM_REF_PREPARE_TX,
295 IWL_MVM_REF_PROTECT_TDLS,
296 IWL_MVM_REF_CHECK_CTKILL,
297 IWL_MVM_REF_PRPH_READ,
298 IWL_MVM_REF_PRPH_WRITE,
299 IWL_MVM_REF_NMI,
300 IWL_MVM_REF_TM_CMD,
301 IWL_MVM_REF_EXIT_WORK,
302 IWL_MVM_REF_PROTECT_CSA,
303 IWL_MVM_REF_FW_DBG_COLLECT,
304 IWL_MVM_REF_INIT_UCODE,
305 IWL_MVM_REF_SENDING_CMD,
306 IWL_MVM_REF_RX,
307
308 /* update debugfs.c when changing this */
309
310 IWL_MVM_REF_COUNT,
311 };
312
313 enum iwl_bt_force_ant_mode {
314 BT_FORCE_ANT_DIS = 0,
315 BT_FORCE_ANT_AUTO,
316 BT_FORCE_ANT_BT,
317 BT_FORCE_ANT_WIFI,
318
319 BT_FORCE_ANT_MAX,
320 };
321
322 /**
323 * struct iwl_mvm_vif_bf_data - beacon filtering related data
324 * @bf_enabled: indicates if beacon filtering is enabled
325 * @ba_enabled: indicated if beacon abort is enabled
326 * @ave_beacon_signal: average beacon signal
327 * @last_cqm_event: rssi of the last cqm event
328 * @bt_coex_min_thold: minimum threshold for BT coex
329 * @bt_coex_max_thold: maximum threshold for BT coex
330 * @last_bt_coex_event: rssi of the last BT coex event
331 */
332 struct iwl_mvm_vif_bf_data {
333 bool bf_enabled;
334 bool ba_enabled;
335 int ave_beacon_signal;
336 int last_cqm_event;
337 int bt_coex_min_thold;
338 int bt_coex_max_thold;
339 int last_bt_coex_event;
340 };
341
342 /**
343 * struct iwl_mvm_vif - data per Virtual Interface, it is a MAC context
344 * @id: between 0 and 3
345 * @color: to solve races upon MAC addition and removal
346 * @ap_sta_id: the sta_id of the AP - valid only if VIF type is STA
347 * @bssid: BSSID for this (client) interface
348 * @associated: indicates that we're currently associated, used only for
349 * managing the firmware state in iwl_mvm_bss_info_changed_station()
350 * @ap_assoc_sta_count: count of stations associated to us - valid only
351 * if VIF type is AP
352 * @uploaded: indicates the MAC context has been added to the device
353 * @ap_ibss_active: indicates that AP/IBSS is configured and that the interface
354 * should get quota etc.
355 * @pm_enabled - Indicate if MAC power management is allowed
356 * @monitor_active: indicates that monitor context is configured, and that the
357 * interface should get quota etc.
358 * @low_latency_traffic: indicates low latency traffic was detected
359 * @low_latency_dbgfs: low latency mode set from debugfs
360 * @low_latency_vcmd: low latency mode set from vendor command
361 * @ps_disabled: indicates that this interface requires PS to be disabled
362 * @queue_params: QoS params for this MAC
363 * @bcast_sta: station used for broadcast packets. Used by the following
364 * vifs: P2P_DEVICE, GO and AP.
365 * @beacon_skb: the skb used to hold the AP/GO beacon template
366 * @smps_requests: the SMPS requests of different parts of the driver,
367 * combined on update to yield the overall request to mac80211.
368 * @beacon_stats: beacon statistics, containing the # of received beacons,
369 * # of received beacons accumulated over FW restart, and the current
370 * average signal of beacons retrieved from the firmware
371 * @csa_failed: CSA failed to schedule time event, report an error later
372 * @features: hw features active for this vif
373 */
374 struct iwl_mvm_vif {
375 struct iwl_mvm *mvm;
376 u16 id;
377 u16 color;
378 u8 ap_sta_id;
379
380 u8 bssid[ETH_ALEN];
381 bool associated;
382 u8 ap_assoc_sta_count;
383
384 u16 cab_queue;
385
386 bool uploaded;
387 bool ap_ibss_active;
388 bool pm_enabled;
389 bool monitor_active;
390 bool low_latency_traffic, low_latency_dbgfs, low_latency_vcmd;
391 bool ps_disabled;
392 struct iwl_mvm_vif_bf_data bf_data;
393
394 struct {
395 u32 num_beacons, accu_num_beacons;
396 u8 avg_signal;
397 } beacon_stats;
398
399 u32 ap_beacon_time;
400
401 enum iwl_tsf_id tsf_id;
402
403 /*
404 * QoS data from mac80211, need to store this here
405 * as mac80211 has a separate callback but we need
406 * to have the data for the MAC context
407 */
408 struct ieee80211_tx_queue_params queue_params[IEEE80211_NUM_ACS];
409 struct iwl_mvm_time_event_data time_event_data;
410 struct iwl_mvm_time_event_data hs_time_event_data;
411
412 struct iwl_mvm_int_sta bcast_sta;
413 struct iwl_mvm_int_sta mcast_sta;
414
415 /*
416 * Assigned while mac80211 has the interface in a channel context,
417 * or, for P2P Device, while it exists.
418 */
419 struct iwl_mvm_phy_ctxt *phy_ctxt;
420
421 #ifdef CONFIG_PM
422 /* WoWLAN GTK rekey data */
423 struct {
424 u8 kck[NL80211_KCK_LEN], kek[NL80211_KEK_LEN];
425 __le64 replay_ctr;
426 bool valid;
427 } rekey_data;
428
429 int tx_key_idx;
430
431 bool seqno_valid;
432 u16 seqno;
433 #endif
434
435 #if IS_ENABLED(CONFIG_IPV6)
436 /* IPv6 addresses for WoWLAN */
437 struct in6_addr target_ipv6_addrs[IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_MAX];
438 unsigned long tentative_addrs[BITS_TO_LONGS(IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_MAX)];
439 int num_target_ipv6_addrs;
440 #endif
441
442 #ifdef CONFIG_IWLWIFI_DEBUGFS
443 struct dentry *dbgfs_dir;
444 struct dentry *dbgfs_slink;
445 struct iwl_dbgfs_pm dbgfs_pm;
446 struct iwl_dbgfs_bf dbgfs_bf;
447 struct iwl_mac_power_cmd mac_pwr_cmd;
448 int dbgfs_quota_min;
449 #endif
450
451 enum ieee80211_smps_mode smps_requests[NUM_IWL_MVM_SMPS_REQ];
452
453 /* FW identified misbehaving AP */
454 u8 uapsd_misbehaving_bssid[ETH_ALEN];
455
456 /* Indicates that CSA countdown may be started */
457 bool csa_countdown;
458 bool csa_failed;
459 u16 csa_target_freq;
460
461 /* TCP Checksum Offload */
462 netdev_features_t features;
463
464 /*
465 * link quality measurement - used to check whether this interface
466 * is in the middle of a link quality measurement
467 */
468 bool lqm_active;
469 };
470
471 static inline struct iwl_mvm_vif *
472 iwl_mvm_vif_from_mac80211(struct ieee80211_vif *vif)
473 {
474 if (!vif)
475 return NULL;
476 return (void *)vif->drv_priv;
477 }
478
479 extern const u8 tid_to_mac80211_ac[];
480
481 #define IWL_MVM_SCAN_STOPPING_SHIFT 8
482
483 enum iwl_scan_status {
484 IWL_MVM_SCAN_REGULAR = BIT(0),
485 IWL_MVM_SCAN_SCHED = BIT(1),
486 IWL_MVM_SCAN_NETDETECT = BIT(2),
487
488 IWL_MVM_SCAN_STOPPING_REGULAR = BIT(8),
489 IWL_MVM_SCAN_STOPPING_SCHED = BIT(9),
490 IWL_MVM_SCAN_STOPPING_NETDETECT = BIT(10),
491
492 IWL_MVM_SCAN_REGULAR_MASK = IWL_MVM_SCAN_REGULAR |
493 IWL_MVM_SCAN_STOPPING_REGULAR,
494 IWL_MVM_SCAN_SCHED_MASK = IWL_MVM_SCAN_SCHED |
495 IWL_MVM_SCAN_STOPPING_SCHED,
496 IWL_MVM_SCAN_NETDETECT_MASK = IWL_MVM_SCAN_NETDETECT |
497 IWL_MVM_SCAN_STOPPING_NETDETECT,
498
499 IWL_MVM_SCAN_STOPPING_MASK = 0xff << IWL_MVM_SCAN_STOPPING_SHIFT,
500 IWL_MVM_SCAN_MASK = 0xff,
501 };
502
503 enum iwl_mvm_scan_type {
504 IWL_SCAN_TYPE_NOT_SET,
505 IWL_SCAN_TYPE_UNASSOC,
506 IWL_SCAN_TYPE_WILD,
507 IWL_SCAN_TYPE_MILD,
508 IWL_SCAN_TYPE_FRAGMENTED,
509 };
510
511 enum iwl_mvm_sched_scan_pass_all_states {
512 SCHED_SCAN_PASS_ALL_DISABLED,
513 SCHED_SCAN_PASS_ALL_ENABLED,
514 SCHED_SCAN_PASS_ALL_FOUND,
515 };
516
517 /**
518 * struct iwl_nvm_section - describes an NVM section in memory.
519 *
520 * This struct holds an NVM section read from the NIC using NVM_ACCESS_CMD,
521 * and saved for later use by the driver. Not all NVM sections are saved
522 * this way, only the needed ones.
523 */
524 struct iwl_nvm_section {
525 u16 length;
526 const u8 *data;
527 };
528
529 /**
530 * struct iwl_mvm_tt_mgnt - Thermal Throttling Management structure
531 * @ct_kill_exit: worker to exit thermal kill
532 * @dynamic_smps: Is thermal throttling enabled dynamic_smps?
533 * @tx_backoff: The current thremal throttling tx backoff in uSec.
534 * @min_backoff: The minimal tx backoff due to power restrictions
535 * @params: Parameters to configure the thermal throttling algorithm.
536 * @throttle: Is thermal throttling is active?
537 */
538 struct iwl_mvm_tt_mgmt {
539 struct delayed_work ct_kill_exit;
540 bool dynamic_smps;
541 u32 tx_backoff;
542 u32 min_backoff;
543 struct iwl_tt_params params;
544 bool throttle;
545 };
546
547 #ifdef CONFIG_THERMAL
548 /**
549 *struct iwl_mvm_thermal_device - thermal zone related data
550 * @temp_trips: temperature thresholds for report
551 * @fw_trips_index: keep indexes to original array - temp_trips
552 * @tzone: thermal zone device data
553 */
554 struct iwl_mvm_thermal_device {
555 s16 temp_trips[IWL_MAX_DTS_TRIPS];
556 u8 fw_trips_index[IWL_MAX_DTS_TRIPS];
557 struct thermal_zone_device *tzone;
558 };
559
560 /*
561 * struct iwl_mvm_cooling_device
562 * @cur_state: current state
563 * @cdev: struct thermal cooling device
564 */
565 struct iwl_mvm_cooling_device {
566 u32 cur_state;
567 struct thermal_cooling_device *cdev;
568 };
569 #endif
570
571 #define IWL_MVM_NUM_LAST_FRAMES_UCODE_RATES 8
572
573 struct iwl_mvm_frame_stats {
574 u32 legacy_frames;
575 u32 ht_frames;
576 u32 vht_frames;
577 u32 bw_20_frames;
578 u32 bw_40_frames;
579 u32 bw_80_frames;
580 u32 bw_160_frames;
581 u32 sgi_frames;
582 u32 ngi_frames;
583 u32 siso_frames;
584 u32 mimo2_frames;
585 u32 agg_frames;
586 u32 ampdu_count;
587 u32 success_frames;
588 u32 fail_frames;
589 u32 last_rates[IWL_MVM_NUM_LAST_FRAMES_UCODE_RATES];
590 int last_frame_idx;
591 };
592
593 enum {
594 D0I3_DEFER_WAKEUP,
595 D0I3_PENDING_WAKEUP,
596 };
597
598 #define IWL_MVM_DEBUG_SET_TEMPERATURE_DISABLE 0xff
599 #define IWL_MVM_DEBUG_SET_TEMPERATURE_MIN -100
600 #define IWL_MVM_DEBUG_SET_TEMPERATURE_MAX 200
601
602 enum iwl_mvm_tdls_cs_state {
603 IWL_MVM_TDLS_SW_IDLE = 0,
604 IWL_MVM_TDLS_SW_REQ_SENT,
605 IWL_MVM_TDLS_SW_RESP_RCVD,
606 IWL_MVM_TDLS_SW_REQ_RCVD,
607 IWL_MVM_TDLS_SW_ACTIVE,
608 };
609
610 #define MAX_NUM_LMAC 2
611 struct iwl_mvm_shared_mem_cfg {
612 int num_lmacs;
613 int num_txfifo_entries;
614 struct {
615 u32 txfifo_size[TX_FIFO_MAX_NUM];
616 u32 rxfifo1_size;
617 } lmac[MAX_NUM_LMAC];
618 u32 rxfifo2_size;
619 u32 internal_txfifo_addr;
620 u32 internal_txfifo_size[TX_FIFO_INTERNAL_MAX_NUM];
621 };
622
623 /**
624 * struct iwl_mvm_reorder_buffer - per ra/tid/queue reorder buffer
625 * @head_sn: reorder window head sn
626 * @num_stored: number of mpdus stored in the buffer
627 * @buf_size: the reorder buffer size as set by the last addba request
628 * @sta_id: sta id of this reorder buffer
629 * @queue: queue of this reorder buffer
630 * @last_amsdu: track last ASMDU SN for duplication detection
631 * @last_sub_index: track ASMDU sub frame index for duplication detection
632 * @entries: list of skbs stored
633 * @reorder_time: time the packet was stored in the reorder buffer
634 * @reorder_timer: timer for frames are in the reorder buffer. For AMSDU
635 * it is the time of last received sub-frame
636 * @removed: prevent timer re-arming
637 * @valid: reordering is valid for this queue
638 * @lock: protect reorder buffer internal state
639 * @mvm: mvm pointer, needed for frame timer context
640 */
641 struct iwl_mvm_reorder_buffer {
642 u16 head_sn;
643 u16 num_stored;
644 u8 buf_size;
645 u8 sta_id;
646 int queue;
647 u16 last_amsdu;
648 u8 last_sub_index;
649 struct sk_buff_head entries[IEEE80211_MAX_AMPDU_BUF];
650 unsigned long reorder_time[IEEE80211_MAX_AMPDU_BUF];
651 struct timer_list reorder_timer;
652 bool removed;
653 bool valid;
654 spinlock_t lock;
655 struct iwl_mvm *mvm;
656 } ____cacheline_aligned_in_smp;
657
658 /**
659 * struct iwl_mvm_baid_data - BA session data
660 * @sta_id: station id
661 * @tid: tid of the session
662 * @baid baid of the session
663 * @timeout: the timeout set in the addba request
664 * @last_rx: last rx jiffies, updated only if timeout passed from last update
665 * @session_timer: timer to check if BA session expired, runs at 2 * timeout
666 * @mvm: mvm pointer, needed for timer context
667 * @reorder_buf: reorder buffer, allocated per queue
668 */
669 struct iwl_mvm_baid_data {
670 struct rcu_head rcu_head;
671 u8 sta_id;
672 u8 tid;
673 u8 baid;
674 u16 timeout;
675 unsigned long last_rx;
676 struct timer_list session_timer;
677 struct iwl_mvm *mvm;
678 struct iwl_mvm_reorder_buffer reorder_buf[];
679 };
680
681 /*
682 * enum iwl_mvm_queue_status - queue status
683 * @IWL_MVM_QUEUE_FREE: the queue is not allocated nor reserved
684 * Basically, this means that this queue can be used for any purpose
685 * @IWL_MVM_QUEUE_RESERVED: queue is reserved but not yet in use
686 * This is the state of a queue that has been dedicated for some RATID
687 * (agg'd or not), but that hasn't yet gone through the actual enablement
688 * of iwl_mvm_enable_txq(), and therefore no traffic can go through it yet.
689 * Note that in this state there is no requirement to already know what TID
690 * should be used with this queue, it is just marked as a queue that will
691 * be used, and shouldn't be allocated to anyone else.
692 * @IWL_MVM_QUEUE_READY: queue is ready to be used
693 * This is the state of a queue that has been fully configured (including
694 * SCD pointers, etc), has a specific RA/TID assigned to it, and can be
695 * used to send traffic.
696 * @IWL_MVM_QUEUE_SHARED: queue is shared, or in a process of becoming shared
697 * This is a state in which a single queue serves more than one TID, all of
698 * which are not aggregated. Note that the queue is only associated to one
699 * RA.
700 * @IWL_MVM_QUEUE_INACTIVE: queue is allocated but no traffic on it
701 * This is a state of a queue that has had traffic on it, but during the
702 * last %IWL_MVM_DQA_QUEUE_TIMEOUT time period there has been no traffic on
703 * it. In this state, when a new queue is needed to be allocated but no
704 * such free queue exists, an inactive queue might be freed and given to
705 * the new RA/TID.
706 * @IWL_MVM_QUEUE_RECONFIGURING: queue is being reconfigured
707 * This is the state of a queue that has had traffic pass through it, but
708 * needs to be reconfigured for some reason, e.g. the queue needs to
709 * become unshared and aggregations re-enabled on.
710 */
711 enum iwl_mvm_queue_status {
712 IWL_MVM_QUEUE_FREE,
713 IWL_MVM_QUEUE_RESERVED,
714 IWL_MVM_QUEUE_READY,
715 IWL_MVM_QUEUE_SHARED,
716 IWL_MVM_QUEUE_INACTIVE,
717 IWL_MVM_QUEUE_RECONFIGURING,
718 };
719
720 #define IWL_MVM_DQA_QUEUE_TIMEOUT (5 * HZ)
721 #define IWL_MVM_INVALID_QUEUE 0xFFFF
722
723 #define IWL_MVM_NUM_CIPHERS 10
724
725 #ifdef CONFIG_ACPI
726 #define IWL_MVM_SAR_TABLE_SIZE 10
727 #define IWL_MVM_SAR_PROFILE_NUM 4
728 #define IWL_MVM_GEO_TABLE_SIZE 6
729
730 struct iwl_mvm_sar_profile {
731 bool enabled;
732 u8 table[IWL_MVM_SAR_TABLE_SIZE];
733 };
734
735 struct iwl_mvm_geo_profile {
736 u8 values[IWL_MVM_GEO_TABLE_SIZE];
737 };
738 #endif
739
740 struct iwl_mvm {
741 /* for logger access */
742 struct device *dev;
743
744 struct iwl_trans *trans;
745 const struct iwl_fw *fw;
746 const struct iwl_cfg *cfg;
747 struct iwl_phy_db *phy_db;
748 struct ieee80211_hw *hw;
749
750 /* for protecting access to iwl_mvm */
751 struct mutex mutex;
752 struct list_head async_handlers_list;
753 spinlock_t async_handlers_lock;
754 struct work_struct async_handlers_wk;
755
756 struct work_struct roc_done_wk;
757
758 unsigned long init_status;
759
760 unsigned long status;
761
762 u32 queue_sync_cookie;
763 atomic_t queue_sync_counter;
764 /*
765 * for beacon filtering -
766 * currently only one interface can be supported
767 */
768 struct iwl_mvm_vif *bf_allowed_vif;
769
770 enum iwl_ucode_type cur_ucode;
771 bool hw_registered;
772 bool calibrating;
773 u32 error_event_table[2];
774 u32 log_event_table;
775 u32 umac_error_event_table;
776 bool support_umac_log;
777 struct iwl_sf_region sf_space;
778
779 u32 ampdu_ref;
780 bool ampdu_toggle;
781
782 struct iwl_notif_wait_data notif_wait;
783
784 union {
785 struct mvm_statistics_rx_v3 rx_stats_v3;
786 struct mvm_statistics_rx rx_stats;
787 };
788
789 struct {
790 u64 rx_time;
791 u64 tx_time;
792 u64 on_time_rf;
793 u64 on_time_scan;
794 } radio_stats, accu_radio_stats;
795
796 u16 hw_queue_to_mac80211[IWL_MAX_TVQM_QUEUES];
797
798 struct {
799 u8 hw_queue_refcount;
800 u8 ra_sta_id; /* The RA this queue is mapped to, if exists */
801 bool reserved; /* Is this the TXQ reserved for a STA */
802 u8 mac80211_ac; /* The mac80211 AC this queue is mapped to */
803 u8 txq_tid; /* The TID "owner" of this queue*/
804 u16 tid_bitmap; /* Bitmap of the TIDs mapped to this queue */
805 /* Timestamp for inactivation per TID of this queue */
806 unsigned long last_frame_time[IWL_MAX_TID_COUNT + 1];
807 enum iwl_mvm_queue_status status;
808 } queue_info[IWL_MAX_HW_QUEUES];
809 spinlock_t queue_info_lock; /* For syncing queue mgmt operations */
810 struct work_struct add_stream_wk; /* To add streams to queues */
811
812 atomic_t mac80211_queue_stop_count[IEEE80211_MAX_QUEUES];
813
814 const char *nvm_file_name;
815 struct iwl_nvm_data *nvm_data;
816 /* NVM sections */
817 struct iwl_nvm_section nvm_sections[NVM_MAX_NUM_SECTIONS];
818
819 struct iwl_fw_runtime fwrt;
820
821 /* EEPROM MAC addresses */
822 struct mac_address addresses[IWL_MVM_MAX_ADDRESSES];
823
824 /* data related to data path */
825 struct iwl_rx_phy_info last_phy_info;
826 struct ieee80211_sta __rcu *fw_id_to_mac_id[IWL_MVM_STATION_COUNT];
827 struct work_struct sta_drained_wk;
828 unsigned long sta_deferred_frames[BITS_TO_LONGS(IWL_MVM_STATION_COUNT)];
829 unsigned long sta_drained[BITS_TO_LONGS(IWL_MVM_STATION_COUNT)];
830 atomic_t pending_frames[IWL_MVM_STATION_COUNT];
831 u32 tfd_drained[IWL_MVM_STATION_COUNT];
832 u8 rx_ba_sessions;
833
834 /* configured by mac80211 */
835 u32 rts_threshold;
836
837 /* Scan status, cmd (pre-allocated) and auxiliary station */
838 unsigned int scan_status;
839 void *scan_cmd;
840 struct iwl_mcast_filter_cmd *mcast_filter_cmd;
841 enum iwl_mvm_scan_type scan_type;
842 enum iwl_mvm_sched_scan_pass_all_states sched_scan_pass_all;
843 struct delayed_work scan_timeout_dwork;
844
845 /* max number of simultaneous scans the FW supports */
846 unsigned int max_scans;
847
848 /* ts of the beginning of a non-collect fw dbg data period */
849 unsigned long fw_dbg_non_collect_ts_start[FW_DBG_TRIGGER_MAX - 1];
850
851 /* UMAC scan tracking */
852 u32 scan_uid_status[IWL_MVM_MAX_UMAC_SCANS];
853
854 /* start time of last scan in TSF of the mac that requested the scan */
855 u64 scan_start;
856
857 /* the vif that requested the current scan */
858 struct iwl_mvm_vif *scan_vif;
859
860 /* rx chain antennas set through debugfs for the scan command */
861 u8 scan_rx_ant;
862
863 #ifdef CONFIG_IWLWIFI_BCAST_FILTERING
864 /* broadcast filters to configure for each associated station */
865 const struct iwl_fw_bcast_filter *bcast_filters;
866 #ifdef CONFIG_IWLWIFI_DEBUGFS
867 struct {
868 bool override;
869 struct iwl_bcast_filter_cmd cmd;
870 } dbgfs_bcast_filtering;
871 #endif
872 #endif
873
874 /* Internal station */
875 struct iwl_mvm_int_sta aux_sta;
876 struct iwl_mvm_int_sta snif_sta;
877
878 bool last_ebs_successful;
879
880 u8 scan_last_antenna_idx; /* to toggle TX between antennas */
881 u8 mgmt_last_antenna_idx;
882
883 /* last smart fifo state that was successfully sent to firmware */
884 enum iwl_sf_state sf_state;
885
886 #ifdef CONFIG_IWLWIFI_DEBUGFS
887 struct dentry *debugfs_dir;
888 u32 dbgfs_sram_offset, dbgfs_sram_len;
889 u32 dbgfs_prph_reg_addr;
890 bool disable_power_off;
891 bool disable_power_off_d3;
892
893 bool scan_iter_notif_enabled;
894
895 struct debugfs_blob_wrapper nvm_hw_blob;
896 struct debugfs_blob_wrapper nvm_sw_blob;
897 struct debugfs_blob_wrapper nvm_calib_blob;
898 struct debugfs_blob_wrapper nvm_prod_blob;
899 struct debugfs_blob_wrapper nvm_phy_sku_blob;
900
901 struct iwl_mvm_frame_stats drv_rx_stats;
902 spinlock_t drv_stats_lock;
903 u16 dbgfs_rx_phyinfo;
904 #endif
905
906 struct iwl_mvm_phy_ctxt phy_ctxts[NUM_PHY_CTX];
907
908 struct list_head time_event_list;
909 spinlock_t time_event_lock;
910
911 /*
912 * A bitmap indicating the index of the key in use. The firmware
913 * can hold 16 keys at most. Reflect this fact.
914 */
915 unsigned long fw_key_table[BITS_TO_LONGS(STA_KEY_MAX_NUM)];
916 u8 fw_key_deleted[STA_KEY_MAX_NUM];
917
918 /* references taken by the driver and spinlock protecting them */
919 spinlock_t refs_lock;
920 u8 refs[IWL_MVM_REF_COUNT];
921
922 u8 vif_count;
923
924 /* -1 for always, 0 for never, >0 for that many times */
925 s8 fw_restart;
926 u8 fw_dbg_conf;
927 struct delayed_work fw_dump_wk;
928 const struct iwl_mvm_dump_desc *fw_dump_desc;
929 const struct iwl_fw_dbg_trigger_tlv *fw_dump_trig;
930
931 #ifdef CONFIG_IWLWIFI_LEDS
932 struct led_classdev led;
933 #endif
934
935 struct ieee80211_vif *p2p_device_vif;
936
937 #ifdef CONFIG_PM
938 struct wiphy_wowlan_support wowlan;
939 int gtk_ivlen, gtk_icvlen, ptk_ivlen, ptk_icvlen;
940
941 /* sched scan settings for net detect */
942 struct ieee80211_scan_ies nd_ies;
943 struct cfg80211_match_set *nd_match_sets;
944 int n_nd_match_sets;
945 struct ieee80211_channel **nd_channels;
946 int n_nd_channels;
947 bool net_detect;
948 #ifdef CONFIG_IWLWIFI_DEBUGFS
949 bool d3_wake_sysassert;
950 bool d3_test_active;
951 bool store_d3_resume_sram;
952 void *d3_resume_sram;
953 u32 d3_test_pme_ptr;
954 struct ieee80211_vif *keep_vif;
955 u32 last_netdetect_scans; /* no. of scans in the last net-detect wake */
956 #endif
957 #endif
958
959 /* d0i3 */
960 u8 d0i3_ap_sta_id;
961 bool d0i3_offloading;
962 struct work_struct d0i3_exit_work;
963 struct sk_buff_head d0i3_tx;
964 /* protect d0i3_suspend_flags */
965 struct mutex d0i3_suspend_mutex;
966 unsigned long d0i3_suspend_flags;
967 /* sync d0i3_tx queue and IWL_MVM_STATUS_IN_D0I3 status flag */
968 spinlock_t d0i3_tx_lock;
969 wait_queue_head_t d0i3_exit_waitq;
970 wait_queue_head_t rx_sync_waitq;
971
972 /* BT-Coex */
973 struct iwl_bt_coex_profile_notif last_bt_notif;
974 struct iwl_bt_coex_ci_cmd last_bt_ci_cmd;
975
976 u32 last_ant_isol;
977 u8 last_corun_lut;
978 u8 bt_tx_prio;
979 enum iwl_bt_force_ant_mode bt_force_ant_mode;
980
981 /* Aux ROC */
982 struct list_head aux_roc_te_list;
983
984 /* Thermal Throttling and CTkill */
985 struct iwl_mvm_tt_mgmt thermal_throttle;
986 #ifdef CONFIG_THERMAL
987 struct iwl_mvm_thermal_device tz_device;
988 struct iwl_mvm_cooling_device cooling_dev;
989 #endif
990
991 s32 temperature; /* Celsius */
992 /*
993 * Debug option to set the NIC temperature. This option makes the
994 * driver think this is the actual NIC temperature, and ignore the
995 * real temperature that is received from the fw
996 */
997 bool temperature_test; /* Debug test temperature is enabled */
998
999 struct iwl_time_quota_cmd last_quota_cmd;
1000
1001 #ifdef CONFIG_NL80211_TESTMODE
1002 u32 noa_duration;
1003 struct ieee80211_vif *noa_vif;
1004 #endif
1005
1006 /* Tx queues */
1007 u16 aux_queue;
1008 u16 probe_queue;
1009 u16 p2p_dev_queue;
1010
1011 u8 first_agg_queue;
1012 u8 last_agg_queue;
1013
1014 /* Indicate if device power save is allowed */
1015 u8 ps_disabled; /* u8 instead of bool to ease debugfs_create_* usage */
1016 unsigned int max_amsdu_len; /* used for debugfs only */
1017
1018 struct ieee80211_vif __rcu *csa_vif;
1019 struct ieee80211_vif __rcu *csa_tx_blocked_vif;
1020 u8 csa_tx_block_bcn_timeout;
1021
1022 /* system time of last beacon (for AP/GO interface) */
1023 u32 ap_last_beacon_gp2;
1024
1025 /* indicates that we transmitted the last beacon */
1026 bool ibss_manager;
1027
1028 bool lar_regdom_set;
1029 enum iwl_mcc_source mcc_src;
1030
1031 /* TDLS channel switch data */
1032 struct {
1033 struct delayed_work dwork;
1034 enum iwl_mvm_tdls_cs_state state;
1035
1036 /*
1037 * Current cs sta - might be different from periodic cs peer
1038 * station. Value is meaningless when the cs-state is idle.
1039 */
1040 u8 cur_sta_id;
1041
1042 /* TDLS periodic channel-switch peer */
1043 struct {
1044 u8 sta_id;
1045 u8 op_class;
1046 bool initiator; /* are we the link initiator */
1047 struct cfg80211_chan_def chandef;
1048 struct sk_buff *skb; /* ch sw template */
1049 u32 ch_sw_tm_ie;
1050
1051 /* timestamp of last ch-sw request sent (GP2 time) */
1052 u32 sent_timestamp;
1053 } peer;
1054 } tdls_cs;
1055
1056 struct iwl_mvm_shared_mem_cfg smem_cfg;
1057
1058 u32 ciphers[IWL_MVM_NUM_CIPHERS];
1059 struct ieee80211_cipher_scheme cs[IWL_UCODE_MAX_CS];
1060 struct iwl_mvm_tof_data tof_data;
1061
1062 struct ieee80211_vif *nan_vif;
1063 #define IWL_MAX_BAID 32
1064 struct iwl_mvm_baid_data __rcu *baid_map[IWL_MAX_BAID];
1065
1066 /*
1067 * Drop beacons from other APs in AP mode when there are no connected
1068 * clients.
1069 */
1070 bool drop_bcn_ap_mode;
1071
1072 struct delayed_work cs_tx_unblock_dwork;
1073 #ifdef CONFIG_ACPI
1074 struct iwl_mvm_sar_profile sar_profiles[IWL_MVM_SAR_PROFILE_NUM];
1075 struct iwl_mvm_geo_profile geo_profiles[IWL_NUM_GEO_PROFILES];
1076 #endif
1077 };
1078
1079 /* Extract MVM priv from op_mode and _hw */
1080 #define IWL_OP_MODE_GET_MVM(_iwl_op_mode) \
1081 ((struct iwl_mvm *)(_iwl_op_mode)->op_mode_specific)
1082
1083 #define IWL_MAC80211_GET_MVM(_hw) \
1084 IWL_OP_MODE_GET_MVM((struct iwl_op_mode *)((_hw)->priv))
1085
1086 /**
1087 * enum iwl_mvm_status - MVM status bits
1088 * @IWL_MVM_STATUS_HW_RFKILL: HW RF-kill is asserted
1089 * @IWL_MVM_STATUS_HW_CTKILL: CT-kill is active
1090 * @IWL_MVM_STATUS_ROC_RUNNING: remain-on-channel is running
1091 * @IWL_MVM_STATUS_IN_HW_RESTART: HW restart is active
1092 * @IWL_MVM_STATUS_IN_D0I3: NIC is in D0i3
1093 * @IWL_MVM_STATUS_ROC_AUX_RUNNING: AUX remain-on-channel is running
1094 * @IWL_MVM_STATUS_D3_RECONFIG: D3 reconfiguration is being done
1095 * @IWL_MVM_STATUS_DUMPING_FW_LOG: FW log is being dumped
1096 * @IWL_MVM_STATUS_FIRMWARE_RUNNING: firmware is running
1097 */
1098 enum iwl_mvm_status {
1099 IWL_MVM_STATUS_HW_RFKILL,
1100 IWL_MVM_STATUS_HW_CTKILL,
1101 IWL_MVM_STATUS_ROC_RUNNING,
1102 IWL_MVM_STATUS_IN_HW_RESTART,
1103 IWL_MVM_STATUS_IN_D0I3,
1104 IWL_MVM_STATUS_ROC_AUX_RUNNING,
1105 IWL_MVM_STATUS_D3_RECONFIG,
1106 IWL_MVM_STATUS_DUMPING_FW_LOG,
1107 IWL_MVM_STATUS_FIRMWARE_RUNNING,
1108 };
1109
1110 /* Keep track of completed init configuration */
1111 enum iwl_mvm_init_status {
1112 IWL_MVM_INIT_STATUS_THERMAL_INIT_COMPLETE = BIT(0),
1113 IWL_MVM_INIT_STATUS_LEDS_INIT_COMPLETE = BIT(1),
1114 IWL_MVM_INIT_STATUS_REG_HW_INIT_COMPLETE = BIT(2),
1115 IWL_MVM_INIT_STATUS_TOF_INIT_COMPLETE = BIT(3),
1116 };
1117
1118 static inline bool iwl_mvm_is_radio_killed(struct iwl_mvm *mvm)
1119 {
1120 return test_bit(IWL_MVM_STATUS_HW_RFKILL, &mvm->status) ||
1121 test_bit(IWL_MVM_STATUS_HW_CTKILL, &mvm->status);
1122 }
1123
1124 static inline bool iwl_mvm_is_radio_hw_killed(struct iwl_mvm *mvm)
1125 {
1126 return test_bit(IWL_MVM_STATUS_HW_RFKILL, &mvm->status);
1127 }
1128
1129 static inline bool iwl_mvm_firmware_running(struct iwl_mvm *mvm)
1130 {
1131 return test_bit(IWL_MVM_STATUS_FIRMWARE_RUNNING, &mvm->status);
1132 }
1133
1134 /* Must be called with rcu_read_lock() held and it can only be
1135 * released when mvmsta is not needed anymore.
1136 */
1137 static inline struct iwl_mvm_sta *
1138 iwl_mvm_sta_from_staid_rcu(struct iwl_mvm *mvm, u8 sta_id)
1139 {
1140 struct ieee80211_sta *sta;
1141
1142 if (sta_id >= ARRAY_SIZE(mvm->fw_id_to_mac_id))
1143 return NULL;
1144
1145 sta = rcu_dereference(mvm->fw_id_to_mac_id[sta_id]);
1146
1147 /* This can happen if the station has been removed right now */
1148 if (IS_ERR_OR_NULL(sta))
1149 return NULL;
1150
1151 return iwl_mvm_sta_from_mac80211(sta);
1152 }
1153
1154 static inline struct iwl_mvm_sta *
1155 iwl_mvm_sta_from_staid_protected(struct iwl_mvm *mvm, u8 sta_id)
1156 {
1157 struct ieee80211_sta *sta;
1158
1159 if (sta_id >= ARRAY_SIZE(mvm->fw_id_to_mac_id))
1160 return NULL;
1161
1162 sta = rcu_dereference_protected(mvm->fw_id_to_mac_id[sta_id],
1163 lockdep_is_held(&mvm->mutex));
1164
1165 /* This can happen if the station has been removed right now */
1166 if (IS_ERR_OR_NULL(sta))
1167 return NULL;
1168
1169 return iwl_mvm_sta_from_mac80211(sta);
1170 }
1171
1172 static inline bool iwl_mvm_is_d0i3_supported(struct iwl_mvm *mvm)
1173 {
1174 return !iwlwifi_mod_params.d0i3_disable &&
1175 fw_has_capa(&mvm->fw->ucode_capa,
1176 IWL_UCODE_TLV_CAPA_D0I3_SUPPORT);
1177 }
1178
1179 static inline bool iwl_mvm_is_dqa_supported(struct iwl_mvm *mvm)
1180 {
1181 return fw_has_capa(&mvm->fw->ucode_capa,
1182 IWL_UCODE_TLV_CAPA_DQA_SUPPORT);
1183 }
1184
1185 static inline bool iwl_mvm_enter_d0i3_on_suspend(struct iwl_mvm *mvm)
1186 {
1187 /* For now we only use this mode to differentiate between
1188 * slave transports, which handle D0i3 entry in suspend by
1189 * themselves in conjunction with runtime PM D0i3. So, this
1190 * function is used to check whether we need to do anything
1191 * when entering suspend or if the transport layer has already
1192 * done it.
1193 */
1194 return (mvm->trans->system_pm_mode == IWL_PLAT_PM_MODE_D0I3) &&
1195 (mvm->trans->runtime_pm_mode != IWL_PLAT_PM_MODE_D0I3);
1196 }
1197
1198 static inline bool iwl_mvm_is_dqa_data_queue(struct iwl_mvm *mvm, u8 queue)
1199 {
1200 return (queue >= IWL_MVM_DQA_MIN_DATA_QUEUE) &&
1201 (queue <= IWL_MVM_DQA_MAX_DATA_QUEUE);
1202 }
1203
1204 static inline bool iwl_mvm_is_dqa_mgmt_queue(struct iwl_mvm *mvm, u8 queue)
1205 {
1206 return (queue >= IWL_MVM_DQA_MIN_MGMT_QUEUE) &&
1207 (queue <= IWL_MVM_DQA_MAX_MGMT_QUEUE);
1208 }
1209
1210 static inline bool iwl_mvm_is_lar_supported(struct iwl_mvm *mvm)
1211 {
1212 bool nvm_lar = mvm->nvm_data->lar_enabled;
1213 bool tlv_lar = fw_has_capa(&mvm->fw->ucode_capa,
1214 IWL_UCODE_TLV_CAPA_LAR_SUPPORT);
1215
1216 if (iwlwifi_mod_params.lar_disable)
1217 return false;
1218
1219 /*
1220 * Enable LAR only if it is supported by the FW (TLV) &&
1221 * enabled in the NVM
1222 */
1223 if (mvm->cfg->ext_nvm)
1224 return nvm_lar && tlv_lar;
1225 else
1226 return tlv_lar;
1227 }
1228
1229 static inline bool iwl_mvm_is_wifi_mcc_supported(struct iwl_mvm *mvm)
1230 {
1231 return fw_has_api(&mvm->fw->ucode_capa,
1232 IWL_UCODE_TLV_API_WIFI_MCC_UPDATE) ||
1233 fw_has_capa(&mvm->fw->ucode_capa,
1234 IWL_UCODE_TLV_CAPA_LAR_MULTI_MCC);
1235 }
1236
1237 static inline bool iwl_mvm_bt_is_plcr_supported(struct iwl_mvm *mvm)
1238 {
1239 return fw_has_capa(&mvm->fw->ucode_capa,
1240 IWL_UCODE_TLV_CAPA_BT_COEX_PLCR) &&
1241 IWL_MVM_BT_COEX_CORUNNING;
1242 }
1243
1244 static inline bool iwl_mvm_bt_is_rrc_supported(struct iwl_mvm *mvm)
1245 {
1246 return fw_has_capa(&mvm->fw->ucode_capa,
1247 IWL_UCODE_TLV_CAPA_BT_COEX_RRC) &&
1248 IWL_MVM_BT_COEX_RRC;
1249 }
1250
1251 static inline bool iwl_mvm_is_csum_supported(struct iwl_mvm *mvm)
1252 {
1253 return fw_has_capa(&mvm->fw->ucode_capa,
1254 IWL_UCODE_TLV_CAPA_CSUM_SUPPORT) &&
1255 !IWL_MVM_HW_CSUM_DISABLE;
1256 }
1257
1258 static inline bool iwl_mvm_is_mplut_supported(struct iwl_mvm *mvm)
1259 {
1260 return fw_has_capa(&mvm->fw->ucode_capa,
1261 IWL_UCODE_TLV_CAPA_BT_MPLUT_SUPPORT) &&
1262 IWL_MVM_BT_COEX_MPLUT;
1263 }
1264
1265 static inline
1266 bool iwl_mvm_is_p2p_scm_uapsd_supported(struct iwl_mvm *mvm)
1267 {
1268 return fw_has_capa(&mvm->fw->ucode_capa,
1269 IWL_UCODE_TLV_CAPA_P2P_SCM_UAPSD) &&
1270 !(iwlwifi_mod_params.uapsd_disable &
1271 IWL_DISABLE_UAPSD_P2P_CLIENT);
1272 }
1273
1274 static inline bool iwl_mvm_has_new_rx_api(struct iwl_mvm *mvm)
1275 {
1276 return fw_has_capa(&mvm->fw->ucode_capa,
1277 IWL_UCODE_TLV_CAPA_MULTI_QUEUE_RX_SUPPORT);
1278 }
1279
1280 static inline bool iwl_mvm_has_new_tx_api(struct iwl_mvm *mvm)
1281 {
1282 /* TODO - replace with TLV once defined */
1283 return mvm->trans->cfg->use_tfh;
1284 }
1285
1286 static inline bool iwl_mvm_is_cdb_supported(struct iwl_mvm *mvm)
1287 {
1288 /*
1289 * TODO:
1290 * The issue of how to determine CDB APIs and usage is still not fully
1291 * defined.
1292 * There is a compilation for CDB and non-CDB FW, but there may
1293 * be also runtime check.
1294 * For now there is a TLV for checking compilation mode, but a
1295 * runtime check will also have to be here - once defined.
1296 */
1297 return fw_has_capa(&mvm->fw->ucode_capa,
1298 IWL_UCODE_TLV_CAPA_CDB_SUPPORT);
1299 }
1300
1301 static inline bool iwl_mvm_has_new_rx_stats_api(struct iwl_mvm *mvm)
1302 {
1303 return fw_has_api(&mvm->fw->ucode_capa,
1304 IWL_UCODE_TLV_API_NEW_RX_STATS);
1305 }
1306
1307 static inline struct agg_tx_status *
1308 iwl_mvm_get_agg_status(struct iwl_mvm *mvm, void *tx_resp)
1309 {
1310 if (iwl_mvm_has_new_tx_api(mvm))
1311 return &((struct iwl_mvm_tx_resp *)tx_resp)->status;
1312 else
1313 return ((struct iwl_mvm_tx_resp_v3 *)tx_resp)->status;
1314 }
1315
1316 static inline bool iwl_mvm_is_tt_in_fw(struct iwl_mvm *mvm)
1317 {
1318 #ifdef CONFIG_THERMAL
1319 /* these two TLV are redundant since the responsibility to CT-kill by
1320 * FW happens only after we send at least one command of
1321 * temperature THs report.
1322 */
1323 return fw_has_capa(&mvm->fw->ucode_capa,
1324 IWL_UCODE_TLV_CAPA_CT_KILL_BY_FW) &&
1325 fw_has_capa(&mvm->fw->ucode_capa,
1326 IWL_UCODE_TLV_CAPA_TEMP_THS_REPORT_SUPPORT);
1327 #else /* CONFIG_THERMAL */
1328 return false;
1329 #endif /* CONFIG_THERMAL */
1330 }
1331
1332 static inline bool iwl_mvm_is_ctdp_supported(struct iwl_mvm *mvm)
1333 {
1334 return fw_has_capa(&mvm->fw->ucode_capa,
1335 IWL_UCODE_TLV_CAPA_CTDP_SUPPORT);
1336 }
1337
1338 extern const u8 iwl_mvm_ac_to_tx_fifo[];
1339
1340 struct iwl_rate_info {
1341 u8 plcp; /* uCode API: IWL_RATE_6M_PLCP, etc. */
1342 u8 plcp_siso; /* uCode API: IWL_RATE_SISO_6M_PLCP, etc. */
1343 u8 plcp_mimo2; /* uCode API: IWL_RATE_MIMO2_6M_PLCP, etc. */
1344 u8 plcp_mimo3; /* uCode API: IWL_RATE_MIMO3_6M_PLCP, etc. */
1345 u8 ieee; /* MAC header: IWL_RATE_6M_IEEE, etc. */
1346 };
1347
1348 void __iwl_mvm_mac_stop(struct iwl_mvm *mvm);
1349 int __iwl_mvm_mac_start(struct iwl_mvm *mvm);
1350
1351 /******************
1352 * MVM Methods
1353 ******************/
1354 /* uCode */
1355 int iwl_run_init_mvm_ucode(struct iwl_mvm *mvm, bool read_nvm);
1356
1357 /* Utils */
1358 int iwl_mvm_legacy_rate_to_mac80211_idx(u32 rate_n_flags,
1359 enum nl80211_band band);
1360 void iwl_mvm_hwrate_to_tx_rate(u32 rate_n_flags,
1361 enum nl80211_band band,
1362 struct ieee80211_tx_rate *r);
1363 u8 iwl_mvm_mac80211_idx_to_hwrate(int rate_idx);
1364 void iwl_mvm_dump_nic_error_log(struct iwl_mvm *mvm);
1365 u8 first_antenna(u8 mask);
1366 u8 iwl_mvm_next_antenna(struct iwl_mvm *mvm, u8 valid, u8 last_idx);
1367 void iwl_mvm_get_sync_time(struct iwl_mvm *mvm, u32 *gp2, u64 *boottime);
1368
1369 /* Tx / Host Commands */
1370 int __must_check iwl_mvm_send_cmd(struct iwl_mvm *mvm,
1371 struct iwl_host_cmd *cmd);
1372 int __must_check iwl_mvm_send_cmd_pdu(struct iwl_mvm *mvm, u32 id,
1373 u32 flags, u16 len, const void *data);
1374 int __must_check iwl_mvm_send_cmd_status(struct iwl_mvm *mvm,
1375 struct iwl_host_cmd *cmd,
1376 u32 *status);
1377 int __must_check iwl_mvm_send_cmd_pdu_status(struct iwl_mvm *mvm, u32 id,
1378 u16 len, const void *data,
1379 u32 *status);
1380 int iwl_mvm_tx_skb(struct iwl_mvm *mvm, struct sk_buff *skb,
1381 struct ieee80211_sta *sta);
1382 int iwl_mvm_tx_skb_non_sta(struct iwl_mvm *mvm, struct sk_buff *skb);
1383 void iwl_mvm_set_tx_cmd(struct iwl_mvm *mvm, struct sk_buff *skb,
1384 struct iwl_tx_cmd *tx_cmd,
1385 struct ieee80211_tx_info *info, u8 sta_id);
1386 void iwl_mvm_set_tx_cmd_rate(struct iwl_mvm *mvm, struct iwl_tx_cmd *tx_cmd,
1387 struct ieee80211_tx_info *info,
1388 struct ieee80211_sta *sta, __le16 fc);
1389 #ifdef CONFIG_IWLWIFI_DEBUG
1390 const char *iwl_mvm_get_tx_fail_reason(u32 status);
1391 #else
1392 static inline const char *iwl_mvm_get_tx_fail_reason(u32 status) { return ""; }
1393 #endif
1394 int iwl_mvm_flush_tx_path(struct iwl_mvm *mvm, u32 tfd_msk, u32 flags);
1395 int iwl_mvm_flush_sta(struct iwl_mvm *mvm, void *sta, bool internal, u32 flags);
1396 int iwl_mvm_flush_sta_tids(struct iwl_mvm *mvm, u32 sta_id,
1397 u16 tids, u32 flags);
1398
1399 void iwl_mvm_async_handlers_purge(struct iwl_mvm *mvm);
1400
1401 static inline void iwl_mvm_set_tx_cmd_ccmp(struct ieee80211_tx_info *info,
1402 struct iwl_tx_cmd *tx_cmd)
1403 {
1404 struct ieee80211_key_conf *keyconf = info->control.hw_key;
1405
1406 tx_cmd->sec_ctl = TX_CMD_SEC_CCM;
1407 memcpy(tx_cmd->key, keyconf->key, keyconf->keylen);
1408 }
1409
1410 static inline void iwl_mvm_wait_for_async_handlers(struct iwl_mvm *mvm)
1411 {
1412 flush_work(&mvm->async_handlers_wk);
1413 }
1414
1415 /* Statistics */
1416 void iwl_mvm_handle_rx_statistics(struct iwl_mvm *mvm,
1417 struct iwl_rx_packet *pkt);
1418 void iwl_mvm_rx_statistics(struct iwl_mvm *mvm,
1419 struct iwl_rx_cmd_buffer *rxb);
1420 int iwl_mvm_request_statistics(struct iwl_mvm *mvm, bool clear);
1421 void iwl_mvm_accu_radio_stats(struct iwl_mvm *mvm);
1422
1423 /* NVM */
1424 int iwl_nvm_init(struct iwl_mvm *mvm, bool read_nvm_from_nic);
1425 int iwl_mvm_nvm_get_from_fw(struct iwl_mvm *mvm);
1426 int iwl_mvm_load_nvm_to_nic(struct iwl_mvm *mvm);
1427 int iwl_mvm_read_external_nvm(struct iwl_mvm *mvm);
1428
1429 static inline u8 iwl_mvm_get_valid_tx_ant(struct iwl_mvm *mvm)
1430 {
1431 return mvm->nvm_data && mvm->nvm_data->valid_tx_ant ?
1432 mvm->fw->valid_tx_ant & mvm->nvm_data->valid_tx_ant :
1433 mvm->fw->valid_tx_ant;
1434 }
1435
1436 static inline u8 iwl_mvm_get_valid_rx_ant(struct iwl_mvm *mvm)
1437 {
1438 return mvm->nvm_data && mvm->nvm_data->valid_rx_ant ?
1439 mvm->fw->valid_rx_ant & mvm->nvm_data->valid_rx_ant :
1440 mvm->fw->valid_rx_ant;
1441 }
1442
1443 static inline u32 iwl_mvm_get_phy_config(struct iwl_mvm *mvm)
1444 {
1445 u32 phy_config = ~(FW_PHY_CFG_TX_CHAIN |
1446 FW_PHY_CFG_RX_CHAIN);
1447 u32 valid_rx_ant = iwl_mvm_get_valid_rx_ant(mvm);
1448 u32 valid_tx_ant = iwl_mvm_get_valid_tx_ant(mvm);
1449
1450 phy_config |= valid_tx_ant << FW_PHY_CFG_TX_CHAIN_POS |
1451 valid_rx_ant << FW_PHY_CFG_RX_CHAIN_POS;
1452
1453 return mvm->fw->phy_config & phy_config;
1454 }
1455
1456 int iwl_mvm_up(struct iwl_mvm *mvm);
1457 int iwl_mvm_load_d3_fw(struct iwl_mvm *mvm);
1458
1459 int iwl_mvm_mac_setup_register(struct iwl_mvm *mvm);
1460 bool iwl_mvm_bcast_filter_build_cmd(struct iwl_mvm *mvm,
1461 struct iwl_bcast_filter_cmd *cmd);
1462
1463 /*
1464 * FW notifications / CMD responses handlers
1465 * Convention: iwl_mvm_rx_<NAME OF THE CMD>
1466 */
1467 void iwl_mvm_rx_rx_phy_cmd(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1468 void iwl_mvm_rx_rx_mpdu(struct iwl_mvm *mvm, struct napi_struct *napi,
1469 struct iwl_rx_cmd_buffer *rxb);
1470 void iwl_mvm_rx_mpdu_mq(struct iwl_mvm *mvm, struct napi_struct *napi,
1471 struct iwl_rx_cmd_buffer *rxb, int queue);
1472 void iwl_mvm_rx_frame_release(struct iwl_mvm *mvm, struct napi_struct *napi,
1473 struct iwl_rx_cmd_buffer *rxb, int queue);
1474 int iwl_mvm_notify_rx_queue(struct iwl_mvm *mvm, u32 rxq_mask,
1475 const u8 *data, u32 count);
1476 void iwl_mvm_rx_queue_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb,
1477 int queue);
1478 void iwl_mvm_rx_tx_cmd(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1479 void iwl_mvm_mfu_assert_dump_notif(struct iwl_mvm *mvm,
1480 struct iwl_rx_cmd_buffer *rxb);
1481 void iwl_mvm_rx_ba_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1482 void iwl_mvm_rx_ant_coupling_notif(struct iwl_mvm *mvm,
1483 struct iwl_rx_cmd_buffer *rxb);
1484 void iwl_mvm_rx_fw_error(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1485 void iwl_mvm_rx_card_state_notif(struct iwl_mvm *mvm,
1486 struct iwl_rx_cmd_buffer *rxb);
1487 void iwl_mvm_rx_mfuart_notif(struct iwl_mvm *mvm,
1488 struct iwl_rx_cmd_buffer *rxb);
1489 void iwl_mvm_rx_shared_mem_cfg_notif(struct iwl_mvm *mvm,
1490 struct iwl_rx_cmd_buffer *rxb);
1491
1492 /* MVM PHY */
1493 int iwl_mvm_phy_ctxt_add(struct iwl_mvm *mvm, struct iwl_mvm_phy_ctxt *ctxt,
1494 struct cfg80211_chan_def *chandef,
1495 u8 chains_static, u8 chains_dynamic);
1496 int iwl_mvm_phy_ctxt_changed(struct iwl_mvm *mvm, struct iwl_mvm_phy_ctxt *ctxt,
1497 struct cfg80211_chan_def *chandef,
1498 u8 chains_static, u8 chains_dynamic);
1499 void iwl_mvm_phy_ctxt_ref(struct iwl_mvm *mvm,
1500 struct iwl_mvm_phy_ctxt *ctxt);
1501 void iwl_mvm_phy_ctxt_unref(struct iwl_mvm *mvm,
1502 struct iwl_mvm_phy_ctxt *ctxt);
1503 int iwl_mvm_phy_ctx_count(struct iwl_mvm *mvm);
1504 u8 iwl_mvm_get_channel_width(struct cfg80211_chan_def *chandef);
1505 u8 iwl_mvm_get_ctrl_pos(struct cfg80211_chan_def *chandef);
1506
1507 /* MAC (virtual interface) programming */
1508 int iwl_mvm_mac_ctxt_init(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1509 void iwl_mvm_mac_ctxt_release(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1510 int iwl_mvm_mac_ctxt_add(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1511 int iwl_mvm_mac_ctxt_changed(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1512 bool force_assoc_off, const u8 *bssid_override);
1513 int iwl_mvm_mac_ctxt_remove(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1514 u32 iwl_mvm_mac_get_queues_mask(struct ieee80211_vif *vif);
1515 int iwl_mvm_mac_ctxt_beacon_changed(struct iwl_mvm *mvm,
1516 struct ieee80211_vif *vif);
1517 void iwl_mvm_rx_beacon_notif(struct iwl_mvm *mvm,
1518 struct iwl_rx_cmd_buffer *rxb);
1519 void iwl_mvm_rx_missed_beacons_notif(struct iwl_mvm *mvm,
1520 struct iwl_rx_cmd_buffer *rxb);
1521 void iwl_mvm_rx_stored_beacon_notif(struct iwl_mvm *mvm,
1522 struct iwl_rx_cmd_buffer *rxb);
1523 void iwl_mvm_mu_mimo_grp_notif(struct iwl_mvm *mvm,
1524 struct iwl_rx_cmd_buffer *rxb);
1525 void iwl_mvm_sta_pm_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1526 void iwl_mvm_window_status_notif(struct iwl_mvm *mvm,
1527 struct iwl_rx_cmd_buffer *rxb);
1528 void iwl_mvm_mac_ctxt_recalc_tsf_id(struct iwl_mvm *mvm,
1529 struct ieee80211_vif *vif);
1530 unsigned long iwl_mvm_get_used_hw_queues(struct iwl_mvm *mvm,
1531 struct ieee80211_vif *exclude_vif);
1532 void iwl_mvm_channel_switch_noa_notif(struct iwl_mvm *mvm,
1533 struct iwl_rx_cmd_buffer *rxb);
1534 /* Bindings */
1535 int iwl_mvm_binding_add_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1536 int iwl_mvm_binding_remove_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1537
1538 /* Quota management */
1539 int iwl_mvm_update_quotas(struct iwl_mvm *mvm, bool force_upload,
1540 struct ieee80211_vif *disabled_vif);
1541
1542 /* Scanning */
1543 int iwl_mvm_reg_scan_start(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1544 struct cfg80211_scan_request *req,
1545 struct ieee80211_scan_ies *ies);
1546 int iwl_mvm_scan_size(struct iwl_mvm *mvm);
1547 int iwl_mvm_scan_stop(struct iwl_mvm *mvm, int type, bool notify);
1548 int iwl_mvm_max_scan_ie_len(struct iwl_mvm *mvm);
1549 void iwl_mvm_report_scan_aborted(struct iwl_mvm *mvm);
1550 void iwl_mvm_scan_timeout_wk(struct work_struct *work);
1551
1552 /* Scheduled scan */
1553 void iwl_mvm_rx_lmac_scan_complete_notif(struct iwl_mvm *mvm,
1554 struct iwl_rx_cmd_buffer *rxb);
1555 void iwl_mvm_rx_lmac_scan_iter_complete_notif(struct iwl_mvm *mvm,
1556 struct iwl_rx_cmd_buffer *rxb);
1557 int iwl_mvm_sched_scan_start(struct iwl_mvm *mvm,
1558 struct ieee80211_vif *vif,
1559 struct cfg80211_sched_scan_request *req,
1560 struct ieee80211_scan_ies *ies,
1561 int type);
1562 void iwl_mvm_rx_scan_match_found(struct iwl_mvm *mvm,
1563 struct iwl_rx_cmd_buffer *rxb);
1564
1565 /* UMAC scan */
1566 int iwl_mvm_config_scan(struct iwl_mvm *mvm);
1567 void iwl_mvm_rx_umac_scan_complete_notif(struct iwl_mvm *mvm,
1568 struct iwl_rx_cmd_buffer *rxb);
1569 void iwl_mvm_rx_umac_scan_iter_complete_notif(struct iwl_mvm *mvm,
1570 struct iwl_rx_cmd_buffer *rxb);
1571
1572 /* MVM debugfs */
1573 #ifdef CONFIG_IWLWIFI_DEBUGFS
1574 int iwl_mvm_dbgfs_register(struct iwl_mvm *mvm, struct dentry *dbgfs_dir);
1575 void iwl_mvm_vif_dbgfs_register(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1576 void iwl_mvm_vif_dbgfs_clean(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1577 #else
1578 static inline int iwl_mvm_dbgfs_register(struct iwl_mvm *mvm,
1579 struct dentry *dbgfs_dir)
1580 {
1581 return 0;
1582 }
1583 static inline void
1584 iwl_mvm_vif_dbgfs_register(struct iwl_mvm *mvm, struct ieee80211_vif *vif)
1585 {
1586 }
1587 static inline void
1588 iwl_mvm_vif_dbgfs_clean(struct iwl_mvm *mvm, struct ieee80211_vif *vif)
1589 {
1590 }
1591 #endif /* CONFIG_IWLWIFI_DEBUGFS */
1592
1593 /* rate scaling */
1594 int iwl_mvm_send_lq_cmd(struct iwl_mvm *mvm, struct iwl_lq_cmd *lq, bool init);
1595 void iwl_mvm_update_frame_stats(struct iwl_mvm *mvm, u32 rate, bool agg);
1596 int rs_pretty_print_rate(char *buf, const u32 rate);
1597 void rs_update_last_rssi(struct iwl_mvm *mvm,
1598 struct iwl_lq_sta *lq_sta,
1599 struct ieee80211_rx_status *rx_status);
1600
1601 /* power management */
1602 int iwl_mvm_power_update_device(struct iwl_mvm *mvm);
1603 int iwl_mvm_power_update_mac(struct iwl_mvm *mvm);
1604 int iwl_mvm_power_update_ps(struct iwl_mvm *mvm);
1605 int iwl_mvm_power_mac_dbgfs_read(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1606 char *buf, int bufsz);
1607
1608 void iwl_mvm_power_vif_assoc(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1609 void iwl_mvm_power_uapsd_misbehaving_ap_notif(struct iwl_mvm *mvm,
1610 struct iwl_rx_cmd_buffer *rxb);
1611
1612 #ifdef CONFIG_IWLWIFI_LEDS
1613 int iwl_mvm_leds_init(struct iwl_mvm *mvm);
1614 void iwl_mvm_leds_exit(struct iwl_mvm *mvm);
1615 #else
1616 static inline int iwl_mvm_leds_init(struct iwl_mvm *mvm)
1617 {
1618 return 0;
1619 }
1620 static inline void iwl_mvm_leds_exit(struct iwl_mvm *mvm)
1621 {
1622 }
1623 #endif
1624
1625 /* D3 (WoWLAN, NetDetect) */
1626 int iwl_mvm_suspend(struct ieee80211_hw *hw, struct cfg80211_wowlan *wowlan);
1627 int iwl_mvm_resume(struct ieee80211_hw *hw);
1628 void iwl_mvm_set_wakeup(struct ieee80211_hw *hw, bool enabled);
1629 void iwl_mvm_set_rekey_data(struct ieee80211_hw *hw,
1630 struct ieee80211_vif *vif,
1631 struct cfg80211_gtk_rekey_data *data);
1632 void iwl_mvm_ipv6_addr_change(struct ieee80211_hw *hw,
1633 struct ieee80211_vif *vif,
1634 struct inet6_dev *idev);
1635 void iwl_mvm_set_default_unicast_key(struct ieee80211_hw *hw,
1636 struct ieee80211_vif *vif, int idx);
1637 extern const struct file_operations iwl_dbgfs_d3_test_ops;
1638 #ifdef CONFIG_PM
1639 int iwl_mvm_wowlan_config_key_params(struct iwl_mvm *mvm,
1640 struct ieee80211_vif *vif,
1641 bool host_awake,
1642 u32 cmd_flags);
1643 void iwl_mvm_d0i3_update_keys(struct iwl_mvm *mvm,
1644 struct ieee80211_vif *vif,
1645 struct iwl_wowlan_status *status);
1646 void iwl_mvm_set_last_nonqos_seq(struct iwl_mvm *mvm,
1647 struct ieee80211_vif *vif);
1648 #else
1649 static inline int iwl_mvm_wowlan_config_key_params(struct iwl_mvm *mvm,
1650 struct ieee80211_vif *vif,
1651 bool host_awake,
1652 u32 cmd_flags)
1653 {
1654 return 0;
1655 }
1656
1657 static inline void iwl_mvm_d0i3_update_keys(struct iwl_mvm *mvm,
1658 struct ieee80211_vif *vif,
1659 struct iwl_wowlan_status *status)
1660 {
1661 }
1662
1663 static inline void
1664 iwl_mvm_set_last_nonqos_seq(struct iwl_mvm *mvm, struct ieee80211_vif *vif)
1665 {
1666 }
1667 #endif
1668 void iwl_mvm_set_wowlan_qos_seq(struct iwl_mvm_sta *mvm_ap_sta,
1669 struct iwl_wowlan_config_cmd *cmd);
1670 int iwl_mvm_send_proto_offload(struct iwl_mvm *mvm,
1671 struct ieee80211_vif *vif,
1672 bool disable_offloading,
1673 bool offload_ns,
1674 u32 cmd_flags);
1675
1676 /* D0i3 */
1677 void iwl_mvm_ref(struct iwl_mvm *mvm, enum iwl_mvm_ref_type ref_type);
1678 void iwl_mvm_unref(struct iwl_mvm *mvm, enum iwl_mvm_ref_type ref_type);
1679 int iwl_mvm_ref_sync(struct iwl_mvm *mvm, enum iwl_mvm_ref_type ref_type);
1680 bool iwl_mvm_ref_taken(struct iwl_mvm *mvm);
1681 void iwl_mvm_d0i3_enable_tx(struct iwl_mvm *mvm, __le16 *qos_seq);
1682 int iwl_mvm_enter_d0i3(struct iwl_op_mode *op_mode);
1683 int iwl_mvm_exit_d0i3(struct iwl_op_mode *op_mode);
1684 int _iwl_mvm_exit_d0i3(struct iwl_mvm *mvm);
1685
1686 /* BT Coex */
1687 int iwl_mvm_send_bt_init_conf(struct iwl_mvm *mvm);
1688 void iwl_mvm_rx_bt_coex_notif(struct iwl_mvm *mvm,
1689 struct iwl_rx_cmd_buffer *rxb);
1690 void iwl_mvm_bt_rssi_event(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1691 enum ieee80211_rssi_event_data);
1692 void iwl_mvm_bt_coex_vif_change(struct iwl_mvm *mvm);
1693 u16 iwl_mvm_coex_agg_time_limit(struct iwl_mvm *mvm,
1694 struct ieee80211_sta *sta);
1695 bool iwl_mvm_bt_coex_is_mimo_allowed(struct iwl_mvm *mvm,
1696 struct ieee80211_sta *sta);
1697 bool iwl_mvm_bt_coex_is_ant_avail(struct iwl_mvm *mvm, u8 ant);
1698 bool iwl_mvm_bt_coex_is_shared_ant_avail(struct iwl_mvm *mvm);
1699 bool iwl_mvm_bt_coex_is_tpc_allowed(struct iwl_mvm *mvm,
1700 enum nl80211_band band);
1701 u8 iwl_mvm_bt_coex_tx_prio(struct iwl_mvm *mvm, struct ieee80211_hdr *hdr,
1702 struct ieee80211_tx_info *info, u8 ac);
1703
1704 /* beacon filtering */
1705 #ifdef CONFIG_IWLWIFI_DEBUGFS
1706 void
1707 iwl_mvm_beacon_filter_debugfs_parameters(struct ieee80211_vif *vif,
1708 struct iwl_beacon_filter_cmd *cmd);
1709 #else
1710 static inline void
1711 iwl_mvm_beacon_filter_debugfs_parameters(struct ieee80211_vif *vif,
1712 struct iwl_beacon_filter_cmd *cmd)
1713 {}
1714 #endif
1715 int iwl_mvm_update_d0i3_power_mode(struct iwl_mvm *mvm,
1716 struct ieee80211_vif *vif,
1717 bool enable, u32 flags);
1718 int iwl_mvm_enable_beacon_filter(struct iwl_mvm *mvm,
1719 struct ieee80211_vif *vif,
1720 u32 flags);
1721 int iwl_mvm_disable_beacon_filter(struct iwl_mvm *mvm,
1722 struct ieee80211_vif *vif,
1723 u32 flags);
1724 /* SMPS */
1725 void iwl_mvm_update_smps(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1726 enum iwl_mvm_smps_type_request req_type,
1727 enum ieee80211_smps_mode smps_request);
1728 bool iwl_mvm_rx_diversity_allowed(struct iwl_mvm *mvm);
1729
1730 /* Low latency */
1731 int iwl_mvm_update_low_latency(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1732 bool value);
1733 /* get SystemLowLatencyMode - only needed for beacon threshold? */
1734 bool iwl_mvm_low_latency(struct iwl_mvm *mvm);
1735 /* get VMACLowLatencyMode */
1736 static inline bool iwl_mvm_vif_low_latency(struct iwl_mvm_vif *mvmvif)
1737 {
1738 /*
1739 * should this consider associated/active/... state?
1740 *
1741 * Normally low-latency should only be active on interfaces
1742 * that are active, but at least with debugfs it can also be
1743 * enabled on interfaces that aren't active. However, when
1744 * interface aren't active then they aren't added into the
1745 * binding, so this has no real impact. For now, just return
1746 * the current desired low-latency state.
1747 */
1748 return mvmvif->low_latency_dbgfs ||
1749 mvmvif->low_latency_traffic ||
1750 mvmvif->low_latency_vcmd;
1751 }
1752
1753 /* hw scheduler queue config */
1754 bool iwl_mvm_enable_txq(struct iwl_mvm *mvm, int queue, int mac80211_queue,
1755 u16 ssn, const struct iwl_trans_txq_scd_cfg *cfg,
1756 unsigned int wdg_timeout);
1757 int iwl_mvm_tvqm_enable_txq(struct iwl_mvm *mvm, int mac80211_queue,
1758 u8 sta_id, u8 tid, unsigned int timeout);
1759
1760 /*
1761 * Disable a TXQ.
1762 * Note that in non-DQA mode the %mac80211_queue and %tid params are ignored.
1763 */
1764 int iwl_mvm_disable_txq(struct iwl_mvm *mvm, int queue, int mac80211_queue,
1765 u8 tid, u8 flags);
1766 int iwl_mvm_find_free_queue(struct iwl_mvm *mvm, u8 sta_id, u8 minq, u8 maxq);
1767
1768 /* Return a bitmask with all the hw supported queues, except for the
1769 * command queue, which can't be flushed.
1770 */
1771 static inline u32 iwl_mvm_flushable_queues(struct iwl_mvm *mvm)
1772 {
1773 u32 cmd_queue = iwl_mvm_is_dqa_supported(mvm) ? IWL_MVM_DQA_CMD_QUEUE :
1774 IWL_MVM_CMD_QUEUE;
1775
1776 return ((BIT(mvm->cfg->base_params->num_of_queues) - 1) &
1777 ~BIT(cmd_queue));
1778 }
1779
1780 static inline
1781 void iwl_mvm_enable_ac_txq(struct iwl_mvm *mvm, int queue, int mac80211_queue,
1782 u8 fifo, u16 ssn, unsigned int wdg_timeout)
1783 {
1784 struct iwl_trans_txq_scd_cfg cfg = {
1785 .fifo = fifo,
1786 .tid = IWL_MAX_TID_COUNT,
1787 .aggregate = false,
1788 .frame_limit = IWL_FRAME_LIMIT,
1789 };
1790
1791 iwl_mvm_enable_txq(mvm, queue, mac80211_queue, ssn, &cfg, wdg_timeout);
1792 }
1793
1794 static inline void iwl_mvm_stop_device(struct iwl_mvm *mvm)
1795 {
1796 iwl_free_fw_paging(&mvm->fwrt);
1797 clear_bit(IWL_MVM_STATUS_FIRMWARE_RUNNING, &mvm->status);
1798 mvm->fw_dbg_conf = FW_DBG_INVALID;
1799 iwl_trans_stop_device(mvm->trans);
1800 }
1801
1802 /* Stop/start all mac queues in a given bitmap */
1803 void iwl_mvm_start_mac_queues(struct iwl_mvm *mvm, unsigned long mq);
1804 void iwl_mvm_stop_mac_queues(struct iwl_mvm *mvm, unsigned long mq);
1805
1806 /* Re-configure the SCD for a queue that has already been configured */
1807 int iwl_mvm_reconfig_scd(struct iwl_mvm *mvm, int queue, int fifo, int sta_id,
1808 int tid, int frame_limit, u16 ssn);
1809
1810 /* Thermal management and CT-kill */
1811 void iwl_mvm_tt_tx_backoff(struct iwl_mvm *mvm, u32 backoff);
1812 void iwl_mvm_tt_temp_changed(struct iwl_mvm *mvm, u32 temp);
1813 void iwl_mvm_temp_notif(struct iwl_mvm *mvm,
1814 struct iwl_rx_cmd_buffer *rxb);
1815 void iwl_mvm_tt_handler(struct iwl_mvm *mvm);
1816 void iwl_mvm_thermal_initialize(struct iwl_mvm *mvm, u32 min_backoff);
1817 void iwl_mvm_thermal_exit(struct iwl_mvm *mvm);
1818 void iwl_mvm_set_hw_ctkill_state(struct iwl_mvm *mvm, bool state);
1819 int iwl_mvm_get_temp(struct iwl_mvm *mvm, s32 *temp);
1820 void iwl_mvm_ct_kill_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1821 int iwl_mvm_send_temp_report_ths_cmd(struct iwl_mvm *mvm);
1822 int iwl_mvm_ctdp_command(struct iwl_mvm *mvm, u32 op, u32 budget);
1823
1824 /* Location Aware Regulatory */
1825 struct iwl_mcc_update_resp *
1826 iwl_mvm_update_mcc(struct iwl_mvm *mvm, const char *alpha2,
1827 enum iwl_mcc_source src_id);
1828 int iwl_mvm_init_mcc(struct iwl_mvm *mvm);
1829 void iwl_mvm_rx_chub_update_mcc(struct iwl_mvm *mvm,
1830 struct iwl_rx_cmd_buffer *rxb);
1831 struct ieee80211_regdomain *iwl_mvm_get_regdomain(struct wiphy *wiphy,
1832 const char *alpha2,
1833 enum iwl_mcc_source src_id,
1834 bool *changed);
1835 struct ieee80211_regdomain *iwl_mvm_get_current_regdomain(struct iwl_mvm *mvm,
1836 bool *changed);
1837 int iwl_mvm_init_fw_regd(struct iwl_mvm *mvm);
1838 void iwl_mvm_update_changed_regdom(struct iwl_mvm *mvm);
1839
1840 /* smart fifo */
1841 int iwl_mvm_sf_update(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1842 bool added_vif);
1843
1844 /* TDLS */
1845
1846 /*
1847 * We use TID 4 (VI) as a FW-used-only TID when TDLS connections are present.
1848 * This TID is marked as used vs the AP and all connected TDLS peers.
1849 */
1850 #define IWL_MVM_TDLS_FW_TID 4
1851
1852 int iwl_mvm_tdls_sta_count(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1853 void iwl_mvm_teardown_tdls_peers(struct iwl_mvm *mvm);
1854 void iwl_mvm_recalc_tdls_state(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1855 bool sta_added);
1856 void iwl_mvm_mac_mgd_protect_tdls_discover(struct ieee80211_hw *hw,
1857 struct ieee80211_vif *vif);
1858 int iwl_mvm_tdls_channel_switch(struct ieee80211_hw *hw,
1859 struct ieee80211_vif *vif,
1860 struct ieee80211_sta *sta, u8 oper_class,
1861 struct cfg80211_chan_def *chandef,
1862 struct sk_buff *tmpl_skb, u32 ch_sw_tm_ie);
1863 void iwl_mvm_tdls_recv_channel_switch(struct ieee80211_hw *hw,
1864 struct ieee80211_vif *vif,
1865 struct ieee80211_tdls_ch_sw_params *params);
1866 void iwl_mvm_tdls_cancel_channel_switch(struct ieee80211_hw *hw,
1867 struct ieee80211_vif *vif,
1868 struct ieee80211_sta *sta);
1869 void iwl_mvm_rx_tdls_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1870 void iwl_mvm_tdls_ch_switch_work(struct work_struct *work);
1871
1872 void iwl_mvm_sync_rx_queues_internal(struct iwl_mvm *mvm,
1873 struct iwl_mvm_internal_rxq_notif *notif,
1874 u32 size);
1875 void iwl_mvm_reorder_timer_expired(unsigned long data);
1876 struct ieee80211_vif *iwl_mvm_get_bss_vif(struct iwl_mvm *mvm);
1877 bool iwl_mvm_is_vif_assoc(struct iwl_mvm *mvm);
1878
1879 void iwl_mvm_inactivity_check(struct iwl_mvm *mvm);
1880
1881 void iwl_mvm_nic_restart(struct iwl_mvm *mvm, bool fw_error);
1882 unsigned int iwl_mvm_get_wd_timeout(struct iwl_mvm *mvm,
1883 struct ieee80211_vif *vif,
1884 bool tdls, bool cmd_q);
1885 void iwl_mvm_connection_loss(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1886 const char *errmsg);
1887
1888 /* Link Quality Measurement */
1889 int iwl_mvm_send_lqm_cmd(struct ieee80211_vif *vif,
1890 enum iwl_lqm_cmd_operatrions operation,
1891 u32 duration, u32 timeout);
1892 bool iwl_mvm_lqm_active(struct iwl_mvm *mvm);
1893
1894 #ifdef CONFIG_ACPI
1895 int iwl_mvm_sar_select_profile(struct iwl_mvm *mvm, int prof_a, int prof_b);
1896 int iwl_mvm_get_sar_geo_profile(struct iwl_mvm *mvm);
1897 #else
1898 static inline
1899 int iwl_mvm_sar_select_profile(struct iwl_mvm *mvm, int prof_a, int prof_b)
1900 {
1901 return -ENOENT;
1902 }
1903
1904 static inline
1905 int iwl_mvm_get_sar_geo_profile(struct iwl_mvm *mvm)
1906 {
1907 return -ENOENT;
1908 }
1909 #endif /* CONFIG_ACPI */
1910
1911 #endif /* __IWL_MVM_H__ */