]> git.proxmox.com Git - mirror_ubuntu-bionic-kernel.git/blob - drivers/net/wireless/intel/iwlwifi/mvm/mvm.h
6a9a25beab3fcdc3f372a541274bfc0046374280
[mirror_ubuntu-bionic-kernel.git] / drivers / net / wireless / intel / iwlwifi / mvm / mvm.h
1 /******************************************************************************
2 *
3 * This file is provided under a dual BSD/GPLv2 license. When using or
4 * redistributing this file, you may do so under either license.
5 *
6 * GPL LICENSE SUMMARY
7 *
8 * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved.
9 * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH
10 * Copyright(c) 2016 - 2017 Intel Deutschland GmbH
11 *
12 * This program is free software; you can redistribute it and/or modify
13 * it under the terms of version 2 of the GNU General Public License as
14 * published by the Free Software Foundation.
15 *
16 * This program is distributed in the hope that it will be useful, but
17 * WITHOUT ANY WARRANTY; without even the implied warranty of
18 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
19 * General Public License for more details.
20 *
21 * You should have received a copy of the GNU General Public License
22 * along with this program; if not, write to the Free Software
23 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110,
24 * USA
25 *
26 * The full GNU General Public License is included in this distribution
27 * in the file called COPYING.
28 *
29 * Contact Information:
30 * Intel Linux Wireless <linuxwifi@intel.com>
31 * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
32 *
33 * BSD LICENSE
34 *
35 * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved.
36 * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH
37 * Copyright(c) 2016 - 2017 Intel Deutschland GmbH
38 * All rights reserved.
39 *
40 * Redistribution and use in source and binary forms, with or without
41 * modification, are permitted provided that the following conditions
42 * are met:
43 *
44 * * Redistributions of source code must retain the above copyright
45 * notice, this list of conditions and the following disclaimer.
46 * * Redistributions in binary form must reproduce the above copyright
47 * notice, this list of conditions and the following disclaimer in
48 * the documentation and/or other materials provided with the
49 * distribution.
50 * * Neither the name Intel Corporation nor the names of its
51 * contributors may be used to endorse or promote products derived
52 * from this software without specific prior written permission.
53 *
54 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
55 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
56 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
57 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
58 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
59 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
60 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
61 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
62 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
63 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
64 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
65 *
66 *****************************************************************************/
67
68 #ifndef __IWL_MVM_H__
69 #define __IWL_MVM_H__
70
71 #include <linux/list.h>
72 #include <linux/spinlock.h>
73 #include <linux/leds.h>
74 #include <linux/in6.h>
75
76 #ifdef CONFIG_THERMAL
77 #include <linux/thermal.h>
78 #endif
79
80 #include "iwl-op-mode.h"
81 #include "iwl-trans.h"
82 #include "fw/notif-wait.h"
83 #include "iwl-eeprom-parse.h"
84 #include "fw/file.h"
85 #include "iwl-config.h"
86 #include "sta.h"
87 #include "fw-api.h"
88 #include "constants.h"
89 #include "tof.h"
90 #include "fw/runtime.h"
91 #include "fw/dbg.h"
92 #include "fw/acpi.h"
93
94 #define IWL_MVM_MAX_ADDRESSES 5
95 /* RSSI offset for WkP */
96 #define IWL_RSSI_OFFSET 50
97 #define IWL_MVM_MISSED_BEACONS_THRESHOLD 8
98 /* A TimeUnit is 1024 microsecond */
99 #define MSEC_TO_TU(_msec) (_msec*1000/1024)
100
101 /* For GO, this value represents the number of TUs before CSA "beacon
102 * 0" TBTT when the CSA time-event needs to be scheduled to start. It
103 * must be big enough to ensure that we switch in time.
104 */
105 #define IWL_MVM_CHANNEL_SWITCH_TIME_GO 40
106
107 /* For client, this value represents the number of TUs before CSA
108 * "beacon 1" TBTT, instead. This is because we don't know when the
109 * GO/AP will be in the new channel, so we switch early enough.
110 */
111 #define IWL_MVM_CHANNEL_SWITCH_TIME_CLIENT 10
112
113 /*
114 * This value (in TUs) is used to fine tune the CSA NoA end time which should
115 * be just before "beacon 0" TBTT.
116 */
117 #define IWL_MVM_CHANNEL_SWITCH_MARGIN 4
118
119 /*
120 * Number of beacons to transmit on a new channel until we unblock tx to
121 * the stations, even if we didn't identify them on a new channel
122 */
123 #define IWL_MVM_CS_UNBLOCK_TX_TIMEOUT 3
124
125 /* offchannel queue towards mac80211 */
126 #define IWL_MVM_OFFCHANNEL_QUEUE 0
127
128 extern const struct ieee80211_ops iwl_mvm_hw_ops;
129
130 /**
131 * struct iwl_mvm_mod_params - module parameters for iwlmvm
132 * @init_dbg: if true, then the NIC won't be stopped if the INIT fw asserted.
133 * We will register to mac80211 to have testmode working. The NIC must not
134 * be up'ed after the INIT fw asserted. This is useful to be able to use
135 * proprietary tools over testmode to debug the INIT fw.
136 * @tfd_q_hang_detect: enabled the detection of hung transmit queues
137 * @power_scheme: one of enum iwl_power_scheme
138 */
139 struct iwl_mvm_mod_params {
140 bool init_dbg;
141 bool tfd_q_hang_detect;
142 int power_scheme;
143 };
144 extern struct iwl_mvm_mod_params iwlmvm_mod_params;
145
146 struct iwl_mvm_phy_ctxt {
147 u16 id;
148 u16 color;
149 u32 ref;
150
151 enum nl80211_chan_width width;
152
153 /*
154 * TODO: This should probably be removed. Currently here only for rate
155 * scaling algorithm
156 */
157 struct ieee80211_channel *channel;
158 };
159
160 struct iwl_mvm_time_event_data {
161 struct ieee80211_vif *vif;
162 struct list_head list;
163 unsigned long end_jiffies;
164 u32 duration;
165 bool running;
166 u32 uid;
167
168 /*
169 * The access to the 'id' field must be done when the
170 * mvm->time_event_lock is held, as it value is used to indicate
171 * if the te is in the time event list or not (when id == TE_MAX)
172 */
173 u32 id;
174 };
175
176 /* Power management */
177
178 /**
179 * enum iwl_power_scheme
180 * @IWL_POWER_LEVEL_CAM - Continuously Active Mode
181 * @IWL_POWER_LEVEL_BPS - Balanced Power Save (default)
182 * @IWL_POWER_LEVEL_LP - Low Power
183 */
184 enum iwl_power_scheme {
185 IWL_POWER_SCHEME_CAM = 1,
186 IWL_POWER_SCHEME_BPS,
187 IWL_POWER_SCHEME_LP
188 };
189
190 #define IWL_CONN_MAX_LISTEN_INTERVAL 10
191 #define IWL_UAPSD_MAX_SP IEEE80211_WMM_IE_STA_QOSINFO_SP_ALL
192
193 #ifdef CONFIG_IWLWIFI_DEBUGFS
194 enum iwl_dbgfs_pm_mask {
195 MVM_DEBUGFS_PM_KEEP_ALIVE = BIT(0),
196 MVM_DEBUGFS_PM_SKIP_OVER_DTIM = BIT(1),
197 MVM_DEBUGFS_PM_SKIP_DTIM_PERIODS = BIT(2),
198 MVM_DEBUGFS_PM_RX_DATA_TIMEOUT = BIT(3),
199 MVM_DEBUGFS_PM_TX_DATA_TIMEOUT = BIT(4),
200 MVM_DEBUGFS_PM_LPRX_ENA = BIT(6),
201 MVM_DEBUGFS_PM_LPRX_RSSI_THRESHOLD = BIT(7),
202 MVM_DEBUGFS_PM_SNOOZE_ENABLE = BIT(8),
203 MVM_DEBUGFS_PM_UAPSD_MISBEHAVING = BIT(9),
204 MVM_DEBUGFS_PM_USE_PS_POLL = BIT(10),
205 };
206
207 struct iwl_dbgfs_pm {
208 u16 keep_alive_seconds;
209 u32 rx_data_timeout;
210 u32 tx_data_timeout;
211 bool skip_over_dtim;
212 u8 skip_dtim_periods;
213 bool lprx_ena;
214 u32 lprx_rssi_threshold;
215 bool snooze_ena;
216 bool uapsd_misbehaving;
217 bool use_ps_poll;
218 int mask;
219 };
220
221 /* beacon filtering */
222
223 enum iwl_dbgfs_bf_mask {
224 MVM_DEBUGFS_BF_ENERGY_DELTA = BIT(0),
225 MVM_DEBUGFS_BF_ROAMING_ENERGY_DELTA = BIT(1),
226 MVM_DEBUGFS_BF_ROAMING_STATE = BIT(2),
227 MVM_DEBUGFS_BF_TEMP_THRESHOLD = BIT(3),
228 MVM_DEBUGFS_BF_TEMP_FAST_FILTER = BIT(4),
229 MVM_DEBUGFS_BF_TEMP_SLOW_FILTER = BIT(5),
230 MVM_DEBUGFS_BF_ENABLE_BEACON_FILTER = BIT(6),
231 MVM_DEBUGFS_BF_DEBUG_FLAG = BIT(7),
232 MVM_DEBUGFS_BF_ESCAPE_TIMER = BIT(8),
233 MVM_DEBUGFS_BA_ESCAPE_TIMER = BIT(9),
234 MVM_DEBUGFS_BA_ENABLE_BEACON_ABORT = BIT(10),
235 };
236
237 struct iwl_dbgfs_bf {
238 u32 bf_energy_delta;
239 u32 bf_roaming_energy_delta;
240 u32 bf_roaming_state;
241 u32 bf_temp_threshold;
242 u32 bf_temp_fast_filter;
243 u32 bf_temp_slow_filter;
244 u32 bf_enable_beacon_filter;
245 u32 bf_debug_flag;
246 u32 bf_escape_timer;
247 u32 ba_escape_timer;
248 u32 ba_enable_beacon_abort;
249 int mask;
250 };
251 #endif
252
253 enum iwl_mvm_smps_type_request {
254 IWL_MVM_SMPS_REQ_BT_COEX,
255 IWL_MVM_SMPS_REQ_TT,
256 IWL_MVM_SMPS_REQ_PROT,
257 NUM_IWL_MVM_SMPS_REQ,
258 };
259
260 enum iwl_mvm_ref_type {
261 IWL_MVM_REF_UCODE_DOWN,
262 IWL_MVM_REF_SCAN,
263 IWL_MVM_REF_ROC,
264 IWL_MVM_REF_ROC_AUX,
265 IWL_MVM_REF_P2P_CLIENT,
266 IWL_MVM_REF_AP_IBSS,
267 IWL_MVM_REF_USER,
268 IWL_MVM_REF_TX,
269 IWL_MVM_REF_TX_AGG,
270 IWL_MVM_REF_ADD_IF,
271 IWL_MVM_REF_START_AP,
272 IWL_MVM_REF_BSS_CHANGED,
273 IWL_MVM_REF_PREPARE_TX,
274 IWL_MVM_REF_PROTECT_TDLS,
275 IWL_MVM_REF_CHECK_CTKILL,
276 IWL_MVM_REF_PRPH_READ,
277 IWL_MVM_REF_PRPH_WRITE,
278 IWL_MVM_REF_NMI,
279 IWL_MVM_REF_TM_CMD,
280 IWL_MVM_REF_EXIT_WORK,
281 IWL_MVM_REF_PROTECT_CSA,
282 IWL_MVM_REF_FW_DBG_COLLECT,
283 IWL_MVM_REF_INIT_UCODE,
284 IWL_MVM_REF_SENDING_CMD,
285 IWL_MVM_REF_RX,
286
287 /* update debugfs.c when changing this */
288
289 IWL_MVM_REF_COUNT,
290 };
291
292 enum iwl_bt_force_ant_mode {
293 BT_FORCE_ANT_DIS = 0,
294 BT_FORCE_ANT_AUTO,
295 BT_FORCE_ANT_BT,
296 BT_FORCE_ANT_WIFI,
297
298 BT_FORCE_ANT_MAX,
299 };
300
301 /**
302 * struct iwl_mvm_vif_bf_data - beacon filtering related data
303 * @bf_enabled: indicates if beacon filtering is enabled
304 * @ba_enabled: indicated if beacon abort is enabled
305 * @ave_beacon_signal: average beacon signal
306 * @last_cqm_event: rssi of the last cqm event
307 * @bt_coex_min_thold: minimum threshold for BT coex
308 * @bt_coex_max_thold: maximum threshold for BT coex
309 * @last_bt_coex_event: rssi of the last BT coex event
310 */
311 struct iwl_mvm_vif_bf_data {
312 bool bf_enabled;
313 bool ba_enabled;
314 int ave_beacon_signal;
315 int last_cqm_event;
316 int bt_coex_min_thold;
317 int bt_coex_max_thold;
318 int last_bt_coex_event;
319 };
320
321 /**
322 * struct iwl_mvm_vif - data per Virtual Interface, it is a MAC context
323 * @id: between 0 and 3
324 * @color: to solve races upon MAC addition and removal
325 * @ap_sta_id: the sta_id of the AP - valid only if VIF type is STA
326 * @bssid: BSSID for this (client) interface
327 * @associated: indicates that we're currently associated, used only for
328 * managing the firmware state in iwl_mvm_bss_info_changed_station()
329 * @ap_assoc_sta_count: count of stations associated to us - valid only
330 * if VIF type is AP
331 * @uploaded: indicates the MAC context has been added to the device
332 * @ap_ibss_active: indicates that AP/IBSS is configured and that the interface
333 * should get quota etc.
334 * @pm_enabled - Indicate if MAC power management is allowed
335 * @monitor_active: indicates that monitor context is configured, and that the
336 * interface should get quota etc.
337 * @low_latency_traffic: indicates low latency traffic was detected
338 * @low_latency_dbgfs: low latency mode set from debugfs
339 * @low_latency_vcmd: low latency mode set from vendor command
340 * @ps_disabled: indicates that this interface requires PS to be disabled
341 * @queue_params: QoS params for this MAC
342 * @bcast_sta: station used for broadcast packets. Used by the following
343 * vifs: P2P_DEVICE, GO and AP.
344 * @beacon_skb: the skb used to hold the AP/GO beacon template
345 * @smps_requests: the SMPS requests of different parts of the driver,
346 * combined on update to yield the overall request to mac80211.
347 * @beacon_stats: beacon statistics, containing the # of received beacons,
348 * # of received beacons accumulated over FW restart, and the current
349 * average signal of beacons retrieved from the firmware
350 * @csa_failed: CSA failed to schedule time event, report an error later
351 * @features: hw features active for this vif
352 */
353 struct iwl_mvm_vif {
354 struct iwl_mvm *mvm;
355 u16 id;
356 u16 color;
357 u8 ap_sta_id;
358
359 u8 bssid[ETH_ALEN];
360 bool associated;
361 u8 ap_assoc_sta_count;
362
363 u16 cab_queue;
364
365 bool uploaded;
366 bool ap_ibss_active;
367 bool pm_enabled;
368 bool monitor_active;
369 bool low_latency_traffic, low_latency_dbgfs, low_latency_vcmd;
370 bool ps_disabled;
371 struct iwl_mvm_vif_bf_data bf_data;
372
373 struct {
374 u32 num_beacons, accu_num_beacons;
375 u8 avg_signal;
376 } beacon_stats;
377
378 u32 ap_beacon_time;
379
380 enum iwl_tsf_id tsf_id;
381
382 /*
383 * QoS data from mac80211, need to store this here
384 * as mac80211 has a separate callback but we need
385 * to have the data for the MAC context
386 */
387 struct ieee80211_tx_queue_params queue_params[IEEE80211_NUM_ACS];
388 struct iwl_mvm_time_event_data time_event_data;
389 struct iwl_mvm_time_event_data hs_time_event_data;
390
391 struct iwl_mvm_int_sta bcast_sta;
392 struct iwl_mvm_int_sta mcast_sta;
393
394 /*
395 * Assigned while mac80211 has the interface in a channel context,
396 * or, for P2P Device, while it exists.
397 */
398 struct iwl_mvm_phy_ctxt *phy_ctxt;
399
400 #ifdef CONFIG_PM
401 /* WoWLAN GTK rekey data */
402 struct {
403 u8 kck[NL80211_KCK_LEN], kek[NL80211_KEK_LEN];
404 __le64 replay_ctr;
405 bool valid;
406 } rekey_data;
407
408 int tx_key_idx;
409
410 bool seqno_valid;
411 u16 seqno;
412 #endif
413
414 #if IS_ENABLED(CONFIG_IPV6)
415 /* IPv6 addresses for WoWLAN */
416 struct in6_addr target_ipv6_addrs[IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_MAX];
417 unsigned long tentative_addrs[BITS_TO_LONGS(IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_MAX)];
418 int num_target_ipv6_addrs;
419 #endif
420
421 #ifdef CONFIG_IWLWIFI_DEBUGFS
422 struct dentry *dbgfs_dir;
423 struct dentry *dbgfs_slink;
424 struct iwl_dbgfs_pm dbgfs_pm;
425 struct iwl_dbgfs_bf dbgfs_bf;
426 struct iwl_mac_power_cmd mac_pwr_cmd;
427 int dbgfs_quota_min;
428 #endif
429
430 enum ieee80211_smps_mode smps_requests[NUM_IWL_MVM_SMPS_REQ];
431
432 /* FW identified misbehaving AP */
433 u8 uapsd_misbehaving_bssid[ETH_ALEN];
434
435 /* Indicates that CSA countdown may be started */
436 bool csa_countdown;
437 bool csa_failed;
438 u16 csa_target_freq;
439
440 /* TCP Checksum Offload */
441 netdev_features_t features;
442 };
443
444 static inline struct iwl_mvm_vif *
445 iwl_mvm_vif_from_mac80211(struct ieee80211_vif *vif)
446 {
447 if (!vif)
448 return NULL;
449 return (void *)vif->drv_priv;
450 }
451
452 extern const u8 tid_to_mac80211_ac[];
453
454 #define IWL_MVM_SCAN_STOPPING_SHIFT 8
455
456 enum iwl_scan_status {
457 IWL_MVM_SCAN_REGULAR = BIT(0),
458 IWL_MVM_SCAN_SCHED = BIT(1),
459 IWL_MVM_SCAN_NETDETECT = BIT(2),
460
461 IWL_MVM_SCAN_STOPPING_REGULAR = BIT(8),
462 IWL_MVM_SCAN_STOPPING_SCHED = BIT(9),
463 IWL_MVM_SCAN_STOPPING_NETDETECT = BIT(10),
464
465 IWL_MVM_SCAN_REGULAR_MASK = IWL_MVM_SCAN_REGULAR |
466 IWL_MVM_SCAN_STOPPING_REGULAR,
467 IWL_MVM_SCAN_SCHED_MASK = IWL_MVM_SCAN_SCHED |
468 IWL_MVM_SCAN_STOPPING_SCHED,
469 IWL_MVM_SCAN_NETDETECT_MASK = IWL_MVM_SCAN_NETDETECT |
470 IWL_MVM_SCAN_STOPPING_NETDETECT,
471
472 IWL_MVM_SCAN_STOPPING_MASK = 0xff << IWL_MVM_SCAN_STOPPING_SHIFT,
473 IWL_MVM_SCAN_MASK = 0xff,
474 };
475
476 enum iwl_mvm_scan_type {
477 IWL_SCAN_TYPE_NOT_SET,
478 IWL_SCAN_TYPE_UNASSOC,
479 IWL_SCAN_TYPE_WILD,
480 IWL_SCAN_TYPE_MILD,
481 IWL_SCAN_TYPE_FRAGMENTED,
482 };
483
484 enum iwl_mvm_sched_scan_pass_all_states {
485 SCHED_SCAN_PASS_ALL_DISABLED,
486 SCHED_SCAN_PASS_ALL_ENABLED,
487 SCHED_SCAN_PASS_ALL_FOUND,
488 };
489
490 /**
491 * struct iwl_nvm_section - describes an NVM section in memory.
492 *
493 * This struct holds an NVM section read from the NIC using NVM_ACCESS_CMD,
494 * and saved for later use by the driver. Not all NVM sections are saved
495 * this way, only the needed ones.
496 */
497 struct iwl_nvm_section {
498 u16 length;
499 const u8 *data;
500 };
501
502 /**
503 * struct iwl_mvm_tt_mgnt - Thermal Throttling Management structure
504 * @ct_kill_exit: worker to exit thermal kill
505 * @dynamic_smps: Is thermal throttling enabled dynamic_smps?
506 * @tx_backoff: The current thremal throttling tx backoff in uSec.
507 * @min_backoff: The minimal tx backoff due to power restrictions
508 * @params: Parameters to configure the thermal throttling algorithm.
509 * @throttle: Is thermal throttling is active?
510 */
511 struct iwl_mvm_tt_mgmt {
512 struct delayed_work ct_kill_exit;
513 bool dynamic_smps;
514 u32 tx_backoff;
515 u32 min_backoff;
516 struct iwl_tt_params params;
517 bool throttle;
518 };
519
520 #ifdef CONFIG_THERMAL
521 /**
522 *struct iwl_mvm_thermal_device - thermal zone related data
523 * @temp_trips: temperature thresholds for report
524 * @fw_trips_index: keep indexes to original array - temp_trips
525 * @tzone: thermal zone device data
526 */
527 struct iwl_mvm_thermal_device {
528 s16 temp_trips[IWL_MAX_DTS_TRIPS];
529 u8 fw_trips_index[IWL_MAX_DTS_TRIPS];
530 struct thermal_zone_device *tzone;
531 };
532
533 /*
534 * struct iwl_mvm_cooling_device
535 * @cur_state: current state
536 * @cdev: struct thermal cooling device
537 */
538 struct iwl_mvm_cooling_device {
539 u32 cur_state;
540 struct thermal_cooling_device *cdev;
541 };
542 #endif
543
544 #define IWL_MVM_NUM_LAST_FRAMES_UCODE_RATES 8
545
546 struct iwl_mvm_frame_stats {
547 u32 legacy_frames;
548 u32 ht_frames;
549 u32 vht_frames;
550 u32 bw_20_frames;
551 u32 bw_40_frames;
552 u32 bw_80_frames;
553 u32 bw_160_frames;
554 u32 sgi_frames;
555 u32 ngi_frames;
556 u32 siso_frames;
557 u32 mimo2_frames;
558 u32 agg_frames;
559 u32 ampdu_count;
560 u32 success_frames;
561 u32 fail_frames;
562 u32 last_rates[IWL_MVM_NUM_LAST_FRAMES_UCODE_RATES];
563 int last_frame_idx;
564 };
565
566 enum {
567 D0I3_DEFER_WAKEUP,
568 D0I3_PENDING_WAKEUP,
569 };
570
571 #define IWL_MVM_DEBUG_SET_TEMPERATURE_DISABLE 0xff
572 #define IWL_MVM_DEBUG_SET_TEMPERATURE_MIN -100
573 #define IWL_MVM_DEBUG_SET_TEMPERATURE_MAX 200
574
575 enum iwl_mvm_tdls_cs_state {
576 IWL_MVM_TDLS_SW_IDLE = 0,
577 IWL_MVM_TDLS_SW_REQ_SENT,
578 IWL_MVM_TDLS_SW_RESP_RCVD,
579 IWL_MVM_TDLS_SW_REQ_RCVD,
580 IWL_MVM_TDLS_SW_ACTIVE,
581 };
582
583 /**
584 * struct iwl_mvm_reorder_buffer - per ra/tid/queue reorder buffer
585 * @head_sn: reorder window head sn
586 * @num_stored: number of mpdus stored in the buffer
587 * @buf_size: the reorder buffer size as set by the last addba request
588 * @queue: queue of this reorder buffer
589 * @last_amsdu: track last ASMDU SN for duplication detection
590 * @last_sub_index: track ASMDU sub frame index for duplication detection
591 * @reorder_timer: timer for frames are in the reorder buffer. For AMSDU
592 * it is the time of last received sub-frame
593 * @removed: prevent timer re-arming
594 * @valid: reordering is valid for this queue
595 * @lock: protect reorder buffer internal state
596 * @mvm: mvm pointer, needed for frame timer context
597 */
598 struct iwl_mvm_reorder_buffer {
599 u16 head_sn;
600 u16 num_stored;
601 u8 buf_size;
602 int queue;
603 u16 last_amsdu;
604 u8 last_sub_index;
605 struct timer_list reorder_timer;
606 bool removed;
607 bool valid;
608 spinlock_t lock;
609 struct iwl_mvm *mvm;
610 } ____cacheline_aligned_in_smp;
611
612 /**
613 * struct _iwl_mvm_reorder_buf_entry - reorder buffer entry per-queue/per-seqno
614 * @frames: list of skbs stored
615 * @reorder_time: time the packet was stored in the reorder buffer
616 */
617 struct _iwl_mvm_reorder_buf_entry {
618 struct sk_buff_head frames;
619 unsigned long reorder_time;
620 };
621
622 /* make this indirection to get the aligned thing */
623 struct iwl_mvm_reorder_buf_entry {
624 struct _iwl_mvm_reorder_buf_entry e;
625 }
626 #ifndef __CHECKER__
627 /* sparse doesn't like this construct: "bad integer constant expression" */
628 __aligned(roundup_pow_of_two(sizeof(struct _iwl_mvm_reorder_buf_entry)))
629 #endif
630 ;
631
632 /**
633 * struct iwl_mvm_baid_data - BA session data
634 * @sta_id: station id
635 * @tid: tid of the session
636 * @baid baid of the session
637 * @timeout: the timeout set in the addba request
638 * @entries_per_queue: # of buffers per queue, this actually gets
639 * aligned up to avoid cache line sharing between queues
640 * @last_rx: last rx jiffies, updated only if timeout passed from last update
641 * @session_timer: timer to check if BA session expired, runs at 2 * timeout
642 * @mvm: mvm pointer, needed for timer context
643 * @reorder_buf: reorder buffer, allocated per queue
644 * @reorder_buf_data: data
645 */
646 struct iwl_mvm_baid_data {
647 struct rcu_head rcu_head;
648 u8 sta_id;
649 u8 tid;
650 u8 baid;
651 u16 timeout;
652 u16 entries_per_queue;
653 unsigned long last_rx;
654 struct timer_list session_timer;
655 struct iwl_mvm_baid_data __rcu **rcu_ptr;
656 struct iwl_mvm *mvm;
657 struct iwl_mvm_reorder_buffer reorder_buf[IWL_MAX_RX_HW_QUEUES];
658 struct iwl_mvm_reorder_buf_entry entries[];
659 };
660
661 static inline struct iwl_mvm_baid_data *
662 iwl_mvm_baid_data_from_reorder_buf(struct iwl_mvm_reorder_buffer *buf)
663 {
664 return (void *)((u8 *)buf -
665 offsetof(struct iwl_mvm_baid_data, reorder_buf) -
666 sizeof(*buf) * buf->queue);
667 }
668
669 /*
670 * enum iwl_mvm_queue_status - queue status
671 * @IWL_MVM_QUEUE_FREE: the queue is not allocated nor reserved
672 * Basically, this means that this queue can be used for any purpose
673 * @IWL_MVM_QUEUE_RESERVED: queue is reserved but not yet in use
674 * This is the state of a queue that has been dedicated for some RATID
675 * (agg'd or not), but that hasn't yet gone through the actual enablement
676 * of iwl_mvm_enable_txq(), and therefore no traffic can go through it yet.
677 * Note that in this state there is no requirement to already know what TID
678 * should be used with this queue, it is just marked as a queue that will
679 * be used, and shouldn't be allocated to anyone else.
680 * @IWL_MVM_QUEUE_READY: queue is ready to be used
681 * This is the state of a queue that has been fully configured (including
682 * SCD pointers, etc), has a specific RA/TID assigned to it, and can be
683 * used to send traffic.
684 * @IWL_MVM_QUEUE_SHARED: queue is shared, or in a process of becoming shared
685 * This is a state in which a single queue serves more than one TID, all of
686 * which are not aggregated. Note that the queue is only associated to one
687 * RA.
688 * @IWL_MVM_QUEUE_INACTIVE: queue is allocated but no traffic on it
689 * This is a state of a queue that has had traffic on it, but during the
690 * last %IWL_MVM_DQA_QUEUE_TIMEOUT time period there has been no traffic on
691 * it. In this state, when a new queue is needed to be allocated but no
692 * such free queue exists, an inactive queue might be freed and given to
693 * the new RA/TID.
694 * @IWL_MVM_QUEUE_RECONFIGURING: queue is being reconfigured
695 * This is the state of a queue that has had traffic pass through it, but
696 * needs to be reconfigured for some reason, e.g. the queue needs to
697 * become unshared and aggregations re-enabled on.
698 */
699 enum iwl_mvm_queue_status {
700 IWL_MVM_QUEUE_FREE,
701 IWL_MVM_QUEUE_RESERVED,
702 IWL_MVM_QUEUE_READY,
703 IWL_MVM_QUEUE_SHARED,
704 IWL_MVM_QUEUE_INACTIVE,
705 IWL_MVM_QUEUE_RECONFIGURING,
706 };
707
708 #define IWL_MVM_DQA_QUEUE_TIMEOUT (5 * HZ)
709 #define IWL_MVM_INVALID_QUEUE 0xFFFF
710
711 #define IWL_MVM_NUM_CIPHERS 10
712
713 struct iwl_mvm_sar_profile {
714 bool enabled;
715 u8 table[ACPI_SAR_TABLE_SIZE];
716 };
717
718 struct iwl_mvm_geo_profile {
719 u8 values[ACPI_GEO_TABLE_SIZE];
720 };
721
722 struct iwl_mvm {
723 /* for logger access */
724 struct device *dev;
725
726 struct iwl_trans *trans;
727 const struct iwl_fw *fw;
728 const struct iwl_cfg *cfg;
729 struct iwl_phy_db *phy_db;
730 struct ieee80211_hw *hw;
731
732 /* for protecting access to iwl_mvm */
733 struct mutex mutex;
734 struct list_head async_handlers_list;
735 spinlock_t async_handlers_lock;
736 struct work_struct async_handlers_wk;
737
738 struct work_struct roc_done_wk;
739
740 unsigned long init_status;
741
742 unsigned long status;
743
744 u32 queue_sync_cookie;
745 atomic_t queue_sync_counter;
746 /*
747 * for beacon filtering -
748 * currently only one interface can be supported
749 */
750 struct iwl_mvm_vif *bf_allowed_vif;
751
752 bool hw_registered;
753 bool calibrating;
754 u32 error_event_table[2];
755 u32 log_event_table;
756 u32 umac_error_event_table;
757 bool support_umac_log;
758
759 u32 ampdu_ref;
760 bool ampdu_toggle;
761
762 struct iwl_notif_wait_data notif_wait;
763
764 union {
765 struct mvm_statistics_rx_v3 rx_stats_v3;
766 struct mvm_statistics_rx rx_stats;
767 };
768
769 struct {
770 u64 rx_time;
771 u64 tx_time;
772 u64 on_time_rf;
773 u64 on_time_scan;
774 } radio_stats, accu_radio_stats;
775
776 u16 hw_queue_to_mac80211[IWL_MAX_TVQM_QUEUES];
777
778 struct {
779 u8 hw_queue_refcount;
780 u8 ra_sta_id; /* The RA this queue is mapped to, if exists */
781 bool reserved; /* Is this the TXQ reserved for a STA */
782 u8 mac80211_ac; /* The mac80211 AC this queue is mapped to */
783 u8 txq_tid; /* The TID "owner" of this queue*/
784 u16 tid_bitmap; /* Bitmap of the TIDs mapped to this queue */
785 /* Timestamp for inactivation per TID of this queue */
786 unsigned long last_frame_time[IWL_MAX_TID_COUNT + 1];
787 enum iwl_mvm_queue_status status;
788 } queue_info[IWL_MAX_HW_QUEUES];
789 spinlock_t queue_info_lock; /* For syncing queue mgmt operations */
790 struct work_struct add_stream_wk; /* To add streams to queues */
791
792 atomic_t mac80211_queue_stop_count[IEEE80211_MAX_QUEUES];
793
794 const char *nvm_file_name;
795 struct iwl_nvm_data *nvm_data;
796 /* NVM sections */
797 struct iwl_nvm_section nvm_sections[NVM_MAX_NUM_SECTIONS];
798
799 struct iwl_fw_runtime fwrt;
800
801 /* EEPROM MAC addresses */
802 struct mac_address addresses[IWL_MVM_MAX_ADDRESSES];
803
804 /* data related to data path */
805 struct iwl_rx_phy_info last_phy_info;
806 struct ieee80211_sta __rcu *fw_id_to_mac_id[IWL_MVM_STATION_COUNT];
807 unsigned long sta_deferred_frames[BITS_TO_LONGS(IWL_MVM_STATION_COUNT)];
808 u8 rx_ba_sessions;
809
810 /* configured by mac80211 */
811 u32 rts_threshold;
812
813 /* Scan status, cmd (pre-allocated) and auxiliary station */
814 unsigned int scan_status;
815 void *scan_cmd;
816 struct iwl_mcast_filter_cmd *mcast_filter_cmd;
817 enum iwl_mvm_scan_type scan_type;
818 enum iwl_mvm_sched_scan_pass_all_states sched_scan_pass_all;
819 struct delayed_work scan_timeout_dwork;
820
821 /* max number of simultaneous scans the FW supports */
822 unsigned int max_scans;
823
824 /* UMAC scan tracking */
825 u32 scan_uid_status[IWL_MVM_MAX_UMAC_SCANS];
826
827 /* start time of last scan in TSF of the mac that requested the scan */
828 u64 scan_start;
829
830 /* the vif that requested the current scan */
831 struct iwl_mvm_vif *scan_vif;
832
833 /* rx chain antennas set through debugfs for the scan command */
834 u8 scan_rx_ant;
835
836 #ifdef CONFIG_IWLWIFI_BCAST_FILTERING
837 /* broadcast filters to configure for each associated station */
838 const struct iwl_fw_bcast_filter *bcast_filters;
839 #ifdef CONFIG_IWLWIFI_DEBUGFS
840 struct {
841 bool override;
842 struct iwl_bcast_filter_cmd cmd;
843 } dbgfs_bcast_filtering;
844 #endif
845 #endif
846
847 /* Internal station */
848 struct iwl_mvm_int_sta aux_sta;
849 struct iwl_mvm_int_sta snif_sta;
850
851 bool last_ebs_successful;
852
853 u8 scan_last_antenna_idx; /* to toggle TX between antennas */
854 u8 mgmt_last_antenna_idx;
855
856 /* last smart fifo state that was successfully sent to firmware */
857 enum iwl_sf_state sf_state;
858
859 #ifdef CONFIG_IWLWIFI_DEBUGFS
860 struct dentry *debugfs_dir;
861 u32 dbgfs_sram_offset, dbgfs_sram_len;
862 u32 dbgfs_prph_reg_addr;
863 bool disable_power_off;
864 bool disable_power_off_d3;
865
866 bool scan_iter_notif_enabled;
867
868 struct debugfs_blob_wrapper nvm_hw_blob;
869 struct debugfs_blob_wrapper nvm_sw_blob;
870 struct debugfs_blob_wrapper nvm_calib_blob;
871 struct debugfs_blob_wrapper nvm_prod_blob;
872 struct debugfs_blob_wrapper nvm_phy_sku_blob;
873
874 struct iwl_mvm_frame_stats drv_rx_stats;
875 spinlock_t drv_stats_lock;
876 u16 dbgfs_rx_phyinfo;
877 #endif
878
879 struct iwl_mvm_phy_ctxt phy_ctxts[NUM_PHY_CTX];
880
881 struct list_head time_event_list;
882 spinlock_t time_event_lock;
883
884 /*
885 * A bitmap indicating the index of the key in use. The firmware
886 * can hold 16 keys at most. Reflect this fact.
887 */
888 unsigned long fw_key_table[BITS_TO_LONGS(STA_KEY_MAX_NUM)];
889 u8 fw_key_deleted[STA_KEY_MAX_NUM];
890
891 /* references taken by the driver and spinlock protecting them */
892 spinlock_t refs_lock;
893 u8 refs[IWL_MVM_REF_COUNT];
894
895 u8 vif_count;
896
897 /* -1 for always, 0 for never, >0 for that many times */
898 s8 fw_restart;
899
900 #ifdef CONFIG_IWLWIFI_LEDS
901 struct led_classdev led;
902 #endif
903
904 struct ieee80211_vif *p2p_device_vif;
905
906 #ifdef CONFIG_PM
907 struct wiphy_wowlan_support wowlan;
908 int gtk_ivlen, gtk_icvlen, ptk_ivlen, ptk_icvlen;
909
910 /* sched scan settings for net detect */
911 struct ieee80211_scan_ies nd_ies;
912 struct cfg80211_match_set *nd_match_sets;
913 int n_nd_match_sets;
914 struct ieee80211_channel **nd_channels;
915 int n_nd_channels;
916 bool net_detect;
917 #ifdef CONFIG_IWLWIFI_DEBUGFS
918 bool d3_wake_sysassert;
919 bool d3_test_active;
920 bool store_d3_resume_sram;
921 void *d3_resume_sram;
922 u32 d3_test_pme_ptr;
923 struct ieee80211_vif *keep_vif;
924 u32 last_netdetect_scans; /* no. of scans in the last net-detect wake */
925 #endif
926 #endif
927
928 /* d0i3 */
929 u8 d0i3_ap_sta_id;
930 bool d0i3_offloading;
931 struct work_struct d0i3_exit_work;
932 struct sk_buff_head d0i3_tx;
933 /* protect d0i3_suspend_flags */
934 struct mutex d0i3_suspend_mutex;
935 unsigned long d0i3_suspend_flags;
936 /* sync d0i3_tx queue and IWL_MVM_STATUS_IN_D0I3 status flag */
937 spinlock_t d0i3_tx_lock;
938 wait_queue_head_t d0i3_exit_waitq;
939 wait_queue_head_t rx_sync_waitq;
940
941 /* BT-Coex */
942 struct iwl_bt_coex_profile_notif last_bt_notif;
943 struct iwl_bt_coex_ci_cmd last_bt_ci_cmd;
944
945 u8 bt_tx_prio;
946 enum iwl_bt_force_ant_mode bt_force_ant_mode;
947
948 /* Aux ROC */
949 struct list_head aux_roc_te_list;
950
951 /* Thermal Throttling and CTkill */
952 struct iwl_mvm_tt_mgmt thermal_throttle;
953 #ifdef CONFIG_THERMAL
954 struct iwl_mvm_thermal_device tz_device;
955 struct iwl_mvm_cooling_device cooling_dev;
956 #endif
957
958 s32 temperature; /* Celsius */
959 /*
960 * Debug option to set the NIC temperature. This option makes the
961 * driver think this is the actual NIC temperature, and ignore the
962 * real temperature that is received from the fw
963 */
964 bool temperature_test; /* Debug test temperature is enabled */
965
966 struct iwl_time_quota_cmd last_quota_cmd;
967
968 #ifdef CONFIG_NL80211_TESTMODE
969 u32 noa_duration;
970 struct ieee80211_vif *noa_vif;
971 #endif
972
973 /* Tx queues */
974 u16 aux_queue;
975 u16 snif_queue;
976 u16 probe_queue;
977 u16 p2p_dev_queue;
978
979 /* Indicate if device power save is allowed */
980 u8 ps_disabled; /* u8 instead of bool to ease debugfs_create_* usage */
981 unsigned int max_amsdu_len; /* used for debugfs only */
982
983 struct ieee80211_vif __rcu *csa_vif;
984 struct ieee80211_vif __rcu *csa_tx_blocked_vif;
985 u8 csa_tx_block_bcn_timeout;
986
987 /* system time of last beacon (for AP/GO interface) */
988 u32 ap_last_beacon_gp2;
989
990 /* indicates that we transmitted the last beacon */
991 bool ibss_manager;
992
993 bool lar_regdom_set;
994 enum iwl_mcc_source mcc_src;
995
996 /* TDLS channel switch data */
997 struct {
998 struct delayed_work dwork;
999 enum iwl_mvm_tdls_cs_state state;
1000
1001 /*
1002 * Current cs sta - might be different from periodic cs peer
1003 * station. Value is meaningless when the cs-state is idle.
1004 */
1005 u8 cur_sta_id;
1006
1007 /* TDLS periodic channel-switch peer */
1008 struct {
1009 u8 sta_id;
1010 u8 op_class;
1011 bool initiator; /* are we the link initiator */
1012 struct cfg80211_chan_def chandef;
1013 struct sk_buff *skb; /* ch sw template */
1014 u32 ch_sw_tm_ie;
1015
1016 /* timestamp of last ch-sw request sent (GP2 time) */
1017 u32 sent_timestamp;
1018 } peer;
1019 } tdls_cs;
1020
1021
1022 u32 ciphers[IWL_MVM_NUM_CIPHERS];
1023 struct ieee80211_cipher_scheme cs[IWL_UCODE_MAX_CS];
1024 struct iwl_mvm_tof_data tof_data;
1025
1026 struct ieee80211_vif *nan_vif;
1027 #define IWL_MAX_BAID 32
1028 struct iwl_mvm_baid_data __rcu *baid_map[IWL_MAX_BAID];
1029
1030 /*
1031 * Drop beacons from other APs in AP mode when there are no connected
1032 * clients.
1033 */
1034 bool drop_bcn_ap_mode;
1035
1036 struct delayed_work cs_tx_unblock_dwork;
1037
1038 /* does a monitor vif exist (only one can exist hence bool) */
1039 bool monitor_on;
1040 #ifdef CONFIG_ACPI
1041 struct iwl_mvm_sar_profile sar_profiles[ACPI_SAR_PROFILE_NUM];
1042 struct iwl_mvm_geo_profile geo_profiles[ACPI_NUM_GEO_PROFILES];
1043 #endif
1044 };
1045
1046 /* Extract MVM priv from op_mode and _hw */
1047 #define IWL_OP_MODE_GET_MVM(_iwl_op_mode) \
1048 ((struct iwl_mvm *)(_iwl_op_mode)->op_mode_specific)
1049
1050 #define IWL_MAC80211_GET_MVM(_hw) \
1051 IWL_OP_MODE_GET_MVM((struct iwl_op_mode *)((_hw)->priv))
1052
1053 /**
1054 * enum iwl_mvm_status - MVM status bits
1055 * @IWL_MVM_STATUS_HW_RFKILL: HW RF-kill is asserted
1056 * @IWL_MVM_STATUS_HW_CTKILL: CT-kill is active
1057 * @IWL_MVM_STATUS_ROC_RUNNING: remain-on-channel is running
1058 * @IWL_MVM_STATUS_HW_RESTART_REQUESTED: HW restart was requested
1059 * @IWL_MVM_STATUS_IN_HW_RESTART: HW restart is active
1060 * @IWL_MVM_STATUS_IN_D0I3: NIC is in D0i3
1061 * @IWL_MVM_STATUS_ROC_AUX_RUNNING: AUX remain-on-channel is running
1062 * @IWL_MVM_STATUS_D3_RECONFIG: D3 reconfiguration is being done
1063 * @IWL_MVM_STATUS_FIRMWARE_RUNNING: firmware is running
1064 */
1065 enum iwl_mvm_status {
1066 IWL_MVM_STATUS_HW_RFKILL,
1067 IWL_MVM_STATUS_HW_CTKILL,
1068 IWL_MVM_STATUS_ROC_RUNNING,
1069 IWL_MVM_STATUS_HW_RESTART_REQUESTED,
1070 IWL_MVM_STATUS_IN_HW_RESTART,
1071 IWL_MVM_STATUS_IN_D0I3,
1072 IWL_MVM_STATUS_ROC_AUX_RUNNING,
1073 IWL_MVM_STATUS_D3_RECONFIG,
1074 IWL_MVM_STATUS_FIRMWARE_RUNNING,
1075 };
1076
1077 /* Keep track of completed init configuration */
1078 enum iwl_mvm_init_status {
1079 IWL_MVM_INIT_STATUS_THERMAL_INIT_COMPLETE = BIT(0),
1080 IWL_MVM_INIT_STATUS_LEDS_INIT_COMPLETE = BIT(1),
1081 IWL_MVM_INIT_STATUS_REG_HW_INIT_COMPLETE = BIT(2),
1082 IWL_MVM_INIT_STATUS_TOF_INIT_COMPLETE = BIT(3),
1083 };
1084
1085 static inline bool iwl_mvm_is_radio_killed(struct iwl_mvm *mvm)
1086 {
1087 return test_bit(IWL_MVM_STATUS_HW_RFKILL, &mvm->status) ||
1088 test_bit(IWL_MVM_STATUS_HW_CTKILL, &mvm->status);
1089 }
1090
1091 static inline bool iwl_mvm_is_radio_hw_killed(struct iwl_mvm *mvm)
1092 {
1093 return test_bit(IWL_MVM_STATUS_HW_RFKILL, &mvm->status);
1094 }
1095
1096 static inline bool iwl_mvm_firmware_running(struct iwl_mvm *mvm)
1097 {
1098 return test_bit(IWL_MVM_STATUS_FIRMWARE_RUNNING, &mvm->status);
1099 }
1100
1101 /* Must be called with rcu_read_lock() held and it can only be
1102 * released when mvmsta is not needed anymore.
1103 */
1104 static inline struct iwl_mvm_sta *
1105 iwl_mvm_sta_from_staid_rcu(struct iwl_mvm *mvm, u8 sta_id)
1106 {
1107 struct ieee80211_sta *sta;
1108
1109 if (sta_id >= ARRAY_SIZE(mvm->fw_id_to_mac_id))
1110 return NULL;
1111
1112 sta = rcu_dereference(mvm->fw_id_to_mac_id[sta_id]);
1113
1114 /* This can happen if the station has been removed right now */
1115 if (IS_ERR_OR_NULL(sta))
1116 return NULL;
1117
1118 return iwl_mvm_sta_from_mac80211(sta);
1119 }
1120
1121 static inline struct iwl_mvm_sta *
1122 iwl_mvm_sta_from_staid_protected(struct iwl_mvm *mvm, u8 sta_id)
1123 {
1124 struct ieee80211_sta *sta;
1125
1126 if (sta_id >= ARRAY_SIZE(mvm->fw_id_to_mac_id))
1127 return NULL;
1128
1129 sta = rcu_dereference_protected(mvm->fw_id_to_mac_id[sta_id],
1130 lockdep_is_held(&mvm->mutex));
1131
1132 /* This can happen if the station has been removed right now */
1133 if (IS_ERR_OR_NULL(sta))
1134 return NULL;
1135
1136 return iwl_mvm_sta_from_mac80211(sta);
1137 }
1138
1139 static inline bool iwl_mvm_is_d0i3_supported(struct iwl_mvm *mvm)
1140 {
1141 return !iwlwifi_mod_params.d0i3_disable &&
1142 fw_has_capa(&mvm->fw->ucode_capa,
1143 IWL_UCODE_TLV_CAPA_D0I3_SUPPORT);
1144 }
1145
1146 static inline bool iwl_mvm_is_adaptive_dwell_supported(struct iwl_mvm *mvm)
1147 {
1148 return fw_has_api(&mvm->fw->ucode_capa,
1149 IWL_UCODE_TLV_API_ADAPTIVE_DWELL);
1150 }
1151
1152 static inline bool iwl_mvm_enter_d0i3_on_suspend(struct iwl_mvm *mvm)
1153 {
1154 /* For now we only use this mode to differentiate between
1155 * slave transports, which handle D0i3 entry in suspend by
1156 * themselves in conjunction with runtime PM D0i3. So, this
1157 * function is used to check whether we need to do anything
1158 * when entering suspend or if the transport layer has already
1159 * done it.
1160 */
1161 return (mvm->trans->system_pm_mode == IWL_PLAT_PM_MODE_D0I3) &&
1162 (mvm->trans->runtime_pm_mode != IWL_PLAT_PM_MODE_D0I3);
1163 }
1164
1165 static inline bool iwl_mvm_is_dqa_data_queue(struct iwl_mvm *mvm, u8 queue)
1166 {
1167 return (queue >= IWL_MVM_DQA_MIN_DATA_QUEUE) &&
1168 (queue <= IWL_MVM_DQA_MAX_DATA_QUEUE);
1169 }
1170
1171 static inline bool iwl_mvm_is_dqa_mgmt_queue(struct iwl_mvm *mvm, u8 queue)
1172 {
1173 return (queue >= IWL_MVM_DQA_MIN_MGMT_QUEUE) &&
1174 (queue <= IWL_MVM_DQA_MAX_MGMT_QUEUE);
1175 }
1176
1177 static inline bool iwl_mvm_is_lar_supported(struct iwl_mvm *mvm)
1178 {
1179 bool nvm_lar = mvm->nvm_data->lar_enabled;
1180 bool tlv_lar = fw_has_capa(&mvm->fw->ucode_capa,
1181 IWL_UCODE_TLV_CAPA_LAR_SUPPORT);
1182
1183 if (iwlwifi_mod_params.lar_disable)
1184 return false;
1185
1186 /*
1187 * Enable LAR only if it is supported by the FW (TLV) &&
1188 * enabled in the NVM
1189 */
1190 if (mvm->cfg->nvm_type == IWL_NVM_EXT)
1191 return nvm_lar && tlv_lar;
1192 else
1193 return tlv_lar;
1194 }
1195
1196 static inline bool iwl_mvm_is_wifi_mcc_supported(struct iwl_mvm *mvm)
1197 {
1198 return fw_has_api(&mvm->fw->ucode_capa,
1199 IWL_UCODE_TLV_API_WIFI_MCC_UPDATE) ||
1200 fw_has_capa(&mvm->fw->ucode_capa,
1201 IWL_UCODE_TLV_CAPA_LAR_MULTI_MCC);
1202 }
1203
1204 static inline bool iwl_mvm_bt_is_rrc_supported(struct iwl_mvm *mvm)
1205 {
1206 return fw_has_capa(&mvm->fw->ucode_capa,
1207 IWL_UCODE_TLV_CAPA_BT_COEX_RRC) &&
1208 IWL_MVM_BT_COEX_RRC;
1209 }
1210
1211 static inline bool iwl_mvm_is_csum_supported(struct iwl_mvm *mvm)
1212 {
1213 return fw_has_capa(&mvm->fw->ucode_capa,
1214 IWL_UCODE_TLV_CAPA_CSUM_SUPPORT) &&
1215 !IWL_MVM_HW_CSUM_DISABLE;
1216 }
1217
1218 static inline bool iwl_mvm_is_mplut_supported(struct iwl_mvm *mvm)
1219 {
1220 return fw_has_capa(&mvm->fw->ucode_capa,
1221 IWL_UCODE_TLV_CAPA_BT_MPLUT_SUPPORT) &&
1222 IWL_MVM_BT_COEX_MPLUT;
1223 }
1224
1225 static inline
1226 bool iwl_mvm_is_p2p_scm_uapsd_supported(struct iwl_mvm *mvm)
1227 {
1228 return fw_has_capa(&mvm->fw->ucode_capa,
1229 IWL_UCODE_TLV_CAPA_P2P_SCM_UAPSD) &&
1230 !(iwlwifi_mod_params.uapsd_disable &
1231 IWL_DISABLE_UAPSD_P2P_CLIENT);
1232 }
1233
1234 static inline bool iwl_mvm_has_new_rx_api(struct iwl_mvm *mvm)
1235 {
1236 return fw_has_capa(&mvm->fw->ucode_capa,
1237 IWL_UCODE_TLV_CAPA_MULTI_QUEUE_RX_SUPPORT);
1238 }
1239
1240 static inline bool iwl_mvm_has_new_tx_api(struct iwl_mvm *mvm)
1241 {
1242 /* TODO - replace with TLV once defined */
1243 return mvm->trans->cfg->use_tfh;
1244 }
1245
1246 static inline bool iwl_mvm_has_unified_ucode(struct iwl_mvm *mvm)
1247 {
1248 /* TODO - better define this */
1249 return mvm->trans->cfg->device_family >= IWL_DEVICE_FAMILY_A000;
1250 }
1251
1252 static inline bool iwl_mvm_is_cdb_supported(struct iwl_mvm *mvm)
1253 {
1254 /*
1255 * TODO:
1256 * The issue of how to determine CDB APIs and usage is still not fully
1257 * defined.
1258 * There is a compilation for CDB and non-CDB FW, but there may
1259 * be also runtime check.
1260 * For now there is a TLV for checking compilation mode, but a
1261 * runtime check will also have to be here - once defined.
1262 */
1263 return fw_has_capa(&mvm->fw->ucode_capa,
1264 IWL_UCODE_TLV_CAPA_CDB_SUPPORT);
1265 }
1266
1267 static inline bool iwl_mvm_has_new_rx_stats_api(struct iwl_mvm *mvm)
1268 {
1269 return fw_has_api(&mvm->fw->ucode_capa,
1270 IWL_UCODE_TLV_API_NEW_RX_STATS);
1271 }
1272
1273 static inline bool iwl_mvm_has_new_ats_coex_api(struct iwl_mvm *mvm)
1274 {
1275 return fw_has_api(&mvm->fw->ucode_capa,
1276 IWL_UCODE_TLV_API_COEX_ATS_EXTERNAL);
1277 }
1278
1279 static inline bool iwl_mvm_has_quota_low_latency(struct iwl_mvm *mvm)
1280 {
1281 return fw_has_api(&mvm->fw->ucode_capa,
1282 IWL_UCODE_TLV_API_QUOTA_LOW_LATENCY);
1283 }
1284
1285 static inline struct agg_tx_status *
1286 iwl_mvm_get_agg_status(struct iwl_mvm *mvm, void *tx_resp)
1287 {
1288 if (iwl_mvm_has_new_tx_api(mvm))
1289 return &((struct iwl_mvm_tx_resp *)tx_resp)->status;
1290 else
1291 return ((struct iwl_mvm_tx_resp_v3 *)tx_resp)->status;
1292 }
1293
1294 static inline bool iwl_mvm_is_tt_in_fw(struct iwl_mvm *mvm)
1295 {
1296 #ifdef CONFIG_THERMAL
1297 /* these two TLV are redundant since the responsibility to CT-kill by
1298 * FW happens only after we send at least one command of
1299 * temperature THs report.
1300 */
1301 return fw_has_capa(&mvm->fw->ucode_capa,
1302 IWL_UCODE_TLV_CAPA_CT_KILL_BY_FW) &&
1303 fw_has_capa(&mvm->fw->ucode_capa,
1304 IWL_UCODE_TLV_CAPA_TEMP_THS_REPORT_SUPPORT);
1305 #else /* CONFIG_THERMAL */
1306 return false;
1307 #endif /* CONFIG_THERMAL */
1308 }
1309
1310 static inline bool iwl_mvm_is_ctdp_supported(struct iwl_mvm *mvm)
1311 {
1312 return fw_has_capa(&mvm->fw->ucode_capa,
1313 IWL_UCODE_TLV_CAPA_CTDP_SUPPORT);
1314 }
1315
1316 extern const u8 iwl_mvm_ac_to_tx_fifo[];
1317 extern const u8 iwl_mvm_ac_to_gen2_tx_fifo[];
1318
1319 static inline u8 iwl_mvm_mac_ac_to_tx_fifo(struct iwl_mvm *mvm,
1320 enum ieee80211_ac_numbers ac)
1321 {
1322 return iwl_mvm_has_new_tx_api(mvm) ?
1323 iwl_mvm_ac_to_gen2_tx_fifo[ac] : iwl_mvm_ac_to_tx_fifo[ac];
1324 }
1325
1326 struct iwl_rate_info {
1327 u8 plcp; /* uCode API: IWL_RATE_6M_PLCP, etc. */
1328 u8 plcp_siso; /* uCode API: IWL_RATE_SISO_6M_PLCP, etc. */
1329 u8 plcp_mimo2; /* uCode API: IWL_RATE_MIMO2_6M_PLCP, etc. */
1330 u8 plcp_mimo3; /* uCode API: IWL_RATE_MIMO3_6M_PLCP, etc. */
1331 u8 ieee; /* MAC header: IWL_RATE_6M_IEEE, etc. */
1332 };
1333
1334 void __iwl_mvm_mac_stop(struct iwl_mvm *mvm);
1335 int __iwl_mvm_mac_start(struct iwl_mvm *mvm);
1336
1337 /******************
1338 * MVM Methods
1339 ******************/
1340 /* uCode */
1341 int iwl_run_init_mvm_ucode(struct iwl_mvm *mvm, bool read_nvm);
1342
1343 /* Utils */
1344 int iwl_mvm_legacy_rate_to_mac80211_idx(u32 rate_n_flags,
1345 enum nl80211_band band);
1346 void iwl_mvm_hwrate_to_tx_rate(u32 rate_n_flags,
1347 enum nl80211_band band,
1348 struct ieee80211_tx_rate *r);
1349 u8 iwl_mvm_mac80211_idx_to_hwrate(int rate_idx);
1350 void iwl_mvm_dump_nic_error_log(struct iwl_mvm *mvm);
1351 u8 first_antenna(u8 mask);
1352 u8 iwl_mvm_next_antenna(struct iwl_mvm *mvm, u8 valid, u8 last_idx);
1353 void iwl_mvm_get_sync_time(struct iwl_mvm *mvm, u32 *gp2, u64 *boottime);
1354
1355 /* Tx / Host Commands */
1356 int __must_check iwl_mvm_send_cmd(struct iwl_mvm *mvm,
1357 struct iwl_host_cmd *cmd);
1358 int __must_check iwl_mvm_send_cmd_pdu(struct iwl_mvm *mvm, u32 id,
1359 u32 flags, u16 len, const void *data);
1360 int __must_check iwl_mvm_send_cmd_status(struct iwl_mvm *mvm,
1361 struct iwl_host_cmd *cmd,
1362 u32 *status);
1363 int __must_check iwl_mvm_send_cmd_pdu_status(struct iwl_mvm *mvm, u32 id,
1364 u16 len, const void *data,
1365 u32 *status);
1366 int iwl_mvm_tx_skb(struct iwl_mvm *mvm, struct sk_buff *skb,
1367 struct ieee80211_sta *sta);
1368 int iwl_mvm_tx_skb_non_sta(struct iwl_mvm *mvm, struct sk_buff *skb);
1369 void iwl_mvm_set_tx_cmd(struct iwl_mvm *mvm, struct sk_buff *skb,
1370 struct iwl_tx_cmd *tx_cmd,
1371 struct ieee80211_tx_info *info, u8 sta_id);
1372 void iwl_mvm_set_tx_cmd_rate(struct iwl_mvm *mvm, struct iwl_tx_cmd *tx_cmd,
1373 struct ieee80211_tx_info *info,
1374 struct ieee80211_sta *sta, __le16 fc);
1375 #ifdef CONFIG_IWLWIFI_DEBUG
1376 const char *iwl_mvm_get_tx_fail_reason(u32 status);
1377 #else
1378 static inline const char *iwl_mvm_get_tx_fail_reason(u32 status) { return ""; }
1379 #endif
1380 int iwl_mvm_flush_tx_path(struct iwl_mvm *mvm, u32 tfd_msk, u32 flags);
1381 int iwl_mvm_flush_sta(struct iwl_mvm *mvm, void *sta, bool internal, u32 flags);
1382 int iwl_mvm_flush_sta_tids(struct iwl_mvm *mvm, u32 sta_id,
1383 u16 tids, u32 flags);
1384
1385 void iwl_mvm_async_handlers_purge(struct iwl_mvm *mvm);
1386
1387 static inline void iwl_mvm_set_tx_cmd_ccmp(struct ieee80211_tx_info *info,
1388 struct iwl_tx_cmd *tx_cmd)
1389 {
1390 struct ieee80211_key_conf *keyconf = info->control.hw_key;
1391
1392 tx_cmd->sec_ctl = TX_CMD_SEC_CCM;
1393 memcpy(tx_cmd->key, keyconf->key, keyconf->keylen);
1394 }
1395
1396 static inline void iwl_mvm_wait_for_async_handlers(struct iwl_mvm *mvm)
1397 {
1398 flush_work(&mvm->async_handlers_wk);
1399 }
1400
1401 /* Statistics */
1402 void iwl_mvm_handle_rx_statistics(struct iwl_mvm *mvm,
1403 struct iwl_rx_packet *pkt);
1404 void iwl_mvm_rx_statistics(struct iwl_mvm *mvm,
1405 struct iwl_rx_cmd_buffer *rxb);
1406 int iwl_mvm_request_statistics(struct iwl_mvm *mvm, bool clear);
1407 void iwl_mvm_accu_radio_stats(struct iwl_mvm *mvm);
1408
1409 /* NVM */
1410 int iwl_nvm_init(struct iwl_mvm *mvm);
1411 int iwl_mvm_load_nvm_to_nic(struct iwl_mvm *mvm);
1412 int iwl_mvm_read_external_nvm(struct iwl_mvm *mvm);
1413
1414 static inline u8 iwl_mvm_get_valid_tx_ant(struct iwl_mvm *mvm)
1415 {
1416 return mvm->nvm_data && mvm->nvm_data->valid_tx_ant ?
1417 mvm->fw->valid_tx_ant & mvm->nvm_data->valid_tx_ant :
1418 mvm->fw->valid_tx_ant;
1419 }
1420
1421 static inline u8 iwl_mvm_get_valid_rx_ant(struct iwl_mvm *mvm)
1422 {
1423 return mvm->nvm_data && mvm->nvm_data->valid_rx_ant ?
1424 mvm->fw->valid_rx_ant & mvm->nvm_data->valid_rx_ant :
1425 mvm->fw->valid_rx_ant;
1426 }
1427
1428 static inline u32 iwl_mvm_get_phy_config(struct iwl_mvm *mvm)
1429 {
1430 u32 phy_config = ~(FW_PHY_CFG_TX_CHAIN |
1431 FW_PHY_CFG_RX_CHAIN);
1432 u32 valid_rx_ant = iwl_mvm_get_valid_rx_ant(mvm);
1433 u32 valid_tx_ant = iwl_mvm_get_valid_tx_ant(mvm);
1434
1435 phy_config |= valid_tx_ant << FW_PHY_CFG_TX_CHAIN_POS |
1436 valid_rx_ant << FW_PHY_CFG_RX_CHAIN_POS;
1437
1438 return mvm->fw->phy_config & phy_config;
1439 }
1440
1441 int iwl_mvm_up(struct iwl_mvm *mvm);
1442 int iwl_mvm_load_d3_fw(struct iwl_mvm *mvm);
1443
1444 int iwl_mvm_mac_setup_register(struct iwl_mvm *mvm);
1445 bool iwl_mvm_bcast_filter_build_cmd(struct iwl_mvm *mvm,
1446 struct iwl_bcast_filter_cmd *cmd);
1447
1448 /*
1449 * FW notifications / CMD responses handlers
1450 * Convention: iwl_mvm_rx_<NAME OF THE CMD>
1451 */
1452 void iwl_mvm_rx_rx_phy_cmd(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1453 void iwl_mvm_rx_rx_mpdu(struct iwl_mvm *mvm, struct napi_struct *napi,
1454 struct iwl_rx_cmd_buffer *rxb);
1455 void iwl_mvm_rx_mpdu_mq(struct iwl_mvm *mvm, struct napi_struct *napi,
1456 struct iwl_rx_cmd_buffer *rxb, int queue);
1457 void iwl_mvm_rx_frame_release(struct iwl_mvm *mvm, struct napi_struct *napi,
1458 struct iwl_rx_cmd_buffer *rxb, int queue);
1459 int iwl_mvm_notify_rx_queue(struct iwl_mvm *mvm, u32 rxq_mask,
1460 const u8 *data, u32 count);
1461 void iwl_mvm_rx_queue_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb,
1462 int queue);
1463 void iwl_mvm_rx_tx_cmd(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1464 void iwl_mvm_mfu_assert_dump_notif(struct iwl_mvm *mvm,
1465 struct iwl_rx_cmd_buffer *rxb);
1466 void iwl_mvm_rx_ba_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1467 void iwl_mvm_rx_ant_coupling_notif(struct iwl_mvm *mvm,
1468 struct iwl_rx_cmd_buffer *rxb);
1469 void iwl_mvm_rx_fw_error(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1470 void iwl_mvm_rx_card_state_notif(struct iwl_mvm *mvm,
1471 struct iwl_rx_cmd_buffer *rxb);
1472 void iwl_mvm_rx_mfuart_notif(struct iwl_mvm *mvm,
1473 struct iwl_rx_cmd_buffer *rxb);
1474 void iwl_mvm_rx_shared_mem_cfg_notif(struct iwl_mvm *mvm,
1475 struct iwl_rx_cmd_buffer *rxb);
1476
1477 /* MVM PHY */
1478 int iwl_mvm_phy_ctxt_add(struct iwl_mvm *mvm, struct iwl_mvm_phy_ctxt *ctxt,
1479 struct cfg80211_chan_def *chandef,
1480 u8 chains_static, u8 chains_dynamic);
1481 int iwl_mvm_phy_ctxt_changed(struct iwl_mvm *mvm, struct iwl_mvm_phy_ctxt *ctxt,
1482 struct cfg80211_chan_def *chandef,
1483 u8 chains_static, u8 chains_dynamic);
1484 void iwl_mvm_phy_ctxt_ref(struct iwl_mvm *mvm,
1485 struct iwl_mvm_phy_ctxt *ctxt);
1486 void iwl_mvm_phy_ctxt_unref(struct iwl_mvm *mvm,
1487 struct iwl_mvm_phy_ctxt *ctxt);
1488 int iwl_mvm_phy_ctx_count(struct iwl_mvm *mvm);
1489 u8 iwl_mvm_get_channel_width(struct cfg80211_chan_def *chandef);
1490 u8 iwl_mvm_get_ctrl_pos(struct cfg80211_chan_def *chandef);
1491
1492 /* MAC (virtual interface) programming */
1493 int iwl_mvm_mac_ctxt_init(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1494 int iwl_mvm_mac_ctxt_add(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1495 int iwl_mvm_mac_ctxt_changed(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1496 bool force_assoc_off, const u8 *bssid_override);
1497 int iwl_mvm_mac_ctxt_remove(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1498 u32 iwl_mvm_mac_get_queues_mask(struct ieee80211_vif *vif);
1499 int iwl_mvm_mac_ctxt_beacon_changed(struct iwl_mvm *mvm,
1500 struct ieee80211_vif *vif);
1501 void iwl_mvm_rx_beacon_notif(struct iwl_mvm *mvm,
1502 struct iwl_rx_cmd_buffer *rxb);
1503 void iwl_mvm_rx_missed_beacons_notif(struct iwl_mvm *mvm,
1504 struct iwl_rx_cmd_buffer *rxb);
1505 void iwl_mvm_rx_stored_beacon_notif(struct iwl_mvm *mvm,
1506 struct iwl_rx_cmd_buffer *rxb);
1507 void iwl_mvm_mu_mimo_grp_notif(struct iwl_mvm *mvm,
1508 struct iwl_rx_cmd_buffer *rxb);
1509 void iwl_mvm_sta_pm_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1510 void iwl_mvm_window_status_notif(struct iwl_mvm *mvm,
1511 struct iwl_rx_cmd_buffer *rxb);
1512 void iwl_mvm_mac_ctxt_recalc_tsf_id(struct iwl_mvm *mvm,
1513 struct ieee80211_vif *vif);
1514 unsigned long iwl_mvm_get_used_hw_queues(struct iwl_mvm *mvm,
1515 struct ieee80211_vif *exclude_vif);
1516 void iwl_mvm_channel_switch_noa_notif(struct iwl_mvm *mvm,
1517 struct iwl_rx_cmd_buffer *rxb);
1518 /* Bindings */
1519 int iwl_mvm_binding_add_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1520 int iwl_mvm_binding_remove_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1521
1522 /* Quota management */
1523 static inline size_t iwl_mvm_quota_cmd_size(struct iwl_mvm *mvm)
1524 {
1525 return iwl_mvm_has_quota_low_latency(mvm) ?
1526 sizeof(struct iwl_time_quota_cmd) :
1527 sizeof(struct iwl_time_quota_cmd_v1);
1528 }
1529
1530 static inline struct iwl_time_quota_data
1531 *iwl_mvm_quota_cmd_get_quota(struct iwl_mvm *mvm,
1532 struct iwl_time_quota_cmd *cmd,
1533 int i)
1534 {
1535 struct iwl_time_quota_data_v1 *quotas;
1536
1537 if (iwl_mvm_has_quota_low_latency(mvm))
1538 return &cmd->quotas[i];
1539
1540 quotas = (struct iwl_time_quota_data_v1 *)cmd->quotas;
1541 return (struct iwl_time_quota_data *)&quotas[i];
1542 }
1543
1544 int iwl_mvm_update_quotas(struct iwl_mvm *mvm, bool force_upload,
1545 struct ieee80211_vif *disabled_vif);
1546
1547 /* Scanning */
1548 int iwl_mvm_reg_scan_start(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1549 struct cfg80211_scan_request *req,
1550 struct ieee80211_scan_ies *ies);
1551 int iwl_mvm_scan_size(struct iwl_mvm *mvm);
1552 int iwl_mvm_scan_stop(struct iwl_mvm *mvm, int type, bool notify);
1553 int iwl_mvm_max_scan_ie_len(struct iwl_mvm *mvm);
1554 void iwl_mvm_report_scan_aborted(struct iwl_mvm *mvm);
1555 void iwl_mvm_scan_timeout_wk(struct work_struct *work);
1556
1557 /* Scheduled scan */
1558 void iwl_mvm_rx_lmac_scan_complete_notif(struct iwl_mvm *mvm,
1559 struct iwl_rx_cmd_buffer *rxb);
1560 void iwl_mvm_rx_lmac_scan_iter_complete_notif(struct iwl_mvm *mvm,
1561 struct iwl_rx_cmd_buffer *rxb);
1562 int iwl_mvm_sched_scan_start(struct iwl_mvm *mvm,
1563 struct ieee80211_vif *vif,
1564 struct cfg80211_sched_scan_request *req,
1565 struct ieee80211_scan_ies *ies,
1566 int type);
1567 void iwl_mvm_rx_scan_match_found(struct iwl_mvm *mvm,
1568 struct iwl_rx_cmd_buffer *rxb);
1569
1570 /* UMAC scan */
1571 int iwl_mvm_config_scan(struct iwl_mvm *mvm);
1572 void iwl_mvm_rx_umac_scan_complete_notif(struct iwl_mvm *mvm,
1573 struct iwl_rx_cmd_buffer *rxb);
1574 void iwl_mvm_rx_umac_scan_iter_complete_notif(struct iwl_mvm *mvm,
1575 struct iwl_rx_cmd_buffer *rxb);
1576
1577 /* MVM debugfs */
1578 #ifdef CONFIG_IWLWIFI_DEBUGFS
1579 int iwl_mvm_dbgfs_register(struct iwl_mvm *mvm, struct dentry *dbgfs_dir);
1580 void iwl_mvm_vif_dbgfs_register(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1581 void iwl_mvm_vif_dbgfs_clean(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1582 #else
1583 static inline int iwl_mvm_dbgfs_register(struct iwl_mvm *mvm,
1584 struct dentry *dbgfs_dir)
1585 {
1586 return 0;
1587 }
1588 static inline void
1589 iwl_mvm_vif_dbgfs_register(struct iwl_mvm *mvm, struct ieee80211_vif *vif)
1590 {
1591 }
1592 static inline void
1593 iwl_mvm_vif_dbgfs_clean(struct iwl_mvm *mvm, struct ieee80211_vif *vif)
1594 {
1595 }
1596 #endif /* CONFIG_IWLWIFI_DEBUGFS */
1597
1598 /* rate scaling */
1599 int iwl_mvm_send_lq_cmd(struct iwl_mvm *mvm, struct iwl_lq_cmd *lq, bool init);
1600 void iwl_mvm_update_frame_stats(struct iwl_mvm *mvm, u32 rate, bool agg);
1601 int rs_pretty_print_rate(char *buf, const u32 rate);
1602 void rs_update_last_rssi(struct iwl_mvm *mvm,
1603 struct iwl_lq_sta *lq_sta,
1604 struct ieee80211_rx_status *rx_status);
1605
1606 /* power management */
1607 int iwl_mvm_power_update_device(struct iwl_mvm *mvm);
1608 int iwl_mvm_power_update_mac(struct iwl_mvm *mvm);
1609 int iwl_mvm_power_update_ps(struct iwl_mvm *mvm);
1610 int iwl_mvm_power_mac_dbgfs_read(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1611 char *buf, int bufsz);
1612
1613 void iwl_mvm_power_vif_assoc(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1614 void iwl_mvm_power_uapsd_misbehaving_ap_notif(struct iwl_mvm *mvm,
1615 struct iwl_rx_cmd_buffer *rxb);
1616
1617 #ifdef CONFIG_IWLWIFI_LEDS
1618 int iwl_mvm_leds_init(struct iwl_mvm *mvm);
1619 void iwl_mvm_leds_exit(struct iwl_mvm *mvm);
1620 void iwl_mvm_leds_sync(struct iwl_mvm *mvm);
1621 #else
1622 static inline int iwl_mvm_leds_init(struct iwl_mvm *mvm)
1623 {
1624 return 0;
1625 }
1626 static inline void iwl_mvm_leds_exit(struct iwl_mvm *mvm)
1627 {
1628 }
1629 static inline void iwl_mvm_leds_sync(struct iwl_mvm *mvm)
1630 {
1631 }
1632 #endif
1633
1634 /* D3 (WoWLAN, NetDetect) */
1635 int iwl_mvm_suspend(struct ieee80211_hw *hw, struct cfg80211_wowlan *wowlan);
1636 int iwl_mvm_resume(struct ieee80211_hw *hw);
1637 void iwl_mvm_set_wakeup(struct ieee80211_hw *hw, bool enabled);
1638 void iwl_mvm_set_rekey_data(struct ieee80211_hw *hw,
1639 struct ieee80211_vif *vif,
1640 struct cfg80211_gtk_rekey_data *data);
1641 void iwl_mvm_ipv6_addr_change(struct ieee80211_hw *hw,
1642 struct ieee80211_vif *vif,
1643 struct inet6_dev *idev);
1644 void iwl_mvm_set_default_unicast_key(struct ieee80211_hw *hw,
1645 struct ieee80211_vif *vif, int idx);
1646 extern const struct file_operations iwl_dbgfs_d3_test_ops;
1647 #ifdef CONFIG_PM
1648 int iwl_mvm_wowlan_config_key_params(struct iwl_mvm *mvm,
1649 struct ieee80211_vif *vif,
1650 bool host_awake,
1651 u32 cmd_flags);
1652 void iwl_mvm_d0i3_update_keys(struct iwl_mvm *mvm,
1653 struct ieee80211_vif *vif,
1654 struct iwl_wowlan_status *status);
1655 void iwl_mvm_set_last_nonqos_seq(struct iwl_mvm *mvm,
1656 struct ieee80211_vif *vif);
1657 #else
1658 static inline int iwl_mvm_wowlan_config_key_params(struct iwl_mvm *mvm,
1659 struct ieee80211_vif *vif,
1660 bool host_awake,
1661 u32 cmd_flags)
1662 {
1663 return 0;
1664 }
1665
1666 static inline void iwl_mvm_d0i3_update_keys(struct iwl_mvm *mvm,
1667 struct ieee80211_vif *vif,
1668 struct iwl_wowlan_status *status)
1669 {
1670 }
1671
1672 static inline void
1673 iwl_mvm_set_last_nonqos_seq(struct iwl_mvm *mvm, struct ieee80211_vif *vif)
1674 {
1675 }
1676 #endif
1677 void iwl_mvm_set_wowlan_qos_seq(struct iwl_mvm_sta *mvm_ap_sta,
1678 struct iwl_wowlan_config_cmd *cmd);
1679 int iwl_mvm_send_proto_offload(struct iwl_mvm *mvm,
1680 struct ieee80211_vif *vif,
1681 bool disable_offloading,
1682 bool offload_ns,
1683 u32 cmd_flags);
1684
1685 /* D0i3 */
1686 void iwl_mvm_ref(struct iwl_mvm *mvm, enum iwl_mvm_ref_type ref_type);
1687 void iwl_mvm_unref(struct iwl_mvm *mvm, enum iwl_mvm_ref_type ref_type);
1688 int iwl_mvm_ref_sync(struct iwl_mvm *mvm, enum iwl_mvm_ref_type ref_type);
1689 bool iwl_mvm_ref_taken(struct iwl_mvm *mvm);
1690 void iwl_mvm_d0i3_enable_tx(struct iwl_mvm *mvm, __le16 *qos_seq);
1691 int iwl_mvm_enter_d0i3(struct iwl_op_mode *op_mode);
1692 int iwl_mvm_exit_d0i3(struct iwl_op_mode *op_mode);
1693 int _iwl_mvm_exit_d0i3(struct iwl_mvm *mvm);
1694
1695 /* BT Coex */
1696 int iwl_mvm_send_bt_init_conf(struct iwl_mvm *mvm);
1697 void iwl_mvm_rx_bt_coex_notif(struct iwl_mvm *mvm,
1698 struct iwl_rx_cmd_buffer *rxb);
1699 void iwl_mvm_bt_rssi_event(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1700 enum ieee80211_rssi_event_data);
1701 void iwl_mvm_bt_coex_vif_change(struct iwl_mvm *mvm);
1702 u16 iwl_mvm_coex_agg_time_limit(struct iwl_mvm *mvm,
1703 struct ieee80211_sta *sta);
1704 bool iwl_mvm_bt_coex_is_mimo_allowed(struct iwl_mvm *mvm,
1705 struct ieee80211_sta *sta);
1706 bool iwl_mvm_bt_coex_is_ant_avail(struct iwl_mvm *mvm, u8 ant);
1707 bool iwl_mvm_bt_coex_is_shared_ant_avail(struct iwl_mvm *mvm);
1708 bool iwl_mvm_bt_coex_is_tpc_allowed(struct iwl_mvm *mvm,
1709 enum nl80211_band band);
1710 u8 iwl_mvm_bt_coex_tx_prio(struct iwl_mvm *mvm, struct ieee80211_hdr *hdr,
1711 struct ieee80211_tx_info *info, u8 ac);
1712
1713 /* beacon filtering */
1714 #ifdef CONFIG_IWLWIFI_DEBUGFS
1715 void
1716 iwl_mvm_beacon_filter_debugfs_parameters(struct ieee80211_vif *vif,
1717 struct iwl_beacon_filter_cmd *cmd);
1718 #else
1719 static inline void
1720 iwl_mvm_beacon_filter_debugfs_parameters(struct ieee80211_vif *vif,
1721 struct iwl_beacon_filter_cmd *cmd)
1722 {}
1723 #endif
1724 int iwl_mvm_update_d0i3_power_mode(struct iwl_mvm *mvm,
1725 struct ieee80211_vif *vif,
1726 bool enable, u32 flags);
1727 int iwl_mvm_enable_beacon_filter(struct iwl_mvm *mvm,
1728 struct ieee80211_vif *vif,
1729 u32 flags);
1730 int iwl_mvm_disable_beacon_filter(struct iwl_mvm *mvm,
1731 struct ieee80211_vif *vif,
1732 u32 flags);
1733 /* SMPS */
1734 void iwl_mvm_update_smps(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1735 enum iwl_mvm_smps_type_request req_type,
1736 enum ieee80211_smps_mode smps_request);
1737 bool iwl_mvm_rx_diversity_allowed(struct iwl_mvm *mvm);
1738
1739 /* Low latency */
1740 int iwl_mvm_update_low_latency(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1741 bool value);
1742 /* get SystemLowLatencyMode - only needed for beacon threshold? */
1743 bool iwl_mvm_low_latency(struct iwl_mvm *mvm);
1744 /* get VMACLowLatencyMode */
1745 static inline bool iwl_mvm_vif_low_latency(struct iwl_mvm_vif *mvmvif)
1746 {
1747 /*
1748 * should this consider associated/active/... state?
1749 *
1750 * Normally low-latency should only be active on interfaces
1751 * that are active, but at least with debugfs it can also be
1752 * enabled on interfaces that aren't active. However, when
1753 * interface aren't active then they aren't added into the
1754 * binding, so this has no real impact. For now, just return
1755 * the current desired low-latency state.
1756 */
1757 return mvmvif->low_latency_dbgfs ||
1758 mvmvif->low_latency_traffic ||
1759 mvmvif->low_latency_vcmd;
1760 }
1761
1762 /* hw scheduler queue config */
1763 bool iwl_mvm_enable_txq(struct iwl_mvm *mvm, int queue, int mac80211_queue,
1764 u16 ssn, const struct iwl_trans_txq_scd_cfg *cfg,
1765 unsigned int wdg_timeout);
1766 int iwl_mvm_tvqm_enable_txq(struct iwl_mvm *mvm, int mac80211_queue,
1767 u8 sta_id, u8 tid, unsigned int timeout);
1768
1769 int iwl_mvm_disable_txq(struct iwl_mvm *mvm, int queue, int mac80211_queue,
1770 u8 tid, u8 flags);
1771 int iwl_mvm_find_free_queue(struct iwl_mvm *mvm, u8 sta_id, u8 minq, u8 maxq);
1772
1773 /* Return a bitmask with all the hw supported queues, except for the
1774 * command queue, which can't be flushed.
1775 */
1776 static inline u32 iwl_mvm_flushable_queues(struct iwl_mvm *mvm)
1777 {
1778 return ((BIT(mvm->cfg->base_params->num_of_queues) - 1) &
1779 ~BIT(IWL_MVM_DQA_CMD_QUEUE));
1780 }
1781
1782 static inline void iwl_mvm_stop_device(struct iwl_mvm *mvm)
1783 {
1784 iwl_free_fw_paging(&mvm->fwrt);
1785 clear_bit(IWL_MVM_STATUS_FIRMWARE_RUNNING, &mvm->status);
1786 iwl_fw_dump_conf_clear(&mvm->fwrt);
1787 iwl_trans_stop_device(mvm->trans);
1788 }
1789
1790 /* Stop/start all mac queues in a given bitmap */
1791 void iwl_mvm_start_mac_queues(struct iwl_mvm *mvm, unsigned long mq);
1792 void iwl_mvm_stop_mac_queues(struct iwl_mvm *mvm, unsigned long mq);
1793
1794 /* Re-configure the SCD for a queue that has already been configured */
1795 int iwl_mvm_reconfig_scd(struct iwl_mvm *mvm, int queue, int fifo, int sta_id,
1796 int tid, int frame_limit, u16 ssn);
1797
1798 /* Thermal management and CT-kill */
1799 void iwl_mvm_tt_tx_backoff(struct iwl_mvm *mvm, u32 backoff);
1800 void iwl_mvm_tt_temp_changed(struct iwl_mvm *mvm, u32 temp);
1801 void iwl_mvm_temp_notif(struct iwl_mvm *mvm,
1802 struct iwl_rx_cmd_buffer *rxb);
1803 void iwl_mvm_tt_handler(struct iwl_mvm *mvm);
1804 void iwl_mvm_thermal_initialize(struct iwl_mvm *mvm, u32 min_backoff);
1805 void iwl_mvm_thermal_exit(struct iwl_mvm *mvm);
1806 void iwl_mvm_set_hw_ctkill_state(struct iwl_mvm *mvm, bool state);
1807 int iwl_mvm_get_temp(struct iwl_mvm *mvm, s32 *temp);
1808 void iwl_mvm_ct_kill_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1809 void iwl_mvm_enter_ctkill(struct iwl_mvm *mvm);
1810 int iwl_mvm_send_temp_report_ths_cmd(struct iwl_mvm *mvm);
1811 int iwl_mvm_ctdp_command(struct iwl_mvm *mvm, u32 op, u32 budget);
1812
1813 /* Location Aware Regulatory */
1814 struct iwl_mcc_update_resp *
1815 iwl_mvm_update_mcc(struct iwl_mvm *mvm, const char *alpha2,
1816 enum iwl_mcc_source src_id);
1817 int iwl_mvm_init_mcc(struct iwl_mvm *mvm);
1818 void iwl_mvm_rx_chub_update_mcc(struct iwl_mvm *mvm,
1819 struct iwl_rx_cmd_buffer *rxb);
1820 struct ieee80211_regdomain *iwl_mvm_get_regdomain(struct wiphy *wiphy,
1821 const char *alpha2,
1822 enum iwl_mcc_source src_id,
1823 bool *changed);
1824 struct ieee80211_regdomain *iwl_mvm_get_current_regdomain(struct iwl_mvm *mvm,
1825 bool *changed);
1826 int iwl_mvm_init_fw_regd(struct iwl_mvm *mvm);
1827 void iwl_mvm_update_changed_regdom(struct iwl_mvm *mvm);
1828
1829 /* smart fifo */
1830 int iwl_mvm_sf_update(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1831 bool added_vif);
1832
1833 /* TDLS */
1834
1835 /*
1836 * We use TID 4 (VI) as a FW-used-only TID when TDLS connections are present.
1837 * This TID is marked as used vs the AP and all connected TDLS peers.
1838 */
1839 #define IWL_MVM_TDLS_FW_TID 4
1840
1841 int iwl_mvm_tdls_sta_count(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1842 void iwl_mvm_teardown_tdls_peers(struct iwl_mvm *mvm);
1843 void iwl_mvm_recalc_tdls_state(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1844 bool sta_added);
1845 void iwl_mvm_mac_mgd_protect_tdls_discover(struct ieee80211_hw *hw,
1846 struct ieee80211_vif *vif);
1847 int iwl_mvm_tdls_channel_switch(struct ieee80211_hw *hw,
1848 struct ieee80211_vif *vif,
1849 struct ieee80211_sta *sta, u8 oper_class,
1850 struct cfg80211_chan_def *chandef,
1851 struct sk_buff *tmpl_skb, u32 ch_sw_tm_ie);
1852 void iwl_mvm_tdls_recv_channel_switch(struct ieee80211_hw *hw,
1853 struct ieee80211_vif *vif,
1854 struct ieee80211_tdls_ch_sw_params *params);
1855 void iwl_mvm_tdls_cancel_channel_switch(struct ieee80211_hw *hw,
1856 struct ieee80211_vif *vif,
1857 struct ieee80211_sta *sta);
1858 void iwl_mvm_rx_tdls_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1859 void iwl_mvm_tdls_ch_switch_work(struct work_struct *work);
1860
1861 void iwl_mvm_sync_rx_queues_internal(struct iwl_mvm *mvm,
1862 struct iwl_mvm_internal_rxq_notif *notif,
1863 u32 size);
1864 void iwl_mvm_reorder_timer_expired(struct timer_list *t);
1865 struct ieee80211_vif *iwl_mvm_get_bss_vif(struct iwl_mvm *mvm);
1866 bool iwl_mvm_is_vif_assoc(struct iwl_mvm *mvm);
1867
1868 void iwl_mvm_inactivity_check(struct iwl_mvm *mvm);
1869
1870 void iwl_mvm_nic_restart(struct iwl_mvm *mvm, bool fw_error);
1871 unsigned int iwl_mvm_get_wd_timeout(struct iwl_mvm *mvm,
1872 struct ieee80211_vif *vif,
1873 bool tdls, bool cmd_q);
1874 void iwl_mvm_connection_loss(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1875 const char *errmsg);
1876 void iwl_mvm_event_frame_timeout_callback(struct iwl_mvm *mvm,
1877 struct ieee80211_vif *vif,
1878 const struct ieee80211_sta *sta,
1879 u16 tid);
1880
1881 int iwl_mvm_sar_select_profile(struct iwl_mvm *mvm, int prof_a, int prof_b);
1882 int iwl_mvm_get_sar_geo_profile(struct iwl_mvm *mvm);
1883
1884 #endif /* __IWL_MVM_H__ */