]>
Commit | Line | Data |
---|---|---|
8ca151b5 JB |
1 | /****************************************************************************** |
2 | * | |
3 | * This file is provided under a dual BSD/GPLv2 license. When using or | |
4 | * redistributing this file, you may do so under either license. | |
5 | * | |
6 | * GPL LICENSE SUMMARY | |
7 | * | |
51368bf7 | 8 | * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved. |
8b4139dc | 9 | * Copyright(c) 2013 - 2014 Intel Mobile Communications GmbH |
8ca151b5 JB |
10 | * |
11 | * This program is free software; you can redistribute it and/or modify | |
12 | * it under the terms of version 2 of the GNU General Public License as | |
13 | * published by the Free Software Foundation. | |
14 | * | |
15 | * This program is distributed in the hope that it will be useful, but | |
16 | * WITHOUT ANY WARRANTY; without even the implied warranty of | |
17 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU | |
18 | * General Public License for more details. | |
19 | * | |
20 | * You should have received a copy of the GNU General Public License | |
21 | * along with this program; if not, write to the Free Software | |
22 | * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110, | |
23 | * USA | |
24 | * | |
25 | * The full GNU General Public License is included in this distribution | |
410dc5aa | 26 | * in the file called COPYING. |
8ca151b5 JB |
27 | * |
28 | * Contact Information: | |
29 | * Intel Linux Wireless <ilw@linux.intel.com> | |
30 | * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497 | |
31 | * | |
32 | * BSD LICENSE | |
33 | * | |
51368bf7 | 34 | * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved. |
8b4139dc | 35 | * Copyright(c) 2013 - 2014 Intel Mobile Communications GmbH |
8ca151b5 JB |
36 | * All rights reserved. |
37 | * | |
38 | * Redistribution and use in source and binary forms, with or without | |
39 | * modification, are permitted provided that the following conditions | |
40 | * are met: | |
41 | * | |
42 | * * Redistributions of source code must retain the above copyright | |
43 | * notice, this list of conditions and the following disclaimer. | |
44 | * * Redistributions in binary form must reproduce the above copyright | |
45 | * notice, this list of conditions and the following disclaimer in | |
46 | * the documentation and/or other materials provided with the | |
47 | * distribution. | |
48 | * * Neither the name Intel Corporation nor the names of its | |
49 | * contributors may be used to endorse or promote products derived | |
50 | * from this software without specific prior written permission. | |
51 | * | |
52 | * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS | |
53 | * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT | |
54 | * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR | |
55 | * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT | |
56 | * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, | |
57 | * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT | |
58 | * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, | |
59 | * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY | |
60 | * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT | |
61 | * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE | |
62 | * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | |
63 | * | |
64 | *****************************************************************************/ | |
65 | ||
66 | #include <linux/jiffies.h> | |
67 | #include <net/mac80211.h> | |
68 | ||
69 | #include "iwl-notif-wait.h" | |
70 | #include "iwl-trans.h" | |
71 | #include "fw-api.h" | |
72 | #include "time-event.h" | |
73 | #include "mvm.h" | |
74 | #include "iwl-io.h" | |
75 | #include "iwl-prph.h" | |
76 | ||
e635c797 IP |
77 | /* |
78 | * For the high priority TE use a time event type that has similar priority to | |
79 | * the FW's action scan priority. | |
456f6ddf | 80 | */ |
e635c797 IP |
81 | #define IWL_MVM_ROC_TE_TYPE_NORMAL TE_P2P_DEVICE_DISCOVERABLE |
82 | #define IWL_MVM_ROC_TE_TYPE_MGMT_TX TE_P2P_CLIENT_ASSOC | |
456f6ddf | 83 | |
8ca151b5 JB |
84 | void iwl_mvm_te_clear_data(struct iwl_mvm *mvm, |
85 | struct iwl_mvm_time_event_data *te_data) | |
86 | { | |
87 | lockdep_assert_held(&mvm->time_event_lock); | |
88 | ||
89 | if (te_data->id == TE_MAX) | |
90 | return; | |
91 | ||
92 | list_del(&te_data->list); | |
93 | te_data->running = false; | |
94 | te_data->uid = 0; | |
95 | te_data->id = TE_MAX; | |
96 | te_data->vif = NULL; | |
97 | } | |
98 | ||
99 | void iwl_mvm_roc_done_wk(struct work_struct *wk) | |
100 | { | |
101 | struct iwl_mvm *mvm = container_of(wk, struct iwl_mvm, roc_done_wk); | |
b112889c AM |
102 | u32 queues = 0; |
103 | ||
104 | /* | |
105 | * Clear the ROC_RUNNING /ROC_AUX_RUNNING status bit. | |
106 | * This will cause the TX path to drop offchannel transmissions. | |
107 | * That would also be done by mac80211, but it is racy, in particular | |
108 | * in the case that the time event actually completed in the firmware | |
109 | * (which is handled in iwl_mvm_te_handle_notif). | |
110 | */ | |
111 | if (test_and_clear_bit(IWL_MVM_STATUS_ROC_RUNNING, &mvm->status)) | |
112 | queues |= BIT(IWL_MVM_OFFCHANNEL_QUEUE); | |
113 | if (test_and_clear_bit(IWL_MVM_STATUS_ROC_AUX_RUNNING, &mvm->status)) | |
114 | queues |= BIT(mvm->aux_queue); | |
115 | ||
116 | iwl_mvm_unref(mvm, IWL_MVM_REF_ROC); | |
8ca151b5 JB |
117 | |
118 | synchronize_net(); | |
119 | ||
120 | /* | |
121 | * Flush the offchannel queue -- this is called when the time | |
122 | * event finishes or is cancelled, so that frames queued for it | |
123 | * won't get stuck on the queue and be transmitted in the next | |
124 | * time event. | |
125 | * We have to send the command asynchronously since this cannot | |
126 | * be under the mutex for locking reasons, but that's not an | |
127 | * issue as it will have to complete before the next command is | |
128 | * executed, and a new time event means a new command. | |
129 | */ | |
b112889c | 130 | iwl_mvm_flush_tx_path(mvm, queues, false); |
8ca151b5 JB |
131 | } |
132 | ||
133 | static void iwl_mvm_roc_finished(struct iwl_mvm *mvm) | |
134 | { | |
8ca151b5 JB |
135 | /* |
136 | * Of course, our status bit is just as racy as mac80211, so in | |
137 | * addition, fire off the work struct which will drop all frames | |
138 | * from the hardware queues that made it through the race. First | |
139 | * it will of course synchronize the TX path to make sure that | |
140 | * any *new* TX will be rejected. | |
141 | */ | |
142 | schedule_work(&mvm->roc_done_wk); | |
143 | } | |
144 | ||
7f0a7c67 AO |
145 | static void iwl_mvm_csa_noa_start(struct iwl_mvm *mvm) |
146 | { | |
147 | struct ieee80211_vif *csa_vif; | |
148 | ||
149 | rcu_read_lock(); | |
150 | ||
151 | csa_vif = rcu_dereference(mvm->csa_vif); | |
152 | if (!csa_vif || !csa_vif->csa_active) | |
153 | goto out_unlock; | |
154 | ||
155 | IWL_DEBUG_TE(mvm, "CSA NOA started\n"); | |
156 | ||
157 | /* | |
158 | * CSA NoA is started but we still have beacons to | |
159 | * transmit on the current channel. | |
160 | * So we just do nothing here and the switch | |
161 | * will be performed on the last TBTT. | |
162 | */ | |
163 | if (!ieee80211_csa_is_complete(csa_vif)) { | |
164 | IWL_WARN(mvm, "CSA NOA started too early\n"); | |
165 | goto out_unlock; | |
166 | } | |
167 | ||
168 | ieee80211_csa_finish(csa_vif); | |
169 | ||
170 | rcu_read_unlock(); | |
171 | ||
172 | RCU_INIT_POINTER(mvm->csa_vif, NULL); | |
173 | ||
174 | return; | |
175 | ||
176 | out_unlock: | |
177 | rcu_read_unlock(); | |
178 | } | |
179 | ||
05739794 JB |
180 | static bool iwl_mvm_te_check_disconnect(struct iwl_mvm *mvm, |
181 | struct ieee80211_vif *vif, | |
182 | const char *errmsg) | |
183 | { | |
184 | if (vif->type != NL80211_IFTYPE_STATION) | |
185 | return false; | |
186 | if (vif->bss_conf.assoc && vif->bss_conf.dtim_period) | |
187 | return false; | |
188 | if (errmsg) | |
189 | IWL_ERR(mvm, "%s\n", errmsg); | |
190 | ieee80211_connection_loss(vif); | |
191 | return true; | |
192 | } | |
193 | ||
8ca151b5 JB |
194 | /* |
195 | * Handles a FW notification for an event that is known to the driver. | |
196 | * | |
197 | * @mvm: the mvm component | |
198 | * @te_data: the time event data | |
199 | * @notif: the notification data corresponding the time event data. | |
200 | */ | |
201 | static void iwl_mvm_te_handle_notif(struct iwl_mvm *mvm, | |
202 | struct iwl_mvm_time_event_data *te_data, | |
203 | struct iwl_time_event_notif *notif) | |
204 | { | |
205 | lockdep_assert_held(&mvm->time_event_lock); | |
206 | ||
207 | IWL_DEBUG_TE(mvm, "Handle time event notif - UID = 0x%x action %d\n", | |
208 | le32_to_cpu(notif->unique_id), | |
209 | le32_to_cpu(notif->action)); | |
210 | ||
211 | /* | |
212 | * The FW sends the start/end time event notifications even for events | |
213 | * that it fails to schedule. This is indicated in the status field of | |
214 | * the notification. This happens in cases that the scheduler cannot | |
215 | * find a schedule that can handle the event (for example requesting a | |
216 | * P2P Device discoveribility, while there are other higher priority | |
217 | * events in the system). | |
218 | */ | |
9fc3fe96 EG |
219 | if (!le32_to_cpu(notif->status)) { |
220 | bool start = le32_to_cpu(notif->action) & | |
221 | TE_V2_NOTIF_HOST_EVENT_START; | |
222 | IWL_WARN(mvm, "Time Event %s notification failure\n", | |
223 | start ? "start" : "end"); | |
05739794 JB |
224 | if (iwl_mvm_te_check_disconnect(mvm, te_data->vif, NULL)) { |
225 | iwl_mvm_te_clear_data(mvm, te_data); | |
226 | return; | |
227 | } | |
228 | } | |
8ca151b5 | 229 | |
f8f03c3e | 230 | if (le32_to_cpu(notif->action) & TE_V2_NOTIF_HOST_EVENT_END) { |
8ca151b5 JB |
231 | IWL_DEBUG_TE(mvm, |
232 | "TE ended - current time %lu, estimated end %lu\n", | |
233 | jiffies, te_data->end_jiffies); | |
234 | ||
235 | if (te_data->vif->type == NL80211_IFTYPE_P2P_DEVICE) { | |
236 | ieee80211_remain_on_channel_expired(mvm->hw); | |
237 | iwl_mvm_roc_finished(mvm); | |
238 | } | |
239 | ||
240 | /* | |
241 | * By now, we should have finished association | |
242 | * and know the dtim period. | |
243 | */ | |
05739794 | 244 | iwl_mvm_te_check_disconnect(mvm, te_data->vif, |
2e515bf0 | 245 | "No association and the time event is over already..."); |
8ca151b5 | 246 | iwl_mvm_te_clear_data(mvm, te_data); |
f8f03c3e | 247 | } else if (le32_to_cpu(notif->action) & TE_V2_NOTIF_HOST_EVENT_START) { |
8ca151b5 | 248 | te_data->running = true; |
e7f1935c | 249 | te_data->end_jiffies = TU_TO_EXP_TIME(te_data->duration); |
8ca151b5 JB |
250 | |
251 | if (te_data->vif->type == NL80211_IFTYPE_P2P_DEVICE) { | |
252 | set_bit(IWL_MVM_STATUS_ROC_RUNNING, &mvm->status); | |
9f45c36d | 253 | iwl_mvm_ref(mvm, IWL_MVM_REF_ROC); |
8ca151b5 | 254 | ieee80211_ready_on_channel(mvm->hw); |
7f0a7c67 AO |
255 | } else if (te_data->vif->type == NL80211_IFTYPE_AP) { |
256 | if (le32_to_cpu(notif->status)) | |
257 | iwl_mvm_csa_noa_start(mvm); | |
258 | else | |
259 | IWL_DEBUG_TE(mvm, "CSA NOA failed to start\n"); | |
260 | ||
261 | /* we don't need it anymore */ | |
262 | iwl_mvm_te_clear_data(mvm, te_data); | |
8ca151b5 JB |
263 | } |
264 | } else { | |
265 | IWL_WARN(mvm, "Got TE with unknown action\n"); | |
266 | } | |
267 | } | |
268 | ||
b112889c AM |
269 | /* |
270 | * Handle A Aux ROC time event | |
271 | */ | |
272 | static int iwl_mvm_aux_roc_te_handle_notif(struct iwl_mvm *mvm, | |
273 | struct iwl_time_event_notif *notif) | |
274 | { | |
275 | struct iwl_mvm_time_event_data *te_data, *tmp; | |
276 | bool aux_roc_te = false; | |
277 | ||
278 | list_for_each_entry_safe(te_data, tmp, &mvm->aux_roc_te_list, list) { | |
279 | if (le32_to_cpu(notif->unique_id) == te_data->uid) { | |
280 | aux_roc_te = true; | |
281 | break; | |
282 | } | |
283 | } | |
284 | if (!aux_roc_te) /* Not a Aux ROC time event */ | |
285 | return -EINVAL; | |
286 | ||
287 | if (!le32_to_cpu(notif->status)) { | |
288 | IWL_DEBUG_TE(mvm, | |
289 | "ERROR: Aux ROC Time Event %s notification failure\n", | |
290 | (le32_to_cpu(notif->action) & | |
291 | TE_V2_NOTIF_HOST_EVENT_START) ? "start" : "end"); | |
292 | return -EINVAL; | |
293 | } | |
294 | ||
295 | IWL_DEBUG_TE(mvm, | |
296 | "Aux ROC time event notification - UID = 0x%x action %d\n", | |
297 | le32_to_cpu(notif->unique_id), | |
298 | le32_to_cpu(notif->action)); | |
299 | ||
300 | if (le32_to_cpu(notif->action) == TE_V2_NOTIF_HOST_EVENT_END) { | |
301 | /* End TE, notify mac80211 */ | |
302 | ieee80211_remain_on_channel_expired(mvm->hw); | |
303 | iwl_mvm_roc_finished(mvm); /* flush aux queue */ | |
304 | list_del(&te_data->list); /* remove from list */ | |
305 | te_data->running = false; | |
306 | te_data->vif = NULL; | |
307 | te_data->uid = 0; | |
a6cc5163 | 308 | te_data->id = TE_MAX; |
b112889c | 309 | } else if (le32_to_cpu(notif->action) == TE_V2_NOTIF_HOST_EVENT_START) { |
b112889c AM |
310 | set_bit(IWL_MVM_STATUS_ROC_AUX_RUNNING, &mvm->status); |
311 | te_data->running = true; | |
312 | ieee80211_ready_on_channel(mvm->hw); /* Start TE */ | |
313 | } else { | |
314 | IWL_DEBUG_TE(mvm, | |
315 | "ERROR: Unknown Aux ROC Time Event (action = %d)\n", | |
316 | le32_to_cpu(notif->action)); | |
317 | return -EINVAL; | |
318 | } | |
319 | ||
320 | return 0; | |
321 | } | |
322 | ||
8ca151b5 JB |
323 | /* |
324 | * The Rx handler for time event notifications | |
325 | */ | |
326 | int iwl_mvm_rx_time_event_notif(struct iwl_mvm *mvm, | |
327 | struct iwl_rx_cmd_buffer *rxb, | |
328 | struct iwl_device_cmd *cmd) | |
329 | { | |
330 | struct iwl_rx_packet *pkt = rxb_addr(rxb); | |
331 | struct iwl_time_event_notif *notif = (void *)pkt->data; | |
332 | struct iwl_mvm_time_event_data *te_data, *tmp; | |
333 | ||
334 | IWL_DEBUG_TE(mvm, "Time event notification - UID = 0x%x action %d\n", | |
335 | le32_to_cpu(notif->unique_id), | |
336 | le32_to_cpu(notif->action)); | |
337 | ||
338 | spin_lock_bh(&mvm->time_event_lock); | |
b112889c AM |
339 | /* This time event is triggered for Aux ROC request */ |
340 | if (!iwl_mvm_aux_roc_te_handle_notif(mvm, notif)) | |
341 | goto unlock; | |
342 | ||
8ca151b5 JB |
343 | list_for_each_entry_safe(te_data, tmp, &mvm->time_event_list, list) { |
344 | if (le32_to_cpu(notif->unique_id) == te_data->uid) | |
345 | iwl_mvm_te_handle_notif(mvm, te_data, notif); | |
346 | } | |
b112889c | 347 | unlock: |
8ca151b5 JB |
348 | spin_unlock_bh(&mvm->time_event_lock); |
349 | ||
350 | return 0; | |
351 | } | |
352 | ||
d20d37bc LK |
353 | static bool iwl_mvm_te_notif(struct iwl_notif_wait_data *notif_wait, |
354 | struct iwl_rx_packet *pkt, void *data) | |
355 | { | |
356 | struct iwl_mvm *mvm = | |
357 | container_of(notif_wait, struct iwl_mvm, notif_wait); | |
358 | struct iwl_mvm_time_event_data *te_data = data; | |
359 | struct iwl_time_event_notif *resp; | |
360 | int resp_len = iwl_rx_packet_payload_len(pkt); | |
361 | ||
362 | if (WARN_ON(pkt->hdr.cmd != TIME_EVENT_NOTIFICATION)) | |
363 | return true; | |
364 | ||
365 | if (WARN_ON_ONCE(resp_len != sizeof(*resp))) { | |
366 | IWL_ERR(mvm, "Invalid TIME_EVENT_NOTIFICATION response\n"); | |
367 | return true; | |
368 | } | |
369 | ||
370 | resp = (void *)pkt->data; | |
371 | ||
372 | /* te_data->uid is already set in the TIME_EVENT_CMD response */ | |
373 | if (le32_to_cpu(resp->unique_id) != te_data->uid) | |
374 | return false; | |
375 | ||
376 | IWL_DEBUG_TE(mvm, "TIME_EVENT_NOTIFICATION response - UID = 0x%x\n", | |
377 | te_data->uid); | |
378 | if (!resp->status) | |
379 | IWL_ERR(mvm, | |
380 | "TIME_EVENT_NOTIFICATION received but not executed\n"); | |
381 | ||
382 | return true; | |
383 | } | |
384 | ||
ffdf968d JB |
385 | static bool iwl_mvm_time_event_response(struct iwl_notif_wait_data *notif_wait, |
386 | struct iwl_rx_packet *pkt, void *data) | |
8ca151b5 JB |
387 | { |
388 | struct iwl_mvm *mvm = | |
389 | container_of(notif_wait, struct iwl_mvm, notif_wait); | |
390 | struct iwl_mvm_time_event_data *te_data = data; | |
8ca151b5 | 391 | struct iwl_time_event_resp *resp; |
65b30348 | 392 | int resp_len = iwl_rx_packet_payload_len(pkt); |
8ca151b5 | 393 | |
ffdf968d JB |
394 | if (WARN_ON(pkt->hdr.cmd != TIME_EVENT_CMD)) |
395 | return true; | |
8ca151b5 | 396 | |
65b30348 | 397 | if (WARN_ON_ONCE(resp_len != sizeof(*resp))) { |
ffdf968d JB |
398 | IWL_ERR(mvm, "Invalid TIME_EVENT_CMD response\n"); |
399 | return true; | |
400 | } | |
8ca151b5 | 401 | |
ffdf968d | 402 | resp = (void *)pkt->data; |
e3722822 JB |
403 | |
404 | /* we should never get a response to another TIME_EVENT_CMD here */ | |
405 | if (WARN_ON_ONCE(le32_to_cpu(resp->id) != te_data->id)) | |
406 | return false; | |
407 | ||
ffdf968d JB |
408 | te_data->uid = le32_to_cpu(resp->unique_id); |
409 | IWL_DEBUG_TE(mvm, "TIME_EVENT_CMD response - UID = 0x%x\n", | |
410 | te_data->uid); | |
411 | return true; | |
412 | } | |
8ca151b5 | 413 | |
ffdf968d JB |
414 | static int iwl_mvm_time_event_send_add(struct iwl_mvm *mvm, |
415 | struct ieee80211_vif *vif, | |
416 | struct iwl_mvm_time_event_data *te_data, | |
a373f67c | 417 | struct iwl_time_event_cmd *te_cmd) |
ffdf968d JB |
418 | { |
419 | static const u8 time_event_response[] = { TIME_EVENT_CMD }; | |
420 | struct iwl_notification_wait wait_time_event; | |
421 | int ret; | |
422 | ||
423 | lockdep_assert_held(&mvm->mutex); | |
424 | ||
93630dc3 JB |
425 | IWL_DEBUG_TE(mvm, "Add new TE, duration %d TU\n", |
426 | le32_to_cpu(te_cmd->duration)); | |
427 | ||
ffdf968d JB |
428 | spin_lock_bh(&mvm->time_event_lock); |
429 | if (WARN_ON(te_data->id != TE_MAX)) { | |
430 | spin_unlock_bh(&mvm->time_event_lock); | |
431 | return -EIO; | |
432 | } | |
433 | te_data->vif = vif; | |
434 | te_data->duration = le32_to_cpu(te_cmd->duration); | |
435 | te_data->id = le32_to_cpu(te_cmd->id); | |
436 | list_add_tail(&te_data->list, &mvm->time_event_list); | |
437 | spin_unlock_bh(&mvm->time_event_lock); | |
438 | ||
439 | /* | |
440 | * Use a notification wait, which really just processes the | |
441 | * command response and doesn't wait for anything, in order | |
442 | * to be able to process the response and get the UID inside | |
443 | * the RX path. Using CMD_WANT_SKB doesn't work because it | |
444 | * stores the buffer and then wakes up this thread, by which | |
445 | * time another notification (that the time event started) | |
446 | * might already be processed unsuccessfully. | |
447 | */ | |
448 | iwl_init_notification_wait(&mvm->notif_wait, &wait_time_event, | |
449 | time_event_response, | |
450 | ARRAY_SIZE(time_event_response), | |
451 | iwl_mvm_time_event_response, te_data); | |
452 | ||
a1022927 | 453 | ret = iwl_mvm_send_cmd_pdu(mvm, TIME_EVENT_CMD, 0, |
a373f67c | 454 | sizeof(*te_cmd), te_cmd); |
ffdf968d JB |
455 | if (ret) { |
456 | IWL_ERR(mvm, "Couldn't send TIME_EVENT_CMD: %d\n", ret); | |
457 | iwl_remove_notification(&mvm->notif_wait, &wait_time_event); | |
458 | goto out_clear_te; | |
459 | } | |
8ca151b5 | 460 | |
ffdf968d JB |
461 | /* No need to wait for anything, so just pass 1 (0 isn't valid) */ |
462 | ret = iwl_wait_notification(&mvm->notif_wait, &wait_time_event, 1); | |
463 | /* should never fail */ | |
464 | WARN_ON_ONCE(ret); | |
465 | ||
466 | if (ret) { | |
467 | out_clear_te: | |
468 | spin_lock_bh(&mvm->time_event_lock); | |
469 | iwl_mvm_te_clear_data(mvm, te_data); | |
470 | spin_unlock_bh(&mvm->time_event_lock); | |
471 | } | |
472 | return ret; | |
8ca151b5 JB |
473 | } |
474 | ||
475 | void iwl_mvm_protect_session(struct iwl_mvm *mvm, | |
476 | struct ieee80211_vif *vif, | |
016d27e1 | 477 | u32 duration, u32 min_duration, |
d20d37bc | 478 | u32 max_delay, bool wait_for_notif) |
8ca151b5 JB |
479 | { |
480 | struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); | |
481 | struct iwl_mvm_time_event_data *te_data = &mvmvif->time_event_data; | |
d20d37bc LK |
482 | const u8 te_notif_response[] = { TIME_EVENT_NOTIFICATION }; |
483 | struct iwl_notification_wait wait_te_notif; | |
a373f67c | 484 | struct iwl_time_event_cmd time_cmd = {}; |
8ca151b5 JB |
485 | |
486 | lockdep_assert_held(&mvm->mutex); | |
487 | ||
488 | if (te_data->running && | |
e7f1935c | 489 | time_after(te_data->end_jiffies, TU_TO_EXP_TIME(min_duration))) { |
8ca151b5 JB |
490 | IWL_DEBUG_TE(mvm, "We have enough time in the current TE: %u\n", |
491 | jiffies_to_msecs(te_data->end_jiffies - jiffies)); | |
492 | return; | |
493 | } | |
494 | ||
495 | if (te_data->running) { | |
496 | IWL_DEBUG_TE(mvm, "extend 0x%x: only %u ms left\n", | |
497 | te_data->uid, | |
498 | jiffies_to_msecs(te_data->end_jiffies - jiffies)); | |
499 | /* | |
500 | * we don't have enough time | |
501 | * cancel the current TE and issue a new one | |
502 | * Of course it would be better to remove the old one only | |
503 | * when the new one is added, but we don't care if we are off | |
504 | * channel for a bit. All we need to do, is not to return | |
505 | * before we actually begin to be on the channel. | |
506 | */ | |
507 | iwl_mvm_stop_session_protection(mvm, vif); | |
508 | } | |
509 | ||
8ca151b5 JB |
510 | time_cmd.action = cpu_to_le32(FW_CTXT_ACTION_ADD); |
511 | time_cmd.id_and_color = | |
512 | cpu_to_le32(FW_CMD_ID_AND_COLOR(mvmvif->id, mvmvif->color)); | |
513 | time_cmd.id = cpu_to_le32(TE_BSS_STA_AGGRESSIVE_ASSOC); | |
514 | ||
515 | time_cmd.apply_time = | |
516 | cpu_to_le32(iwl_read_prph(mvm->trans, DEVICE_SYSTEM_TIME_REG)); | |
ffdf968d | 517 | |
f8f03c3e | 518 | time_cmd.max_frags = TE_V2_FRAG_NONE; |
016d27e1 | 519 | time_cmd.max_delay = cpu_to_le32(max_delay); |
8ca151b5 JB |
520 | /* TODO: why do we need to interval = bi if it is not periodic? */ |
521 | time_cmd.interval = cpu_to_le32(1); | |
8ca151b5 | 522 | time_cmd.duration = cpu_to_le32(duration); |
f8f03c3e EL |
523 | time_cmd.repeat = 1; |
524 | time_cmd.policy = cpu_to_le16(TE_V2_NOTIF_HOST_EVENT_START | | |
1f6bf078 EG |
525 | TE_V2_NOTIF_HOST_EVENT_END | |
526 | T2_V2_START_IMMEDIATELY); | |
8ca151b5 | 527 | |
d20d37bc LK |
528 | if (!wait_for_notif) { |
529 | iwl_mvm_time_event_send_add(mvm, vif, te_data, &time_cmd); | |
530 | return; | |
531 | } | |
532 | ||
533 | /* | |
534 | * Create notification_wait for the TIME_EVENT_NOTIFICATION to use | |
535 | * right after we send the time event | |
536 | */ | |
537 | iwl_init_notification_wait(&mvm->notif_wait, &wait_te_notif, | |
538 | te_notif_response, | |
539 | ARRAY_SIZE(te_notif_response), | |
540 | iwl_mvm_te_notif, te_data); | |
541 | ||
542 | /* If TE was sent OK - wait for the notification that started */ | |
543 | if (iwl_mvm_time_event_send_add(mvm, vif, te_data, &time_cmd)) { | |
544 | IWL_ERR(mvm, "Failed to add TE to protect session\n"); | |
545 | iwl_remove_notification(&mvm->notif_wait, &wait_te_notif); | |
546 | } else if (iwl_wait_notification(&mvm->notif_wait, &wait_te_notif, | |
547 | TU_TO_JIFFIES(max_delay))) { | |
548 | IWL_ERR(mvm, "Failed to protect session until TE\n"); | |
549 | } | |
8ca151b5 JB |
550 | } |
551 | ||
552 | /* | |
553 | * Explicit request to remove a time event. The removal of a time event needs to | |
554 | * be synchronized with the flow of a time event's end notification, which also | |
555 | * removes the time event from the op mode data structures. | |
556 | */ | |
557 | void iwl_mvm_remove_time_event(struct iwl_mvm *mvm, | |
558 | struct iwl_mvm_vif *mvmvif, | |
559 | struct iwl_mvm_time_event_data *te_data) | |
560 | { | |
a373f67c | 561 | struct iwl_time_event_cmd time_cmd = {}; |
8ca151b5 JB |
562 | u32 id, uid; |
563 | int ret; | |
564 | ||
565 | /* | |
566 | * It is possible that by the time we got to this point the time | |
567 | * event was already removed. | |
568 | */ | |
569 | spin_lock_bh(&mvm->time_event_lock); | |
570 | ||
571 | /* Save time event uid before clearing its data */ | |
572 | uid = te_data->uid; | |
573 | id = te_data->id; | |
574 | ||
575 | /* | |
576 | * The clear_data function handles time events that were already removed | |
577 | */ | |
578 | iwl_mvm_te_clear_data(mvm, te_data); | |
579 | spin_unlock_bh(&mvm->time_event_lock); | |
580 | ||
581 | /* | |
582 | * It is possible that by the time we try to remove it, the time event | |
583 | * has already ended and removed. In such a case there is no need to | |
584 | * send a removal command. | |
585 | */ | |
586 | if (id == TE_MAX) { | |
587 | IWL_DEBUG_TE(mvm, "TE 0x%x has already ended\n", uid); | |
588 | return; | |
589 | } | |
590 | ||
591 | /* When we remove a TE, the UID is to be set in the id field */ | |
592 | time_cmd.id = cpu_to_le32(uid); | |
593 | time_cmd.action = cpu_to_le32(FW_CTXT_ACTION_REMOVE); | |
594 | time_cmd.id_and_color = | |
595 | cpu_to_le32(FW_CMD_ID_AND_COLOR(mvmvif->id, mvmvif->color)); | |
596 | ||
597 | IWL_DEBUG_TE(mvm, "Removing TE 0x%x\n", le32_to_cpu(time_cmd.id)); | |
a1022927 | 598 | ret = iwl_mvm_send_cmd_pdu(mvm, TIME_EVENT_CMD, 0, |
a373f67c | 599 | sizeof(time_cmd), &time_cmd); |
8ca151b5 JB |
600 | if (WARN_ON(ret)) |
601 | return; | |
602 | } | |
603 | ||
604 | void iwl_mvm_stop_session_protection(struct iwl_mvm *mvm, | |
605 | struct ieee80211_vif *vif) | |
606 | { | |
607 | struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); | |
608 | struct iwl_mvm_time_event_data *te_data = &mvmvif->time_event_data; | |
609 | ||
610 | lockdep_assert_held(&mvm->mutex); | |
611 | iwl_mvm_remove_time_event(mvm, mvmvif, te_data); | |
612 | } | |
613 | ||
8ca151b5 | 614 | int iwl_mvm_start_p2p_roc(struct iwl_mvm *mvm, struct ieee80211_vif *vif, |
e635c797 | 615 | int duration, enum ieee80211_roc_type type) |
8ca151b5 JB |
616 | { |
617 | struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); | |
618 | struct iwl_mvm_time_event_data *te_data = &mvmvif->time_event_data; | |
a373f67c | 619 | struct iwl_time_event_cmd time_cmd = {}; |
8ca151b5 JB |
620 | |
621 | lockdep_assert_held(&mvm->mutex); | |
622 | if (te_data->running) { | |
623 | IWL_WARN(mvm, "P2P_DEVICE remain on channel already running\n"); | |
624 | return -EBUSY; | |
625 | } | |
626 | ||
627 | /* | |
628 | * Flush the done work, just in case it's still pending, so that | |
629 | * the work it does can complete and we can accept new frames. | |
630 | */ | |
631 | flush_work(&mvm->roc_done_wk); | |
632 | ||
8ca151b5 JB |
633 | time_cmd.action = cpu_to_le32(FW_CTXT_ACTION_ADD); |
634 | time_cmd.id_and_color = | |
635 | cpu_to_le32(FW_CMD_ID_AND_COLOR(mvmvif->id, mvmvif->color)); | |
e635c797 IP |
636 | |
637 | switch (type) { | |
638 | case IEEE80211_ROC_TYPE_NORMAL: | |
639 | time_cmd.id = cpu_to_le32(IWL_MVM_ROC_TE_TYPE_NORMAL); | |
640 | break; | |
641 | case IEEE80211_ROC_TYPE_MGMT_TX: | |
642 | time_cmd.id = cpu_to_le32(IWL_MVM_ROC_TE_TYPE_MGMT_TX); | |
643 | break; | |
644 | default: | |
645 | WARN_ONCE(1, "Got an invalid ROC type\n"); | |
646 | return -EINVAL; | |
647 | } | |
8ca151b5 JB |
648 | |
649 | time_cmd.apply_time = cpu_to_le32(0); | |
8ca151b5 JB |
650 | time_cmd.interval = cpu_to_le32(1); |
651 | ||
652 | /* | |
e635c797 | 653 | * The P2P Device TEs can have lower priority than other events |
8ca151b5 | 654 | * that are being scheduled by the driver/fw, and thus it might not be |
e635c797 IP |
655 | * scheduled. To improve the chances of it being scheduled, allow them |
656 | * to be fragmented, and in addition allow them to be delayed. | |
8ca151b5 | 657 | */ |
f8f03c3e | 658 | time_cmd.max_frags = min(MSEC_TO_TU(duration)/50, TE_V2_FRAG_ENDLESS); |
8ca151b5 JB |
659 | time_cmd.max_delay = cpu_to_le32(MSEC_TO_TU(duration/2)); |
660 | time_cmd.duration = cpu_to_le32(MSEC_TO_TU(duration)); | |
f8f03c3e EL |
661 | time_cmd.repeat = 1; |
662 | time_cmd.policy = cpu_to_le16(TE_V2_NOTIF_HOST_EVENT_START | | |
1f6bf078 EG |
663 | TE_V2_NOTIF_HOST_EVENT_END | |
664 | T2_V2_START_IMMEDIATELY); | |
8ca151b5 | 665 | |
ffdf968d | 666 | return iwl_mvm_time_event_send_add(mvm, vif, te_data, &time_cmd); |
8ca151b5 JB |
667 | } |
668 | ||
669 | void iwl_mvm_stop_p2p_roc(struct iwl_mvm *mvm) | |
670 | { | |
671 | struct iwl_mvm_vif *mvmvif; | |
672 | struct iwl_mvm_time_event_data *te_data; | |
673 | ||
674 | lockdep_assert_held(&mvm->mutex); | |
675 | ||
676 | /* | |
677 | * Iterate over the list of time events and find the time event that is | |
678 | * associated with a P2P_DEVICE interface. | |
679 | * This assumes that a P2P_DEVICE interface can have only a single time | |
680 | * event at any given time and this time event coresponds to a ROC | |
681 | * request | |
682 | */ | |
683 | mvmvif = NULL; | |
684 | spin_lock_bh(&mvm->time_event_lock); | |
685 | list_for_each_entry(te_data, &mvm->time_event_list, list) { | |
686 | if (te_data->vif->type == NL80211_IFTYPE_P2P_DEVICE) { | |
687 | mvmvif = iwl_mvm_vif_from_mac80211(te_data->vif); | |
688 | break; | |
689 | } | |
690 | } | |
691 | spin_unlock_bh(&mvm->time_event_lock); | |
692 | ||
693 | if (!mvmvif) { | |
694 | IWL_WARN(mvm, "P2P_DEVICE no remain on channel event\n"); | |
695 | return; | |
696 | } | |
697 | ||
698 | iwl_mvm_remove_time_event(mvm, mvmvif, te_data); | |
699 | ||
700 | iwl_mvm_roc_finished(mvm); | |
701 | } | |
7f0a7c67 | 702 | |
f991e17b LC |
703 | int iwl_mvm_schedule_csa_period(struct iwl_mvm *mvm, |
704 | struct ieee80211_vif *vif, | |
705 | u32 duration, u32 apply_time) | |
7f0a7c67 AO |
706 | { |
707 | struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); | |
708 | struct iwl_mvm_time_event_data *te_data = &mvmvif->time_event_data; | |
709 | struct iwl_time_event_cmd time_cmd = {}; | |
710 | ||
711 | lockdep_assert_held(&mvm->mutex); | |
712 | ||
713 | if (te_data->running) { | |
f991e17b | 714 | IWL_DEBUG_TE(mvm, "CS period is already scheduled\n"); |
7f0a7c67 AO |
715 | return -EBUSY; |
716 | } | |
717 | ||
718 | time_cmd.action = cpu_to_le32(FW_CTXT_ACTION_ADD); | |
719 | time_cmd.id_and_color = | |
720 | cpu_to_le32(FW_CMD_ID_AND_COLOR(mvmvif->id, mvmvif->color)); | |
f991e17b | 721 | time_cmd.id = cpu_to_le32(TE_CHANNEL_SWITCH_PERIOD); |
7f0a7c67 AO |
722 | time_cmd.apply_time = cpu_to_le32(apply_time); |
723 | time_cmd.max_frags = TE_V2_FRAG_NONE; | |
724 | time_cmd.duration = cpu_to_le32(duration); | |
725 | time_cmd.repeat = 1; | |
726 | time_cmd.interval = cpu_to_le32(1); | |
727 | time_cmd.policy = cpu_to_le16(TE_V2_NOTIF_HOST_EVENT_START | | |
728 | TE_V2_ABSENCE); | |
729 | ||
730 | return iwl_mvm_time_event_send_add(mvm, vif, te_data, &time_cmd); | |
731 | } |