]>
Commit | Line | Data |
---|---|---|
8ca151b5 JB |
1 | /****************************************************************************** |
2 | * | |
3 | * This file is provided under a dual BSD/GPLv2 license. When using or | |
4 | * redistributing this file, you may do so under either license. | |
5 | * | |
6 | * GPL LICENSE SUMMARY | |
7 | * | |
8 | * Copyright(c) 2012 - 2013 Intel Corporation. All rights reserved. | |
9 | * | |
10 | * This program is free software; you can redistribute it and/or modify | |
11 | * it under the terms of version 2 of the GNU General Public License as | |
12 | * published by the Free Software Foundation. | |
13 | * | |
14 | * This program is distributed in the hope that it will be useful, but | |
15 | * WITHOUT ANY WARRANTY; without even the implied warranty of | |
16 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU | |
17 | * General Public License for more details. | |
18 | * | |
19 | * You should have received a copy of the GNU General Public License | |
20 | * along with this program; if not, write to the Free Software | |
21 | * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110, | |
22 | * USA | |
23 | * | |
24 | * The full GNU General Public License is included in this distribution | |
25 | * in the file called LICENSE.GPL. | |
26 | * | |
27 | * Contact Information: | |
28 | * Intel Linux Wireless <ilw@linux.intel.com> | |
29 | * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497 | |
30 | * | |
31 | * BSD LICENSE | |
32 | * | |
33 | * Copyright(c) 2012 - 2013 Intel Corporation. All rights reserved. | |
34 | * All rights reserved. | |
35 | * | |
36 | * Redistribution and use in source and binary forms, with or without | |
37 | * modification, are permitted provided that the following conditions | |
38 | * are met: | |
39 | * | |
40 | * * Redistributions of source code must retain the above copyright | |
41 | * notice, this list of conditions and the following disclaimer. | |
42 | * * Redistributions in binary form must reproduce the above copyright | |
43 | * notice, this list of conditions and the following disclaimer in | |
44 | * the documentation and/or other materials provided with the | |
45 | * distribution. | |
46 | * * Neither the name Intel Corporation nor the names of its | |
47 | * contributors may be used to endorse or promote products derived | |
48 | * from this software without specific prior written permission. | |
49 | * | |
50 | * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS | |
51 | * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT | |
52 | * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR | |
53 | * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT | |
54 | * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, | |
55 | * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT | |
56 | * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, | |
57 | * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY | |
58 | * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT | |
59 | * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE | |
60 | * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | |
61 | * | |
62 | *****************************************************************************/ | |
63 | ||
64 | #include <linux/jiffies.h> | |
65 | #include <net/mac80211.h> | |
66 | ||
67 | #include "iwl-notif-wait.h" | |
68 | #include "iwl-trans.h" | |
69 | #include "fw-api.h" | |
70 | #include "time-event.h" | |
71 | #include "mvm.h" | |
72 | #include "iwl-io.h" | |
73 | #include "iwl-prph.h" | |
74 | ||
75 | /* A TimeUnit is 1024 microsecond */ | |
76 | #define TU_TO_JIFFIES(_tu) (usecs_to_jiffies((_tu) * 1024)) | |
77 | #define MSEC_TO_TU(_msec) (_msec*1000/1024) | |
78 | ||
456f6ddf IP |
79 | /* For ROC use a TE type which has priority high enough to be scheduled when |
80 | * there is a concurrent BSS or GO/AP. Currently, use a TE type that has | |
81 | * priority similar to the TE priority used for action scans by the FW. | |
82 | * TODO: This needs to be changed, based on the reason for the ROC, i.e., use | |
83 | * TE_P2P_DEVICE_DISCOVERABLE for remain on channel without mgmt skb, and use | |
84 | * TE_P2P_DEVICE_ACTION_SCAN | |
85 | */ | |
86 | #define IWL_MVM_ROC_TE_TYPE TE_P2P_DEVICE_ACTION_SCAN | |
87 | ||
8ca151b5 JB |
88 | void iwl_mvm_te_clear_data(struct iwl_mvm *mvm, |
89 | struct iwl_mvm_time_event_data *te_data) | |
90 | { | |
91 | lockdep_assert_held(&mvm->time_event_lock); | |
92 | ||
93 | if (te_data->id == TE_MAX) | |
94 | return; | |
95 | ||
96 | list_del(&te_data->list); | |
97 | te_data->running = false; | |
98 | te_data->uid = 0; | |
99 | te_data->id = TE_MAX; | |
100 | te_data->vif = NULL; | |
101 | } | |
102 | ||
103 | void iwl_mvm_roc_done_wk(struct work_struct *wk) | |
104 | { | |
105 | struct iwl_mvm *mvm = container_of(wk, struct iwl_mvm, roc_done_wk); | |
106 | ||
107 | synchronize_net(); | |
108 | ||
109 | /* | |
110 | * Flush the offchannel queue -- this is called when the time | |
111 | * event finishes or is cancelled, so that frames queued for it | |
112 | * won't get stuck on the queue and be transmitted in the next | |
113 | * time event. | |
114 | * We have to send the command asynchronously since this cannot | |
115 | * be under the mutex for locking reasons, but that's not an | |
116 | * issue as it will have to complete before the next command is | |
117 | * executed, and a new time event means a new command. | |
118 | */ | |
119 | iwl_mvm_flush_tx_path(mvm, BIT(IWL_OFFCHANNEL_QUEUE), false); | |
120 | } | |
121 | ||
122 | static void iwl_mvm_roc_finished(struct iwl_mvm *mvm) | |
123 | { | |
124 | /* | |
125 | * First, clear the ROC_RUNNING status bit. This will cause the TX | |
126 | * path to drop offchannel transmissions. That would also be done | |
127 | * by mac80211, but it is racy, in particular in the case that the | |
128 | * time event actually completed in the firmware (which is handled | |
129 | * in iwl_mvm_te_handle_notif). | |
130 | */ | |
131 | clear_bit(IWL_MVM_STATUS_ROC_RUNNING, &mvm->status); | |
132 | ||
133 | /* | |
134 | * Of course, our status bit is just as racy as mac80211, so in | |
135 | * addition, fire off the work struct which will drop all frames | |
136 | * from the hardware queues that made it through the race. First | |
137 | * it will of course synchronize the TX path to make sure that | |
138 | * any *new* TX will be rejected. | |
139 | */ | |
140 | schedule_work(&mvm->roc_done_wk); | |
141 | } | |
142 | ||
143 | /* | |
144 | * Handles a FW notification for an event that is known to the driver. | |
145 | * | |
146 | * @mvm: the mvm component | |
147 | * @te_data: the time event data | |
148 | * @notif: the notification data corresponding the time event data. | |
149 | */ | |
150 | static void iwl_mvm_te_handle_notif(struct iwl_mvm *mvm, | |
151 | struct iwl_mvm_time_event_data *te_data, | |
152 | struct iwl_time_event_notif *notif) | |
153 | { | |
154 | lockdep_assert_held(&mvm->time_event_lock); | |
155 | ||
156 | IWL_DEBUG_TE(mvm, "Handle time event notif - UID = 0x%x action %d\n", | |
157 | le32_to_cpu(notif->unique_id), | |
158 | le32_to_cpu(notif->action)); | |
159 | ||
160 | /* | |
161 | * The FW sends the start/end time event notifications even for events | |
162 | * that it fails to schedule. This is indicated in the status field of | |
163 | * the notification. This happens in cases that the scheduler cannot | |
164 | * find a schedule that can handle the event (for example requesting a | |
165 | * P2P Device discoveribility, while there are other higher priority | |
166 | * events in the system). | |
167 | */ | |
168 | WARN_ONCE(!le32_to_cpu(notif->status), | |
169 | "Failed to schedule time event\n"); | |
170 | ||
171 | if (le32_to_cpu(notif->action) == TE_NOTIF_HOST_END) { | |
172 | IWL_DEBUG_TE(mvm, | |
173 | "TE ended - current time %lu, estimated end %lu\n", | |
174 | jiffies, te_data->end_jiffies); | |
175 | ||
176 | if (te_data->vif->type == NL80211_IFTYPE_P2P_DEVICE) { | |
177 | ieee80211_remain_on_channel_expired(mvm->hw); | |
178 | iwl_mvm_roc_finished(mvm); | |
179 | } | |
180 | ||
181 | /* | |
182 | * By now, we should have finished association | |
183 | * and know the dtim period. | |
184 | */ | |
185 | if (te_data->vif->type == NL80211_IFTYPE_STATION && | |
186 | (!te_data->vif->bss_conf.assoc || | |
210a544e | 187 | !te_data->vif->bss_conf.dtim_period)) { |
8ca151b5 JB |
188 | IWL_ERR(mvm, |
189 | "No assocation and the time event is over already...\n"); | |
210a544e JB |
190 | ieee80211_connection_loss(te_data->vif); |
191 | } | |
8ca151b5 JB |
192 | |
193 | iwl_mvm_te_clear_data(mvm, te_data); | |
194 | } else if (le32_to_cpu(notif->action) == TE_NOTIF_HOST_START) { | |
195 | te_data->running = true; | |
196 | te_data->end_jiffies = jiffies + | |
197 | TU_TO_JIFFIES(te_data->duration); | |
198 | ||
199 | if (te_data->vif->type == NL80211_IFTYPE_P2P_DEVICE) { | |
200 | set_bit(IWL_MVM_STATUS_ROC_RUNNING, &mvm->status); | |
201 | ieee80211_ready_on_channel(mvm->hw); | |
202 | } | |
203 | } else { | |
204 | IWL_WARN(mvm, "Got TE with unknown action\n"); | |
205 | } | |
206 | } | |
207 | ||
208 | /* | |
209 | * The Rx handler for time event notifications | |
210 | */ | |
211 | int iwl_mvm_rx_time_event_notif(struct iwl_mvm *mvm, | |
212 | struct iwl_rx_cmd_buffer *rxb, | |
213 | struct iwl_device_cmd *cmd) | |
214 | { | |
215 | struct iwl_rx_packet *pkt = rxb_addr(rxb); | |
216 | struct iwl_time_event_notif *notif = (void *)pkt->data; | |
217 | struct iwl_mvm_time_event_data *te_data, *tmp; | |
218 | ||
219 | IWL_DEBUG_TE(mvm, "Time event notification - UID = 0x%x action %d\n", | |
220 | le32_to_cpu(notif->unique_id), | |
221 | le32_to_cpu(notif->action)); | |
222 | ||
223 | spin_lock_bh(&mvm->time_event_lock); | |
224 | list_for_each_entry_safe(te_data, tmp, &mvm->time_event_list, list) { | |
225 | if (le32_to_cpu(notif->unique_id) == te_data->uid) | |
226 | iwl_mvm_te_handle_notif(mvm, te_data, notif); | |
227 | } | |
228 | spin_unlock_bh(&mvm->time_event_lock); | |
229 | ||
230 | return 0; | |
231 | } | |
232 | ||
ffdf968d JB |
233 | static bool iwl_mvm_time_event_response(struct iwl_notif_wait_data *notif_wait, |
234 | struct iwl_rx_packet *pkt, void *data) | |
8ca151b5 JB |
235 | { |
236 | struct iwl_mvm *mvm = | |
237 | container_of(notif_wait, struct iwl_mvm, notif_wait); | |
238 | struct iwl_mvm_time_event_data *te_data = data; | |
8ca151b5 | 239 | struct iwl_time_event_resp *resp; |
ffdf968d | 240 | int resp_len = le32_to_cpu(pkt->len_n_flags) & FH_RSCSR_FRAME_SIZE_MSK; |
8ca151b5 | 241 | |
ffdf968d JB |
242 | if (WARN_ON(pkt->hdr.cmd != TIME_EVENT_CMD)) |
243 | return true; | |
8ca151b5 | 244 | |
ffdf968d JB |
245 | if (WARN_ON_ONCE(resp_len != sizeof(pkt->hdr) + sizeof(*resp))) { |
246 | IWL_ERR(mvm, "Invalid TIME_EVENT_CMD response\n"); | |
247 | return true; | |
248 | } | |
8ca151b5 | 249 | |
ffdf968d JB |
250 | resp = (void *)pkt->data; |
251 | te_data->uid = le32_to_cpu(resp->unique_id); | |
252 | IWL_DEBUG_TE(mvm, "TIME_EVENT_CMD response - UID = 0x%x\n", | |
253 | te_data->uid); | |
254 | return true; | |
255 | } | |
8ca151b5 | 256 | |
ffdf968d JB |
257 | static int iwl_mvm_time_event_send_add(struct iwl_mvm *mvm, |
258 | struct ieee80211_vif *vif, | |
259 | struct iwl_mvm_time_event_data *te_data, | |
260 | struct iwl_time_event_cmd *te_cmd) | |
261 | { | |
262 | static const u8 time_event_response[] = { TIME_EVENT_CMD }; | |
263 | struct iwl_notification_wait wait_time_event; | |
264 | int ret; | |
265 | ||
266 | lockdep_assert_held(&mvm->mutex); | |
267 | ||
93630dc3 JB |
268 | IWL_DEBUG_TE(mvm, "Add new TE, duration %d TU\n", |
269 | le32_to_cpu(te_cmd->duration)); | |
270 | ||
ffdf968d JB |
271 | spin_lock_bh(&mvm->time_event_lock); |
272 | if (WARN_ON(te_data->id != TE_MAX)) { | |
273 | spin_unlock_bh(&mvm->time_event_lock); | |
274 | return -EIO; | |
275 | } | |
276 | te_data->vif = vif; | |
277 | te_data->duration = le32_to_cpu(te_cmd->duration); | |
278 | te_data->id = le32_to_cpu(te_cmd->id); | |
279 | list_add_tail(&te_data->list, &mvm->time_event_list); | |
280 | spin_unlock_bh(&mvm->time_event_lock); | |
281 | ||
282 | /* | |
283 | * Use a notification wait, which really just processes the | |
284 | * command response and doesn't wait for anything, in order | |
285 | * to be able to process the response and get the UID inside | |
286 | * the RX path. Using CMD_WANT_SKB doesn't work because it | |
287 | * stores the buffer and then wakes up this thread, by which | |
288 | * time another notification (that the time event started) | |
289 | * might already be processed unsuccessfully. | |
290 | */ | |
291 | iwl_init_notification_wait(&mvm->notif_wait, &wait_time_event, | |
292 | time_event_response, | |
293 | ARRAY_SIZE(time_event_response), | |
294 | iwl_mvm_time_event_response, te_data); | |
295 | ||
296 | ret = iwl_mvm_send_cmd_pdu(mvm, TIME_EVENT_CMD, CMD_SYNC, | |
297 | sizeof(*te_cmd), te_cmd); | |
298 | if (ret) { | |
299 | IWL_ERR(mvm, "Couldn't send TIME_EVENT_CMD: %d\n", ret); | |
300 | iwl_remove_notification(&mvm->notif_wait, &wait_time_event); | |
301 | goto out_clear_te; | |
302 | } | |
8ca151b5 | 303 | |
ffdf968d JB |
304 | /* No need to wait for anything, so just pass 1 (0 isn't valid) */ |
305 | ret = iwl_wait_notification(&mvm->notif_wait, &wait_time_event, 1); | |
306 | /* should never fail */ | |
307 | WARN_ON_ONCE(ret); | |
308 | ||
309 | if (ret) { | |
310 | out_clear_te: | |
311 | spin_lock_bh(&mvm->time_event_lock); | |
312 | iwl_mvm_te_clear_data(mvm, te_data); | |
313 | spin_unlock_bh(&mvm->time_event_lock); | |
314 | } | |
315 | return ret; | |
8ca151b5 JB |
316 | } |
317 | ||
318 | void iwl_mvm_protect_session(struct iwl_mvm *mvm, | |
319 | struct ieee80211_vif *vif, | |
320 | u32 duration, u32 min_duration) | |
321 | { | |
322 | struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); | |
323 | struct iwl_mvm_time_event_data *te_data = &mvmvif->time_event_data; | |
8ca151b5 | 324 | struct iwl_time_event_cmd time_cmd = {}; |
8ca151b5 JB |
325 | |
326 | lockdep_assert_held(&mvm->mutex); | |
327 | ||
328 | if (te_data->running && | |
329 | time_after(te_data->end_jiffies, | |
330 | jiffies + TU_TO_JIFFIES(min_duration))) { | |
331 | IWL_DEBUG_TE(mvm, "We have enough time in the current TE: %u\n", | |
332 | jiffies_to_msecs(te_data->end_jiffies - jiffies)); | |
333 | return; | |
334 | } | |
335 | ||
336 | if (te_data->running) { | |
337 | IWL_DEBUG_TE(mvm, "extend 0x%x: only %u ms left\n", | |
338 | te_data->uid, | |
339 | jiffies_to_msecs(te_data->end_jiffies - jiffies)); | |
340 | /* | |
341 | * we don't have enough time | |
342 | * cancel the current TE and issue a new one | |
343 | * Of course it would be better to remove the old one only | |
344 | * when the new one is added, but we don't care if we are off | |
345 | * channel for a bit. All we need to do, is not to return | |
346 | * before we actually begin to be on the channel. | |
347 | */ | |
348 | iwl_mvm_stop_session_protection(mvm, vif); | |
349 | } | |
350 | ||
8ca151b5 JB |
351 | time_cmd.action = cpu_to_le32(FW_CTXT_ACTION_ADD); |
352 | time_cmd.id_and_color = | |
353 | cpu_to_le32(FW_CMD_ID_AND_COLOR(mvmvif->id, mvmvif->color)); | |
354 | time_cmd.id = cpu_to_le32(TE_BSS_STA_AGGRESSIVE_ASSOC); | |
355 | ||
356 | time_cmd.apply_time = | |
357 | cpu_to_le32(iwl_read_prph(mvm->trans, DEVICE_SYSTEM_TIME_REG)); | |
ffdf968d | 358 | |
8ca151b5 JB |
359 | time_cmd.dep_policy = TE_INDEPENDENT; |
360 | time_cmd.is_present = cpu_to_le32(1); | |
361 | time_cmd.max_frags = cpu_to_le32(TE_FRAG_NONE); | |
362 | time_cmd.max_delay = cpu_to_le32(500); | |
363 | /* TODO: why do we need to interval = bi if it is not periodic? */ | |
364 | time_cmd.interval = cpu_to_le32(1); | |
365 | time_cmd.interval_reciprocal = cpu_to_le32(iwl_mvm_reciprocal(1)); | |
366 | time_cmd.duration = cpu_to_le32(duration); | |
367 | time_cmd.repeat = cpu_to_le32(1); | |
368 | time_cmd.notify = cpu_to_le32(TE_NOTIF_HOST_START | TE_NOTIF_HOST_END); | |
369 | ||
ffdf968d | 370 | iwl_mvm_time_event_send_add(mvm, vif, te_data, &time_cmd); |
8ca151b5 JB |
371 | } |
372 | ||
373 | /* | |
374 | * Explicit request to remove a time event. The removal of a time event needs to | |
375 | * be synchronized with the flow of a time event's end notification, which also | |
376 | * removes the time event from the op mode data structures. | |
377 | */ | |
378 | void iwl_mvm_remove_time_event(struct iwl_mvm *mvm, | |
379 | struct iwl_mvm_vif *mvmvif, | |
380 | struct iwl_mvm_time_event_data *te_data) | |
381 | { | |
382 | struct iwl_time_event_cmd time_cmd = {}; | |
383 | u32 id, uid; | |
384 | int ret; | |
385 | ||
386 | /* | |
387 | * It is possible that by the time we got to this point the time | |
388 | * event was already removed. | |
389 | */ | |
390 | spin_lock_bh(&mvm->time_event_lock); | |
391 | ||
392 | /* Save time event uid before clearing its data */ | |
393 | uid = te_data->uid; | |
394 | id = te_data->id; | |
395 | ||
396 | /* | |
397 | * The clear_data function handles time events that were already removed | |
398 | */ | |
399 | iwl_mvm_te_clear_data(mvm, te_data); | |
400 | spin_unlock_bh(&mvm->time_event_lock); | |
401 | ||
402 | /* | |
403 | * It is possible that by the time we try to remove it, the time event | |
404 | * has already ended and removed. In such a case there is no need to | |
405 | * send a removal command. | |
406 | */ | |
407 | if (id == TE_MAX) { | |
408 | IWL_DEBUG_TE(mvm, "TE 0x%x has already ended\n", uid); | |
409 | return; | |
410 | } | |
411 | ||
412 | /* When we remove a TE, the UID is to be set in the id field */ | |
413 | time_cmd.id = cpu_to_le32(uid); | |
414 | time_cmd.action = cpu_to_le32(FW_CTXT_ACTION_REMOVE); | |
415 | time_cmd.id_and_color = | |
416 | cpu_to_le32(FW_CMD_ID_AND_COLOR(mvmvif->id, mvmvif->color)); | |
417 | ||
418 | IWL_DEBUG_TE(mvm, "Removing TE 0x%x\n", le32_to_cpu(time_cmd.id)); | |
419 | ret = iwl_mvm_send_cmd_pdu(mvm, TIME_EVENT_CMD, CMD_ASYNC, | |
420 | sizeof(time_cmd), &time_cmd); | |
421 | if (WARN_ON(ret)) | |
422 | return; | |
423 | } | |
424 | ||
425 | void iwl_mvm_stop_session_protection(struct iwl_mvm *mvm, | |
426 | struct ieee80211_vif *vif) | |
427 | { | |
428 | struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); | |
429 | struct iwl_mvm_time_event_data *te_data = &mvmvif->time_event_data; | |
430 | ||
431 | lockdep_assert_held(&mvm->mutex); | |
432 | iwl_mvm_remove_time_event(mvm, mvmvif, te_data); | |
433 | } | |
434 | ||
8ca151b5 JB |
435 | int iwl_mvm_start_p2p_roc(struct iwl_mvm *mvm, struct ieee80211_vif *vif, |
436 | int duration) | |
437 | { | |
438 | struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); | |
439 | struct iwl_mvm_time_event_data *te_data = &mvmvif->time_event_data; | |
8ca151b5 | 440 | struct iwl_time_event_cmd time_cmd = {}; |
8ca151b5 JB |
441 | |
442 | lockdep_assert_held(&mvm->mutex); | |
443 | if (te_data->running) { | |
444 | IWL_WARN(mvm, "P2P_DEVICE remain on channel already running\n"); | |
445 | return -EBUSY; | |
446 | } | |
447 | ||
448 | /* | |
449 | * Flush the done work, just in case it's still pending, so that | |
450 | * the work it does can complete and we can accept new frames. | |
451 | */ | |
452 | flush_work(&mvm->roc_done_wk); | |
453 | ||
8ca151b5 JB |
454 | time_cmd.action = cpu_to_le32(FW_CTXT_ACTION_ADD); |
455 | time_cmd.id_and_color = | |
456 | cpu_to_le32(FW_CMD_ID_AND_COLOR(mvmvif->id, mvmvif->color)); | |
456f6ddf | 457 | time_cmd.id = cpu_to_le32(IWL_MVM_ROC_TE_TYPE); |
8ca151b5 JB |
458 | |
459 | time_cmd.apply_time = cpu_to_le32(0); | |
460 | time_cmd.dep_policy = cpu_to_le32(TE_INDEPENDENT); | |
461 | time_cmd.is_present = cpu_to_le32(1); | |
462 | ||
463 | time_cmd.interval = cpu_to_le32(1); | |
464 | ||
465 | /* | |
456f6ddf | 466 | * IWL_MVM_ROC_TE_TYPE can have lower priority than other events |
8ca151b5 JB |
467 | * that are being scheduled by the driver/fw, and thus it might not be |
468 | * scheduled. To improve the chances of it being scheduled, allow it to | |
469 | * be fragmented. | |
470 | * In addition, for the same reasons, allow to delay the scheduling of | |
471 | * the time event. | |
472 | */ | |
473 | time_cmd.max_frags = cpu_to_le32(MSEC_TO_TU(duration)/20); | |
474 | time_cmd.max_delay = cpu_to_le32(MSEC_TO_TU(duration/2)); | |
475 | time_cmd.duration = cpu_to_le32(MSEC_TO_TU(duration)); | |
476 | time_cmd.repeat = cpu_to_le32(1); | |
477 | time_cmd.notify = cpu_to_le32(TE_NOTIF_HOST_START | TE_NOTIF_HOST_END); | |
478 | ||
ffdf968d | 479 | return iwl_mvm_time_event_send_add(mvm, vif, te_data, &time_cmd); |
8ca151b5 JB |
480 | } |
481 | ||
482 | void iwl_mvm_stop_p2p_roc(struct iwl_mvm *mvm) | |
483 | { | |
484 | struct iwl_mvm_vif *mvmvif; | |
485 | struct iwl_mvm_time_event_data *te_data; | |
486 | ||
487 | lockdep_assert_held(&mvm->mutex); | |
488 | ||
489 | /* | |
490 | * Iterate over the list of time events and find the time event that is | |
491 | * associated with a P2P_DEVICE interface. | |
492 | * This assumes that a P2P_DEVICE interface can have only a single time | |
493 | * event at any given time and this time event coresponds to a ROC | |
494 | * request | |
495 | */ | |
496 | mvmvif = NULL; | |
497 | spin_lock_bh(&mvm->time_event_lock); | |
498 | list_for_each_entry(te_data, &mvm->time_event_list, list) { | |
499 | if (te_data->vif->type == NL80211_IFTYPE_P2P_DEVICE) { | |
500 | mvmvif = iwl_mvm_vif_from_mac80211(te_data->vif); | |
501 | break; | |
502 | } | |
503 | } | |
504 | spin_unlock_bh(&mvm->time_event_lock); | |
505 | ||
506 | if (!mvmvif) { | |
507 | IWL_WARN(mvm, "P2P_DEVICE no remain on channel event\n"); | |
508 | return; | |
509 | } | |
510 | ||
511 | iwl_mvm_remove_time_event(mvm, mvmvif, te_data); | |
512 | ||
513 | iwl_mvm_roc_finished(mvm); | |
514 | } |