1/******************************************************************************
2 *
3 * This file is provided under a dual BSD/GPLv2 license.  When using or
4 * redistributing this file, you may do so under either license.
5 *
6 * GPL LICENSE SUMMARY
7 *
8 * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved.
9 * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH
10 *
11 * This program is free software; you can redistribute it and/or modify
12 * it under the terms of version 2 of the GNU General Public License as
13 * published by the Free Software Foundation.
14 *
15 * This program is distributed in the hope that it will be useful, but
16 * WITHOUT ANY WARRANTY; without even the implied warranty of
17 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
18 * General Public License for more details.
19 *
20 * You should have received a copy of the GNU General Public License
21 * along with this program; if not, write to the Free Software
22 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110,
23 * USA
24 *
25 * The full GNU General Public License is included in this distribution
26 * in the file called COPYING.
27 *
28 * Contact Information:
29 *  Intel Linux Wireless <ilw@linux.intel.com>
30 * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
31 *
32 * BSD LICENSE
33 *
34 * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved.
35 * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH
36 * All rights reserved.
37 *
38 * Redistribution and use in source and binary forms, with or without
39 * modification, are permitted provided that the following conditions
40 * are met:
41 *
42 *  * Redistributions of source code must retain the above copyright
43 *    notice, this list of conditions and the following disclaimer.
44 *  * Redistributions in binary form must reproduce the above copyright
45 *    notice, this list of conditions and the following disclaimer in
46 *    the documentation and/or other materials provided with the
47 *    distribution.
48 *  * Neither the name Intel Corporation nor the names of its
49 *    contributors may be used to endorse or promote products derived
50 *    from this software without specific prior written permission.
51 *
52 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
53 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
54 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
55 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
56 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
57 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
58 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
59 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
60 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
61 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
62 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
63 *
64 *****************************************************************************/
65
66#include <linux/jiffies.h>
67#include <net/mac80211.h>
68
69#include "iwl-notif-wait.h"
70#include "iwl-trans.h"
71#include "fw-api.h"
72#include "time-event.h"
73#include "mvm.h"
74#include "iwl-io.h"
75#include "iwl-prph.h"
76
77/*
78 * For the high priority TE use a time event type that has similar priority to
79 * the FW's action scan priority.
80 */
81#define IWL_MVM_ROC_TE_TYPE_NORMAL TE_P2P_DEVICE_DISCOVERABLE
82#define IWL_MVM_ROC_TE_TYPE_MGMT_TX TE_P2P_CLIENT_ASSOC
83
84void iwl_mvm_te_clear_data(struct iwl_mvm *mvm,
85			   struct iwl_mvm_time_event_data *te_data)
86{
87	lockdep_assert_held(&mvm->time_event_lock);
88
89	if (te_data->id == TE_MAX)
90		return;
91
92	list_del(&te_data->list);
93	te_data->running = false;
94	te_data->uid = 0;
95	te_data->id = TE_MAX;
96	te_data->vif = NULL;
97}
98
99void iwl_mvm_roc_done_wk(struct work_struct *wk)
100{
101	struct iwl_mvm *mvm = container_of(wk, struct iwl_mvm, roc_done_wk);
102	u32 queues = 0;
103
104	/*
105	 * Clear the ROC_RUNNING /ROC_AUX_RUNNING status bit.
106	 * This will cause the TX path to drop offchannel transmissions.
107	 * That would also be done by mac80211, but it is racy, in particular
108	 * in the case that the time event actually completed in the firmware
109	 * (which is handled in iwl_mvm_te_handle_notif).
110	 */
111	if (test_and_clear_bit(IWL_MVM_STATUS_ROC_RUNNING, &mvm->status)) {
112		queues |= BIT(IWL_MVM_OFFCHANNEL_QUEUE);
113		iwl_mvm_unref(mvm, IWL_MVM_REF_ROC);
114	}
115	if (test_and_clear_bit(IWL_MVM_STATUS_ROC_AUX_RUNNING, &mvm->status)) {
116		queues |= BIT(mvm->aux_queue);
117		iwl_mvm_unref(mvm, IWL_MVM_REF_ROC_AUX);
118	}
119
120	synchronize_net();
121
122	/*
123	 * Flush the offchannel queue -- this is called when the time
124	 * event finishes or is canceled, so that frames queued for it
125	 * won't get stuck on the queue and be transmitted in the next
126	 * time event.
127	 * We have to send the command asynchronously since this cannot
128	 * be under the mutex for locking reasons, but that's not an
129	 * issue as it will have to complete before the next command is
130	 * executed, and a new time event means a new command.
131	 */
132	iwl_mvm_flush_tx_path(mvm, queues, false);
133}
134
135static void iwl_mvm_roc_finished(struct iwl_mvm *mvm)
136{
137	/*
138	 * Of course, our status bit is just as racy as mac80211, so in
139	 * addition, fire off the work struct which will drop all frames
140	 * from the hardware queues that made it through the race. First
141	 * it will of course synchronize the TX path to make sure that
142	 * any *new* TX will be rejected.
143	 */
144	schedule_work(&mvm->roc_done_wk);
145}
146
147static void iwl_mvm_csa_noa_start(struct iwl_mvm *mvm)
148{
149	struct ieee80211_vif *csa_vif;
150
151	rcu_read_lock();
152
153	csa_vif = rcu_dereference(mvm->csa_vif);
154	if (!csa_vif || !csa_vif->csa_active)
155		goto out_unlock;
156
157	IWL_DEBUG_TE(mvm, "CSA NOA started\n");
158
159	/*
160	 * CSA NoA is started but we still have beacons to
161	 * transmit on the current channel.
162	 * So we just do nothing here and the switch
163	 * will be performed on the last TBTT.
164	 */
165	if (!ieee80211_csa_is_complete(csa_vif)) {
166		IWL_WARN(mvm, "CSA NOA started too early\n");
167		goto out_unlock;
168	}
169
170	ieee80211_csa_finish(csa_vif);
171
172	rcu_read_unlock();
173
174	RCU_INIT_POINTER(mvm->csa_vif, NULL);
175
176	return;
177
178out_unlock:
179	rcu_read_unlock();
180}
181
182static bool iwl_mvm_te_check_disconnect(struct iwl_mvm *mvm,
183					struct ieee80211_vif *vif,
184					const char *errmsg)
185{
186	if (vif->type != NL80211_IFTYPE_STATION)
187		return false;
188	if (vif->bss_conf.assoc && vif->bss_conf.dtim_period)
189		return false;
190	if (errmsg)
191		IWL_ERR(mvm, "%s\n", errmsg);
192
193	iwl_mvm_connection_loss(mvm, vif, errmsg);
194	return true;
195}
196
197static void
198iwl_mvm_te_handle_notify_csa(struct iwl_mvm *mvm,
199			     struct iwl_mvm_time_event_data *te_data,
200			     struct iwl_time_event_notif *notif)
201{
202	struct ieee80211_vif *vif = te_data->vif;
203	struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif);
204
205	if (!notif->status)
206		IWL_DEBUG_TE(mvm, "CSA time event failed to start\n");
207
208	switch (te_data->vif->type) {
209	case NL80211_IFTYPE_AP:
210		if (!notif->status)
211			mvmvif->csa_failed = true;
212		iwl_mvm_csa_noa_start(mvm);
213		break;
214	case NL80211_IFTYPE_STATION:
215		if (!notif->status) {
216			iwl_mvm_connection_loss(mvm, vif,
217						"CSA TE failed to start");
218			break;
219		}
220		iwl_mvm_csa_client_absent(mvm, te_data->vif);
221		ieee80211_chswitch_done(te_data->vif, true);
222		break;
223	default:
224		/* should never happen */
225		WARN_ON_ONCE(1);
226		break;
227	}
228
229	/* we don't need it anymore */
230	iwl_mvm_te_clear_data(mvm, te_data);
231}
232
233static void iwl_mvm_te_check_trigger(struct iwl_mvm *mvm,
234				     struct iwl_time_event_notif *notif,
235				     struct iwl_mvm_time_event_data *te_data)
236{
237	struct iwl_fw_dbg_trigger_tlv *trig;
238	struct iwl_fw_dbg_trigger_time_event *te_trig;
239	int i;
240
241	if (!iwl_fw_dbg_trigger_enabled(mvm->fw, FW_DBG_TRIGGER_TIME_EVENT))
242		return;
243
244	trig = iwl_fw_dbg_get_trigger(mvm->fw, FW_DBG_TRIGGER_TIME_EVENT);
245	te_trig = (void *)trig->data;
246
247	if (!iwl_fw_dbg_trigger_check_stop(mvm, te_data->vif, trig))
248		return;
249
250	for (i = 0; i < ARRAY_SIZE(te_trig->time_events); i++) {
251		u32 trig_te_id = le32_to_cpu(te_trig->time_events[i].id);
252		u32 trig_action_bitmap =
253			le32_to_cpu(te_trig->time_events[i].action_bitmap);
254		u32 trig_status_bitmap =
255			le32_to_cpu(te_trig->time_events[i].status_bitmap);
256
257		if (trig_te_id != te_data->id ||
258		    !(trig_action_bitmap & le32_to_cpu(notif->action)) ||
259		    !(trig_status_bitmap & BIT(le32_to_cpu(notif->status))))
260			continue;
261
262		iwl_mvm_fw_dbg_collect_trig(mvm, trig,
263					    "Time event %d Action 0x%x received status: %d",
264					    te_data->id,
265					    le32_to_cpu(notif->action),
266					    le32_to_cpu(notif->status));
267		break;
268	}
269}
270
271/*
272 * Handles a FW notification for an event that is known to the driver.
273 *
274 * @mvm: the mvm component
275 * @te_data: the time event data
276 * @notif: the notification data corresponding the time event data.
277 */
278static void iwl_mvm_te_handle_notif(struct iwl_mvm *mvm,
279				    struct iwl_mvm_time_event_data *te_data,
280				    struct iwl_time_event_notif *notif)
281{
282	lockdep_assert_held(&mvm->time_event_lock);
283
284	IWL_DEBUG_TE(mvm, "Handle time event notif - UID = 0x%x action %d\n",
285		     le32_to_cpu(notif->unique_id),
286		     le32_to_cpu(notif->action));
287
288	iwl_mvm_te_check_trigger(mvm, notif, te_data);
289
290	/*
291	 * The FW sends the start/end time event notifications even for events
292	 * that it fails to schedule. This is indicated in the status field of
293	 * the notification. This happens in cases that the scheduler cannot
294	 * find a schedule that can handle the event (for example requesting a
295	 * P2P Device discoveribility, while there are other higher priority
296	 * events in the system).
297	 */
298	if (!le32_to_cpu(notif->status)) {
299		const char *msg;
300
301		if (notif->action & cpu_to_le32(TE_V2_NOTIF_HOST_EVENT_START))
302			msg = "Time Event start notification failure";
303		else
304			msg = "Time Event end notification failure";
305
306		IWL_DEBUG_TE(mvm, "%s\n", msg);
307
308		if (iwl_mvm_te_check_disconnect(mvm, te_data->vif, msg)) {
309			iwl_mvm_te_clear_data(mvm, te_data);
310			return;
311		}
312	}
313
314	if (le32_to_cpu(notif->action) & TE_V2_NOTIF_HOST_EVENT_END) {
315		IWL_DEBUG_TE(mvm,
316			     "TE ended - current time %lu, estimated end %lu\n",
317			     jiffies, te_data->end_jiffies);
318
319		switch (te_data->vif->type) {
320		case NL80211_IFTYPE_P2P_DEVICE:
321			ieee80211_remain_on_channel_expired(mvm->hw);
322			iwl_mvm_roc_finished(mvm);
323			break;
324		case NL80211_IFTYPE_STATION:
325			/*
326			 * By now, we should have finished association
327			 * and know the dtim period.
328			 */
329			iwl_mvm_te_check_disconnect(mvm, te_data->vif,
330				"No association and the time event is over already...");
331			break;
332		default:
333			break;
334		}
335
336		iwl_mvm_te_clear_data(mvm, te_data);
337	} else if (le32_to_cpu(notif->action) & TE_V2_NOTIF_HOST_EVENT_START) {
338		te_data->running = true;
339		te_data->end_jiffies = TU_TO_EXP_TIME(te_data->duration);
340
341		if (te_data->vif->type == NL80211_IFTYPE_P2P_DEVICE) {
342			set_bit(IWL_MVM_STATUS_ROC_RUNNING, &mvm->status);
343			iwl_mvm_ref(mvm, IWL_MVM_REF_ROC);
344			ieee80211_ready_on_channel(mvm->hw);
345		} else if (te_data->id == TE_CHANNEL_SWITCH_PERIOD) {
346			iwl_mvm_te_handle_notify_csa(mvm, te_data, notif);
347		}
348	} else {
349		IWL_WARN(mvm, "Got TE with unknown action\n");
350	}
351}
352
353/*
354 * Handle A Aux ROC time event
355 */
356static int iwl_mvm_aux_roc_te_handle_notif(struct iwl_mvm *mvm,
357					   struct iwl_time_event_notif *notif)
358{
359	struct iwl_mvm_time_event_data *te_data, *tmp;
360	bool aux_roc_te = false;
361
362	list_for_each_entry_safe(te_data, tmp, &mvm->aux_roc_te_list, list) {
363		if (le32_to_cpu(notif->unique_id) == te_data->uid) {
364			aux_roc_te = true;
365			break;
366		}
367	}
368	if (!aux_roc_te) /* Not a Aux ROC time event */
369		return -EINVAL;
370
371	iwl_mvm_te_check_trigger(mvm, notif, te_data);
372
373	if (!le32_to_cpu(notif->status)) {
374		IWL_DEBUG_TE(mvm,
375			     "ERROR: Aux ROC Time Event %s notification failure\n",
376			     (le32_to_cpu(notif->action) &
377			      TE_V2_NOTIF_HOST_EVENT_START) ? "start" : "end");
378		return -EINVAL;
379	}
380
381	IWL_DEBUG_TE(mvm,
382		     "Aux ROC time event notification  - UID = 0x%x action %d\n",
383		     le32_to_cpu(notif->unique_id),
384		     le32_to_cpu(notif->action));
385
386	if (le32_to_cpu(notif->action) == TE_V2_NOTIF_HOST_EVENT_END) {
387		/* End TE, notify mac80211 */
388		ieee80211_remain_on_channel_expired(mvm->hw);
389		iwl_mvm_roc_finished(mvm); /* flush aux queue */
390		list_del(&te_data->list); /* remove from list */
391		te_data->running = false;
392		te_data->vif = NULL;
393		te_data->uid = 0;
394		te_data->id = TE_MAX;
395	} else if (le32_to_cpu(notif->action) == TE_V2_NOTIF_HOST_EVENT_START) {
396		set_bit(IWL_MVM_STATUS_ROC_AUX_RUNNING, &mvm->status);
397		te_data->running = true;
398		iwl_mvm_ref(mvm, IWL_MVM_REF_ROC_AUX);
399		ieee80211_ready_on_channel(mvm->hw); /* Start TE */
400	} else {
401		IWL_DEBUG_TE(mvm,
402			     "ERROR: Unknown Aux ROC Time Event (action = %d)\n",
403			     le32_to_cpu(notif->action));
404		return -EINVAL;
405	}
406
407	return 0;
408}
409
410/*
411 * The Rx handler for time event notifications
412 */
413int iwl_mvm_rx_time_event_notif(struct iwl_mvm *mvm,
414				struct iwl_rx_cmd_buffer *rxb,
415				struct iwl_device_cmd *cmd)
416{
417	struct iwl_rx_packet *pkt = rxb_addr(rxb);
418	struct iwl_time_event_notif *notif = (void *)pkt->data;
419	struct iwl_mvm_time_event_data *te_data, *tmp;
420
421	IWL_DEBUG_TE(mvm, "Time event notification - UID = 0x%x action %d\n",
422		     le32_to_cpu(notif->unique_id),
423		     le32_to_cpu(notif->action));
424
425	spin_lock_bh(&mvm->time_event_lock);
426	/* This time event is triggered for Aux ROC request */
427	if (!iwl_mvm_aux_roc_te_handle_notif(mvm, notif))
428		goto unlock;
429
430	list_for_each_entry_safe(te_data, tmp, &mvm->time_event_list, list) {
431		if (le32_to_cpu(notif->unique_id) == te_data->uid)
432			iwl_mvm_te_handle_notif(mvm, te_data, notif);
433	}
434unlock:
435	spin_unlock_bh(&mvm->time_event_lock);
436
437	return 0;
438}
439
440static bool iwl_mvm_te_notif(struct iwl_notif_wait_data *notif_wait,
441			     struct iwl_rx_packet *pkt, void *data)
442{
443	struct iwl_mvm *mvm =
444		container_of(notif_wait, struct iwl_mvm, notif_wait);
445	struct iwl_mvm_time_event_data *te_data = data;
446	struct iwl_time_event_notif *resp;
447	int resp_len = iwl_rx_packet_payload_len(pkt);
448
449	if (WARN_ON(pkt->hdr.cmd != TIME_EVENT_NOTIFICATION))
450		return true;
451
452	if (WARN_ON_ONCE(resp_len != sizeof(*resp))) {
453		IWL_ERR(mvm, "Invalid TIME_EVENT_NOTIFICATION response\n");
454		return true;
455	}
456
457	resp = (void *)pkt->data;
458
459	/* te_data->uid is already set in the TIME_EVENT_CMD response */
460	if (le32_to_cpu(resp->unique_id) != te_data->uid)
461		return false;
462
463	IWL_DEBUG_TE(mvm, "TIME_EVENT_NOTIFICATION response - UID = 0x%x\n",
464		     te_data->uid);
465	if (!resp->status)
466		IWL_ERR(mvm,
467			"TIME_EVENT_NOTIFICATION received but not executed\n");
468
469	return true;
470}
471
472static bool iwl_mvm_time_event_response(struct iwl_notif_wait_data *notif_wait,
473					struct iwl_rx_packet *pkt, void *data)
474{
475	struct iwl_mvm *mvm =
476		container_of(notif_wait, struct iwl_mvm, notif_wait);
477	struct iwl_mvm_time_event_data *te_data = data;
478	struct iwl_time_event_resp *resp;
479	int resp_len = iwl_rx_packet_payload_len(pkt);
480
481	if (WARN_ON(pkt->hdr.cmd != TIME_EVENT_CMD))
482		return true;
483
484	if (WARN_ON_ONCE(resp_len != sizeof(*resp))) {
485		IWL_ERR(mvm, "Invalid TIME_EVENT_CMD response\n");
486		return true;
487	}
488
489	resp = (void *)pkt->data;
490
491	/* we should never get a response to another TIME_EVENT_CMD here */
492	if (WARN_ON_ONCE(le32_to_cpu(resp->id) != te_data->id))
493		return false;
494
495	te_data->uid = le32_to_cpu(resp->unique_id);
496	IWL_DEBUG_TE(mvm, "TIME_EVENT_CMD response - UID = 0x%x\n",
497		     te_data->uid);
498	return true;
499}
500
501static int iwl_mvm_time_event_send_add(struct iwl_mvm *mvm,
502				       struct ieee80211_vif *vif,
503				       struct iwl_mvm_time_event_data *te_data,
504				       struct iwl_time_event_cmd *te_cmd)
505{
506	static const u8 time_event_response[] = { TIME_EVENT_CMD };
507	struct iwl_notification_wait wait_time_event;
508	int ret;
509
510	lockdep_assert_held(&mvm->mutex);
511
512	IWL_DEBUG_TE(mvm, "Add new TE, duration %d TU\n",
513		     le32_to_cpu(te_cmd->duration));
514
515	spin_lock_bh(&mvm->time_event_lock);
516	if (WARN_ON(te_data->id != TE_MAX)) {
517		spin_unlock_bh(&mvm->time_event_lock);
518		return -EIO;
519	}
520	te_data->vif = vif;
521	te_data->duration = le32_to_cpu(te_cmd->duration);
522	te_data->id = le32_to_cpu(te_cmd->id);
523	list_add_tail(&te_data->list, &mvm->time_event_list);
524	spin_unlock_bh(&mvm->time_event_lock);
525
526	/*
527	 * Use a notification wait, which really just processes the
528	 * command response and doesn't wait for anything, in order
529	 * to be able to process the response and get the UID inside
530	 * the RX path. Using CMD_WANT_SKB doesn't work because it
531	 * stores the buffer and then wakes up this thread, by which
532	 * time another notification (that the time event started)
533	 * might already be processed unsuccessfully.
534	 */
535	iwl_init_notification_wait(&mvm->notif_wait, &wait_time_event,
536				   time_event_response,
537				   ARRAY_SIZE(time_event_response),
538				   iwl_mvm_time_event_response, te_data);
539
540	ret = iwl_mvm_send_cmd_pdu(mvm, TIME_EVENT_CMD, 0,
541					    sizeof(*te_cmd), te_cmd);
542	if (ret) {
543		IWL_ERR(mvm, "Couldn't send TIME_EVENT_CMD: %d\n", ret);
544		iwl_remove_notification(&mvm->notif_wait, &wait_time_event);
545		goto out_clear_te;
546	}
547
548	/* No need to wait for anything, so just pass 1 (0 isn't valid) */
549	ret = iwl_wait_notification(&mvm->notif_wait, &wait_time_event, 1);
550	/* should never fail */
551	WARN_ON_ONCE(ret);
552
553	if (ret) {
554 out_clear_te:
555		spin_lock_bh(&mvm->time_event_lock);
556		iwl_mvm_te_clear_data(mvm, te_data);
557		spin_unlock_bh(&mvm->time_event_lock);
558	}
559	return ret;
560}
561
562void iwl_mvm_protect_session(struct iwl_mvm *mvm,
563			     struct ieee80211_vif *vif,
564			     u32 duration, u32 min_duration,
565			     u32 max_delay, bool wait_for_notif)
566{
567	struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif);
568	struct iwl_mvm_time_event_data *te_data = &mvmvif->time_event_data;
569	const u8 te_notif_response[] = { TIME_EVENT_NOTIFICATION };
570	struct iwl_notification_wait wait_te_notif;
571	struct iwl_time_event_cmd time_cmd = {};
572
573	lockdep_assert_held(&mvm->mutex);
574
575	if (te_data->running &&
576	    time_after(te_data->end_jiffies, TU_TO_EXP_TIME(min_duration))) {
577		IWL_DEBUG_TE(mvm, "We have enough time in the current TE: %u\n",
578			     jiffies_to_msecs(te_data->end_jiffies - jiffies));
579		return;
580	}
581
582	if (te_data->running) {
583		IWL_DEBUG_TE(mvm, "extend 0x%x: only %u ms left\n",
584			     te_data->uid,
585			     jiffies_to_msecs(te_data->end_jiffies - jiffies));
586		/*
587		 * we don't have enough time
588		 * cancel the current TE and issue a new one
589		 * Of course it would be better to remove the old one only
590		 * when the new one is added, but we don't care if we are off
591		 * channel for a bit. All we need to do, is not to return
592		 * before we actually begin to be on the channel.
593		 */
594		iwl_mvm_stop_session_protection(mvm, vif);
595	}
596
597	time_cmd.action = cpu_to_le32(FW_CTXT_ACTION_ADD);
598	time_cmd.id_and_color =
599		cpu_to_le32(FW_CMD_ID_AND_COLOR(mvmvif->id, mvmvif->color));
600	time_cmd.id = cpu_to_le32(TE_BSS_STA_AGGRESSIVE_ASSOC);
601
602	time_cmd.apply_time =
603		cpu_to_le32(iwl_read_prph(mvm->trans, DEVICE_SYSTEM_TIME_REG));
604
605	time_cmd.max_frags = TE_V2_FRAG_NONE;
606	time_cmd.max_delay = cpu_to_le32(max_delay);
607	/* TODO: why do we need to interval = bi if it is not periodic? */
608	time_cmd.interval = cpu_to_le32(1);
609	time_cmd.duration = cpu_to_le32(duration);
610	time_cmd.repeat = 1;
611	time_cmd.policy = cpu_to_le16(TE_V2_NOTIF_HOST_EVENT_START |
612				      TE_V2_NOTIF_HOST_EVENT_END |
613				      T2_V2_START_IMMEDIATELY);
614
615	if (!wait_for_notif) {
616		iwl_mvm_time_event_send_add(mvm, vif, te_data, &time_cmd);
617		return;
618	}
619
620	/*
621	 * Create notification_wait for the TIME_EVENT_NOTIFICATION to use
622	 * right after we send the time event
623	 */
624	iwl_init_notification_wait(&mvm->notif_wait, &wait_te_notif,
625				   te_notif_response,
626				   ARRAY_SIZE(te_notif_response),
627				   iwl_mvm_te_notif, te_data);
628
629	/* If TE was sent OK - wait for the notification that started */
630	if (iwl_mvm_time_event_send_add(mvm, vif, te_data, &time_cmd)) {
631		IWL_ERR(mvm, "Failed to add TE to protect session\n");
632		iwl_remove_notification(&mvm->notif_wait, &wait_te_notif);
633	} else if (iwl_wait_notification(&mvm->notif_wait, &wait_te_notif,
634					 TU_TO_JIFFIES(max_delay))) {
635		IWL_ERR(mvm, "Failed to protect session until TE\n");
636	}
637}
638
639static bool __iwl_mvm_remove_time_event(struct iwl_mvm *mvm,
640					struct iwl_mvm_time_event_data *te_data,
641					u32 *uid)
642{
643	u32 id;
644
645	/*
646	 * It is possible that by the time we got to this point the time
647	 * event was already removed.
648	 */
649	spin_lock_bh(&mvm->time_event_lock);
650
651	/* Save time event uid before clearing its data */
652	*uid = te_data->uid;
653	id = te_data->id;
654
655	/*
656	 * The clear_data function handles time events that were already removed
657	 */
658	iwl_mvm_te_clear_data(mvm, te_data);
659	spin_unlock_bh(&mvm->time_event_lock);
660
661	/*
662	 * It is possible that by the time we try to remove it, the time event
663	 * has already ended and removed. In such a case there is no need to
664	 * send a removal command.
665	 */
666	if (id == TE_MAX) {
667		IWL_DEBUG_TE(mvm, "TE 0x%x has already ended\n", *uid);
668		return false;
669	}
670
671	return true;
672}
673
674/*
675 * Explicit request to remove a aux roc time event. The removal of a time
676 * event needs to be synchronized with the flow of a time event's end
677 * notification, which also removes the time event from the op mode
678 * data structures.
679 */
680static void iwl_mvm_remove_aux_roc_te(struct iwl_mvm *mvm,
681				      struct iwl_mvm_vif *mvmvif,
682				      struct iwl_mvm_time_event_data *te_data)
683{
684	struct iwl_hs20_roc_req aux_cmd = {};
685	u32 uid;
686	int ret;
687
688	if (!__iwl_mvm_remove_time_event(mvm, te_data, &uid))
689		return;
690
691	aux_cmd.event_unique_id = cpu_to_le32(uid);
692	aux_cmd.action = cpu_to_le32(FW_CTXT_ACTION_REMOVE);
693	aux_cmd.id_and_color =
694		cpu_to_le32(FW_CMD_ID_AND_COLOR(mvmvif->id, mvmvif->color));
695	IWL_DEBUG_TE(mvm, "Removing BSS AUX ROC TE 0x%x\n",
696		     le32_to_cpu(aux_cmd.event_unique_id));
697	ret = iwl_mvm_send_cmd_pdu(mvm, HOT_SPOT_CMD, 0,
698				   sizeof(aux_cmd), &aux_cmd);
699
700	if (WARN_ON(ret))
701		return;
702}
703
704/*
705 * Explicit request to remove a time event. The removal of a time event needs to
706 * be synchronized with the flow of a time event's end notification, which also
707 * removes the time event from the op mode data structures.
708 */
709void iwl_mvm_remove_time_event(struct iwl_mvm *mvm,
710			       struct iwl_mvm_vif *mvmvif,
711			       struct iwl_mvm_time_event_data *te_data)
712{
713	struct iwl_time_event_cmd time_cmd = {};
714	u32 uid;
715	int ret;
716
717	if (!__iwl_mvm_remove_time_event(mvm, te_data, &uid))
718		return;
719
720	/* When we remove a TE, the UID is to be set in the id field */
721	time_cmd.id = cpu_to_le32(uid);
722	time_cmd.action = cpu_to_le32(FW_CTXT_ACTION_REMOVE);
723	time_cmd.id_and_color =
724		cpu_to_le32(FW_CMD_ID_AND_COLOR(mvmvif->id, mvmvif->color));
725
726	IWL_DEBUG_TE(mvm, "Removing TE 0x%x\n", le32_to_cpu(time_cmd.id));
727	ret = iwl_mvm_send_cmd_pdu(mvm, TIME_EVENT_CMD, 0,
728				   sizeof(time_cmd), &time_cmd);
729	if (WARN_ON(ret))
730		return;
731}
732
733void iwl_mvm_stop_session_protection(struct iwl_mvm *mvm,
734				     struct ieee80211_vif *vif)
735{
736	struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif);
737	struct iwl_mvm_time_event_data *te_data = &mvmvif->time_event_data;
738
739	lockdep_assert_held(&mvm->mutex);
740	iwl_mvm_remove_time_event(mvm, mvmvif, te_data);
741}
742
743int iwl_mvm_start_p2p_roc(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
744			  int duration, enum ieee80211_roc_type type)
745{
746	struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif);
747	struct iwl_mvm_time_event_data *te_data = &mvmvif->time_event_data;
748	struct iwl_time_event_cmd time_cmd = {};
749
750	lockdep_assert_held(&mvm->mutex);
751	if (te_data->running) {
752		IWL_WARN(mvm, "P2P_DEVICE remain on channel already running\n");
753		return -EBUSY;
754	}
755
756	/*
757	 * Flush the done work, just in case it's still pending, so that
758	 * the work it does can complete and we can accept new frames.
759	 */
760	flush_work(&mvm->roc_done_wk);
761
762	time_cmd.action = cpu_to_le32(FW_CTXT_ACTION_ADD);
763	time_cmd.id_and_color =
764		cpu_to_le32(FW_CMD_ID_AND_COLOR(mvmvif->id, mvmvif->color));
765
766	switch (type) {
767	case IEEE80211_ROC_TYPE_NORMAL:
768		time_cmd.id = cpu_to_le32(IWL_MVM_ROC_TE_TYPE_NORMAL);
769		break;
770	case IEEE80211_ROC_TYPE_MGMT_TX:
771		time_cmd.id = cpu_to_le32(IWL_MVM_ROC_TE_TYPE_MGMT_TX);
772		break;
773	default:
774		WARN_ONCE(1, "Got an invalid ROC type\n");
775		return -EINVAL;
776	}
777
778	time_cmd.apply_time = cpu_to_le32(0);
779	time_cmd.interval = cpu_to_le32(1);
780
781	/*
782	 * The P2P Device TEs can have lower priority than other events
783	 * that are being scheduled by the driver/fw, and thus it might not be
784	 * scheduled. To improve the chances of it being scheduled, allow them
785	 * to be fragmented, and in addition allow them to be delayed.
786	 */
787	time_cmd.max_frags = min(MSEC_TO_TU(duration)/50, TE_V2_FRAG_ENDLESS);
788	time_cmd.max_delay = cpu_to_le32(MSEC_TO_TU(duration/2));
789	time_cmd.duration = cpu_to_le32(MSEC_TO_TU(duration));
790	time_cmd.repeat = 1;
791	time_cmd.policy = cpu_to_le16(TE_V2_NOTIF_HOST_EVENT_START |
792				      TE_V2_NOTIF_HOST_EVENT_END |
793				      T2_V2_START_IMMEDIATELY);
794
795	return iwl_mvm_time_event_send_add(mvm, vif, te_data, &time_cmd);
796}
797
798void iwl_mvm_stop_roc(struct iwl_mvm *mvm)
799{
800	struct iwl_mvm_vif *mvmvif;
801	struct iwl_mvm_time_event_data *te_data;
802	bool is_p2p = false;
803
804	lockdep_assert_held(&mvm->mutex);
805
806	mvmvif = NULL;
807	spin_lock_bh(&mvm->time_event_lock);
808
809	/*
810	 * Iterate over the list of time events and find the time event that is
811	 * associated with a P2P_DEVICE interface.
812	 * This assumes that a P2P_DEVICE interface can have only a single time
813	 * event at any given time and this time event coresponds to a ROC
814	 * request
815	 */
816	list_for_each_entry(te_data, &mvm->time_event_list, list) {
817		if (te_data->vif->type == NL80211_IFTYPE_P2P_DEVICE) {
818			mvmvif = iwl_mvm_vif_from_mac80211(te_data->vif);
819			is_p2p = true;
820			goto remove_te;
821		}
822	}
823
824	/*
825	 * Iterate over the list of aux roc time events and find the time
826	 * event that is associated with a BSS interface.
827	 * This assumes that a BSS interface can have only a single time
828	 * event at any given time and this time event corresponds to a ROC
829	 * request
830	 */
831	list_for_each_entry(te_data, &mvm->aux_roc_te_list, list) {
832		mvmvif = iwl_mvm_vif_from_mac80211(te_data->vif);
833		goto remove_te;
834	}
835
836remove_te:
837	spin_unlock_bh(&mvm->time_event_lock);
838
839	if (!mvmvif) {
840		IWL_WARN(mvm, "No remain on channel event\n");
841		return;
842	}
843
844	if (is_p2p)
845		iwl_mvm_remove_time_event(mvm, mvmvif, te_data);
846	else
847		iwl_mvm_remove_aux_roc_te(mvm, mvmvif, te_data);
848
849	iwl_mvm_roc_finished(mvm);
850}
851
852int iwl_mvm_schedule_csa_period(struct iwl_mvm *mvm,
853				struct ieee80211_vif *vif,
854				u32 duration, u32 apply_time)
855{
856	struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif);
857	struct iwl_mvm_time_event_data *te_data = &mvmvif->time_event_data;
858	struct iwl_time_event_cmd time_cmd = {};
859
860	lockdep_assert_held(&mvm->mutex);
861
862	if (te_data->running) {
863		IWL_DEBUG_TE(mvm, "CS period is already scheduled\n");
864		return -EBUSY;
865	}
866
867	time_cmd.action = cpu_to_le32(FW_CTXT_ACTION_ADD);
868	time_cmd.id_and_color =
869		cpu_to_le32(FW_CMD_ID_AND_COLOR(mvmvif->id, mvmvif->color));
870	time_cmd.id = cpu_to_le32(TE_CHANNEL_SWITCH_PERIOD);
871	time_cmd.apply_time = cpu_to_le32(apply_time);
872	time_cmd.max_frags = TE_V2_FRAG_NONE;
873	time_cmd.duration = cpu_to_le32(duration);
874	time_cmd.repeat = 1;
875	time_cmd.interval = cpu_to_le32(1);
876	time_cmd.policy = cpu_to_le16(TE_V2_NOTIF_HOST_EVENT_START |
877				      TE_V2_ABSENCE);
878
879	return iwl_mvm_time_event_send_add(mvm, vif, te_data, &time_cmd);
880}
881