mwifiex: add custom IE framework
[deliverable/linux.git] / drivers / net / wireless / iwlwifi / iwl-agn-lib.c
1 /******************************************************************************
2 *
3 * GPL LICENSE SUMMARY
4 *
5 * Copyright(c) 2008 - 2012 Intel Corporation. All rights reserved.
6 *
7 * This program is free software; you can redistribute it and/or modify
8 * it under the terms of version 2 of the GNU General Public License as
9 * published by the Free Software Foundation.
10 *
11 * This program is distributed in the hope that it will be useful, but
12 * WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14 * General Public License for more details.
15 *
16 * You should have received a copy of the GNU General Public License
17 * along with this program; if not, write to the Free Software
18 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110,
19 * USA
20 *
21 * The full GNU General Public License is included in this distribution
22 * in the file called LICENSE.GPL.
23 *
24 * Contact Information:
25 * Intel Linux Wireless <ilw@linux.intel.com>
26 * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
27 *
28 *****************************************************************************/
29 #include <linux/etherdevice.h>
30 #include <linux/kernel.h>
31 #include <linux/module.h>
32 #include <linux/init.h>
33 #include <linux/sched.h>
34 #include <net/mac80211.h>
35
36 #include "iwl-dev.h"
37 #include "iwl-io.h"
38 #include "iwl-agn-hw.h"
39 #include "iwl-agn.h"
40 #include "iwl-trans.h"
41 #include "iwl-modparams.h"
42
43 int iwlagn_hw_valid_rtc_data_addr(u32 addr)
44 {
45 return (addr >= IWLAGN_RTC_DATA_LOWER_BOUND) &&
46 (addr < IWLAGN_RTC_DATA_UPPER_BOUND);
47 }
48
49 int iwlagn_send_tx_power(struct iwl_priv *priv)
50 {
51 struct iwlagn_tx_power_dbm_cmd tx_power_cmd;
52 u8 tx_ant_cfg_cmd;
53
54 if (WARN_ONCE(test_bit(STATUS_SCAN_HW, &priv->status),
55 "TX Power requested while scanning!\n"))
56 return -EAGAIN;
57
58 /* half dBm need to multiply */
59 tx_power_cmd.global_lmt = (s8)(2 * priv->tx_power_user_lmt);
60
61 if (priv->tx_power_lmt_in_half_dbm &&
62 priv->tx_power_lmt_in_half_dbm < tx_power_cmd.global_lmt) {
63 /*
64 * For the newer devices which using enhanced/extend tx power
65 * table in EEPROM, the format is in half dBm. driver need to
66 * convert to dBm format before report to mac80211.
67 * By doing so, there is a possibility of 1/2 dBm resolution
68 * lost. driver will perform "round-up" operation before
69 * reporting, but it will cause 1/2 dBm tx power over the
70 * regulatory limit. Perform the checking here, if the
71 * "tx_power_user_lmt" is higher than EEPROM value (in
72 * half-dBm format), lower the tx power based on EEPROM
73 */
74 tx_power_cmd.global_lmt = priv->tx_power_lmt_in_half_dbm;
75 }
76 tx_power_cmd.flags = IWLAGN_TX_POWER_NO_CLOSED;
77 tx_power_cmd.srv_chan_lmt = IWLAGN_TX_POWER_AUTO;
78
79 if (IWL_UCODE_API(priv->fw->ucode_ver) == 1)
80 tx_ant_cfg_cmd = REPLY_TX_POWER_DBM_CMD_V1;
81 else
82 tx_ant_cfg_cmd = REPLY_TX_POWER_DBM_CMD;
83
84 return iwl_dvm_send_cmd_pdu(priv, tx_ant_cfg_cmd, CMD_SYNC,
85 sizeof(tx_power_cmd), &tx_power_cmd);
86 }
87
88 void iwlagn_temperature(struct iwl_priv *priv)
89 {
90 lockdep_assert_held(&priv->statistics.lock);
91
92 /* store temperature from correct statistics (in Celsius) */
93 priv->temperature = le32_to_cpu(priv->statistics.common.temperature);
94 iwl_tt_handler(priv);
95 }
96
97 int iwlagn_hwrate_to_mac80211_idx(u32 rate_n_flags, enum ieee80211_band band)
98 {
99 int idx = 0;
100 int band_offset = 0;
101
102 /* HT rate format: mac80211 wants an MCS number, which is just LSB */
103 if (rate_n_flags & RATE_MCS_HT_MSK) {
104 idx = (rate_n_flags & 0xff);
105 return idx;
106 /* Legacy rate format, search for match in table */
107 } else {
108 if (band == IEEE80211_BAND_5GHZ)
109 band_offset = IWL_FIRST_OFDM_RATE;
110 for (idx = band_offset; idx < IWL_RATE_COUNT_LEGACY; idx++)
111 if (iwl_rates[idx].plcp == (rate_n_flags & 0xFF))
112 return idx - band_offset;
113 }
114
115 return -1;
116 }
117
118 int iwlagn_manage_ibss_station(struct iwl_priv *priv,
119 struct ieee80211_vif *vif, bool add)
120 {
121 struct iwl_vif_priv *vif_priv = (void *)vif->drv_priv;
122
123 if (add)
124 return iwlagn_add_bssid_station(priv, vif_priv->ctx,
125 vif->bss_conf.bssid,
126 &vif_priv->ibss_bssid_sta_id);
127 return iwl_remove_station(priv, vif_priv->ibss_bssid_sta_id,
128 vif->bss_conf.bssid);
129 }
130
131 /**
132 * iwlagn_txfifo_flush: send REPLY_TXFIFO_FLUSH command to uCode
133 *
134 * pre-requirements:
135 * 1. acquire mutex before calling
136 * 2. make sure rf is on and not in exit state
137 */
138 int iwlagn_txfifo_flush(struct iwl_priv *priv, u16 flush_control)
139 {
140 struct iwl_txfifo_flush_cmd flush_cmd;
141 struct iwl_host_cmd cmd = {
142 .id = REPLY_TXFIFO_FLUSH,
143 .len = { sizeof(struct iwl_txfifo_flush_cmd), },
144 .flags = CMD_SYNC,
145 .data = { &flush_cmd, },
146 };
147
148 might_sleep();
149
150 memset(&flush_cmd, 0, sizeof(flush_cmd));
151 if (flush_control & BIT(IWL_RXON_CTX_BSS))
152 flush_cmd.fifo_control = IWL_SCD_VO_MSK | IWL_SCD_VI_MSK |
153 IWL_SCD_BE_MSK | IWL_SCD_BK_MSK |
154 IWL_SCD_MGMT_MSK;
155 if ((flush_control & BIT(IWL_RXON_CTX_PAN)) &&
156 (priv->valid_contexts != BIT(IWL_RXON_CTX_BSS)))
157 flush_cmd.fifo_control |= IWL_PAN_SCD_VO_MSK |
158 IWL_PAN_SCD_VI_MSK | IWL_PAN_SCD_BE_MSK |
159 IWL_PAN_SCD_BK_MSK | IWL_PAN_SCD_MGMT_MSK |
160 IWL_PAN_SCD_MULTICAST_MSK;
161
162 if (priv->hw_params.sku & EEPROM_SKU_CAP_11N_ENABLE)
163 flush_cmd.fifo_control |= IWL_AGG_TX_QUEUE_MSK;
164
165 IWL_DEBUG_INFO(priv, "fifo queue control: 0X%x\n",
166 flush_cmd.fifo_control);
167 flush_cmd.flush_control = cpu_to_le16(flush_control);
168
169 return iwl_dvm_send_cmd(priv, &cmd);
170 }
171
172 void iwlagn_dev_txfifo_flush(struct iwl_priv *priv, u16 flush_control)
173 {
174 mutex_lock(&priv->mutex);
175 ieee80211_stop_queues(priv->hw);
176 if (iwlagn_txfifo_flush(priv, IWL_DROP_ALL)) {
177 IWL_ERR(priv, "flush request fail\n");
178 goto done;
179 }
180 IWL_DEBUG_INFO(priv, "wait transmit/flush all frames\n");
181 iwl_trans_wait_tx_queue_empty(priv->trans);
182 done:
183 ieee80211_wake_queues(priv->hw);
184 mutex_unlock(&priv->mutex);
185 }
186
187 /*
188 * BT coex
189 */
190 /* Notmal TDM */
191 static const __le32 iwlagn_def_3w_lookup[IWLAGN_BT_DECISION_LUT_SIZE] = {
192 cpu_to_le32(0xaaaaaaaa),
193 cpu_to_le32(0xaaaaaaaa),
194 cpu_to_le32(0xaeaaaaaa),
195 cpu_to_le32(0xaaaaaaaa),
196 cpu_to_le32(0xcc00ff28),
197 cpu_to_le32(0x0000aaaa),
198 cpu_to_le32(0xcc00aaaa),
199 cpu_to_le32(0x0000aaaa),
200 cpu_to_le32(0xc0004000),
201 cpu_to_le32(0x00004000),
202 cpu_to_le32(0xf0005000),
203 cpu_to_le32(0xf0005000),
204 };
205
206
207 /* Loose Coex */
208 static const __le32 iwlagn_loose_lookup[IWLAGN_BT_DECISION_LUT_SIZE] = {
209 cpu_to_le32(0xaaaaaaaa),
210 cpu_to_le32(0xaaaaaaaa),
211 cpu_to_le32(0xaeaaaaaa),
212 cpu_to_le32(0xaaaaaaaa),
213 cpu_to_le32(0xcc00ff28),
214 cpu_to_le32(0x0000aaaa),
215 cpu_to_le32(0xcc00aaaa),
216 cpu_to_le32(0x0000aaaa),
217 cpu_to_le32(0x00000000),
218 cpu_to_le32(0x00000000),
219 cpu_to_le32(0xf0005000),
220 cpu_to_le32(0xf0005000),
221 };
222
223 /* Full concurrency */
224 static const __le32 iwlagn_concurrent_lookup[IWLAGN_BT_DECISION_LUT_SIZE] = {
225 cpu_to_le32(0xaaaaaaaa),
226 cpu_to_le32(0xaaaaaaaa),
227 cpu_to_le32(0xaaaaaaaa),
228 cpu_to_le32(0xaaaaaaaa),
229 cpu_to_le32(0xaaaaaaaa),
230 cpu_to_le32(0xaaaaaaaa),
231 cpu_to_le32(0xaaaaaaaa),
232 cpu_to_le32(0xaaaaaaaa),
233 cpu_to_le32(0x00000000),
234 cpu_to_le32(0x00000000),
235 cpu_to_le32(0x00000000),
236 cpu_to_le32(0x00000000),
237 };
238
239 void iwlagn_send_advance_bt_config(struct iwl_priv *priv)
240 {
241 struct iwl_basic_bt_cmd basic = {
242 .max_kill = IWLAGN_BT_MAX_KILL_DEFAULT,
243 .bt3_timer_t7_value = IWLAGN_BT3_T7_DEFAULT,
244 .bt3_prio_sample_time = IWLAGN_BT3_PRIO_SAMPLE_DEFAULT,
245 .bt3_timer_t2_value = IWLAGN_BT3_T2_DEFAULT,
246 };
247 struct iwl_bt_cmd_v1 bt_cmd_v1;
248 struct iwl_bt_cmd_v2 bt_cmd_v2;
249 int ret;
250
251 BUILD_BUG_ON(sizeof(iwlagn_def_3w_lookup) !=
252 sizeof(basic.bt3_lookup_table));
253
254 if (priv->cfg->bt_params) {
255 /*
256 * newer generation of devices (2000 series and newer)
257 * use the version 2 of the bt command
258 * we need to make sure sending the host command
259 * with correct data structure to avoid uCode assert
260 */
261 if (priv->cfg->bt_params->bt_session_2) {
262 bt_cmd_v2.prio_boost = cpu_to_le32(
263 priv->cfg->bt_params->bt_prio_boost);
264 bt_cmd_v2.tx_prio_boost = 0;
265 bt_cmd_v2.rx_prio_boost = 0;
266 } else {
267 bt_cmd_v1.prio_boost =
268 priv->cfg->bt_params->bt_prio_boost;
269 bt_cmd_v1.tx_prio_boost = 0;
270 bt_cmd_v1.rx_prio_boost = 0;
271 }
272 } else {
273 IWL_ERR(priv, "failed to construct BT Coex Config\n");
274 return;
275 }
276
277 /*
278 * Possible situations when BT needs to take over for receive,
279 * at the same time where STA needs to response to AP's frame(s),
280 * reduce the tx power of the required response frames, by that,
281 * allow the concurrent BT receive & WiFi transmit
282 * (BT - ANT A, WiFi -ANT B), without interference to one another
283 *
284 * Reduced tx power apply to control frames only (ACK/Back/CTS)
285 * when indicated by the BT config command
286 */
287 basic.kill_ack_mask = priv->kill_ack_mask;
288 basic.kill_cts_mask = priv->kill_cts_mask;
289 if (priv->reduced_txpower)
290 basic.reduce_txpower = IWLAGN_BT_REDUCED_TX_PWR;
291 basic.valid = priv->bt_valid;
292
293 /*
294 * Configure BT coex mode to "no coexistence" when the
295 * user disabled BT coexistence, we have no interface
296 * (might be in monitor mode), or the interface is in
297 * IBSS mode (no proper uCode support for coex then).
298 */
299 if (!iwlwifi_mod_params.bt_coex_active ||
300 priv->iw_mode == NL80211_IFTYPE_ADHOC) {
301 basic.flags = IWLAGN_BT_FLAG_COEX_MODE_DISABLED;
302 } else {
303 basic.flags = IWLAGN_BT_FLAG_COEX_MODE_3W <<
304 IWLAGN_BT_FLAG_COEX_MODE_SHIFT;
305
306 if (!priv->bt_enable_pspoll)
307 basic.flags |= IWLAGN_BT_FLAG_SYNC_2_BT_DISABLE;
308 else
309 basic.flags &= ~IWLAGN_BT_FLAG_SYNC_2_BT_DISABLE;
310
311 if (priv->bt_ch_announce)
312 basic.flags |= IWLAGN_BT_FLAG_CHANNEL_INHIBITION;
313 IWL_DEBUG_COEX(priv, "BT coex flag: 0X%x\n", basic.flags);
314 }
315 priv->bt_enable_flag = basic.flags;
316 if (priv->bt_full_concurrent)
317 memcpy(basic.bt3_lookup_table, iwlagn_concurrent_lookup,
318 sizeof(iwlagn_concurrent_lookup));
319 else
320 memcpy(basic.bt3_lookup_table, iwlagn_def_3w_lookup,
321 sizeof(iwlagn_def_3w_lookup));
322
323 IWL_DEBUG_COEX(priv, "BT coex %s in %s mode\n",
324 basic.flags ? "active" : "disabled",
325 priv->bt_full_concurrent ?
326 "full concurrency" : "3-wire");
327
328 if (priv->cfg->bt_params->bt_session_2) {
329 memcpy(&bt_cmd_v2.basic, &basic,
330 sizeof(basic));
331 ret = iwl_dvm_send_cmd_pdu(priv, REPLY_BT_CONFIG,
332 CMD_SYNC, sizeof(bt_cmd_v2), &bt_cmd_v2);
333 } else {
334 memcpy(&bt_cmd_v1.basic, &basic,
335 sizeof(basic));
336 ret = iwl_dvm_send_cmd_pdu(priv, REPLY_BT_CONFIG,
337 CMD_SYNC, sizeof(bt_cmd_v1), &bt_cmd_v1);
338 }
339 if (ret)
340 IWL_ERR(priv, "failed to send BT Coex Config\n");
341
342 }
343
344 void iwlagn_bt_adjust_rssi_monitor(struct iwl_priv *priv, bool rssi_ena)
345 {
346 struct iwl_rxon_context *ctx, *found_ctx = NULL;
347 bool found_ap = false;
348
349 lockdep_assert_held(&priv->mutex);
350
351 /* Check whether AP or GO mode is active. */
352 if (rssi_ena) {
353 for_each_context(priv, ctx) {
354 if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_AP &&
355 iwl_is_associated_ctx(ctx)) {
356 found_ap = true;
357 break;
358 }
359 }
360 }
361
362 /*
363 * If disable was received or If GO/AP mode, disable RSSI
364 * measurements.
365 */
366 if (!rssi_ena || found_ap) {
367 if (priv->cur_rssi_ctx) {
368 ctx = priv->cur_rssi_ctx;
369 ieee80211_disable_rssi_reports(ctx->vif);
370 priv->cur_rssi_ctx = NULL;
371 }
372 return;
373 }
374
375 /*
376 * If rssi measurements need to be enabled, consider all cases now.
377 * Figure out how many contexts are active.
378 */
379 for_each_context(priv, ctx) {
380 if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_STATION &&
381 iwl_is_associated_ctx(ctx)) {
382 found_ctx = ctx;
383 break;
384 }
385 }
386
387 /*
388 * rssi monitor already enabled for the correct interface...nothing
389 * to do.
390 */
391 if (found_ctx == priv->cur_rssi_ctx)
392 return;
393
394 /*
395 * Figure out if rssi monitor is currently enabled, and needs
396 * to be changed. If rssi monitor is already enabled, disable
397 * it first else just enable rssi measurements on the
398 * interface found above.
399 */
400 if (priv->cur_rssi_ctx) {
401 ctx = priv->cur_rssi_ctx;
402 if (ctx->vif)
403 ieee80211_disable_rssi_reports(ctx->vif);
404 }
405
406 priv->cur_rssi_ctx = found_ctx;
407
408 if (!found_ctx)
409 return;
410
411 ieee80211_enable_rssi_reports(found_ctx->vif,
412 IWLAGN_BT_PSP_MIN_RSSI_THRESHOLD,
413 IWLAGN_BT_PSP_MAX_RSSI_THRESHOLD);
414 }
415
416 static bool iwlagn_bt_traffic_is_sco(struct iwl_bt_uart_msg *uart_msg)
417 {
418 return BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3 >>
419 BT_UART_MSG_FRAME3SCOESCO_POS;
420 }
421
422 static void iwlagn_bt_traffic_change_work(struct work_struct *work)
423 {
424 struct iwl_priv *priv =
425 container_of(work, struct iwl_priv, bt_traffic_change_work);
426 struct iwl_rxon_context *ctx;
427 int smps_request = -1;
428
429 if (priv->bt_enable_flag == IWLAGN_BT_FLAG_COEX_MODE_DISABLED) {
430 /* bt coex disabled */
431 return;
432 }
433
434 /*
435 * Note: bt_traffic_load can be overridden by scan complete and
436 * coex profile notifications. Ignore that since only bad consequence
437 * can be not matching debug print with actual state.
438 */
439 IWL_DEBUG_COEX(priv, "BT traffic load changes: %d\n",
440 priv->bt_traffic_load);
441
442 switch (priv->bt_traffic_load) {
443 case IWL_BT_COEX_TRAFFIC_LOAD_NONE:
444 if (priv->bt_status)
445 smps_request = IEEE80211_SMPS_DYNAMIC;
446 else
447 smps_request = IEEE80211_SMPS_AUTOMATIC;
448 break;
449 case IWL_BT_COEX_TRAFFIC_LOAD_LOW:
450 smps_request = IEEE80211_SMPS_DYNAMIC;
451 break;
452 case IWL_BT_COEX_TRAFFIC_LOAD_HIGH:
453 case IWL_BT_COEX_TRAFFIC_LOAD_CONTINUOUS:
454 smps_request = IEEE80211_SMPS_STATIC;
455 break;
456 default:
457 IWL_ERR(priv, "Invalid BT traffic load: %d\n",
458 priv->bt_traffic_load);
459 break;
460 }
461
462 mutex_lock(&priv->mutex);
463
464 /*
465 * We can not send command to firmware while scanning. When the scan
466 * complete we will schedule this work again. We do check with mutex
467 * locked to prevent new scan request to arrive. We do not check
468 * STATUS_SCANNING to avoid race when queue_work two times from
469 * different notifications, but quit and not perform any work at all.
470 */
471 if (test_bit(STATUS_SCAN_HW, &priv->status))
472 goto out;
473
474 iwl_update_chain_flags(priv);
475
476 if (smps_request != -1) {
477 priv->current_ht_config.smps = smps_request;
478 for_each_context(priv, ctx) {
479 if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_STATION)
480 ieee80211_request_smps(ctx->vif, smps_request);
481 }
482 }
483
484 /*
485 * Dynamic PS poll related functionality. Adjust RSSI measurements if
486 * necessary.
487 */
488 iwlagn_bt_coex_rssi_monitor(priv);
489 out:
490 mutex_unlock(&priv->mutex);
491 }
492
493 /*
494 * If BT sco traffic, and RSSI monitor is enabled, move measurements to the
495 * correct interface or disable it if this is the last interface to be
496 * removed.
497 */
498 void iwlagn_bt_coex_rssi_monitor(struct iwl_priv *priv)
499 {
500 if (priv->bt_is_sco &&
501 priv->bt_traffic_load == IWL_BT_COEX_TRAFFIC_LOAD_CONTINUOUS)
502 iwlagn_bt_adjust_rssi_monitor(priv, true);
503 else
504 iwlagn_bt_adjust_rssi_monitor(priv, false);
505 }
506
507 static void iwlagn_print_uartmsg(struct iwl_priv *priv,
508 struct iwl_bt_uart_msg *uart_msg)
509 {
510 IWL_DEBUG_COEX(priv, "Message Type = 0x%X, SSN = 0x%X, "
511 "Update Req = 0x%X\n",
512 (BT_UART_MSG_FRAME1MSGTYPE_MSK & uart_msg->frame1) >>
513 BT_UART_MSG_FRAME1MSGTYPE_POS,
514 (BT_UART_MSG_FRAME1SSN_MSK & uart_msg->frame1) >>
515 BT_UART_MSG_FRAME1SSN_POS,
516 (BT_UART_MSG_FRAME1UPDATEREQ_MSK & uart_msg->frame1) >>
517 BT_UART_MSG_FRAME1UPDATEREQ_POS);
518
519 IWL_DEBUG_COEX(priv, "Open connections = 0x%X, Traffic load = 0x%X, "
520 "Chl_SeqN = 0x%X, In band = 0x%X\n",
521 (BT_UART_MSG_FRAME2OPENCONNECTIONS_MSK & uart_msg->frame2) >>
522 BT_UART_MSG_FRAME2OPENCONNECTIONS_POS,
523 (BT_UART_MSG_FRAME2TRAFFICLOAD_MSK & uart_msg->frame2) >>
524 BT_UART_MSG_FRAME2TRAFFICLOAD_POS,
525 (BT_UART_MSG_FRAME2CHLSEQN_MSK & uart_msg->frame2) >>
526 BT_UART_MSG_FRAME2CHLSEQN_POS,
527 (BT_UART_MSG_FRAME2INBAND_MSK & uart_msg->frame2) >>
528 BT_UART_MSG_FRAME2INBAND_POS);
529
530 IWL_DEBUG_COEX(priv, "SCO/eSCO = 0x%X, Sniff = 0x%X, A2DP = 0x%X, "
531 "ACL = 0x%X, Master = 0x%X, OBEX = 0x%X\n",
532 (BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3) >>
533 BT_UART_MSG_FRAME3SCOESCO_POS,
534 (BT_UART_MSG_FRAME3SNIFF_MSK & uart_msg->frame3) >>
535 BT_UART_MSG_FRAME3SNIFF_POS,
536 (BT_UART_MSG_FRAME3A2DP_MSK & uart_msg->frame3) >>
537 BT_UART_MSG_FRAME3A2DP_POS,
538 (BT_UART_MSG_FRAME3ACL_MSK & uart_msg->frame3) >>
539 BT_UART_MSG_FRAME3ACL_POS,
540 (BT_UART_MSG_FRAME3MASTER_MSK & uart_msg->frame3) >>
541 BT_UART_MSG_FRAME3MASTER_POS,
542 (BT_UART_MSG_FRAME3OBEX_MSK & uart_msg->frame3) >>
543 BT_UART_MSG_FRAME3OBEX_POS);
544
545 IWL_DEBUG_COEX(priv, "Idle duration = 0x%X\n",
546 (BT_UART_MSG_FRAME4IDLEDURATION_MSK & uart_msg->frame4) >>
547 BT_UART_MSG_FRAME4IDLEDURATION_POS);
548
549 IWL_DEBUG_COEX(priv, "Tx Activity = 0x%X, Rx Activity = 0x%X, "
550 "eSCO Retransmissions = 0x%X\n",
551 (BT_UART_MSG_FRAME5TXACTIVITY_MSK & uart_msg->frame5) >>
552 BT_UART_MSG_FRAME5TXACTIVITY_POS,
553 (BT_UART_MSG_FRAME5RXACTIVITY_MSK & uart_msg->frame5) >>
554 BT_UART_MSG_FRAME5RXACTIVITY_POS,
555 (BT_UART_MSG_FRAME5ESCORETRANSMIT_MSK & uart_msg->frame5) >>
556 BT_UART_MSG_FRAME5ESCORETRANSMIT_POS);
557
558 IWL_DEBUG_COEX(priv, "Sniff Interval = 0x%X, Discoverable = 0x%X\n",
559 (BT_UART_MSG_FRAME6SNIFFINTERVAL_MSK & uart_msg->frame6) >>
560 BT_UART_MSG_FRAME6SNIFFINTERVAL_POS,
561 (BT_UART_MSG_FRAME6DISCOVERABLE_MSK & uart_msg->frame6) >>
562 BT_UART_MSG_FRAME6DISCOVERABLE_POS);
563
564 IWL_DEBUG_COEX(priv, "Sniff Activity = 0x%X, Page = "
565 "0x%X, Inquiry = 0x%X, Connectable = 0x%X\n",
566 (BT_UART_MSG_FRAME7SNIFFACTIVITY_MSK & uart_msg->frame7) >>
567 BT_UART_MSG_FRAME7SNIFFACTIVITY_POS,
568 (BT_UART_MSG_FRAME7PAGE_MSK & uart_msg->frame7) >>
569 BT_UART_MSG_FRAME7PAGE_POS,
570 (BT_UART_MSG_FRAME7INQUIRY_MSK & uart_msg->frame7) >>
571 BT_UART_MSG_FRAME7INQUIRY_POS,
572 (BT_UART_MSG_FRAME7CONNECTABLE_MSK & uart_msg->frame7) >>
573 BT_UART_MSG_FRAME7CONNECTABLE_POS);
574 }
575
576 static bool iwlagn_set_kill_msk(struct iwl_priv *priv,
577 struct iwl_bt_uart_msg *uart_msg)
578 {
579 bool need_update = false;
580 u8 kill_msk = IWL_BT_KILL_REDUCE;
581 static const __le32 bt_kill_ack_msg[3] = {
582 IWLAGN_BT_KILL_ACK_MASK_DEFAULT,
583 IWLAGN_BT_KILL_ACK_CTS_MASK_SCO,
584 IWLAGN_BT_KILL_ACK_CTS_MASK_REDUCE};
585 static const __le32 bt_kill_cts_msg[3] = {
586 IWLAGN_BT_KILL_CTS_MASK_DEFAULT,
587 IWLAGN_BT_KILL_ACK_CTS_MASK_SCO,
588 IWLAGN_BT_KILL_ACK_CTS_MASK_REDUCE};
589
590 if (!priv->reduced_txpower)
591 kill_msk = (BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3)
592 ? IWL_BT_KILL_OVERRIDE : IWL_BT_KILL_DEFAULT;
593 if (priv->kill_ack_mask != bt_kill_ack_msg[kill_msk] ||
594 priv->kill_cts_mask != bt_kill_cts_msg[kill_msk]) {
595 priv->bt_valid |= IWLAGN_BT_VALID_KILL_ACK_MASK;
596 priv->kill_ack_mask = bt_kill_ack_msg[kill_msk];
597 priv->bt_valid |= IWLAGN_BT_VALID_KILL_CTS_MASK;
598 priv->kill_cts_mask = bt_kill_cts_msg[kill_msk];
599 need_update = true;
600 }
601 return need_update;
602 }
603
604 /*
605 * Upon RSSI changes, sends a bt config command with following changes
606 * 1. enable/disable "reduced control frames tx power
607 * 2. update the "kill)ack_mask" and "kill_cts_mask"
608 *
609 * If "reduced tx power" is enabled, uCode shall
610 * 1. ACK/Back/CTS rate shall reduced to 6Mbps
611 * 2. not use duplciate 20/40MHz mode
612 */
613 static bool iwlagn_fill_txpower_mode(struct iwl_priv *priv,
614 struct iwl_bt_uart_msg *uart_msg)
615 {
616 bool need_update = false;
617 struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS];
618 int ave_rssi;
619
620 ave_rssi = ieee80211_ave_rssi(ctx->vif);
621 if (!ave_rssi) {
622 /* no rssi data, no changes to reduce tx power */
623 IWL_DEBUG_COEX(priv, "no rssi data available\n");
624 return need_update;
625 }
626 if (!priv->reduced_txpower &&
627 !iwl_is_associated(priv, IWL_RXON_CTX_PAN) &&
628 (ave_rssi > BT_ENABLE_REDUCED_TXPOWER_THRESHOLD) &&
629 (uart_msg->frame3 & (BT_UART_MSG_FRAME3ACL_MSK |
630 BT_UART_MSG_FRAME3OBEX_MSK)) &&
631 !(uart_msg->frame3 & (BT_UART_MSG_FRAME3SCOESCO_MSK |
632 BT_UART_MSG_FRAME3SNIFF_MSK | BT_UART_MSG_FRAME3A2DP_MSK))) {
633 /* enabling reduced tx power */
634 priv->reduced_txpower = true;
635 priv->bt_valid |= IWLAGN_BT_VALID_REDUCED_TX_PWR;
636 need_update = true;
637 } else if (priv->reduced_txpower &&
638 (iwl_is_associated(priv, IWL_RXON_CTX_PAN) ||
639 (ave_rssi < BT_DISABLE_REDUCED_TXPOWER_THRESHOLD) ||
640 (uart_msg->frame3 & (BT_UART_MSG_FRAME3SCOESCO_MSK |
641 BT_UART_MSG_FRAME3SNIFF_MSK | BT_UART_MSG_FRAME3A2DP_MSK)) ||
642 !(uart_msg->frame3 & (BT_UART_MSG_FRAME3ACL_MSK |
643 BT_UART_MSG_FRAME3OBEX_MSK)))) {
644 /* disable reduced tx power */
645 priv->reduced_txpower = false;
646 priv->bt_valid |= IWLAGN_BT_VALID_REDUCED_TX_PWR;
647 need_update = true;
648 }
649
650 return need_update;
651 }
652
653 int iwlagn_bt_coex_profile_notif(struct iwl_priv *priv,
654 struct iwl_rx_cmd_buffer *rxb,
655 struct iwl_device_cmd *cmd)
656 {
657 struct iwl_rx_packet *pkt = rxb_addr(rxb);
658 struct iwl_bt_coex_profile_notif *coex = (void *)pkt->data;
659 struct iwl_bt_uart_msg *uart_msg = &coex->last_bt_uart_msg;
660
661 if (priv->bt_enable_flag == IWLAGN_BT_FLAG_COEX_MODE_DISABLED) {
662 /* bt coex disabled */
663 return 0;
664 }
665
666 IWL_DEBUG_COEX(priv, "BT Coex notification:\n");
667 IWL_DEBUG_COEX(priv, " status: %d\n", coex->bt_status);
668 IWL_DEBUG_COEX(priv, " traffic load: %d\n", coex->bt_traffic_load);
669 IWL_DEBUG_COEX(priv, " CI compliance: %d\n",
670 coex->bt_ci_compliance);
671 iwlagn_print_uartmsg(priv, uart_msg);
672
673 priv->last_bt_traffic_load = priv->bt_traffic_load;
674 priv->bt_is_sco = iwlagn_bt_traffic_is_sco(uart_msg);
675
676 if (priv->iw_mode != NL80211_IFTYPE_ADHOC) {
677 if (priv->bt_status != coex->bt_status ||
678 priv->last_bt_traffic_load != coex->bt_traffic_load) {
679 if (coex->bt_status) {
680 /* BT on */
681 if (!priv->bt_ch_announce)
682 priv->bt_traffic_load =
683 IWL_BT_COEX_TRAFFIC_LOAD_HIGH;
684 else
685 priv->bt_traffic_load =
686 coex->bt_traffic_load;
687 } else {
688 /* BT off */
689 priv->bt_traffic_load =
690 IWL_BT_COEX_TRAFFIC_LOAD_NONE;
691 }
692 priv->bt_status = coex->bt_status;
693 queue_work(priv->workqueue,
694 &priv->bt_traffic_change_work);
695 }
696 }
697
698 /* schedule to send runtime bt_config */
699 /* check reduce power before change ack/cts kill mask */
700 if (iwlagn_fill_txpower_mode(priv, uart_msg) ||
701 iwlagn_set_kill_msk(priv, uart_msg))
702 queue_work(priv->workqueue, &priv->bt_runtime_config);
703
704
705 /* FIXME: based on notification, adjust the prio_boost */
706
707 priv->bt_ci_compliance = coex->bt_ci_compliance;
708 return 0;
709 }
710
711 void iwlagn_bt_rx_handler_setup(struct iwl_priv *priv)
712 {
713 priv->rx_handlers[REPLY_BT_COEX_PROFILE_NOTIF] =
714 iwlagn_bt_coex_profile_notif;
715 }
716
717 void iwlagn_bt_setup_deferred_work(struct iwl_priv *priv)
718 {
719 INIT_WORK(&priv->bt_traffic_change_work,
720 iwlagn_bt_traffic_change_work);
721 }
722
723 void iwlagn_bt_cancel_deferred_work(struct iwl_priv *priv)
724 {
725 cancel_work_sync(&priv->bt_traffic_change_work);
726 }
727
728 static bool is_single_rx_stream(struct iwl_priv *priv)
729 {
730 return priv->current_ht_config.smps == IEEE80211_SMPS_STATIC ||
731 priv->current_ht_config.single_chain_sufficient;
732 }
733
734 #define IWL_NUM_RX_CHAINS_MULTIPLE 3
735 #define IWL_NUM_RX_CHAINS_SINGLE 2
736 #define IWL_NUM_IDLE_CHAINS_DUAL 2
737 #define IWL_NUM_IDLE_CHAINS_SINGLE 1
738
739 /*
740 * Determine how many receiver/antenna chains to use.
741 *
742 * More provides better reception via diversity. Fewer saves power
743 * at the expense of throughput, but only when not in powersave to
744 * start with.
745 *
746 * MIMO (dual stream) requires at least 2, but works better with 3.
747 * This does not determine *which* chains to use, just how many.
748 */
749 static int iwl_get_active_rx_chain_count(struct iwl_priv *priv)
750 {
751 if (priv->cfg->bt_params &&
752 priv->cfg->bt_params->advanced_bt_coexist &&
753 (priv->bt_full_concurrent ||
754 priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH)) {
755 /*
756 * only use chain 'A' in bt high traffic load or
757 * full concurrency mode
758 */
759 return IWL_NUM_RX_CHAINS_SINGLE;
760 }
761 /* # of Rx chains to use when expecting MIMO. */
762 if (is_single_rx_stream(priv))
763 return IWL_NUM_RX_CHAINS_SINGLE;
764 else
765 return IWL_NUM_RX_CHAINS_MULTIPLE;
766 }
767
768 /*
769 * When we are in power saving mode, unless device support spatial
770 * multiplexing power save, use the active count for rx chain count.
771 */
772 static int iwl_get_idle_rx_chain_count(struct iwl_priv *priv, int active_cnt)
773 {
774 /* # Rx chains when idling, depending on SMPS mode */
775 switch (priv->current_ht_config.smps) {
776 case IEEE80211_SMPS_STATIC:
777 case IEEE80211_SMPS_DYNAMIC:
778 return IWL_NUM_IDLE_CHAINS_SINGLE;
779 case IEEE80211_SMPS_AUTOMATIC:
780 case IEEE80211_SMPS_OFF:
781 return active_cnt;
782 default:
783 WARN(1, "invalid SMPS mode %d",
784 priv->current_ht_config.smps);
785 return active_cnt;
786 }
787 }
788
789 /* up to 4 chains */
790 static u8 iwl_count_chain_bitmap(u32 chain_bitmap)
791 {
792 u8 res;
793 res = (chain_bitmap & BIT(0)) >> 0;
794 res += (chain_bitmap & BIT(1)) >> 1;
795 res += (chain_bitmap & BIT(2)) >> 2;
796 res += (chain_bitmap & BIT(3)) >> 3;
797 return res;
798 }
799
800 /**
801 * iwlagn_set_rxon_chain - Set up Rx chain usage in "staging" RXON image
802 *
803 * Selects how many and which Rx receivers/antennas/chains to use.
804 * This should not be used for scan command ... it puts data in wrong place.
805 */
806 void iwlagn_set_rxon_chain(struct iwl_priv *priv, struct iwl_rxon_context *ctx)
807 {
808 bool is_single = is_single_rx_stream(priv);
809 bool is_cam = !test_bit(STATUS_POWER_PMI, &priv->status);
810 u8 idle_rx_cnt, active_rx_cnt, valid_rx_cnt;
811 u32 active_chains;
812 u16 rx_chain;
813
814 /* Tell uCode which antennas are actually connected.
815 * Before first association, we assume all antennas are connected.
816 * Just after first association, iwl_chain_noise_calibration()
817 * checks which antennas actually *are* connected. */
818 if (priv->chain_noise_data.active_chains)
819 active_chains = priv->chain_noise_data.active_chains;
820 else
821 active_chains = priv->hw_params.valid_rx_ant;
822
823 if (priv->cfg->bt_params &&
824 priv->cfg->bt_params->advanced_bt_coexist &&
825 (priv->bt_full_concurrent ||
826 priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH)) {
827 /*
828 * only use chain 'A' in bt high traffic load or
829 * full concurrency mode
830 */
831 active_chains = first_antenna(active_chains);
832 }
833
834 rx_chain = active_chains << RXON_RX_CHAIN_VALID_POS;
835
836 /* How many receivers should we use? */
837 active_rx_cnt = iwl_get_active_rx_chain_count(priv);
838 idle_rx_cnt = iwl_get_idle_rx_chain_count(priv, active_rx_cnt);
839
840
841 /* correct rx chain count according hw settings
842 * and chain noise calibration
843 */
844 valid_rx_cnt = iwl_count_chain_bitmap(active_chains);
845 if (valid_rx_cnt < active_rx_cnt)
846 active_rx_cnt = valid_rx_cnt;
847
848 if (valid_rx_cnt < idle_rx_cnt)
849 idle_rx_cnt = valid_rx_cnt;
850
851 rx_chain |= active_rx_cnt << RXON_RX_CHAIN_MIMO_CNT_POS;
852 rx_chain |= idle_rx_cnt << RXON_RX_CHAIN_CNT_POS;
853
854 ctx->staging.rx_chain = cpu_to_le16(rx_chain);
855
856 if (!is_single && (active_rx_cnt >= IWL_NUM_RX_CHAINS_SINGLE) && is_cam)
857 ctx->staging.rx_chain |= RXON_RX_CHAIN_MIMO_FORCE_MSK;
858 else
859 ctx->staging.rx_chain &= ~RXON_RX_CHAIN_MIMO_FORCE_MSK;
860
861 IWL_DEBUG_ASSOC(priv, "rx_chain=0x%X active=%d idle=%d\n",
862 ctx->staging.rx_chain,
863 active_rx_cnt, idle_rx_cnt);
864
865 WARN_ON(active_rx_cnt == 0 || idle_rx_cnt == 0 ||
866 active_rx_cnt < idle_rx_cnt);
867 }
868
869 u8 iwl_toggle_tx_ant(struct iwl_priv *priv, u8 ant, u8 valid)
870 {
871 int i;
872 u8 ind = ant;
873
874 if (priv->band == IEEE80211_BAND_2GHZ &&
875 priv->bt_traffic_load >= IWL_BT_COEX_TRAFFIC_LOAD_HIGH)
876 return 0;
877
878 for (i = 0; i < RATE_ANT_NUM - 1; i++) {
879 ind = (ind + 1) < RATE_ANT_NUM ? ind + 1 : 0;
880 if (valid & BIT(ind))
881 return ind;
882 }
883 return ant;
884 }
885
886 #ifdef CONFIG_PM_SLEEP
887 static void iwlagn_convert_p1k(u16 *p1k, __le16 *out)
888 {
889 int i;
890
891 for (i = 0; i < IWLAGN_P1K_SIZE; i++)
892 out[i] = cpu_to_le16(p1k[i]);
893 }
894
895 struct wowlan_key_data {
896 struct iwl_rxon_context *ctx;
897 struct iwlagn_wowlan_rsc_tsc_params_cmd *rsc_tsc;
898 struct iwlagn_wowlan_tkip_params_cmd *tkip;
899 const u8 *bssid;
900 bool error, use_rsc_tsc, use_tkip;
901 };
902
903
904 static void iwlagn_wowlan_program_keys(struct ieee80211_hw *hw,
905 struct ieee80211_vif *vif,
906 struct ieee80211_sta *sta,
907 struct ieee80211_key_conf *key,
908 void *_data)
909 {
910 struct iwl_priv *priv = IWL_MAC80211_GET_DVM(hw);
911 struct wowlan_key_data *data = _data;
912 struct iwl_rxon_context *ctx = data->ctx;
913 struct aes_sc *aes_sc, *aes_tx_sc = NULL;
914 struct tkip_sc *tkip_sc, *tkip_tx_sc = NULL;
915 struct iwlagn_p1k_cache *rx_p1ks;
916 u8 *rx_mic_key;
917 struct ieee80211_key_seq seq;
918 u32 cur_rx_iv32 = 0;
919 u16 p1k[IWLAGN_P1K_SIZE];
920 int ret, i;
921
922 mutex_lock(&priv->mutex);
923
924 if ((key->cipher == WLAN_CIPHER_SUITE_WEP40 ||
925 key->cipher == WLAN_CIPHER_SUITE_WEP104) &&
926 !sta && !ctx->key_mapping_keys)
927 ret = iwl_set_default_wep_key(priv, ctx, key);
928 else
929 ret = iwl_set_dynamic_key(priv, ctx, key, sta);
930
931 if (ret) {
932 IWL_ERR(priv, "Error setting key during suspend!\n");
933 data->error = true;
934 }
935
936 switch (key->cipher) {
937 case WLAN_CIPHER_SUITE_TKIP:
938 if (sta) {
939 tkip_sc = data->rsc_tsc->all_tsc_rsc.tkip.unicast_rsc;
940 tkip_tx_sc = &data->rsc_tsc->all_tsc_rsc.tkip.tsc;
941
942 rx_p1ks = data->tkip->rx_uni;
943
944 ieee80211_get_key_tx_seq(key, &seq);
945 tkip_tx_sc->iv16 = cpu_to_le16(seq.tkip.iv16);
946 tkip_tx_sc->iv32 = cpu_to_le32(seq.tkip.iv32);
947
948 ieee80211_get_tkip_p1k_iv(key, seq.tkip.iv32, p1k);
949 iwlagn_convert_p1k(p1k, data->tkip->tx.p1k);
950
951 memcpy(data->tkip->mic_keys.tx,
952 &key->key[NL80211_TKIP_DATA_OFFSET_TX_MIC_KEY],
953 IWLAGN_MIC_KEY_SIZE);
954
955 rx_mic_key = data->tkip->mic_keys.rx_unicast;
956 } else {
957 tkip_sc =
958 data->rsc_tsc->all_tsc_rsc.tkip.multicast_rsc;
959 rx_p1ks = data->tkip->rx_multi;
960 rx_mic_key = data->tkip->mic_keys.rx_mcast;
961 }
962
963 /*
964 * For non-QoS this relies on the fact that both the uCode and
965 * mac80211 use TID 0 (as they need to to avoid replay attacks)
966 * for checking the IV in the frames.
967 */
968 for (i = 0; i < IWLAGN_NUM_RSC; i++) {
969 ieee80211_get_key_rx_seq(key, i, &seq);
970 tkip_sc[i].iv16 = cpu_to_le16(seq.tkip.iv16);
971 tkip_sc[i].iv32 = cpu_to_le32(seq.tkip.iv32);
972 /* wrapping isn't allowed, AP must rekey */
973 if (seq.tkip.iv32 > cur_rx_iv32)
974 cur_rx_iv32 = seq.tkip.iv32;
975 }
976
977 ieee80211_get_tkip_rx_p1k(key, data->bssid, cur_rx_iv32, p1k);
978 iwlagn_convert_p1k(p1k, rx_p1ks[0].p1k);
979 ieee80211_get_tkip_rx_p1k(key, data->bssid,
980 cur_rx_iv32 + 1, p1k);
981 iwlagn_convert_p1k(p1k, rx_p1ks[1].p1k);
982
983 memcpy(rx_mic_key,
984 &key->key[NL80211_TKIP_DATA_OFFSET_RX_MIC_KEY],
985 IWLAGN_MIC_KEY_SIZE);
986
987 data->use_tkip = true;
988 data->use_rsc_tsc = true;
989 break;
990 case WLAN_CIPHER_SUITE_CCMP:
991 if (sta) {
992 u8 *pn = seq.ccmp.pn;
993
994 aes_sc = data->rsc_tsc->all_tsc_rsc.aes.unicast_rsc;
995 aes_tx_sc = &data->rsc_tsc->all_tsc_rsc.aes.tsc;
996
997 ieee80211_get_key_tx_seq(key, &seq);
998 aes_tx_sc->pn = cpu_to_le64(
999 (u64)pn[5] |
1000 ((u64)pn[4] << 8) |
1001 ((u64)pn[3] << 16) |
1002 ((u64)pn[2] << 24) |
1003 ((u64)pn[1] << 32) |
1004 ((u64)pn[0] << 40));
1005 } else
1006 aes_sc = data->rsc_tsc->all_tsc_rsc.aes.multicast_rsc;
1007
1008 /*
1009 * For non-QoS this relies on the fact that both the uCode and
1010 * mac80211 use TID 0 for checking the IV in the frames.
1011 */
1012 for (i = 0; i < IWLAGN_NUM_RSC; i++) {
1013 u8 *pn = seq.ccmp.pn;
1014
1015 ieee80211_get_key_rx_seq(key, i, &seq);
1016 aes_sc->pn = cpu_to_le64(
1017 (u64)pn[5] |
1018 ((u64)pn[4] << 8) |
1019 ((u64)pn[3] << 16) |
1020 ((u64)pn[2] << 24) |
1021 ((u64)pn[1] << 32) |
1022 ((u64)pn[0] << 40));
1023 }
1024 data->use_rsc_tsc = true;
1025 break;
1026 }
1027
1028 mutex_unlock(&priv->mutex);
1029 }
1030
1031 int iwlagn_send_patterns(struct iwl_priv *priv,
1032 struct cfg80211_wowlan *wowlan)
1033 {
1034 struct iwlagn_wowlan_patterns_cmd *pattern_cmd;
1035 struct iwl_host_cmd cmd = {
1036 .id = REPLY_WOWLAN_PATTERNS,
1037 .dataflags[0] = IWL_HCMD_DFL_NOCOPY,
1038 .flags = CMD_SYNC,
1039 };
1040 int i, err;
1041
1042 if (!wowlan->n_patterns)
1043 return 0;
1044
1045 cmd.len[0] = sizeof(*pattern_cmd) +
1046 wowlan->n_patterns * sizeof(struct iwlagn_wowlan_pattern);
1047
1048 pattern_cmd = kmalloc(cmd.len[0], GFP_KERNEL);
1049 if (!pattern_cmd)
1050 return -ENOMEM;
1051
1052 pattern_cmd->n_patterns = cpu_to_le32(wowlan->n_patterns);
1053
1054 for (i = 0; i < wowlan->n_patterns; i++) {
1055 int mask_len = DIV_ROUND_UP(wowlan->patterns[i].pattern_len, 8);
1056
1057 memcpy(&pattern_cmd->patterns[i].mask,
1058 wowlan->patterns[i].mask, mask_len);
1059 memcpy(&pattern_cmd->patterns[i].pattern,
1060 wowlan->patterns[i].pattern,
1061 wowlan->patterns[i].pattern_len);
1062 pattern_cmd->patterns[i].mask_size = mask_len;
1063 pattern_cmd->patterns[i].pattern_size =
1064 wowlan->patterns[i].pattern_len;
1065 }
1066
1067 cmd.data[0] = pattern_cmd;
1068 err = iwl_dvm_send_cmd(priv, &cmd);
1069 kfree(pattern_cmd);
1070 return err;
1071 }
1072
1073 int iwlagn_suspend(struct iwl_priv *priv, struct cfg80211_wowlan *wowlan)
1074 {
1075 struct iwlagn_wowlan_wakeup_filter_cmd wakeup_filter_cmd;
1076 struct iwl_rxon_cmd rxon;
1077 struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS];
1078 struct iwlagn_wowlan_kek_kck_material_cmd kek_kck_cmd;
1079 struct iwlagn_wowlan_tkip_params_cmd tkip_cmd = {};
1080 struct iwlagn_d3_config_cmd d3_cfg_cmd = {};
1081 struct wowlan_key_data key_data = {
1082 .ctx = ctx,
1083 .bssid = ctx->active.bssid_addr,
1084 .use_rsc_tsc = false,
1085 .tkip = &tkip_cmd,
1086 .use_tkip = false,
1087 };
1088 int ret, i;
1089 u16 seq;
1090
1091 key_data.rsc_tsc = kzalloc(sizeof(*key_data.rsc_tsc), GFP_KERNEL);
1092 if (!key_data.rsc_tsc)
1093 return -ENOMEM;
1094
1095 memset(&wakeup_filter_cmd, 0, sizeof(wakeup_filter_cmd));
1096
1097 /*
1098 * We know the last used seqno, and the uCode expects to know that
1099 * one, it will increment before TX.
1100 */
1101 seq = le16_to_cpu(priv->last_seq_ctl) & IEEE80211_SCTL_SEQ;
1102 wakeup_filter_cmd.non_qos_seq = cpu_to_le16(seq);
1103
1104 /*
1105 * For QoS counters, we store the one to use next, so subtract 0x10
1106 * since the uCode will add 0x10 before using the value.
1107 */
1108 for (i = 0; i < IWL_MAX_TID_COUNT; i++) {
1109 seq = priv->tid_data[IWL_AP_ID][i].seq_number;
1110 seq -= 0x10;
1111 wakeup_filter_cmd.qos_seq[i] = cpu_to_le16(seq);
1112 }
1113
1114 if (wowlan->disconnect)
1115 wakeup_filter_cmd.enabled |=
1116 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_BEACON_MISS |
1117 IWLAGN_WOWLAN_WAKEUP_LINK_CHANGE);
1118 if (wowlan->magic_pkt)
1119 wakeup_filter_cmd.enabled |=
1120 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_MAGIC_PACKET);
1121 if (wowlan->gtk_rekey_failure)
1122 wakeup_filter_cmd.enabled |=
1123 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_GTK_REKEY_FAIL);
1124 if (wowlan->eap_identity_req)
1125 wakeup_filter_cmd.enabled |=
1126 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_EAP_IDENT_REQ);
1127 if (wowlan->four_way_handshake)
1128 wakeup_filter_cmd.enabled |=
1129 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_4WAY_HANDSHAKE);
1130 if (wowlan->n_patterns)
1131 wakeup_filter_cmd.enabled |=
1132 cpu_to_le32(IWLAGN_WOWLAN_WAKEUP_PATTERN_MATCH);
1133
1134 if (wowlan->rfkill_release)
1135 d3_cfg_cmd.wakeup_flags |=
1136 cpu_to_le32(IWLAGN_D3_WAKEUP_RFKILL);
1137
1138 iwl_scan_cancel_timeout(priv, 200);
1139
1140 memcpy(&rxon, &ctx->active, sizeof(rxon));
1141
1142 priv->ucode_loaded = false;
1143 iwl_trans_stop_device(priv->trans);
1144
1145 priv->wowlan = true;
1146
1147 ret = iwl_load_ucode_wait_alive(priv, IWL_UCODE_WOWLAN);
1148 if (ret)
1149 goto out;
1150
1151 /* now configure WoWLAN ucode */
1152 ret = iwl_alive_start(priv);
1153 if (ret)
1154 goto out;
1155
1156 memcpy(&ctx->staging, &rxon, sizeof(rxon));
1157 ret = iwlagn_commit_rxon(priv, ctx);
1158 if (ret)
1159 goto out;
1160
1161 ret = iwl_power_update_mode(priv, true);
1162 if (ret)
1163 goto out;
1164
1165 if (!iwlwifi_mod_params.sw_crypto) {
1166 /* mark all keys clear */
1167 priv->ucode_key_table = 0;
1168 ctx->key_mapping_keys = 0;
1169
1170 /*
1171 * This needs to be unlocked due to lock ordering
1172 * constraints. Since we're in the suspend path
1173 * that isn't really a problem though.
1174 */
1175 mutex_unlock(&priv->mutex);
1176 ieee80211_iter_keys(priv->hw, ctx->vif,
1177 iwlagn_wowlan_program_keys,
1178 &key_data);
1179 mutex_lock(&priv->mutex);
1180 if (key_data.error) {
1181 ret = -EIO;
1182 goto out;
1183 }
1184
1185 if (key_data.use_rsc_tsc) {
1186 struct iwl_host_cmd rsc_tsc_cmd = {
1187 .id = REPLY_WOWLAN_TSC_RSC_PARAMS,
1188 .flags = CMD_SYNC,
1189 .data[0] = key_data.rsc_tsc,
1190 .dataflags[0] = IWL_HCMD_DFL_NOCOPY,
1191 .len[0] = sizeof(*key_data.rsc_tsc),
1192 };
1193
1194 ret = iwl_dvm_send_cmd(priv, &rsc_tsc_cmd);
1195 if (ret)
1196 goto out;
1197 }
1198
1199 if (key_data.use_tkip) {
1200 ret = iwl_dvm_send_cmd_pdu(priv,
1201 REPLY_WOWLAN_TKIP_PARAMS,
1202 CMD_SYNC, sizeof(tkip_cmd),
1203 &tkip_cmd);
1204 if (ret)
1205 goto out;
1206 }
1207
1208 if (priv->have_rekey_data) {
1209 memset(&kek_kck_cmd, 0, sizeof(kek_kck_cmd));
1210 memcpy(kek_kck_cmd.kck, priv->kck, NL80211_KCK_LEN);
1211 kek_kck_cmd.kck_len = cpu_to_le16(NL80211_KCK_LEN);
1212 memcpy(kek_kck_cmd.kek, priv->kek, NL80211_KEK_LEN);
1213 kek_kck_cmd.kek_len = cpu_to_le16(NL80211_KEK_LEN);
1214 kek_kck_cmd.replay_ctr = priv->replay_ctr;
1215
1216 ret = iwl_dvm_send_cmd_pdu(priv,
1217 REPLY_WOWLAN_KEK_KCK_MATERIAL,
1218 CMD_SYNC, sizeof(kek_kck_cmd),
1219 &kek_kck_cmd);
1220 if (ret)
1221 goto out;
1222 }
1223 }
1224
1225 ret = iwl_dvm_send_cmd_pdu(priv, REPLY_D3_CONFIG, CMD_SYNC,
1226 sizeof(d3_cfg_cmd), &d3_cfg_cmd);
1227 if (ret)
1228 goto out;
1229
1230 ret = iwl_dvm_send_cmd_pdu(priv, REPLY_WOWLAN_WAKEUP_FILTER,
1231 CMD_SYNC, sizeof(wakeup_filter_cmd),
1232 &wakeup_filter_cmd);
1233 if (ret)
1234 goto out;
1235
1236 ret = iwlagn_send_patterns(priv, wowlan);
1237 out:
1238 kfree(key_data.rsc_tsc);
1239 return ret;
1240 }
1241 #endif
1242
1243 int iwl_dvm_send_cmd(struct iwl_priv *priv, struct iwl_host_cmd *cmd)
1244 {
1245 if (iwl_is_rfkill(priv) || iwl_is_ctkill(priv)) {
1246 IWL_WARN(priv, "Not sending command - %s KILL\n",
1247 iwl_is_rfkill(priv) ? "RF" : "CT");
1248 return -EIO;
1249 }
1250
1251 if (test_bit(STATUS_FW_ERROR, &priv->status)) {
1252 IWL_ERR(priv, "Command %s failed: FW Error\n",
1253 iwl_dvm_get_cmd_string(cmd->id));
1254 return -EIO;
1255 }
1256
1257 /*
1258 * Synchronous commands from this op-mode must hold
1259 * the mutex, this ensures we don't try to send two
1260 * (or more) synchronous commands at a time.
1261 */
1262 if (cmd->flags & CMD_SYNC)
1263 lockdep_assert_held(&priv->mutex);
1264
1265 if (priv->ucode_owner == IWL_OWNERSHIP_TM &&
1266 !(cmd->flags & CMD_ON_DEMAND)) {
1267 IWL_DEBUG_HC(priv, "tm own the uCode, no regular hcmd send\n");
1268 return -EIO;
1269 }
1270
1271 return iwl_trans_send_cmd(priv->trans, cmd);
1272 }
1273
1274 int iwl_dvm_send_cmd_pdu(struct iwl_priv *priv, u8 id,
1275 u32 flags, u16 len, const void *data)
1276 {
1277 struct iwl_host_cmd cmd = {
1278 .id = id,
1279 .len = { len, },
1280 .data = { data, },
1281 .flags = flags,
1282 };
1283
1284 return iwl_dvm_send_cmd(priv, &cmd);
1285 }
This page took 0.058406 seconds and 5 git commands to generate.