1 /******************************************************************************
3 * Copyright(c) 2003 - 2012 Intel Corporation. All rights reserved.
5 * Portions of this file are derived from the ipw3945 project, as well
6 * as portions of the ieee80211 subsystem header files.
8 * This program is free software; you can redistribute it and/or modify it
9 * under the terms of version 2 of the GNU General Public License as
10 * published by the Free Software Foundation.
12 * This program is distributed in the hope that it will be useful, but WITHOUT
13 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
14 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
17 * You should have received a copy of the GNU General Public License along with
18 * this program; if not, write to the Free Software Foundation, Inc.,
19 * 51 Franklin Street, Fifth Floor, Boston, MA 02110, USA
21 * The full GNU General Public License is included in this distribution in the
22 * file called LICENSE.
24 * Contact Information:
25 * Intel Linux Wireless <ilw@linux.intel.com>
26 * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
28 *****************************************************************************/
30 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
32 #include <linux/kernel.h>
33 #include <linux/module.h>
34 #include <linux/init.h>
35 #include <linux/slab.h>
36 #include <linux/delay.h>
37 #include <linux/sched.h>
38 #include <linux/skbuff.h>
39 #include <linux/netdevice.h>
40 #include <linux/etherdevice.h>
41 #include <linux/if_arp.h>
43 #include <net/mac80211.h>
45 #include <asm/div64.h>
47 #include "iwl-eeprom.h"
51 #include "iwl-agn-calib.h"
53 #include "iwl-shared.h"
54 #include "iwl-trans.h"
55 #include "iwl-op-mode.h"
57 /******************************************************************************
61 ******************************************************************************/
64 * module name, copyright, version, etc.
66 #define DRV_DESCRIPTION "Intel(R) Wireless WiFi Link AGN driver for Linux"
68 #ifdef CONFIG_IWLWIFI_DEBUG
74 #define DRV_VERSION IWLWIFI_VERSION VD
77 MODULE_DESCRIPTION(DRV_DESCRIPTION
);
78 MODULE_VERSION(DRV_VERSION
);
79 MODULE_AUTHOR(DRV_COPYRIGHT
" " DRV_AUTHOR
);
80 MODULE_LICENSE("GPL");
81 MODULE_ALIAS("iwlagn");
83 void iwl_update_chain_flags(struct iwl_priv
*priv
)
85 struct iwl_rxon_context
*ctx
;
87 for_each_context(priv
, ctx
) {
88 iwlagn_set_rxon_chain(priv
, ctx
);
89 if (ctx
->active
.rx_chain
!= ctx
->staging
.rx_chain
)
90 iwlagn_commit_rxon(priv
, ctx
);
94 /* Parse the beacon frame to find the TIM element and set tim_idx & tim_size */
95 static void iwl_set_beacon_tim(struct iwl_priv
*priv
,
96 struct iwl_tx_beacon_cmd
*tx_beacon_cmd
,
97 u8
*beacon
, u32 frame_size
)
100 struct ieee80211_mgmt
*mgmt
= (struct ieee80211_mgmt
*)beacon
;
103 * The index is relative to frame start but we start looking at the
104 * variable-length part of the beacon.
106 tim_idx
= mgmt
->u
.beacon
.variable
- beacon
;
108 /* Parse variable-length elements of beacon to find WLAN_EID_TIM */
109 while ((tim_idx
< (frame_size
- 2)) &&
110 (beacon
[tim_idx
] != WLAN_EID_TIM
))
111 tim_idx
+= beacon
[tim_idx
+1] + 2;
113 /* If TIM field was found, set variables */
114 if ((tim_idx
< (frame_size
- 1)) && (beacon
[tim_idx
] == WLAN_EID_TIM
)) {
115 tx_beacon_cmd
->tim_idx
= cpu_to_le16(tim_idx
);
116 tx_beacon_cmd
->tim_size
= beacon
[tim_idx
+1];
118 IWL_WARN(priv
, "Unable to find TIM Element in beacon\n");
121 int iwlagn_send_beacon_cmd(struct iwl_priv
*priv
)
123 struct iwl_tx_beacon_cmd
*tx_beacon_cmd
;
124 struct iwl_host_cmd cmd
= {
125 .id
= REPLY_TX_BEACON
,
128 struct ieee80211_tx_info
*info
;
134 * We have to set up the TX command, the TX Beacon command, and the
138 lockdep_assert_held(&priv
->mutex
);
140 if (!priv
->beacon_ctx
) {
141 IWL_ERR(priv
, "trying to build beacon w/o beacon context!\n");
145 if (WARN_ON(!priv
->beacon_skb
))
148 /* Allocate beacon command */
149 if (!priv
->beacon_cmd
)
150 priv
->beacon_cmd
= kzalloc(sizeof(*tx_beacon_cmd
), GFP_KERNEL
);
151 tx_beacon_cmd
= priv
->beacon_cmd
;
155 frame_size
= priv
->beacon_skb
->len
;
157 /* Set up TX command fields */
158 tx_beacon_cmd
->tx
.len
= cpu_to_le16((u16
)frame_size
);
159 tx_beacon_cmd
->tx
.sta_id
= priv
->beacon_ctx
->bcast_sta_id
;
160 tx_beacon_cmd
->tx
.stop_time
.life_time
= TX_CMD_LIFE_TIME_INFINITE
;
161 tx_beacon_cmd
->tx
.tx_flags
= TX_CMD_FLG_SEQ_CTL_MSK
|
162 TX_CMD_FLG_TSF_MSK
| TX_CMD_FLG_STA_RATE_MSK
;
164 /* Set up TX beacon command fields */
165 iwl_set_beacon_tim(priv
, tx_beacon_cmd
, priv
->beacon_skb
->data
,
168 /* Set up packet rate and flags */
169 info
= IEEE80211_SKB_CB(priv
->beacon_skb
);
172 * Let's set up the rate at least somewhat correctly;
173 * it will currently not actually be used by the uCode,
174 * it uses the broadcast station's rate instead.
176 if (info
->control
.rates
[0].idx
< 0 ||
177 info
->control
.rates
[0].flags
& IEEE80211_TX_RC_MCS
)
180 rate
= info
->control
.rates
[0].idx
;
182 priv
->mgmt_tx_ant
= iwl_toggle_tx_ant(priv
, priv
->mgmt_tx_ant
,
183 hw_params(priv
).valid_tx_ant
);
184 rate_flags
= iwl_ant_idx_to_flags(priv
->mgmt_tx_ant
);
186 /* In mac80211, rates for 5 GHz start at 0 */
187 if (info
->band
== IEEE80211_BAND_5GHZ
)
188 rate
+= IWL_FIRST_OFDM_RATE
;
189 else if (rate
>= IWL_FIRST_CCK_RATE
&& rate
<= IWL_LAST_CCK_RATE
)
190 rate_flags
|= RATE_MCS_CCK_MSK
;
192 tx_beacon_cmd
->tx
.rate_n_flags
=
193 iwl_hw_set_rate_n_flags(rate
, rate_flags
);
196 cmd
.len
[0] = sizeof(*tx_beacon_cmd
);
197 cmd
.data
[0] = tx_beacon_cmd
;
198 cmd
.dataflags
[0] = IWL_HCMD_DFL_NOCOPY
;
199 cmd
.len
[1] = frame_size
;
200 cmd
.data
[1] = priv
->beacon_skb
->data
;
201 cmd
.dataflags
[1] = IWL_HCMD_DFL_NOCOPY
;
203 return iwl_dvm_send_cmd(priv
, &cmd
);
206 static void iwl_bg_beacon_update(struct work_struct
*work
)
208 struct iwl_priv
*priv
=
209 container_of(work
, struct iwl_priv
, beacon_update
);
210 struct sk_buff
*beacon
;
212 mutex_lock(&priv
->mutex
);
213 if (!priv
->beacon_ctx
) {
214 IWL_ERR(priv
, "updating beacon w/o beacon context!\n");
218 if (priv
->beacon_ctx
->vif
->type
!= NL80211_IFTYPE_AP
) {
220 * The ucode will send beacon notifications even in
221 * IBSS mode, but we don't want to process them. But
222 * we need to defer the type check to here due to
223 * requiring locking around the beacon_ctx access.
228 /* Pull updated AP beacon from mac80211. will fail if not in AP mode */
229 beacon
= ieee80211_beacon_get(priv
->hw
, priv
->beacon_ctx
->vif
);
231 IWL_ERR(priv
, "update beacon failed -- keeping old\n");
235 /* new beacon skb is allocated every time; dispose previous.*/
236 dev_kfree_skb(priv
->beacon_skb
);
238 priv
->beacon_skb
= beacon
;
240 iwlagn_send_beacon_cmd(priv
);
242 mutex_unlock(&priv
->mutex
);
245 static void iwl_bg_bt_runtime_config(struct work_struct
*work
)
247 struct iwl_priv
*priv
=
248 container_of(work
, struct iwl_priv
, bt_runtime_config
);
250 if (test_bit(STATUS_EXIT_PENDING
, &priv
->status
))
253 /* dont send host command if rf-kill is on */
254 if (!iwl_is_ready_rf(priv
))
256 iwlagn_send_advance_bt_config(priv
);
259 static void iwl_bg_bt_full_concurrency(struct work_struct
*work
)
261 struct iwl_priv
*priv
=
262 container_of(work
, struct iwl_priv
, bt_full_concurrency
);
263 struct iwl_rxon_context
*ctx
;
265 mutex_lock(&priv
->mutex
);
267 if (test_bit(STATUS_EXIT_PENDING
, &priv
->status
))
270 /* dont send host command if rf-kill is on */
271 if (!iwl_is_ready_rf(priv
))
274 IWL_DEBUG_INFO(priv
, "BT coex in %s mode\n",
275 priv
->bt_full_concurrent
?
276 "full concurrency" : "3-wire");
279 * LQ & RXON updated cmds must be sent before BT Config cmd
280 * to avoid 3-wire collisions
282 for_each_context(priv
, ctx
) {
283 iwlagn_set_rxon_chain(priv
, ctx
);
284 iwlagn_commit_rxon(priv
, ctx
);
287 iwlagn_send_advance_bt_config(priv
);
289 mutex_unlock(&priv
->mutex
);
293 * iwl_bg_statistics_periodic - Timer callback to queue statistics
295 * This callback is provided in order to send a statistics request.
297 * This timer function is continually reset to execute within
298 * REG_RECALIB_PERIOD seconds since the last STATISTICS_NOTIFICATION
299 * was received. We need to ensure we receive the statistics in order
300 * to update the temperature used for calibrating the TXPOWER.
302 static void iwl_bg_statistics_periodic(unsigned long data
)
304 struct iwl_priv
*priv
= (struct iwl_priv
*)data
;
306 if (test_bit(STATUS_EXIT_PENDING
, &priv
->status
))
309 /* dont send host command if rf-kill is on */
310 if (!iwl_is_ready_rf(priv
))
313 iwl_send_statistics_request(priv
, CMD_ASYNC
, false);
317 static void iwl_print_cont_event_trace(struct iwl_priv
*priv
, u32 base
,
318 u32 start_idx
, u32 num_events
,
319 u32 capacity
, u32 mode
)
322 u32 ptr
; /* SRAM byte address of log data */
323 u32 ev
, time
, data
; /* event log data */
324 unsigned long reg_flags
;
327 ptr
= base
+ (4 * sizeof(u32
)) + (start_idx
* 2 * sizeof(u32
));
329 ptr
= base
+ (4 * sizeof(u32
)) + (start_idx
* 3 * sizeof(u32
));
331 /* Make sure device is powered up for SRAM reads */
332 spin_lock_irqsave(&trans(priv
)->reg_lock
, reg_flags
);
333 if (unlikely(!iwl_grab_nic_access(trans(priv
)))) {
334 spin_unlock_irqrestore(&trans(priv
)->reg_lock
, reg_flags
);
338 /* Set starting address; reads will auto-increment */
339 iwl_write32(trans(priv
), HBUS_TARG_MEM_RADDR
, ptr
);
342 * Refuse to read more than would have fit into the log from
343 * the current start_idx. This used to happen due to the race
344 * described below, but now WARN because the code below should
345 * prevent it from happening here.
347 if (WARN_ON(num_events
> capacity
- start_idx
))
348 num_events
= capacity
- start_idx
;
351 * "time" is actually "data" for mode 0 (no timestamp).
352 * place event id # at far right for easier visual parsing.
354 for (i
= 0; i
< num_events
; i
++) {
355 ev
= iwl_read32(trans(priv
), HBUS_TARG_MEM_RDAT
);
356 time
= iwl_read32(trans(priv
), HBUS_TARG_MEM_RDAT
);
358 trace_iwlwifi_dev_ucode_cont_event(
359 trans(priv
)->dev
, 0, time
, ev
);
361 data
= iwl_read32(trans(priv
), HBUS_TARG_MEM_RDAT
);
362 trace_iwlwifi_dev_ucode_cont_event(
363 trans(priv
)->dev
, time
, data
, ev
);
366 /* Allow device to power down */
367 iwl_release_nic_access(trans(priv
));
368 spin_unlock_irqrestore(&trans(priv
)->reg_lock
, reg_flags
);
371 static void iwl_continuous_event_trace(struct iwl_priv
*priv
)
373 u32 capacity
; /* event log capacity in # entries */
380 u32 base
; /* SRAM byte address of event log header */
381 u32 mode
; /* 0 - no timestamp, 1 - timestamp recorded */
382 u32 num_wraps
; /* # times uCode wrapped to top of log */
383 u32 next_entry
; /* index of next entry to be written by uCode */
385 base
= priv
->device_pointers
.log_event_table
;
386 if (iwlagn_hw_valid_rtc_data_addr(base
)) {
387 iwl_read_targ_mem_words(trans(priv
), base
, &read
, sizeof(read
));
389 capacity
= read
.capacity
;
391 num_wraps
= read
.wrap_counter
;
392 next_entry
= read
.write_counter
;
397 * Unfortunately, the uCode doesn't use temporary variables.
398 * Therefore, it can happen that we read next_entry == capacity,
399 * which really means next_entry == 0.
401 if (unlikely(next_entry
== capacity
))
404 * Additionally, the uCode increases the write pointer before
405 * the wraps counter, so if the write pointer is smaller than
406 * the old write pointer (wrap occurred) but we read that no
407 * wrap occurred, we actually read between the next_entry and
408 * num_wraps update (this does happen in practice!!) -- take
409 * that into account by increasing num_wraps.
411 if (unlikely(next_entry
< priv
->event_log
.next_entry
&&
412 num_wraps
== priv
->event_log
.num_wraps
))
415 if (num_wraps
== priv
->event_log
.num_wraps
) {
416 iwl_print_cont_event_trace(
417 priv
, base
, priv
->event_log
.next_entry
,
418 next_entry
- priv
->event_log
.next_entry
,
421 priv
->event_log
.non_wraps_count
++;
423 if (num_wraps
- priv
->event_log
.num_wraps
> 1)
424 priv
->event_log
.wraps_more_count
++;
426 priv
->event_log
.wraps_once_count
++;
428 trace_iwlwifi_dev_ucode_wrap_event(trans(priv
)->dev
,
429 num_wraps
- priv
->event_log
.num_wraps
,
430 next_entry
, priv
->event_log
.next_entry
);
432 if (next_entry
< priv
->event_log
.next_entry
) {
433 iwl_print_cont_event_trace(
434 priv
, base
, priv
->event_log
.next_entry
,
435 capacity
- priv
->event_log
.next_entry
,
438 iwl_print_cont_event_trace(
439 priv
, base
, 0, next_entry
, capacity
, mode
);
441 iwl_print_cont_event_trace(
442 priv
, base
, next_entry
,
443 capacity
- next_entry
,
446 iwl_print_cont_event_trace(
447 priv
, base
, 0, next_entry
, capacity
, mode
);
451 priv
->event_log
.num_wraps
= num_wraps
;
452 priv
->event_log
.next_entry
= next_entry
;
456 * iwl_bg_ucode_trace - Timer callback to log ucode event
458 * The timer is continually set to execute every
459 * UCODE_TRACE_PERIOD milliseconds after the last timer expired
460 * this function is to perform continuous uCode event logging operation
463 static void iwl_bg_ucode_trace(unsigned long data
)
465 struct iwl_priv
*priv
= (struct iwl_priv
*)data
;
467 if (test_bit(STATUS_EXIT_PENDING
, &priv
->status
))
470 if (priv
->event_log
.ucode_trace
) {
471 iwl_continuous_event_trace(priv
);
472 /* Reschedule the timer to occur in UCODE_TRACE_PERIOD */
473 mod_timer(&priv
->ucode_trace
,
474 jiffies
+ msecs_to_jiffies(UCODE_TRACE_PERIOD
));
478 static void iwl_bg_tx_flush(struct work_struct
*work
)
480 struct iwl_priv
*priv
=
481 container_of(work
, struct iwl_priv
, tx_flush
);
483 if (test_bit(STATUS_EXIT_PENDING
, &priv
->status
))
486 /* do nothing if rf-kill is on */
487 if (!iwl_is_ready_rf(priv
))
490 IWL_DEBUG_INFO(priv
, "device request: flush all tx frames\n");
491 iwlagn_dev_txfifo_flush(priv
, IWL_DROP_ALL
);
495 * queue/FIFO/AC mapping definitions
498 #define IWL_TX_FIFO_BK 0 /* shared */
499 #define IWL_TX_FIFO_BE 1
500 #define IWL_TX_FIFO_VI 2 /* shared */
501 #define IWL_TX_FIFO_VO 3
502 #define IWL_TX_FIFO_BK_IPAN IWL_TX_FIFO_BK
503 #define IWL_TX_FIFO_BE_IPAN 4
504 #define IWL_TX_FIFO_VI_IPAN IWL_TX_FIFO_VI
505 #define IWL_TX_FIFO_VO_IPAN 5
506 /* re-uses the VO FIFO, uCode will properly flush/schedule */
507 #define IWL_TX_FIFO_AUX 5
508 #define IWL_TX_FIFO_UNUSED -1
510 #define IWLAGN_CMD_FIFO_NUM 7
513 * This queue number is required for proper operation
514 * because the ucode will stop/start the scheduler as
517 #define IWL_IPAN_MCAST_QUEUE 8
519 static const u8 iwlagn_default_queue_to_tx_fifo
[] = {
527 static const u8 iwlagn_ipan_queue_to_tx_fifo
[] = {
541 static const u8 iwlagn_bss_ac_to_fifo
[] = {
548 static const u8 iwlagn_bss_ac_to_queue
[] = {
552 static const u8 iwlagn_pan_ac_to_fifo
[] = {
559 static const u8 iwlagn_pan_ac_to_queue
[] = {
563 static const u8 iwlagn_bss_queue_to_ac
[] = {
570 static const u8 iwlagn_pan_queue_to_ac
[] = {
581 static void iwl_init_context(struct iwl_priv
*priv
, u32 ucode_flags
)
586 * The default context is always valid,
587 * the PAN context depends on uCode.
589 priv
->valid_contexts
= BIT(IWL_RXON_CTX_BSS
);
590 if (ucode_flags
& IWL_UCODE_TLV_FLAGS_PAN
)
591 priv
->valid_contexts
|= BIT(IWL_RXON_CTX_PAN
);
593 for (i
= 0; i
< NUM_IWL_RXON_CTX
; i
++)
594 priv
->contexts
[i
].ctxid
= i
;
596 priv
->contexts
[IWL_RXON_CTX_BSS
].always_active
= true;
597 priv
->contexts
[IWL_RXON_CTX_BSS
].is_active
= true;
598 priv
->contexts
[IWL_RXON_CTX_BSS
].rxon_cmd
= REPLY_RXON
;
599 priv
->contexts
[IWL_RXON_CTX_BSS
].rxon_timing_cmd
= REPLY_RXON_TIMING
;
600 priv
->contexts
[IWL_RXON_CTX_BSS
].rxon_assoc_cmd
= REPLY_RXON_ASSOC
;
601 priv
->contexts
[IWL_RXON_CTX_BSS
].qos_cmd
= REPLY_QOS_PARAM
;
602 priv
->contexts
[IWL_RXON_CTX_BSS
].ap_sta_id
= IWL_AP_ID
;
603 priv
->contexts
[IWL_RXON_CTX_BSS
].wep_key_cmd
= REPLY_WEPKEY
;
604 priv
->contexts
[IWL_RXON_CTX_BSS
].bcast_sta_id
= IWLAGN_BROADCAST_ID
;
605 priv
->contexts
[IWL_RXON_CTX_BSS
].exclusive_interface_modes
=
606 BIT(NL80211_IFTYPE_ADHOC
);
607 priv
->contexts
[IWL_RXON_CTX_BSS
].interface_modes
=
608 BIT(NL80211_IFTYPE_STATION
);
609 priv
->contexts
[IWL_RXON_CTX_BSS
].ap_devtype
= RXON_DEV_TYPE_AP
;
610 priv
->contexts
[IWL_RXON_CTX_BSS
].ibss_devtype
= RXON_DEV_TYPE_IBSS
;
611 priv
->contexts
[IWL_RXON_CTX_BSS
].station_devtype
= RXON_DEV_TYPE_ESS
;
612 priv
->contexts
[IWL_RXON_CTX_BSS
].unused_devtype
= RXON_DEV_TYPE_ESS
;
613 memcpy(priv
->contexts
[IWL_RXON_CTX_BSS
].ac_to_queue
,
614 iwlagn_bss_ac_to_queue
, sizeof(iwlagn_bss_ac_to_queue
));
615 memcpy(priv
->contexts
[IWL_RXON_CTX_BSS
].ac_to_fifo
,
616 iwlagn_bss_ac_to_fifo
, sizeof(iwlagn_bss_ac_to_fifo
));
618 priv
->contexts
[IWL_RXON_CTX_PAN
].rxon_cmd
= REPLY_WIPAN_RXON
;
619 priv
->contexts
[IWL_RXON_CTX_PAN
].rxon_timing_cmd
=
620 REPLY_WIPAN_RXON_TIMING
;
621 priv
->contexts
[IWL_RXON_CTX_PAN
].rxon_assoc_cmd
=
622 REPLY_WIPAN_RXON_ASSOC
;
623 priv
->contexts
[IWL_RXON_CTX_PAN
].qos_cmd
= REPLY_WIPAN_QOS_PARAM
;
624 priv
->contexts
[IWL_RXON_CTX_PAN
].ap_sta_id
= IWL_AP_ID_PAN
;
625 priv
->contexts
[IWL_RXON_CTX_PAN
].wep_key_cmd
= REPLY_WIPAN_WEPKEY
;
626 priv
->contexts
[IWL_RXON_CTX_PAN
].bcast_sta_id
= IWLAGN_PAN_BCAST_ID
;
627 priv
->contexts
[IWL_RXON_CTX_PAN
].station_flags
= STA_FLG_PAN_STATION
;
628 priv
->contexts
[IWL_RXON_CTX_PAN
].interface_modes
=
629 BIT(NL80211_IFTYPE_STATION
) | BIT(NL80211_IFTYPE_AP
);
631 if (ucode_flags
& IWL_UCODE_TLV_FLAGS_P2P
)
632 priv
->contexts
[IWL_RXON_CTX_PAN
].interface_modes
|=
633 BIT(NL80211_IFTYPE_P2P_CLIENT
) |
634 BIT(NL80211_IFTYPE_P2P_GO
);
636 priv
->contexts
[IWL_RXON_CTX_PAN
].ap_devtype
= RXON_DEV_TYPE_CP
;
637 priv
->contexts
[IWL_RXON_CTX_PAN
].station_devtype
= RXON_DEV_TYPE_2STA
;
638 priv
->contexts
[IWL_RXON_CTX_PAN
].unused_devtype
= RXON_DEV_TYPE_P2P
;
639 memcpy(priv
->contexts
[IWL_RXON_CTX_PAN
].ac_to_queue
,
640 iwlagn_pan_ac_to_queue
, sizeof(iwlagn_pan_ac_to_queue
));
641 memcpy(priv
->contexts
[IWL_RXON_CTX_PAN
].ac_to_fifo
,
642 iwlagn_pan_ac_to_fifo
, sizeof(iwlagn_pan_ac_to_fifo
));
643 priv
->contexts
[IWL_RXON_CTX_PAN
].mcast_queue
= IWL_IPAN_MCAST_QUEUE
;
645 BUILD_BUG_ON(NUM_IWL_RXON_CTX
!= 2);
648 static void iwl_rf_kill_ct_config(struct iwl_priv
*priv
)
650 struct iwl_ct_kill_config cmd
;
651 struct iwl_ct_kill_throttling_config adv_cmd
;
654 iwl_write32(trans(priv
), CSR_UCODE_DRV_GP1_CLR
,
655 CSR_UCODE_DRV_GP1_REG_BIT_CT_KILL_EXIT
);
657 priv
->thermal_throttle
.ct_kill_toggle
= false;
659 if (cfg(priv
)->base_params
->support_ct_kill_exit
) {
660 adv_cmd
.critical_temperature_enter
=
661 cpu_to_le32(hw_params(priv
).ct_kill_threshold
);
662 adv_cmd
.critical_temperature_exit
=
663 cpu_to_le32(hw_params(priv
).ct_kill_exit_threshold
);
665 ret
= iwl_dvm_send_cmd_pdu(priv
,
666 REPLY_CT_KILL_CONFIG_CMD
,
667 CMD_SYNC
, sizeof(adv_cmd
), &adv_cmd
);
669 IWL_ERR(priv
, "REPLY_CT_KILL_CONFIG_CMD failed\n");
671 IWL_DEBUG_INFO(priv
, "REPLY_CT_KILL_CONFIG_CMD "
672 "succeeded, critical temperature enter is %d,"
674 hw_params(priv
).ct_kill_threshold
,
675 hw_params(priv
).ct_kill_exit_threshold
);
677 cmd
.critical_temperature_R
=
678 cpu_to_le32(hw_params(priv
).ct_kill_threshold
);
680 ret
= iwl_dvm_send_cmd_pdu(priv
,
681 REPLY_CT_KILL_CONFIG_CMD
,
682 CMD_SYNC
, sizeof(cmd
), &cmd
);
684 IWL_ERR(priv
, "REPLY_CT_KILL_CONFIG_CMD failed\n");
686 IWL_DEBUG_INFO(priv
, "REPLY_CT_KILL_CONFIG_CMD "
688 "critical temperature is %d\n",
689 hw_params(priv
).ct_kill_threshold
);
693 static int iwlagn_send_calib_cfg_rt(struct iwl_priv
*priv
, u32 cfg
)
695 struct iwl_calib_cfg_cmd calib_cfg_cmd
;
696 struct iwl_host_cmd cmd
= {
697 .id
= CALIBRATION_CFG_CMD
,
698 .len
= { sizeof(struct iwl_calib_cfg_cmd
), },
699 .data
= { &calib_cfg_cmd
, },
702 memset(&calib_cfg_cmd
, 0, sizeof(calib_cfg_cmd
));
703 calib_cfg_cmd
.ucd_calib_cfg
.once
.is_enable
= IWL_CALIB_RT_CFG_ALL
;
704 calib_cfg_cmd
.ucd_calib_cfg
.once
.start
= cpu_to_le32(cfg
);
706 return iwl_dvm_send_cmd(priv
, &cmd
);
710 static int iwlagn_send_tx_ant_config(struct iwl_priv
*priv
, u8 valid_tx_ant
)
712 struct iwl_tx_ant_config_cmd tx_ant_cmd
= {
713 .valid
= cpu_to_le32(valid_tx_ant
),
716 if (IWL_UCODE_API(priv
->fw
->ucode_ver
) > 1) {
717 IWL_DEBUG_HC(priv
, "select valid tx ant: %u\n", valid_tx_ant
);
718 return iwl_dvm_send_cmd_pdu(priv
,
719 TX_ANT_CONFIGURATION_CMD
,
721 sizeof(struct iwl_tx_ant_config_cmd
),
724 IWL_DEBUG_HC(priv
, "TX_ANT_CONFIGURATION_CMD not supported\n");
730 * iwl_alive_start - called after REPLY_ALIVE notification received
731 * from protocol/runtime uCode (initialization uCode's
732 * Alive gets handled by iwl_init_alive_start()).
734 int iwl_alive_start(struct iwl_priv
*priv
)
737 struct iwl_rxon_context
*ctx
= &priv
->contexts
[IWL_RXON_CTX_BSS
];
739 IWL_DEBUG_INFO(priv
, "Runtime Alive received.\n");
741 /* After the ALIVE response, we can send host commands to the uCode */
742 set_bit(STATUS_ALIVE
, &priv
->status
);
744 /* Enable watchdog to monitor the driver tx queues */
745 iwl_setup_watchdog(priv
);
747 if (iwl_is_rfkill(priv
))
750 if (priv
->event_log
.ucode_trace
) {
751 /* start collecting data now */
752 mod_timer(&priv
->ucode_trace
, jiffies
);
755 /* download priority table before any calibration request */
756 if (cfg(priv
)->bt_params
&&
757 cfg(priv
)->bt_params
->advanced_bt_coexist
) {
758 /* Configure Bluetooth device coexistence support */
759 if (cfg(priv
)->bt_params
->bt_sco_disable
)
760 priv
->bt_enable_pspoll
= false;
762 priv
->bt_enable_pspoll
= true;
764 priv
->bt_valid
= IWLAGN_BT_ALL_VALID_MSK
;
765 priv
->kill_ack_mask
= IWLAGN_BT_KILL_ACK_MASK_DEFAULT
;
766 priv
->kill_cts_mask
= IWLAGN_BT_KILL_CTS_MASK_DEFAULT
;
767 iwlagn_send_advance_bt_config(priv
);
768 priv
->bt_valid
= IWLAGN_BT_VALID_ENABLE_FLAGS
;
769 priv
->cur_rssi_ctx
= NULL
;
771 iwl_send_prio_tbl(priv
);
773 /* FIXME: w/a to force change uCode BT state machine */
774 ret
= iwl_send_bt_env(priv
, IWL_BT_COEX_ENV_OPEN
,
775 BT_COEX_PRIO_TBL_EVT_INIT_CALIB2
);
778 ret
= iwl_send_bt_env(priv
, IWL_BT_COEX_ENV_CLOSE
,
779 BT_COEX_PRIO_TBL_EVT_INIT_CALIB2
);
784 * default is 2-wire BT coexexistence support
786 iwl_send_bt_config(priv
);
790 * Perform runtime calibrations, including DC calibration.
792 iwlagn_send_calib_cfg_rt(priv
, IWL_CALIB_CFG_DC_IDX
);
794 ieee80211_wake_queues(priv
->hw
);
796 priv
->active_rate
= IWL_RATES_MASK
;
798 /* Configure Tx antenna selection based on H/W config */
799 iwlagn_send_tx_ant_config(priv
, hw_params(priv
).valid_tx_ant
);
801 if (iwl_is_associated_ctx(ctx
) && !priv
->wowlan
) {
802 struct iwl_rxon_cmd
*active_rxon
=
803 (struct iwl_rxon_cmd
*)&ctx
->active
;
804 /* apply any changes in staging */
805 ctx
->staging
.filter_flags
|= RXON_FILTER_ASSOC_MSK
;
806 active_rxon
->filter_flags
&= ~RXON_FILTER_ASSOC_MSK
;
808 struct iwl_rxon_context
*tmp
;
809 /* Initialize our rx_config data */
810 for_each_context(priv
, tmp
)
811 iwl_connection_init_rx_config(priv
, tmp
);
813 iwlagn_set_rxon_chain(priv
, ctx
);
817 /* WoWLAN ucode will not reply in the same way, skip it */
818 iwl_reset_run_time_calib(priv
);
821 set_bit(STATUS_READY
, &priv
->status
);
823 /* Configure the adapter for unassociated operation */
824 ret
= iwlagn_commit_rxon(priv
, ctx
);
828 /* At this point, the NIC is initialized and operational */
829 iwl_rf_kill_ct_config(priv
);
831 IWL_DEBUG_INFO(priv
, "ALIVE processing complete.\n");
833 return iwl_power_update_mode(priv
, true);
837 * iwl_clear_driver_stations - clear knowledge of all stations from driver
838 * @priv: iwl priv struct
840 * This is called during iwl_down() to make sure that in the case
841 * we're coming there from a hardware restart mac80211 will be
842 * able to reconfigure stations -- if we're getting there in the
843 * normal down flow then the stations will already be cleared.
845 static void iwl_clear_driver_stations(struct iwl_priv
*priv
)
847 struct iwl_rxon_context
*ctx
;
849 spin_lock_bh(&priv
->sta_lock
);
850 memset(priv
->stations
, 0, sizeof(priv
->stations
));
851 priv
->num_stations
= 0;
853 priv
->ucode_key_table
= 0;
855 for_each_context(priv
, ctx
) {
857 * Remove all key information that is not stored as part
858 * of station information since mac80211 may not have had
859 * a chance to remove all the keys. When device is
860 * reconfigured by mac80211 after an error all keys will
863 memset(ctx
->wep_keys
, 0, sizeof(ctx
->wep_keys
));
864 ctx
->key_mapping_keys
= 0;
867 spin_unlock_bh(&priv
->sta_lock
);
870 void iwl_down(struct iwl_priv
*priv
)
874 IWL_DEBUG_INFO(priv
, DRV_NAME
" is going down\n");
876 lockdep_assert_held(&priv
->mutex
);
878 iwl_scan_cancel_timeout(priv
, 200);
881 * If active, scanning won't cancel it, so say it expired.
882 * No race since we hold the mutex here and a new one
883 * can't come in at this time.
885 ieee80211_remain_on_channel_expired(priv
->hw
);
888 test_and_set_bit(STATUS_EXIT_PENDING
, &priv
->status
);
890 /* Stop TX queues watchdog. We need to have STATUS_EXIT_PENDING bit set
891 * to prevent rearm timer */
892 del_timer_sync(&priv
->watchdog
);
894 iwl_clear_ucode_stations(priv
, NULL
);
895 iwl_dealloc_bcast_stations(priv
);
896 iwl_clear_driver_stations(priv
);
898 /* reset BT coex data */
900 priv
->cur_rssi_ctx
= NULL
;
902 if (cfg(priv
)->bt_params
)
903 priv
->bt_traffic_load
=
904 cfg(priv
)->bt_params
->bt_init_traffic_load
;
906 priv
->bt_traffic_load
= 0;
907 priv
->bt_full_concurrent
= false;
908 priv
->bt_ci_compliance
= 0;
910 /* Wipe out the EXIT_PENDING status bit if we are not actually
911 * exiting the module */
913 clear_bit(STATUS_EXIT_PENDING
, &priv
->status
);
915 if (priv
->mac80211_registered
)
916 ieee80211_stop_queues(priv
->hw
);
918 priv
->ucode_loaded
= false;
919 iwl_trans_stop_device(trans(priv
));
921 /* Clear out all status bits but a few that are stable across reset */
922 priv
->status
&= test_bit(STATUS_RF_KILL_HW
, &priv
->status
) <<
924 test_bit(STATUS_GEO_CONFIGURED
, &priv
->status
) <<
925 STATUS_GEO_CONFIGURED
|
926 test_bit(STATUS_FW_ERROR
, &priv
->status
) <<
928 test_bit(STATUS_EXIT_PENDING
, &priv
->status
) <<
931 dev_kfree_skb(priv
->beacon_skb
);
932 priv
->beacon_skb
= NULL
;
935 /*****************************************************************************
937 * Workqueue callbacks
939 *****************************************************************************/
941 static void iwl_bg_run_time_calib_work(struct work_struct
*work
)
943 struct iwl_priv
*priv
= container_of(work
, struct iwl_priv
,
944 run_time_calib_work
);
946 mutex_lock(&priv
->mutex
);
948 if (test_bit(STATUS_EXIT_PENDING
, &priv
->status
) ||
949 test_bit(STATUS_SCANNING
, &priv
->status
)) {
950 mutex_unlock(&priv
->mutex
);
954 if (priv
->start_calib
) {
955 iwl_chain_noise_calibration(priv
);
956 iwl_sensitivity_calibration(priv
);
959 mutex_unlock(&priv
->mutex
);
962 void iwlagn_prepare_restart(struct iwl_priv
*priv
)
964 struct iwl_rxon_context
*ctx
;
965 bool bt_full_concurrent
;
972 lockdep_assert_held(&priv
->mutex
);
974 for_each_context(priv
, ctx
)
979 * __iwl_down() will clear the BT status variables,
980 * which is correct, but when we restart we really
981 * want to keep them so restore them afterwards.
983 * The restart process will later pick them up and
984 * re-configure the hw when we reconfigure the BT
987 bt_full_concurrent
= priv
->bt_full_concurrent
;
988 bt_ci_compliance
= priv
->bt_ci_compliance
;
989 bt_load
= priv
->bt_traffic_load
;
990 bt_status
= priv
->bt_status
;
991 bt_is_sco
= priv
->bt_is_sco
;
995 priv
->bt_full_concurrent
= bt_full_concurrent
;
996 priv
->bt_ci_compliance
= bt_ci_compliance
;
997 priv
->bt_traffic_load
= bt_load
;
998 priv
->bt_status
= bt_status
;
999 priv
->bt_is_sco
= bt_is_sco
;
1001 /* reset all queues */
1002 for (i
= 0; i
< IEEE80211_NUM_ACS
; i
++)
1003 atomic_set(&priv
->ac_stop_count
[i
], 0);
1005 for (i
= IWLAGN_FIRST_AMPDU_QUEUE
; i
< IWL_MAX_HW_QUEUES
; i
++)
1006 priv
->queue_to_ac
[i
] = IWL_INVALID_AC
;
1008 memset(priv
->agg_q_alloc
, 0, sizeof(priv
->agg_q_alloc
));
1011 static void iwl_bg_restart(struct work_struct
*data
)
1013 struct iwl_priv
*priv
= container_of(data
, struct iwl_priv
, restart
);
1015 if (test_bit(STATUS_EXIT_PENDING
, &priv
->status
))
1018 if (test_and_clear_bit(STATUS_FW_ERROR
, &priv
->status
)) {
1019 mutex_lock(&priv
->mutex
);
1020 iwlagn_prepare_restart(priv
);
1021 mutex_unlock(&priv
->mutex
);
1022 iwl_cancel_deferred_work(priv
);
1023 ieee80211_restart_hw(priv
->hw
);
1032 void iwlagn_disable_roc(struct iwl_priv
*priv
)
1034 struct iwl_rxon_context
*ctx
= &priv
->contexts
[IWL_RXON_CTX_PAN
];
1036 lockdep_assert_held(&priv
->mutex
);
1038 if (!priv
->hw_roc_setup
)
1041 ctx
->staging
.dev_type
= RXON_DEV_TYPE_P2P
;
1042 ctx
->staging
.filter_flags
&= ~RXON_FILTER_ASSOC_MSK
;
1044 priv
->hw_roc_channel
= NULL
;
1046 memset(ctx
->staging
.node_addr
, 0, ETH_ALEN
);
1048 iwlagn_commit_rxon(priv
, ctx
);
1050 ctx
->is_active
= false;
1051 priv
->hw_roc_setup
= false;
1054 static void iwlagn_disable_roc_work(struct work_struct
*work
)
1056 struct iwl_priv
*priv
= container_of(work
, struct iwl_priv
,
1057 hw_roc_disable_work
.work
);
1059 mutex_lock(&priv
->mutex
);
1060 iwlagn_disable_roc(priv
);
1061 mutex_unlock(&priv
->mutex
);
1064 /*****************************************************************************
1066 * driver setup and teardown
1068 *****************************************************************************/
1070 static void iwl_setup_deferred_work(struct iwl_priv
*priv
)
1072 priv
->workqueue
= create_singlethread_workqueue(DRV_NAME
);
1074 INIT_WORK(&priv
->restart
, iwl_bg_restart
);
1075 INIT_WORK(&priv
->beacon_update
, iwl_bg_beacon_update
);
1076 INIT_WORK(&priv
->run_time_calib_work
, iwl_bg_run_time_calib_work
);
1077 INIT_WORK(&priv
->tx_flush
, iwl_bg_tx_flush
);
1078 INIT_WORK(&priv
->bt_full_concurrency
, iwl_bg_bt_full_concurrency
);
1079 INIT_WORK(&priv
->bt_runtime_config
, iwl_bg_bt_runtime_config
);
1080 INIT_DELAYED_WORK(&priv
->hw_roc_disable_work
,
1081 iwlagn_disable_roc_work
);
1083 iwl_setup_scan_deferred_work(priv
);
1085 if (cfg(priv
)->bt_params
)
1086 iwlagn_bt_setup_deferred_work(priv
);
1088 init_timer(&priv
->statistics_periodic
);
1089 priv
->statistics_periodic
.data
= (unsigned long)priv
;
1090 priv
->statistics_periodic
.function
= iwl_bg_statistics_periodic
;
1092 init_timer(&priv
->ucode_trace
);
1093 priv
->ucode_trace
.data
= (unsigned long)priv
;
1094 priv
->ucode_trace
.function
= iwl_bg_ucode_trace
;
1096 init_timer(&priv
->watchdog
);
1097 priv
->watchdog
.data
= (unsigned long)priv
;
1098 priv
->watchdog
.function
= iwl_bg_watchdog
;
1101 void iwl_cancel_deferred_work(struct iwl_priv
*priv
)
1103 if (cfg(priv
)->bt_params
)
1104 iwlagn_bt_cancel_deferred_work(priv
);
1106 cancel_work_sync(&priv
->run_time_calib_work
);
1107 cancel_work_sync(&priv
->beacon_update
);
1109 iwl_cancel_scan_deferred_work(priv
);
1111 cancel_work_sync(&priv
->bt_full_concurrency
);
1112 cancel_work_sync(&priv
->bt_runtime_config
);
1113 cancel_delayed_work_sync(&priv
->hw_roc_disable_work
);
1115 del_timer_sync(&priv
->statistics_periodic
);
1116 del_timer_sync(&priv
->ucode_trace
);
1119 static void iwl_init_hw_rates(struct ieee80211_rate
*rates
)
1123 for (i
= 0; i
< IWL_RATE_COUNT_LEGACY
; i
++) {
1124 rates
[i
].bitrate
= iwl_rates
[i
].ieee
* 5;
1125 rates
[i
].hw_value
= i
; /* Rate scaling will work on indexes */
1126 rates
[i
].hw_value_short
= i
;
1128 if ((i
>= IWL_FIRST_CCK_RATE
) && (i
<= IWL_LAST_CCK_RATE
)) {
1130 * If CCK != 1M then set short preamble rate flag.
1133 (iwl_rates
[i
].plcp
== IWL_RATE_1M_PLCP
) ?
1134 0 : IEEE80211_RATE_SHORT_PREAMBLE
;
1139 #define MAX_BIT_RATE_40_MHZ 150 /* Mbps */
1140 #define MAX_BIT_RATE_20_MHZ 72 /* Mbps */
1141 static void iwl_init_ht_hw_capab(const struct iwl_priv
*priv
,
1142 struct ieee80211_sta_ht_cap
*ht_info
,
1143 enum ieee80211_band band
)
1145 u16 max_bit_rate
= 0;
1146 u8 rx_chains_num
= hw_params(priv
).rx_chains_num
;
1147 u8 tx_chains_num
= hw_params(priv
).tx_chains_num
;
1150 memset(&ht_info
->mcs
, 0, sizeof(ht_info
->mcs
));
1152 ht_info
->ht_supported
= true;
1154 if (cfg(priv
)->ht_params
&&
1155 cfg(priv
)->ht_params
->ht_greenfield_support
)
1156 ht_info
->cap
|= IEEE80211_HT_CAP_GRN_FLD
;
1157 ht_info
->cap
|= IEEE80211_HT_CAP_SGI_20
;
1158 max_bit_rate
= MAX_BIT_RATE_20_MHZ
;
1159 if (hw_params(priv
).ht40_channel
& BIT(band
)) {
1160 ht_info
->cap
|= IEEE80211_HT_CAP_SUP_WIDTH_20_40
;
1161 ht_info
->cap
|= IEEE80211_HT_CAP_SGI_40
;
1162 ht_info
->mcs
.rx_mask
[4] = 0x01;
1163 max_bit_rate
= MAX_BIT_RATE_40_MHZ
;
1166 if (iwlagn_mod_params
.amsdu_size_8K
)
1167 ht_info
->cap
|= IEEE80211_HT_CAP_MAX_AMSDU
;
1169 ht_info
->ampdu_factor
= CFG_HT_RX_AMPDU_FACTOR_DEF
;
1170 ht_info
->ampdu_density
= CFG_HT_MPDU_DENSITY_DEF
;
1172 ht_info
->mcs
.rx_mask
[0] = 0xFF;
1173 if (rx_chains_num
>= 2)
1174 ht_info
->mcs
.rx_mask
[1] = 0xFF;
1175 if (rx_chains_num
>= 3)
1176 ht_info
->mcs
.rx_mask
[2] = 0xFF;
1178 /* Highest supported Rx data rate */
1179 max_bit_rate
*= rx_chains_num
;
1180 WARN_ON(max_bit_rate
& ~IEEE80211_HT_MCS_RX_HIGHEST_MASK
);
1181 ht_info
->mcs
.rx_highest
= cpu_to_le16(max_bit_rate
);
1183 /* Tx MCS capabilities */
1184 ht_info
->mcs
.tx_params
= IEEE80211_HT_MCS_TX_DEFINED
;
1185 if (tx_chains_num
!= rx_chains_num
) {
1186 ht_info
->mcs
.tx_params
|= IEEE80211_HT_MCS_TX_RX_DIFF
;
1187 ht_info
->mcs
.tx_params
|= ((tx_chains_num
- 1) <<
1188 IEEE80211_HT_MCS_TX_MAX_STREAMS_SHIFT
);
1193 * iwl_init_geos - Initialize mac80211's geo/channel info based from eeprom
1195 static int iwl_init_geos(struct iwl_priv
*priv
)
1197 struct iwl_channel_info
*ch
;
1198 struct ieee80211_supported_band
*sband
;
1199 struct ieee80211_channel
*channels
;
1200 struct ieee80211_channel
*geo_ch
;
1201 struct ieee80211_rate
*rates
;
1203 s8 max_tx_power
= IWLAGN_TX_POWER_TARGET_POWER_MIN
;
1205 if (priv
->bands
[IEEE80211_BAND_2GHZ
].n_bitrates
||
1206 priv
->bands
[IEEE80211_BAND_5GHZ
].n_bitrates
) {
1207 IWL_DEBUG_INFO(priv
, "Geography modes already initialized.\n");
1208 set_bit(STATUS_GEO_CONFIGURED
, &priv
->status
);
1212 channels
= kcalloc(priv
->channel_count
,
1213 sizeof(struct ieee80211_channel
), GFP_KERNEL
);
1217 rates
= kcalloc(IWL_RATE_COUNT_LEGACY
, sizeof(struct ieee80211_rate
),
1224 /* 5.2GHz channels start after the 2.4GHz channels */
1225 sband
= &priv
->bands
[IEEE80211_BAND_5GHZ
];
1226 sband
->channels
= &channels
[ARRAY_SIZE(iwl_eeprom_band_1
)];
1228 sband
->bitrates
= &rates
[IWL_FIRST_OFDM_RATE
];
1229 sband
->n_bitrates
= IWL_RATE_COUNT_LEGACY
- IWL_FIRST_OFDM_RATE
;
1231 if (hw_params(priv
).sku
& EEPROM_SKU_CAP_11N_ENABLE
)
1232 iwl_init_ht_hw_capab(priv
, &sband
->ht_cap
,
1233 IEEE80211_BAND_5GHZ
);
1235 sband
= &priv
->bands
[IEEE80211_BAND_2GHZ
];
1236 sband
->channels
= channels
;
1238 sband
->bitrates
= rates
;
1239 sband
->n_bitrates
= IWL_RATE_COUNT_LEGACY
;
1241 if (hw_params(priv
).sku
& EEPROM_SKU_CAP_11N_ENABLE
)
1242 iwl_init_ht_hw_capab(priv
, &sband
->ht_cap
,
1243 IEEE80211_BAND_2GHZ
);
1245 priv
->ieee_channels
= channels
;
1246 priv
->ieee_rates
= rates
;
1248 for (i
= 0; i
< priv
->channel_count
; i
++) {
1249 ch
= &priv
->channel_info
[i
];
1251 /* FIXME: might be removed if scan is OK */
1252 if (!is_channel_valid(ch
))
1255 sband
= &priv
->bands
[ch
->band
];
1257 geo_ch
= &sband
->channels
[sband
->n_channels
++];
1259 geo_ch
->center_freq
=
1260 ieee80211_channel_to_frequency(ch
->channel
, ch
->band
);
1261 geo_ch
->max_power
= ch
->max_power_avg
;
1262 geo_ch
->max_antenna_gain
= 0xff;
1263 geo_ch
->hw_value
= ch
->channel
;
1265 if (is_channel_valid(ch
)) {
1266 if (!(ch
->flags
& EEPROM_CHANNEL_IBSS
))
1267 geo_ch
->flags
|= IEEE80211_CHAN_NO_IBSS
;
1269 if (!(ch
->flags
& EEPROM_CHANNEL_ACTIVE
))
1270 geo_ch
->flags
|= IEEE80211_CHAN_PASSIVE_SCAN
;
1272 if (ch
->flags
& EEPROM_CHANNEL_RADAR
)
1273 geo_ch
->flags
|= IEEE80211_CHAN_RADAR
;
1275 geo_ch
->flags
|= ch
->ht40_extension_channel
;
1277 if (ch
->max_power_avg
> max_tx_power
)
1278 max_tx_power
= ch
->max_power_avg
;
1280 geo_ch
->flags
|= IEEE80211_CHAN_DISABLED
;
1283 IWL_DEBUG_INFO(priv
, "Channel %d Freq=%d[%sGHz] %s flag=0x%X\n",
1284 ch
->channel
, geo_ch
->center_freq
,
1285 is_channel_a_band(ch
) ? "5.2" : "2.4",
1286 geo_ch
->flags
& IEEE80211_CHAN_DISABLED
?
1287 "restricted" : "valid",
1291 priv
->tx_power_device_lmt
= max_tx_power
;
1292 priv
->tx_power_user_lmt
= max_tx_power
;
1293 priv
->tx_power_next
= max_tx_power
;
1295 if ((priv
->bands
[IEEE80211_BAND_5GHZ
].n_channels
== 0) &&
1296 hw_params(priv
).sku
& EEPROM_SKU_CAP_BAND_52GHZ
) {
1297 IWL_INFO(priv
, "Incorrectly detected BG card as ABG. "
1298 "Please send your %s to maintainer.\n",
1299 trans(priv
)->hw_id_str
);
1300 hw_params(priv
).sku
&= ~EEPROM_SKU_CAP_BAND_52GHZ
;
1303 IWL_INFO(priv
, "Tunable channels: %d 802.11bg, %d 802.11a channels\n",
1304 priv
->bands
[IEEE80211_BAND_2GHZ
].n_channels
,
1305 priv
->bands
[IEEE80211_BAND_5GHZ
].n_channels
);
1307 set_bit(STATUS_GEO_CONFIGURED
, &priv
->status
);
1313 * iwl_free_geos - undo allocations in iwl_init_geos
1315 static void iwl_free_geos(struct iwl_priv
*priv
)
1317 kfree(priv
->ieee_channels
);
1318 kfree(priv
->ieee_rates
);
1319 clear_bit(STATUS_GEO_CONFIGURED
, &priv
->status
);
1322 static int iwl_init_drv(struct iwl_priv
*priv
)
1326 spin_lock_init(&priv
->sta_lock
);
1328 mutex_init(&priv
->mutex
);
1330 INIT_LIST_HEAD(&priv
->calib_results
);
1332 priv
->ieee_channels
= NULL
;
1333 priv
->ieee_rates
= NULL
;
1334 priv
->band
= IEEE80211_BAND_2GHZ
;
1336 priv
->plcp_delta_threshold
=
1337 cfg(priv
)->base_params
->plcp_delta_threshold
;
1339 priv
->iw_mode
= NL80211_IFTYPE_STATION
;
1340 priv
->current_ht_config
.smps
= IEEE80211_SMPS_STATIC
;
1341 priv
->missed_beacon_threshold
= IWL_MISSED_BEACON_THRESHOLD_DEF
;
1342 priv
->agg_tids_count
= 0;
1344 priv
->ucode_owner
= IWL_OWNERSHIP_DRIVER
;
1346 /* initialize force reset */
1347 priv
->force_reset
[IWL_RF_RESET
].reset_duration
=
1348 IWL_DELAY_NEXT_FORCE_RF_RESET
;
1349 priv
->force_reset
[IWL_FW_RESET
].reset_duration
=
1350 IWL_DELAY_NEXT_FORCE_FW_RELOAD
;
1352 priv
->rx_statistics_jiffies
= jiffies
;
1354 /* Choose which receivers/antennas to use */
1355 iwlagn_set_rxon_chain(priv
, &priv
->contexts
[IWL_RXON_CTX_BSS
]);
1357 iwl_init_scan_params(priv
);
1360 if (cfg(priv
)->bt_params
&&
1361 cfg(priv
)->bt_params
->advanced_bt_coexist
) {
1362 priv
->kill_ack_mask
= IWLAGN_BT_KILL_ACK_MASK_DEFAULT
;
1363 priv
->kill_cts_mask
= IWLAGN_BT_KILL_CTS_MASK_DEFAULT
;
1364 priv
->bt_valid
= IWLAGN_BT_ALL_VALID_MSK
;
1365 priv
->bt_on_thresh
= BT_ON_THRESHOLD_DEF
;
1366 priv
->bt_duration
= BT_DURATION_LIMIT_DEF
;
1367 priv
->dynamic_frag_thresh
= BT_FRAG_THRESHOLD_DEF
;
1370 ret
= iwl_init_channel_map(priv
);
1372 IWL_ERR(priv
, "initializing regulatory failed: %d\n", ret
);
1376 ret
= iwl_init_geos(priv
);
1378 IWL_ERR(priv
, "initializing geos failed: %d\n", ret
);
1379 goto err_free_channel_map
;
1381 iwl_init_hw_rates(priv
->ieee_rates
);
1385 err_free_channel_map
:
1386 iwl_free_channel_map(priv
);
1391 static void iwl_uninit_drv(struct iwl_priv
*priv
)
1393 iwl_free_geos(priv
);
1394 iwl_free_channel_map(priv
);
1395 kfree(priv
->scan_cmd
);
1396 kfree(priv
->beacon_cmd
);
1397 kfree(rcu_dereference_raw(priv
->noa_data
));
1398 iwl_calib_free_results(priv
);
1399 #ifdef CONFIG_IWLWIFI_DEBUGFS
1400 kfree(priv
->wowlan_sram
);
1404 /* Size of one Rx buffer in host DRAM */
1405 #define IWL_RX_BUF_SIZE_4K (4 * 1024)
1406 #define IWL_RX_BUF_SIZE_8K (8 * 1024)
1408 static void iwl_set_hw_params(struct iwl_priv
*priv
)
1410 if (cfg(priv
)->ht_params
)
1411 hw_params(priv
).use_rts_for_aggregation
=
1412 cfg(priv
)->ht_params
->use_rts_for_aggregation
;
1414 if (iwlagn_mod_params
.amsdu_size_8K
)
1415 hw_params(priv
).rx_page_order
=
1416 get_order(IWL_RX_BUF_SIZE_8K
);
1418 hw_params(priv
).rx_page_order
=
1419 get_order(IWL_RX_BUF_SIZE_4K
);
1421 if (iwlagn_mod_params
.disable_11n
& IWL_DISABLE_HT_ALL
)
1422 hw_params(priv
).sku
&= ~EEPROM_SKU_CAP_11N_ENABLE
;
1424 hw_params(priv
).wd_timeout
= cfg(priv
)->base_params
->wd_timeout
;
1426 /* Device-specific setup */
1427 cfg(priv
)->lib
->set_hw_params(priv
);
1432 static void iwl_debug_config(struct iwl_priv
*priv
)
1434 dev_printk(KERN_INFO
, trans(priv
)->dev
, "CONFIG_IWLWIFI_DEBUG "
1435 #ifdef CONFIG_IWLWIFI_DEBUG
1440 dev_printk(KERN_INFO
, trans(priv
)->dev
, "CONFIG_IWLWIFI_DEBUGFS "
1441 #ifdef CONFIG_IWLWIFI_DEBUGFS
1446 dev_printk(KERN_INFO
, trans(priv
)->dev
, "CONFIG_IWLWIFI_DEVICE_TRACING "
1447 #ifdef CONFIG_IWLWIFI_DEVICE_TRACING
1453 dev_printk(KERN_INFO
, trans(priv
)->dev
, "CONFIG_IWLWIFI_DEVICE_TESTMODE "
1454 #ifdef CONFIG_IWLWIFI_DEVICE_TESTMODE
1459 dev_printk(KERN_INFO
, trans(priv
)->dev
, "CONFIG_IWLWIFI_P2P "
1460 #ifdef CONFIG_IWLWIFI_P2P
1467 static struct iwl_op_mode
*iwl_op_mode_dvm_start(struct iwl_trans
*trans
,
1468 const struct iwl_fw
*fw
)
1470 struct iwl_priv
*priv
;
1471 struct ieee80211_hw
*hw
;
1472 struct iwl_op_mode
*op_mode
;
1475 struct iwl_trans_config trans_cfg
;
1476 static const u8 no_reclaim_cmds
[] = {
1480 REPLY_COMPRESSED_BA
,
1481 STATISTICS_NOTIFICATION
,
1488 /************************
1489 * 1. Allocating HW data
1490 ************************/
1491 hw
= iwl_alloc_all();
1493 pr_err("%s: Cannot allocate network device\n",
1499 op_mode
->ops
= &iwl_dvm_ops
;
1500 priv
= IWL_OP_MODE_GET_DVM(op_mode
);
1501 priv
->shrd
= trans
->shrd
;
1505 * Populate the state variables that the transport layer needs
1508 trans_cfg
.op_mode
= op_mode
;
1509 trans_cfg
.no_reclaim_cmds
= no_reclaim_cmds
;
1510 trans_cfg
.n_no_reclaim_cmds
= ARRAY_SIZE(no_reclaim_cmds
);
1512 ucode_flags
= fw
->ucode_capa
.flags
;
1514 #ifndef CONFIG_IWLWIFI_P2P
1515 ucode_flags
&= ~IWL_UCODE_TLV_FLAGS_PAN
;
1518 if (ucode_flags
& IWL_UCODE_TLV_FLAGS_PAN
) {
1519 priv
->sta_key_max_num
= STA_KEY_MAX_NUM_PAN
;
1520 trans_cfg
.cmd_queue
= IWL_IPAN_CMD_QUEUE_NUM
;
1521 trans_cfg
.queue_to_fifo
= iwlagn_ipan_queue_to_tx_fifo
;
1522 trans_cfg
.n_queue_to_fifo
=
1523 ARRAY_SIZE(iwlagn_ipan_queue_to_tx_fifo
);
1524 q_to_ac
= iwlagn_pan_queue_to_ac
;
1525 n_q_to_ac
= ARRAY_SIZE(iwlagn_pan_queue_to_ac
);
1527 priv
->sta_key_max_num
= STA_KEY_MAX_NUM
;
1528 trans_cfg
.cmd_queue
= IWL_DEFAULT_CMD_QUEUE_NUM
;
1529 trans_cfg
.queue_to_fifo
= iwlagn_default_queue_to_tx_fifo
;
1530 trans_cfg
.n_queue_to_fifo
=
1531 ARRAY_SIZE(iwlagn_default_queue_to_tx_fifo
);
1532 q_to_ac
= iwlagn_bss_queue_to_ac
;
1533 n_q_to_ac
= ARRAY_SIZE(iwlagn_bss_queue_to_ac
);
1536 /* Configure transport layer */
1537 iwl_trans_configure(trans(priv
), &trans_cfg
);
1539 /* At this point both hw and priv are allocated. */
1541 SET_IEEE80211_DEV(priv
->hw
, trans(priv
)->dev
);
1543 /* show what debugging capabilities we have */
1544 iwl_debug_config(priv
);
1546 IWL_DEBUG_INFO(priv
, "*** LOAD DRIVER ***\n");
1548 /* is antenna coupling more than 35dB ? */
1549 priv
->bt_ant_couple_ok
=
1550 (iwlagn_mod_params
.ant_coupling
>
1551 IWL_BT_ANTENNA_COUPLING_THRESHOLD
) ?
1554 /* enable/disable bt channel inhibition */
1555 priv
->bt_ch_announce
= iwlagn_mod_params
.bt_ch_announce
;
1556 IWL_DEBUG_INFO(priv
, "BT channel inhibition is %s\n",
1557 (priv
->bt_ch_announce
) ? "On" : "Off");
1559 if (iwl_alloc_traffic_mem(priv
))
1560 IWL_ERR(priv
, "Not enough memory to generate traffic log\n");
1562 /* these spin locks will be used in apm_ops.init and EEPROM access
1563 * we should init now
1565 spin_lock_init(&trans(priv
)->reg_lock
);
1566 spin_lock_init(&priv
->statistics
.lock
);
1568 /***********************
1569 * 2. Read REV register
1570 ***********************/
1571 IWL_INFO(priv
, "Detected %s, REV=0x%X\n",
1572 cfg(priv
)->name
, trans(priv
)->hw_rev
);
1574 if (iwl_trans_start_hw(trans(priv
)))
1575 goto out_free_traffic_mem
;
1580 /* Read the EEPROM */
1581 if (iwl_eeprom_init(trans(priv
), trans(priv
)->hw_rev
)) {
1582 IWL_ERR(priv
, "Unable to init EEPROM\n");
1583 goto out_free_traffic_mem
;
1585 /* Reset chip to save power until we load uCode during "up". */
1586 iwl_trans_stop_hw(trans(priv
));
1588 if (iwl_eeprom_check_version(priv
))
1589 goto out_free_eeprom
;
1591 if (iwl_eeprom_init_hw_params(priv
))
1592 goto out_free_eeprom
;
1594 /* extract MAC Address */
1595 iwl_eeprom_get_mac(priv
->shrd
, priv
->addresses
[0].addr
);
1596 IWL_DEBUG_INFO(priv
, "MAC address: %pM\n", priv
->addresses
[0].addr
);
1597 priv
->hw
->wiphy
->addresses
= priv
->addresses
;
1598 priv
->hw
->wiphy
->n_addresses
= 1;
1599 num_mac
= iwl_eeprom_query16(priv
->shrd
, EEPROM_NUM_MAC_ADDRESS
);
1601 memcpy(priv
->addresses
[1].addr
, priv
->addresses
[0].addr
,
1603 priv
->addresses
[1].addr
[5]++;
1604 priv
->hw
->wiphy
->n_addresses
++;
1607 /************************
1608 * 4. Setup HW constants
1609 ************************/
1610 iwl_set_hw_params(priv
);
1612 if (!(hw_params(priv
).sku
& EEPROM_SKU_CAP_IPAN_ENABLE
)) {
1613 IWL_DEBUG_INFO(priv
, "Your EEPROM disabled PAN");
1614 ucode_flags
&= ~IWL_UCODE_TLV_FLAGS_PAN
;
1616 * if not PAN, then don't support P2P -- might be a uCode
1617 * packaging bug or due to the eeprom check above
1619 ucode_flags
&= ~IWL_UCODE_TLV_FLAGS_P2P
;
1620 priv
->sta_key_max_num
= STA_KEY_MAX_NUM
;
1621 trans_cfg
.cmd_queue
= IWL_DEFAULT_CMD_QUEUE_NUM
;
1622 trans_cfg
.queue_to_fifo
= iwlagn_default_queue_to_tx_fifo
;
1623 trans_cfg
.n_queue_to_fifo
=
1624 ARRAY_SIZE(iwlagn_default_queue_to_tx_fifo
);
1625 q_to_ac
= iwlagn_bss_queue_to_ac
;
1626 n_q_to_ac
= ARRAY_SIZE(iwlagn_bss_queue_to_ac
);
1628 /* Configure transport layer again*/
1629 iwl_trans_configure(trans(priv
), &trans_cfg
);
1632 /*******************
1634 *******************/
1635 for (i
= 0; i
< IEEE80211_NUM_ACS
; i
++)
1636 atomic_set(&priv
->ac_stop_count
[i
], 0);
1638 for (i
= 0; i
< IWL_MAX_HW_QUEUES
; i
++) {
1640 priv
->queue_to_ac
[i
] = q_to_ac
[i
];
1642 priv
->queue_to_ac
[i
] = IWL_INVALID_AC
;
1645 WARN_ON(trans_cfg
.queue_to_fifo
[trans_cfg
.cmd_queue
] !=
1646 IWLAGN_CMD_FIFO_NUM
);
1648 if (iwl_init_drv(priv
))
1649 goto out_free_eeprom
;
1651 /* At this point both hw and priv are initialized. */
1653 /********************
1655 ********************/
1656 iwl_setup_deferred_work(priv
);
1657 iwl_setup_rx_handlers(priv
);
1658 iwl_testmode_init(priv
);
1660 iwl_power_initialize(priv
);
1661 iwl_tt_initialize(priv
);
1663 snprintf(priv
->hw
->wiphy
->fw_version
,
1664 sizeof(priv
->hw
->wiphy
->fw_version
),
1665 "%s", fw
->fw_version
);
1667 priv
->new_scan_threshold_behaviour
=
1668 !!(ucode_flags
& IWL_UCODE_TLV_FLAGS_NEWSCAN
);
1670 priv
->phy_calib_chain_noise_reset_cmd
=
1671 fw
->ucode_capa
.standard_phy_calibration_size
;
1672 priv
->phy_calib_chain_noise_gain_cmd
=
1673 fw
->ucode_capa
.standard_phy_calibration_size
+ 1;
1675 /* initialize all valid contexts */
1676 iwl_init_context(priv
, ucode_flags
);
1678 /**************************************************
1679 * This is still part of probe() in a sense...
1681 * 7. Setup and register with mac80211 and debugfs
1682 **************************************************/
1683 if (iwlagn_mac_setup_register(priv
, &fw
->ucode_capa
))
1684 goto out_destroy_workqueue
;
1686 if (iwl_dbgfs_register(priv
, DRV_NAME
))
1688 "failed to create debugfs files. Ignoring error\n");
1692 out_destroy_workqueue
:
1693 destroy_workqueue(priv
->workqueue
);
1694 priv
->workqueue
= NULL
;
1695 iwl_uninit_drv(priv
);
1697 iwl_eeprom_free(priv
->shrd
);
1698 out_free_traffic_mem
:
1699 iwl_free_traffic_mem(priv
);
1700 ieee80211_free_hw(priv
->hw
);
1706 static void iwl_op_mode_dvm_stop(struct iwl_op_mode
*op_mode
)
1708 struct iwl_priv
*priv
= IWL_OP_MODE_GET_DVM(op_mode
);
1710 IWL_DEBUG_INFO(priv
, "*** UNLOAD DRIVER ***\n");
1712 iwl_dbgfs_unregister(priv
);
1714 iwl_testmode_cleanup(priv
);
1715 iwlagn_mac_unregister(priv
);
1719 /*This will stop the queues, move the device to low power state */
1720 priv
->ucode_loaded
= false;
1721 iwl_trans_stop_device(trans(priv
));
1723 iwl_eeprom_free(priv
->shrd
);
1725 /*netif_stop_queue(dev); */
1726 flush_workqueue(priv
->workqueue
);
1728 /* ieee80211_unregister_hw calls iwlagn_mac_stop, which flushes
1729 * priv->workqueue... so we can't take down the workqueue
1731 destroy_workqueue(priv
->workqueue
);
1732 priv
->workqueue
= NULL
;
1733 iwl_free_traffic_mem(priv
);
1735 iwl_uninit_drv(priv
);
1737 dev_kfree_skb(priv
->beacon_skb
);
1739 ieee80211_free_hw(priv
->hw
);
1742 static const char * const desc_lookup_text
[] = {
1747 "NMI_INTERRUPT_WDG",
1751 "HW_ERROR_TUNE_LOCK",
1752 "HW_ERROR_TEMPERATURE",
1753 "ILLEGAL_CHAN_FREQ",
1756 "NMI_INTERRUPT_HOST",
1757 "NMI_INTERRUPT_ACTION_PT",
1758 "NMI_INTERRUPT_UNKNOWN",
1759 "UCODE_VERSION_MISMATCH",
1760 "HW_ERROR_ABS_LOCK",
1761 "HW_ERROR_CAL_LOCK_FAIL",
1762 "NMI_INTERRUPT_INST_ACTION_PT",
1763 "NMI_INTERRUPT_DATA_ACTION_PT",
1765 "NMI_INTERRUPT_TRM",
1766 "NMI_INTERRUPT_BREAK_POINT",
1773 static struct { char *name
; u8 num
; } advanced_lookup
[] = {
1774 { "NMI_INTERRUPT_WDG", 0x34 },
1775 { "SYSASSERT", 0x35 },
1776 { "UCODE_VERSION_MISMATCH", 0x37 },
1777 { "BAD_COMMAND", 0x38 },
1778 { "NMI_INTERRUPT_DATA_ACTION_PT", 0x3C },
1779 { "FATAL_ERROR", 0x3D },
1780 { "NMI_TRM_HW_ERR", 0x46 },
1781 { "NMI_INTERRUPT_TRM", 0x4C },
1782 { "NMI_INTERRUPT_BREAK_POINT", 0x54 },
1783 { "NMI_INTERRUPT_WDG_RXF_FULL", 0x5C },
1784 { "NMI_INTERRUPT_WDG_NO_RBD_RXF_FULL", 0x64 },
1785 { "NMI_INTERRUPT_HOST", 0x66 },
1786 { "NMI_INTERRUPT_ACTION_PT", 0x7C },
1787 { "NMI_INTERRUPT_UNKNOWN", 0x84 },
1788 { "NMI_INTERRUPT_INST_ACTION_PT", 0x86 },
1789 { "ADVANCED_SYSASSERT", 0 },
1792 static const char *desc_lookup(u32 num
)
1795 int max
= ARRAY_SIZE(desc_lookup_text
);
1798 return desc_lookup_text
[num
];
1800 max
= ARRAY_SIZE(advanced_lookup
) - 1;
1801 for (i
= 0; i
< max
; i
++) {
1802 if (advanced_lookup
[i
].num
== num
)
1805 return advanced_lookup
[i
].name
;
1808 #define ERROR_START_OFFSET (1 * sizeof(u32))
1809 #define ERROR_ELEM_SIZE (7 * sizeof(u32))
1811 static void iwl_dump_nic_error_log(struct iwl_priv
*priv
)
1813 struct iwl_trans
*trans
= trans(priv
);
1815 struct iwl_error_event_table table
;
1817 base
= priv
->device_pointers
.error_event_table
;
1818 if (priv
->cur_ucode
== IWL_UCODE_INIT
) {
1820 base
= priv
->fw
->init_errlog_ptr
;
1823 base
= priv
->fw
->inst_errlog_ptr
;
1826 if (!iwlagn_hw_valid_rtc_data_addr(base
)) {
1828 "Not valid error log pointer 0x%08X for %s uCode\n",
1830 (priv
->cur_ucode
== IWL_UCODE_INIT
)
1835 /*TODO: Update dbgfs with ISR error stats obtained below */
1836 iwl_read_targ_mem_words(trans
, base
, &table
, sizeof(table
));
1838 if (ERROR_START_OFFSET
<= table
.valid
* ERROR_ELEM_SIZE
) {
1839 IWL_ERR(trans
, "Start IWL Error Log Dump:\n");
1840 IWL_ERR(trans
, "Status: 0x%08lX, count: %d\n",
1841 priv
->shrd
->status
, table
.valid
);
1844 trace_iwlwifi_dev_ucode_error(trans
->dev
, table
.error_id
, table
.tsf_low
,
1845 table
.data1
, table
.data2
, table
.line
,
1846 table
.blink1
, table
.blink2
, table
.ilink1
,
1847 table
.ilink2
, table
.bcon_time
, table
.gp1
,
1848 table
.gp2
, table
.gp3
, table
.ucode_ver
,
1849 table
.hw_ver
, table
.brd_ver
);
1850 IWL_ERR(priv
, "0x%08X | %-28s\n", table
.error_id
,
1851 desc_lookup(table
.error_id
));
1852 IWL_ERR(priv
, "0x%08X | uPc\n", table
.pc
);
1853 IWL_ERR(priv
, "0x%08X | branchlink1\n", table
.blink1
);
1854 IWL_ERR(priv
, "0x%08X | branchlink2\n", table
.blink2
);
1855 IWL_ERR(priv
, "0x%08X | interruptlink1\n", table
.ilink1
);
1856 IWL_ERR(priv
, "0x%08X | interruptlink2\n", table
.ilink2
);
1857 IWL_ERR(priv
, "0x%08X | data1\n", table
.data1
);
1858 IWL_ERR(priv
, "0x%08X | data2\n", table
.data2
);
1859 IWL_ERR(priv
, "0x%08X | line\n", table
.line
);
1860 IWL_ERR(priv
, "0x%08X | beacon time\n", table
.bcon_time
);
1861 IWL_ERR(priv
, "0x%08X | tsf low\n", table
.tsf_low
);
1862 IWL_ERR(priv
, "0x%08X | tsf hi\n", table
.tsf_hi
);
1863 IWL_ERR(priv
, "0x%08X | time gp1\n", table
.gp1
);
1864 IWL_ERR(priv
, "0x%08X | time gp2\n", table
.gp2
);
1865 IWL_ERR(priv
, "0x%08X | time gp3\n", table
.gp3
);
1866 IWL_ERR(priv
, "0x%08X | uCode version\n", table
.ucode_ver
);
1867 IWL_ERR(priv
, "0x%08X | hw version\n", table
.hw_ver
);
1868 IWL_ERR(priv
, "0x%08X | board version\n", table
.brd_ver
);
1869 IWL_ERR(priv
, "0x%08X | hcmd\n", table
.hcmd
);
1870 IWL_ERR(priv
, "0x%08X | isr0\n", table
.isr0
);
1871 IWL_ERR(priv
, "0x%08X | isr1\n", table
.isr1
);
1872 IWL_ERR(priv
, "0x%08X | isr2\n", table
.isr2
);
1873 IWL_ERR(priv
, "0x%08X | isr3\n", table
.isr3
);
1874 IWL_ERR(priv
, "0x%08X | isr4\n", table
.isr4
);
1875 IWL_ERR(priv
, "0x%08X | isr_pref\n", table
.isr_pref
);
1876 IWL_ERR(priv
, "0x%08X | wait_event\n", table
.wait_event
);
1877 IWL_ERR(priv
, "0x%08X | l2p_control\n", table
.l2p_control
);
1878 IWL_ERR(priv
, "0x%08X | l2p_duration\n", table
.l2p_duration
);
1879 IWL_ERR(priv
, "0x%08X | l2p_mhvalid\n", table
.l2p_mhvalid
);
1880 IWL_ERR(priv
, "0x%08X | l2p_addr_match\n", table
.l2p_addr_match
);
1881 IWL_ERR(priv
, "0x%08X | lmpm_pmg_sel\n", table
.lmpm_pmg_sel
);
1882 IWL_ERR(priv
, "0x%08X | timestamp\n", table
.u_timestamp
);
1883 IWL_ERR(priv
, "0x%08X | flow_handler\n", table
.flow_handler
);
1886 #define EVENT_START_OFFSET (4 * sizeof(u32))
1889 * iwl_print_event_log - Dump error event log to syslog
1892 static int iwl_print_event_log(struct iwl_priv
*priv
, u32 start_idx
,
1893 u32 num_events
, u32 mode
,
1894 int pos
, char **buf
, size_t bufsz
)
1897 u32 base
; /* SRAM byte address of event log header */
1898 u32 event_size
; /* 2 u32s, or 3 u32s if timestamp recorded */
1899 u32 ptr
; /* SRAM byte address of log data */
1900 u32 ev
, time
, data
; /* event log data */
1901 unsigned long reg_flags
;
1903 struct iwl_trans
*trans
= trans(priv
);
1905 if (num_events
== 0)
1908 base
= priv
->device_pointers
.log_event_table
;
1909 if (priv
->cur_ucode
== IWL_UCODE_INIT
) {
1911 base
= priv
->fw
->init_evtlog_ptr
;
1914 base
= priv
->fw
->inst_evtlog_ptr
;
1918 event_size
= 2 * sizeof(u32
);
1920 event_size
= 3 * sizeof(u32
);
1922 ptr
= base
+ EVENT_START_OFFSET
+ (start_idx
* event_size
);
1924 /* Make sure device is powered up for SRAM reads */
1925 spin_lock_irqsave(&trans
->reg_lock
, reg_flags
);
1926 if (unlikely(!iwl_grab_nic_access(trans
)))
1929 /* Set starting address; reads will auto-increment */
1930 iwl_write32(trans
, HBUS_TARG_MEM_RADDR
, ptr
);
1932 /* "time" is actually "data" for mode 0 (no timestamp).
1933 * place event id # at far right for easier visual parsing. */
1934 for (i
= 0; i
< num_events
; i
++) {
1935 ev
= iwl_read32(trans
, HBUS_TARG_MEM_RDAT
);
1936 time
= iwl_read32(trans
, HBUS_TARG_MEM_RDAT
);
1940 pos
+= scnprintf(*buf
+ pos
, bufsz
- pos
,
1941 "EVT_LOG:0x%08x:%04u\n",
1944 trace_iwlwifi_dev_ucode_event(trans
->dev
, 0,
1946 IWL_ERR(priv
, "EVT_LOG:0x%08x:%04u\n",
1950 data
= iwl_read32(trans
, HBUS_TARG_MEM_RDAT
);
1952 pos
+= scnprintf(*buf
+ pos
, bufsz
- pos
,
1953 "EVT_LOGT:%010u:0x%08x:%04u\n",
1956 IWL_ERR(priv
, "EVT_LOGT:%010u:0x%08x:%04u\n",
1958 trace_iwlwifi_dev_ucode_event(trans
->dev
, time
,
1964 /* Allow device to power down */
1965 iwl_release_nic_access(trans
);
1967 spin_unlock_irqrestore(&trans
->reg_lock
, reg_flags
);
1972 * iwl_print_last_event_logs - Dump the newest # of event log to syslog
1974 static int iwl_print_last_event_logs(struct iwl_priv
*priv
, u32 capacity
,
1975 u32 num_wraps
, u32 next_entry
,
1977 int pos
, char **buf
, size_t bufsz
)
1980 * display the newest DEFAULT_LOG_ENTRIES entries
1981 * i.e the entries just before the next ont that uCode would fill.
1984 if (next_entry
< size
) {
1985 pos
= iwl_print_event_log(priv
,
1986 capacity
- (size
- next_entry
),
1987 size
- next_entry
, mode
,
1989 pos
= iwl_print_event_log(priv
, 0,
1993 pos
= iwl_print_event_log(priv
, next_entry
- size
,
1994 size
, mode
, pos
, buf
, bufsz
);
1996 if (next_entry
< size
) {
1997 pos
= iwl_print_event_log(priv
, 0, next_entry
,
1998 mode
, pos
, buf
, bufsz
);
2000 pos
= iwl_print_event_log(priv
, next_entry
- size
,
2001 size
, mode
, pos
, buf
, bufsz
);
2007 #define DEFAULT_DUMP_EVENT_LOG_ENTRIES (20)
2009 int iwl_dump_nic_event_log(struct iwl_priv
*priv
, bool full_log
,
2010 char **buf
, bool display
)
2012 u32 base
; /* SRAM byte address of event log header */
2013 u32 capacity
; /* event log capacity in # entries */
2014 u32 mode
; /* 0 - no timestamp, 1 - timestamp recorded */
2015 u32 num_wraps
; /* # times uCode wrapped to top of log */
2016 u32 next_entry
; /* index of next entry to be written by uCode */
2017 u32 size
; /* # entries that we'll print */
2021 struct iwl_trans
*trans
= trans(priv
);
2023 base
= priv
->device_pointers
.log_event_table
;
2024 if (priv
->cur_ucode
== IWL_UCODE_INIT
) {
2025 logsize
= priv
->fw
->init_evtlog_size
;
2027 base
= priv
->fw
->init_evtlog_ptr
;
2029 logsize
= priv
->fw
->inst_evtlog_size
;
2031 base
= priv
->fw
->inst_evtlog_ptr
;
2034 if (!iwlagn_hw_valid_rtc_data_addr(base
)) {
2036 "Invalid event log pointer 0x%08X for %s uCode\n",
2038 (priv
->cur_ucode
== IWL_UCODE_INIT
)
2043 /* event log header */
2044 capacity
= iwl_read_targ_mem(trans
, base
);
2045 mode
= iwl_read_targ_mem(trans
, base
+ (1 * sizeof(u32
)));
2046 num_wraps
= iwl_read_targ_mem(trans
, base
+ (2 * sizeof(u32
)));
2047 next_entry
= iwl_read_targ_mem(trans
, base
+ (3 * sizeof(u32
)));
2049 if (capacity
> logsize
) {
2050 IWL_ERR(priv
, "Log capacity %d is bogus, limit to %d "
2051 "entries\n", capacity
, logsize
);
2055 if (next_entry
> logsize
) {
2056 IWL_ERR(priv
, "Log write index %d is bogus, limit to %d\n",
2057 next_entry
, logsize
);
2058 next_entry
= logsize
;
2061 size
= num_wraps
? capacity
: next_entry
;
2063 /* bail out if nothing in log */
2065 IWL_ERR(trans
, "Start IWL Event Log Dump: nothing in log\n");
2069 #ifdef CONFIG_IWLWIFI_DEBUG
2070 if (!(iwl_have_debug_level(IWL_DL_FW_ERRORS
)) && !full_log
)
2071 size
= (size
> DEFAULT_DUMP_EVENT_LOG_ENTRIES
)
2072 ? DEFAULT_DUMP_EVENT_LOG_ENTRIES
: size
;
2074 size
= (size
> DEFAULT_DUMP_EVENT_LOG_ENTRIES
)
2075 ? DEFAULT_DUMP_EVENT_LOG_ENTRIES
: size
;
2077 IWL_ERR(priv
, "Start IWL Event Log Dump: display last %u entries\n",
2080 #ifdef CONFIG_IWLWIFI_DEBUG
2083 bufsz
= capacity
* 48;
2086 *buf
= kmalloc(bufsz
, GFP_KERNEL
);
2090 if (iwl_have_debug_level(IWL_DL_FW_ERRORS
) || full_log
) {
2092 * if uCode has wrapped back to top of log,
2093 * start at the oldest entry,
2094 * i.e the next one that uCode would fill.
2097 pos
= iwl_print_event_log(priv
, next_entry
,
2098 capacity
- next_entry
, mode
,
2100 /* (then/else) start at top of log */
2101 pos
= iwl_print_event_log(priv
, 0,
2102 next_entry
, mode
, pos
, buf
, bufsz
);
2104 pos
= iwl_print_last_event_logs(priv
, capacity
, num_wraps
,
2105 next_entry
, size
, mode
,
2108 pos
= iwl_print_last_event_logs(priv
, capacity
, num_wraps
,
2109 next_entry
, size
, mode
,
2115 static void iwl_nic_error(struct iwl_op_mode
*op_mode
)
2117 struct iwl_priv
*priv
= IWL_OP_MODE_GET_DVM(op_mode
);
2119 IWL_ERR(priv
, "Loaded firmware version: %s\n",
2120 priv
->fw
->fw_version
);
2122 iwl_dump_nic_error_log(priv
);
2123 iwl_dump_nic_event_log(priv
, false, NULL
, false);
2125 iwlagn_fw_error(priv
, false);
2128 static void iwl_cmd_queue_full(struct iwl_op_mode
*op_mode
)
2130 struct iwl_priv
*priv
= IWL_OP_MODE_GET_DVM(op_mode
);
2132 if (!iwl_check_for_ct_kill(priv
)) {
2133 IWL_ERR(priv
, "Restarting adapter queue is full\n");
2134 iwlagn_fw_error(priv
, false);
2138 static void iwl_nic_config(struct iwl_op_mode
*op_mode
)
2140 struct iwl_priv
*priv
= IWL_OP_MODE_GET_DVM(op_mode
);
2142 cfg(priv
)->lib
->nic_config(priv
);
2145 static void iwl_stop_sw_queue(struct iwl_op_mode
*op_mode
, int queue
)
2147 struct iwl_priv
*priv
= IWL_OP_MODE_GET_DVM(op_mode
);
2148 int ac
= priv
->queue_to_ac
[queue
];
2150 if (WARN_ON_ONCE(ac
== IWL_INVALID_AC
))
2153 if (atomic_inc_return(&priv
->ac_stop_count
[ac
]) > 1) {
2154 IWL_DEBUG_TX_QUEUES(priv
,
2155 "queue %d (AC %d) already stopped\n",
2160 set_bit(ac
, &priv
->transport_queue_stop
);
2161 ieee80211_stop_queue(priv
->hw
, ac
);
2164 static void iwl_wake_sw_queue(struct iwl_op_mode
*op_mode
, int queue
)
2166 struct iwl_priv
*priv
= IWL_OP_MODE_GET_DVM(op_mode
);
2167 int ac
= priv
->queue_to_ac
[queue
];
2169 if (WARN_ON_ONCE(ac
== IWL_INVALID_AC
))
2172 if (atomic_dec_return(&priv
->ac_stop_count
[ac
]) > 0) {
2173 IWL_DEBUG_TX_QUEUES(priv
,
2174 "queue %d (AC %d) already awake\n",
2179 clear_bit(ac
, &priv
->transport_queue_stop
);
2181 if (!priv
->passive_no_rx
)
2182 ieee80211_wake_queue(priv
->hw
, ac
);
2185 void iwlagn_lift_passive_no_rx(struct iwl_priv
*priv
)
2189 if (!priv
->passive_no_rx
)
2192 for (ac
= IEEE80211_AC_VO
; ac
< IEEE80211_NUM_ACS
; ac
++) {
2193 if (!test_bit(ac
, &priv
->transport_queue_stop
)) {
2194 IWL_DEBUG_TX_QUEUES(priv
, "Wake queue %d");
2195 ieee80211_wake_queue(priv
->hw
, ac
);
2197 IWL_DEBUG_TX_QUEUES(priv
, "Don't wake queue %d");
2201 priv
->passive_no_rx
= false;
2204 const struct iwl_op_mode_ops iwl_dvm_ops
= {
2205 .start
= iwl_op_mode_dvm_start
,
2206 .stop
= iwl_op_mode_dvm_stop
,
2207 .rx
= iwl_rx_dispatch
,
2208 .queue_full
= iwl_stop_sw_queue
,
2209 .queue_not_full
= iwl_wake_sw_queue
,
2210 .hw_rf_kill
= iwl_set_hw_rfkill_state
,
2211 .free_skb
= iwl_free_skb
,
2212 .nic_error
= iwl_nic_error
,
2213 .cmd_queue_full
= iwl_cmd_queue_full
,
2214 .nic_config
= iwl_nic_config
,
2217 /*****************************************************************************
2219 * driver and module entry point
2221 *****************************************************************************/
2223 struct kmem_cache
*iwl_tx_cmd_pool
;
2225 static int __init
iwl_init(void)
2229 pr_info(DRV_DESCRIPTION
", " DRV_VERSION
"\n");
2230 pr_info(DRV_COPYRIGHT
"\n");
2232 iwl_tx_cmd_pool
= kmem_cache_create("iwl_dev_cmd",
2233 sizeof(struct iwl_device_cmd
),
2234 sizeof(void *), 0, NULL
);
2235 if (!iwl_tx_cmd_pool
)
2238 ret
= iwlagn_rate_control_register();
2240 pr_err("Unable to register rate control algorithm: %d\n", ret
);
2241 goto error_rc_register
;
2244 ret
= iwl_pci_register_driver();
2246 goto error_pci_register
;
2250 iwlagn_rate_control_unregister();
2252 kmem_cache_destroy(iwl_tx_cmd_pool
);
2256 static void __exit
iwl_exit(void)
2258 iwl_pci_unregister_driver();
2259 iwlagn_rate_control_unregister();
2260 kmem_cache_destroy(iwl_tx_cmd_pool
);
2263 module_exit(iwl_exit
);
2264 module_init(iwl_init
);
2266 #ifdef CONFIG_IWLWIFI_DEBUG
2267 module_param_named(debug
, iwlagn_mod_params
.debug_level
, uint
,
2269 MODULE_PARM_DESC(debug
, "debug output mask");
2272 module_param_named(swcrypto
, iwlagn_mod_params
.sw_crypto
, int, S_IRUGO
);
2273 MODULE_PARM_DESC(swcrypto
, "using crypto in software (default 0 [hardware])");
2274 module_param_named(11n_disable
, iwlagn_mod_params
.disable_11n
, uint
, S_IRUGO
);
2275 MODULE_PARM_DESC(11n_disable
,
2276 "disable 11n functionality, bitmap: 1: full, 2: agg TX, 4: agg RX");
2277 module_param_named(amsdu_size_8K
, iwlagn_mod_params
.amsdu_size_8K
,
2279 MODULE_PARM_DESC(amsdu_size_8K
, "enable 8K amsdu size");
2280 module_param_named(fw_restart
, iwlagn_mod_params
.restart_fw
, int, S_IRUGO
);
2281 MODULE_PARM_DESC(fw_restart
, "restart firmware in case of error");
2283 module_param_named(ucode_alternative
,
2284 iwlagn_mod_params
.wanted_ucode_alternative
,
2286 MODULE_PARM_DESC(ucode_alternative
,
2287 "specify ucode alternative to use from ucode file");
2289 module_param_named(antenna_coupling
, iwlagn_mod_params
.ant_coupling
,
2291 MODULE_PARM_DESC(antenna_coupling
,
2292 "specify antenna coupling in dB (defualt: 0 dB)");
2294 module_param_named(bt_ch_inhibition
, iwlagn_mod_params
.bt_ch_announce
,
2296 MODULE_PARM_DESC(bt_ch_inhibition
,
2297 "Enable BT channel inhibition (default: enable)");
2299 module_param_named(plcp_check
, iwlagn_mod_params
.plcp_check
, bool, S_IRUGO
);
2300 MODULE_PARM_DESC(plcp_check
, "Check plcp health (default: 1 [enabled])");
2302 module_param_named(ack_check
, iwlagn_mod_params
.ack_check
, bool, S_IRUGO
);
2303 MODULE_PARM_DESC(ack_check
, "Check ack health (default: 0 [disabled])");
2305 module_param_named(wd_disable
, iwlagn_mod_params
.wd_disable
, int, S_IRUGO
);
2306 MODULE_PARM_DESC(wd_disable
,
2307 "Disable stuck queue watchdog timer 0=system default, "
2308 "1=disable, 2=enable (default: 0)");
2311 * set bt_coex_active to true, uCode will do kill/defer
2312 * every time the priority line is asserted (BT is sending signals on the
2313 * priority line in the PCIx).
2314 * set bt_coex_active to false, uCode will ignore the BT activity and
2315 * perform the normal operation
2317 * User might experience transmit issue on some platform due to WiFi/BT
2318 * co-exist problem. The possible behaviors are:
2319 * Able to scan and finding all the available AP
2320 * Not able to associate with any AP
2321 * On those platforms, WiFi communication can be restored by set
2322 * "bt_coex_active" module parameter to "false"
2324 * default: bt_coex_active = true (BT_COEX_ENABLE)
2326 module_param_named(bt_coex_active
, iwlagn_mod_params
.bt_coex_active
,
2328 MODULE_PARM_DESC(bt_coex_active
, "enable wifi/bt co-exist (default: enable)");
2330 module_param_named(led_mode
, iwlagn_mod_params
.led_mode
, int, S_IRUGO
);
2331 MODULE_PARM_DESC(led_mode
, "0=system default, "
2332 "1=On(RF On)/Off(RF Off), 2=blinking, 3=Off (default: 0)");
2334 module_param_named(power_save
, iwlagn_mod_params
.power_save
,
2336 MODULE_PARM_DESC(power_save
,
2337 "enable WiFi power management (default: disable)");
2339 module_param_named(power_level
, iwlagn_mod_params
.power_level
,
2341 MODULE_PARM_DESC(power_level
,
2342 "default power save level (range from 1 - 5, default: 1)");
2344 module_param_named(auto_agg
, iwlagn_mod_params
.auto_agg
,
2346 MODULE_PARM_DESC(auto_agg
,
2347 "enable agg w/o check traffic load (default: enable)");
2350 * For now, keep using power level 1 instead of automatically
2353 module_param_named(no_sleep_autoadjust
, iwlagn_mod_params
.no_sleep_autoadjust
,
2355 MODULE_PARM_DESC(no_sleep_autoadjust
,
2356 "don't automatically adjust sleep level "
2357 "according to maximum network latency (default: true)");