2 * Copyright (c) 2004-2011 Atheros Communications Inc.
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
11 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
13 * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
14 * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
23 struct ath6kl_sta
*ath6kl_find_sta(struct ath6kl
*ar
, u8
*node_addr
)
25 struct ath6kl_sta
*conn
= NULL
;
28 max_conn
= (ar
->nw_type
== AP_NETWORK
) ? AP_MAX_NUM_STA
: 0;
30 for (i
= 0; i
< max_conn
; i
++) {
31 if (memcmp(node_addr
, ar
->sta_list
[i
].mac
, ETH_ALEN
) == 0) {
32 conn
= &ar
->sta_list
[i
];
40 struct ath6kl_sta
*ath6kl_find_sta_by_aid(struct ath6kl
*ar
, u8 aid
)
42 struct ath6kl_sta
*conn
= NULL
;
45 for (ctr
= 0; ctr
< AP_MAX_NUM_STA
; ctr
++) {
46 if (ar
->sta_list
[ctr
].aid
== aid
) {
47 conn
= &ar
->sta_list
[ctr
];
54 static void ath6kl_add_new_sta(struct ath6kl
*ar
, u8
*mac
, u16 aid
, u8
*wpaie
,
55 u8 ielen
, u8 keymgmt
, u8 ucipher
, u8 auth
)
57 struct ath6kl_sta
*sta
;
62 sta
= &ar
->sta_list
[free_slot
];
63 memcpy(sta
->mac
, mac
, ETH_ALEN
);
64 if (ielen
<= ATH6KL_MAX_IE
)
65 memcpy(sta
->wpa_ie
, wpaie
, ielen
);
67 sta
->keymgmt
= keymgmt
;
68 sta
->ucipher
= ucipher
;
71 ar
->sta_list_index
= ar
->sta_list_index
| (1 << free_slot
);
72 ar
->ap_stats
.sta
[free_slot
].aid
= cpu_to_le32(aid
);
75 static void ath6kl_sta_cleanup(struct ath6kl
*ar
, u8 i
)
77 struct ath6kl_sta
*sta
= &ar
->sta_list
[i
];
79 /* empty the queued pkts in the PS queue if any */
80 spin_lock_bh(&sta
->psq_lock
);
81 skb_queue_purge(&sta
->psq
);
82 spin_unlock_bh(&sta
->psq_lock
);
84 memset(&ar
->ap_stats
.sta
[sta
->aid
- 1], 0,
85 sizeof(struct wmi_per_sta_stat
));
86 memset(sta
->mac
, 0, ETH_ALEN
);
87 memset(sta
->wpa_ie
, 0, ATH6KL_MAX_IE
);
91 ar
->sta_list_index
= ar
->sta_list_index
& ~(1 << i
);
95 static u8
ath6kl_remove_sta(struct ath6kl
*ar
, u8
*mac
, u16 reason
)
99 if (is_zero_ether_addr(mac
))
102 if (is_broadcast_ether_addr(mac
)) {
103 ath6kl_dbg(ATH6KL_DBG_TRC
, "deleting all station\n");
105 for (i
= 0; i
< AP_MAX_NUM_STA
; i
++) {
106 if (!is_zero_ether_addr(ar
->sta_list
[i
].mac
)) {
107 ath6kl_sta_cleanup(ar
, i
);
112 for (i
= 0; i
< AP_MAX_NUM_STA
; i
++) {
113 if (memcmp(ar
->sta_list
[i
].mac
, mac
, ETH_ALEN
) == 0) {
114 ath6kl_dbg(ATH6KL_DBG_TRC
,
115 "deleting station %pM aid=%d reason=%d\n",
116 mac
, ar
->sta_list
[i
].aid
, reason
);
117 ath6kl_sta_cleanup(ar
, i
);
127 enum htc_endpoint_id
ath6kl_ac2_endpoint_id(void *devt
, u8 ac
)
129 struct ath6kl
*ar
= devt
;
130 return ar
->ac2ep_map
[ac
];
133 struct ath6kl_cookie
*ath6kl_alloc_cookie(struct ath6kl
*ar
)
135 struct ath6kl_cookie
*cookie
;
137 cookie
= ar
->cookie_list
;
138 if (cookie
!= NULL
) {
139 ar
->cookie_list
= cookie
->arc_list_next
;
146 void ath6kl_cookie_init(struct ath6kl
*ar
)
150 ar
->cookie_list
= NULL
;
151 ar
->cookie_count
= 0;
153 memset(ar
->cookie_mem
, 0, sizeof(ar
->cookie_mem
));
155 for (i
= 0; i
< MAX_COOKIE_NUM
; i
++)
156 ath6kl_free_cookie(ar
, &ar
->cookie_mem
[i
]);
159 void ath6kl_cookie_cleanup(struct ath6kl
*ar
)
161 ar
->cookie_list
= NULL
;
162 ar
->cookie_count
= 0;
165 void ath6kl_free_cookie(struct ath6kl
*ar
, struct ath6kl_cookie
*cookie
)
172 cookie
->arc_list_next
= ar
->cookie_list
;
173 ar
->cookie_list
= cookie
;
177 /* set the window address register (using 4-byte register access ). */
178 static int ath6kl_set_addrwin_reg(struct ath6kl
*ar
, u32 reg_addr
, u32 addr
)
185 * Write bytes 1,2,3 of the register to set the upper address bytes,
186 * the LSB is written last to initiate the access cycle
189 for (i
= 1; i
<= 3; i
++) {
191 * Fill the buffer with the address byte value we want to
192 * hit 4 times. No need to worry about endianness as the
193 * same byte is copied to all four bytes of addr_val at
196 memset((u8
*)&addr_val
, ((u8
*)&addr
)[i
], 4);
199 * Hit each byte of the register address with a 4-byte
200 * write operation to the same address, this is a harmless
203 status
= hif_read_write_sync(ar
, reg_addr
+ i
, (u8
*)&addr_val
,
204 4, HIF_WR_SYNC_BYTE_FIX
);
210 ath6kl_err("failed to write initial bytes of 0x%x to window reg: 0x%X\n",
216 * Write the address register again, this time write the whole
217 * 4-byte value. The effect here is that the LSB write causes the
218 * cycle to start, the extra 3 byte write to bytes 1,2,3 has no
219 * effect since we are writing the same values again
221 addr_val
= cpu_to_le32(addr
);
222 status
= hif_read_write_sync(ar
, reg_addr
,
224 4, HIF_WR_SYNC_BYTE_INC
);
227 ath6kl_err("failed to write 0x%x to window reg: 0x%X\n",
236 * Read from the hardware through its diagnostic window. No cooperation
237 * from the firmware is required for this.
239 int ath6kl_diag_read32(struct ath6kl
*ar
, u32 address
, u32
*value
)
243 /* set window register to start read cycle */
244 ret
= ath6kl_set_addrwin_reg(ar
, WINDOW_READ_ADDR_ADDRESS
, address
);
249 ret
= hif_read_write_sync(ar
, WINDOW_DATA_ADDRESS
, (u8
*) value
,
250 sizeof(*value
), HIF_RD_SYNC_BYTE_INC
);
252 ath6kl_warn("failed to read32 through diagnose window: %d\n",
261 * Write to the ATH6KL through its diagnostic window. No cooperation from
262 * the Target is required for this.
264 int ath6kl_diag_write32(struct ath6kl
*ar
, u32 address
, __le32 value
)
269 ret
= hif_read_write_sync(ar
, WINDOW_DATA_ADDRESS
, (u8
*) &value
,
270 sizeof(value
), HIF_WR_SYNC_BYTE_INC
);
272 ath6kl_err("failed to write 0x%x during diagnose window to 0x%d\n",
277 /* set window register, which starts the write cycle */
278 return ath6kl_set_addrwin_reg(ar
, WINDOW_WRITE_ADDR_ADDRESS
,
282 int ath6kl_diag_read(struct ath6kl
*ar
, u32 address
, void *data
, u32 length
)
284 u32 count
, *buf
= data
;
287 if (WARN_ON(length
% 4))
290 for (count
= 0; count
< length
/ 4; count
++, address
+= 4) {
291 ret
= ath6kl_diag_read32(ar
, address
, &buf
[count
]);
299 int ath6kl_diag_write(struct ath6kl
*ar
, u32 address
, void *data
, u32 length
)
305 if (WARN_ON(length
% 4))
308 for (count
= 0; count
< length
/ 4; count
++, address
+= 4) {
309 ret
= ath6kl_diag_write32(ar
, address
, buf
[count
]);
317 int ath6kl_read_fwlogs(struct ath6kl
*ar
)
319 struct ath6kl_dbglog_hdr debug_hdr
;
320 struct ath6kl_dbglog_buf debug_buf
;
321 u32 address
, length
, dropped
, firstbuf
, debug_hdr_addr
;
325 buf
= kmalloc(ATH6KL_FWLOG_PAYLOAD_SIZE
, GFP_KERNEL
);
329 address
= TARG_VTOP(ar
->target_type
,
330 ath6kl_get_hi_item_addr(ar
,
331 HI_ITEM(hi_dbglog_hdr
)));
333 ret
= ath6kl_diag_read32(ar
, address
, &debug_hdr_addr
);
337 /* Get the contents of the ring buffer */
338 if (debug_hdr_addr
== 0) {
339 ath6kl_warn("Invalid address for debug_hdr_addr\n");
344 address
= TARG_VTOP(ar
->target_type
, debug_hdr_addr
);
345 ath6kl_diag_read(ar
, address
, &debug_hdr
, sizeof(debug_hdr
));
347 address
= TARG_VTOP(ar
->target_type
,
348 le32_to_cpu(debug_hdr
.dbuf_addr
));
350 dropped
= le32_to_cpu(debug_hdr
.dropped
);
351 ath6kl_diag_read(ar
, address
, &debug_buf
, sizeof(debug_buf
));
356 address
= TARG_VTOP(ar
->target_type
,
357 le32_to_cpu(debug_buf
.buffer_addr
));
358 length
= le32_to_cpu(debug_buf
.length
);
360 if (length
!= 0 && (le32_to_cpu(debug_buf
.length
) <=
361 le32_to_cpu(debug_buf
.bufsize
))) {
362 length
= ALIGN(length
, 4);
364 ret
= ath6kl_diag_read(ar
, address
,
369 ath6kl_debug_fwlog_event(ar
, buf
, length
);
372 address
= TARG_VTOP(ar
->target_type
,
373 le32_to_cpu(debug_buf
.next
));
374 ath6kl_diag_read(ar
, address
, &debug_buf
, sizeof(debug_buf
));
380 if (WARN_ON(loop
== 0)) {
384 } while (address
!= firstbuf
);
392 /* FIXME: move to a better place, target.h? */
393 #define AR6003_RESET_CONTROL_ADDRESS 0x00004000
394 #define AR6004_RESET_CONTROL_ADDRESS 0x00004000
396 static void ath6kl_reset_device(struct ath6kl
*ar
, u32 target_type
,
397 bool wait_fot_compltn
, bool cold_reset
)
403 if (target_type
!= TARGET_TYPE_AR6003
&&
404 target_type
!= TARGET_TYPE_AR6004
)
407 data
= cold_reset
? cpu_to_le32(RESET_CONTROL_COLD_RST
) :
408 cpu_to_le32(RESET_CONTROL_MBOX_RST
);
410 switch (target_type
) {
411 case TARGET_TYPE_AR6003
:
412 address
= AR6003_RESET_CONTROL_ADDRESS
;
414 case TARGET_TYPE_AR6004
:
415 address
= AR6004_RESET_CONTROL_ADDRESS
;
418 address
= AR6003_RESET_CONTROL_ADDRESS
;
422 status
= ath6kl_diag_write32(ar
, address
, data
);
425 ath6kl_err("failed to reset target\n");
428 void ath6kl_stop_endpoint(struct net_device
*dev
, bool keep_profile
,
431 struct ath6kl
*ar
= ath6kl_priv(dev
);
432 static u8 bcast_mac
[] = {0xff, 0xff, 0xff, 0xff, 0xff, 0xff};
435 netif_stop_queue(dev
);
437 /* disable the target and the interrupts associated with it */
438 if (test_bit(WMI_READY
, &ar
->flag
)) {
439 discon_issued
= (test_bit(CONNECTED
, &ar
->flag
) ||
440 test_bit(CONNECT_PEND
, &ar
->flag
));
441 ath6kl_disconnect(ar
);
443 ath6kl_init_profile_info(ar
);
445 del_timer(&ar
->disconnect_timer
);
447 clear_bit(WMI_READY
, &ar
->flag
);
448 ath6kl_wmi_shutdown(ar
->wmi
);
449 clear_bit(WMI_ENABLED
, &ar
->flag
);
453 * After wmi_shudown all WMI events will be dropped. We
454 * need to cleanup the buffers allocated in AP mode and
455 * give disconnect notification to stack, which usually
456 * happens in the disconnect_event. Simulate the disconnect
457 * event by calling the function directly. Sometimes
458 * disconnect_event will be received when the debug logs
462 ath6kl_disconnect_event(ar
, DISCONNECT_CMD
,
463 (ar
->nw_type
& AP_NETWORK
) ?
464 bcast_mac
: ar
->bssid
,
467 ar
->user_key_ctrl
= 0;
470 ath6kl_dbg(ATH6KL_DBG_TRC
,
471 "%s: wmi is not ready 0x%p 0x%p\n",
472 __func__
, ar
, ar
->wmi
);
474 /* Shut down WMI if we have started it */
475 if (test_bit(WMI_ENABLED
, &ar
->flag
)) {
476 ath6kl_dbg(ATH6KL_DBG_TRC
,
477 "%s: shut down wmi\n", __func__
);
478 ath6kl_wmi_shutdown(ar
->wmi
);
479 clear_bit(WMI_ENABLED
, &ar
->flag
);
484 if (ar
->htc_target
) {
485 ath6kl_dbg(ATH6KL_DBG_TRC
, "%s: shut down htc\n", __func__
);
486 ath6kl_htc_stop(ar
->htc_target
);
490 * Try to reset the device if we can. The driver may have been
491 * configure NOT to reset the target during a debug session.
493 ath6kl_dbg(ATH6KL_DBG_TRC
,
494 "attempting to reset target on instance destroy\n");
495 ath6kl_reset_device(ar
, ar
->target_type
, true, true);
498 static void ath6kl_install_static_wep_keys(struct ath6kl
*ar
)
503 for (index
= WMI_MIN_KEY_INDEX
; index
<= WMI_MAX_KEY_INDEX
; index
++) {
504 if (ar
->wep_key_list
[index
].key_len
) {
505 keyusage
= GROUP_USAGE
;
506 if (index
== ar
->def_txkey_index
)
507 keyusage
|= TX_USAGE
;
509 ath6kl_wmi_addkey_cmd(ar
->wmi
,
513 ar
->wep_key_list
[index
].key_len
,
515 ar
->wep_key_list
[index
].key
,
516 KEY_OP_INIT_VAL
, NULL
,
522 void ath6kl_connect_ap_mode_bss(struct ath6kl
*ar
, u16 channel
)
524 struct ath6kl_req_key
*ik
;
526 u8 key_rsc
[ATH6KL_KEY_SEQ_LEN
];
528 ik
= &ar
->ap_mode_bkey
;
530 ath6kl_dbg(ATH6KL_DBG_WLAN_CFG
, "AP mode started on %u MHz\n", channel
);
532 switch (ar
->auth_mode
) {
534 if (ar
->prwise_crypto
== WEP_CRYPT
)
535 ath6kl_install_static_wep_keys(ar
);
539 case (WPA_PSK_AUTH
| WPA2_PSK_AUTH
):
543 ath6kl_dbg(ATH6KL_DBG_WLAN_CFG
, "Delayed addkey for "
544 "the initial group key for AP mode\n");
545 memset(key_rsc
, 0, sizeof(key_rsc
));
546 res
= ath6kl_wmi_addkey_cmd(
547 ar
->wmi
, ik
->key_index
, ik
->key_type
,
548 GROUP_USAGE
, ik
->key_len
, key_rsc
, ik
->key
,
549 KEY_OP_INIT_VAL
, NULL
, SYNC_BOTH_WMIFLAG
);
551 ath6kl_dbg(ATH6KL_DBG_WLAN_CFG
, "Delayed "
552 "addkey failed: %d\n", res
);
557 ath6kl_wmi_bssfilter_cmd(ar
->wmi
, NONE_BSS_FILTER
, 0);
558 set_bit(CONNECTED
, &ar
->flag
);
559 netif_carrier_on(ar
->net_dev
);
562 void ath6kl_connect_ap_mode_sta(struct ath6kl
*ar
, u16 aid
, u8
*mac_addr
,
563 u8 keymgmt
, u8 ucipher
, u8 auth
,
564 u8 assoc_req_len
, u8
*assoc_info
)
566 u8
*ies
= NULL
, *wpa_ie
= NULL
, *pos
;
568 struct station_info sinfo
;
570 ath6kl_dbg(ATH6KL_DBG_TRC
, "new station %pM aid=%d\n", mac_addr
, aid
);
572 if (assoc_req_len
> sizeof(struct ieee80211_hdr_3addr
)) {
573 struct ieee80211_mgmt
*mgmt
=
574 (struct ieee80211_mgmt
*) assoc_info
;
575 if (ieee80211_is_assoc_req(mgmt
->frame_control
) &&
576 assoc_req_len
>= sizeof(struct ieee80211_hdr_3addr
) +
577 sizeof(mgmt
->u
.assoc_req
)) {
578 ies
= mgmt
->u
.assoc_req
.variable
;
579 ies_len
= assoc_info
+ assoc_req_len
- ies
;
580 } else if (ieee80211_is_reassoc_req(mgmt
->frame_control
) &&
581 assoc_req_len
>= sizeof(struct ieee80211_hdr_3addr
)
582 + sizeof(mgmt
->u
.reassoc_req
)) {
583 ies
= mgmt
->u
.reassoc_req
.variable
;
584 ies_len
= assoc_info
+ assoc_req_len
- ies
;
589 while (pos
&& pos
+ 1 < ies
+ ies_len
) {
590 if (pos
+ 2 + pos
[1] > ies
+ ies_len
)
592 if (pos
[0] == WLAN_EID_RSN
)
593 wpa_ie
= pos
; /* RSN IE */
594 else if (pos
[0] == WLAN_EID_VENDOR_SPECIFIC
&&
596 pos
[2] == 0x00 && pos
[3] == 0x50 && pos
[4] == 0xf2) {
598 wpa_ie
= pos
; /* WPA IE */
599 else if (pos
[5] == 0x04) {
600 wpa_ie
= pos
; /* WPS IE */
601 break; /* overrides WPA/RSN IE */
607 ath6kl_add_new_sta(ar
, mac_addr
, aid
, wpa_ie
,
608 wpa_ie
? 2 + wpa_ie
[1] : 0,
609 keymgmt
, ucipher
, auth
);
611 /* send event to application */
612 memset(&sinfo
, 0, sizeof(sinfo
));
614 /* TODO: sinfo.generation */
616 sinfo
.assoc_req_ies
= ies
;
617 sinfo
.assoc_req_ies_len
= ies_len
;
618 sinfo
.filled
|= STATION_INFO_ASSOC_REQ_IES
;
620 cfg80211_new_sta(ar
->net_dev
, mac_addr
, &sinfo
, GFP_KERNEL
);
622 netif_wake_queue(ar
->net_dev
);
625 /* Functions for Tx credit handling */
626 void ath6k_credit_init(struct htc_credit_state_info
*cred_info
,
627 struct list_head
*ep_list
,
630 struct htc_endpoint_credit_dist
*cur_ep_dist
;
633 cred_info
->cur_free_credits
= tot_credits
;
634 cred_info
->total_avail_credits
= tot_credits
;
636 list_for_each_entry(cur_ep_dist
, ep_list
, list
) {
637 if (cur_ep_dist
->endpoint
== ENDPOINT_0
)
640 cur_ep_dist
->cred_min
= cur_ep_dist
->cred_per_msg
;
643 if ((cur_ep_dist
->svc_id
== WMI_DATA_BK_SVC
) ||
644 (cur_ep_dist
->svc_id
== WMI_DATA_BE_SVC
)) {
645 ath6kl_deposit_credit_to_ep(cred_info
,
647 cur_ep_dist
->cred_min
);
648 cur_ep_dist
->dist_flags
|= HTC_EP_ACTIVE
;
651 if (cur_ep_dist
->svc_id
== WMI_CONTROL_SVC
) {
652 ath6kl_deposit_credit_to_ep(cred_info
, cur_ep_dist
,
653 cur_ep_dist
->cred_min
);
655 * Control service is always marked active, it
656 * never goes inactive EVER.
658 cur_ep_dist
->dist_flags
|= HTC_EP_ACTIVE
;
659 } else if (cur_ep_dist
->svc_id
== WMI_DATA_BK_SVC
)
660 /* this is the lowest priority data endpoint */
661 cred_info
->lowestpri_ep_dist
= cur_ep_dist
->list
;
664 * Streams have to be created (explicit | implicit) for all
665 * kinds of traffic. BE endpoints are also inactive in the
666 * beginning. When BE traffic starts it creates implicit
667 * streams that redistributes credits.
669 * Note: all other endpoints have minimums set but are
670 * initially given NO credits. credits will be distributed
671 * as traffic activity demands
675 WARN_ON(cred_info
->cur_free_credits
<= 0);
677 list_for_each_entry(cur_ep_dist
, ep_list
, list
) {
678 if (cur_ep_dist
->endpoint
== ENDPOINT_0
)
681 if (cur_ep_dist
->svc_id
== WMI_CONTROL_SVC
)
682 cur_ep_dist
->cred_norm
= cur_ep_dist
->cred_per_msg
;
685 * For the remaining data endpoints, we assume that
686 * each cred_per_msg are the same. We use a simple
687 * calculation here, we take the remaining credits
688 * and determine how many max messages this can
689 * cover and then set each endpoint's normal value
690 * equal to 3/4 this amount.
692 count
= (cred_info
->cur_free_credits
/
693 cur_ep_dist
->cred_per_msg
)
694 * cur_ep_dist
->cred_per_msg
;
695 count
= (count
* 3) >> 2;
696 count
= max(count
, cur_ep_dist
->cred_per_msg
);
697 cur_ep_dist
->cred_norm
= count
;
703 /* initialize and setup credit distribution */
704 int ath6k_setup_credit_dist(void *htc_handle
,
705 struct htc_credit_state_info
*cred_info
)
707 u16 servicepriority
[5];
709 memset(cred_info
, 0, sizeof(struct htc_credit_state_info
));
711 servicepriority
[0] = WMI_CONTROL_SVC
; /* highest */
712 servicepriority
[1] = WMI_DATA_VO_SVC
;
713 servicepriority
[2] = WMI_DATA_VI_SVC
;
714 servicepriority
[3] = WMI_DATA_BE_SVC
;
715 servicepriority
[4] = WMI_DATA_BK_SVC
; /* lowest */
717 /* set priority list */
718 ath6kl_htc_set_credit_dist(htc_handle
, cred_info
, servicepriority
, 5);
723 /* reduce an ep's credits back to a set limit */
724 static void ath6k_reduce_credits(struct htc_credit_state_info
*cred_info
,
725 struct htc_endpoint_credit_dist
*ep_dist
,
730 ep_dist
->cred_assngd
= limit
;
732 if (ep_dist
->credits
<= limit
)
735 credits
= ep_dist
->credits
- limit
;
736 ep_dist
->credits
-= credits
;
737 cred_info
->cur_free_credits
+= credits
;
740 static void ath6k_credit_update(struct htc_credit_state_info
*cred_info
,
741 struct list_head
*epdist_list
)
743 struct htc_endpoint_credit_dist
*cur_dist_list
;
745 list_for_each_entry(cur_dist_list
, epdist_list
, list
) {
746 if (cur_dist_list
->endpoint
== ENDPOINT_0
)
749 if (cur_dist_list
->cred_to_dist
> 0) {
750 cur_dist_list
->credits
+=
751 cur_dist_list
->cred_to_dist
;
752 cur_dist_list
->cred_to_dist
= 0;
753 if (cur_dist_list
->credits
>
754 cur_dist_list
->cred_assngd
)
755 ath6k_reduce_credits(cred_info
,
757 cur_dist_list
->cred_assngd
);
759 if (cur_dist_list
->credits
>
760 cur_dist_list
->cred_norm
)
761 ath6k_reduce_credits(cred_info
, cur_dist_list
,
762 cur_dist_list
->cred_norm
);
764 if (!(cur_dist_list
->dist_flags
& HTC_EP_ACTIVE
)) {
765 if (cur_dist_list
->txq_depth
== 0)
766 ath6k_reduce_credits(cred_info
,
774 * HTC has an endpoint that needs credits, ep_dist is the endpoint in
777 void ath6k_seek_credits(struct htc_credit_state_info
*cred_info
,
778 struct htc_endpoint_credit_dist
*ep_dist
)
780 struct htc_endpoint_credit_dist
*curdist_list
;
784 if (ep_dist
->svc_id
== WMI_CONTROL_SVC
)
787 if ((ep_dist
->svc_id
== WMI_DATA_VI_SVC
) ||
788 (ep_dist
->svc_id
== WMI_DATA_VO_SVC
))
789 if ((ep_dist
->cred_assngd
>= ep_dist
->cred_norm
))
793 * For all other services, we follow a simple algorithm of:
795 * 1. checking the free pool for credits
796 * 2. checking lower priority endpoints for credits to take
799 credits
= min(cred_info
->cur_free_credits
, ep_dist
->seek_cred
);
801 if (credits
>= ep_dist
->seek_cred
)
805 * We don't have enough in the free pool, try taking away from
806 * lower priority services The rule for taking away credits:
808 * 1. Only take from lower priority endpoints
809 * 2. Only take what is allocated above the minimum (never
810 * starve an endpoint completely)
811 * 3. Only take what you need.
814 list_for_each_entry_reverse(curdist_list
,
815 &cred_info
->lowestpri_ep_dist
,
817 if (curdist_list
== ep_dist
)
820 need
= ep_dist
->seek_cred
- cred_info
->cur_free_credits
;
822 if ((curdist_list
->cred_assngd
- need
) >=
823 curdist_list
->cred_min
) {
825 * The current one has been allocated more than
826 * it's minimum and it has enough credits assigned
827 * above it's minimum to fulfill our need try to
828 * take away just enough to fulfill our need.
830 ath6k_reduce_credits(cred_info
, curdist_list
,
831 curdist_list
->cred_assngd
- need
);
833 if (cred_info
->cur_free_credits
>=
838 if (curdist_list
->endpoint
== ENDPOINT_0
)
842 credits
= min(cred_info
->cur_free_credits
, ep_dist
->seek_cred
);
845 /* did we find some credits? */
847 ath6kl_deposit_credit_to_ep(cred_info
, ep_dist
, credits
);
849 ep_dist
->seek_cred
= 0;
852 /* redistribute credits based on activity change */
853 static void ath6k_redistribute_credits(struct htc_credit_state_info
*info
,
854 struct list_head
*ep_dist_list
)
856 struct htc_endpoint_credit_dist
*curdist_list
;
858 list_for_each_entry(curdist_list
, ep_dist_list
, list
) {
859 if (curdist_list
->endpoint
== ENDPOINT_0
)
862 if ((curdist_list
->svc_id
== WMI_DATA_BK_SVC
) ||
863 (curdist_list
->svc_id
== WMI_DATA_BE_SVC
))
864 curdist_list
->dist_flags
|= HTC_EP_ACTIVE
;
866 if ((curdist_list
->svc_id
!= WMI_CONTROL_SVC
) &&
867 !(curdist_list
->dist_flags
& HTC_EP_ACTIVE
)) {
868 if (curdist_list
->txq_depth
== 0)
869 ath6k_reduce_credits(info
,
872 ath6k_reduce_credits(info
,
874 curdist_list
->cred_min
);
881 * This function is invoked whenever endpoints require credit
882 * distributions. A lock is held while this function is invoked, this
883 * function shall NOT block. The ep_dist_list is a list of distribution
884 * structures in prioritized order as defined by the call to the
885 * htc_set_credit_dist() api.
887 void ath6k_credit_distribute(struct htc_credit_state_info
*cred_info
,
888 struct list_head
*ep_dist_list
,
889 enum htc_credit_dist_reason reason
)
892 case HTC_CREDIT_DIST_SEND_COMPLETE
:
893 ath6k_credit_update(cred_info
, ep_dist_list
);
895 case HTC_CREDIT_DIST_ACTIVITY_CHANGE
:
896 ath6k_redistribute_credits(cred_info
, ep_dist_list
);
902 WARN_ON(cred_info
->cur_free_credits
> cred_info
->total_avail_credits
);
903 WARN_ON(cred_info
->cur_free_credits
< 0);
906 void disconnect_timer_handler(unsigned long ptr
)
908 struct net_device
*dev
= (struct net_device
*)ptr
;
909 struct ath6kl
*ar
= ath6kl_priv(dev
);
911 ath6kl_init_profile_info(ar
);
912 ath6kl_disconnect(ar
);
915 void ath6kl_disconnect(struct ath6kl
*ar
)
917 if (test_bit(CONNECTED
, &ar
->flag
) ||
918 test_bit(CONNECT_PEND
, &ar
->flag
)) {
919 ath6kl_wmi_disconnect_cmd(ar
->wmi
);
921 * Disconnect command is issued, clear the connect pending
922 * flag. The connected flag will be cleared in
923 * disconnect event notification.
925 clear_bit(CONNECT_PEND
, &ar
->flag
);
929 void ath6kl_deep_sleep_enable(struct ath6kl
*ar
)
931 switch (ar
->sme_state
) {
933 cfg80211_connect_result(ar
->net_dev
, ar
->bssid
, NULL
, 0,
935 WLAN_STATUS_UNSPECIFIED_FAILURE
,
941 * FIXME: oddly enough smeState is in DISCONNECTED during
942 * suspend, why? Need to send disconnected event in that
945 cfg80211_disconnected(ar
->net_dev
, 0, NULL
, 0, GFP_KERNEL
);
949 if (test_bit(CONNECTED
, &ar
->flag
) ||
950 test_bit(CONNECT_PEND
, &ar
->flag
))
951 ath6kl_wmi_disconnect_cmd(ar
->wmi
);
953 ar
->sme_state
= SME_DISCONNECTED
;
955 /* disable scanning */
956 if (ath6kl_wmi_scanparams_cmd(ar
->wmi
, 0xFFFF, 0, 0, 0, 0, 0, 0, 0,
958 printk(KERN_WARNING
"ath6kl: failed to disable scan "
961 ath6kl_cfg80211_scan_complete_event(ar
, -ECANCELED
);
963 /* save the current power mode before enabling power save */
964 ar
->wmi
->saved_pwr_mode
= ar
->wmi
->pwr_mode
;
966 if (ath6kl_wmi_powermode_cmd(ar
->wmi
, REC_POWER
) != 0)
967 ath6kl_warn("ath6kl_deep_sleep_enable: "
968 "wmi_powermode_cmd failed\n");
971 /* WMI Event handlers */
973 static const char *get_hw_id_string(u32 id
)
976 case AR6003_REV1_VERSION
:
978 case AR6003_REV2_VERSION
:
980 case AR6003_REV3_VERSION
:
987 void ath6kl_ready_event(void *devt
, u8
*datap
, u32 sw_ver
, u32 abi_ver
)
989 struct ath6kl
*ar
= devt
;
990 struct net_device
*dev
= ar
->net_dev
;
992 memcpy(dev
->dev_addr
, datap
, ETH_ALEN
);
993 ath6kl_dbg(ATH6KL_DBG_TRC
, "%s: mac addr = %pM\n",
994 __func__
, dev
->dev_addr
);
996 ar
->version
.wlan_ver
= sw_ver
;
997 ar
->version
.abi_ver
= abi_ver
;
999 snprintf(ar
->wdev
->wiphy
->fw_version
,
1000 sizeof(ar
->wdev
->wiphy
->fw_version
),
1002 (ar
->version
.wlan_ver
& 0xf0000000) >> 28,
1003 (ar
->version
.wlan_ver
& 0x0f000000) >> 24,
1004 (ar
->version
.wlan_ver
& 0x00ff0000) >> 16,
1005 (ar
->version
.wlan_ver
& 0x0000ffff));
1007 /* indicate to the waiting thread that the ready event was received */
1008 set_bit(WMI_READY
, &ar
->flag
);
1009 wake_up(&ar
->event_wq
);
1011 ath6kl_info("hw %s fw %s%s\n",
1012 get_hw_id_string(ar
->wdev
->wiphy
->hw_version
),
1013 ar
->wdev
->wiphy
->fw_version
,
1014 test_bit(TESTMODE
, &ar
->flag
) ? " testmode" : "");
1017 void ath6kl_scan_complete_evt(struct ath6kl
*ar
, int status
)
1019 ath6kl_cfg80211_scan_complete_event(ar
, status
);
1021 if (!ar
->usr_bss_filter
) {
1022 clear_bit(CLEAR_BSSFILTER_ON_BEACON
, &ar
->flag
);
1023 ath6kl_wmi_bssfilter_cmd(ar
->wmi
, NONE_BSS_FILTER
, 0);
1026 ath6kl_dbg(ATH6KL_DBG_WLAN_SCAN
, "scan complete: %d\n", status
);
1029 void ath6kl_connect_event(struct ath6kl
*ar
, u16 channel
, u8
*bssid
,
1030 u16 listen_int
, u16 beacon_int
,
1031 enum network_type net_type
, u8 beacon_ie_len
,
1032 u8 assoc_req_len
, u8 assoc_resp_len
,
1035 ath6kl_cfg80211_connect_event(ar
, channel
, bssid
,
1036 listen_int
, beacon_int
,
1037 net_type
, beacon_ie_len
,
1038 assoc_req_len
, assoc_resp_len
,
1041 memcpy(ar
->bssid
, bssid
, sizeof(ar
->bssid
));
1042 ar
->bss_ch
= channel
;
1044 if ((ar
->nw_type
== INFRA_NETWORK
))
1045 ath6kl_wmi_listeninterval_cmd(ar
->wmi
, ar
->listen_intvl_t
,
1046 ar
->listen_intvl_b
);
1048 netif_wake_queue(ar
->net_dev
);
1050 /* Update connect & link status atomically */
1051 spin_lock_bh(&ar
->lock
);
1052 set_bit(CONNECTED
, &ar
->flag
);
1053 clear_bit(CONNECT_PEND
, &ar
->flag
);
1054 netif_carrier_on(ar
->net_dev
);
1055 spin_unlock_bh(&ar
->lock
);
1057 aggr_reset_state(ar
->aggr_cntxt
);
1058 ar
->reconnect_flag
= 0;
1060 if ((ar
->nw_type
== ADHOC_NETWORK
) && ar
->ibss_ps_enable
) {
1061 memset(ar
->node_map
, 0, sizeof(ar
->node_map
));
1063 ar
->next_ep_id
= ENDPOINT_2
;
1066 if (!ar
->usr_bss_filter
) {
1067 set_bit(CLEAR_BSSFILTER_ON_BEACON
, &ar
->flag
);
1068 ath6kl_wmi_bssfilter_cmd(ar
->wmi
, CURRENT_BSS_FILTER
, 0);
1072 void ath6kl_tkip_micerr_event(struct ath6kl
*ar
, u8 keyid
, bool ismcast
)
1074 struct ath6kl_sta
*sta
;
1077 * For AP case, keyid will have aid of STA which sent pkt with
1078 * MIC error. Use this aid to get MAC & send it to hostapd.
1080 if (ar
->nw_type
== AP_NETWORK
) {
1081 sta
= ath6kl_find_sta_by_aid(ar
, (keyid
>> 2));
1085 ath6kl_dbg(ATH6KL_DBG_TRC
,
1086 "ap tkip mic error received from aid=%d\n", keyid
);
1088 memset(tsc
, 0, sizeof(tsc
)); /* FIX: get correct TSC */
1089 cfg80211_michael_mic_failure(ar
->net_dev
, sta
->mac
,
1090 NL80211_KEYTYPE_PAIRWISE
, keyid
,
1093 ath6kl_cfg80211_tkip_micerr_event(ar
, keyid
, ismcast
);
1097 static void ath6kl_update_target_stats(struct ath6kl
*ar
, u8
*ptr
, u32 len
)
1099 struct wmi_target_stats
*tgt_stats
=
1100 (struct wmi_target_stats
*) ptr
;
1101 struct target_stats
*stats
= &ar
->target_stats
;
1102 struct tkip_ccmp_stats
*ccmp_stats
;
1105 if (len
< sizeof(*tgt_stats
))
1108 ath6kl_dbg(ATH6KL_DBG_TRC
, "updating target stats\n");
1110 stats
->tx_pkt
+= le32_to_cpu(tgt_stats
->stats
.tx
.pkt
);
1111 stats
->tx_byte
+= le32_to_cpu(tgt_stats
->stats
.tx
.byte
);
1112 stats
->tx_ucast_pkt
+= le32_to_cpu(tgt_stats
->stats
.tx
.ucast_pkt
);
1113 stats
->tx_ucast_byte
+= le32_to_cpu(tgt_stats
->stats
.tx
.ucast_byte
);
1114 stats
->tx_mcast_pkt
+= le32_to_cpu(tgt_stats
->stats
.tx
.mcast_pkt
);
1115 stats
->tx_mcast_byte
+= le32_to_cpu(tgt_stats
->stats
.tx
.mcast_byte
);
1116 stats
->tx_bcast_pkt
+= le32_to_cpu(tgt_stats
->stats
.tx
.bcast_pkt
);
1117 stats
->tx_bcast_byte
+= le32_to_cpu(tgt_stats
->stats
.tx
.bcast_byte
);
1118 stats
->tx_rts_success_cnt
+=
1119 le32_to_cpu(tgt_stats
->stats
.tx
.rts_success_cnt
);
1121 for (ac
= 0; ac
< WMM_NUM_AC
; ac
++)
1122 stats
->tx_pkt_per_ac
[ac
] +=
1123 le32_to_cpu(tgt_stats
->stats
.tx
.pkt_per_ac
[ac
]);
1125 stats
->tx_err
+= le32_to_cpu(tgt_stats
->stats
.tx
.err
);
1126 stats
->tx_fail_cnt
+= le32_to_cpu(tgt_stats
->stats
.tx
.fail_cnt
);
1127 stats
->tx_retry_cnt
+= le32_to_cpu(tgt_stats
->stats
.tx
.retry_cnt
);
1128 stats
->tx_mult_retry_cnt
+=
1129 le32_to_cpu(tgt_stats
->stats
.tx
.mult_retry_cnt
);
1130 stats
->tx_rts_fail_cnt
+=
1131 le32_to_cpu(tgt_stats
->stats
.tx
.rts_fail_cnt
);
1132 stats
->tx_ucast_rate
=
1133 ath6kl_wmi_get_rate(a_sle32_to_cpu(tgt_stats
->stats
.tx
.ucast_rate
));
1135 stats
->rx_pkt
+= le32_to_cpu(tgt_stats
->stats
.rx
.pkt
);
1136 stats
->rx_byte
+= le32_to_cpu(tgt_stats
->stats
.rx
.byte
);
1137 stats
->rx_ucast_pkt
+= le32_to_cpu(tgt_stats
->stats
.rx
.ucast_pkt
);
1138 stats
->rx_ucast_byte
+= le32_to_cpu(tgt_stats
->stats
.rx
.ucast_byte
);
1139 stats
->rx_mcast_pkt
+= le32_to_cpu(tgt_stats
->stats
.rx
.mcast_pkt
);
1140 stats
->rx_mcast_byte
+= le32_to_cpu(tgt_stats
->stats
.rx
.mcast_byte
);
1141 stats
->rx_bcast_pkt
+= le32_to_cpu(tgt_stats
->stats
.rx
.bcast_pkt
);
1142 stats
->rx_bcast_byte
+= le32_to_cpu(tgt_stats
->stats
.rx
.bcast_byte
);
1143 stats
->rx_frgment_pkt
+= le32_to_cpu(tgt_stats
->stats
.rx
.frgment_pkt
);
1144 stats
->rx_err
+= le32_to_cpu(tgt_stats
->stats
.rx
.err
);
1145 stats
->rx_crc_err
+= le32_to_cpu(tgt_stats
->stats
.rx
.crc_err
);
1146 stats
->rx_key_cache_miss
+=
1147 le32_to_cpu(tgt_stats
->stats
.rx
.key_cache_miss
);
1148 stats
->rx_decrypt_err
+= le32_to_cpu(tgt_stats
->stats
.rx
.decrypt_err
);
1149 stats
->rx_dupl_frame
+= le32_to_cpu(tgt_stats
->stats
.rx
.dupl_frame
);
1150 stats
->rx_ucast_rate
=
1151 ath6kl_wmi_get_rate(a_sle32_to_cpu(tgt_stats
->stats
.rx
.ucast_rate
));
1153 ccmp_stats
= &tgt_stats
->stats
.tkip_ccmp_stats
;
1155 stats
->tkip_local_mic_fail
+=
1156 le32_to_cpu(ccmp_stats
->tkip_local_mic_fail
);
1157 stats
->tkip_cnter_measures_invoked
+=
1158 le32_to_cpu(ccmp_stats
->tkip_cnter_measures_invoked
);
1159 stats
->tkip_fmt_err
+= le32_to_cpu(ccmp_stats
->tkip_fmt_err
);
1161 stats
->ccmp_fmt_err
+= le32_to_cpu(ccmp_stats
->ccmp_fmt_err
);
1162 stats
->ccmp_replays
+= le32_to_cpu(ccmp_stats
->ccmp_replays
);
1164 stats
->pwr_save_fail_cnt
+=
1165 le32_to_cpu(tgt_stats
->pm_stats
.pwr_save_failure_cnt
);
1166 stats
->noise_floor_calib
=
1167 a_sle32_to_cpu(tgt_stats
->noise_floor_calib
);
1169 stats
->cs_bmiss_cnt
+=
1170 le32_to_cpu(tgt_stats
->cserv_stats
.cs_bmiss_cnt
);
1171 stats
->cs_low_rssi_cnt
+=
1172 le32_to_cpu(tgt_stats
->cserv_stats
.cs_low_rssi_cnt
);
1173 stats
->cs_connect_cnt
+=
1174 le16_to_cpu(tgt_stats
->cserv_stats
.cs_connect_cnt
);
1175 stats
->cs_discon_cnt
+=
1176 le16_to_cpu(tgt_stats
->cserv_stats
.cs_discon_cnt
);
1178 stats
->cs_ave_beacon_rssi
=
1179 a_sle16_to_cpu(tgt_stats
->cserv_stats
.cs_ave_beacon_rssi
);
1181 stats
->cs_last_roam_msec
=
1182 tgt_stats
->cserv_stats
.cs_last_roam_msec
;
1183 stats
->cs_snr
= tgt_stats
->cserv_stats
.cs_snr
;
1184 stats
->cs_rssi
= a_sle16_to_cpu(tgt_stats
->cserv_stats
.cs_rssi
);
1186 stats
->lq_val
= le32_to_cpu(tgt_stats
->lq_val
);
1188 stats
->wow_pkt_dropped
+=
1189 le32_to_cpu(tgt_stats
->wow_stats
.wow_pkt_dropped
);
1190 stats
->wow_host_pkt_wakeups
+=
1191 tgt_stats
->wow_stats
.wow_host_pkt_wakeups
;
1192 stats
->wow_host_evt_wakeups
+=
1193 tgt_stats
->wow_stats
.wow_host_evt_wakeups
;
1194 stats
->wow_evt_discarded
+=
1195 le16_to_cpu(tgt_stats
->wow_stats
.wow_evt_discarded
);
1197 if (test_bit(STATS_UPDATE_PEND
, &ar
->flag
)) {
1198 clear_bit(STATS_UPDATE_PEND
, &ar
->flag
);
1199 wake_up(&ar
->event_wq
);
1203 static void ath6kl_add_le32(__le32
*var
, __le32 val
)
1205 *var
= cpu_to_le32(le32_to_cpu(*var
) + le32_to_cpu(val
));
1208 void ath6kl_tgt_stats_event(struct ath6kl
*ar
, u8
*ptr
, u32 len
)
1210 struct wmi_ap_mode_stat
*p
= (struct wmi_ap_mode_stat
*) ptr
;
1211 struct wmi_ap_mode_stat
*ap
= &ar
->ap_stats
;
1212 struct wmi_per_sta_stat
*st_ap
, *st_p
;
1215 if (ar
->nw_type
== AP_NETWORK
) {
1216 if (len
< sizeof(*p
))
1219 for (ac
= 0; ac
< AP_MAX_NUM_STA
; ac
++) {
1220 st_ap
= &ap
->sta
[ac
];
1223 ath6kl_add_le32(&st_ap
->tx_bytes
, st_p
->tx_bytes
);
1224 ath6kl_add_le32(&st_ap
->tx_pkts
, st_p
->tx_pkts
);
1225 ath6kl_add_le32(&st_ap
->tx_error
, st_p
->tx_error
);
1226 ath6kl_add_le32(&st_ap
->tx_discard
, st_p
->tx_discard
);
1227 ath6kl_add_le32(&st_ap
->rx_bytes
, st_p
->rx_bytes
);
1228 ath6kl_add_le32(&st_ap
->rx_pkts
, st_p
->rx_pkts
);
1229 ath6kl_add_le32(&st_ap
->rx_error
, st_p
->rx_error
);
1230 ath6kl_add_le32(&st_ap
->rx_discard
, st_p
->rx_discard
);
1234 ath6kl_update_target_stats(ar
, ptr
, len
);
1238 void ath6kl_wakeup_event(void *dev
)
1240 struct ath6kl
*ar
= (struct ath6kl
*) dev
;
1242 wake_up(&ar
->event_wq
);
1245 void ath6kl_txpwr_rx_evt(void *devt
, u8 tx_pwr
)
1247 struct ath6kl
*ar
= (struct ath6kl
*) devt
;
1249 ar
->tx_pwr
= tx_pwr
;
1250 wake_up(&ar
->event_wq
);
1253 void ath6kl_pspoll_event(struct ath6kl
*ar
, u8 aid
)
1255 struct ath6kl_sta
*conn
;
1256 struct sk_buff
*skb
;
1257 bool psq_empty
= false;
1259 conn
= ath6kl_find_sta_by_aid(ar
, aid
);
1264 * Send out a packet queued on ps queue. When the ps queue
1265 * becomes empty update the PVB for this station.
1267 spin_lock_bh(&conn
->psq_lock
);
1268 psq_empty
= skb_queue_empty(&conn
->psq
);
1269 spin_unlock_bh(&conn
->psq_lock
);
1272 /* TODO: Send out a NULL data frame */
1275 spin_lock_bh(&conn
->psq_lock
);
1276 skb
= skb_dequeue(&conn
->psq
);
1277 spin_unlock_bh(&conn
->psq_lock
);
1279 conn
->sta_flags
|= STA_PS_POLLED
;
1280 ath6kl_data_tx(skb
, ar
->net_dev
);
1281 conn
->sta_flags
&= ~STA_PS_POLLED
;
1283 spin_lock_bh(&conn
->psq_lock
);
1284 psq_empty
= skb_queue_empty(&conn
->psq
);
1285 spin_unlock_bh(&conn
->psq_lock
);
1288 ath6kl_wmi_set_pvb_cmd(ar
->wmi
, conn
->aid
, 0);
1291 void ath6kl_dtimexpiry_event(struct ath6kl
*ar
)
1293 bool mcastq_empty
= false;
1294 struct sk_buff
*skb
;
1297 * If there are no associated STAs, ignore the DTIM expiry event.
1298 * There can be potential race conditions where the last associated
1299 * STA may disconnect & before the host could clear the 'Indicate
1300 * DTIM' request to the firmware, the firmware would have just
1301 * indicated a DTIM expiry event. The race is between 'clear DTIM
1302 * expiry cmd' going from the host to the firmware & the DTIM
1303 * expiry event happening from the firmware to the host.
1305 if (!ar
->sta_list_index
)
1308 spin_lock_bh(&ar
->mcastpsq_lock
);
1309 mcastq_empty
= skb_queue_empty(&ar
->mcastpsq
);
1310 spin_unlock_bh(&ar
->mcastpsq_lock
);
1315 /* set the STA flag to dtim_expired for the frame to go out */
1316 set_bit(DTIM_EXPIRED
, &ar
->flag
);
1318 spin_lock_bh(&ar
->mcastpsq_lock
);
1319 while ((skb
= skb_dequeue(&ar
->mcastpsq
)) != NULL
) {
1320 spin_unlock_bh(&ar
->mcastpsq_lock
);
1322 ath6kl_data_tx(skb
, ar
->net_dev
);
1324 spin_lock_bh(&ar
->mcastpsq_lock
);
1326 spin_unlock_bh(&ar
->mcastpsq_lock
);
1328 clear_bit(DTIM_EXPIRED
, &ar
->flag
);
1330 /* clear the LSB of the BitMapCtl field of the TIM IE */
1331 ath6kl_wmi_set_pvb_cmd(ar
->wmi
, MCAST_AID
, 0);
1334 void ath6kl_disconnect_event(struct ath6kl
*ar
, u8 reason
, u8
*bssid
,
1335 u8 assoc_resp_len
, u8
*assoc_info
,
1336 u16 prot_reason_status
)
1338 if (ar
->nw_type
== AP_NETWORK
) {
1339 if (!ath6kl_remove_sta(ar
, bssid
, prot_reason_status
))
1342 /* if no more associated STAs, empty the mcast PS q */
1343 if (ar
->sta_list_index
== 0) {
1344 spin_lock_bh(&ar
->mcastpsq_lock
);
1345 skb_queue_purge(&ar
->mcastpsq
);
1346 spin_unlock_bh(&ar
->mcastpsq_lock
);
1348 /* clear the LSB of the TIM IE's BitMapCtl field */
1349 if (test_bit(WMI_READY
, &ar
->flag
))
1350 ath6kl_wmi_set_pvb_cmd(ar
->wmi
, MCAST_AID
, 0);
1353 if (!is_broadcast_ether_addr(bssid
)) {
1354 /* send event to application */
1355 cfg80211_del_sta(ar
->net_dev
, bssid
, GFP_KERNEL
);
1358 if (memcmp(ar
->net_dev
->dev_addr
, bssid
, ETH_ALEN
) == 0) {
1359 memset(ar
->wep_key_list
, 0, sizeof(ar
->wep_key_list
));
1360 clear_bit(CONNECTED
, &ar
->flag
);
1365 ath6kl_cfg80211_disconnect_event(ar
, reason
, bssid
,
1366 assoc_resp_len
, assoc_info
,
1367 prot_reason_status
);
1369 aggr_reset_state(ar
->aggr_cntxt
);
1371 del_timer(&ar
->disconnect_timer
);
1373 ath6kl_dbg(ATH6KL_DBG_WLAN_CONNECT
,
1374 "disconnect reason is %d\n", reason
);
1377 * If the event is due to disconnect cmd from the host, only they
1378 * the target would stop trying to connect. Under any other
1379 * condition, target would keep trying to connect.
1381 if (reason
== DISCONNECT_CMD
) {
1382 if (!ar
->usr_bss_filter
&& test_bit(WMI_READY
, &ar
->flag
))
1383 ath6kl_wmi_bssfilter_cmd(ar
->wmi
, NONE_BSS_FILTER
, 0);
1385 set_bit(CONNECT_PEND
, &ar
->flag
);
1386 if (((reason
== ASSOC_FAILED
) &&
1387 (prot_reason_status
== 0x11)) ||
1388 ((reason
== ASSOC_FAILED
) && (prot_reason_status
== 0x0)
1389 && (ar
->reconnect_flag
== 1))) {
1390 set_bit(CONNECTED
, &ar
->flag
);
1395 /* update connect & link status atomically */
1396 spin_lock_bh(&ar
->lock
);
1397 clear_bit(CONNECTED
, &ar
->flag
);
1398 netif_carrier_off(ar
->net_dev
);
1399 spin_unlock_bh(&ar
->lock
);
1401 if ((reason
!= CSERV_DISCONNECT
) || (ar
->reconnect_flag
!= 1))
1402 ar
->reconnect_flag
= 0;
1404 if (reason
!= CSERV_DISCONNECT
)
1405 ar
->user_key_ctrl
= 0;
1407 netif_stop_queue(ar
->net_dev
);
1408 memset(ar
->bssid
, 0, sizeof(ar
->bssid
));
1411 ath6kl_tx_data_cleanup(ar
);
1414 static int ath6kl_open(struct net_device
*dev
)
1416 struct ath6kl
*ar
= ath6kl_priv(dev
);
1418 spin_lock_bh(&ar
->lock
);
1420 set_bit(WLAN_ENABLED
, &ar
->flag
);
1422 if (test_bit(CONNECTED
, &ar
->flag
)) {
1423 netif_carrier_on(dev
);
1424 netif_wake_queue(dev
);
1426 netif_carrier_off(dev
);
1428 spin_unlock_bh(&ar
->lock
);
1433 static int ath6kl_close(struct net_device
*dev
)
1435 struct ath6kl
*ar
= ath6kl_priv(dev
);
1437 netif_stop_queue(dev
);
1439 ath6kl_disconnect(ar
);
1441 if (test_bit(WMI_READY
, &ar
->flag
)) {
1442 if (ath6kl_wmi_scanparams_cmd(ar
->wmi
, 0xFFFF, 0, 0, 0, 0, 0, 0,
1446 clear_bit(WLAN_ENABLED
, &ar
->flag
);
1449 ath6kl_cfg80211_scan_complete_event(ar
, -ECANCELED
);
1454 static struct net_device_stats
*ath6kl_get_stats(struct net_device
*dev
)
1456 struct ath6kl
*ar
= ath6kl_priv(dev
);
1458 return &ar
->net_stats
;
1461 static struct net_device_ops ath6kl_netdev_ops
= {
1462 .ndo_open
= ath6kl_open
,
1463 .ndo_stop
= ath6kl_close
,
1464 .ndo_start_xmit
= ath6kl_data_tx
,
1465 .ndo_get_stats
= ath6kl_get_stats
,
1468 void init_netdev(struct net_device
*dev
)
1470 dev
->netdev_ops
= &ath6kl_netdev_ops
;
1471 dev
->watchdog_timeo
= ATH6KL_TX_TIMEOUT
;
1473 dev
->needed_headroom
= ETH_HLEN
;
1474 dev
->needed_headroom
+= sizeof(struct ath6kl_llc_snap_hdr
) +
1475 sizeof(struct wmi_data_hdr
) + HTC_HDR_LENGTH
1476 + WMI_MAX_TX_META_SZ
+ ATH6KL_HTC_ALIGN_BYTES
;