tx.c revision 1c33db782d1d0d9be83feacbb065cd4956f485e7
1/*
2 * This file is part of wl1271
3 *
4 * Copyright (C) 2009 Nokia Corporation
5 *
6 * Contact: Luciano Coelho <luciano.coelho@nokia.com>
7 *
8 * This program is free software; you can redistribute it and/or
9 * modify it under the terms of the GNU General Public License
10 * version 2 as published by the Free Software Foundation.
11 *
12 * This program is distributed in the hope that it will be useful, but
13 * WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
15 * General Public License for more details.
16 *
17 * You should have received a copy of the GNU General Public License
18 * along with this program; if not, write to the Free Software
19 * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
20 * 02110-1301 USA
21 *
22 */
23
24#include <linux/kernel.h>
25#include <linux/module.h>
26#include <linux/etherdevice.h>
27
28#include "wlcore.h"
29#include "debug.h"
30#include "io.h"
31#include "ps.h"
32#include "tx.h"
33#include "event.h"
34#include "hw_ops.h"
35
36/*
37 * TODO: this is here just for now, it must be removed when the data
38 * operations are in place.
39 */
40#include "../wl12xx/reg.h"
41
42static int wl1271_set_default_wep_key(struct wl1271 *wl,
43				      struct wl12xx_vif *wlvif, u8 id)
44{
45	int ret;
46	bool is_ap = (wlvif->bss_type == BSS_TYPE_AP_BSS);
47
48	if (is_ap)
49		ret = wl12xx_cmd_set_default_wep_key(wl, id,
50						     wlvif->ap.bcast_hlid);
51	else
52		ret = wl12xx_cmd_set_default_wep_key(wl, id, wlvif->sta.hlid);
53
54	if (ret < 0)
55		return ret;
56
57	wl1271_debug(DEBUG_CRYPT, "default wep key idx: %d", (int)id);
58	return 0;
59}
60
61static int wl1271_alloc_tx_id(struct wl1271 *wl, struct sk_buff *skb)
62{
63	int id;
64
65	id = find_first_zero_bit(wl->tx_frames_map, wl->num_tx_desc);
66	if (id >= wl->num_tx_desc)
67		return -EBUSY;
68
69	__set_bit(id, wl->tx_frames_map);
70	wl->tx_frames[id] = skb;
71	wl->tx_frames_cnt++;
72	return id;
73}
74
75void wl1271_free_tx_id(struct wl1271 *wl, int id)
76{
77	if (__test_and_clear_bit(id, wl->tx_frames_map)) {
78		if (unlikely(wl->tx_frames_cnt == wl->num_tx_desc))
79			clear_bit(WL1271_FLAG_FW_TX_BUSY, &wl->flags);
80
81		wl->tx_frames[id] = NULL;
82		wl->tx_frames_cnt--;
83	}
84}
85EXPORT_SYMBOL(wl1271_free_tx_id);
86
87static void wl1271_tx_ap_update_inconnection_sta(struct wl1271 *wl,
88						 struct sk_buff *skb)
89{
90	struct ieee80211_hdr *hdr;
91
92	/*
93	 * add the station to the known list before transmitting the
94	 * authentication response. this way it won't get de-authed by FW
95	 * when transmitting too soon.
96	 */
97	hdr = (struct ieee80211_hdr *)(skb->data +
98				       sizeof(struct wl1271_tx_hw_descr));
99	if (ieee80211_is_auth(hdr->frame_control))
100		wl1271_acx_set_inconnection_sta(wl, hdr->addr1);
101}
102
103static void wl1271_tx_regulate_link(struct wl1271 *wl,
104				    struct wl12xx_vif *wlvif,
105				    u8 hlid)
106{
107	bool fw_ps, single_link;
108	u8 tx_pkts;
109
110	if (WARN_ON(!test_bit(hlid, wlvif->links_map)))
111		return;
112
113	fw_ps = test_bit(hlid, (unsigned long *)&wl->ap_fw_ps_map);
114	tx_pkts = wl->links[hlid].allocated_pkts;
115	single_link = (wl->active_link_count == 1);
116
117	/*
118	 * if in FW PS and there is enough data in FW we can put the link
119	 * into high-level PS and clean out its TX queues.
120	 * Make an exception if this is the only connected link. In this
121	 * case FW-memory congestion is less of a problem.
122	 */
123	if (!single_link && fw_ps && tx_pkts >= WL1271_PS_STA_MAX_PACKETS)
124		wl12xx_ps_link_start(wl, wlvif, hlid, true);
125}
126
127bool wl12xx_is_dummy_packet(struct wl1271 *wl, struct sk_buff *skb)
128{
129	return wl->dummy_packet == skb;
130}
131EXPORT_SYMBOL(wl12xx_is_dummy_packet);
132
133static u8 wl12xx_tx_get_hlid_ap(struct wl1271 *wl, struct wl12xx_vif *wlvif,
134				struct sk_buff *skb, struct ieee80211_sta *sta)
135{
136	if (sta) {
137		struct wl1271_station *wl_sta;
138
139		wl_sta = (struct wl1271_station *)sta->drv_priv;
140		return wl_sta->hlid;
141	} else {
142		struct ieee80211_hdr *hdr;
143
144		if (!test_bit(WLVIF_FLAG_AP_STARTED, &wlvif->flags))
145			return wl->system_hlid;
146
147		hdr = (struct ieee80211_hdr *)skb->data;
148		if (is_multicast_ether_addr(ieee80211_get_DA(hdr)))
149			return wlvif->ap.bcast_hlid;
150		else
151			return wlvif->ap.global_hlid;
152	}
153}
154
155u8 wl12xx_tx_get_hlid(struct wl1271 *wl, struct wl12xx_vif *wlvif,
156		      struct sk_buff *skb, struct ieee80211_sta *sta)
157{
158	struct ieee80211_tx_info *control;
159
160	if (wlvif->bss_type == BSS_TYPE_AP_BSS)
161		return wl12xx_tx_get_hlid_ap(wl, wlvif, skb, sta);
162
163	control = IEEE80211_SKB_CB(skb);
164	if (control->flags & IEEE80211_TX_CTL_TX_OFFCHAN) {
165		wl1271_debug(DEBUG_TX, "tx offchannel");
166		return wlvif->dev_hlid;
167	}
168
169	return wlvif->sta.hlid;
170}
171
172unsigned int wlcore_calc_packet_alignment(struct wl1271 *wl,
173					  unsigned int packet_length)
174{
175	if ((wl->quirks & WLCORE_QUIRK_TX_PAD_LAST_FRAME) ||
176	    !(wl->quirks & WLCORE_QUIRK_TX_BLOCKSIZE_ALIGN))
177		return ALIGN(packet_length, WL1271_TX_ALIGN_TO);
178	else
179		return ALIGN(packet_length, WL12XX_BUS_BLOCK_SIZE);
180}
181EXPORT_SYMBOL(wlcore_calc_packet_alignment);
182
183static int wl1271_tx_allocate(struct wl1271 *wl, struct wl12xx_vif *wlvif,
184			      struct sk_buff *skb, u32 extra, u32 buf_offset,
185			      u8 hlid, bool is_gem)
186{
187	struct wl1271_tx_hw_descr *desc;
188	u32 total_len = skb->len + sizeof(struct wl1271_tx_hw_descr) + extra;
189	u32 total_blocks;
190	int id, ret = -EBUSY, ac;
191	u32 spare_blocks;
192
193	if (buf_offset + total_len > wl->aggr_buf_size)
194		return -EAGAIN;
195
196	spare_blocks = wlcore_hw_get_spare_blocks(wl, is_gem);
197
198	/* allocate free identifier for the packet */
199	id = wl1271_alloc_tx_id(wl, skb);
200	if (id < 0)
201		return id;
202
203	total_blocks = wlcore_hw_calc_tx_blocks(wl, total_len, spare_blocks);
204
205	if (total_blocks <= wl->tx_blocks_available) {
206		desc = (struct wl1271_tx_hw_descr *)skb_push(
207			skb, total_len - skb->len);
208
209		wlcore_hw_set_tx_desc_blocks(wl, desc, total_blocks,
210					     spare_blocks);
211
212		desc->id = id;
213
214		wl->tx_blocks_available -= total_blocks;
215		wl->tx_allocated_blocks += total_blocks;
216
217		/* If the FW was empty before, arm the Tx watchdog */
218		if (wl->tx_allocated_blocks == total_blocks)
219			wl12xx_rearm_tx_watchdog_locked(wl);
220
221		ac = wl1271_tx_get_queue(skb_get_queue_mapping(skb));
222		wl->tx_allocated_pkts[ac]++;
223
224		if (test_bit(hlid, wl->links_map))
225			wl->links[hlid].allocated_pkts++;
226
227		ret = 0;
228
229		wl1271_debug(DEBUG_TX,
230			     "tx_allocate: size: %d, blocks: %d, id: %d",
231			     total_len, total_blocks, id);
232	} else {
233		wl1271_free_tx_id(wl, id);
234	}
235
236	return ret;
237}
238
239static void wl1271_tx_fill_hdr(struct wl1271 *wl, struct wl12xx_vif *wlvif,
240			       struct sk_buff *skb, u32 extra,
241			       struct ieee80211_tx_info *control, u8 hlid)
242{
243	struct timespec ts;
244	struct wl1271_tx_hw_descr *desc;
245	int ac, rate_idx;
246	s64 hosttime;
247	u16 tx_attr = 0;
248	__le16 frame_control;
249	struct ieee80211_hdr *hdr;
250	u8 *frame_start;
251	bool is_dummy;
252
253	desc = (struct wl1271_tx_hw_descr *) skb->data;
254	frame_start = (u8 *)(desc + 1);
255	hdr = (struct ieee80211_hdr *)(frame_start + extra);
256	frame_control = hdr->frame_control;
257
258	/* relocate space for security header */
259	if (extra) {
260		int hdrlen = ieee80211_hdrlen(frame_control);
261		memmove(frame_start, hdr, hdrlen);
262		skb_set_network_header(skb, skb_network_offset(skb) + extra);
263	}
264
265	/* configure packet life time */
266	getnstimeofday(&ts);
267	hosttime = (timespec_to_ns(&ts) >> 10);
268	desc->start_time = cpu_to_le32(hosttime - wl->time_offset);
269
270	is_dummy = wl12xx_is_dummy_packet(wl, skb);
271	if (is_dummy || !wlvif || wlvif->bss_type != BSS_TYPE_AP_BSS)
272		desc->life_time = cpu_to_le16(TX_HW_MGMT_PKT_LIFETIME_TU);
273	else
274		desc->life_time = cpu_to_le16(TX_HW_AP_MODE_PKT_LIFETIME_TU);
275
276	/* queue */
277	ac = wl1271_tx_get_queue(skb_get_queue_mapping(skb));
278	desc->tid = skb->priority;
279
280	if (is_dummy) {
281		/*
282		 * FW expects the dummy packet to have an invalid session id -
283		 * any session id that is different than the one set in the join
284		 */
285		tx_attr = (SESSION_COUNTER_INVALID <<
286			   TX_HW_ATTR_OFST_SESSION_COUNTER) &
287			   TX_HW_ATTR_SESSION_COUNTER;
288
289		tx_attr |= TX_HW_ATTR_TX_DUMMY_REQ;
290	} else if (wlvif) {
291		u8 session_id = wl->session_ids[hlid];
292
293		if ((wl->quirks & WLCORE_QUIRK_AP_ZERO_SESSION_ID) &&
294		    (wlvif->bss_type == BSS_TYPE_AP_BSS))
295			session_id = 0;
296
297		/* configure the tx attributes */
298		tx_attr = session_id << TX_HW_ATTR_OFST_SESSION_COUNTER;
299	}
300
301	desc->hlid = hlid;
302	if (is_dummy || !wlvif)
303		rate_idx = 0;
304	else if (wlvif->bss_type != BSS_TYPE_AP_BSS) {
305		/*
306		 * if the packets are data packets
307		 * send them with AP rate policies (EAPOLs are an exception),
308		 * otherwise use default basic rates
309		 */
310		if (skb->protocol == cpu_to_be16(ETH_P_PAE))
311			rate_idx = wlvif->sta.basic_rate_idx;
312		else if (control->flags & IEEE80211_TX_CTL_NO_CCK_RATE)
313			rate_idx = wlvif->sta.p2p_rate_idx;
314		else if (ieee80211_is_data(frame_control))
315			rate_idx = wlvif->sta.ap_rate_idx;
316		else
317			rate_idx = wlvif->sta.basic_rate_idx;
318	} else {
319		if (hlid == wlvif->ap.global_hlid)
320			rate_idx = wlvif->ap.mgmt_rate_idx;
321		else if (hlid == wlvif->ap.bcast_hlid ||
322			 skb->protocol == cpu_to_be16(ETH_P_PAE) ||
323			 !ieee80211_is_data(frame_control))
324			/*
325			 * send non-data, bcast and EAPOLs using the
326			 * min basic rate
327			 */
328			rate_idx = wlvif->ap.bcast_rate_idx;
329		else
330			rate_idx = wlvif->ap.ucast_rate_idx[ac];
331	}
332
333	tx_attr |= rate_idx << TX_HW_ATTR_OFST_RATE_POLICY;
334
335	/* for WEP shared auth - no fw encryption is needed */
336	if (ieee80211_is_auth(frame_control) &&
337	    ieee80211_has_protected(frame_control))
338		tx_attr |= TX_HW_ATTR_HOST_ENCRYPT;
339
340	desc->tx_attr = cpu_to_le16(tx_attr);
341
342	wlcore_hw_set_tx_desc_csum(wl, desc, skb);
343	wlcore_hw_set_tx_desc_data_len(wl, desc, skb);
344}
345
346/* caller must hold wl->mutex */
347static int wl1271_prepare_tx_frame(struct wl1271 *wl, struct wl12xx_vif *wlvif,
348				   struct sk_buff *skb, u32 buf_offset, u8 hlid)
349{
350	struct ieee80211_tx_info *info;
351	u32 extra = 0;
352	int ret = 0;
353	u32 total_len;
354	bool is_dummy;
355	bool is_gem = false;
356
357	if (!skb) {
358		wl1271_error("discarding null skb");
359		return -EINVAL;
360	}
361
362	if (hlid == WL12XX_INVALID_LINK_ID) {
363		wl1271_error("invalid hlid. dropping skb 0x%p", skb);
364		return -EINVAL;
365	}
366
367	info = IEEE80211_SKB_CB(skb);
368
369	is_dummy = wl12xx_is_dummy_packet(wl, skb);
370
371	if ((wl->quirks & WLCORE_QUIRK_TKIP_HEADER_SPACE) &&
372	    info->control.hw_key &&
373	    info->control.hw_key->cipher == WLAN_CIPHER_SUITE_TKIP)
374		extra = WL1271_EXTRA_SPACE_TKIP;
375
376	if (info->control.hw_key) {
377		bool is_wep;
378		u8 idx = info->control.hw_key->hw_key_idx;
379		u32 cipher = info->control.hw_key->cipher;
380
381		is_wep = (cipher == WLAN_CIPHER_SUITE_WEP40) ||
382			 (cipher == WLAN_CIPHER_SUITE_WEP104);
383
384		if (unlikely(is_wep && wlvif->default_key != idx)) {
385			ret = wl1271_set_default_wep_key(wl, wlvif, idx);
386			if (ret < 0)
387				return ret;
388			wlvif->default_key = idx;
389		}
390
391		is_gem = (cipher == WL1271_CIPHER_SUITE_GEM);
392	}
393
394	ret = wl1271_tx_allocate(wl, wlvif, skb, extra, buf_offset, hlid,
395				 is_gem);
396	if (ret < 0)
397		return ret;
398
399	wl1271_tx_fill_hdr(wl, wlvif, skb, extra, info, hlid);
400
401	if (!is_dummy && wlvif && wlvif->bss_type == BSS_TYPE_AP_BSS) {
402		wl1271_tx_ap_update_inconnection_sta(wl, skb);
403		wl1271_tx_regulate_link(wl, wlvif, hlid);
404	}
405
406	/*
407	 * The length of each packet is stored in terms of
408	 * words. Thus, we must pad the skb data to make sure its
409	 * length is aligned.  The number of padding bytes is computed
410	 * and set in wl1271_tx_fill_hdr.
411	 * In special cases, we want to align to a specific block size
412	 * (eg. for wl128x with SDIO we align to 256).
413	 */
414	total_len = wlcore_calc_packet_alignment(wl, skb->len);
415
416	memcpy(wl->aggr_buf + buf_offset, skb->data, skb->len);
417	memset(wl->aggr_buf + buf_offset + skb->len, 0, total_len - skb->len);
418
419	/* Revert side effects in the dummy packet skb, so it can be reused */
420	if (is_dummy)
421		skb_pull(skb, sizeof(struct wl1271_tx_hw_descr));
422
423	return total_len;
424}
425
426u32 wl1271_tx_enabled_rates_get(struct wl1271 *wl, u32 rate_set,
427				enum ieee80211_band rate_band)
428{
429	struct ieee80211_supported_band *band;
430	u32 enabled_rates = 0;
431	int bit;
432
433	band = wl->hw->wiphy->bands[rate_band];
434	for (bit = 0; bit < band->n_bitrates; bit++) {
435		if (rate_set & 0x1)
436			enabled_rates |= band->bitrates[bit].hw_value;
437		rate_set >>= 1;
438	}
439
440	/* MCS rates indication are on bits 16 - 31 */
441	rate_set >>= HW_HT_RATES_OFFSET - band->n_bitrates;
442
443	for (bit = 0; bit < 16; bit++) {
444		if (rate_set & 0x1)
445			enabled_rates |= (CONF_HW_BIT_RATE_MCS_0 << bit);
446		rate_set >>= 1;
447	}
448
449	return enabled_rates;
450}
451
452void wl1271_handle_tx_low_watermark(struct wl1271 *wl)
453{
454	int i;
455	struct wl12xx_vif *wlvif;
456
457	wl12xx_for_each_wlvif(wl, wlvif) {
458		for (i = 0; i < NUM_TX_QUEUES; i++) {
459			if (wlcore_is_queue_stopped_by_reason(wl, wlvif, i,
460					WLCORE_QUEUE_STOP_REASON_WATERMARK) &&
461			    wlvif->tx_queue_count[i] <=
462					WL1271_TX_QUEUE_LOW_WATERMARK)
463				/* firmware buffer has space, restart queues */
464				wlcore_wake_queue(wl, wlvif, i,
465					WLCORE_QUEUE_STOP_REASON_WATERMARK);
466		}
467	}
468}
469
470static int wlcore_select_ac(struct wl1271 *wl)
471{
472	int i, q = -1, ac;
473	u32 min_pkts = 0xffffffff;
474
475	/*
476	 * Find a non-empty ac where:
477	 * 1. There are packets to transmit
478	 * 2. The FW has the least allocated blocks
479	 *
480	 * We prioritize the ACs according to VO>VI>BE>BK
481	 */
482	for (i = 0; i < NUM_TX_QUEUES; i++) {
483		ac = wl1271_tx_get_queue(i);
484		if (wl->tx_queue_count[ac] &&
485		    wl->tx_allocated_pkts[ac] < min_pkts) {
486			q = ac;
487			min_pkts = wl->tx_allocated_pkts[q];
488		}
489	}
490
491	return q;
492}
493
494static struct sk_buff *wlcore_lnk_dequeue(struct wl1271 *wl,
495					  struct wl1271_link *lnk, u8 q)
496{
497	struct sk_buff *skb;
498	unsigned long flags;
499
500	skb = skb_dequeue(&lnk->tx_queue[q]);
501	if (skb) {
502		spin_lock_irqsave(&wl->wl_lock, flags);
503		WARN_ON_ONCE(wl->tx_queue_count[q] <= 0);
504		wl->tx_queue_count[q]--;
505		if (lnk->wlvif) {
506			WARN_ON_ONCE(lnk->wlvif->tx_queue_count[q] <= 0);
507			lnk->wlvif->tx_queue_count[q]--;
508		}
509		spin_unlock_irqrestore(&wl->wl_lock, flags);
510	}
511
512	return skb;
513}
514
515static struct sk_buff *wlcore_lnk_dequeue_high_prio(struct wl1271 *wl,
516						    u8 hlid, u8 ac,
517						    u8 *low_prio_hlid)
518{
519	struct wl1271_link *lnk = &wl->links[hlid];
520
521	if (!wlcore_hw_lnk_high_prio(wl, hlid, lnk)) {
522		if (*low_prio_hlid == WL12XX_INVALID_LINK_ID &&
523		    !skb_queue_empty(&lnk->tx_queue[ac]) &&
524		    wlcore_hw_lnk_low_prio(wl, hlid, lnk))
525			/* we found the first non-empty low priority queue */
526			*low_prio_hlid = hlid;
527
528		return NULL;
529	}
530
531	return wlcore_lnk_dequeue(wl, lnk, ac);
532}
533
534static struct sk_buff *wlcore_vif_dequeue_high_prio(struct wl1271 *wl,
535						    struct wl12xx_vif *wlvif,
536						    u8 ac, u8 *hlid,
537						    u8 *low_prio_hlid)
538{
539	struct sk_buff *skb = NULL;
540	int i, h, start_hlid;
541
542	/* start from the link after the last one */
543	start_hlid = (wlvif->last_tx_hlid + 1) % WL12XX_MAX_LINKS;
544
545	/* dequeue according to AC, round robin on each link */
546	for (i = 0; i < WL12XX_MAX_LINKS; i++) {
547		h = (start_hlid + i) % WL12XX_MAX_LINKS;
548
549		/* only consider connected stations */
550		if (!test_bit(h, wlvif->links_map))
551			continue;
552
553		skb = wlcore_lnk_dequeue_high_prio(wl, h, ac,
554						   low_prio_hlid);
555		if (!skb)
556			continue;
557
558		wlvif->last_tx_hlid = h;
559		break;
560	}
561
562	if (!skb)
563		wlvif->last_tx_hlid = 0;
564
565	*hlid = wlvif->last_tx_hlid;
566	return skb;
567}
568
569static struct sk_buff *wl1271_skb_dequeue(struct wl1271 *wl, u8 *hlid)
570{
571	unsigned long flags;
572	struct wl12xx_vif *wlvif = wl->last_wlvif;
573	struct sk_buff *skb = NULL;
574	int ac;
575	u8 low_prio_hlid = WL12XX_INVALID_LINK_ID;
576
577	ac = wlcore_select_ac(wl);
578	if (ac < 0)
579		goto out;
580
581	/* continue from last wlvif (round robin) */
582	if (wlvif) {
583		wl12xx_for_each_wlvif_continue(wl, wlvif) {
584			if (!wlvif->tx_queue_count[ac])
585				continue;
586
587			skb = wlcore_vif_dequeue_high_prio(wl, wlvif, ac, hlid,
588							   &low_prio_hlid);
589			if (!skb)
590				continue;
591
592			wl->last_wlvif = wlvif;
593			break;
594		}
595	}
596
597	/* dequeue from the system HLID before the restarting wlvif list */
598	if (!skb) {
599		skb = wlcore_lnk_dequeue_high_prio(wl, wl->system_hlid,
600						   ac, &low_prio_hlid);
601		if (skb) {
602			*hlid = wl->system_hlid;
603			wl->last_wlvif = NULL;
604		}
605	}
606
607	/* Do a new pass over the wlvif list. But no need to continue
608	 * after last_wlvif. The previous pass should have found it. */
609	if (!skb) {
610		wl12xx_for_each_wlvif(wl, wlvif) {
611			if (!wlvif->tx_queue_count[ac])
612				goto next;
613
614			skb = wlcore_vif_dequeue_high_prio(wl, wlvif, ac, hlid,
615							   &low_prio_hlid);
616			if (skb) {
617				wl->last_wlvif = wlvif;
618				break;
619			}
620
621next:
622			if (wlvif == wl->last_wlvif)
623				break;
624		}
625	}
626
627	/* no high priority skbs found - but maybe a low priority one? */
628	if (!skb && low_prio_hlid != WL12XX_INVALID_LINK_ID) {
629		struct wl1271_link *lnk = &wl->links[low_prio_hlid];
630		skb = wlcore_lnk_dequeue(wl, lnk, ac);
631
632		WARN_ON(!skb); /* we checked this before */
633		*hlid = low_prio_hlid;
634
635		/* ensure proper round robin in the vif/link levels */
636		wl->last_wlvif = lnk->wlvif;
637		if (lnk->wlvif)
638			lnk->wlvif->last_tx_hlid = low_prio_hlid;
639
640	}
641
642	if (!skb &&
643	    test_and_clear_bit(WL1271_FLAG_DUMMY_PACKET_PENDING, &wl->flags)) {
644		int q;
645
646		skb = wl->dummy_packet;
647		*hlid = wl->system_hlid;
648		q = wl1271_tx_get_queue(skb_get_queue_mapping(skb));
649		spin_lock_irqsave(&wl->wl_lock, flags);
650		WARN_ON_ONCE(wl->tx_queue_count[q] <= 0);
651		wl->tx_queue_count[q]--;
652		spin_unlock_irqrestore(&wl->wl_lock, flags);
653	}
654
655out:
656	return skb;
657}
658
659static void wl1271_skb_queue_head(struct wl1271 *wl, struct wl12xx_vif *wlvif,
660				  struct sk_buff *skb, u8 hlid)
661{
662	unsigned long flags;
663	int q = wl1271_tx_get_queue(skb_get_queue_mapping(skb));
664
665	if (wl12xx_is_dummy_packet(wl, skb)) {
666		set_bit(WL1271_FLAG_DUMMY_PACKET_PENDING, &wl->flags);
667	} else {
668		skb_queue_head(&wl->links[hlid].tx_queue[q], skb);
669
670		/* make sure we dequeue the same packet next time */
671		wlvif->last_tx_hlid = (hlid + WL12XX_MAX_LINKS - 1) %
672				      WL12XX_MAX_LINKS;
673	}
674
675	spin_lock_irqsave(&wl->wl_lock, flags);
676	wl->tx_queue_count[q]++;
677	if (wlvif)
678		wlvif->tx_queue_count[q]++;
679	spin_unlock_irqrestore(&wl->wl_lock, flags);
680}
681
682static bool wl1271_tx_is_data_present(struct sk_buff *skb)
683{
684	struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)(skb->data);
685
686	return ieee80211_is_data_present(hdr->frame_control);
687}
688
689void wl12xx_rearm_rx_streaming(struct wl1271 *wl, unsigned long *active_hlids)
690{
691	struct wl12xx_vif *wlvif;
692	u32 timeout;
693	u8 hlid;
694
695	if (!wl->conf.rx_streaming.interval)
696		return;
697
698	if (!wl->conf.rx_streaming.always &&
699	    !test_bit(WL1271_FLAG_SOFT_GEMINI, &wl->flags))
700		return;
701
702	timeout = wl->conf.rx_streaming.duration;
703	wl12xx_for_each_wlvif_sta(wl, wlvif) {
704		bool found = false;
705		for_each_set_bit(hlid, active_hlids, WL12XX_MAX_LINKS) {
706			if (test_bit(hlid, wlvif->links_map)) {
707				found  = true;
708				break;
709			}
710		}
711
712		if (!found)
713			continue;
714
715		/* enable rx streaming */
716		if (!test_bit(WLVIF_FLAG_RX_STREAMING_STARTED, &wlvif->flags))
717			ieee80211_queue_work(wl->hw,
718					     &wlvif->rx_streaming_enable_work);
719
720		mod_timer(&wlvif->rx_streaming_timer,
721			  jiffies + msecs_to_jiffies(timeout));
722	}
723}
724
725/*
726 * Returns failure values only in case of failed bus ops within this function.
727 * wl1271_prepare_tx_frame retvals won't be returned in order to avoid
728 * triggering recovery by higher layers when not necessary.
729 * In case a FW command fails within wl1271_prepare_tx_frame fails a recovery
730 * will be queued in wl1271_cmd_send. -EAGAIN/-EBUSY from prepare_tx_frame
731 * can occur and are legitimate so don't propagate. -EINVAL will emit a WARNING
732 * within prepare_tx_frame code but there's nothing we should do about those
733 * as well.
734 */
735int wlcore_tx_work_locked(struct wl1271 *wl)
736{
737	struct wl12xx_vif *wlvif;
738	struct sk_buff *skb;
739	struct wl1271_tx_hw_descr *desc;
740	u32 buf_offset = 0, last_len = 0;
741	bool sent_packets = false;
742	unsigned long active_hlids[BITS_TO_LONGS(WL12XX_MAX_LINKS)] = {0};
743	int ret = 0;
744	int bus_ret = 0;
745	u8 hlid;
746
747	if (unlikely(wl->state != WLCORE_STATE_ON))
748		return 0;
749
750	while ((skb = wl1271_skb_dequeue(wl, &hlid))) {
751		struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb);
752		bool has_data = false;
753
754		wlvif = NULL;
755		if (!wl12xx_is_dummy_packet(wl, skb))
756			wlvif = wl12xx_vif_to_data(info->control.vif);
757		else
758			hlid = wl->system_hlid;
759
760		has_data = wlvif && wl1271_tx_is_data_present(skb);
761		ret = wl1271_prepare_tx_frame(wl, wlvif, skb, buf_offset,
762					      hlid);
763		if (ret == -EAGAIN) {
764			/*
765			 * Aggregation buffer is full.
766			 * Flush buffer and try again.
767			 */
768			wl1271_skb_queue_head(wl, wlvif, skb, hlid);
769
770			buf_offset = wlcore_hw_pre_pkt_send(wl, buf_offset,
771							    last_len);
772			bus_ret = wlcore_write_data(wl, REG_SLV_MEM_DATA,
773					     wl->aggr_buf, buf_offset, true);
774			if (bus_ret < 0)
775				goto out;
776
777			sent_packets = true;
778			buf_offset = 0;
779			continue;
780		} else if (ret == -EBUSY) {
781			/*
782			 * Firmware buffer is full.
783			 * Queue back last skb, and stop aggregating.
784			 */
785			wl1271_skb_queue_head(wl, wlvif, skb, hlid);
786			/* No work left, avoid scheduling redundant tx work */
787			set_bit(WL1271_FLAG_FW_TX_BUSY, &wl->flags);
788			goto out_ack;
789		} else if (ret < 0) {
790			if (wl12xx_is_dummy_packet(wl, skb))
791				/*
792				 * fw still expects dummy packet,
793				 * so re-enqueue it
794				 */
795				wl1271_skb_queue_head(wl, wlvif, skb, hlid);
796			else
797				ieee80211_free_txskb(wl->hw, skb);
798			goto out_ack;
799		}
800		last_len = ret;
801		buf_offset += last_len;
802		wl->tx_packets_count++;
803		if (has_data) {
804			desc = (struct wl1271_tx_hw_descr *) skb->data;
805			__set_bit(desc->hlid, active_hlids);
806		}
807	}
808
809out_ack:
810	if (buf_offset) {
811		buf_offset = wlcore_hw_pre_pkt_send(wl, buf_offset, last_len);
812		bus_ret = wlcore_write_data(wl, REG_SLV_MEM_DATA, wl->aggr_buf,
813					     buf_offset, true);
814		if (bus_ret < 0)
815			goto out;
816
817		sent_packets = true;
818	}
819	if (sent_packets) {
820		/*
821		 * Interrupt the firmware with the new packets. This is only
822		 * required for older hardware revisions
823		 */
824		if (wl->quirks & WLCORE_QUIRK_END_OF_TRANSACTION) {
825			bus_ret = wlcore_write32(wl, WL12XX_HOST_WR_ACCESS,
826					     wl->tx_packets_count);
827			if (bus_ret < 0)
828				goto out;
829		}
830
831		wl1271_handle_tx_low_watermark(wl);
832	}
833	wl12xx_rearm_rx_streaming(wl, active_hlids);
834
835out:
836	return bus_ret;
837}
838
839void wl1271_tx_work(struct work_struct *work)
840{
841	struct wl1271 *wl = container_of(work, struct wl1271, tx_work);
842	int ret;
843
844	mutex_lock(&wl->mutex);
845	ret = wl1271_ps_elp_wakeup(wl);
846	if (ret < 0)
847		goto out;
848
849	ret = wlcore_tx_work_locked(wl);
850	if (ret < 0) {
851		wl12xx_queue_recovery_work(wl);
852		goto out;
853	}
854
855	wl1271_ps_elp_sleep(wl);
856out:
857	mutex_unlock(&wl->mutex);
858}
859
860static u8 wl1271_tx_get_rate_flags(u8 rate_class_index)
861{
862	u8 flags = 0;
863
864	/*
865	 * TODO: use wl12xx constants when this code is moved to wl12xx, as
866	 * only it uses Tx-completion.
867	 */
868	if (rate_class_index <= 8)
869		flags |= IEEE80211_TX_RC_MCS;
870
871	/*
872	 * TODO: use wl12xx constants when this code is moved to wl12xx, as
873	 * only it uses Tx-completion.
874	 */
875	if (rate_class_index == 0)
876		flags |= IEEE80211_TX_RC_SHORT_GI;
877
878	return flags;
879}
880
881static void wl1271_tx_complete_packet(struct wl1271 *wl,
882				      struct wl1271_tx_hw_res_descr *result)
883{
884	struct ieee80211_tx_info *info;
885	struct ieee80211_vif *vif;
886	struct wl12xx_vif *wlvif;
887	struct sk_buff *skb;
888	int id = result->id;
889	int rate = -1;
890	u8 rate_flags = 0;
891	u8 retries = 0;
892
893	/* check for id legality */
894	if (unlikely(id >= wl->num_tx_desc || wl->tx_frames[id] == NULL)) {
895		wl1271_warning("TX result illegal id: %d", id);
896		return;
897	}
898
899	skb = wl->tx_frames[id];
900	info = IEEE80211_SKB_CB(skb);
901
902	if (wl12xx_is_dummy_packet(wl, skb)) {
903		wl1271_free_tx_id(wl, id);
904		return;
905	}
906
907	/* info->control is valid as long as we don't update info->status */
908	vif = info->control.vif;
909	wlvif = wl12xx_vif_to_data(vif);
910
911	/* update the TX status info */
912	if (result->status == TX_SUCCESS) {
913		if (!(info->flags & IEEE80211_TX_CTL_NO_ACK))
914			info->flags |= IEEE80211_TX_STAT_ACK;
915		rate = wlcore_rate_to_idx(wl, result->rate_class_index,
916					  wlvif->band);
917		rate_flags = wl1271_tx_get_rate_flags(result->rate_class_index);
918		retries = result->ack_failures;
919	} else if (result->status == TX_RETRY_EXCEEDED) {
920		wl->stats.excessive_retries++;
921		retries = result->ack_failures;
922	}
923
924	info->status.rates[0].idx = rate;
925	info->status.rates[0].count = retries;
926	info->status.rates[0].flags = rate_flags;
927	info->status.ack_signal = -1;
928
929	wl->stats.retry_count += result->ack_failures;
930
931	/*
932	 * update sequence number only when relevant, i.e. only in
933	 * sessions of TKIP, AES and GEM (not in open or WEP sessions)
934	 */
935	if (info->control.hw_key &&
936	    (info->control.hw_key->cipher == WLAN_CIPHER_SUITE_TKIP ||
937	     info->control.hw_key->cipher == WLAN_CIPHER_SUITE_CCMP ||
938	     info->control.hw_key->cipher == WL1271_CIPHER_SUITE_GEM)) {
939		u8 fw_lsb = result->tx_security_sequence_number_lsb;
940		u8 cur_lsb = wlvif->tx_security_last_seq_lsb;
941
942		/*
943		 * update security sequence number, taking care of potential
944		 * wrap-around
945		 */
946		wlvif->tx_security_seq += (fw_lsb - cur_lsb) & 0xff;
947		wlvif->tx_security_last_seq_lsb = fw_lsb;
948	}
949
950	/* remove private header from packet */
951	skb_pull(skb, sizeof(struct wl1271_tx_hw_descr));
952
953	/* remove TKIP header space if present */
954	if ((wl->quirks & WLCORE_QUIRK_TKIP_HEADER_SPACE) &&
955	    info->control.hw_key &&
956	    info->control.hw_key->cipher == WLAN_CIPHER_SUITE_TKIP) {
957		int hdrlen = ieee80211_get_hdrlen_from_skb(skb);
958		memmove(skb->data + WL1271_EXTRA_SPACE_TKIP, skb->data,
959			hdrlen);
960		skb_pull(skb, WL1271_EXTRA_SPACE_TKIP);
961	}
962
963	wl1271_debug(DEBUG_TX, "tx status id %u skb 0x%p failures %u rate 0x%x"
964		     " status 0x%x",
965		     result->id, skb, result->ack_failures,
966		     result->rate_class_index, result->status);
967
968	/* return the packet to the stack */
969	skb_queue_tail(&wl->deferred_tx_queue, skb);
970	queue_work(wl->freezable_wq, &wl->netstack_work);
971	wl1271_free_tx_id(wl, result->id);
972}
973
974/* Called upon reception of a TX complete interrupt */
975int wlcore_tx_complete(struct wl1271 *wl)
976{
977	struct wl1271_acx_mem_map *memmap = wl->target_mem_map;
978	u32 count, fw_counter;
979	u32 i;
980	int ret;
981
982	/* read the tx results from the chipset */
983	ret = wlcore_read(wl, le32_to_cpu(memmap->tx_result),
984			  wl->tx_res_if, sizeof(*wl->tx_res_if), false);
985	if (ret < 0)
986		goto out;
987
988	fw_counter = le32_to_cpu(wl->tx_res_if->tx_result_fw_counter);
989
990	/* write host counter to chipset (to ack) */
991	ret = wlcore_write32(wl, le32_to_cpu(memmap->tx_result) +
992			     offsetof(struct wl1271_tx_hw_res_if,
993				      tx_result_host_counter), fw_counter);
994	if (ret < 0)
995		goto out;
996
997	count = fw_counter - wl->tx_results_count;
998	wl1271_debug(DEBUG_TX, "tx_complete received, packets: %d", count);
999
1000	/* verify that the result buffer is not getting overrun */
1001	if (unlikely(count > TX_HW_RESULT_QUEUE_LEN))
1002		wl1271_warning("TX result overflow from chipset: %d", count);
1003
1004	/* process the results */
1005	for (i = 0; i < count; i++) {
1006		struct wl1271_tx_hw_res_descr *result;
1007		u8 offset = wl->tx_results_count & TX_HW_RESULT_QUEUE_LEN_MASK;
1008
1009		/* process the packet */
1010		result =  &(wl->tx_res_if->tx_results_queue[offset]);
1011		wl1271_tx_complete_packet(wl, result);
1012
1013		wl->tx_results_count++;
1014	}
1015
1016out:
1017	return ret;
1018}
1019EXPORT_SYMBOL(wlcore_tx_complete);
1020
1021void wl1271_tx_reset_link_queues(struct wl1271 *wl, u8 hlid)
1022{
1023	struct sk_buff *skb;
1024	int i;
1025	unsigned long flags;
1026	struct ieee80211_tx_info *info;
1027	int total[NUM_TX_QUEUES];
1028	struct wl1271_link *lnk = &wl->links[hlid];
1029
1030	for (i = 0; i < NUM_TX_QUEUES; i++) {
1031		total[i] = 0;
1032		while ((skb = skb_dequeue(&lnk->tx_queue[i]))) {
1033			wl1271_debug(DEBUG_TX, "link freeing skb 0x%p", skb);
1034
1035			if (!wl12xx_is_dummy_packet(wl, skb)) {
1036				info = IEEE80211_SKB_CB(skb);
1037				info->status.rates[0].idx = -1;
1038				info->status.rates[0].count = 0;
1039				ieee80211_tx_status_ni(wl->hw, skb);
1040			}
1041
1042			total[i]++;
1043		}
1044	}
1045
1046	spin_lock_irqsave(&wl->wl_lock, flags);
1047	for (i = 0; i < NUM_TX_QUEUES; i++) {
1048		wl->tx_queue_count[i] -= total[i];
1049		if (lnk->wlvif)
1050			lnk->wlvif->tx_queue_count[i] -= total[i];
1051	}
1052	spin_unlock_irqrestore(&wl->wl_lock, flags);
1053
1054	wl1271_handle_tx_low_watermark(wl);
1055}
1056
1057/* caller must hold wl->mutex and TX must be stopped */
1058void wl12xx_tx_reset_wlvif(struct wl1271 *wl, struct wl12xx_vif *wlvif)
1059{
1060	int i;
1061
1062	/* TX failure */
1063	for_each_set_bit(i, wlvif->links_map, WL12XX_MAX_LINKS) {
1064		if (wlvif->bss_type == BSS_TYPE_AP_BSS) {
1065			/* this calls wl12xx_free_link */
1066			wl1271_free_sta(wl, wlvif, i);
1067		} else {
1068			u8 hlid = i;
1069			wlvif->sta.ba_rx_bitmap = 0;
1070			wl12xx_free_link(wl, wlvif, &hlid);
1071		}
1072	}
1073	wlvif->last_tx_hlid = 0;
1074
1075	for (i = 0; i < NUM_TX_QUEUES; i++)
1076		wlvif->tx_queue_count[i] = 0;
1077}
1078/* caller must hold wl->mutex and TX must be stopped */
1079void wl12xx_tx_reset(struct wl1271 *wl)
1080{
1081	int i;
1082	struct sk_buff *skb;
1083	struct ieee80211_tx_info *info;
1084
1085	/* only reset the queues if something bad happened */
1086	if (wl1271_tx_total_queue_count(wl) != 0) {
1087		for (i = 0; i < WL12XX_MAX_LINKS; i++)
1088			wl1271_tx_reset_link_queues(wl, i);
1089
1090		for (i = 0; i < NUM_TX_QUEUES; i++)
1091			wl->tx_queue_count[i] = 0;
1092	}
1093
1094	/*
1095	 * Make sure the driver is at a consistent state, in case this
1096	 * function is called from a context other than interface removal.
1097	 * This call will always wake the TX queues.
1098	 */
1099	wl1271_handle_tx_low_watermark(wl);
1100
1101	for (i = 0; i < wl->num_tx_desc; i++) {
1102		if (wl->tx_frames[i] == NULL)
1103			continue;
1104
1105		skb = wl->tx_frames[i];
1106		wl1271_free_tx_id(wl, i);
1107		wl1271_debug(DEBUG_TX, "freeing skb 0x%p", skb);
1108
1109		if (!wl12xx_is_dummy_packet(wl, skb)) {
1110			/*
1111			 * Remove private headers before passing the skb to
1112			 * mac80211
1113			 */
1114			info = IEEE80211_SKB_CB(skb);
1115			skb_pull(skb, sizeof(struct wl1271_tx_hw_descr));
1116			if ((wl->quirks & WLCORE_QUIRK_TKIP_HEADER_SPACE) &&
1117			    info->control.hw_key &&
1118			    info->control.hw_key->cipher ==
1119			    WLAN_CIPHER_SUITE_TKIP) {
1120				int hdrlen = ieee80211_get_hdrlen_from_skb(skb);
1121				memmove(skb->data + WL1271_EXTRA_SPACE_TKIP,
1122					skb->data, hdrlen);
1123				skb_pull(skb, WL1271_EXTRA_SPACE_TKIP);
1124			}
1125
1126			info->status.rates[0].idx = -1;
1127			info->status.rates[0].count = 0;
1128
1129			ieee80211_tx_status_ni(wl->hw, skb);
1130		}
1131	}
1132}
1133
1134#define WL1271_TX_FLUSH_TIMEOUT 500000
1135
1136/* caller must *NOT* hold wl->mutex */
1137void wl1271_tx_flush(struct wl1271 *wl)
1138{
1139	unsigned long timeout, start_time;
1140	int i;
1141	start_time = jiffies;
1142	timeout = start_time + usecs_to_jiffies(WL1271_TX_FLUSH_TIMEOUT);
1143
1144	/* only one flush should be in progress, for consistent queue state */
1145	mutex_lock(&wl->flush_mutex);
1146
1147	mutex_lock(&wl->mutex);
1148	if (wl->tx_frames_cnt == 0 && wl1271_tx_total_queue_count(wl) == 0) {
1149		mutex_unlock(&wl->mutex);
1150		goto out;
1151	}
1152
1153	wlcore_stop_queues(wl, WLCORE_QUEUE_STOP_REASON_FLUSH);
1154
1155	while (!time_after(jiffies, timeout)) {
1156		wl1271_debug(DEBUG_MAC80211, "flushing tx buffer: %d %d",
1157			     wl->tx_frames_cnt,
1158			     wl1271_tx_total_queue_count(wl));
1159
1160		/* force Tx and give the driver some time to flush data */
1161		mutex_unlock(&wl->mutex);
1162		if (wl1271_tx_total_queue_count(wl))
1163			wl1271_tx_work(&wl->tx_work);
1164		msleep(20);
1165		mutex_lock(&wl->mutex);
1166
1167		if ((wl->tx_frames_cnt == 0) &&
1168		    (wl1271_tx_total_queue_count(wl) == 0)) {
1169			wl1271_debug(DEBUG_MAC80211, "tx flush took %d ms",
1170				     jiffies_to_msecs(jiffies - start_time));
1171			goto out_wake;
1172		}
1173	}
1174
1175	wl1271_warning("Unable to flush all TX buffers, "
1176		       "timed out (timeout %d ms",
1177		       WL1271_TX_FLUSH_TIMEOUT / 1000);
1178
1179	/* forcibly flush all Tx buffers on our queues */
1180	for (i = 0; i < WL12XX_MAX_LINKS; i++)
1181		wl1271_tx_reset_link_queues(wl, i);
1182
1183out_wake:
1184	wlcore_wake_queues(wl, WLCORE_QUEUE_STOP_REASON_FLUSH);
1185	mutex_unlock(&wl->mutex);
1186out:
1187	mutex_unlock(&wl->flush_mutex);
1188}
1189EXPORT_SYMBOL_GPL(wl1271_tx_flush);
1190
1191u32 wl1271_tx_min_rate_get(struct wl1271 *wl, u32 rate_set)
1192{
1193	if (WARN_ON(!rate_set))
1194		return 0;
1195
1196	return BIT(__ffs(rate_set));
1197}
1198EXPORT_SYMBOL_GPL(wl1271_tx_min_rate_get);
1199
1200void wlcore_stop_queue_locked(struct wl1271 *wl, struct wl12xx_vif *wlvif,
1201			      u8 queue, enum wlcore_queue_stop_reason reason)
1202{
1203	int hwq = wlcore_tx_get_mac80211_queue(wlvif, queue);
1204	bool stopped = !!wl->queue_stop_reasons[hwq];
1205
1206	/* queue should not be stopped for this reason */
1207	WARN_ON_ONCE(test_and_set_bit(reason, &wl->queue_stop_reasons[hwq]));
1208
1209	if (stopped)
1210		return;
1211
1212	ieee80211_stop_queue(wl->hw, hwq);
1213}
1214
1215void wlcore_stop_queue(struct wl1271 *wl, struct wl12xx_vif *wlvif, u8 queue,
1216		       enum wlcore_queue_stop_reason reason)
1217{
1218	unsigned long flags;
1219
1220	spin_lock_irqsave(&wl->wl_lock, flags);
1221	wlcore_stop_queue_locked(wl, wlvif, queue, reason);
1222	spin_unlock_irqrestore(&wl->wl_lock, flags);
1223}
1224
1225void wlcore_wake_queue(struct wl1271 *wl, struct wl12xx_vif *wlvif, u8 queue,
1226		       enum wlcore_queue_stop_reason reason)
1227{
1228	unsigned long flags;
1229	int hwq = wlcore_tx_get_mac80211_queue(wlvif, queue);
1230
1231	spin_lock_irqsave(&wl->wl_lock, flags);
1232
1233	/* queue should not be clear for this reason */
1234	WARN_ON_ONCE(!test_and_clear_bit(reason, &wl->queue_stop_reasons[hwq]));
1235
1236	if (wl->queue_stop_reasons[hwq])
1237		goto out;
1238
1239	ieee80211_wake_queue(wl->hw, hwq);
1240
1241out:
1242	spin_unlock_irqrestore(&wl->wl_lock, flags);
1243}
1244
1245void wlcore_stop_queues(struct wl1271 *wl,
1246			enum wlcore_queue_stop_reason reason)
1247{
1248	int i;
1249	unsigned long flags;
1250
1251	spin_lock_irqsave(&wl->wl_lock, flags);
1252
1253	/* mark all possible queues as stopped */
1254        for (i = 0; i < WLCORE_NUM_MAC_ADDRESSES * NUM_TX_QUEUES; i++)
1255                WARN_ON_ONCE(test_and_set_bit(reason,
1256					      &wl->queue_stop_reasons[i]));
1257
1258	/* use the global version to make sure all vifs in mac80211 we don't
1259	 * know are stopped.
1260	 */
1261	ieee80211_stop_queues(wl->hw);
1262
1263	spin_unlock_irqrestore(&wl->wl_lock, flags);
1264}
1265
1266void wlcore_wake_queues(struct wl1271 *wl,
1267			enum wlcore_queue_stop_reason reason)
1268{
1269	int i;
1270	unsigned long flags;
1271
1272	spin_lock_irqsave(&wl->wl_lock, flags);
1273
1274	/* mark all possible queues as awake */
1275        for (i = 0; i < WLCORE_NUM_MAC_ADDRESSES * NUM_TX_QUEUES; i++)
1276		WARN_ON_ONCE(!test_and_clear_bit(reason,
1277						 &wl->queue_stop_reasons[i]));
1278
1279	/* use the global version to make sure all vifs in mac80211 we don't
1280	 * know are woken up.
1281	 */
1282	ieee80211_wake_queues(wl->hw);
1283
1284	spin_unlock_irqrestore(&wl->wl_lock, flags);
1285}
1286
1287bool wlcore_is_queue_stopped_by_reason(struct wl1271 *wl,
1288				       struct wl12xx_vif *wlvif, u8 queue,
1289				       enum wlcore_queue_stop_reason reason)
1290{
1291	int hwq = wlcore_tx_get_mac80211_queue(wlvif, queue);
1292	return test_bit(reason, &wl->queue_stop_reasons[hwq]);
1293}
1294
1295bool wlcore_is_queue_stopped(struct wl1271 *wl, struct wl12xx_vif *wlvif,
1296			     u8 queue)
1297{
1298	int hwq = wlcore_tx_get_mac80211_queue(wlvif, queue);
1299	return !!wl->queue_stop_reasons[hwq];
1300}
1301