2008-02-06 05:42:23 +08:00
|
|
|
/*
|
2010-08-07 02:45:38 +08:00
|
|
|
Copyright (C) 2010 Willow Garage <http://www.willowgarage.com>
|
|
|
|
Copyright (C) 2004 - 2010 Ivo van Doorn <IvDoorn@gmail.com>
|
2009-11-08 23:39:55 +08:00
|
|
|
Copyright (C) 2004 - 2009 Gertjan van Wingerde <gwingerde@gmail.com>
|
2008-02-06 05:42:23 +08:00
|
|
|
<http://rt2x00.serialmonkey.com>
|
|
|
|
|
|
|
|
This program is free software; you can redistribute it and/or modify
|
|
|
|
it under the terms of the GNU General Public License as published by
|
|
|
|
the Free Software Foundation; either version 2 of the License, or
|
|
|
|
(at your option) any later version.
|
|
|
|
|
|
|
|
This program is distributed in the hope that it will be useful,
|
|
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
GNU General Public License for more details.
|
|
|
|
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
|
|
along with this program; if not, write to the
|
|
|
|
Free Software Foundation, Inc.,
|
|
|
|
59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
|
|
|
|
*/
|
|
|
|
|
|
|
|
/*
|
|
|
|
Module: rt2x00lib
|
|
|
|
Abstract: rt2x00 queue specific routines.
|
|
|
|
*/
|
|
|
|
|
include cleanup: Update gfp.h and slab.h includes to prepare for breaking implicit slab.h inclusion from percpu.h
percpu.h is included by sched.h and module.h and thus ends up being
included when building most .c files. percpu.h includes slab.h which
in turn includes gfp.h making everything defined by the two files
universally available and complicating inclusion dependencies.
percpu.h -> slab.h dependency is about to be removed. Prepare for
this change by updating users of gfp and slab facilities include those
headers directly instead of assuming availability. As this conversion
needs to touch large number of source files, the following script is
used as the basis of conversion.
http://userweb.kernel.org/~tj/misc/slabh-sweep.py
The script does the followings.
* Scan files for gfp and slab usages and update includes such that
only the necessary includes are there. ie. if only gfp is used,
gfp.h, if slab is used, slab.h.
* When the script inserts a new include, it looks at the include
blocks and try to put the new include such that its order conforms
to its surrounding. It's put in the include block which contains
core kernel includes, in the same order that the rest are ordered -
alphabetical, Christmas tree, rev-Xmas-tree or at the end if there
doesn't seem to be any matching order.
* If the script can't find a place to put a new include (mostly
because the file doesn't have fitting include block), it prints out
an error message indicating which .h file needs to be added to the
file.
The conversion was done in the following steps.
1. The initial automatic conversion of all .c files updated slightly
over 4000 files, deleting around 700 includes and adding ~480 gfp.h
and ~3000 slab.h inclusions. The script emitted errors for ~400
files.
2. Each error was manually checked. Some didn't need the inclusion,
some needed manual addition while adding it to implementation .h or
embedding .c file was more appropriate for others. This step added
inclusions to around 150 files.
3. The script was run again and the output was compared to the edits
from #2 to make sure no file was left behind.
4. Several build tests were done and a couple of problems were fixed.
e.g. lib/decompress_*.c used malloc/free() wrappers around slab
APIs requiring slab.h to be added manually.
5. The script was run on all .h files but without automatically
editing them as sprinkling gfp.h and slab.h inclusions around .h
files could easily lead to inclusion dependency hell. Most gfp.h
inclusion directives were ignored as stuff from gfp.h was usually
wildly available and often used in preprocessor macros. Each
slab.h inclusion directive was examined and added manually as
necessary.
6. percpu.h was updated not to include slab.h.
7. Build test were done on the following configurations and failures
were fixed. CONFIG_GCOV_KERNEL was turned off for all tests (as my
distributed build env didn't work with gcov compiles) and a few
more options had to be turned off depending on archs to make things
build (like ipr on powerpc/64 which failed due to missing writeq).
* x86 and x86_64 UP and SMP allmodconfig and a custom test config.
* powerpc and powerpc64 SMP allmodconfig
* sparc and sparc64 SMP allmodconfig
* ia64 SMP allmodconfig
* s390 SMP allmodconfig
* alpha SMP allmodconfig
* um on x86_64 SMP allmodconfig
8. percpu.h modifications were reverted so that it could be applied as
a separate patch and serve as bisection point.
Given the fact that I had only a couple of failures from tests on step
6, I'm fairly confident about the coverage of this conversion patch.
If there is a breakage, it's likely to be something in one of the arch
headers which should be easily discoverable easily on most builds of
the specific arch.
Signed-off-by: Tejun Heo <tj@kernel.org>
Guess-its-ok-by: Christoph Lameter <cl@linux-foundation.org>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Lee Schermerhorn <Lee.Schermerhorn@hp.com>
2010-03-24 16:04:11 +08:00
|
|
|
#include <linux/slab.h>
|
2008-02-06 05:42:23 +08:00
|
|
|
#include <linux/kernel.h>
|
|
|
|
#include <linux/module.h>
|
2008-06-17 01:56:31 +08:00
|
|
|
#include <linux/dma-mapping.h>
|
2008-02-06 05:42:23 +08:00
|
|
|
|
|
|
|
#include "rt2x00.h"
|
|
|
|
#include "rt2x00lib.h"
|
|
|
|
|
2012-04-19 19:24:10 +08:00
|
|
|
struct sk_buff *rt2x00queue_alloc_rxskb(struct queue_entry *entry, gfp_t gfp)
|
2008-06-07 04:54:12 +08:00
|
|
|
{
|
2013-04-17 20:30:47 +08:00
|
|
|
struct data_queue *queue = entry->queue;
|
|
|
|
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
|
2008-06-17 01:56:31 +08:00
|
|
|
struct sk_buff *skb;
|
|
|
|
struct skb_frame_desc *skbdesc;
|
2008-08-04 22:37:44 +08:00
|
|
|
unsigned int frame_size;
|
|
|
|
unsigned int head_size = 0;
|
|
|
|
unsigned int tail_size = 0;
|
2008-06-07 04:54:12 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* The frame size includes descriptor size, because the
|
|
|
|
* hardware directly receive the frame into the skbuffer.
|
|
|
|
*/
|
2013-04-17 20:30:47 +08:00
|
|
|
frame_size = queue->data_size + queue->desc_size + queue->winfo_size;
|
2008-06-07 04:54:12 +08:00
|
|
|
|
|
|
|
/*
|
2008-07-04 20:56:07 +08:00
|
|
|
* The payload should be aligned to a 4-byte boundary,
|
|
|
|
* this means we need at least 3 bytes for moving the frame
|
|
|
|
* into the correct offset.
|
2008-06-07 04:54:12 +08:00
|
|
|
*/
|
2008-08-04 22:37:44 +08:00
|
|
|
head_size = 4;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* For IV/EIV/ICV assembly we must make sure there is
|
|
|
|
* at least 8 bytes bytes available in headroom for IV/EIV
|
2008-12-04 00:29:48 +08:00
|
|
|
* and 8 bytes for ICV data as tailroon.
|
2008-08-04 22:37:44 +08:00
|
|
|
*/
|
2011-04-18 21:27:06 +08:00
|
|
|
if (test_bit(CAPABILITY_HW_CRYPTO, &rt2x00dev->cap_flags)) {
|
2008-08-04 22:37:44 +08:00
|
|
|
head_size += 8;
|
2008-12-04 00:29:48 +08:00
|
|
|
tail_size += 8;
|
2008-08-04 22:37:44 +08:00
|
|
|
}
|
2008-06-07 04:54:12 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Allocate skbuffer.
|
|
|
|
*/
|
2012-04-19 19:24:10 +08:00
|
|
|
skb = __dev_alloc_skb(frame_size + head_size + tail_size, gfp);
|
2008-06-07 04:54:12 +08:00
|
|
|
if (!skb)
|
|
|
|
return NULL;
|
|
|
|
|
2008-08-04 22:37:44 +08:00
|
|
|
/*
|
|
|
|
* Make sure we not have a frame with the requested bytes
|
|
|
|
* available in the head and tail.
|
|
|
|
*/
|
|
|
|
skb_reserve(skb, head_size);
|
2008-06-07 04:54:12 +08:00
|
|
|
skb_put(skb, frame_size);
|
|
|
|
|
2008-06-17 01:56:31 +08:00
|
|
|
/*
|
|
|
|
* Populate skbdesc.
|
|
|
|
*/
|
|
|
|
skbdesc = get_skb_frame_desc(skb);
|
|
|
|
memset(skbdesc, 0, sizeof(*skbdesc));
|
|
|
|
skbdesc->entry = entry;
|
|
|
|
|
2011-04-18 21:27:06 +08:00
|
|
|
if (test_bit(REQUIRE_DMA, &rt2x00dev->cap_flags)) {
|
2013-02-13 21:27:05 +08:00
|
|
|
dma_addr_t skb_dma;
|
|
|
|
|
|
|
|
skb_dma = dma_map_single(rt2x00dev->dev, skb->data, skb->len,
|
|
|
|
DMA_FROM_DEVICE);
|
|
|
|
if (unlikely(dma_mapping_error(rt2x00dev->dev, skb_dma))) {
|
|
|
|
dev_kfree_skb_any(skb);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
skbdesc->skb_dma = skb_dma;
|
2008-06-17 01:56:31 +08:00
|
|
|
skbdesc->flags |= SKBDESC_DMA_MAPPED_RX;
|
|
|
|
}
|
|
|
|
|
2008-06-07 04:54:12 +08:00
|
|
|
return skb;
|
|
|
|
}
|
2008-06-17 01:56:08 +08:00
|
|
|
|
2013-02-13 21:27:05 +08:00
|
|
|
int rt2x00queue_map_txskb(struct queue_entry *entry)
|
2008-06-17 01:56:08 +08:00
|
|
|
{
|
2010-10-11 21:37:25 +08:00
|
|
|
struct device *dev = entry->queue->rt2x00dev->dev;
|
|
|
|
struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
|
2008-06-17 01:56:31 +08:00
|
|
|
|
2008-08-30 03:04:50 +08:00
|
|
|
skbdesc->skb_dma =
|
2010-10-11 21:37:25 +08:00
|
|
|
dma_map_single(dev, entry->skb->data, entry->skb->len, DMA_TO_DEVICE);
|
2013-02-13 21:27:05 +08:00
|
|
|
|
|
|
|
if (unlikely(dma_mapping_error(dev, skbdesc->skb_dma)))
|
|
|
|
return -ENOMEM;
|
|
|
|
|
2008-06-17 01:56:31 +08:00
|
|
|
skbdesc->flags |= SKBDESC_DMA_MAPPED_TX;
|
2013-02-13 21:27:05 +08:00
|
|
|
return 0;
|
2008-06-17 01:56:31 +08:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(rt2x00queue_map_txskb);
|
|
|
|
|
2010-10-11 21:37:25 +08:00
|
|
|
void rt2x00queue_unmap_skb(struct queue_entry *entry)
|
2008-06-17 01:56:31 +08:00
|
|
|
{
|
2010-10-11 21:37:25 +08:00
|
|
|
struct device *dev = entry->queue->rt2x00dev->dev;
|
|
|
|
struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
|
2008-06-17 01:56:31 +08:00
|
|
|
|
|
|
|
if (skbdesc->flags & SKBDESC_DMA_MAPPED_RX) {
|
2010-10-11 21:37:25 +08:00
|
|
|
dma_unmap_single(dev, skbdesc->skb_dma, entry->skb->len,
|
2008-06-17 01:56:31 +08:00
|
|
|
DMA_FROM_DEVICE);
|
|
|
|
skbdesc->flags &= ~SKBDESC_DMA_MAPPED_RX;
|
2010-10-09 19:33:43 +08:00
|
|
|
} else if (skbdesc->flags & SKBDESC_DMA_MAPPED_TX) {
|
2010-10-11 21:37:25 +08:00
|
|
|
dma_unmap_single(dev, skbdesc->skb_dma, entry->skb->len,
|
2008-06-17 01:56:31 +08:00
|
|
|
DMA_TO_DEVICE);
|
|
|
|
skbdesc->flags &= ~SKBDESC_DMA_MAPPED_TX;
|
|
|
|
}
|
|
|
|
}
|
2010-06-03 16:51:45 +08:00
|
|
|
EXPORT_SYMBOL_GPL(rt2x00queue_unmap_skb);
|
2008-06-17 01:56:31 +08:00
|
|
|
|
2010-10-11 21:37:25 +08:00
|
|
|
void rt2x00queue_free_skb(struct queue_entry *entry)
|
2008-06-17 01:56:31 +08:00
|
|
|
{
|
2010-10-11 21:37:25 +08:00
|
|
|
if (!entry->skb)
|
2008-07-05 21:11:57 +08:00
|
|
|
return;
|
|
|
|
|
2010-10-11 21:37:25 +08:00
|
|
|
rt2x00queue_unmap_skb(entry);
|
|
|
|
dev_kfree_skb_any(entry->skb);
|
|
|
|
entry->skb = NULL;
|
2008-06-17 01:56:08 +08:00
|
|
|
}
|
2008-06-07 04:54:12 +08:00
|
|
|
|
2009-08-30 02:30:45 +08:00
|
|
|
void rt2x00queue_align_frame(struct sk_buff *skb)
|
2009-04-26 22:08:50 +08:00
|
|
|
{
|
|
|
|
unsigned int frame_length = skb->len;
|
2009-08-30 02:30:45 +08:00
|
|
|
unsigned int align = ALIGN_SIZE(skb, 0);
|
2009-04-26 22:08:50 +08:00
|
|
|
|
|
|
|
if (!align)
|
|
|
|
return;
|
|
|
|
|
2009-08-30 02:30:45 +08:00
|
|
|
skb_push(skb, align);
|
|
|
|
memmove(skb->data, skb->data + align, frame_length);
|
|
|
|
skb_trim(skb, frame_length);
|
|
|
|
}
|
|
|
|
|
|
|
|
void rt2x00queue_insert_l2pad(struct sk_buff *skb, unsigned int header_length)
|
|
|
|
{
|
2009-12-05 06:47:03 +08:00
|
|
|
unsigned int payload_length = skb->len - header_length;
|
2009-08-30 02:30:45 +08:00
|
|
|
unsigned int header_align = ALIGN_SIZE(skb, 0);
|
|
|
|
unsigned int payload_align = ALIGN_SIZE(skb, header_length);
|
2009-12-05 06:47:07 +08:00
|
|
|
unsigned int l2pad = payload_length ? L2PAD_SIZE(header_length) : 0;
|
2009-08-30 02:30:45 +08:00
|
|
|
|
2009-12-05 06:47:03 +08:00
|
|
|
/*
|
|
|
|
* Adjust the header alignment if the payload needs to be moved more
|
|
|
|
* than the header.
|
|
|
|
*/
|
|
|
|
if (payload_align > header_align)
|
|
|
|
header_align += 4;
|
|
|
|
|
|
|
|
/* There is nothing to do if no alignment is needed */
|
|
|
|
if (!header_align)
|
|
|
|
return;
|
2009-08-30 02:30:45 +08:00
|
|
|
|
2009-12-05 06:47:03 +08:00
|
|
|
/* Reserve the amount of space needed in front of the frame */
|
|
|
|
skb_push(skb, header_align);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Move the header.
|
|
|
|
*/
|
|
|
|
memmove(skb->data, skb->data + header_align, header_length);
|
|
|
|
|
|
|
|
/* Move the payload, if present and if required */
|
|
|
|
if (payload_length && payload_align)
|
2009-08-30 02:30:45 +08:00
|
|
|
memmove(skb->data + header_length + l2pad,
|
2009-11-25 06:11:32 +08:00
|
|
|
skb->data + header_length + l2pad + payload_align,
|
2009-12-05 06:47:03 +08:00
|
|
|
payload_length);
|
|
|
|
|
|
|
|
/* Trim the skb to the correct size */
|
|
|
|
skb_trim(skb, header_length + l2pad + payload_length);
|
2009-04-26 22:08:50 +08:00
|
|
|
}
|
|
|
|
|
2009-08-30 02:30:45 +08:00
|
|
|
void rt2x00queue_remove_l2pad(struct sk_buff *skb, unsigned int header_length)
|
|
|
|
{
|
2010-12-13 19:33:12 +08:00
|
|
|
/*
|
|
|
|
* L2 padding is only present if the skb contains more than just the
|
|
|
|
* IEEE 802.11 header.
|
|
|
|
*/
|
|
|
|
unsigned int l2pad = (skb->len > header_length) ?
|
|
|
|
L2PAD_SIZE(header_length) : 0;
|
2009-08-30 02:30:45 +08:00
|
|
|
|
2009-12-05 06:47:02 +08:00
|
|
|
if (!l2pad)
|
2009-08-30 02:30:45 +08:00
|
|
|
return;
|
|
|
|
|
2010-12-13 19:33:12 +08:00
|
|
|
memmove(skb->data + l2pad, skb->data, header_length);
|
|
|
|
skb_pull(skb, l2pad);
|
2009-08-30 02:30:45 +08:00
|
|
|
}
|
|
|
|
|
2011-07-07 04:57:00 +08:00
|
|
|
static void rt2x00queue_create_tx_descriptor_seq(struct rt2x00_dev *rt2x00dev,
|
|
|
|
struct sk_buff *skb,
|
2008-12-20 17:58:33 +08:00
|
|
|
struct txentry_desc *txdesc)
|
|
|
|
{
|
2011-07-07 04:57:00 +08:00
|
|
|
struct ieee80211_tx_info *tx_info = IEEE80211_SKB_CB(skb);
|
|
|
|
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)skb->data;
|
2008-12-20 17:58:33 +08:00
|
|
|
struct rt2x00_intf *intf = vif_to_intf(tx_info->control.vif);
|
2012-06-01 17:29:40 +08:00
|
|
|
u16 seqno;
|
2008-12-20 17:58:33 +08:00
|
|
|
|
2011-03-04 02:39:56 +08:00
|
|
|
if (!(tx_info->flags & IEEE80211_TX_CTL_ASSIGN_SEQ))
|
2008-12-20 17:58:33 +08:00
|
|
|
return;
|
|
|
|
|
2011-03-04 02:42:01 +08:00
|
|
|
__set_bit(ENTRY_TXD_GENERATE_SEQ, &txdesc->flags);
|
|
|
|
|
2012-04-02 19:21:06 +08:00
|
|
|
if (!test_bit(REQUIRE_SW_SEQNO, &rt2x00dev->cap_flags)) {
|
|
|
|
/*
|
|
|
|
* rt2800 has a H/W (or F/W) bug, device incorrectly increase
|
|
|
|
* seqno on retransmited data (non-QOS) frames. To workaround
|
|
|
|
* the problem let's generate seqno in software if QOS is
|
|
|
|
* disabled.
|
|
|
|
*/
|
|
|
|
if (test_bit(CONFIG_QOS_DISABLED, &rt2x00dev->flags))
|
|
|
|
__clear_bit(ENTRY_TXD_GENERATE_SEQ, &txdesc->flags);
|
|
|
|
else
|
|
|
|
/* H/W will generate sequence number */
|
|
|
|
return;
|
|
|
|
}
|
2011-03-04 02:42:01 +08:00
|
|
|
|
2008-12-20 17:58:33 +08:00
|
|
|
/*
|
2011-03-04 02:42:01 +08:00
|
|
|
* The hardware is not able to insert a sequence number. Assign a
|
|
|
|
* software generated one here.
|
2008-12-20 17:58:33 +08:00
|
|
|
*
|
|
|
|
* This is wrong because beacons are not getting sequence
|
|
|
|
* numbers assigned properly.
|
|
|
|
*
|
|
|
|
* A secondary problem exists for drivers that cannot toggle
|
|
|
|
* sequence counting per-frame, since those will override the
|
|
|
|
* sequence counter given by mac80211.
|
|
|
|
*/
|
|
|
|
if (test_bit(ENTRY_TXD_FIRST_FRAGMENT, &txdesc->flags))
|
2012-06-01 17:29:40 +08:00
|
|
|
seqno = atomic_add_return(0x10, &intf->seqno);
|
|
|
|
else
|
|
|
|
seqno = atomic_read(&intf->seqno);
|
2008-12-20 17:58:33 +08:00
|
|
|
|
2012-06-01 17:29:40 +08:00
|
|
|
hdr->seq_ctrl &= cpu_to_le16(IEEE80211_SCTL_FRAG);
|
|
|
|
hdr->seq_ctrl |= cpu_to_le16(seqno);
|
2008-12-20 17:58:33 +08:00
|
|
|
}
|
|
|
|
|
2011-07-07 04:57:00 +08:00
|
|
|
static void rt2x00queue_create_tx_descriptor_plcp(struct rt2x00_dev *rt2x00dev,
|
|
|
|
struct sk_buff *skb,
|
2008-12-20 17:58:33 +08:00
|
|
|
struct txentry_desc *txdesc,
|
|
|
|
const struct rt2x00_rate *hwrate)
|
|
|
|
{
|
2011-07-07 04:57:00 +08:00
|
|
|
struct ieee80211_tx_info *tx_info = IEEE80211_SKB_CB(skb);
|
2008-12-20 17:58:33 +08:00
|
|
|
struct ieee80211_tx_rate *txrate = &tx_info->control.rates[0];
|
|
|
|
unsigned int data_length;
|
|
|
|
unsigned int duration;
|
|
|
|
unsigned int residual;
|
|
|
|
|
2011-03-04 02:43:25 +08:00
|
|
|
/*
|
|
|
|
* Determine with what IFS priority this frame should be send.
|
|
|
|
* Set ifs to IFS_SIFS when the this is not the first fragment,
|
|
|
|
* or this fragment came after RTS/CTS.
|
|
|
|
*/
|
|
|
|
if (test_bit(ENTRY_TXD_FIRST_FRAGMENT, &txdesc->flags))
|
|
|
|
txdesc->u.plcp.ifs = IFS_BACKOFF;
|
|
|
|
else
|
|
|
|
txdesc->u.plcp.ifs = IFS_SIFS;
|
|
|
|
|
2008-12-20 17:58:33 +08:00
|
|
|
/* Data length + CRC + Crypto overhead (IV/EIV/ICV/MIC) */
|
2011-07-07 04:57:00 +08:00
|
|
|
data_length = skb->len + 4;
|
|
|
|
data_length += rt2x00crypto_tx_overhead(rt2x00dev, skb);
|
2008-12-20 17:58:33 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* PLCP setup
|
|
|
|
* Length calculation depends on OFDM/CCK rate.
|
|
|
|
*/
|
2011-03-04 02:42:35 +08:00
|
|
|
txdesc->u.plcp.signal = hwrate->plcp;
|
|
|
|
txdesc->u.plcp.service = 0x04;
|
2008-12-20 17:58:33 +08:00
|
|
|
|
|
|
|
if (hwrate->flags & DEV_RATE_OFDM) {
|
2011-03-04 02:42:35 +08:00
|
|
|
txdesc->u.plcp.length_high = (data_length >> 6) & 0x3f;
|
|
|
|
txdesc->u.plcp.length_low = data_length & 0x3f;
|
2008-12-20 17:58:33 +08:00
|
|
|
} else {
|
|
|
|
/*
|
|
|
|
* Convert length to microseconds.
|
|
|
|
*/
|
|
|
|
residual = GET_DURATION_RES(data_length, hwrate->bitrate);
|
|
|
|
duration = GET_DURATION(data_length, hwrate->bitrate);
|
|
|
|
|
|
|
|
if (residual != 0) {
|
|
|
|
duration++;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Check if we need to set the Length Extension
|
|
|
|
*/
|
|
|
|
if (hwrate->bitrate == 110 && residual <= 30)
|
2011-03-04 02:42:35 +08:00
|
|
|
txdesc->u.plcp.service |= 0x80;
|
2008-12-20 17:58:33 +08:00
|
|
|
}
|
|
|
|
|
2011-03-04 02:42:35 +08:00
|
|
|
txdesc->u.plcp.length_high = (duration >> 8) & 0xff;
|
|
|
|
txdesc->u.plcp.length_low = duration & 0xff;
|
2008-12-20 17:58:33 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* When preamble is enabled we should set the
|
|
|
|
* preamble bit for the signal.
|
|
|
|
*/
|
|
|
|
if (txrate->flags & IEEE80211_TX_RC_USE_SHORT_PREAMBLE)
|
2011-03-04 02:42:35 +08:00
|
|
|
txdesc->u.plcp.signal |= 0x08;
|
2008-12-20 17:58:33 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2011-07-07 04:57:00 +08:00
|
|
|
static void rt2x00queue_create_tx_descriptor_ht(struct rt2x00_dev *rt2x00dev,
|
|
|
|
struct sk_buff *skb,
|
2011-04-18 21:33:41 +08:00
|
|
|
struct txentry_desc *txdesc,
|
2012-07-24 03:33:42 +08:00
|
|
|
struct ieee80211_sta *sta,
|
2011-04-18 21:33:41 +08:00
|
|
|
const struct rt2x00_rate *hwrate)
|
|
|
|
{
|
2011-07-07 04:57:00 +08:00
|
|
|
struct ieee80211_tx_info *tx_info = IEEE80211_SKB_CB(skb);
|
2011-04-18 21:33:41 +08:00
|
|
|
struct ieee80211_tx_rate *txrate = &tx_info->control.rates[0];
|
2011-07-07 04:57:00 +08:00
|
|
|
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)skb->data;
|
2011-09-08 20:37:19 +08:00
|
|
|
struct rt2x00_sta *sta_priv = NULL;
|
2011-04-18 21:33:41 +08:00
|
|
|
|
2012-07-24 03:33:42 +08:00
|
|
|
if (sta) {
|
2011-04-18 21:33:41 +08:00
|
|
|
txdesc->u.ht.mpdu_density =
|
2012-07-24 03:33:42 +08:00
|
|
|
sta->ht_cap.ampdu_density;
|
2011-04-18 21:33:41 +08:00
|
|
|
|
2012-07-24 03:33:42 +08:00
|
|
|
sta_priv = sta_to_rt2x00_sta(sta);
|
2011-09-08 20:37:19 +08:00
|
|
|
txdesc->u.ht.wcid = sta_priv->wcid;
|
|
|
|
}
|
|
|
|
|
2011-04-18 21:33:41 +08:00
|
|
|
/*
|
|
|
|
* If IEEE80211_TX_RC_MCS is set txrate->idx just contains the
|
|
|
|
* mcs rate to be used
|
|
|
|
*/
|
|
|
|
if (txrate->flags & IEEE80211_TX_RC_MCS) {
|
|
|
|
txdesc->u.ht.mcs = txrate->idx;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* MIMO PS should be set to 1 for STA's using dynamic SM PS
|
|
|
|
* when using more then one tx stream (>MCS7).
|
|
|
|
*/
|
2012-07-24 03:33:42 +08:00
|
|
|
if (sta && txdesc->u.ht.mcs > 7 &&
|
2013-02-12 21:21:00 +08:00
|
|
|
sta->smps_mode == IEEE80211_SMPS_DYNAMIC)
|
2011-04-18 21:33:41 +08:00
|
|
|
__set_bit(ENTRY_TXD_HT_MIMO_PS, &txdesc->flags);
|
|
|
|
} else {
|
|
|
|
txdesc->u.ht.mcs = rt2x00_get_rate_mcs(hwrate->mcs);
|
|
|
|
if (txrate->flags & IEEE80211_TX_RC_USE_SHORT_PREAMBLE)
|
|
|
|
txdesc->u.ht.mcs |= 0x08;
|
|
|
|
}
|
|
|
|
|
2012-04-04 22:15:33 +08:00
|
|
|
if (test_bit(CONFIG_HT_DISABLED, &rt2x00dev->flags)) {
|
|
|
|
if (!(tx_info->flags & IEEE80211_TX_CTL_FIRST_FRAGMENT))
|
|
|
|
txdesc->u.ht.txop = TXOP_SIFS;
|
|
|
|
else
|
|
|
|
txdesc->u.ht.txop = TXOP_BACKOFF;
|
|
|
|
|
|
|
|
/* Left zero on all other settings. */
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
txdesc->u.ht.ba_size = 7; /* FIXME: What value is needed? */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Only one STBC stream is supported for now.
|
|
|
|
*/
|
|
|
|
if (tx_info->flags & IEEE80211_TX_CTL_STBC)
|
|
|
|
txdesc->u.ht.stbc = 1;
|
|
|
|
|
2011-04-18 21:33:41 +08:00
|
|
|
/*
|
|
|
|
* This frame is eligible for an AMPDU, however, don't aggregate
|
|
|
|
* frames that are intended to probe a specific tx rate.
|
|
|
|
*/
|
|
|
|
if (tx_info->flags & IEEE80211_TX_CTL_AMPDU &&
|
|
|
|
!(tx_info->flags & IEEE80211_TX_CTL_RATE_CTRL_PROBE))
|
|
|
|
__set_bit(ENTRY_TXD_HT_AMPDU, &txdesc->flags);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Set 40Mhz mode if necessary (for legacy rates this will
|
|
|
|
* duplicate the frame to both channels).
|
|
|
|
*/
|
|
|
|
if (txrate->flags & IEEE80211_TX_RC_40_MHZ_WIDTH ||
|
|
|
|
txrate->flags & IEEE80211_TX_RC_DUP_DATA)
|
|
|
|
__set_bit(ENTRY_TXD_HT_BW_40, &txdesc->flags);
|
|
|
|
if (txrate->flags & IEEE80211_TX_RC_SHORT_GI)
|
|
|
|
__set_bit(ENTRY_TXD_HT_SHORT_GI, &txdesc->flags);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Determine IFS values
|
|
|
|
* - Use TXOP_BACKOFF for management frames except beacons
|
|
|
|
* - Use TXOP_SIFS for fragment bursts
|
|
|
|
* - Use TXOP_HTTXOP for everything else
|
|
|
|
*
|
|
|
|
* Note: rt2800 devices won't use CTS protection (if used)
|
|
|
|
* for frames not transmitted with TXOP_HTTXOP
|
|
|
|
*/
|
|
|
|
if (ieee80211_is_mgmt(hdr->frame_control) &&
|
|
|
|
!ieee80211_is_beacon(hdr->frame_control))
|
|
|
|
txdesc->u.ht.txop = TXOP_BACKOFF;
|
|
|
|
else if (!(tx_info->flags & IEEE80211_TX_CTL_FIRST_FRAGMENT))
|
|
|
|
txdesc->u.ht.txop = TXOP_SIFS;
|
|
|
|
else
|
|
|
|
txdesc->u.ht.txop = TXOP_HTTXOP;
|
|
|
|
}
|
|
|
|
|
2011-07-07 04:57:00 +08:00
|
|
|
static void rt2x00queue_create_tx_descriptor(struct rt2x00_dev *rt2x00dev,
|
|
|
|
struct sk_buff *skb,
|
2012-07-24 03:33:42 +08:00
|
|
|
struct txentry_desc *txdesc,
|
|
|
|
struct ieee80211_sta *sta)
|
2008-05-10 19:46:13 +08:00
|
|
|
{
|
2011-07-07 04:57:00 +08:00
|
|
|
struct ieee80211_tx_info *tx_info = IEEE80211_SKB_CB(skb);
|
|
|
|
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)skb->data;
|
2011-03-04 02:43:49 +08:00
|
|
|
struct ieee80211_tx_rate *txrate = &tx_info->control.rates[0];
|
|
|
|
struct ieee80211_rate *rate;
|
|
|
|
const struct rt2x00_rate *hwrate = NULL;
|
2008-05-10 19:46:13 +08:00
|
|
|
|
|
|
|
memset(txdesc, 0, sizeof(*txdesc));
|
|
|
|
|
2009-04-26 22:08:50 +08:00
|
|
|
/*
|
2010-05-04 04:43:05 +08:00
|
|
|
* Header and frame information.
|
2009-04-26 22:08:50 +08:00
|
|
|
*/
|
2011-07-07 04:57:00 +08:00
|
|
|
txdesc->length = skb->len;
|
|
|
|
txdesc->header_length = ieee80211_get_hdrlen_from_skb(skb);
|
2009-04-26 22:08:50 +08:00
|
|
|
|
2008-05-10 19:46:13 +08:00
|
|
|
/*
|
|
|
|
* Check whether this frame is to be acked.
|
|
|
|
*/
|
2008-05-15 18:55:29 +08:00
|
|
|
if (!(tx_info->flags & IEEE80211_TX_CTL_NO_ACK))
|
2008-05-10 19:46:13 +08:00
|
|
|
__set_bit(ENTRY_TXD_ACK, &txdesc->flags);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Check if this is a RTS/CTS frame
|
|
|
|
*/
|
2008-06-17 01:54:57 +08:00
|
|
|
if (ieee80211_is_rts(hdr->frame_control) ||
|
|
|
|
ieee80211_is_cts(hdr->frame_control)) {
|
2008-05-10 19:46:13 +08:00
|
|
|
__set_bit(ENTRY_TXD_BURST, &txdesc->flags);
|
2008-06-17 01:54:57 +08:00
|
|
|
if (ieee80211_is_rts(hdr->frame_control))
|
2008-05-10 19:46:13 +08:00
|
|
|
__set_bit(ENTRY_TXD_RTS_FRAME, &txdesc->flags);
|
2008-05-15 18:55:29 +08:00
|
|
|
else
|
2008-05-10 19:46:13 +08:00
|
|
|
__set_bit(ENTRY_TXD_CTS_FRAME, &txdesc->flags);
|
2008-05-15 18:55:29 +08:00
|
|
|
if (tx_info->control.rts_cts_rate_idx >= 0)
|
2008-05-15 18:55:27 +08:00
|
|
|
rate =
|
2008-05-15 18:55:29 +08:00
|
|
|
ieee80211_get_rts_cts_rate(rt2x00dev->hw, tx_info);
|
2008-05-10 19:46:13 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Determine retry information.
|
|
|
|
*/
|
2008-10-21 18:40:02 +08:00
|
|
|
txdesc->retry_limit = tx_info->control.rates[0].count - 1;
|
2008-12-03 01:20:04 +08:00
|
|
|
if (txdesc->retry_limit >= rt2x00dev->long_retry)
|
2008-05-10 19:46:13 +08:00
|
|
|
__set_bit(ENTRY_TXD_RETRY_MODE, &txdesc->flags);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Check if more fragments are pending
|
|
|
|
*/
|
2010-06-15 04:10:09 +08:00
|
|
|
if (ieee80211_has_morefrags(hdr->frame_control)) {
|
2008-05-10 19:46:13 +08:00
|
|
|
__set_bit(ENTRY_TXD_BURST, &txdesc->flags);
|
|
|
|
__set_bit(ENTRY_TXD_MORE_FRAG, &txdesc->flags);
|
|
|
|
}
|
|
|
|
|
2010-06-15 04:10:09 +08:00
|
|
|
/*
|
|
|
|
* Check if more frames (!= fragments) are pending
|
|
|
|
*/
|
|
|
|
if (tx_info->flags & IEEE80211_TX_CTL_MORE_FRAMES)
|
|
|
|
__set_bit(ENTRY_TXD_BURST, &txdesc->flags);
|
|
|
|
|
2008-05-10 19:46:13 +08:00
|
|
|
/*
|
|
|
|
* Beacons and probe responses require the tsf timestamp
|
2011-02-20 20:56:07 +08:00
|
|
|
* to be inserted into the frame.
|
2008-05-10 19:46:13 +08:00
|
|
|
*/
|
2011-02-20 20:56:07 +08:00
|
|
|
if (ieee80211_is_beacon(hdr->frame_control) ||
|
|
|
|
ieee80211_is_probe_resp(hdr->frame_control))
|
2008-05-10 19:46:13 +08:00
|
|
|
__set_bit(ENTRY_TXD_REQ_TIMESTAMP, &txdesc->flags);
|
|
|
|
|
2008-12-20 17:58:33 +08:00
|
|
|
if ((tx_info->flags & IEEE80211_TX_CTL_FIRST_FRAGMENT) &&
|
2011-03-04 02:43:25 +08:00
|
|
|
!test_bit(ENTRY_TXD_RTS_FRAME, &txdesc->flags))
|
2008-05-10 19:46:13 +08:00
|
|
|
__set_bit(ENTRY_TXD_FIRST_FRAGMENT, &txdesc->flags);
|
|
|
|
|
2008-12-20 17:59:02 +08:00
|
|
|
/*
|
|
|
|
* Determine rate modulation.
|
|
|
|
*/
|
2011-03-04 02:43:49 +08:00
|
|
|
if (txrate->flags & IEEE80211_TX_RC_GREEN_FIELD)
|
|
|
|
txdesc->rate_mode = RATE_MODE_HT_GREENFIELD;
|
|
|
|
else if (txrate->flags & IEEE80211_TX_RC_MCS)
|
|
|
|
txdesc->rate_mode = RATE_MODE_HT_MIX;
|
|
|
|
else {
|
|
|
|
rate = ieee80211_get_tx_rate(rt2x00dev->hw, tx_info);
|
|
|
|
hwrate = rt2x00_get_rate(rate->hw_value);
|
|
|
|
if (hwrate->flags & DEV_RATE_OFDM)
|
|
|
|
txdesc->rate_mode = RATE_MODE_OFDM;
|
|
|
|
else
|
|
|
|
txdesc->rate_mode = RATE_MODE_CCK;
|
|
|
|
}
|
2008-05-10 19:46:13 +08:00
|
|
|
|
2008-12-20 17:58:33 +08:00
|
|
|
/*
|
|
|
|
* Apply TX descriptor handling by components
|
|
|
|
*/
|
2011-07-07 04:57:00 +08:00
|
|
|
rt2x00crypto_create_tx_descriptor(rt2x00dev, skb, txdesc);
|
|
|
|
rt2x00queue_create_tx_descriptor_seq(rt2x00dev, skb, txdesc);
|
2011-03-04 02:42:35 +08:00
|
|
|
|
2011-04-18 21:27:06 +08:00
|
|
|
if (test_bit(REQUIRE_HT_TX_DESC, &rt2x00dev->cap_flags))
|
2011-07-07 04:57:00 +08:00
|
|
|
rt2x00queue_create_tx_descriptor_ht(rt2x00dev, skb, txdesc,
|
2012-07-24 03:33:42 +08:00
|
|
|
sta, hwrate);
|
2011-03-04 02:42:35 +08:00
|
|
|
else
|
2011-07-07 04:57:00 +08:00
|
|
|
rt2x00queue_create_tx_descriptor_plcp(rt2x00dev, skb, txdesc,
|
|
|
|
hwrate);
|
2008-05-10 19:46:13 +08:00
|
|
|
}
|
|
|
|
|
2010-06-30 03:41:05 +08:00
|
|
|
static int rt2x00queue_write_tx_data(struct queue_entry *entry,
|
|
|
|
struct txentry_desc *txdesc)
|
|
|
|
{
|
|
|
|
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* This should not happen, we already checked the entry
|
|
|
|
* was ours. When the hardware disagrees there has been
|
|
|
|
* a queue corruption!
|
|
|
|
*/
|
|
|
|
if (unlikely(rt2x00dev->ops->lib->get_entry_state &&
|
|
|
|
rt2x00dev->ops->lib->get_entry_state(entry))) {
|
2013-04-19 23:33:40 +08:00
|
|
|
rt2x00_err(rt2x00dev,
|
|
|
|
"Corrupt queue %d, accessing entry which is not ours\n"
|
|
|
|
"Please file bug report to %s\n",
|
|
|
|
entry->queue->qid, DRV_PROJECT);
|
2010-06-30 03:41:05 +08:00
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Add the requested extra tx headroom in front of the skb.
|
|
|
|
*/
|
|
|
|
skb_push(entry->skb, rt2x00dev->ops->extra_tx_headroom);
|
|
|
|
memset(entry->skb->data, 0, rt2x00dev->ops->extra_tx_headroom);
|
|
|
|
|
|
|
|
/*
|
2010-06-30 03:42:23 +08:00
|
|
|
* Call the driver's write_tx_data function, if it exists.
|
2010-06-30 03:41:05 +08:00
|
|
|
*/
|
2010-06-30 03:42:23 +08:00
|
|
|
if (rt2x00dev->ops->lib->write_tx_data)
|
|
|
|
rt2x00dev->ops->lib->write_tx_data(entry, txdesc);
|
2010-06-30 03:41:05 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Map the skb to DMA.
|
|
|
|
*/
|
2013-02-13 21:27:05 +08:00
|
|
|
if (test_bit(REQUIRE_DMA, &rt2x00dev->cap_flags) &&
|
|
|
|
rt2x00queue_map_txskb(entry))
|
|
|
|
return -ENOMEM;
|
2010-06-30 03:41:05 +08:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2008-07-09 21:12:44 +08:00
|
|
|
static void rt2x00queue_write_tx_descriptor(struct queue_entry *entry,
|
|
|
|
struct txentry_desc *txdesc)
|
2008-05-10 19:46:13 +08:00
|
|
|
{
|
2008-06-07 04:53:14 +08:00
|
|
|
struct data_queue *queue = entry->queue;
|
2008-05-10 19:46:13 +08:00
|
|
|
|
2010-08-24 01:53:39 +08:00
|
|
|
queue->rt2x00dev->ops->lib->write_tx_desc(entry, txdesc);
|
2008-05-10 19:46:13 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* All processing on the frame has been completed, this means
|
|
|
|
* it is now ready to be dumped to userspace through debugfs.
|
|
|
|
*/
|
2010-08-24 01:53:39 +08:00
|
|
|
rt2x00debug_dump_frame(queue->rt2x00dev, DUMP_FRAME_TX, entry->skb);
|
2010-05-10 03:24:22 +08:00
|
|
|
}
|
|
|
|
|
2010-11-06 22:48:23 +08:00
|
|
|
static void rt2x00queue_kick_tx_queue(struct data_queue *queue,
|
2010-05-10 03:24:22 +08:00
|
|
|
struct txentry_desc *txdesc)
|
|
|
|
{
|
2008-05-10 19:46:13 +08:00
|
|
|
/*
|
2008-06-07 04:53:14 +08:00
|
|
|
* Check if we need to kick the queue, there are however a few rules
|
2010-05-10 03:24:22 +08:00
|
|
|
* 1) Don't kick unless this is the last in frame in a burst.
|
2008-06-07 04:53:14 +08:00
|
|
|
* When the burst flag is set, this frame is always followed
|
|
|
|
* by another frame which in some way are related to eachother.
|
|
|
|
* This is true for fragments, RTS or CTS-to-self frames.
|
2010-05-10 03:24:22 +08:00
|
|
|
* 2) Rule 1 can be broken when the available entries
|
2008-06-07 04:53:14 +08:00
|
|
|
* in the queue are less then a certain threshold.
|
2008-05-10 19:46:13 +08:00
|
|
|
*/
|
2008-06-07 04:53:14 +08:00
|
|
|
if (rt2x00queue_threshold(queue) ||
|
|
|
|
!test_bit(ENTRY_TXD_BURST, &txdesc->flags))
|
2010-12-13 19:34:54 +08:00
|
|
|
queue->rt2x00dev->ops->lib->kick_queue(queue);
|
2008-05-10 19:46:13 +08:00
|
|
|
}
|
|
|
|
|
2013-01-18 00:34:32 +08:00
|
|
|
static void rt2x00queue_bar_check(struct queue_entry *entry)
|
|
|
|
{
|
|
|
|
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
|
|
|
|
struct ieee80211_bar *bar = (void *) (entry->skb->data +
|
|
|
|
rt2x00dev->ops->extra_tx_headroom);
|
|
|
|
struct rt2x00_bar_list_entry *bar_entry;
|
|
|
|
|
|
|
|
if (likely(!ieee80211_is_back_req(bar->frame_control)))
|
|
|
|
return;
|
|
|
|
|
|
|
|
bar_entry = kmalloc(sizeof(*bar_entry), GFP_ATOMIC);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If the alloc fails we still send the BAR out but just don't track
|
|
|
|
* it in our bar list. And as a result we will report it to mac80211
|
|
|
|
* back as failed.
|
|
|
|
*/
|
|
|
|
if (!bar_entry)
|
|
|
|
return;
|
|
|
|
|
|
|
|
bar_entry->entry = entry;
|
|
|
|
bar_entry->block_acked = 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Copy the relevant parts of the 802.11 BAR into out check list
|
|
|
|
* such that we can use RCU for less-overhead in the RX path since
|
|
|
|
* sending BARs and processing the according BlockAck should be
|
|
|
|
* the exception.
|
|
|
|
*/
|
|
|
|
memcpy(bar_entry->ra, bar->ra, sizeof(bar->ra));
|
|
|
|
memcpy(bar_entry->ta, bar->ta, sizeof(bar->ta));
|
|
|
|
bar_entry->control = bar->control;
|
|
|
|
bar_entry->start_seq_num = bar->start_seq_num;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Insert BAR into our BAR check list.
|
|
|
|
*/
|
|
|
|
spin_lock_bh(&rt2x00dev->bar_list_lock);
|
|
|
|
list_add_tail_rcu(&bar_entry->list, &rt2x00dev->bar_list);
|
|
|
|
spin_unlock_bh(&rt2x00dev->bar_list_lock);
|
|
|
|
}
|
|
|
|
|
2009-11-19 08:08:30 +08:00
|
|
|
int rt2x00queue_write_tx_frame(struct data_queue *queue, struct sk_buff *skb,
|
|
|
|
bool local)
|
2008-06-07 04:50:28 +08:00
|
|
|
{
|
2008-10-21 18:40:02 +08:00
|
|
|
struct ieee80211_tx_info *tx_info;
|
2011-07-07 04:56:24 +08:00
|
|
|
struct queue_entry *entry;
|
2008-06-07 04:50:28 +08:00
|
|
|
struct txentry_desc txdesc;
|
2008-06-17 01:56:54 +08:00
|
|
|
struct skb_frame_desc *skbdesc;
|
2008-10-21 18:40:02 +08:00
|
|
|
u8 rate_idx, rate_flags;
|
2011-07-07 04:56:24 +08:00
|
|
|
int ret = 0;
|
|
|
|
|
2008-06-07 04:50:28 +08:00
|
|
|
/*
|
|
|
|
* Copy all TX descriptor information into txdesc,
|
|
|
|
* after that we are free to use the skb->cb array
|
|
|
|
* for our information.
|
|
|
|
*/
|
2012-07-24 03:33:42 +08:00
|
|
|
rt2x00queue_create_tx_descriptor(queue->rt2x00dev, skb, &txdesc, NULL);
|
2008-06-07 04:50:28 +08:00
|
|
|
|
2008-06-17 01:56:54 +08:00
|
|
|
/*
|
2008-10-21 18:40:02 +08:00
|
|
|
* All information is retrieved from the skb->cb array,
|
2008-08-04 22:37:44 +08:00
|
|
|
* now we should claim ownership of the driver part of that
|
2008-10-21 18:40:02 +08:00
|
|
|
* array, preserving the bitrate index and flags.
|
2008-06-17 01:56:54 +08:00
|
|
|
*/
|
2008-10-21 18:40:02 +08:00
|
|
|
tx_info = IEEE80211_SKB_CB(skb);
|
|
|
|
rate_idx = tx_info->control.rates[0].idx;
|
|
|
|
rate_flags = tx_info->control.rates[0].flags;
|
2008-11-12 07:01:37 +08:00
|
|
|
skbdesc = get_skb_frame_desc(skb);
|
2008-06-17 01:56:54 +08:00
|
|
|
memset(skbdesc, 0, sizeof(*skbdesc));
|
2008-10-21 18:40:02 +08:00
|
|
|
skbdesc->tx_rate_idx = rate_idx;
|
|
|
|
skbdesc->tx_rate_flags = rate_flags;
|
2008-06-17 01:56:54 +08:00
|
|
|
|
2009-11-19 08:08:30 +08:00
|
|
|
if (local)
|
|
|
|
skbdesc->flags |= SKBDESC_NOT_MAC80211;
|
|
|
|
|
2008-08-04 22:37:44 +08:00
|
|
|
/*
|
|
|
|
* When hardware encryption is supported, and this frame
|
|
|
|
* is to be encrypted, we should strip the IV/EIV data from
|
tree-wide: Assorted spelling fixes
In particular, several occurances of funny versions of 'success',
'unknown', 'therefore', 'acknowledge', 'argument', 'achieve', 'address',
'beginning', 'desirable', 'separate' and 'necessary' are fixed.
Signed-off-by: Daniel Mack <daniel@caiaq.de>
Cc: Joe Perches <joe@perches.com>
Cc: Junio C Hamano <gitster@pobox.com>
Signed-off-by: Jiri Kosina <jkosina@suse.cz>
2010-02-03 08:01:28 +08:00
|
|
|
* the frame so we can provide it to the driver separately.
|
2008-08-04 22:37:44 +08:00
|
|
|
*/
|
|
|
|
if (test_bit(ENTRY_TXD_ENCRYPT, &txdesc.flags) &&
|
2008-12-03 01:20:42 +08:00
|
|
|
!test_bit(ENTRY_TXD_ENCRYPT_IV, &txdesc.flags)) {
|
2011-04-18 21:27:06 +08:00
|
|
|
if (test_bit(REQUIRE_COPY_IV, &queue->rt2x00dev->cap_flags))
|
2009-04-26 22:08:30 +08:00
|
|
|
rt2x00crypto_tx_copy_iv(skb, &txdesc);
|
2008-12-03 01:20:42 +08:00
|
|
|
else
|
2009-04-26 22:08:30 +08:00
|
|
|
rt2x00crypto_tx_remove_iv(skb, &txdesc);
|
2008-12-03 01:20:42 +08:00
|
|
|
}
|
2008-08-04 22:37:44 +08:00
|
|
|
|
2009-08-09 05:53:47 +08:00
|
|
|
/*
|
2011-03-31 09:57:33 +08:00
|
|
|
* When DMA allocation is required we should guarantee to the
|
2009-08-09 05:53:47 +08:00
|
|
|
* driver that the DMA is aligned to a 4-byte boundary.
|
|
|
|
* However some drivers require L2 padding to pad the payload
|
|
|
|
* rather then the header. This could be a requirement for
|
|
|
|
* PCI and USB devices, while header alignment only is valid
|
|
|
|
* for PCI devices.
|
|
|
|
*/
|
2011-04-18 21:27:06 +08:00
|
|
|
if (test_bit(REQUIRE_L2PAD, &queue->rt2x00dev->cap_flags))
|
2011-07-07 04:57:37 +08:00
|
|
|
rt2x00queue_insert_l2pad(skb, txdesc.header_length);
|
2011-04-18 21:27:06 +08:00
|
|
|
else if (test_bit(REQUIRE_DMA, &queue->rt2x00dev->cap_flags))
|
2011-07-07 04:57:37 +08:00
|
|
|
rt2x00queue_align_frame(skb);
|
|
|
|
|
2012-03-09 19:39:54 +08:00
|
|
|
/*
|
|
|
|
* That function must be called with bh disabled.
|
|
|
|
*/
|
2011-07-07 04:57:37 +08:00
|
|
|
spin_lock(&queue->tx_lock);
|
|
|
|
|
|
|
|
if (unlikely(rt2x00queue_full(queue))) {
|
2013-04-19 23:33:40 +08:00
|
|
|
rt2x00_err(queue->rt2x00dev, "Dropping frame due to full tx queue %d\n",
|
|
|
|
queue->qid);
|
2011-07-07 04:57:37 +08:00
|
|
|
ret = -ENOBUFS;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
entry = rt2x00queue_get_entry(queue, Q_INDEX);
|
|
|
|
|
|
|
|
if (unlikely(test_and_set_bit(ENTRY_OWNER_DEVICE_DATA,
|
|
|
|
&entry->flags))) {
|
2013-04-19 23:33:40 +08:00
|
|
|
rt2x00_err(queue->rt2x00dev,
|
|
|
|
"Arrived at non-free entry in the non-full queue %d\n"
|
|
|
|
"Please file bug report to %s\n",
|
|
|
|
queue->qid, DRV_PROJECT);
|
2011-07-07 04:57:37 +08:00
|
|
|
ret = -EINVAL;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
skbdesc->entry = entry;
|
|
|
|
entry->skb = skb;
|
2009-04-26 22:08:50 +08:00
|
|
|
|
2008-08-04 22:37:44 +08:00
|
|
|
/*
|
|
|
|
* It could be possible that the queue was corrupted and this
|
2008-11-12 07:01:37 +08:00
|
|
|
* call failed. Since we always return NETDEV_TX_OK to mac80211,
|
|
|
|
* this frame will simply be dropped.
|
2008-08-04 22:37:44 +08:00
|
|
|
*/
|
2010-06-30 03:41:05 +08:00
|
|
|
if (unlikely(rt2x00queue_write_tx_data(entry, &txdesc))) {
|
2008-08-30 03:04:26 +08:00
|
|
|
clear_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags);
|
2008-08-04 22:37:44 +08:00
|
|
|
entry->skb = NULL;
|
2011-07-07 04:56:24 +08:00
|
|
|
ret = -EIO;
|
|
|
|
goto out;
|
2008-06-07 04:50:28 +08:00
|
|
|
}
|
|
|
|
|
2013-01-18 00:34:32 +08:00
|
|
|
/*
|
|
|
|
* Put BlockAckReqs into our check list for driver BA processing.
|
|
|
|
*/
|
|
|
|
rt2x00queue_bar_check(entry);
|
|
|
|
|
2008-08-30 03:04:26 +08:00
|
|
|
set_bit(ENTRY_DATA_PENDING, &entry->flags);
|
2008-06-07 04:50:28 +08:00
|
|
|
|
2011-04-18 21:29:38 +08:00
|
|
|
rt2x00queue_index_inc(entry, Q_INDEX);
|
2008-06-07 04:50:28 +08:00
|
|
|
rt2x00queue_write_tx_descriptor(entry, &txdesc);
|
2010-11-06 22:48:23 +08:00
|
|
|
rt2x00queue_kick_tx_queue(queue, &txdesc);
|
2008-06-07 04:50:28 +08:00
|
|
|
|
2011-07-07 04:56:24 +08:00
|
|
|
out:
|
|
|
|
spin_unlock(&queue->tx_lock);
|
|
|
|
return ret;
|
2008-06-07 04:50:28 +08:00
|
|
|
}
|
|
|
|
|
2011-01-30 20:16:03 +08:00
|
|
|
int rt2x00queue_clear_beacon(struct rt2x00_dev *rt2x00dev,
|
|
|
|
struct ieee80211_vif *vif)
|
|
|
|
{
|
|
|
|
struct rt2x00_intf *intf = vif_to_intf(vif);
|
|
|
|
|
|
|
|
if (unlikely(!intf->beacon))
|
|
|
|
return -ENOBUFS;
|
|
|
|
|
|
|
|
mutex_lock(&intf->beacon_skb_mutex);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Clean up the beacon skb.
|
|
|
|
*/
|
|
|
|
rt2x00queue_free_skb(intf->beacon);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Clear beacon (single bssid devices don't need to clear the beacon
|
|
|
|
* since the beacon queue will get stopped anyway).
|
|
|
|
*/
|
|
|
|
if (rt2x00dev->ops->lib->clear_beacon)
|
|
|
|
rt2x00dev->ops->lib->clear_beacon(intf->beacon);
|
|
|
|
|
|
|
|
mutex_unlock(&intf->beacon_skb_mutex);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2011-01-30 20:16:28 +08:00
|
|
|
int rt2x00queue_update_beacon_locked(struct rt2x00_dev *rt2x00dev,
|
|
|
|
struct ieee80211_vif *vif)
|
2008-07-09 21:12:44 +08:00
|
|
|
{
|
|
|
|
struct rt2x00_intf *intf = vif_to_intf(vif);
|
|
|
|
struct skb_frame_desc *skbdesc;
|
|
|
|
struct txentry_desc txdesc;
|
|
|
|
|
|
|
|
if (unlikely(!intf->beacon))
|
|
|
|
return -ENOBUFS;
|
|
|
|
|
2009-08-09 05:55:18 +08:00
|
|
|
/*
|
|
|
|
* Clean up the beacon skb.
|
|
|
|
*/
|
2010-10-11 21:37:25 +08:00
|
|
|
rt2x00queue_free_skb(intf->beacon);
|
2009-08-09 05:55:18 +08:00
|
|
|
|
2008-07-09 21:12:44 +08:00
|
|
|
intf->beacon->skb = ieee80211_beacon_get(rt2x00dev->hw, vif);
|
2011-01-30 20:16:28 +08:00
|
|
|
if (!intf->beacon->skb)
|
2008-07-09 21:12:44 +08:00
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Copy all TX descriptor information into txdesc,
|
|
|
|
* after that we are free to use the skb->cb array
|
|
|
|
* for our information.
|
|
|
|
*/
|
2012-07-24 03:33:42 +08:00
|
|
|
rt2x00queue_create_tx_descriptor(rt2x00dev, intf->beacon->skb, &txdesc, NULL);
|
2008-07-09 21:12:44 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Fill in skb descriptor
|
|
|
|
*/
|
|
|
|
skbdesc = get_skb_frame_desc(intf->beacon->skb);
|
|
|
|
memset(skbdesc, 0, sizeof(*skbdesc));
|
|
|
|
skbdesc->entry = intf->beacon;
|
|
|
|
|
|
|
|
/*
|
2011-01-30 20:16:03 +08:00
|
|
|
* Send beacon to hardware.
|
2008-07-09 21:12:44 +08:00
|
|
|
*/
|
2010-05-09 05:40:25 +08:00
|
|
|
rt2x00dev->ops->lib->write_beacon(intf->beacon, &txdesc);
|
2008-07-09 21:12:44 +08:00
|
|
|
|
2011-01-30 20:16:28 +08:00
|
|
|
return 0;
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
int rt2x00queue_update_beacon(struct rt2x00_dev *rt2x00dev,
|
|
|
|
struct ieee80211_vif *vif)
|
|
|
|
{
|
|
|
|
struct rt2x00_intf *intf = vif_to_intf(vif);
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
mutex_lock(&intf->beacon_skb_mutex);
|
|
|
|
ret = rt2x00queue_update_beacon_locked(rt2x00dev, vif);
|
2009-08-09 05:55:18 +08:00
|
|
|
mutex_unlock(&intf->beacon_skb_mutex);
|
|
|
|
|
2011-01-30 20:16:28 +08:00
|
|
|
return ret;
|
2008-07-09 21:12:44 +08:00
|
|
|
}
|
|
|
|
|
2011-04-18 21:27:43 +08:00
|
|
|
bool rt2x00queue_for_each_entry(struct data_queue *queue,
|
2010-08-24 01:54:21 +08:00
|
|
|
enum queue_index start,
|
|
|
|
enum queue_index end,
|
2013-03-15 16:57:56 +08:00
|
|
|
void *data,
|
|
|
|
bool (*fn)(struct queue_entry *entry,
|
|
|
|
void *data))
|
2010-08-24 01:54:21 +08:00
|
|
|
{
|
|
|
|
unsigned long irqflags;
|
|
|
|
unsigned int index_start;
|
|
|
|
unsigned int index_end;
|
|
|
|
unsigned int i;
|
|
|
|
|
|
|
|
if (unlikely(start >= Q_INDEX_MAX || end >= Q_INDEX_MAX)) {
|
2013-04-19 23:33:40 +08:00
|
|
|
rt2x00_err(queue->rt2x00dev,
|
|
|
|
"Entry requested from invalid index range (%d - %d)\n",
|
|
|
|
start, end);
|
2011-04-18 21:27:43 +08:00
|
|
|
return true;
|
2010-08-24 01:54:21 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Only protect the range we are going to loop over,
|
|
|
|
* if during our loop a extra entry is set to pending
|
|
|
|
* it should not be kicked during this run, since it
|
|
|
|
* is part of another TX operation.
|
|
|
|
*/
|
2010-11-06 22:48:05 +08:00
|
|
|
spin_lock_irqsave(&queue->index_lock, irqflags);
|
2010-08-24 01:54:21 +08:00
|
|
|
index_start = queue->index[start];
|
|
|
|
index_end = queue->index[end];
|
2010-11-06 22:48:05 +08:00
|
|
|
spin_unlock_irqrestore(&queue->index_lock, irqflags);
|
2010-08-24 01:54:21 +08:00
|
|
|
|
|
|
|
/*
|
2011-03-31 09:57:33 +08:00
|
|
|
* Start from the TX done pointer, this guarantees that we will
|
2010-08-24 01:54:21 +08:00
|
|
|
* send out all frames in the correct order.
|
|
|
|
*/
|
|
|
|
if (index_start < index_end) {
|
2011-04-18 21:27:43 +08:00
|
|
|
for (i = index_start; i < index_end; i++) {
|
2013-03-15 16:57:56 +08:00
|
|
|
if (fn(&queue->entries[i], data))
|
2011-04-18 21:27:43 +08:00
|
|
|
return true;
|
|
|
|
}
|
2010-08-24 01:54:21 +08:00
|
|
|
} else {
|
2011-04-18 21:27:43 +08:00
|
|
|
for (i = index_start; i < queue->limit; i++) {
|
2013-03-15 16:57:56 +08:00
|
|
|
if (fn(&queue->entries[i], data))
|
2011-04-18 21:27:43 +08:00
|
|
|
return true;
|
|
|
|
}
|
2010-08-24 01:54:21 +08:00
|
|
|
|
2011-04-18 21:27:43 +08:00
|
|
|
for (i = 0; i < index_end; i++) {
|
2013-03-15 16:57:56 +08:00
|
|
|
if (fn(&queue->entries[i], data))
|
2011-04-18 21:27:43 +08:00
|
|
|
return true;
|
|
|
|
}
|
2010-08-24 01:54:21 +08:00
|
|
|
}
|
2011-04-18 21:27:43 +08:00
|
|
|
|
|
|
|
return false;
|
2010-08-24 01:54:21 +08:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(rt2x00queue_for_each_entry);
|
|
|
|
|
2008-02-06 05:42:23 +08:00
|
|
|
struct queue_entry *rt2x00queue_get_entry(struct data_queue *queue,
|
|
|
|
enum queue_index index)
|
|
|
|
{
|
|
|
|
struct queue_entry *entry;
|
2008-03-10 05:44:30 +08:00
|
|
|
unsigned long irqflags;
|
2008-02-06 05:42:23 +08:00
|
|
|
|
|
|
|
if (unlikely(index >= Q_INDEX_MAX)) {
|
2013-04-19 23:33:40 +08:00
|
|
|
rt2x00_err(queue->rt2x00dev, "Entry requested from invalid index type (%d)\n",
|
|
|
|
index);
|
2008-02-06 05:42:23 +08:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2010-11-06 22:48:05 +08:00
|
|
|
spin_lock_irqsave(&queue->index_lock, irqflags);
|
2008-02-06 05:42:23 +08:00
|
|
|
|
|
|
|
entry = &queue->entries[queue->index[index]];
|
|
|
|
|
2010-11-06 22:48:05 +08:00
|
|
|
spin_unlock_irqrestore(&queue->index_lock, irqflags);
|
2008-02-06 05:42:23 +08:00
|
|
|
|
|
|
|
return entry;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(rt2x00queue_get_entry);
|
|
|
|
|
2011-04-18 21:29:38 +08:00
|
|
|
void rt2x00queue_index_inc(struct queue_entry *entry, enum queue_index index)
|
2008-02-06 05:42:23 +08:00
|
|
|
{
|
2011-04-18 21:29:38 +08:00
|
|
|
struct data_queue *queue = entry->queue;
|
2008-03-10 05:44:30 +08:00
|
|
|
unsigned long irqflags;
|
|
|
|
|
2008-02-06 05:42:23 +08:00
|
|
|
if (unlikely(index >= Q_INDEX_MAX)) {
|
2013-04-19 23:33:40 +08:00
|
|
|
rt2x00_err(queue->rt2x00dev,
|
|
|
|
"Index change on invalid index type (%d)\n", index);
|
2008-02-06 05:42:23 +08:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2010-11-06 22:48:05 +08:00
|
|
|
spin_lock_irqsave(&queue->index_lock, irqflags);
|
2008-02-06 05:42:23 +08:00
|
|
|
|
|
|
|
queue->index[index]++;
|
|
|
|
if (queue->index[index] >= queue->limit)
|
|
|
|
queue->index[index] = 0;
|
|
|
|
|
2011-04-18 21:29:38 +08:00
|
|
|
entry->last_action = jiffies;
|
2010-08-31 03:15:19 +08:00
|
|
|
|
2008-02-03 22:55:21 +08:00
|
|
|
if (index == Q_INDEX) {
|
|
|
|
queue->length++;
|
|
|
|
} else if (index == Q_INDEX_DONE) {
|
|
|
|
queue->length--;
|
2008-10-18 03:16:17 +08:00
|
|
|
queue->count++;
|
2008-02-03 22:55:21 +08:00
|
|
|
}
|
2008-02-06 05:42:23 +08:00
|
|
|
|
2010-11-06 22:48:05 +08:00
|
|
|
spin_unlock_irqrestore(&queue->index_lock, irqflags);
|
2008-02-06 05:42:23 +08:00
|
|
|
}
|
|
|
|
|
2010-12-13 19:35:17 +08:00
|
|
|
void rt2x00queue_pause_queue(struct data_queue *queue)
|
|
|
|
{
|
|
|
|
if (!test_bit(DEVICE_STATE_PRESENT, &queue->rt2x00dev->flags) ||
|
|
|
|
!test_bit(QUEUE_STARTED, &queue->flags) ||
|
|
|
|
test_and_set_bit(QUEUE_PAUSED, &queue->flags))
|
|
|
|
return;
|
|
|
|
|
|
|
|
switch (queue->qid) {
|
2010-12-13 19:36:38 +08:00
|
|
|
case QID_AC_VO:
|
|
|
|
case QID_AC_VI:
|
2010-12-13 19:35:17 +08:00
|
|
|
case QID_AC_BE:
|
|
|
|
case QID_AC_BK:
|
|
|
|
/*
|
|
|
|
* For TX queues, we have to disable the queue
|
|
|
|
* inside mac80211.
|
|
|
|
*/
|
|
|
|
ieee80211_stop_queue(queue->rt2x00dev->hw, queue->qid);
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(rt2x00queue_pause_queue);
|
|
|
|
|
|
|
|
void rt2x00queue_unpause_queue(struct data_queue *queue)
|
|
|
|
{
|
|
|
|
if (!test_bit(DEVICE_STATE_PRESENT, &queue->rt2x00dev->flags) ||
|
|
|
|
!test_bit(QUEUE_STARTED, &queue->flags) ||
|
|
|
|
!test_and_clear_bit(QUEUE_PAUSED, &queue->flags))
|
|
|
|
return;
|
|
|
|
|
|
|
|
switch (queue->qid) {
|
2010-12-13 19:36:38 +08:00
|
|
|
case QID_AC_VO:
|
|
|
|
case QID_AC_VI:
|
2010-12-13 19:35:17 +08:00
|
|
|
case QID_AC_BE:
|
|
|
|
case QID_AC_BK:
|
|
|
|
/*
|
|
|
|
* For TX queues, we have to enable the queue
|
|
|
|
* inside mac80211.
|
|
|
|
*/
|
|
|
|
ieee80211_wake_queue(queue->rt2x00dev->hw, queue->qid);
|
|
|
|
break;
|
2010-12-13 19:35:40 +08:00
|
|
|
case QID_RX:
|
|
|
|
/*
|
|
|
|
* For RX we need to kick the queue now in order to
|
|
|
|
* receive frames.
|
|
|
|
*/
|
|
|
|
queue->rt2x00dev->ops->lib->kick_queue(queue);
|
2010-12-13 19:35:17 +08:00
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(rt2x00queue_unpause_queue);
|
|
|
|
|
|
|
|
void rt2x00queue_start_queue(struct data_queue *queue)
|
|
|
|
{
|
|
|
|
mutex_lock(&queue->status_lock);
|
|
|
|
|
|
|
|
if (!test_bit(DEVICE_STATE_PRESENT, &queue->rt2x00dev->flags) ||
|
|
|
|
test_and_set_bit(QUEUE_STARTED, &queue->flags)) {
|
|
|
|
mutex_unlock(&queue->status_lock);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
set_bit(QUEUE_PAUSED, &queue->flags);
|
|
|
|
|
|
|
|
queue->rt2x00dev->ops->lib->start_queue(queue);
|
|
|
|
|
|
|
|
rt2x00queue_unpause_queue(queue);
|
|
|
|
|
|
|
|
mutex_unlock(&queue->status_lock);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(rt2x00queue_start_queue);
|
|
|
|
|
|
|
|
void rt2x00queue_stop_queue(struct data_queue *queue)
|
|
|
|
{
|
|
|
|
mutex_lock(&queue->status_lock);
|
|
|
|
|
|
|
|
if (!test_and_clear_bit(QUEUE_STARTED, &queue->flags)) {
|
|
|
|
mutex_unlock(&queue->status_lock);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
rt2x00queue_pause_queue(queue);
|
|
|
|
|
|
|
|
queue->rt2x00dev->ops->lib->stop_queue(queue);
|
|
|
|
|
|
|
|
mutex_unlock(&queue->status_lock);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(rt2x00queue_stop_queue);
|
|
|
|
|
2010-12-13 19:35:40 +08:00
|
|
|
void rt2x00queue_flush_queue(struct data_queue *queue, bool drop)
|
|
|
|
{
|
|
|
|
bool started;
|
|
|
|
bool tx_queue =
|
2010-12-13 19:36:38 +08:00
|
|
|
(queue->qid == QID_AC_VO) ||
|
2010-12-13 19:35:40 +08:00
|
|
|
(queue->qid == QID_AC_VI) ||
|
2010-12-13 19:36:38 +08:00
|
|
|
(queue->qid == QID_AC_BE) ||
|
|
|
|
(queue->qid == QID_AC_BK);
|
2010-12-13 19:35:40 +08:00
|
|
|
|
|
|
|
mutex_lock(&queue->status_lock);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If the queue has been started, we must stop it temporarily
|
|
|
|
* to prevent any new frames to be queued on the device. If
|
|
|
|
* we are not dropping the pending frames, the queue must
|
|
|
|
* only be stopped in the software and not the hardware,
|
|
|
|
* otherwise the queue will never become empty on its own.
|
|
|
|
*/
|
|
|
|
started = test_bit(QUEUE_STARTED, &queue->flags);
|
|
|
|
if (started) {
|
|
|
|
/*
|
|
|
|
* Pause the queue
|
|
|
|
*/
|
|
|
|
rt2x00queue_pause_queue(queue);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If we are not supposed to drop any pending
|
|
|
|
* frames, this means we must force a start (=kick)
|
|
|
|
* to the queue to make sure the hardware will
|
|
|
|
* start transmitting.
|
|
|
|
*/
|
|
|
|
if (!drop && tx_queue)
|
|
|
|
queue->rt2x00dev->ops->lib->kick_queue(queue);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2011-04-18 21:31:02 +08:00
|
|
|
* Check if driver supports flushing, if that is the case we can
|
|
|
|
* defer the flushing to the driver. Otherwise we must use the
|
|
|
|
* alternative which just waits for the queue to become empty.
|
2010-12-13 19:35:40 +08:00
|
|
|
*/
|
2011-04-18 21:31:02 +08:00
|
|
|
if (likely(queue->rt2x00dev->ops->lib->flush_queue))
|
|
|
|
queue->rt2x00dev->ops->lib->flush_queue(queue, drop);
|
2010-12-13 19:35:40 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* The queue flush has failed...
|
|
|
|
*/
|
|
|
|
if (unlikely(!rt2x00queue_empty(queue)))
|
2013-04-19 23:33:40 +08:00
|
|
|
rt2x00_warn(queue->rt2x00dev, "Queue %d failed to flush\n",
|
|
|
|
queue->qid);
|
2010-12-13 19:35:40 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Restore the queue to the previous status
|
|
|
|
*/
|
|
|
|
if (started)
|
|
|
|
rt2x00queue_unpause_queue(queue);
|
|
|
|
|
|
|
|
mutex_unlock(&queue->status_lock);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(rt2x00queue_flush_queue);
|
|
|
|
|
2010-12-13 19:35:17 +08:00
|
|
|
void rt2x00queue_start_queues(struct rt2x00_dev *rt2x00dev)
|
|
|
|
{
|
|
|
|
struct data_queue *queue;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* rt2x00queue_start_queue will call ieee80211_wake_queue
|
|
|
|
* for each queue after is has been properly initialized.
|
|
|
|
*/
|
|
|
|
tx_queue_for_each(rt2x00dev, queue)
|
|
|
|
rt2x00queue_start_queue(queue);
|
|
|
|
|
|
|
|
rt2x00queue_start_queue(rt2x00dev->rx);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(rt2x00queue_start_queues);
|
|
|
|
|
|
|
|
void rt2x00queue_stop_queues(struct rt2x00_dev *rt2x00dev)
|
|
|
|
{
|
|
|
|
struct data_queue *queue;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* rt2x00queue_stop_queue will call ieee80211_stop_queue
|
|
|
|
* as well, but we are completely shutting doing everything
|
|
|
|
* now, so it is much safer to stop all TX queues at once,
|
|
|
|
* and use rt2x00queue_stop_queue for cleaning up.
|
|
|
|
*/
|
|
|
|
ieee80211_stop_queues(rt2x00dev->hw);
|
|
|
|
|
|
|
|
tx_queue_for_each(rt2x00dev, queue)
|
|
|
|
rt2x00queue_stop_queue(queue);
|
|
|
|
|
|
|
|
rt2x00queue_stop_queue(rt2x00dev->rx);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(rt2x00queue_stop_queues);
|
|
|
|
|
2010-12-13 19:35:40 +08:00
|
|
|
void rt2x00queue_flush_queues(struct rt2x00_dev *rt2x00dev, bool drop)
|
|
|
|
{
|
|
|
|
struct data_queue *queue;
|
|
|
|
|
|
|
|
tx_queue_for_each(rt2x00dev, queue)
|
|
|
|
rt2x00queue_flush_queue(queue, drop);
|
|
|
|
|
|
|
|
rt2x00queue_flush_queue(rt2x00dev->rx, drop);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(rt2x00queue_flush_queues);
|
|
|
|
|
2008-02-06 05:42:23 +08:00
|
|
|
static void rt2x00queue_reset(struct data_queue *queue)
|
|
|
|
{
|
2008-03-10 05:44:30 +08:00
|
|
|
unsigned long irqflags;
|
2010-08-31 03:15:19 +08:00
|
|
|
unsigned int i;
|
2008-03-10 05:44:30 +08:00
|
|
|
|
2010-11-06 22:48:05 +08:00
|
|
|
spin_lock_irqsave(&queue->index_lock, irqflags);
|
2008-02-06 05:42:23 +08:00
|
|
|
|
|
|
|
queue->count = 0;
|
|
|
|
queue->length = 0;
|
2010-08-31 03:15:19 +08:00
|
|
|
|
2011-04-18 21:29:38 +08:00
|
|
|
for (i = 0; i < Q_INDEX_MAX; i++)
|
2010-08-31 03:15:19 +08:00
|
|
|
queue->index[i] = 0;
|
2008-02-06 05:42:23 +08:00
|
|
|
|
2010-11-06 22:48:05 +08:00
|
|
|
spin_unlock_irqrestore(&queue->index_lock, irqflags);
|
2008-02-06 05:42:23 +08:00
|
|
|
}
|
|
|
|
|
2008-11-08 22:25:33 +08:00
|
|
|
void rt2x00queue_init_queues(struct rt2x00_dev *rt2x00dev)
|
2008-02-06 05:42:23 +08:00
|
|
|
{
|
|
|
|
struct data_queue *queue;
|
|
|
|
unsigned int i;
|
|
|
|
|
2008-11-08 22:25:33 +08:00
|
|
|
queue_for_each(rt2x00dev, queue) {
|
2008-02-06 05:42:23 +08:00
|
|
|
rt2x00queue_reset(queue);
|
|
|
|
|
2010-12-13 19:36:00 +08:00
|
|
|
for (i = 0; i < queue->limit; i++)
|
2008-11-08 22:25:33 +08:00
|
|
|
rt2x00dev->ops->lib->clear_entry(&queue->entries[i]);
|
2008-02-06 05:42:23 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-06-04 19:40:39 +08:00
|
|
|
static int rt2x00queue_alloc_entries(struct data_queue *queue)
|
2008-02-06 05:42:23 +08:00
|
|
|
{
|
|
|
|
struct queue_entry *entries;
|
|
|
|
unsigned int entry_size;
|
|
|
|
unsigned int i;
|
|
|
|
|
|
|
|
rt2x00queue_reset(queue);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Allocate all queue entries.
|
|
|
|
*/
|
2013-06-04 19:40:38 +08:00
|
|
|
entry_size = sizeof(*entries) + queue->priv_size;
|
2010-08-11 15:02:48 +08:00
|
|
|
entries = kcalloc(queue->limit, entry_size, GFP_KERNEL);
|
2008-02-06 05:42:23 +08:00
|
|
|
if (!entries)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
#define QUEUE_ENTRY_PRIV_OFFSET(__base, __index, __limit, __esize, __psize) \
|
2010-11-06 22:47:25 +08:00
|
|
|
(((char *)(__base)) + ((__limit) * (__esize)) + \
|
|
|
|
((__index) * (__psize)))
|
2008-02-06 05:42:23 +08:00
|
|
|
|
|
|
|
for (i = 0; i < queue->limit; i++) {
|
|
|
|
entries[i].flags = 0;
|
|
|
|
entries[i].queue = queue;
|
|
|
|
entries[i].skb = NULL;
|
|
|
|
entries[i].entry_idx = i;
|
|
|
|
entries[i].priv_data =
|
|
|
|
QUEUE_ENTRY_PRIV_OFFSET(entries, i, queue->limit,
|
2013-06-04 19:40:38 +08:00
|
|
|
sizeof(*entries), queue->priv_size);
|
2008-02-06 05:42:23 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
#undef QUEUE_ENTRY_PRIV_OFFSET
|
|
|
|
|
|
|
|
queue->entries = entries;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2010-10-11 21:37:25 +08:00
|
|
|
static void rt2x00queue_free_skbs(struct data_queue *queue)
|
2008-06-17 01:56:08 +08:00
|
|
|
{
|
|
|
|
unsigned int i;
|
|
|
|
|
|
|
|
if (!queue->entries)
|
|
|
|
return;
|
|
|
|
|
|
|
|
for (i = 0; i < queue->limit; i++) {
|
2010-10-11 21:37:25 +08:00
|
|
|
rt2x00queue_free_skb(&queue->entries[i]);
|
2008-06-17 01:56:08 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2010-10-11 21:37:25 +08:00
|
|
|
static int rt2x00queue_alloc_rxskbs(struct data_queue *queue)
|
2008-06-17 01:56:08 +08:00
|
|
|
{
|
|
|
|
unsigned int i;
|
|
|
|
struct sk_buff *skb;
|
|
|
|
|
|
|
|
for (i = 0; i < queue->limit; i++) {
|
2012-04-19 19:24:10 +08:00
|
|
|
skb = rt2x00queue_alloc_rxskb(&queue->entries[i], GFP_KERNEL);
|
2008-06-17 01:56:08 +08:00
|
|
|
if (!skb)
|
2008-06-21 04:10:53 +08:00
|
|
|
return -ENOMEM;
|
2008-06-17 01:56:08 +08:00
|
|
|
queue->entries[i].skb = skb;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2008-02-06 05:42:23 +08:00
|
|
|
int rt2x00queue_initialize(struct rt2x00_dev *rt2x00dev)
|
|
|
|
{
|
|
|
|
struct data_queue *queue;
|
|
|
|
int status;
|
|
|
|
|
2013-06-04 19:40:39 +08:00
|
|
|
status = rt2x00queue_alloc_entries(rt2x00dev->rx);
|
2008-02-06 05:42:23 +08:00
|
|
|
if (status)
|
|
|
|
goto exit;
|
|
|
|
|
|
|
|
tx_queue_for_each(rt2x00dev, queue) {
|
2013-06-04 19:40:39 +08:00
|
|
|
status = rt2x00queue_alloc_entries(queue);
|
2008-02-06 05:42:23 +08:00
|
|
|
if (status)
|
|
|
|
goto exit;
|
|
|
|
}
|
|
|
|
|
2013-06-04 19:40:39 +08:00
|
|
|
status = rt2x00queue_alloc_entries(rt2x00dev->bcn);
|
2008-02-06 05:42:23 +08:00
|
|
|
if (status)
|
|
|
|
goto exit;
|
|
|
|
|
2011-04-18 21:27:06 +08:00
|
|
|
if (test_bit(REQUIRE_ATIM_QUEUE, &rt2x00dev->cap_flags)) {
|
2013-06-04 19:40:39 +08:00
|
|
|
status = rt2x00queue_alloc_entries(rt2x00dev->atim);
|
2008-06-17 01:56:08 +08:00
|
|
|
if (status)
|
|
|
|
goto exit;
|
|
|
|
}
|
2008-02-06 05:42:23 +08:00
|
|
|
|
2010-10-11 21:37:25 +08:00
|
|
|
status = rt2x00queue_alloc_rxskbs(rt2x00dev->rx);
|
2008-02-06 05:42:23 +08:00
|
|
|
if (status)
|
|
|
|
goto exit;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
exit:
|
2013-04-19 23:33:40 +08:00
|
|
|
rt2x00_err(rt2x00dev, "Queue entries allocation failed\n");
|
2008-02-06 05:42:23 +08:00
|
|
|
|
|
|
|
rt2x00queue_uninitialize(rt2x00dev);
|
|
|
|
|
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
|
|
|
void rt2x00queue_uninitialize(struct rt2x00_dev *rt2x00dev)
|
|
|
|
{
|
|
|
|
struct data_queue *queue;
|
|
|
|
|
2010-10-11 21:37:25 +08:00
|
|
|
rt2x00queue_free_skbs(rt2x00dev->rx);
|
2008-06-17 01:56:08 +08:00
|
|
|
|
2008-02-06 05:42:23 +08:00
|
|
|
queue_for_each(rt2x00dev, queue) {
|
|
|
|
kfree(queue->entries);
|
|
|
|
queue->entries = NULL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-06-04 05:39:53 +08:00
|
|
|
static const struct data_queue_desc *
|
|
|
|
rt2x00queue_get_qdesc_by_qid(struct rt2x00_dev *rt2x00dev,
|
|
|
|
enum data_queue_qid qid)
|
|
|
|
{
|
|
|
|
switch (qid) {
|
|
|
|
case QID_RX:
|
|
|
|
return rt2x00dev->ops->rx;
|
|
|
|
|
|
|
|
case QID_AC_BE:
|
|
|
|
case QID_AC_BK:
|
|
|
|
case QID_AC_VO:
|
|
|
|
case QID_AC_VI:
|
|
|
|
return rt2x00dev->ops->tx;
|
|
|
|
|
|
|
|
case QID_BEACON:
|
|
|
|
return rt2x00dev->ops->bcn;
|
|
|
|
|
|
|
|
case QID_ATIM:
|
|
|
|
return rt2x00dev->ops->atim;
|
|
|
|
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2008-02-11 05:51:41 +08:00
|
|
|
static void rt2x00queue_init(struct rt2x00_dev *rt2x00dev,
|
|
|
|
struct data_queue *queue, enum data_queue_qid qid)
|
|
|
|
{
|
2010-12-13 19:35:17 +08:00
|
|
|
mutex_init(&queue->status_lock);
|
2011-07-07 04:56:24 +08:00
|
|
|
spin_lock_init(&queue->tx_lock);
|
2010-11-06 22:48:05 +08:00
|
|
|
spin_lock_init(&queue->index_lock);
|
2008-02-11 05:51:41 +08:00
|
|
|
|
|
|
|
queue->rt2x00dev = rt2x00dev;
|
|
|
|
queue->qid = qid;
|
2008-08-30 03:05:45 +08:00
|
|
|
queue->txop = 0;
|
2008-02-11 05:51:41 +08:00
|
|
|
queue->aifs = 2;
|
|
|
|
queue->cw_min = 5;
|
|
|
|
queue->cw_max = 10;
|
2013-06-04 05:39:53 +08:00
|
|
|
|
2013-06-04 19:40:42 +08:00
|
|
|
if (rt2x00dev->ops->queue_init) {
|
|
|
|
rt2x00dev->ops->queue_init(queue);
|
|
|
|
} else {
|
|
|
|
const struct data_queue_desc *qdesc;
|
|
|
|
|
|
|
|
qdesc = rt2x00queue_get_qdesc_by_qid(rt2x00dev, qid);
|
|
|
|
BUG_ON(!qdesc);
|
2013-06-04 05:39:53 +08:00
|
|
|
|
2013-06-04 19:40:42 +08:00
|
|
|
queue->limit = qdesc->entry_num;
|
|
|
|
queue->data_size = qdesc->data_size;
|
|
|
|
queue->desc_size = qdesc->desc_size;
|
|
|
|
queue->winfo_size = qdesc->winfo_size;
|
|
|
|
queue->priv_size = qdesc->priv_size;
|
|
|
|
}
|
2013-06-04 19:40:41 +08:00
|
|
|
|
|
|
|
queue->threshold = DIV_ROUND_UP(queue->limit, 10);
|
2008-02-11 05:51:41 +08:00
|
|
|
}
|
|
|
|
|
2008-02-06 05:42:23 +08:00
|
|
|
int rt2x00queue_allocate(struct rt2x00_dev *rt2x00dev)
|
|
|
|
{
|
|
|
|
struct data_queue *queue;
|
|
|
|
enum data_queue_qid qid;
|
|
|
|
unsigned int req_atim =
|
2011-04-18 21:27:06 +08:00
|
|
|
!!test_bit(REQUIRE_ATIM_QUEUE, &rt2x00dev->cap_flags);
|
2008-02-06 05:42:23 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* We need the following queues:
|
|
|
|
* RX: 1
|
2008-05-10 19:43:33 +08:00
|
|
|
* TX: ops->tx_queues
|
2008-02-06 05:42:23 +08:00
|
|
|
* Beacon: 1
|
|
|
|
* Atim: 1 (if required)
|
|
|
|
*/
|
2008-05-10 19:43:33 +08:00
|
|
|
rt2x00dev->data_queues = 2 + rt2x00dev->ops->tx_queues + req_atim;
|
2008-02-06 05:42:23 +08:00
|
|
|
|
2010-08-11 15:02:48 +08:00
|
|
|
queue = kcalloc(rt2x00dev->data_queues, sizeof(*queue), GFP_KERNEL);
|
2008-02-06 05:42:23 +08:00
|
|
|
if (!queue) {
|
2013-04-19 23:33:40 +08:00
|
|
|
rt2x00_err(rt2x00dev, "Queue allocation failed\n");
|
2008-02-06 05:42:23 +08:00
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Initialize pointers
|
|
|
|
*/
|
|
|
|
rt2x00dev->rx = queue;
|
|
|
|
rt2x00dev->tx = &queue[1];
|
2008-05-10 19:43:33 +08:00
|
|
|
rt2x00dev->bcn = &queue[1 + rt2x00dev->ops->tx_queues];
|
2011-03-04 02:46:09 +08:00
|
|
|
rt2x00dev->atim = req_atim ? &queue[2 + rt2x00dev->ops->tx_queues] : NULL;
|
2008-02-06 05:42:23 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Initialize queue parameters.
|
|
|
|
* RX: qid = QID_RX
|
2010-12-13 19:36:38 +08:00
|
|
|
* TX: qid = QID_AC_VO + index
|
2008-02-06 05:42:23 +08:00
|
|
|
* TX: cw_min: 2^5 = 32.
|
|
|
|
* TX: cw_max: 2^10 = 1024.
|
2008-06-04 02:29:05 +08:00
|
|
|
* BCN: qid = QID_BEACON
|
|
|
|
* ATIM: qid = QID_ATIM
|
2008-02-06 05:42:23 +08:00
|
|
|
*/
|
2008-02-11 05:51:41 +08:00
|
|
|
rt2x00queue_init(rt2x00dev, rt2x00dev->rx, QID_RX);
|
2008-02-06 05:42:23 +08:00
|
|
|
|
2010-12-13 19:36:38 +08:00
|
|
|
qid = QID_AC_VO;
|
2008-02-11 05:51:41 +08:00
|
|
|
tx_queue_for_each(rt2x00dev, queue)
|
|
|
|
rt2x00queue_init(rt2x00dev, queue, qid++);
|
2008-02-06 05:42:23 +08:00
|
|
|
|
2011-03-04 02:46:09 +08:00
|
|
|
rt2x00queue_init(rt2x00dev, rt2x00dev->bcn, QID_BEACON);
|
2008-02-06 05:42:23 +08:00
|
|
|
if (req_atim)
|
2011-03-04 02:46:09 +08:00
|
|
|
rt2x00queue_init(rt2x00dev, rt2x00dev->atim, QID_ATIM);
|
2008-02-06 05:42:23 +08:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
void rt2x00queue_free(struct rt2x00_dev *rt2x00dev)
|
|
|
|
{
|
|
|
|
kfree(rt2x00dev->rx);
|
|
|
|
rt2x00dev->rx = NULL;
|
|
|
|
rt2x00dev->tx = NULL;
|
|
|
|
rt2x00dev->bcn = NULL;
|
|
|
|
}
|