264 lines
8.3 KiB
Diff
264 lines
8.3 KiB
Diff
--- a/ath/if_ath.c
|
|
+++ b/ath/if_ath.c
|
|
@@ -3239,7 +3239,6 @@ ath_hardstart(struct sk_buff *skb, struc
|
|
struct ath_softc *sc = dev->priv;
|
|
struct ieee80211_node *ni = NULL;
|
|
struct ath_buf *bf = NULL;
|
|
- struct ether_header *eh;
|
|
ath_bufhead bf_head;
|
|
struct ath_buf *tbf, *tempbf;
|
|
struct sk_buff *tskb;
|
|
@@ -3251,6 +3250,7 @@ ath_hardstart(struct sk_buff *skb, struc
|
|
*/
|
|
int requeue = 0;
|
|
#ifdef ATH_SUPERG_FF
|
|
+ struct ether_header *eh;
|
|
unsigned int pktlen;
|
|
struct ieee80211com *ic = &sc->sc_ic;
|
|
struct ath_node *an;
|
|
@@ -3316,27 +3316,9 @@ ath_hardstart(struct sk_buff *skb, struc
|
|
requeue = 1;
|
|
goto hardstart_fail;
|
|
}
|
|
-#endif
|
|
|
|
- /* If the skb data is shared, we will copy it so we can strip padding
|
|
- * without affecting any other bridge ports. */
|
|
- if (skb_cloned(skb)) {
|
|
- /* Remember the original SKB so we can free up our references */
|
|
- struct sk_buff *skb_new;
|
|
- skb_new = skb_copy(skb, GFP_ATOMIC);
|
|
- if (skb_new == NULL) {
|
|
- DPRINTF(sc, ATH_DEBUG_XMIT,
|
|
- "Dropping; skb_copy failure.\n");
|
|
- /* No free RAM, do not requeue! */
|
|
- goto hardstart_fail;
|
|
- }
|
|
- ieee80211_skb_copy_noderef(skb, skb_new);
|
|
- ieee80211_dev_kfree_skb(&skb);
|
|
- skb = skb_new;
|
|
- }
|
|
eh = (struct ether_header *)skb->data;
|
|
|
|
-#ifdef ATH_SUPERG_FF
|
|
/* NB: use this lock to protect an->an_tx_ffbuf (and txq->axq_stageq)
|
|
* in athff_can_aggregate() call too. */
|
|
ATH_TXQ_LOCK_IRQ(txq);
|
|
--- a/net80211/ieee80211_output.c
|
|
+++ b/net80211/ieee80211_output.c
|
|
@@ -283,7 +283,7 @@ ieee80211_hardstart(struct sk_buff *skb,
|
|
* normal vap. */
|
|
if (vap->iv_xrvap && (ni == vap->iv_bss) &&
|
|
vap->iv_xrvap->iv_sta_assoc) {
|
|
- struct sk_buff *skb1 = skb_copy(skb, GFP_ATOMIC);
|
|
+ struct sk_buff *skb1 = skb_clone(skb, GFP_ATOMIC);
|
|
if (skb1) {
|
|
memset(SKB_CB(skb1), 0, sizeof(struct ieee80211_cb));
|
|
#ifdef IEEE80211_DEBUG_REFCNT
|
|
@@ -566,7 +566,7 @@ ieee80211_skbhdr_adjust(struct ieee80211
|
|
struct ieee80211_key *key, struct sk_buff *skb, int ismulticast)
|
|
{
|
|
/* XXX pre-calculate per node? */
|
|
- int need_headroom = LLC_SNAPFRAMELEN + hdrsize + IEEE80211_ADDR_LEN;
|
|
+ int need_headroom = LLC_SNAPFRAMELEN + hdrsize;
|
|
int need_tailroom = 0;
|
|
#ifdef ATH_SUPERG_FF
|
|
int isff = ATH_FF_MAGIC_PRESENT(skb);
|
|
@@ -608,109 +608,56 @@ ieee80211_skbhdr_adjust(struct ieee80211
|
|
need_tailroom += cip->ic_miclen;
|
|
}
|
|
|
|
- if (skb_shared(skb)) {
|
|
- /* Take our own reference to the node in the clone */
|
|
- ieee80211_ref_node(SKB_CB(skb)->ni);
|
|
- /* Unshare the node, decrementing users in the old skb */
|
|
- skb = skb_unshare(skb, GFP_ATOMIC);
|
|
+ need_headroom -= skb_headroom(skb);
|
|
+ if (isff)
|
|
+ need_tailroom -= skb_tailroom(skb2);
|
|
+ else
|
|
+ need_tailroom -= skb_tailroom(skb);
|
|
+
|
|
+ if (need_headroom < 0)
|
|
+ need_headroom = 0;
|
|
+ if (need_tailroom < 0)
|
|
+ need_tailroom = 0;
|
|
+
|
|
+ if (skb_cloned(skb) || (need_headroom > 0) ||
|
|
+ (!isff && (need_tailroom > 0))) {
|
|
+
|
|
+ if (pskb_expand_head(skb, need_headroom, need_tailroom, GFP_ATOMIC)) {
|
|
+ IEEE80211_DPRINTF(vap, IEEE80211_MSG_OUTPUT,
|
|
+ "%s: cannot expand storage (tail)\n", __func__);
|
|
+ goto error;
|
|
+ }
|
|
}
|
|
|
|
#ifdef ATH_SUPERG_FF
|
|
if (isff) {
|
|
- if (skb == NULL) {
|
|
- IEEE80211_DPRINTF(vap, IEEE80211_MSG_OUTPUT,
|
|
- "%s: cannot unshare for encapsulation\n",
|
|
- __func__);
|
|
- vap->iv_stats.is_tx_nobuf++;
|
|
- ieee80211_dev_kfree_skb(&skb2);
|
|
-
|
|
- return NULL;
|
|
- }
|
|
+ inter_headroom -= skb_headroom(skb2);
|
|
+ if (inter_headroom < 0)
|
|
+ inter_headroom = 0;
|
|
+ if ((skb_cloned(skb2) ||
|
|
+ (inter_headroom > 0) || (need_tailroom > 0))) {
|
|
|
|
- /* first skb header */
|
|
- if (skb_headroom(skb) < need_headroom) {
|
|
- struct sk_buff *tmp = skb;
|
|
- skb = skb_realloc_headroom(skb, need_headroom);
|
|
- if (skb == NULL) {
|
|
+ if (pskb_expand_head(skb2, inter_headroom,
|
|
+ need_tailroom, GFP_ATOMIC)) {
|
|
IEEE80211_DPRINTF(vap, IEEE80211_MSG_OUTPUT,
|
|
- "%s: cannot expand storage (head1)\n",
|
|
- __func__);
|
|
- vap->iv_stats.is_tx_nobuf++;
|
|
- ieee80211_dev_kfree_skb(&skb2);
|
|
- return NULL;
|
|
- } else
|
|
- ieee80211_skb_copy_noderef(tmp, skb);
|
|
- ieee80211_dev_kfree_skb(&tmp);
|
|
- /* NB: cb[] area was copied, but not next ptr. must do that
|
|
- * prior to return on success. */
|
|
- }
|
|
-
|
|
- /* second skb with header and tail adjustments possible */
|
|
- if (skb_tailroom(skb2) < need_tailroom) {
|
|
- int n = 0;
|
|
- if (inter_headroom > skb_headroom(skb2))
|
|
- n = inter_headroom - skb_headroom(skb2);
|
|
- if (pskb_expand_head(skb2, n,
|
|
- need_tailroom - skb_tailroom(skb2), GFP_ATOMIC)) {
|
|
- ieee80211_dev_kfree_skb(&skb2);
|
|
- IEEE80211_DPRINTF(vap, IEEE80211_MSG_OUTPUT,
|
|
- "%s: cannot expand storage (tail2)\n",
|
|
- __func__);
|
|
- vap->iv_stats.is_tx_nobuf++;
|
|
- /* this shouldn't happen, but don't send first ff either */
|
|
- ieee80211_dev_kfree_skb(&skb);
|
|
+ "%s: cannot expand storage (tail)\n", __func__);
|
|
+ goto error;
|
|
}
|
|
- } else if (skb_headroom(skb2) < inter_headroom) {
|
|
- struct sk_buff *tmp = skb2;
|
|
-
|
|
- skb2 = skb_realloc_headroom(skb2, inter_headroom);
|
|
- if (skb2 == NULL) {
|
|
- IEEE80211_DPRINTF(vap, IEEE80211_MSG_OUTPUT,
|
|
- "%s: cannot expand storage (head2)\n",
|
|
- __func__);
|
|
- vap->iv_stats.is_tx_nobuf++;
|
|
- /* this shouldn't happen, but don't send first ff either */
|
|
- ieee80211_dev_kfree_skb(&skb);
|
|
- skb = NULL;
|
|
- } else
|
|
- ieee80211_skb_copy_noderef(tmp, skb);
|
|
- ieee80211_dev_kfree_skb(&tmp);
|
|
}
|
|
- if (skb) {
|
|
- skb->next = skb2;
|
|
- }
|
|
- return skb;
|
|
+ skb->next = skb2;
|
|
}
|
|
#endif /* ATH_SUPERG_FF */
|
|
- if (skb == NULL) {
|
|
- IEEE80211_DPRINTF(vap, IEEE80211_MSG_OUTPUT,
|
|
- "%s: cannot unshare for encapsulation\n", __func__);
|
|
- vap->iv_stats.is_tx_nobuf++;
|
|
- } else if (skb_tailroom(skb) < need_tailroom) {
|
|
- int n = 0;
|
|
- if (need_headroom > skb_headroom(skb))
|
|
- n = need_headroom - skb_headroom(skb);
|
|
- if (pskb_expand_head(skb, n, need_tailroom -
|
|
- skb_tailroom(skb), GFP_ATOMIC)) {
|
|
- IEEE80211_DPRINTF(vap, IEEE80211_MSG_OUTPUT,
|
|
- "%s: cannot expand storage (tail)\n", __func__);
|
|
- vap->iv_stats.is_tx_nobuf++;
|
|
- ieee80211_dev_kfree_skb(&skb);
|
|
- }
|
|
- } else if (skb_headroom(skb) < need_headroom) {
|
|
- struct sk_buff *tmp = skb;
|
|
- skb = skb_realloc_headroom(skb, need_headroom);
|
|
- /* Increment reference count after copy */
|
|
- if (skb == NULL) {
|
|
- IEEE80211_DPRINTF(vap, IEEE80211_MSG_OUTPUT,
|
|
- "%s: cannot expand storage (head)\n", __func__);
|
|
- vap->iv_stats.is_tx_nobuf++;
|
|
- } else
|
|
- ieee80211_skb_copy_noderef(tmp, skb);
|
|
- ieee80211_dev_kfree_skb(&tmp);
|
|
- }
|
|
|
|
return skb;
|
|
+
|
|
+error:
|
|
+ vap->iv_stats.is_tx_nobuf++;
|
|
+ ieee80211_dev_kfree_skb(&skb);
|
|
+#ifdef ATH_SUPERG_FF
|
|
+ if (skb2)
|
|
+ ieee80211_dev_kfree_skb(&skb2);
|
|
+#endif
|
|
+ return NULL;
|
|
}
|
|
|
|
#define KEY_UNDEFINED(k) ((k).wk_cipher == &ieee80211_cipher_none)
|
|
--- a/net80211/ieee80211_input.c
|
|
+++ b/net80211/ieee80211_input.c
|
|
@@ -204,7 +204,6 @@ ieee80211_input(struct ieee80211vap * va
|
|
struct ieee80211_frame *wh;
|
|
struct ieee80211_key *key;
|
|
struct ether_header *eh;
|
|
- struct sk_buff *skb2;
|
|
#ifdef ATH_SUPERG_FF
|
|
struct llc *llc;
|
|
#endif
|
|
@@ -244,20 +243,6 @@ ieee80211_input(struct ieee80211vap * va
|
|
vap->iv_stats.is_rx_tooshort++;
|
|
goto out;
|
|
}
|
|
- /* Clone the SKB... we assume somewhere in this driver that we 'own'
|
|
- * the skbuff passed into hard start and we do a lot of messing with it
|
|
- * but bridges under some cases will not clone for the first pass of skb
|
|
- * to a bridge port, but will then clone for subsequent ones. This is
|
|
- * odd behavior but it means that if we have trashed the skb we are given
|
|
- * then other ports get clones of the residual garbage.
|
|
- */
|
|
- if ((skb2 = skb_copy(skb, GFP_ATOMIC)) == NULL) {
|
|
- vap->iv_devstats.tx_dropped++;
|
|
- goto out;
|
|
- }
|
|
- ieee80211_skb_copy_noderef(skb, skb2);
|
|
- ieee80211_dev_kfree_skb(&skb);
|
|
- skb = skb2;
|
|
|
|
/*
|
|
* Bit of a cheat here, we use a pointer for a 3-address
|
|
@@ -738,7 +723,7 @@ ieee80211_input(struct ieee80211vap * va
|
|
/* ether_type must be length as FF frames are always LLC/SNAP encap'd */
|
|
frame_len = ntohs(eh_tmp->ether_type);
|
|
|
|
- skb1 = skb_copy(skb, GFP_ATOMIC);
|
|
+ skb1 = skb_clone(skb, GFP_ATOMIC);
|
|
if (skb1 == NULL)
|
|
goto err;
|
|
ieee80211_skb_copy_noderef(skb, skb1);
|
|
@@ -1137,7 +1122,7 @@ ieee80211_deliver_data(struct ieee80211_
|
|
|
|
if (ETHER_IS_MULTICAST(eh->ether_dhost) && !netif_queue_stopped(dev)) {
|
|
/* Create a SKB for the BSS to send out. */
|
|
- skb1 = skb_copy(skb, GFP_ATOMIC);
|
|
+ skb1 = skb_clone(skb, GFP_ATOMIC);
|
|
if (skb1)
|
|
SKB_CB(skb1)->ni = ieee80211_ref_node(vap->iv_bss);
|
|
}
|