net: rename vlan_tx_* helpers since "tx" is misleading there
The same macros are used for rx as well. So rename it. Signed-off-by: Jiri Pirko <jiri@resnulli.us> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
d8b9605d26
commit
df8a39defa
|
@ -279,8 +279,8 @@ Possible BPF extensions are shown in the following table:
|
|||
hatype skb->dev->type
|
||||
rxhash skb->hash
|
||||
cpu raw_smp_processor_id()
|
||||
vlan_tci vlan_tx_tag_get(skb)
|
||||
vlan_pr vlan_tx_tag_present(skb)
|
||||
vlan_tci skb_vlan_tag_get(skb)
|
||||
vlan_pr skb_vlan_tag_present(skb)
|
||||
rand prandom_u32()
|
||||
|
||||
These extensions can also be prefixed with '#'.
|
||||
|
|
|
@ -373,11 +373,11 @@ static int nes_nic_send(struct sk_buff *skb, struct net_device *netdev)
|
|||
wqe_fragment_length = (__le16 *)&nic_sqe->wqe_words[NES_NIC_SQ_WQE_LENGTH_0_TAG_IDX];
|
||||
|
||||
/* setup the VLAN tag if present */
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
nes_debug(NES_DBG_NIC_TX, "%s: VLAN packet to send... VLAN = %08X\n",
|
||||
netdev->name, vlan_tx_tag_get(skb));
|
||||
netdev->name, skb_vlan_tag_get(skb));
|
||||
wqe_misc = NES_NIC_SQ_WQE_TAGVALUE_ENABLE;
|
||||
wqe_fragment_length[0] = (__force __le16) vlan_tx_tag_get(skb);
|
||||
wqe_fragment_length[0] = (__force __le16) skb_vlan_tag_get(skb);
|
||||
} else
|
||||
wqe_misc = 0;
|
||||
|
||||
|
@ -576,11 +576,12 @@ static int nes_netdev_start_xmit(struct sk_buff *skb, struct net_device *netdev)
|
|||
wqe_fragment_length =
|
||||
(__le16 *)&nic_sqe->wqe_words[NES_NIC_SQ_WQE_LENGTH_0_TAG_IDX];
|
||||
/* setup the VLAN tag if present */
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
nes_debug(NES_DBG_NIC_TX, "%s: VLAN packet to send... VLAN = %08X\n",
|
||||
netdev->name, vlan_tx_tag_get(skb) );
|
||||
netdev->name,
|
||||
skb_vlan_tag_get(skb));
|
||||
wqe_misc = NES_NIC_SQ_WQE_TAGVALUE_ENABLE;
|
||||
wqe_fragment_length[0] = (__force __le16) vlan_tx_tag_get(skb);
|
||||
wqe_fragment_length[0] = (__force __le16) skb_vlan_tag_get(skb);
|
||||
} else
|
||||
wqe_misc = 0;
|
||||
|
||||
|
|
|
@ -769,11 +769,11 @@ typhoon_start_tx(struct sk_buff *skb, struct net_device *dev)
|
|||
first_txd->processFlags |= TYPHOON_TX_PF_IP_CHKSUM;
|
||||
}
|
||||
|
||||
if(vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
first_txd->processFlags |=
|
||||
TYPHOON_TX_PF_INSERT_VLAN | TYPHOON_TX_PF_VLAN_PRIORITY;
|
||||
first_txd->processFlags |=
|
||||
cpu_to_le32(htons(vlan_tx_tag_get(skb)) <<
|
||||
cpu_to_le32(htons(skb_vlan_tag_get(skb)) <<
|
||||
TYPHOON_TX_PF_VLAN_TAG_SHIFT);
|
||||
}
|
||||
|
||||
|
|
|
@ -2429,9 +2429,9 @@ static netdev_tx_t ace_start_xmit(struct sk_buff *skb,
|
|||
flagsize = (skb->len << 16) | (BD_FLG_END);
|
||||
if (skb->ip_summed == CHECKSUM_PARTIAL)
|
||||
flagsize |= BD_FLG_TCP_UDP_SUM;
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
flagsize |= BD_FLG_VLAN_TAG;
|
||||
vlan_tag = vlan_tx_tag_get(skb);
|
||||
vlan_tag = skb_vlan_tag_get(skb);
|
||||
}
|
||||
desc = ap->tx_ring + idx;
|
||||
idx = (idx + 1) % ACE_TX_RING_ENTRIES(ap);
|
||||
|
@ -2450,9 +2450,9 @@ static netdev_tx_t ace_start_xmit(struct sk_buff *skb,
|
|||
flagsize = (skb_headlen(skb) << 16);
|
||||
if (skb->ip_summed == CHECKSUM_PARTIAL)
|
||||
flagsize |= BD_FLG_TCP_UDP_SUM;
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
flagsize |= BD_FLG_VLAN_TAG;
|
||||
vlan_tag = vlan_tx_tag_get(skb);
|
||||
vlan_tag = skb_vlan_tag_get(skb);
|
||||
}
|
||||
|
||||
ace_load_tx_bd(ap, ap->tx_ring + idx, mapping, flagsize, vlan_tag);
|
||||
|
|
|
@ -1299,11 +1299,11 @@ static netdev_tx_t amd8111e_start_xmit(struct sk_buff *skb,
|
|||
lp->tx_ring[tx_index].tx_flags = 0;
|
||||
|
||||
#if AMD8111E_VLAN_TAG_USED
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
lp->tx_ring[tx_index].tag_ctrl_cmd |=
|
||||
cpu_to_le16(TCC_VLAN_INSERT);
|
||||
lp->tx_ring[tx_index].tag_ctrl_info =
|
||||
cpu_to_le16(vlan_tx_tag_get(skb));
|
||||
cpu_to_le16(skb_vlan_tag_get(skb));
|
||||
|
||||
}
|
||||
#endif
|
||||
|
|
|
@ -1165,8 +1165,8 @@ static void xgbe_prep_tx_tstamp(struct xgbe_prv_data *pdata,
|
|||
|
||||
static void xgbe_prep_vlan(struct sk_buff *skb, struct xgbe_packet_data *packet)
|
||||
{
|
||||
if (vlan_tx_tag_present(skb))
|
||||
packet->vlan_ctag = vlan_tx_tag_get(skb);
|
||||
if (skb_vlan_tag_present(skb))
|
||||
packet->vlan_ctag = skb_vlan_tag_get(skb);
|
||||
}
|
||||
|
||||
static int xgbe_prep_tso(struct sk_buff *skb, struct xgbe_packet_data *packet)
|
||||
|
@ -1247,9 +1247,9 @@ static void xgbe_packet_info(struct xgbe_prv_data *pdata,
|
|||
XGMAC_SET_BITS(packet->attributes, TX_PACKET_ATTRIBUTES,
|
||||
CSUM_ENABLE, 1);
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
/* VLAN requires an extra descriptor if tag is different */
|
||||
if (vlan_tx_tag_get(skb) != ring->tx.cur_vlan_ctag)
|
||||
if (skb_vlan_tag_get(skb) != ring->tx.cur_vlan_ctag)
|
||||
/* We can share with the TSO context descriptor */
|
||||
if (!context_desc) {
|
||||
context_desc = 1;
|
||||
|
|
|
@ -2235,8 +2235,8 @@ static netdev_tx_t atl1c_xmit_frame(struct sk_buff *skb,
|
|||
return NETDEV_TX_OK;
|
||||
}
|
||||
|
||||
if (unlikely(vlan_tx_tag_present(skb))) {
|
||||
u16 vlan = vlan_tx_tag_get(skb);
|
||||
if (unlikely(skb_vlan_tag_present(skb))) {
|
||||
u16 vlan = skb_vlan_tag_get(skb);
|
||||
__le16 tag;
|
||||
|
||||
vlan = cpu_to_le16(vlan);
|
||||
|
|
|
@ -1892,8 +1892,8 @@ static netdev_tx_t atl1e_xmit_frame(struct sk_buff *skb,
|
|||
|
||||
tpd = atl1e_get_tpd(adapter);
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
u16 vlan_tag = vlan_tx_tag_get(skb);
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
u16 vlan_tag = skb_vlan_tag_get(skb);
|
||||
u16 atl1e_vlan_tag;
|
||||
|
||||
tpd->word3 |= 1 << TPD_INS_VL_TAG_SHIFT;
|
||||
|
|
|
@ -2415,8 +2415,8 @@ static netdev_tx_t atl1_xmit_frame(struct sk_buff *skb,
|
|||
(u16) atomic_read(&tpd_ring->next_to_use));
|
||||
memset(ptpd, 0, sizeof(struct tx_packet_desc));
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
vlan_tag = vlan_tx_tag_get(skb);
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
vlan_tag = skb_vlan_tag_get(skb);
|
||||
vlan_tag = (vlan_tag << 4) | (vlan_tag >> 13) |
|
||||
((vlan_tag >> 9) & 0x8);
|
||||
ptpd->word3 |= 1 << TPD_INS_VL_TAG_SHIFT;
|
||||
|
|
|
@ -887,8 +887,8 @@ static netdev_tx_t atl2_xmit_frame(struct sk_buff *skb,
|
|||
offset = ((u32)(skb->len-copy_len + 3) & ~3);
|
||||
}
|
||||
#ifdef NETIF_F_HW_VLAN_CTAG_TX
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
u16 vlan_tag = vlan_tx_tag_get(skb);
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
u16 vlan_tag = skb_vlan_tag_get(skb);
|
||||
vlan_tag = (vlan_tag << 4) |
|
||||
(vlan_tag >> 13) |
|
||||
((vlan_tag >> 9) & 0x8);
|
||||
|
|
|
@ -6597,9 +6597,9 @@ bnx2_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
|||
vlan_tag_flags |= TX_BD_FLAGS_TCP_UDP_CKSUM;
|
||||
}
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
vlan_tag_flags |=
|
||||
(TX_BD_FLAGS_VLAN_TAG | (vlan_tx_tag_get(skb) << 16));
|
||||
(TX_BD_FLAGS_VLAN_TAG | (skb_vlan_tag_get(skb) << 16));
|
||||
}
|
||||
|
||||
if ((mss = skb_shinfo(skb)->gso_size)) {
|
||||
|
|
|
@ -3865,9 +3865,9 @@ netdev_tx_t bnx2x_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
|||
"sending pkt %u @%p next_idx %u bd %u @%p\n",
|
||||
pkt_prod, tx_buf, txdata->tx_pkt_prod, bd_prod, tx_start_bd);
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
tx_start_bd->vlan_or_ethertype =
|
||||
cpu_to_le16(vlan_tx_tag_get(skb));
|
||||
cpu_to_le16(skb_vlan_tag_get(skb));
|
||||
tx_start_bd->bd_flags.as_bitfield |=
|
||||
(X_ETH_OUTBAND_VLAN << ETH_TX_BD_FLAGS_VLAN_MODE_SHIFT);
|
||||
} else {
|
||||
|
|
|
@ -8002,9 +8002,9 @@ static netdev_tx_t tg3_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
|||
!mss && skb->len > VLAN_ETH_FRAME_LEN)
|
||||
base_flags |= TXD_FLAG_JMB_PKT;
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
base_flags |= TXD_FLAG_VLAN;
|
||||
vlan = vlan_tx_tag_get(skb);
|
||||
vlan = skb_vlan_tag_get(skb);
|
||||
}
|
||||
|
||||
if ((unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)) &&
|
||||
|
|
|
@ -2824,8 +2824,8 @@ bnad_txq_wi_prepare(struct bnad *bnad, struct bna_tcb *tcb,
|
|||
u32 gso_size;
|
||||
u16 vlan_tag = 0;
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
vlan_tag = (u16)vlan_tx_tag_get(skb);
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
vlan_tag = (u16)skb_vlan_tag_get(skb);
|
||||
flags |= (BNA_TXQ_WI_CF_INS_PRIO | BNA_TXQ_WI_CF_INS_VLAN);
|
||||
}
|
||||
if (test_bit(BNAD_RF_CEE_RUNNING, &bnad->run_flags)) {
|
||||
|
|
|
@ -1860,9 +1860,9 @@ netdev_tx_t t1_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
|||
}
|
||||
cpl->iff = dev->if_port;
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
cpl->vlan_valid = 1;
|
||||
cpl->vlan = htons(vlan_tx_tag_get(skb));
|
||||
cpl->vlan = htons(skb_vlan_tag_get(skb));
|
||||
st->vlan_insert++;
|
||||
} else
|
||||
cpl->vlan_valid = 0;
|
||||
|
|
|
@ -1148,8 +1148,8 @@ static void write_tx_pkt_wr(struct adapter *adap, struct sk_buff *skb,
|
|||
cpl->len = htonl(skb->len);
|
||||
cntrl = V_TXPKT_INTF(pi->port_id);
|
||||
|
||||
if (vlan_tx_tag_present(skb))
|
||||
cntrl |= F_TXPKT_VLAN_VLD | V_TXPKT_VLAN(vlan_tx_tag_get(skb));
|
||||
if (skb_vlan_tag_present(skb))
|
||||
cntrl |= F_TXPKT_VLAN_VLD | V_TXPKT_VLAN(skb_vlan_tag_get(skb));
|
||||
|
||||
tso_info = V_LSO_MSS(skb_shinfo(skb)->gso_size);
|
||||
if (tso_info) {
|
||||
|
@ -1282,7 +1282,7 @@ netdev_tx_t t3_eth_xmit(struct sk_buff *skb, struct net_device *dev)
|
|||
qs->port_stats[SGE_PSTAT_TX_CSUM]++;
|
||||
if (skb_shinfo(skb)->gso_size)
|
||||
qs->port_stats[SGE_PSTAT_TSO]++;
|
||||
if (vlan_tx_tag_present(skb))
|
||||
if (skb_vlan_tag_present(skb))
|
||||
qs->port_stats[SGE_PSTAT_VLANINS]++;
|
||||
|
||||
/*
|
||||
|
|
|
@ -1154,9 +1154,9 @@ out_free: dev_kfree_skb_any(skb);
|
|||
cntrl = TXPKT_L4CSUM_DIS | TXPKT_IPCSUM_DIS;
|
||||
}
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
q->vlan_ins++;
|
||||
cntrl |= TXPKT_VLAN_VLD | TXPKT_VLAN(vlan_tx_tag_get(skb));
|
||||
cntrl |= TXPKT_VLAN_VLD | TXPKT_VLAN(skb_vlan_tag_get(skb));
|
||||
}
|
||||
|
||||
cpl->ctrl0 = htonl(TXPKT_OPCODE(CPL_TX_PKT_XT) |
|
||||
|
|
|
@ -1326,9 +1326,9 @@ int t4vf_eth_xmit(struct sk_buff *skb, struct net_device *dev)
|
|||
* If there's a VLAN tag present, add that to the list of things to
|
||||
* do in this Work Request.
|
||||
*/
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
txq->vlan_ins++;
|
||||
cntrl |= TXPKT_VLAN_VLD | TXPKT_VLAN(vlan_tx_tag_get(skb));
|
||||
cntrl |= TXPKT_VLAN_VLD | TXPKT_VLAN(skb_vlan_tag_get(skb));
|
||||
}
|
||||
|
||||
/*
|
||||
|
|
|
@ -520,10 +520,10 @@ static inline void enic_queue_wq_skb(struct enic *enic,
|
|||
int loopback = 0;
|
||||
int err;
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
/* VLAN tag from trunking driver */
|
||||
vlan_tag_insert = 1;
|
||||
vlan_tag = vlan_tx_tag_get(skb);
|
||||
vlan_tag = skb_vlan_tag_get(skb);
|
||||
} else if (enic->loop_enable) {
|
||||
vlan_tag = enic->loop_tag;
|
||||
loopback = 1;
|
||||
|
|
|
@ -694,7 +694,7 @@ static inline u16 be_get_tx_vlan_tag(struct be_adapter *adapter,
|
|||
u8 vlan_prio;
|
||||
u16 vlan_tag;
|
||||
|
||||
vlan_tag = vlan_tx_tag_get(skb);
|
||||
vlan_tag = skb_vlan_tag_get(skb);
|
||||
vlan_prio = (vlan_tag & VLAN_PRIO_MASK) >> VLAN_PRIO_SHIFT;
|
||||
/* If vlan priority provided by OS is NOT in available bmap */
|
||||
if (!(adapter->vlan_prio_bmap & (1 << vlan_prio)))
|
||||
|
@ -745,7 +745,7 @@ static void wrb_fill_hdr(struct be_adapter *adapter, struct be_eth_hdr_wrb *hdr,
|
|||
SET_TX_WRB_HDR_BITS(udpcs, hdr, 1);
|
||||
}
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
SET_TX_WRB_HDR_BITS(vlan, hdr, 1);
|
||||
vlan_tag = be_get_tx_vlan_tag(adapter, skb);
|
||||
SET_TX_WRB_HDR_BITS(vlan_tag, hdr, vlan_tag);
|
||||
|
@ -864,7 +864,7 @@ static struct sk_buff *be_insert_vlan_in_pkt(struct be_adapter *adapter,
|
|||
if (unlikely(!skb))
|
||||
return skb;
|
||||
|
||||
if (vlan_tx_tag_present(skb))
|
||||
if (skb_vlan_tag_present(skb))
|
||||
vlan_tag = be_get_tx_vlan_tag(adapter, skb);
|
||||
|
||||
if (qnq_async_evt_rcvd(adapter) && adapter->pvid) {
|
||||
|
@ -923,7 +923,7 @@ static bool be_ipv6_exthdr_check(struct sk_buff *skb)
|
|||
|
||||
static int be_vlan_tag_tx_chk(struct be_adapter *adapter, struct sk_buff *skb)
|
||||
{
|
||||
return vlan_tx_tag_present(skb) || adapter->pvid || adapter->qnq_vid;
|
||||
return skb_vlan_tag_present(skb) || adapter->pvid || adapter->qnq_vid;
|
||||
}
|
||||
|
||||
static int be_ipv6_tx_stall_chk(struct be_adapter *adapter, struct sk_buff *skb)
|
||||
|
@ -946,7 +946,7 @@ static struct sk_buff *be_lancer_xmit_workarounds(struct be_adapter *adapter,
|
|||
eth_hdr_len = ntohs(skb->protocol) == ETH_P_8021Q ?
|
||||
VLAN_ETH_HLEN : ETH_HLEN;
|
||||
if (skb->len <= 60 &&
|
||||
(lancer_chip(adapter) || vlan_tx_tag_present(skb)) &&
|
||||
(lancer_chip(adapter) || skb_vlan_tag_present(skb)) &&
|
||||
is_ipv4_pkt(skb)) {
|
||||
ip = (struct iphdr *)ip_hdr(skb);
|
||||
pskb_trim(skb, eth_hdr_len + ntohs(ip->tot_len));
|
||||
|
@ -964,7 +964,7 @@ static struct sk_buff *be_lancer_xmit_workarounds(struct be_adapter *adapter,
|
|||
* Manually insert VLAN in pkt.
|
||||
*/
|
||||
if (skb->ip_summed != CHECKSUM_PARTIAL &&
|
||||
vlan_tx_tag_present(skb)) {
|
||||
skb_vlan_tag_present(skb)) {
|
||||
skb = be_insert_vlan_in_pkt(adapter, skb, skip_hw_vlan);
|
||||
if (unlikely(!skb))
|
||||
goto err;
|
||||
|
|
|
@ -2170,7 +2170,7 @@ static inline void gfar_tx_checksum(struct sk_buff *skb, struct txfcb *fcb,
|
|||
void inline gfar_tx_vlan(struct sk_buff *skb, struct txfcb *fcb)
|
||||
{
|
||||
fcb->flags |= TXFCB_VLN;
|
||||
fcb->vlctl = vlan_tx_tag_get(skb);
|
||||
fcb->vlctl = skb_vlan_tag_get(skb);
|
||||
}
|
||||
|
||||
static inline struct txbd8 *skip_txbd(struct txbd8 *bdp, int stride,
|
||||
|
@ -2230,7 +2230,7 @@ static int gfar_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
|||
regs = tx_queue->grp->regs;
|
||||
|
||||
do_csum = (CHECKSUM_PARTIAL == skb->ip_summed);
|
||||
do_vlan = vlan_tx_tag_present(skb);
|
||||
do_vlan = skb_vlan_tag_present(skb);
|
||||
do_tstamp = (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) &&
|
||||
priv->hwts_tx_en;
|
||||
|
||||
|
|
|
@ -2064,9 +2064,9 @@ static int ehea_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
|||
memset(swqe, 0, SWQE_HEADER_SIZE);
|
||||
atomic_dec(&pr->swqe_avail);
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
swqe->tx_control |= EHEA_SWQE_VLAN_INSERT;
|
||||
swqe->vlan_tag = vlan_tx_tag_get(skb);
|
||||
swqe->vlan_tag = skb_vlan_tag_get(skb);
|
||||
}
|
||||
|
||||
pr->tx_packets++;
|
||||
|
|
|
@ -3226,9 +3226,10 @@ static netdev_tx_t e1000_xmit_frame(struct sk_buff *skb,
|
|||
return NETDEV_TX_BUSY;
|
||||
}
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
tx_flags |= E1000_TX_FLAGS_VLAN;
|
||||
tx_flags |= (vlan_tx_tag_get(skb) << E1000_TX_FLAGS_VLAN_SHIFT);
|
||||
tx_flags |= (skb_vlan_tag_get(skb) <<
|
||||
E1000_TX_FLAGS_VLAN_SHIFT);
|
||||
}
|
||||
|
||||
first = tx_ring->next_to_use;
|
||||
|
|
|
@ -5463,8 +5463,8 @@ static int e1000_transfer_dhcp_info(struct e1000_adapter *adapter,
|
|||
struct e1000_hw *hw = &adapter->hw;
|
||||
u16 length, offset;
|
||||
|
||||
if (vlan_tx_tag_present(skb) &&
|
||||
!((vlan_tx_tag_get(skb) == adapter->hw.mng_cookie.vlan_id) &&
|
||||
if (skb_vlan_tag_present(skb) &&
|
||||
!((skb_vlan_tag_get(skb) == adapter->hw.mng_cookie.vlan_id) &&
|
||||
(adapter->hw.mng_cookie.status &
|
||||
E1000_MNG_DHCP_COOKIE_STATUS_VLAN)))
|
||||
return 0;
|
||||
|
@ -5603,9 +5603,10 @@ static netdev_tx_t e1000_xmit_frame(struct sk_buff *skb,
|
|||
if (e1000_maybe_stop_tx(tx_ring, count + 2))
|
||||
return NETDEV_TX_BUSY;
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
tx_flags |= E1000_TX_FLAGS_VLAN;
|
||||
tx_flags |= (vlan_tx_tag_get(skb) << E1000_TX_FLAGS_VLAN_SHIFT);
|
||||
tx_flags |= (skb_vlan_tag_get(skb) <<
|
||||
E1000_TX_FLAGS_VLAN_SHIFT);
|
||||
}
|
||||
|
||||
first = tx_ring->next_to_use;
|
||||
|
|
|
@ -965,8 +965,8 @@ static void fm10k_tx_map(struct fm10k_ring *tx_ring,
|
|||
tx_desc = FM10K_TX_DESC(tx_ring, i);
|
||||
|
||||
/* add HW VLAN tag */
|
||||
if (vlan_tx_tag_present(skb))
|
||||
tx_desc->vlan = cpu_to_le16(vlan_tx_tag_get(skb));
|
||||
if (skb_vlan_tag_present(skb))
|
||||
tx_desc->vlan = cpu_to_le16(skb_vlan_tag_get(skb));
|
||||
else
|
||||
tx_desc->vlan = 0;
|
||||
|
||||
|
|
|
@ -609,7 +609,7 @@ static netdev_tx_t fm10k_xmit_frame(struct sk_buff *skb, struct net_device *dev)
|
|||
int err;
|
||||
|
||||
if ((skb->protocol == htons(ETH_P_8021Q)) &&
|
||||
!vlan_tx_tag_present(skb)) {
|
||||
!skb_vlan_tag_present(skb)) {
|
||||
/* FM10K only supports hardware tagging, any tags in frame
|
||||
* are considered 2nd level or "outer" tags
|
||||
*/
|
||||
|
|
|
@ -1772,8 +1772,8 @@ static int i40e_tx_prepare_vlan_flags(struct sk_buff *skb,
|
|||
u32 tx_flags = 0;
|
||||
|
||||
/* if we have a HW VLAN tag being added, default to the HW one */
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
tx_flags |= vlan_tx_tag_get(skb) << I40E_TX_FLAGS_VLAN_SHIFT;
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
tx_flags |= skb_vlan_tag_get(skb) << I40E_TX_FLAGS_VLAN_SHIFT;
|
||||
tx_flags |= I40E_TX_FLAGS_HW_VLAN;
|
||||
/* else if it is a SW VLAN, check the next protocol and store the tag */
|
||||
} else if (protocol == htons(ETH_P_8021Q)) {
|
||||
|
|
|
@ -1122,8 +1122,8 @@ static int i40e_tx_prepare_vlan_flags(struct sk_buff *skb,
|
|||
u32 tx_flags = 0;
|
||||
|
||||
/* if we have a HW VLAN tag being added, default to the HW one */
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
tx_flags |= vlan_tx_tag_get(skb) << I40E_TX_FLAGS_VLAN_SHIFT;
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
tx_flags |= skb_vlan_tag_get(skb) << I40E_TX_FLAGS_VLAN_SHIFT;
|
||||
tx_flags |= I40E_TX_FLAGS_HW_VLAN;
|
||||
/* else if it is a SW VLAN, check the next protocol and store the tag */
|
||||
} else if (protocol == htons(ETH_P_8021Q)) {
|
||||
|
|
|
@ -5035,9 +5035,9 @@ netdev_tx_t igb_xmit_frame_ring(struct sk_buff *skb,
|
|||
|
||||
skb_tx_timestamp(skb);
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
tx_flags |= IGB_TX_FLAGS_VLAN;
|
||||
tx_flags |= (vlan_tx_tag_get(skb) << IGB_TX_FLAGS_VLAN_SHIFT);
|
||||
tx_flags |= (skb_vlan_tag_get(skb) << IGB_TX_FLAGS_VLAN_SHIFT);
|
||||
}
|
||||
|
||||
/* record initial flags and protocol */
|
||||
|
|
|
@ -2234,9 +2234,10 @@ static netdev_tx_t igbvf_xmit_frame_ring_adv(struct sk_buff *skb,
|
|||
return NETDEV_TX_BUSY;
|
||||
}
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
tx_flags |= IGBVF_TX_FLAGS_VLAN;
|
||||
tx_flags |= (vlan_tx_tag_get(skb) << IGBVF_TX_FLAGS_VLAN_SHIFT);
|
||||
tx_flags |= (skb_vlan_tag_get(skb) <<
|
||||
IGBVF_TX_FLAGS_VLAN_SHIFT);
|
||||
}
|
||||
|
||||
if (skb->protocol == htons(ETH_P_IP))
|
||||
|
|
|
@ -1532,9 +1532,9 @@ ixgb_xmit_frame(struct sk_buff *skb, struct net_device *netdev)
|
|||
DESC_NEEDED)))
|
||||
return NETDEV_TX_BUSY;
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
tx_flags |= IXGB_TX_FLAGS_VLAN;
|
||||
vlan_id = vlan_tx_tag_get(skb);
|
||||
vlan_id = skb_vlan_tag_get(skb);
|
||||
}
|
||||
|
||||
first = adapter->tx_ring.next_to_use;
|
||||
|
|
|
@ -7217,8 +7217,8 @@ netdev_tx_t ixgbe_xmit_frame_ring(struct sk_buff *skb,
|
|||
first->gso_segs = 1;
|
||||
|
||||
/* if we have a HW VLAN tag being added default to the HW one */
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
tx_flags |= vlan_tx_tag_get(skb) << IXGBE_TX_FLAGS_VLAN_SHIFT;
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
tx_flags |= skb_vlan_tag_get(skb) << IXGBE_TX_FLAGS_VLAN_SHIFT;
|
||||
tx_flags |= IXGBE_TX_FLAGS_HW_VLAN;
|
||||
/* else if it is a SW VLAN check the next protocol and store the tag */
|
||||
} else if (protocol == htons(ETH_P_8021Q)) {
|
||||
|
|
|
@ -3452,8 +3452,8 @@ static int ixgbevf_xmit_frame(struct sk_buff *skb, struct net_device *netdev)
|
|||
first->bytecount = skb->len;
|
||||
first->gso_segs = 1;
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
tx_flags |= vlan_tx_tag_get(skb);
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
tx_flags |= skb_vlan_tag_get(skb);
|
||||
tx_flags <<= IXGBE_TX_FLAGS_VLAN_SHIFT;
|
||||
tx_flags |= IXGBE_TX_FLAGS_VLAN;
|
||||
}
|
||||
|
|
|
@ -2154,9 +2154,9 @@ jme_tx_csum(struct jme_adapter *jme, struct sk_buff *skb, u8 *flags)
|
|||
static inline void
|
||||
jme_tx_vlan(struct sk_buff *skb, __le16 *vlan, u8 *flags)
|
||||
{
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
*flags |= TXFLAG_TAGON;
|
||||
*vlan = cpu_to_le16(vlan_tx_tag_get(skb));
|
||||
*vlan = cpu_to_le16(skb_vlan_tag_get(skb));
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -1895,14 +1895,14 @@ static netdev_tx_t sky2_xmit_frame(struct sk_buff *skb,
|
|||
ctrl = 0;
|
||||
|
||||
/* Add VLAN tag, can piggyback on LRGLEN or ADDR64 */
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
if (!le) {
|
||||
le = get_tx_le(sky2, &slot);
|
||||
le->addr = 0;
|
||||
le->opcode = OP_VLAN|HW_OWNER;
|
||||
} else
|
||||
le->opcode |= OP_VLAN;
|
||||
le->length = cpu_to_be16(vlan_tx_tag_get(skb));
|
||||
le->length = cpu_to_be16(skb_vlan_tag_get(skb));
|
||||
ctrl |= INS_VLAN;
|
||||
}
|
||||
|
||||
|
@ -2594,7 +2594,7 @@ static struct sk_buff *sky2_receive(struct net_device *dev,
|
|||
sky2->rx_next = (sky2->rx_next + 1) % sky2->rx_pending;
|
||||
prefetch(sky2->rx_ring + sky2->rx_next);
|
||||
|
||||
if (vlan_tx_tag_present(re->skb))
|
||||
if (skb_vlan_tag_present(re->skb))
|
||||
count -= VLAN_HLEN; /* Account for vlan tag */
|
||||
|
||||
/* This chip has hardware problems that generates bogus status.
|
||||
|
|
|
@ -682,8 +682,8 @@ u16 mlx4_en_select_queue(struct net_device *dev, struct sk_buff *skb,
|
|||
if (dev->num_tc)
|
||||
return skb_tx_hash(dev, skb);
|
||||
|
||||
if (vlan_tx_tag_present(skb))
|
||||
up = vlan_tx_tag_get(skb) >> VLAN_PRIO_SHIFT;
|
||||
if (skb_vlan_tag_present(skb))
|
||||
up = skb_vlan_tag_get(skb) >> VLAN_PRIO_SHIFT;
|
||||
|
||||
return fallback(dev, skb) % rings_p_up + up * rings_p_up;
|
||||
}
|
||||
|
@ -742,8 +742,8 @@ netdev_tx_t mlx4_en_xmit(struct sk_buff *skb, struct net_device *dev)
|
|||
goto tx_drop;
|
||||
}
|
||||
|
||||
if (vlan_tx_tag_present(skb))
|
||||
vlan_tag = vlan_tx_tag_get(skb);
|
||||
if (skb_vlan_tag_present(skb))
|
||||
vlan_tag = skb_vlan_tag_get(skb);
|
||||
|
||||
|
||||
netdev_txq_bql_enqueue_prefetchw(ring->tx_queue);
|
||||
|
@ -930,7 +930,7 @@ netdev_tx_t mlx4_en_xmit(struct sk_buff *skb, struct net_device *dev)
|
|||
real_size = (real_size / 16) & 0x3f;
|
||||
|
||||
if (ring->bf_enabled && desc_size <= MAX_BF && !bounce &&
|
||||
!vlan_tx_tag_present(skb) && send_doorbell) {
|
||||
!skb_vlan_tag_present(skb) && send_doorbell) {
|
||||
tx_desc->ctrl.bf_qpn = ring->doorbell_qpn |
|
||||
cpu_to_be32(real_size);
|
||||
|
||||
|
@ -952,7 +952,7 @@ netdev_tx_t mlx4_en_xmit(struct sk_buff *skb, struct net_device *dev)
|
|||
} else {
|
||||
tx_desc->ctrl.vlan_tag = cpu_to_be16(vlan_tag);
|
||||
tx_desc->ctrl.ins_vlan = MLX4_WQE_CTRL_INS_VLAN *
|
||||
!!vlan_tx_tag_present(skb);
|
||||
!!skb_vlan_tag_present(skb);
|
||||
tx_desc->ctrl.fence_size = real_size;
|
||||
|
||||
/* Ensure new descriptor hits memory
|
||||
|
|
|
@ -1122,12 +1122,12 @@ static netdev_tx_t ns83820_hard_start_xmit(struct sk_buff *skb,
|
|||
}
|
||||
|
||||
#ifdef NS83820_VLAN_ACCEL_SUPPORT
|
||||
if(vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
/* fetch the vlan tag info out of the
|
||||
* ancillary data if the vlan code
|
||||
* is using hw vlan acceleration
|
||||
*/
|
||||
short tag = vlan_tx_tag_get(skb);
|
||||
short tag = skb_vlan_tag_get(skb);
|
||||
extsts |= (EXTSTS_VPKT | htons(tag));
|
||||
}
|
||||
#endif
|
||||
|
|
|
@ -4045,8 +4045,8 @@ static netdev_tx_t s2io_xmit(struct sk_buff *skb, struct net_device *dev)
|
|||
}
|
||||
|
||||
queue = 0;
|
||||
if (vlan_tx_tag_present(skb))
|
||||
vlan_tag = vlan_tx_tag_get(skb);
|
||||
if (skb_vlan_tag_present(skb))
|
||||
vlan_tag = skb_vlan_tag_get(skb);
|
||||
if (sp->config.tx_steering_type == TX_DEFAULT_STEERING) {
|
||||
if (skb->protocol == htons(ETH_P_IP)) {
|
||||
struct iphdr *ip;
|
||||
|
|
|
@ -890,8 +890,8 @@ vxge_xmit(struct sk_buff *skb, struct net_device *dev)
|
|||
dev->name, __func__, __LINE__,
|
||||
fifo_hw, dtr, dtr_priv);
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
u16 vlan_tag = vlan_tx_tag_get(skb);
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
u16 vlan_tag = skb_vlan_tag_get(skb);
|
||||
vxge_hw_fifo_txdl_vlan_set(dtr, vlan_tag);
|
||||
}
|
||||
|
||||
|
|
|
@ -2462,9 +2462,9 @@ static netdev_tx_t nv_start_xmit_optimized(struct sk_buff *skb,
|
|||
NV_TX2_CHECKSUM_L3 | NV_TX2_CHECKSUM_L4 : 0;
|
||||
|
||||
/* vlan tag */
|
||||
if (vlan_tx_tag_present(skb))
|
||||
if (skb_vlan_tag_present(skb))
|
||||
start_tx->txvlan = cpu_to_le32(NV_TX3_VLAN_TAG_PRESENT |
|
||||
vlan_tx_tag_get(skb));
|
||||
skb_vlan_tag_get(skb));
|
||||
else
|
||||
start_tx->txvlan = 0;
|
||||
|
||||
|
|
|
@ -1893,9 +1893,9 @@ netxen_tso_check(struct net_device *netdev,
|
|||
protocol = vh->h_vlan_encapsulated_proto;
|
||||
flags = FLAGS_VLAN_TAGGED;
|
||||
|
||||
} else if (vlan_tx_tag_present(skb)) {
|
||||
} else if (skb_vlan_tag_present(skb)) {
|
||||
flags = FLAGS_VLAN_OOB;
|
||||
vid = vlan_tx_tag_get(skb);
|
||||
vid = skb_vlan_tag_get(skb);
|
||||
netxen_set_tx_vlan_tci(first_desc, vid);
|
||||
vlan_oob = 1;
|
||||
}
|
||||
|
|
|
@ -321,8 +321,8 @@ static void qlcnic_send_filter(struct qlcnic_adapter *adapter,
|
|||
if (protocol == ETH_P_8021Q) {
|
||||
vh = (struct vlan_ethhdr *)skb->data;
|
||||
vlan_id = ntohs(vh->h_vlan_TCI);
|
||||
} else if (vlan_tx_tag_present(skb)) {
|
||||
vlan_id = vlan_tx_tag_get(skb);
|
||||
} else if (skb_vlan_tag_present(skb)) {
|
||||
vlan_id = skb_vlan_tag_get(skb);
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -473,9 +473,9 @@ static int qlcnic_tx_pkt(struct qlcnic_adapter *adapter,
|
|||
flags = QLCNIC_FLAGS_VLAN_TAGGED;
|
||||
vlan_tci = ntohs(vh->h_vlan_TCI);
|
||||
protocol = ntohs(vh->h_vlan_encapsulated_proto);
|
||||
} else if (vlan_tx_tag_present(skb)) {
|
||||
} else if (skb_vlan_tag_present(skb)) {
|
||||
flags = QLCNIC_FLAGS_VLAN_OOB;
|
||||
vlan_tci = vlan_tx_tag_get(skb);
|
||||
vlan_tci = skb_vlan_tag_get(skb);
|
||||
}
|
||||
if (unlikely(adapter->tx_pvid)) {
|
||||
if (vlan_tci && !(adapter->flags & QLCNIC_TAGGING_ENABLED))
|
||||
|
|
|
@ -2660,11 +2660,11 @@ static netdev_tx_t qlge_send(struct sk_buff *skb, struct net_device *ndev)
|
|||
|
||||
mac_iocb_ptr->frame_len = cpu_to_le16((u16) skb->len);
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
netif_printk(qdev, tx_queued, KERN_DEBUG, qdev->ndev,
|
||||
"Adding a vlan tag %d.\n", vlan_tx_tag_get(skb));
|
||||
"Adding a vlan tag %d.\n", skb_vlan_tag_get(skb));
|
||||
mac_iocb_ptr->flags3 |= OB_MAC_IOCB_V;
|
||||
mac_iocb_ptr->vlan_tci = cpu_to_le16(vlan_tx_tag_get(skb));
|
||||
mac_iocb_ptr->vlan_tci = cpu_to_le16(skb_vlan_tag_get(skb));
|
||||
}
|
||||
tso = ql_tso(skb, (struct ob_mac_tso_iocb_req *)mac_iocb_ptr);
|
||||
if (tso < 0) {
|
||||
|
|
|
@ -708,8 +708,8 @@ static void cp_tx (struct cp_private *cp)
|
|||
|
||||
static inline u32 cp_tx_vlan_tag(struct sk_buff *skb)
|
||||
{
|
||||
return vlan_tx_tag_present(skb) ?
|
||||
TxVlanTag | swab16(vlan_tx_tag_get(skb)) : 0x00;
|
||||
return skb_vlan_tag_present(skb) ?
|
||||
TxVlanTag | swab16(skb_vlan_tag_get(skb)) : 0x00;
|
||||
}
|
||||
|
||||
static void unwind_tx_frag_mapping(struct cp_private *cp, struct sk_buff *skb,
|
||||
|
|
|
@ -2073,8 +2073,8 @@ static int rtl8169_set_features(struct net_device *dev,
|
|||
|
||||
static inline u32 rtl8169_tx_vlan_tag(struct sk_buff *skb)
|
||||
{
|
||||
return (vlan_tx_tag_present(skb)) ?
|
||||
TxVlanTag | swab16(vlan_tx_tag_get(skb)) : 0x00;
|
||||
return (skb_vlan_tag_present(skb)) ?
|
||||
TxVlanTag | swab16(skb_vlan_tag_get(skb)) : 0x00;
|
||||
}
|
||||
|
||||
static void rtl8169_rx_vlan_tag(struct RxDesc *desc, struct sk_buff *skb)
|
||||
|
|
|
@ -1272,7 +1272,7 @@ static netdev_tx_t sxgbe_xmit(struct sk_buff *skb, struct net_device *dev)
|
|||
if (unlikely(skb_is_gso(skb) && tqueue->prev_mss != cur_mss))
|
||||
ctxt_desc_req = 1;
|
||||
|
||||
if (unlikely(vlan_tx_tag_present(skb) ||
|
||||
if (unlikely(skb_vlan_tag_present(skb) ||
|
||||
((skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) &&
|
||||
tqueue->hwts_tx_en)))
|
||||
ctxt_desc_req = 1;
|
||||
|
|
|
@ -1650,9 +1650,9 @@ static netdev_tx_t bdx_tx_transmit(struct sk_buff *skb,
|
|||
txd_mss);
|
||||
}
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
/*Cut VLAN ID to 12 bits */
|
||||
txd_vlan_id = vlan_tx_tag_get(skb) & BITS_MASK(12);
|
||||
txd_vlan_id = skb_vlan_tag_get(skb) & BITS_MASK(12);
|
||||
txd_vtag = 1;
|
||||
}
|
||||
|
||||
|
|
|
@ -1781,8 +1781,8 @@ static netdev_tx_t rhine_start_tx(struct sk_buff *skb,
|
|||
rp->tx_ring[entry].desc_length =
|
||||
cpu_to_le32(TXDESC | (skb->len >= ETH_ZLEN ? skb->len : ETH_ZLEN));
|
||||
|
||||
if (unlikely(vlan_tx_tag_present(skb))) {
|
||||
u16 vid_pcp = vlan_tx_tag_get(skb);
|
||||
if (unlikely(skb_vlan_tag_present(skb))) {
|
||||
u16 vid_pcp = skb_vlan_tag_get(skb);
|
||||
|
||||
/* drop CFI/DEI bit, register needs VID and PCP */
|
||||
vid_pcp = (vid_pcp & VLAN_VID_MASK) |
|
||||
|
@ -1803,7 +1803,7 @@ static netdev_tx_t rhine_start_tx(struct sk_buff *skb,
|
|||
|
||||
/* Non-x86 Todo: explicitly flush cache lines here. */
|
||||
|
||||
if (vlan_tx_tag_present(skb))
|
||||
if (skb_vlan_tag_present(skb))
|
||||
/* Tx queues are bits 7-0 (first Tx queue: bit 7) */
|
||||
BYTE_REG_BITS_ON(1 << 7, ioaddr + TQWake);
|
||||
|
||||
|
|
|
@ -2611,8 +2611,8 @@ static netdev_tx_t velocity_xmit(struct sk_buff *skb,
|
|||
|
||||
td_ptr->tdesc1.cmd = TCPLS_NORMAL + (tdinfo->nskb_dma + 1) * 16;
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
td_ptr->tdesc1.vlan = cpu_to_le16(vlan_tx_tag_get(skb));
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
td_ptr->tdesc1.vlan = cpu_to_le16(skb_vlan_tag_get(skb));
|
||||
td_ptr->tdesc1.TCR |= TCR0_VETAG;
|
||||
}
|
||||
|
||||
|
|
|
@ -645,7 +645,7 @@ static void macvtap_skb_to_vnet_hdr(struct macvtap_queue *q,
|
|||
|
||||
if (skb->ip_summed == CHECKSUM_PARTIAL) {
|
||||
vnet_hdr->flags = VIRTIO_NET_HDR_F_NEEDS_CSUM;
|
||||
if (vlan_tx_tag_present(skb))
|
||||
if (skb_vlan_tag_present(skb))
|
||||
vnet_hdr->csum_start = cpu_to_macvtap16(q,
|
||||
skb_checksum_start_offset(skb) + VLAN_HLEN);
|
||||
else
|
||||
|
@ -821,13 +821,13 @@ static ssize_t macvtap_put_user(struct macvtap_queue *q,
|
|||
total = vnet_hdr_len;
|
||||
total += skb->len;
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
struct {
|
||||
__be16 h_vlan_proto;
|
||||
__be16 h_vlan_TCI;
|
||||
} veth;
|
||||
veth.h_vlan_proto = skb->vlan_proto;
|
||||
veth.h_vlan_TCI = htons(vlan_tx_tag_get(skb));
|
||||
veth.h_vlan_TCI = htons(skb_vlan_tag_get(skb));
|
||||
|
||||
vlan_offset = offsetof(struct vlan_ethhdr, h_vlan_proto);
|
||||
total += VLAN_HLEN;
|
||||
|
|
|
@ -1260,7 +1260,7 @@ static ssize_t tun_put_user(struct tun_struct *tun,
|
|||
int vlan_hlen = 0;
|
||||
int vnet_hdr_sz = 0;
|
||||
|
||||
if (vlan_tx_tag_present(skb))
|
||||
if (skb_vlan_tag_present(skb))
|
||||
vlan_hlen = VLAN_HLEN;
|
||||
|
||||
if (tun->flags & IFF_VNET_HDR)
|
||||
|
@ -1337,7 +1337,7 @@ static ssize_t tun_put_user(struct tun_struct *tun,
|
|||
} veth;
|
||||
|
||||
veth.h_vlan_proto = skb->vlan_proto;
|
||||
veth.h_vlan_TCI = htons(vlan_tx_tag_get(skb));
|
||||
veth.h_vlan_TCI = htons(skb_vlan_tag_get(skb));
|
||||
|
||||
vlan_offset = offsetof(struct vlan_ethhdr, h_vlan_proto);
|
||||
|
||||
|
|
|
@ -1421,10 +1421,10 @@ static int msdn_giant_send_check(struct sk_buff *skb)
|
|||
|
||||
static inline void rtl_tx_vlan_tag(struct tx_desc *desc, struct sk_buff *skb)
|
||||
{
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
u32 opts2;
|
||||
|
||||
opts2 = TX_VLAN_TAG | swab16(vlan_tx_tag_get(skb));
|
||||
opts2 = TX_VLAN_TAG | swab16(skb_vlan_tag_get(skb));
|
||||
desc->opts2 |= cpu_to_le32(opts2);
|
||||
}
|
||||
}
|
||||
|
|
|
@ -1038,9 +1038,9 @@ vmxnet3_tq_xmit(struct sk_buff *skb, struct vmxnet3_tx_queue *tq,
|
|||
le32_add_cpu(&tq->shared->txNumDeferred, 1);
|
||||
}
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
gdesc->txd.ti = 1;
|
||||
gdesc->txd.tci = vlan_tx_tag_get(skb);
|
||||
gdesc->txd.tci = skb_vlan_tag_get(skb);
|
||||
}
|
||||
|
||||
/* finally flips the GEN bit of the SOP desc. */
|
||||
|
|
|
@ -1561,7 +1561,7 @@ static int vxlan6_xmit_skb(struct vxlan_sock *vs,
|
|||
|
||||
min_headroom = LL_RESERVED_SPACE(dst->dev) + dst->header_len
|
||||
+ VXLAN_HLEN + sizeof(struct ipv6hdr)
|
||||
+ (vlan_tx_tag_present(skb) ? VLAN_HLEN : 0);
|
||||
+ (skb_vlan_tag_present(skb) ? VLAN_HLEN : 0);
|
||||
|
||||
/* Need space for new headers (invalidates iph ptr) */
|
||||
err = skb_cow_head(skb, min_headroom);
|
||||
|
@ -1607,7 +1607,7 @@ int vxlan_xmit_skb(struct vxlan_sock *vs,
|
|||
|
||||
min_headroom = LL_RESERVED_SPACE(rt->dst.dev) + rt->dst.header_len
|
||||
+ VXLAN_HLEN + sizeof(struct iphdr)
|
||||
+ (vlan_tx_tag_present(skb) ? VLAN_HLEN : 0);
|
||||
+ (skb_vlan_tag_present(skb) ? VLAN_HLEN : 0);
|
||||
|
||||
/* Need space for new headers (invalidates iph ptr) */
|
||||
err = skb_cow_head(skb, min_headroom);
|
||||
|
|
|
@ -2800,12 +2800,12 @@ static void qeth_l3_fill_header(struct qeth_card *card, struct qeth_hdr *hdr,
|
|||
* before we're going to overwrite this location with next hop ip.
|
||||
* v6 uses passthrough, v4 sets the tag in the QDIO header.
|
||||
*/
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
if ((ipv == 4) || (card->info.type == QETH_CARD_TYPE_IQD))
|
||||
hdr->hdr.l3.ext_flags = QETH_HDR_EXT_VLAN_FRAME;
|
||||
else
|
||||
hdr->hdr.l3.ext_flags = QETH_HDR_EXT_INCLUDE_VLAN_TAG;
|
||||
hdr->hdr.l3.vlan_id = vlan_tx_tag_get(skb);
|
||||
hdr->hdr.l3.vlan_id = skb_vlan_tag_get(skb);
|
||||
}
|
||||
|
||||
hdr->hdr.l3.length = skb->len - sizeof(struct qeth_hdr);
|
||||
|
@ -2986,7 +2986,7 @@ static int qeth_l3_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
|||
skb_pull(new_skb, ETH_HLEN);
|
||||
}
|
||||
|
||||
if (ipv != 4 && vlan_tx_tag_present(new_skb)) {
|
||||
if (ipv != 4 && skb_vlan_tag_present(new_skb)) {
|
||||
skb_push(new_skb, VLAN_HLEN);
|
||||
skb_copy_to_linear_data(new_skb, new_skb->data + 4, 4);
|
||||
skb_copy_to_linear_data_offset(new_skb, 4,
|
||||
|
@ -2995,7 +2995,7 @@ static int qeth_l3_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
|||
new_skb->data + 12, 4);
|
||||
tag = (u16 *)(new_skb->data + 12);
|
||||
*tag = __constant_htons(ETH_P_8021Q);
|
||||
*(tag + 1) = htons(vlan_tx_tag_get(new_skb));
|
||||
*(tag + 1) = htons(skb_vlan_tag_get(new_skb));
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -469,7 +469,7 @@ static int peek_head_len(struct sock *sk)
|
|||
head = skb_peek(&sk->sk_receive_queue);
|
||||
if (likely(head)) {
|
||||
len = head->len;
|
||||
if (vlan_tx_tag_present(head))
|
||||
if (skb_vlan_tag_present(head))
|
||||
len += VLAN_HLEN;
|
||||
}
|
||||
|
||||
|
|
|
@ -78,9 +78,9 @@ static inline bool is_vlan_dev(struct net_device *dev)
|
|||
return dev->priv_flags & IFF_802_1Q_VLAN;
|
||||
}
|
||||
|
||||
#define vlan_tx_tag_present(__skb) ((__skb)->vlan_tci & VLAN_TAG_PRESENT)
|
||||
#define vlan_tx_tag_get(__skb) ((__skb)->vlan_tci & ~VLAN_TAG_PRESENT)
|
||||
#define vlan_tx_tag_get_id(__skb) ((__skb)->vlan_tci & VLAN_VID_MASK)
|
||||
#define skb_vlan_tag_present(__skb) ((__skb)->vlan_tci & VLAN_TAG_PRESENT)
|
||||
#define skb_vlan_tag_get(__skb) ((__skb)->vlan_tci & ~VLAN_TAG_PRESENT)
|
||||
#define skb_vlan_tag_get_id(__skb) ((__skb)->vlan_tci & VLAN_VID_MASK)
|
||||
|
||||
/**
|
||||
* struct vlan_pcpu_stats - VLAN percpu rx/tx stats
|
||||
|
@ -376,7 +376,7 @@ static inline struct sk_buff *vlan_insert_tag_set_proto(struct sk_buff *skb,
|
|||
static inline struct sk_buff *__vlan_hwaccel_push_inside(struct sk_buff *skb)
|
||||
{
|
||||
skb = vlan_insert_tag_set_proto(skb, skb->vlan_proto,
|
||||
vlan_tx_tag_get(skb));
|
||||
skb_vlan_tag_get(skb));
|
||||
if (likely(skb))
|
||||
skb->vlan_tci = 0;
|
||||
return skb;
|
||||
|
@ -393,7 +393,7 @@ static inline struct sk_buff *__vlan_hwaccel_push_inside(struct sk_buff *skb)
|
|||
*/
|
||||
static inline struct sk_buff *vlan_hwaccel_push_inside(struct sk_buff *skb)
|
||||
{
|
||||
if (vlan_tx_tag_present(skb))
|
||||
if (skb_vlan_tag_present(skb))
|
||||
skb = __vlan_hwaccel_push_inside(skb);
|
||||
return skb;
|
||||
}
|
||||
|
@ -442,8 +442,8 @@ static inline int __vlan_get_tag(const struct sk_buff *skb, u16 *vlan_tci)
|
|||
static inline int __vlan_hwaccel_get_tag(const struct sk_buff *skb,
|
||||
u16 *vlan_tci)
|
||||
{
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
*vlan_tci = vlan_tx_tag_get(skb);
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
*vlan_tci = skb_vlan_tag_get(skb);
|
||||
return 0;
|
||||
} else {
|
||||
*vlan_tci = 0;
|
||||
|
@ -480,7 +480,7 @@ static inline __be16 vlan_get_protocol(const struct sk_buff *skb)
|
|||
{
|
||||
__be16 protocol = 0;
|
||||
|
||||
if (vlan_tx_tag_present(skb) ||
|
||||
if (skb_vlan_tag_present(skb) ||
|
||||
skb->protocol != cpu_to_be16(ETH_P_8021Q))
|
||||
protocol = skb->protocol;
|
||||
else {
|
||||
|
|
|
@ -121,7 +121,7 @@ static inline __be16 tc_skb_protocol(const struct sk_buff *skb)
|
|||
* vlan accelerated path. In that case, use skb->vlan_proto
|
||||
* as the original vlan header was already stripped.
|
||||
*/
|
||||
if (vlan_tx_tag_present(skb))
|
||||
if (skb_vlan_tag_present(skb))
|
||||
return skb->vlan_proto;
|
||||
return skb->protocol;
|
||||
}
|
||||
|
|
|
@ -40,9 +40,9 @@ TRACE_EVENT(net_dev_start_xmit,
|
|||
__assign_str(name, dev->name);
|
||||
__entry->queue_mapping = skb->queue_mapping;
|
||||
__entry->skbaddr = skb;
|
||||
__entry->vlan_tagged = vlan_tx_tag_present(skb);
|
||||
__entry->vlan_tagged = skb_vlan_tag_present(skb);
|
||||
__entry->vlan_proto = ntohs(skb->vlan_proto);
|
||||
__entry->vlan_tci = vlan_tx_tag_get(skb);
|
||||
__entry->vlan_tci = skb_vlan_tag_get(skb);
|
||||
__entry->protocol = ntohs(skb->protocol);
|
||||
__entry->ip_summed = skb->ip_summed;
|
||||
__entry->len = skb->len;
|
||||
|
@ -174,9 +174,9 @@ DECLARE_EVENT_CLASS(net_dev_rx_verbose_template,
|
|||
#endif
|
||||
__entry->queue_mapping = skb->queue_mapping;
|
||||
__entry->skbaddr = skb;
|
||||
__entry->vlan_tagged = vlan_tx_tag_present(skb);
|
||||
__entry->vlan_tagged = skb_vlan_tag_present(skb);
|
||||
__entry->vlan_proto = ntohs(skb->vlan_proto);
|
||||
__entry->vlan_tci = vlan_tx_tag_get(skb);
|
||||
__entry->vlan_tci = skb_vlan_tag_get(skb);
|
||||
__entry->protocol = ntohs(skb->protocol);
|
||||
__entry->ip_summed = skb->ip_summed;
|
||||
__entry->hash = skb->hash;
|
||||
|
|
|
@ -9,7 +9,7 @@ bool vlan_do_receive(struct sk_buff **skbp)
|
|||
{
|
||||
struct sk_buff *skb = *skbp;
|
||||
__be16 vlan_proto = skb->vlan_proto;
|
||||
u16 vlan_id = vlan_tx_tag_get_id(skb);
|
||||
u16 vlan_id = skb_vlan_tag_get_id(skb);
|
||||
struct net_device *vlan_dev;
|
||||
struct vlan_pcpu_stats *rx_stats;
|
||||
|
||||
|
|
|
@ -66,17 +66,17 @@ static int brnf_pass_vlan_indev __read_mostly = 0;
|
|||
#endif
|
||||
|
||||
#define IS_IP(skb) \
|
||||
(!vlan_tx_tag_present(skb) && skb->protocol == htons(ETH_P_IP))
|
||||
(!skb_vlan_tag_present(skb) && skb->protocol == htons(ETH_P_IP))
|
||||
|
||||
#define IS_IPV6(skb) \
|
||||
(!vlan_tx_tag_present(skb) && skb->protocol == htons(ETH_P_IPV6))
|
||||
(!skb_vlan_tag_present(skb) && skb->protocol == htons(ETH_P_IPV6))
|
||||
|
||||
#define IS_ARP(skb) \
|
||||
(!vlan_tx_tag_present(skb) && skb->protocol == htons(ETH_P_ARP))
|
||||
(!skb_vlan_tag_present(skb) && skb->protocol == htons(ETH_P_ARP))
|
||||
|
||||
static inline __be16 vlan_proto(const struct sk_buff *skb)
|
||||
{
|
||||
if (vlan_tx_tag_present(skb))
|
||||
if (skb_vlan_tag_present(skb))
|
||||
return skb->protocol;
|
||||
else if (skb->protocol == htons(ETH_P_8021Q))
|
||||
return vlan_eth_hdr(skb)->h_vlan_encapsulated_proto;
|
||||
|
@ -436,11 +436,11 @@ static struct net_device *brnf_get_logical_dev(struct sk_buff *skb, const struct
|
|||
struct net_device *vlan, *br;
|
||||
|
||||
br = bridge_parent(dev);
|
||||
if (brnf_pass_vlan_indev == 0 || !vlan_tx_tag_present(skb))
|
||||
if (brnf_pass_vlan_indev == 0 || !skb_vlan_tag_present(skb))
|
||||
return br;
|
||||
|
||||
vlan = __vlan_find_dev_deep_rcu(br, skb->vlan_proto,
|
||||
vlan_tx_tag_get(skb) & VLAN_VID_MASK);
|
||||
skb_vlan_tag_get(skb) & VLAN_VID_MASK);
|
||||
|
||||
return vlan ? vlan : br;
|
||||
}
|
||||
|
|
|
@ -628,8 +628,8 @@ static inline int br_vlan_get_tag(const struct sk_buff *skb, u16 *vid)
|
|||
{
|
||||
int err = 0;
|
||||
|
||||
if (vlan_tx_tag_present(skb))
|
||||
*vid = vlan_tx_tag_get(skb) & VLAN_VID_MASK;
|
||||
if (skb_vlan_tag_present(skb))
|
||||
*vid = skb_vlan_tag_get(skb) & VLAN_VID_MASK;
|
||||
else {
|
||||
*vid = 0;
|
||||
err = -EINVAL;
|
||||
|
|
|
@ -187,7 +187,7 @@ bool br_allowed_ingress(struct net_bridge *br, struct net_port_vlans *v,
|
|||
* sent from vlan device on the bridge device, it does not have
|
||||
* HW accelerated vlan tag.
|
||||
*/
|
||||
if (unlikely(!vlan_tx_tag_present(skb) &&
|
||||
if (unlikely(!skb_vlan_tag_present(skb) &&
|
||||
skb->protocol == proto)) {
|
||||
skb = skb_vlan_untag(skb);
|
||||
if (unlikely(!skb))
|
||||
|
@ -200,7 +200,7 @@ bool br_allowed_ingress(struct net_bridge *br, struct net_port_vlans *v,
|
|||
/* Protocol-mismatch, empty out vlan_tci for new tag */
|
||||
skb_push(skb, ETH_HLEN);
|
||||
skb = vlan_insert_tag_set_proto(skb, skb->vlan_proto,
|
||||
vlan_tx_tag_get(skb));
|
||||
skb_vlan_tag_get(skb));
|
||||
if (unlikely(!skb))
|
||||
return false;
|
||||
|
||||
|
|
|
@ -45,8 +45,8 @@ ebt_vlan_mt(const struct sk_buff *skb, struct xt_action_param *par)
|
|||
/* VLAN encapsulated Type/Length field, given from orig frame */
|
||||
__be16 encap;
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
TCI = vlan_tx_tag_get(skb);
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
TCI = skb_vlan_tag_get(skb);
|
||||
encap = skb->protocol;
|
||||
} else {
|
||||
const struct vlan_hdr *fp;
|
||||
|
|
|
@ -133,7 +133,7 @@ ebt_basic_match(const struct ebt_entry *e, const struct sk_buff *skb,
|
|||
__be16 ethproto;
|
||||
int verdict, i;
|
||||
|
||||
if (vlan_tx_tag_present(skb))
|
||||
if (skb_vlan_tag_present(skb))
|
||||
ethproto = htons(ETH_P_8021Q);
|
||||
else
|
||||
ethproto = h->h_proto;
|
||||
|
|
|
@ -2578,7 +2578,7 @@ netdev_features_t netif_skb_features(struct sk_buff *skb)
|
|||
if (skb->encapsulation)
|
||||
features &= dev->hw_enc_features;
|
||||
|
||||
if (!vlan_tx_tag_present(skb)) {
|
||||
if (!skb_vlan_tag_present(skb)) {
|
||||
if (unlikely(protocol == htons(ETH_P_8021Q) ||
|
||||
protocol == htons(ETH_P_8021AD))) {
|
||||
struct vlan_ethhdr *veh = (struct vlan_ethhdr *)skb->data;
|
||||
|
@ -2659,7 +2659,7 @@ struct sk_buff *dev_hard_start_xmit(struct sk_buff *first, struct net_device *de
|
|||
static struct sk_buff *validate_xmit_vlan(struct sk_buff *skb,
|
||||
netdev_features_t features)
|
||||
{
|
||||
if (vlan_tx_tag_present(skb) &&
|
||||
if (skb_vlan_tag_present(skb) &&
|
||||
!vlan_hw_offload_capable(features, skb->vlan_proto))
|
||||
skb = __vlan_hwaccel_push_inside(skb);
|
||||
return skb;
|
||||
|
@ -3676,7 +3676,7 @@ static int __netif_receive_skb_core(struct sk_buff *skb, bool pfmemalloc)
|
|||
if (pfmemalloc && !skb_pfmemalloc_protocol(skb))
|
||||
goto drop;
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
if (pt_prev) {
|
||||
ret = deliver_skb(skb, pt_prev, orig_dev);
|
||||
pt_prev = NULL;
|
||||
|
@ -3708,8 +3708,8 @@ static int __netif_receive_skb_core(struct sk_buff *skb, bool pfmemalloc)
|
|||
}
|
||||
}
|
||||
|
||||
if (unlikely(vlan_tx_tag_present(skb))) {
|
||||
if (vlan_tx_tag_get_id(skb))
|
||||
if (unlikely(skb_vlan_tag_present(skb))) {
|
||||
if (skb_vlan_tag_get_id(skb))
|
||||
skb->pkt_type = PACKET_OTHERHOST;
|
||||
/* Note: we might in the future use prio bits
|
||||
* and set skb->priority like in vlan_do_receive()
|
||||
|
|
|
@ -77,7 +77,7 @@ static int netpoll_start_xmit(struct sk_buff *skb, struct net_device *dev,
|
|||
|
||||
features = netif_skb_features(skb);
|
||||
|
||||
if (vlan_tx_tag_present(skb) &&
|
||||
if (skb_vlan_tag_present(skb) &&
|
||||
!vlan_hw_offload_capable(features, skb->vlan_proto)) {
|
||||
skb = __vlan_hwaccel_push_inside(skb);
|
||||
if (unlikely(!skb)) {
|
||||
|
|
|
@ -4197,7 +4197,7 @@ struct sk_buff *skb_vlan_untag(struct sk_buff *skb)
|
|||
struct vlan_hdr *vhdr;
|
||||
u16 vlan_tci;
|
||||
|
||||
if (unlikely(vlan_tx_tag_present(skb))) {
|
||||
if (unlikely(skb_vlan_tag_present(skb))) {
|
||||
/* vlan_tci is already set-up so leave this for another time */
|
||||
return skb;
|
||||
}
|
||||
|
@ -4283,7 +4283,7 @@ int skb_vlan_pop(struct sk_buff *skb)
|
|||
__be16 vlan_proto;
|
||||
int err;
|
||||
|
||||
if (likely(vlan_tx_tag_present(skb))) {
|
||||
if (likely(skb_vlan_tag_present(skb))) {
|
||||
skb->vlan_tci = 0;
|
||||
} else {
|
||||
if (unlikely((skb->protocol != htons(ETH_P_8021Q) &&
|
||||
|
@ -4313,7 +4313,7 @@ EXPORT_SYMBOL(skb_vlan_pop);
|
|||
|
||||
int skb_vlan_push(struct sk_buff *skb, __be16 vlan_proto, u16 vlan_tci)
|
||||
{
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
unsigned int offset = skb->data - skb_mac_header(skb);
|
||||
int err;
|
||||
|
||||
|
@ -4323,7 +4323,7 @@ int skb_vlan_push(struct sk_buff *skb, __be16 vlan_proto, u16 vlan_tci)
|
|||
*/
|
||||
__skb_push(skb, offset);
|
||||
err = __vlan_insert_tag(skb, skb->vlan_proto,
|
||||
vlan_tx_tag_get(skb));
|
||||
skb_vlan_tag_get(skb));
|
||||
if (err)
|
||||
return err;
|
||||
skb->protocol = skb->vlan_proto;
|
||||
|
|
|
@ -119,7 +119,7 @@ int geneve_xmit_skb(struct geneve_sock *gs, struct rtable *rt,
|
|||
|
||||
min_headroom = LL_RESERVED_SPACE(rt->dst.dev) + rt->dst.header_len
|
||||
+ GENEVE_BASE_HLEN + opt_len + sizeof(struct iphdr)
|
||||
+ (vlan_tx_tag_present(skb) ? VLAN_HLEN : 0);
|
||||
+ (skb_vlan_tag_present(skb) ? VLAN_HLEN : 0);
|
||||
|
||||
err = skb_cow_head(skb, min_headroom);
|
||||
if (unlikely(err)) {
|
||||
|
|
|
@ -212,7 +212,7 @@ static int pop_vlan(struct sk_buff *skb, struct sw_flow_key *key)
|
|||
int err;
|
||||
|
||||
err = skb_vlan_pop(skb);
|
||||
if (vlan_tx_tag_present(skb))
|
||||
if (skb_vlan_tag_present(skb))
|
||||
invalidate_flow_key(key);
|
||||
else
|
||||
key->eth.tci = 0;
|
||||
|
@ -222,7 +222,7 @@ static int pop_vlan(struct sk_buff *skb, struct sw_flow_key *key)
|
|||
static int push_vlan(struct sk_buff *skb, struct sw_flow_key *key,
|
||||
const struct ovs_action_push_vlan *vlan)
|
||||
{
|
||||
if (vlan_tx_tag_present(skb))
|
||||
if (skb_vlan_tag_present(skb))
|
||||
invalidate_flow_key(key);
|
||||
else
|
||||
key->eth.tci = vlan->vlan_tci;
|
||||
|
|
|
@ -419,7 +419,7 @@ static int queue_userspace_packet(struct datapath *dp, struct sk_buff *skb,
|
|||
if (!dp_ifindex)
|
||||
return -ENODEV;
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
nskb = skb_clone(skb, GFP_ATOMIC);
|
||||
if (!nskb)
|
||||
return -ENOMEM;
|
||||
|
|
|
@ -70,7 +70,7 @@ void ovs_flow_stats_update(struct sw_flow *flow, __be16 tcp_flags,
|
|||
{
|
||||
struct flow_stats *stats;
|
||||
int node = numa_node_id();
|
||||
int len = skb->len + (vlan_tx_tag_present(skb) ? VLAN_HLEN : 0);
|
||||
int len = skb->len + (skb_vlan_tag_present(skb) ? VLAN_HLEN : 0);
|
||||
|
||||
stats = rcu_dereference(flow->stats[node]);
|
||||
|
||||
|
@ -472,7 +472,7 @@ static int key_extract(struct sk_buff *skb, struct sw_flow_key *key)
|
|||
*/
|
||||
|
||||
key->eth.tci = 0;
|
||||
if (vlan_tx_tag_present(skb))
|
||||
if (skb_vlan_tag_present(skb))
|
||||
key->eth.tci = htons(skb->vlan_tci);
|
||||
else if (eth->h_proto == htons(ETH_P_8021Q))
|
||||
if (unlikely(parse_vlan(skb, key)))
|
||||
|
|
|
@ -166,7 +166,7 @@ static int gre_tnl_send(struct vport *vport, struct sk_buff *skb)
|
|||
|
||||
min_headroom = LL_RESERVED_SPACE(rt->dst.dev) + rt->dst.header_len
|
||||
+ tunnel_hlen + sizeof(struct iphdr)
|
||||
+ (vlan_tx_tag_present(skb) ? VLAN_HLEN : 0);
|
||||
+ (skb_vlan_tag_present(skb) ? VLAN_HLEN : 0);
|
||||
if (skb_headroom(skb) < min_headroom || skb_header_cloned(skb)) {
|
||||
int head_delta = SKB_DATA_ALIGN(min_headroom -
|
||||
skb_headroom(skb) +
|
||||
|
|
|
@ -480,7 +480,8 @@ void ovs_vport_receive(struct vport *vport, struct sk_buff *skb,
|
|||
stats = this_cpu_ptr(vport->percpu_stats);
|
||||
u64_stats_update_begin(&stats->syncp);
|
||||
stats->rx_packets++;
|
||||
stats->rx_bytes += skb->len + (vlan_tx_tag_present(skb) ? VLAN_HLEN : 0);
|
||||
stats->rx_bytes += skb->len +
|
||||
(skb_vlan_tag_present(skb) ? VLAN_HLEN : 0);
|
||||
u64_stats_update_end(&stats->syncp);
|
||||
|
||||
OVS_CB(skb)->input_vport = vport;
|
||||
|
|
|
@ -986,8 +986,8 @@ static void prb_clear_rxhash(struct tpacket_kbdq_core *pkc,
|
|||
static void prb_fill_vlan_info(struct tpacket_kbdq_core *pkc,
|
||||
struct tpacket3_hdr *ppd)
|
||||
{
|
||||
if (vlan_tx_tag_present(pkc->skb)) {
|
||||
ppd->hv1.tp_vlan_tci = vlan_tx_tag_get(pkc->skb);
|
||||
if (skb_vlan_tag_present(pkc->skb)) {
|
||||
ppd->hv1.tp_vlan_tci = skb_vlan_tag_get(pkc->skb);
|
||||
ppd->hv1.tp_vlan_tpid = ntohs(pkc->skb->vlan_proto);
|
||||
ppd->tp_status = TP_STATUS_VLAN_VALID | TP_STATUS_VLAN_TPID_VALID;
|
||||
} else {
|
||||
|
@ -2000,8 +2000,8 @@ static int tpacket_rcv(struct sk_buff *skb, struct net_device *dev,
|
|||
h.h2->tp_net = netoff;
|
||||
h.h2->tp_sec = ts.tv_sec;
|
||||
h.h2->tp_nsec = ts.tv_nsec;
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
h.h2->tp_vlan_tci = vlan_tx_tag_get(skb);
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
h.h2->tp_vlan_tci = skb_vlan_tag_get(skb);
|
||||
h.h2->tp_vlan_tpid = ntohs(skb->vlan_proto);
|
||||
status |= TP_STATUS_VLAN_VALID | TP_STATUS_VLAN_TPID_VALID;
|
||||
} else {
|
||||
|
@ -3010,8 +3010,8 @@ static int packet_recvmsg(struct kiocb *iocb, struct socket *sock,
|
|||
aux.tp_snaplen = skb->len;
|
||||
aux.tp_mac = 0;
|
||||
aux.tp_net = skb_network_offset(skb);
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
aux.tp_vlan_tci = vlan_tx_tag_get(skb);
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
aux.tp_vlan_tci = skb_vlan_tag_get(skb);
|
||||
aux.tp_vlan_tpid = ntohs(skb->vlan_proto);
|
||||
aux.tp_status |= TP_STATUS_VLAN_VALID | TP_STATUS_VLAN_TPID_VALID;
|
||||
} else {
|
||||
|
|
|
@ -176,7 +176,7 @@ META_COLLECTOR(int_vlan_tag)
|
|||
{
|
||||
unsigned short tag;
|
||||
|
||||
tag = vlan_tx_tag_get(skb);
|
||||
tag = skb_vlan_tag_get(skb);
|
||||
if (!tag && __vlan_get_tag(skb, &tag))
|
||||
*err = -1;
|
||||
else
|
||||
|
|
|
@ -708,8 +708,8 @@ unsigned int cfg80211_classify8021d(struct sk_buff *skb,
|
|||
if (skb->priority >= 256 && skb->priority <= 263)
|
||||
return skb->priority - 256;
|
||||
|
||||
if (vlan_tx_tag_present(skb)) {
|
||||
vlan_priority = (vlan_tx_tag_get(skb) & VLAN_PRIO_MASK)
|
||||
if (skb_vlan_tag_present(skb)) {
|
||||
vlan_priority = (skb_vlan_tag_get(skb) & VLAN_PRIO_MASK)
|
||||
>> VLAN_PRIO_SHIFT;
|
||||
if (vlan_priority > 0)
|
||||
return vlan_priority;
|
||||
|
|
Loading…
Reference in New Issue