gso: Support partial splitting at the frag_list pointer
Since commit 8a29111c7
("net: gro: allow to build full sized skb")
gro may build buffers with a frag_list. This can hurt forwarding
because most NICs can't offload such packets, they need to be
segmented in software. This patch splits buffers with a frag_list
at the frag_list pointer into buffers that can be TSO offloaded.
Signed-off-by: Steffen Klassert <steffen.klassert@secunet.com>
Acked-by: Alexander Duyck <alexander.h.duyck@intel.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
e867e87ae8
commit
07b26c9454
|
@ -3097,11 +3097,31 @@ struct sk_buff *skb_segment(struct sk_buff *head_skb,
|
|||
sg = !!(features & NETIF_F_SG);
|
||||
csum = !!can_checksum_protocol(features, proto);
|
||||
|
||||
/* GSO partial only requires that we trim off any excess that
|
||||
* doesn't fit into an MSS sized block, so take care of that
|
||||
* now.
|
||||
*/
|
||||
if (sg && csum && (features & NETIF_F_GSO_PARTIAL)) {
|
||||
if (sg && csum && (mss != GSO_BY_FRAGS)) {
|
||||
if (!(features & NETIF_F_GSO_PARTIAL)) {
|
||||
struct sk_buff *iter;
|
||||
|
||||
if (!list_skb ||
|
||||
!net_gso_ok(features, skb_shinfo(head_skb)->gso_type))
|
||||
goto normal;
|
||||
|
||||
/* Split the buffer at the frag_list pointer.
|
||||
* This is based on the assumption that all
|
||||
* buffers in the chain excluding the last
|
||||
* containing the same amount of data.
|
||||
*/
|
||||
skb_walk_frags(head_skb, iter) {
|
||||
if (skb_headlen(iter))
|
||||
goto normal;
|
||||
|
||||
len -= iter->len;
|
||||
}
|
||||
}
|
||||
|
||||
/* GSO partial only requires that we trim off any excess that
|
||||
* doesn't fit into an MSS sized block, so take care of that
|
||||
* now.
|
||||
*/
|
||||
partial_segs = len / mss;
|
||||
if (partial_segs > 1)
|
||||
mss *= partial_segs;
|
||||
|
@ -3109,6 +3129,7 @@ struct sk_buff *skb_segment(struct sk_buff *head_skb,
|
|||
partial_segs = 0;
|
||||
}
|
||||
|
||||
normal:
|
||||
headroom = skb_headroom(head_skb);
|
||||
pos = skb_headlen(head_skb);
|
||||
|
||||
|
@ -3300,21 +3321,29 @@ struct sk_buff *skb_segment(struct sk_buff *head_skb,
|
|||
*/
|
||||
segs->prev = tail;
|
||||
|
||||
/* Update GSO info on first skb in partial sequence. */
|
||||
if (partial_segs) {
|
||||
struct sk_buff *iter;
|
||||
int type = skb_shinfo(head_skb)->gso_type;
|
||||
unsigned short gso_size = skb_shinfo(head_skb)->gso_size;
|
||||
|
||||
/* Update type to add partial and then remove dodgy if set */
|
||||
type |= SKB_GSO_PARTIAL;
|
||||
type |= (features & NETIF_F_GSO_PARTIAL) / NETIF_F_GSO_PARTIAL * SKB_GSO_PARTIAL;
|
||||
type &= ~SKB_GSO_DODGY;
|
||||
|
||||
/* Update GSO info and prepare to start updating headers on
|
||||
* our way back down the stack of protocols.
|
||||
*/
|
||||
skb_shinfo(segs)->gso_size = skb_shinfo(head_skb)->gso_size;
|
||||
skb_shinfo(segs)->gso_segs = partial_segs;
|
||||
skb_shinfo(segs)->gso_type = type;
|
||||
SKB_GSO_CB(segs)->data_offset = skb_headroom(segs) + doffset;
|
||||
for (iter = segs; iter; iter = iter->next) {
|
||||
skb_shinfo(iter)->gso_size = gso_size;
|
||||
skb_shinfo(iter)->gso_segs = partial_segs;
|
||||
skb_shinfo(iter)->gso_type = type;
|
||||
SKB_GSO_CB(iter)->data_offset = skb_headroom(iter) + doffset;
|
||||
}
|
||||
|
||||
if (tail->len - doffset <= gso_size)
|
||||
skb_shinfo(tail)->gso_size = 0;
|
||||
else if (tail != segs)
|
||||
skb_shinfo(tail)->gso_segs = DIV_ROUND_UP(tail->len - doffset, gso_size);
|
||||
}
|
||||
|
||||
/* Following permits correct backpressure, for protocols
|
||||
|
|
|
@ -1192,7 +1192,7 @@ EXPORT_SYMBOL(inet_sk_rebuild_header);
|
|||
struct sk_buff *inet_gso_segment(struct sk_buff *skb,
|
||||
netdev_features_t features)
|
||||
{
|
||||
bool udpfrag = false, fixedid = false, encap;
|
||||
bool udpfrag = false, fixedid = false, gso_partial, encap;
|
||||
struct sk_buff *segs = ERR_PTR(-EINVAL);
|
||||
const struct net_offload *ops;
|
||||
unsigned int offset = 0;
|
||||
|
@ -1245,6 +1245,8 @@ struct sk_buff *inet_gso_segment(struct sk_buff *skb,
|
|||
if (IS_ERR_OR_NULL(segs))
|
||||
goto out;
|
||||
|
||||
gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL);
|
||||
|
||||
skb = segs;
|
||||
do {
|
||||
iph = (struct iphdr *)(skb_mac_header(skb) + nhoff);
|
||||
|
@ -1259,9 +1261,13 @@ struct sk_buff *inet_gso_segment(struct sk_buff *skb,
|
|||
iph->id = htons(id);
|
||||
id += skb_shinfo(skb)->gso_segs;
|
||||
}
|
||||
tot_len = skb_shinfo(skb)->gso_size +
|
||||
SKB_GSO_CB(skb)->data_offset +
|
||||
skb->head - (unsigned char *)iph;
|
||||
|
||||
if (gso_partial)
|
||||
tot_len = skb_shinfo(skb)->gso_size +
|
||||
SKB_GSO_CB(skb)->data_offset +
|
||||
skb->head - (unsigned char *)iph;
|
||||
else
|
||||
tot_len = skb->len - nhoff;
|
||||
} else {
|
||||
if (!fixedid)
|
||||
iph->id = htons(id++);
|
||||
|
|
|
@ -24,7 +24,7 @@ static struct sk_buff *gre_gso_segment(struct sk_buff *skb,
|
|||
__be16 protocol = skb->protocol;
|
||||
u16 mac_len = skb->mac_len;
|
||||
int gre_offset, outer_hlen;
|
||||
bool need_csum, ufo;
|
||||
bool need_csum, ufo, gso_partial;
|
||||
|
||||
if (!skb->encapsulation)
|
||||
goto out;
|
||||
|
@ -69,6 +69,8 @@ static struct sk_buff *gre_gso_segment(struct sk_buff *skb,
|
|||
goto out;
|
||||
}
|
||||
|
||||
gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL);
|
||||
|
||||
outer_hlen = skb_tnl_header_len(skb);
|
||||
gre_offset = outer_hlen - tnl_hlen;
|
||||
skb = segs;
|
||||
|
@ -96,7 +98,7 @@ static struct sk_buff *gre_gso_segment(struct sk_buff *skb,
|
|||
greh = (struct gre_base_hdr *)skb_transport_header(skb);
|
||||
pcsum = (__sum16 *)(greh + 1);
|
||||
|
||||
if (skb_is_gso(skb)) {
|
||||
if (gso_partial) {
|
||||
unsigned int partial_adj;
|
||||
|
||||
/* Adjust checksum to account for the fact that
|
||||
|
|
|
@ -90,12 +90,6 @@ struct sk_buff *tcp_gso_segment(struct sk_buff *skb,
|
|||
goto out;
|
||||
}
|
||||
|
||||
/* GSO partial only requires splitting the frame into an MSS
|
||||
* multiple and possibly a remainder. So update the mss now.
|
||||
*/
|
||||
if (features & NETIF_F_GSO_PARTIAL)
|
||||
mss = skb->len - (skb->len % mss);
|
||||
|
||||
copy_destructor = gso_skb->destructor == tcp_wfree;
|
||||
ooo_okay = gso_skb->ooo_okay;
|
||||
/* All segments but the first should have ooo_okay cleared */
|
||||
|
@ -108,6 +102,13 @@ struct sk_buff *tcp_gso_segment(struct sk_buff *skb,
|
|||
/* Only first segment might have ooo_okay set */
|
||||
segs->ooo_okay = ooo_okay;
|
||||
|
||||
/* GSO partial and frag_list segmentation only requires splitting
|
||||
* the frame into an MSS multiple and possibly a remainder, both
|
||||
* cases return a GSO skb. So update the mss now.
|
||||
*/
|
||||
if (skb_is_gso(segs))
|
||||
mss *= skb_shinfo(segs)->gso_segs;
|
||||
|
||||
delta = htonl(oldlen + (thlen + mss));
|
||||
|
||||
skb = segs;
|
||||
|
|
|
@ -21,7 +21,7 @@ static struct sk_buff *__skb_udp_tunnel_segment(struct sk_buff *skb,
|
|||
__be16 new_protocol, bool is_ipv6)
|
||||
{
|
||||
int tnl_hlen = skb_inner_mac_header(skb) - skb_transport_header(skb);
|
||||
bool remcsum, need_csum, offload_csum, ufo;
|
||||
bool remcsum, need_csum, offload_csum, ufo, gso_partial;
|
||||
struct sk_buff *segs = ERR_PTR(-EINVAL);
|
||||
struct udphdr *uh = udp_hdr(skb);
|
||||
u16 mac_offset = skb->mac_header;
|
||||
|
@ -88,6 +88,8 @@ static struct sk_buff *__skb_udp_tunnel_segment(struct sk_buff *skb,
|
|||
goto out;
|
||||
}
|
||||
|
||||
gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL);
|
||||
|
||||
outer_hlen = skb_tnl_header_len(skb);
|
||||
udp_offset = outer_hlen - tnl_hlen;
|
||||
skb = segs;
|
||||
|
@ -117,7 +119,7 @@ static struct sk_buff *__skb_udp_tunnel_segment(struct sk_buff *skb,
|
|||
* will be using a length value equal to only one MSS sized
|
||||
* segment instead of the entire frame.
|
||||
*/
|
||||
if (skb_is_gso(skb)) {
|
||||
if (gso_partial) {
|
||||
uh->len = htons(skb_shinfo(skb)->gso_size +
|
||||
SKB_GSO_CB(skb)->data_offset +
|
||||
skb->head - (unsigned char *)uh);
|
||||
|
|
|
@ -69,6 +69,7 @@ static struct sk_buff *ipv6_gso_segment(struct sk_buff *skb,
|
|||
int offset = 0;
|
||||
bool encap, udpfrag;
|
||||
int nhoff;
|
||||
bool gso_partial;
|
||||
|
||||
skb_reset_network_header(skb);
|
||||
nhoff = skb_network_header(skb) - skb_mac_header(skb);
|
||||
|
@ -101,9 +102,11 @@ static struct sk_buff *ipv6_gso_segment(struct sk_buff *skb,
|
|||
if (IS_ERR(segs))
|
||||
goto out;
|
||||
|
||||
gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL);
|
||||
|
||||
for (skb = segs; skb; skb = skb->next) {
|
||||
ipv6h = (struct ipv6hdr *)(skb_mac_header(skb) + nhoff);
|
||||
if (skb_is_gso(skb))
|
||||
if (gso_partial)
|
||||
payload_len = skb_shinfo(skb)->gso_size +
|
||||
SKB_GSO_CB(skb)->data_offset +
|
||||
skb->head - (unsigned char *)(ipv6h + 1);
|
||||
|
|
Loading…
Reference in New Issue