mirror of
https://github.com/torvalds/linux
synced 2024-11-05 18:23:50 +00:00
tcp: let tcp_mtu_probe() build headless packets
tcp_mtu_probe() is still copying payload from skbs in the write queue, using skb_copy_bits(), ignoring potential errors. Modern TCP stack wants to only deal with payload found in page frags, as this is a prereq for TCPDirect (host stack might not have access to the payload) Signed-off-by: Eric Dumazet <edumazet@google.com> Link: https://lore.kernel.org/r/20230607214113.1992947-1-edumazet@google.com Signed-off-by: Jakub Kicinski <kuba@kernel.org>
This commit is contained in:
parent
f84ad5cffd
commit
736013292e
1 changed files with 58 additions and 2 deletions
|
@ -2319,6 +2319,57 @@ static bool tcp_can_coalesce_send_queue_head(struct sock *sk, int len)
|
|||
return true;
|
||||
}
|
||||
|
||||
static int tcp_clone_payload(struct sock *sk, struct sk_buff *to,
|
||||
int probe_size)
|
||||
{
|
||||
skb_frag_t *lastfrag = NULL, *fragto = skb_shinfo(to)->frags;
|
||||
int i, todo, len = 0, nr_frags = 0;
|
||||
const struct sk_buff *skb;
|
||||
|
||||
if (!sk_wmem_schedule(sk, to->truesize + probe_size))
|
||||
return -ENOMEM;
|
||||
|
||||
skb_queue_walk(&sk->sk_write_queue, skb) {
|
||||
const skb_frag_t *fragfrom = skb_shinfo(skb)->frags;
|
||||
|
||||
if (skb_headlen(skb))
|
||||
return -EINVAL;
|
||||
|
||||
for (i = 0; i < skb_shinfo(skb)->nr_frags; i++, fragfrom++) {
|
||||
if (len >= probe_size)
|
||||
goto commit;
|
||||
todo = min_t(int, skb_frag_size(fragfrom),
|
||||
probe_size - len);
|
||||
len += todo;
|
||||
if (lastfrag &&
|
||||
skb_frag_page(fragfrom) == skb_frag_page(lastfrag) &&
|
||||
skb_frag_off(fragfrom) == skb_frag_off(lastfrag) +
|
||||
skb_frag_size(lastfrag)) {
|
||||
skb_frag_size_add(lastfrag, todo);
|
||||
continue;
|
||||
}
|
||||
if (unlikely(nr_frags == MAX_SKB_FRAGS))
|
||||
return -E2BIG;
|
||||
skb_frag_page_copy(fragto, fragfrom);
|
||||
skb_frag_off_copy(fragto, fragfrom);
|
||||
skb_frag_size_set(fragto, todo);
|
||||
nr_frags++;
|
||||
lastfrag = fragto++;
|
||||
}
|
||||
}
|
||||
commit:
|
||||
WARN_ON_ONCE(len != probe_size);
|
||||
for (i = 0; i < nr_frags; i++)
|
||||
skb_frag_ref(to, i);
|
||||
|
||||
skb_shinfo(to)->nr_frags = nr_frags;
|
||||
to->truesize += probe_size;
|
||||
to->len += probe_size;
|
||||
to->data_len += probe_size;
|
||||
__skb_header_release(to);
|
||||
return 0;
|
||||
}
|
||||
|
||||
/* Create a new MTU probe if we are ready.
|
||||
* MTU probe is regularly attempting to increase the path MTU by
|
||||
* deliberately sending larger packets. This discovers routing
|
||||
|
@ -2395,9 +2446,15 @@ static int tcp_mtu_probe(struct sock *sk)
|
|||
return -1;
|
||||
|
||||
/* We're allowed to probe. Build it now. */
|
||||
nskb = tcp_stream_alloc_skb(sk, probe_size, GFP_ATOMIC, false);
|
||||
nskb = tcp_stream_alloc_skb(sk, 0, GFP_ATOMIC, false);
|
||||
if (!nskb)
|
||||
return -1;
|
||||
|
||||
/* build the payload, and be prepared to abort if this fails. */
|
||||
if (tcp_clone_payload(sk, nskb, probe_size)) {
|
||||
consume_skb(nskb);
|
||||
return -1;
|
||||
}
|
||||
sk_wmem_queued_add(sk, nskb->truesize);
|
||||
sk_mem_charge(sk, nskb->truesize);
|
||||
|
||||
|
@ -2415,7 +2472,6 @@ static int tcp_mtu_probe(struct sock *sk)
|
|||
len = 0;
|
||||
tcp_for_write_queue_from_safe(skb, next, sk) {
|
||||
copy = min_t(int, skb->len, probe_size - len);
|
||||
skb_copy_bits(skb, 0, skb_put(nskb, copy), copy);
|
||||
|
||||
if (skb->len <= copy) {
|
||||
/* We've eaten all the data from this skb.
|
||||
|
|
Loading…
Reference in a new issue