mirror of
https://github.com/torvalds/linux
synced 2024-11-05 18:23:50 +00:00
tcp: consolidate PRR packet accounting
This patch series fixes an undo bug in fast recovery: the sender mistakenly undos the cwnd too early but continues fast retransmits until all pending data are acked. This also multiplies the SNMP stat PARTIALUNDO events by the degree of the network reordering. The first patch prepares the fix by consolidating the accounting of newly_acked_sacked in tcp_cwnd_reduction(), instead of updating newly_acked_sacked everytime sacked_out is adjusted. Also pass acked and prior_unsacked as const type because they are readonly in the rest of recovery processing. Signed-off-by: Yuchung Cheng <ycheng@google.com> Acked-by: Neal Cardwell <ncardwell@google.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
76723bca28
commit
6804973ffb
1 changed files with 20 additions and 25 deletions
|
@ -2430,12 +2430,14 @@ static void tcp_init_cwnd_reduction(struct sock *sk, const bool set_ssthresh)
|
|||
TCP_ECN_queue_cwr(tp);
|
||||
}
|
||||
|
||||
static void tcp_cwnd_reduction(struct sock *sk, int newly_acked_sacked,
|
||||
static void tcp_cwnd_reduction(struct sock *sk, const int prior_unsacked,
|
||||
int fast_rexmit)
|
||||
{
|
||||
struct tcp_sock *tp = tcp_sk(sk);
|
||||
int sndcnt = 0;
|
||||
int delta = tp->snd_ssthresh - tcp_packets_in_flight(tp);
|
||||
int newly_acked_sacked = prior_unsacked -
|
||||
(tp->packets_out - tp->sacked_out);
|
||||
|
||||
tp->prr_delivered += newly_acked_sacked;
|
||||
if (tcp_packets_in_flight(tp) > tp->snd_ssthresh) {
|
||||
|
@ -2492,7 +2494,7 @@ static void tcp_try_keep_open(struct sock *sk)
|
|||
}
|
||||
}
|
||||
|
||||
static void tcp_try_to_open(struct sock *sk, int flag, int newly_acked_sacked)
|
||||
static void tcp_try_to_open(struct sock *sk, int flag, const int prior_unsacked)
|
||||
{
|
||||
struct tcp_sock *tp = tcp_sk(sk);
|
||||
|
||||
|
@ -2509,7 +2511,7 @@ static void tcp_try_to_open(struct sock *sk, int flag, int newly_acked_sacked)
|
|||
if (inet_csk(sk)->icsk_ca_state != TCP_CA_Open)
|
||||
tcp_moderate_cwnd(tp);
|
||||
} else {
|
||||
tcp_cwnd_reduction(sk, newly_acked_sacked, 0);
|
||||
tcp_cwnd_reduction(sk, prior_unsacked, 0);
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -2678,15 +2680,14 @@ static void tcp_process_loss(struct sock *sk, int flag, bool is_dupack)
|
|||
* It does _not_ decide what to send, it is made in function
|
||||
* tcp_xmit_retransmit_queue().
|
||||
*/
|
||||
static void tcp_fastretrans_alert(struct sock *sk, int pkts_acked,
|
||||
int prior_sacked, int prior_packets,
|
||||
static void tcp_fastretrans_alert(struct sock *sk, const int acked,
|
||||
const int prior_unsacked,
|
||||
bool is_dupack, int flag)
|
||||
{
|
||||
struct inet_connection_sock *icsk = inet_csk(sk);
|
||||
struct tcp_sock *tp = tcp_sk(sk);
|
||||
int do_lost = is_dupack || ((flag & FLAG_DATA_SACKED) &&
|
||||
(tcp_fackets_out(tp) > tp->reordering));
|
||||
int newly_acked_sacked = 0;
|
||||
int fast_rexmit = 0;
|
||||
|
||||
if (WARN_ON(!tp->packets_out && tp->sacked_out))
|
||||
|
@ -2739,9 +2740,7 @@ static void tcp_fastretrans_alert(struct sock *sk, int pkts_acked,
|
|||
if (tcp_is_reno(tp) && is_dupack)
|
||||
tcp_add_reno_sack(sk);
|
||||
} else
|
||||
do_lost = tcp_try_undo_partial(sk, pkts_acked);
|
||||
newly_acked_sacked = prior_packets - tp->packets_out +
|
||||
tp->sacked_out - prior_sacked;
|
||||
do_lost = tcp_try_undo_partial(sk, acked);
|
||||
break;
|
||||
case TCP_CA_Loss:
|
||||
tcp_process_loss(sk, flag, is_dupack);
|
||||
|
@ -2755,14 +2754,12 @@ static void tcp_fastretrans_alert(struct sock *sk, int pkts_acked,
|
|||
if (is_dupack)
|
||||
tcp_add_reno_sack(sk);
|
||||
}
|
||||
newly_acked_sacked = prior_packets - tp->packets_out +
|
||||
tp->sacked_out - prior_sacked;
|
||||
|
||||
if (icsk->icsk_ca_state <= TCP_CA_Disorder)
|
||||
tcp_try_undo_dsack(sk);
|
||||
|
||||
if (!tcp_time_to_recover(sk, flag)) {
|
||||
tcp_try_to_open(sk, flag, newly_acked_sacked);
|
||||
tcp_try_to_open(sk, flag, prior_unsacked);
|
||||
return;
|
||||
}
|
||||
|
||||
|
@ -2784,7 +2781,7 @@ static void tcp_fastretrans_alert(struct sock *sk, int pkts_acked,
|
|||
|
||||
if (do_lost)
|
||||
tcp_update_scoreboard(sk, fast_rexmit);
|
||||
tcp_cwnd_reduction(sk, newly_acked_sacked, fast_rexmit);
|
||||
tcp_cwnd_reduction(sk, prior_unsacked, fast_rexmit);
|
||||
tcp_xmit_retransmit_queue(sk);
|
||||
}
|
||||
|
||||
|
@ -3268,9 +3265,8 @@ static int tcp_ack(struct sock *sk, const struct sk_buff *skb, int flag)
|
|||
u32 prior_in_flight;
|
||||
u32 prior_fackets;
|
||||
int prior_packets = tp->packets_out;
|
||||
int prior_sacked = tp->sacked_out;
|
||||
int pkts_acked = 0;
|
||||
int previous_packets_out = 0;
|
||||
const int prior_unsacked = tp->packets_out - tp->sacked_out;
|
||||
int acked = 0; /* Number of packets newly acked */
|
||||
|
||||
/* If the ack is older than previous acks
|
||||
* then we can probably ignore it.
|
||||
|
@ -3345,18 +3341,17 @@ static int tcp_ack(struct sock *sk, const struct sk_buff *skb, int flag)
|
|||
goto no_queue;
|
||||
|
||||
/* See if we can take anything off of the retransmit queue. */
|
||||
previous_packets_out = tp->packets_out;
|
||||
acked = tp->packets_out;
|
||||
flag |= tcp_clean_rtx_queue(sk, prior_fackets, prior_snd_una);
|
||||
|
||||
pkts_acked = previous_packets_out - tp->packets_out;
|
||||
acked -= tp->packets_out;
|
||||
|
||||
if (tcp_ack_is_dubious(sk, flag)) {
|
||||
/* Advance CWND, if state allows this. */
|
||||
if ((flag & FLAG_DATA_ACKED) && tcp_may_raise_cwnd(sk, flag))
|
||||
tcp_cong_avoid(sk, ack, prior_in_flight);
|
||||
is_dupack = !(flag & (FLAG_SND_UNA_ADVANCED | FLAG_NOT_DUP));
|
||||
tcp_fastretrans_alert(sk, pkts_acked, prior_sacked,
|
||||
prior_packets, is_dupack, flag);
|
||||
tcp_fastretrans_alert(sk, acked, prior_unsacked,
|
||||
is_dupack, flag);
|
||||
} else {
|
||||
if (flag & FLAG_DATA_ACKED)
|
||||
tcp_cong_avoid(sk, ack, prior_in_flight);
|
||||
|
@ -3378,8 +3373,8 @@ static int tcp_ack(struct sock *sk, const struct sk_buff *skb, int flag)
|
|||
no_queue:
|
||||
/* If data was DSACKed, see if we can undo a cwnd reduction. */
|
||||
if (flag & FLAG_DSACKING_ACK)
|
||||
tcp_fastretrans_alert(sk, pkts_acked, prior_sacked,
|
||||
prior_packets, is_dupack, flag);
|
||||
tcp_fastretrans_alert(sk, acked, prior_unsacked,
|
||||
is_dupack, flag);
|
||||
/* If this ack opens up a zero window, clear backoff. It was
|
||||
* being used to time the probes, and is probably far higher than
|
||||
* it needs to be for normal retransmission.
|
||||
|
@ -3401,8 +3396,8 @@ static int tcp_ack(struct sock *sk, const struct sk_buff *skb, int flag)
|
|||
*/
|
||||
if (TCP_SKB_CB(skb)->sacked) {
|
||||
flag |= tcp_sacktag_write_queue(sk, skb, prior_snd_una);
|
||||
tcp_fastretrans_alert(sk, pkts_acked, prior_sacked,
|
||||
prior_packets, is_dupack, flag);
|
||||
tcp_fastretrans_alert(sk, acked, prior_unsacked,
|
||||
is_dupack, flag);
|
||||
}
|
||||
|
||||
SOCK_DEBUG(sk, "Ack %u before %u:%u\n", ack, tp->snd_una, tp->snd_nxt);
|
||||
|
|
Loading…
Reference in a new issue