mirror of
https://github.com/adulau/aha.git
synced 2024-12-27 11:16:11 +00:00
[TCP]: Cong.ctrl modules: remove unused good_ack from cong_avoid
Signed-off-by: Ilpo Järvinen <ilpo.jarvinen@helsinki.fi> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
ede9f3b186
commit
c3a05c6050
14 changed files with 24 additions and 35 deletions
|
@ -652,7 +652,7 @@ struct tcp_congestion_ops {
|
|||
/* lower bound for congestion window (optional) */
|
||||
u32 (*min_cwnd)(const struct sock *sk);
|
||||
/* do new cwnd calculation (required) */
|
||||
void (*cong_avoid)(struct sock *sk, u32 ack, u32 in_flight, int good_ack);
|
||||
void (*cong_avoid)(struct sock *sk, u32 ack, u32 in_flight);
|
||||
/* call before changing ca_state (optional) */
|
||||
void (*set_state)(struct sock *sk, u8 new_state);
|
||||
/* call when cwnd event occurs (optional) */
|
||||
|
@ -683,7 +683,7 @@ extern void tcp_slow_start(struct tcp_sock *tp);
|
|||
|
||||
extern struct tcp_congestion_ops tcp_init_congestion_ops;
|
||||
extern u32 tcp_reno_ssthresh(struct sock *sk);
|
||||
extern void tcp_reno_cong_avoid(struct sock *sk, u32 ack, u32 in_flight, int flag);
|
||||
extern void tcp_reno_cong_avoid(struct sock *sk, u32 ack, u32 in_flight);
|
||||
extern u32 tcp_reno_min_cwnd(const struct sock *sk);
|
||||
extern struct tcp_congestion_ops tcp_reno;
|
||||
|
||||
|
|
|
@ -136,8 +136,7 @@ static inline void bictcp_update(struct bictcp *ca, u32 cwnd)
|
|||
ca->cnt = 1;
|
||||
}
|
||||
|
||||
static void bictcp_cong_avoid(struct sock *sk, u32 ack,
|
||||
u32 in_flight, int data_acked)
|
||||
static void bictcp_cong_avoid(struct sock *sk, u32 ack, u32 in_flight)
|
||||
{
|
||||
struct tcp_sock *tp = tcp_sk(sk);
|
||||
struct bictcp *ca = inet_csk_ca(sk);
|
||||
|
|
|
@ -324,7 +324,7 @@ EXPORT_SYMBOL_GPL(tcp_slow_start);
|
|||
/* This is Jacobson's slow start and congestion avoidance.
|
||||
* SIGCOMM '88, p. 328.
|
||||
*/
|
||||
void tcp_reno_cong_avoid(struct sock *sk, u32 ack, u32 in_flight, int flag)
|
||||
void tcp_reno_cong_avoid(struct sock *sk, u32 ack, u32 in_flight)
|
||||
{
|
||||
struct tcp_sock *tp = tcp_sk(sk);
|
||||
|
||||
|
|
|
@ -246,8 +246,7 @@ static inline void bictcp_update(struct bictcp *ca, u32 cwnd)
|
|||
ca->cnt = 1;
|
||||
}
|
||||
|
||||
static void bictcp_cong_avoid(struct sock *sk, u32 ack,
|
||||
u32 in_flight, int data_acked)
|
||||
static void bictcp_cong_avoid(struct sock *sk, u32 ack, u32 in_flight)
|
||||
{
|
||||
struct tcp_sock *tp = tcp_sk(sk);
|
||||
struct bictcp *ca = inet_csk_ca(sk);
|
||||
|
|
|
@ -109,8 +109,7 @@ static void hstcp_init(struct sock *sk)
|
|||
tp->snd_cwnd_clamp = min_t(u32, tp->snd_cwnd_clamp, 0xffffffff/128);
|
||||
}
|
||||
|
||||
static void hstcp_cong_avoid(struct sock *sk, u32 adk,
|
||||
u32 in_flight, int data_acked)
|
||||
static void hstcp_cong_avoid(struct sock *sk, u32 adk, u32 in_flight)
|
||||
{
|
||||
struct tcp_sock *tp = tcp_sk(sk);
|
||||
struct hstcp *ca = inet_csk_ca(sk);
|
||||
|
|
|
@ -225,8 +225,7 @@ static u32 htcp_recalc_ssthresh(struct sock *sk)
|
|||
return max((tp->snd_cwnd * ca->beta) >> 7, 2U);
|
||||
}
|
||||
|
||||
static void htcp_cong_avoid(struct sock *sk, u32 ack,
|
||||
u32 in_flight, int data_acked)
|
||||
static void htcp_cong_avoid(struct sock *sk, u32 ack, u32 in_flight)
|
||||
{
|
||||
struct tcp_sock *tp = tcp_sk(sk);
|
||||
struct htcp *ca = inet_csk_ca(sk);
|
||||
|
|
|
@ -85,8 +85,7 @@ static inline u32 hybla_fraction(u32 odds)
|
|||
* o Give cwnd a new value based on the model proposed
|
||||
* o remember increments <1
|
||||
*/
|
||||
static void hybla_cong_avoid(struct sock *sk, u32 ack,
|
||||
u32 in_flight, int flag)
|
||||
static void hybla_cong_avoid(struct sock *sk, u32 ack, u32 in_flight)
|
||||
{
|
||||
struct tcp_sock *tp = tcp_sk(sk);
|
||||
struct hybla *ca = inet_csk_ca(sk);
|
||||
|
@ -103,7 +102,7 @@ static void hybla_cong_avoid(struct sock *sk, u32 ack,
|
|||
return;
|
||||
|
||||
if (!ca->hybla_en)
|
||||
return tcp_reno_cong_avoid(sk, ack, in_flight, flag);
|
||||
return tcp_reno_cong_avoid(sk, ack, in_flight);
|
||||
|
||||
if (ca->rho == 0)
|
||||
hybla_recalc_param(sk);
|
||||
|
|
|
@ -256,8 +256,7 @@ static void tcp_illinois_state(struct sock *sk, u8 new_state)
|
|||
/*
|
||||
* Increase window in response to successful acknowledgment.
|
||||
*/
|
||||
static void tcp_illinois_cong_avoid(struct sock *sk, u32 ack,
|
||||
u32 in_flight, int flag)
|
||||
static void tcp_illinois_cong_avoid(struct sock *sk, u32 ack, u32 in_flight)
|
||||
{
|
||||
struct tcp_sock *tp = tcp_sk(sk);
|
||||
struct illinois *ca = inet_csk_ca(sk);
|
||||
|
|
|
@ -2710,11 +2710,10 @@ static inline void tcp_ack_update_rtt(struct sock *sk, const int flag,
|
|||
tcp_ack_no_tstamp(sk, seq_rtt, flag);
|
||||
}
|
||||
|
||||
static void tcp_cong_avoid(struct sock *sk, u32 ack,
|
||||
u32 in_flight, int good)
|
||||
static void tcp_cong_avoid(struct sock *sk, u32 ack, u32 in_flight)
|
||||
{
|
||||
const struct inet_connection_sock *icsk = inet_csk(sk);
|
||||
icsk->icsk_ca_ops->cong_avoid(sk, ack, in_flight, good);
|
||||
icsk->icsk_ca_ops->cong_avoid(sk, ack, in_flight);
|
||||
tcp_sk(sk)->snd_cwnd_stamp = tcp_time_stamp;
|
||||
}
|
||||
|
||||
|
@ -3238,11 +3237,11 @@ static int tcp_ack(struct sock *sk, struct sk_buff *skb, int flag)
|
|||
/* Advance CWND, if state allows this. */
|
||||
if ((flag & FLAG_DATA_ACKED) && !frto_cwnd &&
|
||||
tcp_may_raise_cwnd(sk, flag))
|
||||
tcp_cong_avoid(sk, ack, prior_in_flight, 0);
|
||||
tcp_cong_avoid(sk, ack, prior_in_flight);
|
||||
tcp_fastretrans_alert(sk, prior_packets - tp->packets_out, flag);
|
||||
} else {
|
||||
if ((flag & FLAG_DATA_ACKED) && !frto_cwnd)
|
||||
tcp_cong_avoid(sk, ack, prior_in_flight, 1);
|
||||
tcp_cong_avoid(sk, ack, prior_in_flight);
|
||||
}
|
||||
|
||||
if ((flag & FLAG_FORWARD_PROGRESS) || !(flag&FLAG_NOT_DUP))
|
||||
|
|
|
@ -115,12 +115,12 @@ static void tcp_lp_init(struct sock *sk)
|
|||
* Will only call newReno CA when away from inference.
|
||||
* From TCP-LP's paper, this will be handled in additive increasement.
|
||||
*/
|
||||
static void tcp_lp_cong_avoid(struct sock *sk, u32 ack, u32 in_flight, int flag)
|
||||
static void tcp_lp_cong_avoid(struct sock *sk, u32 ack, u32 in_flight)
|
||||
{
|
||||
struct lp *lp = inet_csk_ca(sk);
|
||||
|
||||
if (!(lp->flag & LP_WITHIN_INF))
|
||||
tcp_reno_cong_avoid(sk, ack, in_flight, flag);
|
||||
tcp_reno_cong_avoid(sk, ack, in_flight);
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
|
@ -15,8 +15,7 @@
|
|||
#define TCP_SCALABLE_AI_CNT 50U
|
||||
#define TCP_SCALABLE_MD_SCALE 3
|
||||
|
||||
static void tcp_scalable_cong_avoid(struct sock *sk, u32 ack,
|
||||
u32 in_flight, int flag)
|
||||
static void tcp_scalable_cong_avoid(struct sock *sk, u32 ack, u32 in_flight)
|
||||
{
|
||||
struct tcp_sock *tp = tcp_sk(sk);
|
||||
|
||||
|
|
|
@ -162,14 +162,13 @@ void tcp_vegas_cwnd_event(struct sock *sk, enum tcp_ca_event event)
|
|||
}
|
||||
EXPORT_SYMBOL_GPL(tcp_vegas_cwnd_event);
|
||||
|
||||
static void tcp_vegas_cong_avoid(struct sock *sk, u32 ack,
|
||||
u32 in_flight, int flag)
|
||||
static void tcp_vegas_cong_avoid(struct sock *sk, u32 ack, u32 in_flight)
|
||||
{
|
||||
struct tcp_sock *tp = tcp_sk(sk);
|
||||
struct vegas *vegas = inet_csk_ca(sk);
|
||||
|
||||
if (!vegas->doing_vegas_now)
|
||||
return tcp_reno_cong_avoid(sk, ack, in_flight, flag);
|
||||
return tcp_reno_cong_avoid(sk, ack, in_flight);
|
||||
|
||||
/* The key players are v_beg_snd_una and v_beg_snd_nxt.
|
||||
*
|
||||
|
@ -228,7 +227,7 @@ static void tcp_vegas_cong_avoid(struct sock *sk, u32 ack,
|
|||
/* We don't have enough RTT samples to do the Vegas
|
||||
* calculation, so we'll behave like Reno.
|
||||
*/
|
||||
tcp_reno_cong_avoid(sk, ack, in_flight, flag);
|
||||
tcp_reno_cong_avoid(sk, ack, in_flight);
|
||||
} else {
|
||||
u32 rtt, target_cwnd, diff;
|
||||
|
||||
|
|
|
@ -114,14 +114,13 @@ static void tcp_veno_cwnd_event(struct sock *sk, enum tcp_ca_event event)
|
|||
tcp_veno_init(sk);
|
||||
}
|
||||
|
||||
static void tcp_veno_cong_avoid(struct sock *sk, u32 ack,
|
||||
u32 in_flight, int flag)
|
||||
static void tcp_veno_cong_avoid(struct sock *sk, u32 ack, u32 in_flight)
|
||||
{
|
||||
struct tcp_sock *tp = tcp_sk(sk);
|
||||
struct veno *veno = inet_csk_ca(sk);
|
||||
|
||||
if (!veno->doing_veno_now)
|
||||
return tcp_reno_cong_avoid(sk, ack, in_flight, flag);
|
||||
return tcp_reno_cong_avoid(sk, ack, in_flight);
|
||||
|
||||
/* limited by applications */
|
||||
if (!tcp_is_cwnd_limited(sk, in_flight))
|
||||
|
@ -132,7 +131,7 @@ static void tcp_veno_cong_avoid(struct sock *sk, u32 ack,
|
|||
/* We don't have enough rtt samples to do the Veno
|
||||
* calculation, so we'll behave like Reno.
|
||||
*/
|
||||
tcp_reno_cong_avoid(sk, ack, in_flight, flag);
|
||||
tcp_reno_cong_avoid(sk, ack, in_flight);
|
||||
} else {
|
||||
u32 rtt, target_cwnd;
|
||||
|
||||
|
|
|
@ -69,8 +69,7 @@ static void tcp_yeah_pkts_acked(struct sock *sk, u32 pkts_acked, s32 rtt_us)
|
|||
tcp_vegas_pkts_acked(sk, pkts_acked, rtt_us);
|
||||
}
|
||||
|
||||
static void tcp_yeah_cong_avoid(struct sock *sk, u32 ack,
|
||||
u32 in_flight, int flag)
|
||||
static void tcp_yeah_cong_avoid(struct sock *sk, u32 ack, u32 in_flight)
|
||||
{
|
||||
struct tcp_sock *tp = tcp_sk(sk);
|
||||
struct yeah *yeah = inet_csk_ca(sk);
|
||||
|
|
Loading…
Reference in a new issue