diff options
author | Mateusz Guzik <mjg@FreeBSD.org> | 2020-09-01 21:19:14 +0000 |
---|---|---|
committer | Mateusz Guzik <mjg@FreeBSD.org> | 2020-09-01 21:19:14 +0000 |
commit | 662c13053f4bf2d6245ba7e2b66c10d1cd5c1fb9 (patch) | |
tree | 5cffe4473640975abc443ef4b8047b34cdb4a039 /sys/netinet/tcp_stacks | |
parent | 586ee69f09c5060621486b34722a4a806581f93d (diff) | |
download | src-662c13053f4bf2d6245ba7e2b66c10d1cd5c1fb9.tar.gz src-662c13053f4bf2d6245ba7e2b66c10d1cd5c1fb9.zip |
net: clean up empty lines in .c and .h files
Notes
Notes:
svn path=/head/; revision=365071
Diffstat (limited to 'sys/netinet/tcp_stacks')
-rw-r--r-- | sys/netinet/tcp_stacks/bbr.c | 30 | ||||
-rw-r--r-- | sys/netinet/tcp_stacks/rack.c | 38 | ||||
-rw-r--r-- | sys/netinet/tcp_stacks/rack_bbr_common.c | 2 | ||||
-rw-r--r-- | sys/netinet/tcp_stacks/tcp_bbr.h | 9 | ||||
-rw-r--r-- | sys/netinet/tcp_stacks/tcp_rack.h | 1 |
5 files changed, 0 insertions, 80 deletions
diff --git a/sys/netinet/tcp_stacks/bbr.c b/sys/netinet/tcp_stacks/bbr.c index 3d89dd1b8efb..e90ec4828100 100644 --- a/sys/netinet/tcp_stacks/bbr.c +++ b/sys/netinet/tcp_stacks/bbr.c @@ -160,7 +160,6 @@ static int32_t bbr_hardware_pacing_limit = 8000; static int32_t bbr_quanta = 3; /* How much extra quanta do we get? */ static int32_t bbr_no_retran = 0; - static int32_t bbr_error_base_paceout = 10000; /* usec to pace */ static int32_t bbr_max_net_error_cnt = 10; /* Should the following be dynamic too -- loss wise */ @@ -218,7 +217,6 @@ static int32_t bbr_target_cwnd_mult_limit = 8; */ static int32_t bbr_cwnd_min_val = BBR_PROBERTT_NUM_MSS; - static int32_t bbr_cwnd_min_val_hs = BBR_HIGHSPEED_NUM_MSS; static int32_t bbr_gain_to_target = 1; @@ -516,7 +514,6 @@ static void static void bbr_timer_cancel(struct tcp_bbr *bbr, int32_t line, uint32_t cts); - static void bbr_log_pacing_delay_calc(struct tcp_bbr *bbr, uint16_t gain, uint32_t len, uint32_t cts, uint32_t usecs, uint64_t bw, uint32_t override, int mod); @@ -1128,8 +1125,6 @@ bbr_calc_time(uint32_t cts, uint32_t earlier_time) { return (cts - earlier_time); } - - static int sysctl_bbr_clear_lost(SYSCTL_HANDLER_ARGS) { @@ -1806,7 +1801,6 @@ bbr_init_sysctls(void) &bbr_nohdwr_pacing_enobuf, "Total number of enobufs for non-hardware paced flows"); - bbr_flows_whdwr_pacing = counter_u64_alloc(M_WAITOK); SYSCTL_ADD_COUNTER_U64(&bbr_sysctl_ctx, SYSCTL_CHILDREN(bbr_sysctl_root), @@ -1933,8 +1927,6 @@ bbr_log_type_rwnd_collapse(struct tcp_bbr *bbr, int seq, int mode, uint32_t coun } } - - static void bbr_log_type_just_return(struct tcp_bbr *bbr, uint32_t cts, uint32_t tlen, uint8_t hpts_calling, uint8_t reason, uint32_t p_maxseg, int len) @@ -1961,7 +1953,6 @@ bbr_log_type_just_return(struct tcp_bbr *bbr, uint32_t cts, uint32_t tlen, uint8 } } - static void bbr_log_type_enter_rec(struct tcp_bbr *bbr, uint32_t seq) { @@ -2569,7 +2560,6 @@ bbr_log_to_cancel(struct tcp_bbr *bbr, int32_t line, uint32_t cts, uint8_t hpts_ } } - static void bbr_log_tstmp_validation(struct tcp_bbr *bbr, uint64_t peer_delta, uint64_t delta) { @@ -2589,7 +2579,6 @@ bbr_log_tstmp_validation(struct tcp_bbr *bbr, uint64_t peer_delta, uint64_t delt &bbr->rc_inp->inp_socket->so_snd, BBR_LOG_TSTMP_VAL, 0, 0, &log, false, &bbr->rc_tv); - } } @@ -2680,7 +2669,6 @@ bbr_log_type_bbrupd(struct tcp_bbr *bbr, uint8_t flex8, uint32_t cts, } } - static void bbr_log_type_ltbw(struct tcp_bbr *bbr, uint32_t cts, int32_t reason, uint32_t newbw, uint32_t obw, uint32_t diff, @@ -2916,7 +2904,6 @@ bbr_set_epoch(struct tcp_bbr *bbr, uint32_t cts, int32_t line) bbr->r_ctl.rc_rcv_epoch_start = cts; } - static inline void bbr_isit_a_pkt_epoch(struct tcp_bbr *bbr, uint32_t cts, struct bbr_sendmap *rsm, int32_t line, int32_t cum_acked) { @@ -3333,7 +3320,6 @@ bbr_alloc_full_limit(struct tcp_bbr *bbr) return (bbr_alloc(bbr)); } - /* wrapper to allocate a sendmap entry, subject to a specific limit */ static struct bbr_sendmap * bbr_alloc_limit(struct tcp_bbr *bbr, uint8_t limit_type) @@ -3540,7 +3526,6 @@ bbr_get_header_oh(struct tcp_bbr *bbr) return(seg_oh); } - static uint32_t bbr_get_pacing_length(struct tcp_bbr *bbr, uint16_t gain, uint32_t useconds_time, uint64_t bw) { @@ -4299,7 +4284,6 @@ bbr_is_lost(struct tcp_bbr *bbr, struct bbr_sendmap *rsm, uint32_t cts) { uint32_t thresh; - thresh = bbr_calc_thresh_rack(bbr, bbr_get_rtt(bbr, BBR_RTT_RACK), cts, rsm); if ((cts - rsm->r_tim_lastsent[(rsm->r_rtr_cnt - 1)]) >= thresh) { @@ -5085,7 +5069,6 @@ bbr_timeout_rxt(struct tcpcb *tp, struct tcp_bbr *bbr, uint32_t cts) (V_tcp_pmtud_blackhole_detect == 3 && isipv6)) && ((tp->t_state == TCPS_ESTABLISHED) || (tp->t_state == TCPS_FIN_WAIT_1))) { - /* * Idea here is that at each stage of mtu probe (usually, * 1448 -> 1188 -> 524) should be given 2 chances to recover @@ -6154,7 +6137,6 @@ bbr_collapse_rtt(struct tcpcb *tp, struct tcp_bbr *bbr, int32_t rtt) tp->t_softerror = 0; } - static void tcp_bbr_xmit_timer(struct tcp_bbr *bbr, uint32_t rtt_usecs, uint32_t rsm_send_time, uint32_t r_start, uint32_t tsin) { @@ -6301,7 +6283,6 @@ tcp_bbr_xmit_timer_commit(struct tcp_bbr *bbr, struct tcpcb *tp, uint32_t cts) uint32_t rtt, tsin; int32_t rtt_ticks; - if (bbr->rtt_valid == 0) /* No valid sample */ return; @@ -6507,7 +6488,6 @@ bbr_earlier_retran(struct tcpcb *tp, struct tcp_bbr *bbr, struct bbr_sendmap *rs BBR_STAT_ADD(bbr_badfr_bytes, (rsm->r_end - rsm->r_start)); } - static void bbr_set_reduced_rtt(struct tcp_bbr *bbr, uint32_t cts, uint32_t line) { @@ -6762,7 +6742,6 @@ bbr_google_measurement(struct tcp_bbr *bbr, struct bbr_sendmap *rsm, uint32_t rt } } - static void bbr_update_bbr_info(struct tcp_bbr *bbr, struct bbr_sendmap *rsm, uint32_t rtt, uint32_t cts, uint32_t tsin, uint32_t uts, int32_t match, uint32_t rsm_send_time, int32_t ack_type, struct tcpopt *to) @@ -6911,7 +6890,6 @@ bbr_update_rtt(struct tcpcb *tp, struct tcp_bbr *bbr, (ack_type == BBR_CUM_ACKED) && (to->to_flags & TOF_TS) && (to->to_tsecr != 0)) { - t = tcp_tv_to_mssectick(&bbr->rc_tv) - to->to_tsecr; if (t < 1) t = 1; @@ -7308,7 +7286,6 @@ out: return (changed); } - static void inline bbr_peer_reneges(struct tcp_bbr *bbr, struct bbr_sendmap *rsm, tcp_seq th_ack) { @@ -7481,7 +7458,6 @@ bbr_log_ack(struct tcpcb *tp, struct tcpopt *to, struct tcphdr *th, bbr->r_wanted_output = 1; more: if (rsm == NULL) { - if (tp->t_flags & TF_SENTFIN) { /* if we send a FIN we will not hav a map */ goto proc_sack; @@ -8455,7 +8431,6 @@ bbr_process_data(struct mbuf *m, struct tcphdr *th, struct socket *so, tp->rcv_nxt++; } switch (tp->t_state) { - /* * In SYN_RECEIVED and ESTABLISHED STATES enter the * CLOSE_WAIT state. @@ -9194,7 +9169,6 @@ bbr_do_syn_recv(struct mbuf *m, struct tcphdr *th, struct socket *so, if (thflags & TH_ACK) bbr_log_syn(tp, to); if (IS_FASTOPEN(tp->t_flags) && tp->t_tfo_pending) { - tcp_fastopen_decrement_counter(tp->t_tfo_pending); tp->t_tfo_pending = NULL; } @@ -9876,7 +9850,6 @@ bbr_do_lastack(struct mbuf *m, struct tcphdr *th, struct socket *so, tiwin, thflags, nxt_pkt)); } - /* * Return value of 1, the TCB is unlocked and most * likely gone, return value of 0, the TCB is still @@ -10420,7 +10393,6 @@ bbr_substate_change(struct tcp_bbr *bbr, uint32_t cts, int32_t line, int dolog) */ int32_t old_state, old_gain; - old_state = bbr_state_val(bbr); old_gain = bbr->r_ctl.rc_bbr_hptsi_gain; if (bbr_state_val(bbr) == BBR_SUB_LEVEL1) { @@ -11011,7 +10983,6 @@ bbr_should_enter_probe_rtt(struct tcp_bbr *bbr, uint32_t cts) return (0); } - static int32_t bbr_google_startup(struct tcp_bbr *bbr, uint32_t cts, int32_t pkt_epoch) { @@ -11204,7 +11175,6 @@ bbr_state_change(struct tcp_bbr *bbr, uint32_t cts, int32_t epoch, int32_t pkt_e bbr->rc_filled_pipe = 1; bbr->r_ctl.bbr_lost_at_state = bbr->r_ctl.rc_lost; if (SEQ_GT(cts, bbr->r_ctl.rc_bbr_state_time)) { - time_in = cts - bbr->r_ctl.rc_bbr_state_time; counter_u64_add(bbr_state_time[bbr->rc_bbr_state], time_in); } else diff --git a/sys/netinet/tcp_stacks/rack.c b/sys/netinet/tcp_stacks/rack.c index 5e9a568bd6ba..e18528a4463d 100644 --- a/sys/netinet/tcp_stacks/rack.c +++ b/sys/netinet/tcp_stacks/rack.c @@ -259,7 +259,6 @@ static int32_t rack_probertt_clear_is = 1; static int32_t rack_max_drain_hbp = 1; /* Extra drain times gpsrtt for highly buffered paths */ static int32_t rack_hbp_thresh = 3; /* what is the divisor max_rtt/min_rtt to decided a hbp */ - /* Part of pacing */ static int32_t rack_max_per_above = 30; /* When we go to increment stop if above 100+this% */ @@ -501,7 +500,6 @@ static void int32_t rack_clear_counter=0; - static int sysctl_rack_clear(SYSCTL_HANDLER_ARGS) { @@ -568,14 +566,11 @@ sysctl_rack_clear(SYSCTL_HANDLER_ARGS) counter_u64_zero(rack_tlp_does_nada); counter_u64_zero(rack_try_scwnd); counter_u64_zero(rack_collapsed_win); - } rack_clear_counter = 0; return (0); } - - static void rack_init_sysctls(void) { @@ -1671,8 +1666,6 @@ rack_log_retran_reason(struct tcp_rack *rack, struct rack_sendmap *rsm, uint32_t } } - - static void rack_log_to_start(struct tcp_rack *rack, uint32_t cts, uint32_t to, int32_t slot, uint8_t which) { @@ -1768,7 +1761,6 @@ rack_log_rtt_upd(struct tcpcb *tp, struct tcp_rack *rack, uint32_t t, uint32_t l log.u_bbr.lost = rsm->r_end; log.u_bbr.cwnd_gain = rsm->r_rtr_cnt; } else { - /* Its a SYN */ log.u_bbr.pkt_epoch = rack->rc_tp->iss; log.u_bbr.lost = 0; @@ -1833,7 +1825,6 @@ rack_log_rtt_sample(struct tcp_rack *rack, uint32_t rtt) } } - static inline void rack_log_progress_event(struct tcp_rack *rack, struct tcpcb *tp, uint32_t tick, int event, int line) { @@ -2414,7 +2405,6 @@ rack_enough_for_measurement(struct tcpcb *tp, struct tcp_rack *rack, tcp_seq th_ return (0); } - static void rack_log_timely(struct tcp_rack *rack, uint32_t logged, uint64_t cur_bw, uint64_t low_bnd, @@ -2711,7 +2701,6 @@ rack_decrease_highrtt(struct tcp_rack *rack, uint32_t curper, uint32_t rtt) return (perf); } - static void rack_decrease_bw_mul(struct tcp_rack *rack, int timely_says, uint32_t rtt, int32_t rtt_diff) { @@ -3301,7 +3290,6 @@ rack_update_multiplier(struct tcp_rack *rack, int32_t timely_says, uint64_t last /* Log it stays the same */ rack_log_timely(rack, 0, last_bw_est, low_bnd, 0, __LINE__, 11); - } rack->rc_gp_timely_dec_cnt++; /* We are not incrementing really no-count */ @@ -3354,7 +3342,6 @@ rack_update_multiplier(struct tcp_rack *rack, int32_t timely_says, uint64_t last /* Log it stays the same */ rack_log_timely(rack, 0, last_bw_est, up_bnd, 0, __LINE__, 12); - } return; } else { @@ -3966,7 +3953,6 @@ rack_post_recovery(struct tcpcb *tp, struct tcphdr *th) struct tcp_rack *rack; uint32_t orig_cwnd; - orig_cwnd = tp->snd_cwnd; INP_WLOCK_ASSERT(tp->t_inpcb); rack = (struct tcp_rack *)tp->t_fb_ptr; @@ -4102,8 +4088,6 @@ rack_cong_signal(struct tcpcb *tp, struct tcphdr *th, uint32_t type) } } - - static inline void rack_cc_after_idle(struct tcp_rack *rack, struct tcpcb *tp) { @@ -4134,7 +4118,6 @@ rack_cc_after_idle(struct tcp_rack *rack, struct tcpcb *tp) } } - /* * Indicate whether this ack should be delayed. We can delay the ack if * following conditions are met: @@ -4193,7 +4176,6 @@ rack_find_high_nonack(struct tcp_rack *rack, struct rack_sendmap *rsm) return (NULL); } - static uint32_t rack_calc_thresh_rack(struct tcp_rack *rack, uint32_t srtt, uint32_t cts) { @@ -4697,7 +4679,6 @@ rack_exit_persist(struct tcpcb *tp, struct tcp_rack *rack, uint32_t cts) rack_exit_probertt(rack, us_cts); } } - } rack->rc_in_persist = 0; rack->r_ctl.rc_went_idle_time = 0; @@ -4811,7 +4792,6 @@ rack_start_hpts_timer(struct tcp_rack *rack, struct tcpcb *tp, uint32_t cts, slot = HPTS_TICKS_PER_USEC; } } else { - slot -= rack->r_ctl.rc_agg_delayed; rack->r_ctl.rc_agg_delayed = 0; /* Make sure we have 100 useconds at minimum */ @@ -5295,7 +5275,6 @@ need_retran: */ struct rack_sendmap *nrsm; - nrsm = rack_alloc_full_limit(rack); if (nrsm == NULL) { /* @@ -5569,7 +5548,6 @@ rack_cc_conn_init(struct tcpcb *tp) { struct tcp_rack *rack; - rack = (struct tcp_rack *)tp->t_fb_ptr; cc_conn_init(tp); /* @@ -5699,7 +5677,6 @@ rack_timeout_rxt(struct tcpcb *tp, struct tcp_rack *rack, uint32_t cts) (V_tcp_pmtud_blackhole_detect == 3 && isipv6)) && ((tp->t_state == TCPS_ESTABLISHED) || (tp->t_state == TCPS_FIN_WAIT_1))) { - /* * Idea here is that at each stage of mtu probe (usually, * 1448 -> 1188 -> 524) should be given 2 chances to recover @@ -5894,7 +5871,6 @@ rack_timer_cancel(struct tcpcb *tp, struct tcp_rack *rack, uint32_t cts, int lin uint32_t us_cts, flags_on_entry; uint8_t hpts_removed = 0; - flags_on_entry = rack->r_ctl.rc_hpts_flags; us_cts = tcp_get_usecs(&tv); if ((rack->r_ctl.rc_hpts_flags & PACE_PKT_OUTPUT) && @@ -6008,7 +5984,6 @@ rack_update_rsm(struct tcpcb *tp, struct tcp_rack *rack, } } - static uint32_t rack_update_entry(struct tcpcb *tp, struct tcp_rack *rack, struct rack_sendmap *rsm, uint32_t ts, int32_t *lenp) @@ -6081,7 +6056,6 @@ rack_update_entry(struct tcpcb *tp, struct tcp_rack *rack, return (0); } - static void rack_log_output(struct tcpcb *tp, struct tcpopt *to, int32_t len, uint32_t seq_out, uint8_t th_flags, int32_t err, uint32_t ts, @@ -6392,7 +6366,6 @@ tcp_rack_xmit_timer(struct tcp_rack *rack, int32_t rtt, uint32_t len, uint32_t u rack->r_ctl.rack_rs.confidence = confidence; rack->r_ctl.rack_rs.rs_us_rtrcnt = rtrcnt; } - } rack_log_rtt_upd(rack->rc_tp, rack, us_rtt, len, rsm, confidence); rack->r_ctl.rack_rs.rs_flags = RACK_RTT_VALID; @@ -7487,7 +7460,6 @@ rack_log_ack(struct tcpcb *tp, struct tcpopt *to, struct tcphdr *th) int loop_start = 0, moved_two = 0; uint32_t tsused; - INP_WLOCK_ASSERT(tp->t_inpcb); if (th->th_flags & TH_RST) { /* We don't log resets */ @@ -7652,7 +7624,6 @@ more: rsm->r_start = th_ack; if (rack->app_limited_needs_set) rack_need_set_test(tp, rack, rsm, tp->snd_una, __LINE__, RACK_USE_BEG); - } proc_sack: /* Check for reneging */ @@ -7704,7 +7675,6 @@ proc_sack: tcp_record_dsack(sack.start, sack.end); #endif } - } /* * Sort the SACK blocks so we can update the rack scoreboard with @@ -8428,7 +8398,6 @@ rack_process_ack(struct mbuf *m, struct tcphdr *th, struct socket *so, tp = tcp_close(tp); ctf_do_dropwithreset(m, tp, th, BANDLIM_UNLIMITED, tlen); return (1); - } } if (ofia) @@ -8779,7 +8748,6 @@ rack_process_data(struct mbuf *m, struct tcphdr *th, struct socket *so, tp->rcv_nxt++; } switch (tp->t_state) { - /* * In SYN_RECEIVED and ESTABLISHED STATES enter the * CLOSE_WAIT state. @@ -9681,7 +9649,6 @@ rack_do_established(struct mbuf *m, struct tcphdr *th, struct socket *so, */ if ((thflags & TH_ACK) == 0) { if (tp->t_flags & TF_NEEDSYN) { - return (rack_process_data(m, th, so, tp, drop_hdrlen, tlen, tiwin, thflags, nxt_pkt)); @@ -10177,7 +10144,6 @@ rack_do_lastack(struct mbuf *m, struct tcphdr *th, struct socket *so, tiwin, thflags, nxt_pkt)); } - /* * Return value of 1, the TCB is unlocked and most * likely gone, return value of 0, the TCP is still @@ -10625,7 +10591,6 @@ rack_fini(struct tcpcb *tp, int32_t tcb_is_purged) tp->snd_nxt = tp->snd_max; } - static void rack_set_state(struct tcpcb *tp, struct tcp_rack *rack) { @@ -10671,7 +10636,6 @@ rack_set_state(struct tcpcb *tp, struct tcp_rack *rack) }; } - static void rack_timer_audit(struct tcpcb *tp, struct tcp_rack *rack, struct sockbuf *sb) { @@ -11608,7 +11572,6 @@ rack_get_pacing_delay(struct tcp_rack *rack, struct tcpcb *tp, uint32_t len, str rate_wanted, rack->r_ctl.crte->rate, __LINE__, err); } - } } if (rack_limit_time_with_srtt && @@ -14117,7 +14080,6 @@ rack_set_sockopt(struct socket *so, struct sockopt *sopt, int32_t error = 0, optval; uint16_t ca, ss; - switch (sopt->sopt_name) { case TCP_RACK_PROP_RATE: /* URL:prop_rate */ case TCP_RACK_PROP : /* URL:prop */ diff --git a/sys/netinet/tcp_stacks/rack_bbr_common.c b/sys/netinet/tcp_stacks/rack_bbr_common.c index e8b3494081c7..b7d758f5fcd8 100644 --- a/sys/netinet/tcp_stacks/rack_bbr_common.c +++ b/sys/netinet/tcp_stacks/rack_bbr_common.c @@ -161,7 +161,6 @@ again: } #endif - /* * The function ctf_process_inbound_raw() is used by * transport developers to do the steps needed to @@ -674,7 +673,6 @@ ctf_process_rst(struct mbuf *m, struct tcphdr *th, struct socket *so, struct tcp if ((SEQ_GEQ(th->th_seq, (tp->last_ack_sent - 1)) && SEQ_LT(th->th_seq, tp->last_ack_sent + tp->rcv_wnd)) || (tp->rcv_wnd == 0 && tp->last_ack_sent == th->th_seq)) { - KASSERT(tp->t_state != TCPS_SYN_SENT, ("%s: TH_RST for TCPS_SYN_SENT th %p tp %p", __func__, th, tp)); diff --git a/sys/netinet/tcp_stacks/tcp_bbr.h b/sys/netinet/tcp_stacks/tcp_bbr.h index 0ba22e17e4f2..30b897cff218 100644 --- a/sys/netinet/tcp_stacks/tcp_bbr.h +++ b/sys/netinet/tcp_stacks/tcp_bbr.h @@ -226,7 +226,6 @@ struct bbr_rtt_sample { #define BBR_RT_SET_GRADIENT 0x200 #define BBR_RT_TS_VALID 0x400 - struct bbr_log { union { struct bbr_sendmap *rsm; /* For alloc/free */ @@ -294,7 +293,6 @@ struct bbr_log_sysctl_out { #define BBR_SEES_STRETCH_ACK 1 #define BBR_SEES_COMPRESSED_ACKS 2 - /* * As we get each SACK we wade through the * rc_map and mark off what is acked. @@ -345,7 +343,6 @@ struct bbr_log_sysctl_out { #define BBR_TIME_TO_SECONDS(a) (a / USECS_IN_SECOND) #define BBR_TIME_TO_MILLI(a) (a / MS_IN_USEC) - /* BBR keeps time in usec's so we divide by 1000 and round up */ #define BBR_TS_TO_MS(t) ((t+999)/MS_IN_USEC) @@ -504,7 +501,6 @@ struct bbr_opts_stats { uint64_t tcp_policer_det; }; - #ifdef _KERNEL #define BBR_STAT_SIZE (sizeof(struct bbr_stats)/sizeof(uint64_t)) extern counter_u64_t bbr_stat_arry[BBR_STAT_SIZE]; @@ -524,7 +520,6 @@ extern counter_u64_t bbr_opts_arry[BBR_OPTS_SIZE]; #define BBR_RECOVERY_HIGHRTT 3 #define BBR_RECOVERY_EXTREMERTT 4 - struct bbr_control { /*******************************/ /* Cache line 2 from bbr start */ @@ -561,7 +556,6 @@ struct bbr_control { uint32_t cur_rtt; /* Last RTT from ack */ - uint32_t rc_went_idle_time; /* Used for persits to see if its * probe-rtt qualified */ uint32_t rc_pace_max_segs:17, /* How much in any single TSO we send Lock(a) */ @@ -580,7 +574,6 @@ struct bbr_control { * pkt-epoch */ uint32_t r_measurement_count; /* count of measurement applied lock(a) */ - uint32_t rc_last_tlp_seq; /* Last tlp sequence Lock(a) */ uint16_t rc_reorder_shift; /* Socket option value Lock(a) */ uint16_t rc_pkt_delay; /* Socket option value Lock(a) */ @@ -664,7 +657,6 @@ struct bbr_control { uint64_t rc_bbr_lastbtlbw; /* For startup, what was last btlbw I * saw to check the 20% gain Lock(a) */ - uint32_t rc_bbr_cwnd_gain; /* Current cwnd gain Lock(a) */ uint32_t rc_pkt_epoch_loss_rate; /* pkt-epoch loss rate */ @@ -740,7 +732,6 @@ struct bbr_control { }; - struct socket; struct tcp_bbr { /* First cache line 0x00 */ diff --git a/sys/netinet/tcp_stacks/tcp_rack.h b/sys/netinet/tcp_stacks/tcp_rack.h index f60e49b65e6c..c45b7c75cd56 100644 --- a/sys/netinet/tcp_stacks/tcp_rack.h +++ b/sys/netinet/tcp_stacks/tcp_rack.h @@ -99,7 +99,6 @@ struct rack_rtt_sample { #define RACK_LOG_TYPE_ALLOC 0x04 #define RACK_LOG_TYPE_FREE 0x05 - struct rack_log { union { struct rack_sendmap *rsm; /* For alloc/free */ |