Lines Matching defs:icsk
120 void clean_acked_data_enable(struct inet_connection_sock *icsk,
123 icsk->icsk_clean_acked = cad;
128 void clean_acked_data_disable(struct inet_connection_sock *icsk)
131 icsk->icsk_clean_acked = NULL;
229 struct inet_connection_sock *icsk = inet_csk(sk);
230 const unsigned int lss = icsk->icsk_ack.last_seg_size;
233 icsk->icsk_ack.last_seg_size = 0;
239 if (len >= icsk->icsk_ack.rcv_mss) {
244 if (unlikely(len != icsk->icsk_ack.rcv_mss)) {
250 icsk->icsk_ack.rcv_mss = min_t(unsigned int, len,
253 if (unlikely(len > icsk->icsk_ack.rcv_mss +
268 icsk->icsk_ack.pending |= ICSK_ACK_PUSHED;
289 icsk->icsk_ack.last_seg_size = len;
291 icsk->icsk_ack.rcv_mss = len;
295 if (icsk->icsk_ack.pending & ICSK_ACK_PUSHED)
296 icsk->icsk_ack.pending |= ICSK_ACK_PUSHED2;
297 icsk->icsk_ack.pending |= ICSK_ACK_PUSHED;
303 struct inet_connection_sock *icsk = inet_csk(sk);
304 unsigned int quickacks = tcp_sk(sk)->rcv_wnd / (2 * icsk->icsk_ack.rcv_mss);
309 if (quickacks > icsk->icsk_ack.quick)
310 icsk->icsk_ack.quick = quickacks;
315 struct inet_connection_sock *icsk = inet_csk(sk);
319 icsk->icsk_ack.ato = TCP_ATO_MIN;
328 const struct inet_connection_sock *icsk = inet_csk(sk);
332 (icsk->icsk_ack.quick && !inet_csk_in_pingpong_mode(sk));
597 struct inet_connection_sock *icsk = inet_csk(sk);
601 icsk->icsk_ack.quick = 0;
794 struct inet_connection_sock *icsk = inet_csk(sk);
805 if (!icsk->icsk_ack.ato) {
810 icsk->icsk_ack.ato = TCP_ATO_MIN;
812 int m = now - icsk->icsk_ack.lrcvtime;
816 icsk->icsk_ack.ato = (icsk->icsk_ack.ato >> 1) + TCP_ATO_MIN / 2;
817 } else if (m < icsk->icsk_ack.ato) {
818 icsk->icsk_ack.ato = (icsk->icsk_ack.ato >> 1) + m;
819 if (icsk->icsk_ack.ato > icsk->icsk_rto)
820 icsk->icsk_ack.ato = icsk->icsk_rto;
821 } else if (m > icsk->icsk_rto) {
828 icsk->icsk_ack.lrcvtime = now;
2156 const struct inet_connection_sock *icsk = inet_csk(sk);
2159 bool new_recovery = icsk->icsk_ca_state < TCP_CA_Recovery;
2165 if (icsk->icsk_ca_state <= TCP_CA_Disorder ||
2167 (icsk->icsk_ca_state == TCP_CA_Loss && !icsk->icsk_retransmits)) {
2170 tp->snd_ssthresh = icsk->icsk_ca_ops->ssthresh(sk);
2182 if (icsk->icsk_ca_state <= TCP_CA_Disorder &&
2196 (new_recovery || icsk->icsk_retransmits) &&
2516 const struct inet_connection_sock *icsk = inet_csk(sk);
2518 tcp_snd_cwnd_set(tp, icsk->icsk_ca_ops->undo_cwnd(sk));
2732 struct inet_connection_sock *icsk = inet_csk(sk);
2734 icsk->icsk_mtup.search_high = icsk->icsk_mtup.probe_size - 1;
2735 icsk->icsk_mtup.probe_size = 0;
2742 struct inet_connection_sock *icsk = inet_csk(sk);
2748 do_div(val, icsk->icsk_mtup.probe_size);
2756 icsk->icsk_mtup.search_low = icsk->icsk_mtup.probe_size;
2757 icsk->icsk_mtup.probe_size = 0;
2758 tcp_sync_mss(sk, icsk->icsk_pmtu_cookie);
2768 const struct inet_connection_sock *icsk = inet_csk(sk);
2808 if (icsk->icsk_ca_state != TCP_CA_Loss) {
2974 struct inet_connection_sock *icsk = inet_csk(sk);
2998 if (icsk->icsk_ca_state == TCP_CA_Open) {
3002 switch (icsk->icsk_ca_state) {
3023 switch (icsk->icsk_ca_state) {
3035 if (icsk->icsk_ca_state != TCP_CA_Recovery) {
3047 if (!(icsk->icsk_ca_state == TCP_CA_Open ||
3059 if (icsk->icsk_ca_state <= TCP_CA_Disorder)
3069 if (icsk->icsk_ca_state < TCP_CA_CWR &&
3070 icsk->icsk_mtup.probe_size &&
3168 const struct inet_connection_sock *icsk = inet_csk(sk);
3170 icsk->icsk_ca_ops->cong_avoid(sk, ack, acked);
3179 const struct inet_connection_sock *icsk = inet_csk(sk);
3193 if (icsk->icsk_pending == ICSK_TIME_REO_TIMEOUT ||
3194 icsk->icsk_pending == ICSK_TIME_LOSS_PROBE) {
3260 const struct inet_connection_sock *icsk = inet_csk(sk);
3391 if (unlikely(icsk->icsk_mtup.probe_size &&
3427 if (icsk->icsk_ca_ops->pkts_acked) {
3433 icsk->icsk_ca_ops->pkts_acked(sk, &sample);
3441 icsk = inet_csk(sk);
3444 tp->lost_out, icsk->icsk_ca_state);
3449 tp->sacked_out, icsk->icsk_ca_state);
3454 tp->retrans_out, icsk->icsk_ca_state);
3464 struct inet_connection_sock *icsk = inet_csk(sk);
3472 icsk->icsk_backoff = 0;
3473 icsk->icsk_probes_tstamp = 0;
3516 const struct inet_connection_sock *icsk = inet_csk(sk);
3518 if (icsk->icsk_ca_ops->cong_control) {
3519 icsk->icsk_ca_ops->cong_control(sk, rs);
3741 const struct inet_connection_sock *icsk = inet_csk(sk);
3743 if (icsk->icsk_ca_ops->in_ack_event)
3744 icsk->icsk_ca_ops->in_ack_event(sk, flags);
3786 struct inet_connection_sock *icsk = inet_csk(sk);
3833 icsk->icsk_retransmits = 0;
3837 if (icsk->icsk_clean_acked)
3838 icsk->icsk_clean_acked(sk, ack);
3906 icsk->icsk_probes_out = 0;
6077 struct inet_connection_sock *icsk = inet_csk(sk);
6081 icsk->icsk_af_ops->rebuild_header(sk);
6098 if (!icsk->icsk_ca_initialized)
6106 struct inet_connection_sock *icsk = inet_csk(sk);
6109 icsk->icsk_ack.lrcvtime = tcp_jiffies32;
6112 icsk->icsk_af_ops->sk_rx_dst_set(sk, skb);
6224 struct inet_connection_sock *icsk = inet_csk(sk);
6247 if (icsk->icsk_retransmits == 0)
6327 tcp_sync_mss(sk, icsk->icsk_pmtu_cookie);
6351 READ_ONCE(icsk->icsk_accept_queue.rskq_defer_accept) ||
6418 tcp_sync_mss(sk, icsk->icsk_pmtu_cookie);
6498 struct inet_connection_sock *icsk = inet_csk(sk);
6528 acceptable = icsk->icsk_af_ops->conn_request(sk, skb) >= 0;