/kernel/linux/linux-5.10/net/ipv4/ |
H A D | tcp_recovery.c | 50 * FACK: sequence delta to highest sacked sequence (sequence space) 77 if ((scb->sacked & TCPCB_LOST) && in tcp_rack_detect_loss() 78 !(scb->sacked & TCPCB_SACKED_RETRANS)) in tcp_rack_detect_loss() 123 void tcp_rack_advance(struct tcp_sock *tp, u8 sacked, u32 end_seq, in tcp_rack_advance() argument 129 if (rtt_us < tcp_min_rtt(tp) && (sacked & TCPCB_RETRANS)) { in tcp_rack_advance() 130 /* If the sacked packet was retransmitted, it's ambiguous in tcp_rack_advance() 132 * retransmission) was sacked. in tcp_rack_advance() 230 if (TCP_SKB_CB(skb)->sacked & TCPCB_LOST) in tcp_newreno_mark_lost()
|
H A D | tcp_rate.c | 71 /* When an skb is sacked or acked, we fill in the rate sample with the (prior) 96 rs->is_retrans = scb->sacked & TCPCB_RETRANS; in tcp_rate_skb_delivered() 106 /* Mark off the skb delivered once it's sacked to avoid being in tcp_rate_skb_delivered() 110 if (scb->sacked & TCPCB_SACKED_ACKED) in tcp_rate_skb_delivered()
|
H A D | tcp_input.c | 102 #define FLAG_SACK_RENEGING 0x2000 /* snd_una advanced to a sacked seq */ 1026 /* It's reordering when higher sequence was delivered (i.e. sacked) before 1085 __u8 sacked = TCP_SKB_CB(skb)->sacked; in tcp_mark_skb_lost() local 1088 if (sacked & TCPCB_SACKED_ACKED) in tcp_mark_skb_lost() 1092 if (sacked & TCPCB_LOST) { in tcp_mark_skb_lost() 1093 if (sacked & TCPCB_SACKED_RETRANS) { in tcp_mark_skb_lost() 1095 TCP_SKB_CB(skb)->sacked &= ~TCPCB_SACKED_RETRANS; in tcp_mark_skb_lost() 1103 TCP_SKB_CB(skb)->sacked |= TCPCB_LOST; in tcp_mark_skb_lost() 1131 * (L|S|R is logically valid, it could occur when L|R is sacked, 1343 tcp_sacktag_one(struct sock *sk, struct tcp_sacktag_state *state, u8 sacked, u32 start_seq, u32 end_seq, int dup_sack, int pcount, u64 xmit_time) tcp_sacktag_one() argument 3247 u8 sacked = scb->sacked; tcp_clean_rtx_queue() local [all...] |
H A D | tcp_output.c | 396 TCP_SKB_CB(skb)->sacked = 0; in tcp_init_nondata_skb() 1466 if (TCP_SKB_CB(skb)->sacked & TCPCB_SACKED_ACKED) in tcp_adjust_pcount() 1468 if (TCP_SKB_CB(skb)->sacked & TCPCB_SACKED_RETRANS) in tcp_adjust_pcount() 1470 if (TCP_SKB_CB(skb)->sacked & TCPCB_LOST) in tcp_adjust_pcount() 1479 (TCP_SKB_CB(skb)->sacked & TCPCB_SACKED_ACKED)) in tcp_adjust_pcount() 1587 TCP_SKB_CB(buff)->sacked = TCP_SKB_CB(skb)->sacked; in tcp_fragment() 2150 TCP_SKB_CB(buff)->sacked = 0; in tso_fragment() 2407 TCP_SKB_CB(nskb)->sacked = 0; in tcp_mtu_probe() 3079 TCP_SKB_CB(skb)->sacked | in tcp_collapse_retrans() 3333 __u8 sacked; tcp_xmit_retransmit_queue() local [all...] |
H A D | tcp.c | 686 tcb->sacked = 0; in skb_entail() 1329 TCP_SKB_CB(skb)->sacked |= TCPCB_REPAIRED; in tcp_sendmsg_locked()
|
H A D | tcp_ipv4.c | 1942 TCP_SKB_CB(skb)->sacked = 0; in tcp_v4_fill_cb()
|
/kernel/linux/linux-6.6/net/ipv4/ |
H A D | tcp_recovery.c | 45 * FACK: sequence delta to highest sacked sequence (sequence space) 72 if ((scb->sacked & TCPCB_LOST) && in tcp_rack_detect_loss() 73 !(scb->sacked & TCPCB_SACKED_RETRANS)) in tcp_rack_detect_loss() 118 void tcp_rack_advance(struct tcp_sock *tp, u8 sacked, u32 end_seq, in tcp_rack_advance() argument 124 if (rtt_us < tcp_min_rtt(tp) && (sacked & TCPCB_RETRANS)) { in tcp_rack_advance() 125 /* If the sacked packet was retransmitted, it's ambiguous in tcp_rack_advance() 127 * retransmission) was sacked. in tcp_rack_advance() 227 if (TCP_SKB_CB(skb)->sacked & TCPCB_LOST) in tcp_newreno_mark_lost()
|
H A D | tcp_rate.c | 72 /* When an skb is sacked or acked, we fill in the rate sample with the (prior) 98 rs->is_retrans = scb->sacked & TCPCB_RETRANS; in tcp_rate_skb_delivered() 108 /* Mark off the skb delivered once it's sacked to avoid being in tcp_rate_skb_delivered() 112 if (scb->sacked & TCPCB_SACKED_ACKED) in tcp_rate_skb_delivered()
|
H A D | tcp_input.c | 99 #define FLAG_SACK_RENEGING 0x2000 /* snd_una advanced to a sacked seq */ 1045 /* It's reordering when higher sequence was delivered (i.e. sacked) before 1104 __u8 sacked = TCP_SKB_CB(skb)->sacked; in tcp_mark_skb_lost() local 1107 if (sacked & TCPCB_SACKED_ACKED) in tcp_mark_skb_lost() 1111 if (sacked & TCPCB_LOST) { in tcp_mark_skb_lost() 1112 if (sacked & TCPCB_SACKED_RETRANS) { in tcp_mark_skb_lost() 1114 TCP_SKB_CB(skb)->sacked &= ~TCPCB_SACKED_RETRANS; in tcp_mark_skb_lost() 1122 TCP_SKB_CB(skb)->sacked |= TCPCB_LOST; in tcp_mark_skb_lost() 1150 * (L|S|R is logically valid, it could occur when L|R is sacked, 1362 tcp_sacktag_one(struct sock *sk, struct tcp_sacktag_state *state, u8 sacked, u32 start_seq, u32 end_seq, int dup_sack, int pcount, u64 xmit_time) tcp_sacktag_one() argument 3279 u8 sacked = scb->sacked; tcp_clean_rtx_queue() local [all...] |
H A D | tcp_output.c | 1478 if (TCP_SKB_CB(skb)->sacked & TCPCB_SACKED_ACKED) in tcp_adjust_pcount() 1480 if (TCP_SKB_CB(skb)->sacked & TCPCB_SACKED_RETRANS) in tcp_adjust_pcount() 1482 if (TCP_SKB_CB(skb)->sacked & TCPCB_LOST) in tcp_adjust_pcount() 1491 (TCP_SKB_CB(skb)->sacked & TCPCB_SACKED_ACKED)) in tcp_adjust_pcount() 1598 TCP_SKB_CB(buff)->sacked = TCP_SKB_CB(skb)->sacked; in tcp_fragment() 3177 TCP_SKB_CB(skb)->sacked |= TCP_SKB_CB(next_skb)->sacked & TCPCB_EVER_RETRANS; in tcp_collapse_retrans() 3201 if (TCP_SKB_CB(skb)->sacked & TCPCB_SACKED_ACKED) in tcp_can_collapse() 3363 TCP_SKB_CB(skb)->sacked | in __tcp_retransmit_skb() 3423 __u8 sacked; tcp_xmit_retransmit_queue() local [all...] |
H A D | tcp_ipv4.c | 1974 TCP_SKB_CB(skb)->sacked = 0; in tcp_v4_fill_cb()
|
H A D | tcp.c | 1169 TCP_SKB_CB(skb)->sacked |= TCPCB_REPAIRED; in tcp_sendmsg_locked()
|
/kernel/linux/linux-5.10/net/sctp/ |
H A D | outqueue.c | 199 INIT_LIST_HEAD(&q->sacked); in sctp_outq_init() 225 list_for_each_safe(lchunk, temp, &q->sacked) { in __sctp_outq_teardown() 1346 list_for_each_safe(lchunk, temp, &q->sacked) { in sctp_outq_sack() 1400 * list and move chunks that are acked by the Cumulative TSN Ack to q->sacked. 1405 * Instead of printing 'sacked' or 'kept' for each TSN on the 1532 &q->sacked); in sctp_check_transmitted()
|
H A D | socket.c | 180 list_for_each_entry(chunk, &q->sacked, transmitted_list) in sctp_for_each_tx_datachunk()
|
/kernel/linux/linux-6.6/net/sctp/ |
H A D | outqueue.c | 199 INIT_LIST_HEAD(&q->sacked); in sctp_outq_init() 225 list_for_each_safe(lchunk, temp, &q->sacked) { in __sctp_outq_teardown() 1372 list_for_each_safe(lchunk, temp, &q->sacked) { in sctp_outq_sack() 1426 * list and move chunks that are acked by the Cumulative TSN Ack to q->sacked. 1431 * Instead of printing 'sacked' or 'kept' for each TSN on the 1558 &q->sacked); in sctp_check_transmitted()
|
H A D | socket.c | 182 list_for_each_entry(chunk, &q->sacked, transmitted_list) in sctp_for_each_tx_datachunk()
|
/kernel/linux/common_modules/newip/third_party/linux-5.10/net/newip/ |
H A D | tcp_nip_output.c | 351 TCP_SKB_CB(skb)->sacked = 0; in tcp_nip_init_nondata_skb() 1203 TCP_SKB_CB(skb)->sacked |= TCPCB_RETRANS; in tcp_nip_retransmit_skb()
|
H A D | tcp_nip.c | 1421 tcb->sacked = 0; in skb_nip_entail() 1903 TCP_SKB_CB(skb)->sacked = 0; in tcp_nip_fill_cb()
|
/kernel/linux/linux-5.10/include/net/ |
H A D | tcp.h | 848 __u8 sacked; /* State flags for SACK. */ member 2111 extern void tcp_rack_advance(struct tcp_sock *tp, u8 sacked, u32 end_seq,
|
/kernel/linux/linux-6.6/include/net/ |
H A D | tcp.h | 873 __u8 sacked; /* State flags for SACK. */ member 2193 extern void tcp_rack_advance(struct tcp_sock *tp, u8 sacked, u32 end_seq,
|
/kernel/linux/linux-5.10/include/net/sctp/ |
H A D | structs.h | 796 /* Has this transport moved the ctsn since we last sacked */ 1069 /* These are chunks that have been sacked but are above the 1072 struct list_head sacked; member
|
/kernel/linux/linux-6.6/include/net/sctp/ |
H A D | structs.h | 803 /* Has this transport moved the ctsn since we last sacked */ 1094 /* These are chunks that have been sacked but are above the 1097 struct list_head sacked; member
|
/kernel/linux/linux-5.10/net/ipv6/ |
H A D | tcp_ipv6.c | 1600 TCP_SKB_CB(skb)->sacked = 0; in tcp_v6_fill_cb()
|
/kernel/linux/linux-6.6/net/ipv6/ |
H A D | tcp_ipv6.c | 1582 TCP_SKB_CB(skb)->sacked = 0; in tcp_v6_fill_cb()
|