Lines Matching defs:start_seq
1006 static u32 tcp_dsack_seen(struct tcp_sock *tp, u32 start_seq,
1011 if (!before(start_seq, end_seq))
1014 seq_len = end_seq - start_seq;
1197 * Implements also blockage to start_seq wrap-around. Problem lies in the
1198 * fact that though start_seq (s) is before end_seq (i.e., not reversed),
1200 * happens when start_seq resides between end_seq wrap (e_w) and snd_nxt
1211 * crazy SACK blocks. Doing this check for start_seq alone closes somewhat
1230 u32 start_seq, u32 end_seq)
1233 if (after(end_seq, tp->snd_nxt) || !before(start_seq, end_seq))
1236 /* Nasty start_seq wrap-around check (see comments above) */
1237 if (!before(start_seq, tp->snd_nxt))
1241 * start_seq == snd_una is non-sensical (see comments above)
1243 if (after(start_seq, tp->snd_una))
1253 if (!before(start_seq, tp->undo_marker))
1261 * start_seq < undo_marker and end_seq >= undo_marker.
1263 return !before(start_seq, end_seq - tp->max_window);
1271 u32 start_seq_0 = get_unaligned_be32(&sp[0].start_seq);
1279 u32 start_seq_1 = get_unaligned_be32(&sp[1].start_seq);
1314 u32 start_seq, u32 end_seq)
1321 in_sack = !after(start_seq, TCP_SKB_CB(skb)->seq) &&
1325 after(TCP_SKB_CB(skb)->end_seq, start_seq)) {
1327 in_sack = !after(start_seq, TCP_SKB_CB(skb)->seq);
1330 pkt_len = start_seq - TCP_SKB_CB(skb)->seq;
1364 u32 start_seq, u32 end_seq,
1376 before(start_seq, state->reord))
1377 state->reord = start_seq;
1402 if (before(start_seq,
1404 before(start_seq, state->reord))
1405 state->reord = start_seq;
1428 before(start_seq, TCP_SKB_CB(tp->lost_skb_hint)->seq))
1454 u32 start_seq = TCP_SKB_CB(skb)->seq; /* start of newly-SACKed */
1455 u32 end_seq = start_seq + shifted; /* end of newly-SACKed */
1466 start_seq, end_seq, dup_sack, pcount,
1563 u32 start_seq, u32 end_seq,
1594 in_sack = !after(start_seq, TCP_SKB_CB(skb)->seq) &&
1608 if (!after(TCP_SKB_CB(skb)->end_seq, start_seq))
1617 in_sack = !after(start_seq, TCP_SKB_CB(skb)->seq);
1702 u32 start_seq, u32 end_seq,
1719 next_dup->start_seq,
1731 start_seq, end_seq, dup_sack);
1741 start_seq,
1810 if (before(next_dup->start_seq, skip_to_seq)) {
1811 skb = tcp_sacktag_skip(skb, sk, next_dup->start_seq);
1813 next_dup->start_seq, next_dup->end_seq,
1866 sp[used_sacks].start_seq = get_unaligned_be32(&sp_wire[i].start_seq);
1870 sp[used_sacks].start_seq,
1906 if (after(sp[j].start_seq, sp[j + 1].start_seq)) {
1926 while (tcp_sack_cache_ok(tp, cache) && !cache->start_seq &&
1932 u32 start_seq = sp[i].start_seq;
1942 !before(start_seq, cache->end_seq))
1947 after(end_seq, cache->start_seq)) {
1950 if (before(start_seq, cache->start_seq)) {
1951 skb = tcp_sacktag_skip(skb, sk, start_seq);
1954 start_seq,
1955 cache->start_seq,
1983 if (!before(start_seq, tcp_highest_sack_seq(tp))) {
1988 skb = tcp_sacktag_skip(skb, sk, start_seq);
1992 start_seq, end_seq, dup_sack);
2000 tp->recv_sack_cache[i].start_seq = 0;
3278 const u32 start_seq = scb->seq;
3306 if (before(start_seq, reord))
3307 reord = start_seq;
4463 if (!after(seq, sp->end_seq) && !after(sp->start_seq, end_seq)) {
4464 if (before(seq, sp->start_seq))
4465 sp->start_seq = seq;
4488 tp->duplicate_sack[0].start_seq = seq;
4510 if (TCP_SKB_CB(skb)->seq == tcp_sk(sk)->duplicate_sack[0].start_seq &&
4550 if (tcp_sack_extend(sp, swalk->start_seq, swalk->end_seq)) {
4635 sp->start_seq = seq;
4656 if (!before(tp->rcv_nxt, sp->start_seq)) {
4852 tp->selective_acks[0].start_seq = seq;