/kernel/linux/linux-5.10/net/ipv4/ |
H A D | tcp_rate.c | 8 * A rate sample records the rate at which the network delivered packets 16 * compression: packets can temporarily appear to be delivered much quicker 67 TCP_SKB_CB(skb)->tx.delivered = tp->delivered; in tcp_rate_skb_sent() 93 rs->prior_delivered = scb->tx.delivered; in tcp_rate_skb_delivered() 106 /* Mark off the skb delivered once it's sacked to avoid being in tcp_rate_skb_delivered() 115 void tcp_rate_gen(struct sock *sk, u32 delivered, u32 lost, in tcp_rate_gen() argument 122 if (tp->app_limited && after(tp->delivered, tp->app_limited)) in tcp_rate_gen() 129 if (delivered) in tcp_rate_gen() 132 rs->acked_sacked = delivered; /* freshl in tcp_rate_gen() [all...] |
H A D | tcp_bbr.c | 7 * bottleneck_bandwidth = windowed_max(delivered / elapsed, 10 round trips) 93 u32 next_rtt_delivered; /* scb->tx.delivered at end of round */ 106 u32 lt_last_delivered; /* LT intvl start: tp->delivered */ 184 /* If lost/delivered ratio > 20%, interval is "lossy" and we may be policed: */ 495 bbr->next_rtt_delivered = tp->delivered; /* start round now */ in bbr_set_cwnd_to_recover_or_restore() 540 else if (cwnd < target_cwnd || tp->delivered < TCP_INIT_CWND) in bbr_set_cwnd() 638 bbr->lt_last_delivered = tp->delivered; in bbr_reset_lt_bw_sampling_interval() 689 u32 lost, delivered; in bbr_lt_bw_sampling() local 735 /* Calculate packets lost and delivered in sampling interval. */ in bbr_lt_bw_sampling() 737 delivered in bbr_lt_bw_sampling() [all...] |
H A D | tcp_dctcp.c | 71 ca->old_delivered = tp->delivered; in dctcp_reset() 125 u32 delivered = tp->delivered - ca->old_delivered; in dctcp_update_alpha() local 131 delivered_ce /= max(1U, delivered); in dctcp_update_alpha() 200 (tp->delivered - ca->old_delivered); in dctcp_get_info()
|
/kernel/linux/linux-6.6/net/ipv4/ |
H A D | tcp_rate.c | 8 * A rate sample records the rate at which the network delivered packets 16 * compression: packets can temporarily appear to be delivered much quicker 67 TCP_SKB_CB(skb)->tx.delivered = tp->delivered; in tcp_rate_skb_sent() 95 rs->prior_delivered = scb->tx.delivered; in tcp_rate_skb_delivered() 108 /* Mark off the skb delivered once it's sacked to avoid being in tcp_rate_skb_delivered() 117 void tcp_rate_gen(struct sock *sk, u32 delivered, u32 lost, in tcp_rate_gen() argument 124 if (tp->app_limited && after(tp->delivered, tp->app_limited)) in tcp_rate_gen() 131 if (delivered) in tcp_rate_gen() 134 rs->acked_sacked = delivered; /* freshl in tcp_rate_gen() [all...] |
H A D | tcp_bbr.c | 7 * bottleneck_bandwidth = windowed_max(delivered / elapsed, 10 round trips) 95 u32 next_rtt_delivered; /* scb->tx.delivered at end of round */ 108 u32 lt_last_delivered; /* LT intvl start: tp->delivered */ 186 /* If lost/delivered ratio > 20%, interval is "lossy" and we may be policed: */ 497 bbr->next_rtt_delivered = tp->delivered; /* start round now */ in bbr_set_cwnd_to_recover_or_restore() 542 else if (cwnd < target_cwnd || tp->delivered < TCP_INIT_CWND) in bbr_set_cwnd() 640 bbr->lt_last_delivered = tp->delivered; in bbr_reset_lt_bw_sampling_interval() 691 u32 lost, delivered; in bbr_lt_bw_sampling() local 737 /* Calculate packets lost and delivered in sampling interval. */ in bbr_lt_bw_sampling() 739 delivered in bbr_lt_bw_sampling() [all...] |
H A D | tcp_dctcp.c | 74 ca->old_delivered = tp->delivered; in dctcp_reset() 123 u32 delivered = tp->delivered - ca->old_delivered; in dctcp_update_alpha() local 128 if (delivered > 0) { in dctcp_update_alpha() 136 ce_ratio = (delivered_ce << TCP_PLB_SCALE) / delivered; in dctcp_update_alpha() 150 delivered_ce /= max(1U, delivered); in dctcp_update_alpha() 223 (tp->delivered - ca->old_delivered); in dctcp_get_info()
|
/kernel/linux/linux-5.10/tools/testing/selftests/bpf/prog_tests/ |
H A D | tcp_rtt.c | 9 __u32 delivered; member 47 __u32 dsack_dups, __u32 delivered, __u32 delivered_ce, in verify_sk() 70 if (val.delivered != delivered) { in verify_sk() 71 log_err("%s: unexpected bpf_tcp_sock.delivered %d != %d", in verify_sk() 72 msg, val.delivered, delivered); in verify_sk() 130 /*delivered=*/1, in run_test() 144 /*delivered=*/2, in run_test() 46 verify_sk(int map_fd, int client_fd, const char *msg, __u32 invoked, __u32 dsack_dups, __u32 delivered, __u32 delivered_ce, __u32 icsk_retransmits) verify_sk() argument
|
/kernel/linux/linux-6.6/tools/testing/selftests/bpf/prog_tests/ |
H A D | tcp_rtt.c | 10 __u32 delivered; member 47 __u32 dsack_dups, __u32 delivered, __u32 delivered_ce, in verify_sk() 68 if (val.delivered != delivered) { in verify_sk() 69 log_err("%s: unexpected bpf_tcp_sock.delivered %d != %d", in verify_sk() 70 msg, val.delivered, delivered); in verify_sk() 120 /*delivered=*/1, in run_test() 134 /*delivered=*/2, in run_test() 46 verify_sk(int map_fd, int client_fd, const char *msg, __u32 invoked, __u32 dsack_dups, __u32 delivered, __u32 delivered_ce, __u32 icsk_retransmits) verify_sk() argument
|
/kernel/linux/linux-5.10/tools/testing/selftests/bpf/progs/ |
H A D | tcp_rtt.c | 11 __u32 delivered; member 55 storage->delivered = tcp_sk->delivered; in _sockops()
|
H A D | bpf_dctcp.c | 49 ca->old_delivered = tp->delivered; in dctcp_reset() 98 __u32 delivered = tp->delivered - ca->old_delivered; in BPF_PROG() local 104 delivered_ce /= max(1U, delivered); in BPF_PROG()
|
/kernel/linux/linux-6.6/tools/testing/selftests/bpf/progs/ |
H A D | tcp_rtt.c | 10 __u32 delivered; member 54 storage->delivered = tcp_sk->delivered; in _sockops()
|
H A D | bpf_dctcp.c | 56 ca->old_delivered = tp->delivered; in dctcp_reset() 132 __u32 delivered = tp->delivered - ca->old_delivered; in BPF_PROG() local 138 delivered_ce /= max(1U, delivered); in BPF_PROG()
|
H A D | tcp_ca_write_sk_pacing.c | 49 tp->app_limited = (tp->delivered + tcp_packets_in_flight(tp)) ?: 1; in BPF_PROG()
|
/kernel/linux/linux-5.10/drivers/i2c/busses/ |
H A D | i2c-imx-lpi2c.c | 103 unsigned int delivered; member 343 remaining = lpi2c_imx->msglen - lpi2c_imx->delivered; in lpi2c_imx_set_rx_watermark() 360 if (lpi2c_imx->delivered == lpi2c_imx->msglen) in lpi2c_imx_write_txfifo() 363 data = lpi2c_imx->tx_buf[lpi2c_imx->delivered++]; in lpi2c_imx_write_txfifo() 368 if (lpi2c_imx->delivered < lpi2c_imx->msglen) in lpi2c_imx_write_txfifo() 384 lpi2c_imx->rx_buf[lpi2c_imx->delivered++] = data & 0xff; in lpi2c_imx_read_rxfifo() 396 remaining = lpi2c_imx->msglen - lpi2c_imx->delivered; in lpi2c_imx_read_rxfifo() 412 } else if (!(lpi2c_imx->delivered & 0xff)) { in lpi2c_imx_read_rxfifo() 467 lpi2c_imx->delivered = 0; in lpi2c_imx_xfer()
|
/kernel/linux/linux-6.6/drivers/i2c/busses/ |
H A D | i2c-imx-lpi2c.c | 103 unsigned int delivered; member 346 remaining = lpi2c_imx->msglen - lpi2c_imx->delivered; in lpi2c_imx_set_rx_watermark() 363 if (lpi2c_imx->delivered == lpi2c_imx->msglen) in lpi2c_imx_write_txfifo() 366 data = lpi2c_imx->tx_buf[lpi2c_imx->delivered++]; in lpi2c_imx_write_txfifo() 371 if (lpi2c_imx->delivered < lpi2c_imx->msglen) in lpi2c_imx_write_txfifo() 387 lpi2c_imx->rx_buf[lpi2c_imx->delivered++] = data & 0xff; in lpi2c_imx_read_rxfifo() 399 remaining = lpi2c_imx->msglen - lpi2c_imx->delivered; in lpi2c_imx_read_rxfifo() 415 } else if (!(lpi2c_imx->delivered & 0xff)) { in lpi2c_imx_read_rxfifo() 470 lpi2c_imx->delivered = 0; in lpi2c_imx_xfer()
|
/kernel/linux/linux-5.10/net/bridge/ |
H A D | br_forward.c | 302 goto delivered; in br_multicast_flood() 307 goto delivered; in br_multicast_flood() 315 delivered: in br_multicast_flood()
|
/kernel/linux/linux-6.6/net/bridge/ |
H A D | br_forward.c | 316 goto delivered; in br_multicast_flood() 321 goto delivered; in br_multicast_flood() 329 delivered: in br_multicast_flood()
|
/kernel/linux/linux-5.10/samples/bpf/ |
H A D | tcp_dumpstats_kern.c | 62 bpf_printk("dsack_dups=%u delivered=%u\n", in _sockops() 63 tcp_sk->dsack_dups, tcp_sk->delivered); in _sockops()
|
/kernel/linux/linux-6.6/samples/bpf/ |
H A D | tcp_dumpstats_kern.c | 62 bpf_printk("dsack_dups=%u delivered=%u\n", in _sockops() 63 tcp_sk->dsack_dups, tcp_sk->delivered); in _sockops()
|
/kernel/linux/linux-5.10/drivers/acpi/ |
H A D | cppc_acpi.c | 166 fb_ctrs.reference, fb_ctrs.delivered); in show_feedback_ctrs() 1170 u64 delivered, reference, ref_perf, ctr_wrap_time; in cppc_get_perf_ctrs() local 1207 cpc_read(cpunum, delivered_reg, &delivered); in cppc_get_perf_ctrs() 1220 if (!delivered || !reference || !ref_perf) { in cppc_get_perf_ctrs() 1225 perf_fb_ctrs->delivered = delivered; in cppc_get_perf_ctrs() 1323 * So this CPU can be certain that its request will be delivered in cppc_set_perf() 1324 * So in all cases, this CPU knows that its request will be delivered in cppc_set_perf() 1331 * could have delivered the request to Platform by triggering the in cppc_set_perf()
|
/kernel/linux/linux-5.10/tools/testing/selftests/bpf/ |
H A D | bpf_tcp_helpers.h | 60 __u32 delivered; member 131 __u32 prior_delivered; /* tp->delivered at "prior_mstamp" */ 132 __s32 delivered; /* number of packets delivered over interval */ member 133 long interval_us; /* time for tp->delivered to incr "delivered" */ 134 __u32 snd_interval_us; /* snd interval for delivered packets */ 135 __u32 rcv_interval_us; /* rcv interval for delivered packets */ 175 /* call when packets are delivered to update cwnd and pacing rate,
|
/kernel/linux/linux-6.6/tools/testing/selftests/bpf/ |
H A D | bpf_tcp_helpers.h | 70 __u32 delivered; member 142 __u32 prior_delivered; /* tp->delivered at "prior_mstamp" */ 143 __s32 delivered; /* number of packets delivered over interval */ member 144 long interval_us; /* time for tp->delivered to incr "delivered" */ 145 __u32 snd_interval_us; /* snd interval for delivered packets */ 146 __u32 rcv_interval_us; /* rcv interval for delivered packets */ 186 /* call when packets are delivered to update cwnd and pacing rate,
|
/kernel/linux/linux-5.10/include/acpi/ |
H A D | cppc_acpi.h | 120 u64 delivered; member
|
/kernel/linux/linux-6.6/drivers/acpi/ |
H A D | cppc_acpi.c | 178 fb_ctrs.reference, fb_ctrs.delivered); in show_feedback_ctrs() 1316 u64 delivered, reference, ref_perf, ctr_wrap_time; in cppc_get_perf_ctrs() local 1353 cpc_read(cpunum, delivered_reg, &delivered); in cppc_get_perf_ctrs() 1366 if (!delivered || !reference || !ref_perf) { in cppc_get_perf_ctrs() 1371 perf_fb_ctrs->delivered = delivered; in cppc_get_perf_ctrs() 1672 * So this CPU can be certain that its request will be delivered in cppc_set_perf() 1673 * So in all cases, this CPU knows that its request will be delivered in cppc_set_perf() 1680 * could have delivered the request to Platform by triggering the in cppc_set_perf()
|
/kernel/linux/linux-6.6/include/acpi/ |
H A D | cppc_acpi.h | 124 u64 delivered; member
|