/linux-6.1.9/net/ipv4/ |
D | tcp_rate.c | 67 TCP_SKB_CB(skb)->tx.delivered = tp->delivered; in tcp_rate_skb_sent() 95 rs->prior_delivered = scb->tx.delivered; in tcp_rate_skb_delivered() 117 void tcp_rate_gen(struct sock *sk, u32 delivered, u32 lost, in tcp_rate_gen() argument 124 if (tp->app_limited && after(tp->delivered, tp->app_limited)) in tcp_rate_gen() 131 if (delivered) in tcp_rate_gen() 134 rs->acked_sacked = delivered; /* freshly ACKed or SACKed */ in tcp_rate_gen() 142 rs->delivered = -1; in tcp_rate_gen() 146 rs->delivered = tp->delivered - rs->prior_delivered; in tcp_rate_gen() 176 rs->interval_us, rs->delivered, in tcp_rate_gen() 185 ((u64)rs->delivered * tp->rate_interval_us >= in tcp_rate_gen() [all …]
|
D | tcp_bbr.c | 497 bbr->next_rtt_delivered = tp->delivered; /* start round now */ in bbr_set_cwnd_to_recover_or_restore() 542 else if (cwnd < target_cwnd || tp->delivered < TCP_INIT_CWND) in bbr_set_cwnd() 640 bbr->lt_last_delivered = tp->delivered; in bbr_reset_lt_bw_sampling_interval() 691 u32 lost, delivered; in bbr_lt_bw_sampling() local 739 delivered = tp->delivered - bbr->lt_last_delivered; in bbr_lt_bw_sampling() 741 if (!delivered || (lost << BBR_SCALE) < bbr_lt_loss_thresh * delivered) in bbr_lt_bw_sampling() 754 bw = (u64)delivered * BW_UNIT; in bbr_lt_bw_sampling() 767 if (rs->delivered < 0 || rs->interval_us <= 0) in bbr_update_bw() 772 bbr->next_rtt_delivered = tp->delivered; in bbr_update_bw() 784 bw = div64_long((u64)rs->delivered * BW_UNIT, rs->interval_us); in bbr_update_bw() [all …]
|
D | tcp_dctcp.c | 73 ca->old_delivered = tp->delivered; in dctcp_reset() 127 u32 delivered = tp->delivered - ca->old_delivered; in dctcp_update_alpha() local 133 delivered_ce /= max(1U, delivered); in dctcp_update_alpha() 202 (tp->delivered - ca->old_delivered); in dctcp_get_info()
|
D | tcp_input.c | 1111 static void tcp_count_delivered(struct tcp_sock *tp, u32 delivered, in tcp_count_delivered() argument 1114 tp->delivered += delivered; in tcp_count_delivered() 1116 tp->delivered_ce += delivered; in tcp_count_delivered() 2045 s32 delivered; in tcp_add_reno_sack() local 2049 delivered = tp->sacked_out - prior_sacked; in tcp_add_reno_sack() 2050 if (delivered > 0) in tcp_add_reno_sack() 2051 tcp_count_delivered(tp, delivered, ece_ack); in tcp_add_reno_sack() 3355 sack->rate->prior_delivered + 1 == tp->delivered && in tcp_clean_rtx_queue() 3414 (tp->delivered - sack->rate->prior_delivered); in tcp_clean_rtx_queue() 3748 u32 delivered; in tcp_newly_delivered() local [all …]
|
/linux-6.1.9/tools/testing/selftests/bpf/prog_tests/ |
D | tcp_rtt.c | 10 __u32 delivered; member 47 __u32 dsack_dups, __u32 delivered, __u32 delivered_ce, in verify_sk() argument 68 if (val.delivered != delivered) { in verify_sk() 70 msg, val.delivered, delivered); in verify_sk()
|
/linux-6.1.9/tools/testing/selftests/bpf/progs/ |
D | tcp_rtt.c | 10 __u32 delivered; member 54 storage->delivered = tcp_sk->delivered; in _sockops()
|
D | bpf_dctcp.c | 56 ca->old_delivered = tp->delivered; in dctcp_reset() 132 __u32 delivered = tp->delivered - ca->old_delivered; in BPF_PROG() local 138 delivered_ce /= max(1U, delivered); in BPF_PROG()
|
/linux-6.1.9/drivers/i2c/busses/ |
D | i2c-imx-lpi2c.c | 104 unsigned int delivered; member 344 remaining = lpi2c_imx->msglen - lpi2c_imx->delivered; in lpi2c_imx_set_rx_watermark() 361 if (lpi2c_imx->delivered == lpi2c_imx->msglen) in lpi2c_imx_write_txfifo() 364 data = lpi2c_imx->tx_buf[lpi2c_imx->delivered++]; in lpi2c_imx_write_txfifo() 369 if (lpi2c_imx->delivered < lpi2c_imx->msglen) in lpi2c_imx_write_txfifo() 385 lpi2c_imx->rx_buf[lpi2c_imx->delivered++] = data & 0xff; in lpi2c_imx_read_rxfifo() 397 remaining = lpi2c_imx->msglen - lpi2c_imx->delivered; in lpi2c_imx_read_rxfifo() 413 } else if (!(lpi2c_imx->delivered & 0xff)) { in lpi2c_imx_read_rxfifo() 466 lpi2c_imx->delivered = 0; in lpi2c_imx_xfer()
|
/linux-6.1.9/Documentation/admin-guide/acpi/ |
D | cppc_sysfs.rst | 15 to request performance levels and to measure per-cpu delivered performance. 53 * feedback_ctrs : Includes both Reference and delivered performance counter. 55 Delivered counter ticks up proportional to processor's delivered performance. 65 Below describes the steps to compute the average performance delivered by
|
/linux-6.1.9/Documentation/fb/ |
D | metronomefb.rst | 16 which is then delivered to the AMLCD interface by a host specific method. 23 Metronomefb requires waveform information which is delivered via the AMLCD 25 be delivered from userspace via the firmware class interface. The waveform file
|
/linux-6.1.9/Documentation/powerpc/ |
D | pmu-ebb.rst | 29 When a PMU EBB occurs it is delivered to the currently running process. As such 35 EBBs will ever be delivered. 40 delivered to the first process. It's not clear if this is actually useful. 43 When the PMU is configured for EBBs, all PMU interrupts are delivered to the 110 meaningless. Because interrupts are being delivered to the user process the 119 no further PMU EBBs will be delivered.
|
/linux-6.1.9/Documentation/virt/kvm/x86/ |
D | msr.rst | 200 /* Used for 'page not present' events delivered via #PF */ 203 /* Used for 'page ready' events delivered via interrupt notification */ 213 cpl == 0. Bit 2 is 1 if asynchronous page faults are delivered to L1 as 219 'Page not present' events are currently always delivered as synthetic 229 next event can be delivered. 237 of these bytes is a token which was previously delivered as 'page not 249 Note, previously, 'page ready' events were delivered via the same #PF 251 bit 3 (interrupt based delivery) is not set APF events are not delivered. 254 not be delivered. 256 Currently 'page ready' APF events will be always delivered on the
|
/linux-6.1.9/Documentation/virt/kvm/devices/ |
D | xics.rst | 58 delivered, and 255 is the lowest priority. 75 priority of 255 will never be delivered. 84 This bit is set to 1 if the interrupt is masked (cannot be delivered
|
/linux-6.1.9/net/bridge/ |
D | br_forward.c | 310 goto delivered; in br_multicast_flood() 315 goto delivered; in br_multicast_flood() 323 delivered: in br_multicast_flood()
|
/linux-6.1.9/Documentation/core-api/irq/ |
D | irq-affinity.rst | 41 As can be seen from the line above IRQ44 was delivered only to the first four 60 This time around IRQ44 was delivered only to the last four processors.
|
/linux-6.1.9/tools/testing/selftests/bpf/ |
D | bpf_tcp_helpers.h | 70 __u32 delivered; member 143 __s32 delivered; /* number of packets delivered over interval */ member
|
/linux-6.1.9/samples/bpf/ |
D | tcp_dumpstats_kern.c | 63 tcp_sk->dsack_dups, tcp_sk->delivered); in _sockops()
|
/linux-6.1.9/Documentation/ia64/ |
D | mca.rst | 35 * Because MCA/INIT can be delivered at any time, including when the cpu 129 * x86 NMI typically gets delivered to one cpu. MCA/INIT gets sent to 150 What happens when MCA/INIT is delivered what a cpu is running user 167 was delivered? 182 delivered? 198 delivered.
|
/linux-6.1.9/Documentation/filesystems/ |
D | dnotify.rst | 13 being delivered using signals. 32 By default, SIGIO will be delivered to the process and no other useful
|
/linux-6.1.9/Documentation/devicetree/bindings/leds/irled/ |
D | spi-ir-led.txt | 5 are delivered thourgh that.
|
/linux-6.1.9/drivers/acpi/ |
D | cppc_acpi.c | 178 fb_ctrs.reference, fb_ctrs.delivered); in show_feedback_ctrs() 1302 u64 delivered, reference, ref_perf, ctr_wrap_time; in cppc_get_perf_ctrs() local 1339 cpc_read(cpunum, delivered_reg, &delivered); in cppc_get_perf_ctrs() 1352 if (!delivered || !reference || !ref_perf) { in cppc_get_perf_ctrs() 1357 perf_fb_ctrs->delivered = delivered; in cppc_get_perf_ctrs()
|
/linux-6.1.9/Documentation/bpf/ |
D | prog_sk_lookup.rst | 9 delivered locally. 49 Incoming traffic to established (TCP) and connected (UDP) sockets is delivered
|
/linux-6.1.9/include/acpi/ |
D | cppc_acpi.h | 121 u64 delivered; member
|
/linux-6.1.9/Documentation/networking/ |
D | strparser.rst | 21 callback of a TCP socket. Messages are parsed and delivered as they are 25 outside source. Message are parsed and delivered as the sequence is 57 and no new messages are delivered to the upper layer.
|
/linux-6.1.9/Documentation/networking/device_drivers/ethernet/amazon/ |
D | ena.rst | 114 A Keep-Alive event is delivered by the device every second. 116 statistics. If the keep-alive events aren't delivered as expected the WD resets 183 modes. That is, once MSI-X is delivered to the host, its Cause bit is 251 function delivered in the Rx CQ descriptor is set in the received
|