Home
last modified time | relevance | path

Searched refs:cpu_bstats (Results 1 – 18 of 18) sorted by relevance

/linux-6.1.9/net/core/
Dgen_estimator.c46 struct gnet_stats_basic_sync __percpu *cpu_bstats; member
69 gnet_stats_add_basic(b, e->cpu_bstats, e->bstats, e->running); in est_fetch_counters()
131 struct gnet_stats_basic_sync __percpu *cpu_bstats, in gen_new_estimator() argument
166 est->cpu_bstats = cpu_bstats; in gen_new_estimator()
234 struct gnet_stats_basic_sync __percpu *cpu_bstats, in gen_replace_estimator() argument
239 return gen_new_estimator(bstats, cpu_bstats, rate_est, in gen_replace_estimator()
/linux-6.1.9/include/net/
Dgen_stats.h71 struct gnet_stats_basic_sync __percpu *cpu_bstats,
77 struct gnet_stats_basic_sync __percpu *cpu_bstats,
Dact_api.h39 struct gnet_stats_basic_sync __percpu *cpu_bstats; member
224 if (likely(a->cpu_bstats)) { in tcf_action_update_bstats()
225 bstats_update(this_cpu_ptr(a->cpu_bstats), skb); in tcf_action_update_bstats()
Dsch_generic.h107 struct gnet_stats_basic_sync __percpu *cpu_bstats; member
829 bstats_update(this_cpu_ptr(sch->cpu_bstats), skb); in qdisc_bstats_cpu_update()
1258 struct gnet_stats_basic_sync __percpu *cpu_bstats; member
1266 bstats_update(this_cpu_ptr(miniq->cpu_bstats), skb); in mini_qdisc_bstats_cpu_update()
/linux-6.1.9/net/sched/
Dsch_generic.c948 sch->cpu_bstats = in qdisc_alloc()
950 if (!sch->cpu_bstats) in qdisc_alloc()
955 free_percpu(sch->cpu_bstats); in qdisc_alloc()
1035 free_percpu(qdisc->cpu_bstats); in qdisc_free()
1588 miniqp->miniq1.cpu_bstats = qdisc->cpu_bstats; in mini_qdisc_pair_init()
1590 miniqp->miniq2.cpu_bstats = qdisc->cpu_bstats; in mini_qdisc_pair_init()
Dsch_mq.c147 gnet_stats_add_basic(&sch->bstats, qdisc->cpu_bstats, in mq_dump()
234 if (gnet_stats_copy_basic(d, sch->cpu_bstats, &sch->bstats, true) < 0 || in mq_dump_class_stats()
Dsch_mqprio.c405 gnet_stats_add_basic(&sch->bstats, qdisc->cpu_bstats, in mqprio_dump()
522 gnet_stats_add_basic(&bstats, qdisc->cpu_bstats, in mqprio_dump_class_stats()
542 if (gnet_stats_copy_basic(d, sch->cpu_bstats, in mqprio_dump_class_stats()
Dact_api.c123 free_percpu(p->cpu_bstats); in free_tcf()
752 p->cpu_bstats = netdev_alloc_pcpu_stats(struct gnet_stats_basic_sync); in tcf_idr_create()
753 if (!p->cpu_bstats) in tcf_idr_create()
771 err = gen_new_estimator(&p->tcfa_bstats, p->cpu_bstats, in tcf_idr_create()
788 free_percpu(p->cpu_bstats); in tcf_idr_create()
1514 if (a->cpu_bstats) { in tcf_action_update_stats()
1515 _bstats_update(this_cpu_ptr(a->cpu_bstats), bytes, packets); in tcf_action_update_stats()
1563 if (gnet_stats_copy_basic(&d, p->cpu_bstats, in tcf_action_copy_stats()
Dsch_api.c906 struct gnet_stats_basic_sync __percpu *cpu_bstats = NULL; in tc_fill_qdisc() local
960 cpu_bstats = q->cpu_bstats; in tc_fill_qdisc()
964 if (gnet_stats_copy_basic(&d, cpu_bstats, &q->bstats, true) < 0 || in tc_fill_qdisc()
1039 free_percpu(sch->cpu_bstats); in qdisc_clear_nolock()
1041 sch->cpu_bstats = NULL; in qdisc_clear_nolock()
1298 sch->cpu_bstats, in qdisc_create()
1374 sch->cpu_bstats, in qdisc_change()
Dact_police.c114 police->common.cpu_bstats, in tcf_police_init()
254 bstats_update(this_cpu_ptr(police->common.cpu_bstats), skb); in tcf_police_act()
Dact_skbmod.c33 bstats_update(this_cpu_ptr(d->common.cpu_bstats), skb); in tcf_skbmod_act()
Dact_sample.c165 bstats_update(this_cpu_ptr(s->common.cpu_bstats), skb); in tcf_sample_act()
Dact_ife.c720 bstats_update(this_cpu_ptr(ife->common.cpu_bstats), skb); in tcf_ife_decode()
808 bstats_update(this_cpu_ptr(ife->common.cpu_bstats), skb); in tcf_ife_encode()
Dsch_multiq.c340 if (gnet_stats_copy_basic(d, cl_q->cpu_bstats, &cl_q->bstats, true) < 0 || in multiq_dump_class_stats()
Dact_bpf.c43 bstats_update(this_cpu_ptr(prog->common.cpu_bstats), skb); in tcf_bpf_act()
Dsch_prio.c362 if (gnet_stats_copy_basic(d, cl_q->cpu_bstats, in prio_dump_class_stats()
Dact_mpls.c61 bstats_update(this_cpu_ptr(m->common.cpu_bstats), skb); in tcf_mpls_act()
Dact_skbedit.c47 bstats_update(this_cpu_ptr(d->common.cpu_bstats), skb); in tcf_skbedit_act()