/linux-6.1.9/include/net/ |
D | red.h | 146 int qcount; /* Number of packets since last random member 167 v->qcount = -1; in red_set_vars() 286 v->qcount = -1; in red_restart() 389 return !(((qavg - p->qth_min) >> p->Wlog) * v->qcount < v->qR); in red_mark_probability() 420 v->qcount = -1; in red_action() 424 if (++v->qcount) { in red_action() 426 v->qcount = 0; in red_action() 436 v->qcount = -1; in red_action()
|
/linux-6.1.9/drivers/net/ethernet/chelsio/cxgb4/ |
D | cxgb4_tc_mqprio.c | 14 u32 speed, qcount = 0, qoffset = 0; in cxgb4_mqprio_validate() local 49 qcount += mqprio->qopt.count[i]; in cxgb4_mqprio_validate() 81 if (qoffset >= adap->tids.neotids || qcount > adap->tids.neotids) in cxgb4_mqprio_validate() 431 u32 qoffset, qcount, tot_qcount, qid, hwqid; in cxgb4_mqprio_enable_offload() local 446 qcount = mqprio->qopt.count[i]; in cxgb4_mqprio_enable_offload() 447 for (j = 0; j < qcount; j++) { in cxgb4_mqprio_enable_offload() 486 qcount = mqprio->qopt.count[i]; in cxgb4_mqprio_enable_offload() 487 if (qcount) { in cxgb4_mqprio_enable_offload() 490 qcount = pi->nqsets; in cxgb4_mqprio_enable_offload() 494 ret = netdev_set_tc_queue(dev, i, qcount, qoffset); in cxgb4_mqprio_enable_offload() [all …]
|
/linux-6.1.9/drivers/net/ethernet/marvell/octeontx2/nic/ |
D | otx2_vf.c | 530 int err, qcount; in otx2vf_probe() local 552 qcount = num_online_cpus(); in otx2vf_probe() 553 netdev = alloc_etherdev_mqs(sizeof(*vf), qcount, qcount); in otx2vf_probe() 570 hw->rx_queues = qcount; in otx2vf_probe() 571 hw->tx_queues = qcount; in otx2vf_probe() 572 hw->max_queues = qcount; in otx2vf_probe() 573 hw->tot_tx_queues = qcount; in otx2vf_probe() 626 err = otx2_set_real_num_queues(netdev, qcount, qcount); in otx2vf_probe()
|
D | otx2_pf.c | 2706 int err, qcount; in otx2_probe() local 2730 qcount = min_t(int, num_online_cpus(), OTX2_MAX_CQ_CNT); in otx2_probe() 2732 netdev = alloc_etherdev_mqs(sizeof(*pf), qcount, qcount); in otx2_probe() 2749 hw->rx_queues = qcount; in otx2_probe() 2750 hw->tx_queues = qcount; in otx2_probe() 2751 hw->tot_tx_queues = qcount; in otx2_probe() 2752 hw->max_queues = qcount; in otx2_probe()
|
/linux-6.1.9/drivers/net/ethernet/intel/ice/ |
D | ice_dcb_lib.c | 218 u16 qoffset, qcount; in ice_vsi_cfg_dcb_rings() local 239 qcount = vsi->tc_cfg.tc_info[n].qcount_tx; in ice_vsi_cfg_dcb_rings() 240 for (i = qoffset; i < (qoffset + qcount); i++) in ice_vsi_cfg_dcb_rings() 243 qcount = vsi->tc_cfg.tc_info[n].qcount_rx; in ice_vsi_cfg_dcb_rings() 244 for (i = qoffset; i < (qoffset + qcount); i++) in ice_vsi_cfg_dcb_rings() 261 qcount = vsi->mqprio_qopt.qopt.count[n]; in ice_vsi_cfg_dcb_rings() 262 for (i = qoffset; i < (qoffset + qcount); i++) { in ice_vsi_cfg_dcb_rings()
|
D | ice_main.c | 7836 int qcount = mqprio_qopt->qopt.count[i]; in ice_validate_mqprio_qopt() local 7839 if (!qcount) in ice_validate_mqprio_qopt() 7842 if (is_power_of_2(qcount)) { in ice_validate_mqprio_qopt() 7844 qcount > non_power_of_2_qcount) { in ice_validate_mqprio_qopt() 7846 qcount, non_power_of_2_qcount); in ice_validate_mqprio_qopt() 7849 if (qcount > max_rss_q_cnt) in ice_validate_mqprio_qopt() 7850 max_rss_q_cnt = qcount; in ice_validate_mqprio_qopt() 7853 qcount != non_power_of_2_qcount) { in ice_validate_mqprio_qopt() 7855 qcount, non_power_of_2_qcount); in ice_validate_mqprio_qopt() 7858 if (qcount < max_rss_q_cnt) { in ice_validate_mqprio_qopt() [all …]
|
D | ice_lib.c | 1111 u16 qcount, qmap; in ice_chnl_vsi_setup_q_map() local 1115 qcount = min_t(int, vsi->num_rxq, pf->num_lan_msix); in ice_chnl_vsi_setup_q_map() 1117 pow = order_base_2(qcount); in ice_chnl_vsi_setup_q_map() 1126 ctxt->info.q_mapping[1] = cpu_to_le16(qcount); in ice_chnl_vsi_setup_q_map()
|
D | ice_ethtool.c | 3166 u16 qcount, offset; in ice_get_rxfh_context() local 3180 qcount = vsi->mqprio_qopt.qopt.count[rss_context]; in ice_get_rxfh_context() 3214 indir[i] = offset + lut[i] % qcount; in ice_get_rxfh_context()
|
/linux-6.1.9/net/sched/ |
D | sch_choke.c | 226 q->vars.qcount = -1; in choke_enqueue() 239 q->vars.qcount = -1; in choke_enqueue() 249 } else if (++q->vars.qcount) { in choke_enqueue() 251 q->vars.qcount = 0; in choke_enqueue()
|
/linux-6.1.9/fs/gfs2/ |
D | trace_gfs2.h | 275 __field( u64, qcount ) 292 __entry->qcount = gl->gl_stats.stats[GFS2_LKS_QCOUNT]; 307 (long long)__entry->qcount)
|
/linux-6.1.9/drivers/net/ethernet/intel/i40e/ |
D | i40e_main.c | 1928 u16 qcount = 0, max_qcount, qmap, sections = 0; in i40e_vsi_setup_queue_map_mqprio() local 1953 qcount = vsi->mqprio_qopt.qopt.count[i]; in i40e_vsi_setup_queue_map_mqprio() 1954 if (qcount > max_qcount) in i40e_vsi_setup_queue_map_mqprio() 1955 max_qcount = qcount; in i40e_vsi_setup_queue_map_mqprio() 1957 vsi->tc_config.tc_info[i].qcount = qcount; in i40e_vsi_setup_queue_map_mqprio() 1965 vsi->tc_config.tc_info[i].qcount = 1; in i40e_vsi_setup_queue_map_mqprio() 1971 vsi->num_queue_pairs = offset + qcount; in i40e_vsi_setup_queue_map_mqprio() 2022 u16 qcount; in i40e_vsi_setup_queue_map() local 2091 qcount = min_t(int, pf->alloc_rss_size, in i40e_vsi_setup_queue_map() 2100 qcount = num_tc_qps; in i40e_vsi_setup_queue_map() [all …]
|
D | i40e_txrx.c | 3682 u16 qcount; in i40e_lan_select_queue() local 3700 qcount = vsi->tc_config.tc_info[tclass].qcount; in i40e_lan_select_queue() 3701 hash = i40e_swdcb_skb_tx_hash(netdev, skb, qcount); in i40e_lan_select_queue()
|
D | i40e.h | 307 u16 qcount; /* Total Queues */ member
|
D | i40e_debugfs.c | 474 vsi->tc_config.tc_info[i].qcount, in i40e_dbg_dump_vsi_seid()
|
/linux-6.1.9/drivers/net/ethernet/cavium/thunder/ |
D | nicvf_queues.c | 384 int tail, qcount; in nicvf_refill_rbdr() local 401 qcount = nicvf_queue_reg_read(nic, NIC_QSET_RBDR_0_1_STATUS0, rbdr_idx); in nicvf_refill_rbdr() 402 qcount &= 0x7FFFF; in nicvf_refill_rbdr() 404 if (qcount >= (qs->rbdr_len - 1)) in nicvf_refill_rbdr() 407 refill_rb_cnt = qs->rbdr_len - qcount - 1; in nicvf_refill_rbdr()
|
D | nicvf_main.c | 2099 int err, qcount; in nicvf_probe() local 2128 qcount = netif_get_num_default_rss_queues(); in nicvf_probe() 2133 qcount = min_t(int, num_online_cpus(), in nicvf_probe() 2137 netdev = alloc_etherdev_mqs(sizeof(struct nicvf), qcount, qcount); in nicvf_probe() 2151 nic->max_queues = qcount; in nicvf_probe()
|
/linux-6.1.9/drivers/infiniband/hw/bnxt_re/ |
D | qplib_res.h | 228 u8 qcount[MAX_TQM_ALLOC_REQ]; member
|
D | qplib_res.c | 417 if (!tqmctx->qcount[i]) in bnxt_qplib_alloc_tqm_rings() 419 hwq_attr.depth = ctx->qpc_count * tqmctx->qcount[i]; in bnxt_qplib_alloc_tqm_rings()
|
D | main.c | 175 rdev->qplib_ctx.tqm_ctx.qcount[i] = in bnxt_re_limit_pf_res()
|
/linux-6.1.9/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_sriov.c | 2449 int vf_idx, sb_idx, vfq_idx, qcount, first_vf; in bnx2x_enable_sriov() local 2493 qcount = 0; in bnx2x_enable_sriov() 2498 vf->vfqs = &bp->vfdb->vfqs[qcount]; in bnx2x_enable_sriov() 2499 qcount += vf_sb_count(vf); in bnx2x_enable_sriov()
|
/linux-6.1.9/net/core/ |
D | dev.c | 3196 u16 qcount = dev->real_num_tx_queues; in skb_tx_hash() local 3202 qcount = sb_dev->tc_to_txq[tc].count; in skb_tx_hash() 3203 if (unlikely(!qcount)) { in skb_tx_hash() 3207 qcount = dev->real_num_tx_queues; in skb_tx_hash() 3215 while (unlikely(hash >= qcount)) in skb_tx_hash() 3216 hash -= qcount; in skb_tx_hash() 3220 return (u16) reciprocal_scale(skb_get_hash(skb), qcount) + qoffset; in skb_tx_hash()
|
/linux-6.1.9/drivers/net/wireless/broadcom/brcm80211/brcmfmac/ |
D | sdio.c | 538 static int qcount[NUMPRIO]; variable 2846 if (pktq_plen(&bus->txq, prec) > qcount[prec]) in brcmf_sdio_bus_txdata() 2847 qcount[prec] = pktq_plen(&bus->txq, prec); in brcmf_sdio_bus_txdata()
|
/linux-6.1.9/drivers/net/ethernet/intel/iavf/ |
D | iavf_main.c | 3679 u16 qcount = mqprio_qopt->qopt.count[i]; in __iavf_setup_tc() local 3683 netdev_set_tc_queue(netdev, netdev_tc++, qcount, in __iavf_setup_tc()
|