/linux-5.19.10/drivers/net/ethernet/brocade/bna/ |
D | bfa_msgq.c | 311 static void bfa_msgq_rspq_dbell(struct bfa_msgq_rspq *rspq); 322 bfa_fsm_state_decl(rspq, stopped, struct bfa_msgq_rspq, enum rspq_event); 323 bfa_fsm_state_decl(rspq, init_wait, struct bfa_msgq_rspq, 325 bfa_fsm_state_decl(rspq, ready, struct bfa_msgq_rspq, enum rspq_event); 326 bfa_fsm_state_decl(rspq, dbell_wait, struct bfa_msgq_rspq, 330 rspq_sm_stopped_entry(struct bfa_msgq_rspq *rspq) in rspq_sm_stopped_entry() argument 332 rspq->producer_index = 0; in rspq_sm_stopped_entry() 333 rspq->consumer_index = 0; in rspq_sm_stopped_entry() 334 rspq->flags = 0; in rspq_sm_stopped_entry() 338 rspq_sm_stopped(struct bfa_msgq_rspq *rspq, enum rspq_event event) in rspq_sm_stopped() argument [all …]
|
D | bfa_msgq.h | 105 struct bfa_msgq_rspq rspq; member
|
D | bfi.h | 421 struct bfi_msgq rspq; member
|
/linux-5.19.10/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | sge.c | 1567 struct adapter *adapter = rxq->rspq.adapter; in do_gro() 1573 skb = napi_get_frags(&rxq->rspq.napi); in do_gro() 1585 skb_record_rx_queue(skb, rxq->rspq.idx); in do_gro() 1593 ret = napi_gro_frags(&rxq->rspq.napi); in do_gro() 1611 int t4vf_ethrx_handler(struct sge_rspq *rspq, const __be64 *rsp, in t4vf_ethrx_handler() argument 1617 (rspq->netdev->features & NETIF_F_RXCSUM); in t4vf_ethrx_handler() 1618 struct sge_eth_rxq *rxq = container_of(rspq, struct sge_eth_rxq, rspq); in t4vf_ethrx_handler() 1619 struct adapter *adapter = rspq->adapter; in t4vf_ethrx_handler() 1628 (rspq->netdev->features & NETIF_F_GRO) && csum_ok && in t4vf_ethrx_handler() 1644 skb->protocol = eth_type_trans(skb, rspq->netdev); in t4vf_ethrx_handler() [all …]
|
D | cxgb4vf_main.c | 401 &s->ethrxq[rxq].rspq); in request_msix_queue_irqs() 410 free_irq(adapter->msix_info[--msi].vec, &s->ethrxq[rxq].rspq); in request_msix_queue_irqs() 427 &s->ethrxq[rxq].rspq); in free_msix_queue_irqs() 433 static void qenable(struct sge_rspq *rspq) in qenable() argument 435 napi_enable(&rspq->napi); in qenable() 441 t4_write_reg(rspq->adapter, T4VF_SGE_BASE_ADDR + SGE_VF_GTS, in qenable() 443 SEINTARM_V(rspq->intr_params) | in qenable() 444 INGRESSQID_V(rspq->cntxt_id)); in qenable() 456 qenable(&s->ethrxq[rxq].rspq); in enable_rx() 480 napi_disable(&s->ethrxq[rxq].rspq.napi); in quiesce_rx() [all …]
|
D | adapter.h | 213 struct sge_rspq rspq; /* Response Queue */ member
|
D | t4vf_hw.c | 1248 int start, int n, const u16 *rspq, int nrspq) in t4vf_config_rss_range() argument 1250 const u16 *rsp = rspq; in t4vf_config_rss_range() 1251 const u16 *rsp_end = rspq+nrspq; in t4vf_config_rss_range() 1310 rsp = rspq; in t4vf_config_rss_range()
|
/linux-5.19.10/drivers/scsi/bfa/ |
D | bfa_hw_cb.c | 46 bfa_hwcb_rspq_ack_msix(struct bfa_s *bfa, int rspq, u32 ci) in bfa_hwcb_rspq_ack_msix() argument 48 writel(__HFN_INT_RME_Q0 << RME_Q_NUM(bfa_ioc_pcifn(&bfa->ioc), rspq), in bfa_hwcb_rspq_ack_msix() 51 if (bfa_rspq_ci(bfa, rspq) == ci) in bfa_hwcb_rspq_ack_msix() 54 bfa_rspq_ci(bfa, rspq) = ci; in bfa_hwcb_rspq_ack_msix() 55 writel(ci, bfa->iocfc.bfa_regs.rme_q_ci[rspq]); in bfa_hwcb_rspq_ack_msix() 59 bfa_hwcb_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci) in bfa_hwcb_rspq_ack() argument 61 if (bfa_rspq_ci(bfa, rspq) == ci) in bfa_hwcb_rspq_ack() 64 bfa_rspq_ci(bfa, rspq) = ci; in bfa_hwcb_rspq_ack() 65 writel(ci, bfa->iocfc.bfa_regs.rme_q_ci[rspq]); in bfa_hwcb_rspq_ack()
|
D | bfa_hw_ct.c | 67 bfa_hwct_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci) in bfa_hwct_rspq_ack() argument 71 r32 = readl(bfa->iocfc.bfa_regs.rme_q_ctrl[rspq]); in bfa_hwct_rspq_ack() 72 writel(r32, bfa->iocfc.bfa_regs.rme_q_ctrl[rspq]); in bfa_hwct_rspq_ack() 74 bfa_rspq_ci(bfa, rspq) = ci; in bfa_hwct_rspq_ack() 75 writel(ci, bfa->iocfc.bfa_regs.rme_q_ci[rspq]); in bfa_hwct_rspq_ack() 84 bfa_hwct2_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci) in bfa_hwct2_rspq_ack() argument 86 bfa_rspq_ci(bfa, rspq) = ci; in bfa_hwct2_rspq_ack() 87 writel(ci, bfa->iocfc.bfa_regs.rme_q_ci[rspq]); in bfa_hwct2_rspq_ack()
|
D | bfa.h | 179 void (*hw_rspq_ack)(struct bfa_s *bfa, int rspq, u32 ci); 313 void bfa_hwcb_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci); 325 void bfa_hwct_reqq_ack(struct bfa_s *bfa, int rspq); 326 void bfa_hwct_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci); 327 void bfa_hwct2_rspq_ack(struct bfa_s *bfa, int rspq, u32 ci);
|
D | bfi.h | 593 struct bfi_msgq_s rspq; member
|
/linux-5.19.10/drivers/net/ethernet/chelsio/cxgb4/ |
D | cxgb4_debugfs.c | 2693 rx[i].rspq.netdev ? rx[i].rspq.netdev->name : "N/A"); in sge_qinfo_show() 2708 R("RspQ ID:", rspq.abs_id); in sge_qinfo_show() 2709 R("RspQ size:", rspq.size); in sge_qinfo_show() 2710 R("RspQE size:", rspq.iqe_len); in sge_qinfo_show() 2711 R("RspQ CIDX:", rspq.cidx); in sge_qinfo_show() 2712 R("RspQ Gen:", rspq.gen); in sge_qinfo_show() 2713 S3("u", "Intr delay:", qtimer_val(adap, &rx[i].rspq)); in sge_qinfo_show() 2714 S3("u", "Intr pktcnt:", s->counter_val[rx[i].rspq.pktcnt_idx]); in sge_qinfo_show() 2766 rx[i].rspq.netdev ? rx[i].rspq.netdev->name : "N/A"); in sge_qinfo_show() 2767 R("RspQ ID:", rspq.abs_id); in sge_qinfo_show() [all …]
|
D | cxgb4_uld.c | 78 struct sge_ofld_rxq *rxq = container_of(q, struct sge_ofld_rxq, rspq); in uldrx_handler() 146 err = t4_sge_alloc_rxq(adap, &q->rspq, false, in alloc_uld_rxqs() 158 ids[i] = q->rspq.abs_id; in alloc_uld_rxqs() 164 if (q->rspq.desc) in alloc_uld_rxqs() 165 free_rspq_fl(adap, &q->rspq, in alloc_uld_rxqs() 191 cmplqid = rxq_info->uldrxq[i].rspq.cntxt_id; in setup_sge_queues_uld() 206 if (q->rspq.desc) in t4_free_uld_rxqs() 207 free_rspq_fl(adap, &q->rspq, in t4_free_uld_rxqs() 289 init_rspq(adap, &r->rspq, 5, 1, uld_info->rxq_size, 64); in cfg_queues_uld() 290 r->rspq.uld = uld_type; in cfg_queues_uld() [all …]
|
D | cxgb4_tc_mqprio.c | 200 init_rspq(adap, &eorxq->rspq, in cxgb4_mqprio_alloc_hw_resources() 208 ret = t4_sge_alloc_rxq(adap, &eorxq->rspq, false, in cxgb4_mqprio_alloc_hw_resources() 218 eorxq->rspq.cntxt_id); in cxgb4_mqprio_alloc_hw_resources() 225 eorxq->msix->desc, &eorxq->rspq); in cxgb4_mqprio_alloc_hw_resources() 234 cxgb4_enable_rx(adap, &eorxq->rspq); in cxgb4_mqprio_alloc_hw_resources() 244 cxgb4_quiesce_rx(&eorxq->rspq); in cxgb4_mqprio_alloc_hw_resources() 249 free_irq(eorxq->msix->vec, &eorxq->rspq); in cxgb4_mqprio_alloc_hw_resources() 258 if (eorxq->rspq.desc) in cxgb4_mqprio_alloc_hw_resources() 259 free_rspq_fl(adap, &eorxq->rspq, &eorxq->fl); in cxgb4_mqprio_alloc_hw_resources() 285 if (!adap->sge.eohw_rxq[pi->first_qset].rspq.desc) in cxgb4_mqprio_free_hw_resources() [all …]
|
D | sge.c | 2600 flowc->mnemval[3].val = cpu_to_be32(rxq->rspq.abs_id); in cxgb4_ethofld_send_flowc() 3446 struct adapter *adapter = rxq->rspq.adap; in do_gro() 3452 skb = napi_get_frags(&rxq->rspq.napi); in do_gro() 3466 skb_record_rx_queue(skb, rxq->rspq.idx); in do_gro() 3471 if (rxq->rspq.netdev->features & NETIF_F_RXHASH) in do_gro() 3479 ret = napi_gro_frags(&rxq->rspq.napi); in do_gro() 3590 static void t4_tx_completion_handler(struct sge_rspq *rspq, in t4_tx_completion_handler() argument 3595 struct port_info *pi = netdev_priv(rspq->netdev); in t4_tx_completion_handler() 3596 struct adapter *adapter = rspq->adap; in t4_tx_completion_handler() 3619 txq = &s->ethtxq[pi->first_qset + rspq->idx]; in t4_tx_completion_handler() [all …]
|
D | cxgb4_main.c | 739 &s->ethrxq[ethqidx].rspq); in request_msix_queue_irqs() 752 free_irq(minfo->vec, &s->ethrxq[ethqidx].rspq); in request_msix_queue_irqs() 768 free_irq(minfo->vec, &s->ethrxq[i].rspq); in free_msix_queue_irqs() 874 rss[i] = rxq[*queues].rspq.abs_id; in cxgb4_write_rss() 1079 err = t4_sge_alloc_rxq(adap, &q->rspq, false, dev, in setup_sge_queues() 1087 q->rspq.idx = j; in setup_sge_queues() 1095 q->rspq.cntxt_id, in setup_sge_queues() 1107 cmplqid = rxq_info->uldrxq[i].rspq.cntxt_id; in setup_sge_queues() 1127 QUEUENUMBER_V(s->ethrxq[0].rspq.abs_id)); in setup_sge_queues() 1306 cxgb4_quiesce_rx(&mirror_rxq->rspq); in cxgb4_port_mirror_free_rxq() [all …]
|
D | cxgb4_ethtool.c | 906 e->rx_mini_pending = s->ethrxq[pi->first_qset].rspq.size; in get_sge_param() 933 s->ethrxq[pi->first_qset + i].rspq.size = e->rx_mini_pending; in set_sge_param() 955 err = cxgb4_set_rspq_intr_params(&q->rspq, us, cnt); in set_rx_intr_params() 970 q->rspq.adaptive_rx = adaptive_rx; in set_adaptive_rx_setting() 981 return q->rspq.adaptive_rx; in get_adaptive_rx_setting() 1178 const struct sge_rspq *rq = &adap->sge.ethrxq[pi->first_qset].rspq; in get_coalesce()
|
D | cxgb4_ptp.c | 187 c.u.init.absid = cpu_to_be16(receive_q->rspq.abs_id); in cxgb4_ptp_redirect_rx_packet()
|
D | cxgb4.h | 812 struct sge_rspq rspq; member 826 struct sge_rspq rspq; member 1898 int start, int n, const u16 *rspq, unsigned int nrspq);
|
/linux-5.19.10/drivers/net/ethernet/chelsio/cxgb3/ |
D | sge.c | 176 return container_of(q, struct sge_qset, rspq); in rspq_to_qset() 657 memset(&q->rspq, 0, sizeof(q->rspq)); in t3_reset_qset() 712 if (q->rspq.desc) { in t3_free_qset() 714 t3_sge_disable_rspcntxt(adapter, q->rspq.cntxt_id); in t3_free_qset() 717 q->rspq.size * sizeof(struct rsp_desc), in t3_free_qset() 718 q->rspq.desc, q->rspq.phys_addr); in t3_free_qset() 733 qs->rspq.cntxt_id = id; in init_qset_cntxt() 1904 struct sge_rspq *q = &qs->rspq; in ofld_poll() 2331 struct sge_rspq *q = &qs->rspq; in process_responses() 2503 t3_write_reg(adap, A_SG_GTS, V_RSPQ(qs->rspq.cntxt_id) | in napi_rx_handler() [all …]
|
D | adapter.h | 202 struct sge_rspq rspq; member
|
D | common.h | 697 const u8 * cpus, const u16 *rspq); 746 unsigned int size, int rspq, int ovfl_mode,
|
D | t3_hw.c | 2299 unsigned int size, int rspq, int ovfl_mode, in t3_sge_init_cqcntxt() argument 2312 V_CQ_BASE_HI((u32) base_addr) | V_CQ_RSPQ(rspq) | in t3_sge_init_cqcntxt() 2471 const u8 * cpus, const u16 *rspq) in t3_config_rss() argument 2487 if (rspq) in t3_config_rss() 2490 (i << 16) | rspq[q_idx++]); in t3_config_rss() 2491 if (rspq[q_idx] == 0xffff) in t3_config_rss()
|
D | cxgb3_main.c | 411 rspq.polling), 0, in request_msix_data_irqs() 447 while (adap->sge.qs[0].rspq.offload_pkts < init_cnt + n) { in await_mgmt_replies() 460 unsigned long cnt = adap->sge.qs[0].rspq.offload_pkts; in init_tp_parity() 1172 struct sge_rspq *q = &adap->sge.qs[i].rspq; in t3_synchronize_rx() 1278 t3_intr_handler(adap, adap->sge.qs[0].rspq.polling), in cxgb_up() 2629 t3_intr_handler(adapter, qs->rspq.polling) (0, source); in cxgb_netpoll()
|
D | t3_cpl.h | 1441 __u8 rspq:3; member 1445 __u8 rspq:3; member
|