Home
last modified time | relevance | path

Searched refs:rx_cfg (Results 1 – 25 of 31) sorted by relevance

12

/linux-5.19.10/drivers/hsi/
Dhsi_core.c63 kfree(cl->rx_cfg.channels); in hsi_client_release()
86 cl->rx_cfg = info->rx_cfg; in hsi_new_client()
87 if (cl->rx_cfg.channels) { in hsi_new_client()
88 size = cl->rx_cfg.num_channels * sizeof(*cl->rx_cfg.channels); in hsi_new_client()
89 cl->rx_cfg.channels = kmemdup(info->rx_cfg.channels, size, in hsi_new_client()
91 if (!cl->rx_cfg.channels) in hsi_new_client()
217 &cl->rx_cfg.mode); in hsi_add_client_from_dt()
226 cl->rx_cfg.mode = mode; in hsi_add_client_from_dt()
234 cl->rx_cfg.speed = cl->tx_cfg.speed; in hsi_add_client_from_dt()
237 &cl->rx_cfg.flow); in hsi_add_client_from_dt()
[all …]
/linux-5.19.10/drivers/staging/pi433/
Dpi433_if.c91 struct pi433_rx_cfg rx_cfg; member
163 rf69_set_rx_cfg(struct pi433_device *dev, struct pi433_rx_cfg *rx_cfg) in rf69_set_rx_cfg() argument
169 ret = rf69_set_frequency(dev->spi, rx_cfg->frequency); in rf69_set_rx_cfg()
172 ret = rf69_set_modulation(dev->spi, rx_cfg->modulation); in rf69_set_rx_cfg()
175 ret = rf69_set_bit_rate(dev->spi, rx_cfg->bit_rate); in rf69_set_rx_cfg()
178 ret = rf69_set_antenna_impedance(dev->spi, rx_cfg->antenna_impedance); in rf69_set_rx_cfg()
181 ret = rf69_set_rssi_threshold(dev->spi, rx_cfg->rssi_threshold); in rf69_set_rx_cfg()
184 ret = rf69_set_ook_threshold_dec(dev->spi, rx_cfg->threshold_decrement); in rf69_set_rx_cfg()
187 ret = rf69_set_bandwidth(dev->spi, rx_cfg->bw_mantisse, in rf69_set_rx_cfg()
188 rx_cfg->bw_exponent); in rf69_set_rx_cfg()
[all …]
/linux-5.19.10/drivers/net/ethernet/apple/
Dbmac.c896 unsigned short rx_cfg; in bmac_rx_off() local
898 rx_cfg = bmread(dev, RXCFG); in bmac_rx_off()
899 rx_cfg &= ~RxMACEnable; in bmac_rx_off()
900 bmwrite(dev, RXCFG, rx_cfg); in bmac_rx_off()
902 rx_cfg = bmread(dev, RXCFG); in bmac_rx_off()
903 } while (rx_cfg & RxMACEnable); in bmac_rx_off()
909 unsigned short rx_cfg; in bmac_rx_on() local
911 rx_cfg = bmread(dev, RXCFG); in bmac_rx_on()
912 rx_cfg |= RxMACEnable; in bmac_rx_on()
913 if (hash_enable) rx_cfg |= RxHashFilterEnable; in bmac_rx_on()
[all …]
/linux-5.19.10/drivers/net/ethernet/google/gve/
Dgve_ethtool.c91 for (i = 0; i < priv->rx_cfg.num_queues; i++) { in gve_get_strings()
130 (priv->rx_cfg.num_queues * NUM_GVE_RX_CNTS) + in gve_get_sset_count()
162 rx_qid_to_stats_idx = kmalloc_array(priv->rx_cfg.num_queues, in gve_get_ethtool_stats()
174 ring < priv->rx_cfg.num_queues; ring++) { in gve_get_ethtool_stats()
235 GVE_RX_STATS_REPORT_NUM * priv->rx_cfg.num_queues; in gve_get_ethtool_stats()
236 max_stats_idx = NIC_RX_STATS_REPORT_NUM * priv->rx_cfg.num_queues + in gve_get_ethtool_stats()
254 for (ring = 0; ring < priv->rx_cfg.num_queues; ring++) { in gve_get_ethtool_stats()
294 i += priv->rx_cfg.num_queues * NUM_GVE_RX_CNTS; in gve_get_ethtool_stats()
384 cmd->max_rx = priv->rx_cfg.max_queues; in gve_get_channels()
388 cmd->rx_count = priv->rx_cfg.num_queues; in gve_get_channels()
[all …]
Dgve_main.c51 for (ring = 0; ring < priv->rx_cfg.num_queues; ring++) { in gve_get_stats()
140 priv->rx_cfg.num_queues; in gve_alloc_stats_report()
314 priv->rx_cfg.max_queues = min_t(int, priv->rx_cfg.max_queues, in gve_alloc_notify_blocks()
319 priv->rx_cfg.max_queues); in gve_alloc_notify_blocks()
322 if (priv->rx_cfg.num_queues > priv->rx_cfg.max_queues) in gve_alloc_notify_blocks()
323 priv->rx_cfg.num_queues = priv->rx_cfg.max_queues; in gve_alloc_notify_blocks()
597 err = gve_adminq_create_rx_queues(priv, priv->rx_cfg.num_queues); in gve_create_rings()
600 priv->rx_cfg.num_queues); in gve_create_rings()
607 priv->rx_cfg.num_queues); in gve_create_rings()
616 for (i = 0; i < priv->rx_cfg.num_queues; i++) in gve_create_rings()
[all …]
Dgve.h519 struct gve_queue_config rx_cfg; member
783 return priv->rx_cfg.num_queues; in gve_num_rx_qpls()
Dgve_rx_dqo.c305 for (i = 0; i < priv->rx_cfg.num_queues; i++) { in gve_rx_alloc_rings_dqo()
328 for (i = 0; i < priv->rx_cfg.num_queues; i++) in gve_rx_free_rings_dqo()
Dgve_rx.c252 for (i = 0; i < priv->rx_cfg.num_queues; i++) { in gve_rx_alloc_rings()
275 for (i = 0; i < priv->rx_cfg.num_queues; i++) in gve_rx_free_rings_gqi()
/linux-5.19.10/drivers/net/wireless/intel/iwlwifi/pcie/
Dctxt-info.c168 struct iwl_context_info_rbd_cfg *rx_cfg; in iwl_pcie_ctxt_info_init() local
214 rx_cfg = &ctxt_info->rbd_cfg; in iwl_pcie_ctxt_info_init()
215 rx_cfg->free_rbd_addr = cpu_to_le64(trans_pcie->rxq->bd_dma); in iwl_pcie_ctxt_info_init()
216 rx_cfg->used_rbd_addr = cpu_to_le64(trans_pcie->rxq->used_bd_dma); in iwl_pcie_ctxt_info_init()
217 rx_cfg->status_wr_ptr = cpu_to_le64(trans_pcie->rxq->rb_stts_dma); in iwl_pcie_ctxt_info_init()
/linux-5.19.10/drivers/hsi/clients/
Dnokia-modem.c179 ssip.rx_cfg = cl->rx_cfg; in nokia_modem_probe()
202 cmtspeech.rx_cfg = cl->rx_cfg; in nokia_modem_probe()
Dhsi_char.c355 tmp = cl->rx_cfg; in hsc_rx_set()
356 cl->rx_cfg.mode = rxc->mode; in hsc_rx_set()
357 cl->rx_cfg.num_hw_channels = rxc->channels; in hsc_rx_set()
358 cl->rx_cfg.flow = rxc->flow; in hsc_rx_set()
361 cl->rx_cfg = tmp; in hsc_rx_set()
372 rxc->mode = cl->rx_cfg.mode; in hsc_rx_get()
373 rxc->channels = cl->rx_cfg.num_hw_channels; in hsc_rx_get()
374 rxc->flow = cl->rx_cfg.flow; in hsc_rx_get()
425 if (channel->ch >= channel->cl->rx_cfg.num_hw_channels) in hsc_read()
/linux-5.19.10/drivers/net/ethernet/brocade/bna/
Dbna_tx_rx.c1625 cfg_req->rx_cfg.frame_size = bna_enet_mtu_get(&rx->bna->enet); in bna_bfi_rx_enet_start()
1651 cfg_req->rx_cfg.multi_buffer = in bna_bfi_rx_enet_start()
1690 cfg_req->rx_cfg.rxq_type = BFI_ENET_RXQ_LARGE_SMALL; in bna_bfi_rx_enet_start()
1694 cfg_req->rx_cfg.rxq_type = BFI_ENET_RXQ_HDS; in bna_bfi_rx_enet_start()
1695 cfg_req->rx_cfg.hds.type = rx->hds_cfg.hdr_type; in bna_bfi_rx_enet_start()
1696 cfg_req->rx_cfg.hds.force_offset = rx->hds_cfg.forced_offset; in bna_bfi_rx_enet_start()
1697 cfg_req->rx_cfg.hds.max_header_size = rx->hds_cfg.forced_offset; in bna_bfi_rx_enet_start()
1701 cfg_req->rx_cfg.rxq_type = BFI_ENET_RXQ_SINGLE; in bna_bfi_rx_enet_start()
1707 cfg_req->rx_cfg.strip_vlan = rx->rxf.vlan_strip_status; in bna_bfi_rx_enet_start()
1741 bna_rx_res_check(struct bna_rx_mod *rx_mod, struct bna_rx_config *rx_cfg) in bna_rx_res_check() argument
[all …]
Dbna.h346 struct bna_rx_config *rx_cfg,
/linux-5.19.10/include/linux/hsi/
Dhsi.h104 struct hsi_config rx_cfg; member
129 struct hsi_config rx_cfg; member
224 struct hsi_config rx_cfg; member
/linux-5.19.10/drivers/net/ethernet/micrel/
Dksz884x.c1207 u32 rx_cfg; member
3131 u32 rx_cfg; in set_flow_ctrl() local
3134 rx_cfg = hw->rx_cfg; in set_flow_ctrl()
3137 hw->rx_cfg |= DMA_RX_FLOW_ENABLE; in set_flow_ctrl()
3139 hw->rx_cfg &= ~DMA_RX_FLOW_ENABLE; in set_flow_ctrl()
3145 if (rx_cfg != hw->rx_cfg) in set_flow_ctrl()
3146 writel(hw->rx_cfg, hw->io + KS_DMA_RX_CTRL); in set_flow_ctrl()
3703 hw->rx_cfg = (DMA_RX_BROADCAST | DMA_RX_UNICAST | in hw_setup()
3705 hw->rx_cfg |= KS884X_DMA_RX_MULTICAST; in hw_setup()
3708 hw->rx_cfg |= (DMA_RX_CSUM_TCP | DMA_RX_CSUM_IP); in hw_setup()
[all …]
/linux-5.19.10/include/linux/dma/
Dk3-udma-glue.h69 struct k3_ring_cfg rx_cfg; member
/linux-5.19.10/drivers/net/ethernet/neterion/
Ds2io.c698 struct rx_ring_config *rx_cfg = &config->rx_cfg[i]; in init_shared_mem() local
701 if (rx_cfg->num_rxd % (rxd_count[nic->rxd_mode] + 1)) { in init_shared_mem()
707 size += rx_cfg->num_rxd; in init_shared_mem()
708 ring->block_count = rx_cfg->num_rxd / in init_shared_mem()
710 ring->pkt_cnt = rx_cfg->num_rxd - ring->block_count; in init_shared_mem()
718 struct rx_ring_config *rx_cfg = &config->rx_cfg[i]; in init_shared_mem() local
723 ring->rx_curr_get_info.ring_len = rx_cfg->num_rxd - 1; in init_shared_mem()
726 ring->rx_curr_put_info.ring_len = rx_cfg->num_rxd - 1; in init_shared_mem()
730 blk_cnt = rx_cfg->num_rxd / (rxd_count[nic->rxd_mode] + 1); in init_shared_mem()
790 struct rx_ring_config *rx_cfg = &config->rx_cfg[i]; in init_shared_mem() local
[all …]
/linux-5.19.10/drivers/hsi/controllers/
Domap_ssi_port.c490 writel_relaxed(cl->rx_cfg.num_hw_channels, ssr + SSI_SSR_CHANNELS_REG); in ssi_setup()
494 (cl->rx_cfg.mode != SSI_MODE_FRAME)) in ssi_setup()
496 writel_relaxed(cl->rx_cfg.mode, ssr + SSI_SSR_MODE_REG); in ssi_setup()
497 omap_port->channels = max(cl->rx_cfg.num_hw_channels, in ssi_setup()
509 omap_port->ssr.channels = cl->rx_cfg.num_hw_channels; in ssi_setup()
510 omap_port->ssr.mode = cl->rx_cfg.mode; in ssi_setup()
/linux-5.19.10/drivers/net/ethernet/chelsio/cxgb3/
Dt3_hw.c1133 static void t3_gate_rx_traffic(struct cmac *mac, u32 *rx_cfg, in t3_gate_rx_traffic() argument
1140 *rx_cfg = t3_read_reg(mac->adapter, A_XGM_RX_CFG); in t3_gate_rx_traffic()
1155 static void t3_open_rx_traffic(struct cmac *mac, u32 rx_cfg, in t3_open_rx_traffic() argument
1161 rx_cfg); in t3_open_rx_traffic()
1186 u32 rx_cfg, rx_hash_high, rx_hash_low; in t3_link_changed() local
1190 t3_gate_rx_traffic(mac, &rx_cfg, &rx_hash_high, &rx_hash_low); in t3_link_changed()
1199 t3_open_rx_traffic(mac, rx_cfg, rx_hash_high, rx_hash_low); in t3_link_changed()
1239 u32 rx_cfg, rx_hash_high, rx_hash_low; in t3_link_fault() local
1241 t3_gate_rx_traffic(mac, &rx_cfg, &rx_hash_high, &rx_hash_low); in t3_link_fault()
1249 t3_open_rx_traffic(mac, rx_cfg, rx_hash_high, rx_hash_low); in t3_link_fault()
/linux-5.19.10/drivers/net/ethernet/hisilicon/hns3/hns3pf/
Dhclge_debugfs.c2001 u8 rx_cfg; in hclge_get_vlan_rx_offload_cfg() local
2019 rx_cfg = req->vport_vlan_cfg; in hclge_get_vlan_rx_offload_cfg()
2020 vlan_cfg->strip_tag1 = hnae3_get_bit(rx_cfg, HCLGE_REM_TAG1_EN_B); in hclge_get_vlan_rx_offload_cfg()
2021 vlan_cfg->strip_tag2 = hnae3_get_bit(rx_cfg, HCLGE_REM_TAG2_EN_B); in hclge_get_vlan_rx_offload_cfg()
2022 vlan_cfg->drop_tag1 = hnae3_get_bit(rx_cfg, HCLGE_DISCARD_TAG1_EN_B); in hclge_get_vlan_rx_offload_cfg()
2023 vlan_cfg->drop_tag2 = hnae3_get_bit(rx_cfg, HCLGE_DISCARD_TAG2_EN_B); in hclge_get_vlan_rx_offload_cfg()
2024 vlan_cfg->pri_only1 = hnae3_get_bit(rx_cfg, HCLGE_SHOW_TAG1_EN_B); in hclge_get_vlan_rx_offload_cfg()
2025 vlan_cfg->pri_only2 = hnae3_get_bit(rx_cfg, HCLGE_SHOW_TAG2_EN_B); in hclge_get_vlan_rx_offload_cfg()
/linux-5.19.10/drivers/net/ethernet/ti/
Dam65-cpsw-nuss.c1634 struct k3_udma_glue_rx_channel_cfg rx_cfg = { 0 }; in am65_cpsw_nuss_init_rx_chns() local
1644 rx_cfg.swdata_size = AM65_CPSW_NAV_SW_DATA_SIZE; in am65_cpsw_nuss_init_rx_chns()
1645 rx_cfg.flow_id_num = AM65_CPSW_MAX_RX_FLOWS; in am65_cpsw_nuss_init_rx_chns()
1646 rx_cfg.flow_id_base = common->rx_flow_id_base; in am65_cpsw_nuss_init_rx_chns()
1652 rx_chn->rx_chn = k3_udma_glue_request_rx_chn(dev, "rx", &rx_cfg); in am65_cpsw_nuss_init_rx_chns()
1674 for (i = 0; i < rx_cfg.flow_id_num; i++) { in am65_cpsw_nuss_init_rx_chns()
1685 .rx_cfg = rxring_cfg, in am65_cpsw_nuss_init_rx_chns()
1693 rx_flow_cfg.rx_cfg.size = max_desc_num; in am65_cpsw_nuss_init_rx_chns()
/linux-5.19.10/drivers/dma/ti/
Dk3-udma-glue.c701 flow_cfg->rx_cfg.dma_dev = k3_udma_glue_rx_get_dma_device(rx_chn); in k3_udma_glue_cfg_rx_flow()
702 flow_cfg->rxfdq_cfg.dma_dev = flow_cfg->rx_cfg.dma_dev; in k3_udma_glue_cfg_rx_flow()
706 flow_cfg->rx_cfg.asel = rx_chn->common.atype_asel; in k3_udma_glue_cfg_rx_flow()
710 ret = k3_ringacc_ring_cfg(flow->ringrx, &flow_cfg->rx_cfg); in k3_udma_glue_cfg_rx_flow()
/linux-5.19.10/drivers/usb/gadget/udc/
Dbcm63xx_udc.c467 const struct iudma_ch_cfg *rx_cfg = &iudma_defaults[i]; in bcm63xx_fifo_setup() local
473 ((rx_fifo_slot + rx_cfg->n_fifo_slots - 1) << in bcm63xx_fifo_setup()
475 rx_fifo_slot += rx_cfg->n_fifo_slots; in bcm63xx_fifo_setup()
478 is_hs ? rx_cfg->max_pkt_hs : rx_cfg->max_pkt_fs, in bcm63xx_fifo_setup()
/linux-5.19.10/drivers/soc/qcom/
Dqcom-geni-se.c437 bool msb_to_lsb, bool tx_cfg, bool rx_cfg) in geni_se_config_packing() argument
474 if (rx_cfg) { in geni_se_config_packing()
/linux-5.19.10/include/linux/
Dqcom-geni-se.h448 bool msb_to_lsb, bool tx_cfg, bool rx_cfg);

12