Lines Matching refs:tx_fifo

186 	gve_tx_free_fifo(&tx->tx_fifo, space_freed);  in gve_clean_xdp_done()
220 gve_tx_fifo_release(priv, &tx->tx_fifo); in gve_tx_free_ring()
221 gve_unassign_qpl(priv, tx->tx_fifo.qpl->id); in gve_tx_free_ring()
222 tx->tx_fifo.qpl = NULL; in gve_tx_free_ring()
264 tx->tx_fifo.qpl = gve_assign_tx_qpl(priv, idx); in gve_tx_alloc_ring()
265 if (!tx->tx_fifo.qpl) in gve_tx_alloc_ring()
268 if (gve_tx_fifo_init(priv, &tx->tx_fifo)) in gve_tx_alloc_ring()
290 gve_tx_fifo_release(priv, &tx->tx_fifo); in gve_tx_alloc_ring()
293 gve_unassign_qpl(priv, tx->tx_fifo.qpl->id); in gve_tx_alloc_ring()
357 pad_bytes = gve_tx_fifo_pad_alloc_one_frag(&tx->tx_fifo, in gve_skb_fifo_bytes_required()
397 can_alloc = gve_tx_fifo_can_alloc(&tx->tx_fifo, bytes_required); in gve_can_tx()
535 pad_bytes = gve_tx_fifo_pad_alloc_one_frag(&tx->tx_fifo, hlen); in gve_tx_add_skb_copy()
536 hdr_nfrags = gve_tx_alloc_fifo(&tx->tx_fifo, hlen + pad_bytes, in gve_tx_add_skb_copy()
539 payload_nfrags = gve_tx_alloc_fifo(&tx->tx_fifo, skb->len - hlen, in gve_tx_add_skb_copy()
548 tx->tx_fifo.base + info->iov[hdr_nfrags - 1].iov_offset, in gve_tx_add_skb_copy()
550 gve_dma_sync_for_device(&priv->pdev->dev, tx->tx_fifo.qpl->page_buses, in gve_tx_add_skb_copy()
571 tx->tx_fifo.base + info->iov[i].iov_offset, in gve_tx_add_skb_copy()
573 gve_dma_sync_for_device(&priv->pdev->dev, tx->tx_fifo.qpl->page_buses, in gve_tx_add_skb_copy()
732 pad = gve_tx_fifo_pad_alloc_one_frag(&tx->tx_fifo, len); in gve_tx_fill_xdp()
740 nfrags = gve_tx_alloc_fifo(&tx->tx_fifo, pad + len, in gve_tx_fill_xdp()
758 memcpy(tx->tx_fifo.base + info->iov[iovi].iov_offset, in gve_tx_fill_xdp()
761 tx->tx_fifo.qpl->page_buses, in gve_tx_fill_xdp()
859 gve_tx_free_fifo(&tx->tx_fifo, space_freed); in gve_clean_tx_done()