/linux-5.19.10/drivers/net/ethernet/intel/ice/ |
D | ice_txrx.c | 140 static struct netdev_queue *txring_txq(const struct ice_tx_ring *ring) in txring_txq() function 184 netdev_tx_reset_queue(txring_txq(tx_ring)); in ice_clean_tx_ring() 227 netdev_txq_bql_complete_prefetchw(txring_txq(tx_ring)); in ice_clean_tx_irq() 316 netdev_tx_completed_queue(txring_txq(tx_ring), total_pkts, total_bytes); in ice_clean_tx_irq() 325 if (netif_tx_queue_stopped(txring_txq(tx_ring)) && in ice_clean_tx_irq() 327 netif_tx_wake_queue(txring_txq(tx_ring)); in ice_clean_tx_irq() 1539 netif_tx_stop_queue(txring_txq(tx_ring)); in __ice_maybe_stop_tx() 1548 netif_tx_start_queue(txring_txq(tx_ring)); in __ice_maybe_stop_tx() 1700 kick = __netdev_tx_sent_queue(txring_txq(tx_ring), first->bytecount, in ice_tx_map() 2307 netdev_txq_bql_enqueue_prefetchw(txring_txq(tx_ring)); in ice_xmit_frame_ring()
|
/linux-5.19.10/drivers/net/ethernet/intel/iavf/ |
D | iavf_txrx.h | 523 static inline struct netdev_queue *txring_txq(const struct iavf_ring *ring) in txring_txq() function
|
D | iavf_txrx.c | 83 netdev_tx_reset_queue(txring_txq(tx_ring)); in iavf_clean_tx_ring() 304 netdev_tx_completed_queue(txring_txq(tx_ring), in iavf_clean_tx_irq() 2359 netdev_tx_sent_queue(txring_txq(tx_ring), first->bytecount); in iavf_tx_map() 2388 if (netif_xmit_stopped(txring_txq(tx_ring)) || !netdev_xmit_more()) { in iavf_tx_map()
|
/linux-5.19.10/drivers/net/ethernet/intel/i40e/ |
D | i40e_txrx.h | 554 static inline struct netdev_queue *txring_txq(const struct i40e_ring *ring) in txring_txq() function
|
D | i40e_txrx.c | 820 netdev_tx_reset_queue(txring_txq(tx_ring)); in i40e_clean_tx_ring() 1033 netdev_tx_completed_queue(txring_txq(tx_ring), in i40e_clean_tx_irq() 3595 netdev_tx_sent_queue(txring_txq(tx_ring), first->bytecount); in i40e_tx_map() 3636 if (netif_xmit_stopped(txring_txq(tx_ring)) || !netdev_xmit_more()) { in i40e_tx_map()
|
/linux-5.19.10/drivers/net/ethernet/intel/fm10k/ |
D | fm10k_main.c | 999 netdev_tx_sent_queue(txring_txq(tx_ring), first->bytecount); in fm10k_tx_map() 1022 if (netif_xmit_stopped(txring_txq(tx_ring)) || !netdev_xmit_more()) { in fm10k_tx_map() 1299 netdev_tx_completed_queue(txring_txq(tx_ring), in fm10k_clean_tx_irq()
|
D | fm10k.h | 167 static inline struct netdev_queue *txring_txq(const struct fm10k_ring *ring) in txring_txq() function
|
D | fm10k_netdev.c | 187 netdev_tx_reset_queue(txring_txq(tx_ring)); in fm10k_clean_tx_ring()
|
/linux-5.19.10/drivers/net/ethernet/intel/igc/ |
D | igc_main.c | 256 netdev_tx_reset_queue(txring_txq(tx_ring)); in igc_clean_tx_ring() 1263 netdev_tx_sent_queue(txring_txq(tx_ring), first->bytecount); in igc_tx_map() 1291 if (netif_xmit_stopped(txring_txq(tx_ring)) || !netdev_xmit_more()) { in igc_tx_map() 2182 netdev_tx_sent_queue(txring_txq(ring), buffer->bytecount); in igc_xdp_init_tx_descriptor() 2219 nq = txring_txq(ring); in igc_xdp_xmit_back() 2295 nq = txring_txq(ring); in igc_finalize_xdp() 2611 struct netdev_queue *nq = txring_txq(ring); in igc_xdp_xmit_zc() 2651 netdev_tx_sent_queue(txring_txq(ring), xdp_desc.len); in igc_xdp_xmit_zc() 2763 netdev_tx_completed_queue(txring_txq(tx_ring), in igc_clean_tx_irq() 6050 nq = txring_txq(ring); in igc_xdp_xmit()
|
D | igc.h | 565 static inline struct netdev_queue *txring_txq(const struct igc_ring *tx_ring) in txring_txq() function
|
/linux-5.19.10/drivers/net/ethernet/intel/igb/ |
D | igb.h | 795 static inline struct netdev_queue *txring_txq(const struct igb_ring *tx_ring) in txring_txq() function
|
D | igb_main.c | 2927 nq = txring_txq(tx_ring); in igb_xdp_xmit_back() 2960 nq = txring_txq(tx_ring); in igb_xdp_xmit() 4876 netdev_tx_reset_queue(txring_txq(tx_ring)); in igb_clean_tx_ring() 6200 netdev_tx_sent_queue(txring_txq(tx_ring), first->bytecount); in igb_tx_map() 6228 if (netif_xmit_stopped(txring_txq(tx_ring)) || !netdev_xmit_more()) { in igb_tx_map() 6315 netdev_tx_sent_queue(txring_txq(tx_ring), tx_buffer->bytecount); in igb_xmit_xdp_ring() 6334 if (netif_xmit_stopped(txring_txq(tx_ring)) || !netdev_xmit_more()) in igb_xmit_xdp_ring() 8222 netdev_tx_completed_queue(txring_txq(tx_ring), in igb_clean_tx_irq()
|
D | igb_ethtool.c | 1872 netdev_tx_reset_queue(txring_txq(tx_ring)); in igb_clean_test_rings()
|
/linux-5.19.10/drivers/net/ethernet/intel/ixgbe/ |
D | ixgbe.h | 979 static inline struct netdev_queue *txring_txq(const struct ixgbe_ring *ring) in txring_txq() function
|
D | ixgbe_main.c | 1252 netdev_tx_completed_queue(txring_txq(tx_ring), in ixgbe_clean_tx_irq() 6058 netdev_tx_reset_queue(txring_txq(tx_ring)); in ixgbe_clean_tx_ring() 8321 netdev_tx_sent_queue(txring_txq(tx_ring), first->bytecount); in ixgbe_tx_map() 8349 if (netif_xmit_stopped(txring_txq(tx_ring)) || !netdev_xmit_more()) { in ixgbe_tx_map()
|
D | ixgbe_ethtool.c | 2055 netdev_tx_reset_queue(txring_txq(tx_ring)); in ixgbe_clean_test_rings()
|