Searched refs:curr_entry (Results 1 – 6 of 6) sorted by relevance
317 next_dcb_hw = sparx5_fdma_next_dcb(tx, tx->curr_entry); in sparx5_fdma_xmit()324 tx->curr_entry->nextptr = tx->dma + in sparx5_fdma_xmit()327 tx->curr_entry = next_dcb_hw; in sparx5_fdma_xmit()400 tx->curr_entry = devm_kzalloc(sparx5->dev, size, GFP_KERNEL); in sparx5_fdma_tx_alloc()401 if (!tx->curr_entry) in sparx5_fdma_tx_alloc()403 tx->dma = virt_to_phys(tx->curr_entry); in sparx5_fdma_tx_alloc()404 tx->first_entry = tx->curr_entry; in sparx5_fdma_tx_alloc()408 dcb = &tx->curr_entry[idx]; in sparx5_fdma_tx_alloc()434 tx->curr_entry = dcb; in sparx5_fdma_tx_alloc()
141 struct sparx5_tx_dcb_hw *curr_entry; member
984 rx_ring->curr_entry++; in qlge_update_cq()987 rx_ring->curr_entry = rx_ring->cq_base; in qlge_update_cq()2100 net_rsp = (struct qlge_ob_mac_iocb_rsp *)rx_ring->curr_entry; in qlge_clean_outbound_rx_ring()2145 net_rsp = rx_ring->curr_entry; in qlge_clean_inbound_rx_ring()2978 rx_ring->curr_entry = rx_ring->cq_base; in qlge_start_rx_ring()
1472 struct qlge_net_rsp_iocb *curr_entry; /* next entry on queue */ member
1342 u16 curr_entry[NPC_MCAM_MAX_SHIFTS]; member
2846 old_entry = req->curr_entry[index]; in rvu_mbox_handler_npc_mcam_shift_entry()