/linux-5.19.10/drivers/ps3/ |
D | ps3-vuart.c | 72 } tx_list; member 486 spin_lock_irqsave(&priv->tx_list.lock, flags); in ps3_vuart_write() 488 if (list_empty(&priv->tx_list.head)) { in ps3_vuart_write() 493 spin_unlock_irqrestore(&priv->tx_list.lock, flags); in ps3_vuart_write() 511 spin_unlock_irqrestore(&priv->tx_list.lock, flags); in ps3_vuart_write() 523 spin_lock_irqsave(&priv->tx_list.lock, flags); in ps3_vuart_write() 524 list_add_tail(&lb->link, &priv->tx_list.head); in ps3_vuart_write() 526 spin_unlock_irqrestore(&priv->tx_list.lock, flags); in ps3_vuart_write() 728 spin_lock_irqsave(&priv->tx_list.lock, flags); in ps3_vuart_handle_interrupt_tx() 730 list_for_each_entry_safe(lb, n, &priv->tx_list.head, link) { in ps3_vuart_handle_interrupt_tx() [all …]
|
/linux-5.19.10/sound/soc/intel/common/ |
D | sst-ipc.c | 105 list_add_tail(&msg->list, &ipc->tx_list); in ipc_tx_message() 160 while (!list_empty(&ipc->tx_list) && !ipc->pending) { in ipc_tx_msgs() 170 msg = list_first_entry(&ipc->tx_list, struct ipc_message, list); in ipc_tx_msgs() 261 INIT_LIST_HEAD(&ipc->tx_list); in sst_ipc_init()
|
D | sst-ipc.h | 54 struct list_head tx_list; member
|
/linux-5.19.10/drivers/hid/intel-ish-hid/ishtp/ |
D | client-buffers.c | 149 while (!list_empty(&cl->tx_list.list)) { in ishtp_cl_free_tx_ring() 150 tx_buf = list_entry(cl->tx_list.list.next, in ishtp_cl_free_tx_ring() 268 tx_list_empty = list_empty(&cl->tx_list.list); in ishtp_cl_tx_empty()
|
D | client.c | 98 INIT_LIST_HEAD(&cl->tx_list.list); in ishtp_cl_init() 591 have_msg_to_send = !list_empty(&cl->tx_list.list); in ishtp_cl_send() 592 list_add_tail(&cl_msg->list, &cl->tx_list.list); in ishtp_cl_send() 660 if (list_empty(&cl->tx_list.list)) { in ipc_tx_send() 677 cl_msg = list_entry(cl->tx_list.list.next, struct ishtp_cl_tx_ring, in ipc_tx_send() 759 if (list_empty(&cl->tx_list.list)) { in ishtp_cl_send_msg_dma() 764 cl_msg = list_entry(cl->tx_list.list.next, struct ishtp_cl_tx_ring, in ishtp_cl_send_msg_dma()
|
D | client.h | 69 struct ishtp_cl_tx_ring tx_list, tx_free_list; member
|
/linux-5.19.10/drivers/dma/ |
D | mmp_pdma.c | 83 struct list_head tx_list; member 350 list_for_each_entry(child, &desc->tx_list, node) { in mmp_pdma_tx_submit() 355 list_splice_tail_init(&desc->tx_list, &chan->chain_pending); in mmp_pdma_tx_submit() 374 INIT_LIST_HEAD(&desc->tx_list); in mmp_pdma_alloc_descriptor() 504 list_add_tail(&new->node, &first->tx_list); in mmp_pdma_prep_memcpy() 520 mmp_pdma_free_desc_list(chan, &first->tx_list); in mmp_pdma_prep_memcpy() 578 list_add_tail(&new->node, &first->tx_list); in mmp_pdma_prep_slave_sg() 600 mmp_pdma_free_desc_list(chan, &first->tx_list); in mmp_pdma_prep_slave_sg() 673 list_add_tail(&new->node, &first->tx_list); in mmp_pdma_prep_dma_cyclic() 687 mmp_pdma_free_desc_list(chan, &first->tx_list); in mmp_pdma_prep_dma_cyclic()
|
D | txx9dmac.c | 182 if (!list_empty(&desc->tx_list)) in txx9dmac_last_child() 183 desc = list_entry(desc->tx_list.prev, typeof(*desc), desc_node); in txx9dmac_last_child() 198 INIT_LIST_HEAD(&desc->tx_list); in txx9dmac_desc_alloc() 247 list_for_each_entry(child, &desc->tx_list, desc_node) in txx9dmac_sync_desc_for_cpu() 269 list_for_each_entry(child, &desc->tx_list, desc_node) in txx9dmac_desc_put() 273 list_splice_init(&desc->tx_list, &dc->free_list); in txx9dmac_desc_put() 412 list_splice_init(&desc->tx_list, &dc->free_list); in txx9dmac_descriptor_complete() 530 list_for_each_entry(child, &bad_desc->tx_list, desc_node) in txx9dmac_handle_error() 572 list_for_each_entry(child, &desc->tx_list, desc_node) in txx9dmac_scan_descriptors() 782 list_add_tail(&desc->desc_node, &first->tx_list); in txx9dmac_prep_dma_memcpy() [all …]
|
D | xgene-dma.c | 240 struct list_head tx_list; member 473 list_splice_tail_init(&desc->tx_list, &chan->ld_pending); in xgene_dma_tx_submit() 500 INIT_LIST_HEAD(&desc->tx_list); in xgene_dma_alloc_descriptor() 868 list_add_tail(&new->node, &first->tx_list); in xgene_dma_prep_xor() 873 list_splice(&first->tx_list, &new->tx_list); in xgene_dma_prep_xor() 881 xgene_dma_free_desc_list(chan, &first->tx_list); in xgene_dma_prep_xor() 926 list_add_tail(&new->node, &first->tx_list); in xgene_dma_prep_pq() 950 list_splice(&first->tx_list, &new->tx_list); in xgene_dma_prep_pq() 958 xgene_dma_free_desc_list(chan, &first->tx_list); in xgene_dma_prep_pq()
|
D | altera-msgdma.c | 164 struct list_head tx_list; member 220 INIT_LIST_HEAD(&desc->tx_list); in msgdma_get_descriptor() 237 list_for_each_entry_safe(child, next, &desc->tx_list, node) { in msgdma_free_descriptor() 367 list_add_tail(&new->node, &first->tx_list); in msgdma_prep_memcpy() 441 list_add_tail(&new->node, &first->tx_list); in msgdma_prep_slave_sg() 541 list_for_each_entry_safe(sdesc, next, &desc->tx_list, node) in msgdma_copy_desc_to_fifo()
|
D | ep93xx_dma.c | 128 struct list_head tx_list; member 248 while (!list_empty(&desc->tx_list)) { in ep93xx_dma_set_active() 249 struct ep93xx_dma_desc *d = list_first_entry(&desc->tx_list, in ep93xx_dma_set_active() 712 list_splice_init(&desc->tx_list, &edmac->free_list); in ep93xx_dma_desc_put() 925 INIT_LIST_HEAD(&desc->tx_list); in ep93xx_dma_alloc_chan_resources() 1011 list_add_tail(&desc->node, &first->tx_list); in ep93xx_dma_prep_dma_memcpy() 1086 list_add_tail(&desc->node, &first->tx_list); in ep93xx_dma_prep_slave_sg() 1167 list_add_tail(&desc->node, &first->tx_list); in ep93xx_dma_prep_dma_cyclic()
|
D | pch_dma.c | 87 struct list_head tx_list; member 336 if (list_empty(&desc->tx_list)) { in pdc_dostart() 355 list_splice_init(&desc->tx_list, &pd_chan->free_list); in pdc_chain_complete() 434 INIT_LIST_HEAD(&desc->tx_list); in pdc_alloc_desc() 483 list_splice_init(&desc->tx_list, &pd_chan->free_list); in pdc_desc_put() 630 list_add_tail(&desc->desc_node, &first->tx_list); in pd_prep_slave_sg()
|
D | tegra20-apb-dma.c | 170 struct list_head tx_list; member 308 if (!list_empty(&dma_desc->tx_list)) in tegra_dma_desc_put() 309 list_splice_init(&dma_desc->tx_list, &tdc->free_sg_req); in tegra_dma_desc_put() 702 list_splice_tail_init(&dma_desc->tx_list, &tdc->pending_sg_req); in tegra_dma_tx_submit() 1099 INIT_LIST_HEAD(&dma_desc->tx_list); in tegra_dma_prep_slave_sg() 1142 list_add_tail(&sg_req->node, &dma_desc->tx_list); in tegra_dma_prep_slave_sg() 1248 INIT_LIST_HEAD(&dma_desc->tx_list); in tegra_dma_prep_dma_cyclic() 1277 list_add_tail(&sg_req->node, &dma_desc->tx_list); in tegra_dma_prep_dma_cyclic()
|
D | fsldma.c | 403 list_splice_tail_init(&desc->tx_list, &chan->ld_pending); in append_ld_queue() 427 list_for_each_entry(child, &desc->tx_list, node) { in fsl_dma_tx_submit() 469 INIT_LIST_HEAD(&desc->tx_list); in fsl_dma_alloc_descriptor() 804 list_add_tail(&new->node, &first->tx_list); in fsl_dma_prep_memcpy() 819 fsldma_free_desc_list_reverse(chan, &first->tx_list); in fsl_dma_prep_memcpy()
|
/linux-5.19.10/drivers/net/ethernet/cavium/octeon/ |
D | octeon_mgmt.c | 131 struct sk_buff_head tx_list; member 263 spin_lock_irqsave(&p->tx_list.lock, flags); in octeon_mgmt_clean_tx_buffers() 268 spin_unlock_irqrestore(&p->tx_list.lock, flags); in octeon_mgmt_clean_tx_buffers() 279 skb = __skb_dequeue(&p->tx_list); in octeon_mgmt_clean_tx_buffers() 288 spin_unlock_irqrestore(&p->tx_list.lock, flags); in octeon_mgmt_clean_tx_buffers() 1259 skb_queue_purge(&p->tx_list); in octeon_mgmt_stop() 1290 spin_lock_irqsave(&p->tx_list.lock, flags); in octeon_mgmt_xmit() 1293 spin_unlock_irqrestore(&p->tx_list.lock, flags); in octeon_mgmt_xmit() 1295 spin_lock_irqsave(&p->tx_list.lock, flags); in octeon_mgmt_xmit() 1300 spin_unlock_irqrestore(&p->tx_list.lock, flags); in octeon_mgmt_xmit() [all …]
|
/linux-5.19.10/drivers/infiniband/hw/hfi1/ |
D | ipoib_tx.c | 411 &txq->tx_list, in hfi1_ipoib_submit_tx_list() 426 if (!list_empty(&txq->tx_list)) { in hfi1_ipoib_flush_tx_list() 545 list_add_tail(&tx->txreq.list, &txq->tx_list); in hfi1_ipoib_send_dma_list() 598 if (netdev_xmit_more() || !list_empty(&txp.txq->tx_list)) in hfi1_ipoib_send() 631 list_add_tail(&txreq->list, &txq->tx_list); in hfi1_ipoib_sdma_sleep() 718 INIT_LIST_HEAD(&txq->tx_list); in hfi1_ipoib_txreq_init() 773 list_for_each_entry_safe(txreq, txreq_tmp, &txq->tx_list, list) { in hfi1_ipoib_drain_tx_list() 864 list_empty(&txq->tx_list)); in hfi1_ipoib_tx_timeout()
|
D | ipoib.h | 112 struct list_head tx_list; member
|
/linux-5.19.10/drivers/dma/dw/ |
D | core.c | 91 INIT_LIST_HEAD(&desc->tx_list); in dwc_desc_get() 107 list_for_each_entry_safe(child, _next, &desc->tx_list, desc_node) { in dwc_desc_put() 203 dwc->tx_node_active = &first->tx_list; in dwc_dostart() 253 list_for_each_entry(child, &desc->tx_list, desc_node) in dwc_descriptor_complete() 325 head = &desc->tx_list; in dwc_scan_descriptors() 384 list_for_each_entry(child, &desc->tx_list, desc_node) { in dwc_scan_descriptors() 457 list_for_each_entry(child, &bad_desc->tx_list, desc_node) in dwc_handle_error() 595 list_add_tail(&desc->desc_node, &first->tx_list); in dwc_prep_dma_memcpy() 685 list_add_tail(&desc->desc_node, &first->tx_list); in dwc_prep_slave_sg() 734 list_add_tail(&desc->desc_node, &first->tx_list); in dwc_prep_slave_sg()
|
/linux-5.19.10/drivers/dma/xilinx/ |
D | zynqmp_dma.c | 181 struct list_head tx_list; member 382 if (!list_empty(&desc->tx_list)) in zynqmp_dma_tx_submit() 383 desc = list_last_entry(&desc->tx_list, in zynqmp_dma_tx_submit() 415 INIT_LIST_HEAD(&desc->tx_list); in zynqmp_dma_get_descriptor() 435 list_for_each_entry_safe(child, next, &sdesc->tx_list, node) { in zynqmp_dma_free_descriptor() 847 list_add_tail(&new->node, &first->tx_list); in zynqmp_dma_prep_memcpy()
|
/linux-5.19.10/drivers/net/wireless/intersil/p54/ |
D | p54spi.h | 89 struct list_head tx_list; member
|
D | p54spi.c | 434 struct p54s_tx_info, tx_list); in p54spi_wq_tx() 436 list_del_init(&entry->tx_list); in p54spi_wq_tx() 441 tx_list); in p54spi_wq_tx() 472 list_add_tail(&di->tx_list, &priv->tx_pending); in p54spi_op_tx()
|
/linux-5.19.10/include/linux/platform_data/ |
D | dma-iop32x.h | 91 struct list_head tx_list; member
|
/linux-5.19.10/include/net/ |
D | xdp_sock.h | 61 struct list_head tx_list; member
|
/linux-5.19.10/drivers/dma/sh/ |
D | shdma-base.c | 566 LIST_HEAD(tx_list); in shdma_prep_sg() 615 list_add_tail(&new->node, &tx_list); in shdma_prep_sg() 623 list_splice_tail(&tx_list, &schan->ld_free); in shdma_prep_sg() 630 list_for_each_entry(new, &tx_list, node) in shdma_prep_sg() 632 list_splice(&tx_list, &schan->ld_free); in shdma_prep_sg()
|
/linux-5.19.10/drivers/net/ethernet/ti/ |
D | tlan.c | 838 priv->tx_list = priv->rx_list + TLAN_NUM_RX_LISTS; in tlan_init() 1054 tail_list = priv->tx_list + priv->tx_tail; in tlan_start_tx() 1093 (priv->tx_list + (TLAN_NUM_TX_LISTS - 1))->forward in tlan_start_tx() 1096 (priv->tx_list + (priv->tx_tail - 1))->forward in tlan_start_tx() 1228 tlan_print_list(priv->tx_list + i, "TX", i); in tlan_get_stats() 1358 head_list = priv->tx_list + priv->tx_head; in tlan_handle_tx_eof() 1381 head_list = priv->tx_list + priv->tx_head; in tlan_handle_tx_eof() 1392 head_list = priv->tx_list + priv->tx_head; in tlan_handle_tx_eof() 1634 head_list = priv->tx_list + priv->tx_head; in tlan_handle_tx_eoc() 1902 list = priv->tx_list + i; in tlan_reset_lists() [all …]
|