/linux-5.19.10/drivers/gpu/drm/amd/amdgpu/ ! |
D | amdgpu_ih.c | 253 u32 ring_index = ih->rptr >> 2; in amdgpu_ih_decode_iv_helper() local 256 dw[0] = le32_to_cpu(ih->ring[ring_index + 0]); in amdgpu_ih_decode_iv_helper() 257 dw[1] = le32_to_cpu(ih->ring[ring_index + 1]); in amdgpu_ih_decode_iv_helper() 258 dw[2] = le32_to_cpu(ih->ring[ring_index + 2]); in amdgpu_ih_decode_iv_helper() 259 dw[3] = le32_to_cpu(ih->ring[ring_index + 3]); in amdgpu_ih_decode_iv_helper() 260 dw[4] = le32_to_cpu(ih->ring[ring_index + 4]); in amdgpu_ih_decode_iv_helper() 261 dw[5] = le32_to_cpu(ih->ring[ring_index + 5]); in amdgpu_ih_decode_iv_helper() 262 dw[6] = le32_to_cpu(ih->ring[ring_index + 6]); in amdgpu_ih_decode_iv_helper() 263 dw[7] = le32_to_cpu(ih->ring[ring_index + 7]); in amdgpu_ih_decode_iv_helper() 287 uint32_t ring_index; in amdgpu_ih_decode_iv_ts_helper() local [all …]
|
D | si_ih.c | 130 u32 ring_index = ih->rptr >> 2; in si_ih_decode_iv() local 133 dw[0] = le32_to_cpu(ih->ring[ring_index + 0]); in si_ih_decode_iv() 134 dw[1] = le32_to_cpu(ih->ring[ring_index + 1]); in si_ih_decode_iv() 135 dw[2] = le32_to_cpu(ih->ring[ring_index + 2]); in si_ih_decode_iv() 136 dw[3] = le32_to_cpu(ih->ring[ring_index + 3]); in si_ih_decode_iv()
|
D | cik_ih.c | 247 u32 ring_index = ih->rptr >> 2; in cik_ih_decode_iv() local 250 dw[0] = le32_to_cpu(ih->ring[ring_index + 0]); in cik_ih_decode_iv() 251 dw[1] = le32_to_cpu(ih->ring[ring_index + 1]); in cik_ih_decode_iv() 252 dw[2] = le32_to_cpu(ih->ring[ring_index + 2]); in cik_ih_decode_iv() 253 dw[3] = le32_to_cpu(ih->ring[ring_index + 3]); in cik_ih_decode_iv()
|
D | cz_ih.c | 239 u32 ring_index = ih->rptr >> 2; in cz_ih_decode_iv() local 242 dw[0] = le32_to_cpu(ih->ring[ring_index + 0]); in cz_ih_decode_iv() 243 dw[1] = le32_to_cpu(ih->ring[ring_index + 1]); in cz_ih_decode_iv() 244 dw[2] = le32_to_cpu(ih->ring[ring_index + 2]); in cz_ih_decode_iv() 245 dw[3] = le32_to_cpu(ih->ring[ring_index + 3]); in cz_ih_decode_iv()
|
D | iceland_ih.c | 238 u32 ring_index = ih->rptr >> 2; in iceland_ih_decode_iv() local 241 dw[0] = le32_to_cpu(ih->ring[ring_index + 0]); in iceland_ih_decode_iv() 242 dw[1] = le32_to_cpu(ih->ring[ring_index + 1]); in iceland_ih_decode_iv() 243 dw[2] = le32_to_cpu(ih->ring[ring_index + 2]); in iceland_ih_decode_iv() 244 dw[3] = le32_to_cpu(ih->ring[ring_index + 3]); in iceland_ih_decode_iv()
|
D | tonga_ih.c | 241 u32 ring_index = ih->rptr >> 2; in tonga_ih_decode_iv() local 244 dw[0] = le32_to_cpu(ih->ring[ring_index + 0]); in tonga_ih_decode_iv() 245 dw[1] = le32_to_cpu(ih->ring[ring_index + 1]); in tonga_ih_decode_iv() 246 dw[2] = le32_to_cpu(ih->ring[ring_index + 2]); in tonga_ih_decode_iv() 247 dw[3] = le32_to_cpu(ih->ring[ring_index + 3]); in tonga_ih_decode_iv()
|
D | amdgpu_irq.c | 475 u32 ring_index = ih->rptr >> 2; in amdgpu_irq_dispatch() local 483 entry.iv_entry = (const uint32_t *)&ih->ring[ring_index]; in amdgpu_irq_dispatch()
|
/linux-5.19.10/drivers/scsi/qla2xxx/ ! |
D | qla_iocb.c | 98 req->ring_index++; in qla2x00_prep_cont_type0_iocb() 99 if (req->ring_index == req->length) { in qla2x00_prep_cont_type0_iocb() 100 req->ring_index = 0; in qla2x00_prep_cont_type0_iocb() 127 req->ring_index++; in qla2x00_prep_cont_type1_iocb() 128 if (req->ring_index == req->length) { in qla2x00_prep_cont_type1_iocb() 129 req->ring_index = 0; in qla2x00_prep_cont_type1_iocb() 381 if (req->ring_index < cnt) in qla2x00_start_scsi() 382 req->cnt = cnt - req->ring_index; in qla2x00_start_scsi() 385 (req->ring_index - cnt); in qla2x00_start_scsi() 422 req->ring_index++; in qla2x00_start_scsi() [all …]
|
D | qla_nvme.c | 457 if (req->ring_index < cnt) in qla2x00_start_nvme_mq() 458 req->cnt = cnt - req->ring_index; in qla2x00_start_nvme_mq() 460 req->cnt = req->length - (req->ring_index - cnt); in qla2x00_start_nvme_mq() 559 req->ring_index++; in qla2x00_start_nvme_mq() 560 if (req->ring_index == req->length) { in qla2x00_start_nvme_mq() 561 req->ring_index = 0; in qla2x00_start_nvme_mq() 583 req->ring_index++; in qla2x00_start_nvme_mq() 584 if (req->ring_index == req->length) { in qla2x00_start_nvme_mq() 585 req->ring_index = 0; in qla2x00_start_nvme_mq() 596 wrt_reg_dword(req->req_q_in, req->ring_index); in qla2x00_start_nvme_mq()
|
D | qla_inline.h | 361 req->ring_index++; in qla_83xx_start_iocbs() 362 if (req->ring_index == req->length) { in qla_83xx_start_iocbs() 363 req->ring_index = 0; in qla_83xx_start_iocbs() 368 wrt_reg_dword(req->req_q_in, req->ring_index); in qla_83xx_start_iocbs()
|
D | qla_mr.c | 1402 rsp->ring_index = 0; in qlafx00_init_response_q_entries() 2699 lreq_q_out = rsp->ring_index; in qlafx00_process_response_queue() 2707 rsp->ring_index++; in qlafx00_process_response_queue() 2709 if (rsp->ring_index == rsp->length) { in qlafx00_process_response_queue() 2711 rsp->ring_index = 0; in qlafx00_process_response_queue() 2760 wrt_reg_dword(rsp->rsp_q_out, rsp->ring_index); in qlafx00_process_response_queue() 2950 req->ring_index++; in qlafx00_prep_cont_type1_iocb() 2951 if (req->ring_index == req->length) { in qlafx00_prep_cont_type1_iocb() 2952 req->ring_index = 0; in qlafx00_prep_cont_type1_iocb() 3092 if (req->ring_index < cnt) in qlafx00_start_scsi() [all …]
|
D | qla_isr.c | 196 rsp_q->ring_index++; in __qla_consume_iocb() 197 if (rsp_q->ring_index == rsp_q->length) { in __qla_consume_iocb() 198 rsp_q->ring_index = 0; in __qla_consume_iocb() 274 rsp_q->ring_index++; in __qla_copy_purex_to_buffer() 275 if (rsp_q->ring_index == rsp_q->length) { in __qla_copy_purex_to_buffer() 276 rsp_q->ring_index = 0; in __qla_copy_purex_to_buffer() 1063 rsp_q->ring_index++; in qla27xx_copy_fpin_pkt() 1064 if (rsp_q->ring_index == rsp_q->length) { in qla27xx_copy_fpin_pkt() 1065 rsp_q->ring_index = 0; in qla27xx_copy_fpin_pkt() 2780 rsp->ring_index++; in qla2x00_process_response_queue() [all …]
|
D | qla_mid.c | 780 req->ring_index = 0; in qla25xx_create_req_que() 792 req->ring_ptr, req->ring_index, in qla25xx_create_req_que() 797 req->ring_ptr, req->ring_index, req->cnt, in qla25xx_create_req_que()
|
D | qla_edif.c | 3072 if (req->ring_index < cnt) in qla28xx_start_scsi_edif() 3073 req->cnt = cnt - req->ring_index; in qla28xx_start_scsi_edif() 3076 (req->ring_index - cnt); in qla28xx_start_scsi_edif() 3234 req->ring_index++; in qla28xx_start_scsi_edif() 3235 if (req->ring_index == req->length) { in qla28xx_start_scsi_edif() 3236 req->ring_index = 0; in qla28xx_start_scsi_edif() 3244 wrt_reg_dword(req->req_q_in, req->ring_index); in qla28xx_start_scsi_edif()
|
/linux-5.19.10/drivers/gpu/drm/radeon/ ! |
D | rv770_dma.c | 49 int ring_index = rdev->asic->copy.dma_ring_index; in rv770_copy_dma() local 50 struct radeon_ring *ring = &rdev->ring[ring_index]; in rv770_copy_dma()
|
D | evergreen_dma.c | 114 int ring_index = rdev->asic->copy.dma_ring_index; in evergreen_copy_dma() local 115 struct radeon_ring *ring = &rdev->ring[ring_index]; in evergreen_copy_dma()
|
D | si_dma.c | 237 int ring_index = rdev->asic->copy.dma_ring_index; in si_copy_dma() local 238 struct radeon_ring *ring = &rdev->ring[ring_index]; in si_copy_dma()
|
D | r600_dma.c | 450 int ring_index = rdev->asic->copy.dma_ring_index; in r600_copy_dma() local 451 struct radeon_ring *ring = &rdev->ring[ring_index]; in r600_copy_dma()
|
D | cik_sdma.c | 585 int ring_index = rdev->asic->copy.dma_ring_index; in cik_copy_dma() local 586 struct radeon_ring *ring = &rdev->ring[ring_index]; in cik_copy_dma()
|
/linux-5.19.10/drivers/net/wireless/intersil/p54/ ! |
D | p54pci.c | 133 int ring_index, struct p54p_desc *ring, u32 ring_limit, in p54p_refill_rx_ring() argument 140 idx = le32_to_cpu(ring_control->host_idx[ring_index]); in p54p_refill_rx_ring() 181 ring_control->host_idx[ring_index] = cpu_to_le32(idx); in p54p_refill_rx_ring() 185 int ring_index, struct p54p_desc *ring, u32 ring_limit, in p54p_check_rx_ring() argument 194 (*index) = idx = le32_to_cpu(ring_control->device_idx[ring_index]); in p54p_check_rx_ring() 241 p54p_refill_rx_ring(dev, ring_index, ring, ring_limit, rx_buf, *index); in p54p_check_rx_ring() 245 int ring_index, struct p54p_desc *ring, u32 ring_limit, in p54p_check_tx_ring() argument 255 (*index) = idx = le32_to_cpu(ring_control->device_idx[ring_index]); in p54p_check_tx_ring()
|
/linux-5.19.10/drivers/net/ethernet/mellanox/mlx4/ ! |
D | en_tx.c | 432 u16 index, ring_index, stamp_index; in mlx4_en_process_tx_cq() local 456 ring_index = ring_cons & size_mask; in mlx4_en_process_tx_cq() 457 stamp_index = ring_index; in mlx4_en_process_tx_cq() 483 ring_index = (ring_index + last_nr_txbb) & size_mask; in mlx4_en_process_tx_cq() 485 if (unlikely(ring->tx_info[ring_index].ts_requested)) in mlx4_en_process_tx_cq() 492 priv, ring, ring_index, in mlx4_en_process_tx_cq() 498 stamp_index = ring_index; in mlx4_en_process_tx_cq() 501 bytes += ring->tx_info[ring_index].nr_bytes; in mlx4_en_process_tx_cq() 502 } while ((++done < budget) && (ring_index != new_index)); in mlx4_en_process_tx_cq()
|
D | en_ethtool.c | 1261 static int mlx4_en_get_rxfh(struct net_device *dev, u32 *ring_index, u8 *key, in mlx4_en_get_rxfh() argument 1272 if (!ring_index) in mlx4_en_get_rxfh() 1274 ring_index[i] = i % rss_rings; in mlx4_en_get_rxfh() 1283 static int mlx4_en_set_rxfh(struct net_device *dev, const u32 *ring_index, in mlx4_en_set_rxfh() argument 1298 if (!ring_index) in mlx4_en_set_rxfh() 1300 if (i > 0 && !ring_index[i] && !rss_rings) in mlx4_en_set_rxfh() 1303 if (ring_index[i] != (i % (rss_rings ?: n))) in mlx4_en_set_rxfh() 1326 if (ring_index) in mlx4_en_set_rxfh()
|
/linux-5.19.10/drivers/net/wireless/ath/wil6210/ ! |
D | txrx_edma.c | 422 int ring_index = 0; in wil_ring_free_edma() local 443 ring_index = ring - wil->ring_tx; in wil_ring_free_edma() 446 ring_index, ring->size, ring->va, in wil_ring_free_edma() 1121 int ring_index) in wil_tx_desc_map_edma() argument 1132 d->mac.d[0] = (ring_index << WIL_EDMA_DESC_TX_MAC_CFG_0_QID_POS); in wil_tx_desc_map_edma() 1362 int ring_index = ring - wil->ring_tx; in wil_tx_tso_gen_desc() local 1381 len, ring_index); in wil_tx_tso_gen_desc() 1407 int ring_index = ring - wil->ring_tx; in __wil_tx_ring_tso_edma() local 1408 struct wil_ring_tx_data *txdata = &wil->ring_tx_data[ring_index]; in __wil_tx_ring_tso_edma() 1423 ring_index); in __wil_tx_ring_tso_edma() [all …]
|
D | txrx.c | 2038 int ring_index = ring - wil->ring_tx; in __wil_tx_ring() local 2039 struct wil_ring_tx_data *txdata = &wil->ring_tx_data[ring_index]; in __wil_tx_ring() 2043 bool mcast = (ring_index == vif->bcast_ring); in __wil_tx_ring() 2047 skb->len, ring_index, nr_frags); in __wil_tx_ring() 2055 ring_index, 1 + nr_frags); in __wil_tx_ring() 2062 wil_dbg_txrx(wil, "Tx[%2d] skb %d bytes 0x%p -> %pad\n", ring_index, in __wil_tx_ring() 2072 ring_index); in __wil_tx_ring() 2081 ring_index); in __wil_tx_ring() 2094 wil_dbg_txrx(wil, "Tx[%2d] desc[%4d]\n", ring_index, i); in __wil_tx_ring() 2103 ring_index); in __wil_tx_ring() [all …]
|
/linux-5.19.10/drivers/net/ethernet/agere/ ! |
D | et131x.c | 2128 u8 ring_index = rfd->ringindex; in nic_return_rfd() local 2130 struct fbr_lookup *fbr = rx_local->fbr[ring_index]; in nic_return_rfd() 2140 if (ring_index == 0) in nic_return_rfd() 2190 u8 ring_index; in nic_rx_pkts() local 2217 ring_index = (psr->word1 >> 26) & 0x03; in nic_rx_pkts() 2218 fbr = rx_local->fbr[ring_index]; in nic_rx_pkts() 2233 if (ring_index > 1 || buff_index > fbr->num_entries - 1) { in nic_rx_pkts() 2258 rfd->ringindex = ring_index; in nic_rx_pkts()
|