Home
last modified time | relevance | path

Searched refs:dma_list (Results 1 – 19 of 19) sorted by relevance

/linux-6.6.21/drivers/crypto/intel/keembay/
Docs-hcu.c444 const struct ocs_hcu_dma_list *dma_list, in ocs_hcu_ll_dma_start() argument
450 if (!dma_list) in ocs_hcu_ll_dma_start()
471 writel(dma_list->dma_addr, hcu_dev->io_base + OCS_HCU_DMA_NEXT_SRC_DESCR); in ocs_hcu_ll_dma_start()
492 struct ocs_hcu_dma_list *dma_list; in ocs_hcu_dma_list_alloc() local
494 dma_list = kmalloc(sizeof(*dma_list), GFP_KERNEL); in ocs_hcu_dma_list_alloc()
495 if (!dma_list) in ocs_hcu_dma_list_alloc()
499 dma_list->head = dma_alloc_coherent(hcu_dev->dev, in ocs_hcu_dma_list_alloc()
500 sizeof(*dma_list->head) * max_nents, in ocs_hcu_dma_list_alloc()
501 &dma_list->dma_addr, GFP_KERNEL); in ocs_hcu_dma_list_alloc()
502 if (!dma_list->head) { in ocs_hcu_dma_list_alloc()
[all …]
Docs-hcu.h77 struct ocs_hcu_dma_list *dma_list);
80 struct ocs_hcu_dma_list *dma_list,
87 const struct ocs_hcu_dma_list *dma_list);
91 const struct ocs_hcu_dma_list *dma_list,
103 const struct ocs_hcu_dma_list *dma_list,
Dkeembay-ocs-hcu-core.c78 struct ocs_hcu_dma_list *dma_list; member
207 if (rctx->dma_list) { in kmb_ocs_hcu_dma_cleanup()
208 ocs_hcu_dma_list_free(hcu_dev, rctx->dma_list); in kmb_ocs_hcu_dma_cleanup()
209 rctx->dma_list = NULL; in kmb_ocs_hcu_dma_cleanup()
289 rctx->dma_list = ocs_hcu_dma_list_alloc(rctx->hcu_dev, nents); in kmb_ocs_dma_prepare()
290 if (!rctx->dma_list) { in kmb_ocs_dma_prepare()
297 rc = ocs_hcu_dma_list_add_tail(rctx->hcu_dev, rctx->dma_list, in kmb_ocs_dma_prepare()
322 rctx->dma_list, in kmb_ocs_dma_prepare()
439 rctx->dma_list, req->result, rctx->dig_sz); in kmb_ocs_hcu_do_one_request()
464 rctx->dma_list); in kmb_ocs_hcu_do_one_request()
[all …]
Docs-aes.c423 dma_addr_t dma_list) in dma_to_ocs_aes_ll() argument
426 iowrite32(dma_list, in dma_to_ocs_aes_ll()
432 dma_addr_t dma_list) in dma_from_ocs_aes_ll() argument
435 iowrite32(dma_list, in dma_from_ocs_aes_ll()
/linux-6.6.21/drivers/infiniband/hw/mthca/
Dmthca_allocator.c196 u64 *dma_list = NULL; in mthca_buf_alloc() local
217 dma_list = kmalloc_array(npages, sizeof(*dma_list), in mthca_buf_alloc()
219 if (!dma_list) in mthca_buf_alloc()
223 dma_list[i] = t + i * (1 << shift); in mthca_buf_alloc()
229 dma_list = kmalloc_array(npages, sizeof(*dma_list), in mthca_buf_alloc()
231 if (!dma_list) in mthca_buf_alloc()
250 dma_list[i] = t; in mthca_buf_alloc()
258 dma_list, shift, npages, in mthca_buf_alloc()
266 kfree(dma_list); in mthca_buf_alloc()
274 kfree(dma_list); in mthca_buf_alloc()
Dmthca_eq.c471 u64 *dma_list = NULL; in mthca_create_eq() local
490 dma_list = kmalloc_array(npages, sizeof(*dma_list), GFP_KERNEL); in mthca_create_eq()
491 if (!dma_list) in mthca_create_eq()
505 dma_list[i] = t; in mthca_create_eq()
519 dma_list, PAGE_SHIFT, npages, in mthca_create_eq()
551 kfree(dma_list); in mthca_create_eq()
582 kfree(dma_list); in mthca_create_eq()
/linux-6.6.21/drivers/infiniband/core/
Dumem_odp.c83 umem_odp->dma_list = kvcalloc( in ib_init_umem_odp()
84 ndmas, sizeof(*umem_odp->dma_list), GFP_KERNEL); in ib_init_umem_odp()
85 if (!umem_odp->dma_list) { in ib_init_umem_odp()
100 kvfree(umem_odp->dma_list); in ib_init_umem_odp()
277 kvfree(umem_odp->dma_list); in ib_umem_odp_release()
303 dma_addr_t *dma_addr = &umem_odp->dma_list[dma_index]; in ib_umem_odp_map_dma_single_page()
427 WARN_ON(umem_odp->dma_list[dma_index]); in ib_umem_odp_map_dma_and_lock()
486 dma = umem_odp->dma_list[idx]; in ib_umem_odp_unmap_dma_pages()
510 umem_odp->dma_list[idx] = 0; in ib_umem_odp_unmap_dma_pages()
/linux-6.6.21/drivers/misc/genwqe/
Dcard_utils.c229 static void genwqe_unmap_pages(struct genwqe_dev *cd, dma_addr_t *dma_list, in genwqe_unmap_pages() argument
235 for (i = 0; (i < num_pages) && (dma_list[i] != 0x0); i++) { in genwqe_unmap_pages()
236 dma_unmap_page(&pci_dev->dev, dma_list[i], PAGE_SIZE, in genwqe_unmap_pages()
238 dma_list[i] = 0x0; in genwqe_unmap_pages()
244 dma_addr_t *dma_list) in genwqe_map_pages() argument
253 dma_list[i] = 0x0; in genwqe_map_pages()
266 dma_list[i] = daddr; in genwqe_map_pages()
271 genwqe_unmap_pages(cd, dma_list, num_pages); in genwqe_map_pages()
375 dma_addr_t *dma_list) in genwqe_setup_sgl() argument
410 daddr = dma_list[p] + map_offs; in genwqe_setup_sgl()
[all …]
Dcard_base.h171 dma_addr_t *dma_list; /* list of dma addresses per page */ member
372 dma_addr_t *dma_list);
Dcard_dev.c965 &m->dma_list[page_offs]); in ddcb_cmd_fixups()
/linux-6.6.21/arch/powerpc/platforms/cell/spufs/
Dspu_utils.h38 struct dma_list_elem dma_list[15] __attribute__ ((aligned(8))); variable
92 dma_list[i].size = 16384; in build_dma_list()
93 dma_list[i].ea_low = ea_low; in build_dma_list()
Dspu_save.c50 unsigned int list = (unsigned int)&dma_list[0]; in save_upper_240kb()
51 unsigned int size = sizeof(dma_list); in save_upper_240kb()
Dspu_restore.c50 unsigned int list = (unsigned int)&dma_list[0]; in restore_upper_240kb()
51 unsigned int size = sizeof(dma_list); in restore_upper_240kb()
/linux-6.6.21/drivers/staging/vme_user/
Dvme.c931 struct vme_dma_list *dma_list; in vme_new_dma_list() local
938 dma_list = kmalloc(sizeof(*dma_list), GFP_KERNEL); in vme_new_dma_list()
939 if (!dma_list) in vme_new_dma_list()
942 INIT_LIST_HEAD(&dma_list->entries); in vme_new_dma_list()
943 dma_list->parent = list_entry(resource->entry, in vme_new_dma_list()
946 mutex_init(&dma_list->mtx); in vme_new_dma_list()
948 return dma_list; in vme_new_dma_list()
/linux-6.6.21/drivers/vfio/
Dvfio_iommu_type1.c67 struct rb_root dma_list; member
168 struct rb_node *node = iommu->dma_list.rb_node; in vfio_find_dma()
188 struct rb_node *node = iommu->dma_list.rb_node; in vfio_find_dma_first_node()
211 struct rb_node **link = &iommu->dma_list.rb_node, *parent = NULL; in vfio_link_dma()
225 rb_insert_color(&new->node, &iommu->dma_list); in vfio_link_dma()
230 rb_erase(&old->node, &iommu->dma_list); in vfio_unlink_dma()
277 for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) { in vfio_iommu_populate_bitmap_full()
288 for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) { in vfio_dma_bitmap_alloc_all()
313 for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) { in vfio_dma_bitmap_free_all()
1217 for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) { in vfio_iova_dirty_bitmap()
[all …]
/linux-6.6.21/include/rdma/
Dib_umem_odp.h25 dma_addr_t *dma_list; member
/linux-6.6.21/drivers/net/ethernet/mellanox/mlx4/
Deq.c977 u64 *dma_list = NULL; in mlx4_create_eq() local
998 dma_list = kmalloc_array(npages, sizeof(*dma_list), GFP_KERNEL); in mlx4_create_eq()
999 if (!dma_list) in mlx4_create_eq()
1015 dma_list[i] = t; in mlx4_create_eq()
1033 err = mlx4_write_mtt(dev, &eq->mtt, 0, npages, dma_list); in mlx4_create_eq()
1053 kfree(dma_list); in mlx4_create_eq()
1082 kfree(dma_list); in mlx4_create_eq()
/linux-6.6.21/sound/pci/
Dmaestro3.c740 struct m3_list dma_list; member
1734 s->index_list[2] = &chip->dma_list; in snd_m3_substream_open()
2156 chip->dma_list.curlen = 0; in snd_m3_assp_init()
2157 chip->dma_list.mem_addr = KDATA_DMA_XFER0; in snd_m3_assp_init()
2158 chip->dma_list.max = MAX_VIRTUAL_DMA_CHANNELS; in snd_m3_assp_init()
/linux-6.6.21/drivers/infiniband/hw/mlx5/
Dodp.c169 pa = odp->dma_list[idx + i]; in populate_mtt()
271 if (umem_odp->dma_list[idx] & in mlx5_ib_invalidate_range()